method
stringlengths 13
441k
| clean_method
stringlengths 7
313k
| doc
stringlengths 17
17.3k
| comment
stringlengths 3
1.42k
| method_name
stringlengths 1
273
| extra
dict | imports
list | imports_info
stringlengths 19
34.8k
| cluster_imports_info
stringlengths 15
3.66k
| libraries
list | libraries_info
stringlengths 6
661
| id
int64 0
2.92M
|
---|---|---|---|---|---|---|---|---|---|---|---|
int insertSelective(TbContentCategory record); | int insertSelective(TbContentCategory record); | /**
* This method was generated by MyBatis Generator.
* This method corresponds to the database table tb_content_category
*
* @mbg.generated Fri Apr 28 17:20:11 CST 2017
*/ | This method was generated by MyBatis Generator. This method corresponds to the database table tb_content_category | insertSelective | {
"repo_name": "fundmarkhua/fmall",
"path": "fm-manage/fm-manage-dao/src/main/java/com/fmall/mapper/TbContentCategoryMapper.java",
"license": "apache-2.0",
"size": 3204
} | [
"com.fmall.pojo.TbContentCategory"
] | import com.fmall.pojo.TbContentCategory; | import com.fmall.pojo.*; | [
"com.fmall.pojo"
] | com.fmall.pojo; | 2,010,425 |
public static void formatCodeAttributesAndDescriptions(List<CodeList> codeLists) {
codeLists.forEach(codeList -> {
formatCodeAttributesAndDescriptions(codeList);
codeList.getValues().forEach(ClsUtils::formatCodeAttributesAndDescriptions);
});
}
/**
* Formats both code attributes, fields and their descriptions for each language. For detailed description of the
* implementation refer to {@link ClsUtils#formatCodeAttributes(Code)} and {@link ClsUtils#formatDescriptions} | static void function(List<CodeList> codeLists) { codeLists.forEach(codeList -> { formatCodeAttributesAndDescriptions(codeList); codeList.getValues().forEach(ClsUtils::formatCodeAttributesAndDescriptions); }); } /** * Formats both code attributes, fields and their descriptions for each language. For detailed description of the * implementation refer to {@link ClsUtils#formatCodeAttributes(Code)} and {@link ClsUtils#formatDescriptions} | /**
* Normalizes the code lists by trimming the unnecessary white spaces at the start and the end of the descriptions
* and extras.
*
* @param codeLists
* a collection of code lists to be formatted
*/ | Normalizes the code lists by trimming the unnecessary white spaces at the start and the end of the descriptions and extras | formatCodeAttributesAndDescriptions | {
"repo_name": "SirmaITT/conservation-space-1.7.0",
"path": "docker/sirma-platform/platform/seip-parent/platform/cls/cls-impl/src/main/java/com/sirma/itt/emf/cls/util/ClsUtils.java",
"license": "lgpl-3.0",
"size": 3601
} | [
"com.sirma.sep.cls.model.Code",
"com.sirma.sep.cls.model.CodeList",
"java.util.List"
] | import com.sirma.sep.cls.model.Code; import com.sirma.sep.cls.model.CodeList; import java.util.List; | import com.sirma.sep.cls.model.*; import java.util.*; | [
"com.sirma.sep",
"java.util"
] | com.sirma.sep; java.util; | 1,933,330 |
public TransHopMeta findTransHop( StepMeta from, StepMeta to, boolean disabledToo ) {
for ( int i = 0; i < nrTransHops(); i++ ) {
TransHopMeta hi = getTransHop( i );
if ( hi.isEnabled() || disabledToo ) {
if ( hi.getFromStep() != null
&& hi.getToStep() != null && hi.getFromStep().equals( from ) && hi.getToStep().equals( to ) ) {
return hi;
}
}
}
return null;
} | TransHopMeta function( StepMeta from, StepMeta to, boolean disabledToo ) { for ( int i = 0; i < nrTransHops(); i++ ) { TransHopMeta hi = getTransHop( i ); if ( hi.isEnabled() disabledToo ) { if ( hi.getFromStep() != null && hi.getToStep() != null && hi.getFromStep().equals( from ) && hi.getToStep().equals( to ) ) { return hi; } } } return null; } | /**
* Search all hops for a hop where a certain step is at the start and another is at the end.
*
* @param from
* The step at the start of the hop.
* @param to
* The step at the end of the hop.
* @param disabledToo
* the disabled too
* @return The hop or null if no hop was found.
*/ | Search all hops for a hop where a certain step is at the start and another is at the end | findTransHop | {
"repo_name": "airy-ict/pentaho-kettle",
"path": "engine/src/org/pentaho/di/trans/TransMeta.java",
"license": "apache-2.0",
"size": 219546
} | [
"org.pentaho.di.trans.step.StepMeta"
] | import org.pentaho.di.trans.step.StepMeta; | import org.pentaho.di.trans.step.*; | [
"org.pentaho.di"
] | org.pentaho.di; | 457,623 |
@Override
public ResetableIterator GetEnumerator() throws OperationFailedException, LockingException, GeneralFailureException, CacheException {
_statusLatch.WaitForAny((byte) (NodeStatus.Initializing | NodeStatus.Running));
if (_internalCache == null) {
throw new UnsupportedOperationException();
}
if ((_statusLatch.IsAnyBitsSet(NodeStatus.Initializing))) {
Object tempVar = getCluster().getCoordinator().clone();
{
return Clustered_GetEnumerator((Address) ((tempVar instanceof Address) ? tempVar : null));
}
}
return new LazyKeysetEnumerator(this, (Object[]) handleKeyList(), false);
} | ResetableIterator function() throws OperationFailedException, LockingException, GeneralFailureException, CacheException { _statusLatch.WaitForAny((byte) (NodeStatus.Initializing NodeStatus.Running)); if (_internalCache == null) { throw new UnsupportedOperationException(); } if ((_statusLatch.IsAnyBitsSet(NodeStatus.Initializing))) { Object tempVar = getCluster().getCoordinator().clone(); { return Clustered_GetEnumerator((Address) ((tempVar instanceof Address) ? tempVar : null)); } } return new LazyKeysetEnumerator(this, (Object[]) handleKeyList(), false); } | /**
* Returns a .NET IEnumerator interface so that a client should be able to
* iterate over the elements of the cache store.
*
* @return IDictionaryEnumerator enumerator.
*/ | Returns a .NET IEnumerator interface so that a client should be able to iterate over the elements of the cache store | GetEnumerator | {
"repo_name": "joseananio/TayzGrid",
"path": "src/tgcache/src/com/alachisoft/tayzgrid/caching/topologies/clustered/ReplicatedServerCache.java",
"license": "apache-2.0",
"size": 207161
} | [
"com.alachisoft.tayzgrid.caching.statistics.NodeStatus",
"com.alachisoft.tayzgrid.caching.util.LazyKeysetEnumerator",
"com.alachisoft.tayzgrid.common.ResetableIterator",
"com.alachisoft.tayzgrid.common.net.Address",
"com.alachisoft.tayzgrid.runtime.exceptions.CacheException",
"com.alachisoft.tayzgrid.runtime.exceptions.GeneralFailureException",
"com.alachisoft.tayzgrid.runtime.exceptions.LockingException",
"com.alachisoft.tayzgrid.runtime.exceptions.OperationFailedException"
] | import com.alachisoft.tayzgrid.caching.statistics.NodeStatus; import com.alachisoft.tayzgrid.caching.util.LazyKeysetEnumerator; import com.alachisoft.tayzgrid.common.ResetableIterator; import com.alachisoft.tayzgrid.common.net.Address; import com.alachisoft.tayzgrid.runtime.exceptions.CacheException; import com.alachisoft.tayzgrid.runtime.exceptions.GeneralFailureException; import com.alachisoft.tayzgrid.runtime.exceptions.LockingException; import com.alachisoft.tayzgrid.runtime.exceptions.OperationFailedException; | import com.alachisoft.tayzgrid.caching.statistics.*; import com.alachisoft.tayzgrid.caching.util.*; import com.alachisoft.tayzgrid.common.*; import com.alachisoft.tayzgrid.common.net.*; import com.alachisoft.tayzgrid.runtime.exceptions.*; | [
"com.alachisoft.tayzgrid"
] | com.alachisoft.tayzgrid; | 2,830,482 |
public static PowerMaxSelectorType fromString(String selector) {
if (!StringUtils.isEmpty(selector)) {
for (PowerMaxSelectorType selectorType : PowerMaxSelectorType.values()) {
if (selectorType.getSelector().equals(selector)) {
return selectorType;
}
}
}
throw new IllegalArgumentException("Invalid label: " + selector);
} | static PowerMaxSelectorType function(String selector) { if (!StringUtils.isEmpty(selector)) { for (PowerMaxSelectorType selectorType : PowerMaxSelectorType.values()) { if (selectorType.getSelector().equals(selector)) { return selectorType; } } } throw new IllegalArgumentException(STR + selector); } | /**
* Get the ENUM value from its string selector
*
* @param selector
* the ENUM label as a string
*
* @return the corresponding ENUM value
*
* @throws IllegalArgumentException
* if no ENUM value corresponds to this string
*/ | Get the ENUM value from its string selector | fromString | {
"repo_name": "vgoldman/openhab",
"path": "bundles/binding/org.openhab.binding.powermax/src/main/java/org/openhab/binding/powermax/internal/PowerMaxSelectorType.java",
"license": "epl-1.0",
"size": 2279
} | [
"org.apache.commons.lang.StringUtils"
] | import org.apache.commons.lang.StringUtils; | import org.apache.commons.lang.*; | [
"org.apache.commons"
] | org.apache.commons; | 524,286 |
public DistributedConfigurationProcessor distributedConfiguration(); | DistributedConfigurationProcessor function(); | /**
* Gets distributed configuration processor.
*
* @return Distributed configuration processor.
*/ | Gets distributed configuration processor | distributedConfiguration | {
"repo_name": "ptupitsyn/ignite",
"path": "modules/core/src/main/java/org/apache/ignite/internal/GridKernalContext.java",
"license": "apache-2.0",
"size": 20940
} | [
"org.apache.ignite.internal.processors.configuration.distributed.DistributedConfigurationProcessor"
] | import org.apache.ignite.internal.processors.configuration.distributed.DistributedConfigurationProcessor; | import org.apache.ignite.internal.processors.configuration.distributed.*; | [
"org.apache.ignite"
] | org.apache.ignite; | 1,045,619 |
public static void saveCertificate(X509Certificate cert, File targetFile) {
File folder = targetFile.getAbsoluteFile().getParentFile();
if (!folder.exists()) {
folder.mkdirs();
}
File tmpFile = new File(folder, Long.toHexString(System.currentTimeMillis()) + ".tmp");
try {
boolean asPem = targetFile.getName().toLowerCase().endsWith(".pem");
if (asPem) {
// PEM encoded X509
PEMWriter pemWriter = null;
try {
pemWriter = new PEMWriter(new FileWriter(tmpFile));
pemWriter.writeObject(cert);
pemWriter.flush();
} finally {
if (pemWriter != null) {
pemWriter.close();
}
}
} else {
// DER encoded X509
FileOutputStream fos = null;
try {
fos = new FileOutputStream(tmpFile);
fos.write(cert.getEncoded());
fos.flush();
} finally {
if (fos != null) {
fos.close();
}
}
}
// rename tmp file to target
if (targetFile.exists()) {
targetFile.delete();
}
tmpFile.renameTo(targetFile);
} catch (Exception e) {
if (tmpFile.exists()) {
tmpFile.delete();
}
throw new RuntimeException("Failed to save certificate " + cert.getSubjectX500Principal().getName(), e);
}
}
| static void function(X509Certificate cert, File targetFile) { File folder = targetFile.getAbsoluteFile().getParentFile(); if (!folder.exists()) { folder.mkdirs(); } File tmpFile = new File(folder, Long.toHexString(System.currentTimeMillis()) + ".tmp"); try { boolean asPem = targetFile.getName().toLowerCase().endsWith(".pem"); if (asPem) { PEMWriter pemWriter = null; try { pemWriter = new PEMWriter(new FileWriter(tmpFile)); pemWriter.writeObject(cert); pemWriter.flush(); } finally { if (pemWriter != null) { pemWriter.close(); } } } else { FileOutputStream fos = null; try { fos = new FileOutputStream(tmpFile); fos.write(cert.getEncoded()); fos.flush(); } finally { if (fos != null) { fos.close(); } } } if (targetFile.exists()) { targetFile.delete(); } tmpFile.renameTo(targetFile); } catch (Exception e) { if (tmpFile.exists()) { tmpFile.delete(); } throw new RuntimeException(STR + cert.getSubjectX500Principal().getName(), e); } } | /**
* Saves the certificate to the file system. If the destination filename
* ends with the pem extension, the certificate is written in the PEM format,
* otherwise the certificate is written in the DER format.
*
* @param cert
* @param targetFile
*/ | Saves the certificate to the file system. If the destination filename ends with the pem extension, the certificate is written in the PEM format, otherwise the certificate is written in the DER format | saveCertificate | {
"repo_name": "ahnjune881214/gitblit",
"path": "src/main/java/com/gitblit/utils/X509Utils.java",
"license": "apache-2.0",
"size": 41191
} | [
"java.io.File",
"java.io.FileOutputStream",
"java.io.FileWriter",
"java.security.cert.X509Certificate",
"org.bouncycastle.openssl.PEMWriter"
] | import java.io.File; import java.io.FileOutputStream; import java.io.FileWriter; import java.security.cert.X509Certificate; import org.bouncycastle.openssl.PEMWriter; | import java.io.*; import java.security.cert.*; import org.bouncycastle.openssl.*; | [
"java.io",
"java.security",
"org.bouncycastle.openssl"
] | java.io; java.security; org.bouncycastle.openssl; | 76,199 |
public void unregisterWorkflowDef(String name, Integer version) {
Preconditions.checkArgument(StringUtils.isNotBlank(name), "Workflow name cannot be blank");
Preconditions.checkNotNull(version, "Version cannot be null");
delete("metadata/workflow/{name}/{version}", name, version);
}
// Task Metadata Operations | void function(String name, Integer version) { Preconditions.checkArgument(StringUtils.isNotBlank(name), STR); Preconditions.checkNotNull(version, STR); delete(STR, name, version); } | /**
* Removes the workflow definition of a workflow from the conductor server.
* It does not remove associated workflows. Use with caution.
*
* @param name Name of the workflow to be unregistered.
* @param version Version of the workflow definition to be unregistered.
*/ | Removes the workflow definition of a workflow from the conductor server. It does not remove associated workflows. Use with caution | unregisterWorkflowDef | {
"repo_name": "grfeng/conductor",
"path": "client/src/main/java/com/netflix/conductor/client/http/MetadataClient.java",
"license": "apache-2.0",
"size": 6994
} | [
"com.google.common.base.Preconditions",
"org.apache.commons.lang.StringUtils"
] | import com.google.common.base.Preconditions; import org.apache.commons.lang.StringUtils; | import com.google.common.base.*; import org.apache.commons.lang.*; | [
"com.google.common",
"org.apache.commons"
] | com.google.common; org.apache.commons; | 439,808 |
List<CompleteEarthquake> getAllEarthquakesByMagnitude(Double magnitude); | List<CompleteEarthquake> getAllEarthquakesByMagnitude(Double magnitude); | /**
* Gets all the earthquakes with a specific magnitude.
*
* @param magnitude The magnitude searched.
*
* @return The list of earthquakes of that magnitude.
*/ | Gets all the earthquakes with a specific magnitude | getAllEarthquakesByMagnitude | {
"repo_name": "AlenaHa/tw-NEDa",
"path": "tw-NEDa-backend/src/main/java/org/neda/service/EarthquakeService.java",
"license": "mit",
"size": 2462
} | [
"java.util.List",
"org.neda.entity.CompleteEarthquake"
] | import java.util.List; import org.neda.entity.CompleteEarthquake; | import java.util.*; import org.neda.entity.*; | [
"java.util",
"org.neda.entity"
] | java.util; org.neda.entity; | 1,965,155 |
protected TimeSource activateTimeSource()
{
synchronized (millisEstimatorLock)
{
if (millisEstimator == null)
{
millisEstimator = new EstimatorTimeSource(new SystemTimeSource(), 10, logger);
}
}
return millisEstimator;
} | TimeSource function() { synchronized (millisEstimatorLock) { if (millisEstimator == null) { millisEstimator = new EstimatorTimeSource(new SystemTimeSource(), 10, logger); } } return millisEstimator; } | /**
* Returns the TimeSource for this Cache. The TimeSource is used for any situation where the current time is required, for
* example the input date for a cache entry or on getting the current time when doing expiration.
*
* Instantiates the TimeSource if it is not yet there.
*
* @return The TimeSource for this Cache.
*/ | Returns the TimeSource for this Cache. The TimeSource is used for any situation where the current time is required, for example the input date for a cache entry or on getting the current time when doing expiration. Instantiates the TimeSource if it is not yet there | activateTimeSource | {
"repo_name": "trivago/triava",
"path": "src/main/java/com/trivago/triava/tcache/Cache.java",
"license": "apache-2.0",
"size": 54279
} | [
"com.trivago.triava.time.EstimatorTimeSource",
"com.trivago.triava.time.SystemTimeSource",
"com.trivago.triava.time.TimeSource"
] | import com.trivago.triava.time.EstimatorTimeSource; import com.trivago.triava.time.SystemTimeSource; import com.trivago.triava.time.TimeSource; | import com.trivago.triava.time.*; | [
"com.trivago.triava"
] | com.trivago.triava; | 1,732,221 |
@Override
public void onAttach(IScriptable target) {
} | void function(IScriptable target) { } | /**
* Called when the script is attached to a target using {@link org.gearvrf.script.IScriptManager#attachScriptFile(IScriptable, IScriptFile)}.
*/ | Called when the script is attached to a target using <code>org.gearvrf.script.IScriptManager#attachScriptFile(IScriptable, IScriptFile)</code> | onAttach | {
"repo_name": "xcaostagit/GearVRf",
"path": "GVRf/Framework/framework/src/main/java/org/gearvrf/GVRMain.java",
"license": "apache-2.0",
"size": 13044
} | [
"org.gearvrf.script.IScriptable"
] | import org.gearvrf.script.IScriptable; | import org.gearvrf.script.*; | [
"org.gearvrf.script"
] | org.gearvrf.script; | 2,634,965 |
@ServiceMethod(returns = ReturnType.SINGLE)
private Mono<Response<Flux<ByteBuffer>>> createOrUpdateByIdWithResponseAsync(
String resourceId, String apiVersion, GenericResourceInner parameters, Context context) {
if (this.client.getEndpoint() == null) {
return Mono
.error(
new IllegalArgumentException(
"Parameter this.client.getEndpoint() is required and cannot be null."));
}
if (resourceId == null) {
return Mono.error(new IllegalArgumentException("Parameter resourceId is required and cannot be null."));
}
if (apiVersion == null) {
return Mono.error(new IllegalArgumentException("Parameter apiVersion is required and cannot be null."));
}
if (parameters == null) {
return Mono.error(new IllegalArgumentException("Parameter parameters is required and cannot be null."));
} else {
parameters.validate();
}
final String accept = "application/json";
context = this.client.mergeContext(context);
return service
.createOrUpdateById(this.client.getEndpoint(), resourceId, apiVersion, parameters, accept, context);
} | @ServiceMethod(returns = ReturnType.SINGLE) Mono<Response<Flux<ByteBuffer>>> function( String resourceId, String apiVersion, GenericResourceInner parameters, Context context) { if (this.client.getEndpoint() == null) { return Mono .error( new IllegalArgumentException( STR)); } if (resourceId == null) { return Mono.error(new IllegalArgumentException(STR)); } if (apiVersion == null) { return Mono.error(new IllegalArgumentException(STR)); } if (parameters == null) { return Mono.error(new IllegalArgumentException(STR)); } else { parameters.validate(); } final String accept = STR; context = this.client.mergeContext(context); return service .createOrUpdateById(this.client.getEndpoint(), resourceId, apiVersion, parameters, accept, context); } | /**
* Create a resource by ID.
*
* @param resourceId The fully qualified ID of the resource, including the resource name and resource type. Use the
* format,
* /subscriptions/{guid}/resourceGroups/{resource-group-name}/{resource-provider-namespace}/{resource-type}/{resource-name}.
* @param apiVersion The API version to use for the operation.
* @param parameters Create or update resource parameters.
* @param context The context to associate with this operation.
* @throws IllegalArgumentException thrown if parameters fail the validation.
* @throws ManagementException thrown if the request is rejected by server.
* @throws RuntimeException all other wrapped checked exceptions if the request fails to be sent.
* @return resource information.
*/ | Create a resource by ID | createOrUpdateByIdWithResponseAsync | {
"repo_name": "Azure/azure-sdk-for-java",
"path": "sdk/resourcemanagerhybrid/azure-resourcemanager-resources/src/main/java/com/azure/resourcemanager/resources/implementation/ResourcesClientImpl.java",
"license": "mit",
"size": 224761
} | [
"com.azure.core.annotation.ReturnType",
"com.azure.core.annotation.ServiceMethod",
"com.azure.core.http.rest.Response",
"com.azure.core.util.Context",
"com.azure.resourcemanager.resources.fluent.models.GenericResourceInner",
"java.nio.ByteBuffer"
] | import com.azure.core.annotation.ReturnType; import com.azure.core.annotation.ServiceMethod; import com.azure.core.http.rest.Response; import com.azure.core.util.Context; import com.azure.resourcemanager.resources.fluent.models.GenericResourceInner; import java.nio.ByteBuffer; | import com.azure.core.annotation.*; import com.azure.core.http.rest.*; import com.azure.core.util.*; import com.azure.resourcemanager.resources.fluent.models.*; import java.nio.*; | [
"com.azure.core",
"com.azure.resourcemanager",
"java.nio"
] | com.azure.core; com.azure.resourcemanager; java.nio; | 1,613,216 |
public synchronized void close() throws MessagingException {
if (!super.isConnected()) // Already closed.
return;
IMAPProtocol protocol = null;
try {
boolean isEmpty;
synchronized (pool) {
// If there's no authenticated connections available
// don't create a new one
isEmpty = pool.authenticatedConnections.isEmpty();
}
if (isEmpty) {
if (pool.debug)
out.println("DEBUG: close() - no connections ");
cleanup();
return;
}
protocol = getStoreProtocol();
synchronized (pool) {
pool.authenticatedConnections.removeElement(protocol);
}
protocol.logout();
} catch (ProtocolException pex) {
// Hmm .. will this ever happen ?
throw new MessagingException(pex.getMessage(), pex);
} finally {
releaseStoreProtocol(protocol);
}
} | synchronized void function() throws MessagingException { if (!super.isConnected()) return; IMAPProtocol protocol = null; try { boolean isEmpty; synchronized (pool) { isEmpty = pool.authenticatedConnections.isEmpty(); } if (isEmpty) { if (pool.debug) out.println(STR); cleanup(); return; } protocol = getStoreProtocol(); synchronized (pool) { pool.authenticatedConnections.removeElement(protocol); } protocol.logout(); } catch (ProtocolException pex) { throw new MessagingException(pex.getMessage(), pex); } finally { releaseStoreProtocol(protocol); } } | /**
* Close this Store.
*/ | Close this Store | close | {
"repo_name": "arthurzaczek/kolab-android",
"path": "javamail/com/sun/mail/imap/IMAPStore.java",
"license": "gpl-3.0",
"size": 60428
} | [
"com.sun.mail.iap.ProtocolException",
"com.sun.mail.imap.protocol.IMAPProtocol",
"javax.mail.MessagingException"
] | import com.sun.mail.iap.ProtocolException; import com.sun.mail.imap.protocol.IMAPProtocol; import javax.mail.MessagingException; | import com.sun.mail.iap.*; import com.sun.mail.imap.protocol.*; import javax.mail.*; | [
"com.sun.mail",
"javax.mail"
] | com.sun.mail; javax.mail; | 2,676,638 |
public Optional<Unit> getDislodgedUnit(final Province province) {
final ProvinceData pd = provArray.get(province.getIndex());
if (pd != null) {
return Optional.ofNullable(pd.getDislodgedUnit());
}
return Optional.empty();
}// getDislodgedUnit()
// last occupier | Optional<Unit> function(final Province province) { final ProvinceData pd = provArray.get(province.getIndex()); if (pd != null) { return Optional.ofNullable(pd.getDislodgedUnit()); } return Optional.empty(); } | /**
* Get the dislodged unit in this Province. Returns null if no dislodged unit exists.
*/ | Get the dislodged unit in this Province. Returns null if no dislodged unit exists | getDislodgedUnit | {
"repo_name": "takaki/jdip",
"path": "src/main/java/dip/world/Position.java",
"license": "gpl-2.0",
"size": 22008
} | [
"java.util.Optional"
] | import java.util.Optional; | import java.util.*; | [
"java.util"
] | java.util; | 1,942,126 |
public static void initTorque() {
try {
Torque.init(props);
}
catch (Exception e) {
e.printStackTrace();
}
}
| static void function() { try { Torque.init(props); } catch (Exception e) { e.printStackTrace(); } } | /**
* This method initialize the Torque.
*/ | This method initialize the Torque | initTorque | {
"repo_name": "trackplus/Genji",
"path": "src/test/java/com/aurel/track/dbase/DatabaseHandler.java",
"license": "gpl-3.0",
"size": 17827
} | [
"org.apache.torque.Torque"
] | import org.apache.torque.Torque; | import org.apache.torque.*; | [
"org.apache.torque"
] | org.apache.torque; | 1,394,906 |
final String id = UUID.randomUUID().toString();
return new Person(id, name, birth);
} | final String id = UUID.randomUUID().toString(); return new Person(id, name, birth); } | /**
* Creates a new person with the given name and birth date. An id is
* automatically generated.
*
* @param name
* the person's name
* @param birth
* the person's birth
* @return the newly created person
*/ | Creates a new person with the given name and birth date. An id is automatically generated | createNew | {
"repo_name": "rduerig/nansai",
"path": "nansai/src/com/github/nansai/data/Person.java",
"license": "mit",
"size": 2442
} | [
"java.util.UUID"
] | import java.util.UUID; | import java.util.*; | [
"java.util"
] | java.util; | 254,583 |
protected static PyDouble readDoubleZero(final MarshalBuffer byteBuffer)
{
final PyFactory.PyDoubleBuilder builder = getFactory().getPyDoubleBuilder();
builder.value(0.0);
return (builder.build());
} | static PyDouble function(final MarshalBuffer byteBuffer) { final PyFactory.PyDoubleBuilder builder = getFactory().getPyDoubleBuilder(); builder.value(0.0); return (builder.build()); } | /**
* Returns <tt>PyDouble</tt>.
*
* @param byteBuffer
* byteBuffer
* @return <tt>PyDouble</tt>
*/ | Returns PyDouble | readDoubleZero | {
"repo_name": "jevetools/unmarshal",
"path": "com.jevetools.unmarshal.api.impl/src/com/jevetools/unmarshal/api/impl/RecursiveImpl.java",
"license": "bsd-3-clause",
"size": 31538
} | [
"com.jevetools.unmarshal.python.api.PyDouble",
"com.jevetools.unmarshal.python.api.PyFactory"
] | import com.jevetools.unmarshal.python.api.PyDouble; import com.jevetools.unmarshal.python.api.PyFactory; | import com.jevetools.unmarshal.python.api.*; | [
"com.jevetools.unmarshal"
] | com.jevetools.unmarshal; | 2,684,670 |
PropertyValue execute(PropertyValue property); | PropertyValue execute(PropertyValue property); | /**
* Returns a changed property value based in the property before drilling.
*
* @param property current property
* @return property after drilling
*/ | Returns a changed property value based in the property before drilling | execute | {
"repo_name": "smee/gradoop",
"path": "gradoop-flink/src/main/java/org/gradoop/flink/model/impl/operators/drilling/functions/drillfunctions/DrillFunction.java",
"license": "apache-2.0",
"size": 1189
} | [
"org.gradoop.common.model.impl.properties.PropertyValue"
] | import org.gradoop.common.model.impl.properties.PropertyValue; | import org.gradoop.common.model.impl.properties.*; | [
"org.gradoop.common"
] | org.gradoop.common; | 934,122 |
public boolean containsAll(Collection<?> arg0) {
for (Object o : arg0) {
if (!contains(o)) {
return false;
}
}
return true;
} | boolean function(Collection<?> arg0) { for (Object o : arg0) { if (!contains(o)) { return false; } } return true; } | /**
* Set inclusion test
*
* @param arg0 Objects to check
* @return true if the list contains all objects in arg0, false otherwise
*/ | Set inclusion test | containsAll | {
"repo_name": "ut-osa/laminar",
"path": "jikesrvm-3.0.0/rvm/src/org/jikesrvm/util/LinkedListRVM.java",
"license": "bsd-3-clause",
"size": 7043
} | [
"java.util.Collection"
] | import java.util.Collection; | import java.util.*; | [
"java.util"
] | java.util; | 2,224,278 |
if (env.getActiveProfiles().length == 0) {
log.warn("No Spring profile configured, running with default configuration");
} else {
log.info("Running with Spring profile(s) : {}", Arrays.toString(env.getActiveProfiles()));
Collection activeProfiles = Arrays.asList(env.getActiveProfiles());
if (activeProfiles.contains("dev") && activeProfiles.contains("prod")) {
log.error("You have misconfigured your application! " +
"It should not run with both the 'dev' and 'prod' profiles at the same time.");
}
if (activeProfiles.contains("prod") && activeProfiles.contains("fast")) {
log.error("You have misconfigured your application! " +
"It should not run with both the 'prod' and 'fast' profiles at the same time.");
}
if (activeProfiles.contains("dev") && activeProfiles.contains("cloud")) {
log.error("You have misconfigured your application! " +
"It should not run with both the 'dev' and 'cloud' profiles at the same time.");
}
}
} | if (env.getActiveProfiles().length == 0) { log.warn(STR); } else { log.info(STR, Arrays.toString(env.getActiveProfiles())); Collection activeProfiles = Arrays.asList(env.getActiveProfiles()); if (activeProfiles.contains("dev") && activeProfiles.contains("prod")) { log.error(STR + STR); } if (activeProfiles.contains("prod") && activeProfiles.contains("fast")) { log.error(STR + STR); } if (activeProfiles.contains("dev") && activeProfiles.contains("cloud")) { log.error(STR + STR); } } } | /**
* Initializes jhipster.
* <p/>
* Spring profiles can be configured with a program arguments --spring.profiles.active=your-active-profile
* <p/>
* <p>
* You can find more information on how profiles work with JHipster on <a href="http://jhipster.github.io/profiles.html">http://jhipster.github.io/profiles.html</a>.
* </p>
*/ | Initializes jhipster. Spring profiles can be configured with a program arguments --spring.profiles.active=your-active-profile You can find more information on how profiles work with JHipster on HREF | initApplication | {
"repo_name": "canmahmut/shopping-list",
"path": "src/main/java/de/codenorm/certification/Application.java",
"license": "gpl-2.0",
"size": 4067
} | [
"java.util.Arrays",
"java.util.Collection"
] | import java.util.Arrays; import java.util.Collection; | import java.util.*; | [
"java.util"
] | java.util; | 1,385,735 |
@Override
public Map<K, ICacheElement<K, V>> getMultiple(Set<K> keys)
{
return new HashMap<K, ICacheElement<K, V>>();
} | Map<K, ICacheElement<K, V>> function(Set<K> keys) { return new HashMap<K, ICacheElement<K, V>>(); } | /**
* Gets multiple items from the cache based on the given set of keys.
* <p>
* @param keys
* @return a map of K key to ICacheElement<K, V> element, or an empty map if there is
* no data in cache for any of these keys
*/ | Gets multiple items from the cache based on the given set of keys. | getMultiple | {
"repo_name": "mohanaraosv/commons-jcs",
"path": "commons-jcs-core/src/test/java/org/apache/commons/jcs/engine/control/CompositeCacheDiskUsageUnitTest.java",
"license": "apache-2.0",
"size": 16607
} | [
"java.util.HashMap",
"java.util.Map",
"java.util.Set",
"org.apache.commons.jcs.engine.behavior.ICacheElement"
] | import java.util.HashMap; import java.util.Map; import java.util.Set; import org.apache.commons.jcs.engine.behavior.ICacheElement; | import java.util.*; import org.apache.commons.jcs.engine.behavior.*; | [
"java.util",
"org.apache.commons"
] | java.util; org.apache.commons; | 1,096,680 |
@TargetApi(21)
public Bundler putSizeF(String key, SizeF value) {
bundle.putSizeF(key, value);
return this;
} | @TargetApi(21) Bundler function(String key, SizeF value) { bundle.putSizeF(key, value); return this; } | /**
* Inserts a SizeF value into the mapping of this Bundle, replacing
* any existing value for the given key. Either key or value may be null.
*
* @param key a String, or null
* @param value a SizeF object, or null
*/ | Inserts a SizeF value into the mapping of this Bundle, replacing any existing value for the given key. Either key or value may be null | putSizeF | {
"repo_name": "HKMOpen/SmartTabLayout",
"path": "utils-v4/src/main/java/com/ogaclejapan/smarttablayout/utils/v4/Bundler.java",
"license": "apache-2.0",
"size": 14195
} | [
"android.annotation.TargetApi",
"android.util.SizeF"
] | import android.annotation.TargetApi; import android.util.SizeF; | import android.annotation.*; import android.util.*; | [
"android.annotation",
"android.util"
] | android.annotation; android.util; | 2,622,118 |
private void deleteUnassignedMassnObjectReferences() {
// ensures that there are no references to the deleted gup_unterhaltungsmassnahme object
try {
final DBServer server = getDbServer();
final Statement st = server.getActiveDBConnection().getConnection().createStatement();
st.executeUpdate(CLEAR_PLANUNGSABSCHNITT_ARRAY_QUERY);
st.executeUpdate(CLEAR_LOS_ARRAY_QUERY);
} catch (Exception e) {
log.error("Error while executing los trigger.", e);
}
} | void function() { try { final DBServer server = getDbServer(); final Statement st = server.getActiveDBConnection().getConnection().createStatement(); st.executeUpdate(CLEAR_PLANUNGSABSCHNITT_ARRAY_QUERY); st.executeUpdate(CLEAR_LOS_ARRAY_QUERY); } catch (Exception e) { log.error(STR, e); } } | /**
* DOCUMENT ME!
*/ | DOCUMENT ME | deleteUnassignedMassnObjectReferences | {
"repo_name": "cismet/cids-custom-wrrl-db-mv-server",
"path": "src/main/java/de/cismet/cids/custom/wrrl_db_mv/server/trigger/GupTrigger.java",
"license": "lgpl-3.0",
"size": 9171
} | [
"java.sql.Statement"
] | import java.sql.Statement; | import java.sql.*; | [
"java.sql"
] | java.sql; | 1,903,765 |
protected String addNewHTMLPage() {
final String newId = ((CPManager) CoreSpringFactory.getBean(CPManager.class)).addBlankPage(cp, translate("cptreecontroller.newpage.title"),
currentPage.getIdentifier());
final CPPage newPage = new CPPage(newId, cp);
// Create an html file
final VFSContainer root = cp.getRootDir();
final VFSLeaf htmlFile = root.createChildLeaf(newId + ".html");
newPage.setFile(htmlFile);
updatePage(newPage);
return newId;
} | String function() { final String newId = ((CPManager) CoreSpringFactory.getBean(CPManager.class)).addBlankPage(cp, translate(STR), currentPage.getIdentifier()); final CPPage newPage = new CPPage(newId, cp); final VFSContainer root = cp.getRootDir(); final VFSLeaf htmlFile = root.createChildLeaf(newId + ".html"); newPage.setFile(htmlFile); updatePage(newPage); return newId; } | /**
* Adds a new page to the CP
*
* @return
*/ | Adds a new page to the CP | addNewHTMLPage | {
"repo_name": "huihoo/olat",
"path": "olat7.8/src/main/java/org/olat/presentation/ims/cp/CPTreeController.java",
"license": "apache-2.0",
"size": 19440
} | [
"org.olat.data.commons.vfs.VFSContainer",
"org.olat.data.commons.vfs.VFSLeaf",
"org.olat.lms.ims.cp.CPManager",
"org.olat.lms.ims.cp.CPPage",
"org.olat.system.spring.CoreSpringFactory"
] | import org.olat.data.commons.vfs.VFSContainer; import org.olat.data.commons.vfs.VFSLeaf; import org.olat.lms.ims.cp.CPManager; import org.olat.lms.ims.cp.CPPage; import org.olat.system.spring.CoreSpringFactory; | import org.olat.data.commons.vfs.*; import org.olat.lms.ims.cp.*; import org.olat.system.spring.*; | [
"org.olat.data",
"org.olat.lms",
"org.olat.system"
] | org.olat.data; org.olat.lms; org.olat.system; | 981,432 |
void metaInf(Action<? super MetaInfNormalization> configuration); | void metaInf(Action<? super MetaInfNormalization> configuration); | /**
* Configures the normalization strategy for the {@code META-INF} directory in archives.
*
* @since 6.6
*/ | Configures the normalization strategy for the META-INF directory in archives | metaInf | {
"repo_name": "gradle/gradle",
"path": "subprojects/core-api/src/main/java/org/gradle/normalization/RuntimeClasspathNormalization.java",
"license": "apache-2.0",
"size": 2864
} | [
"org.gradle.api.Action"
] | import org.gradle.api.Action; | import org.gradle.api.*; | [
"org.gradle.api"
] | org.gradle.api; | 453,227 |
public void setDisplayConnectors(boolean dc,
Color c) {
setDisplayConnectors(dc);
m_connectorColor = c;
} | void function(boolean dc, Color c) { setDisplayConnectors(dc); m_connectorColor = c; } | /**
* Turn on/off the connector points
*
* @param dc a <code>boolean</code> value
* @param c the Color to use
*/ | Turn on/off the connector points | setDisplayConnectors | {
"repo_name": "goddesss/DataModeling",
"path": "src/weka/gui/beans/BeanVisual.java",
"license": "gpl-2.0",
"size": 12172
} | [
"java.awt.Color"
] | import java.awt.Color; | import java.awt.*; | [
"java.awt"
] | java.awt; | 472,703 |
private IdentityProviderCreateContractProperties innerProperties() {
return this.innerProperties;
} | IdentityProviderCreateContractProperties function() { return this.innerProperties; } | /**
* Get the innerProperties property: Identity Provider contract properties.
*
* @return the innerProperties value.
*/ | Get the innerProperties property: Identity Provider contract properties | innerProperties | {
"repo_name": "Azure/azure-sdk-for-java",
"path": "sdk/apimanagement/azure-resourcemanager-apimanagement/src/main/java/com/azure/resourcemanager/apimanagement/models/IdentityProviderCreateContract.java",
"license": "mit",
"size": 11091
} | [
"com.azure.resourcemanager.apimanagement.fluent.models.IdentityProviderCreateContractProperties"
] | import com.azure.resourcemanager.apimanagement.fluent.models.IdentityProviderCreateContractProperties; | import com.azure.resourcemanager.apimanagement.fluent.models.*; | [
"com.azure.resourcemanager"
] | com.azure.resourcemanager; | 1,319,825 |
Method testMethod = method.getMethod();
if (testMethod.getName().endsWith("Bookmark")) {
((AutomatedFunctionalTestBase) test).enableBookmarkMode();
} else if (testMethod.getName().endsWith("Nav")) {
((AutomatedFunctionalTestBase) test).enableNavigationMode();
}
return super.methodInvoker(method, test);
} | Method testMethod = method.getMethod(); if (testMethod.getName().endsWith(STR)) { ((AutomatedFunctionalTestBase) test).enableBookmarkMode(); } else if (testMethod.getName().endsWith("Nav")) { ((AutomatedFunctionalTestBase) test).enableNavigationMode(); } return super.methodInvoker(method, test); } | /**
* Test methods ending with Bookmark will have {@see AutomatedFunctionalTestBase#enableBookmarkMode} called,
* test methods ending with Nav will have {@see AutomatedFunctionalTestBase#enableNavigationMode} called.
*
* @param method test method to check for ending in Bookmark or Nav
* @param test which extends AutomatedFunctionalTestBase
* @return {@see BlockJUnit4ClassRunner#methodInvoker}
*/ | Test methods ending with Bookmark will have AutomatedFunctionalTestBase#enableBookmarkMode called, test methods ending with Nav will have AutomatedFunctionalTestBase#enableNavigationMode called | methodInvoker | {
"repo_name": "geothomasp/kualico-rice-kc",
"path": "rice-tools-test/src/main/java/org/kuali/rice/testtools/selenium/AutomatedFunctionalTestRunner.java",
"license": "apache-2.0",
"size": 2383
} | [
"java.lang.reflect.Method"
] | import java.lang.reflect.Method; | import java.lang.reflect.*; | [
"java.lang"
] | java.lang; | 2,754,190 |
public Observable<ServiceResponse<RouteTableInner>> getByResourceGroupWithServiceResponseAsync(String resourceGroupName, String routeTableName, String expand) {
if (resourceGroupName == null) {
throw new IllegalArgumentException("Parameter resourceGroupName is required and cannot be null.");
}
if (routeTableName == null) {
throw new IllegalArgumentException("Parameter routeTableName is required and cannot be null.");
}
if (this.client.subscriptionId() == null) {
throw new IllegalArgumentException("Parameter this.client.subscriptionId() is required and cannot be null.");
} | Observable<ServiceResponse<RouteTableInner>> function(String resourceGroupName, String routeTableName, String expand) { if (resourceGroupName == null) { throw new IllegalArgumentException(STR); } if (routeTableName == null) { throw new IllegalArgumentException(STR); } if (this.client.subscriptionId() == null) { throw new IllegalArgumentException(STR); } | /**
* Gets the specified route table.
*
* @param resourceGroupName The name of the resource group.
* @param routeTableName The name of the route table.
* @param expand Expands referenced resources.
* @throws IllegalArgumentException thrown if parameters fail the validation
* @return the observable to the RouteTableInner object
*/ | Gets the specified route table | getByResourceGroupWithServiceResponseAsync | {
"repo_name": "navalev/azure-sdk-for-java",
"path": "sdk/network/mgmt-v2019_04_01/src/main/java/com/microsoft/azure/management/network/v2019_04_01/implementation/RouteTablesInner.java",
"license": "mit",
"size": 75681
} | [
"com.microsoft.rest.ServiceResponse"
] | import com.microsoft.rest.ServiceResponse; | import com.microsoft.rest.*; | [
"com.microsoft.rest"
] | com.microsoft.rest; | 2,830,088 |
private void validateArguments(TestContext context, PropertyList propertyList, List<Metadata> metadataList,
SubEngineActionType actionType) throws WebServiceException {
if (context == null) {
throw new WebServiceException("TestContext must not be null.");
}
if (propertyList == null) {
throw new WebServiceException("PropertyList must not be null.");
}
if (metadataList == null || metadataList.isEmpty()) {
throw new WebServiceException("MetadataList must not be null or empty.");
}
if (actionType == null) {
throw new WebServiceException("ActionType must not be null.");
}
if (!(actionType instanceof WebServiceActionType)) {
throw new WebServiceException("ActionType must be a WebServiceActionType.");
}
}
/**
* Returns the metadata for the WebService-call
*
* @param metadataList
* the list of {@link Metadata} | void function(TestContext context, PropertyList propertyList, List<Metadata> metadataList, SubEngineActionType actionType) throws WebServiceException { if (context == null) { throw new WebServiceException(STR); } if (propertyList == null) { throw new WebServiceException(STR); } if (metadataList == null metadataList.isEmpty()) { throw new WebServiceException(STR); } if (actionType == null) { throw new WebServiceException(STR); } if (!(actionType instanceof WebServiceActionType)) { throw new WebServiceException(STR); } } /** * Returns the metadata for the WebService-call * * @param metadataList * the list of {@link Metadata} | /**
* Initial validation of metadata and action type.
*
* @param context
* the test context
* @param propertyList
* the list of properties
* @param metadataList
* the metadata list
* @param actionType
* the action type
* @throws WebServiceException
* thrown, if an error occurs
*/ | Initial validation of metadata and action type | validateArguments | {
"repo_name": "NABUCCO/org.nabucco.testautomation.engine.proxy.ws",
"path": "org.nabucco.testautomation.engine.proxy.ws/src/main/org/nabucco/testautomation/engine/proxy/ws/command/soap/server/SoapServerImpl.java",
"license": "epl-1.0",
"size": 8090
} | [
"java.util.List",
"org.nabucco.testautomation.engine.base.context.TestContext",
"org.nabucco.testautomation.engine.proxy.SubEngineActionType",
"org.nabucco.testautomation.engine.proxy.ws.WebServiceActionType",
"org.nabucco.testautomation.engine.proxy.ws.exception.WebServiceException",
"org.nabucco.testautomation.property.facade.datatype.PropertyList",
"org.nabucco.testautomation.script.facade.datatype.metadata.Metadata"
] | import java.util.List; import org.nabucco.testautomation.engine.base.context.TestContext; import org.nabucco.testautomation.engine.proxy.SubEngineActionType; import org.nabucco.testautomation.engine.proxy.ws.WebServiceActionType; import org.nabucco.testautomation.engine.proxy.ws.exception.WebServiceException; import org.nabucco.testautomation.property.facade.datatype.PropertyList; import org.nabucco.testautomation.script.facade.datatype.metadata.Metadata; | import java.util.*; import org.nabucco.testautomation.engine.base.context.*; import org.nabucco.testautomation.engine.proxy.*; import org.nabucco.testautomation.engine.proxy.ws.*; import org.nabucco.testautomation.engine.proxy.ws.exception.*; import org.nabucco.testautomation.property.facade.datatype.*; import org.nabucco.testautomation.script.facade.datatype.metadata.*; | [
"java.util",
"org.nabucco.testautomation"
] | java.util; org.nabucco.testautomation; | 190,265 |
public List<GHVerifiedKey> getPublicVerifiedKeys() throws IOException {
return Collections.unmodifiableList(Arrays.asList(root.retrieve().to(
"/users/" + getLogin() + "/keys", GHVerifiedKey[].class)));
} | List<GHVerifiedKey> function() throws IOException { return Collections.unmodifiableList(Arrays.asList(root.retrieve().to( STR + getLogin() + "/keys", GHVerifiedKey[].class))); } | /**
* Returns the read-only list of all the public verified keys of the current user.
*
* Differently from the getPublicKeys() method, the retrieval of the user's
* verified public keys does not require any READ/WRITE OAuth Scope to the
* user's profile.
*
* @return
* Always non-null.
*/ | Returns the read-only list of all the public verified keys of the current user. Differently from the getPublicKeys() method, the retrieval of the user's verified public keys does not require any READ/WRITE OAuth Scope to the user's profile | getPublicVerifiedKeys | {
"repo_name": "PankajHingane-REISysIn/KnowledgeArticleExportImport",
"path": "src/main/java/org/kohsuke/github/GHMyself.java",
"license": "mit",
"size": 6661
} | [
"java.io.IOException",
"java.util.Arrays",
"java.util.Collections",
"java.util.List"
] | import java.io.IOException; import java.util.Arrays; import java.util.Collections; import java.util.List; | import java.io.*; import java.util.*; | [
"java.io",
"java.util"
] | java.io; java.util; | 990,922 |
public ServiceFuture<StreamingJobInner> getByResourceGroupAsync(String resourceGroupName, String jobName, final ServiceCallback<StreamingJobInner> serviceCallback) {
return ServiceFuture.fromHeaderResponse(getByResourceGroupWithServiceResponseAsync(resourceGroupName, jobName), serviceCallback);
} | ServiceFuture<StreamingJobInner> function(String resourceGroupName, String jobName, final ServiceCallback<StreamingJobInner> serviceCallback) { return ServiceFuture.fromHeaderResponse(getByResourceGroupWithServiceResponseAsync(resourceGroupName, jobName), serviceCallback); } | /**
* Gets details about the specified streaming job.
*
* @param resourceGroupName The name of the resource group that contains the resource. You can obtain this value from the Azure Resource Manager API or the portal.
* @param jobName The name of the streaming job.
* @param serviceCallback the async ServiceCallback to handle successful and failed responses.
* @throws IllegalArgumentException thrown if parameters fail the validation
* @return the {@link ServiceFuture} object
*/ | Gets details about the specified streaming job | getByResourceGroupAsync | {
"repo_name": "hovsepm/azure-sdk-for-java",
"path": "streamanalytics/resource-manager/v2016_03_01/src/main/java/com/microsoft/azure/management/streamanalytics/v2016_03_01/implementation/StreamingJobsInner.java",
"license": "mit",
"size": 143599
} | [
"com.microsoft.rest.ServiceCallback",
"com.microsoft.rest.ServiceFuture"
] | import com.microsoft.rest.ServiceCallback; import com.microsoft.rest.ServiceFuture; | import com.microsoft.rest.*; | [
"com.microsoft.rest"
] | com.microsoft.rest; | 1,389,112 |
public native List<String> getListInterestsInNetwork() throws
RepaConnectException, ClassNotFoundException; | native List<String> function() throws RepaConnectException, ClassNotFoundException; | /**
* Get a list of interest collected by daemon in NETWORK
* return the interests in list passed by parameter
*
* @return list of interest
*
* @throws RepaConnectException - if an error occurs when connect the socket(include number exception)
* @throws ClassNotFoundException - class not found in JNI
*/ | Get a list of interest collected by daemon in NETWORK return the interests in list passed by parameter | getListInterestsInNetwork | {
"repo_name": "hmoraes/repa-java-api",
"path": "src/ufrj/coppe/lcp/repa/RepaSocket.java",
"license": "gpl-2.0",
"size": 7827
} | [
"java.util.List"
] | import java.util.List; | import java.util.*; | [
"java.util"
] | java.util; | 1,193,119 |
public Collection<Point2D> vertices() {
return vertices;
}
| Collection<Point2D> function() { return vertices; } | /**
* Returns the points of the polygon. The result is a pointer to the inner
* collection of vertices.
*/ | Returns the points of the polygon. The result is a pointer to the inner collection of vertices | vertices | {
"repo_name": "pokowaka/android-geom",
"path": "geom/src/main/java/math/geom2d/polygon/SimplePolygon2D.java",
"license": "lgpl-2.1",
"size": 18458
} | [
"java.util.Collection",
"math.geom2d.Point2D"
] | import java.util.Collection; import math.geom2d.Point2D; | import java.util.*; import math.geom2d.*; | [
"java.util",
"math.geom2d"
] | java.util; math.geom2d; | 2,143,044 |
public static Trade adaptTrade(ANXTrade anxTrade) {
OrderType orderType = adaptSide(anxTrade.getTradeType());
BigDecimal amount = anxTrade.getAmount();
BigDecimal price = anxTrade.getPrice();
CurrencyPair currencyPair = adaptCurrencyPair(anxTrade.getItem(), anxTrade.getPriceCurrency());
Date dateTime = DateUtils.fromMillisUtc(anxTrade.getTid());
final String tradeId = String.valueOf(anxTrade.getTid());
return new Trade(orderType, amount, currencyPair, price, dateTime, tradeId);
} | static Trade function(ANXTrade anxTrade) { OrderType orderType = adaptSide(anxTrade.getTradeType()); BigDecimal amount = anxTrade.getAmount(); BigDecimal price = anxTrade.getPrice(); CurrencyPair currencyPair = adaptCurrencyPair(anxTrade.getItem(), anxTrade.getPriceCurrency()); Date dateTime = DateUtils.fromMillisUtc(anxTrade.getTid()); final String tradeId = String.valueOf(anxTrade.getTid()); return new Trade(orderType, amount, currencyPair, price, dateTime, tradeId); } | /**
* Adapts a ANXTrade to a Trade Object
*
* @param anxTrade
* @return
*/ | Adapts a ANXTrade to a Trade Object | adaptTrade | {
"repo_name": "codeck/XChange",
"path": "xchange-anx/src/main/java/com/xeiam/xchange/anx/v2/ANXAdapters.java",
"license": "mit",
"size": 9535
} | [
"com.xeiam.xchange.anx.v2.dto.marketdata.ANXTrade",
"com.xeiam.xchange.currency.CurrencyPair",
"com.xeiam.xchange.dto.Order",
"com.xeiam.xchange.dto.marketdata.Trade",
"com.xeiam.xchange.utils.DateUtils",
"java.math.BigDecimal",
"java.util.Date"
] | import com.xeiam.xchange.anx.v2.dto.marketdata.ANXTrade; import com.xeiam.xchange.currency.CurrencyPair; import com.xeiam.xchange.dto.Order; import com.xeiam.xchange.dto.marketdata.Trade; import com.xeiam.xchange.utils.DateUtils; import java.math.BigDecimal; import java.util.Date; | import com.xeiam.xchange.anx.v2.dto.marketdata.*; import com.xeiam.xchange.currency.*; import com.xeiam.xchange.dto.*; import com.xeiam.xchange.dto.marketdata.*; import com.xeiam.xchange.utils.*; import java.math.*; import java.util.*; | [
"com.xeiam.xchange",
"java.math",
"java.util"
] | com.xeiam.xchange; java.math; java.util; | 2,845,782 |
dbc.precondition(collection != null, "cannot call all with a null collection");
final Function<Iterator<E>, R> consumer = new ConsumeIntoCollection<>(new ConstantSupplier<R>(collection));
return consumer.apply(iterator);
} | dbc.precondition(collection != null, STR); final Function<Iterator<E>, R> consumer = new ConsumeIntoCollection<>(new ConstantSupplier<R>(collection)); return consumer.apply(iterator); } | /**
* Yields all elements of the iterator (in the provided collection).
*
* @param <R> the returned collection type
* @param <E> the collection element type
* @param iterator the iterator that will be consumed
* @param collection the collection where the iterator is consumed
* @return the collection filled with iterator values
*/ | Yields all elements of the iterator (in the provided collection) | all | {
"repo_name": "emaze/emaze-dysfunctional",
"path": "src/main/java/net/emaze/dysfunctional/Consumers.java",
"license": "bsd-3-clause",
"size": 28336
} | [
"java.util.Iterator",
"java.util.function.Function",
"net.emaze.dysfunctional.consumers.ConsumeIntoCollection",
"net.emaze.dysfunctional.dispatching.delegates.ConstantSupplier"
] | import java.util.Iterator; import java.util.function.Function; import net.emaze.dysfunctional.consumers.ConsumeIntoCollection; import net.emaze.dysfunctional.dispatching.delegates.ConstantSupplier; | import java.util.*; import java.util.function.*; import net.emaze.dysfunctional.consumers.*; import net.emaze.dysfunctional.dispatching.delegates.*; | [
"java.util",
"net.emaze.dysfunctional"
] | java.util; net.emaze.dysfunctional; | 77,904 |
public long getCapacity() throws IOException {
return volumes.getCapacity();
} | long function() throws IOException { return volumes.getCapacity(); } | /**
* Return total capacity, used and unused
*/ | Return total capacity, used and unused | getCapacity | {
"repo_name": "ryanobjc/hadoop-cloudera",
"path": "src/hdfs/org/apache/hadoop/hdfs/server/datanode/FSDataset.java",
"license": "apache-2.0",
"size": 63084
} | [
"java.io.IOException"
] | import java.io.IOException; | import java.io.*; | [
"java.io"
] | java.io; | 538,395 |
@Override
public void onCancelResetChange() {
startActivity(new Intent(AuthActivity.this, ViewfinderActivity.class));
} | void function() { startActivity(new Intent(AuthActivity.this, ViewfinderActivity.class)); } | /**
* Called when the user clicks the cancel button. Goes to the inbox.
*/ | Called when the user clicks the cancel button. Goes to the inbox | onCancelResetChange | {
"repo_name": "0359xiaodong/viewfinder",
"path": "clients/android/src/co/viewfinder/AuthActivity.java",
"license": "apache-2.0",
"size": 17072
} | [
"android.content.Intent"
] | import android.content.Intent; | import android.content.*; | [
"android.content"
] | android.content; | 1,941,763 |
protected void childBooleanValue(Production node, Node child)
throws ParseException {
node.addChild(child);
} | void function(Production node, Node child) throws ParseException { node.addChild(child); } | /**
* Called when adding a child to a parse tree node.
*
* @param node the parent node
* @param child the child node, or null
*
* @throws ParseException if the node analysis discovered errors
*/ | Called when adding a child to a parse tree node | childBooleanValue | {
"repo_name": "richb-hanover/mibble-2.9.2",
"path": "src/java/net/percederberg/mibble/asn1/Asn1Analyzer.java",
"license": "gpl-2.0",
"size": 275483
} | [
"net.percederberg.grammatica.parser.Node",
"net.percederberg.grammatica.parser.ParseException",
"net.percederberg.grammatica.parser.Production"
] | import net.percederberg.grammatica.parser.Node; import net.percederberg.grammatica.parser.ParseException; import net.percederberg.grammatica.parser.Production; | import net.percederberg.grammatica.parser.*; | [
"net.percederberg.grammatica"
] | net.percederberg.grammatica; | 447,722 |
public void setTimeZone(TimeZone timeZone) {
this.timeZone = timeZone;
} | void function(TimeZone timeZone) { this.timeZone = timeZone; } | /**
* Sets the time zone for the field.
*
* @param timeZone
* the new time zone to use
* @since 8.2
*/ | Sets the time zone for the field | setTimeZone | {
"repo_name": "Darsstar/framework",
"path": "client/src/main/java/com/vaadin/client/ui/VAbstractTextualDate.java",
"license": "apache-2.0",
"size": 16452
} | [
"com.google.gwt.i18n.client.TimeZone"
] | import com.google.gwt.i18n.client.TimeZone; | import com.google.gwt.i18n.client.*; | [
"com.google.gwt"
] | com.google.gwt; | 2,874,864 |
public static boolean validatePHRExtractTemplateId(PHRExtract phrExtract, DiagnosticChain diagnostics, Map<Object, Object> context) {
if (VALIDATE_PHR_EXTRACT_TEMPLATE_ID__DIAGNOSTIC_CHAIN_MAP__EOCL_INV == null) {
OCL.Helper helper = EOCL_ENV.createOCLHelper();
helper.setContext(IHEPackage.Literals.PHR_EXTRACT);
try {
VALIDATE_PHR_EXTRACT_TEMPLATE_ID__DIAGNOSTIC_CHAIN_MAP__EOCL_INV = helper.createInvariant(VALIDATE_PHR_EXTRACT_TEMPLATE_ID__DIAGNOSTIC_CHAIN_MAP__EOCL_EXP);
}
catch (ParserException pe) {
throw new UnsupportedOperationException(pe.getLocalizedMessage());
}
}
if (!EOCL_ENV.createQuery(VALIDATE_PHR_EXTRACT_TEMPLATE_ID__DIAGNOSTIC_CHAIN_MAP__EOCL_INV).check(phrExtract)) {
if (diagnostics != null) {
diagnostics.add
(new BasicDiagnostic
(Diagnostic.ERROR,
IHEValidator.DIAGNOSTIC_SOURCE,
IHEValidator.PHR_EXTRACT__PHR_EXTRACT_TEMPLATE_ID,
IHEPlugin.INSTANCE.getString("PHRExtractTemplateId"),
new Object [] { phrExtract }));
}
return false;
}
return true;
}
| static boolean function(PHRExtract phrExtract, DiagnosticChain diagnostics, Map<Object, Object> context) { if (VALIDATE_PHR_EXTRACT_TEMPLATE_ID__DIAGNOSTIC_CHAIN_MAP__EOCL_INV == null) { OCL.Helper helper = EOCL_ENV.createOCLHelper(); helper.setContext(IHEPackage.Literals.PHR_EXTRACT); try { VALIDATE_PHR_EXTRACT_TEMPLATE_ID__DIAGNOSTIC_CHAIN_MAP__EOCL_INV = helper.createInvariant(VALIDATE_PHR_EXTRACT_TEMPLATE_ID__DIAGNOSTIC_CHAIN_MAP__EOCL_EXP); } catch (ParserException pe) { throw new UnsupportedOperationException(pe.getLocalizedMessage()); } } if (!EOCL_ENV.createQuery(VALIDATE_PHR_EXTRACT_TEMPLATE_ID__DIAGNOSTIC_CHAIN_MAP__EOCL_INV).check(phrExtract)) { if (diagnostics != null) { diagnostics.add (new BasicDiagnostic (Diagnostic.ERROR, IHEValidator.DIAGNOSTIC_SOURCE, IHEValidator.PHR_EXTRACT__PHR_EXTRACT_TEMPLATE_ID, IHEPlugin.INSTANCE.getString(STR), new Object [] { phrExtract })); } return false; } return true; } | /**
* <!-- begin-user-doc -->
* <!-- end-user-doc -->
* <!-- begin-model-doc -->
* self.templateId->exists(id : datatypes::II | id.root = '1.3.6.1.4.1.19376.1.5.3.1.1.5')
* @param phrExtract The receiving '<em><b>PHR Extract</b></em>' model object.
* @param diagnostics The chain of diagnostics to which problems are to be appended.
* @param context The cache of context-specific information.
* <!-- end-model-doc -->
* @generated
*/ | self.templateId->exists(id : datatypes::II | id.root = '1.3.6.1.4.1.19376.1.5.3.1.1.5') | validatePHRExtractTemplateId | {
"repo_name": "drbgfc/mdht",
"path": "cda/deprecated/org.openhealthtools.mdht.uml.cda.ihe/src/org/openhealthtools/mdht/uml/cda/ihe/operations/PHRExtractOperations.java",
"license": "epl-1.0",
"size": 4053
} | [
"java.util.Map",
"org.eclipse.emf.common.util.BasicDiagnostic",
"org.eclipse.emf.common.util.Diagnostic",
"org.eclipse.emf.common.util.DiagnosticChain",
"org.eclipse.ocl.ParserException",
"org.openhealthtools.mdht.uml.cda.ihe.IHEPackage",
"org.openhealthtools.mdht.uml.cda.ihe.IHEPlugin",
"org.openhealthtools.mdht.uml.cda.ihe.PHRExtract",
"org.openhealthtools.mdht.uml.cda.ihe.util.IHEValidator"
] | import java.util.Map; import org.eclipse.emf.common.util.BasicDiagnostic; import org.eclipse.emf.common.util.Diagnostic; import org.eclipse.emf.common.util.DiagnosticChain; import org.eclipse.ocl.ParserException; import org.openhealthtools.mdht.uml.cda.ihe.IHEPackage; import org.openhealthtools.mdht.uml.cda.ihe.IHEPlugin; import org.openhealthtools.mdht.uml.cda.ihe.PHRExtract; import org.openhealthtools.mdht.uml.cda.ihe.util.IHEValidator; | import java.util.*; import org.eclipse.emf.common.util.*; import org.eclipse.ocl.*; import org.openhealthtools.mdht.uml.cda.ihe.*; import org.openhealthtools.mdht.uml.cda.ihe.util.*; | [
"java.util",
"org.eclipse.emf",
"org.eclipse.ocl",
"org.openhealthtools.mdht"
] | java.util; org.eclipse.emf; org.eclipse.ocl; org.openhealthtools.mdht; | 2,603,975 |
public Map<Long, Map<Long, List<TreeImageDisplay>>> getData()
{
return data;
}
| Map<Long, Map<Long, List<TreeImageDisplay>>> function() { return data; } | /**
* Returns the data.
*
* @return See above.
*/ | Returns the data | getData | {
"repo_name": "knabar/openmicroscopy",
"path": "components/insight/SRC/org/openmicroscopy/shoola/agents/events/treeviewer/ExperimenterLoadedDataEvent.java",
"license": "gpl-2.0",
"size": 2149
} | [
"java.util.List",
"java.util.Map",
"org.openmicroscopy.shoola.agents.util.browser.TreeImageDisplay"
] | import java.util.List; import java.util.Map; import org.openmicroscopy.shoola.agents.util.browser.TreeImageDisplay; | import java.util.*; import org.openmicroscopy.shoola.agents.util.browser.*; | [
"java.util",
"org.openmicroscopy.shoola"
] | java.util; org.openmicroscopy.shoola; | 2,726,016 |
public ArrayList<ArrayList<String>> getEcotypes() {
return ecotypes;
} | ArrayList<ArrayList<String>> function() { return ecotypes; } | /**
* Get the ecotypes.
*
* @return The ecotypes.
*/ | Get the ecotypes | getEcotypes | {
"repo_name": "weiway/New-Ecotype-Simulation",
"path": "src/java/Demarcation.java",
"license": "gpl-3.0",
"size": 9219
} | [
"java.util.ArrayList"
] | import java.util.ArrayList; | import java.util.*; | [
"java.util"
] | java.util; | 2,701,544 |
public ServiceCall<Void> addPetAsync(final ServiceCallback<Void> serviceCallback) {
return ServiceCall.create(addPetWithServiceResponseAsync(), serviceCallback);
} | ServiceCall<Void> function(final ServiceCallback<Void> serviceCallback) { return ServiceCall.create(addPetWithServiceResponseAsync(), serviceCallback); } | /**
* Add a new pet to the store.
* Adds a new pet to the store. You may receive an HTTP invalid input if your pet is invalid.
*
* @param serviceCallback the async ServiceCallback to handle successful and failed responses.
* @return the {@link ServiceCall} object
*/ | Add a new pet to the store. Adds a new pet to the store. You may receive an HTTP invalid input if your pet is invalid | addPetAsync | {
"repo_name": "yugangw-msft/autorest",
"path": "Samples/petstore/Java/implementation/SwaggerPetstoreImpl.java",
"license": "mit",
"size": 99949
} | [
"com.microsoft.rest.ServiceCall",
"com.microsoft.rest.ServiceCallback"
] | import com.microsoft.rest.ServiceCall; import com.microsoft.rest.ServiceCallback; | import com.microsoft.rest.*; | [
"com.microsoft.rest"
] | com.microsoft.rest; | 2,094,067 |
protected synchronized void saveNobleData()
{
if ((_nobles == null) || _nobles.isEmpty())
{
return;
}
try (Connection con = DatabaseFactory.getInstance().getConnection())
{
for (Entry<Integer, StatsSet> entry : _nobles.entrySet())
{
final StatsSet nobleInfo = entry.getValue();
if (nobleInfo == null)
{
continue;
}
final int charId = entry.getKey();
final int classId = nobleInfo.getInt(CLASS_ID);
final int points = nobleInfo.getInt(POINTS);
final int compDone = nobleInfo.getInt(COMP_DONE);
final int compWon = nobleInfo.getInt(COMP_WON);
final int compLost = nobleInfo.getInt(COMP_LOST);
final int compDrawn = nobleInfo.getInt(COMP_DRAWN);
final int compDoneWeek = nobleInfo.getInt(COMP_DONE_WEEK);
final int compDoneWeekClassed = nobleInfo.getInt(COMP_DONE_WEEK_CLASSED);
final int compDoneWeekNonClassed = nobleInfo.getInt(COMP_DONE_WEEK_NON_CLASSED);
final int compDoneWeekTeam = nobleInfo.getInt(COMP_DONE_WEEK_TEAM);
final boolean toSave = nobleInfo.getBoolean("to_save");
try (PreparedStatement statement = con.prepareStatement(toSave ? OLYMPIAD_SAVE_NOBLES : OLYMPIAD_UPDATE_NOBLES))
{
if (toSave)
{
statement.setInt(1, charId);
statement.setInt(2, classId);
statement.setInt(3, points);
statement.setInt(4, compDone);
statement.setInt(5, compWon);
statement.setInt(6, compLost);
statement.setInt(7, compDrawn);
statement.setInt(8, compDoneWeek);
statement.setInt(9, compDoneWeekClassed);
statement.setInt(10, compDoneWeekNonClassed);
statement.setInt(11, compDoneWeekTeam);
nobleInfo.set("to_save", false);
}
else
{
statement.setInt(1, points);
statement.setInt(2, compDone);
statement.setInt(3, compWon);
statement.setInt(4, compLost);
statement.setInt(5, compDrawn);
statement.setInt(6, compDoneWeek);
statement.setInt(7, compDoneWeekClassed);
statement.setInt(8, compDoneWeekNonClassed);
statement.setInt(9, compDoneWeekTeam);
statement.setInt(10, charId);
}
statement.execute();
}
}
}
catch (SQLException e)
{
LOGGER.log(Level.SEVERE, getClass().getSimpleName() + ": Failed to save noblesse data to database: ", e);
}
}
| synchronized void function() { if ((_nobles == null) _nobles.isEmpty()) { return; } try (Connection con = DatabaseFactory.getInstance().getConnection()) { for (Entry<Integer, StatsSet> entry : _nobles.entrySet()) { final StatsSet nobleInfo = entry.getValue(); if (nobleInfo == null) { continue; } final int charId = entry.getKey(); final int classId = nobleInfo.getInt(CLASS_ID); final int points = nobleInfo.getInt(POINTS); final int compDone = nobleInfo.getInt(COMP_DONE); final int compWon = nobleInfo.getInt(COMP_WON); final int compLost = nobleInfo.getInt(COMP_LOST); final int compDrawn = nobleInfo.getInt(COMP_DRAWN); final int compDoneWeek = nobleInfo.getInt(COMP_DONE_WEEK); final int compDoneWeekClassed = nobleInfo.getInt(COMP_DONE_WEEK_CLASSED); final int compDoneWeekNonClassed = nobleInfo.getInt(COMP_DONE_WEEK_NON_CLASSED); final int compDoneWeekTeam = nobleInfo.getInt(COMP_DONE_WEEK_TEAM); final boolean toSave = nobleInfo.getBoolean(STR); try (PreparedStatement statement = con.prepareStatement(toSave ? OLYMPIAD_SAVE_NOBLES : OLYMPIAD_UPDATE_NOBLES)) { if (toSave) { statement.setInt(1, charId); statement.setInt(2, classId); statement.setInt(3, points); statement.setInt(4, compDone); statement.setInt(5, compWon); statement.setInt(6, compLost); statement.setInt(7, compDrawn); statement.setInt(8, compDoneWeek); statement.setInt(9, compDoneWeekClassed); statement.setInt(10, compDoneWeekNonClassed); statement.setInt(11, compDoneWeekTeam); nobleInfo.set(STR, false); } else { statement.setInt(1, points); statement.setInt(2, compDone); statement.setInt(3, compWon); statement.setInt(4, compLost); statement.setInt(5, compDrawn); statement.setInt(6, compDoneWeek); statement.setInt(7, compDoneWeekClassed); statement.setInt(8, compDoneWeekNonClassed); statement.setInt(9, compDoneWeekTeam); statement.setInt(10, charId); } statement.execute(); } } } catch (SQLException e) { LOGGER.log(Level.SEVERE, getClass().getSimpleName() + STR, e); } } | /**
* Save noblesse data to database
*/ | Save noblesse data to database | saveNobleData | {
"repo_name": "rubenswagner/L2J-Global",
"path": "java/com/l2jglobal/gameserver/model/olympiad/Olympiad.java",
"license": "gpl-3.0",
"size": 38507
} | [
"com.l2jglobal.commons.database.DatabaseFactory",
"com.l2jglobal.gameserver.model.StatsSet",
"java.sql.Connection",
"java.sql.PreparedStatement",
"java.sql.SQLException",
"java.util.Map",
"java.util.logging.Level"
] | import com.l2jglobal.commons.database.DatabaseFactory; import com.l2jglobal.gameserver.model.StatsSet; import java.sql.Connection; import java.sql.PreparedStatement; import java.sql.SQLException; import java.util.Map; import java.util.logging.Level; | import com.l2jglobal.commons.database.*; import com.l2jglobal.gameserver.model.*; import java.sql.*; import java.util.*; import java.util.logging.*; | [
"com.l2jglobal.commons",
"com.l2jglobal.gameserver",
"java.sql",
"java.util"
] | com.l2jglobal.commons; com.l2jglobal.gameserver; java.sql; java.util; | 213,416 |
L3VpnIfs l3VpnIfs(); | L3VpnIfs l3VpnIfs(); | /**
* Returns the attribute l3VpnIfs.
*
* @return value of l3VpnIfs
*/ | Returns the attribute l3VpnIfs | l3VpnIfs | {
"repo_name": "mengmoya/onos",
"path": "apps/l3vpn/nel3vpn/nemgr/src/main/java/org/onosproject/yang/gen/v1/ne/l3vpn/api/rev20141225/nel3vpnapi/l3vpninstances/L3VpnInstance.java",
"license": "apache-2.0",
"size": 3370
} | [
"org.onosproject.yang.gen.v1.ne.l3vpn.comm.rev20141225.nel3vpncomm.l3vpnifs.L3VpnIfs"
] | import org.onosproject.yang.gen.v1.ne.l3vpn.comm.rev20141225.nel3vpncomm.l3vpnifs.L3VpnIfs; | import org.onosproject.yang.gen.v1.ne.l3vpn.comm.rev20141225.nel3vpncomm.l3vpnifs.*; | [
"org.onosproject.yang"
] | org.onosproject.yang; | 2,586,848 |
@Test
public void testNodeNumbers() throws CloneNotSupportedException {
Individual ind = Individual.fromString("(+ x (* x (- (% x x) 1)))",
config);
config.setMutationProbability(1.0);
Individual mutant = PointMutation.mutate(ind, context);
assertEquals(ind.getRoot().subtreeToNumberedString(), mutant.getRoot()
.subtreeToNumberedString());
} | void function() throws CloneNotSupportedException { Individual ind = Individual.fromString(STR, config); config.setMutationProbability(1.0); Individual mutant = PointMutation.mutate(ind, context); assertEquals(ind.getRoot().subtreeToNumberedString(), mutant.getRoot() .subtreeToNumberedString()); } | /**
* Makes sure that all nodes have their numbers set correctly after
* mutation.
*
* @throws CloneNotSupportedException
*/ | Makes sure that all nodes have their numbers set correctly after mutation | testNodeNumbers | {
"repo_name": "burks-pub/gecco2015",
"path": "src/test/java/ec/research/gp/simple/operators/PointMutationTest.java",
"license": "bsd-2-clause",
"size": 4072
} | [
"ec.research.gp.simple.operators.PointMutation",
"ec.research.gp.simple.representation.Individual",
"org.junit.Assert"
] | import ec.research.gp.simple.operators.PointMutation; import ec.research.gp.simple.representation.Individual; import org.junit.Assert; | import ec.research.gp.simple.operators.*; import ec.research.gp.simple.representation.*; import org.junit.*; | [
"ec.research.gp",
"org.junit"
] | ec.research.gp; org.junit; | 1,704,685 |
public void setMembers(SortedMap<Tenor, ExternalId> members) {
JodaBeanUtils.notNull(members, "members");
this._members.clear();
this._members.putAll(members);
} | void function(SortedMap<Tenor, ExternalId> members) { JodaBeanUtils.notNull(members, STR); this._members.clear(); this._members.putAll(members); } | /**
* Sets the members.
* @param members the new value of the property, not null
*/ | Sets the members | setMembers | {
"repo_name": "jeorme/OG-Platform",
"path": "projects/OG-FinancialTypes/src/main/java/com/opengamma/financial/security/index/IndexFamily.java",
"license": "apache-2.0",
"size": 6771
} | [
"com.opengamma.id.ExternalId",
"com.opengamma.util.time.Tenor",
"java.util.SortedMap",
"org.joda.beans.JodaBeanUtils"
] | import com.opengamma.id.ExternalId; import com.opengamma.util.time.Tenor; import java.util.SortedMap; import org.joda.beans.JodaBeanUtils; | import com.opengamma.id.*; import com.opengamma.util.time.*; import java.util.*; import org.joda.beans.*; | [
"com.opengamma.id",
"com.opengamma.util",
"java.util",
"org.joda.beans"
] | com.opengamma.id; com.opengamma.util; java.util; org.joda.beans; | 2,796,762 |
public static byte[] sha256hash160(byte[] input) {
try {
byte[] sha256 = MessageDigest.getInstance("SHA-256").digest(input);
RIPEMD160Digest digest = new RIPEMD160Digest();
digest.update(sha256, 0, sha256.length);
byte[] out = new byte[20];
digest.doFinal(out, 0);
return out;
} catch (NoSuchAlgorithmException e) {
throw new RuntimeException(e); // Cannot happen.
}
} | static byte[] function(byte[] input) { try { byte[] sha256 = MessageDigest.getInstance(STR).digest(input); RIPEMD160Digest digest = new RIPEMD160Digest(); digest.update(sha256, 0, sha256.length); byte[] out = new byte[20]; digest.doFinal(out, 0); return out; } catch (NoSuchAlgorithmException e) { throw new RuntimeException(e); } } | /**
* Calculates RIPEMD160(SHA256(input)). This is used in Address calculations.
*/ | Calculates RIPEMD160(SHA256(input)). This is used in Address calculations | sha256hash160 | {
"repo_name": "pavel4n/wowdoge.org",
"path": "core/src/main/java/com/google/dogecoin/core/Utils.java",
"license": "apache-2.0",
"size": 23380
} | [
"java.security.MessageDigest",
"java.security.NoSuchAlgorithmException",
"org.spongycastle.crypto.digests.RIPEMD160Digest"
] | import java.security.MessageDigest; import java.security.NoSuchAlgorithmException; import org.spongycastle.crypto.digests.RIPEMD160Digest; | import java.security.*; import org.spongycastle.crypto.digests.*; | [
"java.security",
"org.spongycastle.crypto"
] | java.security; org.spongycastle.crypto; | 1,160,871 |
Optional<String> getMachineReadableLog(); | Optional<String> getMachineReadableLog(); | /**
* For more information how MachineReadableLog works check {@link
* com.facebook.buck.event.listener.MachineReadableLoggerListener}.
*
* @return the contents of the machine readable logs. Each line marks a different key to json.
*/ | For more information how MachineReadableLog works check <code>com.facebook.buck.event.listener.MachineReadableLoggerListener</code> | getMachineReadableLog | {
"repo_name": "zpao/buck",
"path": "src/com/facebook/buck/doctor/config/DoctorEndpointRequest.java",
"license": "apache-2.0",
"size": 1502
} | [
"java.util.Optional"
] | import java.util.Optional; | import java.util.*; | [
"java.util"
] | java.util; | 1,711,427 |
@ServiceMethod(returns = ReturnType.COLLECTION)
PagedFlux<MetricNamespaceInner> listAsync(String resourceUri, String startTime); | @ServiceMethod(returns = ReturnType.COLLECTION) PagedFlux<MetricNamespaceInner> listAsync(String resourceUri, String startTime); | /**
* Lists the metric namespaces for the resource.
*
* @param resourceUri The identifier of the resource.
* @param startTime The ISO 8601 conform Date start time from which to query for metric namespaces.
* @throws IllegalArgumentException thrown if parameters fail the validation.
* @throws com.azure.core.management.exception.ManagementException thrown if the request is rejected by server.
* @throws RuntimeException all other wrapped checked exceptions if the request fails to be sent.
* @return represents collection of metric namespaces.
*/ | Lists the metric namespaces for the resource | listAsync | {
"repo_name": "Azure/azure-sdk-for-java",
"path": "sdk/resourcemanagerhybrid/azure-resourcemanager-monitor/src/main/java/com/azure/resourcemanager/monitor/fluent/MetricNamespacesClient.java",
"license": "mit",
"size": 3433
} | [
"com.azure.core.annotation.ReturnType",
"com.azure.core.annotation.ServiceMethod",
"com.azure.core.http.rest.PagedFlux",
"com.azure.resourcemanager.monitor.fluent.models.MetricNamespaceInner"
] | import com.azure.core.annotation.ReturnType; import com.azure.core.annotation.ServiceMethod; import com.azure.core.http.rest.PagedFlux; import com.azure.resourcemanager.monitor.fluent.models.MetricNamespaceInner; | import com.azure.core.annotation.*; import com.azure.core.http.rest.*; import com.azure.resourcemanager.monitor.fluent.models.*; | [
"com.azure.core",
"com.azure.resourcemanager"
] | com.azure.core; com.azure.resourcemanager; | 1,078,027 |
public static <T> T resolveReferenceParameter(CamelContext context, String value, Class<T> type) {
return resolveReferenceParameter(context, value, type, true);
} | static <T> T function(CamelContext context, String value, Class<T> type) { return resolveReferenceParameter(context, value, type, true); } | /**
* Resolves a reference parameter by making a lookup in the registry.
*
* @param <T> type of object to lookup.
* @param context Camel context to use for lookup.
* @param value reference parameter value.
* @param type type of object to lookup.
* @return lookup result.
* @throws IllegalArgumentException if referenced object was not found in registry.
*/ | Resolves a reference parameter by making a lookup in the registry | resolveReferenceParameter | {
"repo_name": "punkhorn/camel-upstream",
"path": "core/camel-support/src/main/java/org/apache/camel/support/EndpointHelper.java",
"license": "apache-2.0",
"size": 16622
} | [
"org.apache.camel.CamelContext"
] | import org.apache.camel.CamelContext; | import org.apache.camel.*; | [
"org.apache.camel"
] | org.apache.camel; | 2,059,823 |
public SubscriptionPolicy[] getSubscriptionPolicies(int tenantID) throws APIManagementException {
List<SubscriptionPolicy> policies = new ArrayList<SubscriptionPolicy>();
Connection conn = null;
PreparedStatement ps = null;
ResultSet rs = null;
String sqlQuery = SQLConstants.GET_SUBSCRIPTION_POLICIES;
if (forceCaseInsensitiveComparisons) {
sqlQuery = SQLConstants.GET_SUBSCRIPTION_POLICIES;
}
try {
conn = APIMgtDBUtil.getConnection();
ps = conn.prepareStatement(sqlQuery);
ps.setInt(1, tenantID);
rs = ps.executeQuery();
while (rs.next()) {
SubscriptionPolicy subPolicy = new SubscriptionPolicy(
rs.getString(ThrottlePolicyConstants.COLUMN_NAME));
setCommonPolicyDetails(subPolicy, rs);
subPolicy.setRateLimitCount(rs.getInt(ThrottlePolicyConstants.COLUMN_RATE_LIMIT_COUNT));
subPolicy.setRateLimitTimeUnit(rs.getString(ThrottlePolicyConstants.COLUMN_RATE_LIMIT_TIME_UNIT));
subPolicy.setStopOnQuotaReach(rs.getBoolean(ThrottlePolicyConstants.COLUMN_STOP_ON_QUOTA_REACH));
subPolicy.setBillingPlan(rs.getString(ThrottlePolicyConstants.COLUMN_BILLING_PLAN));
InputStream binary = rs.getBinaryStream(ThrottlePolicyConstants.COLUMN_CUSTOM_ATTRIB);
if(binary != null){
byte[] customAttrib = APIUtil.toByteArray(binary);
subPolicy.setCustomAttributes(customAttrib);
}
policies.add(subPolicy);
}
} catch (SQLException e) {
handleException("Error while executing SQL", e);
} catch (IOException e) {
handleException("Error while converting input stream to byte array", e);
} finally {
APIMgtDBUtil.closeAllConnections(ps, conn, rs);
}
return policies.toArray(new SubscriptionPolicy[policies.size()]);
} | SubscriptionPolicy[] function(int tenantID) throws APIManagementException { List<SubscriptionPolicy> policies = new ArrayList<SubscriptionPolicy>(); Connection conn = null; PreparedStatement ps = null; ResultSet rs = null; String sqlQuery = SQLConstants.GET_SUBSCRIPTION_POLICIES; if (forceCaseInsensitiveComparisons) { sqlQuery = SQLConstants.GET_SUBSCRIPTION_POLICIES; } try { conn = APIMgtDBUtil.getConnection(); ps = conn.prepareStatement(sqlQuery); ps.setInt(1, tenantID); rs = ps.executeQuery(); while (rs.next()) { SubscriptionPolicy subPolicy = new SubscriptionPolicy( rs.getString(ThrottlePolicyConstants.COLUMN_NAME)); setCommonPolicyDetails(subPolicy, rs); subPolicy.setRateLimitCount(rs.getInt(ThrottlePolicyConstants.COLUMN_RATE_LIMIT_COUNT)); subPolicy.setRateLimitTimeUnit(rs.getString(ThrottlePolicyConstants.COLUMN_RATE_LIMIT_TIME_UNIT)); subPolicy.setStopOnQuotaReach(rs.getBoolean(ThrottlePolicyConstants.COLUMN_STOP_ON_QUOTA_REACH)); subPolicy.setBillingPlan(rs.getString(ThrottlePolicyConstants.COLUMN_BILLING_PLAN)); InputStream binary = rs.getBinaryStream(ThrottlePolicyConstants.COLUMN_CUSTOM_ATTRIB); if(binary != null){ byte[] customAttrib = APIUtil.toByteArray(binary); subPolicy.setCustomAttributes(customAttrib); } policies.add(subPolicy); } } catch (SQLException e) { handleException(STR, e); } catch (IOException e) { handleException(STR, e); } finally { APIMgtDBUtil.closeAllConnections(ps, conn, rs); } return policies.toArray(new SubscriptionPolicy[policies.size()]); } | /**
* Get all subscription level policeis belongs to specific tenant
*
* @param tenantID tenantID filters the polices belongs to specific tenant
* @return subscriptionPolicy array list
*/ | Get all subscription level policeis belongs to specific tenant | getSubscriptionPolicies | {
"repo_name": "knPerera/carbon-apimgt",
"path": "components/apimgt/org.wso2.carbon.apimgt.impl/src/main/java/org/wso2/carbon/apimgt/impl/dao/ApiMgtDAO.java",
"license": "apache-2.0",
"size": 493075
} | [
"java.io.IOException",
"java.io.InputStream",
"java.sql.Connection",
"java.sql.PreparedStatement",
"java.sql.ResultSet",
"java.sql.SQLException",
"java.util.ArrayList",
"java.util.List",
"org.wso2.carbon.apimgt.api.APIManagementException",
"org.wso2.carbon.apimgt.api.model.policy.SubscriptionPolicy",
"org.wso2.carbon.apimgt.impl.ThrottlePolicyConstants",
"org.wso2.carbon.apimgt.impl.dao.constants.SQLConstants",
"org.wso2.carbon.apimgt.impl.utils.APIMgtDBUtil",
"org.wso2.carbon.apimgt.impl.utils.APIUtil"
] | import java.io.IOException; import java.io.InputStream; import java.sql.Connection; import java.sql.PreparedStatement; import java.sql.ResultSet; import java.sql.SQLException; import java.util.ArrayList; import java.util.List; import org.wso2.carbon.apimgt.api.APIManagementException; import org.wso2.carbon.apimgt.api.model.policy.SubscriptionPolicy; import org.wso2.carbon.apimgt.impl.ThrottlePolicyConstants; import org.wso2.carbon.apimgt.impl.dao.constants.SQLConstants; import org.wso2.carbon.apimgt.impl.utils.APIMgtDBUtil; import org.wso2.carbon.apimgt.impl.utils.APIUtil; | import java.io.*; import java.sql.*; import java.util.*; import org.wso2.carbon.apimgt.api.*; import org.wso2.carbon.apimgt.api.model.policy.*; import org.wso2.carbon.apimgt.impl.*; import org.wso2.carbon.apimgt.impl.dao.constants.*; import org.wso2.carbon.apimgt.impl.utils.*; | [
"java.io",
"java.sql",
"java.util",
"org.wso2.carbon"
] | java.io; java.sql; java.util; org.wso2.carbon; | 1,052,160 |
ClassLoader getClassLoader(String componentName, String componentVersion, ComponentReference componentReference); | ClassLoader getClassLoader(String componentName, String componentVersion, ComponentReference componentReference); | /**
* Returns class loader of the given component.
*
* @param componentName full component name
* @param componentVersion component version
* @param componentReference component reference
* @return class loader for specified component
*/ | Returns class loader of the given component | getClassLoader | {
"repo_name": "rasika90/carbon-uuf",
"path": "components/uuf-core/src/main/java/org/wso2/carbon/uuf/internal/core/create/ClassLoaderProvider.java",
"license": "apache-2.0",
"size": 1200
} | [
"org.wso2.carbon.uuf.reference.ComponentReference"
] | import org.wso2.carbon.uuf.reference.ComponentReference; | import org.wso2.carbon.uuf.reference.*; | [
"org.wso2.carbon"
] | org.wso2.carbon; | 1,870,300 |
static double[] yearFractions(List<Period> periods, ZonedDateTime valuationTime) {
double[] yearFractions = new double[periods.size()];
int index = 0;
for (Period period : periods) {
yearFractions[index++] = TimeCalculator.getTimeBetween(valuationTime, valuationTime.plus(period));
}
return yearFractions;
} | static double[] yearFractions(List<Period> periods, ZonedDateTime valuationTime) { double[] yearFractions = new double[periods.size()]; int index = 0; for (Period period : periods) { yearFractions[index++] = TimeCalculator.getTimeBetween(valuationTime, valuationTime.plus(period)); } return yearFractions; } | /**
* Converts the input values into an array of year fractions for passing to the analytics.
* If periods is a double array it is returned unchanged. If it's a list of periods then the year fraction
* is calculated for each period relative to the valuation time (using {@link TimeCalculator}).
* @param periods The input axis values, a list of periods or a double array
* @param valuationTime The valuation time
* @return An array of year fractions for the axis values
*/ | Converts the input values into an array of year fractions for passing to the analytics. If periods is a double array it is returned unchanged. If it's a list of periods then the year fraction is calculated for each period relative to the valuation time (using <code>TimeCalculator</code>) | yearFractions | {
"repo_name": "DevStreet/FinanceAnalytics",
"path": "projects/OG-Integration/src/main/java/com/opengamma/integration/marketdata/manipulator/dsl/VolatilitySurfaceShiftManipulator.java",
"license": "apache-2.0",
"size": 19859
} | [
"com.opengamma.analytics.util.time.TimeCalculator",
"java.util.List",
"org.threeten.bp.Period",
"org.threeten.bp.ZonedDateTime"
] | import com.opengamma.analytics.util.time.TimeCalculator; import java.util.List; import org.threeten.bp.Period; import org.threeten.bp.ZonedDateTime; | import com.opengamma.analytics.util.time.*; import java.util.*; import org.threeten.bp.*; | [
"com.opengamma.analytics",
"java.util",
"org.threeten.bp"
] | com.opengamma.analytics; java.util; org.threeten.bp; | 952,722 |
@Test
public void testPersistentPRWithGatewaySenderPersistenceEnabled_Restart2() {
// create locator on local site
Integer lnPort = (Integer) vm0.invoke(() -> WANTestBase.createFirstLocatorWithDSId(1));
// create locator on remote site
Integer nyPort = (Integer) vm1.invoke(() -> WANTestBase.createFirstRemoteLocator(2, lnPort));
// create cache in local site
createCacheInVMs(lnPort, vm4, vm5, vm6, vm7);
// create senders with disk store
String diskStore1 = (String) vm4.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2,
true, 100, 10, false, true, null, null, false));
String diskStore2 = (String) vm5.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2,
true, 100, 10, false, true, null, null, false));
String diskStore3 = (String) vm6.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2,
true, 100, 10, false, true, null, null, false));
String diskStore4 = (String) vm7.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2,
true, 100, 10, false, true, null, null, false));
LogWriterUtils.getLogWriter()
.info("The DS are: " + diskStore1 + "," + diskStore2 + "," + diskStore3 + "," + diskStore4);
// create PR on local site
vm4.invoke(() -> WANTestBase.createPersistentPartitionedRegion(getTestMethodName(), "ln", 1,
100, isOffHeap()));
vm5.invoke(() -> WANTestBase.createPersistentPartitionedRegion(getTestMethodName(), "ln", 1,
100, isOffHeap()));
vm6.invoke(() -> WANTestBase.createPersistentPartitionedRegion(getTestMethodName(), "ln", 1,
100, isOffHeap()));
vm7.invoke(() -> WANTestBase.createPersistentPartitionedRegion(getTestMethodName(), "ln", 1,
100, isOffHeap()));
// start the senders on local site
startSenderInVMs("ln", vm4, vm5, vm6, vm7);
// wait for senders to become running
vm4.invoke(waitForSenderRunnable());
vm5.invoke(waitForSenderRunnable());
vm6.invoke(waitForSenderRunnable());
vm7.invoke(waitForSenderRunnable());
// pause the senders
vm4.invoke(pauseSenderRunnable());
vm5.invoke(pauseSenderRunnable());
vm6.invoke(pauseSenderRunnable());
vm7.invoke(pauseSenderRunnable());
// start puts in region on local site
vm4.invoke(() -> WANTestBase.doPuts(getTestMethodName(), 300));
LogWriterUtils.getLogWriter().info("Completed puts in the region");
// --------------------close and rebuild local site
// -------------------------------------------------
// kill the senders
vm4.invoke(killSenderRunnable());
vm5.invoke(killSenderRunnable());
vm6.invoke(killSenderRunnable());
vm7.invoke(killSenderRunnable());
LogWriterUtils.getLogWriter().info("Killed all the senders.");
// restart the vm
createCacheInVMs(lnPort, vm4, vm5, vm6, vm7);
LogWriterUtils.getLogWriter().info("Created back the cache");
// create senders with disk store
vm4.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2, true, 100, 10, false, true,
null, diskStore1, true));
vm5.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2, true, 100, 10, false, true,
null, diskStore2, true));
vm6.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2, true, 100, 10, false, true,
null, diskStore3, true));
vm7.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2, true, 100, 10, false, true,
null, diskStore4, true));
LogWriterUtils.getLogWriter().info("Created the senders back from the disk store.");
// create PR on local site
AsyncInvocation inv1 = vm4.invokeAsync(() -> WANTestBase
.createPersistentPartitionedRegion(getTestMethodName(), "ln", 1, 100, isOffHeap()));
AsyncInvocation inv2 = vm5.invokeAsync(() -> WANTestBase
.createPersistentPartitionedRegion(getTestMethodName(), "ln", 1, 100, isOffHeap()));
AsyncInvocation inv3 = vm6.invokeAsync(() -> WANTestBase
.createPersistentPartitionedRegion(getTestMethodName(), "ln", 1, 100, isOffHeap()));
AsyncInvocation inv4 = vm7.invokeAsync(() -> WANTestBase
.createPersistentPartitionedRegion(getTestMethodName(), "ln", 1, 100, isOffHeap()));
try {
inv1.join();
inv2.join();
inv3.join();
inv4.join();
} catch (InterruptedException e) {
e.printStackTrace();
fail();
}
LogWriterUtils.getLogWriter().info("Created back the partitioned regions");
// start the senders in async mode. This will ensure that the
// node of shadow PR that went down last will come up first
startSenderInVMsAsync("ln", vm4, vm5, vm6, vm7);
LogWriterUtils.getLogWriter().info("Waiting for senders running.");
// wait for senders running
vm4.invoke(waitForSenderRunnable());
vm5.invoke(waitForSenderRunnable());
vm6.invoke(waitForSenderRunnable());
vm7.invoke(waitForSenderRunnable());
LogWriterUtils.getLogWriter().info("Creating the receiver.");
// create receiver on remote site
createCacheInVMs(nyPort, vm2, vm3);
// create PR on remote site
LogWriterUtils.getLogWriter().info("Creating the partitioned region at receiver. ");
vm2.invoke(() -> WANTestBase.createPersistentPartitionedRegion(getTestMethodName(), null, 1,
100, isOffHeap()));
vm3.invoke(() -> WANTestBase.createPersistentPartitionedRegion(getTestMethodName(), null, 1,
100, isOffHeap()));
createReceiverInVMs(vm2, vm3);
vm4.invoke(pauseSenderRunnable());
vm5.invoke(pauseSenderRunnable());
vm6.invoke(pauseSenderRunnable());
vm7.invoke(pauseSenderRunnable());
LogWriterUtils.getLogWriter().info("Doing some extra puts. ");
// start puts in region on local site
vm4.invoke(() -> WANTestBase.doPutsAfter300(getTestMethodName(), 1000));
// ----------------------------------------------------------------------------------------------------
vm4.invoke(() -> WANTestBase.resumeSender("ln"));
vm5.invoke(() -> WANTestBase.resumeSender("ln"));
vm6.invoke(() -> WANTestBase.resumeSender("ln"));
vm7.invoke(() -> WANTestBase.resumeSender("ln"));
LogWriterUtils.getLogWriter().info("Validating the region size at the receiver end. ");
vm2.invoke(() -> WANTestBase.validateRegionSize(getTestMethodName(), 1000));
vm3.invoke(() -> WANTestBase.validateRegionSize(getTestMethodName(), 1000));
} | void function() { Integer lnPort = (Integer) vm0.invoke(() -> WANTestBase.createFirstLocatorWithDSId(1)); Integer nyPort = (Integer) vm1.invoke(() -> WANTestBase.createFirstRemoteLocator(2, lnPort)); createCacheInVMs(lnPort, vm4, vm5, vm6, vm7); String diskStore1 = (String) vm4.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2, true, 100, 10, false, true, null, null, false)); String diskStore2 = (String) vm5.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2, true, 100, 10, false, true, null, null, false)); String diskStore3 = (String) vm6.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2, true, 100, 10, false, true, null, null, false)); String diskStore4 = (String) vm7.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2, true, 100, 10, false, true, null, null, false)); LogWriterUtils.getLogWriter() .info(STR + diskStore1 + "," + diskStore2 + "," + diskStore3 + "," + diskStore4); vm4.invoke(() -> WANTestBase.createPersistentPartitionedRegion(getTestMethodName(), "ln", 1, 100, isOffHeap())); vm5.invoke(() -> WANTestBase.createPersistentPartitionedRegion(getTestMethodName(), "ln", 1, 100, isOffHeap())); vm6.invoke(() -> WANTestBase.createPersistentPartitionedRegion(getTestMethodName(), "ln", 1, 100, isOffHeap())); vm7.invoke(() -> WANTestBase.createPersistentPartitionedRegion(getTestMethodName(), "ln", 1, 100, isOffHeap())); startSenderInVMs("ln", vm4, vm5, vm6, vm7); vm4.invoke(waitForSenderRunnable()); vm5.invoke(waitForSenderRunnable()); vm6.invoke(waitForSenderRunnable()); vm7.invoke(waitForSenderRunnable()); vm4.invoke(pauseSenderRunnable()); vm5.invoke(pauseSenderRunnable()); vm6.invoke(pauseSenderRunnable()); vm7.invoke(pauseSenderRunnable()); vm4.invoke(() -> WANTestBase.doPuts(getTestMethodName(), 300)); LogWriterUtils.getLogWriter().info(STR); vm4.invoke(killSenderRunnable()); vm5.invoke(killSenderRunnable()); vm6.invoke(killSenderRunnable()); vm7.invoke(killSenderRunnable()); LogWriterUtils.getLogWriter().info(STR); createCacheInVMs(lnPort, vm4, vm5, vm6, vm7); LogWriterUtils.getLogWriter().info(STR); vm4.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2, true, 100, 10, false, true, null, diskStore1, true)); vm5.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2, true, 100, 10, false, true, null, diskStore2, true)); vm6.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2, true, 100, 10, false, true, null, diskStore3, true)); vm7.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2, true, 100, 10, false, true, null, diskStore4, true)); LogWriterUtils.getLogWriter().info(STR); AsyncInvocation inv1 = vm4.invokeAsync(() -> WANTestBase .createPersistentPartitionedRegion(getTestMethodName(), "ln", 1, 100, isOffHeap())); AsyncInvocation inv2 = vm5.invokeAsync(() -> WANTestBase .createPersistentPartitionedRegion(getTestMethodName(), "ln", 1, 100, isOffHeap())); AsyncInvocation inv3 = vm6.invokeAsync(() -> WANTestBase .createPersistentPartitionedRegion(getTestMethodName(), "ln", 1, 100, isOffHeap())); AsyncInvocation inv4 = vm7.invokeAsync(() -> WANTestBase .createPersistentPartitionedRegion(getTestMethodName(), "ln", 1, 100, isOffHeap())); try { inv1.join(); inv2.join(); inv3.join(); inv4.join(); } catch (InterruptedException e) { e.printStackTrace(); fail(); } LogWriterUtils.getLogWriter().info(STR); startSenderInVMsAsync("ln", vm4, vm5, vm6, vm7); LogWriterUtils.getLogWriter().info(STR); vm4.invoke(waitForSenderRunnable()); vm5.invoke(waitForSenderRunnable()); vm6.invoke(waitForSenderRunnable()); vm7.invoke(waitForSenderRunnable()); LogWriterUtils.getLogWriter().info(STR); createCacheInVMs(nyPort, vm2, vm3); LogWriterUtils.getLogWriter().info(STR); vm2.invoke(() -> WANTestBase.createPersistentPartitionedRegion(getTestMethodName(), null, 1, 100, isOffHeap())); vm3.invoke(() -> WANTestBase.createPersistentPartitionedRegion(getTestMethodName(), null, 1, 100, isOffHeap())); createReceiverInVMs(vm2, vm3); vm4.invoke(pauseSenderRunnable()); vm5.invoke(pauseSenderRunnable()); vm6.invoke(pauseSenderRunnable()); vm7.invoke(pauseSenderRunnable()); LogWriterUtils.getLogWriter().info(STR); vm4.invoke(() -> WANTestBase.doPutsAfter300(getTestMethodName(), 1000)); vm4.invoke(() -> WANTestBase.resumeSender("ln")); vm5.invoke(() -> WANTestBase.resumeSender("ln")); vm6.invoke(() -> WANTestBase.resumeSender("ln")); vm7.invoke(() -> WANTestBase.resumeSender("ln")); LogWriterUtils.getLogWriter().info(STR); vm2.invoke(() -> WANTestBase.validateRegionSize(getTestMethodName(), 1000)); vm3.invoke(() -> WANTestBase.validateRegionSize(getTestMethodName(), 1000)); } | /**
* Enable persistence for PR and GatewaySender. Pause the sender and do some puts in local region.
* Close the local site and rebuild the region and sender from disk store. Dispatcher should not
* start dispatching events recovered from persistent sender. Check if the remote site receives
* all the events.
*/ | Enable persistence for PR and GatewaySender. Pause the sender and do some puts in local region. Close the local site and rebuild the region and sender from disk store. Dispatcher should not start dispatching events recovered from persistent sender. Check if the remote site receives all the events | testPersistentPRWithGatewaySenderPersistenceEnabled_Restart2 | {
"repo_name": "pdxrunner/geode",
"path": "geode-wan/src/distributedTest/java/org/apache/geode/internal/cache/wan/parallel/ParallelWANPersistenceEnabledGatewaySenderDUnitTest.java",
"license": "apache-2.0",
"size": 70917
} | [
"org.apache.geode.internal.cache.wan.WANTestBase",
"org.apache.geode.test.dunit.AsyncInvocation",
"org.apache.geode.test.dunit.LogWriterUtils",
"org.junit.Assert"
] | import org.apache.geode.internal.cache.wan.WANTestBase; import org.apache.geode.test.dunit.AsyncInvocation; import org.apache.geode.test.dunit.LogWriterUtils; import org.junit.Assert; | import org.apache.geode.internal.cache.wan.*; import org.apache.geode.test.dunit.*; import org.junit.*; | [
"org.apache.geode",
"org.junit"
] | org.apache.geode; org.junit; | 2,704,928 |
public NodeConnectorStatistics readNodeConnector(String container,
NodeConnector nodeConnector, boolean cached); | NodeConnectorStatistics function(String container, NodeConnector nodeConnector, boolean cached); | /**
* Returns the hardware view of the specified network node connector
* for the given container
* @param node
* @return
*/ | Returns the hardware view of the specified network node connector for the given container | readNodeConnector | {
"repo_name": "lbchen/ODL",
"path": "opendaylight/protocol_plugins/openflow/src/main/java/org/opendaylight/controller/protocol_plugin/openflow/IReadServiceFilter.java",
"license": "epl-1.0",
"size": 3448
} | [
"org.opendaylight.controller.sal.core.NodeConnector",
"org.opendaylight.controller.sal.reader.NodeConnectorStatistics"
] | import org.opendaylight.controller.sal.core.NodeConnector; import org.opendaylight.controller.sal.reader.NodeConnectorStatistics; | import org.opendaylight.controller.sal.core.*; import org.opendaylight.controller.sal.reader.*; | [
"org.opendaylight.controller"
] | org.opendaylight.controller; | 1,266,446 |
public void pushToSource(Bundle bundle) {
mPipeSource.push(bundle);
} | void function(Bundle bundle) { mPipeSource.push(bundle); } | /**
* Push a {@link dk.itu.spcl.jlpf.common.Bundle} object into the source to be processed by the filters
*
* @param bundle Wrapped object to be processed
*/ | Push a <code>dk.itu.spcl.jlpf.common.Bundle</code> object into the source to be processed by the filters | pushToSource | {
"repo_name": "centosGit/JLPF",
"path": "JLPF/src/main/java/dk/itu/spcl/jlpf/core/Computable.java",
"license": "gpl-3.0",
"size": 5177
} | [
"dk.itu.spcl.jlpf.common.Bundle"
] | import dk.itu.spcl.jlpf.common.Bundle; | import dk.itu.spcl.jlpf.common.*; | [
"dk.itu.spcl"
] | dk.itu.spcl; | 845,200 |
public void testNonExistentConfigFile()
{
ThreadPoolManager.setPropsFileName( "somefilethatdoesntexist" );
ThreadPoolManager mgr = ThreadPoolManager.getInstance();
assertNotNull( mgr );
ThreadPoolExecutor pool = mgr.getPool( "doesntexist" );
assertNotNull( "Should have gotten back a pool configured like the default", pool );
int max = pool.getMaximumPoolSize();
System.out.println( max );
// it will load from the default file
int expected = Integer.parseInt( PropertyLoader.loadProperties( "cache.ccf" )
.getProperty( "thread_pool.default.maximumPoolSize" ) );
// "Max should be " + expected",
assertEquals( max, expected );
} | void function() { ThreadPoolManager.setPropsFileName( STR ); ThreadPoolManager mgr = ThreadPoolManager.getInstance(); assertNotNull( mgr ); ThreadPoolExecutor pool = mgr.getPool( STR ); assertNotNull( STR, pool ); int max = pool.getMaximumPoolSize(); System.out.println( max ); int expected = Integer.parseInt( PropertyLoader.loadProperties( STR ) .getProperty( STR ) ); assertEquals( max, expected ); } | /**
* Makes ure we can get a non existent pool from the non exitent config
* file.
*/ | Makes ure we can get a non existent pool from the non exitent config file | testNonExistentConfigFile | {
"repo_name": "tikue/jcs2-snapshot",
"path": "src/test/org/apache/commons/jcs/utils/threadpool/ThreadPoolManagerUnitTest.java",
"license": "apache-2.0",
"size": 7126
} | [
"java.util.concurrent.ThreadPoolExecutor",
"org.apache.commons.jcs.utils.props.PropertyLoader"
] | import java.util.concurrent.ThreadPoolExecutor; import org.apache.commons.jcs.utils.props.PropertyLoader; | import java.util.concurrent.*; import org.apache.commons.jcs.utils.props.*; | [
"java.util",
"org.apache.commons"
] | java.util; org.apache.commons; | 501,554 |
protected DBObject augmentObjectCacheKey(DBObject key) {
return key;
} | DBObject function(DBObject key) { return key; } | /**
* This method allows subclasses an opportunity to add any important data to a cache key, such
* as the __acl from the AccessControlMongoDAO.
*
* TODO: revisit these methods...
*
* @param key key to augment
* @return augmented key
*/ | This method allows subclasses an opportunity to add any important data to a cache key, such as the __acl from the AccessControlMongoDAO | augmentObjectCacheKey | {
"repo_name": "jeppetto/jeppetto",
"path": "jeppetto-dao-mongo/src/main/java/org/iternine/jeppetto/dao/mongodb/MongoDBQueryModelDAO.java",
"license": "apache-2.0",
"size": 51971
} | [
"com.mongodb.DBObject"
] | import com.mongodb.DBObject; | import com.mongodb.*; | [
"com.mongodb"
] | com.mongodb; | 836,143 |
private ByteBuffer encodePrincipalName( short version, String principalName )
{
String[] split = principalName.split( "@" );
String nameComponentPart = split[0];
String realm = split[1];
String[] nameComponents = nameComponentPart.split( "/" );
// Compute the size of the buffer
List<byte[]> strings = new ArrayList<byte[]>();
// Initialize the size with the number of components' size
int size = 2;
size += encodeCountedString( strings, realm );
// compute NameComponents
for ( String nameComponent : nameComponents )
{
size += encodeCountedString( strings, nameComponent );
}
ByteBuffer buffer = ByteBuffer.allocate( size );
// Now, write the data into the buffer
// store the numComponents
if ( version == Keytab.VERSION_0X501 )
{
// increment for version 0x0501
buffer.putShort( ( short ) ( nameComponents.length + 1 ) );
}
else
{
// Version = OxO502
buffer.putShort( ( short ) ( nameComponents.length ) );
}
// Store the realm and the nameComponents
for ( byte[] string : strings )
{
buffer.putShort( ( short ) ( string.length ) );
buffer.put( string );
}
buffer.flip();
return buffer;
} | ByteBuffer function( short version, String principalName ) { String[] split = principalName.split( "@" ); String nameComponentPart = split[0]; String realm = split[1]; String[] nameComponents = nameComponentPart.split( "/" ); List<byte[]> strings = new ArrayList<byte[]>(); int size = 2; size += encodeCountedString( strings, realm ); for ( String nameComponent : nameComponents ) { size += encodeCountedString( strings, nameComponent ); } ByteBuffer buffer = ByteBuffer.allocate( size ); if ( version == Keytab.VERSION_0X501 ) { buffer.putShort( ( short ) ( nameComponents.length + 1 ) ); } else { buffer.putShort( ( short ) ( nameComponents.length ) ); } for ( byte[] string : strings ) { buffer.putShort( ( short ) ( string.length ) ); buffer.put( string ); } buffer.flip(); return buffer; } | /**
* Encode a principal name.
*
* @param buffer
* @param principalName
*/ | Encode a principal name | encodePrincipalName | {
"repo_name": "darranl/directory-server",
"path": "kerberos-codec/src/main/java/org/apache/directory/server/kerberos/shared/keytab/KeytabEncoder.java",
"license": "apache-2.0",
"size": 7820
} | [
"java.nio.ByteBuffer",
"java.util.ArrayList",
"java.util.List"
] | import java.nio.ByteBuffer; import java.util.ArrayList; import java.util.List; | import java.nio.*; import java.util.*; | [
"java.nio",
"java.util"
] | java.nio; java.util; | 235,611 |
public BirthmarkService getProvider(); | BirthmarkService function(); | /**
* returns service provider interface of this extractor.
*/ | returns service provider interface of this extractor | getProvider | {
"repo_name": "tamada/stigmata",
"path": "src/main/java/com/github/stigmata/BirthmarkExtractor.java",
"license": "apache-2.0",
"size": 2109
} | [
"com.github.stigmata.spi.BirthmarkService"
] | import com.github.stigmata.spi.BirthmarkService; | import com.github.stigmata.spi.*; | [
"com.github.stigmata"
] | com.github.stigmata; | 72,522 |
int j, J, p, q, px, top, n, Gp[], Gi[], Bp[], Bi[];
double Gx[], Bx[];
if (!Dcs_util.CS_CSC(G) || !Dcs_util.CS_CSC(B) || xi == null || x == null)
return (-1);
Gp = G.p;
Gi = G.i;
Gx = G.x;
n = G.n;
Bp = B.p;
Bi = B.i;
Bx = B.x;
top = Dcs_reach.cs_reach(G, B, k, xi, pinv);
for (p = top; p < n; p++)
x[xi[p]] = 0;
for (p = Bp[k]; p < Bp[k + 1]; p++)
x[Bi[p]] = Bx[p];
for (px = top; px < n; px++) {
j = xi[px];
J = pinv != null ? (pinv[j]) : j;
if (J < 0)
continue;
x[j] /= Gx[lo ? (Gp[J]) : (Gp[J + 1] - 1)];
p = lo ? (Gp[J] + 1) : (Gp[J]);
q = lo ? (Gp[J + 1]) : (Gp[J + 1] - 1);
for (; p < q; p++) {
x[Gi[p]] -= Gx[p] * x[j];
}
}
return (top);
}
| int j, J, p, q, px, top, n, Gp[], Gi[], Bp[], Bi[]; double Gx[], Bx[]; if (!Dcs_util.CS_CSC(G) !Dcs_util.CS_CSC(B) xi == null x == null) return (-1); Gp = G.p; Gi = G.i; Gx = G.x; n = G.n; Bp = B.p; Bi = B.i; Bx = B.x; top = Dcs_reach.cs_reach(G, B, k, xi, pinv); for (p = top; p < n; p++) x[xi[p]] = 0; for (p = Bp[k]; p < Bp[k + 1]; p++) x[Bi[p]] = Bx[p]; for (px = top; px < n; px++) { j = xi[px]; J = pinv != null ? (pinv[j]) : j; if (J < 0) continue; x[j] /= Gx[lo ? (Gp[J]) : (Gp[J + 1] - 1)]; p = lo ? (Gp[J] + 1) : (Gp[J]); q = lo ? (Gp[J + 1]) : (Gp[J + 1] - 1); for (; p < q; p++) { x[Gi[p]] -= Gx[p] * x[j]; } } return (top); } | /**
* solve Gx=b(:,k), where G is either upper (lo=false) or lower (lo=true)
* triangular.
*
* @param G
* lower or upper triangular matrix in column-compressed form
* @param B
* right hand side, b=B(:,k)
* @param k
* use kth column of B as right hand side
* @param xi
* size 2*n, nonzero pattern of x in xi[top..n-1]
* @param x
* size n, x in x[xi[top..n-1]]
* @param pinv
* mapping of rows to columns of G, ignored if null
* @param lo
* true if lower triangular, false if upper
* @return top, -1 in error
*/ | solve Gx=b(:,k), where G is either upper (lo=false) or lower (lo=true) triangular | cs_spsolve | {
"repo_name": "rwl/CSparseJ",
"path": "src/main/java/edu/emory/mathcs/csparsej/tdouble/Dcs_spsolve.java",
"license": "lgpl-2.1",
"size": 3425
} | [
"edu.emory.mathcs.csparsej.tdouble.Dcs_common"
] | import edu.emory.mathcs.csparsej.tdouble.Dcs_common; | import edu.emory.mathcs.csparsej.tdouble.*; | [
"edu.emory.mathcs"
] | edu.emory.mathcs; | 1,965,594 |
WindowAndroid getWindowAndroid(); | WindowAndroid getWindowAndroid(); | /**
* Gets a {@link WindowAndroid} which is passed on to {@link ReparentingTask}, used in the
* reparenting process.
*/ | Gets a <code>WindowAndroid</code> which is passed on to <code>ReparentingTask</code>, used in the reparenting process | getWindowAndroid | {
"repo_name": "endlessm/chromium-browser",
"path": "chrome/android/java/src/org/chromium/chrome/browser/tab_activity_glue/ReparentingTask.java",
"license": "bsd-3-clause",
"size": 7924
} | [
"org.chromium.ui.base.WindowAndroid"
] | import org.chromium.ui.base.WindowAndroid; | import org.chromium.ui.base.*; | [
"org.chromium.ui"
] | org.chromium.ui; | 2,660,630 |
private void setupNameEnumerator() {
try {
_handle = NDNHandle.open();
_nameEnumerator = new NDNNameEnumerator(_handle, this);
} catch (ConfigurationException e) {
Log.warningStackTrace(e);
} catch (IOException e) {
Log.warningStackTrace(e);
}
} | void function() { try { _handle = NDNHandle.open(); _nameEnumerator = new NDNNameEnumerator(_handle, this); } catch (ConfigurationException e) { Log.warningStackTrace(e); } catch (IOException e) { Log.warningStackTrace(e); } } | /**
* Method to get an instance of a NDNHandle and NDNNameEnumerator.
*
* @return void
*/ | Method to get an instance of a NDNHandle and NDNNameEnumerator | setupNameEnumerator | {
"repo_name": "gujianxiao/gatewayForMulticom",
"path": "javasrc/src/main/org/ndnx/ndn/utils/explorer/ContentExplorer.java",
"license": "lgpl-2.1",
"size": 39307
} | [
"java.io.IOException",
"org.ndnx.ndn.NDNHandle",
"org.ndnx.ndn.config.ConfigurationException",
"org.ndnx.ndn.impl.support.Log",
"org.ndnx.ndn.profiles.nameenum.NDNNameEnumerator"
] | import java.io.IOException; import org.ndnx.ndn.NDNHandle; import org.ndnx.ndn.config.ConfigurationException; import org.ndnx.ndn.impl.support.Log; import org.ndnx.ndn.profiles.nameenum.NDNNameEnumerator; | import java.io.*; import org.ndnx.ndn.*; import org.ndnx.ndn.config.*; import org.ndnx.ndn.impl.support.*; import org.ndnx.ndn.profiles.nameenum.*; | [
"java.io",
"org.ndnx.ndn"
] | java.io; org.ndnx.ndn; | 1,184,338 |
@Override
public negative_factors_trend remove(Serializable primaryKey)
throws NoSuchnegative_factors_trendException, SystemException {
Session session = null;
try {
session = openSession();
negative_factors_trend negative_factors_trend = (negative_factors_trend)session.get(negative_factors_trendImpl.class,
primaryKey);
if (negative_factors_trend == null) {
if (_log.isWarnEnabled()) {
_log.warn(_NO_SUCH_ENTITY_WITH_PRIMARY_KEY + primaryKey);
}
throw new NoSuchnegative_factors_trendException(_NO_SUCH_ENTITY_WITH_PRIMARY_KEY +
primaryKey);
}
return remove(negative_factors_trend);
}
catch (NoSuchnegative_factors_trendException nsee) {
throw nsee;
}
catch (Exception e) {
throw processException(e);
}
finally {
closeSession(session);
}
} | negative_factors_trend function(Serializable primaryKey) throws NoSuchnegative_factors_trendException, SystemException { Session session = null; try { session = openSession(); negative_factors_trend negative_factors_trend = (negative_factors_trend)session.get(negative_factors_trendImpl.class, primaryKey); if (negative_factors_trend == null) { if (_log.isWarnEnabled()) { _log.warn(_NO_SUCH_ENTITY_WITH_PRIMARY_KEY + primaryKey); } throw new NoSuchnegative_factors_trendException(_NO_SUCH_ENTITY_WITH_PRIMARY_KEY + primaryKey); } return remove(negative_factors_trend); } catch (NoSuchnegative_factors_trendException nsee) { throw nsee; } catch (Exception e) { throw processException(e); } finally { closeSession(session); } } | /**
* Removes the negative_factors_trend with the primary key from the database. Also notifies the appropriate model listeners.
*
* @param primaryKey the primary key of the negative_factors_trend
* @return the negative_factors_trend that was removed
* @throws com.iucn.whp.dbservice.NoSuchnegative_factors_trendException if a negative_factors_trend with the primary key could not be found
* @throws SystemException if a system exception occurred
*/ | Removes the negative_factors_trend with the primary key from the database. Also notifies the appropriate model listeners | remove | {
"repo_name": "iucn-whp/world-heritage-outlook",
"path": "portlets/iucn-dbservice-portlet/docroot/WEB-INF/src/com/iucn/whp/dbservice/service/persistence/negative_factors_trendPersistenceImpl.java",
"license": "gpl-2.0",
"size": 36424
} | [
"com.liferay.portal.kernel.dao.orm.Session",
"com.liferay.portal.kernel.exception.SystemException",
"java.io.Serializable"
] | import com.liferay.portal.kernel.dao.orm.Session; import com.liferay.portal.kernel.exception.SystemException; import java.io.Serializable; | import com.liferay.portal.kernel.dao.orm.*; import com.liferay.portal.kernel.exception.*; import java.io.*; | [
"com.liferay.portal",
"java.io"
] | com.liferay.portal; java.io; | 2,347,355 |
public boolean allowSetPassword(Context context,
HttpServletRequest request,
String username)
throws SQLException
{
// XXX is this right?
return false;
} | boolean function(Context context, HttpServletRequest request, String username) throws SQLException { return false; } | /**
* Cannot change LDAP password through dspace, right?
*/ | Cannot change LDAP password through dspace, right | allowSetPassword | {
"repo_name": "rnathanday/dryad-repo",
"path": "dspace-api/src/main/java/org/dspace/authenticate/LDAPAuthentication.java",
"license": "bsd-3-clause",
"size": 17159
} | [
"java.sql.SQLException",
"javax.servlet.http.HttpServletRequest",
"org.dspace.core.Context"
] | import java.sql.SQLException; import javax.servlet.http.HttpServletRequest; import org.dspace.core.Context; | import java.sql.*; import javax.servlet.http.*; import org.dspace.core.*; | [
"java.sql",
"javax.servlet",
"org.dspace.core"
] | java.sql; javax.servlet; org.dspace.core; | 1,967,101 |
//----------//
// getPeaks //
//----------//
public List<PeakEntry<Double>> getDoublePeaks (int minCount)
{
final List<PeakEntry<Double>> peaks = new ArrayList<>();
K start = null;
K stop = null;
K best = null;
Integer bestCount = null;
boolean isAbove = false;
for (Entry<K, Integer> entry : map.entrySet()) {
if (entry.getValue() >= minCount) {
if ((bestCount == null) || (bestCount < entry.getValue())) {
best = entry.getKey();
bestCount = entry.getValue();
}
if (isAbove) { // Above -> Above
stop = entry.getKey();
} else { // Below -> Above
stop = start = entry.getKey();
isAbove = true;
}
} else {
if (isAbove) { // Above -> Below
peaks.add(
new PeakEntry<>(
createDoublePeak(start, best, stop, minCount),
(double) bestCount / totalCount));
stop = start = best = null;
bestCount = null;
isAbove = false;
} else { // Below -> Below
}
}
}
// Last range
if (isAbove) {
peaks.add(
new PeakEntry<>(
createDoublePeak(start, best, stop, minCount),
(double) bestCount / totalCount));
}
// Sort by decreasing count values
Collections.sort(peaks, reverseDoublePeakComparator);
return peaks;
}
| List<PeakEntry<Double>> function (int minCount) { final List<PeakEntry<Double>> peaks = new ArrayList<>(); K start = null; K stop = null; K best = null; Integer bestCount = null; boolean isAbove = false; for (Entry<K, Integer> entry : map.entrySet()) { if (entry.getValue() >= minCount) { if ((bestCount == null) (bestCount < entry.getValue())) { best = entry.getKey(); bestCount = entry.getValue(); } if (isAbove) { stop = entry.getKey(); } else { stop = start = entry.getKey(); isAbove = true; } } else { if (isAbove) { peaks.add( new PeakEntry<>( createDoublePeak(start, best, stop, minCount), (double) bestCount / totalCount)); stop = start = best = null; bestCount = null; isAbove = false; } else { } } } if (isAbove) { peaks.add( new PeakEntry<>( createDoublePeak(start, best, stop, minCount), (double) bestCount / totalCount)); } Collections.sort(peaks, reverseDoublePeakComparator); return peaks; } | /**
* Report the sequence of bucket peaks whose count is equal to or
* greater than the specified minCount value.
*
* @param minCount the desired minimum count value
* @return the (perhaps empty but not null) sequence of peaks of buckets
*/ | Report the sequence of bucket peaks whose count is equal to or greater than the specified minCount value | getDoublePeaks | {
"repo_name": "jlpoolen/libreveris",
"path": "src/main/omr/math/Histogram.java",
"license": "lgpl-3.0",
"size": 21357
} | [
"java.util.ArrayList",
"java.util.Collections",
"java.util.List",
"java.util.Map"
] | import java.util.ArrayList; import java.util.Collections; import java.util.List; import java.util.Map; | import java.util.*; | [
"java.util"
] | java.util; | 547,218 |
FactoryFinderResolver getFactoryFinderResolver(); | FactoryFinderResolver getFactoryFinderResolver(); | /**
* Gets the factory finder resolver to use
*
* @return the factory finder resolver
*/ | Gets the factory finder resolver to use | getFactoryFinderResolver | {
"repo_name": "christophd/camel",
"path": "core/camel-api/src/main/java/org/apache/camel/ExtendedCamelContext.java",
"license": "apache-2.0",
"size": 29137
} | [
"org.apache.camel.spi.FactoryFinderResolver"
] | import org.apache.camel.spi.FactoryFinderResolver; | import org.apache.camel.spi.*; | [
"org.apache.camel"
] | org.apache.camel; | 2,479,391 |
Collection getGroups(); | Collection getGroups(); | /**
*
* Access the groups defined for this assignment.
*
* @return A Collection (String) of group refs (authorization group ids) defined for this message; empty if none are defined.
*/ | Access the groups defined for this assignment | getGroups | {
"repo_name": "ouit0408/sakai",
"path": "assignment/assignment-api/api/src/java/org/sakaiproject/assignment/api/Assignment.java",
"license": "apache-2.0",
"size": 11954
} | [
"java.util.Collection"
] | import java.util.Collection; | import java.util.*; | [
"java.util"
] | java.util; | 935,044 |
@JsonProperty("_links")
public NoteLinks getLinks() {
return Links;
} | @JsonProperty(STR) NoteLinks function() { return Links; } | /**
* The links of a note
*
* @return
* The Links
*/ | The links of a note | getLinks | {
"repo_name": "adrobisch/putput-data",
"path": "putput-data-api/src/main/java/org/putput/api/model/Note.java",
"license": "lgpl-3.0",
"size": 4731
} | [
"com.fasterxml.jackson.annotation.JsonProperty"
] | import com.fasterxml.jackson.annotation.JsonProperty; | import com.fasterxml.jackson.annotation.*; | [
"com.fasterxml.jackson"
] | com.fasterxml.jackson; | 733,667 |
double getDouble(int columnIndex) throws OntoDriverException; | double getDouble(int columnIndex) throws OntoDriverException; | /**
* Retrieves value from column at the specified index and returns it as {@code double}.
*
* @param columnIndex Column index, the first column has index 0
* @return {@code double} value
* @throws IllegalStateException If called on a closed result set
* @throws OntoDriverException If the {@code columnIndex} is not a valid column index, the value cannot be cast to
* {@code double} or there occurs some other error
*/ | Retrieves value from column at the specified index and returns it as double | getDouble | {
"repo_name": "kbss-cvut/jopa",
"path": "ontodriver-api/src/main/java/cz/cvut/kbss/ontodriver/iteration/ResultRow.java",
"license": "lgpl-3.0",
"size": 14731
} | [
"cz.cvut.kbss.ontodriver.exception.OntoDriverException"
] | import cz.cvut.kbss.ontodriver.exception.OntoDriverException; | import cz.cvut.kbss.ontodriver.exception.*; | [
"cz.cvut.kbss"
] | cz.cvut.kbss; | 1,862,489 |
public static RefactoringStatus checkMethodInHierarchy(ITypeBinding type, String methodName, ITypeBinding returnType, ITypeBinding[] parameters) {
RefactoringStatus result= new RefactoringStatus();
IMethodBinding method= Bindings.findMethodInHierarchy(type, methodName, parameters);
if (method != null) {
boolean returnTypeClash= false;
ITypeBinding methodReturnType= method.getReturnType();
if (returnType != null && methodReturnType != null) {
String returnTypeKey= returnType.getKey();
String methodReturnTypeKey= methodReturnType.getKey();
if (returnTypeKey == null && methodReturnTypeKey == null) {
returnTypeClash= returnType != methodReturnType;
} else if (returnTypeKey != null && methodReturnTypeKey != null) {
returnTypeClash= !returnTypeKey.equals(methodReturnTypeKey);
}
}
ITypeBinding dc= method.getDeclaringClass();
if (returnTypeClash) {
result.addError(Messages.format(RefactoringCoreMessages.Checks_methodName_returnTypeClash,
new Object[] {BasicElementLabels.getJavaElementName(methodName), BasicElementLabels.getJavaElementName(dc.getName())}),
JavaStatusContext.create(method));
} else {
if (method.isConstructor()) {
result.addWarning(Messages.format(RefactoringCoreMessages.Checks_methodName_constructor,
new Object[] { BasicElementLabels.getJavaElementName(dc.getName()) }));
} else {
result.addError(Messages.format(RefactoringCoreMessages.Checks_methodName_overrides,
new Object[] { BasicElementLabels.getJavaElementName(methodName), BasicElementLabels.getJavaElementName(dc.getName()) }),
JavaStatusContext.create(method));
}
}
}
return result;
}
//---- Selection checks -------------------------------------------------------------------- | static RefactoringStatus function(ITypeBinding type, String methodName, ITypeBinding returnType, ITypeBinding[] parameters) { RefactoringStatus result= new RefactoringStatus(); IMethodBinding method= Bindings.findMethodInHierarchy(type, methodName, parameters); if (method != null) { boolean returnTypeClash= false; ITypeBinding methodReturnType= method.getReturnType(); if (returnType != null && methodReturnType != null) { String returnTypeKey= returnType.getKey(); String methodReturnTypeKey= methodReturnType.getKey(); if (returnTypeKey == null && methodReturnTypeKey == null) { returnTypeClash= returnType != methodReturnType; } else if (returnTypeKey != null && methodReturnTypeKey != null) { returnTypeClash= !returnTypeKey.equals(methodReturnTypeKey); } } ITypeBinding dc= method.getDeclaringClass(); if (returnTypeClash) { result.addError(Messages.format(RefactoringCoreMessages.Checks_methodName_returnTypeClash, new Object[] {BasicElementLabels.getJavaElementName(methodName), BasicElementLabels.getJavaElementName(dc.getName())}), JavaStatusContext.create(method)); } else { if (method.isConstructor()) { result.addWarning(Messages.format(RefactoringCoreMessages.Checks_methodName_constructor, new Object[] { BasicElementLabels.getJavaElementName(dc.getName()) })); } else { result.addError(Messages.format(RefactoringCoreMessages.Checks_methodName_overrides, new Object[] { BasicElementLabels.getJavaElementName(methodName), BasicElementLabels.getJavaElementName(dc.getName()) }), JavaStatusContext.create(method)); } } } return result; } | /**
* Checks if the new method somehow conflicts with an already existing method in
* the hierarchy. The following checks are done:
* <ul>
* <li> if the new method overrides a method defined in the given type or in one of its
* super classes. </li>
* </ul>
* @param type
* @param methodName
* @param returnType
* @param parameters
* @return the status
*/ | Checks if the new method somehow conflicts with an already existing method in the hierarchy. The following checks are done: if the new method overrides a method defined in the given type or in one of its super classes. | checkMethodInHierarchy | {
"repo_name": "elucash/eclipse-oxygen",
"path": "org.eclipse.jdt.ui/src/org/eclipse/jdt/internal/corext/refactoring/Checks.java",
"license": "epl-1.0",
"size": 34153
} | [
"org.eclipse.jdt.core.dom.IMethodBinding",
"org.eclipse.jdt.core.dom.ITypeBinding",
"org.eclipse.jdt.internal.core.manipulation.util.BasicElementLabels",
"org.eclipse.jdt.internal.corext.dom.Bindings",
"org.eclipse.jdt.internal.corext.refactoring.base.JavaStatusContext",
"org.eclipse.jdt.internal.corext.util.Messages",
"org.eclipse.ltk.core.refactoring.RefactoringStatus"
] | import org.eclipse.jdt.core.dom.IMethodBinding; import org.eclipse.jdt.core.dom.ITypeBinding; import org.eclipse.jdt.internal.core.manipulation.util.BasicElementLabels; import org.eclipse.jdt.internal.corext.dom.Bindings; import org.eclipse.jdt.internal.corext.refactoring.base.JavaStatusContext; import org.eclipse.jdt.internal.corext.util.Messages; import org.eclipse.ltk.core.refactoring.RefactoringStatus; | import org.eclipse.jdt.core.dom.*; import org.eclipse.jdt.internal.core.manipulation.util.*; import org.eclipse.jdt.internal.corext.dom.*; import org.eclipse.jdt.internal.corext.refactoring.base.*; import org.eclipse.jdt.internal.corext.util.*; import org.eclipse.ltk.core.refactoring.*; | [
"org.eclipse.jdt",
"org.eclipse.ltk"
] | org.eclipse.jdt; org.eclipse.ltk; | 2,911,768 |
return discoveryEventBus;
}
private DiscoveryEventUtils() {
//Utility class private constructor intentionally left blank.
}
public static final class SearchStartedEvent {
private final Type type;
public SearchStartedEvent(Type type) {
this.type = type;
} | return discoveryEventBus; } private DiscoveryEventUtils() { } public static final class SearchStartedEvent { private final Type type; public SearchStartedEvent(Type type) { this.type = type; } | /**
* Get the discovery event bus.
*
* @return The discovery event bus.
*/ | Get the discovery event bus | getDiscoveryEventBus | {
"repo_name": "eugene7646/autopsy",
"path": "Core/src/org/sleuthkit/autopsy/discovery/search/DiscoveryEventUtils.java",
"license": "apache-2.0",
"size": 16041
} | [
"org.sleuthkit.autopsy.discovery.search.SearchData"
] | import org.sleuthkit.autopsy.discovery.search.SearchData; | import org.sleuthkit.autopsy.discovery.search.*; | [
"org.sleuthkit.autopsy"
] | org.sleuthkit.autopsy; | 12,132 |
@Test
public void canIntegrateEntityMetadataCheckTrue()
{
String entityName = "testEntity";
DataServiceImpl dataServiceImpl = Mockito.mock(DataServiceImpl.class);
when(dataServiceImpl.hasRepository(entityName)).thenReturn(Boolean.TRUE);
DefaultEntityMetaData existingEntityMetaData = new DefaultEntityMetaData(entityName);
existingEntityMetaData.addAttribute("ID").setIdAttribute(true);
DefaultEntityMetaData newEntityMetaData = new DefaultEntityMetaData(entityName);
newEntityMetaData.addAttribute("ID").setIdAttribute(true);
when(dataServiceImpl.getEntityMetaData(entityName)).thenReturn(existingEntityMetaData);
MetaDataServiceImpl metaDataService = new MetaDataServiceImpl(dataServiceImpl);
assertTrue(metaDataService.canIntegrateEntityMetadataCheck(newEntityMetaData));
} | void function() { String entityName = STR; DataServiceImpl dataServiceImpl = Mockito.mock(DataServiceImpl.class); when(dataServiceImpl.hasRepository(entityName)).thenReturn(Boolean.TRUE); DefaultEntityMetaData existingEntityMetaData = new DefaultEntityMetaData(entityName); existingEntityMetaData.addAttribute("ID").setIdAttribute(true); DefaultEntityMetaData newEntityMetaData = new DefaultEntityMetaData(entityName); newEntityMetaData.addAttribute("ID").setIdAttribute(true); when(dataServiceImpl.getEntityMetaData(entityName)).thenReturn(existingEntityMetaData); MetaDataServiceImpl metaDataService = new MetaDataServiceImpl(dataServiceImpl); assertTrue(metaDataService.canIntegrateEntityMetadataCheck(newEntityMetaData)); } | /**
* Test that a new entity has at least all attributes as of the existing entity. If not it results false.
*/ | Test that a new entity has at least all attributes as of the existing entity. If not it results false | canIntegrateEntityMetadataCheckTrue | {
"repo_name": "adini121/molgenis",
"path": "molgenis-data/src/test/java/org/molgenis/data/meta/MetaDataServiceImplTest.java",
"license": "lgpl-3.0",
"size": 11313
} | [
"org.mockito.Mockito",
"org.molgenis.data.support.DataServiceImpl",
"org.molgenis.data.support.DefaultEntityMetaData",
"org.testng.Assert"
] | import org.mockito.Mockito; import org.molgenis.data.support.DataServiceImpl; import org.molgenis.data.support.DefaultEntityMetaData; import org.testng.Assert; | import org.mockito.*; import org.molgenis.data.support.*; import org.testng.*; | [
"org.mockito",
"org.molgenis.data",
"org.testng"
] | org.mockito; org.molgenis.data; org.testng; | 76,850 |
public Turn(Section theSection)
{
setSection(theSection);
}
public Turn(Section theSection, Node turnNode)
{
setSection(theSection);
try
{
setSpeakerId(turnNode.getAttributes().getNamedItem("speaker").getNodeValue());
} catch(Exception exception) {}
try
{
setStartTime(turnNode.getAttributes().getNamedItem("startTime").getNodeValue());
setIdFromTime();
} catch(Exception exception) {}
try
{
setEndTime(turnNode.getAttributes().getNamedItem("endTime").getNodeValue());
} catch(Exception exception) {}
// traverse nodes
NodeList children = turnNode.getChildNodes();
Sync lastSync = new Sync(this, null); // previous sync
Sync sync = new Sync(this, null); // current sync
for (int j = 0; j < children.getLength(); j++)
{
Node child = children.item(j);
if (child.getNodeType() == Node.ELEMENT_NODE
&& child.getNodeName().equalsIgnoreCase("Sync"))
{ // Sync node
lastSync = sync;
sync = new Sync(this, child);
lastSync.setEndTime(sync.getTime());
addSync(sync);
//System.out.println("start Sync: " + sync);
}
else if (child.getNodeType() == Node.ELEMENT_NODE
&& child.getNodeName().equalsIgnoreCase("Who"))
{ // Who node - multiple speakers
try
{
// there should be multiple speakers named
StringTokenizer names = new StringTokenizer(getSpeakerId(), " ");
int speakerIndex = Integer.parseInt(
child.getAttributes().getNamedItem("nb").getNodeValue());
// go to the indexed speaker
while (--speakerIndex > 0)
{
names.nextToken();
}
String strSpeaker = names.nextToken();
sync.addWho(strSpeaker);
}
catch (Exception ex)
{
//System.out.println("Turn "+getStartTime()+": Failed to parse speaker for Who: " + child.toString() + "\n" + ex.toString());
}
}
else if (child.getNodeType() == Node.ELEMENT_NODE
&& child.getNodeName().equalsIgnoreCase("Event"))
{ // Event node - non-speech event
Event event = new Event(sync, child);
//System.out.println("Event: " + event);
}
else if (child.getNodeType() == Node.ELEMENT_NODE
&& child.getNodeName().equalsIgnoreCase("Comment"))
{ // Event node - non-speech event
Comment comment = new Comment(sync, child);
//System.out.println("Comment: " + comment);
}
else if (child.getNodeType() == Node.TEXT_NODE)
{ // text node
String strText = child.getNodeValue();
if (strText.trim().length() > 0)
{
sync.appendText(strText);
//System.out.println("Text: " + strText);
//System.out.println("Sync now: " + sync);
}
}
else
{
//System.out.println("Ignoring element: " + child.toString());
}
} // next child
// set the end time of the last sync
sync.setEndTime(getEndTime());
} // end of constructor
| public Turn(Section theSection) { setSection(theSection); } public Turn(Section theSection, Node turnNode) { setSection(theSection); try { setSpeakerId(turnNode.getAttributes().getNamedItem(STR).getNodeValue()); } catch(Exception exception) {} try { setStartTime(turnNode.getAttributes().getNamedItem(STR).getNodeValue()); setIdFromTime(); } catch(Exception exception) {} try { setEndTime(turnNode.getAttributes().getNamedItem(STR).getNodeValue()); } catch(Exception exception) {} NodeList children = turnNode.getChildNodes(); Sync lastSync = new Sync(this, null); Sync sync = new Sync(this, null); for (int j = 0; j < children.getLength(); j++) { Node child = children.item(j); if (child.getNodeType() == Node.ELEMENT_NODE && child.getNodeName().equalsIgnoreCase("Sync")) { lastSync = sync; sync = new Sync(this, child); lastSync.setEndTime(sync.getTime()); addSync(sync); } else if (child.getNodeType() == Node.ELEMENT_NODE && child.getNodeName().equalsIgnoreCase("Who")) { try { StringTokenizer names = new StringTokenizer(getSpeakerId(), " "); int speakerIndex = Integer.parseInt( child.getAttributes().getNamedItem("nb").getNodeValue()); while (--speakerIndex > 0) { names.nextToken(); } String strSpeaker = names.nextToken(); sync.addWho(strSpeaker); } catch (Exception ex) { } } else if (child.getNodeType() == Node.ELEMENT_NODE && child.getNodeName().equalsIgnoreCase("Event")) { Event event = new Event(sync, child); } else if (child.getNodeType() == Node.ELEMENT_NODE && child.getNodeName().equalsIgnoreCase(STR)) { Comment comment = new Comment(sync, child); } else if (child.getNodeType() == Node.TEXT_NODE) { String strText = child.getNodeValue(); if (strText.trim().length() > 0) { sync.appendText(strText); } } else { } } sync.setEndTime(getEndTime()); } | /**
* Syncs mutator - Ordered list of syncronized transcript parts (A list of {@link nz.ac.canterbury.ling.transcriber.Sync} objects)
* @param vNewSyncs Ordered list of syncronized transcript parts (A list of {@link nz.ac.canterbury.ling.transcriber.Sync} objects)
*/ | Syncs mutator - Ordered list of syncronized transcript parts (A list of <code>nz.ac.canterbury.ling.transcriber.Sync</code> objects) | setSyncs | {
"repo_name": "nzilbb/ag",
"path": "formatter/transcriber/src/main/java/nzilbb/formatter/transcriber/Turn.java",
"license": "gpl-3.0",
"size": 11585
} | [
"java.util.StringTokenizer",
"org.w3c.dom.Node",
"org.w3c.dom.NodeList"
] | import java.util.StringTokenizer; import org.w3c.dom.Node; import org.w3c.dom.NodeList; | import java.util.*; import org.w3c.dom.*; | [
"java.util",
"org.w3c.dom"
] | java.util; org.w3c.dom; | 953,574 |
private Object processFixed(Schema schemaNode) {
int size = schemaNode.getFixedSize();
byte[] bytes = new byte[size];
for (int i = 0; i < size; i++) {
bytes[i] = (byte) 0;
}
GenericFixed result = new GenericData.Fixed(schemaNode, bytes);
return result;
} | Object function(Schema schemaNode) { int size = schemaNode.getFixedSize(); byte[] bytes = new byte[size]; for (int i = 0; i < size; i++) { bytes[i] = (byte) 0; } GenericFixed result = new GenericData.Fixed(schemaNode, bytes); return result; } | /**
* Processes fixed type.
*
* @param schemaNode schema for current type.
* @return generated value for input record type.
*/ | Processes fixed type | processFixed | {
"repo_name": "sashadidukh/kaa",
"path": "common/core/src/main/java/org/kaaproject/kaa/server/common/core/algorithms/generation/DefaultRecordGenerationAlgorithmImpl.java",
"license": "apache-2.0",
"size": 10897
} | [
"org.apache.avro.Schema",
"org.apache.avro.generic.GenericData",
"org.apache.avro.generic.GenericFixed"
] | import org.apache.avro.Schema; import org.apache.avro.generic.GenericData; import org.apache.avro.generic.GenericFixed; | import org.apache.avro.*; import org.apache.avro.generic.*; | [
"org.apache.avro"
] | org.apache.avro; | 1,366,780 |
@Override
public Object accept(ICOSVisitor visitor) throws IOException
{
return visitor.visitFromInt(this);
} | Object function(ICOSVisitor visitor) throws IOException { return visitor.visitFromInt(this); } | /**
* visitor pattern double dispatch method.
*
* @param visitor The object to notify when visiting this object.
* @return any object, depending on the visitor implementation, or null
* @throws IOException If an error occurs while visiting this object.
*/ | visitor pattern double dispatch method | accept | {
"repo_name": "ZhenyaM/veraPDF-pdfbox",
"path": "pdfbox/src/main/java/org/apache/pdfbox/cos/COSInteger.java",
"license": "apache-2.0",
"size": 4907
} | [
"java.io.IOException"
] | import java.io.IOException; | import java.io.*; | [
"java.io"
] | java.io; | 2,098,378 |
public static boolean existsCloudProvider(String cloudName) {
String xPathToProp = "/Project/Cloud/Provider[@name='" + cloudName + "']";
XPathResult res = (XPathResult) evaluator.evaluate(xPathToProp,
projectDoc,
null,
XPathResult.FIRST_ORDERED_NODE_TYPE,
null);
Node n = res.getSingleNodeValue();
return n != null;
} | static boolean function(String cloudName) { String xPathToProp = STR + cloudName + "']"; XPathResult res = (XPathResult) evaluator.evaluate(xPathToProp, projectDoc, null, XPathResult.FIRST_ORDERED_NODE_TYPE, null); Node n = res.getSingleNodeValue(); return n != null; } | /**
* Checks if a any cloud Provider with that name appears on the project file
*
* @param cloudName Name of the cloud provider
* @return true if the provider exists
*/ | Checks if a any cloud Provider with that name appears on the project file | existsCloudProvider | {
"repo_name": "ElsevierSoftwareX/SOFTX-D-15-00010",
"path": "compss/compss-rt/rt/src/main/java/integratedtoolkit/util/ProjectManager.java",
"license": "apache-2.0",
"size": 18800
} | [
"org.w3c.dom.Node",
"org.w3c.dom.xpath.XPathResult"
] | import org.w3c.dom.Node; import org.w3c.dom.xpath.XPathResult; | import org.w3c.dom.*; import org.w3c.dom.xpath.*; | [
"org.w3c.dom"
] | org.w3c.dom; | 2,686,208 |
EReference getIfExpr_Elsifs(); | EReference getIfExpr_Elsifs(); | /**
* Returns the meta object for the containment reference list '{@link org.eclectic.frontend.core.IfExpr#getElsifs <em>Elsifs</em>}'.
* <!-- begin-user-doc -->
* <!-- end-user-doc -->
* @return the meta object for the containment reference list '<em>Elsifs</em>'.
* @see org.eclectic.frontend.core.IfExpr#getElsifs()
* @see #getIfExpr()
* @generated
*/ | Returns the meta object for the containment reference list '<code>org.eclectic.frontend.core.IfExpr#getElsifs Elsifs</code>'. | getIfExpr_Elsifs | {
"repo_name": "jesusc/eclectic",
"path": "plugins/org.eclectic.frontend.asm/src-gen/org/eclectic/frontend/core/CorePackage.java",
"license": "gpl-3.0",
"size": 187193
} | [
"org.eclipse.emf.ecore.EReference"
] | import org.eclipse.emf.ecore.EReference; | import org.eclipse.emf.ecore.*; | [
"org.eclipse.emf"
] | org.eclipse.emf; | 1,635,916 |
public Map<String, String> getParasMap() {
return parasMap;
} | Map<String, String> function() { return parasMap; } | /**
* get paras map
*
* @return
*/ | get paras map | getParasMap | {
"repo_name": "solaris0403/SeleneDemo",
"path": "common_lib/src/main/java/com/tony/selene/common/trinea/android/common/entity/HttpRequest.java",
"license": "gpl-2.0",
"size": 3981
} | [
"java.util.Map"
] | import java.util.Map; | import java.util.*; | [
"java.util"
] | java.util; | 556,708 |
public Settings indexSettings() {
return this.indexSettings;
} | Settings function() { return this.indexSettings; } | /**
* Returns settings that should be added/changed in all restored indices
*/ | Returns settings that should be added/changed in all restored indices | indexSettings | {
"repo_name": "anti-social/elasticsearch",
"path": "src/main/java/org/elasticsearch/action/admin/cluster/snapshots/restore/RestoreSnapshotRequest.java",
"license": "apache-2.0",
"size": 24240
} | [
"org.elasticsearch.common.settings.Settings"
] | import org.elasticsearch.common.settings.Settings; | import org.elasticsearch.common.settings.*; | [
"org.elasticsearch.common"
] | org.elasticsearch.common; | 463,090 |
ImmutableValue<Boolean> powered(); | ImmutableValue<Boolean> powered(); | /**
* Gets the {@link ImmutableValue} for the powered state of the {@link Structure}.
*
* @return The value for the powered state
*/ | Gets the <code>ImmutableValue</code> for the powered state of the <code>Structure</code> | powered | {
"repo_name": "JBYoshi/SpongeAPI",
"path": "src/main/java/org/spongepowered/api/data/manipulator/immutable/tileentity/ImmutableStructureData.java",
"license": "mit",
"size": 3828
} | [
"org.spongepowered.api.data.value.immutable.ImmutableValue"
] | import org.spongepowered.api.data.value.immutable.ImmutableValue; | import org.spongepowered.api.data.value.immutable.*; | [
"org.spongepowered.api"
] | org.spongepowered.api; | 2,028,013 |
private Object execute(Object o, Method m, Object... args) {
Object result = null;
try {
m.setAccessible(true); // suppress Java language access
result = m.invoke(o, args);
} catch (IllegalAccessException e) {
propagateCause(e);
} catch (InvocationTargetException e) {
propagateCause(e);
}
return result;
} | Object function(Object o, Method m, Object... args) { Object result = null; try { m.setAccessible(true); result = m.invoke(o, args); } catch (IllegalAccessException e) { propagateCause(e); } catch (InvocationTargetException e) { propagateCause(e); } return result; } | /**
* Invokes the method.
*
* @return result of execution
*/ | Invokes the method | execute | {
"repo_name": "daveray/Hystrix",
"path": "hystrix-contrib/hystrix-javanica/src/main/java/com/netflix/hystrix/contrib/javanica/command/MethodExecutionAction.java",
"license": "apache-2.0",
"size": 2402
} | [
"java.lang.reflect.InvocationTargetException",
"java.lang.reflect.Method"
] | import java.lang.reflect.InvocationTargetException; import java.lang.reflect.Method; | import java.lang.reflect.*; | [
"java.lang"
] | java.lang; | 65,523 |
protected synchronized void closeOutputStream(final StreamingOutputStream streamingOutputStream) {
try {
final Iterator<WeakReference<StreamingOutputStream>> it = this.connectedOutputStreams.iterator();
while (it.hasNext()) {
StreamingOutputStream current = null;
final WeakReference<StreamingOutputStream> next = it.next();
if ((current = next.get()) == null || current == streamingOutputStream) {
it.remove();
}
}
} finally {
final StreamingChunk currentChunk = streamingOutputStream.getCurrentChunk();
if (currentChunk != null) {
currentChunk.setCanGrow(false);
}
streamingOutputStream.setCurrentChunk(null);
}
} | synchronized void function(final StreamingOutputStream streamingOutputStream) { try { final Iterator<WeakReference<StreamingOutputStream>> it = this.connectedOutputStreams.iterator(); while (it.hasNext()) { StreamingOutputStream current = null; final WeakReference<StreamingOutputStream> next = it.next(); if ((current = next.get()) == null current == streamingOutputStream) { it.remove(); } } } finally { final StreamingChunk currentChunk = streamingOutputStream.getCurrentChunk(); if (currentChunk != null) { currentChunk.setCanGrow(false); } streamingOutputStream.setCurrentChunk(null); } } | /**
* removes given StreamingOutputStream from this StreamingController. also
* set canGrow to false on its StreamingChunk
*
* @param streamingOutputStream
*/ | removes given StreamingOutputStream from this StreamingController. also set canGrow to false on its StreamingChunk | closeOutputStream | {
"repo_name": "friedlwo/AppWoksUtils",
"path": "src/org/appwork/utils/io/streamingio/Streaming.java",
"license": "artistic-2.0",
"size": 15117
} | [
"java.lang.ref.WeakReference",
"java.util.Iterator"
] | import java.lang.ref.WeakReference; import java.util.Iterator; | import java.lang.ref.*; import java.util.*; | [
"java.lang",
"java.util"
] | java.lang; java.util; | 1,510,342 |
@Test
public void testRemoveEventListenerNullType()
{
assertFalse("Wrong result",
list.removeEventListener(null, new ListenerTestImpl()));
} | void function() { assertFalse(STR, list.removeEventListener(null, new ListenerTestImpl())); } | /**
* Tests that removeEventListener() can handle a null event type.
*/ | Tests that removeEventListener() can handle a null event type | testRemoveEventListenerNullType | {
"repo_name": "mohanaraosv/commons-configuration",
"path": "src/test/java/org/apache/commons/configuration2/event/TestEventListenerList.java",
"license": "apache-2.0",
"size": 20553
} | [
"org.junit.Assert"
] | import org.junit.Assert; | import org.junit.*; | [
"org.junit"
] | org.junit; | 2,555,614 |
@JsonProperty("permissions")
public BranchImplpermissions getPermissions() {
return permissions;
} | @JsonProperty(STR) BranchImplpermissions function() { return permissions; } | /**
* Get permissions
* @return permissions
**/ | Get permissions | getPermissions | {
"repo_name": "cliffano/swaggy-jenkins",
"path": "clients/jaxrs-cxf-client/generated/src/gen/java/org/openapitools/model/BranchImpl.java",
"license": "mit",
"size": 7889
} | [
"com.fasterxml.jackson.annotation.JsonProperty",
"org.openapitools.model.BranchImplpermissions"
] | import com.fasterxml.jackson.annotation.JsonProperty; import org.openapitools.model.BranchImplpermissions; | import com.fasterxml.jackson.annotation.*; import org.openapitools.model.*; | [
"com.fasterxml.jackson",
"org.openapitools.model"
] | com.fasterxml.jackson; org.openapitools.model; | 328,696 |
public void drawResults(String tabId, String resultString, String contentType) {
QueryTab tab = (QueryTab)view.getTabs().getTab(tabId);
view.getTabs().selectTab(tabId);
if (tab == null) {
view.getElements().onError("No tab to draw results in");
}
tab.getResultContainer().processResult(resultString, contentType);
} | void function(String tabId, String resultString, String contentType) { QueryTab tab = (QueryTab)view.getTabs().getTab(tabId); view.getTabs().selectTab(tabId); if (tab == null) { view.getElements().onError(STR); } tab.getResultContainer().processResult(resultString, contentType); } | /**
* Draw json or xml results
* Keep this method in the view object, so that it is easily callable from js
*
* @param tabId Tab id to draw results in.
* Pass this, so when you switch tabs just after clicking the query button, the results still gets drawn in the proper tab
* @param resultString
* @param contentType Content type of query result
*/ | Draw json or xml results Keep this method in the view object, so that it is easily callable from js | drawResults | {
"repo_name": "JervenBolleman/yasgui",
"path": "src/main/java/com/data2semantics/yasgui/client/helpers/CallableJsMethods.java",
"license": "mit",
"size": 8400
} | [
"com.data2semantics.yasgui.client.tab.QueryTab"
] | import com.data2semantics.yasgui.client.tab.QueryTab; | import com.data2semantics.yasgui.client.tab.*; | [
"com.data2semantics.yasgui"
] | com.data2semantics.yasgui; | 1,268,764 |
InboundPacket mapInboundPacket(PiPacketOperation packetOperation)
throws PiInterpreterException; | InboundPacket mapInboundPacket(PiPacketOperation packetOperation) throws PiInterpreterException; | /**
* Returns an inbound packet equivalent to the given PI packet operation.
*
* @param packetOperation packet operation
* @return inbound packet
* @throws PiInterpreterException if the packet operation cannot be mapped
* to an inbound packet
*/ | Returns an inbound packet equivalent to the given PI packet operation | mapInboundPacket | {
"repo_name": "kuujo/onos",
"path": "core/api/src/main/java/org/onosproject/net/pi/model/PiPipelineInterpreter.java",
"license": "apache-2.0",
"size": 5652
} | [
"org.onosproject.net.packet.InboundPacket",
"org.onosproject.net.pi.runtime.PiPacketOperation"
] | import org.onosproject.net.packet.InboundPacket; import org.onosproject.net.pi.runtime.PiPacketOperation; | import org.onosproject.net.packet.*; import org.onosproject.net.pi.runtime.*; | [
"org.onosproject.net"
] | org.onosproject.net; | 1,600,412 |
private void onAppsUpdated() {
// Sort the list of apps
mApps.clear();
mApps.addAll(mComponentToAppMap.values());
Collections.sort(mApps, mAppNameComparator.getAppInfoComparator());
// As a special case for some languages (currently only Simplified Chinese), we may need to
// coalesce sections
Locale curLocale = mLauncher.getResources().getConfiguration().locale;
TreeMap<String, ArrayList<AppInfo>> sectionMap = null;
boolean localeRequiresSectionSorting = curLocale.equals(Locale.SIMPLIFIED_CHINESE);
if (localeRequiresSectionSorting) {
// Compute the section headers. We use a TreeMap with the section name comparator to
// ensure that the sections are ordered when we iterate over it later
sectionMap = new TreeMap<>(mAppNameComparator.getSectionNameComparator());
for (AppInfo info : mApps) {
// Add the section to the cache
String sectionName = getAndUpdateCachedSectionName(info.title);
// Add it to the mapping
ArrayList<AppInfo> sectionApps = sectionMap.get(sectionName);
if (sectionApps == null) {
sectionApps = new ArrayList<>();
sectionMap.put(sectionName, sectionApps);
}
sectionApps.add(info);
}
// Add each of the section apps to the list in order
List<AppInfo> allApps = new ArrayList<>(mApps.size());
for (Map.Entry<String, ArrayList<AppInfo>> entry : sectionMap.entrySet()) {
allApps.addAll(entry.getValue());
}
mApps.clear();
mApps.addAll(allApps);
} else {
// Just compute the section headers for use below
for (AppInfo info : mApps) {
// Add the section to the cache
getAndUpdateCachedSectionName(info.title);
}
}
// Recompose the set of adapter items from the current set of apps
updateAdapterItems();
} | void function() { mApps.clear(); mApps.addAll(mComponentToAppMap.values()); Collections.sort(mApps, mAppNameComparator.getAppInfoComparator()); Locale curLocale = mLauncher.getResources().getConfiguration().locale; TreeMap<String, ArrayList<AppInfo>> sectionMap = null; boolean localeRequiresSectionSorting = curLocale.equals(Locale.SIMPLIFIED_CHINESE); if (localeRequiresSectionSorting) { sectionMap = new TreeMap<>(mAppNameComparator.getSectionNameComparator()); for (AppInfo info : mApps) { String sectionName = getAndUpdateCachedSectionName(info.title); ArrayList<AppInfo> sectionApps = sectionMap.get(sectionName); if (sectionApps == null) { sectionApps = new ArrayList<>(); sectionMap.put(sectionName, sectionApps); } sectionApps.add(info); } List<AppInfo> allApps = new ArrayList<>(mApps.size()); for (Map.Entry<String, ArrayList<AppInfo>> entry : sectionMap.entrySet()) { allApps.addAll(entry.getValue()); } mApps.clear(); mApps.addAll(allApps); } else { for (AppInfo info : mApps) { getAndUpdateCachedSectionName(info.title); } } updateAdapterItems(); } | /**
* Updates internals when the set of apps are updated.
*/ | Updates internals when the set of apps are updated | onAppsUpdated | {
"repo_name": "bojanvu23/android_packages_apps_Trebuchet_Gradle",
"path": "Trebuchet/src/main/java/com/lite/android/launcher3/allapps/AlphabeticalAppsList.java",
"license": "apache-2.0",
"size": 26573
} | [
"com.lite.android.launcher3.AppInfo",
"java.util.ArrayList",
"java.util.Collections",
"java.util.List",
"java.util.Locale",
"java.util.Map",
"java.util.TreeMap"
] | import com.lite.android.launcher3.AppInfo; import java.util.ArrayList; import java.util.Collections; import java.util.List; import java.util.Locale; import java.util.Map; import java.util.TreeMap; | import com.lite.android.launcher3.*; import java.util.*; | [
"com.lite.android",
"java.util"
] | com.lite.android; java.util; | 1,654,747 |
CompleteMultipartUploadResult uploadObject(final UploadObjectRequest req)
throws IOException, InterruptedException, ExecutionException {
// Set up the pipeline for concurrent encrypt and upload
// Set up a thread pool for this pipeline
ExecutorService es = req.getExecutorService();
final boolean defaultExecutorService = es == null;
if (es == null)
es = Executors.newFixedThreadPool(clientConfiguration.getMaxConnections());
UploadObjectObserver observer = req.getUploadObjectObserver();
if (observer == null)
observer = new UploadObjectObserver();
// initialize the observer
observer.init(req, this, this, es);
// Initiate upload
observer.onUploadInitiation(req);
final List<PartETag> partETags = new ArrayList<PartETag>();
MultiFileOutputStream mfos = req.getMultiFileOutputStream();
if (mfos == null)
mfos = new MultiFileOutputStream();
try {
// initialize the multi-file output stream
mfos.init(observer, req.getPartSize(), req.getDiskLimit());
// Kicks off the encryption-upload pipeline;
// Note mfos is automatically closed upon method completion.
putLocalObject(req, mfos);
// block till all part have been uploaded
for (Future<UploadPartResult> future: observer.getFutures()) {
UploadPartResult partResult = future.get();
partETags.add(new PartETag(partResult.getPartNumber(), partResult.getETag()));
}
} finally {
if (defaultExecutorService)
es.shutdownNow(); // shut down the locally created thread pool
mfos.cleanup(); // delete left-over temp files
}
// Complete upload
return observer.onCompletion(partETags);
} | CompleteMultipartUploadResult uploadObject(final UploadObjectRequest req) throws IOException, InterruptedException, ExecutionException { ExecutorService es = req.getExecutorService(); final boolean defaultExecutorService = es == null; if (es == null) es = Executors.newFixedThreadPool(clientConfiguration.getMaxConnections()); UploadObjectObserver observer = req.getUploadObjectObserver(); if (observer == null) observer = new UploadObjectObserver(); observer.init(req, this, this, es); observer.onUploadInitiation(req); final List<PartETag> partETags = new ArrayList<PartETag>(); MultiFileOutputStream mfos = req.getMultiFileOutputStream(); if (mfos == null) mfos = new MultiFileOutputStream(); try { mfos.init(observer, req.getPartSize(), req.getDiskLimit()); putLocalObject(req, mfos); for (Future<UploadPartResult> future: observer.getFutures()) { UploadPartResult partResult = future.get(); partETags.add(new PartETag(partResult.getPartNumber(), partResult.getETag())); } } finally { if (defaultExecutorService) es.shutdownNow(); mfos.cleanup(); } return observer.onCompletion(partETags); } | /**
* Used for performance testing purposes only. Hence package private.
* This method is subject to removal anytime without notice.
*/ | Used for performance testing purposes only. Hence package private. This method is subject to removal anytime without notice | uploadObject | {
"repo_name": "sdole/aws-sdk-java",
"path": "aws-java-sdk-s3/src/main/java/com/amazonaws/services/s3/AmazonS3Client.java",
"license": "apache-2.0",
"size": 194278
} | [
"com.amazonaws.services.s3.internal.MultiFileOutputStream",
"com.amazonaws.services.s3.model.CompleteMultipartUploadResult",
"com.amazonaws.services.s3.model.PartETag",
"com.amazonaws.services.s3.model.UploadObjectRequest",
"com.amazonaws.services.s3.model.UploadPartResult",
"java.io.IOException",
"java.util.ArrayList",
"java.util.List",
"java.util.concurrent.ExecutionException",
"java.util.concurrent.ExecutorService",
"java.util.concurrent.Executors",
"java.util.concurrent.Future"
] | import com.amazonaws.services.s3.internal.MultiFileOutputStream; import com.amazonaws.services.s3.model.CompleteMultipartUploadResult; import com.amazonaws.services.s3.model.PartETag; import com.amazonaws.services.s3.model.UploadObjectRequest; import com.amazonaws.services.s3.model.UploadPartResult; import java.io.IOException; import java.util.ArrayList; import java.util.List; import java.util.concurrent.ExecutionException; import java.util.concurrent.ExecutorService; import java.util.concurrent.Executors; import java.util.concurrent.Future; | import com.amazonaws.services.s3.internal.*; import com.amazonaws.services.s3.model.*; import java.io.*; import java.util.*; import java.util.concurrent.*; | [
"com.amazonaws.services",
"java.io",
"java.util"
] | com.amazonaws.services; java.io; java.util; | 2,384,328 |
public final void tunnelProxy(HttpHost proxy, boolean secure) {
if (proxy == null) {
throw new IllegalArgumentException("Proxy host may not be null.");
}
if (!this.connected) {
throw new IllegalStateException("No tunnel unless connected.");
}
if (this.proxyChain == null) {
throw new IllegalStateException("No proxy tunnel without proxy.");
}
// prepare an extended proxy chain
HttpHost[] proxies = new HttpHost[this.proxyChain.length+1];
System.arraycopy(this.proxyChain, 0,
proxies, 0, this.proxyChain.length);
proxies[proxies.length-1] = proxy;
this.proxyChain = proxies;
this.secure = secure;
}
| final void function(HttpHost proxy, boolean secure) { if (proxy == null) { throw new IllegalArgumentException(STR); } if (!this.connected) { throw new IllegalStateException(STR); } if (this.proxyChain == null) { throw new IllegalStateException(STR); } HttpHost[] proxies = new HttpHost[this.proxyChain.length+1]; System.arraycopy(this.proxyChain, 0, proxies, 0, this.proxyChain.length); proxies[proxies.length-1] = proxy; this.proxyChain = proxies; this.secure = secure; } | /**
* Tracks tunnelling to a proxy in a proxy chain.
* This will extend the tracked proxy chain, but it does not mark
* the route as tunnelled. Only end-to-end tunnels are considered there.
*
* @param proxy the proxy tunnelled to
* @param secure <code>true</code> if the route is secure,
* <code>false</code> otherwise
*/ | Tracks tunnelling to a proxy in a proxy chain. This will extend the tracked proxy chain, but it does not mark the route as tunnelled. Only end-to-end tunnels are considered there | tunnelProxy | {
"repo_name": "onedanshow/Screen-Courter",
"path": "lib/src/org/apache/http/conn/routing/RouteTracker.java",
"license": "gpl-3.0",
"size": 12915
} | [
"org.apache.http.HttpHost"
] | import org.apache.http.HttpHost; | import org.apache.http.*; | [
"org.apache.http"
] | org.apache.http; | 2,737,979 |
@Override
public void fit(ObservationReal... oa) {
fit(Arrays.asList(oa));
} | void function(ObservationReal... oa) { fit(Arrays.asList(oa)); } | /**
* Fits this observation distribution function to a (non empty) set of
* observations. This method performs one iteration of an
* expectation-maximisation algorithm.
*
* @param oa A set of observations compatible with this function.
*/ | Fits this observation distribution function to a (non empty) set of observations. This method performs one iteration of an expectation-maximisation algorithm | fit | {
"repo_name": "KommuSoft/jahmm",
"path": "jahmm/src/jahmm/observables/OpdfGaussianMixture.java",
"license": "bsd-3-clause",
"size": 10497
} | [
"java.util.Arrays"
] | import java.util.Arrays; | import java.util.*; | [
"java.util"
] | java.util; | 2,439,871 |
private void startGemFire() {
Cache c = GemFireCacheImpl.getInstance();
if (c == null) {
CacheFactory cacheFactory = new CacheFactory();
if (logLevel != null)
cacheFactory.set("log-level", logLevel);
this.cache = cacheFactory.create();
} else
this.cache = c;
this.logger = this.cache.getLogger();
} | void function() { Cache c = GemFireCacheImpl.getInstance(); if (c == null) { CacheFactory cacheFactory = new CacheFactory(); if (logLevel != null) cacheFactory.set(STR, logLevel); this.cache = cacheFactory.create(); } else this.cache = c; this.logger = this.cache.getLogger(); } | /**
* Initializes the {@link Cache}, and creates Redis necessities
* Region and protects declares that {@link Region} to be protected.
* Also, every {@link GemFireRedisServer} will check for entries already in the
* meta data Region.
*/ | Initializes the <code>Cache</code>, and creates Redis necessities Region and protects declares that <code>Region</code> to be protected. Also, every <code>GemFireRedisServer</code> will check for entries already in the meta data Region | startGemFire | {
"repo_name": "kidaa/incubator-geode",
"path": "gemfire-core/src/main/java/com/gemstone/gemfire/redis/GemFireRedisServer.java",
"license": "apache-2.0",
"size": 26150
} | [
"com.gemstone.gemfire.cache.Cache",
"com.gemstone.gemfire.cache.CacheFactory",
"com.gemstone.gemfire.internal.cache.GemFireCacheImpl"
] | import com.gemstone.gemfire.cache.Cache; import com.gemstone.gemfire.cache.CacheFactory; import com.gemstone.gemfire.internal.cache.GemFireCacheImpl; | import com.gemstone.gemfire.cache.*; import com.gemstone.gemfire.internal.cache.*; | [
"com.gemstone.gemfire"
] | com.gemstone.gemfire; | 1,109,070 |
private void cancel() {
setResult(Activity.RESULT_CANCELED);
finish();
} | void function() { setResult(Activity.RESULT_CANCELED); finish(); } | /**
* Method that cancels the activity
*/ | Method that cancels the activity | cancel | {
"repo_name": "AnimeROM/android_package_AnimeManager",
"path": "src/com/animerom/filemanager/activities/PickerActivity.java",
"license": "apache-2.0",
"size": 21324
} | [
"android.app.Activity"
] | import android.app.Activity; | import android.app.*; | [
"android.app"
] | android.app; | 1,437,100 |
public static SpriteDefinition createSprite(int left, int top, int width, int height, int hotX, int hotY)
{
SpriteDefinition _result = new SpriteDefinition();
Box _sourceBox = new Box().withX(left).withY(top).withWidth(width).withHeight(height);
Point _hotPoint = new Point().withX(hotX).withY(hotY);
_result.withSourceBox(_sourceBox).withHotPoint(_hotPoint);
return _result;
} | static SpriteDefinition function(int left, int top, int width, int height, int hotX, int hotY) { SpriteDefinition _result = new SpriteDefinition(); Box _sourceBox = new Box().withX(left).withY(top).withWidth(width).withHeight(height); Point _hotPoint = new Point().withX(hotX).withY(hotY); _result.withSourceBox(_sourceBox).withHotPoint(_hotPoint); return _result; } | /**
* Macro to create a fully defined anonymous sprite.
*
* @param left
* x coordinate of the top-left corner.
* @param top
* y coordinate of the top-left corner.
* @param width
* width of the bloc.
* @param height
* height of the bloc.
* @param hotX
* x coordinate of the hot point.
* @param hotY
* y coordinate of the hot point.
* @return a sprite definition.
*/ | Macro to create a fully defined anonymous sprite | createSprite | {
"repo_name": "sporniket/game",
"path": "canvas/src/main/java/com/sporniket/libre/game/canvas/sprite/SpriteDefinitionUtils.java",
"license": "lgpl-3.0",
"size": 5036
} | [
"com.sporniket.libre.game.canvas.Box",
"com.sporniket.libre.game.canvas.Point"
] | import com.sporniket.libre.game.canvas.Box; import com.sporniket.libre.game.canvas.Point; | import com.sporniket.libre.game.canvas.*; | [
"com.sporniket.libre"
] | com.sporniket.libre; | 338,644 |
public void testAutoCloseServerCursorProperty() throws Exception {
String url = "jdbc:ignite:thin://127.0.0.1?" + JdbcThinUtils.PARAM_AUTO_CLOSE_SERVER_CURSOR;
String err = "Failed to parse boolean property [name=" + JdbcThinUtils.PARAM_AUTO_CLOSE_SERVER_CURSOR;
assertInvalid(url + "=0", err);
assertInvalid(url + "=1", err);
assertInvalid(url + "=false1", err);
assertInvalid(url + "=true1", err);
try (Connection conn = DriverManager.getConnection("jdbc:ignite:thin://127.0.0.1")) {
assertFalse(io(conn).autoCloseServerCursor());
}
try (Connection conn = DriverManager.getConnection(url + "=true")) {
assertTrue(io(conn).autoCloseServerCursor());
}
try (Connection conn = DriverManager.getConnection(url + "=True")) {
assertTrue(io(conn).autoCloseServerCursor());
}
try (Connection conn = DriverManager.getConnection(url + "=false")) {
assertFalse(io(conn).autoCloseServerCursor());
}
try (Connection conn = DriverManager.getConnection(url + "=False")) {
assertFalse(io(conn).autoCloseServerCursor());
}
} | void function() throws Exception { String url = STRFailed to parse boolean property [name=STR=0STR=1STR=false1STR=true1STRjdbc:ignite:thin: assertFalse(io(conn).autoCloseServerCursor()); } try (Connection conn = DriverManager.getConnection(url + "=true")) { assertTrue(io(conn).autoCloseServerCursor()); } try (Connection conn = DriverManager.getConnection(url + "=True")) { assertTrue(io(conn).autoCloseServerCursor()); } try (Connection conn = DriverManager.getConnection(url + STR)) { assertFalse(io(conn).autoCloseServerCursor()); } try (Connection conn = DriverManager.getConnection(url + STR)) { assertFalse(io(conn).autoCloseServerCursor()); } } | /**
* Test autoCloseServerCursor property handling.
*
* @throws Exception If failed.
*/ | Test autoCloseServerCursor property handling | testAutoCloseServerCursorProperty | {
"repo_name": "a1vanov/ignite",
"path": "modules/clients/src/test/java/org/apache/ignite/jdbc/thin/JdbcThinConnectionSelfTest.java",
"license": "apache-2.0",
"size": 12351
} | [
"java.sql.Connection",
"java.sql.DriverManager"
] | import java.sql.Connection; import java.sql.DriverManager; | import java.sql.*; | [
"java.sql"
] | java.sql; | 994,326 |
public void setOverheadSize(JobCollection collection)
{
OverheadSize = collection.getSettings().getBooleanProperty(Keys.KEY_Input_useReadOverhead) ? 2048000 : 0;
} | void function(JobCollection collection) { OverheadSize = collection.getSettings().getBooleanProperty(Keys.KEY_Input_useReadOverhead) ? 2048000 : 0; } | /**
* getOverhead to collect more samples
*/ | getOverhead to collect more samples | setOverheadSize | {
"repo_name": "esurharun/projectx",
"path": "src/net/sourceforge/dvb/projectx/parser/StreamParserBase.java",
"license": "gpl-2.0",
"size": 17428
} | [
"net.sourceforge.dvb.projectx.common.JobCollection",
"net.sourceforge.dvb.projectx.common.Keys"
] | import net.sourceforge.dvb.projectx.common.JobCollection; import net.sourceforge.dvb.projectx.common.Keys; | import net.sourceforge.dvb.projectx.common.*; | [
"net.sourceforge.dvb"
] | net.sourceforge.dvb; | 1,084,476 |
public List<String> getAvailablePaymentActionsForReturn(String returnId, String paymentId) throws Exception
{
MozuClient<List<String>> client = com.mozu.api.clients.commerce.ReturnClient.getAvailablePaymentActionsForReturnClient( returnId, paymentId);
client.setContext(_apiContext);
client.executeRequest();
return client.getResult();
} | List<String> function(String returnId, String paymentId) throws Exception { MozuClient<List<String>> client = com.mozu.api.clients.commerce.ReturnClient.getAvailablePaymentActionsForReturnClient( returnId, paymentId); client.setContext(_apiContext); client.executeRequest(); return client.getResult(); } | /**
* Retrieves a list of the payment actions available to perform for the specified return when a return results in a refund to the customer.
* <p><pre><code>
* Return return = new Return();
* string string = return.getAvailablePaymentActionsForReturn( returnId, paymentId);
* </code></pre></p>
* @param paymentId Unique identifier of the payment for which to perform the action.
* @param returnId Unique identifier of the return whose items you want to get.
* @return List<string>
* @see string
*/ | Retrieves a list of the payment actions available to perform for the specified return when a return results in a refund to the customer. <code><code> Return return = new Return(); string string = return.getAvailablePaymentActionsForReturn( returnId, paymentId); </code></code> | getAvailablePaymentActionsForReturn | {
"repo_name": "johngatti/mozu-java",
"path": "mozu-java-core/src/main/java/com/mozu/api/resources/commerce/ReturnResource.java",
"license": "mit",
"size": 28030
} | [
"com.mozu.api.MozuClient",
"java.util.List"
] | import com.mozu.api.MozuClient; import java.util.List; | import com.mozu.api.*; import java.util.*; | [
"com.mozu.api",
"java.util"
] | com.mozu.api; java.util; | 2,315,390 |
public boolean areaNeedSync(ARPluginContext context) throws ARException
{
return false;
} | boolean function(ARPluginContext context) throws ARException { return false; } | /**
* Plugin synchronization with the plugin server.
*
* @param context plugin context
* @return true is the plugin was synch
* @throws ARException in case of any error will occur
*/ | Plugin synchronization with the plugin server | areaNeedSync | {
"repo_name": "stefandmn/AREasy",
"path": "src/java/org/areasy/runtime/plugins/area/RuntimeArea.java",
"license": "lgpl-3.0",
"size": 8928
} | [
"com.bmc.arsys.api.ARException",
"com.bmc.arsys.pluginsvr.plugins.ARPluginContext"
] | import com.bmc.arsys.api.ARException; import com.bmc.arsys.pluginsvr.plugins.ARPluginContext; | import com.bmc.arsys.api.*; import com.bmc.arsys.pluginsvr.plugins.*; | [
"com.bmc.arsys"
] | com.bmc.arsys; | 322,426 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.