method
stringlengths
13
441k
clean_method
stringlengths
7
313k
doc
stringlengths
17
17.3k
comment
stringlengths
3
1.42k
method_name
stringlengths
1
273
extra
dict
imports
list
imports_info
stringlengths
19
34.8k
cluster_imports_info
stringlengths
15
3.66k
libraries
list
libraries_info
stringlengths
6
661
id
int64
0
2.92M
int insertSelective(TbContentCategory record);
int insertSelective(TbContentCategory record);
/** * This method was generated by MyBatis Generator. * This method corresponds to the database table tb_content_category * * @mbg.generated Fri Apr 28 17:20:11 CST 2017 */
This method was generated by MyBatis Generator. This method corresponds to the database table tb_content_category
insertSelective
{ "repo_name": "fundmarkhua/fmall", "path": "fm-manage/fm-manage-dao/src/main/java/com/fmall/mapper/TbContentCategoryMapper.java", "license": "apache-2.0", "size": 3204 }
[ "com.fmall.pojo.TbContentCategory" ]
import com.fmall.pojo.TbContentCategory;
import com.fmall.pojo.*;
[ "com.fmall.pojo" ]
com.fmall.pojo;
2,010,425
public static void formatCodeAttributesAndDescriptions(List<CodeList> codeLists) { codeLists.forEach(codeList -> { formatCodeAttributesAndDescriptions(codeList); codeList.getValues().forEach(ClsUtils::formatCodeAttributesAndDescriptions); }); } /** * Formats both code attributes, fields and their descriptions for each language. For detailed description of the * implementation refer to {@link ClsUtils#formatCodeAttributes(Code)} and {@link ClsUtils#formatDescriptions}
static void function(List<CodeList> codeLists) { codeLists.forEach(codeList -> { formatCodeAttributesAndDescriptions(codeList); codeList.getValues().forEach(ClsUtils::formatCodeAttributesAndDescriptions); }); } /** * Formats both code attributes, fields and their descriptions for each language. For detailed description of the * implementation refer to {@link ClsUtils#formatCodeAttributes(Code)} and {@link ClsUtils#formatDescriptions}
/** * Normalizes the code lists by trimming the unnecessary white spaces at the start and the end of the descriptions * and extras. * * @param codeLists * a collection of code lists to be formatted */
Normalizes the code lists by trimming the unnecessary white spaces at the start and the end of the descriptions and extras
formatCodeAttributesAndDescriptions
{ "repo_name": "SirmaITT/conservation-space-1.7.0", "path": "docker/sirma-platform/platform/seip-parent/platform/cls/cls-impl/src/main/java/com/sirma/itt/emf/cls/util/ClsUtils.java", "license": "lgpl-3.0", "size": 3601 }
[ "com.sirma.sep.cls.model.Code", "com.sirma.sep.cls.model.CodeList", "java.util.List" ]
import com.sirma.sep.cls.model.Code; import com.sirma.sep.cls.model.CodeList; import java.util.List;
import com.sirma.sep.cls.model.*; import java.util.*;
[ "com.sirma.sep", "java.util" ]
com.sirma.sep; java.util;
1,933,330
public TransHopMeta findTransHop( StepMeta from, StepMeta to, boolean disabledToo ) { for ( int i = 0; i < nrTransHops(); i++ ) { TransHopMeta hi = getTransHop( i ); if ( hi.isEnabled() || disabledToo ) { if ( hi.getFromStep() != null && hi.getToStep() != null && hi.getFromStep().equals( from ) && hi.getToStep().equals( to ) ) { return hi; } } } return null; }
TransHopMeta function( StepMeta from, StepMeta to, boolean disabledToo ) { for ( int i = 0; i < nrTransHops(); i++ ) { TransHopMeta hi = getTransHop( i ); if ( hi.isEnabled() disabledToo ) { if ( hi.getFromStep() != null && hi.getToStep() != null && hi.getFromStep().equals( from ) && hi.getToStep().equals( to ) ) { return hi; } } } return null; }
/** * Search all hops for a hop where a certain step is at the start and another is at the end. * * @param from * The step at the start of the hop. * @param to * The step at the end of the hop. * @param disabledToo * the disabled too * @return The hop or null if no hop was found. */
Search all hops for a hop where a certain step is at the start and another is at the end
findTransHop
{ "repo_name": "airy-ict/pentaho-kettle", "path": "engine/src/org/pentaho/di/trans/TransMeta.java", "license": "apache-2.0", "size": 219546 }
[ "org.pentaho.di.trans.step.StepMeta" ]
import org.pentaho.di.trans.step.StepMeta;
import org.pentaho.di.trans.step.*;
[ "org.pentaho.di" ]
org.pentaho.di;
457,623
@Override public ResetableIterator GetEnumerator() throws OperationFailedException, LockingException, GeneralFailureException, CacheException { _statusLatch.WaitForAny((byte) (NodeStatus.Initializing | NodeStatus.Running)); if (_internalCache == null) { throw new UnsupportedOperationException(); } if ((_statusLatch.IsAnyBitsSet(NodeStatus.Initializing))) { Object tempVar = getCluster().getCoordinator().clone(); { return Clustered_GetEnumerator((Address) ((tempVar instanceof Address) ? tempVar : null)); } } return new LazyKeysetEnumerator(this, (Object[]) handleKeyList(), false); }
ResetableIterator function() throws OperationFailedException, LockingException, GeneralFailureException, CacheException { _statusLatch.WaitForAny((byte) (NodeStatus.Initializing NodeStatus.Running)); if (_internalCache == null) { throw new UnsupportedOperationException(); } if ((_statusLatch.IsAnyBitsSet(NodeStatus.Initializing))) { Object tempVar = getCluster().getCoordinator().clone(); { return Clustered_GetEnumerator((Address) ((tempVar instanceof Address) ? tempVar : null)); } } return new LazyKeysetEnumerator(this, (Object[]) handleKeyList(), false); }
/** * Returns a .NET IEnumerator interface so that a client should be able to * iterate over the elements of the cache store. * * @return IDictionaryEnumerator enumerator. */
Returns a .NET IEnumerator interface so that a client should be able to iterate over the elements of the cache store
GetEnumerator
{ "repo_name": "joseananio/TayzGrid", "path": "src/tgcache/src/com/alachisoft/tayzgrid/caching/topologies/clustered/ReplicatedServerCache.java", "license": "apache-2.0", "size": 207161 }
[ "com.alachisoft.tayzgrid.caching.statistics.NodeStatus", "com.alachisoft.tayzgrid.caching.util.LazyKeysetEnumerator", "com.alachisoft.tayzgrid.common.ResetableIterator", "com.alachisoft.tayzgrid.common.net.Address", "com.alachisoft.tayzgrid.runtime.exceptions.CacheException", "com.alachisoft.tayzgrid.runtime.exceptions.GeneralFailureException", "com.alachisoft.tayzgrid.runtime.exceptions.LockingException", "com.alachisoft.tayzgrid.runtime.exceptions.OperationFailedException" ]
import com.alachisoft.tayzgrid.caching.statistics.NodeStatus; import com.alachisoft.tayzgrid.caching.util.LazyKeysetEnumerator; import com.alachisoft.tayzgrid.common.ResetableIterator; import com.alachisoft.tayzgrid.common.net.Address; import com.alachisoft.tayzgrid.runtime.exceptions.CacheException; import com.alachisoft.tayzgrid.runtime.exceptions.GeneralFailureException; import com.alachisoft.tayzgrid.runtime.exceptions.LockingException; import com.alachisoft.tayzgrid.runtime.exceptions.OperationFailedException;
import com.alachisoft.tayzgrid.caching.statistics.*; import com.alachisoft.tayzgrid.caching.util.*; import com.alachisoft.tayzgrid.common.*; import com.alachisoft.tayzgrid.common.net.*; import com.alachisoft.tayzgrid.runtime.exceptions.*;
[ "com.alachisoft.tayzgrid" ]
com.alachisoft.tayzgrid;
2,830,482
public static PowerMaxSelectorType fromString(String selector) { if (!StringUtils.isEmpty(selector)) { for (PowerMaxSelectorType selectorType : PowerMaxSelectorType.values()) { if (selectorType.getSelector().equals(selector)) { return selectorType; } } } throw new IllegalArgumentException("Invalid label: " + selector); }
static PowerMaxSelectorType function(String selector) { if (!StringUtils.isEmpty(selector)) { for (PowerMaxSelectorType selectorType : PowerMaxSelectorType.values()) { if (selectorType.getSelector().equals(selector)) { return selectorType; } } } throw new IllegalArgumentException(STR + selector); }
/** * Get the ENUM value from its string selector * * @param selector * the ENUM label as a string * * @return the corresponding ENUM value * * @throws IllegalArgumentException * if no ENUM value corresponds to this string */
Get the ENUM value from its string selector
fromString
{ "repo_name": "vgoldman/openhab", "path": "bundles/binding/org.openhab.binding.powermax/src/main/java/org/openhab/binding/powermax/internal/PowerMaxSelectorType.java", "license": "epl-1.0", "size": 2279 }
[ "org.apache.commons.lang.StringUtils" ]
import org.apache.commons.lang.StringUtils;
import org.apache.commons.lang.*;
[ "org.apache.commons" ]
org.apache.commons;
524,286
public DistributedConfigurationProcessor distributedConfiguration();
DistributedConfigurationProcessor function();
/** * Gets distributed configuration processor. * * @return Distributed configuration processor. */
Gets distributed configuration processor
distributedConfiguration
{ "repo_name": "ptupitsyn/ignite", "path": "modules/core/src/main/java/org/apache/ignite/internal/GridKernalContext.java", "license": "apache-2.0", "size": 20940 }
[ "org.apache.ignite.internal.processors.configuration.distributed.DistributedConfigurationProcessor" ]
import org.apache.ignite.internal.processors.configuration.distributed.DistributedConfigurationProcessor;
import org.apache.ignite.internal.processors.configuration.distributed.*;
[ "org.apache.ignite" ]
org.apache.ignite;
1,045,619
public static void saveCertificate(X509Certificate cert, File targetFile) { File folder = targetFile.getAbsoluteFile().getParentFile(); if (!folder.exists()) { folder.mkdirs(); } File tmpFile = new File(folder, Long.toHexString(System.currentTimeMillis()) + ".tmp"); try { boolean asPem = targetFile.getName().toLowerCase().endsWith(".pem"); if (asPem) { // PEM encoded X509 PEMWriter pemWriter = null; try { pemWriter = new PEMWriter(new FileWriter(tmpFile)); pemWriter.writeObject(cert); pemWriter.flush(); } finally { if (pemWriter != null) { pemWriter.close(); } } } else { // DER encoded X509 FileOutputStream fos = null; try { fos = new FileOutputStream(tmpFile); fos.write(cert.getEncoded()); fos.flush(); } finally { if (fos != null) { fos.close(); } } } // rename tmp file to target if (targetFile.exists()) { targetFile.delete(); } tmpFile.renameTo(targetFile); } catch (Exception e) { if (tmpFile.exists()) { tmpFile.delete(); } throw new RuntimeException("Failed to save certificate " + cert.getSubjectX500Principal().getName(), e); } }
static void function(X509Certificate cert, File targetFile) { File folder = targetFile.getAbsoluteFile().getParentFile(); if (!folder.exists()) { folder.mkdirs(); } File tmpFile = new File(folder, Long.toHexString(System.currentTimeMillis()) + ".tmp"); try { boolean asPem = targetFile.getName().toLowerCase().endsWith(".pem"); if (asPem) { PEMWriter pemWriter = null; try { pemWriter = new PEMWriter(new FileWriter(tmpFile)); pemWriter.writeObject(cert); pemWriter.flush(); } finally { if (pemWriter != null) { pemWriter.close(); } } } else { FileOutputStream fos = null; try { fos = new FileOutputStream(tmpFile); fos.write(cert.getEncoded()); fos.flush(); } finally { if (fos != null) { fos.close(); } } } if (targetFile.exists()) { targetFile.delete(); } tmpFile.renameTo(targetFile); } catch (Exception e) { if (tmpFile.exists()) { tmpFile.delete(); } throw new RuntimeException(STR + cert.getSubjectX500Principal().getName(), e); } }
/** * Saves the certificate to the file system. If the destination filename * ends with the pem extension, the certificate is written in the PEM format, * otherwise the certificate is written in the DER format. * * @param cert * @param targetFile */
Saves the certificate to the file system. If the destination filename ends with the pem extension, the certificate is written in the PEM format, otherwise the certificate is written in the DER format
saveCertificate
{ "repo_name": "ahnjune881214/gitblit", "path": "src/main/java/com/gitblit/utils/X509Utils.java", "license": "apache-2.0", "size": 41191 }
[ "java.io.File", "java.io.FileOutputStream", "java.io.FileWriter", "java.security.cert.X509Certificate", "org.bouncycastle.openssl.PEMWriter" ]
import java.io.File; import java.io.FileOutputStream; import java.io.FileWriter; import java.security.cert.X509Certificate; import org.bouncycastle.openssl.PEMWriter;
import java.io.*; import java.security.cert.*; import org.bouncycastle.openssl.*;
[ "java.io", "java.security", "org.bouncycastle.openssl" ]
java.io; java.security; org.bouncycastle.openssl;
76,199
public void unregisterWorkflowDef(String name, Integer version) { Preconditions.checkArgument(StringUtils.isNotBlank(name), "Workflow name cannot be blank"); Preconditions.checkNotNull(version, "Version cannot be null"); delete("metadata/workflow/{name}/{version}", name, version); } // Task Metadata Operations
void function(String name, Integer version) { Preconditions.checkArgument(StringUtils.isNotBlank(name), STR); Preconditions.checkNotNull(version, STR); delete(STR, name, version); }
/** * Removes the workflow definition of a workflow from the conductor server. * It does not remove associated workflows. Use with caution. * * @param name Name of the workflow to be unregistered. * @param version Version of the workflow definition to be unregistered. */
Removes the workflow definition of a workflow from the conductor server. It does not remove associated workflows. Use with caution
unregisterWorkflowDef
{ "repo_name": "grfeng/conductor", "path": "client/src/main/java/com/netflix/conductor/client/http/MetadataClient.java", "license": "apache-2.0", "size": 6994 }
[ "com.google.common.base.Preconditions", "org.apache.commons.lang.StringUtils" ]
import com.google.common.base.Preconditions; import org.apache.commons.lang.StringUtils;
import com.google.common.base.*; import org.apache.commons.lang.*;
[ "com.google.common", "org.apache.commons" ]
com.google.common; org.apache.commons;
439,808
List<CompleteEarthquake> getAllEarthquakesByMagnitude(Double magnitude);
List<CompleteEarthquake> getAllEarthquakesByMagnitude(Double magnitude);
/** * Gets all the earthquakes with a specific magnitude. * * @param magnitude The magnitude searched. * * @return The list of earthquakes of that magnitude. */
Gets all the earthquakes with a specific magnitude
getAllEarthquakesByMagnitude
{ "repo_name": "AlenaHa/tw-NEDa", "path": "tw-NEDa-backend/src/main/java/org/neda/service/EarthquakeService.java", "license": "mit", "size": 2462 }
[ "java.util.List", "org.neda.entity.CompleteEarthquake" ]
import java.util.List; import org.neda.entity.CompleteEarthquake;
import java.util.*; import org.neda.entity.*;
[ "java.util", "org.neda.entity" ]
java.util; org.neda.entity;
1,965,155
protected TimeSource activateTimeSource() { synchronized (millisEstimatorLock) { if (millisEstimator == null) { millisEstimator = new EstimatorTimeSource(new SystemTimeSource(), 10, logger); } } return millisEstimator; }
TimeSource function() { synchronized (millisEstimatorLock) { if (millisEstimator == null) { millisEstimator = new EstimatorTimeSource(new SystemTimeSource(), 10, logger); } } return millisEstimator; }
/** * Returns the TimeSource for this Cache. The TimeSource is used for any situation where the current time is required, for * example the input date for a cache entry or on getting the current time when doing expiration. * * Instantiates the TimeSource if it is not yet there. * * @return The TimeSource for this Cache. */
Returns the TimeSource for this Cache. The TimeSource is used for any situation where the current time is required, for example the input date for a cache entry or on getting the current time when doing expiration. Instantiates the TimeSource if it is not yet there
activateTimeSource
{ "repo_name": "trivago/triava", "path": "src/main/java/com/trivago/triava/tcache/Cache.java", "license": "apache-2.0", "size": 54279 }
[ "com.trivago.triava.time.EstimatorTimeSource", "com.trivago.triava.time.SystemTimeSource", "com.trivago.triava.time.TimeSource" ]
import com.trivago.triava.time.EstimatorTimeSource; import com.trivago.triava.time.SystemTimeSource; import com.trivago.triava.time.TimeSource;
import com.trivago.triava.time.*;
[ "com.trivago.triava" ]
com.trivago.triava;
1,732,221
@Override public void onAttach(IScriptable target) { }
void function(IScriptable target) { }
/** * Called when the script is attached to a target using {@link org.gearvrf.script.IScriptManager#attachScriptFile(IScriptable, IScriptFile)}. */
Called when the script is attached to a target using <code>org.gearvrf.script.IScriptManager#attachScriptFile(IScriptable, IScriptFile)</code>
onAttach
{ "repo_name": "xcaostagit/GearVRf", "path": "GVRf/Framework/framework/src/main/java/org/gearvrf/GVRMain.java", "license": "apache-2.0", "size": 13044 }
[ "org.gearvrf.script.IScriptable" ]
import org.gearvrf.script.IScriptable;
import org.gearvrf.script.*;
[ "org.gearvrf.script" ]
org.gearvrf.script;
2,634,965
@ServiceMethod(returns = ReturnType.SINGLE) private Mono<Response<Flux<ByteBuffer>>> createOrUpdateByIdWithResponseAsync( String resourceId, String apiVersion, GenericResourceInner parameters, Context context) { if (this.client.getEndpoint() == null) { return Mono .error( new IllegalArgumentException( "Parameter this.client.getEndpoint() is required and cannot be null.")); } if (resourceId == null) { return Mono.error(new IllegalArgumentException("Parameter resourceId is required and cannot be null.")); } if (apiVersion == null) { return Mono.error(new IllegalArgumentException("Parameter apiVersion is required and cannot be null.")); } if (parameters == null) { return Mono.error(new IllegalArgumentException("Parameter parameters is required and cannot be null.")); } else { parameters.validate(); } final String accept = "application/json"; context = this.client.mergeContext(context); return service .createOrUpdateById(this.client.getEndpoint(), resourceId, apiVersion, parameters, accept, context); }
@ServiceMethod(returns = ReturnType.SINGLE) Mono<Response<Flux<ByteBuffer>>> function( String resourceId, String apiVersion, GenericResourceInner parameters, Context context) { if (this.client.getEndpoint() == null) { return Mono .error( new IllegalArgumentException( STR)); } if (resourceId == null) { return Mono.error(new IllegalArgumentException(STR)); } if (apiVersion == null) { return Mono.error(new IllegalArgumentException(STR)); } if (parameters == null) { return Mono.error(new IllegalArgumentException(STR)); } else { parameters.validate(); } final String accept = STR; context = this.client.mergeContext(context); return service .createOrUpdateById(this.client.getEndpoint(), resourceId, apiVersion, parameters, accept, context); }
/** * Create a resource by ID. * * @param resourceId The fully qualified ID of the resource, including the resource name and resource type. Use the * format, * /subscriptions/{guid}/resourceGroups/{resource-group-name}/{resource-provider-namespace}/{resource-type}/{resource-name}. * @param apiVersion The API version to use for the operation. * @param parameters Create or update resource parameters. * @param context The context to associate with this operation. * @throws IllegalArgumentException thrown if parameters fail the validation. * @throws ManagementException thrown if the request is rejected by server. * @throws RuntimeException all other wrapped checked exceptions if the request fails to be sent. * @return resource information. */
Create a resource by ID
createOrUpdateByIdWithResponseAsync
{ "repo_name": "Azure/azure-sdk-for-java", "path": "sdk/resourcemanagerhybrid/azure-resourcemanager-resources/src/main/java/com/azure/resourcemanager/resources/implementation/ResourcesClientImpl.java", "license": "mit", "size": 224761 }
[ "com.azure.core.annotation.ReturnType", "com.azure.core.annotation.ServiceMethod", "com.azure.core.http.rest.Response", "com.azure.core.util.Context", "com.azure.resourcemanager.resources.fluent.models.GenericResourceInner", "java.nio.ByteBuffer" ]
import com.azure.core.annotation.ReturnType; import com.azure.core.annotation.ServiceMethod; import com.azure.core.http.rest.Response; import com.azure.core.util.Context; import com.azure.resourcemanager.resources.fluent.models.GenericResourceInner; import java.nio.ByteBuffer;
import com.azure.core.annotation.*; import com.azure.core.http.rest.*; import com.azure.core.util.*; import com.azure.resourcemanager.resources.fluent.models.*; import java.nio.*;
[ "com.azure.core", "com.azure.resourcemanager", "java.nio" ]
com.azure.core; com.azure.resourcemanager; java.nio;
1,613,216
public synchronized void close() throws MessagingException { if (!super.isConnected()) // Already closed. return; IMAPProtocol protocol = null; try { boolean isEmpty; synchronized (pool) { // If there's no authenticated connections available // don't create a new one isEmpty = pool.authenticatedConnections.isEmpty(); } if (isEmpty) { if (pool.debug) out.println("DEBUG: close() - no connections "); cleanup(); return; } protocol = getStoreProtocol(); synchronized (pool) { pool.authenticatedConnections.removeElement(protocol); } protocol.logout(); } catch (ProtocolException pex) { // Hmm .. will this ever happen ? throw new MessagingException(pex.getMessage(), pex); } finally { releaseStoreProtocol(protocol); } }
synchronized void function() throws MessagingException { if (!super.isConnected()) return; IMAPProtocol protocol = null; try { boolean isEmpty; synchronized (pool) { isEmpty = pool.authenticatedConnections.isEmpty(); } if (isEmpty) { if (pool.debug) out.println(STR); cleanup(); return; } protocol = getStoreProtocol(); synchronized (pool) { pool.authenticatedConnections.removeElement(protocol); } protocol.logout(); } catch (ProtocolException pex) { throw new MessagingException(pex.getMessage(), pex); } finally { releaseStoreProtocol(protocol); } }
/** * Close this Store. */
Close this Store
close
{ "repo_name": "arthurzaczek/kolab-android", "path": "javamail/com/sun/mail/imap/IMAPStore.java", "license": "gpl-3.0", "size": 60428 }
[ "com.sun.mail.iap.ProtocolException", "com.sun.mail.imap.protocol.IMAPProtocol", "javax.mail.MessagingException" ]
import com.sun.mail.iap.ProtocolException; import com.sun.mail.imap.protocol.IMAPProtocol; import javax.mail.MessagingException;
import com.sun.mail.iap.*; import com.sun.mail.imap.protocol.*; import javax.mail.*;
[ "com.sun.mail", "javax.mail" ]
com.sun.mail; javax.mail;
2,676,638
public Optional<Unit> getDislodgedUnit(final Province province) { final ProvinceData pd = provArray.get(province.getIndex()); if (pd != null) { return Optional.ofNullable(pd.getDislodgedUnit()); } return Optional.empty(); }// getDislodgedUnit() // last occupier
Optional<Unit> function(final Province province) { final ProvinceData pd = provArray.get(province.getIndex()); if (pd != null) { return Optional.ofNullable(pd.getDislodgedUnit()); } return Optional.empty(); }
/** * Get the dislodged unit in this Province. Returns null if no dislodged unit exists. */
Get the dislodged unit in this Province. Returns null if no dislodged unit exists
getDislodgedUnit
{ "repo_name": "takaki/jdip", "path": "src/main/java/dip/world/Position.java", "license": "gpl-2.0", "size": 22008 }
[ "java.util.Optional" ]
import java.util.Optional;
import java.util.*;
[ "java.util" ]
java.util;
1,942,126
public static void initTorque() { try { Torque.init(props); } catch (Exception e) { e.printStackTrace(); } }
static void function() { try { Torque.init(props); } catch (Exception e) { e.printStackTrace(); } }
/** * This method initialize the Torque. */
This method initialize the Torque
initTorque
{ "repo_name": "trackplus/Genji", "path": "src/test/java/com/aurel/track/dbase/DatabaseHandler.java", "license": "gpl-3.0", "size": 17827 }
[ "org.apache.torque.Torque" ]
import org.apache.torque.Torque;
import org.apache.torque.*;
[ "org.apache.torque" ]
org.apache.torque;
1,394,906
final String id = UUID.randomUUID().toString(); return new Person(id, name, birth); }
final String id = UUID.randomUUID().toString(); return new Person(id, name, birth); }
/** * Creates a new person with the given name and birth date. An id is * automatically generated. * * @param name * the person's name * @param birth * the person's birth * @return the newly created person */
Creates a new person with the given name and birth date. An id is automatically generated
createNew
{ "repo_name": "rduerig/nansai", "path": "nansai/src/com/github/nansai/data/Person.java", "license": "mit", "size": 2442 }
[ "java.util.UUID" ]
import java.util.UUID;
import java.util.*;
[ "java.util" ]
java.util;
254,583
protected static PyDouble readDoubleZero(final MarshalBuffer byteBuffer) { final PyFactory.PyDoubleBuilder builder = getFactory().getPyDoubleBuilder(); builder.value(0.0); return (builder.build()); }
static PyDouble function(final MarshalBuffer byteBuffer) { final PyFactory.PyDoubleBuilder builder = getFactory().getPyDoubleBuilder(); builder.value(0.0); return (builder.build()); }
/** * Returns <tt>PyDouble</tt>. * * @param byteBuffer * byteBuffer * @return <tt>PyDouble</tt> */
Returns PyDouble
readDoubleZero
{ "repo_name": "jevetools/unmarshal", "path": "com.jevetools.unmarshal.api.impl/src/com/jevetools/unmarshal/api/impl/RecursiveImpl.java", "license": "bsd-3-clause", "size": 31538 }
[ "com.jevetools.unmarshal.python.api.PyDouble", "com.jevetools.unmarshal.python.api.PyFactory" ]
import com.jevetools.unmarshal.python.api.PyDouble; import com.jevetools.unmarshal.python.api.PyFactory;
import com.jevetools.unmarshal.python.api.*;
[ "com.jevetools.unmarshal" ]
com.jevetools.unmarshal;
2,684,670
PropertyValue execute(PropertyValue property);
PropertyValue execute(PropertyValue property);
/** * Returns a changed property value based in the property before drilling. * * @param property current property * @return property after drilling */
Returns a changed property value based in the property before drilling
execute
{ "repo_name": "smee/gradoop", "path": "gradoop-flink/src/main/java/org/gradoop/flink/model/impl/operators/drilling/functions/drillfunctions/DrillFunction.java", "license": "apache-2.0", "size": 1189 }
[ "org.gradoop.common.model.impl.properties.PropertyValue" ]
import org.gradoop.common.model.impl.properties.PropertyValue;
import org.gradoop.common.model.impl.properties.*;
[ "org.gradoop.common" ]
org.gradoop.common;
934,122
public boolean containsAll(Collection<?> arg0) { for (Object o : arg0) { if (!contains(o)) { return false; } } return true; }
boolean function(Collection<?> arg0) { for (Object o : arg0) { if (!contains(o)) { return false; } } return true; }
/** * Set inclusion test * * @param arg0 Objects to check * @return true if the list contains all objects in arg0, false otherwise */
Set inclusion test
containsAll
{ "repo_name": "ut-osa/laminar", "path": "jikesrvm-3.0.0/rvm/src/org/jikesrvm/util/LinkedListRVM.java", "license": "bsd-3-clause", "size": 7043 }
[ "java.util.Collection" ]
import java.util.Collection;
import java.util.*;
[ "java.util" ]
java.util;
2,224,278
if (env.getActiveProfiles().length == 0) { log.warn("No Spring profile configured, running with default configuration"); } else { log.info("Running with Spring profile(s) : {}", Arrays.toString(env.getActiveProfiles())); Collection activeProfiles = Arrays.asList(env.getActiveProfiles()); if (activeProfiles.contains("dev") && activeProfiles.contains("prod")) { log.error("You have misconfigured your application! " + "It should not run with both the 'dev' and 'prod' profiles at the same time."); } if (activeProfiles.contains("prod") && activeProfiles.contains("fast")) { log.error("You have misconfigured your application! " + "It should not run with both the 'prod' and 'fast' profiles at the same time."); } if (activeProfiles.contains("dev") && activeProfiles.contains("cloud")) { log.error("You have misconfigured your application! " + "It should not run with both the 'dev' and 'cloud' profiles at the same time."); } } }
if (env.getActiveProfiles().length == 0) { log.warn(STR); } else { log.info(STR, Arrays.toString(env.getActiveProfiles())); Collection activeProfiles = Arrays.asList(env.getActiveProfiles()); if (activeProfiles.contains("dev") && activeProfiles.contains("prod")) { log.error(STR + STR); } if (activeProfiles.contains("prod") && activeProfiles.contains("fast")) { log.error(STR + STR); } if (activeProfiles.contains("dev") && activeProfiles.contains("cloud")) { log.error(STR + STR); } } }
/** * Initializes jhipster. * <p/> * Spring profiles can be configured with a program arguments --spring.profiles.active=your-active-profile * <p/> * <p> * You can find more information on how profiles work with JHipster on <a href="http://jhipster.github.io/profiles.html">http://jhipster.github.io/profiles.html</a>. * </p> */
Initializes jhipster. Spring profiles can be configured with a program arguments --spring.profiles.active=your-active-profile You can find more information on how profiles work with JHipster on HREF
initApplication
{ "repo_name": "canmahmut/shopping-list", "path": "src/main/java/de/codenorm/certification/Application.java", "license": "gpl-2.0", "size": 4067 }
[ "java.util.Arrays", "java.util.Collection" ]
import java.util.Arrays; import java.util.Collection;
import java.util.*;
[ "java.util" ]
java.util;
1,385,735
@Override public Map<K, ICacheElement<K, V>> getMultiple(Set<K> keys) { return new HashMap<K, ICacheElement<K, V>>(); }
Map<K, ICacheElement<K, V>> function(Set<K> keys) { return new HashMap<K, ICacheElement<K, V>>(); }
/** * Gets multiple items from the cache based on the given set of keys. * <p> * @param keys * @return a map of K key to ICacheElement<K, V> element, or an empty map if there is * no data in cache for any of these keys */
Gets multiple items from the cache based on the given set of keys.
getMultiple
{ "repo_name": "mohanaraosv/commons-jcs", "path": "commons-jcs-core/src/test/java/org/apache/commons/jcs/engine/control/CompositeCacheDiskUsageUnitTest.java", "license": "apache-2.0", "size": 16607 }
[ "java.util.HashMap", "java.util.Map", "java.util.Set", "org.apache.commons.jcs.engine.behavior.ICacheElement" ]
import java.util.HashMap; import java.util.Map; import java.util.Set; import org.apache.commons.jcs.engine.behavior.ICacheElement;
import java.util.*; import org.apache.commons.jcs.engine.behavior.*;
[ "java.util", "org.apache.commons" ]
java.util; org.apache.commons;
1,096,680
@TargetApi(21) public Bundler putSizeF(String key, SizeF value) { bundle.putSizeF(key, value); return this; }
@TargetApi(21) Bundler function(String key, SizeF value) { bundle.putSizeF(key, value); return this; }
/** * Inserts a SizeF value into the mapping of this Bundle, replacing * any existing value for the given key. Either key or value may be null. * * @param key a String, or null * @param value a SizeF object, or null */
Inserts a SizeF value into the mapping of this Bundle, replacing any existing value for the given key. Either key or value may be null
putSizeF
{ "repo_name": "HKMOpen/SmartTabLayout", "path": "utils-v4/src/main/java/com/ogaclejapan/smarttablayout/utils/v4/Bundler.java", "license": "apache-2.0", "size": 14195 }
[ "android.annotation.TargetApi", "android.util.SizeF" ]
import android.annotation.TargetApi; import android.util.SizeF;
import android.annotation.*; import android.util.*;
[ "android.annotation", "android.util" ]
android.annotation; android.util;
2,622,118
private void deleteUnassignedMassnObjectReferences() { // ensures that there are no references to the deleted gup_unterhaltungsmassnahme object try { final DBServer server = getDbServer(); final Statement st = server.getActiveDBConnection().getConnection().createStatement(); st.executeUpdate(CLEAR_PLANUNGSABSCHNITT_ARRAY_QUERY); st.executeUpdate(CLEAR_LOS_ARRAY_QUERY); } catch (Exception e) { log.error("Error while executing los trigger.", e); } }
void function() { try { final DBServer server = getDbServer(); final Statement st = server.getActiveDBConnection().getConnection().createStatement(); st.executeUpdate(CLEAR_PLANUNGSABSCHNITT_ARRAY_QUERY); st.executeUpdate(CLEAR_LOS_ARRAY_QUERY); } catch (Exception e) { log.error(STR, e); } }
/** * DOCUMENT ME! */
DOCUMENT ME
deleteUnassignedMassnObjectReferences
{ "repo_name": "cismet/cids-custom-wrrl-db-mv-server", "path": "src/main/java/de/cismet/cids/custom/wrrl_db_mv/server/trigger/GupTrigger.java", "license": "lgpl-3.0", "size": 9171 }
[ "java.sql.Statement" ]
import java.sql.Statement;
import java.sql.*;
[ "java.sql" ]
java.sql;
1,903,765
protected String addNewHTMLPage() { final String newId = ((CPManager) CoreSpringFactory.getBean(CPManager.class)).addBlankPage(cp, translate("cptreecontroller.newpage.title"), currentPage.getIdentifier()); final CPPage newPage = new CPPage(newId, cp); // Create an html file final VFSContainer root = cp.getRootDir(); final VFSLeaf htmlFile = root.createChildLeaf(newId + ".html"); newPage.setFile(htmlFile); updatePage(newPage); return newId; }
String function() { final String newId = ((CPManager) CoreSpringFactory.getBean(CPManager.class)).addBlankPage(cp, translate(STR), currentPage.getIdentifier()); final CPPage newPage = new CPPage(newId, cp); final VFSContainer root = cp.getRootDir(); final VFSLeaf htmlFile = root.createChildLeaf(newId + ".html"); newPage.setFile(htmlFile); updatePage(newPage); return newId; }
/** * Adds a new page to the CP * * @return */
Adds a new page to the CP
addNewHTMLPage
{ "repo_name": "huihoo/olat", "path": "olat7.8/src/main/java/org/olat/presentation/ims/cp/CPTreeController.java", "license": "apache-2.0", "size": 19440 }
[ "org.olat.data.commons.vfs.VFSContainer", "org.olat.data.commons.vfs.VFSLeaf", "org.olat.lms.ims.cp.CPManager", "org.olat.lms.ims.cp.CPPage", "org.olat.system.spring.CoreSpringFactory" ]
import org.olat.data.commons.vfs.VFSContainer; import org.olat.data.commons.vfs.VFSLeaf; import org.olat.lms.ims.cp.CPManager; import org.olat.lms.ims.cp.CPPage; import org.olat.system.spring.CoreSpringFactory;
import org.olat.data.commons.vfs.*; import org.olat.lms.ims.cp.*; import org.olat.system.spring.*;
[ "org.olat.data", "org.olat.lms", "org.olat.system" ]
org.olat.data; org.olat.lms; org.olat.system;
981,432
void metaInf(Action<? super MetaInfNormalization> configuration);
void metaInf(Action<? super MetaInfNormalization> configuration);
/** * Configures the normalization strategy for the {@code META-INF} directory in archives. * * @since 6.6 */
Configures the normalization strategy for the META-INF directory in archives
metaInf
{ "repo_name": "gradle/gradle", "path": "subprojects/core-api/src/main/java/org/gradle/normalization/RuntimeClasspathNormalization.java", "license": "apache-2.0", "size": 2864 }
[ "org.gradle.api.Action" ]
import org.gradle.api.Action;
import org.gradle.api.*;
[ "org.gradle.api" ]
org.gradle.api;
453,227
public void setDisplayConnectors(boolean dc, Color c) { setDisplayConnectors(dc); m_connectorColor = c; }
void function(boolean dc, Color c) { setDisplayConnectors(dc); m_connectorColor = c; }
/** * Turn on/off the connector points * * @param dc a <code>boolean</code> value * @param c the Color to use */
Turn on/off the connector points
setDisplayConnectors
{ "repo_name": "goddesss/DataModeling", "path": "src/weka/gui/beans/BeanVisual.java", "license": "gpl-2.0", "size": 12172 }
[ "java.awt.Color" ]
import java.awt.Color;
import java.awt.*;
[ "java.awt" ]
java.awt;
472,703
private IdentityProviderCreateContractProperties innerProperties() { return this.innerProperties; }
IdentityProviderCreateContractProperties function() { return this.innerProperties; }
/** * Get the innerProperties property: Identity Provider contract properties. * * @return the innerProperties value. */
Get the innerProperties property: Identity Provider contract properties
innerProperties
{ "repo_name": "Azure/azure-sdk-for-java", "path": "sdk/apimanagement/azure-resourcemanager-apimanagement/src/main/java/com/azure/resourcemanager/apimanagement/models/IdentityProviderCreateContract.java", "license": "mit", "size": 11091 }
[ "com.azure.resourcemanager.apimanagement.fluent.models.IdentityProviderCreateContractProperties" ]
import com.azure.resourcemanager.apimanagement.fluent.models.IdentityProviderCreateContractProperties;
import com.azure.resourcemanager.apimanagement.fluent.models.*;
[ "com.azure.resourcemanager" ]
com.azure.resourcemanager;
1,319,825
Method testMethod = method.getMethod(); if (testMethod.getName().endsWith("Bookmark")) { ((AutomatedFunctionalTestBase) test).enableBookmarkMode(); } else if (testMethod.getName().endsWith("Nav")) { ((AutomatedFunctionalTestBase) test).enableNavigationMode(); } return super.methodInvoker(method, test); }
Method testMethod = method.getMethod(); if (testMethod.getName().endsWith(STR)) { ((AutomatedFunctionalTestBase) test).enableBookmarkMode(); } else if (testMethod.getName().endsWith("Nav")) { ((AutomatedFunctionalTestBase) test).enableNavigationMode(); } return super.methodInvoker(method, test); }
/** * Test methods ending with Bookmark will have {@see AutomatedFunctionalTestBase#enableBookmarkMode} called, * test methods ending with Nav will have {@see AutomatedFunctionalTestBase#enableNavigationMode} called. * * @param method test method to check for ending in Bookmark or Nav * @param test which extends AutomatedFunctionalTestBase * @return {@see BlockJUnit4ClassRunner#methodInvoker} */
Test methods ending with Bookmark will have AutomatedFunctionalTestBase#enableBookmarkMode called, test methods ending with Nav will have AutomatedFunctionalTestBase#enableNavigationMode called
methodInvoker
{ "repo_name": "geothomasp/kualico-rice-kc", "path": "rice-tools-test/src/main/java/org/kuali/rice/testtools/selenium/AutomatedFunctionalTestRunner.java", "license": "apache-2.0", "size": 2383 }
[ "java.lang.reflect.Method" ]
import java.lang.reflect.Method;
import java.lang.reflect.*;
[ "java.lang" ]
java.lang;
2,754,190
public Observable<ServiceResponse<RouteTableInner>> getByResourceGroupWithServiceResponseAsync(String resourceGroupName, String routeTableName, String expand) { if (resourceGroupName == null) { throw new IllegalArgumentException("Parameter resourceGroupName is required and cannot be null."); } if (routeTableName == null) { throw new IllegalArgumentException("Parameter routeTableName is required and cannot be null."); } if (this.client.subscriptionId() == null) { throw new IllegalArgumentException("Parameter this.client.subscriptionId() is required and cannot be null."); }
Observable<ServiceResponse<RouteTableInner>> function(String resourceGroupName, String routeTableName, String expand) { if (resourceGroupName == null) { throw new IllegalArgumentException(STR); } if (routeTableName == null) { throw new IllegalArgumentException(STR); } if (this.client.subscriptionId() == null) { throw new IllegalArgumentException(STR); }
/** * Gets the specified route table. * * @param resourceGroupName The name of the resource group. * @param routeTableName The name of the route table. * @param expand Expands referenced resources. * @throws IllegalArgumentException thrown if parameters fail the validation * @return the observable to the RouteTableInner object */
Gets the specified route table
getByResourceGroupWithServiceResponseAsync
{ "repo_name": "navalev/azure-sdk-for-java", "path": "sdk/network/mgmt-v2019_04_01/src/main/java/com/microsoft/azure/management/network/v2019_04_01/implementation/RouteTablesInner.java", "license": "mit", "size": 75681 }
[ "com.microsoft.rest.ServiceResponse" ]
import com.microsoft.rest.ServiceResponse;
import com.microsoft.rest.*;
[ "com.microsoft.rest" ]
com.microsoft.rest;
2,830,088
private void validateArguments(TestContext context, PropertyList propertyList, List<Metadata> metadataList, SubEngineActionType actionType) throws WebServiceException { if (context == null) { throw new WebServiceException("TestContext must not be null."); } if (propertyList == null) { throw new WebServiceException("PropertyList must not be null."); } if (metadataList == null || metadataList.isEmpty()) { throw new WebServiceException("MetadataList must not be null or empty."); } if (actionType == null) { throw new WebServiceException("ActionType must not be null."); } if (!(actionType instanceof WebServiceActionType)) { throw new WebServiceException("ActionType must be a WebServiceActionType."); } } /** * Returns the metadata for the WebService-call * * @param metadataList * the list of {@link Metadata}
void function(TestContext context, PropertyList propertyList, List<Metadata> metadataList, SubEngineActionType actionType) throws WebServiceException { if (context == null) { throw new WebServiceException(STR); } if (propertyList == null) { throw new WebServiceException(STR); } if (metadataList == null metadataList.isEmpty()) { throw new WebServiceException(STR); } if (actionType == null) { throw new WebServiceException(STR); } if (!(actionType instanceof WebServiceActionType)) { throw new WebServiceException(STR); } } /** * Returns the metadata for the WebService-call * * @param metadataList * the list of {@link Metadata}
/** * Initial validation of metadata and action type. * * @param context * the test context * @param propertyList * the list of properties * @param metadataList * the metadata list * @param actionType * the action type * @throws WebServiceException * thrown, if an error occurs */
Initial validation of metadata and action type
validateArguments
{ "repo_name": "NABUCCO/org.nabucco.testautomation.engine.proxy.ws", "path": "org.nabucco.testautomation.engine.proxy.ws/src/main/org/nabucco/testautomation/engine/proxy/ws/command/soap/server/SoapServerImpl.java", "license": "epl-1.0", "size": 8090 }
[ "java.util.List", "org.nabucco.testautomation.engine.base.context.TestContext", "org.nabucco.testautomation.engine.proxy.SubEngineActionType", "org.nabucco.testautomation.engine.proxy.ws.WebServiceActionType", "org.nabucco.testautomation.engine.proxy.ws.exception.WebServiceException", "org.nabucco.testautomation.property.facade.datatype.PropertyList", "org.nabucco.testautomation.script.facade.datatype.metadata.Metadata" ]
import java.util.List; import org.nabucco.testautomation.engine.base.context.TestContext; import org.nabucco.testautomation.engine.proxy.SubEngineActionType; import org.nabucco.testautomation.engine.proxy.ws.WebServiceActionType; import org.nabucco.testautomation.engine.proxy.ws.exception.WebServiceException; import org.nabucco.testautomation.property.facade.datatype.PropertyList; import org.nabucco.testautomation.script.facade.datatype.metadata.Metadata;
import java.util.*; import org.nabucco.testautomation.engine.base.context.*; import org.nabucco.testautomation.engine.proxy.*; import org.nabucco.testautomation.engine.proxy.ws.*; import org.nabucco.testautomation.engine.proxy.ws.exception.*; import org.nabucco.testautomation.property.facade.datatype.*; import org.nabucco.testautomation.script.facade.datatype.metadata.*;
[ "java.util", "org.nabucco.testautomation" ]
java.util; org.nabucco.testautomation;
190,265
public List<GHVerifiedKey> getPublicVerifiedKeys() throws IOException { return Collections.unmodifiableList(Arrays.asList(root.retrieve().to( "/users/" + getLogin() + "/keys", GHVerifiedKey[].class))); }
List<GHVerifiedKey> function() throws IOException { return Collections.unmodifiableList(Arrays.asList(root.retrieve().to( STR + getLogin() + "/keys", GHVerifiedKey[].class))); }
/** * Returns the read-only list of all the public verified keys of the current user. * * Differently from the getPublicKeys() method, the retrieval of the user's * verified public keys does not require any READ/WRITE OAuth Scope to the * user's profile. * * @return * Always non-null. */
Returns the read-only list of all the public verified keys of the current user. Differently from the getPublicKeys() method, the retrieval of the user's verified public keys does not require any READ/WRITE OAuth Scope to the user's profile
getPublicVerifiedKeys
{ "repo_name": "PankajHingane-REISysIn/KnowledgeArticleExportImport", "path": "src/main/java/org/kohsuke/github/GHMyself.java", "license": "mit", "size": 6661 }
[ "java.io.IOException", "java.util.Arrays", "java.util.Collections", "java.util.List" ]
import java.io.IOException; import java.util.Arrays; import java.util.Collections; import java.util.List;
import java.io.*; import java.util.*;
[ "java.io", "java.util" ]
java.io; java.util;
990,922
public ServiceFuture<StreamingJobInner> getByResourceGroupAsync(String resourceGroupName, String jobName, final ServiceCallback<StreamingJobInner> serviceCallback) { return ServiceFuture.fromHeaderResponse(getByResourceGroupWithServiceResponseAsync(resourceGroupName, jobName), serviceCallback); }
ServiceFuture<StreamingJobInner> function(String resourceGroupName, String jobName, final ServiceCallback<StreamingJobInner> serviceCallback) { return ServiceFuture.fromHeaderResponse(getByResourceGroupWithServiceResponseAsync(resourceGroupName, jobName), serviceCallback); }
/** * Gets details about the specified streaming job. * * @param resourceGroupName The name of the resource group that contains the resource. You can obtain this value from the Azure Resource Manager API or the portal. * @param jobName The name of the streaming job. * @param serviceCallback the async ServiceCallback to handle successful and failed responses. * @throws IllegalArgumentException thrown if parameters fail the validation * @return the {@link ServiceFuture} object */
Gets details about the specified streaming job
getByResourceGroupAsync
{ "repo_name": "hovsepm/azure-sdk-for-java", "path": "streamanalytics/resource-manager/v2016_03_01/src/main/java/com/microsoft/azure/management/streamanalytics/v2016_03_01/implementation/StreamingJobsInner.java", "license": "mit", "size": 143599 }
[ "com.microsoft.rest.ServiceCallback", "com.microsoft.rest.ServiceFuture" ]
import com.microsoft.rest.ServiceCallback; import com.microsoft.rest.ServiceFuture;
import com.microsoft.rest.*;
[ "com.microsoft.rest" ]
com.microsoft.rest;
1,389,112
public native List<String> getListInterestsInNetwork() throws RepaConnectException, ClassNotFoundException;
native List<String> function() throws RepaConnectException, ClassNotFoundException;
/** * Get a list of interest collected by daemon in NETWORK * return the interests in list passed by parameter * * @return list of interest * * @throws RepaConnectException - if an error occurs when connect the socket(include number exception) * @throws ClassNotFoundException - class not found in JNI */
Get a list of interest collected by daemon in NETWORK return the interests in list passed by parameter
getListInterestsInNetwork
{ "repo_name": "hmoraes/repa-java-api", "path": "src/ufrj/coppe/lcp/repa/RepaSocket.java", "license": "gpl-2.0", "size": 7827 }
[ "java.util.List" ]
import java.util.List;
import java.util.*;
[ "java.util" ]
java.util;
1,193,119
public Collection<Point2D> vertices() { return vertices; }
Collection<Point2D> function() { return vertices; }
/** * Returns the points of the polygon. The result is a pointer to the inner * collection of vertices. */
Returns the points of the polygon. The result is a pointer to the inner collection of vertices
vertices
{ "repo_name": "pokowaka/android-geom", "path": "geom/src/main/java/math/geom2d/polygon/SimplePolygon2D.java", "license": "lgpl-2.1", "size": 18458 }
[ "java.util.Collection", "math.geom2d.Point2D" ]
import java.util.Collection; import math.geom2d.Point2D;
import java.util.*; import math.geom2d.*;
[ "java.util", "math.geom2d" ]
java.util; math.geom2d;
2,143,044
public static Trade adaptTrade(ANXTrade anxTrade) { OrderType orderType = adaptSide(anxTrade.getTradeType()); BigDecimal amount = anxTrade.getAmount(); BigDecimal price = anxTrade.getPrice(); CurrencyPair currencyPair = adaptCurrencyPair(anxTrade.getItem(), anxTrade.getPriceCurrency()); Date dateTime = DateUtils.fromMillisUtc(anxTrade.getTid()); final String tradeId = String.valueOf(anxTrade.getTid()); return new Trade(orderType, amount, currencyPair, price, dateTime, tradeId); }
static Trade function(ANXTrade anxTrade) { OrderType orderType = adaptSide(anxTrade.getTradeType()); BigDecimal amount = anxTrade.getAmount(); BigDecimal price = anxTrade.getPrice(); CurrencyPair currencyPair = adaptCurrencyPair(anxTrade.getItem(), anxTrade.getPriceCurrency()); Date dateTime = DateUtils.fromMillisUtc(anxTrade.getTid()); final String tradeId = String.valueOf(anxTrade.getTid()); return new Trade(orderType, amount, currencyPair, price, dateTime, tradeId); }
/** * Adapts a ANXTrade to a Trade Object * * @param anxTrade * @return */
Adapts a ANXTrade to a Trade Object
adaptTrade
{ "repo_name": "codeck/XChange", "path": "xchange-anx/src/main/java/com/xeiam/xchange/anx/v2/ANXAdapters.java", "license": "mit", "size": 9535 }
[ "com.xeiam.xchange.anx.v2.dto.marketdata.ANXTrade", "com.xeiam.xchange.currency.CurrencyPair", "com.xeiam.xchange.dto.Order", "com.xeiam.xchange.dto.marketdata.Trade", "com.xeiam.xchange.utils.DateUtils", "java.math.BigDecimal", "java.util.Date" ]
import com.xeiam.xchange.anx.v2.dto.marketdata.ANXTrade; import com.xeiam.xchange.currency.CurrencyPair; import com.xeiam.xchange.dto.Order; import com.xeiam.xchange.dto.marketdata.Trade; import com.xeiam.xchange.utils.DateUtils; import java.math.BigDecimal; import java.util.Date;
import com.xeiam.xchange.anx.v2.dto.marketdata.*; import com.xeiam.xchange.currency.*; import com.xeiam.xchange.dto.*; import com.xeiam.xchange.dto.marketdata.*; import com.xeiam.xchange.utils.*; import java.math.*; import java.util.*;
[ "com.xeiam.xchange", "java.math", "java.util" ]
com.xeiam.xchange; java.math; java.util;
2,845,782
dbc.precondition(collection != null, "cannot call all with a null collection"); final Function<Iterator<E>, R> consumer = new ConsumeIntoCollection<>(new ConstantSupplier<R>(collection)); return consumer.apply(iterator); }
dbc.precondition(collection != null, STR); final Function<Iterator<E>, R> consumer = new ConsumeIntoCollection<>(new ConstantSupplier<R>(collection)); return consumer.apply(iterator); }
/** * Yields all elements of the iterator (in the provided collection). * * @param <R> the returned collection type * @param <E> the collection element type * @param iterator the iterator that will be consumed * @param collection the collection where the iterator is consumed * @return the collection filled with iterator values */
Yields all elements of the iterator (in the provided collection)
all
{ "repo_name": "emaze/emaze-dysfunctional", "path": "src/main/java/net/emaze/dysfunctional/Consumers.java", "license": "bsd-3-clause", "size": 28336 }
[ "java.util.Iterator", "java.util.function.Function", "net.emaze.dysfunctional.consumers.ConsumeIntoCollection", "net.emaze.dysfunctional.dispatching.delegates.ConstantSupplier" ]
import java.util.Iterator; import java.util.function.Function; import net.emaze.dysfunctional.consumers.ConsumeIntoCollection; import net.emaze.dysfunctional.dispatching.delegates.ConstantSupplier;
import java.util.*; import java.util.function.*; import net.emaze.dysfunctional.consumers.*; import net.emaze.dysfunctional.dispatching.delegates.*;
[ "java.util", "net.emaze.dysfunctional" ]
java.util; net.emaze.dysfunctional;
77,904
public long getCapacity() throws IOException { return volumes.getCapacity(); }
long function() throws IOException { return volumes.getCapacity(); }
/** * Return total capacity, used and unused */
Return total capacity, used and unused
getCapacity
{ "repo_name": "ryanobjc/hadoop-cloudera", "path": "src/hdfs/org/apache/hadoop/hdfs/server/datanode/FSDataset.java", "license": "apache-2.0", "size": 63084 }
[ "java.io.IOException" ]
import java.io.IOException;
import java.io.*;
[ "java.io" ]
java.io;
538,395
@Override public void onCancelResetChange() { startActivity(new Intent(AuthActivity.this, ViewfinderActivity.class)); }
void function() { startActivity(new Intent(AuthActivity.this, ViewfinderActivity.class)); }
/** * Called when the user clicks the cancel button. Goes to the inbox. */
Called when the user clicks the cancel button. Goes to the inbox
onCancelResetChange
{ "repo_name": "0359xiaodong/viewfinder", "path": "clients/android/src/co/viewfinder/AuthActivity.java", "license": "apache-2.0", "size": 17072 }
[ "android.content.Intent" ]
import android.content.Intent;
import android.content.*;
[ "android.content" ]
android.content;
1,941,763
protected void childBooleanValue(Production node, Node child) throws ParseException { node.addChild(child); }
void function(Production node, Node child) throws ParseException { node.addChild(child); }
/** * Called when adding a child to a parse tree node. * * @param node the parent node * @param child the child node, or null * * @throws ParseException if the node analysis discovered errors */
Called when adding a child to a parse tree node
childBooleanValue
{ "repo_name": "richb-hanover/mibble-2.9.2", "path": "src/java/net/percederberg/mibble/asn1/Asn1Analyzer.java", "license": "gpl-2.0", "size": 275483 }
[ "net.percederberg.grammatica.parser.Node", "net.percederberg.grammatica.parser.ParseException", "net.percederberg.grammatica.parser.Production" ]
import net.percederberg.grammatica.parser.Node; import net.percederberg.grammatica.parser.ParseException; import net.percederberg.grammatica.parser.Production;
import net.percederberg.grammatica.parser.*;
[ "net.percederberg.grammatica" ]
net.percederberg.grammatica;
447,722
public void setTimeZone(TimeZone timeZone) { this.timeZone = timeZone; }
void function(TimeZone timeZone) { this.timeZone = timeZone; }
/** * Sets the time zone for the field. * * @param timeZone * the new time zone to use * @since 8.2 */
Sets the time zone for the field
setTimeZone
{ "repo_name": "Darsstar/framework", "path": "client/src/main/java/com/vaadin/client/ui/VAbstractTextualDate.java", "license": "apache-2.0", "size": 16452 }
[ "com.google.gwt.i18n.client.TimeZone" ]
import com.google.gwt.i18n.client.TimeZone;
import com.google.gwt.i18n.client.*;
[ "com.google.gwt" ]
com.google.gwt;
2,874,864
public static boolean validatePHRExtractTemplateId(PHRExtract phrExtract, DiagnosticChain diagnostics, Map<Object, Object> context) { if (VALIDATE_PHR_EXTRACT_TEMPLATE_ID__DIAGNOSTIC_CHAIN_MAP__EOCL_INV == null) { OCL.Helper helper = EOCL_ENV.createOCLHelper(); helper.setContext(IHEPackage.Literals.PHR_EXTRACT); try { VALIDATE_PHR_EXTRACT_TEMPLATE_ID__DIAGNOSTIC_CHAIN_MAP__EOCL_INV = helper.createInvariant(VALIDATE_PHR_EXTRACT_TEMPLATE_ID__DIAGNOSTIC_CHAIN_MAP__EOCL_EXP); } catch (ParserException pe) { throw new UnsupportedOperationException(pe.getLocalizedMessage()); } } if (!EOCL_ENV.createQuery(VALIDATE_PHR_EXTRACT_TEMPLATE_ID__DIAGNOSTIC_CHAIN_MAP__EOCL_INV).check(phrExtract)) { if (diagnostics != null) { diagnostics.add (new BasicDiagnostic (Diagnostic.ERROR, IHEValidator.DIAGNOSTIC_SOURCE, IHEValidator.PHR_EXTRACT__PHR_EXTRACT_TEMPLATE_ID, IHEPlugin.INSTANCE.getString("PHRExtractTemplateId"), new Object [] { phrExtract })); } return false; } return true; }
static boolean function(PHRExtract phrExtract, DiagnosticChain diagnostics, Map<Object, Object> context) { if (VALIDATE_PHR_EXTRACT_TEMPLATE_ID__DIAGNOSTIC_CHAIN_MAP__EOCL_INV == null) { OCL.Helper helper = EOCL_ENV.createOCLHelper(); helper.setContext(IHEPackage.Literals.PHR_EXTRACT); try { VALIDATE_PHR_EXTRACT_TEMPLATE_ID__DIAGNOSTIC_CHAIN_MAP__EOCL_INV = helper.createInvariant(VALIDATE_PHR_EXTRACT_TEMPLATE_ID__DIAGNOSTIC_CHAIN_MAP__EOCL_EXP); } catch (ParserException pe) { throw new UnsupportedOperationException(pe.getLocalizedMessage()); } } if (!EOCL_ENV.createQuery(VALIDATE_PHR_EXTRACT_TEMPLATE_ID__DIAGNOSTIC_CHAIN_MAP__EOCL_INV).check(phrExtract)) { if (diagnostics != null) { diagnostics.add (new BasicDiagnostic (Diagnostic.ERROR, IHEValidator.DIAGNOSTIC_SOURCE, IHEValidator.PHR_EXTRACT__PHR_EXTRACT_TEMPLATE_ID, IHEPlugin.INSTANCE.getString(STR), new Object [] { phrExtract })); } return false; } return true; }
/** * <!-- begin-user-doc --> * <!-- end-user-doc --> * <!-- begin-model-doc --> * self.templateId->exists(id : datatypes::II | id.root = '1.3.6.1.4.1.19376.1.5.3.1.1.5') * @param phrExtract The receiving '<em><b>PHR Extract</b></em>' model object. * @param diagnostics The chain of diagnostics to which problems are to be appended. * @param context The cache of context-specific information. * <!-- end-model-doc --> * @generated */
self.templateId->exists(id : datatypes::II | id.root = '1.3.6.1.4.1.19376.1.5.3.1.1.5')
validatePHRExtractTemplateId
{ "repo_name": "drbgfc/mdht", "path": "cda/deprecated/org.openhealthtools.mdht.uml.cda.ihe/src/org/openhealthtools/mdht/uml/cda/ihe/operations/PHRExtractOperations.java", "license": "epl-1.0", "size": 4053 }
[ "java.util.Map", "org.eclipse.emf.common.util.BasicDiagnostic", "org.eclipse.emf.common.util.Diagnostic", "org.eclipse.emf.common.util.DiagnosticChain", "org.eclipse.ocl.ParserException", "org.openhealthtools.mdht.uml.cda.ihe.IHEPackage", "org.openhealthtools.mdht.uml.cda.ihe.IHEPlugin", "org.openhealthtools.mdht.uml.cda.ihe.PHRExtract", "org.openhealthtools.mdht.uml.cda.ihe.util.IHEValidator" ]
import java.util.Map; import org.eclipse.emf.common.util.BasicDiagnostic; import org.eclipse.emf.common.util.Diagnostic; import org.eclipse.emf.common.util.DiagnosticChain; import org.eclipse.ocl.ParserException; import org.openhealthtools.mdht.uml.cda.ihe.IHEPackage; import org.openhealthtools.mdht.uml.cda.ihe.IHEPlugin; import org.openhealthtools.mdht.uml.cda.ihe.PHRExtract; import org.openhealthtools.mdht.uml.cda.ihe.util.IHEValidator;
import java.util.*; import org.eclipse.emf.common.util.*; import org.eclipse.ocl.*; import org.openhealthtools.mdht.uml.cda.ihe.*; import org.openhealthtools.mdht.uml.cda.ihe.util.*;
[ "java.util", "org.eclipse.emf", "org.eclipse.ocl", "org.openhealthtools.mdht" ]
java.util; org.eclipse.emf; org.eclipse.ocl; org.openhealthtools.mdht;
2,603,975
public Map<Long, Map<Long, List<TreeImageDisplay>>> getData() { return data; }
Map<Long, Map<Long, List<TreeImageDisplay>>> function() { return data; }
/** * Returns the data. * * @return See above. */
Returns the data
getData
{ "repo_name": "knabar/openmicroscopy", "path": "components/insight/SRC/org/openmicroscopy/shoola/agents/events/treeviewer/ExperimenterLoadedDataEvent.java", "license": "gpl-2.0", "size": 2149 }
[ "java.util.List", "java.util.Map", "org.openmicroscopy.shoola.agents.util.browser.TreeImageDisplay" ]
import java.util.List; import java.util.Map; import org.openmicroscopy.shoola.agents.util.browser.TreeImageDisplay;
import java.util.*; import org.openmicroscopy.shoola.agents.util.browser.*;
[ "java.util", "org.openmicroscopy.shoola" ]
java.util; org.openmicroscopy.shoola;
2,726,016
public ArrayList<ArrayList<String>> getEcotypes() { return ecotypes; }
ArrayList<ArrayList<String>> function() { return ecotypes; }
/** * Get the ecotypes. * * @return The ecotypes. */
Get the ecotypes
getEcotypes
{ "repo_name": "weiway/New-Ecotype-Simulation", "path": "src/java/Demarcation.java", "license": "gpl-3.0", "size": 9219 }
[ "java.util.ArrayList" ]
import java.util.ArrayList;
import java.util.*;
[ "java.util" ]
java.util;
2,701,544
public ServiceCall<Void> addPetAsync(final ServiceCallback<Void> serviceCallback) { return ServiceCall.create(addPetWithServiceResponseAsync(), serviceCallback); }
ServiceCall<Void> function(final ServiceCallback<Void> serviceCallback) { return ServiceCall.create(addPetWithServiceResponseAsync(), serviceCallback); }
/** * Add a new pet to the store. * Adds a new pet to the store. You may receive an HTTP invalid input if your pet is invalid. * * @param serviceCallback the async ServiceCallback to handle successful and failed responses. * @return the {@link ServiceCall} object */
Add a new pet to the store. Adds a new pet to the store. You may receive an HTTP invalid input if your pet is invalid
addPetAsync
{ "repo_name": "yugangw-msft/autorest", "path": "Samples/petstore/Java/implementation/SwaggerPetstoreImpl.java", "license": "mit", "size": 99949 }
[ "com.microsoft.rest.ServiceCall", "com.microsoft.rest.ServiceCallback" ]
import com.microsoft.rest.ServiceCall; import com.microsoft.rest.ServiceCallback;
import com.microsoft.rest.*;
[ "com.microsoft.rest" ]
com.microsoft.rest;
2,094,067
protected synchronized void saveNobleData() { if ((_nobles == null) || _nobles.isEmpty()) { return; } try (Connection con = DatabaseFactory.getInstance().getConnection()) { for (Entry<Integer, StatsSet> entry : _nobles.entrySet()) { final StatsSet nobleInfo = entry.getValue(); if (nobleInfo == null) { continue; } final int charId = entry.getKey(); final int classId = nobleInfo.getInt(CLASS_ID); final int points = nobleInfo.getInt(POINTS); final int compDone = nobleInfo.getInt(COMP_DONE); final int compWon = nobleInfo.getInt(COMP_WON); final int compLost = nobleInfo.getInt(COMP_LOST); final int compDrawn = nobleInfo.getInt(COMP_DRAWN); final int compDoneWeek = nobleInfo.getInt(COMP_DONE_WEEK); final int compDoneWeekClassed = nobleInfo.getInt(COMP_DONE_WEEK_CLASSED); final int compDoneWeekNonClassed = nobleInfo.getInt(COMP_DONE_WEEK_NON_CLASSED); final int compDoneWeekTeam = nobleInfo.getInt(COMP_DONE_WEEK_TEAM); final boolean toSave = nobleInfo.getBoolean("to_save"); try (PreparedStatement statement = con.prepareStatement(toSave ? OLYMPIAD_SAVE_NOBLES : OLYMPIAD_UPDATE_NOBLES)) { if (toSave) { statement.setInt(1, charId); statement.setInt(2, classId); statement.setInt(3, points); statement.setInt(4, compDone); statement.setInt(5, compWon); statement.setInt(6, compLost); statement.setInt(7, compDrawn); statement.setInt(8, compDoneWeek); statement.setInt(9, compDoneWeekClassed); statement.setInt(10, compDoneWeekNonClassed); statement.setInt(11, compDoneWeekTeam); nobleInfo.set("to_save", false); } else { statement.setInt(1, points); statement.setInt(2, compDone); statement.setInt(3, compWon); statement.setInt(4, compLost); statement.setInt(5, compDrawn); statement.setInt(6, compDoneWeek); statement.setInt(7, compDoneWeekClassed); statement.setInt(8, compDoneWeekNonClassed); statement.setInt(9, compDoneWeekTeam); statement.setInt(10, charId); } statement.execute(); } } } catch (SQLException e) { LOGGER.log(Level.SEVERE, getClass().getSimpleName() + ": Failed to save noblesse data to database: ", e); } }
synchronized void function() { if ((_nobles == null) _nobles.isEmpty()) { return; } try (Connection con = DatabaseFactory.getInstance().getConnection()) { for (Entry<Integer, StatsSet> entry : _nobles.entrySet()) { final StatsSet nobleInfo = entry.getValue(); if (nobleInfo == null) { continue; } final int charId = entry.getKey(); final int classId = nobleInfo.getInt(CLASS_ID); final int points = nobleInfo.getInt(POINTS); final int compDone = nobleInfo.getInt(COMP_DONE); final int compWon = nobleInfo.getInt(COMP_WON); final int compLost = nobleInfo.getInt(COMP_LOST); final int compDrawn = nobleInfo.getInt(COMP_DRAWN); final int compDoneWeek = nobleInfo.getInt(COMP_DONE_WEEK); final int compDoneWeekClassed = nobleInfo.getInt(COMP_DONE_WEEK_CLASSED); final int compDoneWeekNonClassed = nobleInfo.getInt(COMP_DONE_WEEK_NON_CLASSED); final int compDoneWeekTeam = nobleInfo.getInt(COMP_DONE_WEEK_TEAM); final boolean toSave = nobleInfo.getBoolean(STR); try (PreparedStatement statement = con.prepareStatement(toSave ? OLYMPIAD_SAVE_NOBLES : OLYMPIAD_UPDATE_NOBLES)) { if (toSave) { statement.setInt(1, charId); statement.setInt(2, classId); statement.setInt(3, points); statement.setInt(4, compDone); statement.setInt(5, compWon); statement.setInt(6, compLost); statement.setInt(7, compDrawn); statement.setInt(8, compDoneWeek); statement.setInt(9, compDoneWeekClassed); statement.setInt(10, compDoneWeekNonClassed); statement.setInt(11, compDoneWeekTeam); nobleInfo.set(STR, false); } else { statement.setInt(1, points); statement.setInt(2, compDone); statement.setInt(3, compWon); statement.setInt(4, compLost); statement.setInt(5, compDrawn); statement.setInt(6, compDoneWeek); statement.setInt(7, compDoneWeekClassed); statement.setInt(8, compDoneWeekNonClassed); statement.setInt(9, compDoneWeekTeam); statement.setInt(10, charId); } statement.execute(); } } } catch (SQLException e) { LOGGER.log(Level.SEVERE, getClass().getSimpleName() + STR, e); } }
/** * Save noblesse data to database */
Save noblesse data to database
saveNobleData
{ "repo_name": "rubenswagner/L2J-Global", "path": "java/com/l2jglobal/gameserver/model/olympiad/Olympiad.java", "license": "gpl-3.0", "size": 38507 }
[ "com.l2jglobal.commons.database.DatabaseFactory", "com.l2jglobal.gameserver.model.StatsSet", "java.sql.Connection", "java.sql.PreparedStatement", "java.sql.SQLException", "java.util.Map", "java.util.logging.Level" ]
import com.l2jglobal.commons.database.DatabaseFactory; import com.l2jglobal.gameserver.model.StatsSet; import java.sql.Connection; import java.sql.PreparedStatement; import java.sql.SQLException; import java.util.Map; import java.util.logging.Level;
import com.l2jglobal.commons.database.*; import com.l2jglobal.gameserver.model.*; import java.sql.*; import java.util.*; import java.util.logging.*;
[ "com.l2jglobal.commons", "com.l2jglobal.gameserver", "java.sql", "java.util" ]
com.l2jglobal.commons; com.l2jglobal.gameserver; java.sql; java.util;
213,416
L3VpnIfs l3VpnIfs();
L3VpnIfs l3VpnIfs();
/** * Returns the attribute l3VpnIfs. * * @return value of l3VpnIfs */
Returns the attribute l3VpnIfs
l3VpnIfs
{ "repo_name": "mengmoya/onos", "path": "apps/l3vpn/nel3vpn/nemgr/src/main/java/org/onosproject/yang/gen/v1/ne/l3vpn/api/rev20141225/nel3vpnapi/l3vpninstances/L3VpnInstance.java", "license": "apache-2.0", "size": 3370 }
[ "org.onosproject.yang.gen.v1.ne.l3vpn.comm.rev20141225.nel3vpncomm.l3vpnifs.L3VpnIfs" ]
import org.onosproject.yang.gen.v1.ne.l3vpn.comm.rev20141225.nel3vpncomm.l3vpnifs.L3VpnIfs;
import org.onosproject.yang.gen.v1.ne.l3vpn.comm.rev20141225.nel3vpncomm.l3vpnifs.*;
[ "org.onosproject.yang" ]
org.onosproject.yang;
2,586,848
@Test public void testNodeNumbers() throws CloneNotSupportedException { Individual ind = Individual.fromString("(+ x (* x (- (% x x) 1)))", config); config.setMutationProbability(1.0); Individual mutant = PointMutation.mutate(ind, context); assertEquals(ind.getRoot().subtreeToNumberedString(), mutant.getRoot() .subtreeToNumberedString()); }
void function() throws CloneNotSupportedException { Individual ind = Individual.fromString(STR, config); config.setMutationProbability(1.0); Individual mutant = PointMutation.mutate(ind, context); assertEquals(ind.getRoot().subtreeToNumberedString(), mutant.getRoot() .subtreeToNumberedString()); }
/** * Makes sure that all nodes have their numbers set correctly after * mutation. * * @throws CloneNotSupportedException */
Makes sure that all nodes have their numbers set correctly after mutation
testNodeNumbers
{ "repo_name": "burks-pub/gecco2015", "path": "src/test/java/ec/research/gp/simple/operators/PointMutationTest.java", "license": "bsd-2-clause", "size": 4072 }
[ "ec.research.gp.simple.operators.PointMutation", "ec.research.gp.simple.representation.Individual", "org.junit.Assert" ]
import ec.research.gp.simple.operators.PointMutation; import ec.research.gp.simple.representation.Individual; import org.junit.Assert;
import ec.research.gp.simple.operators.*; import ec.research.gp.simple.representation.*; import org.junit.*;
[ "ec.research.gp", "org.junit" ]
ec.research.gp; org.junit;
1,704,685
public void setMembers(SortedMap<Tenor, ExternalId> members) { JodaBeanUtils.notNull(members, "members"); this._members.clear(); this._members.putAll(members); }
void function(SortedMap<Tenor, ExternalId> members) { JodaBeanUtils.notNull(members, STR); this._members.clear(); this._members.putAll(members); }
/** * Sets the members. * @param members the new value of the property, not null */
Sets the members
setMembers
{ "repo_name": "jeorme/OG-Platform", "path": "projects/OG-FinancialTypes/src/main/java/com/opengamma/financial/security/index/IndexFamily.java", "license": "apache-2.0", "size": 6771 }
[ "com.opengamma.id.ExternalId", "com.opengamma.util.time.Tenor", "java.util.SortedMap", "org.joda.beans.JodaBeanUtils" ]
import com.opengamma.id.ExternalId; import com.opengamma.util.time.Tenor; import java.util.SortedMap; import org.joda.beans.JodaBeanUtils;
import com.opengamma.id.*; import com.opengamma.util.time.*; import java.util.*; import org.joda.beans.*;
[ "com.opengamma.id", "com.opengamma.util", "java.util", "org.joda.beans" ]
com.opengamma.id; com.opengamma.util; java.util; org.joda.beans;
2,796,762
public static byte[] sha256hash160(byte[] input) { try { byte[] sha256 = MessageDigest.getInstance("SHA-256").digest(input); RIPEMD160Digest digest = new RIPEMD160Digest(); digest.update(sha256, 0, sha256.length); byte[] out = new byte[20]; digest.doFinal(out, 0); return out; } catch (NoSuchAlgorithmException e) { throw new RuntimeException(e); // Cannot happen. } }
static byte[] function(byte[] input) { try { byte[] sha256 = MessageDigest.getInstance(STR).digest(input); RIPEMD160Digest digest = new RIPEMD160Digest(); digest.update(sha256, 0, sha256.length); byte[] out = new byte[20]; digest.doFinal(out, 0); return out; } catch (NoSuchAlgorithmException e) { throw new RuntimeException(e); } }
/** * Calculates RIPEMD160(SHA256(input)). This is used in Address calculations. */
Calculates RIPEMD160(SHA256(input)). This is used in Address calculations
sha256hash160
{ "repo_name": "pavel4n/wowdoge.org", "path": "core/src/main/java/com/google/dogecoin/core/Utils.java", "license": "apache-2.0", "size": 23380 }
[ "java.security.MessageDigest", "java.security.NoSuchAlgorithmException", "org.spongycastle.crypto.digests.RIPEMD160Digest" ]
import java.security.MessageDigest; import java.security.NoSuchAlgorithmException; import org.spongycastle.crypto.digests.RIPEMD160Digest;
import java.security.*; import org.spongycastle.crypto.digests.*;
[ "java.security", "org.spongycastle.crypto" ]
java.security; org.spongycastle.crypto;
1,160,871
Optional<String> getMachineReadableLog();
Optional<String> getMachineReadableLog();
/** * For more information how MachineReadableLog works check {@link * com.facebook.buck.event.listener.MachineReadableLoggerListener}. * * @return the contents of the machine readable logs. Each line marks a different key to json. */
For more information how MachineReadableLog works check <code>com.facebook.buck.event.listener.MachineReadableLoggerListener</code>
getMachineReadableLog
{ "repo_name": "zpao/buck", "path": "src/com/facebook/buck/doctor/config/DoctorEndpointRequest.java", "license": "apache-2.0", "size": 1502 }
[ "java.util.Optional" ]
import java.util.Optional;
import java.util.*;
[ "java.util" ]
java.util;
1,711,427
@ServiceMethod(returns = ReturnType.COLLECTION) PagedFlux<MetricNamespaceInner> listAsync(String resourceUri, String startTime);
@ServiceMethod(returns = ReturnType.COLLECTION) PagedFlux<MetricNamespaceInner> listAsync(String resourceUri, String startTime);
/** * Lists the metric namespaces for the resource. * * @param resourceUri The identifier of the resource. * @param startTime The ISO 8601 conform Date start time from which to query for metric namespaces. * @throws IllegalArgumentException thrown if parameters fail the validation. * @throws com.azure.core.management.exception.ManagementException thrown if the request is rejected by server. * @throws RuntimeException all other wrapped checked exceptions if the request fails to be sent. * @return represents collection of metric namespaces. */
Lists the metric namespaces for the resource
listAsync
{ "repo_name": "Azure/azure-sdk-for-java", "path": "sdk/resourcemanagerhybrid/azure-resourcemanager-monitor/src/main/java/com/azure/resourcemanager/monitor/fluent/MetricNamespacesClient.java", "license": "mit", "size": 3433 }
[ "com.azure.core.annotation.ReturnType", "com.azure.core.annotation.ServiceMethod", "com.azure.core.http.rest.PagedFlux", "com.azure.resourcemanager.monitor.fluent.models.MetricNamespaceInner" ]
import com.azure.core.annotation.ReturnType; import com.azure.core.annotation.ServiceMethod; import com.azure.core.http.rest.PagedFlux; import com.azure.resourcemanager.monitor.fluent.models.MetricNamespaceInner;
import com.azure.core.annotation.*; import com.azure.core.http.rest.*; import com.azure.resourcemanager.monitor.fluent.models.*;
[ "com.azure.core", "com.azure.resourcemanager" ]
com.azure.core; com.azure.resourcemanager;
1,078,027
public static <T> T resolveReferenceParameter(CamelContext context, String value, Class<T> type) { return resolveReferenceParameter(context, value, type, true); }
static <T> T function(CamelContext context, String value, Class<T> type) { return resolveReferenceParameter(context, value, type, true); }
/** * Resolves a reference parameter by making a lookup in the registry. * * @param <T> type of object to lookup. * @param context Camel context to use for lookup. * @param value reference parameter value. * @param type type of object to lookup. * @return lookup result. * @throws IllegalArgumentException if referenced object was not found in registry. */
Resolves a reference parameter by making a lookup in the registry
resolveReferenceParameter
{ "repo_name": "punkhorn/camel-upstream", "path": "core/camel-support/src/main/java/org/apache/camel/support/EndpointHelper.java", "license": "apache-2.0", "size": 16622 }
[ "org.apache.camel.CamelContext" ]
import org.apache.camel.CamelContext;
import org.apache.camel.*;
[ "org.apache.camel" ]
org.apache.camel;
2,059,823
public SubscriptionPolicy[] getSubscriptionPolicies(int tenantID) throws APIManagementException { List<SubscriptionPolicy> policies = new ArrayList<SubscriptionPolicy>(); Connection conn = null; PreparedStatement ps = null; ResultSet rs = null; String sqlQuery = SQLConstants.GET_SUBSCRIPTION_POLICIES; if (forceCaseInsensitiveComparisons) { sqlQuery = SQLConstants.GET_SUBSCRIPTION_POLICIES; } try { conn = APIMgtDBUtil.getConnection(); ps = conn.prepareStatement(sqlQuery); ps.setInt(1, tenantID); rs = ps.executeQuery(); while (rs.next()) { SubscriptionPolicy subPolicy = new SubscriptionPolicy( rs.getString(ThrottlePolicyConstants.COLUMN_NAME)); setCommonPolicyDetails(subPolicy, rs); subPolicy.setRateLimitCount(rs.getInt(ThrottlePolicyConstants.COLUMN_RATE_LIMIT_COUNT)); subPolicy.setRateLimitTimeUnit(rs.getString(ThrottlePolicyConstants.COLUMN_RATE_LIMIT_TIME_UNIT)); subPolicy.setStopOnQuotaReach(rs.getBoolean(ThrottlePolicyConstants.COLUMN_STOP_ON_QUOTA_REACH)); subPolicy.setBillingPlan(rs.getString(ThrottlePolicyConstants.COLUMN_BILLING_PLAN)); InputStream binary = rs.getBinaryStream(ThrottlePolicyConstants.COLUMN_CUSTOM_ATTRIB); if(binary != null){ byte[] customAttrib = APIUtil.toByteArray(binary); subPolicy.setCustomAttributes(customAttrib); } policies.add(subPolicy); } } catch (SQLException e) { handleException("Error while executing SQL", e); } catch (IOException e) { handleException("Error while converting input stream to byte array", e); } finally { APIMgtDBUtil.closeAllConnections(ps, conn, rs); } return policies.toArray(new SubscriptionPolicy[policies.size()]); }
SubscriptionPolicy[] function(int tenantID) throws APIManagementException { List<SubscriptionPolicy> policies = new ArrayList<SubscriptionPolicy>(); Connection conn = null; PreparedStatement ps = null; ResultSet rs = null; String sqlQuery = SQLConstants.GET_SUBSCRIPTION_POLICIES; if (forceCaseInsensitiveComparisons) { sqlQuery = SQLConstants.GET_SUBSCRIPTION_POLICIES; } try { conn = APIMgtDBUtil.getConnection(); ps = conn.prepareStatement(sqlQuery); ps.setInt(1, tenantID); rs = ps.executeQuery(); while (rs.next()) { SubscriptionPolicy subPolicy = new SubscriptionPolicy( rs.getString(ThrottlePolicyConstants.COLUMN_NAME)); setCommonPolicyDetails(subPolicy, rs); subPolicy.setRateLimitCount(rs.getInt(ThrottlePolicyConstants.COLUMN_RATE_LIMIT_COUNT)); subPolicy.setRateLimitTimeUnit(rs.getString(ThrottlePolicyConstants.COLUMN_RATE_LIMIT_TIME_UNIT)); subPolicy.setStopOnQuotaReach(rs.getBoolean(ThrottlePolicyConstants.COLUMN_STOP_ON_QUOTA_REACH)); subPolicy.setBillingPlan(rs.getString(ThrottlePolicyConstants.COLUMN_BILLING_PLAN)); InputStream binary = rs.getBinaryStream(ThrottlePolicyConstants.COLUMN_CUSTOM_ATTRIB); if(binary != null){ byte[] customAttrib = APIUtil.toByteArray(binary); subPolicy.setCustomAttributes(customAttrib); } policies.add(subPolicy); } } catch (SQLException e) { handleException(STR, e); } catch (IOException e) { handleException(STR, e); } finally { APIMgtDBUtil.closeAllConnections(ps, conn, rs); } return policies.toArray(new SubscriptionPolicy[policies.size()]); }
/** * Get all subscription level policeis belongs to specific tenant * * @param tenantID tenantID filters the polices belongs to specific tenant * @return subscriptionPolicy array list */
Get all subscription level policeis belongs to specific tenant
getSubscriptionPolicies
{ "repo_name": "knPerera/carbon-apimgt", "path": "components/apimgt/org.wso2.carbon.apimgt.impl/src/main/java/org/wso2/carbon/apimgt/impl/dao/ApiMgtDAO.java", "license": "apache-2.0", "size": 493075 }
[ "java.io.IOException", "java.io.InputStream", "java.sql.Connection", "java.sql.PreparedStatement", "java.sql.ResultSet", "java.sql.SQLException", "java.util.ArrayList", "java.util.List", "org.wso2.carbon.apimgt.api.APIManagementException", "org.wso2.carbon.apimgt.api.model.policy.SubscriptionPolicy", "org.wso2.carbon.apimgt.impl.ThrottlePolicyConstants", "org.wso2.carbon.apimgt.impl.dao.constants.SQLConstants", "org.wso2.carbon.apimgt.impl.utils.APIMgtDBUtil", "org.wso2.carbon.apimgt.impl.utils.APIUtil" ]
import java.io.IOException; import java.io.InputStream; import java.sql.Connection; import java.sql.PreparedStatement; import java.sql.ResultSet; import java.sql.SQLException; import java.util.ArrayList; import java.util.List; import org.wso2.carbon.apimgt.api.APIManagementException; import org.wso2.carbon.apimgt.api.model.policy.SubscriptionPolicy; import org.wso2.carbon.apimgt.impl.ThrottlePolicyConstants; import org.wso2.carbon.apimgt.impl.dao.constants.SQLConstants; import org.wso2.carbon.apimgt.impl.utils.APIMgtDBUtil; import org.wso2.carbon.apimgt.impl.utils.APIUtil;
import java.io.*; import java.sql.*; import java.util.*; import org.wso2.carbon.apimgt.api.*; import org.wso2.carbon.apimgt.api.model.policy.*; import org.wso2.carbon.apimgt.impl.*; import org.wso2.carbon.apimgt.impl.dao.constants.*; import org.wso2.carbon.apimgt.impl.utils.*;
[ "java.io", "java.sql", "java.util", "org.wso2.carbon" ]
java.io; java.sql; java.util; org.wso2.carbon;
1,052,160
ClassLoader getClassLoader(String componentName, String componentVersion, ComponentReference componentReference);
ClassLoader getClassLoader(String componentName, String componentVersion, ComponentReference componentReference);
/** * Returns class loader of the given component. * * @param componentName full component name * @param componentVersion component version * @param componentReference component reference * @return class loader for specified component */
Returns class loader of the given component
getClassLoader
{ "repo_name": "rasika90/carbon-uuf", "path": "components/uuf-core/src/main/java/org/wso2/carbon/uuf/internal/core/create/ClassLoaderProvider.java", "license": "apache-2.0", "size": 1200 }
[ "org.wso2.carbon.uuf.reference.ComponentReference" ]
import org.wso2.carbon.uuf.reference.ComponentReference;
import org.wso2.carbon.uuf.reference.*;
[ "org.wso2.carbon" ]
org.wso2.carbon;
1,870,300
static double[] yearFractions(List<Period> periods, ZonedDateTime valuationTime) { double[] yearFractions = new double[periods.size()]; int index = 0; for (Period period : periods) { yearFractions[index++] = TimeCalculator.getTimeBetween(valuationTime, valuationTime.plus(period)); } return yearFractions; }
static double[] yearFractions(List<Period> periods, ZonedDateTime valuationTime) { double[] yearFractions = new double[periods.size()]; int index = 0; for (Period period : periods) { yearFractions[index++] = TimeCalculator.getTimeBetween(valuationTime, valuationTime.plus(period)); } return yearFractions; }
/** * Converts the input values into an array of year fractions for passing to the analytics. * If periods is a double array it is returned unchanged. If it's a list of periods then the year fraction * is calculated for each period relative to the valuation time (using {@link TimeCalculator}). * @param periods The input axis values, a list of periods or a double array * @param valuationTime The valuation time * @return An array of year fractions for the axis values */
Converts the input values into an array of year fractions for passing to the analytics. If periods is a double array it is returned unchanged. If it's a list of periods then the year fraction is calculated for each period relative to the valuation time (using <code>TimeCalculator</code>)
yearFractions
{ "repo_name": "DevStreet/FinanceAnalytics", "path": "projects/OG-Integration/src/main/java/com/opengamma/integration/marketdata/manipulator/dsl/VolatilitySurfaceShiftManipulator.java", "license": "apache-2.0", "size": 19859 }
[ "com.opengamma.analytics.util.time.TimeCalculator", "java.util.List", "org.threeten.bp.Period", "org.threeten.bp.ZonedDateTime" ]
import com.opengamma.analytics.util.time.TimeCalculator; import java.util.List; import org.threeten.bp.Period; import org.threeten.bp.ZonedDateTime;
import com.opengamma.analytics.util.time.*; import java.util.*; import org.threeten.bp.*;
[ "com.opengamma.analytics", "java.util", "org.threeten.bp" ]
com.opengamma.analytics; java.util; org.threeten.bp;
952,722
@Test public void testPersistentPRWithGatewaySenderPersistenceEnabled_Restart2() { // create locator on local site Integer lnPort = (Integer) vm0.invoke(() -> WANTestBase.createFirstLocatorWithDSId(1)); // create locator on remote site Integer nyPort = (Integer) vm1.invoke(() -> WANTestBase.createFirstRemoteLocator(2, lnPort)); // create cache in local site createCacheInVMs(lnPort, vm4, vm5, vm6, vm7); // create senders with disk store String diskStore1 = (String) vm4.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2, true, 100, 10, false, true, null, null, false)); String diskStore2 = (String) vm5.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2, true, 100, 10, false, true, null, null, false)); String diskStore3 = (String) vm6.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2, true, 100, 10, false, true, null, null, false)); String diskStore4 = (String) vm7.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2, true, 100, 10, false, true, null, null, false)); LogWriterUtils.getLogWriter() .info("The DS are: " + diskStore1 + "," + diskStore2 + "," + diskStore3 + "," + diskStore4); // create PR on local site vm4.invoke(() -> WANTestBase.createPersistentPartitionedRegion(getTestMethodName(), "ln", 1, 100, isOffHeap())); vm5.invoke(() -> WANTestBase.createPersistentPartitionedRegion(getTestMethodName(), "ln", 1, 100, isOffHeap())); vm6.invoke(() -> WANTestBase.createPersistentPartitionedRegion(getTestMethodName(), "ln", 1, 100, isOffHeap())); vm7.invoke(() -> WANTestBase.createPersistentPartitionedRegion(getTestMethodName(), "ln", 1, 100, isOffHeap())); // start the senders on local site startSenderInVMs("ln", vm4, vm5, vm6, vm7); // wait for senders to become running vm4.invoke(waitForSenderRunnable()); vm5.invoke(waitForSenderRunnable()); vm6.invoke(waitForSenderRunnable()); vm7.invoke(waitForSenderRunnable()); // pause the senders vm4.invoke(pauseSenderRunnable()); vm5.invoke(pauseSenderRunnable()); vm6.invoke(pauseSenderRunnable()); vm7.invoke(pauseSenderRunnable()); // start puts in region on local site vm4.invoke(() -> WANTestBase.doPuts(getTestMethodName(), 300)); LogWriterUtils.getLogWriter().info("Completed puts in the region"); // --------------------close and rebuild local site // ------------------------------------------------- // kill the senders vm4.invoke(killSenderRunnable()); vm5.invoke(killSenderRunnable()); vm6.invoke(killSenderRunnable()); vm7.invoke(killSenderRunnable()); LogWriterUtils.getLogWriter().info("Killed all the senders."); // restart the vm createCacheInVMs(lnPort, vm4, vm5, vm6, vm7); LogWriterUtils.getLogWriter().info("Created back the cache"); // create senders with disk store vm4.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2, true, 100, 10, false, true, null, diskStore1, true)); vm5.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2, true, 100, 10, false, true, null, diskStore2, true)); vm6.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2, true, 100, 10, false, true, null, diskStore3, true)); vm7.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2, true, 100, 10, false, true, null, diskStore4, true)); LogWriterUtils.getLogWriter().info("Created the senders back from the disk store."); // create PR on local site AsyncInvocation inv1 = vm4.invokeAsync(() -> WANTestBase .createPersistentPartitionedRegion(getTestMethodName(), "ln", 1, 100, isOffHeap())); AsyncInvocation inv2 = vm5.invokeAsync(() -> WANTestBase .createPersistentPartitionedRegion(getTestMethodName(), "ln", 1, 100, isOffHeap())); AsyncInvocation inv3 = vm6.invokeAsync(() -> WANTestBase .createPersistentPartitionedRegion(getTestMethodName(), "ln", 1, 100, isOffHeap())); AsyncInvocation inv4 = vm7.invokeAsync(() -> WANTestBase .createPersistentPartitionedRegion(getTestMethodName(), "ln", 1, 100, isOffHeap())); try { inv1.join(); inv2.join(); inv3.join(); inv4.join(); } catch (InterruptedException e) { e.printStackTrace(); fail(); } LogWriterUtils.getLogWriter().info("Created back the partitioned regions"); // start the senders in async mode. This will ensure that the // node of shadow PR that went down last will come up first startSenderInVMsAsync("ln", vm4, vm5, vm6, vm7); LogWriterUtils.getLogWriter().info("Waiting for senders running."); // wait for senders running vm4.invoke(waitForSenderRunnable()); vm5.invoke(waitForSenderRunnable()); vm6.invoke(waitForSenderRunnable()); vm7.invoke(waitForSenderRunnable()); LogWriterUtils.getLogWriter().info("Creating the receiver."); // create receiver on remote site createCacheInVMs(nyPort, vm2, vm3); // create PR on remote site LogWriterUtils.getLogWriter().info("Creating the partitioned region at receiver. "); vm2.invoke(() -> WANTestBase.createPersistentPartitionedRegion(getTestMethodName(), null, 1, 100, isOffHeap())); vm3.invoke(() -> WANTestBase.createPersistentPartitionedRegion(getTestMethodName(), null, 1, 100, isOffHeap())); createReceiverInVMs(vm2, vm3); vm4.invoke(pauseSenderRunnable()); vm5.invoke(pauseSenderRunnable()); vm6.invoke(pauseSenderRunnable()); vm7.invoke(pauseSenderRunnable()); LogWriterUtils.getLogWriter().info("Doing some extra puts. "); // start puts in region on local site vm4.invoke(() -> WANTestBase.doPutsAfter300(getTestMethodName(), 1000)); // ---------------------------------------------------------------------------------------------------- vm4.invoke(() -> WANTestBase.resumeSender("ln")); vm5.invoke(() -> WANTestBase.resumeSender("ln")); vm6.invoke(() -> WANTestBase.resumeSender("ln")); vm7.invoke(() -> WANTestBase.resumeSender("ln")); LogWriterUtils.getLogWriter().info("Validating the region size at the receiver end. "); vm2.invoke(() -> WANTestBase.validateRegionSize(getTestMethodName(), 1000)); vm3.invoke(() -> WANTestBase.validateRegionSize(getTestMethodName(), 1000)); }
void function() { Integer lnPort = (Integer) vm0.invoke(() -> WANTestBase.createFirstLocatorWithDSId(1)); Integer nyPort = (Integer) vm1.invoke(() -> WANTestBase.createFirstRemoteLocator(2, lnPort)); createCacheInVMs(lnPort, vm4, vm5, vm6, vm7); String diskStore1 = (String) vm4.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2, true, 100, 10, false, true, null, null, false)); String diskStore2 = (String) vm5.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2, true, 100, 10, false, true, null, null, false)); String diskStore3 = (String) vm6.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2, true, 100, 10, false, true, null, null, false)); String diskStore4 = (String) vm7.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2, true, 100, 10, false, true, null, null, false)); LogWriterUtils.getLogWriter() .info(STR + diskStore1 + "," + diskStore2 + "," + diskStore3 + "," + diskStore4); vm4.invoke(() -> WANTestBase.createPersistentPartitionedRegion(getTestMethodName(), "ln", 1, 100, isOffHeap())); vm5.invoke(() -> WANTestBase.createPersistentPartitionedRegion(getTestMethodName(), "ln", 1, 100, isOffHeap())); vm6.invoke(() -> WANTestBase.createPersistentPartitionedRegion(getTestMethodName(), "ln", 1, 100, isOffHeap())); vm7.invoke(() -> WANTestBase.createPersistentPartitionedRegion(getTestMethodName(), "ln", 1, 100, isOffHeap())); startSenderInVMs("ln", vm4, vm5, vm6, vm7); vm4.invoke(waitForSenderRunnable()); vm5.invoke(waitForSenderRunnable()); vm6.invoke(waitForSenderRunnable()); vm7.invoke(waitForSenderRunnable()); vm4.invoke(pauseSenderRunnable()); vm5.invoke(pauseSenderRunnable()); vm6.invoke(pauseSenderRunnable()); vm7.invoke(pauseSenderRunnable()); vm4.invoke(() -> WANTestBase.doPuts(getTestMethodName(), 300)); LogWriterUtils.getLogWriter().info(STR); vm4.invoke(killSenderRunnable()); vm5.invoke(killSenderRunnable()); vm6.invoke(killSenderRunnable()); vm7.invoke(killSenderRunnable()); LogWriterUtils.getLogWriter().info(STR); createCacheInVMs(lnPort, vm4, vm5, vm6, vm7); LogWriterUtils.getLogWriter().info(STR); vm4.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2, true, 100, 10, false, true, null, diskStore1, true)); vm5.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2, true, 100, 10, false, true, null, diskStore2, true)); vm6.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2, true, 100, 10, false, true, null, diskStore3, true)); vm7.invoke(() -> WANTestBase.createSenderWithDiskStore("ln", 2, true, 100, 10, false, true, null, diskStore4, true)); LogWriterUtils.getLogWriter().info(STR); AsyncInvocation inv1 = vm4.invokeAsync(() -> WANTestBase .createPersistentPartitionedRegion(getTestMethodName(), "ln", 1, 100, isOffHeap())); AsyncInvocation inv2 = vm5.invokeAsync(() -> WANTestBase .createPersistentPartitionedRegion(getTestMethodName(), "ln", 1, 100, isOffHeap())); AsyncInvocation inv3 = vm6.invokeAsync(() -> WANTestBase .createPersistentPartitionedRegion(getTestMethodName(), "ln", 1, 100, isOffHeap())); AsyncInvocation inv4 = vm7.invokeAsync(() -> WANTestBase .createPersistentPartitionedRegion(getTestMethodName(), "ln", 1, 100, isOffHeap())); try { inv1.join(); inv2.join(); inv3.join(); inv4.join(); } catch (InterruptedException e) { e.printStackTrace(); fail(); } LogWriterUtils.getLogWriter().info(STR); startSenderInVMsAsync("ln", vm4, vm5, vm6, vm7); LogWriterUtils.getLogWriter().info(STR); vm4.invoke(waitForSenderRunnable()); vm5.invoke(waitForSenderRunnable()); vm6.invoke(waitForSenderRunnable()); vm7.invoke(waitForSenderRunnable()); LogWriterUtils.getLogWriter().info(STR); createCacheInVMs(nyPort, vm2, vm3); LogWriterUtils.getLogWriter().info(STR); vm2.invoke(() -> WANTestBase.createPersistentPartitionedRegion(getTestMethodName(), null, 1, 100, isOffHeap())); vm3.invoke(() -> WANTestBase.createPersistentPartitionedRegion(getTestMethodName(), null, 1, 100, isOffHeap())); createReceiverInVMs(vm2, vm3); vm4.invoke(pauseSenderRunnable()); vm5.invoke(pauseSenderRunnable()); vm6.invoke(pauseSenderRunnable()); vm7.invoke(pauseSenderRunnable()); LogWriterUtils.getLogWriter().info(STR); vm4.invoke(() -> WANTestBase.doPutsAfter300(getTestMethodName(), 1000)); vm4.invoke(() -> WANTestBase.resumeSender("ln")); vm5.invoke(() -> WANTestBase.resumeSender("ln")); vm6.invoke(() -> WANTestBase.resumeSender("ln")); vm7.invoke(() -> WANTestBase.resumeSender("ln")); LogWriterUtils.getLogWriter().info(STR); vm2.invoke(() -> WANTestBase.validateRegionSize(getTestMethodName(), 1000)); vm3.invoke(() -> WANTestBase.validateRegionSize(getTestMethodName(), 1000)); }
/** * Enable persistence for PR and GatewaySender. Pause the sender and do some puts in local region. * Close the local site and rebuild the region and sender from disk store. Dispatcher should not * start dispatching events recovered from persistent sender. Check if the remote site receives * all the events. */
Enable persistence for PR and GatewaySender. Pause the sender and do some puts in local region. Close the local site and rebuild the region and sender from disk store. Dispatcher should not start dispatching events recovered from persistent sender. Check if the remote site receives all the events
testPersistentPRWithGatewaySenderPersistenceEnabled_Restart2
{ "repo_name": "pdxrunner/geode", "path": "geode-wan/src/distributedTest/java/org/apache/geode/internal/cache/wan/parallel/ParallelWANPersistenceEnabledGatewaySenderDUnitTest.java", "license": "apache-2.0", "size": 70917 }
[ "org.apache.geode.internal.cache.wan.WANTestBase", "org.apache.geode.test.dunit.AsyncInvocation", "org.apache.geode.test.dunit.LogWriterUtils", "org.junit.Assert" ]
import org.apache.geode.internal.cache.wan.WANTestBase; import org.apache.geode.test.dunit.AsyncInvocation; import org.apache.geode.test.dunit.LogWriterUtils; import org.junit.Assert;
import org.apache.geode.internal.cache.wan.*; import org.apache.geode.test.dunit.*; import org.junit.*;
[ "org.apache.geode", "org.junit" ]
org.apache.geode; org.junit;
2,704,928
public NodeConnectorStatistics readNodeConnector(String container, NodeConnector nodeConnector, boolean cached);
NodeConnectorStatistics function(String container, NodeConnector nodeConnector, boolean cached);
/** * Returns the hardware view of the specified network node connector * for the given container * @param node * @return */
Returns the hardware view of the specified network node connector for the given container
readNodeConnector
{ "repo_name": "lbchen/ODL", "path": "opendaylight/protocol_plugins/openflow/src/main/java/org/opendaylight/controller/protocol_plugin/openflow/IReadServiceFilter.java", "license": "epl-1.0", "size": 3448 }
[ "org.opendaylight.controller.sal.core.NodeConnector", "org.opendaylight.controller.sal.reader.NodeConnectorStatistics" ]
import org.opendaylight.controller.sal.core.NodeConnector; import org.opendaylight.controller.sal.reader.NodeConnectorStatistics;
import org.opendaylight.controller.sal.core.*; import org.opendaylight.controller.sal.reader.*;
[ "org.opendaylight.controller" ]
org.opendaylight.controller;
1,266,446
public void pushToSource(Bundle bundle) { mPipeSource.push(bundle); }
void function(Bundle bundle) { mPipeSource.push(bundle); }
/** * Push a {@link dk.itu.spcl.jlpf.common.Bundle} object into the source to be processed by the filters * * @param bundle Wrapped object to be processed */
Push a <code>dk.itu.spcl.jlpf.common.Bundle</code> object into the source to be processed by the filters
pushToSource
{ "repo_name": "centosGit/JLPF", "path": "JLPF/src/main/java/dk/itu/spcl/jlpf/core/Computable.java", "license": "gpl-3.0", "size": 5177 }
[ "dk.itu.spcl.jlpf.common.Bundle" ]
import dk.itu.spcl.jlpf.common.Bundle;
import dk.itu.spcl.jlpf.common.*;
[ "dk.itu.spcl" ]
dk.itu.spcl;
845,200
public void testNonExistentConfigFile() { ThreadPoolManager.setPropsFileName( "somefilethatdoesntexist" ); ThreadPoolManager mgr = ThreadPoolManager.getInstance(); assertNotNull( mgr ); ThreadPoolExecutor pool = mgr.getPool( "doesntexist" ); assertNotNull( "Should have gotten back a pool configured like the default", pool ); int max = pool.getMaximumPoolSize(); System.out.println( max ); // it will load from the default file int expected = Integer.parseInt( PropertyLoader.loadProperties( "cache.ccf" ) .getProperty( "thread_pool.default.maximumPoolSize" ) ); // "Max should be " + expected", assertEquals( max, expected ); }
void function() { ThreadPoolManager.setPropsFileName( STR ); ThreadPoolManager mgr = ThreadPoolManager.getInstance(); assertNotNull( mgr ); ThreadPoolExecutor pool = mgr.getPool( STR ); assertNotNull( STR, pool ); int max = pool.getMaximumPoolSize(); System.out.println( max ); int expected = Integer.parseInt( PropertyLoader.loadProperties( STR ) .getProperty( STR ) ); assertEquals( max, expected ); }
/** * Makes ure we can get a non existent pool from the non exitent config * file. */
Makes ure we can get a non existent pool from the non exitent config file
testNonExistentConfigFile
{ "repo_name": "tikue/jcs2-snapshot", "path": "src/test/org/apache/commons/jcs/utils/threadpool/ThreadPoolManagerUnitTest.java", "license": "apache-2.0", "size": 7126 }
[ "java.util.concurrent.ThreadPoolExecutor", "org.apache.commons.jcs.utils.props.PropertyLoader" ]
import java.util.concurrent.ThreadPoolExecutor; import org.apache.commons.jcs.utils.props.PropertyLoader;
import java.util.concurrent.*; import org.apache.commons.jcs.utils.props.*;
[ "java.util", "org.apache.commons" ]
java.util; org.apache.commons;
501,554
protected DBObject augmentObjectCacheKey(DBObject key) { return key; }
DBObject function(DBObject key) { return key; }
/** * This method allows subclasses an opportunity to add any important data to a cache key, such * as the __acl from the AccessControlMongoDAO. * * TODO: revisit these methods... * * @param key key to augment * @return augmented key */
This method allows subclasses an opportunity to add any important data to a cache key, such as the __acl from the AccessControlMongoDAO
augmentObjectCacheKey
{ "repo_name": "jeppetto/jeppetto", "path": "jeppetto-dao-mongo/src/main/java/org/iternine/jeppetto/dao/mongodb/MongoDBQueryModelDAO.java", "license": "apache-2.0", "size": 51971 }
[ "com.mongodb.DBObject" ]
import com.mongodb.DBObject;
import com.mongodb.*;
[ "com.mongodb" ]
com.mongodb;
836,143
private ByteBuffer encodePrincipalName( short version, String principalName ) { String[] split = principalName.split( "@" ); String nameComponentPart = split[0]; String realm = split[1]; String[] nameComponents = nameComponentPart.split( "/" ); // Compute the size of the buffer List<byte[]> strings = new ArrayList<byte[]>(); // Initialize the size with the number of components' size int size = 2; size += encodeCountedString( strings, realm ); // compute NameComponents for ( String nameComponent : nameComponents ) { size += encodeCountedString( strings, nameComponent ); } ByteBuffer buffer = ByteBuffer.allocate( size ); // Now, write the data into the buffer // store the numComponents if ( version == Keytab.VERSION_0X501 ) { // increment for version 0x0501 buffer.putShort( ( short ) ( nameComponents.length + 1 ) ); } else { // Version = OxO502 buffer.putShort( ( short ) ( nameComponents.length ) ); } // Store the realm and the nameComponents for ( byte[] string : strings ) { buffer.putShort( ( short ) ( string.length ) ); buffer.put( string ); } buffer.flip(); return buffer; }
ByteBuffer function( short version, String principalName ) { String[] split = principalName.split( "@" ); String nameComponentPart = split[0]; String realm = split[1]; String[] nameComponents = nameComponentPart.split( "/" ); List<byte[]> strings = new ArrayList<byte[]>(); int size = 2; size += encodeCountedString( strings, realm ); for ( String nameComponent : nameComponents ) { size += encodeCountedString( strings, nameComponent ); } ByteBuffer buffer = ByteBuffer.allocate( size ); if ( version == Keytab.VERSION_0X501 ) { buffer.putShort( ( short ) ( nameComponents.length + 1 ) ); } else { buffer.putShort( ( short ) ( nameComponents.length ) ); } for ( byte[] string : strings ) { buffer.putShort( ( short ) ( string.length ) ); buffer.put( string ); } buffer.flip(); return buffer; }
/** * Encode a principal name. * * @param buffer * @param principalName */
Encode a principal name
encodePrincipalName
{ "repo_name": "darranl/directory-server", "path": "kerberos-codec/src/main/java/org/apache/directory/server/kerberos/shared/keytab/KeytabEncoder.java", "license": "apache-2.0", "size": 7820 }
[ "java.nio.ByteBuffer", "java.util.ArrayList", "java.util.List" ]
import java.nio.ByteBuffer; import java.util.ArrayList; import java.util.List;
import java.nio.*; import java.util.*;
[ "java.nio", "java.util" ]
java.nio; java.util;
235,611
public BirthmarkService getProvider();
BirthmarkService function();
/** * returns service provider interface of this extractor. */
returns service provider interface of this extractor
getProvider
{ "repo_name": "tamada/stigmata", "path": "src/main/java/com/github/stigmata/BirthmarkExtractor.java", "license": "apache-2.0", "size": 2109 }
[ "com.github.stigmata.spi.BirthmarkService" ]
import com.github.stigmata.spi.BirthmarkService;
import com.github.stigmata.spi.*;
[ "com.github.stigmata" ]
com.github.stigmata;
72,522
int j, J, p, q, px, top, n, Gp[], Gi[], Bp[], Bi[]; double Gx[], Bx[]; if (!Dcs_util.CS_CSC(G) || !Dcs_util.CS_CSC(B) || xi == null || x == null) return (-1); Gp = G.p; Gi = G.i; Gx = G.x; n = G.n; Bp = B.p; Bi = B.i; Bx = B.x; top = Dcs_reach.cs_reach(G, B, k, xi, pinv); for (p = top; p < n; p++) x[xi[p]] = 0; for (p = Bp[k]; p < Bp[k + 1]; p++) x[Bi[p]] = Bx[p]; for (px = top; px < n; px++) { j = xi[px]; J = pinv != null ? (pinv[j]) : j; if (J < 0) continue; x[j] /= Gx[lo ? (Gp[J]) : (Gp[J + 1] - 1)]; p = lo ? (Gp[J] + 1) : (Gp[J]); q = lo ? (Gp[J + 1]) : (Gp[J + 1] - 1); for (; p < q; p++) { x[Gi[p]] -= Gx[p] * x[j]; } } return (top); }
int j, J, p, q, px, top, n, Gp[], Gi[], Bp[], Bi[]; double Gx[], Bx[]; if (!Dcs_util.CS_CSC(G) !Dcs_util.CS_CSC(B) xi == null x == null) return (-1); Gp = G.p; Gi = G.i; Gx = G.x; n = G.n; Bp = B.p; Bi = B.i; Bx = B.x; top = Dcs_reach.cs_reach(G, B, k, xi, pinv); for (p = top; p < n; p++) x[xi[p]] = 0; for (p = Bp[k]; p < Bp[k + 1]; p++) x[Bi[p]] = Bx[p]; for (px = top; px < n; px++) { j = xi[px]; J = pinv != null ? (pinv[j]) : j; if (J < 0) continue; x[j] /= Gx[lo ? (Gp[J]) : (Gp[J + 1] - 1)]; p = lo ? (Gp[J] + 1) : (Gp[J]); q = lo ? (Gp[J + 1]) : (Gp[J + 1] - 1); for (; p < q; p++) { x[Gi[p]] -= Gx[p] * x[j]; } } return (top); }
/** * solve Gx=b(:,k), where G is either upper (lo=false) or lower (lo=true) * triangular. * * @param G * lower or upper triangular matrix in column-compressed form * @param B * right hand side, b=B(:,k) * @param k * use kth column of B as right hand side * @param xi * size 2*n, nonzero pattern of x in xi[top..n-1] * @param x * size n, x in x[xi[top..n-1]] * @param pinv * mapping of rows to columns of G, ignored if null * @param lo * true if lower triangular, false if upper * @return top, -1 in error */
solve Gx=b(:,k), where G is either upper (lo=false) or lower (lo=true) triangular
cs_spsolve
{ "repo_name": "rwl/CSparseJ", "path": "src/main/java/edu/emory/mathcs/csparsej/tdouble/Dcs_spsolve.java", "license": "lgpl-2.1", "size": 3425 }
[ "edu.emory.mathcs.csparsej.tdouble.Dcs_common" ]
import edu.emory.mathcs.csparsej.tdouble.Dcs_common;
import edu.emory.mathcs.csparsej.tdouble.*;
[ "edu.emory.mathcs" ]
edu.emory.mathcs;
1,965,594
WindowAndroid getWindowAndroid();
WindowAndroid getWindowAndroid();
/** * Gets a {@link WindowAndroid} which is passed on to {@link ReparentingTask}, used in the * reparenting process. */
Gets a <code>WindowAndroid</code> which is passed on to <code>ReparentingTask</code>, used in the reparenting process
getWindowAndroid
{ "repo_name": "endlessm/chromium-browser", "path": "chrome/android/java/src/org/chromium/chrome/browser/tab_activity_glue/ReparentingTask.java", "license": "bsd-3-clause", "size": 7924 }
[ "org.chromium.ui.base.WindowAndroid" ]
import org.chromium.ui.base.WindowAndroid;
import org.chromium.ui.base.*;
[ "org.chromium.ui" ]
org.chromium.ui;
2,660,630
private void setupNameEnumerator() { try { _handle = NDNHandle.open(); _nameEnumerator = new NDNNameEnumerator(_handle, this); } catch (ConfigurationException e) { Log.warningStackTrace(e); } catch (IOException e) { Log.warningStackTrace(e); } }
void function() { try { _handle = NDNHandle.open(); _nameEnumerator = new NDNNameEnumerator(_handle, this); } catch (ConfigurationException e) { Log.warningStackTrace(e); } catch (IOException e) { Log.warningStackTrace(e); } }
/** * Method to get an instance of a NDNHandle and NDNNameEnumerator. * * @return void */
Method to get an instance of a NDNHandle and NDNNameEnumerator
setupNameEnumerator
{ "repo_name": "gujianxiao/gatewayForMulticom", "path": "javasrc/src/main/org/ndnx/ndn/utils/explorer/ContentExplorer.java", "license": "lgpl-2.1", "size": 39307 }
[ "java.io.IOException", "org.ndnx.ndn.NDNHandle", "org.ndnx.ndn.config.ConfigurationException", "org.ndnx.ndn.impl.support.Log", "org.ndnx.ndn.profiles.nameenum.NDNNameEnumerator" ]
import java.io.IOException; import org.ndnx.ndn.NDNHandle; import org.ndnx.ndn.config.ConfigurationException; import org.ndnx.ndn.impl.support.Log; import org.ndnx.ndn.profiles.nameenum.NDNNameEnumerator;
import java.io.*; import org.ndnx.ndn.*; import org.ndnx.ndn.config.*; import org.ndnx.ndn.impl.support.*; import org.ndnx.ndn.profiles.nameenum.*;
[ "java.io", "org.ndnx.ndn" ]
java.io; org.ndnx.ndn;
1,184,338
@Override public negative_factors_trend remove(Serializable primaryKey) throws NoSuchnegative_factors_trendException, SystemException { Session session = null; try { session = openSession(); negative_factors_trend negative_factors_trend = (negative_factors_trend)session.get(negative_factors_trendImpl.class, primaryKey); if (negative_factors_trend == null) { if (_log.isWarnEnabled()) { _log.warn(_NO_SUCH_ENTITY_WITH_PRIMARY_KEY + primaryKey); } throw new NoSuchnegative_factors_trendException(_NO_SUCH_ENTITY_WITH_PRIMARY_KEY + primaryKey); } return remove(negative_factors_trend); } catch (NoSuchnegative_factors_trendException nsee) { throw nsee; } catch (Exception e) { throw processException(e); } finally { closeSession(session); } }
negative_factors_trend function(Serializable primaryKey) throws NoSuchnegative_factors_trendException, SystemException { Session session = null; try { session = openSession(); negative_factors_trend negative_factors_trend = (negative_factors_trend)session.get(negative_factors_trendImpl.class, primaryKey); if (negative_factors_trend == null) { if (_log.isWarnEnabled()) { _log.warn(_NO_SUCH_ENTITY_WITH_PRIMARY_KEY + primaryKey); } throw new NoSuchnegative_factors_trendException(_NO_SUCH_ENTITY_WITH_PRIMARY_KEY + primaryKey); } return remove(negative_factors_trend); } catch (NoSuchnegative_factors_trendException nsee) { throw nsee; } catch (Exception e) { throw processException(e); } finally { closeSession(session); } }
/** * Removes the negative_factors_trend with the primary key from the database. Also notifies the appropriate model listeners. * * @param primaryKey the primary key of the negative_factors_trend * @return the negative_factors_trend that was removed * @throws com.iucn.whp.dbservice.NoSuchnegative_factors_trendException if a negative_factors_trend with the primary key could not be found * @throws SystemException if a system exception occurred */
Removes the negative_factors_trend with the primary key from the database. Also notifies the appropriate model listeners
remove
{ "repo_name": "iucn-whp/world-heritage-outlook", "path": "portlets/iucn-dbservice-portlet/docroot/WEB-INF/src/com/iucn/whp/dbservice/service/persistence/negative_factors_trendPersistenceImpl.java", "license": "gpl-2.0", "size": 36424 }
[ "com.liferay.portal.kernel.dao.orm.Session", "com.liferay.portal.kernel.exception.SystemException", "java.io.Serializable" ]
import com.liferay.portal.kernel.dao.orm.Session; import com.liferay.portal.kernel.exception.SystemException; import java.io.Serializable;
import com.liferay.portal.kernel.dao.orm.*; import com.liferay.portal.kernel.exception.*; import java.io.*;
[ "com.liferay.portal", "java.io" ]
com.liferay.portal; java.io;
2,347,355
public boolean allowSetPassword(Context context, HttpServletRequest request, String username) throws SQLException { // XXX is this right? return false; }
boolean function(Context context, HttpServletRequest request, String username) throws SQLException { return false; }
/** * Cannot change LDAP password through dspace, right? */
Cannot change LDAP password through dspace, right
allowSetPassword
{ "repo_name": "rnathanday/dryad-repo", "path": "dspace-api/src/main/java/org/dspace/authenticate/LDAPAuthentication.java", "license": "bsd-3-clause", "size": 17159 }
[ "java.sql.SQLException", "javax.servlet.http.HttpServletRequest", "org.dspace.core.Context" ]
import java.sql.SQLException; import javax.servlet.http.HttpServletRequest; import org.dspace.core.Context;
import java.sql.*; import javax.servlet.http.*; import org.dspace.core.*;
[ "java.sql", "javax.servlet", "org.dspace.core" ]
java.sql; javax.servlet; org.dspace.core;
1,967,101
//----------// // getPeaks // //----------// public List<PeakEntry<Double>> getDoublePeaks (int minCount) { final List<PeakEntry<Double>> peaks = new ArrayList<>(); K start = null; K stop = null; K best = null; Integer bestCount = null; boolean isAbove = false; for (Entry<K, Integer> entry : map.entrySet()) { if (entry.getValue() >= minCount) { if ((bestCount == null) || (bestCount < entry.getValue())) { best = entry.getKey(); bestCount = entry.getValue(); } if (isAbove) { // Above -> Above stop = entry.getKey(); } else { // Below -> Above stop = start = entry.getKey(); isAbove = true; } } else { if (isAbove) { // Above -> Below peaks.add( new PeakEntry<>( createDoublePeak(start, best, stop, minCount), (double) bestCount / totalCount)); stop = start = best = null; bestCount = null; isAbove = false; } else { // Below -> Below } } } // Last range if (isAbove) { peaks.add( new PeakEntry<>( createDoublePeak(start, best, stop, minCount), (double) bestCount / totalCount)); } // Sort by decreasing count values Collections.sort(peaks, reverseDoublePeakComparator); return peaks; }
List<PeakEntry<Double>> function (int minCount) { final List<PeakEntry<Double>> peaks = new ArrayList<>(); K start = null; K stop = null; K best = null; Integer bestCount = null; boolean isAbove = false; for (Entry<K, Integer> entry : map.entrySet()) { if (entry.getValue() >= minCount) { if ((bestCount == null) (bestCount < entry.getValue())) { best = entry.getKey(); bestCount = entry.getValue(); } if (isAbove) { stop = entry.getKey(); } else { stop = start = entry.getKey(); isAbove = true; } } else { if (isAbove) { peaks.add( new PeakEntry<>( createDoublePeak(start, best, stop, minCount), (double) bestCount / totalCount)); stop = start = best = null; bestCount = null; isAbove = false; } else { } } } if (isAbove) { peaks.add( new PeakEntry<>( createDoublePeak(start, best, stop, minCount), (double) bestCount / totalCount)); } Collections.sort(peaks, reverseDoublePeakComparator); return peaks; }
/** * Report the sequence of bucket peaks whose count is equal to or * greater than the specified minCount value. * * @param minCount the desired minimum count value * @return the (perhaps empty but not null) sequence of peaks of buckets */
Report the sequence of bucket peaks whose count is equal to or greater than the specified minCount value
getDoublePeaks
{ "repo_name": "jlpoolen/libreveris", "path": "src/main/omr/math/Histogram.java", "license": "lgpl-3.0", "size": 21357 }
[ "java.util.ArrayList", "java.util.Collections", "java.util.List", "java.util.Map" ]
import java.util.ArrayList; import java.util.Collections; import java.util.List; import java.util.Map;
import java.util.*;
[ "java.util" ]
java.util;
547,218
FactoryFinderResolver getFactoryFinderResolver();
FactoryFinderResolver getFactoryFinderResolver();
/** * Gets the factory finder resolver to use * * @return the factory finder resolver */
Gets the factory finder resolver to use
getFactoryFinderResolver
{ "repo_name": "christophd/camel", "path": "core/camel-api/src/main/java/org/apache/camel/ExtendedCamelContext.java", "license": "apache-2.0", "size": 29137 }
[ "org.apache.camel.spi.FactoryFinderResolver" ]
import org.apache.camel.spi.FactoryFinderResolver;
import org.apache.camel.spi.*;
[ "org.apache.camel" ]
org.apache.camel;
2,479,391
Collection getGroups();
Collection getGroups();
/** * * Access the groups defined for this assignment. * * @return A Collection (String) of group refs (authorization group ids) defined for this message; empty if none are defined. */
Access the groups defined for this assignment
getGroups
{ "repo_name": "ouit0408/sakai", "path": "assignment/assignment-api/api/src/java/org/sakaiproject/assignment/api/Assignment.java", "license": "apache-2.0", "size": 11954 }
[ "java.util.Collection" ]
import java.util.Collection;
import java.util.*;
[ "java.util" ]
java.util;
935,044
@JsonProperty("_links") public NoteLinks getLinks() { return Links; }
@JsonProperty(STR) NoteLinks function() { return Links; }
/** * The links of a note * * @return * The Links */
The links of a note
getLinks
{ "repo_name": "adrobisch/putput-data", "path": "putput-data-api/src/main/java/org/putput/api/model/Note.java", "license": "lgpl-3.0", "size": 4731 }
[ "com.fasterxml.jackson.annotation.JsonProperty" ]
import com.fasterxml.jackson.annotation.JsonProperty;
import com.fasterxml.jackson.annotation.*;
[ "com.fasterxml.jackson" ]
com.fasterxml.jackson;
733,667
double getDouble(int columnIndex) throws OntoDriverException;
double getDouble(int columnIndex) throws OntoDriverException;
/** * Retrieves value from column at the specified index and returns it as {@code double}. * * @param columnIndex Column index, the first column has index 0 * @return {@code double} value * @throws IllegalStateException If called on a closed result set * @throws OntoDriverException If the {@code columnIndex} is not a valid column index, the value cannot be cast to * {@code double} or there occurs some other error */
Retrieves value from column at the specified index and returns it as double
getDouble
{ "repo_name": "kbss-cvut/jopa", "path": "ontodriver-api/src/main/java/cz/cvut/kbss/ontodriver/iteration/ResultRow.java", "license": "lgpl-3.0", "size": 14731 }
[ "cz.cvut.kbss.ontodriver.exception.OntoDriverException" ]
import cz.cvut.kbss.ontodriver.exception.OntoDriverException;
import cz.cvut.kbss.ontodriver.exception.*;
[ "cz.cvut.kbss" ]
cz.cvut.kbss;
1,862,489
public static RefactoringStatus checkMethodInHierarchy(ITypeBinding type, String methodName, ITypeBinding returnType, ITypeBinding[] parameters) { RefactoringStatus result= new RefactoringStatus(); IMethodBinding method= Bindings.findMethodInHierarchy(type, methodName, parameters); if (method != null) { boolean returnTypeClash= false; ITypeBinding methodReturnType= method.getReturnType(); if (returnType != null && methodReturnType != null) { String returnTypeKey= returnType.getKey(); String methodReturnTypeKey= methodReturnType.getKey(); if (returnTypeKey == null && methodReturnTypeKey == null) { returnTypeClash= returnType != methodReturnType; } else if (returnTypeKey != null && methodReturnTypeKey != null) { returnTypeClash= !returnTypeKey.equals(methodReturnTypeKey); } } ITypeBinding dc= method.getDeclaringClass(); if (returnTypeClash) { result.addError(Messages.format(RefactoringCoreMessages.Checks_methodName_returnTypeClash, new Object[] {BasicElementLabels.getJavaElementName(methodName), BasicElementLabels.getJavaElementName(dc.getName())}), JavaStatusContext.create(method)); } else { if (method.isConstructor()) { result.addWarning(Messages.format(RefactoringCoreMessages.Checks_methodName_constructor, new Object[] { BasicElementLabels.getJavaElementName(dc.getName()) })); } else { result.addError(Messages.format(RefactoringCoreMessages.Checks_methodName_overrides, new Object[] { BasicElementLabels.getJavaElementName(methodName), BasicElementLabels.getJavaElementName(dc.getName()) }), JavaStatusContext.create(method)); } } } return result; } //---- Selection checks --------------------------------------------------------------------
static RefactoringStatus function(ITypeBinding type, String methodName, ITypeBinding returnType, ITypeBinding[] parameters) { RefactoringStatus result= new RefactoringStatus(); IMethodBinding method= Bindings.findMethodInHierarchy(type, methodName, parameters); if (method != null) { boolean returnTypeClash= false; ITypeBinding methodReturnType= method.getReturnType(); if (returnType != null && methodReturnType != null) { String returnTypeKey= returnType.getKey(); String methodReturnTypeKey= methodReturnType.getKey(); if (returnTypeKey == null && methodReturnTypeKey == null) { returnTypeClash= returnType != methodReturnType; } else if (returnTypeKey != null && methodReturnTypeKey != null) { returnTypeClash= !returnTypeKey.equals(methodReturnTypeKey); } } ITypeBinding dc= method.getDeclaringClass(); if (returnTypeClash) { result.addError(Messages.format(RefactoringCoreMessages.Checks_methodName_returnTypeClash, new Object[] {BasicElementLabels.getJavaElementName(methodName), BasicElementLabels.getJavaElementName(dc.getName())}), JavaStatusContext.create(method)); } else { if (method.isConstructor()) { result.addWarning(Messages.format(RefactoringCoreMessages.Checks_methodName_constructor, new Object[] { BasicElementLabels.getJavaElementName(dc.getName()) })); } else { result.addError(Messages.format(RefactoringCoreMessages.Checks_methodName_overrides, new Object[] { BasicElementLabels.getJavaElementName(methodName), BasicElementLabels.getJavaElementName(dc.getName()) }), JavaStatusContext.create(method)); } } } return result; }
/** * Checks if the new method somehow conflicts with an already existing method in * the hierarchy. The following checks are done: * <ul> * <li> if the new method overrides a method defined in the given type or in one of its * super classes. </li> * </ul> * @param type * @param methodName * @param returnType * @param parameters * @return the status */
Checks if the new method somehow conflicts with an already existing method in the hierarchy. The following checks are done: if the new method overrides a method defined in the given type or in one of its super classes.
checkMethodInHierarchy
{ "repo_name": "elucash/eclipse-oxygen", "path": "org.eclipse.jdt.ui/src/org/eclipse/jdt/internal/corext/refactoring/Checks.java", "license": "epl-1.0", "size": 34153 }
[ "org.eclipse.jdt.core.dom.IMethodBinding", "org.eclipse.jdt.core.dom.ITypeBinding", "org.eclipse.jdt.internal.core.manipulation.util.BasicElementLabels", "org.eclipse.jdt.internal.corext.dom.Bindings", "org.eclipse.jdt.internal.corext.refactoring.base.JavaStatusContext", "org.eclipse.jdt.internal.corext.util.Messages", "org.eclipse.ltk.core.refactoring.RefactoringStatus" ]
import org.eclipse.jdt.core.dom.IMethodBinding; import org.eclipse.jdt.core.dom.ITypeBinding; import org.eclipse.jdt.internal.core.manipulation.util.BasicElementLabels; import org.eclipse.jdt.internal.corext.dom.Bindings; import org.eclipse.jdt.internal.corext.refactoring.base.JavaStatusContext; import org.eclipse.jdt.internal.corext.util.Messages; import org.eclipse.ltk.core.refactoring.RefactoringStatus;
import org.eclipse.jdt.core.dom.*; import org.eclipse.jdt.internal.core.manipulation.util.*; import org.eclipse.jdt.internal.corext.dom.*; import org.eclipse.jdt.internal.corext.refactoring.base.*; import org.eclipse.jdt.internal.corext.util.*; import org.eclipse.ltk.core.refactoring.*;
[ "org.eclipse.jdt", "org.eclipse.ltk" ]
org.eclipse.jdt; org.eclipse.ltk;
2,911,768
return discoveryEventBus; } private DiscoveryEventUtils() { //Utility class private constructor intentionally left blank. } public static final class SearchStartedEvent { private final Type type; public SearchStartedEvent(Type type) { this.type = type; }
return discoveryEventBus; } private DiscoveryEventUtils() { } public static final class SearchStartedEvent { private final Type type; public SearchStartedEvent(Type type) { this.type = type; }
/** * Get the discovery event bus. * * @return The discovery event bus. */
Get the discovery event bus
getDiscoveryEventBus
{ "repo_name": "eugene7646/autopsy", "path": "Core/src/org/sleuthkit/autopsy/discovery/search/DiscoveryEventUtils.java", "license": "apache-2.0", "size": 16041 }
[ "org.sleuthkit.autopsy.discovery.search.SearchData" ]
import org.sleuthkit.autopsy.discovery.search.SearchData;
import org.sleuthkit.autopsy.discovery.search.*;
[ "org.sleuthkit.autopsy" ]
org.sleuthkit.autopsy;
12,132
@Test public void canIntegrateEntityMetadataCheckTrue() { String entityName = "testEntity"; DataServiceImpl dataServiceImpl = Mockito.mock(DataServiceImpl.class); when(dataServiceImpl.hasRepository(entityName)).thenReturn(Boolean.TRUE); DefaultEntityMetaData existingEntityMetaData = new DefaultEntityMetaData(entityName); existingEntityMetaData.addAttribute("ID").setIdAttribute(true); DefaultEntityMetaData newEntityMetaData = new DefaultEntityMetaData(entityName); newEntityMetaData.addAttribute("ID").setIdAttribute(true); when(dataServiceImpl.getEntityMetaData(entityName)).thenReturn(existingEntityMetaData); MetaDataServiceImpl metaDataService = new MetaDataServiceImpl(dataServiceImpl); assertTrue(metaDataService.canIntegrateEntityMetadataCheck(newEntityMetaData)); }
void function() { String entityName = STR; DataServiceImpl dataServiceImpl = Mockito.mock(DataServiceImpl.class); when(dataServiceImpl.hasRepository(entityName)).thenReturn(Boolean.TRUE); DefaultEntityMetaData existingEntityMetaData = new DefaultEntityMetaData(entityName); existingEntityMetaData.addAttribute("ID").setIdAttribute(true); DefaultEntityMetaData newEntityMetaData = new DefaultEntityMetaData(entityName); newEntityMetaData.addAttribute("ID").setIdAttribute(true); when(dataServiceImpl.getEntityMetaData(entityName)).thenReturn(existingEntityMetaData); MetaDataServiceImpl metaDataService = new MetaDataServiceImpl(dataServiceImpl); assertTrue(metaDataService.canIntegrateEntityMetadataCheck(newEntityMetaData)); }
/** * Test that a new entity has at least all attributes as of the existing entity. If not it results false. */
Test that a new entity has at least all attributes as of the existing entity. If not it results false
canIntegrateEntityMetadataCheckTrue
{ "repo_name": "adini121/molgenis", "path": "molgenis-data/src/test/java/org/molgenis/data/meta/MetaDataServiceImplTest.java", "license": "lgpl-3.0", "size": 11313 }
[ "org.mockito.Mockito", "org.molgenis.data.support.DataServiceImpl", "org.molgenis.data.support.DefaultEntityMetaData", "org.testng.Assert" ]
import org.mockito.Mockito; import org.molgenis.data.support.DataServiceImpl; import org.molgenis.data.support.DefaultEntityMetaData; import org.testng.Assert;
import org.mockito.*; import org.molgenis.data.support.*; import org.testng.*;
[ "org.mockito", "org.molgenis.data", "org.testng" ]
org.mockito; org.molgenis.data; org.testng;
76,850
public Turn(Section theSection) { setSection(theSection); } public Turn(Section theSection, Node turnNode) { setSection(theSection); try { setSpeakerId(turnNode.getAttributes().getNamedItem("speaker").getNodeValue()); } catch(Exception exception) {} try { setStartTime(turnNode.getAttributes().getNamedItem("startTime").getNodeValue()); setIdFromTime(); } catch(Exception exception) {} try { setEndTime(turnNode.getAttributes().getNamedItem("endTime").getNodeValue()); } catch(Exception exception) {} // traverse nodes NodeList children = turnNode.getChildNodes(); Sync lastSync = new Sync(this, null); // previous sync Sync sync = new Sync(this, null); // current sync for (int j = 0; j < children.getLength(); j++) { Node child = children.item(j); if (child.getNodeType() == Node.ELEMENT_NODE && child.getNodeName().equalsIgnoreCase("Sync")) { // Sync node lastSync = sync; sync = new Sync(this, child); lastSync.setEndTime(sync.getTime()); addSync(sync); //System.out.println("start Sync: " + sync); } else if (child.getNodeType() == Node.ELEMENT_NODE && child.getNodeName().equalsIgnoreCase("Who")) { // Who node - multiple speakers try { // there should be multiple speakers named StringTokenizer names = new StringTokenizer(getSpeakerId(), " "); int speakerIndex = Integer.parseInt( child.getAttributes().getNamedItem("nb").getNodeValue()); // go to the indexed speaker while (--speakerIndex > 0) { names.nextToken(); } String strSpeaker = names.nextToken(); sync.addWho(strSpeaker); } catch (Exception ex) { //System.out.println("Turn "+getStartTime()+": Failed to parse speaker for Who: " + child.toString() + "\n" + ex.toString()); } } else if (child.getNodeType() == Node.ELEMENT_NODE && child.getNodeName().equalsIgnoreCase("Event")) { // Event node - non-speech event Event event = new Event(sync, child); //System.out.println("Event: " + event); } else if (child.getNodeType() == Node.ELEMENT_NODE && child.getNodeName().equalsIgnoreCase("Comment")) { // Event node - non-speech event Comment comment = new Comment(sync, child); //System.out.println("Comment: " + comment); } else if (child.getNodeType() == Node.TEXT_NODE) { // text node String strText = child.getNodeValue(); if (strText.trim().length() > 0) { sync.appendText(strText); //System.out.println("Text: " + strText); //System.out.println("Sync now: " + sync); } } else { //System.out.println("Ignoring element: " + child.toString()); } } // next child // set the end time of the last sync sync.setEndTime(getEndTime()); } // end of constructor
public Turn(Section theSection) { setSection(theSection); } public Turn(Section theSection, Node turnNode) { setSection(theSection); try { setSpeakerId(turnNode.getAttributes().getNamedItem(STR).getNodeValue()); } catch(Exception exception) {} try { setStartTime(turnNode.getAttributes().getNamedItem(STR).getNodeValue()); setIdFromTime(); } catch(Exception exception) {} try { setEndTime(turnNode.getAttributes().getNamedItem(STR).getNodeValue()); } catch(Exception exception) {} NodeList children = turnNode.getChildNodes(); Sync lastSync = new Sync(this, null); Sync sync = new Sync(this, null); for (int j = 0; j < children.getLength(); j++) { Node child = children.item(j); if (child.getNodeType() == Node.ELEMENT_NODE && child.getNodeName().equalsIgnoreCase("Sync")) { lastSync = sync; sync = new Sync(this, child); lastSync.setEndTime(sync.getTime()); addSync(sync); } else if (child.getNodeType() == Node.ELEMENT_NODE && child.getNodeName().equalsIgnoreCase("Who")) { try { StringTokenizer names = new StringTokenizer(getSpeakerId(), " "); int speakerIndex = Integer.parseInt( child.getAttributes().getNamedItem("nb").getNodeValue()); while (--speakerIndex > 0) { names.nextToken(); } String strSpeaker = names.nextToken(); sync.addWho(strSpeaker); } catch (Exception ex) { } } else if (child.getNodeType() == Node.ELEMENT_NODE && child.getNodeName().equalsIgnoreCase("Event")) { Event event = new Event(sync, child); } else if (child.getNodeType() == Node.ELEMENT_NODE && child.getNodeName().equalsIgnoreCase(STR)) { Comment comment = new Comment(sync, child); } else if (child.getNodeType() == Node.TEXT_NODE) { String strText = child.getNodeValue(); if (strText.trim().length() > 0) { sync.appendText(strText); } } else { } } sync.setEndTime(getEndTime()); }
/** * Syncs mutator - Ordered list of syncronized transcript parts (A list of {@link nz.ac.canterbury.ling.transcriber.Sync} objects) * @param vNewSyncs Ordered list of syncronized transcript parts (A list of {@link nz.ac.canterbury.ling.transcriber.Sync} objects) */
Syncs mutator - Ordered list of syncronized transcript parts (A list of <code>nz.ac.canterbury.ling.transcriber.Sync</code> objects)
setSyncs
{ "repo_name": "nzilbb/ag", "path": "formatter/transcriber/src/main/java/nzilbb/formatter/transcriber/Turn.java", "license": "gpl-3.0", "size": 11585 }
[ "java.util.StringTokenizer", "org.w3c.dom.Node", "org.w3c.dom.NodeList" ]
import java.util.StringTokenizer; import org.w3c.dom.Node; import org.w3c.dom.NodeList;
import java.util.*; import org.w3c.dom.*;
[ "java.util", "org.w3c.dom" ]
java.util; org.w3c.dom;
953,574
private Object processFixed(Schema schemaNode) { int size = schemaNode.getFixedSize(); byte[] bytes = new byte[size]; for (int i = 0; i < size; i++) { bytes[i] = (byte) 0; } GenericFixed result = new GenericData.Fixed(schemaNode, bytes); return result; }
Object function(Schema schemaNode) { int size = schemaNode.getFixedSize(); byte[] bytes = new byte[size]; for (int i = 0; i < size; i++) { bytes[i] = (byte) 0; } GenericFixed result = new GenericData.Fixed(schemaNode, bytes); return result; }
/** * Processes fixed type. * * @param schemaNode schema for current type. * @return generated value for input record type. */
Processes fixed type
processFixed
{ "repo_name": "sashadidukh/kaa", "path": "common/core/src/main/java/org/kaaproject/kaa/server/common/core/algorithms/generation/DefaultRecordGenerationAlgorithmImpl.java", "license": "apache-2.0", "size": 10897 }
[ "org.apache.avro.Schema", "org.apache.avro.generic.GenericData", "org.apache.avro.generic.GenericFixed" ]
import org.apache.avro.Schema; import org.apache.avro.generic.GenericData; import org.apache.avro.generic.GenericFixed;
import org.apache.avro.*; import org.apache.avro.generic.*;
[ "org.apache.avro" ]
org.apache.avro;
1,366,780
@Override public Object accept(ICOSVisitor visitor) throws IOException { return visitor.visitFromInt(this); }
Object function(ICOSVisitor visitor) throws IOException { return visitor.visitFromInt(this); }
/** * visitor pattern double dispatch method. * * @param visitor The object to notify when visiting this object. * @return any object, depending on the visitor implementation, or null * @throws IOException If an error occurs while visiting this object. */
visitor pattern double dispatch method
accept
{ "repo_name": "ZhenyaM/veraPDF-pdfbox", "path": "pdfbox/src/main/java/org/apache/pdfbox/cos/COSInteger.java", "license": "apache-2.0", "size": 4907 }
[ "java.io.IOException" ]
import java.io.IOException;
import java.io.*;
[ "java.io" ]
java.io;
2,098,378
public static boolean existsCloudProvider(String cloudName) { String xPathToProp = "/Project/Cloud/Provider[@name='" + cloudName + "']"; XPathResult res = (XPathResult) evaluator.evaluate(xPathToProp, projectDoc, null, XPathResult.FIRST_ORDERED_NODE_TYPE, null); Node n = res.getSingleNodeValue(); return n != null; }
static boolean function(String cloudName) { String xPathToProp = STR + cloudName + "']"; XPathResult res = (XPathResult) evaluator.evaluate(xPathToProp, projectDoc, null, XPathResult.FIRST_ORDERED_NODE_TYPE, null); Node n = res.getSingleNodeValue(); return n != null; }
/** * Checks if a any cloud Provider with that name appears on the project file * * @param cloudName Name of the cloud provider * @return true if the provider exists */
Checks if a any cloud Provider with that name appears on the project file
existsCloudProvider
{ "repo_name": "ElsevierSoftwareX/SOFTX-D-15-00010", "path": "compss/compss-rt/rt/src/main/java/integratedtoolkit/util/ProjectManager.java", "license": "apache-2.0", "size": 18800 }
[ "org.w3c.dom.Node", "org.w3c.dom.xpath.XPathResult" ]
import org.w3c.dom.Node; import org.w3c.dom.xpath.XPathResult;
import org.w3c.dom.*; import org.w3c.dom.xpath.*;
[ "org.w3c.dom" ]
org.w3c.dom;
2,686,208
EReference getIfExpr_Elsifs();
EReference getIfExpr_Elsifs();
/** * Returns the meta object for the containment reference list '{@link org.eclectic.frontend.core.IfExpr#getElsifs <em>Elsifs</em>}'. * <!-- begin-user-doc --> * <!-- end-user-doc --> * @return the meta object for the containment reference list '<em>Elsifs</em>'. * @see org.eclectic.frontend.core.IfExpr#getElsifs() * @see #getIfExpr() * @generated */
Returns the meta object for the containment reference list '<code>org.eclectic.frontend.core.IfExpr#getElsifs Elsifs</code>'.
getIfExpr_Elsifs
{ "repo_name": "jesusc/eclectic", "path": "plugins/org.eclectic.frontend.asm/src-gen/org/eclectic/frontend/core/CorePackage.java", "license": "gpl-3.0", "size": 187193 }
[ "org.eclipse.emf.ecore.EReference" ]
import org.eclipse.emf.ecore.EReference;
import org.eclipse.emf.ecore.*;
[ "org.eclipse.emf" ]
org.eclipse.emf;
1,635,916
public Map<String, String> getParasMap() { return parasMap; }
Map<String, String> function() { return parasMap; }
/** * get paras map * * @return */
get paras map
getParasMap
{ "repo_name": "solaris0403/SeleneDemo", "path": "common_lib/src/main/java/com/tony/selene/common/trinea/android/common/entity/HttpRequest.java", "license": "gpl-2.0", "size": 3981 }
[ "java.util.Map" ]
import java.util.Map;
import java.util.*;
[ "java.util" ]
java.util;
556,708
public Settings indexSettings() { return this.indexSettings; }
Settings function() { return this.indexSettings; }
/** * Returns settings that should be added/changed in all restored indices */
Returns settings that should be added/changed in all restored indices
indexSettings
{ "repo_name": "anti-social/elasticsearch", "path": "src/main/java/org/elasticsearch/action/admin/cluster/snapshots/restore/RestoreSnapshotRequest.java", "license": "apache-2.0", "size": 24240 }
[ "org.elasticsearch.common.settings.Settings" ]
import org.elasticsearch.common.settings.Settings;
import org.elasticsearch.common.settings.*;
[ "org.elasticsearch.common" ]
org.elasticsearch.common;
463,090
ImmutableValue<Boolean> powered();
ImmutableValue<Boolean> powered();
/** * Gets the {@link ImmutableValue} for the powered state of the {@link Structure}. * * @return The value for the powered state */
Gets the <code>ImmutableValue</code> for the powered state of the <code>Structure</code>
powered
{ "repo_name": "JBYoshi/SpongeAPI", "path": "src/main/java/org/spongepowered/api/data/manipulator/immutable/tileentity/ImmutableStructureData.java", "license": "mit", "size": 3828 }
[ "org.spongepowered.api.data.value.immutable.ImmutableValue" ]
import org.spongepowered.api.data.value.immutable.ImmutableValue;
import org.spongepowered.api.data.value.immutable.*;
[ "org.spongepowered.api" ]
org.spongepowered.api;
2,028,013
private Object execute(Object o, Method m, Object... args) { Object result = null; try { m.setAccessible(true); // suppress Java language access result = m.invoke(o, args); } catch (IllegalAccessException e) { propagateCause(e); } catch (InvocationTargetException e) { propagateCause(e); } return result; }
Object function(Object o, Method m, Object... args) { Object result = null; try { m.setAccessible(true); result = m.invoke(o, args); } catch (IllegalAccessException e) { propagateCause(e); } catch (InvocationTargetException e) { propagateCause(e); } return result; }
/** * Invokes the method. * * @return result of execution */
Invokes the method
execute
{ "repo_name": "daveray/Hystrix", "path": "hystrix-contrib/hystrix-javanica/src/main/java/com/netflix/hystrix/contrib/javanica/command/MethodExecutionAction.java", "license": "apache-2.0", "size": 2402 }
[ "java.lang.reflect.InvocationTargetException", "java.lang.reflect.Method" ]
import java.lang.reflect.InvocationTargetException; import java.lang.reflect.Method;
import java.lang.reflect.*;
[ "java.lang" ]
java.lang;
65,523
protected synchronized void closeOutputStream(final StreamingOutputStream streamingOutputStream) { try { final Iterator<WeakReference<StreamingOutputStream>> it = this.connectedOutputStreams.iterator(); while (it.hasNext()) { StreamingOutputStream current = null; final WeakReference<StreamingOutputStream> next = it.next(); if ((current = next.get()) == null || current == streamingOutputStream) { it.remove(); } } } finally { final StreamingChunk currentChunk = streamingOutputStream.getCurrentChunk(); if (currentChunk != null) { currentChunk.setCanGrow(false); } streamingOutputStream.setCurrentChunk(null); } }
synchronized void function(final StreamingOutputStream streamingOutputStream) { try { final Iterator<WeakReference<StreamingOutputStream>> it = this.connectedOutputStreams.iterator(); while (it.hasNext()) { StreamingOutputStream current = null; final WeakReference<StreamingOutputStream> next = it.next(); if ((current = next.get()) == null current == streamingOutputStream) { it.remove(); } } } finally { final StreamingChunk currentChunk = streamingOutputStream.getCurrentChunk(); if (currentChunk != null) { currentChunk.setCanGrow(false); } streamingOutputStream.setCurrentChunk(null); } }
/** * removes given StreamingOutputStream from this StreamingController. also * set canGrow to false on its StreamingChunk * * @param streamingOutputStream */
removes given StreamingOutputStream from this StreamingController. also set canGrow to false on its StreamingChunk
closeOutputStream
{ "repo_name": "friedlwo/AppWoksUtils", "path": "src/org/appwork/utils/io/streamingio/Streaming.java", "license": "artistic-2.0", "size": 15117 }
[ "java.lang.ref.WeakReference", "java.util.Iterator" ]
import java.lang.ref.WeakReference; import java.util.Iterator;
import java.lang.ref.*; import java.util.*;
[ "java.lang", "java.util" ]
java.lang; java.util;
1,510,342
@Test public void testRemoveEventListenerNullType() { assertFalse("Wrong result", list.removeEventListener(null, new ListenerTestImpl())); }
void function() { assertFalse(STR, list.removeEventListener(null, new ListenerTestImpl())); }
/** * Tests that removeEventListener() can handle a null event type. */
Tests that removeEventListener() can handle a null event type
testRemoveEventListenerNullType
{ "repo_name": "mohanaraosv/commons-configuration", "path": "src/test/java/org/apache/commons/configuration2/event/TestEventListenerList.java", "license": "apache-2.0", "size": 20553 }
[ "org.junit.Assert" ]
import org.junit.Assert;
import org.junit.*;
[ "org.junit" ]
org.junit;
2,555,614
@JsonProperty("permissions") public BranchImplpermissions getPermissions() { return permissions; }
@JsonProperty(STR) BranchImplpermissions function() { return permissions; }
/** * Get permissions * @return permissions **/
Get permissions
getPermissions
{ "repo_name": "cliffano/swaggy-jenkins", "path": "clients/jaxrs-cxf-client/generated/src/gen/java/org/openapitools/model/BranchImpl.java", "license": "mit", "size": 7889 }
[ "com.fasterxml.jackson.annotation.JsonProperty", "org.openapitools.model.BranchImplpermissions" ]
import com.fasterxml.jackson.annotation.JsonProperty; import org.openapitools.model.BranchImplpermissions;
import com.fasterxml.jackson.annotation.*; import org.openapitools.model.*;
[ "com.fasterxml.jackson", "org.openapitools.model" ]
com.fasterxml.jackson; org.openapitools.model;
328,696
public void drawResults(String tabId, String resultString, String contentType) { QueryTab tab = (QueryTab)view.getTabs().getTab(tabId); view.getTabs().selectTab(tabId); if (tab == null) { view.getElements().onError("No tab to draw results in"); } tab.getResultContainer().processResult(resultString, contentType); }
void function(String tabId, String resultString, String contentType) { QueryTab tab = (QueryTab)view.getTabs().getTab(tabId); view.getTabs().selectTab(tabId); if (tab == null) { view.getElements().onError(STR); } tab.getResultContainer().processResult(resultString, contentType); }
/** * Draw json or xml results * Keep this method in the view object, so that it is easily callable from js * * @param tabId Tab id to draw results in. * Pass this, so when you switch tabs just after clicking the query button, the results still gets drawn in the proper tab * @param resultString * @param contentType Content type of query result */
Draw json or xml results Keep this method in the view object, so that it is easily callable from js
drawResults
{ "repo_name": "JervenBolleman/yasgui", "path": "src/main/java/com/data2semantics/yasgui/client/helpers/CallableJsMethods.java", "license": "mit", "size": 8400 }
[ "com.data2semantics.yasgui.client.tab.QueryTab" ]
import com.data2semantics.yasgui.client.tab.QueryTab;
import com.data2semantics.yasgui.client.tab.*;
[ "com.data2semantics.yasgui" ]
com.data2semantics.yasgui;
1,268,764
InboundPacket mapInboundPacket(PiPacketOperation packetOperation) throws PiInterpreterException;
InboundPacket mapInboundPacket(PiPacketOperation packetOperation) throws PiInterpreterException;
/** * Returns an inbound packet equivalent to the given PI packet operation. * * @param packetOperation packet operation * @return inbound packet * @throws PiInterpreterException if the packet operation cannot be mapped * to an inbound packet */
Returns an inbound packet equivalent to the given PI packet operation
mapInboundPacket
{ "repo_name": "kuujo/onos", "path": "core/api/src/main/java/org/onosproject/net/pi/model/PiPipelineInterpreter.java", "license": "apache-2.0", "size": 5652 }
[ "org.onosproject.net.packet.InboundPacket", "org.onosproject.net.pi.runtime.PiPacketOperation" ]
import org.onosproject.net.packet.InboundPacket; import org.onosproject.net.pi.runtime.PiPacketOperation;
import org.onosproject.net.packet.*; import org.onosproject.net.pi.runtime.*;
[ "org.onosproject.net" ]
org.onosproject.net;
1,600,412
private void onAppsUpdated() { // Sort the list of apps mApps.clear(); mApps.addAll(mComponentToAppMap.values()); Collections.sort(mApps, mAppNameComparator.getAppInfoComparator()); // As a special case for some languages (currently only Simplified Chinese), we may need to // coalesce sections Locale curLocale = mLauncher.getResources().getConfiguration().locale; TreeMap<String, ArrayList<AppInfo>> sectionMap = null; boolean localeRequiresSectionSorting = curLocale.equals(Locale.SIMPLIFIED_CHINESE); if (localeRequiresSectionSorting) { // Compute the section headers. We use a TreeMap with the section name comparator to // ensure that the sections are ordered when we iterate over it later sectionMap = new TreeMap<>(mAppNameComparator.getSectionNameComparator()); for (AppInfo info : mApps) { // Add the section to the cache String sectionName = getAndUpdateCachedSectionName(info.title); // Add it to the mapping ArrayList<AppInfo> sectionApps = sectionMap.get(sectionName); if (sectionApps == null) { sectionApps = new ArrayList<>(); sectionMap.put(sectionName, sectionApps); } sectionApps.add(info); } // Add each of the section apps to the list in order List<AppInfo> allApps = new ArrayList<>(mApps.size()); for (Map.Entry<String, ArrayList<AppInfo>> entry : sectionMap.entrySet()) { allApps.addAll(entry.getValue()); } mApps.clear(); mApps.addAll(allApps); } else { // Just compute the section headers for use below for (AppInfo info : mApps) { // Add the section to the cache getAndUpdateCachedSectionName(info.title); } } // Recompose the set of adapter items from the current set of apps updateAdapterItems(); }
void function() { mApps.clear(); mApps.addAll(mComponentToAppMap.values()); Collections.sort(mApps, mAppNameComparator.getAppInfoComparator()); Locale curLocale = mLauncher.getResources().getConfiguration().locale; TreeMap<String, ArrayList<AppInfo>> sectionMap = null; boolean localeRequiresSectionSorting = curLocale.equals(Locale.SIMPLIFIED_CHINESE); if (localeRequiresSectionSorting) { sectionMap = new TreeMap<>(mAppNameComparator.getSectionNameComparator()); for (AppInfo info : mApps) { String sectionName = getAndUpdateCachedSectionName(info.title); ArrayList<AppInfo> sectionApps = sectionMap.get(sectionName); if (sectionApps == null) { sectionApps = new ArrayList<>(); sectionMap.put(sectionName, sectionApps); } sectionApps.add(info); } List<AppInfo> allApps = new ArrayList<>(mApps.size()); for (Map.Entry<String, ArrayList<AppInfo>> entry : sectionMap.entrySet()) { allApps.addAll(entry.getValue()); } mApps.clear(); mApps.addAll(allApps); } else { for (AppInfo info : mApps) { getAndUpdateCachedSectionName(info.title); } } updateAdapterItems(); }
/** * Updates internals when the set of apps are updated. */
Updates internals when the set of apps are updated
onAppsUpdated
{ "repo_name": "bojanvu23/android_packages_apps_Trebuchet_Gradle", "path": "Trebuchet/src/main/java/com/lite/android/launcher3/allapps/AlphabeticalAppsList.java", "license": "apache-2.0", "size": 26573 }
[ "com.lite.android.launcher3.AppInfo", "java.util.ArrayList", "java.util.Collections", "java.util.List", "java.util.Locale", "java.util.Map", "java.util.TreeMap" ]
import com.lite.android.launcher3.AppInfo; import java.util.ArrayList; import java.util.Collections; import java.util.List; import java.util.Locale; import java.util.Map; import java.util.TreeMap;
import com.lite.android.launcher3.*; import java.util.*;
[ "com.lite.android", "java.util" ]
com.lite.android; java.util;
1,654,747
CompleteMultipartUploadResult uploadObject(final UploadObjectRequest req) throws IOException, InterruptedException, ExecutionException { // Set up the pipeline for concurrent encrypt and upload // Set up a thread pool for this pipeline ExecutorService es = req.getExecutorService(); final boolean defaultExecutorService = es == null; if (es == null) es = Executors.newFixedThreadPool(clientConfiguration.getMaxConnections()); UploadObjectObserver observer = req.getUploadObjectObserver(); if (observer == null) observer = new UploadObjectObserver(); // initialize the observer observer.init(req, this, this, es); // Initiate upload observer.onUploadInitiation(req); final List<PartETag> partETags = new ArrayList<PartETag>(); MultiFileOutputStream mfos = req.getMultiFileOutputStream(); if (mfos == null) mfos = new MultiFileOutputStream(); try { // initialize the multi-file output stream mfos.init(observer, req.getPartSize(), req.getDiskLimit()); // Kicks off the encryption-upload pipeline; // Note mfos is automatically closed upon method completion. putLocalObject(req, mfos); // block till all part have been uploaded for (Future<UploadPartResult> future: observer.getFutures()) { UploadPartResult partResult = future.get(); partETags.add(new PartETag(partResult.getPartNumber(), partResult.getETag())); } } finally { if (defaultExecutorService) es.shutdownNow(); // shut down the locally created thread pool mfos.cleanup(); // delete left-over temp files } // Complete upload return observer.onCompletion(partETags); }
CompleteMultipartUploadResult uploadObject(final UploadObjectRequest req) throws IOException, InterruptedException, ExecutionException { ExecutorService es = req.getExecutorService(); final boolean defaultExecutorService = es == null; if (es == null) es = Executors.newFixedThreadPool(clientConfiguration.getMaxConnections()); UploadObjectObserver observer = req.getUploadObjectObserver(); if (observer == null) observer = new UploadObjectObserver(); observer.init(req, this, this, es); observer.onUploadInitiation(req); final List<PartETag> partETags = new ArrayList<PartETag>(); MultiFileOutputStream mfos = req.getMultiFileOutputStream(); if (mfos == null) mfos = new MultiFileOutputStream(); try { mfos.init(observer, req.getPartSize(), req.getDiskLimit()); putLocalObject(req, mfos); for (Future<UploadPartResult> future: observer.getFutures()) { UploadPartResult partResult = future.get(); partETags.add(new PartETag(partResult.getPartNumber(), partResult.getETag())); } } finally { if (defaultExecutorService) es.shutdownNow(); mfos.cleanup(); } return observer.onCompletion(partETags); }
/** * Used for performance testing purposes only. Hence package private. * This method is subject to removal anytime without notice. */
Used for performance testing purposes only. Hence package private. This method is subject to removal anytime without notice
uploadObject
{ "repo_name": "sdole/aws-sdk-java", "path": "aws-java-sdk-s3/src/main/java/com/amazonaws/services/s3/AmazonS3Client.java", "license": "apache-2.0", "size": 194278 }
[ "com.amazonaws.services.s3.internal.MultiFileOutputStream", "com.amazonaws.services.s3.model.CompleteMultipartUploadResult", "com.amazonaws.services.s3.model.PartETag", "com.amazonaws.services.s3.model.UploadObjectRequest", "com.amazonaws.services.s3.model.UploadPartResult", "java.io.IOException", "java.util.ArrayList", "java.util.List", "java.util.concurrent.ExecutionException", "java.util.concurrent.ExecutorService", "java.util.concurrent.Executors", "java.util.concurrent.Future" ]
import com.amazonaws.services.s3.internal.MultiFileOutputStream; import com.amazonaws.services.s3.model.CompleteMultipartUploadResult; import com.amazonaws.services.s3.model.PartETag; import com.amazonaws.services.s3.model.UploadObjectRequest; import com.amazonaws.services.s3.model.UploadPartResult; import java.io.IOException; import java.util.ArrayList; import java.util.List; import java.util.concurrent.ExecutionException; import java.util.concurrent.ExecutorService; import java.util.concurrent.Executors; import java.util.concurrent.Future;
import com.amazonaws.services.s3.internal.*; import com.amazonaws.services.s3.model.*; import java.io.*; import java.util.*; import java.util.concurrent.*;
[ "com.amazonaws.services", "java.io", "java.util" ]
com.amazonaws.services; java.io; java.util;
2,384,328
public final void tunnelProxy(HttpHost proxy, boolean secure) { if (proxy == null) { throw new IllegalArgumentException("Proxy host may not be null."); } if (!this.connected) { throw new IllegalStateException("No tunnel unless connected."); } if (this.proxyChain == null) { throw new IllegalStateException("No proxy tunnel without proxy."); } // prepare an extended proxy chain HttpHost[] proxies = new HttpHost[this.proxyChain.length+1]; System.arraycopy(this.proxyChain, 0, proxies, 0, this.proxyChain.length); proxies[proxies.length-1] = proxy; this.proxyChain = proxies; this.secure = secure; }
final void function(HttpHost proxy, boolean secure) { if (proxy == null) { throw new IllegalArgumentException(STR); } if (!this.connected) { throw new IllegalStateException(STR); } if (this.proxyChain == null) { throw new IllegalStateException(STR); } HttpHost[] proxies = new HttpHost[this.proxyChain.length+1]; System.arraycopy(this.proxyChain, 0, proxies, 0, this.proxyChain.length); proxies[proxies.length-1] = proxy; this.proxyChain = proxies; this.secure = secure; }
/** * Tracks tunnelling to a proxy in a proxy chain. * This will extend the tracked proxy chain, but it does not mark * the route as tunnelled. Only end-to-end tunnels are considered there. * * @param proxy the proxy tunnelled to * @param secure <code>true</code> if the route is secure, * <code>false</code> otherwise */
Tracks tunnelling to a proxy in a proxy chain. This will extend the tracked proxy chain, but it does not mark the route as tunnelled. Only end-to-end tunnels are considered there
tunnelProxy
{ "repo_name": "onedanshow/Screen-Courter", "path": "lib/src/org/apache/http/conn/routing/RouteTracker.java", "license": "gpl-3.0", "size": 12915 }
[ "org.apache.http.HttpHost" ]
import org.apache.http.HttpHost;
import org.apache.http.*;
[ "org.apache.http" ]
org.apache.http;
2,737,979
@Override public void fit(ObservationReal... oa) { fit(Arrays.asList(oa)); }
void function(ObservationReal... oa) { fit(Arrays.asList(oa)); }
/** * Fits this observation distribution function to a (non empty) set of * observations. This method performs one iteration of an * expectation-maximisation algorithm. * * @param oa A set of observations compatible with this function. */
Fits this observation distribution function to a (non empty) set of observations. This method performs one iteration of an expectation-maximisation algorithm
fit
{ "repo_name": "KommuSoft/jahmm", "path": "jahmm/src/jahmm/observables/OpdfGaussianMixture.java", "license": "bsd-3-clause", "size": 10497 }
[ "java.util.Arrays" ]
import java.util.Arrays;
import java.util.*;
[ "java.util" ]
java.util;
2,439,871
private void startGemFire() { Cache c = GemFireCacheImpl.getInstance(); if (c == null) { CacheFactory cacheFactory = new CacheFactory(); if (logLevel != null) cacheFactory.set("log-level", logLevel); this.cache = cacheFactory.create(); } else this.cache = c; this.logger = this.cache.getLogger(); }
void function() { Cache c = GemFireCacheImpl.getInstance(); if (c == null) { CacheFactory cacheFactory = new CacheFactory(); if (logLevel != null) cacheFactory.set(STR, logLevel); this.cache = cacheFactory.create(); } else this.cache = c; this.logger = this.cache.getLogger(); }
/** * Initializes the {@link Cache}, and creates Redis necessities * Region and protects declares that {@link Region} to be protected. * Also, every {@link GemFireRedisServer} will check for entries already in the * meta data Region. */
Initializes the <code>Cache</code>, and creates Redis necessities Region and protects declares that <code>Region</code> to be protected. Also, every <code>GemFireRedisServer</code> will check for entries already in the meta data Region
startGemFire
{ "repo_name": "kidaa/incubator-geode", "path": "gemfire-core/src/main/java/com/gemstone/gemfire/redis/GemFireRedisServer.java", "license": "apache-2.0", "size": 26150 }
[ "com.gemstone.gemfire.cache.Cache", "com.gemstone.gemfire.cache.CacheFactory", "com.gemstone.gemfire.internal.cache.GemFireCacheImpl" ]
import com.gemstone.gemfire.cache.Cache; import com.gemstone.gemfire.cache.CacheFactory; import com.gemstone.gemfire.internal.cache.GemFireCacheImpl;
import com.gemstone.gemfire.cache.*; import com.gemstone.gemfire.internal.cache.*;
[ "com.gemstone.gemfire" ]
com.gemstone.gemfire;
1,109,070
private void cancel() { setResult(Activity.RESULT_CANCELED); finish(); }
void function() { setResult(Activity.RESULT_CANCELED); finish(); }
/** * Method that cancels the activity */
Method that cancels the activity
cancel
{ "repo_name": "AnimeROM/android_package_AnimeManager", "path": "src/com/animerom/filemanager/activities/PickerActivity.java", "license": "apache-2.0", "size": 21324 }
[ "android.app.Activity" ]
import android.app.Activity;
import android.app.*;
[ "android.app" ]
android.app;
1,437,100
public static SpriteDefinition createSprite(int left, int top, int width, int height, int hotX, int hotY) { SpriteDefinition _result = new SpriteDefinition(); Box _sourceBox = new Box().withX(left).withY(top).withWidth(width).withHeight(height); Point _hotPoint = new Point().withX(hotX).withY(hotY); _result.withSourceBox(_sourceBox).withHotPoint(_hotPoint); return _result; }
static SpriteDefinition function(int left, int top, int width, int height, int hotX, int hotY) { SpriteDefinition _result = new SpriteDefinition(); Box _sourceBox = new Box().withX(left).withY(top).withWidth(width).withHeight(height); Point _hotPoint = new Point().withX(hotX).withY(hotY); _result.withSourceBox(_sourceBox).withHotPoint(_hotPoint); return _result; }
/** * Macro to create a fully defined anonymous sprite. * * @param left * x coordinate of the top-left corner. * @param top * y coordinate of the top-left corner. * @param width * width of the bloc. * @param height * height of the bloc. * @param hotX * x coordinate of the hot point. * @param hotY * y coordinate of the hot point. * @return a sprite definition. */
Macro to create a fully defined anonymous sprite
createSprite
{ "repo_name": "sporniket/game", "path": "canvas/src/main/java/com/sporniket/libre/game/canvas/sprite/SpriteDefinitionUtils.java", "license": "lgpl-3.0", "size": 5036 }
[ "com.sporniket.libre.game.canvas.Box", "com.sporniket.libre.game.canvas.Point" ]
import com.sporniket.libre.game.canvas.Box; import com.sporniket.libre.game.canvas.Point;
import com.sporniket.libre.game.canvas.*;
[ "com.sporniket.libre" ]
com.sporniket.libre;
338,644
public void testAutoCloseServerCursorProperty() throws Exception { String url = "jdbc:ignite:thin://127.0.0.1?" + JdbcThinUtils.PARAM_AUTO_CLOSE_SERVER_CURSOR; String err = "Failed to parse boolean property [name=" + JdbcThinUtils.PARAM_AUTO_CLOSE_SERVER_CURSOR; assertInvalid(url + "=0", err); assertInvalid(url + "=1", err); assertInvalid(url + "=false1", err); assertInvalid(url + "=true1", err); try (Connection conn = DriverManager.getConnection("jdbc:ignite:thin://127.0.0.1")) { assertFalse(io(conn).autoCloseServerCursor()); } try (Connection conn = DriverManager.getConnection(url + "=true")) { assertTrue(io(conn).autoCloseServerCursor()); } try (Connection conn = DriverManager.getConnection(url + "=True")) { assertTrue(io(conn).autoCloseServerCursor()); } try (Connection conn = DriverManager.getConnection(url + "=false")) { assertFalse(io(conn).autoCloseServerCursor()); } try (Connection conn = DriverManager.getConnection(url + "=False")) { assertFalse(io(conn).autoCloseServerCursor()); } }
void function() throws Exception { String url = STRFailed to parse boolean property [name=STR=0STR=1STR=false1STR=true1STRjdbc:ignite:thin: assertFalse(io(conn).autoCloseServerCursor()); } try (Connection conn = DriverManager.getConnection(url + "=true")) { assertTrue(io(conn).autoCloseServerCursor()); } try (Connection conn = DriverManager.getConnection(url + "=True")) { assertTrue(io(conn).autoCloseServerCursor()); } try (Connection conn = DriverManager.getConnection(url + STR)) { assertFalse(io(conn).autoCloseServerCursor()); } try (Connection conn = DriverManager.getConnection(url + STR)) { assertFalse(io(conn).autoCloseServerCursor()); } }
/** * Test autoCloseServerCursor property handling. * * @throws Exception If failed. */
Test autoCloseServerCursor property handling
testAutoCloseServerCursorProperty
{ "repo_name": "a1vanov/ignite", "path": "modules/clients/src/test/java/org/apache/ignite/jdbc/thin/JdbcThinConnectionSelfTest.java", "license": "apache-2.0", "size": 12351 }
[ "java.sql.Connection", "java.sql.DriverManager" ]
import java.sql.Connection; import java.sql.DriverManager;
import java.sql.*;
[ "java.sql" ]
java.sql;
994,326
public void setOverheadSize(JobCollection collection) { OverheadSize = collection.getSettings().getBooleanProperty(Keys.KEY_Input_useReadOverhead) ? 2048000 : 0; }
void function(JobCollection collection) { OverheadSize = collection.getSettings().getBooleanProperty(Keys.KEY_Input_useReadOverhead) ? 2048000 : 0; }
/** * getOverhead to collect more samples */
getOverhead to collect more samples
setOverheadSize
{ "repo_name": "esurharun/projectx", "path": "src/net/sourceforge/dvb/projectx/parser/StreamParserBase.java", "license": "gpl-2.0", "size": 17428 }
[ "net.sourceforge.dvb.projectx.common.JobCollection", "net.sourceforge.dvb.projectx.common.Keys" ]
import net.sourceforge.dvb.projectx.common.JobCollection; import net.sourceforge.dvb.projectx.common.Keys;
import net.sourceforge.dvb.projectx.common.*;
[ "net.sourceforge.dvb" ]
net.sourceforge.dvb;
1,084,476
public List<String> getAvailablePaymentActionsForReturn(String returnId, String paymentId) throws Exception { MozuClient<List<String>> client = com.mozu.api.clients.commerce.ReturnClient.getAvailablePaymentActionsForReturnClient( returnId, paymentId); client.setContext(_apiContext); client.executeRequest(); return client.getResult(); }
List<String> function(String returnId, String paymentId) throws Exception { MozuClient<List<String>> client = com.mozu.api.clients.commerce.ReturnClient.getAvailablePaymentActionsForReturnClient( returnId, paymentId); client.setContext(_apiContext); client.executeRequest(); return client.getResult(); }
/** * Retrieves a list of the payment actions available to perform for the specified return when a return results in a refund to the customer. * <p><pre><code> * Return return = new Return(); * string string = return.getAvailablePaymentActionsForReturn( returnId, paymentId); * </code></pre></p> * @param paymentId Unique identifier of the payment for which to perform the action. * @param returnId Unique identifier of the return whose items you want to get. * @return List<string> * @see string */
Retrieves a list of the payment actions available to perform for the specified return when a return results in a refund to the customer. <code><code> Return return = new Return(); string string = return.getAvailablePaymentActionsForReturn( returnId, paymentId); </code></code>
getAvailablePaymentActionsForReturn
{ "repo_name": "johngatti/mozu-java", "path": "mozu-java-core/src/main/java/com/mozu/api/resources/commerce/ReturnResource.java", "license": "mit", "size": 28030 }
[ "com.mozu.api.MozuClient", "java.util.List" ]
import com.mozu.api.MozuClient; import java.util.List;
import com.mozu.api.*; import java.util.*;
[ "com.mozu.api", "java.util" ]
com.mozu.api; java.util;
2,315,390
public boolean areaNeedSync(ARPluginContext context) throws ARException { return false; }
boolean function(ARPluginContext context) throws ARException { return false; }
/** * Plugin synchronization with the plugin server. * * @param context plugin context * @return true is the plugin was synch * @throws ARException in case of any error will occur */
Plugin synchronization with the plugin server
areaNeedSync
{ "repo_name": "stefandmn/AREasy", "path": "src/java/org/areasy/runtime/plugins/area/RuntimeArea.java", "license": "lgpl-3.0", "size": 8928 }
[ "com.bmc.arsys.api.ARException", "com.bmc.arsys.pluginsvr.plugins.ARPluginContext" ]
import com.bmc.arsys.api.ARException; import com.bmc.arsys.pluginsvr.plugins.ARPluginContext;
import com.bmc.arsys.api.*; import com.bmc.arsys.pluginsvr.plugins.*;
[ "com.bmc.arsys" ]
com.bmc.arsys;
322,426