language
stringclasses 2
values | func_code_string
stringlengths 63
466k
|
---|---|
python | def create_training_example(self,
environment_id,
collection_id,
query_id,
document_id=None,
cross_reference=None,
relevance=None,
**kwargs):
"""
Add example to training data query.
Adds a example to this training data query.
:param str environment_id: The ID of the environment.
:param str collection_id: The ID of the collection.
:param str query_id: The ID of the query used for training.
:param str document_id: The document ID associated with this training example.
:param str cross_reference: The cross reference associated with this training
example.
:param int relevance: The relevance of the training example.
:param dict headers: A `dict` containing the request headers
:return: A `DetailedResponse` containing the result, headers and HTTP status code.
:rtype: DetailedResponse
"""
if environment_id is None:
raise ValueError('environment_id must be provided')
if collection_id is None:
raise ValueError('collection_id must be provided')
if query_id is None:
raise ValueError('query_id must be provided')
headers = {}
if 'headers' in kwargs:
headers.update(kwargs.get('headers'))
sdk_headers = get_sdk_headers('discovery', 'V1',
'create_training_example')
headers.update(sdk_headers)
params = {'version': self.version}
data = {
'document_id': document_id,
'cross_reference': cross_reference,
'relevance': relevance
}
url = '/v1/environments/{0}/collections/{1}/training_data/{2}/examples'.format(
*self._encode_path_vars(environment_id, collection_id, query_id))
response = self.request(
method='POST',
url=url,
headers=headers,
params=params,
json=data,
accept_json=True)
return response |
java | @Override
public Map<String, List<String>> get(URI uri, Map<String, List<String>> requestHeaders) throws IOException {
clearExpired(new Date());
ArrayList<String> cookies = new ArrayList<String>();
for (Cookie cookie : getCookies(uri)) {
cookies.add(cookie.getName() + "=" + cookie.getValue());
}
return singletonMap(SM.COOKIE, singletonList(join(cookies, SEP)));
} |
java | public String[][] getPopupMap()
{
String string[][] = {
{RunProcessInField.LOCAL, "Run locally"},
{RunProcessInField.LOCAL_PROCESS, "Local process"},
{RunProcessInField.REMOTE_PROCESS, "Remote process"},
};
return string;
} |
java | @Indexable(type = IndexableType.REINDEX)
@Override
public CommerceNotificationTemplate addCommerceNotificationTemplate(
CommerceNotificationTemplate commerceNotificationTemplate) {
commerceNotificationTemplate.setNew(true);
return commerceNotificationTemplatePersistence.update(commerceNotificationTemplate);
} |
java | @Override
public Object proceed() throws Exception {
Object rc = null;
//if there are more interceptors left in the chain, call the next one
if (nextInterceptor < interceptors.size()) {
rc = invokeNextInterceptor();
}
else {
//otherwise call proceed on the delegate InvocationContext
rc = delegateInvocationContext.proceed();
}
return rc;
} |
python | def convex_conj(self):
"""The convex conjugate functional of the group L1-norm."""
conj_exp = conj_exponent(self.pointwise_norm.exponent)
return IndicatorGroupL1UnitBall(self.domain, exponent=conj_exp) |
java | public TLSARecord getTLSARecord(URL url) {
String recordValue;
int port = url.getPort();
if (port == -1) {
port = url.getDefaultPort();
}
String tlsaRecordName = String.format("_%s._tcp.%s", port, DNSUtil.ensureDot(url.getHost()));
try {
recordValue = this.dnssecResolver.resolve(tlsaRecordName, Type.TLSA);
} catch (DNSSECException e) {
return null;
}
if (recordValue.equals("")) return null;
// Process TLSA Record
String[] tlsaValues = recordValue.split(" ");
if (tlsaValues.length != 4) return null;
try {
return new TLSARecord(
new Name(tlsaRecordName),
DClass.IN,
0,
Integer.parseInt(tlsaValues[0]),
Integer.parseInt(tlsaValues[1]),
Integer.parseInt(tlsaValues[2]),
BaseEncoding.base16().decode(tlsaValues[3])
);
} catch (TextParseException e) {
return null;
}
} |
java | private boolean isAnomaly(Instance instance, ActiveRule rule) {
//AMRUles is equipped with anomaly detection. If on, compute the anomaly value.
boolean isAnomaly = false;
if (this.noAnomalyDetection == false){
if (rule.getInstancesSeen() >= this.anomalyNumInstThreshold) {
isAnomaly = rule.isAnomaly(instance,
this.univariateAnomalyprobabilityThreshold,
this.multivariateAnomalyProbabilityThreshold,
this.anomalyNumInstThreshold);
}
}
return isAnomaly;
} |
java | @Nullable
private static File _getResourceSource (final String resource, final ClassLoader loader)
{
if (resource != null)
{
URL url;
if (loader != null)
{
url = loader.getResource (resource);
}
else
{
url = ClassLoader.getSystemResource (resource);
}
return UrlUtils.getResourceRoot (url, resource);
}
return null;
} |
java | public Observable<VirtualNetworkRuleInner> createOrUpdateAsync(String resourceGroupName, String accountName, String virtualNetworkRuleName, CreateOrUpdateVirtualNetworkRuleParameters parameters) {
return createOrUpdateWithServiceResponseAsync(resourceGroupName, accountName, virtualNetworkRuleName, parameters).map(new Func1<ServiceResponse<VirtualNetworkRuleInner>, VirtualNetworkRuleInner>() {
@Override
public VirtualNetworkRuleInner call(ServiceResponse<VirtualNetworkRuleInner> response) {
return response.body();
}
});
} |
python | def getOSDesc(interface, ext_list):
"""
Return an OS description header.
interface (int)
Related interface number.
ext_list (list of OSExtCompatDesc or OSExtPropDesc)
List of instances of extended descriptors.
"""
try:
ext_type, = {type(x) for x in ext_list}
except ValueError:
raise TypeError('Extensions of a single type are required.')
if issubclass(ext_type, OSExtCompatDesc):
wIndex = 4
kw = {
'b': OSDescHeaderBCount(
bCount=len(ext_list),
Reserved=0,
),
}
elif issubclass(ext_type, OSExtPropDescHead):
wIndex = 5
kw = {
'wCount': len(ext_list),
}
else:
raise TypeError('Extensions of unexpected type')
ext_list_type = ext_type * len(ext_list)
klass = type(
'OSDesc',
(OSDescHeader, ),
{
'_fields_': [
('ext_list', ext_list_type),
],
},
)
return klass(
interface=interface,
dwLength=ctypes.sizeof(klass),
bcdVersion=1,
wIndex=wIndex,
ext_list=ext_list_type(*ext_list),
**kw
) |
python | def _process_phenstatement(self, limit):
"""
The phenstatements are the genotype-to-phenotype associations,
in the context of an environment.
These are also curated to a publication. So we make oban associations,
adding the pubs as a source. We additionally add the internal key as
a comment for tracking purposes.
:param limit:
:return:
"""
if self.test_mode:
graph = self.testgraph
else:
graph = self.graph
model = Model(graph)
raw = '/'.join((self.rawdir, 'phenstatement'))
LOG.info("processing phenstatement")
line_counter = 0
with open(raw, 'r') as f:
filereader = csv.reader(f, delimiter='\t', quotechar='\"')
f.readline() # read the header row; skip
for line in filereader:
(phenstatement_id, genotype_id, environment_id, phenotype_id,
type_id, pub_id) = line
# 168549 166695 1 8507 60468 151256
# 168550 166695 1 8508 60468 151256
# 168551 166696 1 8509 60468 151256
# 168552 166696 1 8510 60468 151256
line_counter += 1
phenstatement_key = phenstatement_id
phenstatement_id = self._makeInternalIdentifier(
'phenstatement', phenstatement_key)
genotype_key = genotype_id
if self.test_mode and \
int(genotype_key) not in self.test_keys['genotype']:
continue
genotype_id = self.idhash['genotype'][genotype_key]
environment_key = environment_id
environment_id = self.idhash['environment'][environment_key]
phenotype_key = phenotype_id
phenotype_internal_id = self._makeInternalIdentifier(
'phenotype', phenotype_key) # TEMP
phenotype_internal_label = self.label_hash[
phenotype_internal_id]
phenotype_id = self.idhash['phenotype'][phenotype_key]
pub_key = pub_id
pub_id = self.idhash['publication'][pub_key]
# figure out if there is a relevant stage
assoc = G2PAssoc(graph, self.name, genotype_id, phenotype_id)
if phenotype_id in self.phenocv:
stages = set(
s for s in
self.phenocv[phenotype_id] if re.match(r'FBdv', s))
if len(stages) == 1:
s = stages.pop()
assoc.set_stage(s, s)
elif len(stages) > 1:
LOG.warning(
"There's more than one stage specified per "
"phenotype. I don't know what to do. %s",
str(stages))
non_stage_ids = self.phenocv[phenotype_id] - stages
LOG.debug('Other non-stage bits: %s', str(non_stage_ids))
# TODO do something with the other parts
# of a pheno-cv relationship
assoc.set_environment(environment_id)
# TODO check remove unspecified environments?
assoc.add_source(pub_id)
assoc.add_association_to_graph()
assoc_id = assoc.get_association_id()
model.addComment(assoc_id, phenstatement_id)
model.addDescription(assoc_id, phenotype_internal_label)
if not self.test_mode and limit is not None and line_counter > limit:
break
return |
java | public static Integer[] createAndInitializeArray(int size, Integer start) {
Integer[] result = new Integer[size];
for (int i = 0; i < result.length; i++) {
result[i] = start++;
}
return result;
} |
python | def sample_bitstrings(self, n_samples):
"""
Sample bitstrings from the distribution defined by the wavefunction.
:param n_samples: The number of bitstrings to sample
:return: An array of shape (n_samples, n_qubits)
"""
possible_bitstrings = np.array(list(itertools.product((0, 1), repeat=len(self))))
inds = np.random.choice(2 ** len(self), n_samples, p=self.probabilities())
bitstrings = possible_bitstrings[inds, :]
return bitstrings |
java | public <X extends Exception> boolean replaceAllIf(Try.Predicate<? super K, X> predicate, Collection<?> oldValues, E newValue) throws X {
boolean modified = false;
for (Map.Entry<K, V> entry : this.valueMap.entrySet()) {
if (predicate.test(entry.getKey())) {
if (entry.getValue().removeAll(oldValues)) {
entry.getValue().add(newValue);
modified = true;
}
}
}
return modified;
} |
java | @Override
public int compare(JTAResource o1, JTAResource o2) {
if (tc.isEntryEnabled())
Tr.entry(tc, "compare", new Object[] { o1, o2, this });
int result = 0;
int p1 = o1.getPriority();
int p2 = o2.getPriority();
if (p1 < p2)
result = 1;
else if (p1 > p2)
result = -1;
if (tc.isEntryEnabled())
Tr.exit(tc, "compare", result);
return result;
} |
java | public BlockMeta getBlockMeta(long blockId) throws BlockDoesNotExistException {
BlockMeta blockMeta = mBlockIdToBlockMap.get(blockId);
if (blockMeta == null) {
throw new BlockDoesNotExistException(ExceptionMessage.BLOCK_META_NOT_FOUND, blockId);
}
return blockMeta;
} |
java | protected void arrangePreparedAccessContext(LaJobRuntime runtime) {
if (accessContextArranger == null) {
return;
}
final AccessContextResource resource = createAccessContextResource(runtime);
final AccessContext context = accessContextArranger.arrangePreparedAccessContext(resource);
if (context == null) {
String msg = "Cannot return null from access context arranger: " + accessContextArranger + " runtime=" + runtime;
throw new IllegalStateException(msg);
}
PreparedAccessContext.setAccessContextOnThread(context);
} |
java | public void copyCategories(CmsObject cms, CmsResource fromResource, String toResourceSitePath) throws CmsException {
List<CmsCategory> categories = readResourceCategories(cms, fromResource);
for (CmsCategory category : categories) {
addResourceToCategory(cms, toResourceSitePath, category);
}
} |
java | @Override
public String asString() {
ByteArrayOutputStream result = new ByteArrayOutputStream(getIntContentLength(this.response));
saveTo(result);
return result.toString();
} |
java | public static Map createHeaderMap(PageContext pContext)
{
final HttpServletRequest request =
(HttpServletRequest) pContext.getRequest();
return new EnumeratedMap()
{
public Enumeration enumerateKeys()
{
return request.getHeaderNames();
}
public Object getValue(Object pKey)
{
if (pKey instanceof String) {
return request.getHeader((String) pKey);
} else {
return null;
}
}
public boolean isMutable()
{
return false;
}
};
} |
python | def edges(self, source=None, relation=None, target=None):
"""
Return edges filtered by their *source*, *relation*, or *target*.
Edges don't include terminal triples (node types or attributes).
"""
edgematch = lambda e: (
(source is None or source == e.source) and
(relation is None or relation == e.relation) and
(target is None or target == e.target)
)
variables = self.variables()
edges = [t for t in self._triples if t.target in variables]
return list(filter(edgematch, edges)) |
java | public void setOutputFile (final File value)
{
if (value != null && !value.isAbsolute ())
{
throw new IllegalArgumentException ("path is not absolute: " + value);
}
this.outputFile = value;
} |
java | public void serviceName_account_userPrincipalName_PUT(String serviceName, String userPrincipalName, OvhAccount body) throws IOException {
String qPath = "/msServices/{serviceName}/account/{userPrincipalName}";
StringBuilder sb = path(qPath, serviceName, userPrincipalName);
exec(qPath, "PUT", sb.toString(), body);
} |
java | public boolean marketplace_removeListing(Long listingId, CharSequence status)
throws FacebookException, IOException {
assert MARKETPLACE_STATUS_DEFAULT.equals(status) || MARKETPLACE_STATUS_SUCCESS.equals(status) ||
MARKETPLACE_STATUS_NOT_SUCCESS.equals(status) : "Invalid status: " + status;
T result =
this.callMethod(FacebookMethod.MARKETPLACE_REMOVE_LISTING, new Pair<String, CharSequence>("listing_id",
listingId.toString()),
new Pair<String, CharSequence>("status", status));
return this.extractBoolean(result);
} |
python | def geometric_partitions(iterable, floor=1, ceiling=32768):
'''
Partition an iterable into chunks. Returns an iterator over partitions.
'''
partition_size = floor
run_length = multiprocessing.cpu_count()
run_count = 0
try:
while True:
#print("partition_size =", partition_size)
# Split the iterable and replace the original iterator to avoid
# advancing it
partition, iterable = itertools.tee(iterable)
# Yield the first partition, limited to the partition size
yield Queryable(partition).take(partition_size)
# Advance to the start of the next partition, this will raise
# StopIteration if the iterator is exhausted
for i in range(partition_size):
next(iterable)
# If we've reached the end of a run of this size, double the
# partition size
run_count += 1
if run_count >= run_length:
partition_size *= 2
run_count = 0
# Unless we have hit the ceiling
if partition_size > ceiling:
partition_size = ceiling
except StopIteration:
pass |
python | def stage_pywbem_args(self, method, **kwargs):
"""
Set requst method and all args.
Normally called before the cmd is executed to record request
parameters
"""
# pylint: disable=attribute-defined-outside-init
self._pywbem_method = method
self._pywbem_args = kwargs |
python | def star(self, login, repo):
"""Star to login/repo
:param str login: (required), owner of the repo
:param str repo: (required), name of the repo
:return: bool
"""
resp = False
if login and repo:
url = self._build_url('user', 'starred', login, repo)
resp = self._boolean(self._put(url), 204, 404)
return resp |
java | private static void weaveDNAStrands(LinkedList<DNASegment> segments, int firstJulianDay,
HashMap<Integer, DNAStrand> strands, int top, int bottom, int[] dayXs) {
// First, get rid of any colors that ended up with no segments
Iterator<DNAStrand> strandIterator = strands.values().iterator();
while (strandIterator.hasNext()) {
DNAStrand strand = strandIterator.next();
if (strand.count < 1 && strand.allDays == null) {
strandIterator.remove();
continue;
}
strand.points = new float[strand.count * 4];
strand.position = 0;
}
// Go through each segment and compute its points
for (DNASegment segment : segments) {
// Add the points to the strand of that color
DNAStrand strand = strands.get(segment.color);
int dayIndex = segment.day - firstJulianDay;
int dayStartMinute = segment.startMinute % DAY_IN_MINUTES;
int dayEndMinute = segment.endMinute % DAY_IN_MINUTES;
int height = bottom - top;
int workDayHeight = height * 3 / 4;
int remainderHeight = (height - workDayHeight) / 2;
int x = dayXs[dayIndex];
int y0 = 0;
int y1 = 0;
y0 = top + getPixelOffsetFromMinutes(dayStartMinute, workDayHeight, remainderHeight);
y1 = top + getPixelOffsetFromMinutes(dayEndMinute, workDayHeight, remainderHeight);
if (DEBUG) {
Log.d(TAG, "Adding " + Integer.toHexString(segment.color) + " at x,y0,y1: " + x
+ " " + y0 + " " + y1 + " for " + dayStartMinute + " " + dayEndMinute);
}
strand.points[strand.position++] = x;
strand.points[strand.position++] = y0;
strand.points[strand.position++] = x;
strand.points[strand.position++] = y1;
}
} |
python | def geoid(self):
""""Return first child of the column, or self that is marked as a geographic identifier"""
if self.valuetype_class.is_geoid():
return self
for c in self.table.columns:
if c.parent == self.name and c.valuetype_class.is_geoid():
return c |
java | List<PartitionInstance> getAbsent() {
List<PartitionInstance> absentPartitions = new ArrayList<>();
for (Map.Entry<Integer, List<PartitionInstance>> e : nodeIndexToHostedPartitions.entrySet()) {
if (e.getKey() < 0) {
absentPartitions.addAll(e.getValue());
}
}
return absentPartitions;
} |
java | public static void parseLongestText(String text, AhoCorasickDoubleArrayTrie.IHit<CoreDictionary.Attribute> processor)
{
if (trie != null)
{
final int[] lengthArray = new int[text.length()];
final CoreDictionary.Attribute[] attributeArray = new CoreDictionary.Attribute[text.length()];
char[] charArray = text.toCharArray();
DoubleArrayTrie<CoreDictionary.Attribute>.Searcher searcher = dat.getSearcher(charArray, 0);
while (searcher.next())
{
lengthArray[searcher.begin] = searcher.length;
attributeArray[searcher.begin] = searcher.value;
}
trie.parseText(charArray, new AhoCorasickDoubleArrayTrie.IHit<CoreDictionary.Attribute>()
{
@Override
public void hit(int begin, int end, CoreDictionary.Attribute value)
{
int length = end - begin;
if (length > lengthArray[begin])
{
lengthArray[begin] = length;
attributeArray[begin] = value;
}
}
});
for (int i = 0; i < charArray.length;)
{
if (lengthArray[i] == 0)
{
++i;
}
else
{
processor.hit(i, i + lengthArray[i], attributeArray[i]);
i += lengthArray[i];
}
}
}
else
dat.parseLongestText(text, processor);
} |
python | def interpoled_resampling(W, x):
"""Resampling based on an interpolated CDF, as described in Malik and Pitt.
Parameters
----------
W: (N,) array
weights
x: (N,) array
particles
Returns
-------
xrs: (N,) array
the resampled particles
"""
N = W.shape[0]
idx = np.argsort(x)
xs = x[idx]
ws = W[idx]
cs = np.cumsum(avg_n_nplusone(ws))
u = random.rand(N)
xrs = np.empty(N)
where = np.searchsorted(cs, u)
# costs O(N log(N)) but algorithm has O(N log(N)) complexity anyway
for n in range(N):
m = where[n]
if m==0:
xrs[n] = xs[0]
elif m==N:
xrs[n] = xs[-1]
else:
xrs[n] = interpol(cs[m-1], cs[m], xs[m-1], xs[m], u[n])
return xrs |
python | def _construct_sharded(self):
"""Construct command line strings for a sharded cluster."""
current_version = self.getMongoDVersion()
num_mongos = self.args['mongos'] if self.args['mongos'] > 0 else 1
shard_names = self._get_shard_names(self.args)
# create shards as stand-alones or replica sets
nextport = self.args['port'] + num_mongos
for shard in shard_names:
if (self.args['single'] and
LooseVersion(current_version) >= LooseVersion("3.6.0")):
errmsg = " \n * In MongoDB 3.6 and above a Shard must be " \
"made up of a replica set. Please use --replicaset " \
"option when starting a sharded cluster.*"
raise SystemExit(errmsg)
elif (self.args['single'] and
LooseVersion(current_version) < LooseVersion("3.6.0")):
self.shard_connection_str.append(
self._construct_single(
self.dir, nextport, name=shard, extra='--shardsvr'))
nextport += 1
elif self.args['replicaset']:
self.shard_connection_str.append(
self._construct_replset(
self.dir, nextport, shard,
num_nodes=list(range(self.args['nodes'])),
arbiter=self.args['arbiter'], extra='--shardsvr'))
nextport += self.args['nodes']
if self.args['arbiter']:
nextport += 1
# start up config server(s)
config_string = []
# SCCC config servers (MongoDB <3.3.0)
if not self.args['csrs'] and self.args['config'] >= 3:
config_names = ['config1', 'config2', 'config3']
else:
config_names = ['config']
# CSRS config servers (MongoDB >=3.1.0)
if self.args['csrs']:
config_string.append(self._construct_config(self.dir, nextport,
"configRepl", True))
else:
for name in config_names:
self._construct_config(self.dir, nextport, name)
config_string.append('%s:%i' % (self.args['hostname'],
nextport))
nextport += 1
# multiple mongos use <datadir>/mongos/ as subdir for log files
if num_mongos > 1:
mongosdir = os.path.join(self.dir, 'mongos')
if not os.path.exists(mongosdir):
if self.args['verbose']:
print("creating directory: %s" % mongosdir)
os.makedirs(mongosdir)
# start up mongos, but put them to the front of the port range
nextport = self.args['port']
for i in range(num_mongos):
if num_mongos > 1:
mongos_logfile = 'mongos/mongos_%i.log' % nextport
else:
mongos_logfile = 'mongos.log'
self._construct_mongos(os.path.join(self.dir, mongos_logfile),
nextport, ','.join(config_string))
nextport += 1 |
python | def are_worth_chaining(parser, to_type: Type[S], converter: Converter[S, T]) -> bool:
"""
Utility method to check if it makes sense to chain this parser with the given destination type, and the given
converter to create a parsing chain. Returns True if it brings value to chain them.
To bring value,
* the converter's output should not be a parent class of the parser's output. Otherwise
the chain does not even make any progress :)
* The parser has to allow chaining (with converter.can_chain=True)
:param parser:
:param to_type:
:param converter:
:return:
"""
if not parser.can_chain:
# The base parser prevents chaining
return False
elif not is_any_type(to_type) and is_any_type(converter.to_type):
# we gain the capability to generate any type. So it is interesting.
return True
elif issubclass(to_type, converter.to_type):
# Not interesting : the outcome of the chain would be not better than one of the parser alone
return False
# Note: we dont say that chaining a generic parser with a converter is useless. Indeed it might unlock some
# capabilities for the user (new file extensions, etc.) that would not be available with the generic parser
# targetting to_type alone. For example parsing object A from its constructor then converting A to B might
# sometimes be interesting, rather than parsing B from its constructor
else:
# Interesting
return True |
java | private String[] superTypes( ObjectType cmisType ) {
if (cmisType.getBaseTypeId() == BaseTypeId.CMIS_FOLDER) {
return new String[] {JcrConstants.NT_FOLDER};
}
if (cmisType.getBaseTypeId() == BaseTypeId.CMIS_DOCUMENT) {
return new String[] {JcrConstants.NT_FILE};
}
return new String[] {cmisType.getParentType().getId()};
} |
java | public static Version parse(String version) {
Assert.hasText(version, "Version must not be null o empty!");
String[] parts = version.trim().split("\\.");
int[] intParts = new int[parts.length];
for (int i = 0; i < parts.length; i++) {
String input = i == parts.length - 1 ? parts[i].replaceAll("\\D.*", "") : parts[i];
if (StringUtils.hasText(input)) {
try {
intParts[i] = Integer.parseInt(input);
} catch (IllegalArgumentException o_O) {
throw new IllegalArgumentException(String.format(VERSION_PARSE_ERROR, input, version), o_O);
}
}
}
return new Version(intParts);
} |
java | public static String jsonNameVal(final String indent,
final String name,
final String val) {
StringBuilder sb = new StringBuilder();
sb.append(indent);
sb.append("\"");
sb.append(name);
sb.append("\": ");
if (val != null) {
sb.append(jsonEncode(val));
}
return sb.toString();
} |
java | public static <T> QBean<T> fields(Path<? extends T> type, Map<String, ? extends Expression<?>> bindings) {
return new QBean<T>(type.getType(), true, bindings);
} |
python | def parse_value(self, text: str) -> Optional[bool]:
"""Parse boolean value.
Args:
text: String representation of the value.
"""
if text == "true":
return True
if text == "false":
return False |
java | public static boolean isPublicStatic(Field f) {
final int modifiers = f.getModifiers();
return Modifier.isPublic(modifiers) && Modifier.isStatic(modifiers);
} |
java | @Override
@LogarithmicTime(amortized = true)
public Handle<K, V> insert(K key) {
return insert(key, null);
} |
java | public static File create(FileCreateParams params, RequestOptions options)
throws StripeException {
checkNullTypedParams(classUrl(File.class, Stripe.getUploadBase()), params);
return create(params.toMap(), options);
} |
java | protected void doExecuteCommand() {
if (this.commandExecutor instanceof ParameterizableActionCommandExecutor) {
((ParameterizableActionCommandExecutor) this.commandExecutor).execute(getParameters());
}
else {
if (this.commandExecutor != null) {
this.commandExecutor.execute();
}
}
} |
python | def tf(cluster):
"""
Computes the term frequency and stores it as a dictionary
:param cluster: the cluster that contains the metadata
:return: tf dictionary
"""
counts = dict()
words = cluster.split(' ')
for word in words:
counts[word] = counts.get(word, 0) + 1
return counts |
python | def _get_value(self, var):
"""Return value of variable in solution."""
return self._problem._p.get_value(self._problem._variables[var]) |
java | int numNodes(Block b) {
BlockInfo info = blocks.get(b);
return info == null ? 0 : info.numNodes();
} |
java | public Response put(Session session, String path, InputStream inputStream, String fileNodeType,
String contentNodeType, List<String> mixins, List<String> tokens, MultivaluedMap<String, String> allowedAutoVersionPath)
{
try
{
Node node = null;
boolean isVersioned;
try
{
node = (Node)session.getItem(path);
}
catch (PathNotFoundException pexc)
{
nullResourceLocks.checkLock(session, path, tokens);
}
if (node == null)
{
node = session.getRootNode().addNode(TextUtil.relativizePath(path, false), fileNodeType);
// We set the new path
path = node.getPath();
node.addNode("jcr:content", contentNodeType);
updateContent(node, inputStream, mixins);
isVersioned = isVersionSupported(node.getPath(), session.getWorkspace().getName(), allowedAutoVersionPath);
if (isVersioned && node.canAddMixin(VersionHistoryUtils.MIX_VERSIONABLE))
{
node.addMixin(VersionHistoryUtils.MIX_VERSIONABLE);
}
node.getSession().save();
}
else
{
isVersioned = isVersionSupported(node.getPath(), session.getWorkspace().getName(), allowedAutoVersionPath);
if (isVersioned)
{
VersionHistoryUtils.createVersion(node);
updateContent(node, inputStream, mixins);
}
else
{
updateContent(node, inputStream, mixins);
}
}
session.save();
}
catch (LockException exc)
{
if(log.isDebugEnabled())
{
log.debug(exc.getMessage(), exc);
}
return Response.status(HTTPStatus.LOCKED).entity(exc.getMessage()).build();
}
catch (AccessDeniedException exc)
{
if(log.isDebugEnabled())
{
log.debug(exc.getMessage(), exc);
}
return Response.status(HTTPStatus.FORBIDDEN).entity(exc.getMessage()).build();
}
catch (RepositoryException exc)
{
if(log.isDebugEnabled())
{
log.debug(exc.getMessage(), exc);
}
return Response.status(HTTPStatus.CONFLICT).entity(exc.getMessage()).build();
}
catch (Exception exc)
{
if(log.isDebugEnabled())
{
log.debug(exc.getMessage(), exc);
}
return Response.status(HTTPStatus.CONFLICT).entity(exc.getMessage()).build();
}
if (uriBuilder != null)
{
return Response.created(uriBuilder.path(session.getWorkspace().getName()).path(path).build()).build();
}
// to save compatibility if uriBuilder is not provided
return Response.status(HTTPStatus.CREATED).build();
} |
java | public ListZonesResponse listZones(AbstractBceRequest request) {
InternalRequest internalRequest = this.createRequest(request, HttpMethodName.GET, ZONE);
return invokeHttpClient(internalRequest, ListZonesResponse.class);
} |
python | def outputWord(self):
"""Output report to word docx
"""
import docx
from docx.enum.text import WD_ALIGN_PARAGRAPH
doc = docx.Document()
doc.styles['Normal'].paragraph_format.alignment = WD_ALIGN_PARAGRAPH.JUSTIFY
doc.add_heading(self.title, level=0)
if self.addTime:
from time import localtime, strftime
doc.add_heading(strftime("%Y-%m-%d %H:%M:%S", localtime()), level=1)
# Append introduction
if self.p:
doc.add_heading('Introduction',level=1)
for p in renewliner(self.p).split('\n'):
doc.add_paragraph(p)
# Sections
c = count(1)
#Prepare fig and table numbers
self.listFigures(tuple())
self.listTables(tuple())
for section in self.sections:
section.sectionsWord((next(c),),doc=doc)
# Append conclusion
if self.conclusion:
doc.add_heading('Conclusion', level=1)
for p in renewliner(self.conclusion).split('\n'):
doc.add_paragraph(p)
# Generate Word document
doc.save(self.outfile+'.docx') |
java | public void sendWithBulkProcessor(IndexRequest indexRequest) {
Optional.ofNullable(this.bulkProcessor)
.orElseGet(() -> setAndReturnBulkProcessor(BulkProcessor
.builder(jmESClient, bulkProcessorListener).build()))
.add(indexRequest);
} |
java | public void writeHeadersTo(final CacheResponse pResponse) {
String[] headers = getHeaderNames();
for (String header : headers) {
// HACK...
// Strip away internal headers
if (HTTPCache.HEADER_CACHED_TIME.equals(header)) {
continue;
}
// TODO: Replace Last-Modified with X-Cached-At? See CachedEntityImpl
String[] headerValues = getHeaderValues(header);
for (int i = 0; i < headerValues.length; i++) {
String headerValue = headerValues[i];
if (i == 0) {
pResponse.setHeader(header, headerValue);
}
else {
pResponse.addHeader(header, headerValue);
}
}
}
} |
python | def Address(self):
"""
Get the wallet address associated with the token.
Returns:
str: base58 encoded string representing the wallet address.
"""
if self._address is None:
self._address = Crypto.ToAddress(self.ScriptHash)
return self._address |
java | public synchronized void removeObserver(final String notification, final ApptentiveNotificationObserver observer) {
final ApptentiveNotificationObserverList list = findObserverList(notification);
if (list != null) {
list.removeObserver(observer);
}
} |
java | protected static Variable findTargetVariable(VariableExpression ve) {
final Variable accessedVariable = ve.getAccessedVariable() != null ? ve.getAccessedVariable() : ve;
if (accessedVariable != ve) {
if (accessedVariable instanceof VariableExpression)
return findTargetVariable((VariableExpression) accessedVariable);
}
return accessedVariable;
} |
java | public static DiscreteLogLinearFactor fromFeatureGeneratorSparse(DiscreteFactor factor,
FeatureGenerator<Assignment, String> featureGen) {
Iterator<Outcome> iter = factor.outcomeIterator();
Set<String> featureNames = Sets.newHashSet();
while (iter.hasNext()) {
Outcome o = iter.next();
for (Map.Entry<String, Double> f : featureGen.generateFeatures(o.getAssignment()).entrySet()) {
featureNames.add(f.getKey());
}
}
List<String> featureNamesSorted = Lists.newArrayList(featureNames);
Collections.sort(featureNamesSorted);
DiscreteVariable featureType = new DiscreteVariable(FEATURE_VAR_NAME, featureNamesSorted);
VariableNumMap featureVar = VariableNumMap.singleton(
Ints.max(factor.getVars().getVariableNumsArray()) + 1, FEATURE_VAR_NAME, featureType);
TableFactorBuilder builder = new TableFactorBuilder(
factor.getVars().union(featureVar), SparseTensorBuilder.getFactory());
iter = factor.outcomeIterator();
while (iter.hasNext()) {
Outcome o = iter.next();
for (Map.Entry<String, Double> f : featureGen.generateFeatures(o.getAssignment()).entrySet()) {
Assignment featureAssignment = featureVar.outcomeArrayToAssignment(f.getKey());
builder.setWeight(o.getAssignment().union(featureAssignment), f.getValue());
}
}
DiscreteFactor featureFactor = builder.build();
return new DiscreteLogLinearFactor(factor.getVars(), featureVar, featureFactor, factor);
} |
python | def id(self, peer=None, **kwargs):
"""Shows IPFS Node ID info.
Returns the PublicKey, ProtocolVersion, ID, AgentVersion and
Addresses of the connected daemon or some other node.
.. code-block:: python
>>> c.id()
{'ID': 'QmVgNoP89mzpgEAAqK8owYoDEyB97MkcGvoWZir8otE9Uc',
'PublicKey': 'CAASpgIwggEiMA0GCSqGSIb3DQEBAQUAA4IBDwAwggE … BAAE=',
'AgentVersion': 'go-libp2p/3.3.4',
'ProtocolVersion': 'ipfs/0.1.0',
'Addresses': [
'/ip4/127.0.0.1/tcp/4001/ipfs/QmVgNoP89mzpgEAAqK8owYo … E9Uc',
'/ip4/10.1.0.172/tcp/4001/ipfs/QmVgNoP89mzpgEAAqK8owY … E9Uc',
'/ip4/172.18.0.1/tcp/4001/ipfs/QmVgNoP89mzpgEAAqK8owY … E9Uc',
'/ip6/::1/tcp/4001/ipfs/QmVgNoP89mzpgEAAqK8owYoDEyB97 … E9Uc',
'/ip6/fccc:7904:b05b:a579:957b:deef:f066:cad9/tcp/400 … E9Uc',
'/ip6/fd56:1966:efd8::212/tcp/4001/ipfs/QmVgNoP89mzpg … E9Uc',
'/ip6/fd56:1966:efd8:0:def1:34d0:773:48f/tcp/4001/ipf … E9Uc',
'/ip6/2001:db8:1::1/tcp/4001/ipfs/QmVgNoP89mzpgEAAqK8 … E9Uc',
'/ip4/77.116.233.54/tcp/4001/ipfs/QmVgNoP89mzpgEAAqK8 … E9Uc',
'/ip4/77.116.233.54/tcp/10842/ipfs/QmVgNoP89mzpgEAAqK … E9Uc']}
Parameters
----------
peer : str
Peer.ID of the node to look up (local node if ``None``)
Returns
-------
dict : Information about the IPFS node
"""
args = (peer,) if peer is not None else ()
return self._client.request('/id', args, decoder='json', **kwargs) |
python | def peakdelta(v, delta, x=None):
"""
Returns two arrays
function [maxtab, mintab]=peakdelta(v, delta, x)
%PEAKDET Detect peaks in a vector
% [MAXTAB, MINTAB] = peakdelta(V, DELTA) finds the local
% maxima and minima ("peaks") in the vector V.
% MAXTAB and MINTAB consists of two columns. Column 1
% contains indices in V, and column 2 the found values.
%
% With [MAXTAB, MINTAB] = peakdelta(V, DELTA, X) the indices
% in MAXTAB and MINTAB are replaced with the corresponding
% X-values.
%
% A point is considered a maximum peak if it has the maximal
% value, and was preceded (to the left) by a value lower by
% DELTA.
% Eli Billauer, 3.4.05 (Explicitly not copyrighted).
% This function is released to the public domain; Any use is allowed.
"""
maxtab = []
mintab = []
if x is None:
x = arange(len(v))
v = asarray(v)
if len(v) != len(x):
sys.exit('Input vectors v and x must have same length')
if not isscalar(delta):
sys.exit('Input argument delta must be a scalar')
if delta <= 0:
sys.exit('Input argument delta must be positive')
mn, mx = Inf, -Inf
mnpos, mxpos = NaN, NaN
lookformax = True
for i in arange(len(v)):
this = v[i]
if this > mx:
mx = this
mxpos = x[i]
if this < mn:
mn = this
mnpos = x[i]
if lookformax:
if this < mx - delta:
maxtab.append((mxpos, mx))
mn = this
mnpos = x[i]
lookformax = False
else:
if this > mn + delta:
mintab.append((mnpos, mn))
mx = this
mxpos = x[i]
lookformax = True
return array(maxtab), array(mintab) |
python | def table(self):
"""Return a large string of the entire table ready to be printed to the terminal."""
dimensions = max_dimensions(self.table_data, self.padding_left, self.padding_right)[:3]
return flatten(self.gen_table(*dimensions)) |
java | public void setSearchableInfo(SearchableInfo searchable) {
mSearchable = searchable;
if (mSearchable != null) {
updateSearchAutoComplete();
updateQueryHint();
}
// Cache the voice search capability
mVoiceButtonEnabled = hasVoiceSearch();
if (mVoiceButtonEnabled) {
// Disable the microphone on the keyboard, as a mic is displayed near the text box
// TODO: use imeOptions to disable voice input when the new API will be available
mQueryTextView.setPrivateImeOptions(IME_OPTION_NO_MICROPHONE);
}
updateViewsVisibility(isIconified());
} |
java | protected PrintJob createJob(final PrintJobEntry entry) {
//CHECKSTYLE:ON
PrintJob job = this.context.getBean(PrintJob.class);
job.setEntry(entry);
job.setSecurityContext(SecurityContextHolder.getContext());
return job;
} |
python | def _xpathDict(xml, xpath, cls, parent, **kwargs):
""" Returns a default Dict given certain information
:param xml: An xml tree
:type xml: etree
:param xpath: XPath to find children
:type xpath: str
:param cls: Class identifying children
:type cls: inventory.Resource
:param parent: Parent of object
:type parent: CtsCollection
:rtype: collections.defaultdict.<basestring, inventory.Resource>
:returns: Dictionary of children
"""
children = []
for child in xml.xpath(xpath, namespaces=XPATH_NAMESPACES):
children.append(cls.parse(
resource=child,
parent=parent,
**kwargs
))
return children |
python | def delete_asset_content(self, asset_content_id):
"""Deletes content from an ``Asset``.
arg: asset_content_id (osid.id.Id): the ``Id`` of the
``AssetContent``
raise: NotFound - ``asset_content_id`` is not found
raise: NullArgument - ``asset_content_id`` is ``null``
raise: OperationFailed - unable to complete request
raise: PermissionDenied - authorization failure
*compliance: mandatory -- This method must be implemented.*
"""
# Implemented from template for
# osid.repository.AssetAdminSession.delete_asset_content_template
from dlkit.abstract_osid.id.primitives import Id as ABCId
from .objects import AssetContent
collection = JSONClientValidated('repository',
collection='Asset',
runtime=self._runtime)
if not isinstance(asset_content_id, ABCId):
raise errors.InvalidArgument('the argument is not a valid OSID Id')
asset = collection.find_one({'assetContents._id': ObjectId(asset_content_id.get_identifier())})
index = 0
found = False
for i in asset['assetContents']:
if i['_id'] == ObjectId(asset_content_id.get_identifier()):
asset_content_map = asset['assetContents'].pop(index)
index += 1
found = True
if not found:
raise errors.OperationFailed()
AssetContent(
osid_object_map=asset_content_map,
runtime=self._runtime,
proxy=self._proxy)._delete()
collection.save(asset) |
java | private String[] splitSeqName(String sequenceName) {
String[] result = new String[3];
String[] barSplit = sequenceName.split("/");
if (barSplit.length == 2) {
result[0] = barSplit[0];
String[] positions = barSplit[1].split("-");
if (positions.length == 2) {
result[1] = positions[0];
result[2] = positions[1];
}
} else {
result[0] = sequenceName;
result[1] = null;
result[2] = null;
}
return result;
} |
python | def formatted(self):
"""str: The BIC separated in the blocks bank-, country- and location-code."""
formatted = ' '.join([self.bank_code, self.country_code, self.location_code])
if self.branch_code:
formatted += ' ' + self.branch_code
return formatted |
python | def run(self, context=None, stdout=None, stderr=None):
"Like execute, but records a skip if the should_skip method returns True."
if self.should_skip():
self._record_skipped_example(self.formatter)
self.num_skipped += 1
else:
self.execute(context, stdout, stderr)
return self.num_successes, self.num_failures, self.num_skipped |
python | def buy_open_order_quantity(self):
"""
[int] 买方向挂单量
"""
return sum(order.unfilled_quantity for order in self.open_orders if
order.side == SIDE.BUY and order.position_effect == POSITION_EFFECT.OPEN) |
python | def write_single_coil(self, starting_address, value):
"""
Write single Coil to Master device (Function code 5)
starting_address: Coil to be written
value: Coil Value to be written
"""
self.__transactionIdentifier+=1
if (self.__ser is not None):
if (self.__ser.closed):
raise Exception.SerialPortNotOpenedException("serial port not opened")
function_code = 5
length = 6;
transaction_identifier_lsb = self.__transactionIdentifier&0xFF
transaction_identifier_msb = ((self.__transactionIdentifier&0xFF00) >> 8)
length_lsb = length&0xFF
length_msb = (length&0xFF00) >> 8
starting_address_lsb = starting_address&0xFF
starting_address_msb = (starting_address&0xFF00) >> 8
if value:
valueLSB = 0x00
valueMSB = (0xFF00) >> 8
else:
valueLSB = 0x00
valueMSB = (0x00) >> 8
if (self.__ser is not None):
data = bytearray([self.__unitIdentifier, function_code, starting_address_msb, starting_address_lsb, valueMSB, valueLSB, 0, 0])
crc = self.__calculateCRC(data, len(data)-2, 0)
crcLSB = crc&0xFF
crcMSB = (crc&0xFF00) >> 8
data[6] = crcLSB
data[7] = crcMSB
self.__ser.write(data)
bytes_to_read = 8
data = self.__ser.read(bytes_to_read)
b=bytearray(data)
data = b
if (len(data) < bytes_to_read):
raise Exceptions.TimeoutError('Read timeout Exception')
if ((data[1] == 0x85) & (data[2] == 0x01)):
raise Exceptions.function_codeNotSupportedException("Function code not supported by master");
if ((data[1] == 0x85) & (data[2] == 0x02)):
raise Exceptions.starting_addressInvalidException("Address invalid");
if ((data[1] == 0x85) & (data[2] == 0x03)):
raise Exceptions.QuantityInvalidException("Value invalid");
if ((data[1] == 0x85) & (data[2] == 0x04)):
raise Exceptions.ModbusException("error reading");
crc = self.__calculateCRC(data, len(data) - 2, 0)
crcLSB = crc&0xFF
crcMSB = (crc&0xFF00) >> 8
if ((crcLSB != data[len(data)-2]) & (crcMSB != data[len(data)-1])):
raise Exceptions.CRCCheckFailedException("CRC check failed");
if data[1] == self.__unitIdentifier:
return True
else:
return False
else:
protocolIdentifierLSB = 0x00;
protocolIdentifierMSB = 0x00;
length_lsb = 0x06;
length_msb = 0x00;
data = bytearray([transaction_identifier_msb, transaction_identifier_lsb, protocolIdentifierMSB, protocolIdentifierLSB, length_msb, length_lsb, self.__unitIdentifier, function_code, starting_address_msb, starting_address_lsb, valueMSB, valueLSB])
self.__tcpClientSocket.send(data)
bytes_to_read = 12
self.__receivedata = bytearray()
try:
while (len(self.__receivedata) == 0):
pass
except Exception:
raise Exception('Read Timeout')
data = bytearray(self.__receivedata)
if ((data[1+6] == 0x85) & (data[2+6] == 0x01)):
raise Exceptions.function_codeNotSupportedException("Function code not supported by master");
if ((data[1+6] == 0x85) & (data[2+6] == 0x02)):
raise Exceptions.starting_addressInvalidException("Address invalid");
if ((data[1+6] == 0x85) & (data[2+6] == 0x03)):
raise Exceptions.QuantityInvalidException("Value invalid");
if ((data[1+6] == 0x85) & (data[2+6] == 0x04)):
raise Exceptions.ModbusException("error reading");
return True |
java | public static byte[] encryptAES2Base64(byte[] data, byte[] key) {
try {
return Base64.getEncoder().encode(encryptAES(data, key));
} catch (Exception e) {
return null;
}
} |
python | def parse_response(self, response):
"""
Parse the response and build a `scanboo_common.http_client.HttpResponse` object.
For successful responses, convert the json data into a dict.
:param response: the `requests` response
:return: [HttpResponse] response object
"""
status = response.status_code
if response.ok:
data = response.json()
return HttpResponse(ok=response.ok, status=status, errors=None, data=data)
else:
try:
errors = response.json()
except ValueError:
errors = response.content
return HttpResponse(ok=response.ok, status=status, errors=errors, data=None) |
java | private void processEnumJavadoc(Options.OptionInfo oi) {
Enum<?>[] constants = (Enum<?>[]) oi.baseType.getEnumConstants();
if (constants == null) {
return;
}
oi.enumJdoc = new LinkedHashMap<>();
for (Enum<?> constant : constants) {
assert oi.enumJdoc != null : "@AssumeAssertion(nullness): bug in flow?";
oi.enumJdoc.put(constant.name(), "");
}
ClassDoc enumDoc = root.classNamed(oi.baseType.getName());
if (enumDoc == null) {
return;
}
assert oi.enumJdoc != null : "@AssumeAssertion(nullness): bug in flow?";
for (String name : oi.enumJdoc.keySet()) {
for (FieldDoc fd : enumDoc.fields()) {
if (fd.name().equals(name)) {
if (formatJavadoc) {
oi.enumJdoc.put(name, fd.commentText());
} else {
oi.enumJdoc.put(name, javadocToHtml(fd));
}
break;
}
}
}
} |
python | def labelTextWidth(label):
""" Returns the width of label text of the label in pixels.
IMPORTANT: does not work when the labels are styled using style sheets.
Unfortunately it is possible to retrieve the settings (e.g. padding) that were set by the
style sheet without parsing the style sheet as text.
"""
# The Qt source shows that fontMetrics().size calls fontMetrics().boundingRect with
# the TextLongestVariant included in the flags. TextLongestVariant is an internal flag
# which is used to force selecting the longest string in a multi-length string.
# See: http://stackoverflow.com/a/8638114/625350
fontMetrics = label.fontMetrics()
#contentsWidth = label.fontMetrics().boundingRect(label.text()).width()
contentsWidth = fontMetrics.size(label.alignment(), label.text()).width()
# If indent is negative, or if no indent has been set, the label computes the effective indent
# as follows: If frameWidth() is 0, the effective indent becomes 0. If frameWidth() is greater
# than 0, the effective indent becomes half the width of the "x" character of the widget's
# current font().
# See http://doc.qt.io/qt-4.8/qlabel.html#indent-prop
if label.indent() < 0 and label.frameWidth(): # no indent, but we do have a frame
indent = fontMetrics.width('x') / 2 - label.margin()
indent *= 2 # the indent seems to be added to the other side as well
else:
indent = label.indent()
result = contentsWidth + indent + 2 * label.frameWidth() + 2 * label.margin()
if 1:
#print ("contentsMargins: {}".format(label.getContentsMargins()))
#print ("contentsRect: {}".format(label.contentsRect()))
#print ("frameRect: {}".format(label.frameRect()))
print ("contentsWidth: {}".format(contentsWidth))
print ("lineWidth: {}".format(label.lineWidth()))
print ("midLineWidth: {}".format(label.midLineWidth()))
print ("frameWidth: {}".format(label.frameWidth()))
print ("margin: {}".format(label.margin()))
print ("indent: {}".format(label.indent()))
print ("actual indent: {}".format(indent))
print ("RESULT: {}".format(result))
print ()
return result |
python | def list_external_jar_dependencies(self, binary):
"""Returns the external jar dependencies of the given binary.
:param binary: The jvm binary target to list transitive external dependencies for.
:type binary: :class:`pants.backend.jvm.targets.jvm_binary.JvmBinary`
:returns: A list of (jar path, coordinate) tuples.
:rtype: list of (string, :class:`pants.java.jar.M2Coordinate`)
"""
classpath_products = self.context.products.get_data('runtime_classpath')
classpath_entries = classpath_products.get_artifact_classpath_entries_for_targets(
binary.closure(bfs=True, include_scopes=Scopes.JVM_RUNTIME_SCOPES,
respect_intransitive=True))
external_jars = OrderedSet(jar_entry for conf, jar_entry in classpath_entries
if conf == 'default')
return [(entry.path, entry.coordinate) for entry in external_jars
if not entry.is_excluded_by(binary.deploy_excludes)] |
java | public void setViewEnabled(boolean enable) {
if (enable && mView == null) {
mView = makeView();
}
if (mView != null) {
mView.setViewEnabled(enable);
}
} |
python | def set_torrent_download_limit(self, infohash_list, limit):
"""
Set download speed limit of the supplied torrents.
:param infohash_list: Single or list() of infohashes.
:param limit: Speed limit in bytes.
"""
data = self._process_infohash_list(infohash_list)
data.update({'limit': limit})
return self._post('command/setTorrentsDlLimit', data=data) |
python | def _get_true_palette_entry(self, name, digits):
''' Compute truecolor entry, once on the fly.
values must become sequence of decimal int strings: ('1', '2', '3')
'''
values = None
type_digits = type(digits)
is_fbterm = (env.TERM == 'fbterm') # sigh
if 'truecolor' in self._palette_support: # build entry
values = [self._start_codes_true]
if type_digits is str: # convert hex string
if len(digits) == 3:
values.extend(str(int(ch + ch, 16)) for ch in digits)
else: # chunk 'BB00BB', to ints to 'R', 'G', 'B':
values.extend(str(int(digits[i:i+2], 16)) for i in (0, 2 ,4))
else: # tuple of str-digit or int, may not matter to bother:
values.extend(str(digit) for digit in digits)
# downgrade section
elif 'extended' in self._palette_support:
if type_digits is str:
nearest_idx = find_nearest_color_hexstr(digits,
method=self._dg_method)
else: # tuple
if type(digits[0]) is str: # convert to ints
digits = tuple(int(digit) for digit in digits)
nearest_idx = find_nearest_color_index(*digits,
method=self._dg_method)
start_codes = self._start_codes_extended
if is_fbterm:
start_codes = self._start_codes_extended_fbterm
values = [start_codes, str(nearest_idx)]
elif 'basic' in self._palette_support:
if type_digits is str:
nearest_idx = find_nearest_color_hexstr(digits, color_table4,
method=self._dg_method)
else: # tuple
if type(digits[0]) is str: # convert to ints
digits = tuple(int(digit) for digit in digits)
nearest_idx = find_nearest_color_index(*digits,
color_table=color_table4,
method=self._dg_method)
values = self._index_to_ansi_values(nearest_idx)
return (self._create_entry(name, values, fbterm=is_fbterm)
if values else empty) |
python | def retry(self, delay=0, group=None, message=None):
'''Retry this job in a little bit, in the same queue. This is meant
for the times when you detect a transient failure yourself'''
args = ['retry', self.jid, self.queue_name, self.worker_name, delay]
if group is not None and message is not None:
args.append(group)
args.append(message)
return self.client(*args) |
java | @Override
public String newId(final RootDocument rootDocument) {
return newId(mongoTemplate, FamilyDocumentMongo.class,
rootDocument.getFilename(), "F");
} |
java | private void generateEncode(TypeToken<?> outputType, Schema schema) {
TypeToken<?> callOutputType = getCallTypeToken(outputType, schema);
Method encodeMethod = getMethod(void.class, "encode", callOutputType.getRawType(), Encoder.class);
if (!Object.class.equals(callOutputType.getRawType())) {
// Generate the synthetic method for the bridging
Method method = getMethod(void.class, "encode", Object.class, Encoder.class);
GeneratorAdapter mg = new GeneratorAdapter(Opcodes.ACC_PUBLIC + Opcodes.ACC_BRIDGE + Opcodes.ACC_SYNTHETIC,
method, null, new Type[] {Type.getType(IOException.class)},
classWriter);
mg.loadThis();
mg.loadArg(0);
mg.checkCast(Type.getType(callOutputType.getRawType()));
mg.loadArg(1);
mg.invokeVirtual(classType, encodeMethod);
mg.returnValue();
mg.endMethod();
}
// Generate the top level public encode method
String methodSignature = null;
if (callOutputType.getType() instanceof ParameterizedType) {
methodSignature = Signatures.getMethodSignature(encodeMethod, new TypeToken[]{callOutputType, null});
}
GeneratorAdapter mg = new GeneratorAdapter(Opcodes.ACC_PUBLIC, encodeMethod, methodSignature,
new Type[] {Type.getType(IOException.class)}, classWriter);
// Delegate to the actual encode method(value, encoder, schema, Sets.newIdentityHashSet());
mg.loadThis();
mg.loadArg(0);
mg.loadArg(1);
mg.loadThis();
mg.getField(classType, "schema", Type.getType(Schema.class));
// seenRefs Set
mg.invokeStatic(Type.getType(Sets.class), getMethod(Set.class, "newIdentityHashSet"));
mg.invokeVirtual(classType, getEncodeMethod(outputType, schema));
mg.returnValue();
mg.endMethod();
} |
python | def send_mail(self, sender, subject, recipients, message, response_id=None, html_message=False):
"""Send an email using EBUio features. If response_id is set, replies will be send back to the PlugIt server."""
params = {
'sender': sender,
'subject': subject,
'dests': recipients,
'message': message,
'html_message': html_message,
}
if response_id:
params['response_id'] = response_id
return self._request('mail/', postParams=params, verb='POST') |
java | protected TokenList.Token createOp( TokenList.Token left , TokenList.Token op , TokenList.Token right ,
TokenList tokens , Sequence sequence )
{
Operation.Info info = functions.create(op.symbol, left.getVariable(), right.getVariable());
sequence.addOperation(info.op);
// replace the symbols with their output
TokenList.Token t = new TokenList.Token(info.output);
tokens.remove(left);
tokens.remove(right);
tokens.replace(op,t);
return t;
} |
python | def write_out(self, output):
"""Banana banana
"""
for page in self.walk():
ext = self.project.extensions[page.extension_name]
ext.write_out_page(output, page) |
python | def get_saved_search(self, id, **kwargs): # noqa: E501
"""Get a specific saved search # noqa: E501
# noqa: E501
This method makes a synchronous HTTP request by default. To make an
asynchronous HTTP request, please pass async_req=True
>>> thread = api.get_saved_search(id, async_req=True)
>>> result = thread.get()
:param async_req bool
:param str id: (required)
:return: ResponseContainerSavedSearch
If the method is called asynchronously,
returns the request thread.
"""
kwargs['_return_http_data_only'] = True
if kwargs.get('async_req'):
return self.get_saved_search_with_http_info(id, **kwargs) # noqa: E501
else:
(data) = self.get_saved_search_with_http_info(id, **kwargs) # noqa: E501
return data |
java | private static BigInteger bigTenToThe(int n) {
if (n < 0)
return BigInteger.ZERO;
if (n < BIG_TEN_POWERS_TABLE_MAX) {
BigInteger[] pows = BIG_TEN_POWERS_TABLE;
if (n < pows.length)
return pows[n];
else
return expandBigIntegerTenPowers(n);
}
return BigInteger.TEN.pow(n);
} |
java | public static java.util.List<com.liferay.commerce.price.list.model.CommercePriceList> getCommercePriceListsByUuidAndCompanyId(
String uuid, long companyId) {
return getService()
.getCommercePriceListsByUuidAndCompanyId(uuid, companyId);
} |
python | def get_activity_admin_session(self, proxy):
"""Gets the ``OsidSession`` associated with the activity administration service.
arg: proxy (osid.proxy.Proxy): a proxy
return: (osid.learning.ActivityAdminSession) - an
``ActivityAdminSession``
raise: NullArgument - ``proxy`` is ``null``
raise: OperationFailed - unable to complete request
raise: Unimplemented - ``supports_activity_admin()`` is
``false``
*compliance: optional -- This method must be implemented if
``supports_activity_admin()`` is ``true``.*
"""
if not self.supports_activity_admin():
raise errors.Unimplemented()
# pylint: disable=no-member
return sessions.ActivityAdminSession(proxy=proxy, runtime=self._runtime) |
python | def auto_stratify(scores, **kwargs):
"""Generate Strata instance automatically
Parameters
----------
scores : array-like, shape=(n_items,)
ordered array of scores which quantify the classifier confidence for
the items in the pool. High scores indicate a high confidence that
the true label is a "1" (and vice versa for label "0").
**kwargs :
optional keyword arguments. May include 'stratification_method',
'stratification_n_strata', 'stratification_n_bins'.
Returns
-------
Strata instance
"""
if 'stratification_method' in kwargs:
method = kwargs['stratification_method']
else:
method = 'cum_sqrt_F'
if 'stratification_n_strata' in kwargs:
n_strata = kwargs['stratification_n_strata']
else:
n_strata = 'auto'
if 'stratification_n_bins' in kwargs:
n_bins = kwargs['stratification_n_bins']
strata = stratify_by_scores(scores, n_strata, method = method, \
n_bins = n_bins)
else:
strata = stratify_by_scores(scores, n_strata, method = method)
return strata |
java | public static boolean isRegisteredOnServer(Context context) {
final SharedPreferences prefs = getGCMPreferences(context);
boolean isRegistered = prefs.getBoolean(PROPERTY_ON_SERVER, false);
Log.v(TAG, "Is registered on server: " + isRegistered);
if (isRegistered) {
// checks if the information is not stale
long expirationTime =
prefs.getLong(PROPERTY_ON_SERVER_EXPIRATION_TIME, -1);
if (System.currentTimeMillis() > expirationTime) {
Log.v(TAG, "flag expired on: " + new Timestamp(expirationTime));
return false;
}
}
return isRegistered;
} |
python | def add_multiple(self, *users):
"""Add multiple users to the group at once.
Each given user must be a dictionary containing a nickname and either
an email, phone number, or user_id.
:param args users: the users to add
:return: a membership request
:rtype: :class:`MembershipRequest`
"""
guid = uuid.uuid4()
for i, user_ in enumerate(users):
user_['guid'] = '{}-{}'.format(guid, i)
payload = {'members': users}
url = utils.urljoin(self.url, 'add')
response = self.session.post(url, json=payload)
return MembershipRequest(self, *users, group_id=self.group_id,
**response.data) |
java | public void matchSet(Hashtable elements, int combine_op,
int compare_op) throws DBException
{
// Delete the old search criteria
criteria = new Criteria();
// If compare_op is not a binary operator, throw an exception
if ((compare_op & BINARY_OPER_MASK) == 0)
{
throw new DBException();
}
if (combine_op == AND)
{
// combine all value pairs by an AND
// For each of the elements in the hashtable, create a comparison node for the match
for (Enumeration e = elements.keys(); e.hasMoreElements();)
{
// Get the element name from the enumerator
// and its value
String elementName = (String) e.nextElement();
String elementValue = (String) elements.get(elementName);
switch (compare_op)
{
case LIKE:
{
criteria.addLike(elementName, elementValue);
break;
}
case EQUAL:
{
criteria.addEqualTo(elementName, elementValue);
break;
}
case NOT_EQUAL:
{
criteria.addNotEqualTo(elementName, elementValue);
break;
}
case LESS_THAN:
{
criteria.addLessThan(elementName, elementValue);
break;
}
case GREATER_THAN:
{
criteria.addGreaterThan(elementName, elementValue);
break;
}
case GREATER_EQUAL:
{
criteria.addGreaterOrEqualThan(elementName, elementValue);
break;
}
case LESS_EQUAL:
{
criteria.addLessOrEqualThan(elementName, elementValue);
break;
}
default:
{
throw new DBException("Unsupported binary operation in OJBSearchFilter!");
}
} // end of switch
} // end of for
}
else if (combine_op == OR)
{
// combine all value pairs by an OR
// For each of the elements in the hashtable, create a comparison node for the match
for (Enumeration e = elements.keys(); e.hasMoreElements();)
{
// Get the element name from the enumerator
// and its value
String elementName = (String) e.nextElement();
String elementValue = (String) elements.get(elementName);
switch (compare_op)
{
case LIKE:
{
Criteria tempCrit = new Criteria();
tempCrit.addLike(elementName, elementValue);
criteria.addOrCriteria(tempCrit);
break;
}
case EQUAL:
{
Criteria tempCrit = new Criteria();
tempCrit.addEqualTo(elementName, elementValue);
criteria.addOrCriteria(tempCrit);
break;
}
case NOT_EQUAL:
{
Criteria tempCrit = new Criteria();
tempCrit.addNotEqualTo(elementName, elementValue);
criteria.addOrCriteria(tempCrit);
break;
}
case LESS_THAN:
{
Criteria tempCrit = new Criteria();
tempCrit.addLessThan(elementName, elementValue);
criteria.addOrCriteria(tempCrit);
break;
}
case GREATER_THAN:
{
Criteria tempCrit = new Criteria();
tempCrit.addGreaterThan(elementName, elementValue);
criteria.addOrCriteria(tempCrit);
break;
}
case GREATER_EQUAL:
{
Criteria tempCrit = new Criteria();
tempCrit.addGreaterOrEqualThan(elementName, elementValue);
criteria.addOrCriteria(tempCrit);
break;
}
case LESS_EQUAL:
{
Criteria tempCrit = new Criteria();
tempCrit.addLessOrEqualThan(elementName, elementValue);
criteria.addOrCriteria(tempCrit);
break;
}
default:
{
throw new DBException("Unsupported binary operation in OJBSearchFilter!");
}
} // end of switch
} // end of for
}
else
{
// combine_op is not a logical operator, throw an exception
throw new DBException();
}
} |
java | @Nonnull
public final LBiByteFunctionBuilder<R> withHandling(@Nonnull HandlingInstructions<RuntimeException, RuntimeException> handling) {
Null.nonNullArg(handling, "handling");
if (this.handling != null) {
throw new UnsupportedOperationException("Handling is already set for this builder.");
}
this.handling = handling;
return self();
} |
python | def ecef2geodetic(x: float, y: float, z: float,
ell: Ellipsoid = None, deg: bool = True) -> Tuple[float, float, float]:
"""
convert ECEF (meters) to geodetic coordinates
Parameters
----------
x : float or numpy.ndarray of float
target x ECEF coordinate (meters)
y : float or numpy.ndarray of float
target y ECEF coordinate (meters)
z : float or numpy.ndarray of float
target z ECEF coordinate (meters)
ell : Ellipsoid, optional
reference ellipsoid
deg : bool, optional
degrees input/output (False: radians in/out)
Returns
-------
lat : float or numpy.ndarray of float
target geodetic latitude
lon : float or numpy.ndarray of float
target geodetic longitude
h : float or numpy.ndarray of float
target altitude above geodetic ellipsoid (meters)
based on:
You, Rey-Jer. (2000). Transformation of Cartesian to Geodetic Coordinates without Iterations.
Journal of Surveying Engineering. doi: 10.1061/(ASCE)0733-9453
"""
if ell is None:
ell = Ellipsoid()
x = np.asarray(x)
y = np.asarray(y)
z = np.asarray(z)
r = sqrt(x**2 + y**2 + z**2)
E = sqrt(ell.a**2 - ell.b**2)
# eqn. 4a
u = sqrt(0.5 * (r**2 - E**2) + 0.5 * sqrt((r**2 - E**2)**2 + 4 * E**2 * z**2))
Q = hypot(x, y)
huE = hypot(u, E)
# eqn. 4b
with np.errstate(divide='ignore'):
Beta = arctan(huE / u * z / hypot(x, y))
# eqn. 13
eps = ((ell.b * u - ell.a * huE + E**2) * sin(Beta)) / (ell.a * huE * 1 / cos(Beta) - E**2 * cos(Beta))
Beta += eps
# %% final output
lat = arctan(ell.a / ell.b * tan(Beta))
lon = arctan2(y, x)
# eqn. 7
alt = hypot(z - ell.b * sin(Beta),
Q - ell.a * cos(Beta))
# inside ellipsoid?
with np.errstate(invalid='ignore'):
inside = x**2 / ell.a**2 + y**2 / ell.a**2 + z**2 / ell.b**2 < 1
if isinstance(inside, np.ndarray):
alt[inside] = -alt[inside]
elif inside:
alt = -alt
if deg:
lat = degrees(lat)
lon = degrees(lon)
return lat, lon, alt |
python | def info(request, message, extra_tags='', fail_silently=False):
"""Adds a message with the ``INFO`` level."""
add_message(request, constants.INFO, message, extra_tags=extra_tags,
fail_silently=fail_silently) |
java | @Override
public void visit(NodeVisitor v) {
if (v.visit(this)) {
for (AstNode var : variables) {
var.visit(v);
}
}
} |
java | @SuppressWarnings("unchecked")
public static Map<String, Object> normaliseAndValidateQuery(Map<String, Object> query) throws QueryException{
boolean isWildCard = false;
if (query.isEmpty()) {
isWildCard = true;
}
// First expand the query to include a leading compound predicate
// if there isn't one already.
query = addImplicitAnd(query);
// At this point we will have a single entry map, key AND or OR,
// forming the compound predicate.
String compoundOperator = (String) query.keySet().toArray()[0];
List<Object> predicates = new ArrayList<Object>();
if (query.get(compoundOperator) instanceof List) {
// Next make sure all the predicates have an operator -- the EQ
// operator is implicit and we need to add it if there isn't one.
// Take
// [ {"field1": "mike"}, ... ]
// and make
// [ {"field1": { "$eq": "mike"} }, ... ]
predicates = addImplicitEq((List<Object>) query.get(compoundOperator));
// Then all shorthand operators like $ne, if present, need to be converted
// to their logical longhand equivalent.
// Take
// [ { "field1": { "$ne": "mike"} }, ... ]
// and make
// [ { "field1": { "$not" : { "$eq": "mike"} } }, ... ]
predicates = handleShortHandOperators(predicates);
// Now in the event that extraneous $not operators exist in the query,
// these operators must be compressed down to the their logical equivalent.
// Take
// [ { "field1": { "$not" : { $"not" : { "$eq": "mike"} } } }, ... ]
// and make
// [ { "field1": { "$eq": "mike"} }, ... ]
predicates = compressMultipleNotOperators(predicates);
// Here we ensure that all non-whole number arguments included in a $mod
// clause are truncated. This provides for consistent behavior between
// the SQL engine and the unindexed matcher.
// Take
// [ { "field1": { "$mod" : [ 2.6, 1.7] } }, ... ]
// and make
// [ { "field1": { "$mod" : [ 2, 1 ] } }, ... ]
predicates = truncateModArguments(predicates);
}
Map<String, Object> selector = new HashMap<String, Object>();
selector.put(compoundOperator, predicates);
if (!isWildCard) {
validateSelector(selector);
}
return selector;
} |
java | @Override
public MessageRepresentation getMessage(int commandCode, long applicationId, boolean isRequest) {
if (!this.configured) {
return null;
}
MessageRepresentation key = new MessageRepresentationImpl(commandCode, applicationId, isRequest);
return this.commandMap.get(key);
} |
java | double refine_clusters(List<Cluster<K>> clusters)
{
double[] norms = new double[clusters.size()];
int offset = 0;
for (Cluster cluster : clusters)
{
norms[offset++] = cluster.composite_vector().norm();
}
double eval_cluster = 0.0;
int loop_count = 0;
while (loop_count++ < NUM_REFINE_LOOP)
{
List<int[]> items = new ArrayList<int[]>(documents_.size());
for (int i = 0; i < clusters.size(); i++)
{
for (int j = 0; j < clusters.get(i).documents().size(); j++)
{
items.add(new int[]{i, j});
}
}
Collections.shuffle(items);
boolean changed = false;
for (int[] item : items)
{
int cluster_id = item[0];
int item_id = item[1];
Cluster<K> cluster = clusters.get(cluster_id);
Document<K> doc = cluster.documents().get(item_id);
double value_base = refined_vector_value(cluster.composite_vector(), doc.feature(), -1);
double norm_base_moved = Math.pow(norms[cluster_id], 2) + value_base;
norm_base_moved = norm_base_moved > 0 ? Math.sqrt(norm_base_moved) : 0.0;
double eval_max = -1.0;
double norm_max = 0.0;
int max_index = 0;
for (int j = 0; j < clusters.size(); j++)
{
if (cluster_id == j)
continue;
Cluster<K> other = clusters.get(j);
double value_target = refined_vector_value(other.composite_vector(), doc.feature(), 1);
double norm_target_moved = Math.pow(norms[j], 2) + value_target;
norm_target_moved = norm_target_moved > 0 ? Math.sqrt(norm_target_moved) : 0.0;
double eval_moved = norm_base_moved + norm_target_moved - norms[cluster_id] - norms[j];
if (eval_max < eval_moved)
{
eval_max = eval_moved;
norm_max = norm_target_moved;
max_index = j;
}
}
if (eval_max > 0)
{
eval_cluster += eval_max;
clusters.get(max_index).add_document(doc);
clusters.get(cluster_id).remove_document(item_id);
norms[cluster_id] = norm_base_moved;
norms[max_index] = norm_max;
changed = true;
}
}
if (!changed)
break;
for (Cluster<K> cluster : clusters)
{
cluster.refresh();
}
}
return eval_cluster;
} |
java | private static String[] getAllowedMethods(String[] existingMethods) {
int listCapacity = existingMethods.length + 1;
List<String> allowedMethods = new ArrayList<>(listCapacity);
allowedMethods.addAll(Arrays.asList(existingMethods));
allowedMethods.add(METHOD_PATCH);
return allowedMethods.toArray(new String[0]);
} |
java | @Override
public <A, B, C, D> P4<A, B, C, D> asProc4() {
return new P4<>(this, args);
} |
java | public int layerSize(String layerName) {
Layer l = getLayer(layerName);
if(l == null){
throw new IllegalArgumentException("No layer with name \"" + layerName + "\" exists");
}
org.deeplearning4j.nn.conf.layers.Layer conf = l.conf().getLayer();
if (conf == null || !(conf instanceof FeedForwardLayer)) {
return 0;
}
FeedForwardLayer ffl = (FeedForwardLayer) conf;
// FIXME: int cast
return (int) ffl.getNOut();
} |
Subsets and Splits