hash
stringlengths 40
40
| diff
stringlengths 131
114k
| message
stringlengths 7
980
| project
stringlengths 5
67
| split
stringclasses 1
value |
---|---|---|---|---|
b0e5b0af9758af6c3b3a69f7603d929d0c7fcedd | diff --git a/dist/js/jquery.orgchart.js b/dist/js/jquery.orgchart.js
index <HASH>..<HASH> 100755
--- a/dist/js/jquery.orgchart.js
+++ b/dist/js/jquery.orgchart.js
@@ -51,12 +51,8 @@
return addSiblings.apply(this, Array.prototype.splice.call(arguments, 1));
case 'removeNodes':
return removeNodes.apply(this, Array.prototype.splice.call(arguments, 1));
- case 'getHierarchy': {
- if (!$(this).find('.node:first')[0].id) {
- return 'Error: Nodes of orghcart to be exported must have id attribute!';
- }
- return getHierarchy.apply(this, [$(this)]);
- }
+ case 'getHierarchy':
+ return getHierarchy.apply(this, [$(this), this.data('orgchart').options]);
default: // initiation time
var opts = $.extend(defaultOptions, options);
this.data('orgchart', { 'options' : opts });
@@ -225,12 +221,21 @@
return data;
}
- function getHierarchy($orgchart) {
+ function getHierarchy($orgchart, options) {
var $tr = $orgchart.find('tr:first');
- var subObj = { 'id': $tr.find('.node')[0].id };
+
+ var $node = $tr.find('.node')
+ var nodeContent = options.nodeContent;
+ var subObj = {};
+
+ subObj.name = $node.find('.title').text();
+
+ $node[0].id && (subObj.id = $node[0].id);
+ nodeContent && (subObj[nodeContent] = $node.find('.content').text());
+
$tr.siblings(':last').children().each(function() {
if (!subObj.children) { subObj.children = []; }
- subObj.children.push(getHierarchy($(this)));
+ subObj.children.push(getHierarchy($(this), options));
});
return subObj;
}
@@ -434,9 +439,10 @@
var $nodeDiv = $('<div' + (opts.draggable ? ' draggable="true"' : '') + (nodeData[opts.nodeId] ? ' id="' + nodeData[opts.nodeId] + '"' : '') + '>')
.addClass('node' + (level >= opts.depth ? ' slide-up' : ''))
.append('<div class="title">' + nodeData[opts.nodeTitle] + '</div>')
- .append(typeof opts.nodeContent !== 'undefined' ? '<div class="content">' + nodeData[opts.nodeContent] + '</div>' : '');
+ .append(typeof opts.nodeContent !== 'undefined' ? '<div class="content">' + (nodeData[opts.nodeContent] || '') + '</div>' : '');
// append 4 direction arrows
- var flags = nodeData.relationship;
+ var flags = nodeData.relationship || '';
+
if (Number(flags.substr(0,1))) {
$nodeDiv.append('<i class="edge verticalEdge topEdge fa"></i>');
} | Improvement getHierarchy
and fix bug during create node when there is no 'relationship' attribute. | dabeng_OrgChart | train |
6b9d2db1470fa04a5746cb12287e4074141f7f73 | diff --git a/statsd/src/test/java/org/cloudfoundry/identity/statsd/UaaMetricsEmitterTests.java b/statsd/src/test/java/org/cloudfoundry/identity/statsd/UaaMetricsEmitterTests.java
index <HASH>..<HASH> 100644
--- a/statsd/src/test/java/org/cloudfoundry/identity/statsd/UaaMetricsEmitterTests.java
+++ b/statsd/src/test/java/org/cloudfoundry/identity/statsd/UaaMetricsEmitterTests.java
@@ -140,11 +140,11 @@ public class UaaMetricsEmitterTests {
Mockito.verify(statsDClient).gauge(eq("vitals.jvm.heap.init"), gt(0l));
Mockito.verify(statsDClient).gauge(eq("vitals.jvm.heap.committed"), gt(0l));
Mockito.verify(statsDClient).gauge(eq("vitals.jvm.heap.used"), gt(0l));
- Mockito.verify(statsDClient).gauge(eq("vitals.jvm.heap.max"), gt(0l));
+ //Mockito.verify(statsDClient).gauge(eq("vitals.jvm.heap.max"), gt(0l));
Mockito.verify(statsDClient).gauge(eq("vitals.jvm.non-heap.init"), gt(0l));
Mockito.verify(statsDClient).gauge(eq("vitals.jvm.non-heap.committed"), gt(0l));
Mockito.verify(statsDClient).gauge(eq("vitals.jvm.non-heap.used"), gt(0l));
- Mockito.verify(statsDClient).gauge(eq("vitals.jvm.non-heap.max"), gt(0l));
+ //Mockito.verify(statsDClient).gauge(eq("vitals.jvm.non-heap.max"), gt(0l));
}
@Test
diff --git a/statsd/src/test/java/org/cloudfoundry/identity/statsd/integration/UaaMetricsEmitterIT.java b/statsd/src/test/java/org/cloudfoundry/identity/statsd/integration/UaaMetricsEmitterIT.java
index <HASH>..<HASH> 100644
--- a/statsd/src/test/java/org/cloudfoundry/identity/statsd/integration/UaaMetricsEmitterIT.java
+++ b/statsd/src/test/java/org/cloudfoundry/identity/statsd/integration/UaaMetricsEmitterIT.java
@@ -88,9 +88,9 @@ public class UaaMetricsEmitterIT {
"uaa.vitals.jvm.heap.used",
"uaa.vitals.jvm.heap.max",
"uaa.vitals.jvm.non-heap.init",
- "uaa.vitals.jvm.non-heap.committed",
- "uaa.vitals.jvm.non-heap.used",
- "uaa.vitals.jvm.non-heap.max"
+ "uaa.vitals.jvm.non-heap.committed"
+// ,"uaa.vitals.jvm.non-heap.used", //max return -1 and are not emitted
+// "uaa.vitals.jvm.non-heap.max" //max return -1 and are not emitted
);
private static Map<String, String> secondBatch; | Metrics that don't have a value, -1, are not emitted
[#<I>] <URL> | cloudfoundry_uaa | train |
3d49fc131517c93d9857070626f76a7de757d8ec | diff --git a/framework/core/src/Core/Access/DiscussionPolicy.php b/framework/core/src/Core/Access/DiscussionPolicy.php
index <HASH>..<HASH> 100644
--- a/framework/core/src/Core/Access/DiscussionPolicy.php
+++ b/framework/core/src/Core/Access/DiscussionPolicy.php
@@ -109,6 +109,8 @@ class DiscussionPolicy extends AbstractPolicy
*/
public function delete(User $actor, Discussion $discussion)
{
- return $this->rename($actor, $discussion);
+ if ($discussion->start_user_id == $actor->id && $discussion->participants_count <= 1) {
+ return true;
+ }
}
} | Prevent users from being incorrectly able to delete their own discussions | flarum_core | train |
3a57b9c2fee2aced5f263ea7e6844f444657b0c2 | diff --git a/nomad/vault.go b/nomad/vault.go
index <HASH>..<HASH> 100644
--- a/nomad/vault.go
+++ b/nomad/vault.go
@@ -583,15 +583,15 @@ func (v *vaultClient) renew() (bool, error) {
// Attempt to renew the token
secret, err := v.auth.RenewSelf(v.tokenData.CreationTTL)
if err != nil {
-
// Check if there is a permission denied
recoverable := !structs.VaultUnrecoverableError.MatchString(err.Error())
return recoverable, fmt.Errorf("failed to renew the vault token: %v", err)
}
+
if secret == nil {
// It's possible for RenewSelf to return (nil, nil) if the
// response body from Vault is empty.
- return fmt.Errorf("renewal failed: empty response from vault")
+ return true, fmt.Errorf("renewal failed: empty response from vault")
}
// these treated as transient errors, where can keep renewing | nil secrets as recoverable to keep renew attempts | hashicorp_nomad | train |
dff23244eca9fe70bb425158988175273a9ad88f | diff --git a/fusesoc/sections.py b/fusesoc/sections.py
index <HASH>..<HASH> 100644
--- a/fusesoc/sections.py
+++ b/fusesoc/sections.py
@@ -55,6 +55,8 @@ class VerilatorSection(ToolSection):
self.name ='verilator'
+ self.include_dirs = []
+
self._add_listitem('verilator_options')
self._add_listitem('src_files')
self._add_listitem('include_files') | sections.py: Always define include_dirs in VerilatorSection | olofk_fusesoc | train |
cfa71d125168ac4e2f33aee0d132ada1c9ad4a3d | diff --git a/lib/class-wp-json-comments-controller.php b/lib/class-wp-json-comments-controller.php
index <HASH>..<HASH> 100644
--- a/lib/class-wp-json-comments-controller.php
+++ b/lib/class-wp-json-comments-controller.php
@@ -404,14 +404,14 @@ class WP_JSON_Comments_Controller extends WP_JSON_Controller {
return true;
}
- if ( current_user_can( 'edit_comment', $comment->comment_ID ) ) {
- return true;
+ if ( 0 == get_current_user_id() ) {
+ return false;
}
- if ( get_current_user_id() == $comment->user_id ) {
+ if ( ! empty( $comment->user_id ) && get_current_user_id() == $comment->user_id ) {
return true;
}
- return false;
+ return current_user_can( 'edit_comment', $comment->comment_ID );
}
} | Do not allow unauthenticated users to read unapproved comments. | WP-API_WP-API | train |
2bf4f0f84786f16a2fd9eaf6d40c081ac03bc52b | diff --git a/core/src/main/java/com/google/errorprone/bugpatterns/GuiceAssistedInjectScoping.java b/core/src/main/java/com/google/errorprone/bugpatterns/GuiceAssistedInjectScoping.java
index <HASH>..<HASH> 100644
--- a/core/src/main/java/com/google/errorprone/bugpatterns/GuiceAssistedInjectScoping.java
+++ b/core/src/main/java/com/google/errorprone/bugpatterns/GuiceAssistedInjectScoping.java
@@ -20,6 +20,7 @@ import static com.google.errorprone.BugPattern.Category.GUICE;
import static com.google.errorprone.BugPattern.MaturityLevel.MATURE;
import static com.google.errorprone.BugPattern.SeverityLevel.ERROR;
import static com.google.errorprone.matchers.Matchers.annotations;
+import static com.google.errorprone.matchers.Matchers.anyOf;
import static com.google.errorprone.matchers.Matchers.constructor;
import static com.google.errorprone.matchers.Matchers.hasAnnotation;
import static com.google.errorprone.matchers.Matchers.methodHasParameters;
@@ -59,18 +60,21 @@ import com.sun.source.tree.VariableTree;
category = GUICE, severity = ERROR, maturity = MATURE)
public class GuiceAssistedInjectScoping extends DescribingMatcher<ClassTree> {
- private static final String SCOPE_ANNOTATION_STRING = "com.google.inject.ScopeAnnotation";
- private static final String ASSISTED_ANNOTATION_STRING =
- "com.google.inject.assistedinject.Assisted";
- private static final String INJECT_ANNOTATION_STRING = "com.google.inject.Inject";
- private static final String ASSISTED_INJECT_ANNOTATION_STRING =
+ private static final String GUICE_SCOPE_ANNOTATION = "com.google.inject.ScopeAnnotation";
+ private static final String JAVAX_SCOPE_ANNOTATION = "javax.inject.Scope";
+ private static final String ASSISTED_ANNOTATION = "com.google.inject.assistedinject.Assisted";
+ private static final String GUICE_INJECT_ANNOTATION = "com.google.inject.Inject";
+ private static final String JAVAX_INJECT_ANNOTATION = "javax.inject.Inject";
+ private static final String ASSISTED_INJECT_ANNOTATION =
"com.google.inject.assistedinject.AssistedInject";
/**
* Matches classes that have an annotation that itself is annotated with @ScopeAnnotation.
*/
+ @SuppressWarnings("unchecked")
private MultiMatcher<ClassTree, AnnotationTree> classAnnotationMatcher =
- annotations(ANY, hasAnnotation(SCOPE_ANNOTATION_STRING, AnnotationTree.class));
+ annotations(ANY, anyOf(hasAnnotation(GUICE_SCOPE_ANNOTATION, AnnotationTree.class),
+ hasAnnotation(JAVAX_SCOPE_ANNOTATION)));
/**
* Matches if:
@@ -84,16 +88,17 @@ public class GuiceAssistedInjectScoping extends DescribingMatcher<ClassTree> {
@Override
public boolean matches(ClassTree classTree, VisitorState state) {
MultiMatcher<ClassTree, MethodTree> constructorWithInjectMatcher =
- constructor(ANY, hasAnnotation(INJECT_ANNOTATION_STRING, MethodTree.class));
+ constructor(ANY, anyOf(hasAnnotation(GUICE_INJECT_ANNOTATION, MethodTree.class),
+ hasAnnotation(JAVAX_INJECT_ANNOTATION)));
if (constructorWithInjectMatcher.matches(classTree, state)) {
// Check constructor with @Inject annotation for parameter with @Assisted annotation.
return methodHasParameters(ANY,
- hasAnnotation(ASSISTED_ANNOTATION_STRING, VariableTree.class))
+ hasAnnotation(ASSISTED_ANNOTATION, VariableTree.class))
.matches(constructorWithInjectMatcher.getMatchingNode(), state);
}
- return constructor(ANY, hasAnnotation(ASSISTED_INJECT_ANNOTATION_STRING, MethodTree.class))
+ return constructor(ANY, hasAnnotation(ASSISTED_INJECT_ANNOTATION, MethodTree.class))
.matches(classTree, state);
}
};
diff --git a/core/src/test/resources/com/google/errorprone/bugpatterns/GuiceAssistedInjectScopingPositiveCases.java b/core/src/test/resources/com/google/errorprone/bugpatterns/GuiceAssistedInjectScopingPositiveCases.java
index <HASH>..<HASH> 100644
--- a/core/src/test/resources/com/google/errorprone/bugpatterns/GuiceAssistedInjectScopingPositiveCases.java
+++ b/core/src/test/resources/com/google/errorprone/bugpatterns/GuiceAssistedInjectScopingPositiveCases.java
@@ -82,8 +82,16 @@ public class GuiceAssistedInjectScopingPositiveCases {
@AssistedInject
public TestClass5(int i, String unassisted) {
}
-
}
-
-
+
+ /**
+ * JSR330 annotations.
+ */
+ //BUG: Suggestion includes "remove this line"
+ @javax.inject.Singleton
+ public class TestClass6 {
+ @javax.inject.Inject
+ public TestClass6(String unassisted, @Assisted String assisted) {
+ }
+ }
} | Support JSR<I> annotations in GuiceAssistedInjectScoping | google_error-prone | train |
bc92c9e320ae56e6e0bf2c25413be7de58950fbe | diff --git a/cobra/core/model.py b/cobra/core/model.py
index <HASH>..<HASH> 100644
--- a/cobra/core/model.py
+++ b/cobra/core/model.py
@@ -814,7 +814,7 @@ class Model(Object):
else:
assert_optimal(self, message)
- def optimize(self, objective_sense=None, **kwargs):
+ def optimize(self, objective_sense=None, raise_error=False, **kwargs):
"""
Optimize the model using flux balance analysis.
@@ -823,6 +823,9 @@ class Model(Object):
objective_sense : {None, 'maximize' 'minimize'}, optional
Whether fluxes should be maximized or minimized. In case of None,
the previous direction is used.
+ raise_error : bool
+ If true, raise an OptimizationError if solver status is not
+ optimal.
solver : {None, 'glpk', 'cglpk', 'gurobi', 'cplex'}, optional
If unspecified will use the currently defined `self.solver`
otherwise it will use the given solver and update the attribute.
@@ -855,7 +858,7 @@ class Model(Object):
"max": "maximize", "min": "minimize"}[original_direction]
solution = optimize(self, objective_sense=objective_sense,
**kwargs)
- check_solver_status(solution.status)
+ check_solver_status(solution.status, raise_error=raise_error)
return solution
self.solver = solver
@@ -863,7 +866,7 @@ class Model(Object):
{"maximize": "max", "minimize": "min"}.get(
objective_sense, original_direction)
self.slim_optimize()
- solution = get_solution(self)
+ solution = get_solution(self, raise_error=raise_error)
self.objective.direction = original_direction
return solution
diff --git a/cobra/core/solution.py b/cobra/core/solution.py
index <HASH>..<HASH> 100644
--- a/cobra/core/solution.py
+++ b/cobra/core/solution.py
@@ -273,7 +273,7 @@ class LegacySolution(object):
DeprecationWarning)
-def get_solution(model, reactions=None, metabolites=None):
+def get_solution(model, reactions=None, metabolites=None, raise_error=False):
"""
Generate a solution representation of the current solver state.
@@ -287,6 +287,8 @@ def get_solution(model, reactions=None, metabolites=None):
metabolites : list, optional
An iterable of `cobra.Metabolite` objects. Uses `model.metabolites` by
default.
+ raise_error : bool
+ If true, raise an OptimizationError if solver status is not optimal.
Returns
-------
@@ -297,7 +299,7 @@ def get_solution(model, reactions=None, metabolites=None):
This is only intended for the `optlang` solver interfaces and not the
legacy solvers.
"""
- check_solver_status(model.solver.status)
+ check_solver_status(model.solver.status, raise_error=raise_error)
if reactions is None:
reactions = model.reactions
if metabolites is None:
diff --git a/cobra/test/test_model.py b/cobra/test/test_model.py
index <HASH>..<HASH> 100644
--- a/cobra/test/test_model.py
+++ b/cobra/test/test_model.py
@@ -767,6 +767,17 @@ class TestCobraModel:
with pytest.raises(OptimizationError):
model.slim_optimize(error_value=None)
+ @pytest.mark.parametrize("solver", optlang_solvers)
+ def test_optimize(self, model, solver):
+ model.solver = solver
+ with model:
+ assert model.optimize().objective_value > 0.872
+ model.reactions.Biomass_Ecoli_core.lower_bound = 10
+ with pytest.warns(UserWarning):
+ model.optimize()
+ with pytest.raises(OptimizationError):
+ model.optimize(raise_error=True)
+
def test_change_objective(self, model):
# Test for correct optimization behavior
model.optimize()
diff --git a/cobra/util/solver.py b/cobra/util/solver.py
index <HASH>..<HASH> 100644
--- a/cobra/util/solver.py
+++ b/cobra/util/solver.py
@@ -394,11 +394,11 @@ def fix_objective_as_constraint(model, fraction=1, bound=None,
add_cons_vars_to_problem(model, constraint, sloppy=True)
-def check_solver_status(status):
+def check_solver_status(status, raise_error=False):
"""Perform standard checks on a solver's status."""
if status == optlang.interface.OPTIMAL:
return
- elif status == optlang.interface.INFEASIBLE:
+ elif status == optlang.interface.INFEASIBLE and not raise_error:
warn("solver status is '{}'".format(status), UserWarning)
elif status is None:
raise RuntimeError(
diff --git a/release-notes/next-release.md b/release-notes/next-release.md
index <HASH>..<HASH> 100644
--- a/release-notes/next-release.md
+++ b/release-notes/next-release.md
@@ -14,6 +14,8 @@
notebooks.
- New convenience functions `cobra.flux_analysis.find_essential_genes` and
`cobra.flux_analysis.find_essential_reactions`.
+- `Model.optimize` has new parameter `raise_error` to enable option to
+ get trigger exception if no feasible solution could be found.
## Deprecated features | feat: option to raise error from model.optimize
Add a feature to replace this
```
model.slim_optimize()
assert_optimal()
solution = get_solution(model)
```
with just
```
solution = model.optimize(raise_error=True)
```
and avoid the imports assert_optimal and get_solution in other packages | opencobra_cobrapy | train |
2f28f1636ebdb696a438867ab53e2503c121e3d6 | diff --git a/src/Content.php b/src/Content.php
index <HASH>..<HASH> 100644
--- a/src/Content.php
+++ b/src/Content.php
@@ -23,14 +23,17 @@ class Content implements \ArrayAccess
// The last time we weight a searchresult
private $lastWeight = 0;
+ // Whether this is a "real" contenttype or an embedded ones
+ private $isRootType;
public $user;
public $sortorder;
public $config;
public $group;
- public function __construct(Silex\Application $app, $contenttype = '', $values = '')
+ public function __construct(Silex\Application $app, $contenttype = '', $values = '', $isRootType = true)
{
$this->app = $app;
+ $this->isRootType = $isRootType;
if (!empty($contenttype)) {
// Set the contenttype
@@ -341,7 +344,7 @@ class Content implements \ArrayAccess
// Template fields need to be done last
// As the template has to have been selected
- if (is_array($this->contenttype)) {
+ if ($this->isRootType) {
if (empty($values['templatefields'])) {
$this->setValue('templatefields', array());
} else {
@@ -352,6 +355,11 @@ class Content implements \ArrayAccess
public function setValue($key, $value)
{
+ // Don't set templateFields if not a real contenttype
+ if (($key == 'templatefields') && (!$this->isRootType)) {
+ return;
+ }
+
// Check if the value need to be unserialized.
if (is_string($value) && substr($value, 0, 2) == "a:") {
try {
@@ -394,7 +402,6 @@ class Content implements \ArrayAccess
}
if ($key == 'templatefields') {
-
$oldValue = $this->values[$key];
if ((is_string($value)) || (is_array($value))) {
if (is_string($value)) {
@@ -408,7 +415,7 @@ class Content implements \ArrayAccess
}
if ($unserdata !== false) {
- $templateContent = new Content($this->app, '', array());
+ $templateContent = new Content($this->app, $this->getTemplateFieldsContentType(), array(), false);
$value = $templateContent;
$this->populateTemplateFieldsContenttype($value);
$templateContent->setValues($unserdata);
@@ -416,7 +423,6 @@ class Content implements \ArrayAccess
$value = null;
}
}
-
}
if (!isset($this->values['datechanged']) ||
@@ -544,16 +550,16 @@ class Content implements \ArrayAccess
$this->setValues($values);
}
- protected function populateTemplateFieldsContenttype($templatefields) {
+ protected function getTemplateFieldsContentType() {
if (is_array($this->contenttype)) {
- if ((!$this->contenttype['viewless']) && (!empty($templatefields)) && ($templateFieldsConfig = $this->app['config']->get('theme/template_fields'))) {
+ if ($templateFieldsConfig = $this->app['config']->get('theme/template_fields')) {
$template = $this->app['templatechooser']->record($this);
if (array_key_exists($template, $templateFieldsConfig)) {
-
- $templatefields->contenttype = $templateFieldsConfig[$template];
+ return $templateFieldsConfig[$template];
}
}
}
+ return '';
}
public function hasTemplateFields() { | Add flag for contenttypes that shouldn't have templatefields | bolt_bolt | train |
288c35b6d5921e9b066b4f802777e42a71e1f4ab | diff --git a/src/cr/cube/cubepart.py b/src/cr/cube/cubepart.py
index <HASH>..<HASH> 100644
--- a/src/cr/cube/cubepart.py
+++ b/src/cr/cube/cubepart.py
@@ -749,7 +749,13 @@ class _Slice(CubePartition):
@lazyproperty
def payload_order(self):
- """1D np.int64 ndarray of signed int idx respecting the payload order"""
+ """1D np.int64 ndarray of signed int idx respecting the payload order.
+
+ Positive integers indicate the 1-indexed position in payload of regular
+ elements, while negative integers are the subtotal insertions.
+
+ Needed for reordering color palette in exporter.
+ """
return self._assembler.payload_order
@lazyproperty
@@ -830,7 +836,13 @@ class _Slice(CubePartition):
@lazyproperty
def row_order(self):
- """1D np.int64 ndarray of signed int idx for each assembled row of stripe."""
+ """1D np.int64 ndarray of signed int idx for each assembled row of stripe.
+
+ Positive integers indicate the 1-indexed position in payload of regular
+ elements, while negative integers are the subtotal insertions.
+
+ Needed for reordering color palette in exporter.
+ """
return self._assembler.row_order
@lazyproperty
@@ -1527,7 +1539,13 @@ class _Strand(CubePartition):
@lazyproperty
def payload_order(self):
- """1D np.int64 ndarray of signed int idx respecting the payload order"""
+ """1D np.int64 ndarray of signed int idx respecting the payload order.
+
+ Positive integers indicate the 1-indexed position in payload of regular
+ elements, while negative integers are the subtotal insertions.
+
+ Needed for reordering color palette in exporter.
+ """
return self._assembler.payload_order
@lazyproperty
@@ -1586,7 +1604,13 @@ class _Strand(CubePartition):
@lazyproperty
def row_order(self):
- """1D np.int64 ndarray of signed int idx for each assembled row of stripe."""
+ """1D np.int64 ndarray of signed int idx for each assembled row of stripe.
+
+ Positive integers indicate the 1-indexed position in payload of regular
+ elements, while negative integers are the subtotal insertions.
+
+ Needed for reordering color palette in exporter.
+ """
return self._assembler.row_order
@lazyproperty | [<I>]: Address greg's comments | Crunch-io_crunch-cube | train |
dd3cb920364a0826e409f9cf0eacdfebe8dd754f | diff --git a/qbit/spring/src/main/java/io/advantageous/qbit/spring/config/PlatformConfiguration.java b/qbit/spring/src/main/java/io/advantageous/qbit/spring/config/PlatformConfiguration.java
index <HASH>..<HASH> 100644
--- a/qbit/spring/src/main/java/io/advantageous/qbit/spring/config/PlatformConfiguration.java
+++ b/qbit/spring/src/main/java/io/advantageous/qbit/spring/config/PlatformConfiguration.java
@@ -57,7 +57,7 @@ public class PlatformConfiguration {
.setMicroServiceName(props.getPrefix())
.setHealthService(healthServiceAsync);
- if (statsCollector.isPresent()) {
+ if (statsCollector.isPresent() && props.getJvmStatsRefresh() > 0) {
adminBuilder.registerJavaVMStatsJobEveryNSeconds(statsCollector.get(), props.getJvmStatsRefresh());
} | only do jvm stats if we have a refresh prop | advantageous_qbit | train |
c66a8b386f65b4ff715d0fe520a6a0c58d0ef1cd | diff --git a/core/src/main/java/com/graphhopper/storage/GraphHopperStorage.java b/core/src/main/java/com/graphhopper/storage/GraphHopperStorage.java
index <HASH>..<HASH> 100644
--- a/core/src/main/java/com/graphhopper/storage/GraphHopperStorage.java
+++ b/core/src/main/java/com/graphhopper/storage/GraphHopperStorage.java
@@ -247,9 +247,9 @@ public class GraphHopperStorage implements GraphStorage
{
int integ = (int) (distance * INT_DIST_FACTOR);
if (integ < 0)
- throw new IllegalArgumentException("Distance cannot be empty: " +
- distance + ", maybe overflow issue? integer: " + integ);
-
+ throw new IllegalArgumentException("Distance cannot be empty: "
+ + distance + ", maybe overflow issue? integer: " + integ);
+
// Due to rounding errors e.g. when getting the distance from another DataAccess object
// the following exception is not a good idea:
// Allow integ to be 0 only if distance is 0
@@ -264,7 +264,11 @@ public class GraphHopperStorage implements GraphStorage
*/
private double getDist( long pointer )
{
- return edges.getInt(pointer + E_DIST) / INT_DIST_FACTOR;
+ int val = edges.getInt(pointer + E_DIST);
+ if (val == Integer.MAX_VALUE)
+ return Double.POSITIVE_INFINITY;
+
+ return val / INT_DIST_FACTOR;
}
@Override
@@ -630,7 +634,7 @@ public class GraphHopperStorage implements GraphStorage
{
throw new UnsupportedOperationException("Not supported yet.");
}
-
+
@Override
public int getEdge()
{
diff --git a/core/src/test/java/com/graphhopper/storage/AbstractGraphStorageTester.java b/core/src/test/java/com/graphhopper/storage/AbstractGraphStorageTester.java
index <HASH>..<HASH> 100644
--- a/core/src/test/java/com/graphhopper/storage/AbstractGraphStorageTester.java
+++ b/core/src/test/java/com/graphhopper/storage/AbstractGraphStorageTester.java
@@ -80,6 +80,15 @@ public abstract class AbstractGraphStorageTester
}
@Test
+ public void testInfinityWeight()
+ {
+ graph = createGraph();
+ EdgeIteratorState edge = graph.edge(0, 1);
+ edge.setDistance(Double.POSITIVE_INFINITY);
+ assertTrue(Double.isInfinite(edge.getDistance()));
+ }
+
+ @Test
public void testSetNodes()
{
graph = createGraph(); | proper handling of infinit distances | graphhopper_graphhopper | train |
e99f993ee2274c920420f426a2a6b4480ca557f5 | diff --git a/asteval/asteval.py b/asteval/asteval.py
index <HASH>..<HASH> 100644
--- a/asteval/asteval.py
+++ b/asteval/asteval.py
@@ -758,8 +758,8 @@ class Interpreter(object):
return func(*args, **keywords)
except Exception as ex:
self.raise_exception(
- node, msg="Error running function call '%s': %s"
- % (func.__name__, ex))
+ node, msg="Error running function call '%s' with args %s and "
+ "kwargs %s: %s" % (func.__name__, args, keywords, ex))
def on_arg(self, node): # ('test', 'msg')
"""Arg for function definitions."""
@@ -819,6 +819,7 @@ class Procedure(object):
"""TODO: docstring in public method."""
self.__ininit__ = True
self.name = name
+ self.__name__ = self.name
self.__asteval__ = interp
self.raise_exc = self.__asteval__.raise_exception
self.__doc__ = doc | Give Procedure objects a __name__ attribute | newville_asteval | train |
6f00cd999e54e24df602991f5b73d39da3a4869f | diff --git a/normandy/recipes/api/serializers.py b/normandy/recipes/api/serializers.py
index <HASH>..<HASH> 100644
--- a/normandy/recipes/api/serializers.py
+++ b/normandy/recipes/api/serializers.py
@@ -7,7 +7,7 @@ from normandy.base.api.serializers import UserSerializer
from normandy.recipes.api.fields import ActionImplementationHyperlinkField
from normandy.recipes.models import (Action, Recipe, Approval, ApprovalRequest,
ApprovalRequestComment)
-from normandy.recipes.validators import Validator
+from normandy.recipes.validators import JSONSchemaValidator
class ActionSerializer(serializers.ModelSerializer):
@@ -110,7 +110,7 @@ class RecipeSerializer(serializers.ModelSerializer):
except:
raise serializers.ValidationError('Could not find arguments schema.')
- schemaValidator = Validator(schema)
+ schemaValidator = JSONSchemaValidator(schema)
errorResponse = {}
errors = sorted(schemaValidator.iter_errors(value), key=lambda e: e.path)
diff --git a/normandy/recipes/validators.py b/normandy/recipes/validators.py
index <HASH>..<HASH> 100644
--- a/normandy/recipes/validators.py
+++ b/normandy/recipes/validators.py
@@ -20,7 +20,7 @@ def _required(validator, required, instance, schema):
# Construct validator as extension of Json Schema Draft 4.
-Validator = jsonschema.validators.extend(
+JSONSchemaValidator = jsonschema.validators.extend(
validator=jsonschema.validators.Draft4Validator,
validators={
'required': _required | Rename Validator to JSONSchemaValidator | mozilla_normandy | train |
32f0efb087c40f02914943e18a4990c07fc88da3 | diff --git a/.storybook/config.js b/.storybook/config.js
index <HASH>..<HASH> 100644
--- a/.storybook/config.js
+++ b/.storybook/config.js
@@ -1,5 +1,5 @@
import { Text, View } from 'react-native';
-import { configure, addDecorator } from "@storybook/react";
+import { configure, addDecorator } from '@storybook/react';
import { withInfo } from '@storybook/addon-info';
import { withOptions } from '@storybook/addon-options';
import { withKnobs } from '@storybook/addon-knobs';
@@ -7,7 +7,7 @@ import { withKnobs } from '@storybook/addon-knobs';
const req = require.context(
"../packages",
true,
- /^((?!node_modules).)*\.(stories|stories.web)\.js$/
+ /^((?!node_modules).)*\.(stories|stories.web)\.(tsx|js)$/
);
addDecorator(withInfo({
diff --git a/.storybook/webpack.config.js b/.storybook/webpack.config.js
index <HASH>..<HASH> 100644
--- a/.storybook/webpack.config.js
+++ b/.storybook/webpack.config.js
@@ -2,7 +2,7 @@ const path = require("path");
const webpack = require("webpack");
module.exports = async ({ config }, env, defaultConfig) => {
- config.devtool = "eval-source-map"
+ config.devtool = "eval-source-map";
config.resolve = {
...config.resolve,
alias: {
@@ -10,7 +10,7 @@ module.exports = async ({ config }, env, defaultConfig) => {
"react-native": "react-native-web",
"@storybook/react-native": "@storybook/react"
},
- extensions: [".web.js", ".js", ".mjs"],
+ extensions: [".tsx", ".ts", ".web.js", ".js", ".mjs"],
mainFields: ["devModule", "dev", "module", "main"]
};
config.plugins.push(
@@ -19,15 +19,22 @@ module.exports = async ({ config }, env, defaultConfig) => {
manifest: path.resolve("./dist/public/vendor-manifest.json")
})
);
+ config.module.rules.push({
+ test: /\.(tsx?)$/,
+ loader: "babel-loader",
+ options: {
+ configFile: "./babel.config.js"
+ }
+ });
config.module.rules.push(
{
test: /\.(png|jpe?g|gif)$/,
- loader: 'react-native-web-image-loader?name=[hash].[ext]',
+ loader: "react-native-web-image-loader?name=[hash].[ext]"
},
{
test: /\.mjs$/,
include: /node_modules/,
- type: 'javascript/auto'
+ type: "javascript/auto"
},
{
test: /\.(graphql|gql)$/,
@@ -37,5 +44,5 @@ module.exports = async ({ config }, env, defaultConfig) => {
}
);
- return config
+ return config;
};
diff --git a/package.json b/package.json
index <HASH>..<HASH> 100644
--- a/package.json
+++ b/package.json
@@ -88,6 +88,7 @@
"@babel/core": "7.4.4",
"@babel/plugin-transform-runtime": "7.4.4",
"@babel/preset-env": "7.4.4",
+ "@babel/preset-typescript": "7.9.0",
"@storybook/addon-actions": "5.3.18",
"@storybook/addon-info": "5.3.18",
"@storybook/addon-knobs": "5.3.18", | added storybook for ts-components (#<I>) | newsuk_times-components | train |
dac6256b5f6c31797e307fa058747629ee8adfb5 | diff --git a/lib/LineMessageView.js b/lib/LineMessageView.js
index <HASH>..<HASH> 100644
--- a/lib/LineMessageView.js
+++ b/lib/LineMessageView.js
@@ -44,7 +44,11 @@ LineMessageView.content = function () {
LineMessageView.prototype.goToLine = function () {
var char = (this.character !== undefined) ? this.character - 1 : 0;
- var activeFile = Path.relative( atom.project.rootDirectory.path, atom.workspace.getActiveEditor().getUri() );
+ var activeFile;
+ var activeEditor = atom.workspace.getActiveEditor();
+ if (typeof activeEditor !== "undefined" && activeEditor !== null) {
+ activeFile = Path.relative( atom.project.rootDirectory.path, activeEditor.getUri() );
+ }
if (this.file !== undefined && this.file !== activeFile) {
atom.workspace.open(this.file, {initialLine: this.line - 1}); | Support file change when no active editors are open | tcarlsen_atom-message-panel | train |
effded2a214b6220b293402f24b7a5f9c28ddeb8 | diff --git a/lib/Models/TableDataSource.js b/lib/Models/TableDataSource.js
index <HASH>..<HASH> 100644
--- a/lib/Models/TableDataSource.js
+++ b/lib/Models/TableDataSource.js
@@ -278,7 +278,7 @@ function getPositionOfRowNumber(specialColumns, rowNumber) {
return Cartesian3.fromDegrees(
specialColumns.longitude.values[rowNumber],
specialColumns.latitude.values[rowNumber],
- defined(specialColumns.height) ? specialColumns.height.values[rowNumber] : undefined
+ defined(specialColumns.height) && !isNaN(specialColumns.height) ? specialColumns.height.values[rowNumber] : undefined
);
} | Make sure table point height isn't coerced to NaN | TerriaJS_terriajs | train |
e1ef7b463693f3b6094274346b3de942b234d14d | diff --git a/integration/v7/isolated/table_alignment_test.go b/integration/v7/isolated/table_alignment_test.go
index <HASH>..<HASH> 100644
--- a/integration/v7/isolated/table_alignment_test.go
+++ b/integration/v7/isolated/table_alignment_test.go
@@ -33,7 +33,8 @@ var _ = Describe("table alignment", func() {
})
})
- When("output is in language with multibyte characters", func() {
+ // TODO: un-pend when we have a translation for "API version:"
+ PWhen("output is in language with multibyte characters", func() {
BeforeEach(func() {
helpers.LoginCF()
helpers.TargetOrgAndSpace(ReadOnlyOrg, ReadOnlySpace)
@@ -44,10 +45,8 @@ var _ = Describe("table alignment", func() {
It("aligns the table correctly", func() {
username, _ := helpers.GetCredentials()
session := helpers.CFWithEnv(map[string]string{"LANG": "ja-JP.utf8"}, "target")
- Eventually(session).Should(Say("API エンドポイント: %s", apiURL))
- // TODO: "version" here should be translated for all languages. We have translation resources for "api version"
- // (lowercase), which is what this said in V6, but we don't yet have them for "API version" (uppercase).
- Eventually(session).Should(Say(`API version: [\d.]+`))
+ Eventually(session).Should(Say("API エンドポイント %s", apiURL))
+ Eventually(session).Should(Say(`API バージョン: [\d.]+`))
Eventually(session).Should(Say("ユーザー: %s", username))
Eventually(session).Should(Say("組織: %s", ReadOnlyOrg))
Eventually(session).Should(Say("スペース: %s", ReadOnlySpace)) | v7: Temporarily pend integration test needing translation
We recently changed the string from "api endpoint:" to "API endpoint:"
(capitalized), and we do not yet have the corresponding Japanese
translation. Pending this test until we figure that out. | cloudfoundry_cli | train |
67f2f7257765763089f5bc4a2ef51060f1bb6d49 | diff --git a/pyiso.py b/pyiso.py
index <HASH>..<HASH> 100644
--- a/pyiso.py
+++ b/pyiso.py
@@ -1629,9 +1629,6 @@ class PyIso(object):
if not self.initialized:
raise Exception("This object is not yet initialized; call either open() or new() to create an ISO")
- if self.pvd is None:
- raise Exception("This object does not have a Primary Volume Descriptor yet")
-
if not overwrite and os.path.exists(outpath):
raise Exception("Output file already exists") | Remove a redundant check from write() | clalancette_pycdlib | train |
496fecb31d8c184e5325066523b250412a633e2a | diff --git a/config/Common.php b/config/Common.php
index <HASH>..<HASH> 100644
--- a/config/Common.php
+++ b/config/Common.php
@@ -8,42 +8,42 @@ class Common extends Config
{
public function define(Container $di)
{
- $di->set('cli_context', $di->lazyNew('Aura\Cli\Context'));
- $di->set('cli_stdio', $di->lazyNew('Aura\Cli\Stdio'));
+ $di->set('aura/cli-kernel:context', $di->lazyNew('Aura\Cli\Context'));
+ $di->set('aura/cli-kernel:stdio', $di->lazyNew('Aura\Cli\Stdio'));
$di->set(
- 'cli_dispatcher',
+ 'aura/cli-kernel:dispatcher',
$di->lazyNew('Aura\Dispatcher\Dispatcher', array(
'object_param' => 'command',
)
));
$di->set(
- 'cli_help_service',
+ 'aura/cli-kernel:help_service',
$di->lazyNew('Aura\Cli_Kernel\HelpService')
);
$di->params['Aura\Cli_Kernel\CliKernel'] = array(
- 'context' => $di->lazyGet('cli_context'),
- 'stdio' => $di->lazyGet('cli_stdio'),
- 'dispatcher' => $di->lazyGet('cli_dispatcher'),
+ 'context' => $di->lazyGet('aura/cli-kernel:context'),
+ 'stdio' => $di->lazyGet('aura/cli-kernel:stdio'),
+ 'dispatcher' => $di->lazyGet('aura/cli-kernel:dispatcher'),
'logger' => $di->lazyGet('aura/project-kernel:logger'),
);
$di->params['Aura\Cli_Kernel\HelpCommand'] = array(
- 'stdio' => $di->lazyGet('cli_stdio'),
- 'dispatcher' => $di->lazyGet('cli_dispatcher'),
- 'help_service' => $di->lazyGet('cli_help_service'),
+ 'stdio' => $di->lazyGet('aura/cli-kernel:stdio'),
+ 'dispatcher' => $di->lazyGet('aura/cli-kernel:dispatcher'),
+ 'help_service' => $di->lazyGet('aura/cli-kernel:help_service'),
);
}
public function modify(Container $di)
{
- $dispatcher = $di->get('cli_dispatcher');
+ $dispatcher = $di->get('aura/cli-kernel:dispatcher');
$dispatcher->setObject(
'help',
$di->lazyNew('Aura\Cli_Kernel\HelpCommand')
);
- $help_service = $di->get('cli_help_service');
+ $help_service = $di->get('aura/cli-kernel:help_service');
$help_service->set('help', $di->lazyNew('Aura\Cli_Kernel\HelpHelp'));
}
}
diff --git a/tests/kernel/config/Kernel.php b/tests/kernel/config/Kernel.php
index <HASH>..<HASH> 100644
--- a/tests/kernel/config/Kernel.php
+++ b/tests/kernel/config/Kernel.php
@@ -27,8 +27,8 @@ class Kernel extends Config
public function modify(Container $di)
{
- $dispatcher = $di->get('cli_dispatcher');
- $stdio = $di->get('cli_stdio');
+ $dispatcher = $di->get('aura/cli-kernel:dispatcher');
+ $stdio = $di->get('aura/cli-kernel:stdio');
$dispatcher->setObject(
'aura-integration-hello',
@@ -51,7 +51,7 @@ class Kernel extends Config
}
);
- $help_service = $di->get('cli_help_service');
+ $help_service = $di->get('aura/cli-kernel:help_service');
$help_service->set('aura-integration-hello', function () use ($di) {
$help = $di->newInstance('Aura\Cli\Help');
$help->setSummary('Integration test command for hello world.'); | rename 'cli_*' services to 'aura/cli-kernel:*' | auraphp_Aura.Cli_Kernel | train |
fee7683b868bbaf4a628343e4890b5937d34c0a3 | diff --git a/spec/api-session-spec.js b/spec/api-session-spec.js
index <HASH>..<HASH> 100644
--- a/spec/api-session-spec.js
+++ b/spec/api-session-spec.js
@@ -236,9 +236,8 @@ describe('session module', function () {
})
describe('will-download event', function () {
- var w = null
-
beforeEach(function () {
+ if (w != null) w.destroy()
w = new BrowserWindow({
show: false,
width: 400,
@@ -246,10 +245,6 @@ describe('session module', function () {
})
})
- afterEach(function () {
- return closeWindow(w).then(function () { w = null })
- })
-
it('can cancel default download behavior', function (done) {
const mockFile = new Buffer(1024)
const contentDisposition = 'inline; filename="mockFile.txt"' | Reuse window variable and only close from root afterEach | electron_electron | train |
5a08bc23fb36b9895a2402fa170a1338a4562b92 | diff --git a/setup.py b/setup.py
index <HASH>..<HASH> 100644
--- a/setup.py
+++ b/setup.py
@@ -2,7 +2,7 @@ from setuptools import setup, find_packages
setup(
name='django-file-picker',
- version='0.8.0',
+ version='0.9.0',
author='Caktus Consulting Group',
author_email='solutions@caktusgroup.com',
packages=find_packages(exclude=['sample_project']), | update version in setup.py too | caktus_django-file-picker | train |
a3982500839c42e83a0e083b021ce19e2c62df0c | diff --git a/lib/cluster/services/cluster/backends/kubernetes/index.js b/lib/cluster/services/cluster/backends/kubernetes/index.js
index <HASH>..<HASH> 100644
--- a/lib/cluster/services/cluster/backends/kubernetes/index.js
+++ b/lib/cluster/services/cluster/backends/kubernetes/index.js
@@ -6,7 +6,9 @@ const parseError = require('error_parser');
const Promise = require('bluebird');
const k8sState = require('./k8sState');
const stateUtils = require('../state-utils');
-// const Queue = require('queue');
+const slicerServiceTemplate = require('./services/slicer.json');
+const slicerDeploymentTemplate = require('./deployments/slicer.json');
+const workerDeploymentTemplate = require('./deployments/worker.json');
/*
Execution Life Cycle for _status
@@ -17,11 +19,9 @@ const stateUtils = require('../state-utils');
aborted - when a job was running at the point when the cluster shutsdown
*/
-// FIXME: I want to write tests for this but I don't know how to deal with the
-// messaging stuff
-
+// FIXME: See Jared's comment regarding executionService, will leave this
+// as a known issue: https://github.com/terascope/teraslice/issues/750
module.exports = function kubernetesClusterBackend(context, messaging, executionService) {
- const events = context.apis.foundation.getSystemEvents();
const logger = context.apis.foundation.makeLogger({ module: 'kubernetes_cluster_service' });
const slicerAllocationAttempts = context.sysconfig.teraslice.slicer_allocation_attempts;
const clusterState = {};
@@ -29,6 +29,7 @@ module.exports = function kubernetesClusterBackend(context, messaging, execution
// FIXME: clusterStats stores aggregated stats about what gets processed,
// when a slice on any job fails, it gets recorded here. What should this
// do in the k8s case?
+ // See issue: https://github.com/terascope/teraslice/issues/751
const clusterStats = {
slicer: {
processed: 0,
@@ -115,10 +116,10 @@ module.exports = function kubernetesClusterBackend(context, messaging, execution
*/
function allocateSlicer(execution, recoverExecution) {
const slicerName = `teraslice-slicer-${execution.ex_id}`;
- const slicerService = require('./services/slicer.json');
- const slicerDeployment = require('./deployments/slicer.json');
+ const slicerService = _.cloneDeep(slicerServiceTemplate);
+ const slicerDeployment = _.cloneDeep(slicerDeploymentTemplate);
- // FIXME: extract this hard coded port out somewhere sensible
+ // TODO: extract this hard coded port out somewhere sensible
execution.slicer_port = 45680;
execution.slicer_hostname = slicerName;
if (recoverExecution) execution.recover_execution = recoverExecution;
@@ -162,8 +163,7 @@ module.exports = function kubernetesClusterBackend(context, messaging, execution
* @return {Promise} [description]
*/
function allocateWorkers(execution, numWorkers) {
- // FIXME: Verify that this gets a fresh worker.json on every call
- const workerDeployment = require('./deployments/worker.json');
+ const workerDeployment = _.cloneDeep(workerDeploymentTemplate);
const workerName = `teraslice-worker-${execution.ex_id}`;
workerDeployment.metadata.name = workerName;
@@ -187,7 +187,7 @@ module.exports = function kubernetesClusterBackend(context, messaging, execution
// FIXME: These functions should probably do something with the response
- // FIXME: I find is strange that the expected return value here is
+ // NOTE: I find is strange that the expected return value here is
// effectively the same as the function inputs
async function addWorkers(executionContext, numWorkers) {
const response = await k8s.scaleExecution(executionContext.ex_id, numWorkers, 'add');
diff --git a/spec/lib/cluster/services/cluster/backends/kubernetes/k8sState-spec.js b/spec/lib/cluster/services/cluster/backends/kubernetes/k8sState-spec.js
index <HASH>..<HASH> 100644
--- a/spec/lib/cluster/services/cluster/backends/kubernetes/k8sState-spec.js
+++ b/spec/lib/cluster/services/cluster/backends/kubernetes/k8sState-spec.js
@@ -3,7 +3,7 @@
const podsJobRunning = require('./files/job-running-v1-k8s-pods.json');
const k8sState = require('../../../../../../../lib/cluster/services/cluster/backends/kubernetes/k8sState');
-fdescribe('k8sState', () => {
+describe('k8sState', () => {
it('generates cluster state correctly on first call', () => {
const clusterState = {}; | removed fdescribe and made sure we clone templates | terascope_teraslice | train |
a22ae68b5f09f2ffa19e249bd76b2d6a34744d6d | diff --git a/variation/cnv.py b/variation/cnv.py
index <HASH>..<HASH> 100644
--- a/variation/cnv.py
+++ b/variation/cnv.py
@@ -378,7 +378,7 @@ def gcdepth(args):
"""
import hashlib
from jcvi.algorithms.formula import MAD_interval as confidence_interval
- from jcvi.graphics.base import plt, savefig, set2
+ from jcvi.graphics.base import latex, plt, savefig, set2
p = OptionParser(gcdepth.__doc__)
opts, args = p.parse_args(args)
@@ -420,7 +420,7 @@ def gcdepth(args):
ax.legend(handles=[patch], loc="best")
ax.set_xlim(0, 1)
ax.set_ylim(0, 100)
- ax.set_title("{} ({})".format(sample_name.split("_")[0], tag))
+ ax.set_title("{} ({})".format(latex(sample_name), tag))
ax.set_xlabel("GC content")
ax.set_ylabel("Depth")
savefig(sample_name + ".gcdepth.png") | [variation] Update latex symbols in title in cnv.gcdepth() | tanghaibao_jcvi | train |
35c6ae549a8545151aecc73975c7f8a0190e9fb0 | diff --git a/test/application_spawner_spec.rb b/test/application_spawner_spec.rb
index <HASH>..<HASH> 100644
--- a/test/application_spawner_spec.rb
+++ b/test/application_spawner_spec.rb
@@ -82,12 +82,14 @@ describe ApplicationSpawner do
end
end
-if Process.euid == ApplicationSpawner::ROOT_UID
- describe "ApplicationSpawner privilege lowering support" do
- include TestHelper
-
+
+Process.euid == ApplicationSpawner::ROOT_UID &&
+describe("ApplicationSpawner privilege lowering support") do
+ include TestHelper
+
+ describe "regular spawning" do
it_should_behave_like "a spawner that supports lowering of privileges"
-
+
def spawn_stub_application(options = {})
options = {
:lower_privilege => true,
@@ -106,4 +108,24 @@ if Process.euid == ApplicationSpawner::ROOT_UID
end
end
end
+
+ describe "conservative spawning" do
+ it_should_behave_like "a spawner that supports lowering of privileges"
+
+ def spawn_stub_application(options = {})
+ options = {
+ :lower_privilege => true,
+ :lowest_user => CONFIG['lowest_user']
+ }.merge(options)
+ @spawner = ApplicationSpawner.new(@stub.app_root,
+ options[:lower_privilege],
+ options[:lowest_user])
+ begin
+ app = @spawner.spawn_application!
+ yield app
+ ensure
+ app.close if app
+ end
+ end
+ end
end | Add more tests for conservative spawning. | phusion_passenger | train |
3242c5b4a9ce0a595d5beb280a0a2d99fecb57e4 | diff --git a/ghost/admin/controllers/modals/invite-new-user.js b/ghost/admin/controllers/modals/invite-new-user.js
index <HASH>..<HASH> 100644
--- a/ghost/admin/controllers/modals/invite-new-user.js
+++ b/ghost/admin/controllers/modals/invite-new-user.js
@@ -44,7 +44,7 @@ var InviteNewUserController = Ember.Controller.extend({
if (newUser.get('status') === 'invited-pending') {
self.notifications.showWarn('Invitation email was not sent. Please try resending.');
} else {
- self.notifications.showSuccess(notificationText, false);
+ self.notifications.showSuccess(notificationText);
}
}).catch(function (errors) {
newUser.deleteRecord();
diff --git a/ghost/admin/controllers/settings/users/user.js b/ghost/admin/controllers/settings/users/user.js
index <HASH>..<HASH> 100644
--- a/ghost/admin/controllers/settings/users/user.js
+++ b/ghost/admin/controllers/settings/users/user.js
@@ -94,7 +94,7 @@ var SettingsUserController = Ember.ObjectController.extend({
self.notifications.showWarn('Invitation email was not sent. Please try resending.');
} else {
self.get('model').set('status', result.users[0].status);
- self.notifications.showSuccess(notificationText, false);
+ self.notifications.showSuccess(notificationText);
}
}).catch(function (error) {
self.notifications.showAPIError(error);
diff --git a/ghost/admin/routes/reset.js b/ghost/admin/routes/reset.js
index <HASH>..<HASH> 100644
--- a/ghost/admin/routes/reset.js
+++ b/ghost/admin/routes/reset.js
@@ -5,7 +5,7 @@ var ResetRoute = Ember.Route.extend(styleBody, loadingIndicator, {
classNames: ['ghost-reset'],
beforeModel: function () {
if (this.get('session').isAuthenticated) {
- this.notifications.showWarn('You can\'t reset your password while you\'re signed in.', true);
+ this.notifications.showWarn('You can\'t reset your password while you\'re signed in.', { delayed: true });
this.transitionTo(SimpleAuth.Configuration.routeAfterAuthentication);
}
},
diff --git a/ghost/admin/routes/signup.js b/ghost/admin/routes/signup.js
index <HASH>..<HASH> 100644
--- a/ghost/admin/routes/signup.js
+++ b/ghost/admin/routes/signup.js
@@ -5,7 +5,7 @@ var SignupRoute = Ember.Route.extend(styleBody, loadingIndicator, {
classNames: ['ghost-signup'],
beforeModel: function () {
if (this.get('session').isAuthenticated) {
- this.notifications.showWarn('You need to sign out to register as a new user.', true);
+ this.notifications.showWarn('You need to sign out to register as a new user.', { delayed: true });
this.transitionTo(SimpleAuth.Configuration.routeAfterAuthentication);
}
}, | Cleanup from notifications refactor.
Refs #<I> | TryGhost_Ghost | train |
c760a0f151bc8be724df84c601722fae273ebf7e | diff --git a/src/Cache/Engine/MemcachedEngine.php b/src/Cache/Engine/MemcachedEngine.php
index <HASH>..<HASH> 100644
--- a/src/Cache/Engine/MemcachedEngine.php
+++ b/src/Cache/Engine/MemcachedEngine.php
@@ -147,6 +147,12 @@ class MemcachedEngine extends CacheEngine {
}
}
+ if (empty($this->_config['username']) && !empty($config['login'])) {
+ throw new InvalidArgumentException(
+ 'Please pass "username" instead of "login" for connecting to Memcached'
+ );
+ }
+
if ($this->_config['username'] !== null && $this->_config['password'] !== null) {
if (!method_exists($this->_Memcached, 'setSaslAuthData')) {
throw new InvalidArgumentException( | Adding exception for when login is passed instead of username in Memcached | cakephp_cakephp | train |
8da27a8388641f70bdfed03d5c30fc509c007bce | diff --git a/lib/dragonfly/data_storage/file_data_store.rb b/lib/dragonfly/data_storage/file_data_store.rb
index <HASH>..<HASH> 100644
--- a/lib/dragonfly/data_storage/file_data_store.rb
+++ b/lib/dragonfly/data_storage/file_data_store.rb
@@ -57,7 +57,7 @@ module Dragonfly
dirname = File.dirname(path)
basename = File.basename(path, '.*')
extname = File.extname(path)
- "#{dirname}/#{basename}_#{Time.now.usec.to_s(32)}#{extname}"
+ "#{dirname}/#{basename}_#{(Time.now.usec*10 + rand(100)).to_s(32)}#{extname}"
end
private | If FileDataStore#disambiguate was called in quick succession it was not random enough | markevans_dragonfly | train |
e762758babbb27e357393ee90d028257516c3fd0 | diff --git a/hazelcast/src/main/java/com/hazelcast/map/impl/operation/GetOperation.java b/hazelcast/src/main/java/com/hazelcast/map/impl/operation/GetOperation.java
index <HASH>..<HASH> 100644
--- a/hazelcast/src/main/java/com/hazelcast/map/impl/operation/GetOperation.java
+++ b/hazelcast/src/main/java/com/hazelcast/map/impl/operation/GetOperation.java
@@ -18,8 +18,8 @@ package com.hazelcast.map.impl.operation;
import com.hazelcast.core.OperationTimeoutException;
import com.hazelcast.internal.locksupport.LockWaitNotifyKey;
-import com.hazelcast.map.impl.MapDataSerializerHook;
import com.hazelcast.internal.serialization.Data;
+import com.hazelcast.map.impl.MapDataSerializerHook;
import com.hazelcast.spi.impl.operationservice.BlockingOperation;
import com.hazelcast.spi.impl.operationservice.WaitNotifyKey;
@@ -39,15 +39,23 @@ public final class GetOperation extends ReadonlyKeyBasedMapOperation implements
@Override
protected void runInternal() {
Object currentValue = recordStore.get(dataKey, false, getCallerAddress());
- if (!executedLocally() && currentValue instanceof Data) {
- // in case of a 'remote' call (e..g a client call) we prevent making an onheap copy of the offheap data
+ if (noCopyReadAllowed(currentValue)) {
+ // in case of a 'remote' call (e.g a client call) we prevent making
+ // an on-heap copy of the off-heap data
result = (Data) currentValue;
} else {
- // in case of a local call, we do make a copy so we can safely share it with e.g. near cache invalidation
+ // in case of a local call, we do make a copy, so we can safely share
+ // it with e.g. near cache invalidation
result = mapService.getMapServiceContext().toData(currentValue);
}
}
+ private boolean noCopyReadAllowed(Object currentValue) {
+ return currentValue instanceof Data
+ && (!getNodeEngine().getLocalMember().getUuid().equals(getCallerUuid())
+ || !super.executedLocally());
+ }
+
@Override
protected void afterRunInternal() {
mapServiceContext.interceptAfterGet(mapContainer.getInterceptorRegistry(), result); | Prevent making an on-heap copy of the off-heap data for client calls | hazelcast_hazelcast | train |
744b6cc3c93b2d8846651959a569cdac3ecbb196 | diff --git a/src/fkooman/Http/Url.php b/src/fkooman/Http/Url.php
index <HASH>..<HASH> 100644
--- a/src/fkooman/Http/Url.php
+++ b/src/fkooman/Http/Url.php
@@ -144,7 +144,7 @@ class Url
return dirname($this->srv['SCRIPT_NAME']);
}
- public function getRootUri()
+ public function getRootUrl()
{
$s = $this->getScheme();
$h = $this->getHost();
diff --git a/tests/fkooman/Http/UrlTest.php b/tests/fkooman/Http/UrlTest.php
index <HASH>..<HASH> 100644
--- a/tests/fkooman/Http/UrlTest.php
+++ b/tests/fkooman/Http/UrlTest.php
@@ -38,7 +38,7 @@ class UrlTest extends PHPUnit_Framework_TestCase
$this->assertEquals(80, $u->getPort());
$this->assertEquals('/bar/index.php', $u->getRoot());
$this->assertNull($u->getPathInfo());
- $this->assertEquals('http://www.example.org/bar/index.php', $u->getRootUri());
+ $this->assertEquals('http://www.example.org/bar/index.php', $u->getRootUrl());
$this->assertEquals(array('foo' => 'bar'), $u->getQueryArray());
$this->assertEquals('bar', $u->getQueryParameter('foo'));
}
@@ -60,7 +60,7 @@ class UrlTest extends PHPUnit_Framework_TestCase
$this->assertEquals(443, $u->getPort());
$this->assertNull($u->getPathInfo());
$this->assertEquals('/bar/index.php', $u->getRoot());
- $this->assertEquals('https://www.example.org/bar/index.php', $u->getRootUri());
+ $this->assertEquals('https://www.example.org/bar/index.php', $u->getRootUrl());
$this->assertEquals(array('foo' => 'bar'), $u->getQueryArray());
$this->assertEquals('bar', $u->getQueryParameter('foo'));
}
@@ -81,7 +81,7 @@ class UrlTest extends PHPUnit_Framework_TestCase
$this->assertEquals(8080, $u->getPort());
$this->assertNull($u->getPathInfo());
$this->assertEquals('/bar/index.php', $u->getRoot());
- $this->assertEquals('http://www.example.org:8080/bar/index.php', $u->getRootUri());
+ $this->assertEquals('http://www.example.org:8080/bar/index.php', $u->getRootUrl());
$this->assertEquals(array('foo' => 'bar'), $u->getQueryArray());
$this->assertEquals('bar', $u->getQueryParameter('foo'));
}
@@ -103,7 +103,7 @@ class UrlTest extends PHPUnit_Framework_TestCase
$this->assertEquals(80, $u->getPort());
$this->assertEquals('/bar/index.php', $u->getRoot());
$this->assertEquals('/def', $u->getPathInfo());
- $this->assertEquals('http://www.example.org/bar/index.php', $u->getRootUri());
+ $this->assertEquals('http://www.example.org/bar/index.php', $u->getRootUrl());
$this->assertEquals(array('foo' => 'bar'), $u->getQueryArray());
$this->assertEquals('bar', $u->getQueryParameter('foo'));
}
@@ -125,7 +125,7 @@ class UrlTest extends PHPUnit_Framework_TestCase
$this->assertEquals(80, $u->getPort());
$this->assertEquals('/bar', $u->getRoot());
$this->assertEquals('/def', $u->getPathInfo());
- $this->assertEquals('http://www.example.org/bar', $u->getRootUri());
+ $this->assertEquals('http://www.example.org/bar', $u->getRootUrl());
$this->assertEquals(array('foo' => 'bar'), $u->getQueryArray());
$this->assertEquals('bar', $u->getQueryParameter('foo'));
}
@@ -146,7 +146,7 @@ class UrlTest extends PHPUnit_Framework_TestCase
$this->assertEquals(80, $u->getPort());
$this->assertEquals('/bar/index.php', $u->getRoot());
$this->assertNull($u->getPathInfo());
- $this->assertEquals('http://www.example.org/bar/index.php', $u->getRootUri());
+ $this->assertEquals('http://www.example.org/bar/index.php', $u->getRootUrl());
$this->assertEquals(array(), $u->getQueryArray());
$this->assertNull($u->getQueryParameter('foo'));
}
@@ -168,7 +168,7 @@ class UrlTest extends PHPUnit_Framework_TestCase
$this->assertEquals(443, $u->getPort());
$this->assertEquals('/bar/index.php', $u->getRoot());
$this->assertNull($u->getPathInfo());
- $this->assertEquals('https://www.example.org/bar/index.php', $u->getRootUri());
+ $this->assertEquals('https://www.example.org/bar/index.php', $u->getRootUrl());
$this->assertEquals(array('foo' => 'bar'), $u->getQueryArray());
} | rename `getRootUri()` to `getRootUrl()` | fkooman_php-lib-rest | train |
f97006b5cbdd9813e6715eb5960f39e1dd402603 | diff --git a/master/buildbot/status/builder.py b/master/buildbot/status/builder.py
index <HASH>..<HASH> 100644
--- a/master/buildbot/status/builder.py
+++ b/master/buildbot/status/builder.py
@@ -336,7 +336,8 @@ class BuilderStatus(styles.Versioned):
max_buildnum=None,
finished_before=None,
results=None,
- max_search=200):
+ max_search=200,
+ filter_fn=None):
got = 0
branches = set(branches)
for Nb in itertools.count(1):
@@ -363,6 +364,9 @@ class BuilderStatus(styles.Versioned):
if results is not None:
if build.getResults() not in results:
continue
+ if filter_fn is not None:
+ if not filter_fn(build):
+ continue
got += 1
yield build
if num_builds is not None: | Accept a filter callable in BuilderStatus's generateFinishedBuilds | buildbot_buildbot | train |
1cf0b335b49b6c420d154ba449322dd8b21f2a0f | diff --git a/Mbh/Router.php b/Mbh/Router.php
index <HASH>..<HASH> 100644
--- a/Mbh/Router.php
+++ b/Mbh/Router.php
@@ -44,7 +44,7 @@ final class Router implements RouterInterface
{
$this->rootPath = (string) (new Path($rootPath))->normalize()->removeTrailingSlashes();
$this->route = urldecode((string) (new Uri($_SERVER['REQUEST_URI']))->removeQuery());
- $this->requestMethod = strtolower($_SERVER['REQUEST_METHOD']);
+ $this->requestMethod = $_SERVER['REQUEST_METHOD'];
$this->methods = [
'GET',
@@ -262,7 +262,8 @@ final class Router implements RouterInterface
*/
public function map(array $methods, $pattern, $callback = null, $inject = null)
{
- $methods = array_map('strtolower', $methods);
+ // According to RFC methods are defined in uppercase (See RFC 7231)
+ $methods = array_map("strtoupper", $methods);
if (in_array($this->requestMethod, $methods, true)) {
$matches = $this->matchRoute($pattern); | Router working. Preparing for important mod. | MBHFramework_mbh-rest | train |
28d6ae57cd197598b0e2b589e7b7a91ff1a76a4a | diff --git a/cocaine/proxy/jsonrpc.py b/cocaine/proxy/jsonrpc.py
index <HASH>..<HASH> 100644
--- a/cocaine/proxy/jsonrpc.py
+++ b/cocaine/proxy/jsonrpc.py
@@ -58,7 +58,7 @@ class JSONRPC(IPlugin):
return
api = dict((data[0], data) for data in service.api.itervalues())
- if method not in api.keys():
+ if method not in api:
JSONRPC._send_400_error(request, -32601, 'Method not found.')
return
diff --git a/tests/test_jsonrpc.py b/tests/test_jsonrpc.py
index <HASH>..<HASH> 100644
--- a/tests/test_jsonrpc.py
+++ b/tests/test_jsonrpc.py
@@ -58,4 +58,3 @@ class TestJSONRPC(AsyncTestCase):
ResponseStartLine(version='HTTP/1.1', code=400, reason='Bad JSON-RPC request'),
mock.ANY,
'{"message": "The JSON sent is not a valid Request object.", "code": -32600}')
- | refactor: check existence in map instead of list | cocaine_cocaine-tools | train |
9155de035795c00594f5e4b3878bbaf763ff272d | diff --git a/tests/test_lilypond.py b/tests/test_lilypond.py
index <HASH>..<HASH> 100644
--- a/tests/test_lilypond.py
+++ b/tests/test_lilypond.py
@@ -165,6 +165,8 @@ class TestLilyPond(TestCase):
)
def test_octave_check(self):
+ import logging
+ logging.disable(logging.WARN)
self.eq(
r"\relative c'' { c2 d='4 d e2 f }",
[
@@ -173,6 +175,7 @@ class TestLilyPond(TestCase):
(128, None, None),
]
)
+ logging.disable(logging.NOTSET)
def test_acciaccatura(self):
self.eq( | supress warning during octave check | jtauber_sebastian | train |
2b0b6d3364e14028e20ee4cdfee5309d07a48237 | diff --git a/tests/runners.py b/tests/runners.py
index <HASH>..<HASH> 100644
--- a/tests/runners.py
+++ b/tests/runners.py
@@ -66,17 +66,6 @@ class Remote_(Spec):
# with the Runner instead being a Remote. Or do we just replicate the
# basics?
- def may_wrap_command_with_things_like_bash_dash_c(self):
- "may wrap command with things like bash -c"
- # TODO: how? also implies top level run() wants to pass **kwargs to
- # runner somehow, though that's dangerous; maybe allow runner to
- # expose what it expects so run() can correctly determine things.
- # TODO: oughtn't this be part of invoke proper?
- skip()
-
- def does_not_wrap_command_by_default(self):
- skip()
-
# TODO: all other run() tests from fab1... | Yea this part is absolutely Invoke not Fabric | fabric_fabric | train |
6cd194b32856d397b9f3aa5dc96dfadfd1f8f79c | diff --git a/src/test/java/io/sniffy/servlet/SnifferFilterTest.java b/src/test/java/io/sniffy/servlet/SnifferFilterTest.java
index <HASH>..<HASH> 100644
--- a/src/test/java/io/sniffy/servlet/SnifferFilterTest.java
+++ b/src/test/java/io/sniffy/servlet/SnifferFilterTest.java
@@ -1,7 +1,6 @@
package io.sniffy.servlet;
import io.sniffy.BaseTest;
-import org.junit.Ignore;
import org.junit.Test;
import org.junit.runner.RunWith;
import org.mockito.Mock;
@@ -36,7 +35,7 @@ public class SnifferFilterTest extends BaseTest {
MockMvcRequestBuilders.get("/petclinic/foo/bar?baz").contextPath("/petclinic").buildRequest(servletContext);
private SnifferFilter filter = new SnifferFilter();
- protected FilterConfig getFilterConfig() {
+ private FilterConfig getFilterConfig() {
FilterConfig filterConfig = mock(FilterConfig.class);
when(filterConfig.getInitParameter("inject-html")).thenReturn("true");
when(filterConfig.getInitParameter("exclude-pattern")).thenReturn("^/baz/.*$");
@@ -61,6 +60,19 @@ public class SnifferFilterTest extends BaseTest {
}
@Test
+ public void testDestroy() throws IOException, ServletException {
+
+ FilterConfig filterConfig = getFilterConfig();
+ when(filterConfig.getInitParameter("exclude-pattern")).thenReturn("^/foo/ba.*$");
+
+ SnifferFilter filter = new SnifferFilter();
+ filter.init(filterConfig);
+
+ filter.destroy();
+
+ }
+
+ @Test
public void testGetSnifferJs() throws IOException, ServletException {
FilterConfig filterConfig = getFilterConfig();
@@ -81,6 +93,26 @@ public class SnifferFilterTest extends BaseTest {
}
@Test
+ public void testGetSnifferJsMap() throws IOException, ServletException {
+
+ FilterConfig filterConfig = getFilterConfig();
+ when(filterConfig.getInitParameter("exclude-pattern")).thenReturn("^.*(\\.js|\\.css)$");
+
+ SnifferFilter filter = new SnifferFilter();
+ filter.init(filterConfig);
+
+ MockHttpServletRequest httpServletRequest = MockMvcRequestBuilders.
+ get("/petclinic" + SnifferFilter.JAVASCRIPT_MAP_URI).
+ contextPath("/petclinic").buildRequest(servletContext);
+
+ filter.doFilter(httpServletRequest, httpServletResponse, filterChain);
+
+ assertFalse(httpServletResponse.getHeaderNames().contains(HEADER_NUMBER_OF_QUERIES));
+ assertTrue(httpServletResponse.getContentLength() > 100);
+
+ }
+
+ @Test
public void testFilterNoQueries() throws IOException, ServletException {
filter.doFilter(httpServletRequest, httpServletResponse, filterChain);
@@ -121,7 +153,6 @@ public class SnifferFilterTest extends BaseTest {
}
-
@Test
public void testDisabledFilterOneQuery() throws IOException, ServletException {
@@ -138,6 +169,24 @@ public class SnifferFilterTest extends BaseTest {
}
@Test
+ public void testDisabledInConfigFilterOneQuery() throws IOException, ServletException {
+
+ doAnswer(invocation -> {executeStatement(); return null;}).
+ when(filterChain).doFilter(any(), any());
+
+ FilterConfig filterConfig = getFilterConfig();
+ when(filterConfig.getInitParameter("enabled")).thenReturn("false");
+
+ SnifferFilter filter = new SnifferFilter();
+ filter.init(filterConfig);
+
+ filter.doFilter(httpServletRequest, httpServletResponse, filterChain);
+
+ assertFalse(httpServletResponse.containsHeader(HEADER_NUMBER_OF_QUERIES));
+
+ }
+
+ @Test
public void testFilterEnabledByRequestParameter() throws IOException, ServletException {
doAnswer(invocation -> {executeStatement(); return null;}).
when(filterChain).doFilter(any(), any());
@@ -155,7 +204,7 @@ public class SnifferFilterTest extends BaseTest {
when(filterChain).doFilter(any(), any());
SnifferFilter filter = new SnifferFilter();
filter.setEnabled(false);
- httpServletRequest.setCookies(null);
+ httpServletRequest.setCookies((Cookie[]) null);
filter.doFilter(httpServletRequest, httpServletResponse, filterChain);
assertFalse(httpServletResponse.containsHeader(HEADER_NUMBER_OF_QUERIES));
}
@@ -447,7 +496,6 @@ public class SnifferFilterTest extends BaseTest {
}
@Test
- @Ignore("spring test framework bug")
public void testInjectHtmlSetContentLengthIntHeader() throws IOException, ServletException {
String actualContent = "<html><head><title>Title</title></head><body>Hello, World!</body></html>"; | Added more unit tests for SnifferFilter | sniffy_sniffy | train |
d15cd85519de934615ab2dfaa9512a852e2be0bc | diff --git a/src/com/mebigfatguy/fbcontrib/detect/MethodReturnsConstant.java b/src/com/mebigfatguy/fbcontrib/detect/MethodReturnsConstant.java
index <HASH>..<HASH> 100755
--- a/src/com/mebigfatguy/fbcontrib/detect/MethodReturnsConstant.java
+++ b/src/com/mebigfatguy/fbcontrib/detect/MethodReturnsConstant.java
@@ -19,10 +19,13 @@
package com.mebigfatguy.fbcontrib.detect;
import java.util.HashMap;
+import java.util.HashSet;
import java.util.Map;
+import java.util.Set;
import org.apache.bcel.Constants;
import org.apache.bcel.classfile.Code;
+import org.apache.bcel.classfile.JavaClass;
import org.apache.bcel.classfile.Method;
import com.mebigfatguy.fbcontrib.utils.BugType;
@@ -47,6 +50,7 @@ public class MethodReturnsConstant extends BytecodeScanningDetector {
private OpcodeStack stack;
private Integer returnRegister;
private Map<Integer, Object> registerConstants;
+ private Set<Method> overloadedMethods;
private Object returnConstant;
private int returnPC;
@@ -60,15 +64,24 @@ public class MethodReturnsConstant extends BytecodeScanningDetector {
this.bugReporter = bugReporter;
}
+ /**
+ * implements the visitor to collect all methods that are overloads. These methods should be ignored, as you may differentiate constants based on parameter
+ * type, or value.
+ *
+ * @param classContext
+ * the currently parsed class object
+ */
@Override
public void visitClassContext(ClassContext classContext) {
try {
stack = new OpcodeStack();
registerConstants = new HashMap<>();
+ overloadedMethods = collectOverloadedMethods(classContext.getJavaClass());
super.visitClassContext(classContext);
} finally {
stack = null;
registerConstants = null;
+ overloadedMethods = null;
}
}
@@ -81,6 +94,11 @@ public class MethodReturnsConstant extends BytecodeScanningDetector {
@Override
public void visitCode(Code obj) {
Method m = getMethod();
+
+ if (overloadedMethods.contains(m)) {
+ return;
+ }
+
int aFlags = m.getAccessFlags();
if ((((aFlags & Constants.ACC_PRIVATE) != 0) || ((aFlags & Constants.ACC_STATIC) != 0)) && ((aFlags & Constants.ACC_SYNTHETIC) == 0)
&& (!m.getSignature().endsWith(")Z"))) {
@@ -204,4 +222,38 @@ public class MethodReturnsConstant extends BytecodeScanningDetector {
}
}
}
+
+ /**
+ * adds all methods of a class that are overloaded to a set. This method is O(nlogn) so for large classes might be slow. Assuming on average it's better
+ * than other choices. When a match is found in the j index, it is removed from the array, so it is not scanned with the i index
+ *
+ * @param cls
+ * the class to look for overloaded methods
+ * @return the set of methods that are overloaded
+ */
+ private Set<Method> collectOverloadedMethods(JavaClass cls) {
+
+ Set<Method> overloads = new HashSet<>();
+
+ Method[] methods = cls.getMethods();
+ int numMethods = methods.length;
+
+ for (int i = 0; i < numMethods; i++) {
+ boolean foundOverload = false;
+
+ for (int j = i + 1; j < numMethods; j++) {
+
+ if (methods[i].getName().equals(methods[j].getName())) {
+ overloads.add(methods[j]);
+ methods[j--] = methods[--numMethods];
+ foundOverload = true;
+ }
+ }
+ if (foundOverload) {
+ overloads.add(methods[i]);
+ }
+ }
+
+ return overloads;
+ }
} | don't report MRC for overloaded methods | mebigfatguy_fb-contrib | train |
48897d365409235e413f75bec38244c3de67f966 | diff --git a/tests/imbu/unit/imbu_models_test.py b/tests/imbu/unit/imbu_models_test.py
index <HASH>..<HASH> 100644
--- a/tests/imbu/unit/imbu_models_test.py
+++ b/tests/imbu/unit/imbu_models_test.py
@@ -58,11 +58,11 @@ class TestImbu(unittest.TestCase):
shutil.rmtree("fake_cache_root")
- def _setupFakeImbuModelsInstance(self):
+ def _setupFakeImbuModelsInstance(self, retina="en_associative"):
return ImbuModels(
cacheRoot="fake_cache_root",
dataPath=self.dataPath,
- retina="en_associative",
+ retina=retina,
apiKey=os.environ.get("CORTICAL_API_KEY")
)
@@ -99,8 +99,11 @@ class TestImbu(unittest.TestCase):
def _exerciseModelLifecycle(self, modelType, queryTerm="unicorn",
- networkConfigName="imbu_sensor_knn.json"):
- imbu = self._setupFakeImbuModelsInstance()
+ networkConfigName="imbu_sensor_knn.json",
+ retina="en_associative"):
+ """ Create, save, load, and assert consistent results."""
+
+ imbu = self._setupFakeImbuModelsInstance(retina=retina)
checkpointLocation = self._createTempModelCheckpoint()
@@ -129,20 +132,36 @@ class TestImbu(unittest.TestCase):
imbu.query(model, queryTerm))))
- def testCreateSaveLoadCioWordFingerprintModel(self):
+ def testCreateSaveLoadCioWordFingerprint(self):
self._exerciseModelLifecycle("CioWordFingerprint")
- def testCreateSaveLoadCioDocumentFingerprintModel(self):
+ def testCreateSaveLoadCioDocumentFingerprint(self):
self._exerciseModelLifecycle("CioDocumentFingerprint")
- def testCreateSaveLoadHTMNetworkModel(self):
- self._exerciseModelLifecycle("HTMNetwork")
+ def testCreateSaveLoadKeywords(self):
+ self._exerciseModelLifecycle("Keywords")
- def testLoadKeywordsModel(self):
- self._exerciseModelLifecycle("Keywords")
+ def testCreateSaveLoadSensorNetwork(self):
+ self._exerciseModelLifecycle(
+ "HTM_sensor_knn",
+ networkConfigName="imbu_sensor_knn.json")
+
+
+ def testCreateSaveLoadSensorSimpleTPNetwork(self):
+ self._exerciseModelLifecycle(
+ "HTM_sensor_simple_tp_knn",
+ networkConfigName="imbu_sensor_simple_tp_knn.json",
+ retina="en_associative_64_univ")
+
+
+ def testCreateSaveLoadSensorTMSimpleTPNetwork(self):
+ self._exerciseModelLifecycle(
+ "HTM_sensor_tm_simple_tp_knn",
+ networkConfigName="imbu_sensor_tm_simple_tp_knn.json",
+ retina="en_associative_64_univ")
def testMappingModelNamesToModelTypes(self): | unit test lifecycle for all Imbu models | numenta_htmresearch | train |
e1d66e326ea50d545c8a513f91b973147304265a | diff --git a/app-servers/SocksServer.php b/app-servers/SocksServer.php
index <HASH>..<HASH> 100644
--- a/app-servers/SocksServer.php
+++ b/app-servers/SocksServer.php
@@ -11,7 +11,7 @@ class SocksServer extends AsyncServer
{
Daemon::$settings += array(
'mod'.$this->modname.'listen' => 'tcp://0.0.0.0',
- 'mod'.$this->modname.'listenport' => 1080,
+ 'mod'.$this->modname.'listenport' => 1081,
'mod'.$this->modname.'auth' => 0,
'mod'.$this->modname.'username' => 'User',
'mod'.$this->modname.'password' => 'Password',
@@ -39,6 +39,7 @@ class SocksSession extends SocketSession
{
public $ver; // protocol version (X'04' / X'05')
public $state = 0; // (0 - start, 1 - aborted, 2 - handshaked, 3 - authorized, 4 - data exchange)
+ public $slave;
/* @method stdin
@description Called when new data recieved.
@param string New data.
@@ -143,11 +144,45 @@ class SocksSession extends SocketSession
$this->buf = binarySubstr($this->buf,$pl);
$connId = $this->appInstance->connectTo($this->destAddr = $address,$this->destPort = $port);
- $this->slave = $this->appInstance->sessions[$connId] = new SocksServerSlaveSession($connId,$this->appInstance);
- $this->slave->client = $this;
- $this->slave->write($this->buf);
- $this->buf = '';
+ if (!$connId) // Early connection error
+ {
+ $this->write($this->ver."\x05");
+ $this->finish();
+ }
+ else
+ {
+ $this->slave = $this->appInstance->sessions[$connId] = new SocksServerSlaveSession($connId,$this->appInstance);
+ $this->slave->client = $this;
+ $this->slave->write($this->buf);
+ $this->buf = '';
+ $this->state = 4;
+ }
+ }
+ }
+ public function onSlaveReady($code)
+ {
+ $reply =
+ $this->ver // Version
+ .chr($code) // Status
+ ."\x00" // Reserved
+ ;
+ if (Daemon::$useSockets && socket_getsockname(Daemon::$worker->pool[$this->connId],$address,$port))
+ {
+ $reply .=
+ (strpos($address,':') === FALSE?"\x01":"\x04") // IPv4/IPv6
+ .inet_pton($address) // Address
+ ."\x00\x00"//pack('n',$port) // Port
+ ;
}
+ else
+ {
+ $reply .=
+ "\x01"
+ ."\x00\x00\x00\x00"
+ ."\x00\x00"
+ ;
+ }
+ $this->write($reply);
}
/* @method onFinish
@description Event of SocketSession (asyncServer).
@@ -155,13 +190,26 @@ class SocksSession extends SocketSession
*/
public function onFinish()
{
- $this->slave->finish();
+ if ($this->slave) {$this->slave->finish();}
unset($this->slave);
}
}
class SocksServerSlaveSession extends SocketSession
{
public $client;
+ public $ready = FALSE;
+ /* @method onwrite
+ @description Called when the connection is ready to accept new data.
+ @return void
+ */
+ public function onWrite()
+ {
+ if (!$this->ready)
+ {
+ $this->ready = TRUE;
+ $this->client->onSlaveReady(0x00);
+ }
+ }
/* @method stdin
@description Called when new data recieved.
@param string New data.
@@ -177,7 +225,7 @@ class SocksServerSlaveSession extends SocketSession
*/
public function onFinish()
{
- unset($this->client);
$this->client->finish();
+ unset($this->client);
}
} | SocksServer.php replaced with newer version (mis-commit). | kakserpom_phpdaemon | train |
d44cd9f64bf2be963e9fda8a4195386c5cea62bc | diff --git a/lib/isono/rack/proc.rb b/lib/isono/rack/proc.rb
index <HASH>..<HASH> 100644
--- a/lib/isono/rack/proc.rb
+++ b/lib/isono/rack/proc.rb
@@ -15,19 +15,20 @@ module Rack
end
def call(req, res)
- dup.__send__(:_call, req, res)
- end
-
- private
- def _call(req, res)
Thread.current["#{THREAD_LOCAL_KEY}/request"] = req
Thread.current["#{THREAD_LOCAL_KEY}/response"] = res
begin
- # handle response in the block
- @context.extend InjectMethods
- @context.instance_eval(&@blk)
- # send empty message back to client if the response is not handled in block.
- res.response(nil) unless res.responded?
+ # create per-request context object from original.
+ c = @context.dup
+ c.extend InjectMethods
+ begin
+ c.instance_eval(&@blk)
+ # send empty message back to client if the response is not handled in block.
+ res.response(nil) unless res.responded?
+ rescue ::Exception => e
+ res.response(e)
+ raise e
+ end
ensure
Thread.current["#{THREAD_LOCAL_KEY}/request"] = nil
Thread.current["#{THREAD_LOCAL_KEY}/response"] = nil | change Rack::Proc to duplicate context object per request basis. | axsh_isono | train |
c8bd6ffea50664f0e4ac64b3c06abf58627a8610 | diff --git a/src/main/java/org/apache/jmeter/JMeterMojo.java b/src/main/java/org/apache/jmeter/JMeterMojo.java
index <HASH>..<HASH> 100644
--- a/src/main/java/org/apache/jmeter/JMeterMojo.java
+++ b/src/main/java/org/apache/jmeter/JMeterMojo.java
@@ -34,6 +34,7 @@ import org.apache.tools.ant.DirectoryScanner;
*
* @author Tim McCune
* @goal jmeter
+ * @requiresProject true
*/
public class JMeterMojo extends AbstractMojo {
@@ -143,8 +144,7 @@ public class JMeterMojo extends AbstractMojo {
private String reportPostfix;
private File workDir;
- private File saveServiceProps;
- private File upgradeProps;
+ private List<File> temporaryPropertyFiles = new ArrayList<File>();
private File jmeterLog;
private DateFormat fmt = new SimpleDateFormat("yyMMdd");
@@ -170,8 +170,9 @@ public class JMeterMojo extends AbstractMojo {
}
checkForErrors(results);
} finally {
- saveServiceProps.delete();
- upgradeProps.delete();
+ for(File temporaryPropertyFile : temporaryPropertyFiles) {
+ temporaryPropertyFile.delete();
+ }
}
}
@@ -239,7 +240,9 @@ public class JMeterMojo extends AbstractMojo {
private void initSystemProps() throws MojoExecutionException {
workDir = new File("target" + File.separator + "jmeter");
workDir.mkdirs();
- createSaveServiceProps();
+ createTemporaryProperties();
+ resolveJmeterArtifact();
+
jmeterLog = new File(workDir, "jmeter.log");
try {
System.setProperty("log_file", jmeterLog.getCanonicalPath());
@@ -249,6 +252,19 @@ public class JMeterMojo extends AbstractMojo {
}
/**
+ * Resolve JMeter artifact, set necessary System Property.
+ *
+ * This mess is necessary because JMeter must load this info from a file.
+ * Loading resources from classpath won't work.
+ */
+ private void resolveJmeterArtifact() {
+ //set search path for JMeter. JMeter loads function classes from this path.
+ System.setProperty("search_paths", repoDir.toString() + "/org/apache/jmeter/jmeter/2.4/jmeter-2.4.jar");
+ }
+
+ /**
+ * Create temporary property files and set necessary System Properties.
+ *
* This mess is necessary because JMeter must load this info from a file.
* Loading resources from classpath won't work.
*
@@ -256,25 +272,24 @@ public class JMeterMojo extends AbstractMojo {
* Exception
*/
@SuppressWarnings("unchecked")
- private void createSaveServiceProps() throws MojoExecutionException {
- saveServiceProps = new File(workDir, "saveservice.properties");
- upgradeProps = new File(workDir, "upgrade.properties");
- try {
- FileWriter out = new FileWriter(saveServiceProps);
- IOUtils.copy(Thread.currentThread().getContextClassLoader().getResourceAsStream("saveservice.properties"), out);
- out.flush();
- out.close();
- System.setProperty("saveservice_properties", File.separator + "target" + File.separator + "jmeter" + File.separator + "saveservice.properties");
-
- out = new FileWriter(upgradeProps);
- IOUtils.copy(Thread.currentThread().getContextClassLoader().getResourceAsStream("upgrade.properties"), out);
- out.flush();
- out.close();
- System.setProperty("upgrade_properties", File.separator + "target" + File.separator + "jmeter" + File.separator + "upgrade.properties");
-
- System.setProperty("search_paths", repoDir.toString() + "/org/apache/jmeter/jmeter/2.4/jmeter-2.4.jar");
- } catch (IOException e) {
- throw new MojoExecutionException("Could not create temporary saveservice.properties", e);
+ private void createTemporaryProperties() throws MojoExecutionException {
+ String jmeterTargetDir = File.separator + "target" + File.separator + "jmeter" + File.separator;
+ File saveServiceProps = new File(workDir, "saveservice.properties");
+ System.setProperty("saveservice_properties", jmeterTargetDir + saveServiceProps.getName());
+ temporaryPropertyFiles.add(saveServiceProps);
+ File upgradeProps = new File(workDir, "upgrade.properties");
+ System.setProperty("upgrade_properties", jmeterTargetDir + upgradeProps.getName());
+ temporaryPropertyFiles.add(upgradeProps);
+
+ for (File propertyFile : temporaryPropertyFiles) {
+ try {
+ FileWriter out = new FileWriter(propertyFile);
+ IOUtils.copy(Thread.currentThread().getContextClassLoader().getResourceAsStream(propertyFile.getName()), out);
+ out.flush();
+ out.close();
+ } catch (IOException e) {
+ throw new MojoExecutionException("Could not create temporary property file "+propertyFile.getName()+" in directory "+jmeterTargetDir, e);
+ }
}
} | refactored handling of temporary files and resolving of JMeter artifact | jmeter-maven-plugin_jmeter-maven-plugin | train |
423b17f8a904379d6520cff69b13dffc1493f222 | diff --git a/test/functional/ft_30_smtp_participant.rb b/test/functional/ft_30_smtp_participant.rb
index <HASH>..<HASH> 100644
--- a/test/functional/ft_30_smtp_participant.rb
+++ b/test/functional/ft_30_smtp_participant.rb
@@ -35,8 +35,8 @@ class NftSmtpParticipantTest < Test::Unit::TestCase
end
end
- trapfile = '/tmp/ruote_mailtrap.txt'
- FileUtils.rm(trapfile) rescue nil
+ trapfile = Ruote::WIN ? 'ruote_mailtrap.txt' : '/tmp/ruote_mailtrap.txt'
+ FileUtils.rm_f(trapfile)
Thread.new do
Trap.new('127.0.0.1', 2525, true, trapfile) | adapted SmtpParticipant test to windows | jmettraux_ruote | train |
313bc1e339fa4292d9ef5bc74acc82436eab7e1e | diff --git a/daemon/disk_usage.go b/daemon/disk_usage.go
index <HASH>..<HASH> 100644
--- a/daemon/disk_usage.go
+++ b/daemon/disk_usage.go
@@ -99,7 +99,6 @@ func (daemon *Daemon) SystemDiskUsage(ctx context.Context) (*types.DiskUsage, er
for platform := range daemon.stores {
layerRefs := daemon.getLayerRefs(platform)
allLayers := daemon.stores[platform].layerStore.Map()
- var allLayersSize int64
for _, l := range allLayers {
select {
case <-ctx.Done(): | Fix variable shadowing causing LayersSize to be reported as 0 | moby_moby | train |
94a7c20d77c8afe8acc97209983ee68663ae70dd | diff --git a/uaa/src/test/java/org/cloudfoundry/identity/uaa/integration/feature/OIDCLoginIT.java b/uaa/src/test/java/org/cloudfoundry/identity/uaa/integration/feature/OIDCLoginIT.java
index <HASH>..<HASH> 100644
--- a/uaa/src/test/java/org/cloudfoundry/identity/uaa/integration/feature/OIDCLoginIT.java
+++ b/uaa/src/test/java/org/cloudfoundry/identity/uaa/integration/feature/OIDCLoginIT.java
@@ -125,7 +125,21 @@ public class OIDCLoginIT {
Assert.assertThat(webDriver.findElement(By.linkText("My OIDC Provider")).getAttribute("href"), Matchers.containsString("scope=openid+cloud_controller.read"));
}
+ @Test
+ public void scopesIncludedInAuthorizeRequest_When_Issuer_Set() throws Exception {
+ createOIDCProviderWithRequestedScopes("https://oidc10.identity.cf-app.com/oauth/token");
+ try {
+ webDriver.get(appUrl);
+ } finally {
+ IntegrationTestUtils.takeScreenShot(webDriver);
+ }
+ Assert.assertThat(webDriver.findElement(By.linkText("My OIDC Provider")).getAttribute("href"), Matchers.containsString("scope=openid+cloud_controller.read"));
+ }
+
private void createOIDCProviderWithRequestedScopes() throws Exception {
+ createOIDCProviderWithRequestedScopes(null);
+ }
+ private void createOIDCProviderWithRequestedScopes(String issuer) throws Exception {
IdentityProvider<AbstractXOAuthIdentityProviderDefinition> identityProvider = new IdentityProvider<>();
identityProvider.setName("my oidc provider");
identityProvider.setIdentityZoneId(OriginKeys.UAA);
@@ -139,6 +153,7 @@ public class OIDCLoginIT {
config.setSkipSslValidation(true);
config.setRelyingPartyId("identity");
config.setRelyingPartySecret("identitysecret");
+ config.setIssuer(issuer);
List<String> requestedScopes = new ArrayList<>();
requestedScopes.add("openid");
requestedScopes.add("cloud_controller.read"); | Add a test where we configure the issuer. Note this one is the same issuer, but the server gets it from the config.
[#<I>] <URL> | cloudfoundry_uaa | train |
4a918740331bfec899060ddf2e06a438dc4c8ba9 | diff --git a/tests/Unit/Suites/DataPool/SearchEngine/AbstractSearchEngineTest.php b/tests/Unit/Suites/DataPool/SearchEngine/AbstractSearchEngineTest.php
index <HASH>..<HASH> 100644
--- a/tests/Unit/Suites/DataPool/SearchEngine/AbstractSearchEngineTest.php
+++ b/tests/Unit/Suites/DataPool/SearchEngine/AbstractSearchEngineTest.php
@@ -858,44 +858,6 @@ abstract class AbstractSearchEngineTest extends \PHPUnit_Framework_TestCase
$this->assertOrder($expectedOrder, $searchEngineResponse->getProductIds());
}
- public function testSearchResultsAreSortedByStringValuesCaseInsensitively()
- {
- $productAId = ProductId::fromString('A');
- $productBId = ProductId::fromString('B');
- $productCId = ProductId::fromString('C');
-
- $fieldCode = 'foo';
- $fieldValue = 0;
-
- $documentA = $this->createSearchDocument([$fieldCode => 1], $productAId);
- $documentB = $this->createSearchDocument([$fieldCode => 3], $productBId);
- $documentC = $this->createSearchDocument([$fieldCode => 2], $productCId);
- $stubSearchDocumentCollection = $this->createStubSearchDocumentCollection($documentA, $documentB, $documentC);
-
- $this->searchEngine->addSearchDocumentCollection($stubSearchDocumentCollection);
-
- $criteria = SearchCriterionGreaterOrEqualThan::create($fieldCode, $fieldValue);
- $selectedFilters = [];
- $facetFilterRequest = new FacetFilterRequest;
- $rowsPerPage = 100;
- $pageNumber = 0;
- $sortOrderConfig = $this->createStubSortOrderConfig($fieldCode, SortOrderDirection::ASC);
-
- $searchEngineResponse = $this->searchEngine->getSearchDocumentsMatchingCriteria(
- $criteria,
- $selectedFilters,
- $this->testContext,
- $facetFilterRequest,
- $rowsPerPage,
- $pageNumber,
- $sortOrderConfig
- );
-
- $expectedOrder = [$productAId, $productCId, $productBId];
-
- $this->assertOrder($expectedOrder, $searchEngineResponse->getProductIds());
- }
-
/**
* @param FacetFieldTransformationRegistry $facetFieldTransformationRegistry
* @return SearchEngine | Issue #<I>: Remove obsolete test | lizards-and-pumpkins_catalog | train |
973dad46ccf80af3abbc3000bd0443e7abf2d459 | diff --git a/cmd/burrow/commands/configure.go b/cmd/burrow/commands/configure.go
index <HASH>..<HASH> 100644
--- a/cmd/burrow/commands/configure.go
+++ b/cmd/burrow/commands/configure.go
@@ -283,7 +283,7 @@ func Configure(output Output) func(cmd *cli.Cmd) {
if err != nil {
output.Fatalf("Could not form GenesisDoc JSON: %v", err)
}
- err = ioutil.WriteFile(*separateGenesisDoc, genesisDocJSON, 0700)
+ err = ioutil.WriteFile(*separateGenesisDoc, genesisDocJSON, 0644)
if err != nil {
output.Fatalf("Could not write GenesisDoc JSON: %v", err)
} | genesis.json should not be executable | hyperledger_burrow | train |
a780883b3c74ad29d992b2a1e5e61b3a3eb5edb4 | diff --git a/CHANGES.md b/CHANGES.md
index <HASH>..<HASH> 100644
--- a/CHANGES.md
+++ b/CHANGES.md
@@ -5,7 +5,7 @@
pixel grid together. Now also requires datasets to always be linked
in order to be shown in the same viewer. [#335, #337]
-- Fix compatibility with the latest developer version of glue. [#339]
+- Fix compatibility with the latest developer version of glue. [#339, #342]
- Fix bug with hidden layers in the 3D scatter viewer becoming visible after
saving and loading session file. [#340]
diff --git a/glue_vispy_viewers/common/selection_tools.py b/glue_vispy_viewers/common/selection_tools.py
index <HASH>..<HASH> 100644
--- a/glue_vispy_viewers/common/selection_tools.py
+++ b/glue_vispy_viewers/common/selection_tools.py
@@ -8,7 +8,12 @@ import numpy as np
from glue.core import Data
from glue.config import viewer_tool
-from glue.viewers.common.qt.tool import CheckableTool
+
+try:
+ from glue.viewers.common.tool import CheckableTool
+except ImportError: # glue-core <0.15
+ from glue.viewers.common.qt.tool import CheckableTool
+
from glue.core.command import ApplySubsetState
from glue.core.roi import RectangularROI, CircularROI, PolygonalROI, Projected3dROI
diff --git a/glue_vispy_viewers/common/tools.py b/glue_vispy_viewers/common/tools.py
index <HASH>..<HASH> 100644
--- a/glue_vispy_viewers/common/tools.py
+++ b/glue_vispy_viewers/common/tools.py
@@ -3,7 +3,12 @@ from __future__ import absolute_import, division, print_function
import os
from qtpy import QtGui, compat
-from glue.viewers.common.qt.tool import Tool, CheckableTool
+
+try:
+ from glue.viewers.common.tool import Tool, CheckableTool
+except ImportError: # glue-core <0.15
+ from glue.viewers.common.qt.tool import Tool, CheckableTool
+
from glue.config import viewer_tool
from ..extern.vispy import app, io | Update Tool/CheckableTool import for glue-core <I> | glue-viz_glue-vispy-viewers | train |
1960384505a491e83554718e6735ff8825ec7afd | diff --git a/openfisca_core/taxbenefitsystems/tax_benefit_system.py b/openfisca_core/taxbenefitsystems/tax_benefit_system.py
index <HASH>..<HASH> 100644
--- a/openfisca_core/taxbenefitsystems/tax_benefit_system.py
+++ b/openfisca_core/taxbenefitsystems/tax_benefit_system.py
@@ -149,7 +149,7 @@ class TaxBenefitSystem:
variable: The variable to add. Must be a subclass of Variable.
Raises:
- VariableNameConflictError: if a variable with the same name have previously been added to the tax and benefit system.
+ openfisca_core.errors.VariableNameConflictError: if a variable with the same name have previously been added to the tax and benefit system.
"""
return self.load_variable(variable, update = False) | Use fully qualified name in docstring | openfisca_openfisca-core | train |
bdb68363da89c85a06305f913ceff9428cf47e8d | diff --git a/src/test/java/io/vlingo/xoom/actors/plugin/mailbox/sharedringbuffer/RingBufferMailboxActorTest.java b/src/test/java/io/vlingo/xoom/actors/plugin/mailbox/sharedringbuffer/RingBufferMailboxActorTest.java
index <HASH>..<HASH> 100644
--- a/src/test/java/io/vlingo/xoom/actors/plugin/mailbox/sharedringbuffer/RingBufferMailboxActorTest.java
+++ b/src/test/java/io/vlingo/xoom/actors/plugin/mailbox/sharedringbuffer/RingBufferMailboxActorTest.java
@@ -31,10 +31,10 @@ public class RingBufferMailboxActorTest extends ActorsTest {
private static final int ThroughputWarmUpCount = 4194304;
@Test
- public void testBasicDispatch() throws Exception {
+ public void testBasicDispatch() {
init(MailboxSize);
- final TestResults testResults = new TestResults(MaxCount);
+ final TestResults testResults = new TestResults(MaxCount + 1);
final CountTaker countTaker =
world.actorFor( | Update the number of expected operations in the test to reflect the actual operations
Otherwise the test tends to fail occasionally as it does not wait long enough for all the operations to complete. The test calls setMaximum() on testResults which increases the number of operations by one. | vlingo_vlingo-actors | train |
596acb7f79c33ba0c338d62ef9942a6f9cee81f1 | diff --git a/isso/utils/http.py b/isso/utils/http.py
index <HASH>..<HASH> 100644
--- a/isso/utils/http.py
+++ b/isso/utils/http.py
@@ -7,6 +7,7 @@ try:
except ImportError:
import http.client as httplib
+from isso import dist
from isso.wsgi import urlsplit
@@ -21,6 +22,10 @@ class curl(object):
return resp.status
"""
+ headers = {
+ "User-Agent": "Isso/{0} (+http://posativ.org/isso)".format(dist.version)
+ }
+
def __init__(self, method, host, path, timeout=3):
self.method = method
self.host = host
@@ -35,7 +40,7 @@ class curl(object):
self.con = http(host, port, timeout=self.timeout)
try:
- self.con.request(self.method, self.path)
+ self.con.request(self.method, self.path, headers=self.headers)
except (httplib.HTTPException, socket.error):
return None | request with User-Agent 'Isso/...', #<I> | posativ_isso | train |
5617f2a27f871acefafdda99de70304fe5301020 | diff --git a/consumergroup/consumer_group.go b/consumergroup/consumer_group.go
index <HASH>..<HASH> 100644
--- a/consumergroup/consumer_group.go
+++ b/consumergroup/consumer_group.go
@@ -194,7 +194,12 @@ func NewConsumerGroup(client *sarama.Client, zoo *ZK, name string, topic string,
// Returns an error if any, but may also return a NoCheckout error to indicate
// that no partition was available. You should add an artificial delay keep your CPU cool.
func (cg *ConsumerGroup) Checkout(callback func(*PartitionConsumer) error) error {
- cg.checkout <- true
+ select {
+ case <-cg.stopper:
+ return NoCheckout
+ default:
+ cg.checkout <- true
+ }
var claimed *PartitionConsumer
select { | Fix deadlock when trying to read from checkout channel | wvanbergen_kafka | train |
097240f60215b866d24aebd02cc4159bdc6e7451 | diff --git a/activerecord/lib/active_record/nested_attributes.rb b/activerecord/lib/active_record/nested_attributes.rb
index <HASH>..<HASH> 100644
--- a/activerecord/lib/active_record/nested_attributes.rb
+++ b/activerecord/lib/active_record/nested_attributes.rb
@@ -321,7 +321,7 @@ module ActiveRecord
if check_existing_record && (record = send(association_name)) &&
(options[:update_only] || record.id.to_s == attributes['id'].to_s)
- assign_to_or_mark_for_destruction(record, attributes, options[:allow_destroy])
+ assign_to_or_mark_for_destruction(record, attributes, options[:allow_destroy]) unless call_reject_if(association_name, attributes)
elsif attributes['id']
existing_record = self.class.reflect_on_association(association_name).klass.find(attributes['id'])
@@ -399,11 +399,11 @@ module ActiveRecord
elsif existing_records.count == 0 #Existing record but not yet associated
existing_record = self.class.reflect_on_association(association_name).klass.find(attributes['id'])
- association.send(:add_record_to_target_with_callbacks, existing_record) unless association.loaded?
+ association.send(:add_record_to_target_with_callbacks, existing_record) if !association.loaded? && !call_reject_if(association_name, attributes)
assign_to_or_mark_for_destruction(existing_record, attributes, options[:allow_destroy])
elsif existing_record = existing_records.detect { |record| record.id.to_s == attributes['id'].to_s }
- association.send(:add_record_to_target_with_callbacks, existing_record) unless association.loaded?
+ association.send(:add_record_to_target_with_callbacks, existing_record) if !association.loaded? && !call_reject_if(association_name, attributes)
assign_to_or_mark_for_destruction(existing_record, attributes, options[:allow_destroy])
end
diff --git a/activerecord/test/cases/nested_attributes_test.rb b/activerecord/test/cases/nested_attributes_test.rb
index <HASH>..<HASH> 100644
--- a/activerecord/test/cases/nested_attributes_test.rb
+++ b/activerecord/test/cases/nested_attributes_test.rb
@@ -114,6 +114,15 @@ class TestNestedAttributesInGeneral < ActiveRecord::TestCase
pirate.ship_attributes = { :name => 'Hello Pearl' }
assert_difference('Ship.count') { pirate.save! }
end
+
+ def test_reject_if_with_a_proc_which_returns_true_always
+ Pirate.accepts_nested_attributes_for :ship, :reject_if => proc {|attributes| true }
+ pirate = Pirate.new(:catchphrase => "Stop wastin' me time")
+ ship = pirate.create_ship(:name => 's1')
+ pirate.update_attributes({:ship_attributes => { :name => 's2', :id => ship.id } })
+ assert_equal 's1', ship.reload.name
+ end
+
end
class TestNestedAttributesOnAHasOneAssociation < ActiveRecord::TestCase | reject_id option should be respected while using nested_attributes
[#<I> state:resolved] | rails_rails | train |
c03889941365c07f55be6f440a0b806b0e9fb4a5 | diff --git a/lib/util/Components.js b/lib/util/Components.js
index <HASH>..<HASH> 100644
--- a/lib/util/Components.js
+++ b/lib/util/Components.js
@@ -237,7 +237,7 @@ function componentRule(rule, context) {
* @returns {Boolean} True if React.createElement called
*/
isReactCreateElement: function(node) {
- return (
+ var calledOnReact = (
node &&
node.callee &&
node.callee.object &&
@@ -245,6 +245,17 @@ function componentRule(rule, context) {
node.callee.property &&
node.callee.property.name === 'createElement'
);
+
+ var calledDirectly = (
+ node &&
+ node.callee &&
+ node.callee.name === 'createElement'
+ );
+
+ if (this.hasDestructuredReactCreateElement()) {
+ return calledDirectly || calledOnReact;
+ }
+ return calledOnReact;
},
/**
@@ -290,10 +301,7 @@ function componentRule(rule, context) {
node[property] &&
node[property].type === 'JSXElement'
;
- var returnsReactCreateElement =
- this.hasDestructuredReactCreateElement() ||
- this.isReactCreateElement(node[property])
- ;
+ var returnsReactCreateElement = this.isReactCreateElement(node[property]);
return Boolean(
returnsConditionalJSX ||
diff --git a/tests/lib/rules/no-multi-comp.js b/tests/lib/rules/no-multi-comp.js
index <HASH>..<HASH> 100644
--- a/tests/lib/rules/no-multi-comp.js
+++ b/tests/lib/rules/no-multi-comp.js
@@ -10,6 +10,7 @@
var rule = require('../../../lib/rules/no-multi-comp');
var RuleTester = require('eslint').RuleTester;
+var assign = require('object.assign');
var parserOptions = {
ecmaVersion: 6,
@@ -89,6 +90,21 @@ ruleTester.run('no-multi-comp', rule, {
options: [{
ignoreStateless: true
}]
+ }, {
+ // multiple non-components
+ code: [
+ 'import React, { createElement } from "react"',
+ 'const helperFoo = () => {',
+ ' return true;',
+ '};',
+ 'function helperBar() {',
+ ' return false;',
+ '};',
+ 'function RealComponent() {',
+ ' return createElement("img");',
+ '};'
+ ].join('\n'),
+ parserOptions: assign({sourceType: 'module'}, parserOptions)
}],
invalid: [{ | Bug fix for false positives with no-multi-comp
Previously, when createElement was destructured from React, it would
cause any function defined in the file to be flagged as a React
Component. This change makes it such that the function must call
createElement to be considered a component.
Fixes #<I>
---
Review: Use object.assign with sourceType: module in the added test over
babel-eslint. | ytanruengsri_eslint-plugin-react-ssr | train |
1ebf4c97717e9511a16d84afbd88fe261a69b9e2 | diff --git a/ontrack-boot/src/main/java/net/nemerosa/ontrack/boot/ui/StructureAPI.java b/ontrack-boot/src/main/java/net/nemerosa/ontrack/boot/ui/StructureAPI.java
index <HASH>..<HASH> 100644
--- a/ontrack-boot/src/main/java/net/nemerosa/ontrack/boot/ui/StructureAPI.java
+++ b/ontrack-boot/src/main/java/net/nemerosa/ontrack/boot/ui/StructureAPI.java
@@ -28,6 +28,8 @@ public interface StructureAPI {
ResourceCollection<Branch> getBranchListForProject(ID projectId);
+ Form newBranchForm();
+
Resource<Branch> newBranch(ID projectId, NameDescription nameDescription);
Resource<Branch> getBranch(ID branchId);
diff --git a/ontrack-boot/src/main/java/net/nemerosa/ontrack/boot/ui/StructureAPIController.java b/ontrack-boot/src/main/java/net/nemerosa/ontrack/boot/ui/StructureAPIController.java
index <HASH>..<HASH> 100644
--- a/ontrack-boot/src/main/java/net/nemerosa/ontrack/boot/ui/StructureAPIController.java
+++ b/ontrack-boot/src/main/java/net/nemerosa/ontrack/boot/ui/StructureAPIController.java
@@ -94,6 +94,12 @@ public class StructureAPIController extends AbstractResourceController implement
@Override
@RequestMapping(value = "projects/{projectId}/branches/create", method = RequestMethod.GET)
+ public Form newBranchForm() {
+ return Branch.form();
+ }
+
+ @Override
+ @RequestMapping(value = "projects/{projectId}/branches/create", method = RequestMethod.POST)
public Resource<Branch> newBranch(@PathVariable ID projectId, @RequestBody NameDescription nameDescription) {
// Gets the project
Project project = structureRepository.getProject(projectId);
@@ -134,11 +140,12 @@ public class StructureAPIController extends AbstractResourceController implement
}
private Resource<Branch> toBranchResourceWithActions(Branch branch) {
- return toBranchResource(branch);
- // TODO Update link
- // TODO Delete link
- // TODO View link
- // TODO Builds link
+ return toBranchResource(branch)
+ // TODO Update link (with authorisation)
+ // TODO Delete link
+ // TODO View link
+ // TODO Builds link
+ ;
}
private Resource<Branch> toBranchResource(Branch branch) {
diff --git a/ontrack-model/src/main/java/net/nemerosa/ontrack/model/structure/Branch.java b/ontrack-model/src/main/java/net/nemerosa/ontrack/model/structure/Branch.java
index <HASH>..<HASH> 100644
--- a/ontrack-model/src/main/java/net/nemerosa/ontrack/model/structure/Branch.java
+++ b/ontrack-model/src/main/java/net/nemerosa/ontrack/model/structure/Branch.java
@@ -2,6 +2,7 @@ package net.nemerosa.ontrack.model.structure;
import com.fasterxml.jackson.annotation.JsonView;
import lombok.Data;
+import net.nemerosa.ontrack.model.form.Form;
@Data
public class Branch implements Entity {
@@ -25,4 +26,7 @@ public class Branch implements Entity {
return new Branch(id, name, description, project);
}
+ public static Form form() {
+ return Form.nameAndDescription();
+ }
} | Branches: creation of a branch (form) | nemerosa_ontrack | train |
9d2462f6295c3ada78381ccb27d70e27f0e07877 | diff --git a/mungers/e2e/e2e.go b/mungers/e2e/e2e.go
index <HASH>..<HASH> 100644
--- a/mungers/e2e/e2e.go
+++ b/mungers/e2e/e2e.go
@@ -18,10 +18,7 @@ package e2e
import (
"bufio"
- "encoding/json"
"fmt"
- "io/ioutil"
- "net/http"
"strings"
"sync"
@@ -98,41 +95,9 @@ func (e *E2ETester) Stable() bool {
}
const (
- successString = "SUCCESS"
expectedXMLHeader = "<?xml version=\"1.0\" encoding=\"UTF-8\"?>"
)
-type finishedFile struct {
- Result string `json:"result"`
- Timestamp uint64 `json:"timestamp"`
-}
-
-func checkFinishedStatus(job string, lastBuildNumber int) bool {
- response, err := utils.GetFileFromJenkinsGoogleBucket(job, lastBuildNumber, "finished.json")
- if err != nil {
- glog.Errorf("Error while getting data for %v/%v/%v: %v", job, lastBuildNumber, "finished.json", err)
- return false
- }
-
- defer response.Body.Close()
- if response.StatusCode != http.StatusOK {
- glog.Infof("Got a non-success response %v while reading data for %v/%v/%v", response.StatusCode, job, lastBuildNumber, "finished.json")
- return false
- }
- result := finishedFile{}
- body, err := ioutil.ReadAll(response.Body)
- if err != nil {
- glog.Errorf("Failed to read the response for %v/%v/%v: %v", job, lastBuildNumber, "finished.json", err)
- return false
- }
- err = json.Unmarshal(body, &result)
- if err != nil {
- glog.Errorf("Failed to unmarshal %v: %v", string(body), err)
- return false
- }
- return result.Result == successString
-}
-
// GCSBasedStable is a version of Stable function that depends on files stored in GCS instead of Jenkis
func (e *E2ETester) GCSBasedStable() bool {
for _, job := range e.JobNames {
@@ -142,7 +107,7 @@ func (e *E2ETester) GCSBasedStable() bool {
glog.Errorf("Error while getting data for %v: %v", job, err)
continue
}
- if !checkFinishedStatus(job, lastBuildNumber) {
+ if stable, err := utils.CheckFinishedStatus(job, lastBuildNumber); !stable || err != nil {
return false
}
}
@@ -162,7 +127,7 @@ func (e *E2ETester) GCSWeakStable() bool {
glog.Errorf("Error while getting data for %v: %v", job, err)
continue
}
- if checkFinishedStatus(job, lastBuildNumber) {
+ if stable, err := utils.CheckFinishedStatus(job, lastBuildNumber); stable && err == nil {
continue
}
@@ -209,7 +174,10 @@ func (e *E2ETester) GCSWeakStable() bool {
// If we're here it means that we weren't able to find a test that failed, which means that the reason of build failure is comming from the infrastructure
// Check results of previous two builds.
- if !checkFinishedStatus(job, lastBuildNumber-1) && !checkFinishedStatus(job, lastBuildNumber-2) {
+ if stable, err := utils.CheckFinishedStatus(job, lastBuildNumber-1); !stable || err != nil {
+ return false
+ }
+ if stable, err := utils.CheckFinishedStatus(job, lastBuildNumber-2); !stable || err != nil {
return false
}
} | Move CheckFinishedStatus to test-utils | kubernetes_test-infra | train |
c4ad7100a94c864d1bc5309cd69aec7364b782ed | diff --git a/revapi-java/src/main/java/org/revapi/java/checks/common/SerializationChecker.java b/revapi-java/src/main/java/org/revapi/java/checks/common/SerializationChecker.java
index <HASH>..<HASH> 100644
--- a/revapi-java/src/main/java/org/revapi/java/checks/common/SerializationChecker.java
+++ b/revapi-java/src/main/java/org/revapi/java/checks/common/SerializationChecker.java
@@ -315,6 +315,10 @@ public class SerializationChecker extends CheckBase {
for (TypeMirror st : Util.getAllSuperClasses(types, type.asType())) {
Element ste = types.asElement(st);
+ if (ste == null) {
+ // a missing class
+ continue;
+ }
ElementFilter.fieldsIn(ste.getEnclosedElements()).stream().filter(serializableFields).sorted(bySimpleName)
.map(e -> types.asMemberOf((DeclaredType) st, e)).forEach(fields::add);
}
@@ -585,7 +589,7 @@ public class SerializationChecker extends CheckBase {
}
/**
- * Adapted from {@link java.io.ObjectStreamClass.MemberSignature}
+ * Adapted from {@code java.io.ObjectStreamClass.MemberSignature}
*
* <p>
* Class for computing and caching field/constructor/method signatures during serialVersionUID calculation. | Now that missing types are handled more strictly, we actually might
encounter them when looking up superclasses. | revapi_revapi | train |
6e19c688af3728e2c095ded24df89c8f11259ebd | diff --git a/go/engine/bootstrap.go b/go/engine/bootstrap.go
index <HASH>..<HASH> 100644
--- a/go/engine/bootstrap.go
+++ b/go/engine/bootstrap.go
@@ -75,6 +75,11 @@ func (e *Bootstrap) Run(ctx *Context) error {
return gerr
}
+ if !e.status.LoggedIn {
+ e.G().Log.Debug("not logged in, not running syncer")
+ return nil
+ }
+
// get user summaries
ts := libkb.NewTracker2Syncer(e.G(), e.status.Uid, true)
if e.G().ConnectivityMonitor.IsConnected(context.Background()) == libkb.ConnectivityMonitorYes { | Fix bootstrap when not logged in | keybase_client | train |
91dd1a523bc1ba5fbf46fb6d694176fc277086c3 | diff --git a/ht/hx.py b/ht/hx.py
index <HASH>..<HASH> 100644
--- a/ht/hx.py
+++ b/ht/hx.py
@@ -519,10 +519,11 @@ possible for this configuration; the maximum effectiveness possible is %s.' % (m
NTU = -(1. + Cr*Cr)**-0.5*log((E - 1.)/(E + 1.))
return shells*NTU
elif subtype == 'crossflow':
+ # Can't use a bisect solver here because at high NTU there's a derivative of 0
+ # due to the integral term not changing when it's very near one
guess = NTU_from_effectiveness(effectiveness, Cr, 'crossflow approximate')
def to_solve(NTU, Cr, effectiveness):
return effectiveness_from_NTU(NTU, Cr, subtype='crossflow') - effectiveness
- #return ridder(to_solve, a=1E-3, b=1E3, args=(Cr, effectiveness))
return newton(to_solve, guess, args=(Cr, effectiveness))
elif subtype == 'crossflow approximate':
# This will fail if NTU is more than 10,000 or less than 1E-7, but
@@ -968,12 +969,21 @@ def temperature_effectiveness_basic(R1, NTU1, subtype='crossflow'):
P_1 = \frac{1 - \exp[-NTU_1(1-R_1)]}{1 - R_1 \exp[-NTU_1(1-R_1)]}
For cross-flow (single-pass) heat exchangers with both fluids unmixed
- (this configuration is symmetric):
+ (this configuration is symmetric), there are two solutions available;
+ a frequently cited approximation and an exact solution which uses
+ a numerical integration developed in [4]_. The approximate solution is:
.. math::
P_1 \approx 1 - \exp\left[\frac{NTU_1^{0.22}}{R_1}
(\exp(-R_1 NTU_1^{0.78})-1)\right]
+ The exact solution for crossflow (single pass, fluids unmixed) is:
+
+ .. math::
+ \epsilon = \frac{1}{R_1} - \frac{\exp(-R_1 \cdot NTU_1)}{2(R_1 NTU_1)^2}
+ \int_0^{2 NTU_1\sqrt{R_1}} \left(1 + NTU_1 - \frac{v^2}{4R_1 NTU_1}
+ \right)\exp\left(-\frac{v^2}{4R_1 NTU_1}\right)v I_0(v) dv
+
For cross-flow (single-pass) heat exchangers with fluid 1 mixed, fluid 2
unmixed:
@@ -1010,8 +1020,8 @@ def temperature_effectiveness_basic(R1, NTU1, subtype='crossflow'):
method, calculated with respect to stream 1 [-]
subtype : float
The type of heat exchanger; one of 'counterflow', 'parallel',
- 'crossflow', 'crossflow, mixed 1', 'crossflow, mixed 2',
- 'crossflow, mixed 1&2'.
+ 'crossflow', 'crossflow approximate', 'crossflow, mixed 1',
+ 'crossflow, mixed 2', 'crossflow, mixed 1&2'.
Returns
-------
@@ -1038,16 +1048,25 @@ def temperature_effectiveness_basic(R1, NTU1, subtype='crossflow'):
CRC Press, 2013.
.. [3] Rohsenow, Warren and James Hartnett and Young Cho. Handbook of Heat
Transfer, 3E. New York: McGraw-Hill, 1998.
+ .. [4] Triboix, Alain. "Exact and Approximate Formulas for Cross Flow Heat
+ Exchangers with Unmixed Fluids." International Communications in Heat
+ and Mass Transfer 36, no. 2 (February 1, 2009): 121-24.
+ doi:10.1016/j.icheatmasstransfer.2008.10.012.
'''
if subtype == 'counterflow':
P1 = (1 - exp(-NTU1*(1 - R1)))/(1 - R1*exp(-NTU1*(1-R1)))
elif subtype == 'parallel':
P1 = (1 - exp(-NTU1*(1 + R1)))/(1 + R1)
- elif subtype == 'crossflow':
+ elif subtype == 'crossflow approximate':
# This isn't technically accurate, an infinite sum is required
# It has been computed from two different sources
# but is found not to be within the 1% claimed of this equation
P1 = 1 - exp(NTU1**0.22/R1*(exp(-R1*NTU1**0.78) - 1.))
+ elif subtype == 'crossflow':
+ def to_int(v, NTU1, R1):
+ return (1. + NTU1 - v*v/(4.*R1*NTU1))*exp(-v*v/(4.*R1*NTU1))*v*iv(0, v)
+ int_term = quad(to_int, 0, 2.*NTU1*R1**0.5, args=(NTU1, R1))[0]
+ P1 = 1./R1 - exp(-R1*NTU1)/(2.*(R1*NTU1)**2)*int_term
elif subtype == 'crossflow, mixed 1':
# Not symmetric
K = 1 - exp(-R1*NTU1)
diff --git a/tests/test_hx.py b/tests/test_hx.py
index <HASH>..<HASH> 100644
--- a/tests/test_hx.py
+++ b/tests/test_hx.py
@@ -435,8 +435,10 @@ def test_temperature_effectiveness_basic():
assert_allclose(P1, 0.173382601503)
P1 = temperature_effectiveness_basic(R1=3.5107078039927404, NTU1=0.29786672449248663, subtype='parallel')
assert_allclose(P1, 0.163852912049)
- P1 = temperature_effectiveness_basic(R1=3.5107078039927404, NTU1=0.29786672449248663, subtype='crossflow')
+ P1 = temperature_effectiveness_basic(R1=3.5107078039927404, NTU1=0.29786672449248663, subtype='crossflow approximate')
assert_allclose(P1, 0.149974594007)
+ P1 = temperature_effectiveness_basic(R1=3.5107078039927404, NTU1=0.29786672449248663, subtype='crossflow')
+ assert_allclose(P1, 0.1698702121873175)
P1 = temperature_effectiveness_basic(R1=3.5107078039927404, NTU1=0.29786672449248663, subtype='crossflow, mixed 1')
assert_allclose(P1, 0.168678230894)
P1 = temperature_effectiveness_basic(R1=3.5107078039927404, NTU1=0.29786672449248663, subtype='crossflow, mixed 2') | Add the new exact solution for crossflow (fluids unmixed, single pass) to temperature_effectiveness_basic | CalebBell_ht | train |
b009763ea5f270f618f2c869cbfadd2c952bde41 | diff --git a/salt/grains/disks.py b/salt/grains/disks.py
index <HASH>..<HASH> 100644
--- a/salt/grains/disks.py
+++ b/salt/grains/disks.py
@@ -51,9 +51,8 @@ class _camconsts:
SECTOR_SIZE = 'sector size'
MEDIA_RPM = 'media RPM'
-
-_identify_attribs = {key: _camconsts.__dict__[key] for key in
- _camconsts.__dict__ if not key.startswith('__')}.values()
+_identify_attribs = [_camconsts.__dict__[key] for key in
+ _camconsts.__dict__ if not key.startswith('__')]
def _freebsd_disks(): | Simplify attribs list with list comprehension | saltstack_salt | train |
a56b0e813b129dbdaa9d293bf328d2d08aca7228 | diff --git a/src/index.js b/src/index.js
index <HASH>..<HASH> 100644
--- a/src/index.js
+++ b/src/index.js
@@ -283,6 +283,8 @@ class FlavorUtils {
current.__id = row._id;
current.__rev = row._rev;
current.__version = row.version;
+ current.__keywords = row.keywords;
+ current.__meta = row.meta;
return;
} | add keywords and meta when getting view tree | cheminfo-js_flavor-utils | train |
897ba3e181cf048e9a33a0b5008025ff7336e685 | diff --git a/guice/hibernate/src/main/java/com/peterphi/std/guice/hibernate/module/HibernateSessionFactoryProvider.java b/guice/hibernate/src/main/java/com/peterphi/std/guice/hibernate/module/HibernateSessionFactoryProvider.java
index <HASH>..<HASH> 100755
--- a/guice/hibernate/src/main/java/com/peterphi/std/guice/hibernate/module/HibernateSessionFactoryProvider.java
+++ b/guice/hibernate/src/main/java/com/peterphi/std/guice/hibernate/module/HibernateSessionFactoryProvider.java
@@ -37,6 +37,9 @@ class HibernateSessionFactoryProvider implements Provider<SessionFactory>, Stopp
Provider<ServiceRegistry> serviceRegistryProvider,
Configuration config)
{
+ if (config == null)
+ throw new IllegalArgumentException("Must provide non-null hibernate Configuration!");
+
this.serviceRegistryProvider = serviceRegistryProvider;
this.config = config;
@@ -50,7 +53,15 @@ class HibernateSessionFactoryProvider implements Provider<SessionFactory>, Stopp
{
final ServiceRegistry serviceRegistry = serviceRegistryProvider.get();
- this.sessionFactory = config.buildSessionFactory(serviceRegistry);
+ try
+ {
+ this.sessionFactory = config.buildSessionFactory(serviceRegistry);
+ }
+ catch (Throwable t) {
+ log.warn("Error setting up hibernate session factory",t);
+
+ throw t;
+ }
log.trace("Hibernate Setup Complete.");
} | Improve error logging when hibernate SessionFactory construction fails | petergeneric_stdlib | train |
ec8faf01112660cbecc50b069afeda7f3b20b5c9 | diff --git a/client/state/comments/reducer.js b/client/state/comments/reducer.js
index <HASH>..<HASH> 100644
--- a/client/state/comments/reducer.js
+++ b/client/state/comments/reducer.js
@@ -17,6 +17,7 @@ import {
values,
omit,
startsWith,
+ isInteger,
} from 'lodash';
/**
@@ -367,7 +368,7 @@ export const activeReplies = createReducer(
function updateCount( counts, rawStatus, value = 1 ) {
const status = rawStatus === 'unapproved' ? 'pending' : rawStatus;
- if ( ! counts || ! counts[ status ] ) {
+ if ( ! counts || ! isInteger( counts[ status ] ) ) {
return undefined;
}
const newCounts = {
diff --git a/client/state/comments/test/reducer.js b/client/state/comments/test/reducer.js
index <HASH>..<HASH> 100644
--- a/client/state/comments/test/reducer.js
+++ b/client/state/comments/test/reducer.js
@@ -637,8 +637,8 @@ describe( 'reducer', () => {
type: COMMENTS_CHANGE_STATUS,
siteId: 2916284,
postId: 234,
- status: 'unapproved',
- meta: { comment: { previousStatus: 'approved' } },
+ status: 'trash',
+ meta: { comment: { previousStatus: 'unapproved' } },
};
const state = deepFreeze( {
2916284: {
@@ -666,22 +666,22 @@ describe( 'reducer', () => {
expect( nextState ).toEqual( {
2916284: {
site: {
- all: 11,
- approved: 4,
- pending: 7,
+ all: 10,
+ approved: 5,
+ pending: 5,
postTrashed: 0,
spam: 1,
- totalComments: 12,
- trash: 0,
+ totalComments: 11,
+ trash: 1,
},
234: {
- all: 5,
- approved: 1,
- pending: 4,
+ all: 4,
+ approved: 2,
+ pending: 2,
postTrashed: 0,
spam: 1,
- totalComments: 6,
- trash: 0,
+ totalComments: 5,
+ trash: 1,
},
},
} ); | Comments: update check so we can add numbers to 0 | Automattic_wp-calypso | train |
08ae06fc31efef0791a11cef23bebe530c1fdbac | diff --git a/plugins/CoreHome/javascripts/calendar.js b/plugins/CoreHome/javascripts/calendar.js
index <HASH>..<HASH> 100644
--- a/plugins/CoreHome/javascripts/calendar.js
+++ b/plugins/CoreHome/javascripts/calendar.js
@@ -187,7 +187,7 @@
result.stepMonths = selectedPeriod == 'year' ? 12 : 1;
result.onSelect = function () { updateDate.apply(this, arguments); };
return result;
- };
+ }
$(document).ready(function () {
diff --git a/plugins/CoreHome/javascripts/datatable.js b/plugins/CoreHome/javascripts/datatable.js
index <HASH>..<HASH> 100644
--- a/plugins/CoreHome/javascripts/datatable.js
+++ b/plugins/CoreHome/javascripts/datatable.js
@@ -1420,7 +1420,6 @@ dataTable.prototype =
for (var name in paramOverride) {
self.param[name] = paramOverride[name];
}
- ;
self.reloadAjaxDataTable(true);
});
diff --git a/plugins/Dashboard/javascripts/widgetMenu.js b/plugins/Dashboard/javascripts/widgetMenu.js
index <HASH>..<HASH> 100644
--- a/plugins/Dashboard/javascripts/widgetMenu.js
+++ b/plugins/Dashboard/javascripts/widgetMenu.js
@@ -191,7 +191,7 @@ widgetsHelper.getEmptyWidgetHtml = function (uniqueId, widgetName) {
}
return $('.' + settings.categorylistClass, widgetPreview);
- };
+ }
/**
* Returns the div to show widget list in
@@ -219,7 +219,7 @@ widgetsHelper.getEmptyWidgetHtml = function (uniqueId, widgetName) {
}
return $('.' + settings.widgetlistClass, widgetPreview);
- };
+ }
/**
* Display the given widgets in a widget list
@@ -290,7 +290,7 @@ widgetsHelper.getEmptyWidgetHtml = function (uniqueId, widgetName) {
}
return $('.' + settings.widgetpreviewClass, widgetPreview);
- };
+ }
/**
* Show widget with the given uniqueId in preview
@@ -335,7 +335,7 @@ widgetsHelper.getEmptyWidgetHtml = function (uniqueId, widgetName) {
}
widgetAjaxRequest = widgetsHelper.loadWidgetAjax(widgetUniqueId, widgetParameters, onWidgetLoadedCallback);
- };
+ }
/**
* Reset function
@@ -346,7 +346,7 @@ widgetsHelper.getEmptyWidgetHtml = function (uniqueId, widgetName) {
$('.' + settings.categorylistClass + ' li', widgetPreview).removeClass(settings.choosenClass);
createWidgetList();
createPreviewElement();
- };
+ }
/**
* Constructor
diff --git a/plugins/SegmentEditor/javascripts/Segmentation.js b/plugins/SegmentEditor/javascripts/Segmentation.js
index <HASH>..<HASH> 100644
--- a/plugins/SegmentEditor/javascripts/Segmentation.js
+++ b/plugins/SegmentEditor/javascripts/Segmentation.js
@@ -226,7 +226,7 @@ Segmentation = (function($) {
$(html).find(".add_new_segment").html(self.translations['SegmentEditor_AddNewSegment']);
}
else {
- $(html).find(".add_new_segment").hide();;
+ $(html).find(".add_new_segment").hide();
}
}
else | removed some unnecassary semicolons | matomo-org_matomo | train |
e434ff33956b8b485308e8d103aa2551b8fd720e | diff --git a/lib/sensu/server.rb b/lib/sensu/server.rb
index <HASH>..<HASH> 100644
--- a/lib/sensu/server.rb
+++ b/lib/sensu/server.rb
@@ -300,8 +300,11 @@ module Sensu
case handler[:type]
when 'pipe'
IO.async_popen(handler[:command], event_data, handler[:timeout]) do |output, status|
- output.split(/\n+/).each do |line|
- @logger.info(line)
+ output.each_line do |line|
+ @logger.info('handler output', {
+ :handler => handler,
+ :output => line
+ })
end
@handlers_in_progress_count -= 1
end
@@ -343,8 +346,11 @@ module Sensu
@handlers_in_progress_count -= 1
when 'extension'
handler.safe_run(event_data) do |output, status|
- output.split(/\n+/).each do |line|
- @logger.info(line)
+ output.each_line do |line|
+ @logger.info('handler extension output', {
+ :extension => handler.definition,
+ :output => line
+ })
end
@handlers_in_progress_count -= 1
end | [timeout] added more context to handler output log lines | sensu_sensu | train |
a5f59f10cde99b16be0aef6cb38c375f06561875 | diff --git a/admin/jqadm/src/Admin/JQAdm/Product/Subscription/Standard.php b/admin/jqadm/src/Admin/JQAdm/Product/Subscription/Standard.php
index <HASH>..<HASH> 100644
--- a/admin/jqadm/src/Admin/JQAdm/Product/Subscription/Standard.php
+++ b/admin/jqadm/src/Admin/JQAdm/Product/Subscription/Standard.php
@@ -331,11 +331,11 @@ class Standard
preg_match( $regex, $list['attribute.code'], $matches );
- $list['Y'] = (int) ( $matches[2] ?? 0 );
- $list['M'] = (int) ( $matches[4] ?? 0 );
- $list['W'] = (int) ( $matches[6] ?? 0 );
- $list['D'] = (int) ( $matches[8] ?? 0 );
- $list['H'] = (int) ( $matches[10] ?? 0 );
+ empty( $matches[2] ) ?: $list['Y'] = (int) $matches[2];
+ empty( $matches[4] ) ?: $list['M'] = (int) $matches[4];
+ empty( $matches[6] ) ?: $list['W'] = (int) $matches[6];
+ empty( $matches[8] ) ?: $list['D'] = (int) $matches[8];
+ empty( $matches[10] ) ?: $list['H'] = (int) $matches[10];
$data[] = $list;
} | Compatibility to PHP7 durations | aimeos_ai-admin-jqadm | train |
55854c7de35eabbb52491358ca164cd17a81ca12 | diff --git a/bin/simulate_non_silent_ratio.py b/bin/simulate_non_silent_ratio.py
index <HASH>..<HASH> 100644
--- a/bin/simulate_non_silent_ratio.py
+++ b/bin/simulate_non_silent_ratio.py
@@ -18,7 +18,6 @@ import pandas as pd
import pysam
from multiprocessing import Pool
import argparse
-import datetime
import logging
import copy
diff --git a/bin/simulate_summary.py b/bin/simulate_summary.py
index <HASH>..<HASH> 100644
--- a/bin/simulate_summary.py
+++ b/bin/simulate_summary.py
@@ -20,7 +20,6 @@ import pysam
import csv
from multiprocessing import Pool
import argparse
-import datetime
import logging
import copy
import itertools as it | Removed datetime import since start_logging is now in the utils module | KarchinLab_probabilistic2020 | train |
4778a95aefb7e3d8634270e04c96bb496ed5616d | diff --git a/src/FiniteStateMachine.php b/src/FiniteStateMachine.php
index <HASH>..<HASH> 100644
--- a/src/FiniteStateMachine.php
+++ b/src/FiniteStateMachine.php
@@ -110,6 +110,12 @@ class FiniteStateMachine
throw new Exception('Sleep mode', self::EXCEPTION_SLEEP);
}
$this->_sleep = true;
+ $this->_log[] = array(
+ 'state' => $this->_state,
+ 'reason' => 'sleep',
+ 'symbol' => null,
+ 'timestamp' => $this->getTimestamp(),
+ );
return;
if ($this->_sleep) {
throw new RuntimeException('Could not call method over the sleep mode', self::EXCEPTION_SLEEP);
diff --git a/tests/FiniteStateMachine/SleepTest.php b/tests/FiniteStateMachine/SleepTest.php
index <HASH>..<HASH> 100644
--- a/tests/FiniteStateMachine/SleepTest.php
+++ b/tests/FiniteStateMachine/SleepTest.php
@@ -19,8 +19,10 @@ class Fsm_SleepTest extends FsmTestCase
parent::setUp();
$methods = array(
'isSleep',
+ 'getTimestamp',
);
$this->_fsm = $this->getMockBuilder('TestFiniteStateMachine')->setMethods($methods)->getMock();
+ $this->_fsm->method('getTimestamp')->will($this->returnValue('1.000000'));
$stateSet = array_shift(array_shift($this->provideValidStateSets()));
$this->_fsm->setStateSet($stateSet);
}
@@ -121,4 +123,56 @@ class Fsm_SleepTest extends FsmTestCase
$log = $this->_fsm->sleep();
$this->assertSame($expectedLog, $log);
}
+
+ public function provideSleepLogs()
+ {
+ $log = array(
+ array(
+ 'state' => 'INIT',
+ 'reason' => 'init',
+ 'symbol' => null,
+ 'timestamp' => '1.000000',
+ ),
+ );
+ return array(
+ array(
+ 'state' => 'INIT',
+ 'log' => $log,
+ 'expectedLog' => array_merge($log, array(
+ array(
+ 'state' => 'INIT',
+ 'reason' => 'sleep',
+ 'symbol' => null,
+ 'timestamp' => '1.000000', //see setUp() method
+ ),
+ )),
+ ),
+ array(
+ 'state' => 'CHECKOUT',
+ 'log' => array(),
+ 'expectedLog' => array(
+ array(
+ 'state' => 'CHECKOUT',
+ 'reason' => 'sleep',
+ 'symbol' => null,
+ 'timestamp' => '1.000000', //see setUp() method
+ ),
+ ),
+ ),
+ );
+ }
+
+ /**
+ * @group issue1
+ * @group issue1_sleep_protected
+ * @dataProvider provideSleepLogs
+ */
+ public function test_Sleep_AppendsSleepItemToLog($state, $log, $expectedLog)
+ {
+ $this->setState($state);
+ $this->setLog($log);
+ $this->_fsm->sleep();
+ $log = $this->getLog();
+ $this->assertSame($expectedLog, $log);
+ }
} | #1: Fsm_SleepTest::test_Sleep_AppendsSleepItemToLog() | tourman_fsm | train |
f54ad9e973196b969c81ea82bbdeabd3327f6086 | diff --git a/lib/pdk/util/changelog_generator.rb b/lib/pdk/util/changelog_generator.rb
index <HASH>..<HASH> 100644
--- a/lib/pdk/util/changelog_generator.rb
+++ b/lib/pdk/util/changelog_generator.rb
@@ -9,11 +9,15 @@ module PDK
# Raises if the github_changelog_generator is not available
def self.github_changelog_generator_available!
- require 'bundler'
- return if ::Bundler.rubygems.find_name(GEM).any?
+ check_command = PDK::CLI::Exec::InteractiveCommand.new(PDK::CLI::Exec.bundle_bin, 'show', 'github_changelog_generator')
+ check_command.context = :module
+
+ result = check_command.execute!
+
+ return if result[:exit_code].zero?
raise PDK::CLI::ExitWithError, _(
- 'Unable to generate the changelog as the %{gem} gem is not installed',
+ 'Unable to generate the changelog as the %{gem} gem is not included in this module\'s Gemfile',
) % { gem: GEM }
end
diff --git a/spec/unit/pdk/util/changelog_generator_spec.rb b/spec/unit/pdk/util/changelog_generator_spec.rb
index <HASH>..<HASH> 100644
--- a/spec/unit/pdk/util/changelog_generator_spec.rb
+++ b/spec/unit/pdk/util/changelog_generator_spec.rb
@@ -199,16 +199,19 @@ describe PDK::Util::ChangelogGenerator do
describe '.github_changelog_generator_available!' do
subject(:method) { described_class.github_changelog_generator_available! }
- let(:dummy_spec) { gem_present ? [Gem::Specification.new] : [] }
+ let(:command) { double(PDK::CLI::Exec::InteractiveCommand, :context= => nil) } # rubocop:disable RSpec/VerifiedDoubles
before(:each) do
- allow(::Bundler.rubygems).to receive(:find_name).and_call_original
- allow(::Bundler.rubygems).to receive(:find_name)
- .with('github_changelog_generator').and_return(dummy_spec)
+ expect(PDK::CLI::Exec::InteractiveCommand).to receive(:new).and_return(command)
end
context 'when the gem is available to Bundler' do
- let(:gem_present) { true }
+ let(:command_stdout) { '/path/to/gems/github_changelog_generator-1.15.2' }
+ let(:command_exit_code) { 0 }
+
+ before(:each) do
+ expect(command).to receive(:execute!).and_return(stdout: command_stdout, exit_code: command_exit_code)
+ end
it 'does not raise an error' do
expect { method }.not_to raise_error
@@ -216,10 +219,15 @@ describe PDK::Util::ChangelogGenerator do
end
context 'when the gem is not available to Bundler' do
- let(:gem_present) { false }
+ let(:command_stderr) { 'Could not find gem \'github_changelog_generator\'.' }
+ let(:command_exit_code) { 7 }
+
+ before(:each) do
+ expect(command).to receive(:execute!).and_return(stderr: command_stderr, exit_code: command_exit_code)
+ end
it 'raises an error' do
- expect { method }.to raise_error(PDK::CLI::ExitWithError, %r{not installed})
+ expect { method }.to raise_error(PDK::CLI::ExitWithError, %r{not included})
end
end
end | (#<I>) Check for `github_changelog_generator` in proper bundler context
Previously, the code was checking for the github_changelog_generator gem
in the PDK's execution environment, but then invoking the gem in the
module's execution environment. This change updates the code to check
for the gem in the module's execution environment by invoking `bundle
show`. | puppetlabs_pdk | train |
183c03a28269911db9492ccdab93726381623d0b | diff --git a/main/coreplugins/io.sarl.eclipse/src/io/sarl/eclipse/util/Jdt2Ecore.java b/main/coreplugins/io.sarl.eclipse/src/io/sarl/eclipse/util/Jdt2Ecore.java
index <HASH>..<HASH> 100644
--- a/main/coreplugins/io.sarl.eclipse/src/io/sarl/eclipse/util/Jdt2Ecore.java
+++ b/main/coreplugins/io.sarl.eclipse/src/io/sarl/eclipse/util/Jdt2Ecore.java
@@ -679,10 +679,10 @@ public class Jdt2Ecore {
@Override
public IType next() {
- if (this.current == null) {
+ final IType c = this.current;
+ if (c == null) {
throw new NoSuchElementException();
}
- final IType c = this.current;
final String name = c.getFullyQualifiedName();
this.encountered.add(name);
try {
@@ -693,12 +693,14 @@ public class Jdt2Ecore {
superTypes = new String[] {c.getSuperclassTypeSignature()};
}
for (final String signature : superTypes) {
- final String resolvedSignature = resolveType(c, signature);
- if (!Strings.isNullOrEmpty(resolvedSignature)) {
- this.queue.add(resolvedSignature);
+ if (!Strings.isNullOrEmpty(signature)) {
+ final String resolvedSignature = resolveType(c, signature);
+ if (!Strings.isNullOrEmpty(resolvedSignature)) {
+ this.queue.add(resolvedSignature);
+ }
}
}
- } catch (JavaModelException exception) {
+ } catch (Exception exception) {
//
}
updateCurrent();
diff --git a/main/coreplugins/io.sarl.eclipse/src/io/sarl/eclipse/wizards/elements/AbstractNewSarlElementWizardPage.java b/main/coreplugins/io.sarl.eclipse/src/io/sarl/eclipse/wizards/elements/AbstractNewSarlElementWizardPage.java
index <HASH>..<HASH> 100644
--- a/main/coreplugins/io.sarl.eclipse/src/io/sarl/eclipse/wizards/elements/AbstractNewSarlElementWizardPage.java
+++ b/main/coreplugins/io.sarl.eclipse/src/io/sarl/eclipse/wizards/elements/AbstractNewSarlElementWizardPage.java
@@ -1151,17 +1151,17 @@ public abstract class AbstractNewSarlElementWizardPage extends NewTypeWizardPage
this.injector.injectMembers(extension);
final AbstractSuperTypeSelectionDialog<?> dialog = createSuperClassSelectionDialog(getShell(),
getWizard().getContainer(), project, extension, false);
- if (dialog != null) {
- this.injector.injectMembers(dialog);
- dialog.setTitle(NewWizardMessages.NewTypeWizardPage_SuperClassDialog_title);
- dialog.setMessage(NewWizardMessages.NewTypeWizardPage_SuperClassDialog_message);
- dialog.setInitialPattern(getSuperClass());
+ if (dialog == null) {
+ return super.chooseSuperClass();
+ }
- if (dialog.open() == Window.OK) {
- return (IType) dialog.getFirstResult();
- }
- } else {
- super.chooseSuperClass();
+ this.injector.injectMembers(dialog);
+ dialog.setTitle(NewWizardMessages.NewTypeWizardPage_SuperClassDialog_title);
+ dialog.setMessage(NewWizardMessages.NewTypeWizardPage_SuperClassDialog_message);
+ dialog.setInitialPattern(getSuperClass());
+
+ if (dialog.open() == Window.OK) {
+ return (IType) dialog.getFirstResult();
}
return null;
} | [eclipse] Avoid NPE when creating a type with a super type.
close #<I> | sarl_sarl | train |
50532b807408de0bb2e0f28e8a63d491826ff43c | diff --git a/gwt-material/src/main/java/gwt/material/design/client/ui/MaterialLoader.java b/gwt-material/src/main/java/gwt/material/design/client/ui/MaterialLoader.java
index <HASH>..<HASH> 100644
--- a/gwt-material/src/main/java/gwt/material/design/client/ui/MaterialLoader.java
+++ b/gwt-material/src/main/java/gwt/material/design/client/ui/MaterialLoader.java
@@ -214,4 +214,12 @@ public class MaterialLoader {
public void setScrollDisabled(boolean scrollDisabled) {
this.scrollDisabled = scrollDisabled;
}
+
+ public MaterialProgress getProgress() {
+ return progress;
+ }
+
+ public MaterialPreLoader getPreLoader() {
+ return preLoader;
+ }
} | Add getters for progress and preLoader. | GwtMaterialDesign_gwt-material | train |
95c1bad8673267aaf2f8dea9a0547f9b52a46790 | diff --git a/lib/fog/bluebox/models/blb/lb_application.rb b/lib/fog/bluebox/models/blb/lb_application.rb
index <HASH>..<HASH> 100644
--- a/lib/fog/bluebox/models/blb/lb_application.rb
+++ b/lib/fog/bluebox/models/blb/lb_application.rb
@@ -4,7 +4,7 @@ module Fog
module Bluebox
class BLB
- class Application < Fog::Model
+ class LbApplication < Fog::Model
identity :id
attribute :name
diff --git a/lib/fog/bluebox/models/blb/lb_applications.rb b/lib/fog/bluebox/models/blb/lb_applications.rb
index <HASH>..<HASH> 100644
--- a/lib/fog/bluebox/models/blb/lb_applications.rb
+++ b/lib/fog/bluebox/models/blb/lb_applications.rb
@@ -3,20 +3,13 @@ require 'fog/core/collection'
module Fog
module Bluebox
class BLB
- class Applications < Fog::Collection
+ class LbApplications < Fog::Collection
def all
- data = service.get_blocks.body
+ data = service.get_lb_applications.body
load(data)
end
- def bootstrap(new_attributes = {})
- server = create(new_attributes)
- server.wait_for { ready? }
- server.setup(:key_data => [server.private_key])
- server
- end
-
def get(server_id)
if server_id && server = service.get_block(server_id).body
new(server)
diff --git a/lib/fog/bluebox/models/blb/lb_service.rb b/lib/fog/bluebox/models/blb/lb_service.rb
index <HASH>..<HASH> 100644
--- a/lib/fog/bluebox/models/blb/lb_service.rb
+++ b/lib/fog/bluebox/models/blb/lb_service.rb
@@ -4,7 +4,7 @@ module Fog
module Bluebox
class BLB
- class Service < Fog::Model
+ class LbService < Fog::Model
identity :id
attribute :name
diff --git a/lib/fog/bluebox/models/blb/lb_services.rb b/lib/fog/bluebox/models/blb/lb_services.rb
index <HASH>..<HASH> 100644
--- a/lib/fog/bluebox/models/blb/lb_services.rb
+++ b/lib/fog/bluebox/models/blb/lb_services.rb
@@ -3,20 +3,13 @@ require 'fog/core/collection'
module Fog
module Bluebox
class BLB
- class Applications < Fog::Collection
+ class LbServices < Fog::Collection
def all
- data = service.get_blocks.body
+ data = service.get_lb_services.body
load(data)
end
- def bootstrap(new_attributes = {})
- server = create(new_attributes)
- server.wait_for { ready? }
- server.setup(:key_data => [server.private_key])
- server
- end
-
def get(server_id)
if server_id && server = service.get_block(server_id).body
new(server) | [bluebox|blb] some cleanup of model stubs | fog_fog | train |
63bc32beb93c41914e0ab2e903ec043e11f717aa | diff --git a/law/parameter.py b/law/parameter.py
index <HASH>..<HASH> 100644
--- a/law/parameter.py
+++ b/law/parameter.py
@@ -135,17 +135,18 @@ class DurationParameter(luigi.Parameter):
class CSVParameter(luigi.Parameter):
- """ __init__(*args, cls=luigi.Parameter, unique=False, min_len=None, max_len=None, \
+ """ __init__(*args, cls=luigi.Parameter, unique=False, sort=False, min_len=None, max_len=None, \
choices=None, brace_expand=False, **kwargs)
Parameter that parses a comma-separated value (CSV) and produces a tuple. *cls* can refer to an
other parameter class that will be used to parse and serialize the particular items.
When *unique* is *True*, both parsing and serialization methods make sure that values are
- unique.
+ unique. *sort* can be a boolean or a function for sorting parameter values.
When *min_len* (*max_len*) is set to an integer, an error is raised in case the number of
- elements to serialize or parse is deceeds (exceeds) that value. Just like done in luigi's
- *ChoiceParamater*, *choices* can be a sequence of accepted values.
+ elements to serialize or parse (evaluated after potentially ensuring uniqueness) deceeds
+ (exceeds) that value. Just like done in luigi's *ChoiceParamater*, *choices* can be a sequence
+ of accepted values.
When *brace_expand* is *True*, brace expansion is applied, potentially extending the list of
values.
@@ -202,6 +203,7 @@ class CSVParameter(luigi.Parameter):
def __init__(self, *args, **kwargs):
self._cls = kwargs.pop("cls", luigi.Parameter)
self._unique = kwargs.pop("unique", False)
+ self._sort = kwargs.pop("sort", False)
self._min_len = kwargs.pop("min_len", None)
self._max_len = kwargs.pop("max_len", None)
self._choices = kwargs.pop("choices", None)
@@ -215,6 +217,21 @@ class CSVParameter(luigi.Parameter):
self._inst = self._cls()
+ def _check_unique(self, value):
+ if not self._unique:
+ return value
+
+ return make_unique(value)
+
+ def _check_sort(self, value):
+ if not self._sort:
+ return value
+
+ key = self._sort if callable(self._sort) else None
+ value = sorted(value, key=key)
+
+ return tuple(value)
+
def _check_len(self, value):
str_repr = lambda: self.CSV_SEP.join(str(v) for v in value)
@@ -241,7 +258,6 @@ class CSVParameter(luigi.Parameter):
raise ValueError("invalid parameter value(s) '{}', valid choices are '{}'".format(
str_repr(make_unique(unknown)), str_repr(self._choices)))
-
def parse(self, inp):
""""""
if inp in (None, "", NO_STR):
@@ -257,14 +273,10 @@ class CSVParameter(luigi.Parameter):
else:
ret = (ret,)
- # ensure uniqueness
- if self._unique:
- ret = make_unique(ret)
-
- # check min_len and max_len
+ # apply uniqueness, sort, length and choices checks
+ ret = self._check_unique(ret)
+ ret = self._check_sort(ret)
self._check_len(ret)
-
- # check choices
self._check_choices(ret)
return ret
@@ -274,32 +286,25 @@ class CSVParameter(luigi.Parameter):
if not value:
return ""
else:
- # ensure uniqueness
- if self._unique:
- value = make_unique(value)
-
- # check min_len and max_len
+ # apply uniqueness, sort, length and choices checks
+ value = self._check_unique(value)
+ value = self._check_sort(value)
self._check_len(value)
-
- # check choices
self._check_choices(value)
return self.CSV_SEP.join(str(self._inst.serialize(elem)) for elem in value)
class MultiCSVParameter(CSVParameter):
- """ __init__(*args, cls=luigi.Parameter, unique=False, min_len=None, max_len=None, \
- brace_expand=False, **kwargs)
+ """ __init__(*args, cls=luigi.Parameter, unique=False, sort=False, min_len=None, max_len=None, \
+ choices=None, brace_expand=False, **kwargs)
Parameter that parses several comma-separated values (CSV), separated by colons, and produces a
nested tuple. *cls* can refer to an other parameter class that will be used to parse and
serialize the particular items.
Except for the additional support for multuple CSV sequences, the implementation is based on
- :py:class:`CSVParameter`, which also handles the features controlled by *unique*, *max_len* and
- *min_len*.
-
- When *brace_expand* is *True*, brace expansion is applied, potentially extending the lists of
- values.
+ :py:class:`CSVParameter`, which also handles the features controlled by *unique*, *sort*,
+ *max_len*, *min_len*, *choices* and *brace_expand* per sequence of values.
Example:
@@ -319,6 +324,10 @@ class MultiCSVParameter(CSVParameter):
p.parse("4,5:6,7,8")
# => ValueError
+ p = MultiCSVParameter(cls=luigi.IntParameter, choices=(1, 2))
+ p.parse("1,2:2,3")
+ # => ValueError
+
p = MultiCSVParameter(cls=luigi.IntParameter, brace_expand=True)
p.parse("4,5:6,7,8{8,9}")
# => ((4, 5), (6, 7, 88, 89)) | Add sort option to CSVParameter, simplify value checks. | riga_law | train |
51a3d3beaba31aeb9ed3e76425b7ad7d76416c11 | diff --git a/admin/install.php b/admin/install.php
index <HASH>..<HASH> 100644
--- a/admin/install.php
+++ b/admin/install.php
@@ -16,7 +16,7 @@ if ( !yourls_check_database_version() ) {
if ( !yourls_check_php_version() ) {
$error[] = yourls_s( '%s version is too old. Ask your server admin for an upgrade.', 'PHP' );
- yourls_debug_log( 'PHP version: ' . phpversion() );
+ yourls_debug_log( 'PHP version: ' . PHP_VERSION );
}
// Is YOURLS already installed ?
diff --git a/includes/functions-http.php b/includes/functions-http.php
index <HASH>..<HASH> 100644
--- a/includes/functions-http.php
+++ b/includes/functions-http.php
@@ -296,7 +296,7 @@ function yourls_check_core_version() {
'failed_attempts' => $checks->failed_attempts,
'yourls_site' => defined( 'YOURLS_SITE' ) ? YOURLS_SITE : 'unknown',
'yourls_version' => defined( 'YOURLS_VERSION' ) ? YOURLS_VERSION : 'unknown',
- 'php_version' => phpversion(),
+ 'php_version' => PHP_VERSION,
'mysql_version' => $ydb->mysql_version(),
'locale' => yourls_get_locale(), | Replace phpversion() with PHP_VERSION
According to my tests, we'll speed things by almost <I> seconds every 1 million calls, dude. | YOURLS_YOURLS | train |
9b63ae2b5bcb03f41e8a44f0becf794962db38e1 | diff --git a/sherlock/transient_classifier.py b/sherlock/transient_classifier.py
index <HASH>..<HASH> 100644
--- a/sherlock/transient_classifier.py
+++ b/sherlock/transient_classifier.py
@@ -336,7 +336,7 @@ class transient_classifier():
if self.ra:
return classifications, crossmatches
- if self.updateAnnotations:
+ if self.updateAnnotations and self.settings["database settings"]["transients"]["transient peak magnitude query"]:
self.update_peak_magnitudes()
self.update_classification_annotations_and_summaries(
self.updateAnnotations) | you can now set "transient peak magnitude query" to false in the settings file | thespacedoctor_sherlock | train |
00e009826dc8d05b7f8520b5b573a8634415b7b1 | diff --git a/armstrong/core/arm_layout/backends.py b/armstrong/core/arm_layout/backends.py
index <HASH>..<HASH> 100644
--- a/armstrong/core/arm_layout/backends.py
+++ b/armstrong/core/arm_layout/backends.py
@@ -12,8 +12,7 @@ class BasicLayoutBackend(object):
a._meta.object_name.lower(), name))
return ret
- def render(self, object, name, dictionary=None,
- context_instance=None):
+ def render(self, object, name, dictionary=None, context_instance=None):
dictionary = dictionary or {}
dictionary["object"] = object
template_name = self.get_layout_template_name(object, name)
@@ -24,13 +23,14 @@ class BasicLayoutBackend(object):
return self.render(*args, **kwargs)
-# DEPRECATED: To be removed in ArmLayout 1.4.
+# DEPRECATED: To be removed in ArmLayout 2.0
import warnings
+
def deprecate(cls):
def wrapper(*args, **kwargs):
- msg = "BasicRenderModelBackend is deprecated and will be removed in ArmLayout 1.4. Use BasicLayoutBackend."
- warnings.warn(msg.format(cls.__name__), DeprecationWarning)
+ msg = "BasicRenderModelBackend is deprecated and will be removed in ArmLayout 2.0. Use %s."
+ warnings.warn(msg % cls.__name__, DeprecationWarning)
return cls(*args, **kwargs)
return wrapper
diff --git a/armstrong/core/arm_layout/utils.py b/armstrong/core/arm_layout/utils.py
index <HASH>..<HASH> 100644
--- a/armstrong/core/arm_layout/utils.py
+++ b/armstrong/core/arm_layout/utils.py
@@ -1,5 +1,3 @@
-import warnings
-from django.conf import settings
from armstrong.utils.backends import GenericBackend
NEW = "ARMSTRONG_LAYOUT_BACKEND"
@@ -9,21 +7,25 @@ render_model = (GenericBackend(NEW,
defaults="armstrong.core.arm_layout.backends.BasicLayoutBackend")
.get_backend())
+
+# DEPRECATED: To be removed in ArmLayout 2.0
+import warnings
+from django.conf import settings
+from django.utils.safestring import mark_safe
+from django.template.loader import render_to_string
+
+
if hasattr(settings, OLD):
- msg = "%s is deprecated and will be removed in ArmLayout 1.4. Use %s." % (OLD, NEW)
+ msg = "%s is deprecated and will be removed in ArmLayout 2.0. Use %s." % (OLD, NEW)
warnings.warn(msg, DeprecationWarning)
render_model = (GenericBackend(OLD,
defaults="armstrong.core.arm_layout.backends.BasicLayoutBackend")
.get_backend())
-# DEPRECATED: To be removed in ArmLayout 1.4. Here for backwards compatibility
-from django.utils.safestring import mark_safe
-from django.template.loader import render_to_string
-
def deprecate(func):
def wrapper(*args, **kwargs):
- msg = "Importing `%s` from this module is deprecated and will be removed in ArmLayout 1.4"
+ msg = "Importing `%s` from this module is deprecated and will be removed in ArmLayout 2.0"
warnings.warn(msg % func.__name__, DeprecationWarning)
return func(*args, **kwargs)
return wrapper | Organization tweaks and document removal in version <I> (not <I>). | armstrong_armstrong.core.arm_layout | train |
dcb7d9598209905fdf8cbdc2999e95342cabeee0 | diff --git a/lib/mwlib/src/MW/Setup/Manager/Default.php b/lib/mwlib/src/MW/Setup/Manager/Default.php
index <HASH>..<HASH> 100644
--- a/lib/mwlib/src/MW/Setup/Manager/Default.php
+++ b/lib/mwlib/src/MW/Setup/Manager/Default.php
@@ -107,7 +107,9 @@ class MW_Setup_Manager_Default extends MW_Setup_Manager_Abstract
foreach( $this->_tasks as $name => $task )
{
- $this->_dependencies[$name] = (array) $task->getPreDependencies();
+ foreach( (array) $task->getPreDependencies() as $taskname ) {
+ $this->_dependencies[$name][] = $taskname;
+ }
foreach( (array) $task->getPostDependencies() as $taskname ) {
$this->_dependencies[$taskname][] = $name; | Fixes overwriting post-dependencies to get the correct order of the setup tasks | Arcavias_arcavias-core | train |
84d9b2349b12ed3b202dbad17bd2d6fcf4bfef3a | diff --git a/samples/SLS_Sample.java b/samples/SLS_Sample.java
index <HASH>..<HASH> 100644
--- a/samples/SLS_Sample.java
+++ b/samples/SLS_Sample.java
@@ -4,6 +4,7 @@ import java.util.Map;
public class SLS_Sample {
private String name;
private int age;
+ private int height;
SLS_Sample hasIt(List<SLS_Sample> l, String n) {
@@ -32,7 +33,7 @@ public class SLS_Sample {
boolean fpSetFlag(Map<String, SLS_Sample> m, String name) {
boolean found = false;
for (SLS_Sample s : m.values()) {
- if (s == null || s.name.equals(name)) {
+ if ((s == null) || s.name.equals(name)) {
found = true;
s.age = 1;
}
@@ -52,4 +53,22 @@ public class SLS_Sample {
return total;
}
+ void fpTwoSets(List<SLS_Sample> l, String name, int age, int height) {
+ String n = null;
+ int a = 0;
+ int h = 0;
+
+ for (SLS_Sample s : l) {
+ if (s.name.equals(name)) {
+ n = s.name;
+ } else if (s.age == age) {
+ a = s.age;
+ } else if (s.height == height) {
+ h = s.height;
+ }
+ }
+
+ System.out.println("Found: " + n + " " + a + " " + h);
+ }
+
} | add attempt at seen fp for SLS, needs tweaking | mebigfatguy_fb-contrib | train |
f1a2c021e3c4d4e0916c8bb2e4c13bfed55d177e | diff --git a/actionmailer/lib/action_mailer/base.rb b/actionmailer/lib/action_mailer/base.rb
index <HASH>..<HASH> 100644
--- a/actionmailer/lib/action_mailer/base.rb
+++ b/actionmailer/lib/action_mailer/base.rb
@@ -455,10 +455,6 @@ module ActionMailer
PROTECTED_IVARS = AbstractController::Rendering::DEFAULT_PROTECTED_INSTANCE_VARIABLES + [:@_action_has_layout]
- def _protected_ivars # :nodoc:
- PROTECTED_IVARS
- end
-
helper ActionMailer::MailHelper
class_attribute :delivery_job, default: ::ActionMailer::DeliveryJob
@@ -1030,6 +1026,10 @@ module ActionMailer
"action_mailer"
end
+ def _protected_ivars
+ PROTECTED_IVARS
+ end
+
ActiveSupport.run_load_hooks(:action_mailer, self)
end
end
diff --git a/actionpack/lib/abstract_controller/rendering.rb b/actionpack/lib/abstract_controller/rendering.rb
index <HASH>..<HASH> 100644
--- a/actionpack/lib/abstract_controller/rendering.rb
+++ b/actionpack/lib/abstract_controller/rendering.rb
@@ -120,7 +120,7 @@ module AbstractController
options
end
- def _protected_ivars # :nodoc:
+ def _protected_ivars
DEFAULT_PROTECTED_INSTANCE_VARIABLES
end
end
diff --git a/actionpack/lib/action_controller/base.rb b/actionpack/lib/action_controller/base.rb
index <HASH>..<HASH> 100644
--- a/actionpack/lib/action_controller/base.rb
+++ b/actionpack/lib/action_controller/base.rb
@@ -263,9 +263,10 @@ module ActionController
@_view_renderer @_lookup_context @_routes @_view_runtime @_db_runtime @_helper_proxy
)
- def _protected_ivars # :nodoc:
+ def _protected_ivars
PROTECTED_IVARS
end
+ private :_protected_ivars
ActiveSupport.run_load_hooks(:action_controller_base, self)
ActiveSupport.run_load_hooks(:action_controller, self) | Make _protected_ivars private
This method is only used internally and it being public it was being
retorned in the `action_methods` list. | rails_rails | train |
ff784822fa5a81c3fb1af5e5243bf566326f7d41 | diff --git a/hyperbahn/advertise_test.go b/hyperbahn/advertise_test.go
index <HASH>..<HASH> 100644
--- a/hyperbahn/advertise_test.go
+++ b/hyperbahn/advertise_test.go
@@ -47,6 +47,24 @@ func TestAdvertiseFailed(t *testing.T) {
})
}
+func TestNotListeningChannel(t *testing.T) {
+ withSetup(t, func(hypCh *tchannel.Channel, hyperbahnHostPort string) {
+ adHandler := func(ctx json.Context, req *AdRequest) (*AdResponse, error) {
+ return &AdResponse{1}, nil
+ }
+ json.Register(hypCh, json.Handlers{"ad": adHandler}, nil)
+
+ ch, err := testutils.NewClient(nil)
+ require.NoError(t, err, "testutils NewClient failed")
+
+ client, err := NewClient(ch, configFor(hyperbahnHostPort), nil)
+ assert.NoError(t, err, "hyperbahn NewClient failed")
+ defer client.Close()
+
+ assert.Equal(t, errEphemeralPeer, client.Advertise(), "Advertise without Listen should fail")
+ })
+}
+
type retryTest struct {
// channel used to control the response to an 'ad' call.
respCh chan int | Add test: Advertise should fail before Listen | uber_tchannel-go | train |
860e1fd3ed39994c9404a6e0b6a4f91b888ce39f | diff --git a/core/src/main/java/me/prettyprint/cassandra/model/ExecutingKeyspace.java b/core/src/main/java/me/prettyprint/cassandra/model/ExecutingKeyspace.java
index <HASH>..<HASH> 100644
--- a/core/src/main/java/me/prettyprint/cassandra/model/ExecutingKeyspace.java
+++ b/core/src/main/java/me/prettyprint/cassandra/model/ExecutingKeyspace.java
@@ -18,7 +18,7 @@ import me.prettyprint.hector.api.exceptions.HectorException;
/**
* Thread Safe
* @author Ran Tavory
- *
+ * @author zznate
*/
public class ExecutingKeyspace implements Keyspace {
private static final Map<String, String> EMPTY_CREDENTIALS = Collections.emptyMap();
@@ -34,19 +34,18 @@ public class ExecutingKeyspace implements Keyspace {
public ExecutingKeyspace(String keyspace, HConnectionManager connectionManager,
ConsistencyLevelPolicy consistencyLevelPolicy, FailoverPolicy failoverPolicy) {
this(keyspace, connectionManager, consistencyLevelPolicy, failoverPolicy, EMPTY_CREDENTIALS);
- // TODO add exceptionTranslator w/ default
}
public ExecutingKeyspace(String keyspace, HConnectionManager connectionManager,
ConsistencyLevelPolicy consistencyLevelPolicy, FailoverPolicy failoverPolicy,
Map<String, String> credentials) {
- // TODO allow null keyspace for meta ops
Assert.noneNull(consistencyLevelPolicy, connectionManager);
this.keyspace = keyspace;
this.connectionManager = connectionManager;
this.consistencyLevelPolicy = consistencyLevelPolicy;
this.failoverPolicy = failoverPolicy;
this.credentials = credentials;
+ // TODO make this plug-able
this.exceptionTranslator = new ExceptionsTranslatorImpl();
}
@@ -67,8 +66,6 @@ public class ExecutingKeyspace implements Keyspace {
}
public <T> ExecutionResult<T> doExecute(KeyspaceOperationCallback<T> koc) throws HectorException {
- // doExecute(Operation<T> op) {
- // connectionManager.operateWithFailover(op)
KeyspaceService ks = null;
try {
ks = new KeyspaceServiceImpl(keyspace, consistencyLevelPolicy, connectionManager, failoverPolicy, credentials); | removed extraneous comments, added authorship attrib | hector-client_hector | train |
1f76b3dbfe8ac1c0e8531425e6d54c543c2bbd0c | diff --git a/server.js b/server.js
index <HASH>..<HASH> 100644
--- a/server.js
+++ b/server.js
@@ -259,10 +259,10 @@ function parseMQTTMessage (topic, message) {
// If sending switch data and there is already a level value, send level instead
// SmartThings will turn the device on
if (property === 'switch' && contents === 'on' &&
- history[topicCommand] !== undefined) {
+ history[getTopicFor(device, 'level', TOPIC_COMMAND)] !== undefined) {
winston.info('Passing level instead of switch on');
property = 'level';
- contents = history[topicCommand];
+ contents = history[getTopicFor(device, 'level', TOPIC_COMMAND)];
}
request.post({ | Fixing bug with sending switch instead of level | stjohnjohnson_smartthings-mqtt-bridge | train |
b24d0c9f8ce7f85f3df0263df5b8f4aca53d4328 | diff --git a/pysat/_instrument.py b/pysat/_instrument.py
index <HASH>..<HASH> 100644
--- a/pysat/_instrument.py
+++ b/pysat/_instrument.py
@@ -271,6 +271,7 @@ class Instrument(object):
# use Instrument definition of MetaLabels over the Metadata declaration
self.meta_labels = labels
self.meta = pysat.Meta(labels=self.meta_labels)
+ self.meta.mutable = False
self.labels = pysat.MetaLabels(metadata=self.meta, **labels)
# function processing class, processes data on load | BUG: Ensured meta object immutable upon Instrument instantiation | rstoneback_pysat | train |
d205244a45bfef6d904db8198a955382ccfa4a73 | diff --git a/vendor/autoprefixer.js b/vendor/autoprefixer.js
index <HASH>..<HASH> 100644
--- a/vendor/autoprefixer.js
+++ b/vendor/autoprefixer.js
@@ -1283,6 +1283,9 @@ require.register("autoprefixer/data/prefixes.js", function(exports, require, mod
"font-variant-ligatures": {
browsers: ["bb 10", "chrome 16", "chrome 17", "chrome 18", "chrome 19", "chrome 20", "chrome 21", "chrome 22", "chrome 23", "chrome 24", "chrome 25", "chrome 26", "chrome 27", "chrome 28", "chrome 29", "chrome 30", "ff 4", "ff 5", "ff 6", "ff 7", "ff 8", "ff 9", "ff 10", "ff 11", "ff 12", "ff 13", "ff 14", "ff 15", "ff 16", "ff 17", "ff 18", "ff 19", "ff 20", "ff 21", "ff 22", "ff 23", "ff 24", "ios 7", "opera 15", "opera 16", "safari 7"]
},
+ hyphens: {
+ browsers: ["ff 6", "ff 7", "ff 8", "ff 9", "ff 10", "ff 11", "ff 12", "ff 13", "ff 14", "ff 15", "ff 16", "ff 17", "ff 18", "ff 19", "ff 20", "ff 21", "ff 22", "ff 23", "ff 24", "ie 10", "ie 11", "ios 4.2", "ios 4.3", "ios 5", "ios 5.1", "ios 6", "ios 6.1", "ios 7", "safari 5.1", "safari 6", "safari 7"]
+ },
"justify-content": {
browsers: ["android 2.1", "android 2.2", "android 2.3", "android 3", "android 4", "android 4.1", "android 4.2", "bb 7", "bb 10", "chrome 4", "chrome 5", "chrome 6", "chrome 7", "chrome 8", "chrome 9", "chrome 10", "chrome 11", "chrome 12", "chrome 13", "chrome 14", "chrome 15", "chrome 16", "chrome 17", "chrome 18", "chrome 19", "chrome 20", "chrome 21", "chrome 22", "chrome 23", "chrome 24", "chrome 25", "chrome 26", "chrome 27", "chrome 28", "ff 2", "ff 3", "ff 3.5", "ff 3.6", "ff 4", "ff 5", "ff 6", "ff 7", "ff 8", "ff 9", "ff 10", "ff 11", "ff 12", "ff 13", "ff 14", "ff 15", "ff 16", "ff 17", "ff 18", "ff 19", "ff 20", "ff 21", "ie 10", "ios 3.2", "ios 4", "ios 4.1", "ios 4.2", "ios 4.3", "ios 5", "ios 5.1", "ios 6", "ios 6.1", "ios 7", "opera 15", "opera 16", "safari 3.1", "safari 3.2", "safari 4", "safari 5", "safari 5.1", "safari 6", "safari 7"]
}, | Update autoprefixer.js with hyphens support | ai_autoprefixer-rails | train |
1afdcdf4ca61f52557955e03f813f9c897c1169d | diff --git a/presto-main/src/main/java/com/facebook/presto/operator/scalar/MathFunctions.java b/presto-main/src/main/java/com/facebook/presto/operator/scalar/MathFunctions.java
index <HASH>..<HASH> 100644
--- a/presto-main/src/main/java/com/facebook/presto/operator/scalar/MathFunctions.java
+++ b/presto-main/src/main/java/com/facebook/presto/operator/scalar/MathFunctions.java
@@ -46,6 +46,9 @@ import static com.facebook.presto.spi.type.Decimals.encodeUnscaledValue;
import static com.facebook.presto.spi.type.Decimals.longTenToNth;
import static com.facebook.presto.spi.type.DoubleType.DOUBLE;
import static com.facebook.presto.spi.type.TypeSignature.parseTypeSignature;
+import static com.facebook.presto.spi.type.UnscaledDecimal128Arithmetic.isNegative;
+import static com.facebook.presto.spi.type.UnscaledDecimal128Arithmetic.negate;
+import static com.facebook.presto.spi.type.UnscaledDecimal128Arithmetic.unscaledDecimal;
import static com.facebook.presto.spi.type.VarcharType.VARCHAR;
import static com.facebook.presto.type.DecimalOperators.modulusScalarFunction;
import static com.facebook.presto.type.DecimalOperators.modulusSignatureBuilder;
@@ -130,7 +133,14 @@ public final class MathFunctions
@SqlType("decimal(p, s)")
public static Slice absLong(@SqlType("decimal(p, s)") Slice arg)
{
- return encodeUnscaledValue(decodeUnscaledValue(arg).abs());
+ if (isNegative(arg)) {
+ Slice result = unscaledDecimal(arg);
+ negate(result);
+ return result;
+ }
+ else {
+ return arg;
+ }
}
} | Fast decimal based abs() | prestodb_presto | train |
e667d321299d323064903db707fac147fc791025 | diff --git a/richtextfx/src/main/java/org/fxmisc/richtext/StyledTextArea.java b/richtextfx/src/main/java/org/fxmisc/richtext/StyledTextArea.java
index <HASH>..<HASH> 100644
--- a/richtextfx/src/main/java/org/fxmisc/richtext/StyledTextArea.java
+++ b/richtextfx/src/main/java/org/fxmisc/richtext/StyledTextArea.java
@@ -17,7 +17,6 @@ import java.util.stream.Stream;
import javafx.beans.binding.Binding;
import javafx.beans.binding.Bindings;
-import javafx.beans.binding.BooleanBinding;
import javafx.beans.binding.ObjectBinding;
import javafx.beans.property.BooleanProperty;
import javafx.beans.property.ObjectProperty;
@@ -185,21 +184,18 @@ public class StyledTextArea<PS, S> extends Region
/**
* Indicates when this text area should display a caret.
*/
- private final ObjectProperty<CaretVisibility> showCaret
- = new SimpleObjectProperty<>(this, "showCaret", CaretVisibility.WHEN_EDITABLE);
- public final CaretVisibility getShowCaret() { return showCaret.get(); }
- public final void setShowCaret(CaretVisibility value) { showCaret.set(value); }
- public final ObjectProperty<CaretVisibility> showCaretProperty() { return showCaret; }
+ private final Var<CaretVisibility> showCaret = Var.newSimpleVar(CaretVisibility.AUTO);
+ public final CaretVisibility getShowCaret() { return showCaret.getValue(); }
+ public final void setShowCaret(CaretVisibility value) { showCaret.setValue(value); }
+ public final Var<CaretVisibility> showCaretProperty() { return showCaret; }
public static enum CaretVisibility {
- /** Display a caret when focused. */
- ALWAYS,
- /** Display a caret when focused, enabled, and editable. */
- WHEN_EDITABLE,
- /** Display a caret when focused and enabled. */
- WHEN_ENABLED,
- /** Never display a caret. */
- NEVER
+ /** Caret is displayed. */
+ ON,
+ /** Caret is displayed when area is focused, enabled, and editable. */
+ AUTO,
+ /** Caret is not displayed. */
+ OFF
}
// blinkRate property
@@ -211,7 +207,7 @@ public class StyledTextArea<PS, S> extends Region
public final Duration getBlinkRate() { return blinkRate.get(); }
public final void setBlinkRate(Duration value) { blinkRate.set(value); }
public final ObjectProperty<Duration> blinkRateProperty() { return blinkRate; }
-
+
// undo manager
@Override public UndoManager getUndoManager() { return model.getUndoManager(); }
@Override public void setUndoManager(UndoManagerFactory undoManagerFactory) {
@@ -604,23 +600,20 @@ public class StyledTextArea<PS, S> extends Region
EventStream<Boolean> blinkCaret = EventStreams.valuesOf(showCaretProperty())
.flatMap(mode -> {
switch (mode) {
- case ALWAYS:
- return EventStreams.valuesOf(focusedProperty());
- case NEVER:
- return EventStreams.valuesOf(Val.constant(false));
- case WHEN_ENABLED:
- return EventStreams.valuesOf(focusedProperty()
- .and(disabledProperty().not()));
- default:
- case WHEN_EDITABLE:
- return EventStreams.valuesOf(focusedProperty()
- .and(editableProperty())
- .and(disabledProperty().not()));
- }
+ case ON:
+ return EventStreams.valuesOf(Val.constant(true));
+ case OFF:
+ return EventStreams.valuesOf(Val.constant(false));
+ default:
+ case AUTO:
+ return EventStreams.valuesOf(focusedProperty()
+ .and(editableProperty())
+ .and(disabledProperty().not()));
+ }
});
// the rate at which to display the caret
- EventStream<Duration> blinkRates = EventStreams.valuesOf(blinkRateProperty());
+ EventStream<Duration> blinkRates = EventStreams.valuesOf(blinkRate);
// The caret is visible in periodic intervals,
// but only when blinkCaret is true. | Allow custom dependencies for caret visibility.
To implement one's custom dependency, one can follow the general idea of:
````java
ObservableValue<Boolean> p = // creation code;
area.showCaretProperty().bind(Val.map(p, b -> b ? CaretVisibility.ON : CaretVisibility.OFF));
```` | FXMisc_RichTextFX | train |
46ddf84a0a6431698152ee85e8ac45e2e588c850 | diff --git a/nipap/nipap/nipap.py b/nipap/nipap/nipap.py
index <HASH>..<HASH> 100644
--- a/nipap/nipap/nipap.py
+++ b/nipap/nipap/nipap.py
@@ -2023,17 +2023,14 @@ class Nipap:
del(attr['vrf_name'])
attr['vrf_id'] = vrf['id']
- parent_prefix = self.list_prefix(auth, args['from-prefix'])[0]
- # TODO: what now?
-
# VRF fiddling
if attr['vrf_id'] == 0:
- vrf = None
+ ffp_vrf = None
else:
- vrf = self._get_vrf(auth, attr)
+ ffp_vrf = self._get_vrf(auth, attr)
# get a new prefix
- res = self.find_free_prefix(auth, vrf, args)
+ res = self.find_free_prefix(auth, ffp_vrf, args)
if res != []:
attr['prefix'] = res[0]
else: | Fix from-prefix
Some remnants of code where left here for the improved VRF handling of
from-pool. Unfortunately that broke the from-prefix function. It's now
fixed! | SpriteLink_NIPAP | train |
0fea8c44060d08b3b421f1ddaa809fdffbc89b00 | diff --git a/jest.config.js b/jest.config.js
index <HASH>..<HASH> 100644
--- a/jest.config.js
+++ b/jest.config.js
@@ -9,11 +9,9 @@ module.exports = {
},
testMatch: ['<rootDir>/test/unit/**/*.spec.js'],
testPathIgnorePatterns: ['/node_modules/'],
- setupFilesAfterEnv: [
- './test/setup.js'
- ],
- "transform": {
- "^.+\\.js$": "<rootDir>/node_modules/babel-jest"
+ setupFilesAfterEnv: ['./test/setup.js'],
+ transform: {
+ '^.+\\.js$': '<rootDir>/node_modules/babel-jest'
},
coverageDirectory: 'coverage',
coverageReporters: ['json', 'lcov', 'text-summary', 'clover'],
diff --git a/src/store.js b/src/store.js
index <HASH>..<HASH> 100644
--- a/src/store.js
+++ b/src/store.js
@@ -10,7 +10,7 @@ export function createStore (options) {
export class Store {
constructor (options = {}) {
- if (process.env.NODE_ENV !== 'production') {
+ if (__DEV__) {
assert(typeof Promise !== 'undefined', `vuex requires a Promise polyfill in this browser.`)
assert(this instanceof Store, `store must be called with the new operator.`)
} | fix: UMD bundle containing `process.env` flag (#<I>) | vuejs_vuex | train |
1c3e964f81f7481930f5177404018eee0421ea19 | diff --git a/src/main/java/org/vanilladb/core/storage/buffer/BufferMgr.java b/src/main/java/org/vanilladb/core/storage/buffer/BufferMgr.java
index <HASH>..<HASH> 100644
--- a/src/main/java/org/vanilladb/core/storage/buffer/BufferMgr.java
+++ b/src/main/java/org/vanilladb/core/storage/buffer/BufferMgr.java
@@ -22,6 +22,7 @@ import java.util.List;
import java.util.Map;
import java.util.Map.Entry;
import java.util.Set;
+import java.util.concurrent.atomic.AtomicBoolean;
import java.util.logging.Level;
import java.util.logging.Logger;
@@ -54,12 +55,14 @@ public class BufferMgr implements TransactionLifecycleListener {
protected static final int BUFFER_POOL_SIZE;
private static final long MAX_TIME;
private static final long EPSILON;
+ private static final AtomicBoolean hasWaitingTx = new AtomicBoolean();
static {
MAX_TIME = CoreProperties.getLoader().getPropertyAsLong(BufferMgr.class.getName() + ".MAX_TIME", 10000);
EPSILON = CoreProperties.getLoader().getPropertyAsLong(BufferMgr.class.getName() + ".EPSILON", 50);
BUFFER_POOL_SIZE = CoreProperties.getLoader()
.getPropertyAsInteger(BufferMgr.class.getName() + ".BUFFER_POOL_SIZE", 1024);
+ System.out.println("BufferMgr unpin opt========================================================");
}
class PinningBuffer {
@@ -134,6 +137,7 @@ public class BufferMgr implements TransactionLifecycleListener {
if (buff == null) {
waitedBeforeGotBuffer = true;
synchronized (bufferPool) {
+ hasWaitingTx.set(true);
waitingThreads.add(Thread.currentThread());
while (buff == null && !waitingTooLong(timestamp)) {
@@ -143,6 +147,7 @@ public class BufferMgr implements TransactionLifecycleListener {
}
waitingThreads.remove(Thread.currentThread());
+ hasWaitingTx.set(!waitingThreads.isEmpty());
}
}
@@ -250,8 +255,10 @@ public class BufferMgr implements TransactionLifecycleListener {
bufferPool.unpin(buff);
pinningBuffers.remove(blk);
- synchronized (bufferPool) {
- bufferPool.notifyAll();
+ if (hasWaitingTx.get()) {
+ synchronized (bufferPool) {
+ bufferPool.notifyAll();
+ }
}
}
} | Reduce the call of synchronized block in BufferMgr.unpin | vanilladb_vanillacore | train |
728a9135c883ebd5593ddb39a2529070a7298c10 | diff --git a/plugins/guests/debian/cap/configure_networks.rb b/plugins/guests/debian/cap/configure_networks.rb
index <HASH>..<HASH> 100644
--- a/plugins/guests/debian/cap/configure_networks.rb
+++ b/plugins/guests/debian/cap/configure_networks.rb
@@ -57,8 +57,20 @@ module VagrantPlugins
e_nets[interfaces[network[:interface]]] = e_config
end
end
+
+ # By default, netplan expects the renderer to be systemd-networkd,
+ # but if any device is managed by NetworkManager, then we use that renderer
+ # ref: https://netplan.io/reference
+ renderer = NETPLAN_DEFAULT_RENDERER
+ ethernets.keys.each do |k|
+ if nm_controlled?(comm, k)
+ renderer = "NetworkManager"
+ break
+ end
+ end
+
np_config = {"network" => {"version" => NETPLAN_DEFAULT_VERSION,
- "renderer" => NETPLAN_DEFAULT_RENDERER, "ethernets" => ethernets}}
+ "renderer" => renderer, "ethernets" => ethernets}}
remote_path = upload_tmp_file(comm, np_config.to_yaml)
dest_path = "#{NETPLAN_DIRECTORY}/50-vagrant.yaml"
diff --git a/test/unit/plugins/guests/debian/cap/configure_networks_test.rb b/test/unit/plugins/guests/debian/cap/configure_networks_test.rb
index <HASH>..<HASH> 100644
--- a/test/unit/plugins/guests/debian/cap/configure_networks_test.rb
+++ b/test/unit/plugins/guests/debian/cap/configure_networks_test.rb
@@ -65,6 +65,8 @@ describe "VagrantPlugins::GuestDebian::Cap::ConfigureNetworks" do
end
before do
+ allow(comm).to receive(:test).with("nmcli d show eth1").and_return(false)
+ allow(comm).to receive(:test).with("nmcli d show eth2").and_return(false)
allow(comm).to receive(:test).with("ps -o comm= 1 | grep systemd").and_return(false)
allow(comm).to receive(:test).with("sudo systemctl status systemd-networkd.service").and_return(false)
allow(comm).to receive(:test).with("netplan -h").and_return(false)
@@ -118,7 +120,30 @@ describe "VagrantPlugins::GuestDebian::Cap::ConfigureNetworks" do
expect(comm).to receive(:test).with("netplan -h").and_return(true)
end
+ let(:nm_yml) { "---\nnetwork:\n version: 2\n renderer: NetworkManager\n ethernets:\n eth1:\n dhcp4: true\n eth2:\n addresses:\n - 33.33.33.10/16\n gateway4: 33.33.0.1\n" }
+ let(:networkd_yml) { "---\nnetwork:\n version: 2\n renderer: networkd\n ethernets:\n eth1:\n dhcp4: true\n eth2:\n addresses:\n - 33.33.33.10/16\n gateway4: 33.33.0.1\n" }
+
+ it "uses NetworkManager if detected on device" do
+ allow(cap).to receive(:nm_controlled?).and_return(true)
+ allow(comm).to receive(:test).with("nmcli d show eth1").and_return(true)
+ allow(comm).to receive(:test).with("nmcli d show eth2").and_return(true)
+
+ expect(cap).to receive(:upload_tmp_file).with(comm, nm_yml)
+ .and_return("/tmp/vagrant-network-entry.1234")
+
+ cap.configure_networks(machine, [network_0, network_1])
+
+
+ expect(comm.received_commands[0]).to match("mv -f '/tmp/vagrant-network-entry.*' '/etc/netplan/.*.yaml'")
+ expect(comm.received_commands[0]).to match("chown")
+ expect(comm.received_commands[0]).to match("chmod")
+ expect(comm.received_commands[0]).to match("netplan apply")
+ end
+
it "creates and starts the networks for systemd with netplan" do
+ expect(cap).to receive(:upload_tmp_file).with(comm, networkd_yml)
+ .and_return("/tmp/vagrant-network-entry.1234")
+
cap.configure_networks(machine, [network_0, network_1])
expect(comm.received_commands[0]).to match("mv -f '/tmp/vagrant-network-entry.*' '/etc/netplan/.*.yaml'") | (#<I>) Update netplan config generation to detect NetworkManager
Prior to this commit, when setting up private networks on Ubuntu using
netplan, it assumed that the guest was using systemd, the suggested
default tool to manage networking, and did not take into account devices
that could be managed with NetworkManager. This commit fixes that by
looking at the devices managed on the guest to see if its managed by
NetworkManager, and if so, use that renderer for netplan instead of
networkd. | hashicorp_vagrant | train |
0238e4c2c0b03680cfd4c120d0ce59ca95624487 | diff --git a/lib/pake/pakeColor.class.php b/lib/pake/pakeColor.class.php
index <HASH>..<HASH> 100644
--- a/lib/pake/pakeColor.class.php
+++ b/lib/pake/pakeColor.class.php
@@ -32,7 +32,7 @@ class pakeColor
static function colorize($text = '', $parameters = array())
{
// disable colors if not supported (windows or non tty console)
- if (strtoupper(substr(PHP_OS, 0, 3)) == 'WIN' || !@posix_isatty(STDOUT))
+ if (strtoupper(substr(PHP_OS, 0, 3)) == 'WIN' || !function_exists('posix_isatty') || !@posix_isatty(STDOUT))
{
return $text;
} | fixed a problem with pakeColor when php has no posix functions | indeyets_pake | train |
bd6809d89daa7e4dd779ccef6e234b67cd54c912 | diff --git a/client/src/main/java/com/orientechnologies/orient/client/remote/OStorageRemote.java b/client/src/main/java/com/orientechnologies/orient/client/remote/OStorageRemote.java
index <HASH>..<HASH> 100644
--- a/client/src/main/java/com/orientechnologies/orient/client/remote/OStorageRemote.java
+++ b/client/src/main/java/com/orientechnologies/orient/client/remote/OStorageRemote.java
@@ -483,10 +483,12 @@ public class OStorageRemote extends OStorageAbstract {
network.writeLong(txEntry.record.getIdentity().getClusterPosition());
network.writeInt(txEntry.record.getVersion());
network.writeBytes(txEntry.record.toStream());
-
+ break;
+
case OTransactionEntry.DELETED:
network.writeLong(txEntry.record.getIdentity().getClusterPosition());
network.writeInt(txEntry.record.getVersion());
+ break;
}
} | Fixed issue on remote tx commit. | orientechnologies_orientdb | train |
76bd526a99d2074cfd9a56349b54cc2d46f5821c | diff --git a/src/test/groovy/org/codehaus/gant/ant/tests/Gant_Test.java b/src/test/groovy/org/codehaus/gant/ant/tests/Gant_Test.java
index <HASH>..<HASH> 100644
--- a/src/test/groovy/org/codehaus/gant/ant/tests/Gant_Test.java
+++ b/src/test/groovy/org/codehaus/gant/ant/tests/Gant_Test.java
@@ -102,6 +102,11 @@ public class Gant_Test extends TestCase {
/*
* Run Ant in a separate process. Return the string that results.
*
+ * <p>This method assumes that the command ant is the the path and is executable.</p>
+ *
+ * <p>As at 2008-12-06 Canoo CruiseControl runs with GROOVY_HOME set to /usr/local/java/groovy, and
+ * Codehaus Bamboo runs without GROOVY_HOME being set.</p>
+ *
* @param xmlFile the path to the XML file that Ant is to use.
* @param expectedReturnCode the return code that the Ant execution should return.
* @param withClasspath whether the Ant execution should use the full classpathso as to find all the classes.
@@ -164,9 +169,15 @@ public class Gant_Test extends TestCase {
assertEquals ( createBaseMessage ( ) , runAnt ( path + "/gantTest.xml" , 1 , false ) ) ;
}
//
- // TODO: Find out why this fails on Canoo CruiseControl even though it passes locally and at Codehaus Bamboo.
+ // TODO: Find out why this fails on Canoo CruiseControl even though it passes locally and on Codehaus
+ // Bamboo. Data from builds 139--142 indicate that Ant is failing to executing anything at all. This
+ // implies a configuration issue that is not true for the following test which also fails but ant does
+ // actually start. This implies it is something to do with the path to the XML file that is a problem on
+ // Canoo Cruise Control but not on any other system.
//
public void testRunningAntFromShellSuccessful ( ) {
+ System.err.println ( "AAAA:" + path + "/gantTest.xml" ) ;
+ System.err.println ( "BBBB: " + ( new File ( path , "gantTest.xml" ) ).exists ( ) ) ;
assertEquals ( createBaseMessage ( ) + "\nBUILD SUCCESSFUL\n\n", trimTimeFromSuccessfulBuild ( runAnt ( path + "/gantTest.xml" , 1 , true ) ) ) ;
}
/*
@@ -176,6 +187,12 @@ public class Gant_Test extends TestCase {
//
// TODO: Find out why this test fails on Codehaus Bamboo and Canoo CruiseControl even though it passes
// locally.
+ //
+ // On Canoo Creuise Control, ant starts but then fails in the build target. It looks like an immediate
+ // fail so nothing in the build target is happening. Could be a classpath problem for the Groovy ant
+ // task?
+ //
+ // Codehaus Bamboo seems to not have any access to the detailed logs.
//
public void testBasedirInSubdir ( ) {
final String pathToDirectory = System.getProperty ( "user.dir" ) + System.getProperty ( "file.separator" ) + path ; | Try and get more data about the context from Canoo to see why the test fails.
git-svn-id: <URL> | Gant_Gant | train |
5a0f167c3a2b06e82d045d129b55e1bb86e8d07c | diff --git a/spec/controllers/alchemy/admin/resources_controller_spec.rb b/spec/controllers/alchemy/admin/resources_controller_spec.rb
index <HASH>..<HASH> 100644
--- a/spec/controllers/alchemy/admin/resources_controller_spec.rb
+++ b/spec/controllers/alchemy/admin/resources_controller_spec.rb
@@ -10,8 +10,8 @@ describe Admin::EventsController do
describe '#index' do
let(:params) { Hash.new }
- let!(:peter) { Event.create(name: 'Peter') }
- let!(:lustig) { Event.create(name: 'Lustig') }
+ let!(:peter) { create(:event, name: 'Peter') }
+ let!(:lustig) { create(:event, name: 'Lustig') }
before do
authorize_user(:as_admin)
@@ -33,7 +33,7 @@ describe Admin::EventsController do
end
context "but searching for record with certain association" do
- let(:bauwagen) { Location.create(name: 'Bauwagen') }
+ let(:bauwagen) { create(:location, name: 'Bauwagen') }
let(:params) { {q: {name_or_hidden_name_or_location_name_cont: "Bauwagen"}} }
before do
@@ -52,7 +52,7 @@ describe Admin::EventsController do
describe '#update' do
let(:params) { {q: 'some_query', page: 6} }
- let!(:peter) { Event.create(name: 'Peter') }
+ let!(:peter) { create(:event, name: 'Peter') }
it 'redirects to index, keeping the current location parameters' do
post :update, params: {id: peter.id, event: {name: "Hans"}}.merge(params)
@@ -62,16 +62,17 @@ describe Admin::EventsController do
describe '#create' do
let(:params) { {q: 'some_query', page: 6} }
+ let!(:location) { create(:location) }
it 'redirects to index, keeping the current location parameters' do
- post :create, params: {event: {name: "Hans"}}.merge(params)
+ post :create, params: {event: {name: "Hans", location_id: location.id}}.merge(params)
expect(response.redirect_url).to eq("http://test.host/admin/events?page=6&q=some_query")
end
end
describe '#destroy' do
let(:params) { {q: 'some_query', page: 6} }
- let!(:peter) { Event.create(name: 'Peter') }
+ let!(:peter) { create(:event, name: 'Peter') }
it 'redirects to index, keeping the current location parameters' do
delete :destroy, params: {id: peter.id}.merge(params)
diff --git a/spec/factories.rb b/spec/factories.rb
index <HASH>..<HASH> 100644
--- a/spec/factories.rb
+++ b/spec/factories.rb
@@ -2,6 +2,7 @@ FactoryGirl.define do
factory :event do
name 'My Event'
hidden_name 'not shown'
+ location
starts_at DateTime.new(2012, 03, 02, 8, 15)
ends_at DateTime.new(2012, 03, 02, 19, 30)
lunch_starts_at DateTime.new(2012, 03, 02, 12, 15)
@@ -10,4 +11,8 @@ FactoryGirl.define do
published false
entrance_fee 12.3
end
+
+ factory :location do
+ name 'Awesome Lodge'
+ end
end
diff --git a/spec/features/admin/resources_integration_spec.rb b/spec/features/admin/resources_integration_spec.rb
index <HASH>..<HASH> 100644
--- a/spec/features/admin/resources_integration_spec.rb
+++ b/spec/features/admin/resources_integration_spec.rb
@@ -57,10 +57,13 @@ describe "Resources" do
describe "create resource item" do
context "when form filled with valid data" do
+ let!(:location) { create(:location) }
+
before do
visit '/admin/events/new'
fill_in 'event_name', with: 'My second event'
fill_in 'event_starts_at', with: DateTime.new(2012, 03, 03, 20, 00)
+ select location.name, from: 'Location'
click_on 'Save'
end | Fixes resource specs
The dummy apps `Event` model has a `belongs_to` `Location` relation that is mandatory since Rails <I>. This changes the specs so we always have a location for an event. | AlchemyCMS_alchemy_cms | train |
8a91baa090870aa3e70ea26c716ccde955ec1381 | diff --git a/setuptools/version.py b/setuptools/version.py
index <HASH>..<HASH> 100644
--- a/setuptools/version.py
+++ b/setuptools/version.py
@@ -1 +1 @@
-__version__ = '19.1.1'
+__version__ = '19.1.2' | Bumped to <I> in preparation for next release. | pypa_setuptools | train |
723a642648899cf4828222f307a5d0521c53bbcf | diff --git a/jquery.cytoscapeweb-panzoom.js b/jquery.cytoscapeweb-panzoom.js
index <HASH>..<HASH> 100644
--- a/jquery.cytoscapeweb-panzoom.js
+++ b/jquery.cytoscapeweb-panzoom.js
@@ -2,7 +2,7 @@
var defaults = {
zoomFactor: 0.05,
- zoomDelay: 50,
+ zoomDelay: 16,
minZoom: 0.1,
maxZoom: 10,
panSpeed: 10,
@@ -112,16 +112,16 @@
var handler = function(e){
e.stopPropagation(); // don't trigger dragging of panzoom
e.preventDefault(); // don't cause text selection
+ clearInterval(panInterval);
var pan = handle2pan(e);
if( isNaN(pan.x) || isNaN(pan.y) ){
+ $pIndicator.hide();
return;
}
positionIndicator(pan);
-
- clearInterval(panInterval);
panInterval = setInterval(function(){
$container.cytoscapeweb("get").panBy(pan);
}, options.panSpeed);
@@ -146,32 +146,51 @@
var sliderMax = 100;
var sliderMin = Math.floor( Math.log(options.minZoom)/Math.log(options.maxZoom) * sliderMax );
+ function getSliderVal(){
+ var $handle = $slider.find(".ui-slider-handle");
+ var left = $handle.position().left;
+ var width = $handle.parent().width();
+
+ var range = sliderMax - sliderMin;
+ var min = sliderMin;
+
+ return Math.round( left / width * range + min );
+ }
+
+ function setZoomViaSlider(){
+ var cy = $container.cytoscapeweb("get");
+ var val = getSliderVal();
+
+ var zoom = slider2zoom(val);
+
+ clearTimeout(sliderTimeout);
+ sliderTimeout = null;
+ cy.zoom({
+ level: zoom,
+ renderedPosition: {
+ x: $container.width()/2,
+ y: $container.height()/2
+ }
+ });
+ }
+
$slider.slider({
min: sliderMin,
max: sliderMax,
step: 1,
- val: zoom2slider( $container.cytoscapeweb("get").zoom() ),
- slide: function(){
- var cy = $container.cytoscapeweb("get");
- var val = $slider.slider("value");
- var zoom = slider2zoom(val);
-
- clearTimeout(sliderTimeout);
- sliderTimeout = null;
- cy.zoom({
- level: zoom,
- renderedPosition: {
- x: $container.width()/2,
- y: $container.height()/2
- }
- });
- }
+ val: zoom2slider( $container.cytoscapeweb("get").zoom() )
});
+ function sliderHandler(){
+ setZoomViaSlider();
+ };
+
var sliderMdown = false;
$slider.find(".ui-slider-handle").bind("mousedown", function(){
- sliderMdown = true;
+ sliderMdown = true;
+ $(window).bind("mousemove", sliderHandler);
}).bind("mouseup", function(){
+ $(window).unbind("mousemove", sliderHandler);
sliderMdown = false;
});
@@ -181,7 +200,6 @@
var sliderTimeout;
$container.cytoscapeweb("get").bind("zoom", function(){
-
if( sliderTimeout != null || sliderMdown ){
return;
} | Slider more responsive and behaves a bit better at bounds | cytoscape_cytoscape.js | train |
aa452e1b6fcb1d3ed309088126514fe23db4819e | diff --git a/conn/codec/constants.go b/conn/codec/constants.go
index <HASH>..<HASH> 100644
--- a/conn/codec/constants.go
+++ b/conn/codec/constants.go
@@ -25,7 +25,7 @@ import "errors"
// Codec constants.
const (
HeadLength = 4
- MaxPacketSize = 64 * 1024
+ MaxPacketSize = 1 << 24 //16MB
)
// ErrPacketSizeExcced is the error used for encode/decode.
diff --git a/conn/codec/pomelo_packet_decoder_test.go b/conn/codec/pomelo_packet_decoder_test.go
index <HASH>..<HASH> 100644
--- a/conn/codec/pomelo_packet_decoder_test.go
+++ b/conn/codec/pomelo_packet_decoder_test.go
@@ -19,12 +19,6 @@ var forwardTables = map[string]struct {
"test_kick_type": {[]byte{packet.Kick}, nil},
"test_wrong_packet_type": {[]byte{0x06}, packet.ErrWrongPomeloPacketType},
-
- "test_max_packet_size_handshake_type": {[]byte{packet.Handshake, 0xFF, 0xFF, 0xFF}, ErrPacketSizeExcced},
- "test_max_packet_size_handshake_ack_type": {[]byte{packet.HandshakeAck, 0xFF, 0xFF, 0xFF}, ErrPacketSizeExcced},
- "test_max_packet_size_heartbeat_type": {[]byte{packet.Heartbeat, 0xFF, 0xFF, 0xFF}, ErrPacketSizeExcced},
- "test_max_packet_size_data_type": {[]byte{packet.Data, 0xFF, 0xFF, 0xFF}, ErrPacketSizeExcced},
- "test_max_packet_size_kick_type": {[]byte{packet.Kick, 0xFF, 0xFF, 0xFF}, ErrPacketSizeExcced},
}
var (
diff --git a/conn/codec/pomelo_packet_encoder.go b/conn/codec/pomelo_packet_encoder.go
index <HASH>..<HASH> 100644
--- a/conn/codec/pomelo_packet_encoder.go
+++ b/conn/codec/pomelo_packet_encoder.go
@@ -44,6 +44,10 @@ func (e *PomeloPacketEncoder) Encode(typ packet.Type, data []byte) ([]byte, erro
return nil, packet.ErrWrongPomeloPacketType
}
+ if len(data) > MaxPacketSize {
+ return nil, ErrPacketSizeExcced
+ }
+
p := &packet.Packet{Type: typ, Length: len(data)}
buf := make([]byte, p.Length+HeadLength)
buf[0] = byte(p.Type)
diff --git a/conn/codec/pomelo_packet_encoder_test.go b/conn/codec/pomelo_packet_encoder_test.go
index <HASH>..<HASH> 100644
--- a/conn/codec/pomelo_packet_encoder_test.go
+++ b/conn/codec/pomelo_packet_encoder_test.go
@@ -19,6 +19,8 @@ func helperConcatBytes(packetType packet.Type, length, data []byte) []byte {
return bytes
}
+var tooBigData = make([]byte, 1<<25)
+
var encodeTables = map[string]struct {
packetType packet.Type
length []byte
@@ -27,6 +29,7 @@ var encodeTables = map[string]struct {
}{
"test_encode_handshake": {packet.Handshake, []byte{0x00, 0x00, 0x02}, []byte{0x01, 0x00}, nil},
"test_invalid_packet_type": {0xff, nil, nil, packet.ErrWrongPomeloPacketType},
+ "test_too_big_packet": {packet.Data, nil, tooBigData, ErrPacketSizeExcced},
}
func TestEncode(t *testing.T) {
@@ -37,10 +40,12 @@ func TestEncode(t *testing.T) {
ppe := NewPomeloPacketEncoder()
encoded, err := ppe.Encode(table.packetType, table.data)
-
- expectedEncoded := helperConcatBytes(table.packetType, table.length, table.data)
- assert.Equal(t, expectedEncoded, encoded)
- assert.Equal(t, table.err, err)
+ if table.err != nil {
+ assert.Equal(t, table.err, err)
+ } else {
+ expectedEncoded := helperConcatBytes(table.packetType, table.length, table.data)
+ assert.Equal(t, expectedEncoded, encoded)
+ }
})
}
} | increase max packet size to <I>mb | topfreegames_pitaya | train |
d4b694916a08384d146ab5349c3c5332b59c7f3d | diff --git a/bot/action/toggle.py b/bot/action/toggle.py
index <HASH>..<HASH> 100644
--- a/bot/action/toggle.py
+++ b/bot/action/toggle.py
@@ -1,19 +1,23 @@
-from bot.action.core.action import Action
+from bot.action.core.action import Action, IntermediateAction
from bot.api.domain import Message
-CHANGE_STATUS_VALUES = {
- "on": True,
- "off": False
+ON_VALUE = "on"
+OFF_VALUE = "off"
+
+STATUS_VALUES = {
+ ON_VALUE: True,
+ OFF_VALUE: False
}
class GetSetFeatureAction(Action):
def process(self, event):
feature, new_status = self.parse_args(event.command_args.split())
+ handler = FeatureStatusHandler(event, feature)
if feature is not None and new_status is not None:
- self.set_status(event, feature, new_status.lower())
+ self.set_status(handler, new_status.lower())
elif feature is not None:
- self.send_current_status(event, feature)
+ self.send_current_status(handler)
else:
self.send_usage(event)
@@ -27,25 +31,43 @@ class GetSetFeatureAction(Action):
new_status = args[1]
return feature, new_status
- def set_status(self, event, feature, new_status):
- if new_status in CHANGE_STATUS_VALUES:
- self.save_status(event, feature, new_status)
- self.send_current_status(event, feature, "Done!\n")
+ def set_status(self, handler, new_status):
+ if new_status in STATUS_VALUES:
+ handler.set_status(new_status)
+ self.send_current_status(handler, "Done!\n")
else:
- self.send_usage(event)
+ self.send_usage(handler.event)
- def send_current_status(self, event, feature, prepend=""):
- status = self.get_status(event, feature)
- response = prepend + "Current status of %s: %s" % (feature, status)
- self.api.send_message(Message.create_reply(event.message, response))
+ def send_current_status(self, handler, prepend=""):
+ status = handler.get_status_string()
+ response = prepend + "Current status of %s: %s" % (handler.feature, status)
+ self.api.send_message(Message.create_reply(handler.event.message, response))
def send_usage(self, event):
self.api.send_message(Message.create_reply(event.message, "Usage: " + event.command + " <feature> [on|off]"))
- @staticmethod
- def get_status(event, feature):
- return event.state.get_for("features").get_value(feature, "off")
- @staticmethod
- def save_status(event, feature, new_status):
- event.state.get_for("features").set_value(feature, new_status)
+class ToggleableFeatureAction(IntermediateAction):
+ def __init__(self, feature):
+ super().__init__()
+ self.feature = feature
+
+ def process(self, event):
+ if FeatureStatusHandler(event, self.feature).enabled:
+ self._continue(event)
+
+
+class FeatureStatusHandler:
+ def __init__(self, event, feature):
+ self.event = event
+ self.feature = feature
+
+ @property
+ def enabled(self):
+ return STATUS_VALUES[self.get_status_string()]
+
+ def get_status_string(self):
+ return self.event.state.get_for("features").get_value(self.feature, OFF_VALUE)
+
+ def set_status(self, new_status):
+ self.event.state.get_for("features").set_value(self.feature, new_status)
diff --git a/bot/manager.py b/bot/manager.py
index <HASH>..<HASH> 100644
--- a/bot/manager.py
+++ b/bot/manager.py
@@ -7,7 +7,7 @@ from bot.action.core.command import CommandAction
from bot.action.core.filter import MessageAction, TextMessageAction, NoPendingAction
from bot.action.gapdetector import GapDetectorAction
from bot.action.perchat import PerChatAction
-from bot.action.toggle import GetSetFeatureAction
+from bot.action.toggle import GetSetFeatureAction, ToggleableFeatureAction
from bot.bot import Bot
@@ -29,7 +29,10 @@ class BotManager:
MessageAction().then(
PerChatAction().then(
- PoleAction(),
+ ToggleableFeatureAction("pole").then(
+ PoleAction()
+ ),
+
TextMessageAction().then(
CommandAction("start").then(
AnswerAction( | Add ToggleableFeatureAction to toggle.py and make PoleAction toggleable | alvarogzp_telegram-bot-framework | train |
6387c520b724525861539c43b8da98372fc42534 | diff --git a/cmd/geth/snapshot.go b/cmd/geth/snapshot.go
index <HASH>..<HASH> 100644
--- a/cmd/geth/snapshot.go
+++ b/cmd/geth/snapshot.go
@@ -57,6 +57,7 @@ var (
Category: "MISCELLANEOUS COMMANDS",
Flags: []cli.Flag{
utils.DataDirFlag,
+ utils.AncientFlag,
utils.RopstenFlag,
utils.RinkebyFlag,
utils.GoerliFlag,
@@ -86,6 +87,7 @@ the trie clean cache with default directory will be deleted.
Category: "MISCELLANEOUS COMMANDS",
Flags: []cli.Flag{
utils.DataDirFlag,
+ utils.AncientFlag,
utils.RopstenFlag,
utils.RinkebyFlag,
utils.GoerliFlag,
@@ -105,6 +107,7 @@ In other words, this command does the snapshot to trie conversion.
Category: "MISCELLANEOUS COMMANDS",
Flags: []cli.Flag{
utils.DataDirFlag,
+ utils.AncientFlag,
utils.RopstenFlag,
utils.RinkebyFlag,
utils.GoerliFlag,
@@ -126,6 +129,7 @@ It's also usable without snapshot enabled.
Category: "MISCELLANEOUS COMMANDS",
Flags: []cli.Flag{
utils.DataDirFlag,
+ utils.AncientFlag,
utils.RopstenFlag,
utils.RinkebyFlag,
utils.GoerliFlag, | cmd/geth: add ancient datadir flag to snapshot subcommands (#<I>) | ethereum_go-ethereum | train |
314aae1c345635f3ec1f67c842a38e19ee49b01e | diff --git a/src/components/chips/js/chipsController.js b/src/components/chips/js/chipsController.js
index <HASH>..<HASH> 100644
--- a/src/components/chips/js/chipsController.js
+++ b/src/components/chips/js/chipsController.js
@@ -329,6 +329,7 @@
var scope = this.$scope;
var ctrl = this;
inputElement
+ .attr({ tabindex: 0 })
.on('keydown', function(event) { scope.$apply(function() { ctrl.inputKeydown(event); }); })
.on('focus', function () { scope.$apply(function () { ctrl.selectedChip = null; }); });
};
diff --git a/src/components/chips/js/chipsDirective.js b/src/components/chips/js/chipsDirective.js
index <HASH>..<HASH> 100644
--- a/src/components/chips/js/chipsDirective.js
+++ b/src/components/chips/js/chipsDirective.js
@@ -114,6 +114,7 @@
var CHIP_INPUT_TEMPLATE = '\
<input\
+ tabindex="0"\
placeholder="{{$mdChipsCtrl.getPlaceholder()}}"\
aria-label="{{$mdChipsCtrl.getPlaceholder()}}"\
ng-model="$mdChipsCtrl.chipBuffer"\
@@ -147,7 +148,6 @@
// element propagates to the link function via the attrs argument,
// where various contained-elements can be consumed.
attrs['$mdUserTemplate'] = element.clone();
- attrs['tabindex'] = '0';
return MD_CHIPS_TEMPLATE;
},
require: ['mdChips'],
@@ -219,13 +219,14 @@
*/
return function postLink(scope, element, attrs, controllers) {
$mdTheming(element);
- element.attr('tabindex', '0');
var mdChipsCtrl = controllers[0];
mdChipsCtrl.chipContentsTemplate = chipContentsTemplate;
mdChipsCtrl.chipRemoveTemplate = chipRemoveTemplate;
mdChipsCtrl.chipInputTemplate = chipInputTemplate;
- element.on('focus', function () { mdChipsCtrl.onFocus(); });
+ element
+ .attr({ ariaHidden: true, tabindex: -1 })
+ .on('focus', function () { mdChipsCtrl.onFocus(); });
if (attr.ngModel) {
mdChipsCtrl.configureNgModel(element.controller('ngModel')); | fix(chips): fixes shift+tab interaction with `md-chips` | angular_material | train |
c91b8ce8851a87634f0417e5c07964ef80a11a17 | diff --git a/angr/analyses/backward_slice.py b/angr/analyses/backward_slice.py
index <HASH>..<HASH> 100644
--- a/angr/analyses/backward_slice.py
+++ b/angr/analyses/backward_slice.py
@@ -390,10 +390,10 @@ class BackwardSlice(Analysis):
# Pick all its data dependencies from data dependency graph
if self._ddg is not None and tainted_cl in self._ddg:
if isinstance(self._ddg, networkx.DiGraph):
- predecessors = self._ddg.predecessors(tainted_cl)
+ predecessors = list(self._ddg.predecessors(tainted_cl))
else:
# angr.analyses.DDG
- predecessors = self._ddg.get_predecessors(tainted_cl)
+ predecessors = list(self._ddg.get_predecessors(tainted_cl))
l.debug("Returned %d predecessors for %s from data dependence graph", len(predecessors), tainted_cl)
for p in predecessors: | backwardslice: convert networkx2 iterators to lists; closes #<I> | angr_angr | train |
Subsets and Splits