hash
stringlengths 40
40
| diff
stringlengths 131
114k
| message
stringlengths 7
980
| project
stringlengths 5
67
| split
stringclasses 1
value |
---|---|---|---|---|
62857cfa45e65fd31d6f90532a3a1988a24ed2b0 | diff --git a/ndio/remote/boss/tests/int_test_group.py b/ndio/remote/boss/tests/int_test_group.py
index <HASH>..<HASH> 100644
--- a/ndio/remote/boss/tests/int_test_group.py
+++ b/ndio/remote/boss/tests/int_test_group.py
@@ -59,8 +59,6 @@ class ProjectGroupTest(unittest.TestCase):
self.existing_grp_name = 'int_test_exists'
self.user_name = 'bossadmin'
- self.rmt.group_create(self.existing_grp_name)
-
def cleanup_db(self):
"""Clean up the data model objects used by this test case.
@@ -71,6 +69,7 @@ class ProjectGroupTest(unittest.TestCase):
def setUp(self):
self.initialize()
+ self.rmt.group_create(self.existing_grp_name)
def tearDown(self):
self.cleanup_db() | Moved creation of test group.
No need to create test group in setUpClass(). | jhuapl-boss_intern | train |
b5a2e2cf032e47d1e0dfd752662a4d4068235847 | diff --git a/src/Viserio/Connect/Traits/DetectsLostConnections.php b/src/Viserio/Connect/Traits/DetectsLostConnections.php
index <HASH>..<HASH> 100644
--- a/src/Viserio/Connect/Traits/DetectsLostConnections.php
+++ b/src/Viserio/Connect/Traits/DetectsLostConnections.php
@@ -24,6 +24,7 @@ trait DetectsLostConnections
'is dead or not enabled',
'Error while sending',
'Operation in progress',
+ 'decryption failed or bad record mac',
]);
}
} | Added another lost connection case for pgsql | narrowspark_framework | train |
1cdd38d3c0bed524ac4f48eccf70080b6e54d9e5 | diff --git a/lib/NormalModuleFactory.js b/lib/NormalModuleFactory.js
index <HASH>..<HASH> 100644
--- a/lib/NormalModuleFactory.js
+++ b/lib/NormalModuleFactory.js
@@ -475,22 +475,23 @@ class NormalModuleFactory extends ModuleFactory {
return continueCallback();
}
- normalResolver.resolve(
- contextInfo,
- context,
- unresolvedResource,
- resolveContext,
- (err, resolvedResource, resolvedResourceResolveData) => {
- if (err) {
- if (!isDataURI(unresolvedResource)) return continueCallback(err);
-
- resolvedResource = unresolvedResource;
+ if (isDataURI(unresolvedResource)) {
+ resource = unresolvedResource;
+ continueCallback();
+ } else {
+ normalResolver.resolve(
+ contextInfo,
+ context,
+ unresolvedResource,
+ resolveContext,
+ (err, resolvedResource, resolvedResourceResolveData) => {
+ if (err) return continueCallback(err);
+ resource = resolvedResource;
+ resourceResolveData = resolvedResourceResolveData;
+ continueCallback();
}
- resource = resolvedResource;
- resourceResolveData = resolvedResourceResolveData;
- continueCallback();
- }
- );
+ );
+ }
}
);
} | check for data-uri before resource resolving | webpack_webpack | train |
724c7f63e42cbfd154bcf16f759608d9b18722a2 | diff --git a/src/diamond/collector.py b/src/diamond/collector.py
index <HASH>..<HASH> 100644
--- a/src/diamond/collector.py
+++ b/src/diamond/collector.py
@@ -481,8 +481,6 @@ class Collector(object):
metric_name = 'collector_time_ms'
metric_value = collector_time
self.publish(metric_name, metric_value)
- except Exception, e:
- self.log.exception(e)
finally:
# After collector run, invoke a flush
# method on each handler. | Revert <I>f<I> collector as exceptions are printed by the scheduler thread
Handling the exception in the collector code actually results in signal
exceptions not being passed to the scheduler, thus breaking functionality | python-diamond_Diamond | train |
d805fe21d15f90ebd8846a11bc7f73edd16d771e | diff --git a/lib/active_interaction/version.rb b/lib/active_interaction/version.rb
index <HASH>..<HASH> 100644
--- a/lib/active_interaction/version.rb
+++ b/lib/active_interaction/version.rb
@@ -6,5 +6,5 @@ module ActiveInteraction
# The version number.
#
# @return [Gem::Version]
- VERSION = Gem::Version.new('3.0.1')
+ VERSION = Gem::Version.new('3.1.0')
end | :briefcase: <I> | AaronLasseigne_active_interaction | train |
8108a9bfe2867cc4db79cccecd8b752e185aed0b | diff --git a/http/logical.go b/http/logical.go
index <HASH>..<HASH> 100644
--- a/http/logical.go
+++ b/http/logical.go
@@ -158,8 +158,10 @@ func handleLogicalInternal(core *vault.Core, injectDataIntoTopLevel bool) http.H
// Always forward requests that are using a limited use count token.
// origBody will not be nil if it's a perf standby as it checks
// PerfStandby() but will be nil otherwise.
- if origBody != nil && req.ClientTokenRemainingUses > 0 {
- r.Body = origBody
+ if core.PerfStandby() && req.ClientTokenRemainingUses > 0 {
+ if origBody != nil {
+ r.Body = origBody
+ }
forwardRequest(core, w, r)
return
}
@@ -275,7 +277,9 @@ func handleLogicalInternal(core *vault.Core, injectDataIntoTopLevel bool) http.H
// success.
resp, ok, needsForward := request(core, w, r, req)
if needsForward {
- r.Body = origBody
+ if origBody != nil {
+ r.Body = origBody
+ }
forwardRequest(core, w, r)
return
} | Fix a potential panic due to ioutil.ReadAll not being safe with nil readers | hashicorp_vault | train |
4a72c441d583557282df06a2252e7c06ac83f868 | diff --git a/src/Symfony/Component/DependencyInjection/ParameterBag/FrozenParameterBag.php b/src/Symfony/Component/DependencyInjection/ParameterBag/FrozenParameterBag.php
index <HASH>..<HASH> 100644
--- a/src/Symfony/Component/DependencyInjection/ParameterBag/FrozenParameterBag.php
+++ b/src/Symfony/Component/DependencyInjection/ParameterBag/FrozenParameterBag.php
@@ -69,4 +69,14 @@ class FrozenParameterBag extends ParameterBag
{
throw new LogicException('Impossible to call set() on a frozen ParameterBag.');
}
+
+ /**
+ * {@inheritdoc}
+ *
+ * @api
+ */
+ public function remove($name)
+ {
+ throw new LogicException('Impossible to call remove() on a frozen ParameterBag.');
+ }
} | [DependencyInjection] Freeze also FrozenParameterBag::remove | symfony_symfony | train |
7957a6a48a68cfc05c2b2f3109f653f0bf0754bc | diff --git a/test/RelationshipGraph.js b/test/RelationshipGraph.js
index <HASH>..<HASH> 100644
--- a/test/RelationshipGraph.js
+++ b/test/RelationshipGraph.js
@@ -146,7 +146,7 @@ describe('RelationshipGraph', function() {
});
});
- describe('#ValidateVerifyJSON()' , function() {
+ describe('#ValidateVerifyJSON()', function() {
var graph = d3.select('#test2').relationshipGraph({
'thresholds': [200]
});
@@ -463,8 +463,7 @@ describe('RelationshipGraph', function() {
'parent': 'Warner Bros. Pictures',
'value': '$958,366,855',
'Year': '2013'
- }
- ,
+ },
{
'Movie Title': 'The Hobbit: The Battle of the Five Armies',
'parent': 'Warner Bros. Pictures',
@@ -627,7 +626,7 @@ describe('RelationshipGraph', function() {
var expectedText = ['20th Century Fox (6)', 'Columbia Pictures (3)', 'Lionsgate Films (1)', 'New Line Cinema (3)',
'Paramount Pictures (1)', 'Universal Pictures (7)', 'Walt Disney Studios (16)', 'Warner Bros. Pictures (13)'],
expectedY = [0, 24, 48, 72, 96, 120, 144, 192]//,
- //expectedX = [41, 35, 50, 38, 26, 36, 18, 7];
+ //expectedX = [41, 35, 50, 38, 26, 36, 18, 7];
chai.expect(text.length).to.equal(expectedText.length);
@@ -685,4 +684,76 @@ describe('RelationshipGraph', function() {
});
});
+
+ describe('#VerifySortJson', function() {
+ it('Should be sorted.', function() {
+ var json = [
+ {
+ 'Movie Title': 'Avatar',
+ 'parent': '20th Century Fox',
+ 'value': '$2,787,965,087',
+ 'Year': '2009'
+ },
+ {
+ 'Movie Title': 'Star Wars: The Force Awakens',
+ 'parent': 'Walt Disney Studios',
+ 'value': '$2,066,247,462',
+ 'Year': '2015'
+ },
+ {
+ 'Movie Title': 'Titanic',
+ 'parent': '20th Century Fox',
+ 'value': '$2,186,772,302',
+ 'Year': '1997'
+ }
+ ], expected = [
+ {
+ 'Movie Title': 'Avatar',
+ 'parent': '20th Century Fox',
+ 'value': '$2,787,965,087',
+ 'Year': '2009'
+ },
+ {
+ 'Movie Title': 'Titanic',
+ 'parent': '20th Century Fox',
+ 'value': '$2,186,772,302',
+ 'Year': '1997'
+ },
+ {
+ 'Movie Title': 'Star Wars: The Force Awakens',
+ 'parent': 'Walt Disney Studios',
+ 'value': '$2,066,247,462',
+ 'Year': '2015'
+ }
+ ];
+
+ json.sort(function(child1, child2) {
+ var parent1 = child1.parent.toLowerCase(),
+ parent2 = child2.parent.toLowerCase();
+
+ return (parent1 > parent2) ? 1 : (parent1 < parent2) ? -1 : 0;
+ });
+
+ for (var i = 0; i < json.length; i++) {
+ chai.expect(json[i]['Movie Title']).to.equal(expected[i]['Movie Title']);
+ chai.expect(json[i].parent).to.equal(expected[i].parent);
+ chai.expect(json[i].value).to.equal(expected[i].value);
+ chai.expect(json[i].year).to.equal(expected[i].year);
+ }
+ });
+ });
+
+ describe('#VerifyCustomColors', function() {
+ it('Should have the custom color set.', function() {
+ var custom = ['red', 'green', 'blue'];
+
+ var graph = d3.select('#graph').relationshipGraph({
+ colors: custom
+ });
+
+ for (var i = 0; i < custom.length; i++) {
+ chai.expect(graph.graph.config.colors[i]).to.equal(custom[i]);
+ }
+ });
+ });
}); | Added tests for sorting the JSON and adding custom colors. | hkelly93_d3-relationshipgraph | train |
c0bf0520bb1b80cf3cddaa869d681da6742fca42 | diff --git a/container/api/src/test/java/org/wildfly/swarm/container/ProjectStageTest.java b/container/api/src/test/java/org/wildfly/swarm/container/ProjectStageTest.java
index <HASH>..<HASH> 100644
--- a/container/api/src/test/java/org/wildfly/swarm/container/ProjectStageTest.java
+++ b/container/api/src/test/java/org/wildfly/swarm/container/ProjectStageTest.java
@@ -3,7 +3,7 @@ package org.wildfly.swarm.container;
import java.io.InputStream;
import java.util.List;
-import junit.framework.Assert;
+import org.junit.Assert;
import org.junit.Test;
import org.wildfly.swarm.container.internal.ProjectStageFactory;
import org.wildfly.swarm.spi.api.ProjectStage;
diff --git a/integration-tests/src/test/java/org/wildfly/swarm/integration/SWARM_513/SWARM_513Test.java b/integration-tests/src/test/java/org/wildfly/swarm/integration/SWARM_513/SWARM_513Test.java
index <HASH>..<HASH> 100644
--- a/integration-tests/src/test/java/org/wildfly/swarm/integration/SWARM_513/SWARM_513Test.java
+++ b/integration-tests/src/test/java/org/wildfly/swarm/integration/SWARM_513/SWARM_513Test.java
@@ -5,7 +5,7 @@ import com.squareup.okhttp.OkHttpClient;
import com.squareup.okhttp.Request;
import com.squareup.okhttp.Response;
import com.squareup.okhttp.RequestBody;
-import junit.framework.Assert;
+import org.junit.Assert;
import org.jboss.arquillian.container.test.api.Deployment;
import org.jboss.arquillian.container.test.api.RunAsClient;
import org.jboss.shrinkwrap.api.Archive;
diff --git a/monitor/api/src/test/java/org/wildfly/swarm/monitor/MonitorTest.java b/monitor/api/src/test/java/org/wildfly/swarm/monitor/MonitorTest.java
index <HASH>..<HASH> 100644
--- a/monitor/api/src/test/java/org/wildfly/swarm/monitor/MonitorTest.java
+++ b/monitor/api/src/test/java/org/wildfly/swarm/monitor/MonitorTest.java
@@ -15,7 +15,7 @@
*/
package org.wildfly.swarm.monitor;
-import junit.framework.Assert;
+import org.junit.Assert;
import org.junit.Test;
/** | junit.framework is deprecated | thorntail_thorntail | train |
456f58997cd380970da1e6ea901b4b4d987c4d9f | diff --git a/CHANGELOG.md b/CHANGELOG.md
index <HASH>..<HASH> 100644
--- a/CHANGELOG.md
+++ b/CHANGELOG.md
@@ -23,6 +23,10 @@
- might be a good idea... think on it...
- DOING: update docs
+## 2.0.0a20 (released on 21.01.2020)
+- Fixed UnicodeEncodeError: 'charmap' codec
+ - thanks to [PR-197](https://github.com/yashaka/selene/pull/197) from @ak40u
+
## 2.0.0a19 (released on 16.01.2020)
- removed deprecation from shared.config.counter and reports_folder
- removed backports.functools-lru-cache from project dependencies
diff --git a/selene/__init__.py b/selene/__init__.py
index <HASH>..<HASH> 100644
--- a/selene/__init__.py
+++ b/selene/__init__.py
@@ -211,7 +211,7 @@ Or, by using non-waiting versions, if "you are in a rush:)"::
# """
# todo: add here some type imports like Element, Collection, etc.
-__version__ = '2.0.0a19'
+__version__ = '2.0.0a20'
# --- DEPRECATED, and will be removed soon --- #
diff --git a/selene/support/shared/config.py b/selene/support/shared/config.py
index <HASH>..<HASH> 100644
--- a/selene/support/shared/config.py
+++ b/selene/support/shared/config.py
@@ -126,10 +126,19 @@ class SharedConfig(Config):
stored.quit() # todo: can this raise exception? that we need to supress...
# todo: do we need here pass self.desired_capabilities too?
+
+ set_chrome = lambda: Chrome(
+ executable_path=ChromeDriverManager().install(),
+ options=ChromeOptions())
+
+ set_firefox = lambda: Firefox(
+ executable_path=GeckoDriverManager().install())
+
+ # set_remote = lambda: Remote() # todo: do we really need it? :)
+
new = {
- 'chrome': lambda: Chrome(executable_path=ChromeDriverManager().install(),
- options=ChromeOptions()),
- 'firefox': lambda: Firefox(executable_path=GeckoDriverManager().install())
+ 'chrome': set_chrome,
+ 'firefox': set_firefox
}.get(self.browser_name)()
# todo: think on something like: | bumped <I>a<I>; tiny refactoring of config | yashaka_selene | train |
612e603aa787a76ae77c7e6c34547a70c84aa987 | diff --git a/client/html/templates/common/partials/selection-default.php b/client/html/templates/common/partials/selection-default.php
index <HASH>..<HASH> 100644
--- a/client/html/templates/common/partials/selection-default.php
+++ b/client/html/templates/common/partials/selection-default.php
@@ -18,12 +18,25 @@ $attributes = $this->get( 'selectionAttributeItems', array() );
* default. This setting removes the hint to select an option, so the first one
* is selected by default.
*
+ * The key for each value must be the type code of the attribute, e.g. "width",
+ * "length", "color" or similar types. You can set the layout for all
+ * attributes at once using e.g.
+ *
+ * client/html/catalog/detail/basket/selection/preselect = array(
+ * 'width' => false,
+ * 'color' => true,
+ * )
+ *
+ * Similarly, you can set the pre-selection for a specific attribute only,
+ * leaving the rest untouched:
+ *
+ * client/html/catalog/detail/basket/selection/preselect/color = true
+ *
* @param boolean True to select the first option by default, false to display the select hint
- * @since 2016.05
+ * @since 2016.07
* @category Developer
* @category User
*/
-$preselect = (bool) $this->config( 'client/html/catalog/detail/basket/selection/preselect', false );
/** client/html/catalog/detail/basket/selection/type
* List of layout types for the variant attributes
@@ -81,15 +94,16 @@ $preselect = (bool) $this->config( 'client/html/catalog/detail/basket/selection/
<ul class="selection">
<?php foreach( $this->get( 'selectionAttributeTypeDependencies', array() ) as $code => $attrIds ) : asort( $attrIds ); ?>
<?php $layout = $this->config( 'client/html/catalog/detail/basket/selection/type/' . $code, 'select' ); ?>
- <li class="select-item <?php echo $enc->attr( $layout ) . ' ' . $enc->attr( $code ); ?>">
+<?php $preselect = (bool) $this->config( 'client/html/catalog/detail/basket/selection/preselect/' . $code, false ); ?>
+<li class="select-item <?php echo $enc->attr( $layout ) . ' ' . $enc->attr( $code ); ?>">
<div class="select-name"><?php echo $enc->html( $this->translate( 'client/code', $code ) ); ?></div>
<div class="select-value">
-<?php if( $layout === 'radio' ) : ?>
+<?php if( $layout === 'radio' ) : $first = true; ?>
<ul class="select-list" data-index="<?php echo $index++; ?>">
<?php foreach( $attrIds as $attrId => $position ) : ?>
<?php if( isset( $attributes[$attrId] ) ) : ?>
<li class="select-entry">
- <input class="select-option" id="option-<?php echo $enc->attr( $attrId ); ?>" name="<?php echo $enc->attr( $this->formparam( array( 'b_prod', 0, 'attrvarid', $code ) ) ); ?>" type="radio" value="<?php echo $enc->attr( $attrId ); ?>" />
+ <input class="select-option" id="option-<?php echo $enc->attr( $attrId ); ?>" name="<?php echo $enc->attr( $this->formparam( array( 'b_prod', 0, 'attrvarid', $code ) ) ); ?>" type="radio" value="<?php echo $enc->attr( $attrId ); ?>" <?php echo ( $preselect && $first ? 'checked="checked"' : '' ); $first = false ?>/>
<label class="select-label" for="option-<?php echo $enc->attr( $attrId ); ?>"><!--
<?php foreach( $attributes[$attrId]->getListItems( 'media', 'icon' ) as $listItem ) : ?>
<?php if( ( $item = $listItem->getRefItem() ) !== null ) : ?> | Improved pre-selection of product variant attributes | aimeos_ai-client-html | train |
8825d7edf310ce182d3bef4d70d63bc2658218bd | diff --git a/go/vt/vtgate/vstream_manager_test.go b/go/vt/vtgate/vstream_manager_test.go
index <HASH>..<HASH> 100644
--- a/go/vt/vtgate/vstream_manager_test.go
+++ b/go/vt/vtgate/vstream_manager_test.go
@@ -48,7 +48,6 @@ import (
var mu sync.Mutex
func TestVStreamSkew(t *testing.T) {
- t.Skip()
stream := func(conn *sandboxconn.SandboxConn, shard string, count, idx int64) {
vevents := getVEvents(shard, count, idx)
for _, ev := range vevents {
@@ -123,7 +122,6 @@ func TestVStreamSkew(t *testing.T) {
}
func TestVStreamEvents(t *testing.T) {
- t.Skip()
ctx, cancel := context.WithCancel(context.Background())
defer cancel()
cell := "aa"
@@ -276,7 +274,6 @@ func TestVStreamChunks(t *testing.T) {
}
func TestVStreamMulti(t *testing.T) {
- t.Skip()
ctx, cancel := context.WithCancel(context.Background())
defer cancel()
cell := "aa"
@@ -382,7 +379,6 @@ func TestVStreamRetry(t *testing.T) {
}
func TestVStreamShouldNotSendSourceHeartbeats(t *testing.T) {
- t.Skip()
ctx, cancel := context.WithCancel(context.Background())
defer cancel()
cell := "aa"
@@ -433,7 +429,6 @@ func TestVStreamShouldNotSendSourceHeartbeats(t *testing.T) {
}
func TestVStreamJournalOneToMany(t *testing.T) {
- t.Skip()
ctx, cancel := context.WithCancel(context.Background())
defer cancel()
cell := "aa"
@@ -542,7 +537,6 @@ func TestVStreamJournalOneToMany(t *testing.T) {
}
func TestVStreamJournalManyToOne(t *testing.T) {
- t.Skip()
ctx, cancel := context.WithCancel(context.Background())
defer cancel()
@@ -658,7 +652,6 @@ func TestVStreamJournalManyToOne(t *testing.T) {
}
func TestVStreamJournalNoMatch(t *testing.T) {
- t.Skip()
ctx, cancel := context.WithCancel(context.Background())
defer cancel()
@@ -787,7 +780,6 @@ func TestVStreamJournalNoMatch(t *testing.T) {
}
func TestVStreamJournalPartialMatch(t *testing.T) {
- t.Skip()
ctx, cancel := context.WithCancel(context.Background())
defer cancel()
@@ -1018,7 +1010,6 @@ func TestResolveVStreamParams(t *testing.T) {
}
func TestVStreamIdleHeartbeat(t *testing.T) {
- t.Skip()
cell := "aa"
ks := "TestVStream"
_ = createSandbox(ks) | vstream tests: disable only the consistently failing tests | vitessio_vitess | train |
2bb82be1e555a6ccb1f55365a369dd052017c903 | diff --git a/reordering.py b/reordering.py
index <HASH>..<HASH> 100644
--- a/reordering.py
+++ b/reordering.py
@@ -25,7 +25,6 @@ def ravel(a, order='C'):
return reshape(a=a, newshape=(-1,))
-
class Reorder(Permute):
dterms = 'a',
@@ -38,17 +37,21 @@ class Reorder(Permute):
def compute_dr_wrt(self, wrt):
if wrt is self.a:
- a = self.a
- asz = a.size
- ashape = a.shape
- key = self.unique_reorder_id()
- if key not in self.dr_lookup or key is None:
- JS = self.reorder(np.arange(asz).reshape(ashape))
- IS = np.arange(JS.size)
- data = np.ones_like(IS)
- shape = JS.shape
- self.dr_lookup[key] = sp.csc_matrix((data, (IS, JS.ravel())), shape=(self.r.size, wrt.r.size))
- return self.dr_lookup[key]
+ if True:
+ from scipy.sparse.linalg.interface import LinearOperator
+ return LinearOperator((self.size, wrt.size), lambda x : self.reorder(x.reshape(self.a.shape)).ravel())
+ else:
+ a = self.a
+ asz = a.size
+ ashape = a.shape
+ key = self.unique_reorder_id()
+ if key not in self.dr_lookup or key is None:
+ JS = self.reorder(np.arange(asz).reshape(ashape))
+ IS = np.arange(JS.size)
+ data = np.ones_like(IS)
+ shape = JS.shape
+ self.dr_lookup[key] = sp.csc_matrix((data, (IS, JS.ravel())), shape=(self.r.size, wrt.r.size))
+ return self.dr_lookup[key]
class Sort(Reorder):
dterms = 'a' | reordering.py: make more use of LinearOperator in reordering | mattloper_chumpy | train |
e99f49c2dba565595ebb5b24ccc46f847ec20257 | diff --git a/slacker/__init__.py b/slacker/__init__.py
index <HASH>..<HASH> 100644
--- a/slacker/__init__.py
+++ b/slacker/__init__.py
@@ -85,6 +85,12 @@ class Users(BaseAPI):
class Groups(BaseAPI):
+ def create(self, name):
+ return self.post('groups.create', params={'name': name})
+
+ def create_child(self, channel):
+ return self.post('groups.createChild', params={'channel': channel})
+
def list(self, exclude_archived=None):
return self.get('groups.list',
params={'exclude_archived': exclude_archived}) | Add groups.create and groups.createChild APIs. | os_slacker | train |
412b03f243e93a598073bf4cda6b65084e0d0d74 | diff --git a/spec/lib/pg_search/multisearch/rebuilder_spec.rb b/spec/lib/pg_search/multisearch/rebuilder_spec.rb
index <HASH>..<HASH> 100644
--- a/spec/lib/pg_search/multisearch/rebuilder_spec.rb
+++ b/spec/lib/pg_search/multisearch/rebuilder_spec.rb
@@ -84,10 +84,15 @@ describe PgSearch::Multisearch::Rebuilder do
time = DateTime.parse("2001-01-01")
rebuilder = PgSearch::Multisearch::Rebuilder.new(Model, lambda { time } )
- # Handle change in precision of DateTime objects in SQL in Rails 4.1
+ # Handle change in precision of DateTime objects in SQL in Active Record 4.1
# https://github.com/rails/rails/commit/17f5d8e062909f1fcae25351834d8e89967b645e
+ version_4_1_or_newer = (
+ ActiveRecord::VERSION::MAJOR > 4 ||
+ ActiveRecord::VERSION::MAJOR = 4 && ActiveRecord::VERSION::MINOR >= 1
+ )
+
expected_timestamp =
- if ActiveRecord::VERSION::MAJOR >= 4 && ActiveRecord::VERSION::MINOR >= 1
+ if version_4_1_or_newer
"2001-01-01 00:00:00.000000"
else
"2001-01-01 00:00:00" | Future-proof test for Active Record <I>+ | Casecommons_pg_search | train |
04272a2d5be800d8a11b0b27d49d517330a8b6c8 | diff --git a/lib/airrecord/table.rb b/lib/airrecord/table.rb
index <HASH>..<HASH> 100644
--- a/lib/airrecord/table.rb
+++ b/lib/airrecord/table.rb
@@ -125,7 +125,6 @@ module Airrecord
value = fields[key]
elsif column_mappings[key]
deprecate_symbols if key.is_a? Symbol
- deprecate_symbols
value = fields[column_mappings[key]]
end
@@ -151,7 +150,7 @@ module Airrecord
elsif column_mappings[key]
deprecate_symbols
return if fields[column_mappings[key]] == value # no-op
- @updated_keys << key
+ @updated_keys << column_mappings[key]
fields[column_mappings[key]] = value
else
@updated_keys << key | revert code change, remove duplicate code | sirupsen_airrecord | train |
0377c50b714c245b190384e99705d9394924c609 | diff --git a/wayback-core/src/main/java/org/archive/wayback/resourcestore/FlexResourceStore.java b/wayback-core/src/main/java/org/archive/wayback/resourcestore/FlexResourceStore.java
index <HASH>..<HASH> 100644
--- a/wayback-core/src/main/java/org/archive/wayback/resourcestore/FlexResourceStore.java
+++ b/wayback-core/src/main/java/org/archive/wayback/resourcestore/FlexResourceStore.java
@@ -17,7 +17,6 @@ import org.archive.io.warc.WARCReader;
import org.archive.io.warc.WARCRecord;
import org.archive.util.binsearch.SeekableLineReader;
import org.archive.util.binsearch.SortedTextFile;
-import org.archive.util.binsearch.impl.HTTPSeekableLineReader;
import org.archive.util.iterator.CloseableIterator;
import org.archive.wayback.ResourceStore;
import org.archive.wayback.core.CaptureSearchResult;
@@ -25,7 +24,6 @@ import org.archive.wayback.core.Resource;
import org.archive.wayback.exception.ResourceNotAvailableException;
import org.archive.wayback.resourcestore.resourcefile.ArcResource;
import org.archive.wayback.resourcestore.resourcefile.WarcResource;
-import org.archive.wayback.webapp.PerformanceLogger;
public class FlexResourceStore implements ResourceStore {
@@ -223,8 +221,7 @@ public class FlexResourceStore implements ResourceStore {
}
public Resource getResource(String path, CaptureSearchResult result) throws IOException
- {
- long start = System.currentTimeMillis();
+ {
Resource r = null;
long offset = result.getOffset();
@@ -255,27 +252,31 @@ public class FlexResourceStore implements ResourceStore {
r.parseHeaders();
- if ((customHeader != null) && (slr instanceof HTTPSeekableLineReader)) {
- HTTPSeekableLineReader httpSlr = (HTTPSeekableLineReader)slr;
- String value = httpSlr.getHeaderValue(customHeader);
-
- if (value != null) {
- result.put(CaptureSearchResult.CUSTOM_HEADER_PREFIX + customHeader, value);
- }
- }
+ // Ability to pass on header prefix from data store request
+// if ((customHeader != null) && (slr instanceof HTTPSeekableLineReader)) {
+// HTTPSeekableLineReader httpSlr = (HTTPSeekableLineReader)slr;
+// String value = httpSlr.getHeaderValue(customHeader);
+//
+// if (value != null) {
+// result.put(CaptureSearchResult.CUSTOM_HEADER_PREFIX + customHeader, value);
+// }
+// }
- } catch (IOException io) {
+ } catch (Exception e) {
if (slr != null) {
slr.close();
}
+
r = null;
slr = null;
- throw io;
+
+ if (e instanceof IOException) {
+ throw (IOException)e;
+ } else {
+ throw new IOException(e);
+ }
}
-
- long elapsed = System.currentTimeMillis() - start;
- PerformanceLogger.noteElapsed("{W}arcBlockLoader", elapsed, path);
-
+
return r;
} | FIX: FlexResourceStore catchs all exceptions, including runtime and ensures Resource is closed if its not returned | iipc_openwayback | train |
165d0f89188a3de20dac61f3395b039550c8a243 | diff --git a/sonar-ws-client/src/main/java/org/sonar/wsclient/issue/BulkChangeQuery.java b/sonar-ws-client/src/main/java/org/sonar/wsclient/issue/BulkChangeQuery.java
index <HASH>..<HASH> 100644
--- a/sonar-ws-client/src/main/java/org/sonar/wsclient/issue/BulkChangeQuery.java
+++ b/sonar-ws-client/src/main/java/org/sonar/wsclient/issue/BulkChangeQuery.java
@@ -63,6 +63,11 @@ public class BulkChangeQuery {
return this;
}
+ public BulkChangeQuery sendNotifications(boolean sendNotifications) {
+ params.put("sendNotifications", String.valueOf(sendNotifications));
+ return this;
+ }
+
private BulkChangeQuery addParam(String key, String[] values) {
if (values != null) {
params.put(key, EncodingUtils.toQueryParam(values));
diff --git a/sonar-ws-client/src/test/java/org/sonar/wsclient/issue/BulkChangeQueryTest.java b/sonar-ws-client/src/test/java/org/sonar/wsclient/issue/BulkChangeQueryTest.java
index <HASH>..<HASH> 100644
--- a/sonar-ws-client/src/test/java/org/sonar/wsclient/issue/BulkChangeQueryTest.java
+++ b/sonar-ws-client/src/test/java/org/sonar/wsclient/issue/BulkChangeQueryTest.java
@@ -40,14 +40,15 @@ public class BulkChangeQueryTest {
.actions("assign")
.actionParameter("assign", "assignee", "geoffrey")
.comment("My bulk comment")
- ;
+ .sendNotifications(false);
- assertThat(query.urlParams()).hasSize(4).includes(
+ assertThat(query.urlParams()).hasSize(5).includes(
entry("issues", "ABCD,EFGH"),
entry("actions", "assign"),
entry("assign.assignee", "geoffrey"),
- entry("comment", "My bulk comment")
- );
+ entry("comment", "My bulk comment"),
+ entry("sendNotifications", "false")
+ );
}
@Test | SONAR-<I> Update issue client | SonarSource_sonarqube | train |
d92352d05d8601d643feeb0d58201f6e499348c0 | diff --git a/tests/src/TestBase.php b/tests/src/TestBase.php
index <HASH>..<HASH> 100644
--- a/tests/src/TestBase.php
+++ b/tests/src/TestBase.php
@@ -3,13 +3,14 @@
namespace Grasmash\YamlCli\Tests;
use Symfony\Component\Console\Application;
+use PHPUnit\Framework\TestCase;
/**
* Class BltTestBase.
*
* Base class for all tests that are executed for BLT itself.
*/
-abstract class TestBase extends \PHPUnit_Framework_TestCase
+abstract class TestBase extends TestCase
{
/** @var Application */
@@ -23,8 +24,7 @@ abstract class TestBase extends \PHPUnit_Framework_TestCase
*
* @see https://symfony.com/doc/current/console.html#testing-commands
*/
- public function setUp()
- {
+ protected function setUp(): void {
parent::setUp();
$this->application = new Application();
@@ -33,8 +33,7 @@ abstract class TestBase extends \PHPUnit_Framework_TestCase
/**
* Removes temporary file.
*/
- public function tearDown()
- {
+ protected function tearDown(): void {
parent::tearDown();
// This will only exist if a test called setupTemporaryConfigFiles(). | refactoring TestBase.php. | grasmash_yaml-cli | train |
4b284f7182cb5aca9d15964f4e6db9681231c41d | diff --git a/core/dutycycle/dutyManager.go b/core/dutycycle/dutyManager.go
index <HASH>..<HASH> 100644
--- a/core/dutycycle/dutyManager.go
+++ b/core/dutycycle/dutyManager.go
@@ -125,7 +125,7 @@ func (m *dutyManager) Update(id []byte, freq float64, size uint, datr string, co
var entry dutyEntry
if err == nil {
entry = itf.([]dutyEntry)[0]
- } else if err.(errors.Failure).Nature == errors.Behavioural {
+ } else if err.(errors.Failure).Nature == errors.NotFound {
entry = dutyEntry{
Until: time.Unix(0, 0),
OnAir: make(map[subBand]time.Duration), | [refactor/errors] Update duty cycle manager with NotFound error nature | TheThingsNetwork_ttn | train |
14df2a0f239f72a25963449e16e8f4f10db09fcd | diff --git a/Src/java/cql-to-elm/src/main/java/org/cqframework/cql/cql2elm/Cql2ElmVisitor.java b/Src/java/cql-to-elm/src/main/java/org/cqframework/cql/cql2elm/Cql2ElmVisitor.java
index <HASH>..<HASH> 100755
--- a/Src/java/cql-to-elm/src/main/java/org/cqframework/cql/cql2elm/Cql2ElmVisitor.java
+++ b/Src/java/cql-to-elm/src/main/java/org/cqframework/cql/cql2elm/Cql2ElmVisitor.java
@@ -1766,13 +1766,17 @@ public class Cql2ElmVisitor extends cqlBaseVisitor {
public Object visitConversionExpressionTerm(@NotNull cqlParser.ConversionExpressionTermContext ctx) {
TypeSpecifier targetType = parseTypeSpecifier(ctx.typeSpecifier());
Expression operand = parseExpression(ctx.expression());
- Conversion conversion = libraryBuilder.findConversion(operand.getResultType(), targetType.getResultType(), false);
- if (conversion == null) {
- throw new IllegalArgumentException(String.format("Could not resolve conversion from type %s to type %s.",
- operand.getResultType(), targetType.getResultType()));
+ if (!DataTypes.equal(operand.getResultType(), targetType.getResultType())) {
+ Conversion conversion = libraryBuilder.findConversion(operand.getResultType(), targetType.getResultType(), false);
+ if (conversion == null) {
+ throw new IllegalArgumentException(String.format("Could not resolve conversion from type %s to type %s.",
+ operand.getResultType(), targetType.getResultType()));
+ }
+
+ return libraryBuilder.convertExpression(operand, conversion);
}
- return libraryBuilder.convertExpression(operand, conversion);
+ return operand;
}
@Override | #<I>: Fixed no-op conversions throwing an error. | cqframework_clinical_quality_language | train |
114b4af6f7e82cb8c014d818206050ac71d74ab8 | diff --git a/timeside/server/tests/test_provider.py b/timeside/server/tests/test_provider.py
index <HASH>..<HASH> 100644
--- a/timeside/server/tests/test_provider.py
+++ b/timeside/server/tests/test_provider.py
@@ -17,7 +17,7 @@ class TestProvider(TimeSideTestServer):
for provider in request_providers.data:
if provider['pid'] == 'youtube':
self.youtube_uuid = provider['uuid']
- if provider['pid'] == 'deezer':
+ if provider['pid'] == 'deezer_preview':
self.deezer_uuid = provider['uuid'] | [server] change deezer_preview provider pid in REST API test | Parisson_TimeSide | train |
3891bc33ab2de8fa155f5112154d092d1aac424d | diff --git a/setup.py b/setup.py
index <HASH>..<HASH> 100644
--- a/setup.py
+++ b/setup.py
@@ -21,11 +21,13 @@ except IOError:
readme = ''
changelog = ''
-py_version = sys.version_info[:2]
+py_version = sys.version_info[:3]
-packages = ['mygeotab', 'mygeotab/ext']
+if py_version < (2, 7, 9):
+ raise RuntimeError("This package requres Python 2.7.9+")
-if py_version >= (3, 5):
+packages = ['mygeotab', 'mygeotab/ext']
+if py_version >= (3, 5, 0):
packages.append('mygeotab/async')
setup( | Since all MyGeotab servers use TLS <I>, we must force a version of Python 2 that supports this (<I>+).
Addresses found issue in #<I>. | Geotab_mygeotab-python | train |
e66c9c077bbb05bd237b6f6e098f6b8e3dcea990 | diff --git a/lang/en/completion.php b/lang/en/completion.php
index <HASH>..<HASH> 100644
--- a/lang/en/completion.php
+++ b/lang/en/completion.php
@@ -126,7 +126,7 @@ $string['markedcompleteby']='Marked complete by {$a}';
$string['markingyourselfcomplete']='Marking yourself complete';
$string['moredetails']='More details';
$string['nocriteriaset']='No completion criteria set for this course';
-$string['notenroled']='You are not enroled in this course';
+$string['notenroled']='You are not enrolled in this course';
$string['notyetstarted']='Not yet started';
$string['overallcriteriaaggregation']='Overall criteria type aggregation';
$string['passinggrade']='Passing grade';
@@ -147,7 +147,7 @@ $string['unenrolingfromcourse']='Unenroling from course';
$string['unenrolment']='Unenrolment';
$string['unlockcompletiondelete']='Unlock completion options and delete user completion data';
$string['usealternateselector']='Use the alternate course selector';
-$string['usernotenroled']='User is not enroled in this course';
+$string['usernotenroled']='User is not enrolled in this course';
$string['viewcoursereport']='View course report';
$string['viewingactivity']='Viewing the {$a}';
$string['xdays']='{$a} days'; | MDL-<I> completio: Fix spelling of Enrolled. | moodle_moodle | train |
4ce4c1d58db2ee74d584da759444dddb3419190e | diff --git a/trimesh/interfaces/generic.py b/trimesh/interfaces/generic.py
index <HASH>..<HASH> 100644
--- a/trimesh/interfaces/generic.py
+++ b/trimesh/interfaces/generic.py
@@ -18,7 +18,7 @@ class MeshScript:
# export the meshes to a temporary STL container
for m, f in zip(self.meshes, self.mesh_pre):
- m.export(file_type='stl', file_obj=f.name)
+ m.export(file_type='stl', file_obj=f)
self.replacement = {'mesh_' + str(i) : m.name for i,m in enumerate(self.mesh_pre)}
self.replacement['mesh_pre'] = str([i.name for i in self.mesh_pre]) | Windows does not allow re-opening the file | mikedh_trimesh | train |
9a133d51251e2c048089554327ab3ad4ab72327a | diff --git a/ara/api/tests/tests_play.py b/ara/api/tests/tests_play.py
index <HASH>..<HASH> 100644
--- a/ara/api/tests/tests_play.py
+++ b/ara/api/tests/tests_play.py
@@ -111,7 +111,7 @@ class PlayTestCase(APITestCase):
# Create a playbook and two plays
playbook = factories.PlaybookFactory()
play = factories.PlayFactory(name="first_play", playbook=playbook)
- factories.TaskFactory(name="second_play", playbook=playbook)
+ factories.PlayFactory(name="second_play", playbook=playbook)
# Query for the first play name and expect one result
request = self.client.get("/api/v1/plays?name=%s" % play.name) | tests: fix typo in play unit tests | ansible-community_ara | train |
3a2a0c67683caaf2f6dc8d3e2af1e04ed3e9ab1c | diff --git a/examples/aws/terraform/ansible/ec2.py b/examples/aws/terraform/ansible/ec2.py
index <HASH>..<HASH> 100755
--- a/examples/aws/terraform/ansible/ec2.py
+++ b/examples/aws/terraform/ansible/ec2.py
@@ -95,7 +95,8 @@ def generate_ssh_cfg(cluster, ssh_key=None):
bastion = ''
for reservation in response['Reservations']:
for instance in reservation['Instances']:
- bastion = instance['PublicIpAddress']
+ if instance['State']['Code'] == 16: # code 16 = running
+ bastion = instance['PublicIpAddress']
return ssh_cfg_template.format(bastion=bastion, ssh_key_path=ssh_key_path, identity_file=identity_file)
if args.cluster is None: | test state before getting public IP (#<I>) | gravitational_teleport | train |
716c0cfad85fbcd56bbda1752b6e25395936edb9 | diff --git a/api/charms/client.go b/api/charms/client.go
index <HASH>..<HASH> 100644
--- a/api/charms/client.go
+++ b/api/charms/client.go
@@ -61,7 +61,7 @@ func (c *Client) CharmInfo(charmURL string) (*CharmInfo, error) {
result := &CharmInfo{
Revision: info.Revision,
URL: info.URL,
- Config: convertCharmConfig(info.Config),
+ Config: params.FromCharmOptionMap(info.Config),
Meta: meta,
Actions: convertCharmActions(info.Actions),
Metrics: convertCharmMetrics(info.Metrics),
@@ -70,27 +70,6 @@ func (c *Client) CharmInfo(charmURL string) (*CharmInfo, error) {
return result, nil
}
-func convertCharmConfig(config map[string]params.CharmOption) *charm.Config {
- if len(config) == 0 {
- return nil
- }
- result := &charm.Config{
- Options: make(map[string]charm.Option),
- }
- for key, value := range config {
- result.Options[key] = convertCharmOption(value)
- }
- return result
-}
-
-func convertCharmOption(opt params.CharmOption) charm.Option {
- return charm.Option{
- Type: opt.Type,
- Description: opt.Description,
- Default: opt.Default,
- }
-}
-
func convertCharmMeta(meta *params.CharmMeta) (*charm.Meta, error) {
if meta == nil {
return nil, nil
diff --git a/apiserver/facades/client/charms/client.go b/apiserver/facades/client/charms/client.go
index <HASH>..<HASH> 100644
--- a/apiserver/facades/client/charms/client.go
+++ b/apiserver/facades/client/charms/client.go
@@ -89,7 +89,7 @@ func (a *API) CharmInfo(args params.CharmURL) (params.Charm, error) {
info := params.Charm{
Revision: aCharm.Revision(),
URL: curl.String(),
- Config: convertCharmConfig(aCharm.Config()),
+ Config: params.ToCharmOptionMap(aCharm.Config()),
Meta: convertCharmMeta(aCharm.Meta()),
Actions: convertCharmActions(aCharm.Actions()),
Metrics: convertCharmMetrics(aCharm.Metrics()),
@@ -152,25 +152,6 @@ func (a *API) IsMetered(args params.CharmURL) (params.IsMeteredResult, error) {
return params.IsMeteredResult{Metered: false}, nil
}
-func convertCharmConfig(config *charm.Config) map[string]params.CharmOption {
- if config == nil {
- return nil
- }
- result := make(map[string]params.CharmOption)
- for key, value := range config.Options {
- result[key] = convertCharmOption(value)
- }
- return result
-}
-
-func convertCharmOption(opt charm.Option) params.CharmOption {
- return params.CharmOption{
- Type: opt.Type,
- Description: opt.Description,
- Default: opt.Default,
- }
-}
-
func convertCharmMeta(meta *charm.Meta) *params.CharmMeta {
if meta == nil {
return nil
diff --git a/apiserver/params/charms.go b/apiserver/params/charms.go
index <HASH>..<HASH> 100644
--- a/apiserver/params/charms.go
+++ b/apiserver/params/charms.go
@@ -3,6 +3,8 @@
package params
+import "github.com/juju/charm/v7"
+
// ApplicationCharmResults contains a set of ApplicationCharmResults.
type ApplicationCharmResults struct {
Results []ApplicationCharmResult `json:"results"`
@@ -197,3 +199,43 @@ type ContainerProfileResult struct {
type ContainerProfileResults struct {
Results []ContainerProfileResult `json:"results"`
}
+
+func ToCharmOptionMap(config *charm.Config) map[string]CharmOption {
+ if config == nil {
+ return nil
+ }
+ result := make(map[string]CharmOption)
+ for key, value := range config.Options {
+ result[key] = toParamsCharmOption(value)
+ }
+ return result
+}
+
+func toParamsCharmOption(opt charm.Option) CharmOption {
+ return CharmOption{
+ Type: opt.Type,
+ Description: opt.Description,
+ Default: opt.Default,
+ }
+}
+
+func FromCharmOptionMap(config map[string]CharmOption) *charm.Config {
+ if len(config) == 0 {
+ return nil
+ }
+ result := &charm.Config{
+ Options: make(map[string]charm.Option),
+ }
+ for key, value := range config {
+ result.Options[key] = fromParamsCharmOption(value)
+ }
+ return result
+}
+
+func fromParamsCharmOption(opt CharmOption) charm.Option {
+ return charm.Option{
+ Type: opt.Type,
+ Description: opt.Description,
+ Default: opt.Default,
+ }
+} | Move convertCharmConfig to params.
Renamed to ToCharmOptionMap and FromCharmOptionMap. This functionality
is needed by other libraries. Put in a common place to share. | juju_juju | train |
5fe29a64650eab8da7ef4730db769ce6894a7b6a | diff --git a/internal/util/comparisons.js b/internal/util/comparisons.js
index <HASH>..<HASH> 100644
--- a/internal/util/comparisons.js
+++ b/internal/util/comparisons.js
@@ -190,7 +190,7 @@ function innerDeepEqual(val1, val2, strict, memos) {
return keyCheck(val1, val2, strict, memos, kNoIterator);
}
if (isDate(val1)) {
- if (!isDate(val2) || Date.prototype.getTime.call(val1) !== Date.prototype.getTime.call(val2)) {
+ if (!(isDate(val2) && Date.prototype.getTime.call(val1) === Date.prototype.getTime.call(val2))) {
return false;
}
} else if (isRegExp(val1)) { | Make sure getTime is never called on a non-date object | browserify_commonjs-assert | train |
bd6c07038defa15a9934b7c95850a30fe24b7e32 | diff --git a/src/test/java/com/cronutils/model/time/generator/ExecutionTimeUnixIntegrationTest.java b/src/test/java/com/cronutils/model/time/generator/ExecutionTimeUnixIntegrationTest.java
index <HASH>..<HASH> 100644
--- a/src/test/java/com/cronutils/model/time/generator/ExecutionTimeUnixIntegrationTest.java
+++ b/src/test/java/com/cronutils/model/time/generator/ExecutionTimeUnixIntegrationTest.java
@@ -265,4 +265,15 @@ public class ExecutionTimeUnixIntegrationTest {
ExecutionTime executionTime = ExecutionTime.forCron(cron);
assertEquals(DateTime.parse("2014-12-01T08:00:00Z"), executionTime.nextExecution(date));
}
+
+ /**
+ * Issue #92: Next execution skipping valid date
+ */
+ @Test
+ public void testNextExecution2016() {
+ CronParser parser = new CronParser(CronDefinitionBuilder.instanceDefinitionFor(CronType.UNIX));
+ ExecutionTime executionTime = ExecutionTime.forCron(parser.parse("1 0 * * tue"));
+ DateTime date = DateTime.parse("2016-05-24T01:02:50Z");
+ assertEquals(DateTime.parse("2016-05-31T00:01:00Z"), executionTime.nextExecution(date));
+ }
} | Failing unit test for issue #<I> | jmrozanec_cron-utils | train |
b007283d753c0c7cab66fb9ee51626171b3f72f0 | diff --git a/littletable.py b/littletable.py
index <HASH>..<HASH> 100644
--- a/littletable.py
+++ b/littletable.py
@@ -117,7 +117,7 @@ Here is a simple C{littletable} data storage/retrieval example::
"""
__version__ = "0.13.2"
-__versionTime__ = "3 Nov 2018 00:00 UTC"
+__versionTime__ = "6 Nov 2018 23:02 UTC"
__author__ = "Paul McGuire <[email protected]>"
import sys
@@ -1704,25 +1704,18 @@ class JoinTerm(object):
if __name__ == "__main__":
-
- rawdata = """\
+ import textwrap
+ rawdata = textwrap.dedent("""\
Phoenix:AZ:85001:KPHX
Phoenix:AZ:85001:KPHY
Phoenix:AZ:85001:KPHA
- Dallas:TX:75201:KDFW""".splitlines()
+ Dallas:TX:75201:KDFW""")
# load miniDB
- stations = Table()
+ stations = Table().csv_import(rawdata, delimiter=':', fieldnames=['city', 'state', 'zip', 'stn'])
# stations.create_index("city")
stations.create_index("stn", unique=True)
- data_fields = "city state zip stn".split()
- for d in rawdata:
- rec = DataObject()
- for kk, vv in zip(data_fields, d.split(':')):
- setattr(rec, kk, vv.strip())
- stations.insert(rec)
-
# perform some queries and deletes
for queryargs in [
dict(city="Phoenix"),
@@ -1808,3 +1801,5 @@ if __name__ == "__main__":
print(amfm.by.stn['KPHY'])
except KeyError:
print("no station 'KPHY' in table")
+
+ print(list(stations.all.stn)) | Include inline CSV import and table.all example to sample code in __main__ | ptmcg_littletable | train |
f25f12f84441174aeda5e2545c2846a4ba6eb068 | diff --git a/redisson/src/main/java/org/redisson/connection/pool/ConnectionPool.java b/redisson/src/main/java/org/redisson/connection/pool/ConnectionPool.java
index <HASH>..<HASH> 100644
--- a/redisson/src/main/java/org/redisson/connection/pool/ConnectionPool.java
+++ b/redisson/src/main/java/org/redisson/connection/pool/ConnectionPool.java
@@ -169,8 +169,8 @@ abstract class ConnectionPool<T extends RedisConnection> {
public RFuture<T> get(RedisCommand<?> command) {
for (int j = entries.size() - 1; j >= 0; j--) {
final ClientConnectionsEntry entry = getEntry();
- if (!entry.isFreezed()
- && tryAcquireConnection(entry)) {
+ if ((!entry.isFreezed() || entry.getFreezeReason() == FreezeReason.SYSTEM) &&
+ tryAcquireConnection(entry)) {
return acquireConnection(command, entry);
}
} | add missing condition for tentatively unfreezing node | redisson_redisson | train |
ac257582528f5ab4aed95132688ca24e0db0994f | diff --git a/lib/engine/piston/contextify_task_executor.js b/lib/engine/piston/contextify_task_executor.js
index <HASH>..<HASH> 100644
--- a/lib/engine/piston/contextify_task_executor.js
+++ b/lib/engine/piston/contextify_task_executor.js
@@ -43,15 +43,21 @@ ContextifyTaskExecutor.prototype.execute_task = function(task_request){
child_global_object[name] = child_grafter.graft(child_globals[name]);
});
+ var last_eval;
+ var response;
+
+ try {
+ last_eval = vm.runInContext(task_request.getCode(), child_context);
+ response = new TaskResponse({response:{evaluation: last_eval}});
+ } catch(e) {
+ response = new TaskResponse({response:{error: e.name + ": " + e.message}});
+ }
- var last_eval = vm.runInContext(task_request.getCode(), child_context);
-
- var response = new TaskResponse({response:{evaluation: last_eval}});
this.emit("task executed", response);
};
ContextifyTaskExecutor.prototype.render_child_globals = function(task_request){
- var child_globals = eval(task_request.getContext())();
+ var child_globals = eval(task_request.getContext())(task_request.getLocals());
return child_globals;
};
diff --git a/spec/engine/piston/contextify_task_executor_spec.js b/spec/engine/piston/contextify_task_executor_spec.js
index <HASH>..<HASH> 100644
--- a/spec/engine/piston/contextify_task_executor_spec.js
+++ b/spec/engine/piston/contextify_task_executor_spec.js
@@ -5,6 +5,24 @@ var TaskResponse = require("../../../lib/engine/client/task/response").TaskRespo
describe("ContextifyTaskExecutor", function(){
beforeEach(function(){
this.executor = ContextifyTaskExecutor.create();
+ this.assert_task_response = function(task_request, task_response){
+ var spy = jasmine.createSpy();
+
+ this.executor.on("task executed", spy);
+
+ runs(function(){
+ this.executor.execute_task(task_request);
+ });
+
+ waitsFor(function(){
+ return spy.callCount > 0;
+ });
+
+ runs(function(){
+ expect(spy).toHaveBeenCalledWith(task_response);
+ });
+ };
+
});
it("executes a simple TaskRequest", function(){
@@ -15,20 +33,45 @@ describe("ContextifyTaskExecutor", function(){
locals: {}
});
var simple_task_response = new TaskResponse({response:{evaluation:4}});
- var spy = jasmine.createSpy();
-
- this.executor.on("task executed", spy);
- runs(function(){
- this.executor.execute_task(simple_task);
+ this.assert_task_response(simple_task, simple_task_response);
+ });
+
+ it("executes a TaskRequest with locals", function(){
+ var simple_task = new TaskRequest({
+ task_id: "2",
+ context: "(function(locals){ return { hello:function(){ return 'hello ' + locals.world; } } })",
+ code: "hello()",
+ locals: { world:"world" }
});
+ var simple_task_response = new TaskResponse({response:{evaluation: "hello world"}});
+
+ this.assert_task_response(simple_task, simple_task_response);
+ });
- waitsFor(function(){
- return spy.callCount > 0;
+ it("executes a TaskRequest that contains syntax errors", function(){
+ var simple_task = new TaskRequest({
+ task_id: "2",
+ context: "(function(locals){ return { hello:function(){ return 'hello'; } } })",
+ code: "hello(",
+ locals: { }
});
+ var simple_task_response = new TaskResponse({response:{error: "SyntaxError: Unexpected end of input"}});
+
+ this.assert_task_response(simple_task, simple_task_response);
+ });
- runs(function(){
- expect(spy).toHaveBeenCalledWith(simple_task_response);
- });
+ it("executes a TaskRequest that contains reference errors", function(){
+ var simple_task = new TaskRequest({
+ task_id: "2",
+ context: "(function(locals){ return { hello:function(){ return 'hello'; } } })",
+ code: "foo()",
+ locals: { }
+ });
+ var simple_task_response = new TaskResponse({response:{error: "ReferenceError: foo is not defined"}});
+
+ this.assert_task_response(simple_task, simple_task_response);
});
+
+ xit("executes a TaskRequest with asynchronous code");
});
\ No newline at end of file | WIP: Handle syntax and reference errors during task execution | rehanift_engine.js | train |
1d9fa88f6de61c6975021489b7086d9cf33b945c | diff --git a/gwpy/cli/cliproduct.py b/gwpy/cli/cliproduct.py
index <HASH>..<HASH> 100644
--- a/gwpy/cli/cliproduct.py
+++ b/gwpy/cli/cliproduct.py
@@ -27,8 +27,6 @@ import warnings
import sys
from functools import wraps
-from six import add_metaclass
-
from matplotlib import rcParams
try:
from matplotlib.cm import viridis as DEFAULT_CMAP
@@ -94,8 +92,7 @@ to_s = to_float('s') # pylint: disable=invalid-name
# -- base product class -------------------------------------------------------
-@add_metaclass(abc.ABCMeta)
-class CliProduct(object):
+class CliProduct(object, metaclass=abc.ABCMeta):
"""Base class for all cli plot products
Parameters
@@ -761,15 +758,14 @@ class CliProduct(object):
# -- extensions ---------------------------------------------------------------
-@add_metaclass(abc.ABCMeta)
-class ImageProduct(CliProduct):
+class ImageProduct(CliProduct, metaclass=abc.ABCMeta):
"""Base class for all x/y/color plots
"""
MAX_DATASETS = 1
@classmethod
def init_plot_options(cls, parser):
- super(ImageProduct, cls).init_plot_options(parser)
+ super().init_plot_options(parser)
cls.arg_color_axis(parser)
@classmethod
@@ -797,7 +793,7 @@ class ImageProduct(CliProduct):
def _finalize_arguments(self, args):
if args.cmap is None:
args.cmap = DEFAULT_CMAP.name
- return super(ImageProduct, self)._finalize_arguments(args)
+ return super()._finalize_arguments(args)
@staticmethod
def get_color_label():
@@ -809,7 +805,7 @@ class ImageProduct(CliProduct):
"""Set properties for each axis (scale, limits, label) and create
a colorbar.
"""
- super(ImageProduct, self).set_axes_properties()
+ super().set_axes_properties()
if not self.args.nocolorbar:
self.set_colorbar()
@@ -824,8 +820,7 @@ class ImageProduct(CliProduct):
return # image plots don't have legends
-@add_metaclass(abc.ABCMeta)
-class FFTMixin(object):
+class FFTMixin(object, metaclass=abc.ABCMeta):
"""Mixin for `CliProduct` class that will perform FFTs
This just adds FFT-based command line options
@@ -834,7 +829,7 @@ class FFTMixin(object):
def init_data_options(cls, parser):
"""Set up data input and signal processing options including FFTs
"""
- super(FFTMixin, cls).init_data_options(parser)
+ super().init_data_options(parser)
cls.arg_fft(parser)
@classmethod
@@ -899,11 +894,10 @@ class FFTMixin(object):
args.overlap = recommended_overlap(args.window)
except ValueError:
args.overlap = .5
- return super(FFTMixin, self)._finalize_arguments(args)
+ return super()._finalize_arguments(args)
-@add_metaclass(abc.ABCMeta)
-class TimeDomainProduct(CliProduct):
+class TimeDomainProduct(CliProduct, metaclass=abc.ABCMeta):
"""`CliProduct` with time on the X-axis
"""
@classmethod
@@ -913,7 +907,7 @@ class TimeDomainProduct(CliProduct):
This method includes the standard X-axis options, as well as a new
``--epoch`` option for the time axis.
"""
- group = super(TimeDomainProduct, cls).arg_xaxis(parser)
+ group = super().arg_xaxis(parser)
group.add_argument('--epoch', type=to_gps,
help='center X axis on this GPS time, may be'
'absolute date/time or delta')
@@ -937,7 +931,7 @@ class TimeDomainProduct(CliProduct):
if args.xmax is None:
args.xmax = max(starts) + args.duration
- return super(TimeDomainProduct, self)._finalize_arguments(args)
+ return super()._finalize_arguments(args)
def get_xlabel(self):
"""Default X-axis label for plot
@@ -953,8 +947,7 @@ class TimeDomainProduct(CliProduct):
return ''
-@add_metaclass(abc.ABCMeta)
-class FrequencyDomainProduct(CliProduct):
+class FrequencyDomainProduct(CliProduct, metaclass=abc.ABCMeta):
"""`CliProduct` with frequency on the X-axis
"""
def get_xlabel(self): | Drop use of six.add_metaclass, as it interferes with python3's super() functionality | gwpy_gwpy | train |
022068a23230dd13d8e99852f47eaaf461cf8174 | diff --git a/tests/lib/toolkit/XMLDocumentTest.php b/tests/lib/toolkit/XMLDocumentTest.php
index <HASH>..<HASH> 100644
--- a/tests/lib/toolkit/XMLDocumentTest.php
+++ b/tests/lib/toolkit/XMLDocumentTest.php
@@ -22,6 +22,20 @@ final class XMLDocumentTest extends TestCase
$this->assertEquals('<xml />', $x->generate());
}
+ public function testSetVersion()
+ {
+ $x = (new \XMLDocument('xml'))->setVersion('test')->renderHeader();
+ $this->assertEquals('test', $x->getVersion());
+ $this->assertEquals('<?xml version="test" encoding="utf-8" ?><xml />', $x->generate());
+ }
+
+ public function testSetEncoding()
+ {
+ $x = (new \XMLDocument('xml'))->setEncoding('test')->renderHeader();
+ $this->assertEquals('test', $x->getEncoding());
+ $this->assertEquals('<?xml version="1.0" encoding="test" ?><xml />', $x->generate());
+ }
+
public function testGenerateWithHeader()
{
$x = (new \XMLDocument('xml', 'value')); | Add setVersion() and setEncoding() tests
Picked from <I>d<I>cad
Picked from 0c3ce<I>b<I> | symphonycms_symphony-2 | train |
1043e957ac7c899877758a6129ed1a32f383ef55 | diff --git a/src/Slugify.php b/src/Slugify.php
index <HASH>..<HASH> 100644
--- a/src/Slugify.php
+++ b/src/Slugify.php
@@ -524,6 +524,29 @@ class Slugify implements SlugifyInterface
}
/**
+ * Adds a ruleset to the Slugifier.
+ *
+ * @param string $name Name of the ruleset.
+ * @param string[][] $rules Rules
+ */
+ public function addRuleset($name, array $rules)
+ {
+ $this->rulesets[$name] = $rules;
+
+ return $this;
+ }
+
+ /**
+ * Returns the rulesets.
+ *
+ * @return string[][] Rulesets
+ */
+ public function getRulesets()
+ {
+ return $this->rulesets;
+ }
+
+ /**
* Static method to create new instance of {@see Slugify}.
*
* @return Slugify
diff --git a/tests/SlugifyTest.php b/tests/SlugifyTest.php
index <HASH>..<HASH> 100644
--- a/tests/SlugifyTest.php
+++ b/tests/SlugifyTest.php
@@ -91,6 +91,18 @@ class SlugifyTest extends \PHPUnit_Framework_TestCase
/**
* @test
+ * @covers Cocur\Slugify\Slugify::addRuleset()
+ * @covers Cocur\Slugify\Slugify::getRulesets()
+ */
+ public function addRulesetGetRulesets()
+ {
+ $this->slugify->addRuleset('foo', array('k' => 'key'));
+
+ $this->assertCount(2, $this->slugify->getRulesets());
+ }
+
+ /**
+ * @test
* @covers Cocur\Slugify\Slugify::create()
*/
public function createReturnsAnInstance() | Added possibility to add and get rulesets | cocur_slugify | train |
c4d6b71e40f35821eabc7372151cee7d27f12568 | diff --git a/www/admin/metadata.php b/www/admin/metadata.php
index <HASH>..<HASH> 100644
--- a/www/admin/metadata.php
+++ b/www/admin/metadata.php
@@ -12,36 +12,31 @@ $config = SimpleSAML_Configuration::getInstance();
//$metadata = new SimpleSAML_XML_MetaDataStore($config);
$session = SimpleSAML_Session::getInstance();
-
-
try {
$metadata = SimpleSAML_Metadata_MetaDataStorageHandler::getMetadataHandler();
-
$et = new SimpleSAML_XHTML_Template($config, 'admin-metadatalist.php');
-
if ($config->getValue('enable.saml20-sp') === true) {
$results = array();
- /*
+
$metalist = $metadata->getList('saml20-sp-hosted');
foreach ($metalist AS $entityid => $mentry) {
$results[$entityid] = SimpleSAML_Utilities::checkAssocArrayRules($mentry,
- // TODO: UPDATE Required and optional parameter list
- array('entityid', 'host', 'spNameQualifier', 'NameIDFormat', 'ForceAuthn'),
- array('name', 'description')
+ array('entityid', 'host', 'NameIDFormat', 'ForceAuthn'),
+ array()
);
}
$et->data['metadata.saml20-sp-hosted'] = $results;
- */
+
+ $results = array();
$metalist = $metadata->getList('saml20-idp-remote');
foreach ($metalist AS $entityid => $mentry) {
$results[$entityid] = SimpleSAML_Utilities::checkAssocArrayRules($mentry,
- // TODO: UPDATE Required and optional parameter list
- array('entityid', 'host', 'spNameQualifier', 'NameIDFormat', 'ForceAuthn'),
- array('name', 'description')
+ array('entityid', 'SingleSignOnService', 'SingleLogoutService', 'certFingerprint'),
+ array('name', 'description', 'base64attributes')
);
}
$et->data['metadata.saml20-idp-remote'] = $results;
@@ -53,19 +48,18 @@ try {
$metalist = $metadata->getList('saml20-idp-hosted');
foreach ($metalist AS $entityid => $mentry) {
$results[$entityid] = SimpleSAML_Utilities::checkAssocArrayRules($mentry,
- // TODO: UPDATE Required and optional parameter list
- array('entityid', 'host', 'spNameQualifier', 'NameIDFormat', 'ForceAuthn'),
- array('name', 'description')
+ array('entityid', 'host', 'privatekey', 'certificate', 'auth'),
+ array('requireconsent')
);
}
$et->data['metadata.saml20-idp-hosted'] = $results;
+ $results = array();
$metalist = $metadata->getList('saml20-sp-remote');
foreach ($metalist AS $entityid => $mentry) {
$results[$entityid] = SimpleSAML_Utilities::checkAssocArrayRules($mentry,
- // TODO: UPDATE Required and optional parameter list
- array('entityid', 'host', 'spNameQualifier', 'NameIDFormat', 'ForceAuthn'),
- array('name', 'description')
+ array('entityid', 'spNameQualifier', 'AssertionConsumerService', 'SingleLogoutService', 'NameIDFormat'),
+ array('base64attributes', 'attributemap', 'simplesaml.attributes', 'attributes', 'name', 'description')
);
}
$et->data['metadata.saml20-sp-remote'] = $results;
@@ -73,11 +67,7 @@ try {
}
-
-
-
-
$et->data['header'] = 'Metadata overview'; | Improved the metadata overview admin page by configuring which parameters that are optional and required in each metadata set | simplesamlphp_saml2 | train |
f7ca720f572f3ecc8471810084f98464bd33925b | diff --git a/src/kff.Collection.js b/src/kff.Collection.js
index <HASH>..<HASH> 100644
--- a/src/kff.Collection.js
+++ b/src/kff.Collection.js
@@ -16,7 +16,7 @@ kff.Collection = kff.createClass(
*/
constructor: function(options)
{
- options = options || {};
+ this.options = options = options || {};
this.itemFactory = options.itemFactory || null;
this.itemType = options.itemType || kff.Model;
this.serializeAttrs = options.serializeAttrs || null; | fix(kff.Collection): collection options (metadata) should be saved in the instance for later use (i.e. cloning) | karfcz_kff | train |
f6db0af6fd33eb9d3540c31ec3bc925d05e71d50 | diff --git a/openstack_dashboard/dashboards/project/loadbalancers/workflows.py b/openstack_dashboard/dashboards/project/loadbalancers/workflows.py
index <HASH>..<HASH> 100644
--- a/openstack_dashboard/dashboards/project/loadbalancers/workflows.py
+++ b/openstack_dashboard/dashboards/project/loadbalancers/workflows.py
@@ -25,6 +25,10 @@ from horizon import workflows
from openstack_dashboard import api
+AVAILABLE_PROTOCOLS = ('HTTP', 'HTTPS', 'TCP')
+AVAILABLE_METHODS = ('ROUND_ROBIN', 'LEAST_CONNECTIONS', 'SOURCE_IP')
+
+
class AddPoolAction(workflows.Action):
name = forms.CharField(max_length=80, label=_("Name"))
description = forms.CharField(
@@ -57,14 +61,11 @@ class AddPoolAction(workflows.Action):
self.fields['subnet_id'].choices = subnet_id_choices
protocol_choices = [('', _("Select a Protocol"))]
- protocol_choices.append(('HTTP', 'HTTP'))
- protocol_choices.append(('HTTPS', 'HTTPS'))
+ [protocol_choices.append((p, p)) for p in AVAILABLE_PROTOCOLS]
self.fields['protocol'].choices = protocol_choices
lb_method_choices = [('', _("Select a Method"))]
- lb_method_choices.append(('ROUND_ROBIN', 'ROUND_ROBIN'))
- lb_method_choices.append(('LEAST_CONNECTIONS', 'LEAST_CONNECTIONS'))
- lb_method_choices.append(('SOURCE_IP', 'SOURCE_IP'))
+ [lb_method_choices.append((m, m)) for m in AVAILABLE_METHODS]
self.fields['lb_method'].choices = lb_method_choices
# provider choice
@@ -185,8 +186,7 @@ class AddVipAction(workflows.Action):
args[0]['subnet'])
protocol_choices = [('', _("Select a Protocol"))]
- protocol_choices.append(('HTTP', 'HTTP'))
- protocol_choices.append(('HTTPS', 'HTTPS'))
+ [protocol_choices.append((p, p)) for p in AVAILABLE_PROTOCOLS]
self.fields['protocol'].choices = protocol_choices
session_persistence_choices = [('', _("No Session Persistence"))] | lbaas/horizon - adds tcp protocol choice when create lb
This option was missed
Change-Id: I<I>c<I>a<I>a<I>da9b3dc<I>d<I>a<I>ebe
Closes-Bug: #<I> | openstack_horizon | train |
dcda31fb0dba39f4b1daa54298a6d04fff9dac1b | diff --git a/pkg/sdn/plugin/networkpolicy.go b/pkg/sdn/plugin/networkpolicy.go
index <HASH>..<HASH> 100644
--- a/pkg/sdn/plugin/networkpolicy.go
+++ b/pkg/sdn/plugin/networkpolicy.go
@@ -364,7 +364,6 @@ func (np *networkPolicyPlugin) selectPods(npns *npNamespace, lsel *unversioned.L
func (np *networkPolicyPlugin) parseNetworkPolicy(npns *npNamespace, policy *extensions.NetworkPolicy) (*npPolicy, error) {
npp := &npPolicy{policy: *policy}
- allowAll := false
var destFlows []string
if len(policy.Spec.PodSelector.MatchLabels) > 0 || len(policy.Spec.PodSelector.MatchExpressions) > 0 {
@@ -377,11 +376,6 @@ func (np *networkPolicyPlugin) parseNetworkPolicy(npns *npNamespace, policy *ext
}
for _, rule := range policy.Spec.Ingress {
- if len(rule.Ports) == 0 && len(rule.From) == 0 {
- allowAll = true
- break
- }
-
var portFlows, peerFlows []string
if len(rule.Ports) == 0 {
portFlows = []string{""}
@@ -446,11 +440,7 @@ func (np *networkPolicyPlugin) parseNetworkPolicy(npns *npNamespace, policy *ext
}
}
- if allowAll {
- npp.flows = []string{""}
- } else {
- sort.Strings(npp.flows)
- }
+ sort.Strings(npp.flows)
glog.V(5).Infof("Parsed NetworkPolicy: %#v", npp)
return npp, nil
} | Fix NetworkPolicies allowing from all to *some* (not all) | openshift_origin | train |
2d174843befac50a5e301364281693bcd78d1019 | diff --git a/lib/multi_client/version.rb b/lib/multi_client/version.rb
index <HASH>..<HASH> 100644
--- a/lib/multi_client/version.rb
+++ b/lib/multi_client/version.rb
@@ -1,3 +1,3 @@
module MultiClient
- VERSION = '3.1.1'.freeze
+ VERSION = '3.1.2'.freeze
end | Bumped version to <I> | robotex82_multi_client | train |
06fb9be4310e1f18f03cd2ace0df4f6cf0068709 | diff --git a/examples/full.rb b/examples/full.rb
index <HASH>..<HASH> 100644
--- a/examples/full.rb
+++ b/examples/full.rb
@@ -29,6 +29,9 @@ class UsersController < Cramp::Controller::Base
end
end
+ # Sends a space ( ' ' ) to the client for keeping the connection alive. Default : Every 30 seconds
+ keep_connection_alive :every => 10
+
# Polls every 1 second by default
periodic_timer :poll_user
diff --git a/lib/cramp/controller/base.rb b/lib/cramp/controller/base.rb
index <HASH>..<HASH> 100644
--- a/lib/cramp/controller/base.rb
+++ b/lib/cramp/controller/base.rb
@@ -32,6 +32,11 @@ module Cramp
defined?(@@periodic_timers) ? @@periodic_timers : []
end
+ def self.keep_connection_alive(options = {})
+ options = { :every => 30 }.merge(options)
+ periodic_timer :keep_connection_alive, options
+ end
+
def initialize(env)
@env = env
@timers = []
@@ -93,9 +98,6 @@ module Cramp
end
end
- def start_timer
- end
-
def stop_periodic_timers
@timers.each {|t| t.cancel }
end
@@ -103,6 +105,10 @@ module Cramp
def on_finish
end
+ def keep_connection_alive
+ render " "
+ end
+
end
end
end | Add keep_connection_alive for long polling/streaming | lifo_cramp | train |
2ad03a96b5e9ccbcb9e5b20f91018927397a622e | diff --git a/framework/base/DynamicModel.php b/framework/base/DynamicModel.php
index <HASH>..<HASH> 100644
--- a/framework/base/DynamicModel.php
+++ b/framework/base/DynamicModel.php
@@ -50,6 +50,10 @@ use yii\validators\Validator;
* "dynamic attributes". It basically allows an attribute to be defined dynamically through its constructor
* or [[defineAttribute()]].
*
+ * Note that it does not support the usual validation flow as normal models, i.e. no scenarios are being defined
+ * and therefor none of the attributes is safe. This means that [[load()]] will not be able to perform massive
+ * assignment and return `false`. Attributes should be assigned via constructor instead.
+ *
* @author Qiang Xue <[email protected]>
* @since 2.0
*/ | Update DynamicModel.php
added note about limitations, i.e. no scenarios -> no safe attributes -> no load(). | yiisoft_yii-core | train |
95428ddd6405d53f27867c3d1057e43cbbd2816e | diff --git a/lib/fake_web/registry.rb b/lib/fake_web/registry.rb
index <HASH>..<HASH> 100644
--- a/lib/fake_web/registry.rb
+++ b/lib/fake_web/registry.rb
@@ -58,8 +58,10 @@ module FakeWeb
def normalize_uri(uri)
normalized_uri =
case uri
- when URI then uri
- else
+ when URI then
+ uri.query = sort_query_params(uri.query)
+ uri
+ else
uri = 'http://' + uri unless uri.match('^https?://')
parsed_uri = URI.parse(uri)
parsed_uri.query = sort_query_params(parsed_uri.query)
diff --git a/test/test_query_string.rb b/test/test_query_string.rb
index <HASH>..<HASH> 100644
--- a/test/test_query_string.rb
+++ b/test/test_query_string.rb
@@ -16,6 +16,11 @@ class TestFakeWebQueryString < Test::Unit::TestCase
assert FakeWeb.registered_uri?('http://example.com/?b=1&a=1')
end
+ def test_register_uri_with_query_params_unsorted_from_uri_object
+ FakeWeb.register_uri(URI.join('http://example.com/?b=1&a=1'), :string => 'foo')
+ assert FakeWeb.registered_uri?('http://example.com/?b=1&a=1')
+ end
+
def test_registered_uri_gets_recognized_with_empty_query_params
FakeWeb.register_uri('http://example.com/', :string => 'foo')
assert FakeWeb.registered_uri?('http://example.com/?') | Bug fixed: sorting query parameters from URI objects at register uri time. | chrisk_fakeweb | train |
b19d67ccf239ba2bb0ca55cae20775cf8c51c771 | diff --git a/Documentation/cli/README.md b/Documentation/cli/README.md
index <HASH>..<HASH> 100644
--- a/Documentation/cli/README.md
+++ b/Documentation/cli/README.md
@@ -185,7 +185,7 @@ Set breakpoint condition.
Specifies that the breakpoint, tracepoint or watchpoint should break only if the boolean expression is true.
-See Documentation/cli/expr.md for a description of supported expressions.
+See [Documentation/cli/expr.md](//github.com/go-delve/delve/tree/master/Documentation/cli/expr.md) for a description of supported expressions.
With the -hitcount option a condition on the breakpoint hit count can be set, the following operators are supported
@@ -511,7 +511,7 @@ Evaluate an expression.
[goroutine <n>] [frame <m>] print [%format] <expression>
-See Documentation/cli/expr.md for a description of supported expressions.
+See [Documentation/cli/expr.md](//github.com/go-delve/delve/tree/master/Documentation/cli/expr.md) for a description of supported expressions.
The optional format argument is a format specifier, like the ones used by the fmt package. For example "print %x v" will print v as an hexadecimal number.
@@ -526,7 +526,7 @@ Print contents of CPU registers.
regs [-a]
-Argument -a shows more registers. Individual registers can also be displayed by 'print' and 'display'. See Documentation/cli/expr.md.
+Argument -a shows more registers. Individual registers can also be displayed by 'print' and 'display'. See [Documentation/cli/expr.md.](//github.com/go-delve/delve/tree/master/Documentation/cli/expr.md.)
## restart
@@ -569,7 +569,7 @@ Changes the value of a variable.
[goroutine <n>] [frame <m>] set <variable> = <value>
-See Documentation/cli/expr.md for a description of supported expressions. Only numerical variables and pointers can be changed.
+See [Documentation/cli/expr.md](//github.com/go-delve/delve/tree/master/Documentation/cli/expr.md) for a description of supported expressions. Only numerical variables and pointers can be changed.
## source
diff --git a/pkg/terminal/command.go b/pkg/terminal/command.go
index <HASH>..<HASH> 100644
--- a/pkg/terminal/command.go
+++ b/pkg/terminal/command.go
@@ -124,14 +124,14 @@ Type "help" followed by the name of a command for more information about it.`},
break [name] [locspec]
-See $GOPATH/src/github.com/go-delve/delve/Documentation/cli/locspec.md for the syntax of locspec. If locspec is omitted a breakpoint will be set on the current line.
+See Documentation/cli/locspec.md for the syntax of locspec. If locspec is omitted a breakpoint will be set on the current line.
See also: "help on", "help cond" and "help clear"`},
{aliases: []string{"trace", "t"}, group: breakCmds, cmdFn: tracepoint, allowedPrefixes: onPrefix, helpMsg: `Set tracepoint.
trace [name] [locspec]
-A tracepoint is a breakpoint that does not stop the execution of the program, instead when the tracepoint is hit a notification is displayed. See $GOPATH/src/github.com/go-delve/delve/Documentation/cli/locspec.md for the syntax of locspec. If locspec is omitted a tracepoint will be set on the current line.
+A tracepoint is a breakpoint that does not stop the execution of the program, instead when the tracepoint is hit a notification is displayed. See Documentation/cli/locspec.md for the syntax of locspec. If locspec is omitted a tracepoint will be set on the current line.
See also: "help on", "help cond" and "help clear"`},
{aliases: []string{"watch"}, group: breakCmds, cmdFn: watchpoint, helpMsg: `Set watchpoint.
@@ -428,7 +428,7 @@ Executes the specified command (print, args, locals) in the context of the n-th
source <path>
-If path ends with the .star extension it will be interpreted as a starlark script. See $GOPATH/src/github.com/go-delve/delve/Documentation/cli/starlark.md for the syntax.
+If path ends with the .star extension it will be interpreted as a starlark script. See Documentation/cli/starlark.md for the syntax.
If path is a single '-' character an interactive starlark interpreter will start instead. Type 'exit' to exit.`},
{aliases: []string{"disassemble", "disass"}, cmdFn: disassCommand, helpMsg: `Disassembler.
diff --git a/pkg/terminal/docgen.go b/pkg/terminal/docgen.go
index <HASH>..<HASH> 100644
--- a/pkg/terminal/docgen.go
+++ b/pkg/terminal/docgen.go
@@ -7,13 +7,20 @@ import (
)
func replaceDocPath(s string) string {
- const docpath = "$GOPATH/src/github.com/go-delve/delve/"
+ const docpath = "Documentation/"
+
+ i0 := 0
for {
- start := strings.Index(s, docpath)
+ start := strings.Index(s[i0:], docpath)
if start < 0 {
return s
}
+ start += i0
+ if start-1 >= 0 && s[start-1] != ' ' {
+ i0 = start + len(docpath) + 1
+ continue
+ }
var end int
for end = start + len(docpath); end < len(s); end++ {
if s[end] == ' ' {
@@ -21,8 +28,9 @@ func replaceDocPath(s string) string {
}
}
- text := s[start+len(docpath) : end]
+ text := s[start:end]
s = s[:start] + fmt.Sprintf("[%s](//github.com/go-delve/delve/tree/master/%s)", text, text) + s[end:]
+ i0 = end + 1
}
} | terminal: remove leftover GOPATH references (#<I>)
Remove leftover references to $GOPATH in documentation, change script
that generates markdown documentation to look for substrings that start
with "Documentation/" instead. | go-delve_delve | train |
5c64583e92a8875af5f9d0178012993d587553c7 | diff --git a/lib/geoengineer/gps/yaml_tag.rb b/lib/geoengineer/gps/yaml_tag.rb
index <HASH>..<HASH> 100644
--- a/lib/geoengineer/gps/yaml_tag.rb
+++ b/lib/geoengineer/gps/yaml_tag.rb
@@ -97,8 +97,17 @@ class GeoEngineer::GPS::YamlTag::Sub < GeoEngineer::GPS::YamlTag
result_value = value.dup
all_queries.each do |query|
result = finder.dereference!(query, { auto_load: false })
+
+ # Process result differently per type
+ case result
+ when Array
+ result = result.to_json
+ when Hash
+ result = result.to_json
+ end
result_value.gsub!(/{{\s*#{query}\s*}}/, result)
end
+
result_value.to_json
end | [Fix] allow lists in sub | coinbase_geoengineer | train |
adf9721985be73ad4aefaaf5ec0167245c935855 | diff --git a/check_release_installation.py b/check_release_installation.py
index <HASH>..<HASH> 100644
--- a/check_release_installation.py
+++ b/check_release_installation.py
@@ -90,6 +90,7 @@ class Connection:
def __init__(self, target_host):
self.target_host = target_host
def run_checked(self, command):
+ print(command)
result = self.run(command)
result.assert_succesful()
def run(self, *user_command): | Make check_release_installation.py print each command to be run | andreafrancia_trash-cli | train |
685c948d90dd35ffd9154e4926363ec3e2e868ff | diff --git a/contracts/networks.js b/contracts/networks.js
index <HASH>..<HASH> 100644
--- a/contracts/networks.js
+++ b/contracts/networks.js
@@ -149,5 +149,6 @@ export default [
{ name: 'ropsten', networkId: 3, contracts: null },
{ name: 'rinkeby', networkId: 4, contracts: null },
{ name: 'kovan', networkId: 42, contracts: contractInfo('kovan') },
+ { name: 'test', networkId: 1337, contracts: contractInfo('test') },
{ name: 'test', networkId: TESTNET_ID, contracts: contractInfo('test') }
];
diff --git a/src/eth/SmartContractService.js b/src/eth/SmartContractService.js
index <HASH>..<HASH> 100644
--- a/src/eth/SmartContractService.js
+++ b/src/eth/SmartContractService.js
@@ -122,6 +122,7 @@ export default class SmartContractService extends PrivateService {
if (this._addedContracts) {
const networkName = {
[TESTNET_ID]: 'testnet',
+ [1337]: 'testnet',
[42]: 'kovan',
[1]: 'mainnet'
}[id]; | add geth network id to networks config | makerdao_dai.js | train |
092c5669af2ad6f1346567d4c71f6a60e9d287a5 | diff --git a/library/actor-js/src/main/java/im/actor/model/js/angular/AngularList.java b/library/actor-js/src/main/java/im/actor/model/js/angular/AngularList.java
index <HASH>..<HASH> 100644
--- a/library/actor-js/src/main/java/im/actor/model/js/angular/AngularList.java
+++ b/library/actor-js/src/main/java/im/actor/model/js/angular/AngularList.java
@@ -71,6 +71,15 @@ public class AngularList<T extends JavaScriptObject, V extends BserObject & List
callbacks.remove(callback);
}
+ public void forceReconvert(long id) {
+ for (int i = 0; i < values.size(); i++) {
+ if (values.get(i).getEngineId() == id) {
+ remove(jsValues, i);
+ insert(jsValues, i, entityConverter.convert(values.get(i), messenger));
+ }
+ }
+ }
+
public void forceReconvert() {
clear(jsValues);
diff --git a/library/actor-js/src/main/java/im/actor/model/js/angular/AngularModule.java b/library/actor-js/src/main/java/im/actor/model/js/angular/AngularModule.java
index <HASH>..<HASH> 100644
--- a/library/actor-js/src/main/java/im/actor/model/js/angular/AngularModule.java
+++ b/library/actor-js/src/main/java/im/actor/model/js/angular/AngularModule.java
@@ -167,16 +167,11 @@ public class AngularModule extends BaseModule implements AngularFileLoadedListen
@Override
public void onFileLoaded(long fileId) {
if (dialogsList != null) {
- boolean founded = false;
for (Dialog dialog : dialogsList.getRawItems()) {
if (checkAvatar(dialog.getDialogAvatar(), fileId)) {
- founded = true;
- break;
+ dialogsList.forceReconvert(dialog.getEngineId());
}
}
- if (founded) {
- dialogsList.forceReconvert();
- }
}
for (AngularList<JsMessage, Message> messageList : messagesList.values()) { | wip(js): Improving file loading performance | actorapp_actor-platform | train |
f7b630f81b3789a354f2b97a4773e90e3b60a54f | diff --git a/lib/mail/message.rb b/lib/mail/message.rb
index <HASH>..<HASH> 100644
--- a/lib/mail/message.rb
+++ b/lib/mail/message.rb
@@ -1502,6 +1502,7 @@ module Mail
end
def add_required_fields
+ add_multipart_mixed_header unless parts.empty?
@body = Mail::Body.new('') if body.nil?
add_message_id unless (has_message_id? || self.class == Mail::Part)
add_date unless has_date?
diff --git a/spec/mail/attachments_spec.rb b/spec/mail/attachments_spec.rb
index <HASH>..<HASH> 100644
--- a/spec/mail/attachments_spec.rb
+++ b/spec/mail/attachments_spec.rb
@@ -130,6 +130,15 @@ describe "Attachments" do
end
end
+
+ describe "setting the content type correctly" do
+ it "should set the content type to multipart/mixed if none given and you add an attachment" do
+ mail = Mail.new
+ mail.attachments['test.pdf'] = File.read(fixture('attachments', 'test.pdf'))
+ mail.encoded
+ mail.mime_type.should == 'multipart/mixed'
+ end
+ end
end | Message now adds multipart/mixed as the content type if nothing is set and there are parts to the message | mikel_mail | train |
bd9b2220572653195013d38d56f311f5f668dc37 | diff --git a/lib/patterns.rb b/lib/patterns.rb
index <HASH>..<HASH> 100644
--- a/lib/patterns.rb
+++ b/lib/patterns.rb
@@ -92,8 +92,6 @@ module Patterns
sidekiq-pro
graphql-pro
- # Blacklisted internal Rails dependency misspellings
- # which could be used for malicious purposes.
action-cable
action_cable
action-mailer | Remove comment from gem blacklist array
closed #<I> | rubygems_rubygems.org | train |
a4ce9b955eb7e495700fa841c26436d4ae3b7314 | diff --git a/src/swipe/index.js b/src/swipe/index.js
index <HASH>..<HASH> 100644
--- a/src/swipe/index.js
+++ b/src/swipe/index.js
@@ -1,5 +1,6 @@
// Utils
import { createNamespace } from '../utils';
+import { isHidden } from '../utils/dom/style';
import { preventDefault } from '../utils/dom/event';
import { doubleRaf } from '../utils/dom/raf';
import { range } from '../utils/format/number';
@@ -149,7 +150,7 @@ export default createComponent({
initialize(active = +this.initialSwipe) {
clearTimeout(this.timer);
- if (this.$el) {
+ if (this.$el && !isHidden(this.$el)) {
const rect = this.$el.getBoundingClientRect();
this.computedWidth = +this.width || rect.width;
this.computedHeight = +this.height || rect.height; | fix(Swipe): incorrect width when resize in invisible state (#<I>) | youzan_vant | train |
cda87fc468211987d808be08dc7ab06c059fa3e7 | diff --git a/js/load-image-orientation.js b/js/load-image-orientation.js
index <HASH>..<HASH> 100644
--- a/js/load-image-orientation.js
+++ b/js/load-image-orientation.js
@@ -119,7 +119,7 @@ Exif orientation values to correctly display the letter F:
ctx.scale(1, -1)
break
case 5:
- // vertical flip + 90 rotate right
+ // vertical flip + 90° rotate right
ctx.rotate(0.5 * Math.PI)
ctx.scale(1, -1)
break
@@ -129,7 +129,7 @@ Exif orientation values to correctly display the letter F:
ctx.translate(0, -height)
break
case 7:
- // horizontal flip + 90 rotate right
+ // horizontal flip + 90° rotate right
ctx.rotate(0.5 * Math.PI)
ctx.translate(width, -height)
ctx.scale(-1, 1)
@@ -185,7 +185,7 @@ Exif orientation values to correctly display the letter F:
newOptions.bottom = options.top
break
case 5:
- // vertical flip + 90 rotate right
+ // vertical flip + 90° rotate right
newOptions.left = options.top
newOptions.top = options.left
newOptions.right = options.bottom
@@ -199,7 +199,7 @@ Exif orientation values to correctly display the letter F:
newOptions.bottom = options.left
break
case 7:
- // horizontal flip + 90 rotate right
+ // horizontal flip + 90° rotate right
newOptions.left = options.bottom
newOptions.top = options.right
newOptions.right = options.top | Add missing degree character in source comments. | blueimp_JavaScript-Load-Image | train |
b49fe32d8491558173113beb087980acde99107f | diff --git a/tests/CacheTest.php b/tests/CacheTest.php
index <HASH>..<HASH> 100644
--- a/tests/CacheTest.php
+++ b/tests/CacheTest.php
@@ -3,10 +3,10 @@
namespace ApiClients\Tools\CommandBus;
use ApiClients\Tools\TestUtilities\TestCase;
+use ExceptionalJSON\DecodeErrorException;
use Generator;
use Test\App\Commands\AwesomesauceCommand;
use Test\App\Handlers\AwesomesauceHandler;
-use WyriHaximus\Tactician\CommandHandler\Annotations\Handler;
use function WyriHaximus\iteratorOrArrayToArray;
/**
@@ -33,6 +33,8 @@ final class CacheTest extends TestCase
public function testReadNonExistent(): void
{
+ self::expectException(DecodeErrorException::class);
+
$tmpFile = $this->getTmpDir() . \bin2hex(\random_bytes(13)) . '.json';
\touch($tmpFile); | Set expected exception for CacheTest::testReadNonExistent | php-api-clients_command-bus | train |
59fb563e9ddda315ddb2bf39b1940602f2d39a94 | diff --git a/engines/bastion_katello/app/assets/javascripts/bastion_katello/activation-keys/details/activation-key-associations.controller.js b/engines/bastion_katello/app/assets/javascripts/bastion_katello/activation-keys/details/activation-key-associations.controller.js
index <HASH>..<HASH> 100644
--- a/engines/bastion_katello/app/assets/javascripts/bastion_katello/activation-keys/details/activation-key-associations.controller.js
+++ b/engines/bastion_katello/app/assets/javascripts/bastion_katello/activation-keys/details/activation-key-associations.controller.js
@@ -16,7 +16,7 @@
angular.module('Bastion.activation-keys').controller('ActivationKeyAssociationsController',
['$scope', '$location', 'translate', 'Nutupane', 'ActivationKey', 'ContentHostsHelper', 'CurrentOrganization', 'Host',
function ($scope, $location, translate, Nutupane, ActivationKey, ContentHostsHelper, CurrentOrganization, Host) {
- var contentHostsNutupane, params = {
+ var contentHostsNutupane, nutupaneParams, params = {
'organization_id': CurrentOrganization,
'search': $location.search().search || "",
'page': 1,
@@ -25,7 +25,11 @@ angular.module('Bastion.activation-keys').controller('ActivationKeyAssociationsC
'paged': true
};
- contentHostsNutupane = new Nutupane(Host, params);
+ nutupaneParams = {
+ 'disableAutoLoad': true
+ };
+
+ contentHostsNutupane = new Nutupane(Host, params, undefined, nutupaneParams);
$scope.controllerName = 'hosts';
contentHostsNutupane.searchTransform = function (term) {
var searchQuery, addition = "activation_key_id=" + $scope.$stateParams.activationKeyId;
diff --git a/engines/bastion_katello/app/assets/javascripts/bastion_katello/host-collections/details/host-collection-add-hosts.controller.js b/engines/bastion_katello/app/assets/javascripts/bastion_katello/host-collections/details/host-collection-add-hosts.controller.js
index <HASH>..<HASH> 100644
--- a/engines/bastion_katello/app/assets/javascripts/bastion_katello/host-collections/details/host-collection-add-hosts.controller.js
+++ b/engines/bastion_katello/app/assets/javascripts/bastion_katello/host-collections/details/host-collection-add-hosts.controller.js
@@ -16,7 +16,7 @@
angular.module('Bastion.host-collections').controller('HostCollectionAddHostsController',
['$scope', '$state', '$location', 'translate', 'Nutupane', 'CurrentOrganization', 'Host', 'HostCollection',
function ($scope, $state, $location, translate, Nutupane, CurrentOrganization, Host, HostCollection) {
- var contentNutupane, params;
+ var contentNutupane, params, nutupaneParams;
params = {
'organization_id': CurrentOrganization,
@@ -27,7 +27,10 @@ angular.module('Bastion.host-collections').controller('HostCollectionAddHostsCon
'paged': true
};
- contentNutupane = new Nutupane(Host, params);
+ nutupaneParams = {
+ 'disableAutoLoad': true
+ };
+ contentNutupane = new Nutupane(Host, params, undefined, nutupaneParams);
$scope.controllerName = 'hosts';
contentNutupane.searchTransform = function (term) {
var addition = "-host_collection_id=" + $scope.$stateParams.hostCollectionId;
diff --git a/engines/bastion_katello/app/assets/javascripts/bastion_katello/host-collections/details/host-collection-hosts.controller.js b/engines/bastion_katello/app/assets/javascripts/bastion_katello/host-collections/details/host-collection-hosts.controller.js
index <HASH>..<HASH> 100644
--- a/engines/bastion_katello/app/assets/javascripts/bastion_katello/host-collections/details/host-collection-hosts.controller.js
+++ b/engines/bastion_katello/app/assets/javascripts/bastion_katello/host-collections/details/host-collection-hosts.controller.js
@@ -16,7 +16,7 @@
angular.module('Bastion.host-collections').controller('HostCollectionHostsController',
['$scope', '$location', 'translate', 'Nutupane', 'CurrentOrganization', 'Host', 'HostCollection',
function ($scope, $location, translate, Nutupane, CurrentOrganization, Host, HostCollection) {
- var params;
+ var params, nutupaneParams;
params = {
'organization_id': CurrentOrganization,
@@ -27,7 +27,11 @@ angular.module('Bastion.host-collections').controller('HostCollectionHostsContro
'paged': true
};
- $scope.contentNutupane = new Nutupane(Host, params);
+ nutupaneParams = {
+ 'disableAutoLoad': true
+ };
+
+ $scope.contentNutupane = new Nutupane(Host, params, undefined, nutupaneParams);
$scope.controllerName = 'hosts';
$scope.contentNutupane.searchTransform = function (term) {
var addition = "host_collection_id=" + $scope.$stateParams.hostCollectionId; | fixes #<I> - Host Collection/Host views not filtering correctly | Katello_katello | train |
0583248ecfef5bec8efeaa8e65907f1e9ad0bf1d | diff --git a/lib/paru/selector.rb b/lib/paru/selector.rb
index <HASH>..<HASH> 100644
--- a/lib/paru/selector.rb
+++ b/lib/paru/selector.rb
@@ -187,8 +187,10 @@ module Paru
def is_descendant?(node)
distance = 0
+ parent = nil
begin
distance += 1 if @distance > 0
+ node = parent unless parent.nil?
parent = node.parent
ancestry = parent.type == @type and @classes.all? {|c| parent.has_class? c}
end while not ancestry and not parent.is_root? and distance <= @distance | fix `Relation#is_descendent?` looping forever
if the immediate parent didn't match, this used to loop forever; now it
correctly walks up the tree looking for a matching parent. | htdebeer_paru | train |
8e7b0c3e6390f483fdaacb4c812996a57d97ca87 | diff --git a/gems/aws-sdk-code-generator/lib/aws-sdk-code-generator/dsl/attribute_accessor.rb b/gems/aws-sdk-code-generator/lib/aws-sdk-code-generator/dsl/attribute_accessor.rb
index <HASH>..<HASH> 100644
--- a/gems/aws-sdk-code-generator/lib/aws-sdk-code-generator/dsl/attribute_accessor.rb
+++ b/gems/aws-sdk-code-generator/lib/aws-sdk-code-generator/dsl/attribute_accessor.rb
@@ -4,25 +4,30 @@ module AwsSdkCodeGenerator
include Dsl::CodeObject
- def initialize(name, &block)
+ def initialize(name, api_private: false, &block)
@name = name
- @return_tags = []
+ @tags = []
yield(self) if block
+ api_private! if api_private
end
attr_reader :name
def returns(type, docstring:nil)
- @return_tags << ReturnTag.new(type:type, docstring:docstring)
+ @tags << ReturnTag.new(type:type, docstring:docstring)
+ end
+
+ def api_private!
+ @tags << "# @api private"
end
def documented?
- !@return_tags.empty?
+ [email protected]?
end
def lines
lines = []
- @return_tags.each do |tag|
+ @tags.each do |tag|
lines.concat(tag.lines)
end
lines + ["#{macro} :#{@name}"]
diff --git a/gems/aws-sdk-code-generator/lib/aws-sdk-code-generator/dsl/module.rb b/gems/aws-sdk-code-generator/lib/aws-sdk-code-generator/dsl/module.rb
index <HASH>..<HASH> 100644
--- a/gems/aws-sdk-code-generator/lib/aws-sdk-code-generator/dsl/module.rb
+++ b/gems/aws-sdk-code-generator/lib/aws-sdk-code-generator/dsl/module.rb
@@ -54,8 +54,8 @@ module AwsSdkCodeGenerator
@code_objects << AutoloadStatement.new(const_name, path)
end
- def attr_reader(name, &block)
- a = Dsl::AttributeReader.new(name)
+ def attr_reader(name, options = {}, &block)
+ a = Dsl::AttributeReader.new(name, options)
yield(a) if block
add(a)
end
diff --git a/gems/aws-sdk-code-generator/lib/aws-sdk-code-generator/generators/client_class.rb b/gems/aws-sdk-code-generator/lib/aws-sdk-code-generator/generators/client_class.rb
index <HASH>..<HASH> 100644
--- a/gems/aws-sdk-code-generator/lib/aws-sdk-code-generator/generators/client_class.rb
+++ b/gems/aws-sdk-code-generator/lib/aws-sdk-code-generator/generators/client_class.rb
@@ -46,7 +46,7 @@ module AwsSdkCodeGenerator
apply_waiter_methods(self)
eigenclass do |eigenclass|
eigenclass.docstring('@api private')
- eigenclass.attr_reader('identifier')
+ eigenclass.attr_reader('identifier', api_private: true)
eigenclass.method('errors_module') do |m|
m.code('Errors')
end | Client class .identifier methods are now marked @api private. | aws_aws-sdk-ruby | train |
63b1cf1fe2d3ed62299f57cdd90890de7745b4b8 | diff --git a/auth/ldap/auth.php b/auth/ldap/auth.php
index <HASH>..<HASH> 100644
--- a/auth/ldap/auth.php
+++ b/auth/ldap/auth.php
@@ -478,44 +478,43 @@ class auth_plugin_ldap extends auth_plugin_base {
$droptablesql = array(); /// sql commands to drop the table (because session scope could be a problem for
/// some persistent drivers like ODBTP (mssql) or if this function is invoked
/// from within a PHP application using persistent connections
+ $temptable = $CFG->prefix . 'extuser';
+ $createtemptablesql = '';
// configure a temp table
print "Configuring temp table\n";
switch (strtolower($CFG->dbfamily)) {
case 'mysql':
- $temptable = $CFG->prefix . 'extuser';
$droptablesql[] = 'DROP TEMPORARY TABLE ' . $temptable; // sql command to drop the table (because session scope could be a problem)
- execute_sql_arr($droptablesql, true, false); /// Drop temp table to avoid persistence problems later
- echo "Creating temp table $temptable\n";
- execute_sql('CREATE TEMPORARY TABLE ' . $temptable . ' (username VARCHAR(64), PRIMARY KEY (username)) TYPE=MyISAM', false);
+ $createtemptablesql = 'CREATE TEMPORARY TABLE ' . $temptable . ' (username VARCHAR(64), PRIMARY KEY (username)) TYPE=MyISAM';
break;
case 'postgres':
- $temptable = $CFG->prefix . 'extuser';
$droptablesql[] = 'DROP TABLE ' . $temptable; // sql command to drop the table (because session scope could be a problem)
- execute_sql_arr($droptablesql, true, false); /// Drop temp table to avoid persistence problems later
- echo "Creating temp table $temptable\n";
$bulk_insert_records = 1; // no support for multiple sets of values
- execute_sql('CREATE TEMPORARY TABLE '. $temptable . ' (username VARCHAR(64), PRIMARY KEY (username))', false);
+ $createtemptablesql = 'CREATE TEMPORARY TABLE '. $temptable . ' (username VARCHAR(64), PRIMARY KEY (username))';
break;
case 'mssql':
- $temptable = '#'.$CFG->prefix . 'extuser'; /// MSSQL temp tables begin with #
+ $temptable = '#'. $temptable; /// MSSQL temp tables begin with #
$droptablesql[] = 'DROP TABLE ' . $temptable; // sql command to drop the table (because session scope could be a problem)
- execute_sql_arr($droptablesql, true, false); /// Drop temp table to avoid persistence problems later
- echo "Creating temp table $temptable\n";
$bulk_insert_records = 1; // no support for multiple sets of values
- execute_sql('CREATE TABLE ' . $temptable . ' (username VARCHAR(64), PRIMARY KEY (username))', false);
+ $createtemptablesql = 'CREATE TABLE ' . $temptable . ' (username VARCHAR(64), PRIMARY KEY (username))';
break;
case 'oracle':
- $temptable = $CFG->prefix . 'extuser';
$droptablesql[] = 'TRUNCATE TABLE ' . $temptable; // oracle requires truncate before being able to drop a temp table
$droptablesql[] = 'DROP TABLE ' . $temptable; // sql command to drop the table (because session scope could be a problem)
- execute_sql_arr($droptablesql, true, false); /// Drop temp table to avoid persistence problems later
- echo "Creating temp table $temptable\n";
$bulk_insert_records = 1; // no support for multiple sets of values
- execute_sql('CREATE GLOBAL TEMPORARY TABLE '.$temptable.' (username VARCHAR(64), PRIMARY KEY (username)) ON COMMIT PRESERVE ROWS', false);
+ $createtemptablesql = 'CREATE GLOBAL TEMPORARY TABLE '.$temptable.' (username VARCHAR(64), PRIMARY KEY (username)) ON COMMIT PRESERVE ROWS';
break;
}
+
+ execute_sql_arr($droptablesql, true, false); /// Drop temp table to avoid persistence problems later
+ echo "Creating temp table $temptable\n";
+ if(! execute_sql($createtemptablesql, false) ){
+ print "Failed to create temporary users table - aborting\n";
+ exit;
+ }
+
print "Connecting to ldap...\n";
$ldapconnection = $this->ldap_connect(); | MDL-<I> - abort early when can't create temporary tables when syning users
from LDAP | moodle_moodle | train |
095cf6d37a33da2b64d45802cb81be9476a07de6 | diff --git a/src/DocBlock/Tags/Generic.php b/src/DocBlock/Tags/Generic.php
index <HASH>..<HASH> 100644
--- a/src/DocBlock/Tags/Generic.php
+++ b/src/DocBlock/Tags/Generic.php
@@ -53,7 +53,7 @@ class Generic extends BaseTag implements Factory\StaticMethod
Assert::stringNotEmpty($name);
Assert::notNull($descriptionFactory);
- $description = $descriptionFactory && $body ? $descriptionFactory->create($body, $context) : null;
+ $description = $descriptionFactory && $body !== "" ? $descriptionFactory->create($body, $context) : null;
return new static($name, $description);
} | fixed generic tag to parse description with `0` as expected | phpDocumentor_ReflectionDocBlock | train |
bbc26e716010e48871e692436a4e2b22e259cc01 | diff --git a/deliver/lib/deliver/download_screenshots.rb b/deliver/lib/deliver/download_screenshots.rb
index <HASH>..<HASH> 100644
--- a/deliver/lib/deliver/download_screenshots.rb
+++ b/deliver/lib/deliver/download_screenshots.rb
@@ -13,7 +13,7 @@ module Deliver
end
def self.download(options, folder_path)
- v = options[:use_live_version] ? options[:app].live_version : options[:app].latest_version
+ v = options[:use_live_version] ? options[:app].live_version(platform: options[:platform]) : options[:app].latest_version(platform: options[:platform])
v.screenshots.each do |language, screenshots|
screenshots.each do |screenshot|
diff --git a/deliver/lib/deliver/options.rb b/deliver/lib/deliver/options.rb
index <HASH>..<HASH> 100644
--- a/deliver/lib/deliver/options.rb
+++ b/deliver/lib/deliver/options.rb
@@ -372,6 +372,10 @@ module Deliver
description: "Metadata: Localised privacy url",
optional: true,
is_string: false),
+ FastlaneCore::ConfigItem.new(key: :apple_tv_privacy_policy,
+ description: "Metadata: Localised Apple TV privacy policy text",
+ optional: true,
+ is_string: false),
FastlaneCore::ConfigItem.new(key: :support_url,
description: "Metadata: Localised support url",
optional: true,
diff --git a/deliver/lib/deliver/upload_metadata.rb b/deliver/lib/deliver/upload_metadata.rb
index <HASH>..<HASH> 100644
--- a/deliver/lib/deliver/upload_metadata.rb
+++ b/deliver/lib/deliver/upload_metadata.rb
@@ -10,7 +10,7 @@ module Deliver
NON_LOCALISED_VERSION_VALUES = [:copyright]
# Localised app details values
- LOCALISED_APP_VALUES = [:name, :subtitle, :privacy_url]
+ LOCALISED_APP_VALUES = [:name, :subtitle, :privacy_url, :apple_tv_privacy_policy]
# Non localized app details values
NON_LOCALISED_APP_VALUES = [:primary_category, :secondary_category,
diff --git a/deliver/lib/deliver/upload_screenshots.rb b/deliver/lib/deliver/upload_screenshots.rb
index <HASH>..<HASH> 100644
--- a/deliver/lib/deliver/upload_screenshots.rb
+++ b/deliver/lib/deliver/upload_screenshots.rb
@@ -42,7 +42,7 @@ module Deliver
Helper.show_loading_indicator("Activating #{lng_text} #{enabled_languages.join(', ')}...")
v.save!
# This refreshes the app version from iTC after enabling a localization
- v = app.edit_version
+ v = app.edit_version(platform: options[:platform])
Helper.hide_loading_indicator
end | apple tv privacy policy upload and download handling added (#<I>)
screenshot download and upload issue for platform appletvos when ios and appletvos both are present at appstoreconnect for same bundle id | fastlane_fastlane | train |
675504a68650f0919f609f50ac5c22ea7ff2a098 | diff --git a/caniusepython3/test/test_cli.py b/caniusepython3/test/test_cli.py
index <HASH>..<HASH> 100644
--- a/caniusepython3/test/test_cli.py
+++ b/caniusepython3/test/test_cli.py
@@ -149,11 +149,19 @@ class CLITests(unittest.TestCase):
'its transition:')
self.assertEqual(messages[1], want)
- def test_message_no_blockers(self):
+ @mock.patch('sys.stdout', autospec=True)
+ def test_message_no_blockers_flair_on_utf8_terminal(self, mock_stdout):
+ mock_stdout.encoding = 'UTF-8'
messages = ciu_main.message([])
- self.assertEqual(
- ['You have 0 projects blocking you from using Python 3!'],
- messages)
+ expected = ['\U0001f389 You have 0 projects blocking you from using Python 3!']
+ self.assertEqual(expected, messages)
+
+ @mock.patch('sys.stdout', autospec=True)
+ def test_message_no_blockers(self, mock_stdout):
+ mock_stdout.encoding = None
+ messages = ciu_main.message([])
+ expected = ['You have 0 projects blocking you from using Python 3!']
+ self.assertEqual(expected, messages)
def test_pprint_blockers(self):
simple = [['A']] | Fix failing CLI test on UTF-8 terminals. | brettcannon_caniusepython3 | train |
0a5cc8d7cf641384c4e39df5b6905471b1eef454 | diff --git a/src/sjl.js b/src/sjl.js
index <HASH>..<HASH> 100644
--- a/src/sjl.js
+++ b/src/sjl.js
@@ -212,38 +212,83 @@
});
}
+ /**
+ * Checs if value is a valid number (also checks if isNaN so that you don't have to).
+ * @param value {*}
+ * @returns {Boolean}
+ */
function isNumber (value) {
return classOfIs(value, Number);
}
+ /**
+ * Returns whether a value is a function or not.
+ * @param value {*}
+ * @returns {Boolean}
+ */
function isFunction (value) {
return classOfIs(value, Function);
}
+ /**
+ * Checks if value is an array.
+ * @param value {*}
+ * @returns {boolean}
+ */
function isArray (value) {
return Array.isArray(value);
}
+ /**
+ * Checks if value is a boolean.
+ * @param value {*}
+ * @returns {Boolean}
+ */
function isBoolean (value) {
return classOfIs(value, Boolean);
}
+ /**
+ * Checks whether value is an object or not.
+ * @param value
+ * @returns {Boolean}
+ */
function isObject (value) {
return classOfIs(value, Object);
}
+ /**
+ * Checks whether value is a string or not.
+ * @param value {*}
+ * @returns {Boolean}
+ */
function isString(value) {
return classOfIs(value, String);
}
+ /**
+ * Checks if value is undefined.
+ * @param value {*}
+ * @returns {Boolean}
+ */
function isUndefined (value) {
return classOfIs(value, 'Undefined');
}
+ /**
+ * Checks if value is null.
+ * @param value {*}
+ * @returns {Boolean}
+ */
function isNull (value) {
return classOfIs(value, 'Null');
}
+ /**
+ * Checks if value is a `Symbol`.
+ * @param value {*}
+ * @returns {Boolean}
+ */
function isSymbol (value) {
return classOfIs(value, 'Symbol');
} | Added missing docs to src/sjl.js. | elycruz_sjljs | train |
2a34ebb19e8dfdad36bf6a143a3aa935a7edc77f | diff --git a/indra/preassembler/__init__.py b/indra/preassembler/__init__.py
index <HASH>..<HASH> 100644
--- a/indra/preassembler/__init__.py
+++ b/indra/preassembler/__init__.py
@@ -207,35 +207,39 @@ class Preassembler(object):
if not stmt.supports]
return self.related_stmts
- def check_sequence(self):
- """Iterate over all Statements and check whether references to
+ def check_statements(self):
+ for st in self.stmts:
+ print st
+ self.check_sequence(st)
+
+ @staticmethod
+ def check_sequence(st):
+ """Check whether references to
residues and sequence positions are consistent with sequence
information in the UniProt database"""
# TODO: extend to all Agent modifications
- for st in self.stmts:
- if isinstance(st, Phosphorylation):
- # Skip statements with no specified position
- if st.mod_pos is None:
- continue
- # Is looking at the first element enough?
- sub_id = st.sub.db_refs['UP']
- if not isinstance(sub_id, basestring):
- sub_id = sub_id[0]
- sub = uniprot_client.query_protein(sub_id)
- if st.mod == 'PhosphorylationSerine':
- residue = 'S'
- elif st.mod == 'PhosphorylationThreonine':
- residue = 'T'
- elif st.mod == 'PhosphorylationTyrosine':
- residue = 'Y'
- else:
- continue
- ver = uniprot_client.verify_location(sub, residue,
- location=st.mod_pos)
- if ver is False:
- print '%s:' % st
- print '-> Sequence check failed; position %s on %s is not %s.' %\
- (st.mod_pos, st.sub.name, residue)
+ if isinstance(st, Phosphorylation):
+ # Skip statements with no specified position
+ if st.mod_pos is None:
+ return
+ # Is looking at the first element enough?
+ sub_id = st.sub.db_refs['UP']
+ if not isinstance(sub_id, basestring):
+ sub_id = sub_id[0]
+ sub = uniprot_client.query_protein(sub_id)
+ if st.mod == 'PhosphorylationSerine':
+ residue = 'S'
+ elif st.mod == 'PhosphorylationThreonine':
+ residue = 'T'
+ elif st.mod == 'PhosphorylationTyrosine':
+ residue = 'Y'
+ else:
+ return
+ ver = uniprot_client.verify_location(sub, residue,
+ location=st.mod_pos)
+ if ver is False:
+ print '-> Sequence check failed; position %s on %s is not %s.' %\
+ (st.mod_pos, st.sub.name, residue) | Change check statement loop structure in Preassembler | sorgerlab_indra | train |
fd9bfbf28e045362c83620dd888c5edbb6f39a89 | diff --git a/lib/docx/docx-reader.js b/lib/docx/docx-reader.js
index <HASH>..<HASH> 100644
--- a/lib/docx/docx-reader.js
+++ b/lib/docx/docx-reader.js
@@ -100,26 +100,23 @@ function findPartPaths(docxFile) {
readElement: relationshipsReader.readRelationships,
defaultValue: relationshipsReader.defaultValue
})(docxFile).then(function(documentRelationships) {
- function findPartRelatedToMainDocument(options) {
+ function findPartRelatedToMainDocument(name) {
return findPartPath({
docxFile: docxFile,
relationships: documentRelationships,
- relationshipType: options.relationshipType,
+ relationshipType: "http://schemas.openxmlformats.org/officeDocument/2006/relationships/" + name,
basePath: zipfile.splitPath(mainDocumentPath).dirname,
- fallbackPath: options.fallbackPath
+ fallbackPath: "word/" + name + ".xml"
});
}
return {
mainDocument: mainDocumentPath,
- comments: findPartRelatedToMainDocument({
- relationshipType: "http://schemas.openxmlformats.org/officeDocument/2006/relationships/comments",
- fallbackPath: "word/comments.xml"
- }),
- endnotes: "word/endnotes.xml",
- footnotes: "word/footnotes.xml",
- numbering: "word/numbering.xml",
- styles: "word/styles.xml"
+ comments: findPartRelatedToMainDocument("comments"),
+ endnotes: findPartRelatedToMainDocument("endnotes"),
+ footnotes: findPartRelatedToMainDocument("footnotes"),
+ numbering: findPartRelatedToMainDocument("numbering"),
+ styles: findPartRelatedToMainDocument("styles")
};
});
});
diff --git a/test/docx/docx-reader.tests.js b/test/docx/docx-reader.tests.js
index <HASH>..<HASH> 100644
--- a/test/docx/docx-reader.tests.js
+++ b/test/docx/docx-reader.tests.js
@@ -88,6 +88,7 @@ test("error is thrown when main document part does not exist", function() {
});
+
test("part paths", {
"main document part is found using package relationships": function() {
var relationships = xml.element("r:Relationships", {}, [
@@ -113,35 +114,59 @@ test("part paths", {
return docxReader._findPartPaths(docxFile).then(function(partPaths) {
assert.equal(partPaths.mainDocument, "word/document.xml");
});
+ }
+});
+
+[
+ {
+ name: "comments",
+ type: "http://schemas.openxmlformats.org/officeDocument/2006/relationships/comments",
},
-
- "comments part is found using main document relationships": function() {
+ {
+ name: "endnotes",
+ type: "http://schemas.openxmlformats.org/officeDocument/2006/relationships/endnotes",
+ },
+ {
+ name: "footnotes",
+ type: "http://schemas.openxmlformats.org/officeDocument/2006/relationships/footnotes",
+ },
+ {
+ name: "numbering",
+ type: "http://schemas.openxmlformats.org/officeDocument/2006/relationships/numbering",
+ },
+ {
+ name: "styles",
+ type: "http://schemas.openxmlformats.org/officeDocument/2006/relationships/styles",
+ }
+].forEach(function(options) {
+ test(options.name + " part is found using main document relationships", function() {
var docxFile = createFakeDocxFile({
"_rels/.rels": createPackageRelationships("word/document.xml"),
"word/document.xml": " ",
"word/_rels/document.xml.rels": xml.writeString(xml.element("r:Relationships", {}, [
xml.element("r:Relationship", {
- "Type": "http://schemas.openxmlformats.org/officeDocument/2006/relationships/comments",
- "Target": "comments2.xml"
+ "Type": options.type,
+ "Target": "target-path.xml"
})
]), relationshipNamespaces),
- "word/comments2.xml": " "
+ "word/target-path.xml": " "
});
return docxReader._findPartPaths(docxFile).then(function(partPaths) {
- assert.equal(partPaths.comments, "word/comments2.xml");
+ assert.equal(partPaths[options.name], "word/target-path.xml");
});
- },
-
- "word/comments.xml is used as fallback location for comments part": function() {
- var docxFile = createFakeDocxFile({
+ });
+
+ test("word/" + options.name + ".xml is used as fallback location for " + options.name + " part", function() {
+ var zipContents = {
"_rels/.rels": createPackageRelationships("word/document.xml"),
- "word/document.xml": " ",
- "word/comments.xml": " "
- });
+ "word/document.xml": " "
+ };
+ zipContents["word/" + options.name + ".xml"] = " ";
+ var docxFile = createFakeDocxFile(zipContents);
return docxReader._findPartPaths(docxFile).then(function(partPaths) {
- assert.equal(partPaths.comments, "word/comments.xml");
+ assert.equal(partPaths[options.name], "word/" + options.name + ".xml");
});
- }
+ });
}); | Find parts related to main document using main document relationships | mwilliamson_mammoth.js | train |
47739b0772942057a8ba5bbc037d1e1e064a1b8c | diff --git a/lib/level.js b/lib/level.js
index <HASH>..<HASH> 100644
--- a/lib/level.js
+++ b/lib/level.js
@@ -1,6 +1,6 @@
'use strict';
-exports.ALL = -100000;
+exports.ALL = -Infinity;
/**
* Debug log for execute tracing
@@ -31,4 +31,4 @@ exports.WARN = 2;
*/
exports.ERROR = 3;
-exports.NONE = 100000;
+exports.NONE = Infinity; | use Infinity to improve performance and semantics (#7) | eggjs_egg-logger | train |
5604ce7d661cca2d8e7a9f5c03d5c5a2d11a16e9 | diff --git a/bosh-director/lib/bosh/director/client.rb b/bosh-director/lib/bosh/director/client.rb
index <HASH>..<HASH> 100644
--- a/bosh-director/lib/bosh/director/client.rb
+++ b/bosh-director/lib/bosh/director/client.rb
@@ -22,9 +22,13 @@ module Bosh::Director
end
def method_missing(method_name, *args)
- retries = @retry_methods[method_name] || 0
+ send_message(method_name, *args)
+ end
+
+ def send_message(message_name, *args)
+ retries = @retry_methods[message_name] || 0
begin
- handle_method(method_name, args)
+ handle_method(message_name, args)
rescue RpcTimeout
if retries > 0
retries -= 1
diff --git a/bosh-director/spec/unit/agent_client_spec.rb b/bosh-director/spec/unit/agent_client_spec.rb
index <HASH>..<HASH> 100644
--- a/bosh-director/spec/unit/agent_client_spec.rb
+++ b/bosh-director/spec/unit/agent_client_spec.rb
@@ -2,7 +2,56 @@ require 'spec_helper'
module Bosh::Director
describe AgentClient do
- describe 'long-runnings messages' do
+ shared_examples_for 'a long running message' do |message_name|
+ describe "##{message_name}" do
+ let(:task) do
+ {
+ 'agent_task_id' => 'fake-agent_task_id',
+ 'state' => 'running',
+ 'value' => 'task value'
+ }
+ end
+
+ before do
+ client.stub(send_message: task)
+ client.stub(:get_task) do
+ task['state'] = 'no longer running'
+ task
+ end
+
+ client.stub(:sleep).with(AgentClient::DEFAULT_POLL_INTERVAL)
+ end
+
+ it 'explicitly defines methods for long running messages (to poll)' do
+ expect(client).to respond_to(message_name)
+ end
+
+ it 'decorates the original send_message implementation' do
+ client.public_send(message_name, 'fake', 'args')
+
+ expect(client).to have_received(:send_message).with(message_name, 'fake', 'args')
+ end
+
+ it 'periodically polls the task while it is running' do
+ client.public_send(message_name, 'fake', 'args')
+
+ expect(client).to have_received(:get_task).with('fake-agent_task_id')
+ end
+
+ it 'stops polling once the task is no longer running' do
+ task['state'] = 'something other than running'
+ client.public_send(message_name, 'fake', 'args')
+
+ expect(client).not_to have_received(:get_task)
+ end
+
+ it 'returns the task value' do
+ expect(client.public_send(message_name, 'fake', 'args')).to eq('task value')
+ end
+ end
+ end
+
+ describe 'long running messages' do
let(:vm) do
instance_double('Models::Vm', credentials: nil)
end
@@ -17,19 +66,15 @@ module Bosh::Director
Api::ResourceManager.stub(:new)
end
- it 'explicitly defines methods for long running messages (to poll their tasks)' do
- expect(client.methods).to include(
- :prepare,
- :apply,
- :compile_package,
- :drain,
- :fetch_logs,
- :migrate_disk,
- :mount_disk,
- :stop,
- :unmount_disk,
- )
- end
+ include_examples 'a long running message', :prepare
+ include_examples 'a long running message', :apply
+ include_examples 'a long running message', :compile_package
+ include_examples 'a long running message', :drain
+ include_examples 'a long running message', :fetch_logs
+ include_examples 'a long running message', :migrate_disk
+ include_examples 'a long running message', :mount_disk
+ include_examples 'a long running message', :stop
+ include_examples 'a long running message', :unmount_disk
end
describe 'ping <=> pong' do | Backfill spec for precarious Client <=> AgentClient interaction | cloudfoundry_bosh | train |
ef10e301da290a0569c65932487fc33d8bf06605 | diff --git a/troposphere/kinesis.py b/troposphere/kinesis.py
index <HASH>..<HASH> 100644
--- a/troposphere/kinesis.py
+++ b/troposphere/kinesis.py
@@ -3,10 +3,17 @@
#
# See LICENSE file for full license.
-from . import AWSObject, Tags
+from . import AWSObject, AWSProperty, Tags
from .validators import integer
+class StreamEncryption(AWSProperty):
+ props = {
+ 'EncryptionType': (basestring, True),
+ 'KeyId': (basestring, True),
+ }
+
+
class Stream(AWSObject):
resource_type = "AWS::Kinesis::Stream"
@@ -14,5 +21,6 @@ class Stream(AWSObject):
'Name': (basestring, False),
'RetentionPeriodHours': (integer, False),
'ShardCount': (integer, False),
+ 'StreamEncryption': (StreamEncryption, False),
'Tags': ((Tags, list), False),
} | Add StreamEncryption to AWS::Kinesis::Stream | cloudtools_troposphere | train |
ac8e56288291723bfdcb08353391c2852a8c8185 | diff --git a/can-dom-events-test.js b/can-dom-events-test.js
index <HASH>..<HASH> 100644
--- a/can-dom-events-test.js
+++ b/can-dom-events-test.js
@@ -216,3 +216,30 @@ unit.test('domEvents.addDelegateListener call inner-most handler first (#62)', f
domEvents.dispatch(child, 'click');
done();
});
+
+unit.test('domEvents.addDelegateListener should have a working ev.stopPropagation (#62)', function (assert) {
+ var done = assert.async();
+ var grandparent = document.createElement('div');
+ var parent = document.createElement('p');
+ var child = document.createElement('input');
+
+ grandparent.appendChild(parent);
+ parent.appendChild(child);
+
+ var paragraphClickHandler = function paragraphClickHandler() {
+ domEvents.removeDelegateListener(grandparent, 'click', 'p', paragraphClickHandler);
+ assert.ok(false, 'ev.stopPropagation works');
+ };
+
+ var inputClickHandler = function inputClickHandler(event) {
+ event.stopPropagation();
+ domEvents.removeDelegateListener(grandparent, 'click', 'input', inputClickHandler);
+ assert.ok(true, 'inner-most click handler called first');
+ };
+
+ domEvents.addDelegateListener(grandparent, 'click', 'p', paragraphClickHandler);
+ domEvents.addDelegateListener(grandparent, 'click', 'input', inputClickHandler);
+
+ domEvents.dispatch(child, 'click');
+ done();
+});
diff --git a/helpers/-make-delegate-event-tree.js b/helpers/-make-delegate-event-tree.js
index <HASH>..<HASH> 100644
--- a/helpers/-make-delegate-event-tree.js
+++ b/helpers/-make-delegate-event-tree.js
@@ -20,6 +20,10 @@ function makeDelegator (domEvents) {
"can.setKeyValue": function(eventType, handlersBySelector){
var handler = this.delegated[eventType] = function(ev){
var cur = ev.target;
+ var propagate = true;
+ ev.stopPropagation = function() {
+ propagate = false;
+ };
do {
// document does not implement `.matches` but documentElement does
var el = cur === document ? document.documentElement : cur;
@@ -38,7 +42,7 @@ function makeDelegator (domEvents) {
// we need to continue using `cur` as the loop pointer, otherwhise
// it will never end as documentElement.parentNode === document
cur = cur.parentNode;
- } while (cur && cur !== ev.currentTarget);
+ } while ((cur && cur !== ev.currentTarget) && propagate);
};
this.events[eventType] = handlersBySelector;
domEvents.addEventListener(this.element, eventType, handler, useCapture(eventType)); | making ev.stopPropagation work with delegate listeners | canjs_can-dom-events | train |
cdac35e05752f657a36e95a0deae8d2a1d0e4b46 | diff --git a/pyemma/coordinates/tests/test_readers.py b/pyemma/coordinates/tests/test_readers.py
index <HASH>..<HASH> 100644
--- a/pyemma/coordinates/tests/test_readers.py
+++ b/pyemma/coordinates/tests/test_readers.py
@@ -123,7 +123,7 @@ class TestReaders(six.with_metaclass(GenerateTestMatrix, unittest.TestCase)):
'xtc',
'trr',
'dcd',
- 'h5',
+ 'h5', # pyemma H5Reader
'csv',
)
# transform data or not (identity does not change the data, but pushes it through a StreamingTransformer).
@@ -179,6 +179,16 @@ class TestReaders(six.with_metaclass(GenerateTestMatrix, unittest.TestCase)):
def teardown_class(cls):
shutil.rmtree(cls.tempdir, ignore_errors=True)
+ # H5Reader does not need a topology (and gets only selected in case we do not pass it).
+ def setUp(self):
+ if 'h5' in self._testMethodName:
+ self.pdb_file_bak = self.pdb_file
+ self.pdb_file = None
+
+ def tearDown(self):
+ if 'h5' in self._testMethodName:
+ self.pdb_file = self.pdb_file_bak
+
def _test_lagged_reader(self, file_format, stride, skip, chunksize, lag):
trajs = self.test_trajs[file_format]
reader = coor.source(trajs, top=self.pdb_file, chunksize=chunksize) | [test-readers] do no pass topology, if H5Reader is required. | markovmodel_PyEMMA | train |
3990e564abe116fca96d3dc393cb181f6d098834 | diff --git a/test/secure/publish-to-s3.js b/test/secure/publish-to-s3.js
index <HASH>..<HASH> 100644
--- a/test/secure/publish-to-s3.js
+++ b/test/secure/publish-to-s3.js
@@ -3,15 +3,6 @@ var frauPublisher = require('../../src/publisher'),
eventStream = require('event-stream'),
gUtil = require('gulp-util');
-var publisher = frauPublisher.app({
- id: 'frau-publisher-test',
- creds: {
- key: '***REMOVED***',
- secret: process.env.CREDS_SECRET
- },
- devTag: Math.random().toString(16).slice(2)
-});
-
var content = 'some data';
var filename = 'test.txt';
var file = new gUtil.File({
@@ -26,6 +17,8 @@ var file = new gUtil.File({
describe('publisher', function() {
it('should publish new file', function(cb) {
+ var publisher = createPublisher( Math.random().toString(16).slice(2) );
+
eventStream.readArray( [file] )
.pipe( publisher.getStream() )
.on('end', function() {
@@ -45,4 +38,31 @@ describe('publisher', function() {
}, 500);
});
});
+
+ it('should not overwrite a file', function(cb) {
+ // This test relies on files having already been published with this
+ // devTag. We could remove this dependency by publishing a file and
+ // then trying again. We should do this if necessary, but it didn't
+ // start that way because it seems unnecessarily wasteful.
+ var publisher = createPublisher( 'overwrite-test' );
+
+ eventStream.readArray( [file] )
+ .pipe( publisher.getStream() )
+ .on('error', function() {
+ cb();
+ }).on('end', function() {
+ cb('should not have published');
+ });
+ });
});
+
+function createPublisher(devTag) {
+ return frauPublisher.app({
+ id: 'frau-publisher-test',
+ creds: {
+ key: '***REMOVED***',
+ secret: process.env.CREDS_SECRET
+ },
+ devTag: devTag
+ });
+} | Add an end-to-end overwrite test. | Brightspace_frau-publisher | train |
41177eb1c8b8327ad966b5e12555f9846d51d9de | diff --git a/code/libraries/koowa/libraries/controller/behavior/abstract.php b/code/libraries/koowa/libraries/controller/behavior/abstract.php
index <HASH>..<HASH> 100644
--- a/code/libraries/koowa/libraries/controller/behavior/abstract.php
+++ b/code/libraries/koowa/libraries/controller/behavior/abstract.php
@@ -39,4 +39,25 @@ abstract class KControllerBehaviorAbstract extends KBehaviorAbstract
return $methods;
}
+
+ /**
+ * Command handler
+ *
+ * @param KCommandInterface $command The command
+ * @param KCommandChainInterface $chain The chain executing the command
+ * @return mixed If a handler breaks, returns the break condition. Returns the result of the handler otherwise.
+ */
+ public function execute(KCommandInterface $command, KCommandChainInterface $chain)
+ {
+ $parts = explode('.', $command->getName());
+ $method = '_'.$parts[0].ucfirst($parts[1]);
+
+ if($parts[0] == 'action') {
+ $result = $this->$method($command);
+ } else {
+ $result = parent::execute($command, $chain);
+ }
+
+ return $result;
+ }
} | re #<I> : Added specialised rule for handling 'action' methods in controller behaviors. | timble_kodekit | train |
a1a4b1b2beda0125e496704fff5c0d207db6ad94 | diff --git a/src/file_config/handlers/_common.py b/src/file_config/handlers/_common.py
index <HASH>..<HASH> 100644
--- a/src/file_config/handlers/_common.py
+++ b/src/file_config/handlers/_common.py
@@ -19,13 +19,13 @@ class BaseHandler(abc.ABC):
:rtype: module
"""
- if not hasattr(self, "_imported"):
+ if not hasattr(self, "_imported") or self._imported is None:
self._imported = self._discover_import()
return self._imported
@property
def handler(self):
- if not hasattr(self, "_handler"):
+ if not hasattr(self, "_handler") or self._handler is None:
self._handler = sys.modules[self.imported]
return self._handler
@@ -72,7 +72,7 @@ class BaseHandler(abc.ABC):
return True
return False
- def _discover_import(self):
+ def _discover_import(self, prefer=None):
""" Discovers and imports the best available module from ``packages``.
:raises ModuleNotFoundError: If no module is available
@@ -80,7 +80,11 @@ class BaseHandler(abc.ABC):
:rtype: str
"""
- for module_name in self.packages:
+ available_packages = self.packages
+ if isinstance(prefer, str):
+ available_packages = (prefer,)
+
+ for module_name in available_packages:
spec = importlib.util.find_spec(module_name)
if spec is not None:
importlib.import_module(module_name)
@@ -88,18 +92,31 @@ class BaseHandler(abc.ABC):
if callable(imported_hook):
imported_hook(sys.modules[module_name])
return module_name
- raise ModuleNotFoundError(f"no modules in {self.packages!r} found")
+ raise ModuleNotFoundError(f"no modules in {available_packages!r} found")
def _prefer_package(self, package):
- if package not in self.packages:
- raise ValueError(
- f"prefered package {package!r} does not exist, "
- f"allowed are {self.packages!r}"
- )
- # NOTE: this is a semi-dangerous property override that needs to be done since
- # packages are dynamically defined as part of the subclass but the super needs
- # to be able to override them no matter what
- self.packages = (package,)
+ """ Prefer a serializtion handler over other handlers.
+
+ :param str package: The name of the package to use
+ :raises ValueError: When the given package name is not one of the available
+ supported serializtion packages for this handler
+ :return: The name of the serialization handler
+ :rtype: str
+ """
+
+ if isinstance(package, str) and package != self.imported:
+ if package not in self.packages:
+ raise ValueError(
+ f"preferred package {package!r} does not exist, allowed are "
+ f"{self.packages!r}"
+ )
+ # clear out current serialization handler (if exists)
+ if hasattr(self, "_handler"):
+ del self._handler
+ # manually update imported handlers with a given preference
+ self._imported = self._discover_import(prefer=package)
+ return package
+ return self.imported
def dumps(self, config, instance, prefer=None, **kwargs):
""" An abstract dumps method which dumps an instance into the subclasses format.
@@ -112,10 +129,8 @@ class BaseHandler(abc.ABC):
:rtype: str
"""
- if isinstance(prefer, str):
- self._prefer_package(prefer)
-
- dumps_hook_name = f"on_{self.imported}_dumps"
+ dumper = self._prefer_package(prefer)
+ dumps_hook_name = f"on_{dumper}_dumps"
dumps_hook = getattr(self, dumps_hook_name, None)
if not callable(dumps_hook):
raise ValueError(
@@ -144,10 +159,8 @@ class BaseHandler(abc.ABC):
:rtype: dict
"""
- if isinstance(prefer, str):
- self._prefer_package(prefer)
-
- loads_hook_name = f"on_{self.imported}_loads"
+ loader = self._prefer_package(prefer)
+ loads_hook_name = f"on_{loader}_loads"
loads_hook = getattr(self, loads_hook_name, None)
if not callable(loads_hook):
raise ValueError(
diff --git a/src/file_config/handlers/json.py b/src/file_config/handlers/json.py
index <HASH>..<HASH> 100644
--- a/src/file_config/handlers/json.py
+++ b/src/file_config/handlers/json.py
@@ -9,7 +9,7 @@ class JSONHandler(BaseHandler):
"""
name = "json"
- packages = ("rapidjson", "ujson", "json")
+ packages = ("rapidjson", "json", "ujson")
options = {"indent": None, "sort_keys": False}
def on_json_dumps(self, json, config, dictionary, **kwargs):
diff --git a/tests/handlers/test_yaml.py b/tests/handlers/test_yaml.py
index <HASH>..<HASH> 100644
--- a/tests/handlers/test_yaml.py
+++ b/tests/handlers/test_yaml.py
@@ -30,7 +30,6 @@ def test_yaml_ordereddict():
content = instance.dumps_yaml()
assert content == dedent(
"""\
- foo:
- test: test
+ foo: {test: test}
"""
) | Fixing prefer keyword handler package management | stephen-bunn_file-config | train |
98797911fe3b860c000a38e0c042e0e1881ab7ae | diff --git a/worker/deployer/nested_test.go b/worker/deployer/nested_test.go
index <HASH>..<HASH> 100644
--- a/worker/deployer/nested_test.go
+++ b/worker/deployer/nested_test.go
@@ -366,8 +366,12 @@ func (s *NestedContextSuite) deployThreeUnits(c *gc.C, ctx deployer.Context) {
for {
units := report["units"].(map[string]interface{})
workers := units["workers"].(map[string]interface{})
+
+ first := workers["first/0"].(map[string]interface{})
+ second := workers["second/0"].(map[string]interface{})
third := workers["third/0"].(map[string]interface{})
- if third["state"] == "started" {
+
+ if first["state"] == "started" && second["state"] == "started" && third["state"] == "started" {
break
}
select { | Fixes nested deployer report test to ensure that all workers are
reported as starting, not just the third, which does not guarantee that
the others are up.
This can fail on slow machines and has been observed on arm<I> test jobs. | juju_juju | train |
389468f234c6a82dd5bfffae9f1c8daeee058aa0 | diff --git a/media/src/main/java/com/microsoft/windowsazure/services/media/implementation/VersionHeadersFilter.java b/media/src/main/java/com/microsoft/windowsazure/services/media/implementation/VersionHeadersFilter.java
index <HASH>..<HASH> 100644
--- a/media/src/main/java/com/microsoft/windowsazure/services/media/implementation/VersionHeadersFilter.java
+++ b/media/src/main/java/com/microsoft/windowsazure/services/media/implementation/VersionHeadersFilter.java
@@ -42,7 +42,7 @@ public class VersionHeadersFilter extends IdempotentClientFilter
MultivaluedMap<String, Object> headers = cr.getHeaders();
headers.add("DataServiceVersion", "3.0");
headers.add("MaxDataServiceVersion", "3.0");
- headers.add("x-ms-version", "2.2");
+ headers.add("x-ms-version", "2.5");
return getNext().handle(cr);
}
}
diff --git a/media/src/test/java/com/microsoft/windowsazure/services/media/IntegrationTestBase.java b/media/src/test/java/com/microsoft/windowsazure/services/media/IntegrationTestBase.java
index <HASH>..<HASH> 100644
--- a/media/src/test/java/com/microsoft/windowsazure/services/media/IntegrationTestBase.java
+++ b/media/src/test/java/com/microsoft/windowsazure/services/media/IntegrationTestBase.java
@@ -49,6 +49,8 @@ import com.microsoft.windowsazure.services.media.models.ListResult;
import com.microsoft.windowsazure.services.media.models.Locator;
import com.microsoft.windowsazure.services.media.models.LocatorInfo;
import com.microsoft.windowsazure.services.media.models.LocatorType;
+import com.microsoft.windowsazure.services.media.models.MediaProcessorInfo;
+import com.microsoft.windowsazure.services.media.models.MediaProcessor;
import com.microsoft.windowsazure.services.queue.QueueConfiguration;
import com.microsoft.windowsazure.services.queue.QueueContract;
import com.microsoft.windowsazure.services.queue.QueueService;
@@ -72,9 +74,8 @@ public abstract class IntegrationTestBase
protected static final String validButNonexistAssetId = "nb:cid:UUID:0239f11f-2d36-4e5f-aa35-44d58ccc0973";
protected static final String validButNonexistAccessPolicyId = "nb:pid:UUID:38dcb3a0-ef64-4ad0-bbb5-67a14c6df2f7";
protected static final String validButNonexistLocatorId = "nb:lid:UUID:92a70402-fca9-4aa3-80d7-d4de3792a27a";
-
- protected static final String MEDIA_ENCODER_MEDIA_PROCESSOR_2_2_0_0_ID = "nb:mpid:UUID:70bdc2c3-ebf4-42a9-8542-5afc1e55d217";
-
+
+ protected static String MEDIA_ENCODER_MEDIA_PROCESSOR_2_2_0_0_ID = "nb:mpid:UUID:2e7aa8f3-4961-4e0c-b4db-0e0439e524f5";
protected static final String invalidId = "notAValidId";
@Rule
@@ -101,6 +102,26 @@ public abstract class IntegrationTestBase
cleanupEnvironment();
}
+
+ protected static void getLatestMediaProcessorID() throws Exception {
+ ListResult<MediaProcessorInfo> listMediaProcessorsResult = service.list(MediaProcessor.list());
+ List<MediaProcessorInfo> ps = listMediaProcessorsResult;
+ double latestVersion = 0.0;
+ int position = -1;
+ for (int i = 0; i < ps.size(); i++) {
+ MediaProcessorInfo mediaProcessorInfo = ps.get(i);
+ double d = Double.parseDouble(mediaProcessorInfo.getVersion());
+ if (d > latestVersion) {
+ latestVersion = d;
+ position = i;
+ }
+ }
+
+ if (position != -1)
+ {
+ MEDIA_ENCODER_MEDIA_PROCESSOR_2_2_0_0_ID = ps.get(position).getId();
+ }
+ }
protected static void overrideWithEnv(Configuration config, String key)
{ | replace hardcoded media processor id with dynamic latest one
and upgrade sdk version from <I> to <I>, as media service will deprecate
the <I> version | Azure_azure-sdk-for-java | train |
3d05aa85ea26ef8206a1da053f62e636aba7384a | diff --git a/activity_stream/templatetags/activity_stream_tags.py b/activity_stream/templatetags/activity_stream_tags.py
index <HASH>..<HASH> 100644
--- a/activity_stream/templatetags/activity_stream_tags.py
+++ b/activity_stream/templatetags/activity_stream_tags.py
@@ -31,9 +31,13 @@ def users_activity_stream(context, user, count, offset=0):
offset=0
activity_items = ActivityStreamItem.objects.filter(actor=user,
subjects__isnull=False,
- created_at__lte=datetime.datetime.now()).order_by('-created_at').distinct()[offset:count]
- return {"activity_items": activity_items, "user": context["user"],
- "request":context.get("request")}
+ created_at__lte=datetime.datetime.now())\
+ .order_by('-created_at').distinct()[offset:count]
+
+ return {"activity_items": activity_items,
+ "user": context["user"],
+ "request":context.get("request")
+ }
@register.inclusion_tag("activity_stream/friends_activity_stream.html", takes_context=True)
@@ -48,7 +52,10 @@ def following_activity_stream(context, user, count, offset=0):
following = get_people_i_follow(user, 1000)
following = list(following)
following.append(user)
- activity_items = ActivityStreamItem.objects.filter(actor__in=following, subjects__isnull=False, created_at__lte=datetime.datetime.now()).order_by('-created_at').distinct()[offset:count]
+ activity_items = ActivityStreamItem.objects.filter(actor__in=following, \
+ subjects__isnull=False, \
+ created_at__lte=datetime.datetime.now())\
+ .order_by('-created_at').distinct()[offset:count]
return {"activity_items": activity_items, "user": context["user"],
"request":context.get("request")}
@@ -63,7 +70,8 @@ def global_activity_stream(context, count, offset=0, privacylevel=0):
offset=0
activity_items = ActivityStreamItem.objects.filter(subjects__isnull=False,
- created_at__lte=datetime.datetime.now()).order_by('-created_at').distinct()[offset:count]
+ created_at__lte=datetime.datetime.now())\
+ .order_by('-created_at').distinct()[offset:count]
return {"activity_items": activity_items, "user": context["user"],
"request":context.get("request")}
@@ -100,7 +108,7 @@ class IsFollowingNode(Node):
from_user = self.from_user.resolve(context)
if to_user and from_user:
if to_user.is_authenticated() and from_user.is_authenticated():
- is_following = ActivityFollower.objects.filter(to_user=to_user, from_user=from_user).count()
+ is_following = ActivityFollower.objects.filter(to_user=to_user,from_user=from_user).count()
if is_following:
return self.node_true.render(context)
else:
diff --git a/activity_stream/tests.py b/activity_stream/tests.py
index <HASH>..<HASH> 100644
--- a/activity_stream/tests.py
+++ b/activity_stream/tests.py
@@ -28,12 +28,12 @@ class StoryTest(TestCase):
self.assertTrue(TestSubject.objects.get(pk=photo.id))
activityItem2 = create_activity_item("placed", User.objects.get(username="admin"), photo2)
- items = users_activity_stream(User.objects.get(username="admin"),1000)
+ items = users_activity_stream({}, User.objects.get(username="admin"),1000)
self.assertEquals(len(items['activity_items']), 1)
photo2.delete()
- items = users_activity_stream(User.objects.get(username="admin"),1000)
+ items = users_activity_stream({}, User.objects.get(username="admin"),1000)
self.assertEquals(len(items['activity_items']), 0)
def test_batching(self):
@@ -66,7 +66,7 @@ class StoryTest(TestCase):
self.assertTrue(activityItem)
self.assertEquals(activityItem.is_batched, False)
self.assertEquals(activityItem.subjects.count(), 1)
- items = users_activity_stream(User.objects.get(username="admin"),1000)
+ items = users_activity_stream({}, User.objects.get(username="admin"),1000)
self.assertEquals(len(items['activity_items']), 0)
activityItem.delete() | manually merging changes to fix tests from daonb, making it a bit more pep8 | philippWassibauer_django-activity-stream | train |
191608a3b219a63eca8a6e9b85d556101d2b35ef | diff --git a/lib/claide/command.rb b/lib/claide/command.rb
index <HASH>..<HASH> 100644
--- a/lib/claide/command.rb
+++ b/lib/claide/command.rb
@@ -80,7 +80,7 @@ module CLAide
#
attr_accessor :description
- # @return [Array<String>] The prefixes used t osearch for CLAide plugins.
+ # @return [Array<String>] The prefixes used to search for CLAide plugins.
# Plugins are loaded via their `<plugin_prefix>_plugin.rb` file.
# Defaults to search for `claide` plugins.
#
@@ -233,7 +233,7 @@ module CLAide
#
# The subclass has to combine the result of calling `super` and its own
# list of options. The recommended way of doing this is by concatenating
- # concatenating to this classes’ own options.
+ # to this classes’ own options.
#
# @return [Array<Array>]
# | two typos I noticed today while reading the code | CocoaPods_CLAide | train |
2e47674a466c6e75df2a8a649a5baae24b1b4e4e | diff --git a/tests/func/test_install.py b/tests/func/test_install.py
index <HASH>..<HASH> 100644
--- a/tests/func/test_install.py
+++ b/tests/func/test_install.py
@@ -3,6 +3,7 @@ import pathlib
import sys
import pytest
+from git import GitCommandError
from dvc.exceptions import GitHookAlreadyExistsError
from dvc.utils import file_md5
@@ -35,18 +36,25 @@ class TestInstall:
with pytest.raises(GitHookAlreadyExistsError):
scm.install()
- def test_post_checkout(self, tmp_dir, scm, dvc):
+ def test_pre_commit_hook(self, tmp_dir, scm, dvc, caplog):
+ tmp_dir.dvc_gen("file", "file content", commit="create foo")
+ tmp_dir.gen("file", "file modified")
scm.install()
- tmp_dir.dvc_gen({"file": "file content"}, commit="add")
+ # scm.commit bypasses hooks
+ with pytest.raises(GitCommandError, match=r"modified:\s*file"):
+ scm.repo.git.commit(m="file modified")
+
+ def test_post_checkout(self, tmp_dir, scm, dvc):
+ tmp_dir.dvc_gen({"file": "file content"}, commit="add")
os.unlink("file")
+ scm.install()
+
scm.checkout("new_branch", create_new=True)
assert os.path.isfile("file")
def test_pre_push_hook(self, tmp_dir, scm, dvc, tmp_path_factory):
- scm.install()
-
temp = tmp_path_factory.mktemp("external")
git_remote = temp / "project.git"
storage_path = temp / "dvc_storage"
@@ -64,6 +72,8 @@ class TestInstall:
scm.repo.clone(os.fspath(git_remote))
scm.repo.create_remote("origin", os.fspath(git_remote))
+ scm.install()
+
assert not expected_storage_path.is_file()
scm.repo.git.push("origin", "master")
assert expected_storage_path.is_file()
@@ -74,16 +84,19 @@ class TestInstall:
sys.platform == "win32", reason="Git hooks aren't supported on Windows"
)
def test_merge_driver_no_ancestor(tmp_dir, scm, dvc):
- scm.commit("init")
- scm.install()
- (tmp_dir / ".gitattributes").write_text("*.dvc merge=dvc")
- scm.checkout("one", create_new=True)
- tmp_dir.dvc_gen({"data": {"foo": "foo"}}, commit="one: add data")
+ with tmp_dir.branch("one", new=True):
+ tmp_dir.dvc_gen({"data": {"foo": "foo"}}, commit="one: add data")
- scm.checkout("master")
scm.checkout("two", create_new=True)
+ dvc.checkout() # keep things in sync
+
tmp_dir.dvc_gen({"data": {"bar": "bar"}}, commit="two: add data")
+ # installing hook only before merge, as it runs `dvc` commands which makes
+ # `checkouts` and `commits` above slower
+ scm.install()
+ (tmp_dir / ".gitattributes").write_text("*.dvc merge=dvc")
+
scm.repo.git.merge("one", m="merged", no_gpg_sign=True, no_signoff=True)
# NOTE: dvc shouldn't checkout automatically as it might take a long time
@@ -102,18 +115,21 @@ def test_merge_driver_no_ancestor(tmp_dir, scm, dvc):
sys.platform == "win32", reason="Git hooks aren't supported on Windows"
)
def test_merge_driver(tmp_dir, scm, dvc):
- scm.commit("init")
- scm.install()
- (tmp_dir / ".gitattributes").write_text("*.dvc merge=dvc")
tmp_dir.dvc_gen({"data": {"master": "master"}}, commit="master: add data")
- scm.checkout("one", create_new=True)
- tmp_dir.dvc_gen({"data": {"one": "one"}}, commit="one: add data")
+ with tmp_dir.branch("one", new=True):
+ tmp_dir.dvc_gen({"data": {"one": "one"}}, commit="one: add data")
- scm.checkout("master")
scm.checkout("two", create_new=True)
+ dvc.checkout() # keep things in sync
+
tmp_dir.dvc_gen({"data": {"two": "two"}}, commit="two: add data")
+ # installing hook only before merge, as it runs `dvc` commands on
+ # `checkouts` and `commits` which slows tests down
+ scm.install()
+ (tmp_dir / ".gitattributes").write_text("*.dvc merge=dvc")
+
scm.repo.git.merge("one", m="merged", no_gpg_sign=True, no_signoff=True)
# NOTE: dvc shouldn't checkout automatically as it might take a long time | tests: install hook just before using it (#<I>)
Before, we were installing hooks at the start which was making checkouts
slower. With this change, it should improve these tests by twice as much. | iterative_dvc | train |
fb5eb1fa7787815144828fee0f0938c7273e1e51 | diff --git a/executor/analyze_test.go b/executor/analyze_test.go
index <HASH>..<HASH> 100644
--- a/executor/analyze_test.go
+++ b/executor/analyze_test.go
@@ -371,7 +371,7 @@ func (s *testFastAnalyze) TestAnalyzeFastSample(c *C) {
samples := mockExec.Collectors[i].Samples
c.Assert(len(samples), Equals, 20)
for j := 1; j < 20; j++ {
- cmp, err := samples[j].Value.CompareDatum(tk.Se.GetSessionVars().StmtCtx, &samples[j-1].Value)
+ cmp, err := samples[j].Value.Compare(tk.Se.GetSessionVars().StmtCtx, &samples[j-1].Value, collate.GetBinaryCollator())
c.Assert(err, IsNil)
c.Assert(cmp, Greater, 0)
}
@@ -381,7 +381,7 @@ func (s *testFastAnalyze) TestAnalyzeFastSample(c *C) {
func checkHistogram(sc *stmtctx.StatementContext, hg *statistics.Histogram) (bool, error) {
for i := 0; i < len(hg.Buckets); i++ {
lower, upper := hg.GetLower(i), hg.GetUpper(i)
- cmp, err := upper.CompareDatum(sc, lower)
+ cmp, err := upper.Compare(sc, lower, collate.GetBinaryCollator())
if cmp < 0 || err != nil {
return false, err
}
@@ -389,7 +389,7 @@ func checkHistogram(sc *stmtctx.StatementContext, hg *statistics.Histogram) (boo
continue
}
previousUpper := hg.GetUpper(i - 1)
- cmp, err = lower.CompareDatum(sc, previousUpper)
+ cmp, err = lower.Compare(sc, previousUpper, collate.GetBinaryCollator())
if cmp <= 0 || err != nil {
return false, err
}
diff --git a/statistics/builder.go b/statistics/builder.go
index <HASH>..<HASH> 100644
--- a/statistics/builder.go
+++ b/statistics/builder.go
@@ -23,6 +23,7 @@ import (
"github.com/pingcap/tidb/sessionctx/stmtctx"
"github.com/pingcap/tidb/types"
"github.com/pingcap/tidb/util/codec"
+ "github.com/pingcap/tidb/util/collate"
)
// SortedBuilder is used to build histograms for PK and index.
@@ -67,7 +68,7 @@ func (b *SortedBuilder) Iterate(data types.Datum) error {
b.hist.NDV = 1
return nil
}
- cmp, err := b.hist.GetUpper(int(b.bucketIdx)).CompareDatum(b.sc, &data)
+ cmp, err := b.hist.GetUpper(int(b.bucketIdx)).Compare(b.sc, &data, collate.GetBinaryCollator())
if err != nil {
return errors.Trace(err)
}
@@ -158,7 +159,7 @@ func buildHist(sc *stmtctx.StatementContext, hg *Histogram, samples []*SampleIte
hg.AppendBucket(&samples[0].Value, &samples[0].Value, int64(sampleFactor), int64(ndvFactor))
for i := int64(1); i < sampleNum; i++ {
corrXYSum += float64(i) * float64(samples[i].Ordinal)
- cmp, err := hg.GetUpper(bucketIdx).CompareDatum(sc, &samples[i].Value)
+ cmp, err := hg.GetUpper(bucketIdx).Compare(sc, &samples[i].Value, collate.GetBinaryCollator())
if err != nil {
return 0, errors.Trace(err)
}
diff --git a/util/collate/collate.go b/util/collate/collate.go
index <HASH>..<HASH> 100644
--- a/util/collate/collate.go
+++ b/util/collate/collate.go
@@ -154,6 +154,11 @@ func GetCollator(collate string) Collator {
return binCollatorInstance
}
+// GetBinaryCollator gets the binary collator, it is often used when we want to apply binary compare.
+func GetBinaryCollator() Collator {
+ return binCollatorInstance
+}
+
// GetCollatorByID get the collator according to id, it will return the binary collator if the corresponding collator doesn't exist.
func GetCollatorByID(id int) Collator {
if atomic.LoadInt32(&newCollationEnabled) == 1 { | *: replace compareDatum by compare (#<I>) | pingcap_tidb | train |
d24641e0607b3f3839d10394784258a54c6f3384 | diff --git a/lib/veewee/provider/kvm/box/helper/status.rb b/lib/veewee/provider/kvm/box/helper/status.rb
index <HASH>..<HASH> 100644
--- a/lib/veewee/provider/kvm/box/helper/status.rb
+++ b/lib/veewee/provider/kvm/box/helper/status.rb
@@ -15,11 +15,11 @@ module Veewee
end
def exists_volume?
- [email protected](:name => "#{name}.img").nil?
+ @connection.list_volumes.find { |v| v[:name] == "#{name}.img" }
end
def exists_vm?
- [email protected](:name => name).nil?
+ @connection.list_domains.find { |d| d[:name] == name }
end
end # End Module | Properly determine if a volume or instance exists
Preious to this commit, volumes were determined to exist or not by
calling the Volumes#all method the :name parameter to filter only the
volumes that matched the value of :name. This was not properly
finding the volumes. This commit uses the #list_volumes method and
calls #find on the array returned.
Likewise for instances | jedi4ever_veewee | train |
e5c19e5bddb2738fbab26a4238eface70081f8d5 | diff --git a/src/TickGroup/TickGroup.js b/src/TickGroup/TickGroup.js
index <HASH>..<HASH> 100644
--- a/src/TickGroup/TickGroup.js
+++ b/src/TickGroup/TickGroup.js
@@ -44,7 +44,7 @@ export default class TickGroup extends PureComponent {
constructor(props) {
super(props);
- (this:any).removeTick = this.removeTick.bind(this);
+ (this:any).removeUDID = this.removeUDID.bind(this);
}
state = {
@@ -71,13 +71,13 @@ export default class TickGroup extends PureComponent {
const ticks = scale.ticks ? scale.ticks(tickCount) : [];
this.setState((prevState) => {
- const mapped = ticks.map((t) => ({ val: t }));
+ const mapped = ticks.map((tick) => ({ val: tick }));
const update = dataUpdate({ data: mapped, ...next }, prevState, this.removed);
return { ...update, prevScale: prev.scale, currScale: scale };
});
}
- removeTick(udid) {
+ removeUDID(udid) {
this.removed.set(udid, true);
}
@@ -92,8 +92,8 @@ export default class TickGroup extends PureComponent {
return (
<g className={className}>
- {state.nodes.map((tick) => {
- const udid = keyAccessor(tick);
+ {state.nodes.map((node) => {
+ const udid = keyAccessor(node);
const type = state.udids[udid];
return (
@@ -101,10 +101,10 @@ export default class TickGroup extends PureComponent {
key={udid}
udid={udid}
type={type}
- tick={tick}
- removeTick={this.removeTick}
+ node={node}
prevScale={state.prevScale}
currScale={state.currScale}
+ removeUDID={this.removeUDID}
{...props}
/>
); | modify TickGroup for using withTransitions | sghall_resonance | train |
ef0617af7d3558e99dcd5065633326cc54d28670 | diff --git a/src/Picqer/Financials/Moneybird/Entities/User.php b/src/Picqer/Financials/Moneybird/Entities/User.php
index <HASH>..<HASH> 100644
--- a/src/Picqer/Financials/Moneybird/Entities/User.php
+++ b/src/Picqer/Financials/Moneybird/Entities/User.php
@@ -25,13 +25,6 @@ class User extends Model
'language',
'time_zone',
'permissions',
- 'sales_invoices',
- 'documents',
- 'estimates',
- 'bank',
- 'settings',
- 'ownership',
- 'time_entries',
];
/** | Bugfix: invalid fillable entries
This commit fixes a minor bug where the possible permission entries were
accidentally added as fillable properties to the user object. | picqer_moneybird-php-client | train |
2fdb2ceeb890ad524df5e52db3dca101e9514441 | diff --git a/tests/unit/modules/test_file.py b/tests/unit/modules/test_file.py
index <HASH>..<HASH> 100644
--- a/tests/unit/modules/test_file.py
+++ b/tests/unit/modules/test_file.py
@@ -1,5 +1,3 @@
-# Import python libs
-
import os
import shutil
import tempfile
@@ -15,15 +13,11 @@ import salt.utils.files
import salt.utils.platform
import salt.utils.stringutils
from salt.exceptions import CommandExecutionError, SaltInvocationError
-
-# Import Salt libs
from salt.ext import six
from salt.utils.jinja import SaltCacheLoader
from tests.support.helpers import with_tempfile
from tests.support.mixins import LoaderModuleMockMixin
from tests.support.mock import DEFAULT, MagicMock, Mock, mock_open, patch
-
-# Import Salt Testing libs
from tests.support.runtests import RUNTIME_VARS
from tests.support.unit import TestCase, skipIf | run pre-commit: remove import comments | saltstack_salt | train |
f944c9c2d534185a0389027b04fd96986ba82e7c | diff --git a/ford/sourceform.py b/ford/sourceform.py
index <HASH>..<HASH> 100644
--- a/ford/sourceform.py
+++ b/ford/sourceform.py
@@ -2914,7 +2914,7 @@ class FortranBlockData(FortranContainer):
def process_attribs(self):
for item in self.types:
- if item.name.lower() in self.attr_dict:
+ for attr in self.attr_dict.get(item.name.lower(), []):
if "public" in self.attr_dict[item.name.lower()]:
item.permission = "public"
elif "private" in self.attr_dict[item.name.lower()]: | Fix for undefined variable in `FortranBlockData` | Fortran-FOSS-Programmers_ford | train |
73b5553316f8b80a5833d45518c54191ea293328 | diff --git a/runtimes/assemblyscript/cli-runner.js b/runtimes/assemblyscript/cli-runner.js
index <HASH>..<HASH> 100644
--- a/runtimes/assemblyscript/cli-runner.js
+++ b/runtimes/assemblyscript/cli-runner.js
@@ -14,18 +14,17 @@ async function main(runtime, module, input, expected) {
let exp = JSON.parse(fs.readFileSync(expected));
console.log("expected:");
console.log(JSON.stringify(exp, null, 2));
+ let res;
try {
- let res = await engine.invoke(rt, mod, "qcert_main", arg);
- console.log("output:");
- console.log(JSON.stringify(res, null, 2));
+ res = await engine.invoke(rt, mod, "qcert_main", arg);
} catch(err) {
- console.log("output:");
let res = {
"error": "Eval failed",
"message": err.message
};
- console.log(JSON.stringify(res, null,2));
}
+ console.log("output:");
+ console.log(JSON.stringify(res, null, 2));
}
const rt = process.argv[2]; | fix(wasm) more on cli-runner error | querycert_qcert | train |
3956f9cb9a80337c42269709f179539345b7d580 | diff --git a/joomla/filesystem/stream.php b/joomla/filesystem/stream.php
index <HASH>..<HASH> 100644
--- a/joomla/filesystem/stream.php
+++ b/joomla/filesystem/stream.php
@@ -123,7 +123,7 @@ class JStream extends JObject
* @return boolean
* @since 1.6
*/
- function open($filename, $mode='r', $use_include_path=false, $context=null, $use_prefix=true, $relative=false, $detectprocessingmode=false)
+ function open($filename, $mode='r', $use_include_path=false, $context=null, $use_prefix=false, $relative=false, $detectprocessingmode=false)
{
$filename = $this->_getFilename($filename, $mode, $use_prefix, $relative); | # [#<I>] Temp folder has to include JPATH_ROOT for gz extension upload to work. Thanks Sam.
--HG--
extra : convert_revision : svn%3A6f6e1ebd-4c2b-<I>-<I>f-f<I>bde<I>bce9/development/trunk/libraries%<I> | joomla_joomla-framework | train |
8f5a29a567f4425d0349fa6a8bd869062d799a45 | diff --git a/rqalpha/utils/log_capture.py b/rqalpha/utils/log_capture.py
index <HASH>..<HASH> 100644
--- a/rqalpha/utils/log_capture.py
+++ b/rqalpha/utils/log_capture.py
@@ -35,6 +35,7 @@ class LogCapture:
def __enter__(self):
self._handlers = self._logger.handlers
self._logger.handlers = [self._capture_handler]
+ return self
def __exit__(self, exc_type, exc_val, exc_tb):
self._logger.handlers = self._handlers | bugfix: work as the LogCapture Context Manager
AttributeError: 'NoneType' object has no attribute 'replay'
log_capture.replay() | ricequant_rqalpha | train |
2c3747740ad297d8da55d8c1296511c3776fce86 | diff --git a/src/Behat/Behat/EventDispatcher/Event/AfterStepTested.php b/src/Behat/Behat/EventDispatcher/Event/AfterStepTested.php
index <HASH>..<HASH> 100644
--- a/src/Behat/Behat/EventDispatcher/Event/AfterStepTested.php
+++ b/src/Behat/Behat/EventDispatcher/Event/AfterStepTested.php
@@ -115,7 +115,7 @@ final class AfterStepTested extends StepTested implements AfterTested
*/
public function hasOutput()
{
- return $this->isTeardownHasOutput() || $this->isResultHasException() || $this->isResultCallHasOutput();
+ return $this->teardownHasOutput() || $this->resultHasException() || $this->resultCallHasOutput();
}
/**
@@ -123,7 +123,7 @@ final class AfterStepTested extends StepTested implements AfterTested
*
* @return Boolean
*/
- private function isTeardownHasOutput()
+ private function teardownHasOutput()
{
return $this->teardown->hasOutput();
}
@@ -133,7 +133,7 @@ final class AfterStepTested extends StepTested implements AfterTested
*
* @return Boolean
*/
- private function isResultHasException()
+ private function resultHasException()
{
return $this->result instanceof ExceptionResult && $this->result->getException();
}
@@ -143,7 +143,7 @@ final class AfterStepTested extends StepTested implements AfterTested
*
* @return Boolean
*/
- private function isResultCallHasOutput()
+ private function resultCallHasOutput()
{
if (!$this->result instanceof ExecutedStepResult) {
return false; | Clarify AfterStepTested method names | Behat_Behat | train |
0e0ae6f89819d59174fc9e5565c090ee1b8c330b | diff --git a/lib/cucumber/runtime.rb b/lib/cucumber/runtime.rb
index <HASH>..<HASH> 100644
--- a/lib/cucumber/runtime.rb
+++ b/lib/cucumber/runtime.rb
@@ -46,7 +46,7 @@ module Cucumber
def initialize(configuration = Configuration.default)
@configuration = Configuration.new(configuration)
- @support_code = SupportCode.new(self, @configuration)
+ @support_code = SupportCode::CachesStepMatch.new SupportCode.new(self, @configuration)
@results = Formatter::LegacyApi::Results.new
end
diff --git a/lib/cucumber/runtime/support_code.rb b/lib/cucumber/runtime/support_code.rb
index <HASH>..<HASH> 100644
--- a/lib/cucumber/runtime/support_code.rb
+++ b/lib/cucumber/runtime/support_code.rb
@@ -152,18 +152,7 @@ module Cucumber
end
end
- def step_match(step_name, name_to_report=nil) #:nodoc:
- @match_cache ||= {}
-
- match = @match_cache[[step_name, name_to_report]]
- return match if match
-
- @match_cache[[step_name, name_to_report]] = step_match_without_cache(step_name, name_to_report)
- end
-
- private
-
- def step_match_without_cache(step_name, name_to_report=nil)
+ def step_match(step_name, name_to_report=nil)
matches = matches(step_name, name_to_report)
raise Undefined.new(step_name) if matches.empty?
matches = best_matches(step_name, matches) if matches.size > 1 && guess_step_matches?
@@ -171,6 +160,8 @@ module Cucumber
matches[0]
end
+ private
+
def guess_step_matches?
@configuration.guess?
end
@@ -204,6 +195,17 @@ module Cucumber
Cucumber.logger
end
+ require 'delegate'
+ class CachesStepMatch < SimpleDelegator
+ def step_match(step_name, name_to_report=nil) #:nodoc:
+ @match_cache ||= {}
+
+ match = @match_cache[[step_name, name_to_report]]
+ return match if match
+
+ @match_cache[[step_name, name_to_report]] = super(step_name, name_to_report)
+ end
+ end
end
end
end
diff --git a/lib/cucumber/step_definitions.rb b/lib/cucumber/step_definitions.rb
index <HASH>..<HASH> 100644
--- a/lib/cucumber/step_definitions.rb
+++ b/lib/cucumber/step_definitions.rb
@@ -2,7 +2,7 @@ module Cucumber
class StepDefinitions
def initialize(configuration = Configuration.default)
configuration = Configuration.new(configuration)
- @support_code = Runtime::SupportCode.new(nil, configuration)
+ @support_code = Runtime::SupportCode::CachesStepMatch.new Runtime::SupportCode.new(nil, configuration)
@support_code.load_files_from_paths(configuration.autoload_code_paths)
end
diff --git a/spec/cucumber/runtime/support_code_spec.rb b/spec/cucumber/runtime/support_code_spec.rb
index <HASH>..<HASH> 100644
--- a/spec/cucumber/runtime/support_code_spec.rb
+++ b/spec/cucumber/runtime/support_code_spec.rb
@@ -19,15 +19,19 @@ module Cucumber
expect(format).to eq "it [snows] in [april]"
end
- it "caches step match results" do
- dsl.Given(/it (.*) in (.*)/) { |what, month| }
+ context 'caching' do
+ let(:cached) { Runtime::SupportCode::CachesStepMatch.new(subject) }
+
+ it "caches step match results" do
+ dsl.Given(/it (.*) in (.*)/) { |what, month| }
- step_match = subject.step_match("it snows in april")
+ step_match = cached.step_match("it snows in april")
- expect(@rb).not_to receive(:step_matches)
- second_step_match = subject.step_match("it snows in april")
+ expect(@rb).not_to receive(:step_matches)
+ second_step_match = cached.step_match("it snows in april")
- expect(step_match).to equal(second_step_match)
+ expect(step_match).to equal(second_step_match)
+ end
end
describe "resolving step defintion matches" do | Extract step_match caching to a decorator.
I've left the creation of th cache explicit at construction for the
moment. There may be a better api we can use, but wanted to wait to see
if we can reuse it in plugins first. | cucumber_cucumber-ruby | train |
f43437b75881e51c1d3f72214858357a520172ee | diff --git a/pmxbot/commands.py b/pmxbot/commands.py
index <HASH>..<HASH> 100644
--- a/pmxbot/commands.py
+++ b/pmxbot/commands.py
@@ -806,14 +806,14 @@ def help(client, event, channel, nick, rest):
yield "command not found"
else:
def mk_entries():
- for handler in sorted(_handler_registry,
- key=operator.attrgetter('name')):
- if handler.type_ != 'command':
- continue
- res = "!%s" % handler.name
+ handlers = (handler for handler in _handler_registry
+ if type(handler) is pmxbot.core.CommandHandler)
+ handlers = sorted(handlers, key=operator.attrgetter('name'))
+ for handler in handlers:
+ res = "!" + handler.name
if handler.aliases:
- res += " (%s)" % ', '.join(
- alias.name for alias in handler.aliases)
+ alias_names = (alias.name for alias in handler.aliases)
+ res += " (%s)" % ', '.join(alias_names)
yield res
o = io.StringIO(u" ".join(mk_entries()))
more = o.read(160)
diff --git a/pmxbot/core.py b/pmxbot/core.py
index <HASH>..<HASH> 100644
--- a/pmxbot/core.py
+++ b/pmxbot/core.py
@@ -334,7 +334,7 @@ class CommandHandler(Handler):
def match(self, message, channel):
cmd, _, cmd_args = message.partition(' ')
- return cmd.lower() == '!{name}'.format(vars(self))
+ return cmd.lower() == '!{name}'.format(name = self.name)
def process(self, message):
cmd, _, cmd_args = message.partition(' ') | Fix help to no longer use type_ attribute | yougov_pmxbot | train |
3a574fb8bc0966c99713fd16d8ebe4e755c9320c | diff --git a/lib/rack/tracker/google_analytics/google_analytics.rb b/lib/rack/tracker/google_analytics/google_analytics.rb
index <HASH>..<HASH> 100644
--- a/lib/rack/tracker/google_analytics/google_analytics.rb
+++ b/lib/rack/tracker/google_analytics/google_analytics.rb
@@ -2,18 +2,16 @@ require 'ostruct'
class Rack::Tracker::GoogleAnalytics < Rack::Tracker::Handler
class Event < OpenStruct
+ def write
+ ['send', event].to_json.gsub(/\[|\]/, '')
+ end
+
def event
- {
- hitType: 'event',
- eventCategory: self.event_category,
- eventAction: self.event_action,
- eventLabel: self.event_label,
- eventValue: self.event_value
- }.compact
+ { hitType: 'event' }.merge(attributes).compact
end
- def write
- ['send', event].to_json.gsub(/\[|\]/, '')
+ def attributes
+ Hash[to_h.map { |k,v| ['event' + k.to_s.capitalize, v] }]
end
end
diff --git a/spec/handler/google_analytics_spec.rb b/spec/handler/google_analytics_spec.rb
index <HASH>..<HASH> 100644
--- a/spec/handler/google_analytics_spec.rb
+++ b/spec/handler/google_analytics_spec.rb
@@ -9,7 +9,7 @@ RSpec.describe Rack::Tracker::GoogleAnalytics do
def env
{'tracker' => {
'google_analytics' => [
- Rack::Tracker::GoogleAnalytics::Event.new(event_category: "Users", event_action: "Login", event_label: "Standard")
+ Rack::Tracker::GoogleAnalytics::Event.new(category: "Users", action: "Login", label: "Standard")
]
}}
end
@@ -23,7 +23,7 @@ RSpec.describe Rack::Tracker::GoogleAnalytics do
describe "with a event value" do
def env
{'tracker' => { 'google_analytics' => [
- Rack::Tracker::GoogleAnalytics::Event.new(event_category: "Users", event_action: "Login", event_label: "Standard", event_value: 5)
+ Rack::Tracker::GoogleAnalytics::Event.new(category: "Users", action: "Login", label: "Standard", value: 5)
]}}
end
diff --git a/spec/tracker/controller_spec.rb b/spec/tracker/controller_spec.rb
index <HASH>..<HASH> 100644
--- a/spec/tracker/controller_spec.rb
+++ b/spec/tracker/controller_spec.rb
@@ -9,7 +9,7 @@ class SomeController
def index
tracker do
- google_analytics event_category: 'foo'
+ google_analytics category: 'foo'
end
end
end
@@ -17,7 +17,7 @@ end
RSpec.describe Rack::Tracker::Controller do
context 'controller' do
- let(:event) { Rack::Tracker::GoogleAnalytics::Event.new(event_category: 'foo') }
+ let(:event) { Rack::Tracker::GoogleAnalytics::Event.new(category: 'foo') }
it 'writes the event into env' do
controller = SomeController.new | event attributes should be less verbose | railslove_rack-tracker | train |
fcf29edf3db164ec0e5220efc82598126daef122 | diff --git a/example.py b/example.py
index <HASH>..<HASH> 100644
--- a/example.py
+++ b/example.py
@@ -89,7 +89,7 @@ ticket = None
try:
ticket = client.getObject()
except Exception, e:
- print "Unable to retrieve ticket record: %", e
+ print "Unable to retrieve ticket record: ", e
# Now update the ticket.
update = {
@@ -98,6 +98,6 @@ update = {
try:
update = client.addUpdate(update)
- print "Update ticket 123456. The new update's id is %.", update{'id'}
+ print "Update ticket 123456. The new update's id is %s." % update['id']
except Exception, e:
- print "Unable to update ticket: %", e
+ print "Unable to update ticket: ", e | Fixed print usage in the example script. (Thanks, mahmoudimus!) | softlayer_softlayer-python | train |
e92d2e359b060bed0f48a553e094e4805e44951a | diff --git a/plugins/worker/server/__init__.py b/plugins/worker/server/__init__.py
index <HASH>..<HASH> 100644
--- a/plugins/worker/server/__init__.py
+++ b/plugins/worker/server/__init__.py
@@ -43,6 +43,18 @@ class CustomJobStatus(object):
PUSHING_OUTPUT = 823
CANCELING = 824
+ valid_transitions = {
+ FETCHING_INPUT: [JobStatus.RUNNING],
+ CONVERTING_INPUT: [JobStatus.RUNNING, FETCHING_INPUT],
+ CONVERTING_OUTPUT: [JobStatus.RUNNING],
+ PUSHING_OUTPUT: [JobStatus.RUNNING, CONVERTING_OUTPUT],
+ CANCELING: [JobStatus.INACTIVE, JobStatus.QUEUED, JobStatus.RUNNING],
+ JobStatus.ERROR: [FETCHING_INPUT, CONVERTING_INPUT, CONVERTING_OUTPUT,
+ PUSHING_OUTPUT, CANCELING, JobStatus.QUEUED,
+ JobStatus.RUNNING],
+ JobStatus.CANCELED: [CANCELING]
+ }
+
@classmethod
def isValid(cls, status):
return status in (
@@ -53,6 +65,10 @@ class CustomJobStatus(object):
cls.CANCELING
)
+ @classmethod
+ def validTransition(cls, status):
+ return cls.valid_transitions.get(status)
+
def getCeleryApp():
"""
@@ -145,7 +161,7 @@ def validateApiUrl(doc):
def validateJobStatus(event):
"""Allow our custom job status values."""
- if CustomJobStatus.isValid(event.info):
+ if event.info['handler'] == 'worker_handler' and CustomJobStatus.isValid(event.info):
event.preventDefault().addResponse(True) | Override job state transitions for the worker plugin | girder_girder | train |
34a5fbb1a03e62e3232824fe9c62c91a8ee949a8 | diff --git a/lib/manager/gomod/update.js b/lib/manager/gomod/update.js
index <HASH>..<HASH> 100644
--- a/lib/manager/gomod/update.js
+++ b/lib/manager/gomod/update.js
@@ -6,12 +6,15 @@ module.exports = {
function updateDependency(currentFileContent, upgrade) {
try {
+ logger.debug(`gomod.updateDependency: ${upgrade.newValue}`);
const { depName } = upgrade;
- const depNameNoVersion = depName
+ let depNameNoVersion = depName
.split('/')
.slice(0, 3)
.join('/');
- logger.debug(`gomod.updateDependency: ${upgrade.newValue}`);
+ if (depNameNoVersion.startsWith('gopkg.in')) {
+ depNameNoVersion = depNameNoVersion.replace(/\.v\d+$/, '');
+ }
const lines = currentFileContent.split('\n');
const lineToChange = lines[upgrade.lineNumber];
if ( | fix(gomod): detect gopkg.in major bumps | renovatebot_renovate | train |
46bdde4bc2db8324bc069af1aca0d1a9d7794558 | diff --git a/library/exportsym.py b/library/exportsym.py
index <HASH>..<HASH> 100644
--- a/library/exportsym.py
+++ b/library/exportsym.py
@@ -72,9 +72,9 @@ def createSignal(signal):
return output
def exportSym(db, filename):
-"""
-export canmatrix-object as .sym file (compatible to PEAK-Systems)
-"""
+ """
+ export canmatrix-object as .sym file (compatible to PEAK-Systems)
+ """
global enumDict
global enums | Correct exportSym() docstring indentation | ebroecker_canmatrix | train |
41989a277d8b331be106252eb29816f24229dde7 | diff --git a/setup.py b/setup.py
index <HASH>..<HASH> 100644
--- a/setup.py
+++ b/setup.py
@@ -9,17 +9,18 @@ setup(
name = 'amqp_connection',
version = '1.0',
description = 'Python AMQP connection for worker',
+ license='MIT',
keywords = [
'AMQP',
'connection'
],
- author='Marc-Antoine ARNAUD, Valentin NOEL',
- author_email='[email protected]',
+ author = 'Marc-Antoine ARNAUD, Valentin NOEL',
+ author_email = '[email protected]',
url = 'https://github.com/FTV-Subtil/py_amqp_connection',
packages = [
'amqp_connection'
],
- package_dir={
+ package_dir = {
'amqp_connection': os.path.join(dir_name, 'src')
- }
+ },
) | Clean setup.py and add license description | FTV-Subtil_py_amqp_connection | train |
5f77beba294af462867e9c0dacf2e0b623e824bd | diff --git a/csrf/README.md b/csrf/README.md
index <HASH>..<HASH> 100644
--- a/csrf/README.md
+++ b/csrf/README.md
@@ -1,5 +1,5 @@
## CSRF handler
-Stateless protection against CSRF attacks for Go web applications.
+Offers stateless protection against CSRF attacks for Go web applications.
* Checks [Origin header]() was sent and matches the Host header.
* Falls back to a URL-safe and secure HMAC token stored in a HTTP-only
diff --git a/csrf/csrf.go b/csrf/csrf.go
index <HASH>..<HASH> 100644
--- a/csrf/csrf.go
+++ b/csrf/csrf.go
@@ -18,10 +18,10 @@ import (
// handler is a private struct which contains the handler's configurable options.
type handler struct {
- name string
- domain string
- secret string
- sessionID string
+ name string
+ domain string
+ secret string
+ userID string
}
// Name allows configuring the CSRF cookie name.
@@ -38,10 +38,10 @@ func Secret(s string) option {
}
}
-// SessionID allows to configure a random and unique user session identifier to generate the CSRF token.
-func SessionID(s string) option {
+// UserID allows to configure a random and unique user ID identifier used to generate the CSRF token.
+func UserID(s string) option {
return func(h *handler) {
- h.sessionID = s
+ h.userID = s
}
}
@@ -89,7 +89,7 @@ func Handler(h http.Handler, opts ...option) http.Handler {
// Re-enables browser's XSS filter if it was disabled
w.Header().Set("x-xss-protection", "1; mode=block")
- if csrf.sessionID == "" {
+ if csrf.userID == "" {
http.Error(w, errForbidden, http.StatusForbidden)
return
}
@@ -101,7 +101,7 @@ func Handler(h http.Handler, opts ...option) http.Handler {
case http.MethodDelete:
case http.MethodPost:
default:
- setToken(w, csrf.name, csrf.secret, csrf.sessionID, csrf.domain)
+ setToken(w, csrf.name, csrf.secret, csrf.userID, csrf.domain)
h.ServeHTTP(w, r)
return
}
@@ -120,18 +120,18 @@ func Handler(h http.Handler, opts ...option) http.Handler {
return
}
- if !xsrftoken.Valid(cookie.Value, csrf.secret, csrf.sessionID, "Global") {
+ if !xsrftoken.Valid(cookie.Value, csrf.secret, csrf.userID, "Global") {
http.Error(w, errForbidden, http.StatusForbidden)
return
}
- setToken(w, csrf.name, csrf.secret, csrf.sessionID, csrf.domain)
+ setToken(w, csrf.name, csrf.secret, csrf.userID, csrf.domain)
h.ServeHTTP(w, r)
})
}
-func setToken(w http.ResponseWriter, name, secret, sessionID, domain string) {
- token := xsrftoken.Generate(secret, sessionID, "Global")
+func setToken(w http.ResponseWriter, name, secret, userID, domain string) {
+ token := xsrftoken.Generate(secret, userID, "Global")
cookie := &http.Cookie{
Name: name,
Value: token,
diff --git a/csrf/csrf_test.go b/csrf/csrf_test.go
index <HASH>..<HASH> 100644
--- a/csrf/csrf_test.go
+++ b/csrf/csrf_test.go
@@ -39,19 +39,19 @@ func TestDomainRequired(t *testing.T) {
}
}()
- handler := Handler(requestHandler, SessionID("my session ID!"), Secret("my secret!"))
+ handler := Handler(requestHandler, UserID("my user ID!"), Secret("my secret!"))
ts := httptest.NewServer(handler)
defer ts.Close()
}
func TestCSRFProtection(t *testing.T) {
- handler := Handler(requestHandler, SessionID("my session ID!"), Secret("my secret!"), Domain("localhost"), Name("_csrf"))
+ handler := Handler(requestHandler, UserID("my user ID!"), Secret("my secret!"), Domain("localhost"), Name("_csrf"))
okTS := httptest.NewServer(handler)
defer okTS.Close()
cookie := &http.Cookie{
Name: "_csrf",
- Value: xsrftoken.Generate("my secret!", "my session ID!", "Global"),
+ Value: xsrftoken.Generate("my secret!", "my user ID!", "Global"),
Path: "/",
Domain: "localhost",
Expires: time.Now().Add(xsrftoken.Timeout),
diff --git a/csrf/example_test.go b/csrf/example_test.go
index <HASH>..<HASH> 100644
--- a/csrf/example_test.go
+++ b/csrf/example_test.go
@@ -19,7 +19,7 @@ func ExampleHandler() {
fmt.Fprintf(w, "Welcome to the home page!")
})
- rack := csrf.Handler(mux, csrf.SessionID("session ID"), csrf.Secret("my secret!"), csrf.Domain("localhost"))
+ rack := csrf.Handler(mux, csrf.UserID("user ID"), csrf.Secret("my secret!"), csrf.Domain("localhost"))
http.ListenAndServe(":8080", rack)
} | csrf: changes option name[C to more closely describe what it really is | c4milo_handlers | train |
7d132077c757944f0bf19eadf8e776c5f1758f0d | diff --git a/demos/init.php b/demos/init.php
index <HASH>..<HASH> 100644
--- a/demos/init.php
+++ b/demos/init.php
@@ -41,6 +41,7 @@ if (isset($layout->leftMenu)) {
$form->addItem('Form Multi-column layout', ['form3']);
$form->addItem(['Integration with Columns'], ['form5']);
$form->addItem(['AutoComplete Field', 'icon'=>'yellow star'], ['autocomplete']);
+ $form->addItem(['DropDown Field'], ['dropdown-plus']);
$form->addItem(['Value Selectors'], ['form6']);
$form->addItem(['Conditional Fields'], ['jscondform']); | Link Dropdown demos in Menu
dropdown-plus.php existed, but was not linked in demo menu | atk4_ui | train |
11cb9a795dcf5a9286b910ffc24fe936cb48d98f | diff --git a/server/dashboards.go b/server/dashboards.go
index <HASH>..<HASH> 100644
--- a/server/dashboards.go
+++ b/server/dashboards.go
@@ -52,9 +52,28 @@ func (s *Service) Dashboards(w http.ResponseWriter, r *http.Request) {
// DashboardID returns a single specified dashboard
func (s *Service) DashboardID(w http.ResponseWriter, r *http.Request) {
+ id, err := paramID("id", r)
+ if err != nil {
+ Error(w, http.StatusUnprocessableEntity, err.Error(), s.Logger)
+ return
+ }
+ ctx := r.Context()
+ e, err := s.DashboardsStore.Get(ctx, chronograf.DashboardID(id))
+ if err != nil {
+ notFound(w, id, s.Logger)
+ return
+ }
+
+ res := newDashboardResponse(e)
+ encodeJSON(w, http.StatusOK, res, s.Logger)
}
+// type postDashboardRequest struct {
+// Data interface{} `json:"data"` // Serialization of config.
+// Name string `json:"name,omitempty"` // Exploration name given by user.
+// }
+
// NewDashboard creates and returns a new dashboard object
func (s *Service) NewDashboard(w http.ResponseWriter, r *http.Request) {
@@ -62,7 +81,24 @@ func (s *Service) NewDashboard(w http.ResponseWriter, r *http.Request) {
// RemoveDashboard deletes a dashboard
func (s *Service) RemoveDashboard(w http.ResponseWriter, r *http.Request) {
-
+ id, err := paramID("id", r)
+ if err != nil {
+ Error(w, http.StatusUnprocessableEntity, err.Error(), h.Logger)
+ return
+ }
+
+ ctx := r.Context()
+ e, err := s.DashboardsStore.Get(ctx, chronograf.DashboardID(id))
+ if err != nil {
+ notFound(w, id, s.Logger)
+ return
+ }
+
+ if err := s.DashboardsStore.Delete(ctx, &chronograf.Dashboard{ID: chronograf.DashboardID(id)}); err != nil {
+ unknownErrorWithMessage(w, err, s.Logger)
+ return
+ }
+ w.WriteHeader(http.StatusNoContent)
}
// UpdateDashboard updates a dashboard | add dashboard GET and DELETE handlers | influxdata_influxdb | train |
5543e1fcc71fbc2c781da4587577e070a8711b24 | diff --git a/storage/looputil/loop.go b/storage/looputil/loop.go
index <HASH>..<HASH> 100644
--- a/storage/looputil/loop.go
+++ b/storage/looputil/loop.go
@@ -7,7 +7,7 @@ import (
"bytes"
"os"
"os/exec"
- "path/filepath"
+ "path"
"regexp"
"strconv"
"strings"
@@ -64,7 +64,7 @@ func (m *loopDeviceManager) DetachLoopDevices(rootfs, prefix string) error {
if !strings.HasPrefix(info.backingFile, prefix) {
continue
}
- rootedBackingFile := filepath.Join(rootfs, info.backingFile)
+ rootedBackingFile := path.Join(rootfs, info.backingFile)
st, err := m.stat(rootedBackingFile)
if os.IsNotExist(err) {
continue | storage/looputil: use path, not path/filepath
A bit hackish, but then again so is testing
loop device management on Windows. | juju_juju | train |
111f44ffc450e2db8c30a923dfb664c309c76194 | diff --git a/pymatgen/io/vasp/sets.py b/pymatgen/io/vasp/sets.py
index <HASH>..<HASH> 100644
--- a/pymatgen/io/vasp/sets.py
+++ b/pymatgen/io/vasp/sets.py
@@ -214,6 +214,9 @@ class DictSet(VaspInputSet):
of two ways, e.g. either {"LDAUU":{"O":{"Fe":5}}} to set LDAUU
for Fe to 5 in an oxide, or {"LDAUU":{"Fe":5}} to set LDAUU to
5 regardless of the input structure.
+
+ If a None value is given, that key is unset. For example,
+ {"ENCUT": None} will remove ENCUT from the incar settings.
user_kpoints_settings (dict or Kpoints): Allow user to override kpoints
setting by supplying a dict E.g., {"reciprocal_density": 1000}.
User can also supply Kpoints object. Default is None.
@@ -297,7 +300,14 @@ class DictSet(VaspInputSet):
@property
def incar(self):
settings = dict(self._config_dict["INCAR"])
- settings.update(self.user_incar_settings)
+ for k, v in self.user_incar_settings.items():
+ if v is None:
+ try:
+ del settings[k]
+ except KeyError:
+ settings[k] = v
+ else:
+ settings[k] = v
structure = self.structure
incar = Incar()
comp = structure.composition
diff --git a/pymatgen/io/vasp/tests/test_sets.py b/pymatgen/io/vasp/tests/test_sets.py
index <HASH>..<HASH> 100644
--- a/pymatgen/io/vasp/tests/test_sets.py
+++ b/pymatgen/io/vasp/tests/test_sets.py
@@ -163,6 +163,9 @@ class MITMPRelaxSetTest(unittest.TestCase):
)
self.assertEqual(userset.incar['MAGMOM'], [100, 0.6])
+ noencutset = MPRelaxSet(struct, user_incar_settings={'ENCUT': None})
+ self.assertNotIn("ENCUT", noencutset.incar)
+
# sulfide vs sulfate test
coords = list() | Add None support for user_incar_settings. | materialsproject_pymatgen | train |
Subsets and Splits