hash
stringlengths
40
40
diff
stringlengths
131
114k
message
stringlengths
7
980
project
stringlengths
5
67
split
stringclasses
1 value
62857cfa45e65fd31d6f90532a3a1988a24ed2b0
diff --git a/ndio/remote/boss/tests/int_test_group.py b/ndio/remote/boss/tests/int_test_group.py index <HASH>..<HASH> 100644 --- a/ndio/remote/boss/tests/int_test_group.py +++ b/ndio/remote/boss/tests/int_test_group.py @@ -59,8 +59,6 @@ class ProjectGroupTest(unittest.TestCase): self.existing_grp_name = 'int_test_exists' self.user_name = 'bossadmin' - self.rmt.group_create(self.existing_grp_name) - def cleanup_db(self): """Clean up the data model objects used by this test case. @@ -71,6 +69,7 @@ class ProjectGroupTest(unittest.TestCase): def setUp(self): self.initialize() + self.rmt.group_create(self.existing_grp_name) def tearDown(self): self.cleanup_db()
Moved creation of test group. No need to create test group in setUpClass().
jhuapl-boss_intern
train
b5a2e2cf032e47d1e0dfd752662a4d4068235847
diff --git a/src/Viserio/Connect/Traits/DetectsLostConnections.php b/src/Viserio/Connect/Traits/DetectsLostConnections.php index <HASH>..<HASH> 100644 --- a/src/Viserio/Connect/Traits/DetectsLostConnections.php +++ b/src/Viserio/Connect/Traits/DetectsLostConnections.php @@ -24,6 +24,7 @@ trait DetectsLostConnections 'is dead or not enabled', 'Error while sending', 'Operation in progress', + 'decryption failed or bad record mac', ]); } }
Added another lost connection case for pgsql
narrowspark_framework
train
1cdd38d3c0bed524ac4f48eccf70080b6e54d9e5
diff --git a/lib/NormalModuleFactory.js b/lib/NormalModuleFactory.js index <HASH>..<HASH> 100644 --- a/lib/NormalModuleFactory.js +++ b/lib/NormalModuleFactory.js @@ -475,22 +475,23 @@ class NormalModuleFactory extends ModuleFactory { return continueCallback(); } - normalResolver.resolve( - contextInfo, - context, - unresolvedResource, - resolveContext, - (err, resolvedResource, resolvedResourceResolveData) => { - if (err) { - if (!isDataURI(unresolvedResource)) return continueCallback(err); - - resolvedResource = unresolvedResource; + if (isDataURI(unresolvedResource)) { + resource = unresolvedResource; + continueCallback(); + } else { + normalResolver.resolve( + contextInfo, + context, + unresolvedResource, + resolveContext, + (err, resolvedResource, resolvedResourceResolveData) => { + if (err) return continueCallback(err); + resource = resolvedResource; + resourceResolveData = resolvedResourceResolveData; + continueCallback(); } - resource = resolvedResource; - resourceResolveData = resolvedResourceResolveData; - continueCallback(); - } - ); + ); + } } ); }
check for data-uri before resource resolving
webpack_webpack
train
724c7f63e42cbfd154bcf16f759608d9b18722a2
diff --git a/src/diamond/collector.py b/src/diamond/collector.py index <HASH>..<HASH> 100644 --- a/src/diamond/collector.py +++ b/src/diamond/collector.py @@ -481,8 +481,6 @@ class Collector(object): metric_name = 'collector_time_ms' metric_value = collector_time self.publish(metric_name, metric_value) - except Exception, e: - self.log.exception(e) finally: # After collector run, invoke a flush # method on each handler.
Revert <I>f<I> collector as exceptions are printed by the scheduler thread Handling the exception in the collector code actually results in signal exceptions not being passed to the scheduler, thus breaking functionality
python-diamond_Diamond
train
d805fe21d15f90ebd8846a11bc7f73edd16d771e
diff --git a/lib/active_interaction/version.rb b/lib/active_interaction/version.rb index <HASH>..<HASH> 100644 --- a/lib/active_interaction/version.rb +++ b/lib/active_interaction/version.rb @@ -6,5 +6,5 @@ module ActiveInteraction # The version number. # # @return [Gem::Version] - VERSION = Gem::Version.new('3.0.1') + VERSION = Gem::Version.new('3.1.0') end
:briefcase: <I>
AaronLasseigne_active_interaction
train
8108a9bfe2867cc4db79cccecd8b752e185aed0b
diff --git a/http/logical.go b/http/logical.go index <HASH>..<HASH> 100644 --- a/http/logical.go +++ b/http/logical.go @@ -158,8 +158,10 @@ func handleLogicalInternal(core *vault.Core, injectDataIntoTopLevel bool) http.H // Always forward requests that are using a limited use count token. // origBody will not be nil if it's a perf standby as it checks // PerfStandby() but will be nil otherwise. - if origBody != nil && req.ClientTokenRemainingUses > 0 { - r.Body = origBody + if core.PerfStandby() && req.ClientTokenRemainingUses > 0 { + if origBody != nil { + r.Body = origBody + } forwardRequest(core, w, r) return } @@ -275,7 +277,9 @@ func handleLogicalInternal(core *vault.Core, injectDataIntoTopLevel bool) http.H // success. resp, ok, needsForward := request(core, w, r, req) if needsForward { - r.Body = origBody + if origBody != nil { + r.Body = origBody + } forwardRequest(core, w, r) return }
Fix a potential panic due to ioutil.ReadAll not being safe with nil readers
hashicorp_vault
train
4a72c441d583557282df06a2252e7c06ac83f868
diff --git a/src/Symfony/Component/DependencyInjection/ParameterBag/FrozenParameterBag.php b/src/Symfony/Component/DependencyInjection/ParameterBag/FrozenParameterBag.php index <HASH>..<HASH> 100644 --- a/src/Symfony/Component/DependencyInjection/ParameterBag/FrozenParameterBag.php +++ b/src/Symfony/Component/DependencyInjection/ParameterBag/FrozenParameterBag.php @@ -69,4 +69,14 @@ class FrozenParameterBag extends ParameterBag { throw new LogicException('Impossible to call set() on a frozen ParameterBag.'); } + + /** + * {@inheritdoc} + * + * @api + */ + public function remove($name) + { + throw new LogicException('Impossible to call remove() on a frozen ParameterBag.'); + } }
[DependencyInjection] Freeze also FrozenParameterBag::remove
symfony_symfony
train
7957a6a48a68cfc05c2b2f3109f653f0bf0754bc
diff --git a/test/RelationshipGraph.js b/test/RelationshipGraph.js index <HASH>..<HASH> 100644 --- a/test/RelationshipGraph.js +++ b/test/RelationshipGraph.js @@ -146,7 +146,7 @@ describe('RelationshipGraph', function() { }); }); - describe('#ValidateVerifyJSON()' , function() { + describe('#ValidateVerifyJSON()', function() { var graph = d3.select('#test2').relationshipGraph({ 'thresholds': [200] }); @@ -463,8 +463,7 @@ describe('RelationshipGraph', function() { 'parent': 'Warner Bros. Pictures', 'value': '$958,366,855', 'Year': '2013' - } - , + }, { 'Movie Title': 'The Hobbit: The Battle of the Five Armies', 'parent': 'Warner Bros. Pictures', @@ -627,7 +626,7 @@ describe('RelationshipGraph', function() { var expectedText = ['20th Century Fox (6)', 'Columbia Pictures (3)', 'Lionsgate Films (1)', 'New Line Cinema (3)', 'Paramount Pictures (1)', 'Universal Pictures (7)', 'Walt Disney Studios (16)', 'Warner Bros. Pictures (13)'], expectedY = [0, 24, 48, 72, 96, 120, 144, 192]//, - //expectedX = [41, 35, 50, 38, 26, 36, 18, 7]; + //expectedX = [41, 35, 50, 38, 26, 36, 18, 7]; chai.expect(text.length).to.equal(expectedText.length); @@ -685,4 +684,76 @@ describe('RelationshipGraph', function() { }); }); + + describe('#VerifySortJson', function() { + it('Should be sorted.', function() { + var json = [ + { + 'Movie Title': 'Avatar', + 'parent': '20th Century Fox', + 'value': '$2,787,965,087', + 'Year': '2009' + }, + { + 'Movie Title': 'Star Wars: The Force Awakens', + 'parent': 'Walt Disney Studios', + 'value': '$2,066,247,462', + 'Year': '2015' + }, + { + 'Movie Title': 'Titanic', + 'parent': '20th Century Fox', + 'value': '$2,186,772,302', + 'Year': '1997' + } + ], expected = [ + { + 'Movie Title': 'Avatar', + 'parent': '20th Century Fox', + 'value': '$2,787,965,087', + 'Year': '2009' + }, + { + 'Movie Title': 'Titanic', + 'parent': '20th Century Fox', + 'value': '$2,186,772,302', + 'Year': '1997' + }, + { + 'Movie Title': 'Star Wars: The Force Awakens', + 'parent': 'Walt Disney Studios', + 'value': '$2,066,247,462', + 'Year': '2015' + } + ]; + + json.sort(function(child1, child2) { + var parent1 = child1.parent.toLowerCase(), + parent2 = child2.parent.toLowerCase(); + + return (parent1 > parent2) ? 1 : (parent1 < parent2) ? -1 : 0; + }); + + for (var i = 0; i < json.length; i++) { + chai.expect(json[i]['Movie Title']).to.equal(expected[i]['Movie Title']); + chai.expect(json[i].parent).to.equal(expected[i].parent); + chai.expect(json[i].value).to.equal(expected[i].value); + chai.expect(json[i].year).to.equal(expected[i].year); + } + }); + }); + + describe('#VerifyCustomColors', function() { + it('Should have the custom color set.', function() { + var custom = ['red', 'green', 'blue']; + + var graph = d3.select('#graph').relationshipGraph({ + colors: custom + }); + + for (var i = 0; i < custom.length; i++) { + chai.expect(graph.graph.config.colors[i]).to.equal(custom[i]); + } + }); + }); });
Added tests for sorting the JSON and adding custom colors.
hkelly93_d3-relationshipgraph
train
c0bf0520bb1b80cf3cddaa869d681da6742fca42
diff --git a/container/api/src/test/java/org/wildfly/swarm/container/ProjectStageTest.java b/container/api/src/test/java/org/wildfly/swarm/container/ProjectStageTest.java index <HASH>..<HASH> 100644 --- a/container/api/src/test/java/org/wildfly/swarm/container/ProjectStageTest.java +++ b/container/api/src/test/java/org/wildfly/swarm/container/ProjectStageTest.java @@ -3,7 +3,7 @@ package org.wildfly.swarm.container; import java.io.InputStream; import java.util.List; -import junit.framework.Assert; +import org.junit.Assert; import org.junit.Test; import org.wildfly.swarm.container.internal.ProjectStageFactory; import org.wildfly.swarm.spi.api.ProjectStage; diff --git a/integration-tests/src/test/java/org/wildfly/swarm/integration/SWARM_513/SWARM_513Test.java b/integration-tests/src/test/java/org/wildfly/swarm/integration/SWARM_513/SWARM_513Test.java index <HASH>..<HASH> 100644 --- a/integration-tests/src/test/java/org/wildfly/swarm/integration/SWARM_513/SWARM_513Test.java +++ b/integration-tests/src/test/java/org/wildfly/swarm/integration/SWARM_513/SWARM_513Test.java @@ -5,7 +5,7 @@ import com.squareup.okhttp.OkHttpClient; import com.squareup.okhttp.Request; import com.squareup.okhttp.Response; import com.squareup.okhttp.RequestBody; -import junit.framework.Assert; +import org.junit.Assert; import org.jboss.arquillian.container.test.api.Deployment; import org.jboss.arquillian.container.test.api.RunAsClient; import org.jboss.shrinkwrap.api.Archive; diff --git a/monitor/api/src/test/java/org/wildfly/swarm/monitor/MonitorTest.java b/monitor/api/src/test/java/org/wildfly/swarm/monitor/MonitorTest.java index <HASH>..<HASH> 100644 --- a/monitor/api/src/test/java/org/wildfly/swarm/monitor/MonitorTest.java +++ b/monitor/api/src/test/java/org/wildfly/swarm/monitor/MonitorTest.java @@ -15,7 +15,7 @@ */ package org.wildfly.swarm.monitor; -import junit.framework.Assert; +import org.junit.Assert; import org.junit.Test; /**
junit.framework is deprecated
thorntail_thorntail
train
456f58997cd380970da1e6ea901b4b4d987c4d9f
diff --git a/CHANGELOG.md b/CHANGELOG.md index <HASH>..<HASH> 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -23,6 +23,10 @@ - might be a good idea... think on it... - DOING: update docs +## 2.0.0a20 (released on 21.01.2020) +- Fixed UnicodeEncodeError: 'charmap' codec + - thanks to [PR-197](https://github.com/yashaka/selene/pull/197) from @ak40u + ## 2.0.0a19 (released on 16.01.2020) - removed deprecation from shared.config.counter and reports_folder - removed backports.functools-lru-cache from project dependencies diff --git a/selene/__init__.py b/selene/__init__.py index <HASH>..<HASH> 100644 --- a/selene/__init__.py +++ b/selene/__init__.py @@ -211,7 +211,7 @@ Or, by using non-waiting versions, if "you are in a rush:)":: # """ # todo: add here some type imports like Element, Collection, etc. -__version__ = '2.0.0a19' +__version__ = '2.0.0a20' # --- DEPRECATED, and will be removed soon --- # diff --git a/selene/support/shared/config.py b/selene/support/shared/config.py index <HASH>..<HASH> 100644 --- a/selene/support/shared/config.py +++ b/selene/support/shared/config.py @@ -126,10 +126,19 @@ class SharedConfig(Config): stored.quit() # todo: can this raise exception? that we need to supress... # todo: do we need here pass self.desired_capabilities too? + + set_chrome = lambda: Chrome( + executable_path=ChromeDriverManager().install(), + options=ChromeOptions()) + + set_firefox = lambda: Firefox( + executable_path=GeckoDriverManager().install()) + + # set_remote = lambda: Remote() # todo: do we really need it? :) + new = { - 'chrome': lambda: Chrome(executable_path=ChromeDriverManager().install(), - options=ChromeOptions()), - 'firefox': lambda: Firefox(executable_path=GeckoDriverManager().install()) + 'chrome': set_chrome, + 'firefox': set_firefox }.get(self.browser_name)() # todo: think on something like:
bumped <I>a<I>; tiny refactoring of config
yashaka_selene
train
612e603aa787a76ae77c7e6c34547a70c84aa987
diff --git a/client/html/templates/common/partials/selection-default.php b/client/html/templates/common/partials/selection-default.php index <HASH>..<HASH> 100644 --- a/client/html/templates/common/partials/selection-default.php +++ b/client/html/templates/common/partials/selection-default.php @@ -18,12 +18,25 @@ $attributes = $this->get( 'selectionAttributeItems', array() ); * default. This setting removes the hint to select an option, so the first one * is selected by default. * + * The key for each value must be the type code of the attribute, e.g. "width", + * "length", "color" or similar types. You can set the layout for all + * attributes at once using e.g. + * + * client/html/catalog/detail/basket/selection/preselect = array( + * 'width' => false, + * 'color' => true, + * ) + * + * Similarly, you can set the pre-selection for a specific attribute only, + * leaving the rest untouched: + * + * client/html/catalog/detail/basket/selection/preselect/color = true + * * @param boolean True to select the first option by default, false to display the select hint - * @since 2016.05 + * @since 2016.07 * @category Developer * @category User */ -$preselect = (bool) $this->config( 'client/html/catalog/detail/basket/selection/preselect', false ); /** client/html/catalog/detail/basket/selection/type * List of layout types for the variant attributes @@ -81,15 +94,16 @@ $preselect = (bool) $this->config( 'client/html/catalog/detail/basket/selection/ <ul class="selection"> <?php foreach( $this->get( 'selectionAttributeTypeDependencies', array() ) as $code => $attrIds ) : asort( $attrIds ); ?> <?php $layout = $this->config( 'client/html/catalog/detail/basket/selection/type/' . $code, 'select' ); ?> - <li class="select-item <?php echo $enc->attr( $layout ) . ' ' . $enc->attr( $code ); ?>"> +<?php $preselect = (bool) $this->config( 'client/html/catalog/detail/basket/selection/preselect/' . $code, false ); ?> +<li class="select-item <?php echo $enc->attr( $layout ) . ' ' . $enc->attr( $code ); ?>"> <div class="select-name"><?php echo $enc->html( $this->translate( 'client/code', $code ) ); ?></div> <div class="select-value"> -<?php if( $layout === 'radio' ) : ?> +<?php if( $layout === 'radio' ) : $first = true; ?> <ul class="select-list" data-index="<?php echo $index++; ?>"> <?php foreach( $attrIds as $attrId => $position ) : ?> <?php if( isset( $attributes[$attrId] ) ) : ?> <li class="select-entry"> - <input class="select-option" id="option-<?php echo $enc->attr( $attrId ); ?>" name="<?php echo $enc->attr( $this->formparam( array( 'b_prod', 0, 'attrvarid', $code ) ) ); ?>" type="radio" value="<?php echo $enc->attr( $attrId ); ?>" /> + <input class="select-option" id="option-<?php echo $enc->attr( $attrId ); ?>" name="<?php echo $enc->attr( $this->formparam( array( 'b_prod', 0, 'attrvarid', $code ) ) ); ?>" type="radio" value="<?php echo $enc->attr( $attrId ); ?>" <?php echo ( $preselect && $first ? 'checked="checked"' : '' ); $first = false ?>/> <label class="select-label" for="option-<?php echo $enc->attr( $attrId ); ?>"><!-- <?php foreach( $attributes[$attrId]->getListItems( 'media', 'icon' ) as $listItem ) : ?> <?php if( ( $item = $listItem->getRefItem() ) !== null ) : ?>
Improved pre-selection of product variant attributes
aimeos_ai-client-html
train
8825d7edf310ce182d3bef4d70d63bc2658218bd
diff --git a/go/vt/vtgate/vstream_manager_test.go b/go/vt/vtgate/vstream_manager_test.go index <HASH>..<HASH> 100644 --- a/go/vt/vtgate/vstream_manager_test.go +++ b/go/vt/vtgate/vstream_manager_test.go @@ -48,7 +48,6 @@ import ( var mu sync.Mutex func TestVStreamSkew(t *testing.T) { - t.Skip() stream := func(conn *sandboxconn.SandboxConn, shard string, count, idx int64) { vevents := getVEvents(shard, count, idx) for _, ev := range vevents { @@ -123,7 +122,6 @@ func TestVStreamSkew(t *testing.T) { } func TestVStreamEvents(t *testing.T) { - t.Skip() ctx, cancel := context.WithCancel(context.Background()) defer cancel() cell := "aa" @@ -276,7 +274,6 @@ func TestVStreamChunks(t *testing.T) { } func TestVStreamMulti(t *testing.T) { - t.Skip() ctx, cancel := context.WithCancel(context.Background()) defer cancel() cell := "aa" @@ -382,7 +379,6 @@ func TestVStreamRetry(t *testing.T) { } func TestVStreamShouldNotSendSourceHeartbeats(t *testing.T) { - t.Skip() ctx, cancel := context.WithCancel(context.Background()) defer cancel() cell := "aa" @@ -433,7 +429,6 @@ func TestVStreamShouldNotSendSourceHeartbeats(t *testing.T) { } func TestVStreamJournalOneToMany(t *testing.T) { - t.Skip() ctx, cancel := context.WithCancel(context.Background()) defer cancel() cell := "aa" @@ -542,7 +537,6 @@ func TestVStreamJournalOneToMany(t *testing.T) { } func TestVStreamJournalManyToOne(t *testing.T) { - t.Skip() ctx, cancel := context.WithCancel(context.Background()) defer cancel() @@ -658,7 +652,6 @@ func TestVStreamJournalManyToOne(t *testing.T) { } func TestVStreamJournalNoMatch(t *testing.T) { - t.Skip() ctx, cancel := context.WithCancel(context.Background()) defer cancel() @@ -787,7 +780,6 @@ func TestVStreamJournalNoMatch(t *testing.T) { } func TestVStreamJournalPartialMatch(t *testing.T) { - t.Skip() ctx, cancel := context.WithCancel(context.Background()) defer cancel() @@ -1018,7 +1010,6 @@ func TestResolveVStreamParams(t *testing.T) { } func TestVStreamIdleHeartbeat(t *testing.T) { - t.Skip() cell := "aa" ks := "TestVStream" _ = createSandbox(ks)
vstream tests: disable only the consistently failing tests
vitessio_vitess
train
2bb82be1e555a6ccb1f55365a369dd052017c903
diff --git a/reordering.py b/reordering.py index <HASH>..<HASH> 100644 --- a/reordering.py +++ b/reordering.py @@ -25,7 +25,6 @@ def ravel(a, order='C'): return reshape(a=a, newshape=(-1,)) - class Reorder(Permute): dterms = 'a', @@ -38,17 +37,21 @@ class Reorder(Permute): def compute_dr_wrt(self, wrt): if wrt is self.a: - a = self.a - asz = a.size - ashape = a.shape - key = self.unique_reorder_id() - if key not in self.dr_lookup or key is None: - JS = self.reorder(np.arange(asz).reshape(ashape)) - IS = np.arange(JS.size) - data = np.ones_like(IS) - shape = JS.shape - self.dr_lookup[key] = sp.csc_matrix((data, (IS, JS.ravel())), shape=(self.r.size, wrt.r.size)) - return self.dr_lookup[key] + if True: + from scipy.sparse.linalg.interface import LinearOperator + return LinearOperator((self.size, wrt.size), lambda x : self.reorder(x.reshape(self.a.shape)).ravel()) + else: + a = self.a + asz = a.size + ashape = a.shape + key = self.unique_reorder_id() + if key not in self.dr_lookup or key is None: + JS = self.reorder(np.arange(asz).reshape(ashape)) + IS = np.arange(JS.size) + data = np.ones_like(IS) + shape = JS.shape + self.dr_lookup[key] = sp.csc_matrix((data, (IS, JS.ravel())), shape=(self.r.size, wrt.r.size)) + return self.dr_lookup[key] class Sort(Reorder): dterms = 'a'
reordering.py: make more use of LinearOperator in reordering
mattloper_chumpy
train
e99f49c2dba565595ebb5b24ccc46f847ec20257
diff --git a/slacker/__init__.py b/slacker/__init__.py index <HASH>..<HASH> 100644 --- a/slacker/__init__.py +++ b/slacker/__init__.py @@ -85,6 +85,12 @@ class Users(BaseAPI): class Groups(BaseAPI): + def create(self, name): + return self.post('groups.create', params={'name': name}) + + def create_child(self, channel): + return self.post('groups.createChild', params={'channel': channel}) + def list(self, exclude_archived=None): return self.get('groups.list', params={'exclude_archived': exclude_archived})
Add groups.create and groups.createChild APIs.
os_slacker
train
412b03f243e93a598073bf4cda6b65084e0d0d74
diff --git a/spec/lib/pg_search/multisearch/rebuilder_spec.rb b/spec/lib/pg_search/multisearch/rebuilder_spec.rb index <HASH>..<HASH> 100644 --- a/spec/lib/pg_search/multisearch/rebuilder_spec.rb +++ b/spec/lib/pg_search/multisearch/rebuilder_spec.rb @@ -84,10 +84,15 @@ describe PgSearch::Multisearch::Rebuilder do time = DateTime.parse("2001-01-01") rebuilder = PgSearch::Multisearch::Rebuilder.new(Model, lambda { time } ) - # Handle change in precision of DateTime objects in SQL in Rails 4.1 + # Handle change in precision of DateTime objects in SQL in Active Record 4.1 # https://github.com/rails/rails/commit/17f5d8e062909f1fcae25351834d8e89967b645e + version_4_1_or_newer = ( + ActiveRecord::VERSION::MAJOR > 4 || + ActiveRecord::VERSION::MAJOR = 4 && ActiveRecord::VERSION::MINOR >= 1 + ) + expected_timestamp = - if ActiveRecord::VERSION::MAJOR >= 4 && ActiveRecord::VERSION::MINOR >= 1 + if version_4_1_or_newer "2001-01-01 00:00:00.000000" else "2001-01-01 00:00:00"
Future-proof test for Active Record <I>+
Casecommons_pg_search
train
04272a2d5be800d8a11b0b27d49d517330a8b6c8
diff --git a/lib/airrecord/table.rb b/lib/airrecord/table.rb index <HASH>..<HASH> 100644 --- a/lib/airrecord/table.rb +++ b/lib/airrecord/table.rb @@ -125,7 +125,6 @@ module Airrecord value = fields[key] elsif column_mappings[key] deprecate_symbols if key.is_a? Symbol - deprecate_symbols value = fields[column_mappings[key]] end @@ -151,7 +150,7 @@ module Airrecord elsif column_mappings[key] deprecate_symbols return if fields[column_mappings[key]] == value # no-op - @updated_keys << key + @updated_keys << column_mappings[key] fields[column_mappings[key]] = value else @updated_keys << key
revert code change, remove duplicate code
sirupsen_airrecord
train
0377c50b714c245b190384e99705d9394924c609
diff --git a/wayback-core/src/main/java/org/archive/wayback/resourcestore/FlexResourceStore.java b/wayback-core/src/main/java/org/archive/wayback/resourcestore/FlexResourceStore.java index <HASH>..<HASH> 100644 --- a/wayback-core/src/main/java/org/archive/wayback/resourcestore/FlexResourceStore.java +++ b/wayback-core/src/main/java/org/archive/wayback/resourcestore/FlexResourceStore.java @@ -17,7 +17,6 @@ import org.archive.io.warc.WARCReader; import org.archive.io.warc.WARCRecord; import org.archive.util.binsearch.SeekableLineReader; import org.archive.util.binsearch.SortedTextFile; -import org.archive.util.binsearch.impl.HTTPSeekableLineReader; import org.archive.util.iterator.CloseableIterator; import org.archive.wayback.ResourceStore; import org.archive.wayback.core.CaptureSearchResult; @@ -25,7 +24,6 @@ import org.archive.wayback.core.Resource; import org.archive.wayback.exception.ResourceNotAvailableException; import org.archive.wayback.resourcestore.resourcefile.ArcResource; import org.archive.wayback.resourcestore.resourcefile.WarcResource; -import org.archive.wayback.webapp.PerformanceLogger; public class FlexResourceStore implements ResourceStore { @@ -223,8 +221,7 @@ public class FlexResourceStore implements ResourceStore { } public Resource getResource(String path, CaptureSearchResult result) throws IOException - { - long start = System.currentTimeMillis(); + { Resource r = null; long offset = result.getOffset(); @@ -255,27 +252,31 @@ public class FlexResourceStore implements ResourceStore { r.parseHeaders(); - if ((customHeader != null) && (slr instanceof HTTPSeekableLineReader)) { - HTTPSeekableLineReader httpSlr = (HTTPSeekableLineReader)slr; - String value = httpSlr.getHeaderValue(customHeader); - - if (value != null) { - result.put(CaptureSearchResult.CUSTOM_HEADER_PREFIX + customHeader, value); - } - } + // Ability to pass on header prefix from data store request +// if ((customHeader != null) && (slr instanceof HTTPSeekableLineReader)) { +// HTTPSeekableLineReader httpSlr = (HTTPSeekableLineReader)slr; +// String value = httpSlr.getHeaderValue(customHeader); +// +// if (value != null) { +// result.put(CaptureSearchResult.CUSTOM_HEADER_PREFIX + customHeader, value); +// } +// } - } catch (IOException io) { + } catch (Exception e) { if (slr != null) { slr.close(); } + r = null; slr = null; - throw io; + + if (e instanceof IOException) { + throw (IOException)e; + } else { + throw new IOException(e); + } } - - long elapsed = System.currentTimeMillis() - start; - PerformanceLogger.noteElapsed("{W}arcBlockLoader", elapsed, path); - + return r; }
FIX: FlexResourceStore catchs all exceptions, including runtime and ensures Resource is closed if its not returned
iipc_openwayback
train
165d0f89188a3de20dac61f3395b039550c8a243
diff --git a/sonar-ws-client/src/main/java/org/sonar/wsclient/issue/BulkChangeQuery.java b/sonar-ws-client/src/main/java/org/sonar/wsclient/issue/BulkChangeQuery.java index <HASH>..<HASH> 100644 --- a/sonar-ws-client/src/main/java/org/sonar/wsclient/issue/BulkChangeQuery.java +++ b/sonar-ws-client/src/main/java/org/sonar/wsclient/issue/BulkChangeQuery.java @@ -63,6 +63,11 @@ public class BulkChangeQuery { return this; } + public BulkChangeQuery sendNotifications(boolean sendNotifications) { + params.put("sendNotifications", String.valueOf(sendNotifications)); + return this; + } + private BulkChangeQuery addParam(String key, String[] values) { if (values != null) { params.put(key, EncodingUtils.toQueryParam(values)); diff --git a/sonar-ws-client/src/test/java/org/sonar/wsclient/issue/BulkChangeQueryTest.java b/sonar-ws-client/src/test/java/org/sonar/wsclient/issue/BulkChangeQueryTest.java index <HASH>..<HASH> 100644 --- a/sonar-ws-client/src/test/java/org/sonar/wsclient/issue/BulkChangeQueryTest.java +++ b/sonar-ws-client/src/test/java/org/sonar/wsclient/issue/BulkChangeQueryTest.java @@ -40,14 +40,15 @@ public class BulkChangeQueryTest { .actions("assign") .actionParameter("assign", "assignee", "geoffrey") .comment("My bulk comment") - ; + .sendNotifications(false); - assertThat(query.urlParams()).hasSize(4).includes( + assertThat(query.urlParams()).hasSize(5).includes( entry("issues", "ABCD,EFGH"), entry("actions", "assign"), entry("assign.assignee", "geoffrey"), - entry("comment", "My bulk comment") - ); + entry("comment", "My bulk comment"), + entry("sendNotifications", "false") + ); } @Test
SONAR-<I> Update issue client
SonarSource_sonarqube
train
d92352d05d8601d643feeb0d58201f6e499348c0
diff --git a/tests/src/TestBase.php b/tests/src/TestBase.php index <HASH>..<HASH> 100644 --- a/tests/src/TestBase.php +++ b/tests/src/TestBase.php @@ -3,13 +3,14 @@ namespace Grasmash\YamlCli\Tests; use Symfony\Component\Console\Application; +use PHPUnit\Framework\TestCase; /** * Class BltTestBase. * * Base class for all tests that are executed for BLT itself. */ -abstract class TestBase extends \PHPUnit_Framework_TestCase +abstract class TestBase extends TestCase { /** @var Application */ @@ -23,8 +24,7 @@ abstract class TestBase extends \PHPUnit_Framework_TestCase * * @see https://symfony.com/doc/current/console.html#testing-commands */ - public function setUp() - { + protected function setUp(): void { parent::setUp(); $this->application = new Application(); @@ -33,8 +33,7 @@ abstract class TestBase extends \PHPUnit_Framework_TestCase /** * Removes temporary file. */ - public function tearDown() - { + protected function tearDown(): void { parent::tearDown(); // This will only exist if a test called setupTemporaryConfigFiles().
refactoring TestBase.php.
grasmash_yaml-cli
train
4b284f7182cb5aca9d15964f4e6db9681231c41d
diff --git a/core/dutycycle/dutyManager.go b/core/dutycycle/dutyManager.go index <HASH>..<HASH> 100644 --- a/core/dutycycle/dutyManager.go +++ b/core/dutycycle/dutyManager.go @@ -125,7 +125,7 @@ func (m *dutyManager) Update(id []byte, freq float64, size uint, datr string, co var entry dutyEntry if err == nil { entry = itf.([]dutyEntry)[0] - } else if err.(errors.Failure).Nature == errors.Behavioural { + } else if err.(errors.Failure).Nature == errors.NotFound { entry = dutyEntry{ Until: time.Unix(0, 0), OnAir: make(map[subBand]time.Duration),
[refactor/errors] Update duty cycle manager with NotFound error nature
TheThingsNetwork_ttn
train
14df2a0f239f72a25963449e16e8f4f10db09fcd
diff --git a/Src/java/cql-to-elm/src/main/java/org/cqframework/cql/cql2elm/Cql2ElmVisitor.java b/Src/java/cql-to-elm/src/main/java/org/cqframework/cql/cql2elm/Cql2ElmVisitor.java index <HASH>..<HASH> 100755 --- a/Src/java/cql-to-elm/src/main/java/org/cqframework/cql/cql2elm/Cql2ElmVisitor.java +++ b/Src/java/cql-to-elm/src/main/java/org/cqframework/cql/cql2elm/Cql2ElmVisitor.java @@ -1766,13 +1766,17 @@ public class Cql2ElmVisitor extends cqlBaseVisitor { public Object visitConversionExpressionTerm(@NotNull cqlParser.ConversionExpressionTermContext ctx) { TypeSpecifier targetType = parseTypeSpecifier(ctx.typeSpecifier()); Expression operand = parseExpression(ctx.expression()); - Conversion conversion = libraryBuilder.findConversion(operand.getResultType(), targetType.getResultType(), false); - if (conversion == null) { - throw new IllegalArgumentException(String.format("Could not resolve conversion from type %s to type %s.", - operand.getResultType(), targetType.getResultType())); + if (!DataTypes.equal(operand.getResultType(), targetType.getResultType())) { + Conversion conversion = libraryBuilder.findConversion(operand.getResultType(), targetType.getResultType(), false); + if (conversion == null) { + throw new IllegalArgumentException(String.format("Could not resolve conversion from type %s to type %s.", + operand.getResultType(), targetType.getResultType())); + } + + return libraryBuilder.convertExpression(operand, conversion); } - return libraryBuilder.convertExpression(operand, conversion); + return operand; } @Override
#<I>: Fixed no-op conversions throwing an error.
cqframework_clinical_quality_language
train
114b4af6f7e82cb8c014d818206050ac71d74ab8
diff --git a/timeside/server/tests/test_provider.py b/timeside/server/tests/test_provider.py index <HASH>..<HASH> 100644 --- a/timeside/server/tests/test_provider.py +++ b/timeside/server/tests/test_provider.py @@ -17,7 +17,7 @@ class TestProvider(TimeSideTestServer): for provider in request_providers.data: if provider['pid'] == 'youtube': self.youtube_uuid = provider['uuid'] - if provider['pid'] == 'deezer': + if provider['pid'] == 'deezer_preview': self.deezer_uuid = provider['uuid']
[server] change deezer_preview provider pid in REST API test
Parisson_TimeSide
train
3891bc33ab2de8fa155f5112154d092d1aac424d
diff --git a/setup.py b/setup.py index <HASH>..<HASH> 100644 --- a/setup.py +++ b/setup.py @@ -21,11 +21,13 @@ except IOError: readme = '' changelog = '' -py_version = sys.version_info[:2] +py_version = sys.version_info[:3] -packages = ['mygeotab', 'mygeotab/ext'] +if py_version < (2, 7, 9): + raise RuntimeError("This package requres Python 2.7.9+") -if py_version >= (3, 5): +packages = ['mygeotab', 'mygeotab/ext'] +if py_version >= (3, 5, 0): packages.append('mygeotab/async') setup(
Since all MyGeotab servers use TLS <I>, we must force a version of Python 2 that supports this (<I>+). Addresses found issue in #<I>.
Geotab_mygeotab-python
train
e66c9c077bbb05bd237b6f6e098f6b8e3dcea990
diff --git a/lang/en/completion.php b/lang/en/completion.php index <HASH>..<HASH> 100644 --- a/lang/en/completion.php +++ b/lang/en/completion.php @@ -126,7 +126,7 @@ $string['markedcompleteby']='Marked complete by {$a}'; $string['markingyourselfcomplete']='Marking yourself complete'; $string['moredetails']='More details'; $string['nocriteriaset']='No completion criteria set for this course'; -$string['notenroled']='You are not enroled in this course'; +$string['notenroled']='You are not enrolled in this course'; $string['notyetstarted']='Not yet started'; $string['overallcriteriaaggregation']='Overall criteria type aggregation'; $string['passinggrade']='Passing grade'; @@ -147,7 +147,7 @@ $string['unenrolingfromcourse']='Unenroling from course'; $string['unenrolment']='Unenrolment'; $string['unlockcompletiondelete']='Unlock completion options and delete user completion data'; $string['usealternateselector']='Use the alternate course selector'; -$string['usernotenroled']='User is not enroled in this course'; +$string['usernotenroled']='User is not enrolled in this course'; $string['viewcoursereport']='View course report'; $string['viewingactivity']='Viewing the {$a}'; $string['xdays']='{$a} days';
MDL-<I> completio: Fix spelling of Enrolled.
moodle_moodle
train
4ce4c1d58db2ee74d584da759444dddb3419190e
diff --git a/trimesh/interfaces/generic.py b/trimesh/interfaces/generic.py index <HASH>..<HASH> 100644 --- a/trimesh/interfaces/generic.py +++ b/trimesh/interfaces/generic.py @@ -18,7 +18,7 @@ class MeshScript: # export the meshes to a temporary STL container for m, f in zip(self.meshes, self.mesh_pre): - m.export(file_type='stl', file_obj=f.name) + m.export(file_type='stl', file_obj=f) self.replacement = {'mesh_' + str(i) : m.name for i,m in enumerate(self.mesh_pre)} self.replacement['mesh_pre'] = str([i.name for i in self.mesh_pre])
Windows does not allow re-opening the file
mikedh_trimesh
train
9a133d51251e2c048089554327ab3ad4ab72327a
diff --git a/ara/api/tests/tests_play.py b/ara/api/tests/tests_play.py index <HASH>..<HASH> 100644 --- a/ara/api/tests/tests_play.py +++ b/ara/api/tests/tests_play.py @@ -111,7 +111,7 @@ class PlayTestCase(APITestCase): # Create a playbook and two plays playbook = factories.PlaybookFactory() play = factories.PlayFactory(name="first_play", playbook=playbook) - factories.TaskFactory(name="second_play", playbook=playbook) + factories.PlayFactory(name="second_play", playbook=playbook) # Query for the first play name and expect one result request = self.client.get("/api/v1/plays?name=%s" % play.name)
tests: fix typo in play unit tests
ansible-community_ara
train
3a2a0c67683caaf2f6dc8d3e2af1e04ed3e9ab1c
diff --git a/examples/aws/terraform/ansible/ec2.py b/examples/aws/terraform/ansible/ec2.py index <HASH>..<HASH> 100755 --- a/examples/aws/terraform/ansible/ec2.py +++ b/examples/aws/terraform/ansible/ec2.py @@ -95,7 +95,8 @@ def generate_ssh_cfg(cluster, ssh_key=None): bastion = '' for reservation in response['Reservations']: for instance in reservation['Instances']: - bastion = instance['PublicIpAddress'] + if instance['State']['Code'] == 16: # code 16 = running + bastion = instance['PublicIpAddress'] return ssh_cfg_template.format(bastion=bastion, ssh_key_path=ssh_key_path, identity_file=identity_file) if args.cluster is None:
test state before getting public IP (#<I>)
gravitational_teleport
train
716c0cfad85fbcd56bbda1752b6e25395936edb9
diff --git a/api/charms/client.go b/api/charms/client.go index <HASH>..<HASH> 100644 --- a/api/charms/client.go +++ b/api/charms/client.go @@ -61,7 +61,7 @@ func (c *Client) CharmInfo(charmURL string) (*CharmInfo, error) { result := &CharmInfo{ Revision: info.Revision, URL: info.URL, - Config: convertCharmConfig(info.Config), + Config: params.FromCharmOptionMap(info.Config), Meta: meta, Actions: convertCharmActions(info.Actions), Metrics: convertCharmMetrics(info.Metrics), @@ -70,27 +70,6 @@ func (c *Client) CharmInfo(charmURL string) (*CharmInfo, error) { return result, nil } -func convertCharmConfig(config map[string]params.CharmOption) *charm.Config { - if len(config) == 0 { - return nil - } - result := &charm.Config{ - Options: make(map[string]charm.Option), - } - for key, value := range config { - result.Options[key] = convertCharmOption(value) - } - return result -} - -func convertCharmOption(opt params.CharmOption) charm.Option { - return charm.Option{ - Type: opt.Type, - Description: opt.Description, - Default: opt.Default, - } -} - func convertCharmMeta(meta *params.CharmMeta) (*charm.Meta, error) { if meta == nil { return nil, nil diff --git a/apiserver/facades/client/charms/client.go b/apiserver/facades/client/charms/client.go index <HASH>..<HASH> 100644 --- a/apiserver/facades/client/charms/client.go +++ b/apiserver/facades/client/charms/client.go @@ -89,7 +89,7 @@ func (a *API) CharmInfo(args params.CharmURL) (params.Charm, error) { info := params.Charm{ Revision: aCharm.Revision(), URL: curl.String(), - Config: convertCharmConfig(aCharm.Config()), + Config: params.ToCharmOptionMap(aCharm.Config()), Meta: convertCharmMeta(aCharm.Meta()), Actions: convertCharmActions(aCharm.Actions()), Metrics: convertCharmMetrics(aCharm.Metrics()), @@ -152,25 +152,6 @@ func (a *API) IsMetered(args params.CharmURL) (params.IsMeteredResult, error) { return params.IsMeteredResult{Metered: false}, nil } -func convertCharmConfig(config *charm.Config) map[string]params.CharmOption { - if config == nil { - return nil - } - result := make(map[string]params.CharmOption) - for key, value := range config.Options { - result[key] = convertCharmOption(value) - } - return result -} - -func convertCharmOption(opt charm.Option) params.CharmOption { - return params.CharmOption{ - Type: opt.Type, - Description: opt.Description, - Default: opt.Default, - } -} - func convertCharmMeta(meta *charm.Meta) *params.CharmMeta { if meta == nil { return nil diff --git a/apiserver/params/charms.go b/apiserver/params/charms.go index <HASH>..<HASH> 100644 --- a/apiserver/params/charms.go +++ b/apiserver/params/charms.go @@ -3,6 +3,8 @@ package params +import "github.com/juju/charm/v7" + // ApplicationCharmResults contains a set of ApplicationCharmResults. type ApplicationCharmResults struct { Results []ApplicationCharmResult `json:"results"` @@ -197,3 +199,43 @@ type ContainerProfileResult struct { type ContainerProfileResults struct { Results []ContainerProfileResult `json:"results"` } + +func ToCharmOptionMap(config *charm.Config) map[string]CharmOption { + if config == nil { + return nil + } + result := make(map[string]CharmOption) + for key, value := range config.Options { + result[key] = toParamsCharmOption(value) + } + return result +} + +func toParamsCharmOption(opt charm.Option) CharmOption { + return CharmOption{ + Type: opt.Type, + Description: opt.Description, + Default: opt.Default, + } +} + +func FromCharmOptionMap(config map[string]CharmOption) *charm.Config { + if len(config) == 0 { + return nil + } + result := &charm.Config{ + Options: make(map[string]charm.Option), + } + for key, value := range config { + result.Options[key] = fromParamsCharmOption(value) + } + return result +} + +func fromParamsCharmOption(opt CharmOption) charm.Option { + return charm.Option{ + Type: opt.Type, + Description: opt.Description, + Default: opt.Default, + } +}
Move convertCharmConfig to params. Renamed to ToCharmOptionMap and FromCharmOptionMap. This functionality is needed by other libraries. Put in a common place to share.
juju_juju
train
5fe29a64650eab8da7ef4730db769ce6894a7b6a
diff --git a/internal/util/comparisons.js b/internal/util/comparisons.js index <HASH>..<HASH> 100644 --- a/internal/util/comparisons.js +++ b/internal/util/comparisons.js @@ -190,7 +190,7 @@ function innerDeepEqual(val1, val2, strict, memos) { return keyCheck(val1, val2, strict, memos, kNoIterator); } if (isDate(val1)) { - if (!isDate(val2) || Date.prototype.getTime.call(val1) !== Date.prototype.getTime.call(val2)) { + if (!(isDate(val2) && Date.prototype.getTime.call(val1) === Date.prototype.getTime.call(val2))) { return false; } } else if (isRegExp(val1)) {
Make sure getTime is never called on a non-date object
browserify_commonjs-assert
train
bd6c07038defa15a9934b7c95850a30fe24b7e32
diff --git a/src/test/java/com/cronutils/model/time/generator/ExecutionTimeUnixIntegrationTest.java b/src/test/java/com/cronutils/model/time/generator/ExecutionTimeUnixIntegrationTest.java index <HASH>..<HASH> 100644 --- a/src/test/java/com/cronutils/model/time/generator/ExecutionTimeUnixIntegrationTest.java +++ b/src/test/java/com/cronutils/model/time/generator/ExecutionTimeUnixIntegrationTest.java @@ -265,4 +265,15 @@ public class ExecutionTimeUnixIntegrationTest { ExecutionTime executionTime = ExecutionTime.forCron(cron); assertEquals(DateTime.parse("2014-12-01T08:00:00Z"), executionTime.nextExecution(date)); } + + /** + * Issue #92: Next execution skipping valid date + */ + @Test + public void testNextExecution2016() { + CronParser parser = new CronParser(CronDefinitionBuilder.instanceDefinitionFor(CronType.UNIX)); + ExecutionTime executionTime = ExecutionTime.forCron(parser.parse("1 0 * * tue")); + DateTime date = DateTime.parse("2016-05-24T01:02:50Z"); + assertEquals(DateTime.parse("2016-05-31T00:01:00Z"), executionTime.nextExecution(date)); + } }
Failing unit test for issue #<I>
jmrozanec_cron-utils
train
b007283d753c0c7cab66fb9ee51626171b3f72f0
diff --git a/littletable.py b/littletable.py index <HASH>..<HASH> 100644 --- a/littletable.py +++ b/littletable.py @@ -117,7 +117,7 @@ Here is a simple C{littletable} data storage/retrieval example:: """ __version__ = "0.13.2" -__versionTime__ = "3 Nov 2018 00:00 UTC" +__versionTime__ = "6 Nov 2018 23:02 UTC" __author__ = "Paul McGuire <[email protected]>" import sys @@ -1704,25 +1704,18 @@ class JoinTerm(object): if __name__ == "__main__": - - rawdata = """\ + import textwrap + rawdata = textwrap.dedent("""\ Phoenix:AZ:85001:KPHX Phoenix:AZ:85001:KPHY Phoenix:AZ:85001:KPHA - Dallas:TX:75201:KDFW""".splitlines() + Dallas:TX:75201:KDFW""") # load miniDB - stations = Table() + stations = Table().csv_import(rawdata, delimiter=':', fieldnames=['city', 'state', 'zip', 'stn']) # stations.create_index("city") stations.create_index("stn", unique=True) - data_fields = "city state zip stn".split() - for d in rawdata: - rec = DataObject() - for kk, vv in zip(data_fields, d.split(':')): - setattr(rec, kk, vv.strip()) - stations.insert(rec) - # perform some queries and deletes for queryargs in [ dict(city="Phoenix"), @@ -1808,3 +1801,5 @@ if __name__ == "__main__": print(amfm.by.stn['KPHY']) except KeyError: print("no station 'KPHY' in table") + + print(list(stations.all.stn))
Include inline CSV import and table.all example to sample code in __main__
ptmcg_littletable
train
f25f12f84441174aeda5e2545c2846a4ba6eb068
diff --git a/redisson/src/main/java/org/redisson/connection/pool/ConnectionPool.java b/redisson/src/main/java/org/redisson/connection/pool/ConnectionPool.java index <HASH>..<HASH> 100644 --- a/redisson/src/main/java/org/redisson/connection/pool/ConnectionPool.java +++ b/redisson/src/main/java/org/redisson/connection/pool/ConnectionPool.java @@ -169,8 +169,8 @@ abstract class ConnectionPool<T extends RedisConnection> { public RFuture<T> get(RedisCommand<?> command) { for (int j = entries.size() - 1; j >= 0; j--) { final ClientConnectionsEntry entry = getEntry(); - if (!entry.isFreezed() - && tryAcquireConnection(entry)) { + if ((!entry.isFreezed() || entry.getFreezeReason() == FreezeReason.SYSTEM) && + tryAcquireConnection(entry)) { return acquireConnection(command, entry); } }
add missing condition for tentatively unfreezing node
redisson_redisson
train
ac257582528f5ab4aed95132688ca24e0db0994f
diff --git a/lib/engine/piston/contextify_task_executor.js b/lib/engine/piston/contextify_task_executor.js index <HASH>..<HASH> 100644 --- a/lib/engine/piston/contextify_task_executor.js +++ b/lib/engine/piston/contextify_task_executor.js @@ -43,15 +43,21 @@ ContextifyTaskExecutor.prototype.execute_task = function(task_request){ child_global_object[name] = child_grafter.graft(child_globals[name]); }); + var last_eval; + var response; + + try { + last_eval = vm.runInContext(task_request.getCode(), child_context); + response = new TaskResponse({response:{evaluation: last_eval}}); + } catch(e) { + response = new TaskResponse({response:{error: e.name + ": " + e.message}}); + } - var last_eval = vm.runInContext(task_request.getCode(), child_context); - - var response = new TaskResponse({response:{evaluation: last_eval}}); this.emit("task executed", response); }; ContextifyTaskExecutor.prototype.render_child_globals = function(task_request){ - var child_globals = eval(task_request.getContext())(); + var child_globals = eval(task_request.getContext())(task_request.getLocals()); return child_globals; }; diff --git a/spec/engine/piston/contextify_task_executor_spec.js b/spec/engine/piston/contextify_task_executor_spec.js index <HASH>..<HASH> 100644 --- a/spec/engine/piston/contextify_task_executor_spec.js +++ b/spec/engine/piston/contextify_task_executor_spec.js @@ -5,6 +5,24 @@ var TaskResponse = require("../../../lib/engine/client/task/response").TaskRespo describe("ContextifyTaskExecutor", function(){ beforeEach(function(){ this.executor = ContextifyTaskExecutor.create(); + this.assert_task_response = function(task_request, task_response){ + var spy = jasmine.createSpy(); + + this.executor.on("task executed", spy); + + runs(function(){ + this.executor.execute_task(task_request); + }); + + waitsFor(function(){ + return spy.callCount > 0; + }); + + runs(function(){ + expect(spy).toHaveBeenCalledWith(task_response); + }); + }; + }); it("executes a simple TaskRequest", function(){ @@ -15,20 +33,45 @@ describe("ContextifyTaskExecutor", function(){ locals: {} }); var simple_task_response = new TaskResponse({response:{evaluation:4}}); - var spy = jasmine.createSpy(); - - this.executor.on("task executed", spy); - runs(function(){ - this.executor.execute_task(simple_task); + this.assert_task_response(simple_task, simple_task_response); + }); + + it("executes a TaskRequest with locals", function(){ + var simple_task = new TaskRequest({ + task_id: "2", + context: "(function(locals){ return { hello:function(){ return 'hello ' + locals.world; } } })", + code: "hello()", + locals: { world:"world" } }); + var simple_task_response = new TaskResponse({response:{evaluation: "hello world"}}); + + this.assert_task_response(simple_task, simple_task_response); + }); - waitsFor(function(){ - return spy.callCount > 0; + it("executes a TaskRequest that contains syntax errors", function(){ + var simple_task = new TaskRequest({ + task_id: "2", + context: "(function(locals){ return { hello:function(){ return 'hello'; } } })", + code: "hello(", + locals: { } }); + var simple_task_response = new TaskResponse({response:{error: "SyntaxError: Unexpected end of input"}}); + + this.assert_task_response(simple_task, simple_task_response); + }); - runs(function(){ - expect(spy).toHaveBeenCalledWith(simple_task_response); - }); + it("executes a TaskRequest that contains reference errors", function(){ + var simple_task = new TaskRequest({ + task_id: "2", + context: "(function(locals){ return { hello:function(){ return 'hello'; } } })", + code: "foo()", + locals: { } + }); + var simple_task_response = new TaskResponse({response:{error: "ReferenceError: foo is not defined"}}); + + this.assert_task_response(simple_task, simple_task_response); }); + + xit("executes a TaskRequest with asynchronous code"); }); \ No newline at end of file
WIP: Handle syntax and reference errors during task execution
rehanift_engine.js
train
1d9fa88f6de61c6975021489b7086d9cf33b945c
diff --git a/gwpy/cli/cliproduct.py b/gwpy/cli/cliproduct.py index <HASH>..<HASH> 100644 --- a/gwpy/cli/cliproduct.py +++ b/gwpy/cli/cliproduct.py @@ -27,8 +27,6 @@ import warnings import sys from functools import wraps -from six import add_metaclass - from matplotlib import rcParams try: from matplotlib.cm import viridis as DEFAULT_CMAP @@ -94,8 +92,7 @@ to_s = to_float('s') # pylint: disable=invalid-name # -- base product class ------------------------------------------------------- -@add_metaclass(abc.ABCMeta) -class CliProduct(object): +class CliProduct(object, metaclass=abc.ABCMeta): """Base class for all cli plot products Parameters @@ -761,15 +758,14 @@ class CliProduct(object): # -- extensions --------------------------------------------------------------- -@add_metaclass(abc.ABCMeta) -class ImageProduct(CliProduct): +class ImageProduct(CliProduct, metaclass=abc.ABCMeta): """Base class for all x/y/color plots """ MAX_DATASETS = 1 @classmethod def init_plot_options(cls, parser): - super(ImageProduct, cls).init_plot_options(parser) + super().init_plot_options(parser) cls.arg_color_axis(parser) @classmethod @@ -797,7 +793,7 @@ class ImageProduct(CliProduct): def _finalize_arguments(self, args): if args.cmap is None: args.cmap = DEFAULT_CMAP.name - return super(ImageProduct, self)._finalize_arguments(args) + return super()._finalize_arguments(args) @staticmethod def get_color_label(): @@ -809,7 +805,7 @@ class ImageProduct(CliProduct): """Set properties for each axis (scale, limits, label) and create a colorbar. """ - super(ImageProduct, self).set_axes_properties() + super().set_axes_properties() if not self.args.nocolorbar: self.set_colorbar() @@ -824,8 +820,7 @@ class ImageProduct(CliProduct): return # image plots don't have legends -@add_metaclass(abc.ABCMeta) -class FFTMixin(object): +class FFTMixin(object, metaclass=abc.ABCMeta): """Mixin for `CliProduct` class that will perform FFTs This just adds FFT-based command line options @@ -834,7 +829,7 @@ class FFTMixin(object): def init_data_options(cls, parser): """Set up data input and signal processing options including FFTs """ - super(FFTMixin, cls).init_data_options(parser) + super().init_data_options(parser) cls.arg_fft(parser) @classmethod @@ -899,11 +894,10 @@ class FFTMixin(object): args.overlap = recommended_overlap(args.window) except ValueError: args.overlap = .5 - return super(FFTMixin, self)._finalize_arguments(args) + return super()._finalize_arguments(args) -@add_metaclass(abc.ABCMeta) -class TimeDomainProduct(CliProduct): +class TimeDomainProduct(CliProduct, metaclass=abc.ABCMeta): """`CliProduct` with time on the X-axis """ @classmethod @@ -913,7 +907,7 @@ class TimeDomainProduct(CliProduct): This method includes the standard X-axis options, as well as a new ``--epoch`` option for the time axis. """ - group = super(TimeDomainProduct, cls).arg_xaxis(parser) + group = super().arg_xaxis(parser) group.add_argument('--epoch', type=to_gps, help='center X axis on this GPS time, may be' 'absolute date/time or delta') @@ -937,7 +931,7 @@ class TimeDomainProduct(CliProduct): if args.xmax is None: args.xmax = max(starts) + args.duration - return super(TimeDomainProduct, self)._finalize_arguments(args) + return super()._finalize_arguments(args) def get_xlabel(self): """Default X-axis label for plot @@ -953,8 +947,7 @@ class TimeDomainProduct(CliProduct): return '' -@add_metaclass(abc.ABCMeta) -class FrequencyDomainProduct(CliProduct): +class FrequencyDomainProduct(CliProduct, metaclass=abc.ABCMeta): """`CliProduct` with frequency on the X-axis """ def get_xlabel(self):
Drop use of six.add_metaclass, as it interferes with python3's super() functionality
gwpy_gwpy
train
022068a23230dd13d8e99852f47eaaf461cf8174
diff --git a/tests/lib/toolkit/XMLDocumentTest.php b/tests/lib/toolkit/XMLDocumentTest.php index <HASH>..<HASH> 100644 --- a/tests/lib/toolkit/XMLDocumentTest.php +++ b/tests/lib/toolkit/XMLDocumentTest.php @@ -22,6 +22,20 @@ final class XMLDocumentTest extends TestCase $this->assertEquals('<xml />', $x->generate()); } + public function testSetVersion() + { + $x = (new \XMLDocument('xml'))->setVersion('test')->renderHeader(); + $this->assertEquals('test', $x->getVersion()); + $this->assertEquals('<?xml version="test" encoding="utf-8" ?><xml />', $x->generate()); + } + + public function testSetEncoding() + { + $x = (new \XMLDocument('xml'))->setEncoding('test')->renderHeader(); + $this->assertEquals('test', $x->getEncoding()); + $this->assertEquals('<?xml version="1.0" encoding="test" ?><xml />', $x->generate()); + } + public function testGenerateWithHeader() { $x = (new \XMLDocument('xml', 'value'));
Add setVersion() and setEncoding() tests Picked from <I>d<I>cad Picked from 0c3ce<I>b<I>
symphonycms_symphony-2
train
1043e957ac7c899877758a6129ed1a32f383ef55
diff --git a/src/Slugify.php b/src/Slugify.php index <HASH>..<HASH> 100644 --- a/src/Slugify.php +++ b/src/Slugify.php @@ -524,6 +524,29 @@ class Slugify implements SlugifyInterface } /** + * Adds a ruleset to the Slugifier. + * + * @param string $name Name of the ruleset. + * @param string[][] $rules Rules + */ + public function addRuleset($name, array $rules) + { + $this->rulesets[$name] = $rules; + + return $this; + } + + /** + * Returns the rulesets. + * + * @return string[][] Rulesets + */ + public function getRulesets() + { + return $this->rulesets; + } + + /** * Static method to create new instance of {@see Slugify}. * * @return Slugify diff --git a/tests/SlugifyTest.php b/tests/SlugifyTest.php index <HASH>..<HASH> 100644 --- a/tests/SlugifyTest.php +++ b/tests/SlugifyTest.php @@ -91,6 +91,18 @@ class SlugifyTest extends \PHPUnit_Framework_TestCase /** * @test + * @covers Cocur\Slugify\Slugify::addRuleset() + * @covers Cocur\Slugify\Slugify::getRulesets() + */ + public function addRulesetGetRulesets() + { + $this->slugify->addRuleset('foo', array('k' => 'key')); + + $this->assertCount(2, $this->slugify->getRulesets()); + } + + /** + * @test * @covers Cocur\Slugify\Slugify::create() */ public function createReturnsAnInstance()
Added possibility to add and get rulesets
cocur_slugify
train
c4d6b71e40f35821eabc7372151cee7d27f12568
diff --git a/www/admin/metadata.php b/www/admin/metadata.php index <HASH>..<HASH> 100644 --- a/www/admin/metadata.php +++ b/www/admin/metadata.php @@ -12,36 +12,31 @@ $config = SimpleSAML_Configuration::getInstance(); //$metadata = new SimpleSAML_XML_MetaDataStore($config); $session = SimpleSAML_Session::getInstance(); - - try { $metadata = SimpleSAML_Metadata_MetaDataStorageHandler::getMetadataHandler(); - $et = new SimpleSAML_XHTML_Template($config, 'admin-metadatalist.php'); - if ($config->getValue('enable.saml20-sp') === true) { $results = array(); - /* + $metalist = $metadata->getList('saml20-sp-hosted'); foreach ($metalist AS $entityid => $mentry) { $results[$entityid] = SimpleSAML_Utilities::checkAssocArrayRules($mentry, - // TODO: UPDATE Required and optional parameter list - array('entityid', 'host', 'spNameQualifier', 'NameIDFormat', 'ForceAuthn'), - array('name', 'description') + array('entityid', 'host', 'NameIDFormat', 'ForceAuthn'), + array() ); } $et->data['metadata.saml20-sp-hosted'] = $results; - */ + + $results = array(); $metalist = $metadata->getList('saml20-idp-remote'); foreach ($metalist AS $entityid => $mentry) { $results[$entityid] = SimpleSAML_Utilities::checkAssocArrayRules($mentry, - // TODO: UPDATE Required and optional parameter list - array('entityid', 'host', 'spNameQualifier', 'NameIDFormat', 'ForceAuthn'), - array('name', 'description') + array('entityid', 'SingleSignOnService', 'SingleLogoutService', 'certFingerprint'), + array('name', 'description', 'base64attributes') ); } $et->data['metadata.saml20-idp-remote'] = $results; @@ -53,19 +48,18 @@ try { $metalist = $metadata->getList('saml20-idp-hosted'); foreach ($metalist AS $entityid => $mentry) { $results[$entityid] = SimpleSAML_Utilities::checkAssocArrayRules($mentry, - // TODO: UPDATE Required and optional parameter list - array('entityid', 'host', 'spNameQualifier', 'NameIDFormat', 'ForceAuthn'), - array('name', 'description') + array('entityid', 'host', 'privatekey', 'certificate', 'auth'), + array('requireconsent') ); } $et->data['metadata.saml20-idp-hosted'] = $results; + $results = array(); $metalist = $metadata->getList('saml20-sp-remote'); foreach ($metalist AS $entityid => $mentry) { $results[$entityid] = SimpleSAML_Utilities::checkAssocArrayRules($mentry, - // TODO: UPDATE Required and optional parameter list - array('entityid', 'host', 'spNameQualifier', 'NameIDFormat', 'ForceAuthn'), - array('name', 'description') + array('entityid', 'spNameQualifier', 'AssertionConsumerService', 'SingleLogoutService', 'NameIDFormat'), + array('base64attributes', 'attributemap', 'simplesaml.attributes', 'attributes', 'name', 'description') ); } $et->data['metadata.saml20-sp-remote'] = $results; @@ -73,11 +67,7 @@ try { } - - - - $et->data['header'] = 'Metadata overview';
Improved the metadata overview admin page by configuring which parameters that are optional and required in each metadata set
simplesamlphp_saml2
train
f7ca720f572f3ecc8471810084f98464bd33925b
diff --git a/src/kff.Collection.js b/src/kff.Collection.js index <HASH>..<HASH> 100644 --- a/src/kff.Collection.js +++ b/src/kff.Collection.js @@ -16,7 +16,7 @@ kff.Collection = kff.createClass( */ constructor: function(options) { - options = options || {}; + this.options = options = options || {}; this.itemFactory = options.itemFactory || null; this.itemType = options.itemType || kff.Model; this.serializeAttrs = options.serializeAttrs || null;
fix(kff.Collection): collection options (metadata) should be saved in the instance for later use (i.e. cloning)
karfcz_kff
train
f6db0af6fd33eb9d3540c31ec3bc925d05e71d50
diff --git a/openstack_dashboard/dashboards/project/loadbalancers/workflows.py b/openstack_dashboard/dashboards/project/loadbalancers/workflows.py index <HASH>..<HASH> 100644 --- a/openstack_dashboard/dashboards/project/loadbalancers/workflows.py +++ b/openstack_dashboard/dashboards/project/loadbalancers/workflows.py @@ -25,6 +25,10 @@ from horizon import workflows from openstack_dashboard import api +AVAILABLE_PROTOCOLS = ('HTTP', 'HTTPS', 'TCP') +AVAILABLE_METHODS = ('ROUND_ROBIN', 'LEAST_CONNECTIONS', 'SOURCE_IP') + + class AddPoolAction(workflows.Action): name = forms.CharField(max_length=80, label=_("Name")) description = forms.CharField( @@ -57,14 +61,11 @@ class AddPoolAction(workflows.Action): self.fields['subnet_id'].choices = subnet_id_choices protocol_choices = [('', _("Select a Protocol"))] - protocol_choices.append(('HTTP', 'HTTP')) - protocol_choices.append(('HTTPS', 'HTTPS')) + [protocol_choices.append((p, p)) for p in AVAILABLE_PROTOCOLS] self.fields['protocol'].choices = protocol_choices lb_method_choices = [('', _("Select a Method"))] - lb_method_choices.append(('ROUND_ROBIN', 'ROUND_ROBIN')) - lb_method_choices.append(('LEAST_CONNECTIONS', 'LEAST_CONNECTIONS')) - lb_method_choices.append(('SOURCE_IP', 'SOURCE_IP')) + [lb_method_choices.append((m, m)) for m in AVAILABLE_METHODS] self.fields['lb_method'].choices = lb_method_choices # provider choice @@ -185,8 +186,7 @@ class AddVipAction(workflows.Action): args[0]['subnet']) protocol_choices = [('', _("Select a Protocol"))] - protocol_choices.append(('HTTP', 'HTTP')) - protocol_choices.append(('HTTPS', 'HTTPS')) + [protocol_choices.append((p, p)) for p in AVAILABLE_PROTOCOLS] self.fields['protocol'].choices = protocol_choices session_persistence_choices = [('', _("No Session Persistence"))]
lbaas/horizon - adds tcp protocol choice when create lb This option was missed Change-Id: I<I>c<I>a<I>a<I>da9b3dc<I>d<I>a<I>ebe Closes-Bug: #<I>
openstack_horizon
train
dcda31fb0dba39f4b1daa54298a6d04fff9dac1b
diff --git a/pkg/sdn/plugin/networkpolicy.go b/pkg/sdn/plugin/networkpolicy.go index <HASH>..<HASH> 100644 --- a/pkg/sdn/plugin/networkpolicy.go +++ b/pkg/sdn/plugin/networkpolicy.go @@ -364,7 +364,6 @@ func (np *networkPolicyPlugin) selectPods(npns *npNamespace, lsel *unversioned.L func (np *networkPolicyPlugin) parseNetworkPolicy(npns *npNamespace, policy *extensions.NetworkPolicy) (*npPolicy, error) { npp := &npPolicy{policy: *policy} - allowAll := false var destFlows []string if len(policy.Spec.PodSelector.MatchLabels) > 0 || len(policy.Spec.PodSelector.MatchExpressions) > 0 { @@ -377,11 +376,6 @@ func (np *networkPolicyPlugin) parseNetworkPolicy(npns *npNamespace, policy *ext } for _, rule := range policy.Spec.Ingress { - if len(rule.Ports) == 0 && len(rule.From) == 0 { - allowAll = true - break - } - var portFlows, peerFlows []string if len(rule.Ports) == 0 { portFlows = []string{""} @@ -446,11 +440,7 @@ func (np *networkPolicyPlugin) parseNetworkPolicy(npns *npNamespace, policy *ext } } - if allowAll { - npp.flows = []string{""} - } else { - sort.Strings(npp.flows) - } + sort.Strings(npp.flows) glog.V(5).Infof("Parsed NetworkPolicy: %#v", npp) return npp, nil }
Fix NetworkPolicies allowing from all to *some* (not all)
openshift_origin
train
2d174843befac50a5e301364281693bcd78d1019
diff --git a/lib/multi_client/version.rb b/lib/multi_client/version.rb index <HASH>..<HASH> 100644 --- a/lib/multi_client/version.rb +++ b/lib/multi_client/version.rb @@ -1,3 +1,3 @@ module MultiClient - VERSION = '3.1.1'.freeze + VERSION = '3.1.2'.freeze end
Bumped version to <I>
robotex82_multi_client
train
06fb9be4310e1f18f03cd2ace0df4f6cf0068709
diff --git a/examples/full.rb b/examples/full.rb index <HASH>..<HASH> 100644 --- a/examples/full.rb +++ b/examples/full.rb @@ -29,6 +29,9 @@ class UsersController < Cramp::Controller::Base end end + # Sends a space ( ' ' ) to the client for keeping the connection alive. Default : Every 30 seconds + keep_connection_alive :every => 10 + # Polls every 1 second by default periodic_timer :poll_user diff --git a/lib/cramp/controller/base.rb b/lib/cramp/controller/base.rb index <HASH>..<HASH> 100644 --- a/lib/cramp/controller/base.rb +++ b/lib/cramp/controller/base.rb @@ -32,6 +32,11 @@ module Cramp defined?(@@periodic_timers) ? @@periodic_timers : [] end + def self.keep_connection_alive(options = {}) + options = { :every => 30 }.merge(options) + periodic_timer :keep_connection_alive, options + end + def initialize(env) @env = env @timers = [] @@ -93,9 +98,6 @@ module Cramp end end - def start_timer - end - def stop_periodic_timers @timers.each {|t| t.cancel } end @@ -103,6 +105,10 @@ module Cramp def on_finish end + def keep_connection_alive + render " " + end + end end end
Add keep_connection_alive for long polling/streaming
lifo_cramp
train
2ad03a96b5e9ccbcb9e5b20f91018927397a622e
diff --git a/framework/base/DynamicModel.php b/framework/base/DynamicModel.php index <HASH>..<HASH> 100644 --- a/framework/base/DynamicModel.php +++ b/framework/base/DynamicModel.php @@ -50,6 +50,10 @@ use yii\validators\Validator; * "dynamic attributes". It basically allows an attribute to be defined dynamically through its constructor * or [[defineAttribute()]]. * + * Note that it does not support the usual validation flow as normal models, i.e. no scenarios are being defined + * and therefor none of the attributes is safe. This means that [[load()]] will not be able to perform massive + * assignment and return `false`. Attributes should be assigned via constructor instead. + * * @author Qiang Xue <[email protected]> * @since 2.0 */
Update DynamicModel.php added note about limitations, i.e. no scenarios -> no safe attributes -> no load().
yiisoft_yii-core
train
95428ddd6405d53f27867c3d1057e43cbbd2816e
diff --git a/lib/fake_web/registry.rb b/lib/fake_web/registry.rb index <HASH>..<HASH> 100644 --- a/lib/fake_web/registry.rb +++ b/lib/fake_web/registry.rb @@ -58,8 +58,10 @@ module FakeWeb def normalize_uri(uri) normalized_uri = case uri - when URI then uri - else + when URI then + uri.query = sort_query_params(uri.query) + uri + else uri = 'http://' + uri unless uri.match('^https?://') parsed_uri = URI.parse(uri) parsed_uri.query = sort_query_params(parsed_uri.query) diff --git a/test/test_query_string.rb b/test/test_query_string.rb index <HASH>..<HASH> 100644 --- a/test/test_query_string.rb +++ b/test/test_query_string.rb @@ -16,6 +16,11 @@ class TestFakeWebQueryString < Test::Unit::TestCase assert FakeWeb.registered_uri?('http://example.com/?b=1&a=1') end + def test_register_uri_with_query_params_unsorted_from_uri_object + FakeWeb.register_uri(URI.join('http://example.com/?b=1&a=1'), :string => 'foo') + assert FakeWeb.registered_uri?('http://example.com/?b=1&a=1') + end + def test_registered_uri_gets_recognized_with_empty_query_params FakeWeb.register_uri('http://example.com/', :string => 'foo') assert FakeWeb.registered_uri?('http://example.com/?')
Bug fixed: sorting query parameters from URI objects at register uri time.
chrisk_fakeweb
train
b19d67ccf239ba2bb0ca55cae20775cf8c51c771
diff --git a/Documentation/cli/README.md b/Documentation/cli/README.md index <HASH>..<HASH> 100644 --- a/Documentation/cli/README.md +++ b/Documentation/cli/README.md @@ -185,7 +185,7 @@ Set breakpoint condition. Specifies that the breakpoint, tracepoint or watchpoint should break only if the boolean expression is true. -See Documentation/cli/expr.md for a description of supported expressions. +See [Documentation/cli/expr.md](//github.com/go-delve/delve/tree/master/Documentation/cli/expr.md) for a description of supported expressions. With the -hitcount option a condition on the breakpoint hit count can be set, the following operators are supported @@ -511,7 +511,7 @@ Evaluate an expression. [goroutine <n>] [frame <m>] print [%format] <expression> -See Documentation/cli/expr.md for a description of supported expressions. +See [Documentation/cli/expr.md](//github.com/go-delve/delve/tree/master/Documentation/cli/expr.md) for a description of supported expressions. The optional format argument is a format specifier, like the ones used by the fmt package. For example "print %x v" will print v as an hexadecimal number. @@ -526,7 +526,7 @@ Print contents of CPU registers. regs [-a] -Argument -a shows more registers. Individual registers can also be displayed by 'print' and 'display'. See Documentation/cli/expr.md. +Argument -a shows more registers. Individual registers can also be displayed by 'print' and 'display'. See [Documentation/cli/expr.md.](//github.com/go-delve/delve/tree/master/Documentation/cli/expr.md.) ## restart @@ -569,7 +569,7 @@ Changes the value of a variable. [goroutine <n>] [frame <m>] set <variable> = <value> -See Documentation/cli/expr.md for a description of supported expressions. Only numerical variables and pointers can be changed. +See [Documentation/cli/expr.md](//github.com/go-delve/delve/tree/master/Documentation/cli/expr.md) for a description of supported expressions. Only numerical variables and pointers can be changed. ## source diff --git a/pkg/terminal/command.go b/pkg/terminal/command.go index <HASH>..<HASH> 100644 --- a/pkg/terminal/command.go +++ b/pkg/terminal/command.go @@ -124,14 +124,14 @@ Type "help" followed by the name of a command for more information about it.`}, break [name] [locspec] -See $GOPATH/src/github.com/go-delve/delve/Documentation/cli/locspec.md for the syntax of locspec. If locspec is omitted a breakpoint will be set on the current line. +See Documentation/cli/locspec.md for the syntax of locspec. If locspec is omitted a breakpoint will be set on the current line. See also: "help on", "help cond" and "help clear"`}, {aliases: []string{"trace", "t"}, group: breakCmds, cmdFn: tracepoint, allowedPrefixes: onPrefix, helpMsg: `Set tracepoint. trace [name] [locspec] -A tracepoint is a breakpoint that does not stop the execution of the program, instead when the tracepoint is hit a notification is displayed. See $GOPATH/src/github.com/go-delve/delve/Documentation/cli/locspec.md for the syntax of locspec. If locspec is omitted a tracepoint will be set on the current line. +A tracepoint is a breakpoint that does not stop the execution of the program, instead when the tracepoint is hit a notification is displayed. See Documentation/cli/locspec.md for the syntax of locspec. If locspec is omitted a tracepoint will be set on the current line. See also: "help on", "help cond" and "help clear"`}, {aliases: []string{"watch"}, group: breakCmds, cmdFn: watchpoint, helpMsg: `Set watchpoint. @@ -428,7 +428,7 @@ Executes the specified command (print, args, locals) in the context of the n-th source <path> -If path ends with the .star extension it will be interpreted as a starlark script. See $GOPATH/src/github.com/go-delve/delve/Documentation/cli/starlark.md for the syntax. +If path ends with the .star extension it will be interpreted as a starlark script. See Documentation/cli/starlark.md for the syntax. If path is a single '-' character an interactive starlark interpreter will start instead. Type 'exit' to exit.`}, {aliases: []string{"disassemble", "disass"}, cmdFn: disassCommand, helpMsg: `Disassembler. diff --git a/pkg/terminal/docgen.go b/pkg/terminal/docgen.go index <HASH>..<HASH> 100644 --- a/pkg/terminal/docgen.go +++ b/pkg/terminal/docgen.go @@ -7,13 +7,20 @@ import ( ) func replaceDocPath(s string) string { - const docpath = "$GOPATH/src/github.com/go-delve/delve/" + const docpath = "Documentation/" + + i0 := 0 for { - start := strings.Index(s, docpath) + start := strings.Index(s[i0:], docpath) if start < 0 { return s } + start += i0 + if start-1 >= 0 && s[start-1] != ' ' { + i0 = start + len(docpath) + 1 + continue + } var end int for end = start + len(docpath); end < len(s); end++ { if s[end] == ' ' { @@ -21,8 +28,9 @@ func replaceDocPath(s string) string { } } - text := s[start+len(docpath) : end] + text := s[start:end] s = s[:start] + fmt.Sprintf("[%s](//github.com/go-delve/delve/tree/master/%s)", text, text) + s[end:] + i0 = end + 1 } }
terminal: remove leftover GOPATH references (#<I>) Remove leftover references to $GOPATH in documentation, change script that generates markdown documentation to look for substrings that start with "Documentation/" instead.
go-delve_delve
train
5c64583e92a8875af5f9d0178012993d587553c7
diff --git a/lib/geoengineer/gps/yaml_tag.rb b/lib/geoengineer/gps/yaml_tag.rb index <HASH>..<HASH> 100644 --- a/lib/geoengineer/gps/yaml_tag.rb +++ b/lib/geoengineer/gps/yaml_tag.rb @@ -97,8 +97,17 @@ class GeoEngineer::GPS::YamlTag::Sub < GeoEngineer::GPS::YamlTag result_value = value.dup all_queries.each do |query| result = finder.dereference!(query, { auto_load: false }) + + # Process result differently per type + case result + when Array + result = result.to_json + when Hash + result = result.to_json + end result_value.gsub!(/{{\s*#{query}\s*}}/, result) end + result_value.to_json end
[Fix] allow lists in sub
coinbase_geoengineer
train
adf9721985be73ad4aefaaf5ec0167245c935855
diff --git a/check_release_installation.py b/check_release_installation.py index <HASH>..<HASH> 100644 --- a/check_release_installation.py +++ b/check_release_installation.py @@ -90,6 +90,7 @@ class Connection: def __init__(self, target_host): self.target_host = target_host def run_checked(self, command): + print(command) result = self.run(command) result.assert_succesful() def run(self, *user_command):
Make check_release_installation.py print each command to be run
andreafrancia_trash-cli
train
685c948d90dd35ffd9154e4926363ec3e2e868ff
diff --git a/contracts/networks.js b/contracts/networks.js index <HASH>..<HASH> 100644 --- a/contracts/networks.js +++ b/contracts/networks.js @@ -149,5 +149,6 @@ export default [ { name: 'ropsten', networkId: 3, contracts: null }, { name: 'rinkeby', networkId: 4, contracts: null }, { name: 'kovan', networkId: 42, contracts: contractInfo('kovan') }, + { name: 'test', networkId: 1337, contracts: contractInfo('test') }, { name: 'test', networkId: TESTNET_ID, contracts: contractInfo('test') } ]; diff --git a/src/eth/SmartContractService.js b/src/eth/SmartContractService.js index <HASH>..<HASH> 100644 --- a/src/eth/SmartContractService.js +++ b/src/eth/SmartContractService.js @@ -122,6 +122,7 @@ export default class SmartContractService extends PrivateService { if (this._addedContracts) { const networkName = { [TESTNET_ID]: 'testnet', + [1337]: 'testnet', [42]: 'kovan', [1]: 'mainnet' }[id];
add geth network id to networks config
makerdao_dai.js
train
092c5669af2ad6f1346567d4c71f6a60e9d287a5
diff --git a/library/actor-js/src/main/java/im/actor/model/js/angular/AngularList.java b/library/actor-js/src/main/java/im/actor/model/js/angular/AngularList.java index <HASH>..<HASH> 100644 --- a/library/actor-js/src/main/java/im/actor/model/js/angular/AngularList.java +++ b/library/actor-js/src/main/java/im/actor/model/js/angular/AngularList.java @@ -71,6 +71,15 @@ public class AngularList<T extends JavaScriptObject, V extends BserObject & List callbacks.remove(callback); } + public void forceReconvert(long id) { + for (int i = 0; i < values.size(); i++) { + if (values.get(i).getEngineId() == id) { + remove(jsValues, i); + insert(jsValues, i, entityConverter.convert(values.get(i), messenger)); + } + } + } + public void forceReconvert() { clear(jsValues); diff --git a/library/actor-js/src/main/java/im/actor/model/js/angular/AngularModule.java b/library/actor-js/src/main/java/im/actor/model/js/angular/AngularModule.java index <HASH>..<HASH> 100644 --- a/library/actor-js/src/main/java/im/actor/model/js/angular/AngularModule.java +++ b/library/actor-js/src/main/java/im/actor/model/js/angular/AngularModule.java @@ -167,16 +167,11 @@ public class AngularModule extends BaseModule implements AngularFileLoadedListen @Override public void onFileLoaded(long fileId) { if (dialogsList != null) { - boolean founded = false; for (Dialog dialog : dialogsList.getRawItems()) { if (checkAvatar(dialog.getDialogAvatar(), fileId)) { - founded = true; - break; + dialogsList.forceReconvert(dialog.getEngineId()); } } - if (founded) { - dialogsList.forceReconvert(); - } } for (AngularList<JsMessage, Message> messageList : messagesList.values()) {
wip(js): Improving file loading performance
actorapp_actor-platform
train
f7b630f81b3789a354f2b97a4773e90e3b60a54f
diff --git a/lib/mail/message.rb b/lib/mail/message.rb index <HASH>..<HASH> 100644 --- a/lib/mail/message.rb +++ b/lib/mail/message.rb @@ -1502,6 +1502,7 @@ module Mail end def add_required_fields + add_multipart_mixed_header unless parts.empty? @body = Mail::Body.new('') if body.nil? add_message_id unless (has_message_id? || self.class == Mail::Part) add_date unless has_date? diff --git a/spec/mail/attachments_spec.rb b/spec/mail/attachments_spec.rb index <HASH>..<HASH> 100644 --- a/spec/mail/attachments_spec.rb +++ b/spec/mail/attachments_spec.rb @@ -130,6 +130,15 @@ describe "Attachments" do end end + + describe "setting the content type correctly" do + it "should set the content type to multipart/mixed if none given and you add an attachment" do + mail = Mail.new + mail.attachments['test.pdf'] = File.read(fixture('attachments', 'test.pdf')) + mail.encoded + mail.mime_type.should == 'multipart/mixed' + end + end end
Message now adds multipart/mixed as the content type if nothing is set and there are parts to the message
mikel_mail
train
bd9b2220572653195013d38d56f311f5f668dc37
diff --git a/lib/patterns.rb b/lib/patterns.rb index <HASH>..<HASH> 100644 --- a/lib/patterns.rb +++ b/lib/patterns.rb @@ -92,8 +92,6 @@ module Patterns sidekiq-pro graphql-pro - # Blacklisted internal Rails dependency misspellings - # which could be used for malicious purposes. action-cable action_cable action-mailer
Remove comment from gem blacklist array closed #<I>
rubygems_rubygems.org
train
a4ce9b955eb7e495700fa841c26436d4ae3b7314
diff --git a/src/swipe/index.js b/src/swipe/index.js index <HASH>..<HASH> 100644 --- a/src/swipe/index.js +++ b/src/swipe/index.js @@ -1,5 +1,6 @@ // Utils import { createNamespace } from '../utils'; +import { isHidden } from '../utils/dom/style'; import { preventDefault } from '../utils/dom/event'; import { doubleRaf } from '../utils/dom/raf'; import { range } from '../utils/format/number'; @@ -149,7 +150,7 @@ export default createComponent({ initialize(active = +this.initialSwipe) { clearTimeout(this.timer); - if (this.$el) { + if (this.$el && !isHidden(this.$el)) { const rect = this.$el.getBoundingClientRect(); this.computedWidth = +this.width || rect.width; this.computedHeight = +this.height || rect.height;
fix(Swipe): incorrect width when resize in invisible state (#<I>)
youzan_vant
train
cda87fc468211987d808be08dc7ab06c059fa3e7
diff --git a/js/load-image-orientation.js b/js/load-image-orientation.js index <HASH>..<HASH> 100644 --- a/js/load-image-orientation.js +++ b/js/load-image-orientation.js @@ -119,7 +119,7 @@ Exif orientation values to correctly display the letter F: ctx.scale(1, -1) break case 5: - // vertical flip + 90 rotate right + // vertical flip + 90° rotate right ctx.rotate(0.5 * Math.PI) ctx.scale(1, -1) break @@ -129,7 +129,7 @@ Exif orientation values to correctly display the letter F: ctx.translate(0, -height) break case 7: - // horizontal flip + 90 rotate right + // horizontal flip + 90° rotate right ctx.rotate(0.5 * Math.PI) ctx.translate(width, -height) ctx.scale(-1, 1) @@ -185,7 +185,7 @@ Exif orientation values to correctly display the letter F: newOptions.bottom = options.top break case 5: - // vertical flip + 90 rotate right + // vertical flip + 90° rotate right newOptions.left = options.top newOptions.top = options.left newOptions.right = options.bottom @@ -199,7 +199,7 @@ Exif orientation values to correctly display the letter F: newOptions.bottom = options.left break case 7: - // horizontal flip + 90 rotate right + // horizontal flip + 90° rotate right newOptions.left = options.bottom newOptions.top = options.right newOptions.right = options.top
Add missing degree character in source comments.
blueimp_JavaScript-Load-Image
train
b49fe32d8491558173113beb087980acde99107f
diff --git a/tests/CacheTest.php b/tests/CacheTest.php index <HASH>..<HASH> 100644 --- a/tests/CacheTest.php +++ b/tests/CacheTest.php @@ -3,10 +3,10 @@ namespace ApiClients\Tools\CommandBus; use ApiClients\Tools\TestUtilities\TestCase; +use ExceptionalJSON\DecodeErrorException; use Generator; use Test\App\Commands\AwesomesauceCommand; use Test\App\Handlers\AwesomesauceHandler; -use WyriHaximus\Tactician\CommandHandler\Annotations\Handler; use function WyriHaximus\iteratorOrArrayToArray; /** @@ -33,6 +33,8 @@ final class CacheTest extends TestCase public function testReadNonExistent(): void { + self::expectException(DecodeErrorException::class); + $tmpFile = $this->getTmpDir() . \bin2hex(\random_bytes(13)) . '.json'; \touch($tmpFile);
Set expected exception for CacheTest::testReadNonExistent
php-api-clients_command-bus
train
59fb563e9ddda315ddb2bf39b1940602f2d39a94
diff --git a/engines/bastion_katello/app/assets/javascripts/bastion_katello/activation-keys/details/activation-key-associations.controller.js b/engines/bastion_katello/app/assets/javascripts/bastion_katello/activation-keys/details/activation-key-associations.controller.js index <HASH>..<HASH> 100644 --- a/engines/bastion_katello/app/assets/javascripts/bastion_katello/activation-keys/details/activation-key-associations.controller.js +++ b/engines/bastion_katello/app/assets/javascripts/bastion_katello/activation-keys/details/activation-key-associations.controller.js @@ -16,7 +16,7 @@ angular.module('Bastion.activation-keys').controller('ActivationKeyAssociationsController', ['$scope', '$location', 'translate', 'Nutupane', 'ActivationKey', 'ContentHostsHelper', 'CurrentOrganization', 'Host', function ($scope, $location, translate, Nutupane, ActivationKey, ContentHostsHelper, CurrentOrganization, Host) { - var contentHostsNutupane, params = { + var contentHostsNutupane, nutupaneParams, params = { 'organization_id': CurrentOrganization, 'search': $location.search().search || "", 'page': 1, @@ -25,7 +25,11 @@ angular.module('Bastion.activation-keys').controller('ActivationKeyAssociationsC 'paged': true }; - contentHostsNutupane = new Nutupane(Host, params); + nutupaneParams = { + 'disableAutoLoad': true + }; + + contentHostsNutupane = new Nutupane(Host, params, undefined, nutupaneParams); $scope.controllerName = 'hosts'; contentHostsNutupane.searchTransform = function (term) { var searchQuery, addition = "activation_key_id=" + $scope.$stateParams.activationKeyId; diff --git a/engines/bastion_katello/app/assets/javascripts/bastion_katello/host-collections/details/host-collection-add-hosts.controller.js b/engines/bastion_katello/app/assets/javascripts/bastion_katello/host-collections/details/host-collection-add-hosts.controller.js index <HASH>..<HASH> 100644 --- a/engines/bastion_katello/app/assets/javascripts/bastion_katello/host-collections/details/host-collection-add-hosts.controller.js +++ b/engines/bastion_katello/app/assets/javascripts/bastion_katello/host-collections/details/host-collection-add-hosts.controller.js @@ -16,7 +16,7 @@ angular.module('Bastion.host-collections').controller('HostCollectionAddHostsController', ['$scope', '$state', '$location', 'translate', 'Nutupane', 'CurrentOrganization', 'Host', 'HostCollection', function ($scope, $state, $location, translate, Nutupane, CurrentOrganization, Host, HostCollection) { - var contentNutupane, params; + var contentNutupane, params, nutupaneParams; params = { 'organization_id': CurrentOrganization, @@ -27,7 +27,10 @@ angular.module('Bastion.host-collections').controller('HostCollectionAddHostsCon 'paged': true }; - contentNutupane = new Nutupane(Host, params); + nutupaneParams = { + 'disableAutoLoad': true + }; + contentNutupane = new Nutupane(Host, params, undefined, nutupaneParams); $scope.controllerName = 'hosts'; contentNutupane.searchTransform = function (term) { var addition = "-host_collection_id=" + $scope.$stateParams.hostCollectionId; diff --git a/engines/bastion_katello/app/assets/javascripts/bastion_katello/host-collections/details/host-collection-hosts.controller.js b/engines/bastion_katello/app/assets/javascripts/bastion_katello/host-collections/details/host-collection-hosts.controller.js index <HASH>..<HASH> 100644 --- a/engines/bastion_katello/app/assets/javascripts/bastion_katello/host-collections/details/host-collection-hosts.controller.js +++ b/engines/bastion_katello/app/assets/javascripts/bastion_katello/host-collections/details/host-collection-hosts.controller.js @@ -16,7 +16,7 @@ angular.module('Bastion.host-collections').controller('HostCollectionHostsController', ['$scope', '$location', 'translate', 'Nutupane', 'CurrentOrganization', 'Host', 'HostCollection', function ($scope, $location, translate, Nutupane, CurrentOrganization, Host, HostCollection) { - var params; + var params, nutupaneParams; params = { 'organization_id': CurrentOrganization, @@ -27,7 +27,11 @@ angular.module('Bastion.host-collections').controller('HostCollectionHostsContro 'paged': true }; - $scope.contentNutupane = new Nutupane(Host, params); + nutupaneParams = { + 'disableAutoLoad': true + }; + + $scope.contentNutupane = new Nutupane(Host, params, undefined, nutupaneParams); $scope.controllerName = 'hosts'; $scope.contentNutupane.searchTransform = function (term) { var addition = "host_collection_id=" + $scope.$stateParams.hostCollectionId;
fixes #<I> - Host Collection/Host views not filtering correctly
Katello_katello
train
0583248ecfef5bec8efeaa8e65907f1e9ad0bf1d
diff --git a/lib/paru/selector.rb b/lib/paru/selector.rb index <HASH>..<HASH> 100644 --- a/lib/paru/selector.rb +++ b/lib/paru/selector.rb @@ -187,8 +187,10 @@ module Paru def is_descendant?(node) distance = 0 + parent = nil begin distance += 1 if @distance > 0 + node = parent unless parent.nil? parent = node.parent ancestry = parent.type == @type and @classes.all? {|c| parent.has_class? c} end while not ancestry and not parent.is_root? and distance <= @distance
fix `Relation#is_descendent?` looping forever if the immediate parent didn't match, this used to loop forever; now it correctly walks up the tree looking for a matching parent.
htdebeer_paru
train
8e7b0c3e6390f483fdaacb4c812996a57d97ca87
diff --git a/gems/aws-sdk-code-generator/lib/aws-sdk-code-generator/dsl/attribute_accessor.rb b/gems/aws-sdk-code-generator/lib/aws-sdk-code-generator/dsl/attribute_accessor.rb index <HASH>..<HASH> 100644 --- a/gems/aws-sdk-code-generator/lib/aws-sdk-code-generator/dsl/attribute_accessor.rb +++ b/gems/aws-sdk-code-generator/lib/aws-sdk-code-generator/dsl/attribute_accessor.rb @@ -4,25 +4,30 @@ module AwsSdkCodeGenerator include Dsl::CodeObject - def initialize(name, &block) + def initialize(name, api_private: false, &block) @name = name - @return_tags = [] + @tags = [] yield(self) if block + api_private! if api_private end attr_reader :name def returns(type, docstring:nil) - @return_tags << ReturnTag.new(type:type, docstring:docstring) + @tags << ReturnTag.new(type:type, docstring:docstring) + end + + def api_private! + @tags << "# @api private" end def documented? - !@return_tags.empty? + [email protected]? end def lines lines = [] - @return_tags.each do |tag| + @tags.each do |tag| lines.concat(tag.lines) end lines + ["#{macro} :#{@name}"] diff --git a/gems/aws-sdk-code-generator/lib/aws-sdk-code-generator/dsl/module.rb b/gems/aws-sdk-code-generator/lib/aws-sdk-code-generator/dsl/module.rb index <HASH>..<HASH> 100644 --- a/gems/aws-sdk-code-generator/lib/aws-sdk-code-generator/dsl/module.rb +++ b/gems/aws-sdk-code-generator/lib/aws-sdk-code-generator/dsl/module.rb @@ -54,8 +54,8 @@ module AwsSdkCodeGenerator @code_objects << AutoloadStatement.new(const_name, path) end - def attr_reader(name, &block) - a = Dsl::AttributeReader.new(name) + def attr_reader(name, options = {}, &block) + a = Dsl::AttributeReader.new(name, options) yield(a) if block add(a) end diff --git a/gems/aws-sdk-code-generator/lib/aws-sdk-code-generator/generators/client_class.rb b/gems/aws-sdk-code-generator/lib/aws-sdk-code-generator/generators/client_class.rb index <HASH>..<HASH> 100644 --- a/gems/aws-sdk-code-generator/lib/aws-sdk-code-generator/generators/client_class.rb +++ b/gems/aws-sdk-code-generator/lib/aws-sdk-code-generator/generators/client_class.rb @@ -46,7 +46,7 @@ module AwsSdkCodeGenerator apply_waiter_methods(self) eigenclass do |eigenclass| eigenclass.docstring('@api private') - eigenclass.attr_reader('identifier') + eigenclass.attr_reader('identifier', api_private: true) eigenclass.method('errors_module') do |m| m.code('Errors') end
Client class .identifier methods are now marked @api private.
aws_aws-sdk-ruby
train
63b1cf1fe2d3ed62299f57cdd90890de7745b4b8
diff --git a/auth/ldap/auth.php b/auth/ldap/auth.php index <HASH>..<HASH> 100644 --- a/auth/ldap/auth.php +++ b/auth/ldap/auth.php @@ -478,44 +478,43 @@ class auth_plugin_ldap extends auth_plugin_base { $droptablesql = array(); /// sql commands to drop the table (because session scope could be a problem for /// some persistent drivers like ODBTP (mssql) or if this function is invoked /// from within a PHP application using persistent connections + $temptable = $CFG->prefix . 'extuser'; + $createtemptablesql = ''; // configure a temp table print "Configuring temp table\n"; switch (strtolower($CFG->dbfamily)) { case 'mysql': - $temptable = $CFG->prefix . 'extuser'; $droptablesql[] = 'DROP TEMPORARY TABLE ' . $temptable; // sql command to drop the table (because session scope could be a problem) - execute_sql_arr($droptablesql, true, false); /// Drop temp table to avoid persistence problems later - echo "Creating temp table $temptable\n"; - execute_sql('CREATE TEMPORARY TABLE ' . $temptable . ' (username VARCHAR(64), PRIMARY KEY (username)) TYPE=MyISAM', false); + $createtemptablesql = 'CREATE TEMPORARY TABLE ' . $temptable . ' (username VARCHAR(64), PRIMARY KEY (username)) TYPE=MyISAM'; break; case 'postgres': - $temptable = $CFG->prefix . 'extuser'; $droptablesql[] = 'DROP TABLE ' . $temptable; // sql command to drop the table (because session scope could be a problem) - execute_sql_arr($droptablesql, true, false); /// Drop temp table to avoid persistence problems later - echo "Creating temp table $temptable\n"; $bulk_insert_records = 1; // no support for multiple sets of values - execute_sql('CREATE TEMPORARY TABLE '. $temptable . ' (username VARCHAR(64), PRIMARY KEY (username))', false); + $createtemptablesql = 'CREATE TEMPORARY TABLE '. $temptable . ' (username VARCHAR(64), PRIMARY KEY (username))'; break; case 'mssql': - $temptable = '#'.$CFG->prefix . 'extuser'; /// MSSQL temp tables begin with # + $temptable = '#'. $temptable; /// MSSQL temp tables begin with # $droptablesql[] = 'DROP TABLE ' . $temptable; // sql command to drop the table (because session scope could be a problem) - execute_sql_arr($droptablesql, true, false); /// Drop temp table to avoid persistence problems later - echo "Creating temp table $temptable\n"; $bulk_insert_records = 1; // no support for multiple sets of values - execute_sql('CREATE TABLE ' . $temptable . ' (username VARCHAR(64), PRIMARY KEY (username))', false); + $createtemptablesql = 'CREATE TABLE ' . $temptable . ' (username VARCHAR(64), PRIMARY KEY (username))'; break; case 'oracle': - $temptable = $CFG->prefix . 'extuser'; $droptablesql[] = 'TRUNCATE TABLE ' . $temptable; // oracle requires truncate before being able to drop a temp table $droptablesql[] = 'DROP TABLE ' . $temptable; // sql command to drop the table (because session scope could be a problem) - execute_sql_arr($droptablesql, true, false); /// Drop temp table to avoid persistence problems later - echo "Creating temp table $temptable\n"; $bulk_insert_records = 1; // no support for multiple sets of values - execute_sql('CREATE GLOBAL TEMPORARY TABLE '.$temptable.' (username VARCHAR(64), PRIMARY KEY (username)) ON COMMIT PRESERVE ROWS', false); + $createtemptablesql = 'CREATE GLOBAL TEMPORARY TABLE '.$temptable.' (username VARCHAR(64), PRIMARY KEY (username)) ON COMMIT PRESERVE ROWS'; break; } + + execute_sql_arr($droptablesql, true, false); /// Drop temp table to avoid persistence problems later + echo "Creating temp table $temptable\n"; + if(! execute_sql($createtemptablesql, false) ){ + print "Failed to create temporary users table - aborting\n"; + exit; + } + print "Connecting to ldap...\n"; $ldapconnection = $this->ldap_connect();
MDL-<I> - abort early when can't create temporary tables when syning users from LDAP
moodle_moodle
train
095cf6d37a33da2b64d45802cb81be9476a07de6
diff --git a/src/DocBlock/Tags/Generic.php b/src/DocBlock/Tags/Generic.php index <HASH>..<HASH> 100644 --- a/src/DocBlock/Tags/Generic.php +++ b/src/DocBlock/Tags/Generic.php @@ -53,7 +53,7 @@ class Generic extends BaseTag implements Factory\StaticMethod Assert::stringNotEmpty($name); Assert::notNull($descriptionFactory); - $description = $descriptionFactory && $body ? $descriptionFactory->create($body, $context) : null; + $description = $descriptionFactory && $body !== "" ? $descriptionFactory->create($body, $context) : null; return new static($name, $description); }
fixed generic tag to parse description with `0` as expected
phpDocumentor_ReflectionDocBlock
train
bbc26e716010e48871e692436a4e2b22e259cc01
diff --git a/deliver/lib/deliver/download_screenshots.rb b/deliver/lib/deliver/download_screenshots.rb index <HASH>..<HASH> 100644 --- a/deliver/lib/deliver/download_screenshots.rb +++ b/deliver/lib/deliver/download_screenshots.rb @@ -13,7 +13,7 @@ module Deliver end def self.download(options, folder_path) - v = options[:use_live_version] ? options[:app].live_version : options[:app].latest_version + v = options[:use_live_version] ? options[:app].live_version(platform: options[:platform]) : options[:app].latest_version(platform: options[:platform]) v.screenshots.each do |language, screenshots| screenshots.each do |screenshot| diff --git a/deliver/lib/deliver/options.rb b/deliver/lib/deliver/options.rb index <HASH>..<HASH> 100644 --- a/deliver/lib/deliver/options.rb +++ b/deliver/lib/deliver/options.rb @@ -372,6 +372,10 @@ module Deliver description: "Metadata: Localised privacy url", optional: true, is_string: false), + FastlaneCore::ConfigItem.new(key: :apple_tv_privacy_policy, + description: "Metadata: Localised Apple TV privacy policy text", + optional: true, + is_string: false), FastlaneCore::ConfigItem.new(key: :support_url, description: "Metadata: Localised support url", optional: true, diff --git a/deliver/lib/deliver/upload_metadata.rb b/deliver/lib/deliver/upload_metadata.rb index <HASH>..<HASH> 100644 --- a/deliver/lib/deliver/upload_metadata.rb +++ b/deliver/lib/deliver/upload_metadata.rb @@ -10,7 +10,7 @@ module Deliver NON_LOCALISED_VERSION_VALUES = [:copyright] # Localised app details values - LOCALISED_APP_VALUES = [:name, :subtitle, :privacy_url] + LOCALISED_APP_VALUES = [:name, :subtitle, :privacy_url, :apple_tv_privacy_policy] # Non localized app details values NON_LOCALISED_APP_VALUES = [:primary_category, :secondary_category, diff --git a/deliver/lib/deliver/upload_screenshots.rb b/deliver/lib/deliver/upload_screenshots.rb index <HASH>..<HASH> 100644 --- a/deliver/lib/deliver/upload_screenshots.rb +++ b/deliver/lib/deliver/upload_screenshots.rb @@ -42,7 +42,7 @@ module Deliver Helper.show_loading_indicator("Activating #{lng_text} #{enabled_languages.join(', ')}...") v.save! # This refreshes the app version from iTC after enabling a localization - v = app.edit_version + v = app.edit_version(platform: options[:platform]) Helper.hide_loading_indicator end
apple tv privacy policy upload and download handling added (#<I>) screenshot download and upload issue for platform appletvos when ios and appletvos both are present at appstoreconnect for same bundle id
fastlane_fastlane
train
675504a68650f0919f609f50ac5c22ea7ff2a098
diff --git a/caniusepython3/test/test_cli.py b/caniusepython3/test/test_cli.py index <HASH>..<HASH> 100644 --- a/caniusepython3/test/test_cli.py +++ b/caniusepython3/test/test_cli.py @@ -149,11 +149,19 @@ class CLITests(unittest.TestCase): 'its transition:') self.assertEqual(messages[1], want) - def test_message_no_blockers(self): + @mock.patch('sys.stdout', autospec=True) + def test_message_no_blockers_flair_on_utf8_terminal(self, mock_stdout): + mock_stdout.encoding = 'UTF-8' messages = ciu_main.message([]) - self.assertEqual( - ['You have 0 projects blocking you from using Python 3!'], - messages) + expected = ['\U0001f389 You have 0 projects blocking you from using Python 3!'] + self.assertEqual(expected, messages) + + @mock.patch('sys.stdout', autospec=True) + def test_message_no_blockers(self, mock_stdout): + mock_stdout.encoding = None + messages = ciu_main.message([]) + expected = ['You have 0 projects blocking you from using Python 3!'] + self.assertEqual(expected, messages) def test_pprint_blockers(self): simple = [['A']]
Fix failing CLI test on UTF-8 terminals.
brettcannon_caniusepython3
train
0a5cc8d7cf641384c4e39df5b6905471b1eef454
diff --git a/src/sjl.js b/src/sjl.js index <HASH>..<HASH> 100644 --- a/src/sjl.js +++ b/src/sjl.js @@ -212,38 +212,83 @@ }); } + /** + * Checs if value is a valid number (also checks if isNaN so that you don't have to). + * @param value {*} + * @returns {Boolean} + */ function isNumber (value) { return classOfIs(value, Number); } + /** + * Returns whether a value is a function or not. + * @param value {*} + * @returns {Boolean} + */ function isFunction (value) { return classOfIs(value, Function); } + /** + * Checks if value is an array. + * @param value {*} + * @returns {boolean} + */ function isArray (value) { return Array.isArray(value); } + /** + * Checks if value is a boolean. + * @param value {*} + * @returns {Boolean} + */ function isBoolean (value) { return classOfIs(value, Boolean); } + /** + * Checks whether value is an object or not. + * @param value + * @returns {Boolean} + */ function isObject (value) { return classOfIs(value, Object); } + /** + * Checks whether value is a string or not. + * @param value {*} + * @returns {Boolean} + */ function isString(value) { return classOfIs(value, String); } + /** + * Checks if value is undefined. + * @param value {*} + * @returns {Boolean} + */ function isUndefined (value) { return classOfIs(value, 'Undefined'); } + /** + * Checks if value is null. + * @param value {*} + * @returns {Boolean} + */ function isNull (value) { return classOfIs(value, 'Null'); } + /** + * Checks if value is a `Symbol`. + * @param value {*} + * @returns {Boolean} + */ function isSymbol (value) { return classOfIs(value, 'Symbol'); }
Added missing docs to src/sjl.js.
elycruz_sjljs
train
2a34ebb19e8dfdad36bf6a143a3aa935a7edc77f
diff --git a/indra/preassembler/__init__.py b/indra/preassembler/__init__.py index <HASH>..<HASH> 100644 --- a/indra/preassembler/__init__.py +++ b/indra/preassembler/__init__.py @@ -207,35 +207,39 @@ class Preassembler(object): if not stmt.supports] return self.related_stmts - def check_sequence(self): - """Iterate over all Statements and check whether references to + def check_statements(self): + for st in self.stmts: + print st + self.check_sequence(st) + + @staticmethod + def check_sequence(st): + """Check whether references to residues and sequence positions are consistent with sequence information in the UniProt database""" # TODO: extend to all Agent modifications - for st in self.stmts: - if isinstance(st, Phosphorylation): - # Skip statements with no specified position - if st.mod_pos is None: - continue - # Is looking at the first element enough? - sub_id = st.sub.db_refs['UP'] - if not isinstance(sub_id, basestring): - sub_id = sub_id[0] - sub = uniprot_client.query_protein(sub_id) - if st.mod == 'PhosphorylationSerine': - residue = 'S' - elif st.mod == 'PhosphorylationThreonine': - residue = 'T' - elif st.mod == 'PhosphorylationTyrosine': - residue = 'Y' - else: - continue - ver = uniprot_client.verify_location(sub, residue, - location=st.mod_pos) - if ver is False: - print '%s:' % st - print '-> Sequence check failed; position %s on %s is not %s.' %\ - (st.mod_pos, st.sub.name, residue) + if isinstance(st, Phosphorylation): + # Skip statements with no specified position + if st.mod_pos is None: + return + # Is looking at the first element enough? + sub_id = st.sub.db_refs['UP'] + if not isinstance(sub_id, basestring): + sub_id = sub_id[0] + sub = uniprot_client.query_protein(sub_id) + if st.mod == 'PhosphorylationSerine': + residue = 'S' + elif st.mod == 'PhosphorylationThreonine': + residue = 'T' + elif st.mod == 'PhosphorylationTyrosine': + residue = 'Y' + else: + return + ver = uniprot_client.verify_location(sub, residue, + location=st.mod_pos) + if ver is False: + print '-> Sequence check failed; position %s on %s is not %s.' %\ + (st.mod_pos, st.sub.name, residue)
Change check statement loop structure in Preassembler
sorgerlab_indra
train
fd9bfbf28e045362c83620dd888c5edbb6f39a89
diff --git a/lib/docx/docx-reader.js b/lib/docx/docx-reader.js index <HASH>..<HASH> 100644 --- a/lib/docx/docx-reader.js +++ b/lib/docx/docx-reader.js @@ -100,26 +100,23 @@ function findPartPaths(docxFile) { readElement: relationshipsReader.readRelationships, defaultValue: relationshipsReader.defaultValue })(docxFile).then(function(documentRelationships) { - function findPartRelatedToMainDocument(options) { + function findPartRelatedToMainDocument(name) { return findPartPath({ docxFile: docxFile, relationships: documentRelationships, - relationshipType: options.relationshipType, + relationshipType: "http://schemas.openxmlformats.org/officeDocument/2006/relationships/" + name, basePath: zipfile.splitPath(mainDocumentPath).dirname, - fallbackPath: options.fallbackPath + fallbackPath: "word/" + name + ".xml" }); } return { mainDocument: mainDocumentPath, - comments: findPartRelatedToMainDocument({ - relationshipType: "http://schemas.openxmlformats.org/officeDocument/2006/relationships/comments", - fallbackPath: "word/comments.xml" - }), - endnotes: "word/endnotes.xml", - footnotes: "word/footnotes.xml", - numbering: "word/numbering.xml", - styles: "word/styles.xml" + comments: findPartRelatedToMainDocument("comments"), + endnotes: findPartRelatedToMainDocument("endnotes"), + footnotes: findPartRelatedToMainDocument("footnotes"), + numbering: findPartRelatedToMainDocument("numbering"), + styles: findPartRelatedToMainDocument("styles") }; }); }); diff --git a/test/docx/docx-reader.tests.js b/test/docx/docx-reader.tests.js index <HASH>..<HASH> 100644 --- a/test/docx/docx-reader.tests.js +++ b/test/docx/docx-reader.tests.js @@ -88,6 +88,7 @@ test("error is thrown when main document part does not exist", function() { }); + test("part paths", { "main document part is found using package relationships": function() { var relationships = xml.element("r:Relationships", {}, [ @@ -113,35 +114,59 @@ test("part paths", { return docxReader._findPartPaths(docxFile).then(function(partPaths) { assert.equal(partPaths.mainDocument, "word/document.xml"); }); + } +}); + +[ + { + name: "comments", + type: "http://schemas.openxmlformats.org/officeDocument/2006/relationships/comments", }, - - "comments part is found using main document relationships": function() { + { + name: "endnotes", + type: "http://schemas.openxmlformats.org/officeDocument/2006/relationships/endnotes", + }, + { + name: "footnotes", + type: "http://schemas.openxmlformats.org/officeDocument/2006/relationships/footnotes", + }, + { + name: "numbering", + type: "http://schemas.openxmlformats.org/officeDocument/2006/relationships/numbering", + }, + { + name: "styles", + type: "http://schemas.openxmlformats.org/officeDocument/2006/relationships/styles", + } +].forEach(function(options) { + test(options.name + " part is found using main document relationships", function() { var docxFile = createFakeDocxFile({ "_rels/.rels": createPackageRelationships("word/document.xml"), "word/document.xml": " ", "word/_rels/document.xml.rels": xml.writeString(xml.element("r:Relationships", {}, [ xml.element("r:Relationship", { - "Type": "http://schemas.openxmlformats.org/officeDocument/2006/relationships/comments", - "Target": "comments2.xml" + "Type": options.type, + "Target": "target-path.xml" }) ]), relationshipNamespaces), - "word/comments2.xml": " " + "word/target-path.xml": " " }); return docxReader._findPartPaths(docxFile).then(function(partPaths) { - assert.equal(partPaths.comments, "word/comments2.xml"); + assert.equal(partPaths[options.name], "word/target-path.xml"); }); - }, - - "word/comments.xml is used as fallback location for comments part": function() { - var docxFile = createFakeDocxFile({ + }); + + test("word/" + options.name + ".xml is used as fallback location for " + options.name + " part", function() { + var zipContents = { "_rels/.rels": createPackageRelationships("word/document.xml"), - "word/document.xml": " ", - "word/comments.xml": " " - }); + "word/document.xml": " " + }; + zipContents["word/" + options.name + ".xml"] = " "; + var docxFile = createFakeDocxFile(zipContents); return docxReader._findPartPaths(docxFile).then(function(partPaths) { - assert.equal(partPaths.comments, "word/comments.xml"); + assert.equal(partPaths[options.name], "word/" + options.name + ".xml"); }); - } + }); });
Find parts related to main document using main document relationships
mwilliamson_mammoth.js
train
47739b0772942057a8ba5bbc037d1e1e064a1b8c
diff --git a/lib/level.js b/lib/level.js index <HASH>..<HASH> 100644 --- a/lib/level.js +++ b/lib/level.js @@ -1,6 +1,6 @@ 'use strict'; -exports.ALL = -100000; +exports.ALL = -Infinity; /** * Debug log for execute tracing @@ -31,4 +31,4 @@ exports.WARN = 2; */ exports.ERROR = 3; -exports.NONE = 100000; +exports.NONE = Infinity;
use Infinity to improve performance and semantics (#7)
eggjs_egg-logger
train
5604ce7d661cca2d8e7a9f5c03d5c5a2d11a16e9
diff --git a/bosh-director/lib/bosh/director/client.rb b/bosh-director/lib/bosh/director/client.rb index <HASH>..<HASH> 100644 --- a/bosh-director/lib/bosh/director/client.rb +++ b/bosh-director/lib/bosh/director/client.rb @@ -22,9 +22,13 @@ module Bosh::Director end def method_missing(method_name, *args) - retries = @retry_methods[method_name] || 0 + send_message(method_name, *args) + end + + def send_message(message_name, *args) + retries = @retry_methods[message_name] || 0 begin - handle_method(method_name, args) + handle_method(message_name, args) rescue RpcTimeout if retries > 0 retries -= 1 diff --git a/bosh-director/spec/unit/agent_client_spec.rb b/bosh-director/spec/unit/agent_client_spec.rb index <HASH>..<HASH> 100644 --- a/bosh-director/spec/unit/agent_client_spec.rb +++ b/bosh-director/spec/unit/agent_client_spec.rb @@ -2,7 +2,56 @@ require 'spec_helper' module Bosh::Director describe AgentClient do - describe 'long-runnings messages' do + shared_examples_for 'a long running message' do |message_name| + describe "##{message_name}" do + let(:task) do + { + 'agent_task_id' => 'fake-agent_task_id', + 'state' => 'running', + 'value' => 'task value' + } + end + + before do + client.stub(send_message: task) + client.stub(:get_task) do + task['state'] = 'no longer running' + task + end + + client.stub(:sleep).with(AgentClient::DEFAULT_POLL_INTERVAL) + end + + it 'explicitly defines methods for long running messages (to poll)' do + expect(client).to respond_to(message_name) + end + + it 'decorates the original send_message implementation' do + client.public_send(message_name, 'fake', 'args') + + expect(client).to have_received(:send_message).with(message_name, 'fake', 'args') + end + + it 'periodically polls the task while it is running' do + client.public_send(message_name, 'fake', 'args') + + expect(client).to have_received(:get_task).with('fake-agent_task_id') + end + + it 'stops polling once the task is no longer running' do + task['state'] = 'something other than running' + client.public_send(message_name, 'fake', 'args') + + expect(client).not_to have_received(:get_task) + end + + it 'returns the task value' do + expect(client.public_send(message_name, 'fake', 'args')).to eq('task value') + end + end + end + + describe 'long running messages' do let(:vm) do instance_double('Models::Vm', credentials: nil) end @@ -17,19 +66,15 @@ module Bosh::Director Api::ResourceManager.stub(:new) end - it 'explicitly defines methods for long running messages (to poll their tasks)' do - expect(client.methods).to include( - :prepare, - :apply, - :compile_package, - :drain, - :fetch_logs, - :migrate_disk, - :mount_disk, - :stop, - :unmount_disk, - ) - end + include_examples 'a long running message', :prepare + include_examples 'a long running message', :apply + include_examples 'a long running message', :compile_package + include_examples 'a long running message', :drain + include_examples 'a long running message', :fetch_logs + include_examples 'a long running message', :migrate_disk + include_examples 'a long running message', :mount_disk + include_examples 'a long running message', :stop + include_examples 'a long running message', :unmount_disk end describe 'ping <=> pong' do
Backfill spec for precarious Client <=> AgentClient interaction
cloudfoundry_bosh
train
ef10e301da290a0569c65932487fc33d8bf06605
diff --git a/troposphere/kinesis.py b/troposphere/kinesis.py index <HASH>..<HASH> 100644 --- a/troposphere/kinesis.py +++ b/troposphere/kinesis.py @@ -3,10 +3,17 @@ # # See LICENSE file for full license. -from . import AWSObject, Tags +from . import AWSObject, AWSProperty, Tags from .validators import integer +class StreamEncryption(AWSProperty): + props = { + 'EncryptionType': (basestring, True), + 'KeyId': (basestring, True), + } + + class Stream(AWSObject): resource_type = "AWS::Kinesis::Stream" @@ -14,5 +21,6 @@ class Stream(AWSObject): 'Name': (basestring, False), 'RetentionPeriodHours': (integer, False), 'ShardCount': (integer, False), + 'StreamEncryption': (StreamEncryption, False), 'Tags': ((Tags, list), False), }
Add StreamEncryption to AWS::Kinesis::Stream
cloudtools_troposphere
train
ac8e56288291723bfdcb08353391c2852a8c8185
diff --git a/can-dom-events-test.js b/can-dom-events-test.js index <HASH>..<HASH> 100644 --- a/can-dom-events-test.js +++ b/can-dom-events-test.js @@ -216,3 +216,30 @@ unit.test('domEvents.addDelegateListener call inner-most handler first (#62)', f domEvents.dispatch(child, 'click'); done(); }); + +unit.test('domEvents.addDelegateListener should have a working ev.stopPropagation (#62)', function (assert) { + var done = assert.async(); + var grandparent = document.createElement('div'); + var parent = document.createElement('p'); + var child = document.createElement('input'); + + grandparent.appendChild(parent); + parent.appendChild(child); + + var paragraphClickHandler = function paragraphClickHandler() { + domEvents.removeDelegateListener(grandparent, 'click', 'p', paragraphClickHandler); + assert.ok(false, 'ev.stopPropagation works'); + }; + + var inputClickHandler = function inputClickHandler(event) { + event.stopPropagation(); + domEvents.removeDelegateListener(grandparent, 'click', 'input', inputClickHandler); + assert.ok(true, 'inner-most click handler called first'); + }; + + domEvents.addDelegateListener(grandparent, 'click', 'p', paragraphClickHandler); + domEvents.addDelegateListener(grandparent, 'click', 'input', inputClickHandler); + + domEvents.dispatch(child, 'click'); + done(); +}); diff --git a/helpers/-make-delegate-event-tree.js b/helpers/-make-delegate-event-tree.js index <HASH>..<HASH> 100644 --- a/helpers/-make-delegate-event-tree.js +++ b/helpers/-make-delegate-event-tree.js @@ -20,6 +20,10 @@ function makeDelegator (domEvents) { "can.setKeyValue": function(eventType, handlersBySelector){ var handler = this.delegated[eventType] = function(ev){ var cur = ev.target; + var propagate = true; + ev.stopPropagation = function() { + propagate = false; + }; do { // document does not implement `.matches` but documentElement does var el = cur === document ? document.documentElement : cur; @@ -38,7 +42,7 @@ function makeDelegator (domEvents) { // we need to continue using `cur` as the loop pointer, otherwhise // it will never end as documentElement.parentNode === document cur = cur.parentNode; - } while (cur && cur !== ev.currentTarget); + } while ((cur && cur !== ev.currentTarget) && propagate); }; this.events[eventType] = handlersBySelector; domEvents.addEventListener(this.element, eventType, handler, useCapture(eventType));
making ev.stopPropagation work with delegate listeners
canjs_can-dom-events
train
cdac35e05752f657a36e95a0deae8d2a1d0e4b46
diff --git a/pyemma/coordinates/tests/test_readers.py b/pyemma/coordinates/tests/test_readers.py index <HASH>..<HASH> 100644 --- a/pyemma/coordinates/tests/test_readers.py +++ b/pyemma/coordinates/tests/test_readers.py @@ -123,7 +123,7 @@ class TestReaders(six.with_metaclass(GenerateTestMatrix, unittest.TestCase)): 'xtc', 'trr', 'dcd', - 'h5', + 'h5', # pyemma H5Reader 'csv', ) # transform data or not (identity does not change the data, but pushes it through a StreamingTransformer). @@ -179,6 +179,16 @@ class TestReaders(six.with_metaclass(GenerateTestMatrix, unittest.TestCase)): def teardown_class(cls): shutil.rmtree(cls.tempdir, ignore_errors=True) + # H5Reader does not need a topology (and gets only selected in case we do not pass it). + def setUp(self): + if 'h5' in self._testMethodName: + self.pdb_file_bak = self.pdb_file + self.pdb_file = None + + def tearDown(self): + if 'h5' in self._testMethodName: + self.pdb_file = self.pdb_file_bak + def _test_lagged_reader(self, file_format, stride, skip, chunksize, lag): trajs = self.test_trajs[file_format] reader = coor.source(trajs, top=self.pdb_file, chunksize=chunksize)
[test-readers] do no pass topology, if H5Reader is required.
markovmodel_PyEMMA
train
3990e564abe116fca96d3dc393cb181f6d098834
diff --git a/test/secure/publish-to-s3.js b/test/secure/publish-to-s3.js index <HASH>..<HASH> 100644 --- a/test/secure/publish-to-s3.js +++ b/test/secure/publish-to-s3.js @@ -3,15 +3,6 @@ var frauPublisher = require('../../src/publisher'), eventStream = require('event-stream'), gUtil = require('gulp-util'); -var publisher = frauPublisher.app({ - id: 'frau-publisher-test', - creds: { - key: '***REMOVED***', - secret: process.env.CREDS_SECRET - }, - devTag: Math.random().toString(16).slice(2) -}); - var content = 'some data'; var filename = 'test.txt'; var file = new gUtil.File({ @@ -26,6 +17,8 @@ var file = new gUtil.File({ describe('publisher', function() { it('should publish new file', function(cb) { + var publisher = createPublisher( Math.random().toString(16).slice(2) ); + eventStream.readArray( [file] ) .pipe( publisher.getStream() ) .on('end', function() { @@ -45,4 +38,31 @@ describe('publisher', function() { }, 500); }); }); + + it('should not overwrite a file', function(cb) { + // This test relies on files having already been published with this + // devTag. We could remove this dependency by publishing a file and + // then trying again. We should do this if necessary, but it didn't + // start that way because it seems unnecessarily wasteful. + var publisher = createPublisher( 'overwrite-test' ); + + eventStream.readArray( [file] ) + .pipe( publisher.getStream() ) + .on('error', function() { + cb(); + }).on('end', function() { + cb('should not have published'); + }); + }); }); + +function createPublisher(devTag) { + return frauPublisher.app({ + id: 'frau-publisher-test', + creds: { + key: '***REMOVED***', + secret: process.env.CREDS_SECRET + }, + devTag: devTag + }); +}
Add an end-to-end overwrite test.
Brightspace_frau-publisher
train
41177eb1c8b8327ad966b5e12555f9846d51d9de
diff --git a/code/libraries/koowa/libraries/controller/behavior/abstract.php b/code/libraries/koowa/libraries/controller/behavior/abstract.php index <HASH>..<HASH> 100644 --- a/code/libraries/koowa/libraries/controller/behavior/abstract.php +++ b/code/libraries/koowa/libraries/controller/behavior/abstract.php @@ -39,4 +39,25 @@ abstract class KControllerBehaviorAbstract extends KBehaviorAbstract return $methods; } + + /** + * Command handler + * + * @param KCommandInterface $command The command + * @param KCommandChainInterface $chain The chain executing the command + * @return mixed If a handler breaks, returns the break condition. Returns the result of the handler otherwise. + */ + public function execute(KCommandInterface $command, KCommandChainInterface $chain) + { + $parts = explode('.', $command->getName()); + $method = '_'.$parts[0].ucfirst($parts[1]); + + if($parts[0] == 'action') { + $result = $this->$method($command); + } else { + $result = parent::execute($command, $chain); + } + + return $result; + } }
re #<I> : Added specialised rule for handling 'action' methods in controller behaviors.
timble_kodekit
train
a1a4b1b2beda0125e496704fff5c0d207db6ad94
diff --git a/src/file_config/handlers/_common.py b/src/file_config/handlers/_common.py index <HASH>..<HASH> 100644 --- a/src/file_config/handlers/_common.py +++ b/src/file_config/handlers/_common.py @@ -19,13 +19,13 @@ class BaseHandler(abc.ABC): :rtype: module """ - if not hasattr(self, "_imported"): + if not hasattr(self, "_imported") or self._imported is None: self._imported = self._discover_import() return self._imported @property def handler(self): - if not hasattr(self, "_handler"): + if not hasattr(self, "_handler") or self._handler is None: self._handler = sys.modules[self.imported] return self._handler @@ -72,7 +72,7 @@ class BaseHandler(abc.ABC): return True return False - def _discover_import(self): + def _discover_import(self, prefer=None): """ Discovers and imports the best available module from ``packages``. :raises ModuleNotFoundError: If no module is available @@ -80,7 +80,11 @@ class BaseHandler(abc.ABC): :rtype: str """ - for module_name in self.packages: + available_packages = self.packages + if isinstance(prefer, str): + available_packages = (prefer,) + + for module_name in available_packages: spec = importlib.util.find_spec(module_name) if spec is not None: importlib.import_module(module_name) @@ -88,18 +92,31 @@ class BaseHandler(abc.ABC): if callable(imported_hook): imported_hook(sys.modules[module_name]) return module_name - raise ModuleNotFoundError(f"no modules in {self.packages!r} found") + raise ModuleNotFoundError(f"no modules in {available_packages!r} found") def _prefer_package(self, package): - if package not in self.packages: - raise ValueError( - f"prefered package {package!r} does not exist, " - f"allowed are {self.packages!r}" - ) - # NOTE: this is a semi-dangerous property override that needs to be done since - # packages are dynamically defined as part of the subclass but the super needs - # to be able to override them no matter what - self.packages = (package,) + """ Prefer a serializtion handler over other handlers. + + :param str package: The name of the package to use + :raises ValueError: When the given package name is not one of the available + supported serializtion packages for this handler + :return: The name of the serialization handler + :rtype: str + """ + + if isinstance(package, str) and package != self.imported: + if package not in self.packages: + raise ValueError( + f"preferred package {package!r} does not exist, allowed are " + f"{self.packages!r}" + ) + # clear out current serialization handler (if exists) + if hasattr(self, "_handler"): + del self._handler + # manually update imported handlers with a given preference + self._imported = self._discover_import(prefer=package) + return package + return self.imported def dumps(self, config, instance, prefer=None, **kwargs): """ An abstract dumps method which dumps an instance into the subclasses format. @@ -112,10 +129,8 @@ class BaseHandler(abc.ABC): :rtype: str """ - if isinstance(prefer, str): - self._prefer_package(prefer) - - dumps_hook_name = f"on_{self.imported}_dumps" + dumper = self._prefer_package(prefer) + dumps_hook_name = f"on_{dumper}_dumps" dumps_hook = getattr(self, dumps_hook_name, None) if not callable(dumps_hook): raise ValueError( @@ -144,10 +159,8 @@ class BaseHandler(abc.ABC): :rtype: dict """ - if isinstance(prefer, str): - self._prefer_package(prefer) - - loads_hook_name = f"on_{self.imported}_loads" + loader = self._prefer_package(prefer) + loads_hook_name = f"on_{loader}_loads" loads_hook = getattr(self, loads_hook_name, None) if not callable(loads_hook): raise ValueError( diff --git a/src/file_config/handlers/json.py b/src/file_config/handlers/json.py index <HASH>..<HASH> 100644 --- a/src/file_config/handlers/json.py +++ b/src/file_config/handlers/json.py @@ -9,7 +9,7 @@ class JSONHandler(BaseHandler): """ name = "json" - packages = ("rapidjson", "ujson", "json") + packages = ("rapidjson", "json", "ujson") options = {"indent": None, "sort_keys": False} def on_json_dumps(self, json, config, dictionary, **kwargs): diff --git a/tests/handlers/test_yaml.py b/tests/handlers/test_yaml.py index <HASH>..<HASH> 100644 --- a/tests/handlers/test_yaml.py +++ b/tests/handlers/test_yaml.py @@ -30,7 +30,6 @@ def test_yaml_ordereddict(): content = instance.dumps_yaml() assert content == dedent( """\ - foo: - test: test + foo: {test: test} """ )
Fixing prefer keyword handler package management
stephen-bunn_file-config
train
98797911fe3b860c000a38e0c042e0e1881ab7ae
diff --git a/worker/deployer/nested_test.go b/worker/deployer/nested_test.go index <HASH>..<HASH> 100644 --- a/worker/deployer/nested_test.go +++ b/worker/deployer/nested_test.go @@ -366,8 +366,12 @@ func (s *NestedContextSuite) deployThreeUnits(c *gc.C, ctx deployer.Context) { for { units := report["units"].(map[string]interface{}) workers := units["workers"].(map[string]interface{}) + + first := workers["first/0"].(map[string]interface{}) + second := workers["second/0"].(map[string]interface{}) third := workers["third/0"].(map[string]interface{}) - if third["state"] == "started" { + + if first["state"] == "started" && second["state"] == "started" && third["state"] == "started" { break } select {
Fixes nested deployer report test to ensure that all workers are reported as starting, not just the third, which does not guarantee that the others are up. This can fail on slow machines and has been observed on arm<I> test jobs.
juju_juju
train
389468f234c6a82dd5bfffae9f1c8daeee058aa0
diff --git a/media/src/main/java/com/microsoft/windowsazure/services/media/implementation/VersionHeadersFilter.java b/media/src/main/java/com/microsoft/windowsazure/services/media/implementation/VersionHeadersFilter.java index <HASH>..<HASH> 100644 --- a/media/src/main/java/com/microsoft/windowsazure/services/media/implementation/VersionHeadersFilter.java +++ b/media/src/main/java/com/microsoft/windowsazure/services/media/implementation/VersionHeadersFilter.java @@ -42,7 +42,7 @@ public class VersionHeadersFilter extends IdempotentClientFilter MultivaluedMap<String, Object> headers = cr.getHeaders(); headers.add("DataServiceVersion", "3.0"); headers.add("MaxDataServiceVersion", "3.0"); - headers.add("x-ms-version", "2.2"); + headers.add("x-ms-version", "2.5"); return getNext().handle(cr); } } diff --git a/media/src/test/java/com/microsoft/windowsazure/services/media/IntegrationTestBase.java b/media/src/test/java/com/microsoft/windowsazure/services/media/IntegrationTestBase.java index <HASH>..<HASH> 100644 --- a/media/src/test/java/com/microsoft/windowsazure/services/media/IntegrationTestBase.java +++ b/media/src/test/java/com/microsoft/windowsazure/services/media/IntegrationTestBase.java @@ -49,6 +49,8 @@ import com.microsoft.windowsazure.services.media.models.ListResult; import com.microsoft.windowsazure.services.media.models.Locator; import com.microsoft.windowsazure.services.media.models.LocatorInfo; import com.microsoft.windowsazure.services.media.models.LocatorType; +import com.microsoft.windowsazure.services.media.models.MediaProcessorInfo; +import com.microsoft.windowsazure.services.media.models.MediaProcessor; import com.microsoft.windowsazure.services.queue.QueueConfiguration; import com.microsoft.windowsazure.services.queue.QueueContract; import com.microsoft.windowsazure.services.queue.QueueService; @@ -72,9 +74,8 @@ public abstract class IntegrationTestBase protected static final String validButNonexistAssetId = "nb:cid:UUID:0239f11f-2d36-4e5f-aa35-44d58ccc0973"; protected static final String validButNonexistAccessPolicyId = "nb:pid:UUID:38dcb3a0-ef64-4ad0-bbb5-67a14c6df2f7"; protected static final String validButNonexistLocatorId = "nb:lid:UUID:92a70402-fca9-4aa3-80d7-d4de3792a27a"; - - protected static final String MEDIA_ENCODER_MEDIA_PROCESSOR_2_2_0_0_ID = "nb:mpid:UUID:70bdc2c3-ebf4-42a9-8542-5afc1e55d217"; - + + protected static String MEDIA_ENCODER_MEDIA_PROCESSOR_2_2_0_0_ID = "nb:mpid:UUID:2e7aa8f3-4961-4e0c-b4db-0e0439e524f5"; protected static final String invalidId = "notAValidId"; @Rule @@ -101,6 +102,26 @@ public abstract class IntegrationTestBase cleanupEnvironment(); } + + protected static void getLatestMediaProcessorID() throws Exception { + ListResult<MediaProcessorInfo> listMediaProcessorsResult = service.list(MediaProcessor.list()); + List<MediaProcessorInfo> ps = listMediaProcessorsResult; + double latestVersion = 0.0; + int position = -1; + for (int i = 0; i < ps.size(); i++) { + MediaProcessorInfo mediaProcessorInfo = ps.get(i); + double d = Double.parseDouble(mediaProcessorInfo.getVersion()); + if (d > latestVersion) { + latestVersion = d; + position = i; + } + } + + if (position != -1) + { + MEDIA_ENCODER_MEDIA_PROCESSOR_2_2_0_0_ID = ps.get(position).getId(); + } + } protected static void overrideWithEnv(Configuration config, String key) {
replace hardcoded media processor id with dynamic latest one and upgrade sdk version from <I> to <I>, as media service will deprecate the <I> version
Azure_azure-sdk-for-java
train
3d05aa85ea26ef8206a1da053f62e636aba7384a
diff --git a/activity_stream/templatetags/activity_stream_tags.py b/activity_stream/templatetags/activity_stream_tags.py index <HASH>..<HASH> 100644 --- a/activity_stream/templatetags/activity_stream_tags.py +++ b/activity_stream/templatetags/activity_stream_tags.py @@ -31,9 +31,13 @@ def users_activity_stream(context, user, count, offset=0): offset=0 activity_items = ActivityStreamItem.objects.filter(actor=user, subjects__isnull=False, - created_at__lte=datetime.datetime.now()).order_by('-created_at').distinct()[offset:count] - return {"activity_items": activity_items, "user": context["user"], - "request":context.get("request")} + created_at__lte=datetime.datetime.now())\ + .order_by('-created_at').distinct()[offset:count] + + return {"activity_items": activity_items, + "user": context["user"], + "request":context.get("request") + } @register.inclusion_tag("activity_stream/friends_activity_stream.html", takes_context=True) @@ -48,7 +52,10 @@ def following_activity_stream(context, user, count, offset=0): following = get_people_i_follow(user, 1000) following = list(following) following.append(user) - activity_items = ActivityStreamItem.objects.filter(actor__in=following, subjects__isnull=False, created_at__lte=datetime.datetime.now()).order_by('-created_at').distinct()[offset:count] + activity_items = ActivityStreamItem.objects.filter(actor__in=following, \ + subjects__isnull=False, \ + created_at__lte=datetime.datetime.now())\ + .order_by('-created_at').distinct()[offset:count] return {"activity_items": activity_items, "user": context["user"], "request":context.get("request")} @@ -63,7 +70,8 @@ def global_activity_stream(context, count, offset=0, privacylevel=0): offset=0 activity_items = ActivityStreamItem.objects.filter(subjects__isnull=False, - created_at__lte=datetime.datetime.now()).order_by('-created_at').distinct()[offset:count] + created_at__lte=datetime.datetime.now())\ + .order_by('-created_at').distinct()[offset:count] return {"activity_items": activity_items, "user": context["user"], "request":context.get("request")} @@ -100,7 +108,7 @@ class IsFollowingNode(Node): from_user = self.from_user.resolve(context) if to_user and from_user: if to_user.is_authenticated() and from_user.is_authenticated(): - is_following = ActivityFollower.objects.filter(to_user=to_user, from_user=from_user).count() + is_following = ActivityFollower.objects.filter(to_user=to_user,from_user=from_user).count() if is_following: return self.node_true.render(context) else: diff --git a/activity_stream/tests.py b/activity_stream/tests.py index <HASH>..<HASH> 100644 --- a/activity_stream/tests.py +++ b/activity_stream/tests.py @@ -28,12 +28,12 @@ class StoryTest(TestCase): self.assertTrue(TestSubject.objects.get(pk=photo.id)) activityItem2 = create_activity_item("placed", User.objects.get(username="admin"), photo2) - items = users_activity_stream(User.objects.get(username="admin"),1000) + items = users_activity_stream({}, User.objects.get(username="admin"),1000) self.assertEquals(len(items['activity_items']), 1) photo2.delete() - items = users_activity_stream(User.objects.get(username="admin"),1000) + items = users_activity_stream({}, User.objects.get(username="admin"),1000) self.assertEquals(len(items['activity_items']), 0) def test_batching(self): @@ -66,7 +66,7 @@ class StoryTest(TestCase): self.assertTrue(activityItem) self.assertEquals(activityItem.is_batched, False) self.assertEquals(activityItem.subjects.count(), 1) - items = users_activity_stream(User.objects.get(username="admin"),1000) + items = users_activity_stream({}, User.objects.get(username="admin"),1000) self.assertEquals(len(items['activity_items']), 0) activityItem.delete()
manually merging changes to fix tests from daonb, making it a bit more pep8
philippWassibauer_django-activity-stream
train
191608a3b219a63eca8a6e9b85d556101d2b35ef
diff --git a/lib/claide/command.rb b/lib/claide/command.rb index <HASH>..<HASH> 100644 --- a/lib/claide/command.rb +++ b/lib/claide/command.rb @@ -80,7 +80,7 @@ module CLAide # attr_accessor :description - # @return [Array<String>] The prefixes used t osearch for CLAide plugins. + # @return [Array<String>] The prefixes used to search for CLAide plugins. # Plugins are loaded via their `<plugin_prefix>_plugin.rb` file. # Defaults to search for `claide` plugins. # @@ -233,7 +233,7 @@ module CLAide # # The subclass has to combine the result of calling `super` and its own # list of options. The recommended way of doing this is by concatenating - # concatenating to this classes’ own options. + # to this classes’ own options. # # @return [Array<Array>] #
two typos I noticed today while reading the code
CocoaPods_CLAide
train
2e47674a466c6e75df2a8a649a5baae24b1b4e4e
diff --git a/tests/func/test_install.py b/tests/func/test_install.py index <HASH>..<HASH> 100644 --- a/tests/func/test_install.py +++ b/tests/func/test_install.py @@ -3,6 +3,7 @@ import pathlib import sys import pytest +from git import GitCommandError from dvc.exceptions import GitHookAlreadyExistsError from dvc.utils import file_md5 @@ -35,18 +36,25 @@ class TestInstall: with pytest.raises(GitHookAlreadyExistsError): scm.install() - def test_post_checkout(self, tmp_dir, scm, dvc): + def test_pre_commit_hook(self, tmp_dir, scm, dvc, caplog): + tmp_dir.dvc_gen("file", "file content", commit="create foo") + tmp_dir.gen("file", "file modified") scm.install() - tmp_dir.dvc_gen({"file": "file content"}, commit="add") + # scm.commit bypasses hooks + with pytest.raises(GitCommandError, match=r"modified:\s*file"): + scm.repo.git.commit(m="file modified") + + def test_post_checkout(self, tmp_dir, scm, dvc): + tmp_dir.dvc_gen({"file": "file content"}, commit="add") os.unlink("file") + scm.install() + scm.checkout("new_branch", create_new=True) assert os.path.isfile("file") def test_pre_push_hook(self, tmp_dir, scm, dvc, tmp_path_factory): - scm.install() - temp = tmp_path_factory.mktemp("external") git_remote = temp / "project.git" storage_path = temp / "dvc_storage" @@ -64,6 +72,8 @@ class TestInstall: scm.repo.clone(os.fspath(git_remote)) scm.repo.create_remote("origin", os.fspath(git_remote)) + scm.install() + assert not expected_storage_path.is_file() scm.repo.git.push("origin", "master") assert expected_storage_path.is_file() @@ -74,16 +84,19 @@ class TestInstall: sys.platform == "win32", reason="Git hooks aren't supported on Windows" ) def test_merge_driver_no_ancestor(tmp_dir, scm, dvc): - scm.commit("init") - scm.install() - (tmp_dir / ".gitattributes").write_text("*.dvc merge=dvc") - scm.checkout("one", create_new=True) - tmp_dir.dvc_gen({"data": {"foo": "foo"}}, commit="one: add data") + with tmp_dir.branch("one", new=True): + tmp_dir.dvc_gen({"data": {"foo": "foo"}}, commit="one: add data") - scm.checkout("master") scm.checkout("two", create_new=True) + dvc.checkout() # keep things in sync + tmp_dir.dvc_gen({"data": {"bar": "bar"}}, commit="two: add data") + # installing hook only before merge, as it runs `dvc` commands which makes + # `checkouts` and `commits` above slower + scm.install() + (tmp_dir / ".gitattributes").write_text("*.dvc merge=dvc") + scm.repo.git.merge("one", m="merged", no_gpg_sign=True, no_signoff=True) # NOTE: dvc shouldn't checkout automatically as it might take a long time @@ -102,18 +115,21 @@ def test_merge_driver_no_ancestor(tmp_dir, scm, dvc): sys.platform == "win32", reason="Git hooks aren't supported on Windows" ) def test_merge_driver(tmp_dir, scm, dvc): - scm.commit("init") - scm.install() - (tmp_dir / ".gitattributes").write_text("*.dvc merge=dvc") tmp_dir.dvc_gen({"data": {"master": "master"}}, commit="master: add data") - scm.checkout("one", create_new=True) - tmp_dir.dvc_gen({"data": {"one": "one"}}, commit="one: add data") + with tmp_dir.branch("one", new=True): + tmp_dir.dvc_gen({"data": {"one": "one"}}, commit="one: add data") - scm.checkout("master") scm.checkout("two", create_new=True) + dvc.checkout() # keep things in sync + tmp_dir.dvc_gen({"data": {"two": "two"}}, commit="two: add data") + # installing hook only before merge, as it runs `dvc` commands on + # `checkouts` and `commits` which slows tests down + scm.install() + (tmp_dir / ".gitattributes").write_text("*.dvc merge=dvc") + scm.repo.git.merge("one", m="merged", no_gpg_sign=True, no_signoff=True) # NOTE: dvc shouldn't checkout automatically as it might take a long time
tests: install hook just before using it (#<I>) Before, we were installing hooks at the start which was making checkouts slower. With this change, it should improve these tests by twice as much.
iterative_dvc
train
fb5eb1fa7787815144828fee0f0938c7273e1e51
diff --git a/executor/analyze_test.go b/executor/analyze_test.go index <HASH>..<HASH> 100644 --- a/executor/analyze_test.go +++ b/executor/analyze_test.go @@ -371,7 +371,7 @@ func (s *testFastAnalyze) TestAnalyzeFastSample(c *C) { samples := mockExec.Collectors[i].Samples c.Assert(len(samples), Equals, 20) for j := 1; j < 20; j++ { - cmp, err := samples[j].Value.CompareDatum(tk.Se.GetSessionVars().StmtCtx, &samples[j-1].Value) + cmp, err := samples[j].Value.Compare(tk.Se.GetSessionVars().StmtCtx, &samples[j-1].Value, collate.GetBinaryCollator()) c.Assert(err, IsNil) c.Assert(cmp, Greater, 0) } @@ -381,7 +381,7 @@ func (s *testFastAnalyze) TestAnalyzeFastSample(c *C) { func checkHistogram(sc *stmtctx.StatementContext, hg *statistics.Histogram) (bool, error) { for i := 0; i < len(hg.Buckets); i++ { lower, upper := hg.GetLower(i), hg.GetUpper(i) - cmp, err := upper.CompareDatum(sc, lower) + cmp, err := upper.Compare(sc, lower, collate.GetBinaryCollator()) if cmp < 0 || err != nil { return false, err } @@ -389,7 +389,7 @@ func checkHistogram(sc *stmtctx.StatementContext, hg *statistics.Histogram) (boo continue } previousUpper := hg.GetUpper(i - 1) - cmp, err = lower.CompareDatum(sc, previousUpper) + cmp, err = lower.Compare(sc, previousUpper, collate.GetBinaryCollator()) if cmp <= 0 || err != nil { return false, err } diff --git a/statistics/builder.go b/statistics/builder.go index <HASH>..<HASH> 100644 --- a/statistics/builder.go +++ b/statistics/builder.go @@ -23,6 +23,7 @@ import ( "github.com/pingcap/tidb/sessionctx/stmtctx" "github.com/pingcap/tidb/types" "github.com/pingcap/tidb/util/codec" + "github.com/pingcap/tidb/util/collate" ) // SortedBuilder is used to build histograms for PK and index. @@ -67,7 +68,7 @@ func (b *SortedBuilder) Iterate(data types.Datum) error { b.hist.NDV = 1 return nil } - cmp, err := b.hist.GetUpper(int(b.bucketIdx)).CompareDatum(b.sc, &data) + cmp, err := b.hist.GetUpper(int(b.bucketIdx)).Compare(b.sc, &data, collate.GetBinaryCollator()) if err != nil { return errors.Trace(err) } @@ -158,7 +159,7 @@ func buildHist(sc *stmtctx.StatementContext, hg *Histogram, samples []*SampleIte hg.AppendBucket(&samples[0].Value, &samples[0].Value, int64(sampleFactor), int64(ndvFactor)) for i := int64(1); i < sampleNum; i++ { corrXYSum += float64(i) * float64(samples[i].Ordinal) - cmp, err := hg.GetUpper(bucketIdx).CompareDatum(sc, &samples[i].Value) + cmp, err := hg.GetUpper(bucketIdx).Compare(sc, &samples[i].Value, collate.GetBinaryCollator()) if err != nil { return 0, errors.Trace(err) } diff --git a/util/collate/collate.go b/util/collate/collate.go index <HASH>..<HASH> 100644 --- a/util/collate/collate.go +++ b/util/collate/collate.go @@ -154,6 +154,11 @@ func GetCollator(collate string) Collator { return binCollatorInstance } +// GetBinaryCollator gets the binary collator, it is often used when we want to apply binary compare. +func GetBinaryCollator() Collator { + return binCollatorInstance +} + // GetCollatorByID get the collator according to id, it will return the binary collator if the corresponding collator doesn't exist. func GetCollatorByID(id int) Collator { if atomic.LoadInt32(&newCollationEnabled) == 1 {
*: replace compareDatum by compare (#<I>)
pingcap_tidb
train
d24641e0607b3f3839d10394784258a54c6f3384
diff --git a/lib/veewee/provider/kvm/box/helper/status.rb b/lib/veewee/provider/kvm/box/helper/status.rb index <HASH>..<HASH> 100644 --- a/lib/veewee/provider/kvm/box/helper/status.rb +++ b/lib/veewee/provider/kvm/box/helper/status.rb @@ -15,11 +15,11 @@ module Veewee end def exists_volume? - [email protected](:name => "#{name}.img").nil? + @connection.list_volumes.find { |v| v[:name] == "#{name}.img" } end def exists_vm? - [email protected](:name => name).nil? + @connection.list_domains.find { |d| d[:name] == name } end end # End Module
Properly determine if a volume or instance exists Preious to this commit, volumes were determined to exist or not by calling the Volumes#all method the :name parameter to filter only the volumes that matched the value of :name. This was not properly finding the volumes. This commit uses the #list_volumes method and calls #find on the array returned. Likewise for instances
jedi4ever_veewee
train
e5c19e5bddb2738fbab26a4238eface70081f8d5
diff --git a/src/TickGroup/TickGroup.js b/src/TickGroup/TickGroup.js index <HASH>..<HASH> 100644 --- a/src/TickGroup/TickGroup.js +++ b/src/TickGroup/TickGroup.js @@ -44,7 +44,7 @@ export default class TickGroup extends PureComponent { constructor(props) { super(props); - (this:any).removeTick = this.removeTick.bind(this); + (this:any).removeUDID = this.removeUDID.bind(this); } state = { @@ -71,13 +71,13 @@ export default class TickGroup extends PureComponent { const ticks = scale.ticks ? scale.ticks(tickCount) : []; this.setState((prevState) => { - const mapped = ticks.map((t) => ({ val: t })); + const mapped = ticks.map((tick) => ({ val: tick })); const update = dataUpdate({ data: mapped, ...next }, prevState, this.removed); return { ...update, prevScale: prev.scale, currScale: scale }; }); } - removeTick(udid) { + removeUDID(udid) { this.removed.set(udid, true); } @@ -92,8 +92,8 @@ export default class TickGroup extends PureComponent { return ( <g className={className}> - {state.nodes.map((tick) => { - const udid = keyAccessor(tick); + {state.nodes.map((node) => { + const udid = keyAccessor(node); const type = state.udids[udid]; return ( @@ -101,10 +101,10 @@ export default class TickGroup extends PureComponent { key={udid} udid={udid} type={type} - tick={tick} - removeTick={this.removeTick} + node={node} prevScale={state.prevScale} currScale={state.currScale} + removeUDID={this.removeUDID} {...props} /> );
modify TickGroup for using withTransitions
sghall_resonance
train
ef0617af7d3558e99dcd5065633326cc54d28670
diff --git a/src/Picqer/Financials/Moneybird/Entities/User.php b/src/Picqer/Financials/Moneybird/Entities/User.php index <HASH>..<HASH> 100644 --- a/src/Picqer/Financials/Moneybird/Entities/User.php +++ b/src/Picqer/Financials/Moneybird/Entities/User.php @@ -25,13 +25,6 @@ class User extends Model 'language', 'time_zone', 'permissions', - 'sales_invoices', - 'documents', - 'estimates', - 'bank', - 'settings', - 'ownership', - 'time_entries', ]; /**
Bugfix: invalid fillable entries This commit fixes a minor bug where the possible permission entries were accidentally added as fillable properties to the user object.
picqer_moneybird-php-client
train
2fdb2ceeb890ad524df5e52db3dca101e9514441
diff --git a/tests/unit/modules/test_file.py b/tests/unit/modules/test_file.py index <HASH>..<HASH> 100644 --- a/tests/unit/modules/test_file.py +++ b/tests/unit/modules/test_file.py @@ -1,5 +1,3 @@ -# Import python libs - import os import shutil import tempfile @@ -15,15 +13,11 @@ import salt.utils.files import salt.utils.platform import salt.utils.stringutils from salt.exceptions import CommandExecutionError, SaltInvocationError - -# Import Salt libs from salt.ext import six from salt.utils.jinja import SaltCacheLoader from tests.support.helpers import with_tempfile from tests.support.mixins import LoaderModuleMockMixin from tests.support.mock import DEFAULT, MagicMock, Mock, mock_open, patch - -# Import Salt Testing libs from tests.support.runtests import RUNTIME_VARS from tests.support.unit import TestCase, skipIf
run pre-commit: remove import comments
saltstack_salt
train
f944c9c2d534185a0389027b04fd96986ba82e7c
diff --git a/ford/sourceform.py b/ford/sourceform.py index <HASH>..<HASH> 100644 --- a/ford/sourceform.py +++ b/ford/sourceform.py @@ -2914,7 +2914,7 @@ class FortranBlockData(FortranContainer): def process_attribs(self): for item in self.types: - if item.name.lower() in self.attr_dict: + for attr in self.attr_dict.get(item.name.lower(), []): if "public" in self.attr_dict[item.name.lower()]: item.permission = "public" elif "private" in self.attr_dict[item.name.lower()]:
Fix for undefined variable in `FortranBlockData`
Fortran-FOSS-Programmers_ford
train
73b5553316f8b80a5833d45518c54191ea293328
diff --git a/runtimes/assemblyscript/cli-runner.js b/runtimes/assemblyscript/cli-runner.js index <HASH>..<HASH> 100644 --- a/runtimes/assemblyscript/cli-runner.js +++ b/runtimes/assemblyscript/cli-runner.js @@ -14,18 +14,17 @@ async function main(runtime, module, input, expected) { let exp = JSON.parse(fs.readFileSync(expected)); console.log("expected:"); console.log(JSON.stringify(exp, null, 2)); + let res; try { - let res = await engine.invoke(rt, mod, "qcert_main", arg); - console.log("output:"); - console.log(JSON.stringify(res, null, 2)); + res = await engine.invoke(rt, mod, "qcert_main", arg); } catch(err) { - console.log("output:"); let res = { "error": "Eval failed", "message": err.message }; - console.log(JSON.stringify(res, null,2)); } + console.log("output:"); + console.log(JSON.stringify(res, null, 2)); } const rt = process.argv[2];
fix(wasm) more on cli-runner error
querycert_qcert
train
3956f9cb9a80337c42269709f179539345b7d580
diff --git a/joomla/filesystem/stream.php b/joomla/filesystem/stream.php index <HASH>..<HASH> 100644 --- a/joomla/filesystem/stream.php +++ b/joomla/filesystem/stream.php @@ -123,7 +123,7 @@ class JStream extends JObject * @return boolean * @since 1.6 */ - function open($filename, $mode='r', $use_include_path=false, $context=null, $use_prefix=true, $relative=false, $detectprocessingmode=false) + function open($filename, $mode='r', $use_include_path=false, $context=null, $use_prefix=false, $relative=false, $detectprocessingmode=false) { $filename = $this->_getFilename($filename, $mode, $use_prefix, $relative);
# [#<I>] Temp folder has to include JPATH_ROOT for gz extension upload to work. Thanks Sam. --HG-- extra : convert_revision : svn%3A6f6e1ebd-4c2b-<I>-<I>f-f<I>bde<I>bce9/development/trunk/libraries%<I>
joomla_joomla-framework
train
8f5a29a567f4425d0349fa6a8bd869062d799a45
diff --git a/rqalpha/utils/log_capture.py b/rqalpha/utils/log_capture.py index <HASH>..<HASH> 100644 --- a/rqalpha/utils/log_capture.py +++ b/rqalpha/utils/log_capture.py @@ -35,6 +35,7 @@ class LogCapture: def __enter__(self): self._handlers = self._logger.handlers self._logger.handlers = [self._capture_handler] + return self def __exit__(self, exc_type, exc_val, exc_tb): self._logger.handlers = self._handlers
bugfix: work as the LogCapture Context Manager AttributeError: 'NoneType' object has no attribute 'replay' log_capture.replay()
ricequant_rqalpha
train
2c3747740ad297d8da55d8c1296511c3776fce86
diff --git a/src/Behat/Behat/EventDispatcher/Event/AfterStepTested.php b/src/Behat/Behat/EventDispatcher/Event/AfterStepTested.php index <HASH>..<HASH> 100644 --- a/src/Behat/Behat/EventDispatcher/Event/AfterStepTested.php +++ b/src/Behat/Behat/EventDispatcher/Event/AfterStepTested.php @@ -115,7 +115,7 @@ final class AfterStepTested extends StepTested implements AfterTested */ public function hasOutput() { - return $this->isTeardownHasOutput() || $this->isResultHasException() || $this->isResultCallHasOutput(); + return $this->teardownHasOutput() || $this->resultHasException() || $this->resultCallHasOutput(); } /** @@ -123,7 +123,7 @@ final class AfterStepTested extends StepTested implements AfterTested * * @return Boolean */ - private function isTeardownHasOutput() + private function teardownHasOutput() { return $this->teardown->hasOutput(); } @@ -133,7 +133,7 @@ final class AfterStepTested extends StepTested implements AfterTested * * @return Boolean */ - private function isResultHasException() + private function resultHasException() { return $this->result instanceof ExceptionResult && $this->result->getException(); } @@ -143,7 +143,7 @@ final class AfterStepTested extends StepTested implements AfterTested * * @return Boolean */ - private function isResultCallHasOutput() + private function resultCallHasOutput() { if (!$this->result instanceof ExecutedStepResult) { return false;
Clarify AfterStepTested method names
Behat_Behat
train
0e0ae6f89819d59174fc9e5565c090ee1b8c330b
diff --git a/lib/cucumber/runtime.rb b/lib/cucumber/runtime.rb index <HASH>..<HASH> 100644 --- a/lib/cucumber/runtime.rb +++ b/lib/cucumber/runtime.rb @@ -46,7 +46,7 @@ module Cucumber def initialize(configuration = Configuration.default) @configuration = Configuration.new(configuration) - @support_code = SupportCode.new(self, @configuration) + @support_code = SupportCode::CachesStepMatch.new SupportCode.new(self, @configuration) @results = Formatter::LegacyApi::Results.new end diff --git a/lib/cucumber/runtime/support_code.rb b/lib/cucumber/runtime/support_code.rb index <HASH>..<HASH> 100644 --- a/lib/cucumber/runtime/support_code.rb +++ b/lib/cucumber/runtime/support_code.rb @@ -152,18 +152,7 @@ module Cucumber end end - def step_match(step_name, name_to_report=nil) #:nodoc: - @match_cache ||= {} - - match = @match_cache[[step_name, name_to_report]] - return match if match - - @match_cache[[step_name, name_to_report]] = step_match_without_cache(step_name, name_to_report) - end - - private - - def step_match_without_cache(step_name, name_to_report=nil) + def step_match(step_name, name_to_report=nil) matches = matches(step_name, name_to_report) raise Undefined.new(step_name) if matches.empty? matches = best_matches(step_name, matches) if matches.size > 1 && guess_step_matches? @@ -171,6 +160,8 @@ module Cucumber matches[0] end + private + def guess_step_matches? @configuration.guess? end @@ -204,6 +195,17 @@ module Cucumber Cucumber.logger end + require 'delegate' + class CachesStepMatch < SimpleDelegator + def step_match(step_name, name_to_report=nil) #:nodoc: + @match_cache ||= {} + + match = @match_cache[[step_name, name_to_report]] + return match if match + + @match_cache[[step_name, name_to_report]] = super(step_name, name_to_report) + end + end end end end diff --git a/lib/cucumber/step_definitions.rb b/lib/cucumber/step_definitions.rb index <HASH>..<HASH> 100644 --- a/lib/cucumber/step_definitions.rb +++ b/lib/cucumber/step_definitions.rb @@ -2,7 +2,7 @@ module Cucumber class StepDefinitions def initialize(configuration = Configuration.default) configuration = Configuration.new(configuration) - @support_code = Runtime::SupportCode.new(nil, configuration) + @support_code = Runtime::SupportCode::CachesStepMatch.new Runtime::SupportCode.new(nil, configuration) @support_code.load_files_from_paths(configuration.autoload_code_paths) end diff --git a/spec/cucumber/runtime/support_code_spec.rb b/spec/cucumber/runtime/support_code_spec.rb index <HASH>..<HASH> 100644 --- a/spec/cucumber/runtime/support_code_spec.rb +++ b/spec/cucumber/runtime/support_code_spec.rb @@ -19,15 +19,19 @@ module Cucumber expect(format).to eq "it [snows] in [april]" end - it "caches step match results" do - dsl.Given(/it (.*) in (.*)/) { |what, month| } + context 'caching' do + let(:cached) { Runtime::SupportCode::CachesStepMatch.new(subject) } + + it "caches step match results" do + dsl.Given(/it (.*) in (.*)/) { |what, month| } - step_match = subject.step_match("it snows in april") + step_match = cached.step_match("it snows in april") - expect(@rb).not_to receive(:step_matches) - second_step_match = subject.step_match("it snows in april") + expect(@rb).not_to receive(:step_matches) + second_step_match = cached.step_match("it snows in april") - expect(step_match).to equal(second_step_match) + expect(step_match).to equal(second_step_match) + end end describe "resolving step defintion matches" do
Extract step_match caching to a decorator. I've left the creation of th cache explicit at construction for the moment. There may be a better api we can use, but wanted to wait to see if we can reuse it in plugins first.
cucumber_cucumber-ruby
train
f43437b75881e51c1d3f72214858357a520172ee
diff --git a/pmxbot/commands.py b/pmxbot/commands.py index <HASH>..<HASH> 100644 --- a/pmxbot/commands.py +++ b/pmxbot/commands.py @@ -806,14 +806,14 @@ def help(client, event, channel, nick, rest): yield "command not found" else: def mk_entries(): - for handler in sorted(_handler_registry, - key=operator.attrgetter('name')): - if handler.type_ != 'command': - continue - res = "!%s" % handler.name + handlers = (handler for handler in _handler_registry + if type(handler) is pmxbot.core.CommandHandler) + handlers = sorted(handlers, key=operator.attrgetter('name')) + for handler in handlers: + res = "!" + handler.name if handler.aliases: - res += " (%s)" % ', '.join( - alias.name for alias in handler.aliases) + alias_names = (alias.name for alias in handler.aliases) + res += " (%s)" % ', '.join(alias_names) yield res o = io.StringIO(u" ".join(mk_entries())) more = o.read(160) diff --git a/pmxbot/core.py b/pmxbot/core.py index <HASH>..<HASH> 100644 --- a/pmxbot/core.py +++ b/pmxbot/core.py @@ -334,7 +334,7 @@ class CommandHandler(Handler): def match(self, message, channel): cmd, _, cmd_args = message.partition(' ') - return cmd.lower() == '!{name}'.format(vars(self)) + return cmd.lower() == '!{name}'.format(name = self.name) def process(self, message): cmd, _, cmd_args = message.partition(' ')
Fix help to no longer use type_ attribute
yougov_pmxbot
train
3a574fb8bc0966c99713fd16d8ebe4e755c9320c
diff --git a/lib/rack/tracker/google_analytics/google_analytics.rb b/lib/rack/tracker/google_analytics/google_analytics.rb index <HASH>..<HASH> 100644 --- a/lib/rack/tracker/google_analytics/google_analytics.rb +++ b/lib/rack/tracker/google_analytics/google_analytics.rb @@ -2,18 +2,16 @@ require 'ostruct' class Rack::Tracker::GoogleAnalytics < Rack::Tracker::Handler class Event < OpenStruct + def write + ['send', event].to_json.gsub(/\[|\]/, '') + end + def event - { - hitType: 'event', - eventCategory: self.event_category, - eventAction: self.event_action, - eventLabel: self.event_label, - eventValue: self.event_value - }.compact + { hitType: 'event' }.merge(attributes).compact end - def write - ['send', event].to_json.gsub(/\[|\]/, '') + def attributes + Hash[to_h.map { |k,v| ['event' + k.to_s.capitalize, v] }] end end diff --git a/spec/handler/google_analytics_spec.rb b/spec/handler/google_analytics_spec.rb index <HASH>..<HASH> 100644 --- a/spec/handler/google_analytics_spec.rb +++ b/spec/handler/google_analytics_spec.rb @@ -9,7 +9,7 @@ RSpec.describe Rack::Tracker::GoogleAnalytics do def env {'tracker' => { 'google_analytics' => [ - Rack::Tracker::GoogleAnalytics::Event.new(event_category: "Users", event_action: "Login", event_label: "Standard") + Rack::Tracker::GoogleAnalytics::Event.new(category: "Users", action: "Login", label: "Standard") ] }} end @@ -23,7 +23,7 @@ RSpec.describe Rack::Tracker::GoogleAnalytics do describe "with a event value" do def env {'tracker' => { 'google_analytics' => [ - Rack::Tracker::GoogleAnalytics::Event.new(event_category: "Users", event_action: "Login", event_label: "Standard", event_value: 5) + Rack::Tracker::GoogleAnalytics::Event.new(category: "Users", action: "Login", label: "Standard", value: 5) ]}} end diff --git a/spec/tracker/controller_spec.rb b/spec/tracker/controller_spec.rb index <HASH>..<HASH> 100644 --- a/spec/tracker/controller_spec.rb +++ b/spec/tracker/controller_spec.rb @@ -9,7 +9,7 @@ class SomeController def index tracker do - google_analytics event_category: 'foo' + google_analytics category: 'foo' end end end @@ -17,7 +17,7 @@ end RSpec.describe Rack::Tracker::Controller do context 'controller' do - let(:event) { Rack::Tracker::GoogleAnalytics::Event.new(event_category: 'foo') } + let(:event) { Rack::Tracker::GoogleAnalytics::Event.new(category: 'foo') } it 'writes the event into env' do controller = SomeController.new
event attributes should be less verbose
railslove_rack-tracker
train
fcf29edf3db164ec0e5220efc82598126daef122
diff --git a/example.py b/example.py index <HASH>..<HASH> 100644 --- a/example.py +++ b/example.py @@ -89,7 +89,7 @@ ticket = None try: ticket = client.getObject() except Exception, e: - print "Unable to retrieve ticket record: %", e + print "Unable to retrieve ticket record: ", e # Now update the ticket. update = { @@ -98,6 +98,6 @@ update = { try: update = client.addUpdate(update) - print "Update ticket 123456. The new update's id is %.", update{'id'} + print "Update ticket 123456. The new update's id is %s." % update['id'] except Exception, e: - print "Unable to update ticket: %", e + print "Unable to update ticket: ", e
Fixed print usage in the example script. (Thanks, mahmoudimus!)
softlayer_softlayer-python
train
e92d2e359b060bed0f48a553e094e4805e44951a
diff --git a/plugins/worker/server/__init__.py b/plugins/worker/server/__init__.py index <HASH>..<HASH> 100644 --- a/plugins/worker/server/__init__.py +++ b/plugins/worker/server/__init__.py @@ -43,6 +43,18 @@ class CustomJobStatus(object): PUSHING_OUTPUT = 823 CANCELING = 824 + valid_transitions = { + FETCHING_INPUT: [JobStatus.RUNNING], + CONVERTING_INPUT: [JobStatus.RUNNING, FETCHING_INPUT], + CONVERTING_OUTPUT: [JobStatus.RUNNING], + PUSHING_OUTPUT: [JobStatus.RUNNING, CONVERTING_OUTPUT], + CANCELING: [JobStatus.INACTIVE, JobStatus.QUEUED, JobStatus.RUNNING], + JobStatus.ERROR: [FETCHING_INPUT, CONVERTING_INPUT, CONVERTING_OUTPUT, + PUSHING_OUTPUT, CANCELING, JobStatus.QUEUED, + JobStatus.RUNNING], + JobStatus.CANCELED: [CANCELING] + } + @classmethod def isValid(cls, status): return status in ( @@ -53,6 +65,10 @@ class CustomJobStatus(object): cls.CANCELING ) + @classmethod + def validTransition(cls, status): + return cls.valid_transitions.get(status) + def getCeleryApp(): """ @@ -145,7 +161,7 @@ def validateApiUrl(doc): def validateJobStatus(event): """Allow our custom job status values.""" - if CustomJobStatus.isValid(event.info): + if event.info['handler'] == 'worker_handler' and CustomJobStatus.isValid(event.info): event.preventDefault().addResponse(True)
Override job state transitions for the worker plugin
girder_girder
train
34a5fbb1a03e62e3232824fe9c62c91a8ee949a8
diff --git a/lib/manager/gomod/update.js b/lib/manager/gomod/update.js index <HASH>..<HASH> 100644 --- a/lib/manager/gomod/update.js +++ b/lib/manager/gomod/update.js @@ -6,12 +6,15 @@ module.exports = { function updateDependency(currentFileContent, upgrade) { try { + logger.debug(`gomod.updateDependency: ${upgrade.newValue}`); const { depName } = upgrade; - const depNameNoVersion = depName + let depNameNoVersion = depName .split('/') .slice(0, 3) .join('/'); - logger.debug(`gomod.updateDependency: ${upgrade.newValue}`); + if (depNameNoVersion.startsWith('gopkg.in')) { + depNameNoVersion = depNameNoVersion.replace(/\.v\d+$/, ''); + } const lines = currentFileContent.split('\n'); const lineToChange = lines[upgrade.lineNumber]; if (
fix(gomod): detect gopkg.in major bumps
renovatebot_renovate
train
46bdde4bc2db8324bc069af1aca0d1a9d7794558
diff --git a/library/exportsym.py b/library/exportsym.py index <HASH>..<HASH> 100644 --- a/library/exportsym.py +++ b/library/exportsym.py @@ -72,9 +72,9 @@ def createSignal(signal): return output def exportSym(db, filename): -""" -export canmatrix-object as .sym file (compatible to PEAK-Systems) -""" + """ + export canmatrix-object as .sym file (compatible to PEAK-Systems) + """ global enumDict global enums
Correct exportSym() docstring indentation
ebroecker_canmatrix
train
41989a277d8b331be106252eb29816f24229dde7
diff --git a/setup.py b/setup.py index <HASH>..<HASH> 100644 --- a/setup.py +++ b/setup.py @@ -9,17 +9,18 @@ setup( name = 'amqp_connection', version = '1.0', description = 'Python AMQP connection for worker', + license='MIT', keywords = [ 'AMQP', 'connection' ], - author='Marc-Antoine ARNAUD, Valentin NOEL', - author_email='[email protected]', + author = 'Marc-Antoine ARNAUD, Valentin NOEL', + author_email = '[email protected]', url = 'https://github.com/FTV-Subtil/py_amqp_connection', packages = [ 'amqp_connection' ], - package_dir={ + package_dir = { 'amqp_connection': os.path.join(dir_name, 'src') - } + }, )
Clean setup.py and add license description
FTV-Subtil_py_amqp_connection
train
5f77beba294af462867e9c0dacf2e0b623e824bd
diff --git a/csrf/README.md b/csrf/README.md index <HASH>..<HASH> 100644 --- a/csrf/README.md +++ b/csrf/README.md @@ -1,5 +1,5 @@ ## CSRF handler -Stateless protection against CSRF attacks for Go web applications. +Offers stateless protection against CSRF attacks for Go web applications. * Checks [Origin header]() was sent and matches the Host header. * Falls back to a URL-safe and secure HMAC token stored in a HTTP-only diff --git a/csrf/csrf.go b/csrf/csrf.go index <HASH>..<HASH> 100644 --- a/csrf/csrf.go +++ b/csrf/csrf.go @@ -18,10 +18,10 @@ import ( // handler is a private struct which contains the handler's configurable options. type handler struct { - name string - domain string - secret string - sessionID string + name string + domain string + secret string + userID string } // Name allows configuring the CSRF cookie name. @@ -38,10 +38,10 @@ func Secret(s string) option { } } -// SessionID allows to configure a random and unique user session identifier to generate the CSRF token. -func SessionID(s string) option { +// UserID allows to configure a random and unique user ID identifier used to generate the CSRF token. +func UserID(s string) option { return func(h *handler) { - h.sessionID = s + h.userID = s } } @@ -89,7 +89,7 @@ func Handler(h http.Handler, opts ...option) http.Handler { // Re-enables browser's XSS filter if it was disabled w.Header().Set("x-xss-protection", "1; mode=block") - if csrf.sessionID == "" { + if csrf.userID == "" { http.Error(w, errForbidden, http.StatusForbidden) return } @@ -101,7 +101,7 @@ func Handler(h http.Handler, opts ...option) http.Handler { case http.MethodDelete: case http.MethodPost: default: - setToken(w, csrf.name, csrf.secret, csrf.sessionID, csrf.domain) + setToken(w, csrf.name, csrf.secret, csrf.userID, csrf.domain) h.ServeHTTP(w, r) return } @@ -120,18 +120,18 @@ func Handler(h http.Handler, opts ...option) http.Handler { return } - if !xsrftoken.Valid(cookie.Value, csrf.secret, csrf.sessionID, "Global") { + if !xsrftoken.Valid(cookie.Value, csrf.secret, csrf.userID, "Global") { http.Error(w, errForbidden, http.StatusForbidden) return } - setToken(w, csrf.name, csrf.secret, csrf.sessionID, csrf.domain) + setToken(w, csrf.name, csrf.secret, csrf.userID, csrf.domain) h.ServeHTTP(w, r) }) } -func setToken(w http.ResponseWriter, name, secret, sessionID, domain string) { - token := xsrftoken.Generate(secret, sessionID, "Global") +func setToken(w http.ResponseWriter, name, secret, userID, domain string) { + token := xsrftoken.Generate(secret, userID, "Global") cookie := &http.Cookie{ Name: name, Value: token, diff --git a/csrf/csrf_test.go b/csrf/csrf_test.go index <HASH>..<HASH> 100644 --- a/csrf/csrf_test.go +++ b/csrf/csrf_test.go @@ -39,19 +39,19 @@ func TestDomainRequired(t *testing.T) { } }() - handler := Handler(requestHandler, SessionID("my session ID!"), Secret("my secret!")) + handler := Handler(requestHandler, UserID("my user ID!"), Secret("my secret!")) ts := httptest.NewServer(handler) defer ts.Close() } func TestCSRFProtection(t *testing.T) { - handler := Handler(requestHandler, SessionID("my session ID!"), Secret("my secret!"), Domain("localhost"), Name("_csrf")) + handler := Handler(requestHandler, UserID("my user ID!"), Secret("my secret!"), Domain("localhost"), Name("_csrf")) okTS := httptest.NewServer(handler) defer okTS.Close() cookie := &http.Cookie{ Name: "_csrf", - Value: xsrftoken.Generate("my secret!", "my session ID!", "Global"), + Value: xsrftoken.Generate("my secret!", "my user ID!", "Global"), Path: "/", Domain: "localhost", Expires: time.Now().Add(xsrftoken.Timeout), diff --git a/csrf/example_test.go b/csrf/example_test.go index <HASH>..<HASH> 100644 --- a/csrf/example_test.go +++ b/csrf/example_test.go @@ -19,7 +19,7 @@ func ExampleHandler() { fmt.Fprintf(w, "Welcome to the home page!") }) - rack := csrf.Handler(mux, csrf.SessionID("session ID"), csrf.Secret("my secret!"), csrf.Domain("localhost")) + rack := csrf.Handler(mux, csrf.UserID("user ID"), csrf.Secret("my secret!"), csrf.Domain("localhost")) http.ListenAndServe(":8080", rack) }
csrf: changes option name[C to more closely describe what it really is
c4milo_handlers
train
7d132077c757944f0bf19eadf8e776c5f1758f0d
diff --git a/demos/init.php b/demos/init.php index <HASH>..<HASH> 100644 --- a/demos/init.php +++ b/demos/init.php @@ -41,6 +41,7 @@ if (isset($layout->leftMenu)) { $form->addItem('Form Multi-column layout', ['form3']); $form->addItem(['Integration with Columns'], ['form5']); $form->addItem(['AutoComplete Field', 'icon'=>'yellow star'], ['autocomplete']); + $form->addItem(['DropDown Field'], ['dropdown-plus']); $form->addItem(['Value Selectors'], ['form6']); $form->addItem(['Conditional Fields'], ['jscondform']);
Link Dropdown demos in Menu dropdown-plus.php existed, but was not linked in demo menu
atk4_ui
train
11cb9a795dcf5a9286b910ffc24fe936cb48d98f
diff --git a/server/dashboards.go b/server/dashboards.go index <HASH>..<HASH> 100644 --- a/server/dashboards.go +++ b/server/dashboards.go @@ -52,9 +52,28 @@ func (s *Service) Dashboards(w http.ResponseWriter, r *http.Request) { // DashboardID returns a single specified dashboard func (s *Service) DashboardID(w http.ResponseWriter, r *http.Request) { + id, err := paramID("id", r) + if err != nil { + Error(w, http.StatusUnprocessableEntity, err.Error(), s.Logger) + return + } + ctx := r.Context() + e, err := s.DashboardsStore.Get(ctx, chronograf.DashboardID(id)) + if err != nil { + notFound(w, id, s.Logger) + return + } + + res := newDashboardResponse(e) + encodeJSON(w, http.StatusOK, res, s.Logger) } +// type postDashboardRequest struct { +// Data interface{} `json:"data"` // Serialization of config. +// Name string `json:"name,omitempty"` // Exploration name given by user. +// } + // NewDashboard creates and returns a new dashboard object func (s *Service) NewDashboard(w http.ResponseWriter, r *http.Request) { @@ -62,7 +81,24 @@ func (s *Service) NewDashboard(w http.ResponseWriter, r *http.Request) { // RemoveDashboard deletes a dashboard func (s *Service) RemoveDashboard(w http.ResponseWriter, r *http.Request) { - + id, err := paramID("id", r) + if err != nil { + Error(w, http.StatusUnprocessableEntity, err.Error(), h.Logger) + return + } + + ctx := r.Context() + e, err := s.DashboardsStore.Get(ctx, chronograf.DashboardID(id)) + if err != nil { + notFound(w, id, s.Logger) + return + } + + if err := s.DashboardsStore.Delete(ctx, &chronograf.Dashboard{ID: chronograf.DashboardID(id)}); err != nil { + unknownErrorWithMessage(w, err, s.Logger) + return + } + w.WriteHeader(http.StatusNoContent) } // UpdateDashboard updates a dashboard
add dashboard GET and DELETE handlers
influxdata_influxdb
train
5543e1fcc71fbc2c781da4587577e070a8711b24
diff --git a/storage/looputil/loop.go b/storage/looputil/loop.go index <HASH>..<HASH> 100644 --- a/storage/looputil/loop.go +++ b/storage/looputil/loop.go @@ -7,7 +7,7 @@ import ( "bytes" "os" "os/exec" - "path/filepath" + "path" "regexp" "strconv" "strings" @@ -64,7 +64,7 @@ func (m *loopDeviceManager) DetachLoopDevices(rootfs, prefix string) error { if !strings.HasPrefix(info.backingFile, prefix) { continue } - rootedBackingFile := filepath.Join(rootfs, info.backingFile) + rootedBackingFile := path.Join(rootfs, info.backingFile) st, err := m.stat(rootedBackingFile) if os.IsNotExist(err) { continue
storage/looputil: use path, not path/filepath A bit hackish, but then again so is testing loop device management on Windows.
juju_juju
train
111f44ffc450e2db8c30a923dfb664c309c76194
diff --git a/pymatgen/io/vasp/sets.py b/pymatgen/io/vasp/sets.py index <HASH>..<HASH> 100644 --- a/pymatgen/io/vasp/sets.py +++ b/pymatgen/io/vasp/sets.py @@ -214,6 +214,9 @@ class DictSet(VaspInputSet): of two ways, e.g. either {"LDAUU":{"O":{"Fe":5}}} to set LDAUU for Fe to 5 in an oxide, or {"LDAUU":{"Fe":5}} to set LDAUU to 5 regardless of the input structure. + + If a None value is given, that key is unset. For example, + {"ENCUT": None} will remove ENCUT from the incar settings. user_kpoints_settings (dict or Kpoints): Allow user to override kpoints setting by supplying a dict E.g., {"reciprocal_density": 1000}. User can also supply Kpoints object. Default is None. @@ -297,7 +300,14 @@ class DictSet(VaspInputSet): @property def incar(self): settings = dict(self._config_dict["INCAR"]) - settings.update(self.user_incar_settings) + for k, v in self.user_incar_settings.items(): + if v is None: + try: + del settings[k] + except KeyError: + settings[k] = v + else: + settings[k] = v structure = self.structure incar = Incar() comp = structure.composition diff --git a/pymatgen/io/vasp/tests/test_sets.py b/pymatgen/io/vasp/tests/test_sets.py index <HASH>..<HASH> 100644 --- a/pymatgen/io/vasp/tests/test_sets.py +++ b/pymatgen/io/vasp/tests/test_sets.py @@ -163,6 +163,9 @@ class MITMPRelaxSetTest(unittest.TestCase): ) self.assertEqual(userset.incar['MAGMOM'], [100, 0.6]) + noencutset = MPRelaxSet(struct, user_incar_settings={'ENCUT': None}) + self.assertNotIn("ENCUT", noencutset.incar) + # sulfide vs sulfate test coords = list()
Add None support for user_incar_settings.
materialsproject_pymatgen
train