commit
stringlengths
40
40
old_file
stringlengths
4
264
new_file
stringlengths
4
264
old_contents
stringlengths
0
3.26k
new_contents
stringlengths
1
4.43k
subject
stringlengths
15
624
message
stringlengths
15
4.7k
lang
stringclasses
3 values
license
stringclasses
13 values
repos
stringlengths
5
91.5k
1c78dfa0e0d1905910476b4052e42de287a70b74
runtests.py
runtests.py
#!/usr/bin/env python import os import sys import string def main(): """ Executes the tests. Requires the CherryPy live server to be installed. """ command = "python manage.py test" options = "--exe --with-selenium --with-selenium-fixtures --with-cherrypyliveserver" apps = [] if len(sys.argv) > 1: apps = sys.argv[1:] os.system(command + " " + string.join(apps, " ") + " " + options) if __name__ == "__main__": main()
#!/usr/bin/env python import os import sys import string def main(): """ Executes the tests. Requires the CherryPy live server to be installed. """ command = "python manage.py test" options = "--exe --with-selenium --with-selenium-fixtures --with-cherrypyliveserver --noinput" apps = [] if len(sys.argv) > 1: apps = sys.argv[1:] os.system(command + " " + string.join(apps, " ") + " " + options) if __name__ == "__main__": main()
Update to the run tests script to force database deletion if the test database exists.
Update to the run tests script to force database deletion if the test database exists.
Python
mit
jtakayama/makahiki-draft,jtakayama/ics691-setupbooster,csdl/makahiki,yongwen/makahiki,yongwen/makahiki,jtakayama/makahiki-draft,yongwen/makahiki,justinslee/Wai-Not-Makahiki,csdl/makahiki,jtakayama/makahiki-draft,yongwen/makahiki,csdl/makahiki,csdl/makahiki,jtakayama/makahiki-draft,jtakayama/ics691-setupbooster,jtakayama/ics691-setupbooster
20124d599c6305889315847c15329c02efdd2b8c
migrations/versions/0313_email_access_validated_at.py
migrations/versions/0313_email_access_validated_at.py
""" Revision ID: 0313_email_access_validated_at Revises: 0312_populate_returned_letters Create Date: 2020-01-28 18:03:22.237386 """ from alembic import op import sqlalchemy as sa revision = '0313_email_access_validated_at' down_revision = '0312_populate_returned_letters' def upgrade(): # ### commands auto generated by Alembic - please adjust! ### op.add_column('users', sa.Column('email_access_validated_at', sa.DateTime(), nullable=True)) # if user has email_auth, set email_access_validated_at on last login, else set it at user created_at date. op.execute(""" UPDATE users SET email_access_validated_at = created_at WHERE auth_type = 'sms_auth' """) op.execute(""" UPDATE users SET email_access_validated_at = logged_in_at WHERE auth_type = 'email_auth' """) op.alter_column('users', 'email_access_validated_at', nullable=False) # ### end Alembic commands ### def downgrade(): # ### commands auto generated by Alembic - please adjust! ### op.drop_column('users', 'email_access_validated_at') # ### end Alembic commands ###
""" Revision ID: 0313_email_access_validated_at Revises: 0312_populate_returned_letters Create Date: 2020-01-28 18:03:22.237386 """ from alembic import op import sqlalchemy as sa revision = '0313_email_access_validated_at' down_revision = '0312_populate_returned_letters' def upgrade(): # ### commands auto generated by Alembic - please adjust! ### op.add_column('users', sa.Column('email_access_validated_at', sa.DateTime(), nullable=True)) # if user has email_auth, set email_access_validated_at on last login, else set it at user created_at date. op.execute(""" UPDATE users SET email_access_validated_at = created_at """) op.execute(""" UPDATE users SET email_access_validated_at = logged_in_at WHERE auth_type = 'email_auth' AND email_access_validated_at IS NOT NULL """) op.alter_column('users', 'email_access_validated_at', nullable=False) # ### end Alembic commands ### def downgrade(): # ### commands auto generated by Alembic - please adjust! ### op.drop_column('users', 'email_access_validated_at') # ### end Alembic commands ###
Make sure email_access_validated_at is not null after being populated
Make sure email_access_validated_at is not null after being populated
Python
mit
alphagov/notifications-api,alphagov/notifications-api
d09fb55bd49e266901305b9126077f44f7a1301e
annoying/functions.py
annoying/functions.py
from django.shortcuts import _get_queryset from django.conf import settings def get_object_or_None(klass, *args, **kwargs): """ Uses get() to return an object or None if the object does not exist. klass may be a Model, Manager, or QuerySet object. All other passed arguments and keyword arguments are used in the get() query. Note: Like with get(), a MultipleObjectsReturned will be raised if more than one object is found. """ queryset = _get_queryset(klass) try: return queryset.get(*args, **kwargs) except queryset.model.DoesNotExist: return None def get_config(key, default): """ Get settings from django.conf if exists, return default value otherwise example: ADMIN_EMAIL = get_config('ADMIN_EMAIL', '[email protected]') """ return getattr(settings, key, default)
from django.shortcuts import _get_queryset from django.conf import settings def get_object_or_None(klass, *args, **kwargs): """ Uses get() to return an object or None if the object does not exist. klass may be a Model, Manager, or QuerySet object. All other passed arguments and keyword arguments are used in the get() query. Note: Like with get(), a MultipleObjectsReturned will be raised if more than one object is found. """ queryset = _get_queryset(klass) try: return queryset.get(*args, **kwargs) except queryset.model.DoesNotExist: return None def get_config(key, default=None): """ Get settings from django.conf if exists, return default value otherwise example: ADMIN_EMAIL = get_config('ADMIN_EMAIL', '[email protected]') """ return getattr(settings, key, default)
Set default for get_config to None.
Set default for get_config to None.
Python
bsd-3-clause
skorokithakis/django-annoying,artscoop/django-annoying,kabakchey/django-annoying,skorokithakis/django-annoying,kabakchey/django-annoying,YPCrumble/django-annoying,JshWright/django-annoying
78b2978c3e0e56c4c75a3a6b532e02c995ca69ed
openedx/core/djangoapps/user_api/permissions/views.py
openedx/core/djangoapps/user_api/permissions/views.py
""" NOTE: this API is WIP and has not yet been approved. Do not use this API without talking to Christina or Andy. For more information, see: https://openedx.atlassian.net/wiki/display/TNL/User+API """ from rest_framework.views import APIView from rest_framework.response import Response from rest_framework import status from rest_framework import permissions from django.db import transaction from django.utils.translation import ugettext as _ from openedx.core.lib.api.authentication import ( SessionAuthenticationAllowInactiveUser, OAuth2AuthenticationAllowInactiveUser, ) from openedx.core.lib.api.parsers import MergePatchParser from openedx.core.lib.api.permissions import IsUserInUrlOrStaff from ..errors import UserNotFound, UserNotAuthorized class PermissionsView(APIView): authentication_classes = (OAuth2AuthenticationAllowInactiveUser, SessionAuthenticationAllowInactiveUser) parser_classes = (MergePatchParser,) def get(self, request): """ GET /api/user/v1/ """ try: is_staff = request.user.is_staff except UserNotAuthorized: return Response(status=status.HTTP_403_FORBIDDEN) except UserNotFound: return Response(status=status.HTTP_404_NOT_FOUND) return Response(is_staff)
from rest_framework.views import APIView from rest_framework.response import Response from rest_framework import status from openedx.core.lib.api.authentication import ( SessionAuthenticationAllowInactiveUser, OAuth2AuthenticationAllowInactiveUser, ) from openedx.core.lib.api.parsers import MergePatchParser from ..errors import UserNotFound, UserNotAuthorized class PermissionsView(APIView): authentication_classes = (OAuth2AuthenticationAllowInactiveUser, SessionAuthenticationAllowInactiveUser) parser_classes = (MergePatchParser,) def get(self, request): """ GET /api/user/v1/ """ try: is_staff = request.user.is_staff except UserNotAuthorized: return Response(status=status.HTTP_403_FORBIDDEN) except UserNotFound: return Response(status=status.HTTP_404_NOT_FOUND) return Response(is_staff)
Remove unused import and redundant comment
Remove unused import and redundant comment
Python
agpl-3.0
mbareta/edx-platform-ft,mbareta/edx-platform-ft,mbareta/edx-platform-ft,mbareta/edx-platform-ft
cadee051a462de765bab59ac42d6b372fa49c033
examples/logfile.py
examples/logfile.py
""" Output an Eliot message to a log file using the threaded log writer. """ from __future__ import unicode_literals, print_function from twisted.internet.task import react from eliot.logwriter import ThreadedFileWriter from eliot import Message, Logger, addDestination _logger = Logger() def main(reactor): print("Logging to example-eliot.log...") logWriter = ThreadedFileWriter(open("example-eliot.log", "ab"), reactor) addDestination(logWriter) # Manually start the service. Normally we'd register ThreadedFileWriter # with the usual Twisted Service/Application infrastructure. logWriter.startService() # Log a message: Message.new(value="hello", another=1).write(_logger) # Manually stop the service. done = logWriter.stopService() return done if __name__ == '__main__': react(main, [])
""" Output an Eliot message to a log file using the threaded log writer. """ from __future__ import unicode_literals, print_function from twisted.internet.task import react from eliot.logwriter import ThreadedFileWriter from eliot import Message, Logger _logger = Logger() def main(reactor): print("Logging to example-eliot.log...") logWriter = ThreadedFileWriter(open("example-eliot.log", "ab"), reactor) # Manually start the service, which will add it as a # destination. Normally we'd register ThreadedFileWriter with the usual # Twisted Service/Application infrastructure. logWriter.startService() # Log a message: Message.new(value="hello", another=1).write(_logger) # Manually stop the service. done = logWriter.stopService() return done if __name__ == '__main__': react(main, [])
Fix bug where the service was added as a destination one time too many.
Fix bug where the service was added as a destination one time too many.
Python
apache-2.0
iffy/eliot,ClusterHQ/eliot,ScatterHQ/eliot,ScatterHQ/eliot,ScatterHQ/eliot
9f10dbdabe61ed841c0def319f021a4735f39217
src/sct/templates/__init__.py
src/sct/templates/__init__.py
# -*- coding: utf-8 -*- ''' Copyright 2014 Universitatea de Vest din Timișoara Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. @author: Marian Neagul <[email protected]> @contact: [email protected] @copyright: 2014 Universitatea de Vest din Timișoara '''
# -*- coding: utf-8 -*- """ Copyright 2014 Universitatea de Vest din Timișoara Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. @author: Marian Neagul <[email protected]> @contact: [email protected] @copyright: 2014 Universitatea de Vest din Timișoara """ from sct.templates.hadoop import HadoopServer, HadoopWorker TEMPLATES = { 'hadoop-server': { 'max-node-count': 1, 'cloudinit': HadoopServer }, 'hadoop-worker': { 'max-node-count': None, 'cloudinit': HadoopWorker } } def get_available_templates(): return TEMPLATES.keys() def get_template(name): if name not in TEMPLATES: raise NameError("No such template %s" % name) else: return TEMPLATES.get(name)
Add provisional (needs to be replaced with pkg_resources entry point discovery) template registry
Add provisional (needs to be replaced with pkg_resources entry point discovery) template registry
Python
apache-2.0
mneagul/scape-cloud-toolkit,mneagul/scape-cloud-toolkit,mneagul/scape-cloud-toolkit
a9a794384c6f4c153768cf609f3d8dc657f59daf
campaigns/scrapers.py
campaigns/scrapers.py
import requests import json class KickstarterScraper(object): # TODO: get list of all categories from projects for rendering possible list on main view base_url = "https://www.kickstarter.com/" projects_query_path = "projects/search.json?search={0}&term={1}" @classmethod def scrape_projects(cls, search, term): request_url = cls.base_url + cls.projects_query_path.format(search, term) response = requests.get(request_url).content content = json.loads(response) for item in content: print content return content
import requests import json from bs4 import BeautifulSoup class KickstarterScraper(object): # TODO: get list of all categories from projects for rendering possible list on main view base_url = "https://www.kickstarter.com/" projects_query_path = "projects/search.json?search={0}&term={1}" @classmethod def scrape_projects(cls, search, term): request_url = cls.base_url + cls.projects_query_path.format(search, term) response = requests.get(request_url).content content = json.loads(response) for item in content: print content return content class GiveForwardScraper(object): base_url = "http://www.giveforward.com/" fundraiser_query_path = "fundraisers?query={0}" @classmethod def find_projects(cls, query): response = requests.get(cls.base_url + cls.fundraiser_query_path.format(query)) html = BeautifulSoup(response.content) # button = html.find('input', {'id': 'search_filters_Close To Goal'}) campaigns = html.find_all('div', class_='fr-card-search') almost_funded = [] for indx, campaign in enumerate(campaigns): percent_raised = float(campaign.find('span', class_='meter').get('style').strip('width:').strip('%')) if percent_raised > 90.0 and percent_raised != 100.0: almost_funded.append(campaign) return almost_funded # if __name__ == '__main__': # scraper = GiveForwardScraper # campaigns = scraper.find_projects('cancer') # for campaign in campaigns: # print float(campaign.find('span', class_='meter').get('style').strip('width:').strip('%'))
Add base logic for finding Give Forward campaigns by query that are almost funded
Add base logic for finding Give Forward campaigns by query that are almost funded
Python
mit
lorenanicole/almost_funded,lorenanicole/almost_funded,lorenanicole/almost_funded
19dd4495b09a0019fcce2cfb21b083724033dd7f
handover_service.py
handover_service.py
from flask import Flask app = Flask(__name__) VERSION_PREFIX="/api/v1" @app.route(VERSION_PREFIX + "/handovers") def handovers(): return "handovers\n" @app.route(VERSION_PREFIX + "/drafts") def drafts(): return "drafts\n" if __name__ == "__main__": app.run()
from flask import Flask from flask_restful import Resource, Api app = Flask(__name__) api = Api(app) class Handover(Resource): def get(self): return [{'handover' : 42}] class Draft(Resource): def get(self): return [{'draft' : 1024}] api.add_resource(Handover, '/api/v1/handovers') api.add_resource(Draft, '/api/v1/drafts') if __name__ == "__main__": app.run(debug=True)
Switch skeleton API to use flask-restful
Switch skeleton API to use flask-restful
Python
mit
Duke-GCB/DukeDSHandoverService,Duke-GCB/DukeDSHandoverService,Duke-GCB/DukeDSHandoverService
0534c1cdeb92503a90ef309dee6edddb45234bf7
comrade/users/urls.py
comrade/users/urls.py
from django.conf.urls.defaults import * urlpatterns = patterns('django.contrib.auth.views', url(r'^login/', 'login', name='login'), url(r'^logout/', 'logout', {'next_page':'/'}, name='logout'), url(r'^password/forgot/$', 'password_reset', # LH #269 - ideally this wouldn't be hard coded {'post_reset_redirect':'/accounts/password/forgot/done/'}, name='password_reset'), url(r'^password/forgot/done/$', 'password_reset_done', name='password_reset_done'), url(r'^password/reset/(?P<uidb36>[0-9A-Za-z]+)-(?P<token>.+)/$', 'password_reset_confirm', # LH #269 {'post_reset_redirect':'/accounts/password/reset/done/'}, name='password_reset_confirm'), url(r'^password/reset/done/$', 'password_reset_complete', name='password_reset_complete'), url(r'^password/change/', 'password_change', name='password_change'), url(r'^password/change/done', 'password_change_done', name='password_change'), )
from django.conf.urls.defaults import * from django.core.urlresolvers import reverse from django.utils.functional import lazy reverse_lazy = lazy(reverse, unicode) urlpatterns = patterns('django.contrib.auth.views', url(r'^login/', 'login', name='login'), url(r'^logout/', 'logout', {'next_page':'/'}, name='logout'), url(r'^password/forgot/$', 'password_reset', {'post_reset_redirect':reverse_lazy('users:password_reset_done')}, name='password_reset'), url(r'^password/forgot/done/$', 'password_reset_done', name='password_reset_done'), url(r'^password/reset/(?P<uidb36>[0-9A-Za-z]+)-(?P<token>.+)/$', 'password_reset_confirm', {'post_reset_redirect': reverse_lazy('users:password_reset_complete')}, name='password_reset_confirm'), url(r'^password/reset/done/$', 'password_reset_complete', name='password_reset_complete'), url(r'^password/change/', 'password_change', name='password_change'), url(r'^password/change/done', 'password_change_done', name='password_change'), )
Resolve old Django 1.1 bug in URLs to keep it DRY.
Resolve old Django 1.1 bug in URLs to keep it DRY.
Python
mit
bueda/django-comrade
3572171a917138982cf7e329e5293e1345a9e76d
comics/comics/gws.py
comics/comics/gws.py
from comics.aggregator.crawler import CrawlerBase, CrawlerImage from comics.core.comic_data import ComicDataBase class ComicData(ComicDataBase): name = "Girls With Slingshots" language = "en" url = "http://www.girlswithslingshots.com/" start_date = "2004-09-30" rights = "Danielle Corsetto" class Crawler(CrawlerBase): history_capable_days = 30 schedule = "Mo,Tu,We,Th,Fr" time_zone = "US/Eastern" def crawl(self, pub_date): feed = self.parse_feed("http://www.girlswithslingshots.com/feed/") for entry in feed.for_date(pub_date): page = self.parse_page(entry.link) url = page.src("img#comic") title = entry.title.replace("Girls with Slingshots - ", "") text = page.title("img#comic") return CrawlerImage(url, title, text)
from comics.aggregator.crawler import CrawlerBase, CrawlerImage from comics.core.comic_data import ComicDataBase class ComicData(ComicDataBase): name = "Girls With Slingshots" language = "en" url = "http://www.girlswithslingshots.com/" start_date = "2004-09-30" rights = "Danielle Corsetto" class Crawler(CrawlerBase): history_capable_days = 30 schedule = "Mo,Tu,We,Th,Fr" time_zone = "US/Eastern" def crawl(self, pub_date): feed = self.parse_feed("http://www.girlswithslingshots.com/feed/") for entry in feed.for_date(pub_date): page = self.parse_page(entry.link) url = page.src("img#cc-comic") title = entry.title.replace("Girls with Slingshots - ", "") text = page.title("img#cc-comic") return CrawlerImage(url, title, text)
Update "Girls With Slingshots" after feed change
Update "Girls With Slingshots" after feed change
Python
agpl-3.0
jodal/comics,datagutten/comics,jodal/comics,jodal/comics,datagutten/comics,datagutten/comics,datagutten/comics,jodal/comics
00a3da330668284f700275c7fc3072c792eff374
kolibri/__init__.py
kolibri/__init__.py
""" CAUTION! Keep everything here at at minimum. Do not import stuff. This module is imported in setup.py, so you cannot for instance import a dependency. """ from __future__ import absolute_import from __future__ import print_function from __future__ import unicode_literals from .utils import env from .utils.version import get_version # Setup the environment before loading anything else from the application env.set_env() #: This may not be the exact version as it's subject to modification with #: get_version() - use ``kolibri.__version__`` for the exact version string. VERSION = (0, 12, 5, "beta", 0) __author__ = "Learning Equality" __email__ = "[email protected]" __version__ = str(get_version(VERSION))
""" CAUTION! Keep everything here at at minimum. Do not import stuff. This module is imported in setup.py, so you cannot for instance import a dependency. """ from __future__ import absolute_import from __future__ import print_function from __future__ import unicode_literals from .utils import env from .utils.version import get_version # Setup the environment before loading anything else from the application env.set_env() #: This may not be the exact version as it's subject to modification with #: get_version() - use ``kolibri.__version__`` for the exact version string. VERSION = (0, 12, 5, "final", 0) __author__ = "Learning Equality" __email__ = "[email protected]" __version__ = str(get_version(VERSION))
Update VERSION to 0.12.5 final
Update VERSION to 0.12.5 final
Python
mit
indirectlylit/kolibri,mrpau/kolibri,indirectlylit/kolibri,learningequality/kolibri,mrpau/kolibri,indirectlylit/kolibri,indirectlylit/kolibri,mrpau/kolibri,learningequality/kolibri,learningequality/kolibri,learningequality/kolibri,mrpau/kolibri
e9e4c622ff667e475986e1544ec78b0604b8a511
girder_worker/tasks.py
girder_worker/tasks.py
import core from girder_worker.utils import JobStatus from .app import app def _cleanup(*args, **kwargs): core.events.trigger('cleanup') @app.task(name='girder_worker.run', bind=True, after_return=_cleanup) def run(tasks, *pargs, **kwargs): jobInfo = kwargs.pop('jobInfo', {}) retval = 0 kwargs['_job_manager'] = task.job_manager \ if hasattr(task, 'job_manager') else None kwargs['status'] = JobStatus.RUNNING return core.run(*pargs, **kwargs) @app.task(name='girder_worker.convert') def convert(*pargs, **kwargs): return core.convert(*pargs, **kwargs) @app.task(name='girder_worker.validators') def validators(*pargs, **kwargs): _type, _format = pargs nodes = [] for (node, data) in core.format.conv_graph.nodes(data=True): if ((_type is None) or (_type == node.type)) and \ ((_format is None) or (_format == node.format)): nodes.append({'type': node.type, 'format': node.format, 'validator': data}) return nodes
import core from girder_worker.utils import JobStatus from .app import app def _cleanup(*args, **kwargs): core.events.trigger('cleanup') @app.task(name='girder_worker.run', bind=True, after_return=_cleanup) def run(task, *pargs, **kwargs): kwargs['_job_manager'] = task.job_manager \ if hasattr(task, 'job_manager') else None kwargs['status'] = JobStatus.RUNNING return core.run(*pargs, **kwargs) @app.task(name='girder_worker.convert') def convert(*pargs, **kwargs): return core.convert(*pargs, **kwargs) @app.task(name='girder_worker.validators') def validators(*pargs, **kwargs): _type, _format = pargs nodes = [] for (node, data) in core.format.conv_graph.nodes(data=True): if ((_type is None) or (_type == node.type)) and \ ((_format is None) or (_format == node.format)): nodes.append({'type': node.type, 'format': node.format, 'validator': data}) return nodes
Fix typo from bad conflict resolution during merge
Fix typo from bad conflict resolution during merge
Python
apache-2.0
girder/girder_worker,girder/girder_worker,girder/girder_worker
0a4922dba3367a747d7460b5c1b59c49c67f3026
hcalendar/hcalendar.py
hcalendar/hcalendar.py
#!/usr/bin/env python # -*- coding: utf-8 -*- from __future__ import absolute_import from __future__ import division from __future__ import print_function from __future__ import unicode_literals from .vcalendar import vCalendar from bs4 import BeautifulSoup class hCalendar(object): def __init__(self, markup, value=None, key='id'): if isinstance(markup, BeautifulSoup): self._soup = markup else: self._soup = BeautifulSoup(markup) if value: self._soup = self._soup.find(**{key: value}) self._cals = self._soup.findAll(attrs='vcalendar') if self._cals: self._cals = list(map(vCalendar, self._cals)) else: self._cals = [vCalendar(self._soup)] def __len__(self): return len(self._cals) def __iter__(self): return iter(self._cals) def __getitem__(self, key): return self._cals[key] def getCalendar(self): return self._cals
#!/usr/bin/env python # -*- coding: utf-8 -*- from __future__ import absolute_import from __future__ import division from __future__ import print_function from __future__ import unicode_literals from .vcalendar import vCalendar from bs4 import BeautifulSoup class hCalendar(object): def __init__(self, markup, value=None, key='id'): if isinstance(markup, BeautifulSoup): self._soup = markup else: self._soup = BeautifulSoup(markup, 'html.parser') if value: self._soup = self._soup.find(**{key: value}) self._cals = self._soup.findAll(attrs='vcalendar') if self._cals: self._cals = list(map(vCalendar, self._cals)) else: self._cals = [vCalendar(self._soup)] def __len__(self): return len(self._cals) def __iter__(self): return iter(self._cals) def __getitem__(self, key): return self._cals[key] def getCalendar(self): return self._cals
Add missing parser argument to BeautifulSoup instance
Add missing parser argument to BeautifulSoup instance
Python
mit
mback2k/python-hcalendar
9a32f922e6d5ec6e5bd22eccbe3dceaef7bbd7dc
tailor/tests/utils/charformat_test.py
tailor/tests/utils/charformat_test.py
import unittest from tailor.utils import charformat class MyTestCase(unittest.TestCase): def is_upper_camel_case_test_upper_camel_case_name(self): self.assertTrue(charformat.is_upper_camel_case('HelloWorld')) def is_upper_camel_case_test_lower_camel_case_name(self): self.assertFalse(charformat.is_upper_camel_case('helloWorld')) def is_upper_camel_case_test_blank_name(self): self.assertFalse(charformat.is_upper_camel_case('')) def is_upper_camel_case_test_snake_case_name(self): self.assertFalse(charformat.is_upper_camel_case('Hello_World')) def is_upper_camel_case_test_numeric_name(self): self.assertFalse(charformat.is_upper_camel_case('1ello_world')) if __name__ == '__main__': unittest.main()
import unittest from tailor.utils import charformat class MyTestCase(unittest.TestCase): def is_upper_camel_case_test_upper_camel_case_name(self): self.assertTrue(charformat.is_upper_camel_case('HelloWorld')) def is_upper_camel_case_test_lower_camel_case_name(self): self.assertFalse(charformat.is_upper_camel_case('helloWorld')) def is_upper_camel_case_test_blank_name(self): self.assertFalse(charformat.is_upper_camel_case('')) def is_upper_camel_case_test_snake_case_name(self): self.assertFalse(charformat.is_upper_camel_case('Hello_World')) def is_upper_camel_case_test_numeric_name(self): self.assertFalse(charformat.is_upper_camel_case('1ello_world')) def is_upper_camel_case_test_special_character_name(self): self.assertFalse(charformat.is_upper_camel_case('!ello_world')) if __name__ == '__main__': unittest.main()
Add special character name test case
Add special character name test case
Python
mit
sleekbyte/tailor,sleekbyte/tailor,sleekbyte/tailor,sleekbyte/tailor,sleekbyte/tailor
fd48211548c8c2d5daec0994155ddb7e8d226882
tests/test_anki_sync.py
tests/test_anki_sync.py
import pytest import os import rememberberry from rememberscript import RememberMachine, FileStorage from rememberberry.testing import tmp_data_path, assert_replies, get_isolated_story @pytest.mark.asyncio @tmp_data_path('/tmp/data/', delete=True) async def test_anki_account(): storage = FileStorage() m, storage = get_isolated_story('login_anki', storage) await assert_replies(m.reply(''), 'What is your Anki username?') await assert_replies(m.reply('ajshdkajhsdkajshd'), 'And now the password') await assert_replies(m.reply('jkdhskjhgdksjhg'), 'Authentication with ankiweb failed, try again?', 'What is your Anki username?') await assert_replies(m.reply('[email protected]'), 'And now the password') await assert_replies(m.reply('ankitest'), 'Authentication worked, now I\'ll try to sync your account', 'Syncing anki database', 'Syncing media files (this may take a while)', 'Syncing done', 'Great, you\'re all synced up!', 'enter init')
import pytest import os import rememberberry from rememberscript import RememberMachine, FileStorage from rememberberry.testing import tmp_data_path, assert_replies, get_isolated_story @pytest.mark.asyncio @tmp_data_path('/tmp/data/', delete=True) async def test_anki_account(): storage = FileStorage() storage['username'] = 'alice' m, storage = get_isolated_story('login_anki', storage) await assert_replies(m.reply(''), 'What is your Anki username?') await assert_replies(m.reply('ajshdkajhsdkajshd'), 'And now the password') await assert_replies(m.reply('jkdhskjhgdksjhg'), 'Authentication with ankiweb failed, try again?', 'What is your Anki username?') await assert_replies(m.reply('[email protected]'), 'And now the password') await assert_replies(m.reply('ankitest'), 'Authentication worked, now I\'ll try to sync your account', 'Syncing anki database', 'Syncing media files (this may take a while)', 'Syncing done', 'Great, you\'re all synced up!', 'enter init')
Fix missing username in test
Fix missing username in test
Python
agpl-3.0
rememberberry/rememberberry-server,rememberberry/rememberberry-server
2c38fea1434f8591957c2707359412151c4b6c43
tests/test_timezones.py
tests/test_timezones.py
import unittest import datetime from garage.timezones import TimeZone class TimeZoneTest(unittest.TestCase): def test_time_zone(self): utc = datetime.datetime(2000, 1, 2, 3, 4, 0, 0, TimeZone.UTC) cst = utc.astimezone(TimeZone.CST) print('xxx', utc, cst) self.assertEqual(2000, cst.year) self.assertEqual(1, cst.month) self.assertEqual(2, cst.day) self.assertEqual(11, cst.hour) self.assertEqual(4, cst.minute) self.assertEqual(0, cst.second) self.assertEqual(0, cst.microsecond) if __name__ == '__main__': unittest.main()
import unittest import datetime from garage.timezones import TimeZone class TimeZoneTest(unittest.TestCase): def test_time_zone(self): utc = datetime.datetime(2000, 1, 2, 3, 4, 0, 0, TimeZone.UTC) cst = utc.astimezone(TimeZone.CST) self.assertEqual(2000, cst.year) self.assertEqual(1, cst.month) self.assertEqual(2, cst.day) self.assertEqual(11, cst.hour) self.assertEqual(4, cst.minute) self.assertEqual(0, cst.second) self.assertEqual(0, cst.microsecond) if __name__ == '__main__': unittest.main()
Remove print in unit test
Remove print in unit test
Python
mit
clchiou/garage,clchiou/garage,clchiou/garage,clchiou/garage
b1d3a0c79a52ca1987ea08a546213e1135539927
tools/bots/ddc_tests.py
tools/bots/ddc_tests.py
#!/usr/bin/env python # # Copyright (c) 2016, the Dart project authors. Please see the AUTHORS file # for details. All rights reserved. Use of this source code is governed by a # BSD-style license that can be found in the LICENSE file. import os import os.path import shutil import sys import subprocess import bot import bot_utils utils = bot_utils.GetUtils() BUILD_OS = utils.GuessOS() (bot_name, _) = bot.GetBotName() CHANNEL = bot_utils.GetChannelFromName(bot_name) if __name__ == '__main__': with utils.ChangedWorkingDirectory('pkg/dev_compiler'): dart_exe = utils.CheckedInSdkExecutable() # These two calls mirror pkg/dev_compiler/tool/test.sh. bot.RunProcess([dart_exe, 'tool/build_pkgs.dart', 'test']) bot.RunProcess([dart_exe, 'test/all_tests.dart']) # These mirror pkg/dev_compiler/tool/browser_test.sh. bot.RunProcess(['npm', 'install']) bot.RunProcess(['npm', 'test'])
#!/usr/bin/env python # # Copyright (c) 2016, the Dart project authors. Please see the AUTHORS file # for details. All rights reserved. Use of this source code is governed by a # BSD-style license that can be found in the LICENSE file. import os import os.path import shutil import sys import subprocess import bot import bot_utils utils = bot_utils.GetUtils() BUILD_OS = utils.GuessOS() (bot_name, _) = bot.GetBotName() CHANNEL = bot_utils.GetChannelFromName(bot_name) if __name__ == '__main__': with utils.ChangedWorkingDirectory('pkg/dev_compiler'): dart_exe = utils.CheckedInSdkExecutable() # These two calls mirror pkg/dev_compiler/tool/test.sh. bot.RunProcess([dart_exe, 'tool/build_pkgs.dart', 'test']) bot.RunProcess([dart_exe, 'test/all_tests.dart']) # These mirror pkg/dev_compiler/tool/browser_test.sh. bot.RunProcess(['npm', 'install']) bot.RunProcess(['npm', 'test'], {'CHROME_BIN': 'chrome'})
Set CHROME_BIN on DDC bot
Set CHROME_BIN on DDC bot Noticed the Linux bot is failing on this: https://build.chromium.org/p/client.dart.fyi/builders/ddc-linux-release-be/builds/1724/steps/ddc%20tests/logs/stdio [email protected] Review-Url: https://codereview.chromium.org/2640093002 .
Python
bsd-3-clause
dartino/dart-sdk,dart-archive/dart-sdk,dart-lang/sdk,dartino/dart-sdk,dartino/dart-sdk,dartino/dart-sdk,dartino/dart-sdk,dart-archive/dart-sdk,dart-archive/dart-sdk,dartino/dart-sdk,dartino/dart-sdk,dart-archive/dart-sdk,dartino/dart-sdk,dart-archive/dart-sdk,dart-lang/sdk,dart-archive/dart-sdk,dart-lang/sdk,dart-lang/sdk,dart-archive/dart-sdk,dart-lang/sdk,dart-lang/sdk,dart-lang/sdk,dart-archive/dart-sdk,dart-archive/dart-sdk,dartino/dart-sdk,dart-lang/sdk
c143bc14be8d486d313056c0d1313e03ac438284
examples/ex_aps_parser.py
examples/ex_aps_parser.py
from __future__ import print_function import os import glob import pyingest.parsers.aps as aps import pyingest.parsers.arxiv as arxiv import pyingest.serializers.classic import traceback import json import xmltodict from datetime import datetime input_list = 'bibc.2.out' testfile=[] xmldir = '/proj/ads/fulltext/sources/downloads/cache/APS_HARVEST/harvest.aps.org/v2/journals/articles/' xmltail = '/fulltext.xml' with open(input_list,'rU') as fi: for l in fi.readlines(): doi = l.strip().split('\t')[1] (a,b) = doi.split('/') b = b.replace('.','/') infile = xmldir + a + '/' + b + xmltail testfile.append(infile) for f in testfile: fnord = f[92:] if os.path.isfile(f): print("found! ",fnord) with open(f, 'rU') as fp: parser = aps.APSJATSParser() document = parser.parse(fp) serializer = pyingest.serializers.classic.Tagged() outputfp = open('aps.tag', 'a') serializer.write(document, outputfp) outputfp.close() #except: # print "ERROR!\n%s\n"%f # traceback.print_exc() # pass else: print("not found :( ", fnord)
from __future__ import print_function import os import glob import pyingest.parsers.aps as aps import pyingest.parsers.arxiv as arxiv import pyingest.serializers.classic import traceback import json import xmltodict from datetime import datetime import sys input_list = 'bibc.2.out' testfile=[] xmldir = '/proj/ads/fulltext/sources/downloads/cache/APS_HARVEST/harvest.aps.org/v2/journals/articles/' xmltail = '/fulltext.xml' if sys.version_info > (3,): open_mode = 'r' else: open_mode = 'rU' with open(input_list, open_mode) as fi: for l in fi.readlines(): doi = l.strip().split('\t')[1] (a,b) = doi.split('/') b = b.replace('.', '/') infile = xmldir + a + '/' + b + xmltail testfile.append(infile) for f in testfile: fnord = f[92:] if os.path.isfile(f): print("found! ", fnord) with open(f, open_mode) as fp: parser = aps.APSJATSParser() document = parser.parse(fp) serializer = pyingest.serializers.classic.Tagged() outputfp = open('aps.tag', 'a') serializer.write(document, outputfp) outputfp.close() #except: # print "ERROR!\n%s\n"%f # traceback.print_exc() # pass else: print("not found :( ", fnord)
Use open mode syntax on example file
Use open mode syntax on example file
Python
mit
adsabs/adsabs-pyingest,adsabs/adsabs-pyingest,adsabs/adsabs-pyingest
7d9115aaa429f0a6453c8fcc75c77abc2bdaec93
sort/heap_sort.py
sort/heap_sort.py
def heap_sort(arr): """ Heapsort Complexity: O(n log(n)) """ pass def heapify(arr): pass array = [1,5,65,23,57,1232,-1,-5,-2,242,100,4,423,2,564,9,0,10,43,64] print(array) heap_sort(array) print(array)
Set up basic structure of code
Set up basic structure of code
Python
mit
keon/algorithms,amaozhao/algorithms
bfd166e9679e6fa06e694fd5e587fcf10186d79b
vx_intro.py
vx_intro.py
import vx import math import os import sys _tick_functions = [] def _register_tick_function(f, front=False): if front: _tick_functions.insert(0, f) else: _tick_functions.append(f) def _tick(): for f in _tick_functions: f() vx.my_vx = _tick vx.register_tick_function = _register_tick_function vx.files = sys.argv[1:] import utils import scheduler import keybindings import windows import prompt def _default_start(): if len(vx.files) == 0: win = vx.window(vx.rows, vx.cols, 0, 0) win.blank() win.focus() else: d = math.floor(vx.rows / (len(vx.files))) y = 0 for f in vx.files: win = vx.window(d, vx.cols, y, 0) win.attach_file(f) y += d win.focus() vx.default_start = _default_start sys.path.append(os.path.expanduser('~/.python')) import rc
import vx import math import os import sys _tick_functions = [] def _register_tick_function(f, front=False): if front: _tick_functions.insert(0, f) else: _tick_functions.append(f) def _tick(): for f in _tick_functions: f() vx.my_vx = _tick vx.register_tick_function = _register_tick_function vx.files = sys.argv[1:] import utils import scheduler import keybindings import windows import prompt def _default_start(): if len(vx.files) == 0: win = vx.window(vx.rows, vx.cols, 0, 0) win.blank() win.focus() else: d = math.floor(vx.rows / (len(vx.files))) y = 0 for f in vx.files: win = vx.window(d, vx.cols, y, 0) win.attach_file(f) y += d win.focus() vx.default_start = _default_start sys.path.append(os.path.expanduser('~/.python')) try: import rc except ImportError: pass # just means there was no ~/.python/rc module
Fix a crash if there is no ~/.python/rc.py
Fix a crash if there is no ~/.python/rc.py
Python
mit
philipdexter/vx,philipdexter/vx
bc593f1716a8e36e65cf75a58e524e77d38d5d9c
notation/statistics.py
notation/statistics.py
# encoding: utf-8 # included for ease of use with Python 2 (which has no statistics package) def mean(values): return float(sum(values)) / len(values) def median(values): middle = (len(values) - 1) // 2 if len(values) % 2: return values[middle] else: return mean(values[middle:middle + 2])
# encoding: utf-8 # included for ease of use with Python 2 (which has no statistics package) def mean(values): return float(sum(values)) / len(values) def quantile(p): def bound_quantile(values): ix = int(len(values) * p) if len(values) % 2: return values[ix] elif ix < 1: return values[0] else: return mean(values[ix - 1:ix + 1]) return bound_quantile Q0 = min Q1 = quantile(0.25) Q2 = median = quantile(0.5) Q3 = quantile(0.75) Q4 = max
Add a rudimentary quantile factory function.
Add a rudimentary quantile factory function.
Python
isc
debrouwere/python-ballpark
54c856e987bf570c7bcb8c449726a5d2895c0241
octopus/__init__.py
octopus/__init__.py
__version__ = "trunk" def run (runnable, logging = True): from twisted.internet import reactor if reactor.running: return runnable.run() else: def _complete (result): reactor.stop() def _run (): runnable.run().addBoth(_complete) if logging: import sys from twisted.python import log log.startLogging(sys.stdout) runnable.log += log reactor.callWhenRunning(_run) reactor.run()
__version__ = "trunk" def run (runnable, logging = True): from twisted.internet import reactor if reactor.running: return runnable.run() else: if logging: import sys from twisted.python import log log.startLogging(sys.stdout) runnable.on("log", log.msg) def _complete (result): reactor.stop() if logging: runnable.off("log", log.msg) def _run (): runnable.run().addBoth(_complete) reactor.callWhenRunning(_run) reactor.run()
Fix octopus.run for new events model.
Fix octopus.run for new events model.
Python
mit
richardingham/octopus,richardingham/octopus,richardingham/octopus,richardingham/octopus
fa98f32ce9c2d4e7dff8281bf5e6f154b82599d6
gargoyle/__init__.py
gargoyle/__init__.py
""" gargoyle ~~~~~~~~ :copyright: (c) 2010 DISQUS. :license: Apache License 2.0, see LICENSE for more details. """ __all__ = ('gargoyle', 'ConditionSet', 'autodiscover', 'VERSION') try: VERSION = __import__('pkg_resources') \ .get_distribution('gargoyle').version except Exception, e: VERSION = 'unknown' from gargoyle.manager import gargoyle def autodiscover(): """ Auto-discover INSTALLED_APPS admin.py modules and fail silently when not present. This forces an import on them to register any admin bits they may want. """ import copy from django.conf import settings from django.utils.importlib import import_module for app in settings.INSTALLED_APPS: # Attempt to import the app's gargoyle module. before_import_registry = copy.copy(gargoyle._registry) try: import_module('%s.gargoyle' % app) except: # Reset the model registry to the state before the last import as # this import will have to reoccur on the next request and this # could raise NotRegistered and AlreadyRegistered exceptions gargoyle._registry = before_import_registry # load builtins __import__('gargoyle.builtins')
""" gargoyle ~~~~~~~~ :copyright: (c) 2010 DISQUS. :license: Apache License 2.0, see LICENSE for more details. """ __all__ = ('gargoyle', 'ConditionSet', 'autodiscover', 'VERSION') try: VERSION = __import__('pkg_resources') \ .get_distribution('gargoyle').version except Exception, e: VERSION = 'unknown' from gargoyle.manager import gargoyle def autodiscover(): """ Auto-discover INSTALLED_APPS admin.py modules and fail silently when not present. This forces an import on them to register any admin bits they may want. """ import copy from django.conf import settings from importlib import import_module for app in settings.INSTALLED_APPS: # Attempt to import the app's gargoyle module. before_import_registry = copy.copy(gargoyle._registry) try: import_module('%s.gargoyle' % app) except: # Reset the model registry to the state before the last import as # this import will have to reoccur on the next request and this # could raise NotRegistered and AlreadyRegistered exceptions gargoyle._registry = before_import_registry # load builtins __import__('gargoyle.builtins')
Use python import lib (django import lib will be removed in 1.9).
Use python import lib (django import lib will be removed in 1.9).
Python
apache-2.0
brilliant-org/gargoyle,brilliant-org/gargoyle,brilliant-org/gargoyle
3443c7164e490e0607fff599c497a4fc054f3c48
oslo_cache/_i18n.py
oslo_cache/_i18n.py
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. """oslo.i18n integration module. See http://docs.openstack.org/developer/oslo.i18n/usage.html """ import oslo_i18n _translators = oslo_i18n.TranslatorFactory(domain='oslo.versionedobjects') # The primary translation function using the well-known name "_" _ = _translators.primary # Translators for log levels. # # The abbreviated names are meant to reflect the usual use of a short # name like '_'. The "L" is for "log" and the other letter comes from # the level. _LI = _translators.log_info _LW = _translators.log_warning _LE = _translators.log_error _LC = _translators.log_critical
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. """oslo.i18n integration module. See http://docs.openstack.org/developer/oslo.i18n/usage.html """ import oslo_i18n _translators = oslo_i18n.TranslatorFactory(domain='oslo.cache') # The primary translation function using the well-known name "_" _ = _translators.primary # Translators for log levels. # # The abbreviated names are meant to reflect the usual use of a short # name like '_'. The "L" is for "log" and the other letter comes from # the level. _LI = _translators.log_info _LW = _translators.log_warning _LE = _translators.log_error _LC = _translators.log_critical
Update i18n domain to correct project name
Update i18n domain to correct project name The current oslo_i18n domain name is listed as oslo.versionedobjects Change-Id: I493b66efbd83fb7704fe927866a24b765feb1576
Python
apache-2.0
citrix-openstack-build/oslo.cache,openstack/oslo.cache,openstack/oslo.cache
ec235e290b4428dec2db03a19d678eba52f02fb5
keyring/getpassbackend.py
keyring/getpassbackend.py
"""Specific support for getpass.""" import os import getpass from keyring.core import get_password as original_get_password def get_password(prompt='Password: ', stream=None, service_name='Python', username=None): if username is None: username = getpass.getuser() return original_get_password(service_name, username)
"""Specific support for getpass.""" import os import getpass import keyring.core def get_password(prompt='Password: ', stream=None, service_name='Python', username=None): if username is None: username = getpass.getuser() return keyring.core.get_password(service_name, username)
Use module namespaces to distinguish names instead of 'original_' prefix
Use module namespaces to distinguish names instead of 'original_' prefix
Python
mit
jaraco/keyring
4a711a2709ec5d8a8e04bb0f735fcfaa319cffdf
designate/objects/validation_error.py
designate/objects/validation_error.py
# Copyright 2014 Hewlett-Packard Development Company, L.P. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import six from designate.objects import base class ValidationError(base.DesignateObject): FIELDS = { 'path': {}, 'message': {}, 'validator': {}, 'validator_value': {}, 'raw': {}, } @classmethod def from_js_error(cls, js_error): """Convert a JSON Schema ValidationError instance into a ValidationError instance. """ e = cls() e.path = list(getattr(js_error, 'releative_path', js_error.path)) e.message = six.text_type(js_error) e.validator = js_error.validator e.validator_value = js_error.validator_value e.raw = js_error._contents() return e class ValidationErrorList(base.ListObjectMixin, base.DesignateObject): LIST_ITEM_TYPE = ValidationError
# Copyright 2014 Hewlett-Packard Development Company, L.P. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from designate.objects import base class ValidationError(base.DesignateObject): FIELDS = { 'path': {}, 'message': {}, 'validator': {}, 'validator_value': {}, 'raw': {}, } @classmethod def from_js_error(cls, js_error): """Convert a JSON Schema ValidationError instance into a ValidationError instance. """ e = cls() e.path = list(getattr(js_error, 'releative_path', js_error.path)) e.message = js_error.message e.validator = js_error.validator e.validator_value = js_error.validator_value e.raw = js_error._contents() return e class ValidationErrorList(base.ListObjectMixin, base.DesignateObject): LIST_ITEM_TYPE = ValidationError
Fix the displayed error message in V2 API
Fix the displayed error message in V2 API Change-Id: I07c3f1ed79fa507dbe9b76eb8f5964475516754c
Python
apache-2.0
tonyli71/designate,openstack/designate,ionrock/designate,ionrock/designate,ramsateesh/designate,grahamhayes/designate,cneill/designate-testing,muraliselva10/designate,muraliselva10/designate,cneill/designate-testing,openstack/designate,tonyli71/designate,muraliselva10/designate,grahamhayes/designate,ionrock/designate,tonyli71/designate,grahamhayes/designate,openstack/designate,ramsateesh/designate,cneill/designate-testing,ramsateesh/designate
0ae360b675f2dd0b3607af1bc7b72864e43236b2
userreport/settings_local.EXAMPLE.py
userreport/settings_local.EXAMPLE.py
# Fill in this file and save as settings_local.py PROJECT_NAME = 'SuperTuxKart' PROJECT_URL = 'http://supertuxkart.net/' # SECURITY WARNING: don't run with debug turned on in production! DEBUG = True TEMPLATE_DEBUG = True # Add the name/ip of the server that is running the stats server ALLOWED_HOSTS = ["api.stkaddons.net"] ADMINS = ( ('Your Name', '[email protected]'), ) # Database # https://docs.djangoproject.com/en/1.6/ref/settings/#databases DATABASES = { 'default': { 'ENGINE': 'django.db.backends.mysql', 'NAME': 'stkstats', 'USER': 'stkstats_user', 'PASSWORD': '', 'HOST': 'localhost', 'PORT': '3306', } } # SECURITY WARNING: keep the secret key used in production secret! SECRET_KEY = '#################################################'
# Fill in this file and save as settings_local.py PROJECT_NAME = 'SuperTuxKart' PROJECT_URL = 'http://supertuxkart.net/' # SECURITY WARNING: don't run with debug turned on in production! DEBUG = True TEMPLATE_DEBUG = True # Add the name/ip of the server that is running the stats server ALLOWED_HOSTS = ["addons.supertuxkart.net"] ADMINS = ( ('Your Name', '[email protected]'), ) # Database # https://docs.djangoproject.com/en/1.6/ref/settings/#databases DATABASES = { 'default': { 'ENGINE': 'django.db.backends.mysql', 'NAME': 'stkstats', 'USER': 'stkstats_user', 'PASSWORD': '', 'HOST': 'localhost', 'PORT': '3306', } } # SECURITY WARNING: keep the secret key used in production secret! SECRET_KEY = '#################################################'
Change default example for allowed hosts
Change default example for allowed hosts
Python
mit
leyyin/stk-stats,supertuxkart/stk-stats,leyyin/stk-stats,supertuxkart/stk-stats
6fc2e75426eb34755bf6dbedbd21a4345d9c5738
plugins/websites.py
plugins/websites.py
import re from smartbot import utils class Plugin: def on_message(self, bot, msg, reply): match = re.findall(r"(https?://[^\s]+)", msg["message"], re.IGNORECASE) for i, url in enumerate(match): title = utils.web.get_title(url) if title: reply("[{0}]: {1}".format(i, title)) def on_help(self): return "Echos the titles of websites for any HTTP(S) URL."
import io import re import unittest from smartbot import utils class Plugin: def on_message(self, bot, msg, reply): match = re.findall(r"(https?://[^\s]+)", msg["message"], re.IGNORECASE) for i, url in enumerate(match): title = utils.web.get_title(url) if title: reply("[{0}]: {1}".format(i, title)) def on_help(self): return "Echos the titles of websites for any HTTP(S) URL." class Test(unittest.TestCase): def setUp(self): self.plugin = Plugin() def test_message(self): self.plugin.on_message(None, {"message": "http://tomleese.me.uk"}, lambda x: self.assertEqual("[0]: Tom Leese", x)) def test_help(self): self.assertTrue(self.plugin.on_help())
Add tests for website plugin
Add tests for website plugin
Python
mit
Muzer/smartbot,Cyanogenoid/smartbot,thomasleese/smartbot-old,tomleese/smartbot
58b8b63a8a8e9d1b61d8fc1a0f84f8b2a697efc3
flask_debugtoolbar/panels/versions.py
flask_debugtoolbar/panels/versions.py
import pkg_resources from flask_debugtoolbar.panels import DebugPanel _ = lambda x: x flask_version = pkg_resources.get_distribution('Flask').version class VersionDebugPanel(DebugPanel): """ Panel that displays the Django version. """ name = 'Version' has_content = False def nav_title(self): return _('Versions') def nav_subtitle(self): return 'Flask %s' % flask_version def url(self): return '' def title(self): return _('Versions') def content(self): return None
from flask import __version__ as flask_version from flask_debugtoolbar.panels import DebugPanel _ = lambda x: x class VersionDebugPanel(DebugPanel): """ Panel that displays the Flask version. """ name = 'Version' has_content = False def nav_title(self): return _('Versions') def nav_subtitle(self): return 'Flask %s' % flask_version def url(self): return '' def title(self): return _('Versions') def content(self): return None
Use flask.__version__ instead of pkg_resources.
Use flask.__version__ instead of pkg_resources. This is a simpler way of getting the Flask version.
Python
bsd-3-clause
lepture/flask-debugtoolbar,dianchang/flask-debugtoolbar,lepture/flask-debugtoolbar,dianchang/flask-debugtoolbar,dianchang/flask-debugtoolbar
8852955632b0ef0250ebbe21b5bdefdecdf30e8a
tests/test_dem.py
tests/test_dem.py
import unittest import numpy as np class CalculationMethodsTestCase(unittest.TestCase): def setUp(self): self.dem = DEMGrid() def test_calculate_slope(self): sx, sy = self.dem._calculate_slope() def test_calculate_laplacian(self): del2z = self.dem._calculate_lapalacian() def test_calculate_directional_laplacian(self): alpha = np.pi/4 del2z = self.dem._calculate_lapalacian(alpha) def test_pad_boundary(self): dx = 4 dy = 4 grid = self.dem._griddata pad_x = np.zeros((self.ny, dx/2)) pad_y = np.zeros((self.nx + dx, dy/2)) padgrid = np.vstack([pad_y, np.hstack([pad_x, self.dem._griddata, pad_x]), pad_y]]) self.dem._pad_boundary(dx, dy) assertEqual(self.dem.grid, padgrid, 'Grid padded incorrectly (dx = 2, dy = 2)') dx = 5 dy = 5 grid = self.dem._griddata pad_x = np.zeros((self.ny, np.round(dx/2)) pad_y = np.zeros((self.nx + 2*np.round(dx/2), np.round(dy/2))) padgrid = np.vstack([pad_y, np.hstack([pad_x, self.dem._griddata, pad_x]), pad_y]]) self.dem._pad_boundary(dx, dy) assertEqual(self.dem.grid, padgrid, 'Grid padded incorrectly (dx = 5, dy = 5)')
import unittest import numpy as np class CalculationMethodsTestCase(unittest.TestCase): def setUp(self): self.dem = DEMGrid() def test_calculate_slope(self): sx, sy = self.dem._calculate_slope() def test_calculate_laplacian(self): del2z = self.dem._calculate_lapalacian() def test_calculate_directional_laplacian(self): alpha = np.pi/4 del2z = self.dem._calculate_lapalacian(alpha) def test_pad_boundary(self): dx = 5 dy = 5 grid = self.dem._griddata pad_x = np.zeros((self.ny, np.round(dx/2)) pad_y = np.zeros((self.nx + 2*np.round(dx/2), np.round(dy/2))) padgrid = np.vstack([pad_y, np.hstack([pad_x, self.dem._griddata, pad_x]), pad_y]]) self.dem._pad_boundary(dx, dy) assertEqual(self.dem.grid, padgrid, 'Grid padded incorrectly')
Remove redundant case from padding test
Remove redundant case from padding test
Python
mit
stgl/scarplet,rmsare/scarplet
d9938a50429db16ce60d905bca9844073fe2b0fa
this_app/forms.py
this_app/forms.py
from flask_wtf import FlaskForm from wtforms import StringField, PasswordField from wtforms.validators import Required, Length, Email class SignupForm(FlaskForm): """Render and validate the signup form""" email = StringField("Email", validators=[Required(), Email(), Length(1, 32)]) username = StringField("Username", validators=[Required(), Length(1, 32)]) password = PasswordField("Password", validators=[Required(), Length(1, 32)])
from flask_wtf import FlaskForm from wtforms import StringField, PasswordField, BooleanField from wtforms.validators import DataRequired, Length, Email class SignupForm(FlaskForm): """Render and validate the signup form""" email = StringField("Email", validators=[DataRequired(), Email(message="Invalid email format"), Length(max=32)]) username = StringField("Username", validators=[DataRequired(), Length(2, 32)]) password = PasswordField("Password", validators=[DataRequired(), Length(min=4, max=32)]) class LoginForm(FlaskForm): """Form to let users login""" email = StringField("Username", validators=[DataRequired(), Email(message="Invalid email format"), Length(max=32)]) password = PasswordField("Password", validators=[DataRequired(), Length(4, 32)]) remember = BooleanField("Remember Me")
Use DataRequired to validate form
Use DataRequired to validate form
Python
mit
borenho/flask-bucketlist,borenho/flask-bucketlist
5f2ab0dcaec5a7826ff0652e7c052971083a8398
openid/test/datadriven.py
openid/test/datadriven.py
import unittest class DataDrivenTestCase(unittest.TestCase): cases = [] @classmethod def generateCases(cls): return cls.cases @classmethod def loadTests(cls): tests = [] for case in cls.generateCases(): if isinstance(case, tuple): test = cls(*case) elif isinstance(case, dict): test = cls(**case) else: test = cls(case) tests.append(test) return tests def __init__(self, description): super(DataDrivenTestCase, self).__init__(self, 'runOneTest') self.description = description def shortDescription(self): return '%s for %s' % (self.__class__.__name__, self.description) def loadTests(module_name): loader = unittest.defaultTestLoader this_module = __import__(module_name, {}, {}, [None]) tests = [] for name in dir(this_module): obj = getattr(this_module, name) if isinstance(obj, unittest.TestCase): if hasattr(obj, 'loadTests'): tests.extend(obj.loadTests()) else: tests.append(loader.loadTestsFromTestCase(obj)) return unittest.TestSuite(tests)
import unittest class DataDrivenTestCase(unittest.TestCase): cases = [] @classmethod def generateCases(cls): return cls.cases @classmethod def loadTests(cls): tests = [] for case in cls.generateCases(): if isinstance(case, tuple): test = cls(*case) elif isinstance(case, dict): test = cls(**case) else: test = cls(case) tests.append(test) return tests def __init__(self, description): super(DataDrivenTestCase, self).__init__(self, 'runOneTest') self.description = description def shortDescription(self): return '%s for %s' % (self.__class__.__name__, self.description) def loadTests(module_name): loader = unittest.defaultTestLoader tests = loader.loadTestsFromName(module_name) if not tests: raise AssertionError("No tests for {0}".format(module_name)) return unittest.TestSuite(tests)
Replace ad-hoc pain with builtin methods
Replace ad-hoc pain with builtin methods
Python
apache-2.0
moreati/python3-openid,isagalaev/sm-openid,moreati/python3-openid,moreati/python3-openid,necaris/python3-openid,misli/python3-openid,necaris/python3-openid,misli/python3-openid,misli/python3-openid
82bd501f89d3a228c3de9a2f355266b374c35a54
twork/assembly.py
twork/assembly.py
#!/usr/bin/env python # -*- coding: utf-8 -*- # # Copyright 2012 Zhang ZY<http://idupx.blogspot.com/> # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. ''' set classpath ''' import os import sys CURRENT_PATH = os.path.dirname(os.path.realpath(__file__)) PROJECT_PATH = os.path.realpath(os.path.join(CURRENT_PATH, '..')) if PROJECT_PATH not in sys.path: sys.path.append(PROJECT_PATH) def main(): print 'CURRENT_PATH:', CURRENT_PATH print 'PROJECT_PATH:', PROJECT_PATH if __name__ == '__main__': main()
#!/usr/bin/env python # -*- coding: utf-8 -*- # # Copyright 2012 Zhang ZY<http://idupx.blogspot.com/> # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. ''' set classpath ''' import os import sys CURRENT_PATH = os.path.dirname(os.path.realpath(__file__)) PROJECT_PATH = os.path.realpath(os.path.join(CURRENT_PATH, '..')) if PROJECT_PATH not in sys.path: sys.path.insert(0, PROJECT_PATH) def main(): print 'CURRENT_PATH:', CURRENT_PATH print 'PROJECT_PATH:', PROJECT_PATH if __name__ == '__main__': main()
Add current project path to the first position of sys.modules
Add current project path to the first position of sys.modules
Python
apache-2.0
bufferx/twork,bufferx/twork
847a66ed8eb19206ecc77904dd5db547284b905f
pip/runner.py
pip/runner.py
import sys import os def run(): base = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) ## FIXME: this is kind of crude; if we could create a fake pip ## module, then exec into it and update pip.__path__ properly, we ## wouldn't have to update sys.path: sys.path.insert(0, base) import pip return pip.main() if __name__ == '__main__': run()
import sys import os def run(): base = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) ## FIXME: this is kind of crude; if we could create a fake pip ## module, then exec into it and update pip.__path__ properly, we ## wouldn't have to update sys.path: sys.path.insert(0, base) import pip return pip.main() if __name__ == '__main__': exit = run() if exit: sys.exit(exit)
Make sure exit code is used in -E situation
Make sure exit code is used in -E situation
Python
mit
mindw/pip,pjdelport/pip,patricklaw/pip,alquerci/pip,esc/pip,harrisonfeng/pip,haridsv/pip,prasaianooz/pip,habnabit/pip,ncoghlan/pip,blarghmatey/pip,h4ck3rm1k3/pip,msabramo/pip,Gabriel439/pip,alex/pip,xavfernandez/pip,zvezdan/pip,haridsv/pip,Ivoz/pip,cjerdonek/pip,yati-sagade/pip,harrisonfeng/pip,RonnyPfannschmidt/pip,alex/pip,sigmavirus24/pip,erikrose/pip,qwcode/pip,pjdelport/pip,patricklaw/pip,zvezdan/pip,habnabit/pip,domenkozar/pip,willingc/pip,squidsoup/pip,erikrose/pip,pypa/pip,pfmoore/pip,jmagnusson/pip,mindw/pip,minrk/pip,prasaianooz/pip,davidovich/pip,Carreau/pip,jythontools/pip,techtonik/pip,luzfcb/pip,sbidoul/pip,zenlambda/pip,supriyantomaftuh/pip,blarghmatey/pip,h4ck3rm1k3/pip,pjdelport/pip,xavfernandez/pip,luzfcb/pip,pradyunsg/pip,habnabit/pip,RonnyPfannschmidt/pip,rbtcollins/pip,RonnyPfannschmidt/pip,pypa/pip,yati-sagade/pip,supriyantomaftuh/pip,mindw/pip,pradyunsg/pip,alquerci/pip,rouge8/pip,prasaianooz/pip,jamezpolley/pip,tdsmith/pip,mujiansu/pip,ianw/pip,zvezdan/pip,erikrose/pip,qbdsoft/pip,mujiansu/pip,radiosilence/pip,James-Firth/pip,jasonkying/pip,fiber-space/pip,KarelJakubec/pip,natefoo/pip,qbdsoft/pip,sbidoul/pip,xavfernandez/pip,msabramo/pip,rbtcollins/pip,esc/pip,atdaemon/pip,natefoo/pip,sigmavirus24/pip,willingc/pip,rbtcollins/pip,dstufft/pip,zorosteven/pip,supriyantomaftuh/pip,atdaemon/pip,jamezpolley/pip,tdsmith/pip,ncoghlan/pip,ChristopherHogan/pip,ChristopherHogan/pip,jmagnusson/pip,chaoallsome/pip,luzfcb/pip,squidsoup/pip,wkeyword/pip,jamezpolley/pip,zorosteven/pip,qbdsoft/pip,wkeyword/pip,fiber-space/pip,ianw/pip,esc/pip,wkeyword/pip,davidovich/pip,qwcode/pip,haridsv/pip,graingert/pip,nthall/pip,Carreau/pip,dstufft/pip,alex/pip,Gabriel439/pip,caosmo/pip,ncoghlan/pip,graingert/pip,tdsmith/pip,sigmavirus24/pip,harrisonfeng/pip,graingert/pip,techtonik/pip,blarghmatey/pip,dstufft/pip,rouge8/pip,squidsoup/pip,jythontools/pip,minrk/pip,benesch/pip,davidovich/pip,caosmo/pip,willingc/pip,ChristopherHogan/pip,atdaemon/pip,nthall/pip,h4ck3rm1k3/pip,Ivoz/pip,mujiansu/pip,James-Firth/pip,benesch/pip,yati-sagade/pip,jmagnusson/pip,benesch/pip,techtonik/pip,jasonkying/pip,mattrobenolt/pip,KarelJakubec/pip,jythontools/pip,natefoo/pip,mattrobenolt/pip,caosmo/pip,cjerdonek/pip,pfmoore/pip,KarelJakubec/pip,fiber-space/pip,rouge8/pip,Gabriel439/pip,nthall/pip,zenlambda/pip,zenlambda/pip,chaoallsome/pip,chaoallsome/pip,zorosteven/pip,jasonkying/pip,James-Firth/pip
2bbc289ce21365e18b04cb865328c494b75075fd
numpy/version.py
numpy/version.py
version='0.9.7' import os svn_version_file = os.path.join(os.path.dirname(__file__), 'core','__svn_version__.py') if os.path.isfile(svn_version_file): import imp svn = imp.load_module('numpy.core.__svn_version__', open(svn_version_file), svn_version_file, ('.py','U',1)) version += '.'+svn.version
version='0.9.9' import os svn_version_file = os.path.join(os.path.dirname(__file__), 'core','__svn_version__.py') if os.path.isfile(svn_version_file): import imp svn = imp.load_module('numpy.core.__svn_version__', open(svn_version_file), svn_version_file, ('.py','U',1)) version += '.'+svn.version
Update head revision to 0.9.9
Update head revision to 0.9.9 git-svn-id: 77a43f9646713b91fea7788fad5dfbf67e151ece@2524 94b884b6-d6fd-0310-90d3-974f1d3f35e1
Python
bsd-3-clause
Ademan/NumPy-GSoC,jasonmccampbell/numpy-refactor-sprint,illume/numpy3k,Ademan/NumPy-GSoC,chadnetzer/numpy-gaurdro,jasonmccampbell/numpy-refactor-sprint,Ademan/NumPy-GSoC,efiring/numpy-work,jasonmccampbell/numpy-refactor-sprint,teoliphant/numpy-refactor,chadnetzer/numpy-gaurdro,teoliphant/numpy-refactor,chadnetzer/numpy-gaurdro,illume/numpy3k,illume/numpy3k,efiring/numpy-work,chadnetzer/numpy-gaurdro,efiring/numpy-work,teoliphant/numpy-refactor,illume/numpy3k,teoliphant/numpy-refactor,jasonmccampbell/numpy-refactor-sprint,efiring/numpy-work,Ademan/NumPy-GSoC,teoliphant/numpy-refactor
7a49e7c4344f7d78a84644ade5ca1c3251065f4a
salt/grains/ssds.py
salt/grains/ssds.py
# -*- coding: utf-8 -*- ''' Detect SSDs ''' import os import salt.utils import logging log = logging.getLogger(__name__) def ssds(): ''' Return list of disk devices that are SSD (non-rotational) ''' SSDs = [] for subdir, dirs, files in os.walk('/sys/block'): for dir in dirs: flagfile = subdir + '/' + dir + '/queue/rotational' if os.path.isfile(flagfile): with salt.utils.fopen(flagfile, 'r') as _fp: flag = _fp.read(1) if flag == '0': SSDs.append(dir) log.info(dir + ' is a SSD') elif flag == '1': log.info(dir + ' is no SSD') else: log.warning(flagfile + ' does not report 0 or 1') log.debug(flagfile + ' reports ' + flag) else: log.warning(flagfile + ' does not exist for ' + dir) return {'SSDs': SSDs}
# -*- coding: utf-8 -*- ''' Detect SSDs ''' # Import python libs import glob import salt.utils import logging log = logging.getLogger(__name__) def ssds(): ''' Return list of disk devices that are SSD (non-rotational) ''' ssd_devices = [] for entry in glob.glob('/sys/block/*/queue/rotational'): with salt.utils.fopen(entry) as entry_fp: device = entry.split('/')[3] flag = entry_fp.read(1) if flag == '0': ssd_devices.append(device) log.debug('Device {0} reports itself as an SSD'.format(device)) elif flag == '1': log.debug('Device {0} does not report itself as an SSD'.format(device)) else: log.debug('Unable to identify device {0} as an SSD or not. It does not report 0 or 1'.format(device)) return {'SSDs': ssd_devices}
Use `glob.glob` instead of `os.walk`
Use `glob.glob` instead of `os.walk`
Python
apache-2.0
saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt
89d8ee0b91c9fd579dcf965e9e07f18954625c72
xero/api.py
xero/api.py
from .manager import Manager class Xero(object): """An ORM-like interface to the Xero API""" OBJECT_LIST = (u'Contacts', u'Accounts', u'CreditNotes', u'Currencies', u'Invoices', u'Items', u'Organisation', u'Payments', u'TaxRates', u'TrackingCategories') def __init__(self, credentials): # Iterate through the list of objects we support, for # each of them create an attribute on our self that is # the lowercase name of the object and attach it to an # instance of a Manager object to operate on it for name in self.OBJECT_LIST: setattr(self, name.lower(), Manager(name, credentials.oauth))
from .manager import Manager class Xero(object): """An ORM-like interface to the Xero API""" OBJECT_LIST = (u'Contacts', u'Accounts', u'CreditNotes', u'Currencies', u'Invoices', u'Items', u'Organisation', u'Payments', u'TaxRates', u'TrackingCategories', u'ManualJournals') def __init__(self, credentials): # Iterate through the list of objects we support, for # each of them create an attribute on our self that is # the lowercase name of the object and attach it to an # instance of a Manager object to operate on it for name in self.OBJECT_LIST: setattr(self, name.lower(), Manager(name, credentials.oauth))
Add support for manual journals
Add support for manual journals
Python
bsd-3-clause
wegotpop/pyxero,jarekwg/pyxero,jaymcconnell/pyxero,opendesk/pyxero,thisismyrobot/pyxero,freakboy3742/pyxero,MJMortimer/pyxero,unomena/pyxero,schinckel/pyxero,unomena/pyxeropos,jacobg/pyxero,direvus/pyxero
fb9591c4a2801bfe5f5380c3e33aa44a25db3591
customforms/models.py
customforms/models.py
#!/usr/bin/python # -*- coding: utf-8 -*- from django.utils.translation import ugettext as _ from django.db import models class Form(models.Model): title = models.CharField(_("Title"), max_length=255) def __unicode__(self): return u'%s' % self.title class Meta: ordering = ('title', ) class Question(models.Model): form = models.ForeignKey(Form) title = models.CharField( _("Title"), max_length=255, default=_("Question Title")) help_text = models.TextField(blank=True, null=True) CHOICES = [ ('C', _('Checkbox')), ('R', _('Radio')), ('S', _('Select')), ('T', _('Text')), ] question_type = models.CharField( max_length=1, choices=CHOICES, default="T") required = models.BooleanField(default=False) position = models.PositiveIntegerField(default=0) def __unicode__(self): return u'%s' % (self.title, ) class Meta: ordering = ('form', 'position', ) class Choice(models.Model): question = models.ForeignKey(Question) title = models.CharField(max_length=200,) position = models.PositiveIntegerField(default=0) class Meta: ordering = ('position', ) def __unicode__(self): return u'%s' % (self.title, )
#!/usr/bin/python # -*- coding: utf-8 -*- from django.core.urlresolvers import reverse from django.utils.translation import ugettext as _ from django.db import models class Form(models.Model): title = models.CharField(_("Title"), max_length=255) def __unicode__(self): return u'%s' % self.title class Meta: ordering = ('title', ) def get_absolute_url(self): return reverse('customforms.views.view_form', args=[str(self.id)]) class Question(models.Model): form = models.ForeignKey(Form) title = models.CharField( _("Title"), max_length=255, default=_("Question Title")) help_text = models.TextField(blank=True, null=True) CHOICES = [ ('C', _('Checkbox')), ('R', _('Radio')), ('S', _('Select')), ('T', _('Text')), ] question_type = models.CharField( max_length=1, choices=CHOICES, default="T") required = models.BooleanField(default=False) position = models.PositiveIntegerField(default=0) def __unicode__(self): return u'%s' % (self.title, ) class Meta: ordering = ('form', 'position', ) def get_absolute_url(self): return reverse('customforms.views.view_form', args=[str(self.form.id)]) class Choice(models.Model): question = models.ForeignKey(Question) title = models.CharField(max_length=200,) position = models.PositiveIntegerField(default=0) class Meta: ordering = ('position', ) def __unicode__(self): return u'%s' % (self.title, )
Add absolute URLs to form and question admin
Add absolute URLs to form and question admin
Python
apache-2.0
cschwede/django-customforms
d6ff777c7fb3f645c021da1319bb5d78d13aa9db
meshnet/interface.py
meshnet/interface.py
import serial import struct from siphashc import siphash def _hash(key: str, sender: int, receiver: int, msg_type: int, data: bytes): packed_data = struct.pack(">h>hBs", sender, receiver, msg_type, data) return struct.pack("Q", siphash(key, packed_data)) class SerialMessage(object): def __init__(self): pass def serialize(self): pass class Connection(object): def __init__(self, device): self._device = device self._conn = None def connect(self): self._conn = serial.Serial(self._device, 115200)
import serial import struct from siphashc import siphash def _hash(key: bytes, sender: int, receiver: int, msg_type: int, data: bytes): packed_data = struct.pack(">hhB", sender, receiver, msg_type) + data return struct.pack(">Q", siphash(key, packed_data)) class SerialMessage(object): def __init__(self): pass def serialize(self): pass class Connection(object): def __init__(self, device): self._device = device self._conn = None def connect(self): self._conn = serial.Serial(self._device, 115200)
Fix python siphashing to match c implementation
Fix python siphashing to match c implementation Signed-off-by: Jan Losinski <[email protected]>
Python
bsd-3-clause
janLo/automation_mesh,janLo/automation_mesh,janLo/automation_mesh
b2bab786c4af3dcca7d35b1e6ecff8699e542ec4
pytest_girder/pytest_girder/plugin.py
pytest_girder/pytest_girder/plugin.py
from .fixtures import * # noqa def pytest_addoption(parser): group = parser.getgroup('girder') group.addoption('--mock-db', action='store_true', default=False, help='Whether or not to mock the database using mongomock.') group.addoption('--mongo-uri', action='store', default='mongodb://localhost:27017', help=('The base URI to the MongoDB instance to use for database connections, ' 'default is mongodb://localhost:27017')) group.addoption('--drop-db', action='store', default='both', choices=('both', 'pre', 'post', 'never'), help='When to destroy testing databases, default is both ' '(before and after running tests)')
import os from .fixtures import * # noqa def pytest_configure(config): """ Create the necessary directories for coverage. This is necessary because neither coverage nor pytest-cov have support for making the data_file directory before running. """ covPlugin = config.pluginmanager.get_plugin('_cov') if covPlugin is not None: covPluginConfig = covPlugin.cov_controller.cov.config covDataFileDir = os.path.dirname(covPluginConfig.data_file) try: os.makedirs(covDataFileDir) except OSError: pass def pytest_addoption(parser): group = parser.getgroup('girder') group.addoption('--mock-db', action='store_true', default=False, help='Whether or not to mock the database using mongomock.') group.addoption('--mongo-uri', action='store', default='mongodb://localhost:27017', help=('The base URI to the MongoDB instance to use for database connections, ' 'default is mongodb://localhost:27017')) group.addoption('--drop-db', action='store', default='both', choices=('both', 'pre', 'post', 'never'), help='When to destroy testing databases, default is both ' '(before and after running tests)')
Add a pytest hook for creating the coverage data_file directory
Add a pytest hook for creating the coverage data_file directory
Python
apache-2.0
jbeezley/girder,jbeezley/girder,girder/girder,kotfic/girder,jbeezley/girder,data-exp-lab/girder,Xarthisius/girder,data-exp-lab/girder,girder/girder,RafaelPalomar/girder,jbeezley/girder,girder/girder,kotfic/girder,manthey/girder,kotfic/girder,girder/girder,RafaelPalomar/girder,Xarthisius/girder,RafaelPalomar/girder,Xarthisius/girder,data-exp-lab/girder,manthey/girder,manthey/girder,RafaelPalomar/girder,data-exp-lab/girder,RafaelPalomar/girder,Kitware/girder,manthey/girder,data-exp-lab/girder,Xarthisius/girder,Kitware/girder,Xarthisius/girder,kotfic/girder,Kitware/girder,kotfic/girder,Kitware/girder
b1e6f3eacccb5e575ac47b6a40809f4671510672
rest_flex_fields/utils.py
rest_flex_fields/utils.py
try: # Python 3 from collections.abc import Iterable string_types = (str,) except ImportError: # Python 2 from collections import Iterable string_types = (str, unicode) def is_expanded(request, key): """ Examines request object to return boolean of whether passed field is expanded. """ expand = request.query_params.get("expand", "") expand_fields = [] for e in expand.split(","): expand_fields.extend([e for e in e.split(".")]) return "~all" in expand_fields or key in expand_fields def split_levels(fields): """ Convert dot-notation such as ['a', 'a.b', 'a.d', 'c'] into current-level fields ['a', 'c'] and next-level fields {'a': ['b', 'd']}. """ first_level_fields = [] next_level_fields = {} if not fields: return first_level_fields, next_level_fields assert ( isinstance(fields, Iterable) ), "`fields` must be iterable (e.g. list, tuple, or generator)" if isinstance(fields, string_types): fields = [a.strip() for a in fields.split(",") if a.strip()] for e in fields: if "." in e: first_level, next_level = e.split(".", 1) first_level_fields.append(first_level) next_level_fields.setdefault(first_level, []).append(next_level) else: first_level_fields.append(e) first_level_fields = list(set(first_level_fields)) return first_level_fields, next_level_fields
from collections.abc import Iterable def is_expanded(request, key): """ Examines request object to return boolean of whether passed field is expanded. """ expand = request.query_params.get("expand", "") expand_fields = [] for e in expand.split(","): expand_fields.extend([e for e in e.split(".")]) return "~all" in expand_fields or key in expand_fields def split_levels(fields): """ Convert dot-notation such as ['a', 'a.b', 'a.d', 'c'] into current-level fields ['a', 'c'] and next-level fields {'a': ['b', 'd']}. """ first_level_fields = [] next_level_fields = {} if not fields: return first_level_fields, next_level_fields assert ( isinstance(fields, Iterable) ), "`fields` must be iterable (e.g. list, tuple, or generator)" if isinstance(fields, str): fields = [a.strip() for a in fields.split(",") if a.strip()] for e in fields: if "." in e: first_level, next_level = e.split(".", 1) first_level_fields.append(first_level) next_level_fields.setdefault(first_level, []).append(next_level) else: first_level_fields.append(e) first_level_fields = list(set(first_level_fields)) return first_level_fields, next_level_fields
Drop Python 2 support in split_level utility function
Drop Python 2 support in split_level utility function
Python
mit
rsinger86/drf-flex-fields
cde48bca684e225b2f99be6637380f4ef3365f17
dimod/package_info.py
dimod/package_info.py
__version__ = '1.0.0.dev3' __author__ = 'D-Wave Systems Inc.' __authoremail__ = '[email protected]' __description__ = 'A shared API for binary quadratic model samplers.'
__version__ = '1.0.0.dev4' __author__ = 'D-Wave Systems Inc.' __authoremail__ = '[email protected]' __description__ = 'A shared API for binary quadratic model samplers.'
Update version 1.0.0.dev3 -> 1.0.0.dev4
Update version 1.0.0.dev3 -> 1.0.0.dev4
Python
apache-2.0
dwavesystems/dimod,dwavesystems/dimod
71ea6816eea95e8bf750563718b0dd39114a3c49
pyramid_authsanity/sources.py
pyramid_authsanity/sources.py
from zope.interface import implementer from .interfaces ( IAuthSourceService, ) @implementer(IAuthSourceService) class SessionAuthSource(object): """ An authentication source that uses the current session """ vary = () value_key = 'sanity.value' def __init__(self, context, request): self.request = request self.session = request.session return self def get_value(self): return self.session.get(value_key, [None, None]) def headers_remember(self, value): self.session[value_key] = value return [] def headers_forget(self): if value_key in self.session: del self.session[value_key] return []
from webob.cookies ( SignedCookieProfile, SignedSerializer, ) from zope.interface import implementer from .interfaces ( IAuthSourceService, ) @implementer(IAuthSourceService) class SessionAuthSource(object): """ An authentication source that uses the current session """ vary = () value_key = 'sanity.value' def __init__(self, context, request): self.request = request self.session = request.session return self def get_value(self): return self.session.get(value_key, [None, None]) def headers_remember(self, value): self.session[value_key] = value return [] def headers_forget(self): if value_key in self.session: del self.session[value_key] return [] def CookieAuthSourceFactory( secret, cookie_name='auth', secure=False, max_age=None, httponly=False, path="/", domains=None, timeout=None, reissue_time=None, debug=False, hashalg='sha512', ): """ An authentication source that uses a unique cookie """ @implementer(IAuthSourceService) class CookieAuthSource(object): def __init__(self, context, request): self.domains = domains if self.domains is None: self.domains = [] self.domains.append(request.domain) self.cookie = SignedCookieProfile( secret, 'authsanity', cookie_name, secure=secure, max_age=max_age, httponly=httponly, path=path, domains=domains, hashalg=hashalg, ) # Bind the cookie to the current request self.cookie = self.cookie.bind(request) return self def get_value(self): return self.cookie.get_value() def headers_remember(self, value): return self.cookie.get_headers(value, domains=self.domains) def headers_forget(self): return self.cookie.get_headers('', max_age=0) return CookieAuthSource
Add a cookie based authentication source
Add a cookie based authentication source
Python
isc
usingnamespace/pyramid_authsanity
494f14a69d08e9bfd556fccc6b4e2319db129a38
books/models.py
books/models.py
from django.contrib.auth.models import User from django.db import models from django.db.models import fields class Receipt(models.Model): title = fields.CharField(max_length=255) price = fields.DecimalField(max_digits=10, decimal_places=2) user = models.ForeignKey(User) def __str__(self): return "{}_{}".format(self.title, self.price)
from django.contrib.auth.models import User from django.db import models from django.db.models import fields from django.utils import timezone class Receipt(models.Model): title = fields.CharField(max_length=255) price = fields.DecimalField(max_digits=10, decimal_places=2) created = fields.DateTimeField(auto_now=True) modified = fields.DateTimeField(default=timezone.now()) user = models.ForeignKey(User) def __str__(self): return "{}_{}".format(self.title, self.price)
Add created and modified fields to Receipt
Add created and modified fields to Receipt
Python
mit
trimailov/finance,trimailov/finance,trimailov/finance
b1547647deec6c1edf54c497fa4ed20235ea6902
pymodels/middlelayer/devices/__init__.py
pymodels/middlelayer/devices/__init__.py
from .dcct import DCCT from .li_llrf import LiLLRF from .rf import RF from .sofb import SOFB from .kicker import Kicker from .septum import Septum from .screen import Screen from .bpm import BPM from .ict import ICT from .ict import TranspEff from .egun import HVPS from .egun import Filament
from .dcct import DCCT from .li_llrf import LiLLRF from .rf import RF from .sofb import SOFB from .kicker import Kicker from .septum import Septum from .screen import Screen from .bpm import BPM from .ict import ICT from .ict import TranspEff from .egun import Bias from .egun import Filament from .egun import HVPS
Add missing egun.bias in init
ENH: Add missing egun.bias in init
Python
mit
lnls-fac/sirius
5856e4daaf141e5bf9cdef438378a3757297f9c0
recipe_scrapers/wholefoods.py
recipe_scrapers/wholefoods.py
from ._abstract import AbstractScraper class WholeFoods(AbstractScraper): @classmethod def host(self, domain="com"): return f"www.wholefoodsmarket.{domain}"
from ._abstract import AbstractScraper class WholeFoods(AbstractScraper): @classmethod def host(self, domain="com"): return f"www.wholefoodsmarket.{domain}" def title(self): return self.schema.title() def total_time(self): return self.schema.total_time() def yields(self): return self.schema.yields() def image(self): return self.schema.image() def ingredients(self): return self.schema.ingredients() def instructions(self): return self.schema.instructions() def ratings(self): return self.schema.ratings()
Add wrapper methods for clarity.
Add wrapper methods for clarity.
Python
mit
hhursev/recipe-scraper
b4e8dd76e3095941c9837151b263365f08426ea1
WEIPDCRM/styles/DefaultStyle/views/chart.py
WEIPDCRM/styles/DefaultStyle/views/chart.py
# coding=utf-8 """ DCRM - Darwin Cydia Repository Manager Copyright (C) 2017 WU Zheng <[email protected]> & 0xJacky <[email protected]> This program is free software: you can redistribute it and/or modify it under the terms of the GNU Affero General Public License as published by the Free Software Foundation, either version 3 of the License, or (at your option) any later version. This program is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Affero General Public License for more details. You should have received a copy of the GNU Affero General Public License along with this program. If not, see <http://www.gnu.org/licenses/>. Notice: You have used class-based views, that's awesome. If not necessary, you can try function-based views. You may add lines above as license. """ from django.views.generic import ListView from WEIPDCRM.models.package import Package class ChartView(ListView): model = Package context_object_name = 'package_list' ordering = '-download_times' template_name = 'frontend/chart.html' def get_queryset(self): """ Get 24 packages ordering by download times. :return: QuerySet """ queryset = super(ChartView, self).get_queryset().all()[:24] return queryset
# coding=utf-8 """ DCRM - Darwin Cydia Repository Manager Copyright (C) 2017 WU Zheng <[email protected]> & 0xJacky <[email protected]> This program is free software: you can redistribute it and/or modify it under the terms of the GNU Affero General Public License as published by the Free Software Foundation, either version 3 of the License, or (at your option) any later version. This program is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Affero General Public License for more details. You should have received a copy of the GNU Affero General Public License along with this program. If not, see <http://www.gnu.org/licenses/>. Notice: You have used class-based views, that's awesome. If not necessary, you can try function-based views. You may add lines above as license. """ from django.views.generic import ListView from WEIPDCRM.models.package import Package class ChartView(ListView): model = Package context_object_name = 'package_list' ordering = '-download_count' template_name = 'frontend/chart.html' def get_queryset(self): """ Get 24 packages ordering by download times. :return: QuerySet """ queryset = super(ChartView, self).get_queryset().all()[:24] return queryset
Fix privileges of package frontend.
Fix privileges of package frontend.
Python
agpl-3.0
82Flex/DCRM,82Flex/DCRM,82Flex/DCRM,82Flex/DCRM
5f42f76ffd11e82d51a334b91d64723388ca4a0d
newswall/providers/feed.py
newswall/providers/feed.py
from datetime import datetime import feedparser import time from newswall.providers.base import ProviderBase class Provider(ProviderBase): def update(self): feed = feedparser.parse(self.config['source']) for entry in feed['entries']: self.create_story(entry.link, title=entry.title, body=entry.description, timestamp=datetime.fromtimestamp(time.mktime(entry.date_parsed)), )
""" RSS Feed Provider ================= Required configuration keys:: { "provider": "newswall.providers.feed", "source": "http://twitter.com/statuses/user_timeline/feinheit.rss" } """ from datetime import datetime import feedparser import time from newswall.providers.base import ProviderBase class Provider(ProviderBase): def update(self): feed = feedparser.parse(self.config['source']) for entry in feed['entries']: self.create_story(entry.link, title=entry.title, body=entry.description, timestamp=datetime.fromtimestamp(time.mktime(entry.date_parsed)), )
Add RSS Feed Provider docs
Add RSS Feed Provider docs
Python
bsd-3-clause
michaelkuty/django-newswall,registerguard/django-newswall,matthiask/django-newswall,HerraLampila/django-newswall,registerguard/django-newswall,HerraLampila/django-newswall,michaelkuty/django-newswall,matthiask/django-newswall
931e2d1e8ba3fd6b129a6d74e3a1ad9984c1938a
benchmarks/benchmarks/bench_random.py
benchmarks/benchmarks/bench_random.py
from __future__ import absolute_import, division, print_function from .common import Benchmark import numpy as np class Random(Benchmark): params = ['normal', 'uniform', 'weibull 1', 'binomial 10 0.5', 'poisson 10'] def setup(self, name): items = name.split() name = items.pop(0) params = [float(x) for x in items] self.func = getattr(np.random, name) self.params = tuple(params) + ((100, 100),) def time_rng(self, name): self.func(*self.params) class Shuffle(Benchmark): def setup(self): self.a = np.arange(100000) def time_100000(self): np.random.shuffle(self.a)
from __future__ import absolute_import, division, print_function from .common import Benchmark import numpy as np from numpy.lib import NumpyVersion class Random(Benchmark): params = ['normal', 'uniform', 'weibull 1', 'binomial 10 0.5', 'poisson 10'] def setup(self, name): items = name.split() name = items.pop(0) params = [float(x) for x in items] self.func = getattr(np.random, name) self.params = tuple(params) + ((100, 100),) def time_rng(self, name): self.func(*self.params) class Shuffle(Benchmark): def setup(self): self.a = np.arange(100000) def time_100000(self): np.random.shuffle(self.a) class Randint(Benchmark): def time_randint_fast(self): """Compare to uint32 below""" np.random.randint(0, 2**30, size=10**5) def time_randint_slow(self): """Compare to uint32 below""" np.random.randint(0, 2**30 + 1, size=10**5) class Randint_dtype(Benchmark): high = { 'bool': 1, 'uint8': 2**7, 'uint16': 2**15, 'uint32': 2**31, 'uint64': 2**63 } param_names = ['dtype'] params = ['bool', 'uint8', 'uint16', 'uint32', 'uint64'] def setup(self, name): if NumpyVersion(np.__version__) < '1.11.0.dev0': raise NotImplementedError def time_randint_fast(self, name): high = self.high[name] np.random.randint(0, high, size=10**5, dtype=name) def time_randint_slow(self, name): high = self.high[name] np.random.randint(0, high + 1, size=10**5, dtype=name)
Add benchmark tests for numpy.random.randint.
ENH: Add benchmark tests for numpy.random.randint. This add benchmarks randint. There is one set of benchmarks for the default dtype, 'l', that can be tracked back, and another set for the new dtypes 'bool', 'uint8', 'uint16', 'uint32', and 'uint64'.
Python
bsd-3-clause
shoyer/numpy,Dapid/numpy,jakirkham/numpy,WarrenWeckesser/numpy,chatcannon/numpy,WarrenWeckesser/numpy,b-carter/numpy,anntzer/numpy,ssanderson/numpy,simongibbons/numpy,nbeaver/numpy,SiccarPoint/numpy,numpy/numpy,Eric89GXL/numpy,kiwifb/numpy,seberg/numpy,rgommers/numpy,ESSS/numpy,shoyer/numpy,anntzer/numpy,utke1/numpy,dwillmer/numpy,grlee77/numpy,ddasilva/numpy,charris/numpy,tacaswell/numpy,simongibbons/numpy,endolith/numpy,solarjoe/numpy,numpy/numpy,WarrenWeckesser/numpy,stuarteberg/numpy,SiccarPoint/numpy,mhvk/numpy,ahaldane/numpy,rgommers/numpy,bringingheavendown/numpy,anntzer/numpy,ContinuumIO/numpy,Eric89GXL/numpy,kiwifb/numpy,bringingheavendown/numpy,MSeifert04/numpy,solarjoe/numpy,ahaldane/numpy,jakirkham/numpy,maniteja123/numpy,anntzer/numpy,ssanderson/numpy,tacaswell/numpy,WarrenWeckesser/numpy,ContinuumIO/numpy,maniteja123/numpy,njase/numpy,jakirkham/numpy,maniteja123/numpy,drasmuss/numpy,tynn/numpy,shoyer/numpy,endolith/numpy,madphysicist/numpy,stuarteberg/numpy,madphysicist/numpy,jakirkham/numpy,abalkin/numpy,Dapid/numpy,pbrod/numpy,ContinuumIO/numpy,pdebuyl/numpy,pbrod/numpy,mattip/numpy,gmcastil/numpy,rherault-insa/numpy,stuarteberg/numpy,ESSS/numpy,njase/numpy,jonathanunderwood/numpy,jorisvandenbossche/numpy,gfyoung/numpy,b-carter/numpy,jorisvandenbossche/numpy,grlee77/numpy,jonathanunderwood/numpy,pizzathief/numpy,seberg/numpy,drasmuss/numpy,skwbc/numpy,skwbc/numpy,grlee77/numpy,Eric89GXL/numpy,AustereCuriosity/numpy,gfyoung/numpy,SiccarPoint/numpy,pbrod/numpy,rherault-insa/numpy,dwillmer/numpy,ddasilva/numpy,charris/numpy,simongibbons/numpy,chiffa/numpy,chatcannon/numpy,simongibbons/numpy,argriffing/numpy,mhvk/numpy,shoyer/numpy,njase/numpy,grlee77/numpy,pbrod/numpy,WarrenWeckesser/numpy,pizzathief/numpy,pizzathief/numpy,SiccarPoint/numpy,dwillmer/numpy,MSeifert04/numpy,MSeifert04/numpy,seberg/numpy,joferkington/numpy,MSeifert04/numpy,skwbc/numpy,joferkington/numpy,nbeaver/numpy,pdebuyl/numpy,abalkin/numpy,bertrand-l/numpy,madphysicist/numpy,pdebuyl/numpy,bertrand-l/numpy,rherault-insa/numpy,rgommers/numpy,gmcastil/numpy,dwillmer/numpy,tacaswell/numpy,drasmuss/numpy,seberg/numpy,chiffa/numpy,jakirkham/numpy,endolith/numpy,pbrod/numpy,mhvk/numpy,pdebuyl/numpy,mhvk/numpy,charris/numpy,argriffing/numpy,gfyoung/numpy,chatcannon/numpy,pizzathief/numpy,AustereCuriosity/numpy,stuarteberg/numpy,charris/numpy,MSeifert04/numpy,bringingheavendown/numpy,joferkington/numpy,shoyer/numpy,numpy/numpy,jorisvandenbossche/numpy,Dapid/numpy,simongibbons/numpy,mhvk/numpy,mattip/numpy,jorisvandenbossche/numpy,endolith/numpy,ESSS/numpy,behzadnouri/numpy,chiffa/numpy,kiwifb/numpy,argriffing/numpy,jorisvandenbossche/numpy,joferkington/numpy,behzadnouri/numpy,AustereCuriosity/numpy,utke1/numpy,tynn/numpy,grlee77/numpy,ssanderson/numpy,behzadnouri/numpy,madphysicist/numpy,mattip/numpy,Eric89GXL/numpy,ahaldane/numpy,jonathanunderwood/numpy,abalkin/numpy,ahaldane/numpy,madphysicist/numpy,solarjoe/numpy,utke1/numpy,gmcastil/numpy,ddasilva/numpy,numpy/numpy,tynn/numpy,b-carter/numpy,pizzathief/numpy,mattip/numpy,ahaldane/numpy,bertrand-l/numpy,rgommers/numpy,nbeaver/numpy
ca8e15d50b816c29fc2a0df27d0266826e38b5b8
cellcounter/statistics/serializers.py
cellcounter/statistics/serializers.py
from rest_framework.serializers import ModelSerializer from .models import CountInstance class CountInstanceSerializer(ModelSerializer): class Meta: model = CountInstance
from rest_framework.serializers import ModelSerializer from .models import CountInstance class CountInstanceSerializer(ModelSerializer): class Meta: model = CountInstance fields = ('count_total',)
Update serializer to deal with new model
Update serializer to deal with new model
Python
mit
cellcounter/cellcounter,haematologic/cellcounter,cellcounter/cellcounter,cellcounter/cellcounter,haematologic/cellcounter,haematologic/cellcounter,cellcounter/cellcounter
6f4758b39c257dcabcabc6405cf400e8f6a358ea
cpt/__init__.py
cpt/__init__.py
__version__ = '0.35.0-dev' def get_client_version(): from conans.model.version import Version from conans import __version__ as client_version from os import getenv # It is a mess comparing dev versions, lets assume that the -dev is the further release return Version(client_version.replace("-dev", ""))
__version__ = '0.36.0-dev' def get_client_version(): from conans.model.version import Version from conans import __version__ as client_version from os import getenv # It is a mess comparing dev versions, lets assume that the -dev is the further release return Version(client_version.replace("-dev", ""))
Update develop version to 0.36.0
Update develop version to 0.36.0 Signed-off-by: Uilian Ries <[email protected]>
Python
mit
conan-io/conan-package-tools
3245946ff25889149dc60cf6b1364bd09c953809
faas/puzzleboard-pop/puzzleboard_pop.py
faas/puzzleboard-pop/puzzleboard_pop.py
import json from datetime import datetime import requests from .model.puzzleboard import pop_puzzleboard class HuntwordsPuzzleBoardPopCommand(object): '''Command class that processes puzzleboard-pop message''' def run(self, jreq): '''Command that processes puzzleboard-pop message''' req = json.loads(jreq) pboard = pop_puzzleboard(req['puzzle']) jpboard = json.dumps(dict(pboard)) resp = { 'puzzleboard': jpboard, 'processed': { 'at': f'{datetime.now().isoformat()}', 'status': 'ok' } } send_consumed(pboard) return json.dumps(resp) def send_consumed(pboard): '''Send async request to generate a new copy''' url = '/async-function/puzzleboard-consumed' data = f'{{"puzzle": "{pboard.puzzle.name}" }}' requests.post(url, data)
import json from datetime import datetime import requests from .model.puzzleboard import pop_puzzleboard class HuntwordsPuzzleBoardPopCommand(object): '''Command class that processes puzzleboard-pop message''' def run(self, jreq): '''Command that processes puzzleboard-pop message''' req = json.loads(jreq) pboard = pop_puzzleboard(req['puzzle']) jpboard = json.dumps(dict(pboard)) resp = { 'puzzleboard': jpboard, 'processed': { 'at': f'{datetime.now().isoformat()}', 'status': 'ok' } } send_consumed(pboard) return json.dumps(resp) def send_consumed(pboard): '''Send async request to generate a new copy''' url = 'http://puzzleboard-consumed.openfaas-fn:8080' data = f'{{"puzzle": "{pboard.puzzle.name}" }}' requests.post(url, data)
Change url from relative to internal service endpoint
Change url from relative to internal service endpoint
Python
mit
klmcwhirter/huntwords,klmcwhirter/huntwords,klmcwhirter/huntwords,klmcwhirter/huntwords
608dc0db688be1dabe3c6ba7647807f6697fcefe
tools/misc/python/test-data-in-out.py
tools/misc/python/test-data-in-out.py
# TOOL test-data-in-out.py: "Test data input and output in Python" (Data input output test.) # INPUT input TYPE GENERIC # OUTPUT output # OUTPUT OPTIONAL missing_output.txt import shutil shutil.copyfile('input', 'output')
# TOOL test-data-in-out.py: "Test data input and output in Python" (Data input output test.) # INPUT input TYPE GENERIC # OUTPUT output # OUTPUT OPTIONAL missing_output.txt # IMAGE chipster-tools-python import shutil shutil.copyfile('input', 'output')
Test image definition in SADL
Test image definition in SADL
Python
mit
chipster/chipster-tools,chipster/chipster-tools,chipster/chipster-tools,chipster/chipster-tools
5548e32a32bd1cd5951ce50e74c0fad944a1cf04
ideascube/conf/idb_col_llavedelsaber.py
ideascube/conf/idb_col_llavedelsaber.py
"""Configuration for Llave Del Saber, Colombia""" from .idb import * # noqa from django.utils.translation import ugettext_lazy as _ LANGUAGE_CODE = 'es' DOMAIN = 'bibliotecamovil.lan' ALLOWED_HOSTS = ['.bibliotecamovil.lan', 'localhost'] USER_FORM_FIELDS = USER_FORM_FIELDS + ( (_('Personal informations'), ['extra', 'disabilities']), ) USER_EXTRA_FIELD_LABEL = 'Etnicidad'
"""Configuration for Llave Del Saber, Colombia""" from .idb import * # noqa from django.utils.translation import ugettext_lazy as _ LANGUAGE_CODE = 'es' DOMAIN = 'bibliotecamovil.lan' ALLOWED_HOSTS = ['.bibliotecamovil.lan', 'localhost'] USER_FORM_FIELDS = USER_FORM_FIELDS + ( (_('Personal informations'), ['disabilities']), )
Stop using the extra field for Colombia
Stop using the extra field for Colombia After discussion, this is not something we will have in Ideascube. Fixes #609
Python
agpl-3.0
ideascube/ideascube,ideascube/ideascube,ideascube/ideascube,ideascube/ideascube
c1b433e5ed4c06b956b4d27f6da4e8b1dab54aaf
services/cloudwatch/sample.py
services/cloudwatch/sample.py
''' =================================== Boto 3 - CloudWatch Service Example =================================== This application implements the CloudWatch service that lets you gets information from Amazon Cloud Watch. See the README for more details. ''' import boto3 ''' Define your AWS credentials: ''' AWS_ACCESS_KEY_ID = 'AKIAJM7BQ4WBJJSVU2JQ' AWS_SECRET_ACCESS_KEY = 'Fq9GmwWEsvbcdHuh4McD+ZUmfowPKrnzFmhczV2U' ''' Connection to AWS. ''' client = boto3.client('cloudwatch', aws_access_key_id=AWS_ACCESS_KEY_ID, aws_secret_access_key=AWS_SECRET_ACCESS_KEY) # Main program: if __name__ == '__main__': print_results()
''' =================================== Boto 3 - CloudWatch Service Example =================================== This application implements the CloudWatch service that lets you gets information from Amazon Cloud Watch. See the README for more details. ''' import boto3 ''' Define your AWS credentials: ''' AWS_ACCESS_KEY_ID = '<YOUR ACCESS KEY ID>' AWS_SECRET_ACCESS_KEY = '<YOUR SECRET ACCESS KEY>' ''' Connection to AWS. ''' client = boto3.client('cloudwatch', aws_access_key_id=AWS_ACCESS_KEY_ID, aws_secret_access_key=AWS_SECRET_ACCESS_KEY) # Main program: if __name__ == '__main__': print_results()
Fix issue in cloudwacth service credentials
Fix issue in cloudwacth service credentials
Python
mit
rolandovillca/aws_samples_boto3_sdk
a05a05f24c29dcf039e02b55c18c476dc69757df
shell_manager/problem_repo.py
shell_manager/problem_repo.py
""" Problem repository management for the shell manager. """ import spur, gzip from shutil import copy2 from os.path import join def local_update(repo_path, deb_paths=[]): """ Updates a local deb repository by copying debs and running scanpackages. Args: repo_path: the path to the local repository. dep_paths: list of problem deb paths to copy. """ [copy2(deb_path, repo_path) for deb_path in deb_paths] shell = spur.LocalShell() result = shell.run(["dpkg-scanpackages", ".", "/dev/null"], cwd=repo_path) packages_path = join(repo_path, "Packages.gz") with gzip.open(packages_path, "wb") as packages: packages.write(result.output) print("Updated problem repository.")
""" Problem repository management for the shell manager. """ import spur, gzip from shutil import copy2 from os.path import join def update_repo(args): """ Main entrypoint for repo update operations. """ if args.repo_type == "local": local_update(args.repository, args.package_paths) else: remote_update(args.repository, args.package_paths) def remote_update(repo_ui, deb_paths=[]): """ Pushes packages to a remote deb repository. Args: repo_uri: location of the repository. deb_paths: list of problem deb paths to copy. """ pass def local_update(repo_path, deb_paths=[]): """ Updates a local deb repository by copying debs and running scanpackages. Args: repo_path: the path to the local repository. dep_paths: list of problem deb paths to copy. """ [copy2(deb_path, repo_path) for deb_path in deb_paths] shell = spur.LocalShell() result = shell.run(["dpkg-scanpackages", ".", "/dev/null"], cwd=repo_path) packages_path = join(repo_path, "Packages.gz") with gzip.open(packages_path, "wb") as packages: packages.write(result.output) print("Updated problem repository.")
Update repo entrypoint and remote_update stub.
Update repo entrypoint and remote_update stub.
Python
mit
RitwikGupta/picoCTF-shell-manager,cganas/picoCTF-shell-manager,RitwikGupta/picoCTF-shell-manager,cganas/picoCTF-shell-manager,picoCTF/picoCTF-shell-manager,cganas/picoCTF-shell-manager,cganas/picoCTF-shell-manager,RitwikGupta/picoCTF-shell-manager,picoCTF/picoCTF-shell-manager,picoCTF/picoCTF-shell-manager,picoCTF/picoCTF-shell-manager,RitwikGupta/picoCTF-shell-manager
6f7dba3beccca655b84879ccd0f3071d15536b2f
test/utils.py
test/utils.py
# coding: utf-8 import string import random def generate_string(str_len=6, src=string.ascii_lowercase): return "".join(random.choice(src) for x in xrange(str_len)) def lorem_ipsum(): words_count = random.randint(20, 50) lorem = list([]) for i in xrange(words_count): word_length = random.randint(4, 8) lorem.append(generate_string(str_len=word_length)) return " ".join(lorem)
# coding: utf-8 import string import random def generate_string(str_len=6, src=string.ascii_lowercase): return "".join(random.choice(src) for x in xrange(str_len)) def lorem_ipsum(words_count=30): lorem = list([]) for i in xrange(words_count): word_length = random.randint(4, 8) lorem.append(generate_string(str_len=word_length)) return " ".join(lorem)
Add word_count parameter for lorem_ipsum generator
Add word_count parameter for lorem_ipsum generator
Python
mit
sarutobi/Rynda,sarutobi/flowofkindness,sarutobi/ritmserdtsa,sarutobi/ritmserdtsa,sarutobi/flowofkindness,sarutobi/Rynda,sarutobi/ritmserdtsa,sarutobi/ritmserdtsa,sarutobi/Rynda,sarutobi/flowofkindness,sarutobi/flowofkindness,sarutobi/Rynda
d80f7a89b5bc23802ad5ec9bb8cc6ad523976718
test_gitnl.py
test_gitnl.py
from __future__ import print_function, division, absolute_import import unittest import gitnl class GitnlTestCase(unittest.TestCase): """Tests from 'gitnl.py'.""" def test_push_remotename_branchfrom(self): desired = 'push remotename branchfrom' actual = gitnl.parse_to_git('push my branch branchfrom to a remote called remotename') self.assertEqual(actual, desired) if __name__ == '__main__': unittest.main()
from __future__ import print_function, division, absolute_import import unittest import gitnl class GitnlTestCase(unittest.TestCase): """Tests from 'gitnl.py'.""" def test_push_remotename_branchfrom(self): desired = 'push remotename branchfrom' actual = gitnl.parse_to_git('push my branch branchfrom to a remote called remotename') self.assertEqual(actual, desired) def test_rename_branch(self): desired = 'branch -m old_branch new_branch' actual = gitnl.parse_to_git('branch rename branch old_branch to new_branch') self.assertEqual(actual, desired) if __name__ == '__main__': unittest.main()
Add rename branch locally test
Add rename branch locally test
Python
mit
eteq/gitnl,eteq/gitnl
fb213097e838ddfa40d9f71f1705d7af661cfbdf
tests/unit.py
tests/unit.py
# -*- coding: latin-1 -*- import unittest from github2.issues import Issue from github2.client import Github class ReprTests(unittest.TestCase): """__repr__ must return strings, not unicode objects.""" def test_issue(self): """Issues can have non-ASCII characters in the title.""" i = Issue(title=u'abcdé') self.assertEqual(str, type(repr(i))) class RateLimits(unittest.TestCase): """ How should we handle actual API calls such that tests can run? Perhaps the library should support a ~/.python_github2.conf from which to get the auth? """ def test_delays(self): import datetime USERNAME = '' API_KEY = '' client = Github(username=USERNAME, api_token=API_KEY, requests_per_second=.5) client.users.show('defunkt') start = datetime.datetime.now() client.users.show('mojombo') end = datetime.datetime.now() self.assertGreaterEqual((end - start).total_seconds(), 2.0, "Expected .5 reqs per second to require a 2 second delay between " "calls.")
# -*- coding: latin-1 -*- import unittest from github2.issues import Issue from github2.client import Github class ReprTests(unittest.TestCase): """__repr__ must return strings, not unicode objects.""" def test_issue(self): """Issues can have non-ASCII characters in the title.""" i = Issue(title=u'abcdé') self.assertEqual(str, type(repr(i))) class RateLimits(unittest.TestCase): """ How should we handle actual API calls such that tests can run? Perhaps the library should support a ~/.python_github2.conf from which to get the auth? """ def test_delays(self): import datetime USERNAME = '' API_KEY = '' client = Github(username=USERNAME, api_token=API_KEY, requests_per_second=.5) client.users.show('defunkt') start = datetime.datetime.now() client.users.show('mojombo') end = datetime.datetime.now() delta = end - start delta_seconds = delta.days * 24 * 60 * 60 + delta.seconds self.assertTrue(delta_seconds >= 2, "Expected .5 reqs per second to require a 2 second delay between " "calls.")
Allow tests to be run with Python <2.6.
Allow tests to be run with Python <2.6.
Python
bsd-3-clause
ask/python-github2
d19fa3b085d691780bbdc7b8e5edf9e8b53906e6
todo/views.py
todo/views.py
from todo import app from flask import jsonify, request, url_for from flask import json from todo.database import db_session from todo.models import Entry @app.route("/", methods=["GET", "POST", "DELETE"]) def index(): if request.method == "POST": request_json = request.get_json() entry = Entry(request_json["title"]) db_session.add(entry) db_session.commit() return jsonify(construct_dict(entry, request)) else: if request.method == "DELETE": Entry.query.delete() db_session.commit() response = [] for entry in Entry.query.all(): response.append(construct_dict(entry, request)) return json.dumps(response) @app.route("/<int:entry_id>") def entry(entry_id): return jsonify(construct_dict(Entry.query.filter(Entry.id == entry_id).first(), request)) def construct_dict(entry, request): with request: return dict(title=entry.title, completed=entry.completed, url=url_for("entry", entry_id=entry.id)) @app.teardown_appcontext def shutdown_session(exception=None): db_session.remove()
from todo import app from flask import jsonify, request, url_for from flask import json from todo.database import db_session from todo.models import Entry @app.route("/", methods=["GET", "POST", "DELETE"]) def index(): if request.method == "POST": request_json = request.get_json() entry = Entry(request_json["title"]) db_session.add(entry) db_session.commit() return jsonify(construct_dict(entry)) else: if request.method == "DELETE": Entry.query.delete() db_session.commit() response = [] for entry in Entry.query.all(): response.append(construct_dict(entry)) return json.dumps(response) @app.route("/<int:entry_id>") def entry(entry_id): return jsonify(construct_dict(Entry.query.filter(Entry.id == entry_id).first())) def construct_dict(entry): return dict(title=entry.title, completed=entry.completed, url=url_for("entry", entry_id=entry.id)) @app.teardown_appcontext def shutdown_session(exception=None): db_session.remove()
Revert "Adding request context for proper url generation."
Revert "Adding request context for proper url generation." This reverts commit 3fa12f6b36f7d1d0dd23cf28e79b7c54f1589fbc.
Python
mit
Faerbit/todo-backend-flask
4be7f694220ee969683f07b982f8fcbe61971a04
hairball/plugins/duplicate.py
hairball/plugins/duplicate.py
"""This module provides plugins for basic duplicate code detection.""" from hairball.plugins import HairballPlugin class DuplicateScripts(HairballPlugin): """Plugin that keeps track of which scripts have been used more than once whithin a project.""" def __init__(self): super(DuplicateScripts, self).__init__() self.total_duplicate = 0 self.list_duplicate = [] def finalize(self): """Output the duplicate scripts detected.""" if self.total_duplicate > 0: print("%d duplicate scripts found" % self.total_duplicate) for duplicate in self.list_duplicate: print duplicate def analyze(self, scratch): """Run and return the results from the DuplicateChecks plugin.""" scripts_set = set() for script in self.iter_scripts(scratch): blocks_list = [] for name, _, _ in self.iter_blocks(script.blocks): blocks_list.append(name) blocks_tuple = tuple(blocks_list) if blocks_tuple in scripts_set: if len(blocks_list)>3: self.total_duplicate += 1 self.list_duplicate.append(blocks_list) else: scripts_set.add(blocks_tuple)
"""This module provides plugins for basic duplicate code detection.""" from hairball.plugins import HairballPlugin class DuplicateScripts(HairballPlugin): """Plugin that keeps track of which scripts have been used more than once whithin a project.""" def __init__(self): super(DuplicateScripts, self).__init__() self.total_duplicate = 0 self.list_duplicate = [] def finalize(self): """Output the duplicate scripts detected.""" if self.total_duplicate > 0: print("%d duplicate scripts found" % self.total_duplicate) for duplicate in self.list_duplicate: print duplicate def analyze(self, scratch): """Run and return the results from the DuplicateChecks plugin. Only takes into account scripts with more than 3 blocks""" scripts_set = set() for script in self.iter_scripts(scratch): blocks_list = [] for name, _, _ in self.iter_blocks(script.blocks): blocks_list.append(name) blocks_tuple = tuple(blocks_list) if blocks_tuple in scripts_set: if len(blocks_list)>3: self.total_duplicate += 1 self.list_duplicate.append(blocks_list) else: scripts_set.add(blocks_tuple)
Add comment to explain the length of the scripts taken into account in DuplicateScripts
Add comment to explain the length of the scripts taken into account in DuplicateScripts
Python
bsd-2-clause
ucsb-cs-education/hairball,jemole/hairball,thsunmy/hairball,jemole/hairball,ucsb-cs-education/hairball,thsunmy/hairball
15996286496d913c25290362ba2dba2d349bd5f6
imageManagerUtils/settings.py
imageManagerUtils/settings.py
# Copyright (c) 2017, MIT Licensed, Medicine Yeh # This file helps to read settings from bash script into os.environ import os import sys import subprocess # This path is the location of the caller script MAIN_SCRIPT_PATH = os.path.dirname(os.path.abspath(sys.argv[0])) # Set up the path to settings.sh settings_path = os.path.join(MAIN_SCRIPT_PATH, 'settings.sh') if not os.path.isfile(settings_path): print('Cannot find settings.sh in ' + MAIN_SCRIPT_PATH) exit(1) # This is a tricky way to read bash envs in the script env_str = subprocess.check_output('source {} && env'.format(settings_path), shell=True) # Transform to list of python strings (utf-8 encodings) env_str = env_str.decode('utf-8').split('\n') # Transform from a list to a list of pairs and filter out invalid formats env_list = [kv.split('=') for kv in env_str if len(kv.split('=')) == 2] # Transform from a list to a dictionary env_dict = {kv[0]: kv[1] for kv in env_list} # Update the os.environ globally os.environ.update(env_dict)
# Copyright (c) 2017, MIT Licensed, Medicine Yeh # This file helps to read settings from bash script into os.environ import os import sys import subprocess # This path is the location of the caller script MAIN_SCRIPT_PATH = os.path.dirname(os.path.abspath(sys.argv[0])) # Set up the path to settings.sh settings_path = os.path.join(MAIN_SCRIPT_PATH, 'settings.sh') if not os.path.isfile(settings_path): print('Cannot find settings.sh in ' + MAIN_SCRIPT_PATH) exit(1) # This is a tricky way to read bash envs in the script env_str = subprocess.check_output('source {} && env'.format(settings_path), shell=True, executable='/bin/bash') # Transform to list of python strings (utf-8 encodings) env_str = env_str.decode('utf-8').split('\n') # Transform from a list to a list of pairs and filter out invalid formats env_list = [kv.split('=') for kv in env_str if len(kv.split('=')) == 2] # Transform from a list to a dictionary env_dict = {kv[0]: kv[1] for kv in env_list} # Update the os.environ globally os.environ.update(env_dict)
Fix bug of invoking /bin/sh on several OSs
Fix bug of invoking /bin/sh on several OSs
Python
mit
snippits/qemu_image,snippits/qemu_image,snippits/qemu_image
c027e671d1a47d485755b748f2dffc202c704ff8
goodreadsapi.py
goodreadsapi.py
#!/usr/bin/env python import re from xml.parsers.expat import ExpatError import requests import xmltodict from settings import goodreads_api_key def get_goodreads_ids(comment_msg): # receives goodreads url # returns the id using regex regex = r'goodreads.com/book/show/(\d+)' return set(re.findall(regex, comment_msg)) def get_book_details_by_id(goodreads_id): api_url = 'http://goodreads.com/book/show/{0}?format=xml&key={1}' r = requests.get(api_url.format(goodreads_id, goodreads_api_key)) try: book_data = xmltodict.parse(r.content)['GoodreadsResponse']['book'] except (TypeError, KeyError, ExpatError): return False keys = ['title', 'average_rating', 'ratings_count', 'description', 'num_pages', 'publication_year'] book = {} for k in keys: book[k] = book_data.get(k) if type(book_data['authors']['author']) == list: authors = [author['name'] for author in book_data['authors']['author']] authors = ', '.join(authors) else: authors = book_data['authors']['author']['name'] book['authors'] = authors return book
#!/usr/bin/env python import re from xml.parsers.expat import ExpatError import requests import xmltodict from settings import goodreads_api_key def get_goodreads_ids(comment_msg): # receives goodreads url # returns the id using regex regex = r'goodreads.com/book/show/(\d+)' return set(re.findall(regex, comment_msg)) def get_book_details_by_id(goodreads_id): api_url = 'http://goodreads.com/book/show/{0}?format=xml&key={1}' r = requests.get(api_url.format(goodreads_id, goodreads_api_key)) try: book_data = xmltodict.parse(r.content)['GoodreadsResponse']['book'] except (TypeError, KeyError, ExpatError): return False keys = ['title', 'average_rating', 'ratings_count', 'description', 'num_pages'] book = {} for k in keys: book[k] = book_data.get(k) try: work = book_data['work'] book['publication_year'] = work['original_publication_year']['#text'] except KeyError: book['publication_year'] = book_data.get('publication_year') if type(book_data['authors']['author']) == list: authors = [author['name'] for author in book_data['authors']['author']] authors = ', '.join(authors) else: authors = book_data['authors']['author']['name'] book['authors'] = authors return book
Update goodreads API to `show original_publication_year`
Update goodreads API to `show original_publication_year`
Python
mit
avinassh/Reddit-GoodReads-Bot
59b015bb3e45497b7ec86bf1799e8442a30b65da
py/PMUtil.py
py/PMUtil.py
# PMUtil.py # Phenotype microarray utility functions # # Author: Daniel A Cuevas # Created on 27 Jan. 2015 # Updated on 27 Jan. 2015 from __future__ import absolute_import, division, print_function import sys import time import datetime def timeStamp(): '''Return time stamp''' t = time.time() fmt = '[%Y-%m-%d %H:%M:%S]' return datetime.datetime.fromtimestamp(t).strftime(fmt) def printStatus(msg): '''Print status message''' print('{} {}'.format(timeStamp(), msg), file=sys.stderr) sys.stderr.flush()
# PMUtil.py # Phenotype microarray utility functions # # Author: Daniel A Cuevas # Created on 27 Jan 2015 # Updated on 20 Aug 2015 from __future__ import absolute_import, division, print_function import sys import time import datetime def timeStamp(): '''Return time stamp''' t = time.time() fmt = '[%Y-%m-%d %H:%M:%S]' return datetime.datetime.fromtimestamp(t).strftime(fmt) def printStatus(msg): '''Print status message''' print('{} {}'.format(timeStamp(), msg), file=sys.stderr) sys.stderr.flush() def exitScript(num=1): '''Exit script''' sys.exit(num)
Exit method. - (New) Added exit method.
Exit method. - (New) Added exit method.
Python
mit
dacuevas/PMAnalyzer,dacuevas/PMAnalyzer,dacuevas/PMAnalyzer,dacuevas/PMAnalyzer
a8976ff1c3bdc177ca72becf48c4278f963d2627
gtr/__init__.py
gtr/__init__.py
__all__ = [ "gtr.services.funds.Funds", "gtr.services.organisations.Organisations", "gtr.services.persons.Persons", "gtr.services.projects.Projects" ] __version__ = "0.1.0" from gtr.services.base import _Service from gtr.services.funds import Funds from gtr.services.organisations import Organisations from gtr.services.persons import Persons from gtr.services.projects import Projects
__all__ = [ "gtr.services.funds.Funds", "gtr.services.organisations.Organisations", "gtr.services.persons.Persons", "gtr.services.projects.Projects", "gtr.services.publications.Publications" ] __version__ = "0.1.0" from gtr.services.base import _Service from gtr.services.funds import Funds from gtr.services.organisations import Organisations from gtr.services.persons import Persons from gtr.services.projects import Projects from gtr.services.publications import Publications
Add Publications class to initialisation
Add Publications class to initialisation
Python
apache-2.0
nestauk/gtr
63a26cbf76a3d0135f5b67dd10cc7f383ffa7ebf
helusers/jwt.py
helusers/jwt.py
from django.conf import settings from rest_framework_jwt.authentication import JSONWebTokenAuthentication from rest_framework_jwt.settings import api_settings from .user_utils import get_or_create_user def patch_jwt_settings(): """Patch rest_framework_jwt authentication settings from allauth""" defaults = api_settings.defaults defaults['JWT_PAYLOAD_GET_USER_ID_HANDLER'] = ( __name__ + '.get_user_id_from_payload_handler') if 'allauth.socialaccount' not in settings.INSTALLED_APPS: return from allauth.socialaccount.models import SocialApp try: app = SocialApp.objects.get(provider='helsinki') except SocialApp.DoesNotExist: return defaults['JWT_SECRET_KEY'] = app.secret defaults['JWT_AUDIENCE'] = app.client_id # Disable automatic settings patching for now because it breaks Travis. # patch_jwt_settings() class JWTAuthentication(JSONWebTokenAuthentication): def authenticate_credentials(self, payload): return get_or_create_user(payload) def get_user_id_from_payload_handler(payload): return payload.get('sub')
from django.conf import settings from rest_framework import exceptions from rest_framework_jwt.authentication import JSONWebTokenAuthentication from rest_framework_jwt.settings import api_settings from .user_utils import get_or_create_user def patch_jwt_settings(): """Patch rest_framework_jwt authentication settings from allauth""" defaults = api_settings.defaults defaults['JWT_PAYLOAD_GET_USER_ID_HANDLER'] = ( __name__ + '.get_user_id_from_payload_handler') if 'allauth.socialaccount' not in settings.INSTALLED_APPS: return from allauth.socialaccount.models import SocialApp try: app = SocialApp.objects.get(provider='helsinki') except SocialApp.DoesNotExist: return defaults['JWT_SECRET_KEY'] = app.secret defaults['JWT_AUDIENCE'] = app.client_id # Disable automatic settings patching for now because it breaks Travis. # patch_jwt_settings() class JWTAuthentication(JSONWebTokenAuthentication): def authenticate_credentials(self, payload): user = super().authenticate_credentials(payload) if user and not user.is_active: msg = _('User account is disabled.') raise exceptions.AuthenticationFailed(msg) return get_or_create_user(payload) def get_user_id_from_payload_handler(payload): return payload.get('sub')
Change authenticate_credentials method to raise an exception if the account is disabled
Change authenticate_credentials method to raise an exception if the account is disabled
Python
bsd-2-clause
City-of-Helsinki/django-helusers,City-of-Helsinki/django-helusers
764f8d9d7818076555cde5fcad29f3052b523771
company/autocomplete_light_registry.py
company/autocomplete_light_registry.py
import autocomplete_light from .models import Company class CompanyAutocomplete(autocomplete_light.AutocompleteModelBase): search_fields = ['^name'] model = Company autocomplete_light.register(CompanyAutocomplete)
import autocomplete_light from .models import Company class CompanyAutocomplete(autocomplete_light.AutocompleteModelBase): search_fields = ['name', 'official_name', 'common_name'] model = Company autocomplete_light.register(CompanyAutocomplete)
Add more search fields to autocomplete
Add more search fields to autocomplete
Python
bsd-3-clause
KlubJagiellonski/pola-backend,KlubJagiellonski/pola-backend,KlubJagiellonski/pola-backend,KlubJagiellonski/pola-backend
a06010fcb2f4424d085da1487a6666867a8cbf5b
dbaas/maintenance/admin/maintenance.py
dbaas/maintenance/admin/maintenance.py
# -*- coding: utf-8 -*- from __future__ import absolute_import, unicode_literals from django_services import admin from ..models import Maintenance from ..service.maintenance import MaintenanceService class MaintenanceAdmin(admin.DjangoServicesAdmin): service_class = MaintenanceService search_fields = ("scheduled_for", "description", "maximum_workers", 'status') list_display = ("scheduled_for", "description", "maximum_workers", 'status') fields = ( "description", "scheduled_for", "main_script", "rollback_script", "host_query","maximum_workers", "status", "celery_task_id",) save_on_top = True readonly_fields = ('status', 'celery_task_id') def change_view(self, request, object_id, form_url='', extra_context=None): maintenance = Maintenance.objects.get(id=object_id) if maintenance.celery_task_id: self.readonly_fields = self.fields return super(MaintenanceAdmin, self).change_view(request, object_id, form_url, extra_context=extra_context) def add_view(self, request, form_url='', extra_context=None): return super(MaintenanceAdmin, self).add_view(request, form_url, extra_context)
# -*- coding: utf-8 -*- from __future__ import absolute_import, unicode_literals from django_services import admin from ..models import Maintenance from ..service.maintenance import MaintenanceService from ..forms import MaintenanceForm class MaintenanceAdmin(admin.DjangoServicesAdmin): service_class = MaintenanceService search_fields = ("scheduled_for", "description", "maximum_workers", 'status') list_display = ("scheduled_for", "description", "maximum_workers", 'status') fields = ( "description", "scheduled_for", "main_script", "rollback_script", "host_query","maximum_workers", "status", "celery_task_id",) save_on_top = True readonly_fields = ('status', 'celery_task_id') form = MaintenanceForm def change_view(self, request, object_id, form_url='', extra_context=None): maintenance = Maintenance.objects.get(id=object_id) if maintenance.celery_task_id: self.readonly_fields = self.fields return super(MaintenanceAdmin, self).change_view(request, object_id, form_url, extra_context=extra_context)
Remove add_view and add form for the hole admin
Remove add_view and add form for the hole admin
Python
bsd-3-clause
globocom/database-as-a-service,globocom/database-as-a-service,globocom/database-as-a-service,globocom/database-as-a-service
6f822cf46957d038588e7a71eb91f8ca9f9c95f1
scaffolder/commands/install.py
scaffolder/commands/install.py
#!/usr/bin/env python # -*- coding: utf-8 -*- from optparse import make_option from optparse import OptionParser from scaffolder.core.template import TemplateManager from scaffolder.core.commands import BaseCommand class InstallCommand(BaseCommand): option_list = BaseCommand.option_list + ( make_option( "-t", "--target", dest="target_dir", default='~/.cookiejar', help='Project Templates directory.', metavar="TEMPLATES_DIR" ), ) def __init__(self, name, help='', aliases=(), stdout=None, stderr=None): help = 'install: Installs a Project Template.' parser = OptionParser( version=self.get_version(), option_list=self.get_option_list(), usage='\n %prog {0} ACTION [OPTIONS]'.format(name) ) aliases = ('tmp',) BaseCommand.__init__(self, name, parser=parser, help=help, aliases=aliases) def run(self, *args, **options): src = args[0] tgt = options.get('target_dir') manager = TemplateManager() manager.install(src=src, dest=tgt)
#!/usr/bin/env python # -*- coding: utf-8 -*- from optparse import make_option from optparse import OptionParser from scaffolder import get_minion_path from scaffolder.core.template import TemplateManager from scaffolder.core.commands import BaseCommand class InstallCommand(BaseCommand): option_list = BaseCommand.option_list + ( make_option( "-t", "--target", dest="target_dir", default=get_minion_path('weaver'), help='Project Templates directory.', metavar="TEMPLATES_DIR" ), ) def __init__(self, name, help='', aliases=(), stdout=None, stderr=None): help = 'install: Installs a Project Template.' parser = OptionParser( version=self.get_version(), option_list=self.get_option_list(), usage='\n %prog {0} ACTION [OPTIONS]'.format(name) ) aliases = ('tmp',) BaseCommand.__init__(self, name, parser=parser, help=help, aliases=aliases) def run(self, *args, **options): src = args[0] tgt = options.get('target_dir') manager = TemplateManager() manager.install(src=src, dest=tgt)
Use get_minion_path to get default dir.
InstallCommand: Use get_minion_path to get default dir.
Python
mit
goliatone/minions
95d9bb3a9500d80b5064c5fb4d5bd7b30406d1ae
conanfile.py
conanfile.py
from conans import ConanFile, CMake class GrpccbConan(ConanFile): name = "grpc_cb_core" version = "0.2" license = "Apache-2.0" url = "https://github.com/jinq0123/grpc_cb_core" description = "C++ gRPC core library with callback interface." settings = "os", "compiler", "build_type", "arch" options = {"shared": [True, False]} default_options = "shared=False" requires = "grpc/1.17.2@inexorgame/stable", generators = "cmake", "Premake" # A custom generator: PremakeGen/0.1@memsharded/testing build_requires = "PremakeGen/0.1@memsharded/testing" exports_sources = "src*", "include*", "CMakeLists.txt" def build(self): cmake = CMake(self) self.run('cmake %s %s' % (self.source_folder, cmake.command_line)) self.run("cmake --build . %s" % cmake.build_config) def package(self): self.copy("include/*") self.copy("*.lib", dst="lib", keep_path=False) self.copy("*.dll", dst="bin", keep_path=False) self.copy("*.dylib*", dst="lib", keep_path=False) self.copy("*.so", dst="lib", keep_path=False) self.copy("*.a", dst="lib", keep_path=False) def package_info(self): self.cpp_info.libs = ["grpc_cb_core"]
from conans import ConanFile, CMake class GrpccbConan(ConanFile): name = "grpc_cb_core" version = "0.2" license = "Apache-2.0" url = "https://github.com/jinq0123/grpc_cb_core" description = "C++ gRPC core library with callback interface." settings = "os", "compiler", "build_type", "arch" options = {"shared": [True, False]} default_options = "shared=False" requires = "grpc/1.44.0@", generators = "cmake", "premake" # The builtin premake generator exports_sources = "src*", "include*", "CMakeLists.txt" def build(self): cmake = CMake(self) self.run('cmake %s %s' % (self.source_folder, cmake.command_line)) self.run("cmake --build . %s" % cmake.build_config) def package(self): self.copy("include/*") self.copy("*.lib", dst="lib", keep_path=False) self.copy("*.dll", dst="bin", keep_path=False) self.copy("*.dylib*", dst="lib", keep_path=False) self.copy("*.so", dst="lib", keep_path=False) self.copy("*.a", dst="lib", keep_path=False) def package_info(self): self.cpp_info.libs = ["grpc_cb_core"]
Fix update remote to ConanCenter and grpc to highest buildable/supported version
Fix update remote to ConanCenter and grpc to highest buildable/supported version
Python
apache-2.0
jinq0123/grpc_cb_core,jinq0123/grpc_cb_core,jinq0123/grpc_cb_core
c13a12e6355423d6756b8b514942596c31b0e3a9
conanfile.py
conanfile.py
from conans import ConanFile from conans.tools import download, unzip import os VERSION = "0.0.7" class CMakeModuleCommonConan(ConanFile): name = "cmake-module-common" version = os.environ.get("CONAN_VERSION_OVERRIDE", VERSION) generators = "cmake" url = "http://github.com/polysquare/cmake-module-common" license = "MIT" def source(self): zip_name = "cmake-module-common.zip" download("https://github.com/polysquare/" "cmake-module-common/archive/{version}.zip" "".format(version="v" + VERSION), zip_name) unzip(zip_name) os.unlink(zip_name) def package(self): self.copy(pattern="Find*.cmake", dst="", src="cmake-module-common-" + VERSION, keep_path=True) self.copy(pattern="*.cmake", dst="cmake/cmake-module-common", src="cmake-module-common-" + VERSION, keep_path=True)
from conans import ConanFile from conans.tools import download, unzip import os VERSION = "0.0.7" class CMakeModuleCommonConan(ConanFile): name = "cmake-module-common" version = os.environ.get("CONAN_VERSION_OVERRIDE", VERSION) generators = "cmake" url = "http://github.com/polysquare/cmake-module-common" license = "MIT" requires = ("cmake-unit/master@smspillaz/cmake-unit", "cmake-linter-cmake/master@smspillaz/cmake-linter-cmake", "style-linter-cmake/master@smspillaz/style-linter-cmake") def source(self): zip_name = "cmake-module-common.zip" download("https://github.com/polysquare/" "cmake-module-common/archive/{version}.zip" "".format(version="v" + VERSION), zip_name) unzip(zip_name) os.unlink(zip_name) def package(self): self.copy(pattern="Find*.cmake", dst="", src="cmake-module-common-" + VERSION, keep_path=True) self.copy(pattern="*.cmake", dst="cmake/cmake-module-common", src="cmake-module-common-" + VERSION, keep_path=True)
Make cmake-unit, cmake-linter-cmake and style-linter-cmake normal deps
conan: Make cmake-unit, cmake-linter-cmake and style-linter-cmake normal deps
Python
mit
polysquare/cmake-module-common
306e6939c5b369f4a4ef4bb4d16948dc1f027f53
tests/test_initial_ismaster.py
tests/test_initial_ismaster.py
# Copyright 2015 MongoDB, Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import time from mockupdb import MockupDB, wait_until from pymongo import MongoClient from tests import unittest class TestInitialIsMaster(unittest.TestCase): def test_initial_ismaster(self): server = MockupDB() server.run() self.addCleanup(server.stop) start = time.time() client = MongoClient(server.uri) self.addCleanup(client.close) # A single ismaster is enough for the client to be connected. self.assertIsNone(client.address) server.receives('ismaster').ok() wait_until(lambda: client.address is not None, 'update address', timeout=1) # At least 10 seconds before next heartbeat. server.receives('ismaster').ok() self.assertGreaterEqual(time.time() - start, 10) if __name__ == '__main__': unittest.main()
# Copyright 2015 MongoDB, Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import time from mockupdb import MockupDB, wait_until from pymongo import MongoClient from tests import unittest class TestInitialIsMaster(unittest.TestCase): def test_initial_ismaster(self): server = MockupDB() server.run() self.addCleanup(server.stop) start = time.time() client = MongoClient(server.uri) self.addCleanup(client.close) # A single ismaster is enough for the client to be connected. self.assertFalse(client.nodes) server.receives('ismaster').ok(ismaster=True) wait_until(lambda: client.nodes, 'update nodes', timeout=1) # At least 10 seconds before next heartbeat. server.receives('ismaster').ok(ismaster=True) self.assertGreaterEqual(time.time() - start, 10) if __name__ == '__main__': unittest.main()
Update for PYTHON 985: MongoClient properties now block until connected.
Update for PYTHON 985: MongoClient properties now block until connected.
Python
apache-2.0
ajdavis/pymongo-mockup-tests
af5e90cb544e2e37819302f5750084fc17f7ee12
make_example.py
make_example.py
#!/usr/bin/env python import os import sys import yaml import subprocess class SDBUSPlus(object): def __init__(self, path): self.path = path def __call__(self, *a, **kw): args = [ os.path.join(self.path, 'sdbus++'), '-t', os.path.join(self.path, 'templates') ] subprocess.call(args + list(a), **kw) if __name__ == '__main__': sdbusplus = None for p in os.environ.get('PATH', "").split(os.pathsep): if os.path.exists(os.path.join(p, 'sdbus++')): sdbusplus = SDBUSPlus(p) break if sdbusplus is None: sys.stderr.write('Cannot find sdbus++\n') sys.exit(1) genfiles = { 'server-cpp': lambda x: '%s.cpp' % x, 'server-header': lambda x: os.path.join( os.path.join(*x.split('.')), 'server.hpp') } with open(os.path.join('example', 'interfaces.yaml'), 'r') as fd: interfaces = yaml.load(fd.read()) for i in interfaces: for process, f in genfiles.iteritems(): dest = f(i) parent = os.path.dirname(dest) if parent and not os.path.exists(parent): os.makedirs(parent) with open(dest, 'w') as fd: sdbusplus( '-r', os.path.join('example', 'interfaces'), 'interface', process, i, stdout=fd) # vim: tabstop=8 expandtab shiftwidth=4 softtabstop=4
#!/usr/bin/env python import os import sys import yaml import subprocess if __name__ == '__main__': genfiles = { 'server-cpp': lambda x: '%s.cpp' % x, 'server-header': lambda x: os.path.join( os.path.join(*x.split('.')), 'server.hpp') } with open(os.path.join('example', 'interfaces.yaml'), 'r') as fd: interfaces = yaml.load(fd.read()) for i in interfaces: for process, f in genfiles.iteritems(): dest = f(i) parent = os.path.dirname(dest) if parent and not os.path.exists(parent): os.makedirs(parent) with open(dest, 'w') as fd: subprocess.call([ 'sdbus++', '-r', os.path.join('example', 'interfaces'), 'interface', process, i], stdout=fd) # vim: tabstop=8 expandtab shiftwidth=4 softtabstop=4
Remove sdbus++ template search workaround
Remove sdbus++ template search workaround sdbus++ was fixed upstream to find its templates automatically. Change-Id: I29020b9d1ea4ae8baaca5fe869625a3d96cd6eaf Signed-off-by: Brad Bishop <[email protected]>
Python
apache-2.0
openbmc/phosphor-inventory-manager,openbmc/phosphor-inventory-manager
1e07e9424a1ac69e1e660e6a6f1e58bba15472c1
make_spectra.py
make_spectra.py
# -*- coding: utf-8 -*- import halospectra as hs import randspectra as rs import sys snapnum=sys.argv[1] sim=sys.argv[2] #base="/n/hernquistfs1/mvogelsberger/projects/GFM/Production/Cosmo/Cosmo"+str(sim)+"_V6/L25n512/output/" #savedir="/n/home11/spb/scratch/Cosmo/Cosmo"+str(sim)+"_V6_512/snapdir_"+str(snapnum).rjust(3,'0') base="/home/spb/data/Cosmo/Cosmo"+str(sim)+"_V6/L25n256" savedir="/home/spb/scratch/Cosmo/Cosmo"+str(sim)+"_V6/snapdir_"+str(snapnum).rjust(3,'0') #halo = hs.HaloSpectra(snapnum, base,3, savefile="halo_spectra_DLA.hdf5", savedir=savedir) halo = rs.RandSpectra(snapnum, base,numlos=3000,savedir=savedir, savefile="rand_spectra_DLA.hdf5") halo.get_tau("Si",2,2) halo.get_tau("H",1,1) halo.get_col_density("Z",-1) halo.get_col_density("H",-1) halo.save_file()
# -*- coding: utf-8 -*- import halospectra as hs import randspectra as rs import sys snapnum=sys.argv[1] sim=sys.argv[2] #base="/n/hernquistfs1/mvogelsberger/projects/GFM/Production/Cosmo/Cosmo"+str(sim)+"_V6/L25n512/output/" #savedir="/n/home11/spb/scratch/Cosmo/Cosmo"+str(sim)+"_V6_512/snapdir_"+str(snapnum).rjust(3,'0') base="/home/spb/data/Cosmo/Cosmo"+str(sim)+"_V6/L25n256" savedir="/home/spb/scratch/Cosmo/Cosmo"+str(sim)+"_V6/snapdir_"+str(snapnum).rjust(3,'0') #halo = hs.HaloSpectra(snapnum, base,3, savefile="halo_spectra_DLA.hdf5", savedir=savedir) halo = rs.RandSpectra(snapnum, base,numlos=10000,savedir=savedir, savefile="rand_spectra.hdf5") #halo.get_observer_tau("Si",2) halo.get_tau("H",1,1) #halo.get_col_density("Z",-1) #halo.get_col_density("H",-1) halo.save_file()
Implement saving and loading the observer tau
Implement saving and loading the observer tau
Python
mit
sbird/vw_spectra
8316a60ba2887a511579e8cedb90b3a02fc1889a
dope/util.py
dope/util.py
from uuid import UUID from werkzeug.routing import BaseConverter class UUIDConverter(BaseConverter): to_python = UUID to_url = str
from uuid import UUID from werkzeug.routing import BaseConverter class UUIDConverter(BaseConverter): to_python = UUID def to_url(self, obj): return str(obj).replace('-', '')
Drop dashes from download urls.
Drop dashes from download urls.
Python
mit
mbr/dope,mbr/dope
9d46df1680e3d799971e73ec73043c2a6c0590ce
scripts/build_tar.py
scripts/build_tar.py
#! /usr/bin/python import os import subprocess root_dir = os.path.abspath(os.path.join(os.path.dirname(__file__), "..")) tarfile = os.path.join(root_dir, "src_pkg.tar") def _is_dir_newer(directory, filename): file_mtime = os.stat(filename).st_mtime for dirname, _, filenames in os.walk(directory): for filename in filenames: if filename.endswith(".pyc"): continue if _is_file_newer(os.path.join(dirname, filename), file_mtime): return True return False def _is_file_newer(filename, file_mtime): return os.stat(filename).st_mtime > file_mtime def _tar(): if 0 != subprocess.call("tar cvf {0} flask_app manage.py static".format(tarfile), shell=True, cwd=root_dir): raise Exception("Tar failed") if __name__ == '__main__': if not os.path.exists(tarfile) or \ _is_dir_newer(os.path.join(root_dir, "flask_app"), tarfile) or \ _is_dir_newer(os.path.join(root_dir, "static"), tarfile) or \ _is_file_newer(os.path.join(root_dir, "manage.py"), os.stat(tarfile).st_mtime): _tar()
#! /usr/bin/python import os import subprocess root_dir = os.path.abspath(os.path.join(os.path.dirname(__file__), "..")) tarfile = os.path.join(root_dir, "src_pkg.tar") def _is_dir_newer(directory, filename): file_mtime = os.stat(filename).st_mtime for dirname, _, filenames in os.walk(directory): if _is_file_newer(dirname, file_mtime): return True for filename in filenames: if filename.endswith(".pyc"): continue if _is_file_newer(os.path.join(dirname, filename), file_mtime): return True return False def _is_file_newer(filename, file_mtime): returned = os.stat(filename).st_mtime > file_mtime return returned def _tar(): if 0 != subprocess.call("tar cvf {0} flask_app manage.py static".format(tarfile), shell=True, cwd=root_dir): raise Exception("Tar failed") if __name__ == '__main__': if not os.path.exists(tarfile) or \ _is_dir_newer(os.path.join(root_dir, "flask_app"), tarfile) or \ _is_dir_newer(os.path.join(root_dir, "static"), tarfile) or \ _is_file_newer(os.path.join(root_dir, "manage.py"), os.stat(tarfile).st_mtime): _tar()
Fix building tar in deployment
Fix building tar in deployment
Python
bsd-3-clause
vmalloc/mailboxer,Infinidat/lanister,vmalloc/mailboxer,Infinidat/lanister,getslash/mailboxer,vmalloc/mailboxer,getslash/mailboxer,getslash/mailboxer
a1390619619a364b9fab13504fb5c2464491d449
Largest_Palindrome_Product.py
Largest_Palindrome_Product.py
# Find the largest palindrome made from the product of two n-digit numbers. # Since the result could be very large, you should return the largest palindrome mod 1337. # Example: # Input: 2 # Output: 987 # Explanation: 99 x 91 = 9009, 9009 % 1337 = 987 # Note: # The range of n is [1,8]. def largestPalindrome(n): """ :type n: int :rtype: int """ number = "" for x in range(n): number += "9" minNum = int(number[:-1]) number = int(number) palindrome = 0 for x in range(number, minNum, -2): if (x**2) < palindrome: break for i in range(number, x - 1, -2): product = x * i if product <= palindrome or product % 11 != 0: break elif isPalindrome(product): palindrome = product print(palindrome, palindrome % 1337) break return (palindrome, palindrome % 1337) def isPalindrome(num): """ Return True is number is Palindrome, else return False """ numString = str(num) if numString == numString[::-1]: return True return False n = 8 print(largestPalindrome(n)) # for i in range(upper, int((x*x)**.5), -2): # 990090099 152 99999 9901 99998 76865
# Find the largest palindrome made from the product of two n-digit numbers. # Since the result could be very large, you should return the largest palindrome mod 1337. # Example: # Input: 2 # Output: 987 # Explanation: 99 x 91 = 9009, 9009 % 1337 = 987 # Note: # The range of n is [1,8]. from itertools import product def largestPalindrome(n): """ :type n: int :rtype: int """ number = "" for x in range(n): number += "9" number = int(number) palindrome = 0 for x in range(number, 1, -2): if (x*x) < palindrome: break for i in range(number, x - 1, -2): product = x * i if product < palindrome: break elif isPalindrome(product): palindrome = product break return palindrome % 1337 def isPalindrome(num): """ Return True is number is Palindrome, else return False """ return str(num) == str(num)[::-1] n = 7 print(largestPalindrome(n))
Refactor Largest Palindrome Product for range of n is
Refactor Largest Palindrome Product for range of n is [1,8]
Python
mit
Kunal57/Python_Algorithms
de4af7935c1c8d6751c5a71ad90dd5f531f7a1b0
bin/trigger_upload.py
bin/trigger_upload.py
#!/bin/env python # -*- coding: utf8 -*- """ Triggers an upload process with the specified raw.xz URL. """ import argparse import logging import logging.config import multiprocessing.pool import fedmsg.config import fedimg.uploader logging.config.dictConfig(fedmsg.config.load_config()['logging']) log = logging.getLogger('fedmsg') def trigger_upload(compose_id, url, push_notifications): upload_pool = multiprocessing.pool.ThreadPool(processes=4) fedimg.uploader.upload(upload_pool, [url], compose_id=compose_id, push_notifications=push_notifications) def get_args(): parser = argparse.ArgumentParser( description="Trigger a manual upload process with the " "specified raw.xz URL") parser.add_argument( "-u", "--url", type=str, help=".raw.xz URL", required=True) parser.add_argument( "-c", "--compose-id", type=str, help="compose id of the .raw.xz file", required=True) parser.add_argument( "-p", "--push-notifications", help="Bool to check if we need to push fedmsg notifications", action="store_true", required=False) args = parser.parse_args() return args.url, args.compose_id, args.push_notifications def main(): url, compose_id, push_notifications = get_args() trigger_upload(url, compose_id, push_notifications) if __name__ == '__main__': main()
#!/bin/env python # -*- coding: utf8 -*- """ Triggers an upload process with the specified raw.xz URL. """ import argparse import logging import logging.config import multiprocessing.pool import fedmsg.config import fedimg.uploader logging.config.dictConfig(fedmsg.config.load_config()['logging']) log = logging.getLogger('fedmsg') def trigger_upload(url, compose_id, push_notifications): upload_pool = multiprocessing.pool.ThreadPool(processes=4) fedimg.uploader.upload(upload_pool, [url], compose_id=compose_id, push_notifications=push_notifications) def get_args(): parser = argparse.ArgumentParser( description="Trigger a manual upload process with the " "specified raw.xz URL") parser.add_argument( "-u", "--url", type=str, help=".raw.xz URL", required=True) parser.add_argument( "-c", "--compose-id", type=str, help="compose id of the .raw.xz file", required=True) parser.add_argument( "-p", "--push-notifications", help="Bool to check if we need to push fedmsg notifications", action="store_true", required=False) args = parser.parse_args() return args.url, args.compose_id, args.push_notifications def main(): url, compose_id, push_notifications = get_args() trigger_upload(url, compose_id, push_notifications) if __name__ == '__main__': main()
Fix the script function args
fedimg: Fix the script function args Signed-off-by: Sayan Chowdhury <[email protected]>
Python
agpl-3.0
fedora-infra/fedimg,fedora-infra/fedimg
166bff52496bfb47c5a3a03585bd10fb449b8d77
Lib/curses/__init__.py
Lib/curses/__init__.py
"""curses The main package for curses support for Python. Normally used by importing the package, and perhaps a particular module inside it. import curses from curses import textpad curses.initwin() ... """ __revision__ = "$Id$" from _curses import * from curses.wrapper import wrapper
"""curses The main package for curses support for Python. Normally used by importing the package, and perhaps a particular module inside it. import curses from curses import textpad curses.initwin() ... """ __revision__ = "$Id$" from _curses import * from curses.wrapper import wrapper # Some constants, most notably the ACS_* ones, are only added to the C # _curses module's dictionary after initscr() is called. (Some # versions of SGI's curses don't define values for those constants # until initscr() has been called.) This wrapper function calls the # underlying C initscr(), and then copies the constants from the # _curses module to the curses package's dictionary. Don't do 'from # curses import *' if you'll be needing the ACS_* constants. def initscr(): import _curses, curses stdscr = _curses.initscr() for key, value in _curses.__dict__.items(): if key[0:4] == 'ACS_' or key in ('LINES', 'COLS'): setattr(curses, key, value) return stdscr
Add wrapper for initscr() to copy the ACS_ and LINES,COLS bindings
Add wrapper for initscr() to copy the ACS_ and LINES,COLS bindings
Python
mit
sk-/python2.7-type-annotator,sk-/python2.7-type-annotator,sk-/python2.7-type-annotator
17faea99343e37036b7ee35e5d3273f98a52dba9
Python/tomviz/utils.py
Python/tomviz/utils.py
import numpy as np import vtk.numpy_interface.dataset_adapter as dsa def get_scalars(dataobject): do = dsa.WrapDataObject(dataobject) # get the first rawarray = do.PointData.GetScalars() vtkarray = dsa.vtkDataArrayToVTKArray(rawarray, do) vtkarray.Association = dsa.ArrayAssociation.POINT return vtkarray def set_scalars(dataobject, newscalars): do = dsa.WrapDataObject(dataobject) oldscalars = do.PointData.GetScalars() name = oldscalars.GetName() del oldscalars do.PointData.append(newscalars, name) do.PointData.SetActiveScalars(name)
import numpy as np import vtk.numpy_interface.dataset_adapter as dsa import vtk.util.numpy_support as np_s def get_scalars(dataobject): do = dsa.WrapDataObject(dataobject) # get the first rawarray = do.PointData.GetScalars() vtkarray = dsa.vtkDataArrayToVTKArray(rawarray, do) vtkarray.Association = dsa.ArrayAssociation.POINT return vtkarray def set_scalars(dataobject, newscalars): do = dsa.WrapDataObject(dataobject) oldscalars = do.PointData.GetScalars() name = oldscalars.GetName() del oldscalars # handle the case if the newscalars array has a type that # cannot be passed on to VTK. In which case, we convert to # convert to float64 vtk_typecode = np_s.get_vtk_array_type(newscalars.dtype) if vtk_typecode is None: newscalars = newscalars.astype(np.float64) do.PointData.append(newscalars, name) do.PointData.SetActiveScalars(name)
Fix numpy related errors on Mavericks.
Fix numpy related errors on Mavericks. The problem was due to the fact that operations (like sqrt) can return a float16 arrays which cannot be passed back to VTK directly. Added a temporary conversion to float64. We should potentially handle this in VTK.
Python
bsd-3-clause
cryos/tomviz,thewtex/tomviz,cjh1/tomviz,cryos/tomviz,cryos/tomviz,Hovden/tomviz,Hovden/tomviz,yijiang1/tomviz,cjh1/tomviz,thewtex/tomviz,OpenChemistry/tomviz,mathturtle/tomviz,yijiang1/tomviz,cjh1/tomviz,mathturtle/tomviz,OpenChemistry/tomviz,thewtex/tomviz,OpenChemistry/tomviz,mathturtle/tomviz,OpenChemistry/tomviz
98649d486b9e2eb2c83e594e73cf6bbaa29213e5
examples/simple_server.py
examples/simple_server.py
import argparse import math from pythonosc import dispatcher from pythonosc import osc_server def print_volume_handler(args, volume): print("[{0}] ~ {1}".format(args[0], volume)) def print_compute_handler(args, volume): try: print("[{0}] ~ {1}".format(args[0], args[1](volume))) except ValueError: pass if __name__ == "__main__": parser = argparse.ArgumentParser() parser.add_argument("--ip", default="127.0.0.1", help="The ip to listen on") parser.add_argument("--port", type=int, default=5005, help="The port to listen on") args = parser.parse_args() dispatcher = dispatcher.Dispatcher() dispatcher.map("/debug", print) dispatcher.map("/volume", print_volume_handler, "Volume") dispatcher.map("/logvolume", print_compute_handler, "Log volume", math.log) server = osc_server.ThreadingOSCUDPServer( (args.ip, args.port), dispatcher) print("Serving on {}".format(server.server_address)) server.serve_forever()
import argparse import math from pythonosc import dispatcher from pythonosc import osc_server def print_volume_handler(args, volume): print("[{0}] ~ {1}".format(args[0], volume)) def print_compute_handler(args, volume): try: print("[{0}] ~ {1}".format(args[0], args[1](volume))) except ValueError: pass if __name__ == "__main__": parser = argparse.ArgumentParser() parser.add_argument("--ip", default="0.0.0.0", help="The ip to listen on") parser.add_argument("--port", type=int, default=5005, help="The port to listen on") args = parser.parse_args() dispatcher = dispatcher.Dispatcher() dispatcher.map("/debug", print) dispatcher.map("/volume", print_volume_handler, "Volume") dispatcher.map("/logvolume", print_compute_handler, "Log volume", math.log) server = osc_server.ThreadingOSCUDPServer( (args.ip, args.port), dispatcher) print("Serving on {}".format(server.server_address)) server.serve_forever()
Make the server example listen on 0.0.0.0 by default.
Make the server example listen on 0.0.0.0 by default.
Python
unlicense
mwicat/python2-osc,attwad/python-osc,ragnarula/python-osc,emlyn/python-osc
8b77e1e865d72720a602b7b7cc5912cb852d68cf
settings/dev.py
settings/dev.py
# -*- coding: utf-8 -*- from __future__ import absolute_import, unicode_literals import os from .common import * # noqa DEBUG = True TEMPLATE_DEBUG = DEBUG DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': os.path.join(ROOT_DIR, 'db.sqlite3'), } } ACCOUNT_DEFAULT_HTTP_PROTOCOL = 'http' TEMPLATE_CONTEXT_PROCESSORS += ( "django.core.context_processors.debug", ) EMAIL_BACKEND = 'django.core.mail.backends.console.EmailBackend' INSTALLED_APPS += ('django_extensions',) # settings for celery BROKER_URL = os.environ.get("BROKER_URL", "redis://127.0.0.1:6379/0") CELERY_RESULT_BACKEND = os.environ.get("CELERY_RESULT_BACKEND", 'redis://127.0.0.1:6379/0')
# -*- coding: utf-8 -*- from __future__ import absolute_import, unicode_literals import os from .common import * # noqa DEBUG = True TEMPLATE_DEBUG = DEBUG DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': os.path.join(ROOT_DIR, 'db.sqlite3'), } } ACCOUNT_DEFAULT_HTTP_PROTOCOL = 'http' TEMPLATE_CONTEXT_PROCESSORS += ( "django.core.context_processors.debug", ) EMAIL_BACKEND = 'django.core.mail.backends.console.EmailBackend' INSTALLED_APPS += ('django_extensions',) # settings for celery BROKER_URL = os.environ.get("BROKER_URL", "redis://redis:6379/0") CELERY_RESULT_BACKEND = os.environ.get("CELERY_RESULT_BACKEND", 'redis://redis:6379/0')
Revert back to original settings for Celery Broker
Revert back to original settings for Celery Broker
Python
mit
pythonindia/junction,pythonindia/junction,pythonindia/junction,pythonindia/junction
e753038de039fd23f0d59bb0094f59fc73efe22b
flask_apscheduler/json.py
flask_apscheduler/json.py
import flask import json from datetime import datetime from apscheduler.job import Job from .utils import job_to_dict class JSONEncoder(json.JSONEncoder): def default(self, obj): if isinstance(obj, datetime): return obj.isoformat() if isinstance(obj, Job): return job_to_dict(obj) return super(JSONEncoder, self).default(obj) def dumps(obj, indent=None): return json.dumps(obj, indent=indent, cls=JSONEncoder) def jsonify(data, status=None): indent = None if flask.current_app.config['JSONIFY_PRETTYPRINT_REGULAR'] and not flask.request.is_xhr: indent = 2 return flask.current_app.response_class(dumps(data, indent=indent), status=status, mimetype='application/json')
import datetime import flask import json from apscheduler.job import Job from .utils import job_to_dict loads = json.loads def dumps(obj, indent=None): return json.dumps(obj, indent=indent, cls=JSONEncoder) def jsonify(data, status=None): indent = None if flask.current_app.config['JSONIFY_PRETTYPRINT_REGULAR'] and not flask.request.is_xhr: indent = 2 return flask.current_app.response_class(dumps(data, indent=indent), status=status, mimetype='application/json') class JSONEncoder(json.JSONEncoder): def default(self, obj): if isinstance(obj, datetime.datetime): return obj.isoformat() if isinstance(obj, datetime.date): return obj.isoformat() if isinstance(obj, Job): return job_to_dict(obj) return super(JSONEncoder, self).default(obj)
Set a custom JSON Encoder to serialize date class.
Set a custom JSON Encoder to serialize date class.
Python
apache-2.0
viniciuschiele/flask-apscheduler
edcfe2b156af23943478bc86592b4c8d5dc07e10
flask_mongoengine/json.py
flask_mongoengine/json.py
from flask.json import JSONEncoder from bson import json_util from mongoengine.base import BaseDocument from mongoengine import QuerySet def _make_encoder(superclass): class MongoEngineJSONEncoder(superclass): ''' A JSONEncoder which provides serialization of MongoEngine documents and querysets. ''' def default(self, obj): if isinstance(obj, BaseDocument): return json_util._json_convert(obj.to_mongo()) elif isinstance(obj, QuerySet): return json_util._json_convert(obj.as_pymongo()) return superclass.default(self, obj) return MongoEngineJSONEncoder MongoEngineJSONEncoder = _make_encoder(JSONEncoder) def overide_json_encoder(app): ''' A function to dynamically create a new MongoEngineJSONEncoder class based upon a custom base class. This function allows us to combine MongoEngine serialization with any changes to Flask's JSONEncoder which a user may have made prior to calling init_app. NOTE: This does not cover situations where users override an instance's json_encoder after calling init_app. ''' app.json_encoder = _make_encoder(app.json_encoder)
from flask.json import JSONEncoder from bson import json_util from mongoengine.base import BaseDocument try: from mongoengine.base import BaseQuerySet except ImportError as ie: # support mongoengine < 0.7 from mongoengine.queryset import QuerySet as BaseQuerySet def _make_encoder(superclass): class MongoEngineJSONEncoder(superclass): ''' A JSONEncoder which provides serialization of MongoEngine documents and queryset objects. ''' def default(self, obj): if isinstance(obj, BaseDocument): return json_util._json_convert(obj.to_mongo()) elif isinstance(obj, BaseQuerySet): return json_util._json_convert(obj.as_pymongo()) return superclass.default(self, obj) return MongoEngineJSONEncoder MongoEngineJSONEncoder = _make_encoder(JSONEncoder) def overide_json_encoder(app): ''' A function to dynamically create a new MongoEngineJSONEncoder class based upon a custom base class. This function allows us to combine MongoEngine serialization with any changes to Flask's JSONEncoder which a user may have made prior to calling init_app. NOTE: This does not cover situations where users override an instance's json_encoder after calling init_app. ''' app.json_encoder = _make_encoder(app.json_encoder)
Support older versions of MongoEngine
Support older versions of MongoEngine
Python
bsd-3-clause
gerasim13/flask-mongoengine-1,rochacbruno/flask-mongoengine,quokkaproject/flask-mongoengine,quokkaproject/flask-mongoengine,gerasim13/flask-mongoengine-1,losintikfos/flask-mongoengine,rochacbruno/flask-mongoengine,losintikfos/flask-mongoengine
3d7b5d61b7e985d409cd50c98d4bcbdc8ab9c723
mailer.py
mailer.py
from marrow.mailer import Mailer as MarrowMailer from message import Message import sys class Mailer: MAILER = MarrowMailer(dict(manager=dict(use='immediate'), transport=dict(use='sendmail'))) @staticmethod def send(message): Mailer.MAILER.send(message) @staticmethod def start(): Mailer.MAILER.start() @staticmethod def stop(): Mailer.MAILER.stop() @staticmethod def send_transactions(transactions, to_addr): Mailer.start() message = Message( to=to_addr, subject='New transactions', plain=repr(transactions) ) Mailer.send(message) Mailer.stop() @staticmethod def get_cli_email_addr(): try: return sys.argv[1] except IndexError: return None
from marrow.mailer import Mailer as MarrowMailer from message import Message import sys import os import pwd import socket class Mailer: MAILER = MarrowMailer(dict(manager=dict(use='immediate'), transport=dict(use='sendmail'))) DEFAULT_AUTHOR = pwd.getpwuid(os.getuid()).pw_name + '@' + socket.getfqdn() @staticmethod def send(message): Mailer.MAILER.send(message) @staticmethod def start(): Mailer.MAILER.start() @staticmethod def stop(): Mailer.MAILER.stop() @staticmethod def send_transactions(transactions, to_addr): Mailer.start() message = Message( author=Mailer.DEFAULT_AUTHOR, to=to_addr, subject='New transactions', plain=repr(transactions) ) Mailer.send(message) Mailer.stop() @staticmethod def get_cli_email_addr(): try: return sys.argv[1] except IndexError: return None
Use current user as email author
Use current user as email author
Python
isc
2mv/raapija
65973802a3e68e23f9a903937ef94f8afa277013
ibmcnx/doc/DataSources.py
ibmcnx/doc/DataSources.py
###### # Check ExId (GUID) by Email through JDBC # # Author: Christoph Stoettner # Mail: [email protected] # Documentation: http://scripting101.stoeps.de # # Version: 2.0 # Date: 2014-06-04 # # License: Apache 2.0 # # Check ExId of a User in all Connections Applications import ibmcnx.functions print AdminControl.getCell() cell = "/Cell:" + AdminControl.getCell() + "/" cellid = AdminConfig.getid( cell ) dbs = AdminConfig.list( 'DataSource', str(cellid) ) for db in dbs.splitlines().split('('): t1 = ibmcnx.functions.getDSId( db ) AdminConfig.list( t1 )
###### # Check ExId (GUID) by Email through JDBC # # Author: Christoph Stoettner # Mail: [email protected] # Documentation: http://scripting101.stoeps.de # # Version: 2.0 # Date: 2014-06-04 # # License: Apache 2.0 # # Check ExId of a User in all Connections Applications import ibmcnx.functions print AdminControl.getCell() cell = "/Cell:" + AdminControl.getCell() + "/" cellid = AdminConfig.getid( cell ) dbs = AdminConfig.list( 'DataSource', str(cellid) ) dbs = dbs.splitlines() print dbs for db in dbs.splitlines(): t1 = ibmcnx.functions.getDSId( db ) AdminConfig.list( t1 )
Create documentation of DataSource Settings
: Create documentation of DataSource Settings Task-Url:
Python
apache-2.0
stoeps13/ibmcnx2,stoeps13/ibmcnx2
93f2ff45ff3d61487ed061ae3d1a65051c3d1799
django/contrib/admin/__init__.py
django/contrib/admin/__init__.py
from django.contrib.admin.options import ModelAdmin, HORIZONTAL, VERTICAL from django.contrib.admin.options import StackedInline, TabularInline from django.contrib.admin.sites import AdminSite, site def autodiscover(): """ Auto-discover INSTALLED_APPS admin.py modules and fail silently when not present. This forces an import on them to register any admin bits they may want. """ import copy from django.conf import settings from django.utils.importlib import import_module from django.utils.module_loading import module_has_submodule for app in settings.INSTALLED_APPS: mod = import_module(app) # Attempt to import the app's admin module. try: before_import_registry = copy.copy(site._registry) import_module('%s.admin' % app) except: # Reset the model registry to the state before the last import as # this import will have to reoccur on the next request and this # could raise NotRegistered and AlreadyRegistered exceptions # (see #8245). site._registry = before_import_registry # Decide whether to bubble up this error. If the app just # doesn't have an admin module, we can ignore the error # attempting to import it, otherwise we want it to bubble up. if module_has_submodule(mod, 'admin'): raise
# ACTION_CHECKBOX_NAME is unused, but should stay since its import from here # has been referenced in documentation. from django.contrib.admin.helpers import ACTION_CHECKBOX_NAME from django.contrib.admin.options import ModelAdmin, HORIZONTAL, VERTICAL from django.contrib.admin.options import StackedInline, TabularInline from django.contrib.admin.sites import AdminSite, site def autodiscover(): """ Auto-discover INSTALLED_APPS admin.py modules and fail silently when not present. This forces an import on them to register any admin bits they may want. """ import copy from django.conf import settings from django.utils.importlib import import_module from django.utils.module_loading import module_has_submodule for app in settings.INSTALLED_APPS: mod = import_module(app) # Attempt to import the app's admin module. try: before_import_registry = copy.copy(site._registry) import_module('%s.admin' % app) except: # Reset the model registry to the state before the last import as # this import will have to reoccur on the next request and this # could raise NotRegistered and AlreadyRegistered exceptions # (see #8245). site._registry = before_import_registry # Decide whether to bubble up this error. If the app just # doesn't have an admin module, we can ignore the error # attempting to import it, otherwise we want it to bubble up. if module_has_submodule(mod, 'admin'): raise
Revert the removal of an unused import (in [14175]) that was referenced in documentation. Thanks for noticing, clong.
Revert the removal of an unused import (in [14175]) that was referenced in documentation. Thanks for noticing, clong. git-svn-id: 554f83ef17aa7291f84efa897c1acfc5d0035373@14359 bcc190cf-cafb-0310-a4f2-bffc1f526a37
Python
bsd-3-clause
svn2github/django,svn2github/django,svn2github/django
445f244ddac6001b65f03d058a14178a19919eed
diamondash/config.py
diamondash/config.py
import yaml from diamondash import utils class ConfigError(Exception): """Raised when there is an error parsing a configuration""" class ConfigMetaClass(type): def __new__(mcs, name, bases, dict): cls = type.__new__(mcs, name, bases, dict) defaults = {} for base in bases: if hasattr(base, 'DEFAULTS'): defaults.update(base.DEFAULTS) defaults.update(cls.DEFAULTS) cls.DEFAULTS = defaults return cls class Config(dict): __metaclass__ = ConfigMetaClass DEFAULTS = {} def __init__(self, items): super(Config, self).__init__(self._parse(items)) @classmethod def parse(cls, items): return items @classmethod def _parse(cls, items): items = utils.add_dicts(cls.DEFAULTS, items) return cls.parse(items) @classmethod def from_file(cls, filename, **defaults): items = utils.add_dicts(defaults, yaml.safe_load(open(filename))) return cls(items) @classmethod def for_type(cls, type_name): type_cls = utils.load_class_by_string(type_name) return type_cls.CONFIG_CLS
import yaml from diamondash import utils class ConfigError(Exception): """Raised when there is an error parsing a configuration""" class ConfigMetaClass(type): def __new__(mcs, name, bases, dict): cls = type.__new__(mcs, name, bases, dict) defaults = {} for base in bases: if hasattr(base, 'DEFAULTS'): defaults.update(base.DEFAULTS) defaults.update(cls.DEFAULTS) cls.DEFAULTS = defaults return cls class Config(dict): __metaclass__ = ConfigMetaClass DEFAULTS = {} def __init__(self, items=None): super(Config, self).__init__(self._parse(items or {})) @classmethod def parse(cls, items): return items @classmethod def _parse(cls, items): items = utils.add_dicts(cls.DEFAULTS, items) return cls.parse(items) @classmethod def from_file(cls, filename, **defaults): items = utils.add_dicts(defaults, yaml.safe_load(open(filename))) return cls(items) @classmethod def for_type(cls, type_name): type_cls = utils.load_class_by_string(type_name) return type_cls.CONFIG_CLS
Allow Config to be initialised without any args
Allow Config to be initialised without any args
Python
bsd-3-clause
praekelt/diamondash,praekelt/diamondash,praekelt/diamondash
bfcec696308ee8bfd226a54c17a7e15d49e2aed7
var/spack/repos/builtin/packages/nextflow/package.py
var/spack/repos/builtin/packages/nextflow/package.py
from spack import * from glob import glob import os class Nextflow(Package): """Data-driven computational pipelines""" homepage = "http://www.nextflow.io" version('0.20.1', '0e4e0e3eca1c2c97f9b4bffd944b923a', url='https://github.com/nextflow-io/nextflow/releases/download/v0.20.1/nextflow', expand=False) depends_on('jdk') def unpack(self): pass def install(self, spec, prefix): chmod = which('chmod') mkdirp(prefix.bin) install("nextflow", join_path(prefix.bin, "nextflow")) chmod('+x', join_path(prefix.bin, "nextflow"))
############################################################################## # Copyright (c) 2013-2016, Lawrence Livermore National Security, LLC. # Produced at the Lawrence Livermore National Laboratory. # # This file is part of Spack. # Created by Todd Gamblin, [email protected], All rights reserved. # LLNL-CODE-647188 # # For details, see https://github.com/llnl/spack # Please also see the LICENSE file for our notice and the LGPL. # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU Lesser General Public License (as # published by the Free Software Foundation) version 2.1, February 1999. # # This program is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the IMPLIED WARRANTY OF # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the terms and # conditions of the GNU Lesser General Public License for more details. # # You should have received a copy of the GNU Lesser General Public # License along with this program; if not, write to the Free Software # Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA ############################################################################## from spack import * class Nextflow(Package): """Data-driven computational pipelines""" homepage = "http://www.nextflow.io" version('0.20.1', '0e4e0e3eca1c2c97f9b4bffd944b923a', url='https://github.com/nextflow-io/nextflow/releases/download/v0.20.1/nextflow', expand=False) depends_on('jdk') def unpack(self): pass def install(self, spec, prefix): mkdirp(prefix.bin) install("nextflow", join_path(prefix.bin, "nextflow")) set_executable( join_path(prefix.bin, "nextflow"))
Add standard header, use spack helpers
Add standard header, use spack helpers Added the standard header (stolen from R). Touched up the install to use set_executable rather than doing it myself.
Python
lgpl-2.1
matthiasdiener/spack,mfherbst/spack,lgarren/spack,tmerrick1/spack,TheTimmy/spack,LLNL/spack,tmerrick1/spack,TheTimmy/spack,TheTimmy/spack,matthiasdiener/spack,LLNL/spack,iulian787/spack,matthiasdiener/spack,krafczyk/spack,tmerrick1/spack,EmreAtes/spack,TheTimmy/spack,tmerrick1/spack,iulian787/spack,matthiasdiener/spack,mfherbst/spack,iulian787/spack,tmerrick1/spack,EmreAtes/spack,skosukhin/spack,skosukhin/spack,krafczyk/spack,lgarren/spack,LLNL/spack,matthiasdiener/spack,EmreAtes/spack,iulian787/spack,lgarren/spack,skosukhin/spack,LLNL/spack,krafczyk/spack,mfherbst/spack,TheTimmy/spack,skosukhin/spack,iulian787/spack,lgarren/spack,EmreAtes/spack,krafczyk/spack,mfherbst/spack,LLNL/spack,lgarren/spack,mfherbst/spack,EmreAtes/spack,skosukhin/spack,krafczyk/spack
e81b1ce7536ce32e022fb3132f8468d2472b2e31
atlas/prodtask/management/commands/extendopenended.py
atlas/prodtask/management/commands/extendopenended.py
from django.core.management.base import BaseCommand, CommandError from atlas.prodtask.open_ended import check_open_ended class Command(BaseCommand): args = '<request_id, request_id>' help = 'Extend open ended requests' def handle(self, *args, **options): if not args: try: check_open_ended() except Exception,e: raise CommandError('Some problem during request extension: %s'%e) self.stdout.write('Successfully finished request extension')
from django.core.management.base import BaseCommand, CommandError import time from atlas.prodtask.open_ended import check_open_ended class Command(BaseCommand): args = '<request_id, request_id>' help = 'Extend open ended requests' def handle(self, *args, **options): self.stdout.write('Start open ended at %s'%time.ctime()) if not args: try: check_open_ended() except Exception,e: raise CommandError('Some problem during request extension: %s'%e) self.stdout.write('Successfully finished request extension: %s'%time.ctime())
Improve logging of openended extension
Improve logging of openended extension
Python
apache-2.0
PanDAWMS/panda-bigmon-atlas,PanDAWMS/panda-bigmon-atlas,PanDAWMS/panda-bigmon-atlas,PanDAWMS/panda-bigmon-atlas
6632157febfed7ce99fa1aaecb72393b0301d3aa
geotrek/authent/migrations/0003_auto_20181203_1518.py
geotrek/authent/migrations/0003_auto_20181203_1518.py
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import migrations from django.core.management import call_command from django.conf import settings def add_permissions(apps, schema_editor): if 'geotrek.infrastructure' in settings.INSTALLED_APPS: call_command('update_geotrek_permissions', verbosity=0) UserModel = apps.get_model('auth', 'User') GroupModel = apps.get_model('auth', 'Group') PermissionModel = apps.get_model('auth', 'Permission') ContentTypeModel = apps.get_model("contenttypes", "ContentType") type_permissions = ['add', 'change', 'change_geom', 'delete', 'export', 'read'] content_type_signage = ContentTypeModel.objects.get(model='signage') content_type_infrastructure = ContentTypeModel.objects.get(model='infrastructure') for user in UserModel.objects.all(): for type_perm in type_permissions: if user.user_permissions.filter(codename='%s_infrastructure' % type_perm).exists(): user.user_permissions.add(PermissionModel.objects.get( codename='%s_infrastructure' % type_perm, content_type=content_type_infrastructure)) if user.user_permissions.filter(codename='%s_signage' % type_perm).exists(): user.user_permissions.add(PermissionModel.objects.get( codename='%s_signage' % type_perm, content_type=content_type_signage)) for group in GroupModel.objects.all(): for type_perm in type_permissions: if group.permissions.filter(codename='%s_infrastructure' % type_perm).exists(): group.permissions.add(PermissionModel.objects.get( codename='%s_infrastructure' % type_perm, content_type=content_type_infrastructure)) if group.permissions.filter(codename='%s_signage' % type_perm).exists(): group.permissions.add(PermissionModel.objects.get( codename='%s_signage' % type_perm, content_type=content_type_signage)) PermissionModel.objects.filter(content_type__model='baseinfrastructure').delete() class Migration(migrations.Migration): dependencies = [ ('authent', '0002_auto_20181107_1620'), ] operations = [ migrations.RunPython(add_permissions) ]
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import migrations class Migration(migrations.Migration): dependencies = [ ('authent', '0002_auto_20181107_1620'), ] operations = [ ]
Make empty migration authent 3
Make empty migration authent 3
Python
bsd-2-clause
makinacorpus/Geotrek,GeotrekCE/Geotrek-admin,GeotrekCE/Geotrek-admin,makinacorpus/Geotrek,GeotrekCE/Geotrek-admin,GeotrekCE/Geotrek-admin,makinacorpus/Geotrek,makinacorpus/Geotrek
0324d220872ef063cb39ce62264bd4835f260920
test_project/urls.py
test_project/urls.py
from django.conf.urls import include, url from django.contrib import admin from django.views.generic import RedirectView from test_app.models import DummyModel, MushroomSpot from test_app.views import DummyDocumentOdt, DummyDocumentWeasyprint from mapentity.registry import registry handler403 = 'mapentity.views.handler403' admin.autodiscover() models_urls = registry.register(DummyModel) + registry.register(MushroomSpot) urlpatterns = [ url(r'', include(models_urls, namespace='test_app')), url(r'', include('mapentity.urls', namespace='mapentity', app_name='mapentity')), url(r'^home/$', RedirectView.as_view(url='/', permanent=True), name='home'), url(r'^login/$', 'django.contrib.auth.views.login', name='login'), url(r'^logout/$', 'django.contrib.auth.views.logout', {'next_page': '/'}, name='logout',), url(r'^paperclip/', include('paperclip.urls')), url(r'^admin/', include(admin.site.urls)), url(r'^test/document/dummymodel-(?P<pk>\d+).odt', DummyDocumentOdt.as_view(), name="dummymodel_odt"), url(r'^test/document/dummymodel-(?P<pk>\d+).pdf', DummyDocumentWeasyprint.as_view(), name="dummymodel_pdf"), ]
from django.conf.urls import include, url from django.contrib import admin from django.views.generic import RedirectView from test_app.models import DummyModel, MushroomSpot from test_app.views import DummyDocumentOdt, DummyDocumentWeasyprint from mapentity.registry import registry from django.contrib.auth import views as auth_views handler403 = 'mapentity.views.handler403' admin.autodiscover() models_urls = registry.register(DummyModel) + registry.register(MushroomSpot) urlpatterns = [ url(r'', include(models_urls, namespace='test_app')), url(r'', include('mapentity.urls', namespace='mapentity', app_name='mapentity')), url(r'^home/$', RedirectView.as_view(url='/', permanent=True), name='home'), url(r'^login/$', auth_views.login, name='login'), url(r'^logout/$', auth_views.logout, {'next_page': '/'}, name='logout',), url(r'^paperclip/', include('paperclip.urls')), url(r'^admin/', include(admin.site.urls)), url(r'^test/document/dummymodel-(?P<pk>\d+).odt', DummyDocumentOdt.as_view(), name="dummymodel_odt"), url(r'^test/document/dummymodel-(?P<pk>\d+).pdf', DummyDocumentWeasyprint.as_view(), name="dummymodel_pdf"), ]
Replace str into call in url
Replace str into call in url
Python
bsd-3-clause
makinacorpus/django-mapentity,makinacorpus/django-mapentity,makinacorpus/django-mapentity
a53612d5f276180d204378b9e4974fcd812f6a5b
tests/fake_camera.py
tests/fake_camera.py
from os import listdir from os.path import isfile, join class Camera(object): def __init__(self, path): self.files = [join(path, f) for f in listdir(path)] self.files = sorted([f for f in self.files if isfile(f)]) self.current = 0 def reset(self): self.current = 0 def has_next(self): return self.current < len(self.files) def next(self): img = open(self.files[self.current], 'rb').read() self.current += 1 return img
# -*- coding: utf-8 -*- # Licensed to the Apache Software Foundation (ASF) under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the # "License"); you may not use this file except in compliance # with the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, # software distributed under the License is distributed on an # "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY # KIND, either express or implied. See the License for the # specific language governing permissions and limitations # under the License. from os import listdir from os.path import isfile, join class Camera(object): def __init__(self, path): self.files = [join(path, f) for f in listdir(path)] self.files = sorted([f for f in self.files if isfile(f)]) self.current = 0 def reset(self): self.current = 0 def has_next(self): return self.current < len(self.files) def next(self): img = open(self.files[self.current], 'rb').read() self.current += 1 return img
Add licence header in fake camera test file.
Add licence header in fake camera test file.
Python
apache-2.0
angus-ai/angus-sdk-python
d1ea64d6645f60df38221cbd194c26dff9686dcd
scripts/utils.py
scripts/utils.py
import sys import hashlib def e(s): if type(s) == str: return str return s.encode('utf-8') def d(s): if type(s) == unicode: return s return unicode(s, 'utf-8') def mkid(s): return hashlib.sha1(e(s)).hexdigest()[:2*4] class Logger(object): def __init__(self): self._mode = 'INFO' def progress(self, message): if not sys.stderr.isatty(): return if self._mode == 'PROGRESS': print >>sys.stderr, '\r', print >>sys.stderr, message, self._mode = 'PROGRESS' def info(self, message): if self._mode == 'PROGRESS': print >>sys.stderr print >>sys.stderr, message self._mode = 'INFO'
import sys import hashlib def e(s): if type(s) == str: return str return s.encode('utf-8') def d(s): if type(s) == unicode: return s return unicode(s, 'utf-8') def mkid(s): return hashlib.sha1(e(s)).hexdigest()[:2*4] class Logger(object): def __init__(self): self._mode = 'INFO' def progress(self, message): message = e(message) if not sys.stderr.isatty(): return if self._mode == 'PROGRESS': print >>sys.stderr, '\r', print >>sys.stderr, message, self._mode = 'PROGRESS' def info(self, message): message = e(message) if self._mode == 'PROGRESS': print >>sys.stderr print >>sys.stderr, message self._mode = 'INFO'
Handle logging unicode messages in python2.
Handle logging unicode messages in python2. Former-commit-id: 257d94eb71d5597ff52a18ec1530d73496901ef4
Python
mit
guilherme-pg/citationhunt,eggpi/citationhunt,guilherme-pg/citationhunt,eggpi/citationhunt,eggpi/citationhunt,guilherme-pg/citationhunt,guilherme-pg/citationhunt,eggpi/citationhunt
84a99e9557a323e094c360e748c7d7042980fc59
tests/test_sample.py
tests/test_sample.py
import unittest from tip.algorithms.dummy import dummy_add class TestDummyAdd(unittest.TestCase): def test_lcm(self): r = dummy_add(2, 2) self.assertEqual(r, 4)
import unittest from tip.algorithms.dummy import dummy_add class TestDummyAdd(unittest.TestCase): def test_lcm(self): r = dummy_add(2, 2) self.assertEqual(r, 4)
Test PEP8 integration into Atom
Test PEP8 integration into Atom
Python
unlicense
davidgasquez/tip
1c28341a4cd828de607d9cc4252f444844c0a892
test/bibliopixel/util/udp_test.py
test/bibliopixel/util/udp_test.py
import contextlib, queue, time, unittest from bibliopixel.util import udp TEST_ADDRESS = '127.0.0.1', 5678 TIMEOUT = 0.2 @contextlib.contextmanager def receive_udp(address, results): receiver = udp.QueuedReceiver(address) receiver.start() yield try: while True: results.append(receiver.queue.get(timeout=TIMEOUT)) except queue.Empty: pass class UDPTest(unittest.TestCase): def test_full(self): messages = [s.encode() for s in ('foo', '', 'bar', 'baz', '', 'bing')] expected = [s for s in messages if s] # Note that empty messages are either not sent, or not received. actual = [] with receive_udp(TEST_ADDRESS, actual): sender = udp.QueuedSender(TEST_ADDRESS) sender.start() for m in messages: sender.send(m) self.assertEquals(actual, expected)
import contextlib, queue, time, unittest from bibliopixel.util import udp TEST_ADDRESS = '127.0.0.1', 5678 TIMEOUT = 0.3 @contextlib.contextmanager def receive_udp(address, results): receiver = udp.QueuedReceiver(address) receiver.start() yield try: while True: results.append(receiver.queue.get(timeout=TIMEOUT)) except queue.Empty: pass class UDPTest(unittest.TestCase): def test_full(self): messages = [s.encode() for s in ('foo', '', 'bar', 'baz', '', 'bing')] expected = [s for s in messages if s] # Note that empty messages are either not sent, or not received. actual = [] with receive_udp(TEST_ADDRESS, actual): sender = udp.QueuedSender(TEST_ADDRESS) sender.start() for m in messages: sender.send(m) self.assertEquals(actual, expected)
Tweak up timeout in UDP test
Tweak up timeout in UDP test
Python
mit
rec/BiblioPixel,ManiacalLabs/BiblioPixel,rec/BiblioPixel,rec/BiblioPixel,ManiacalLabs/BiblioPixel,rec/BiblioPixel,ManiacalLabs/BiblioPixel,ManiacalLabs/BiblioPixel
a74e91613be376d6d71fb90c15cab689af661e37
money_conversion/money.py
money_conversion/money.py
from currency_rates import rates class Money(object): def __init__(self, amount, currency): self.amount = amount self.currency = currency.upper() def __repr__(self): return "%.2f %s" % (self.amount, self.currency) def to_currency(self, new_currency): new_currency = new_currency.split('_')[1].upper() amount = self.amount base_currency_rates = rates.get(self.currency) new_amount = amount * base_currency_rates.get(new_currency) return Money(new_amount, new_currency)
from currency_rates import rates class Money(object): def __init__(self, amount, currency): self.amount = amount self.currency = currency.upper() def __repr__(self): return "%.2f %s" % (self.amount, self.currency) def __getattr__(self, currency): def convert(): return self.to_currency(currency) return convert def to_currency(self, currency): currency = currency.split('_')[1].upper() amount = self.amount base_currency_rates = rates.get(self.currency) new_amount = amount * base_currency_rates.get(currency) return Money(new_amount, currency)
Add __getattr__ method in order to be able to call non-defined methods
Add __getattr__ method in order to be able to call non-defined methods
Python
mit
mdsrosa/money-conversion-py
9a698d1428fbe0744c9dba3532b778569dbe1dd4
server.py
server.py
import socket import sys class SimpleServer(object): """Simple server using the socket library""" def __init__(self, blocking=False, connection_oriented=True): """ The constructor initializes socket specifying the blocking status and if it must be a connection oriented socket. :param blocking: A flag that specifies if the socket must be blocking :ptype: Boolean :param connection_oriented: A flag that specifies if the socket must be connection oriented or not :ptype: Boolean """ self.sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM) if not blocking: self.sock.setblocking(0) def connect(self, host, port): """ Connects the server to the "host", and prepares it to listen on "port" :param host: The network layer identifier of an interface :ptype: String or Integer (see help(socket)) :param port: The transport layer identifier of an application :ptype: Integer """ self.sock.connect((host, port))
""" A Simple Server class that allows to configure a socket in a very simple way. It is for studying purposes only. """ import socket import sys __author__ = "Facundo Victor" __license__ = "MIT" __email__ = "[email protected]" class SimpleServer(object): """Simple server using the socket library""" def __init__(self, blocking=False, connection_oriented=True): """ The constructor initializes socket specifying the blocking status and if it must be a connection oriented socket. :param blocking: A flag that specifies if the socket must be blocking :ptype: Boolean :param connection_oriented: A flag that specifies if the socket must be connection oriented or not :ptype: Boolean """ self.sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM) if not blocking: self.sock.setblocking(0) def connect(self, host, port): """ Connects the server to the "host", and prepares it to listen on "port" :param host: The network layer identifier of an interface :ptype: String or Integer (see help(socket)) :param port: The transport layer identifier of an application :ptype: Integer """ server_address = (host, port) self.sock.connect(server_address) print('starting up on %s port %s' % server_address)
Add docstrings and author reference
Add docstrings and author reference
Python
mit
facundovictor/non-blocking-socket-samples
5f501af61b416dae0e46236a8e1f9684dcc66e21
python/decoder_test.py
python/decoder_test.py
import argparse import scanner import numpy as np import cv2 from decode import db @db.loader('frame') def load_frames(buf, metadata): return np.frombuffer(buf, dtype=np.uint8) \ .reshape((metadata.height,metadata.width,3)) def extract_frames(args): job = load_frames(args['dataset'], 'edr') video_paths = job._dataset.video_data.original_video_paths for (vid, frames) in job.as_frame_list(): video_path = video_paths[int(vid)] inp = cv2.VideoCapture(video_path) assert(inp.isOpened()) video_frame_num = -1 for (frame_num, buf) in frames: while video_frame_num != frame_num: _, video_frame = inp.read() video_frame_num += 1 scanner_frame = cv2.cvtColor(buf, cv2.COLOR_RGB2BGR) frame_diff = (scanner_frame - video_frame).sum() if frame_diff != 0: print('Frame {} does not match!'.format(frame_num)) if __name__ == "__main__": p = argparse.ArgumentParser(description='Extract JPEG frames from videos') p.add_argument('dataset', type=str) extract_frames(p.parse_args().__dict__)
import argparse import scanner import numpy as np import cv2 from decode import db @db.loader('frame') def load_frames(buf, metadata): return np.frombuffer(buf, dtype=np.uint8) \ .reshape((metadata.height,metadata.width,3)) def extract_frames(args): job = load_frames(args['dataset'], 'edr') video_paths = job._dataset.video_data.original_video_paths for (vid, frames) in job.as_frame_list(): video_path = video_paths[int(vid)] inp = cv2.VideoCapture(video_path) assert(inp.isOpened()) video_frame_num = -1 for (frame_num, buf) in frames: while video_frame_num != frame_num: _, video_frame = inp.read() video_frame_num += 1 scanner_frame = cv2.cvtColor(buf, cv2.COLOR_RGB2BGR) frame_diff = np.abs(scanner_frame - video_frame) if frame_diff.sum() != 0: print('Frame {} does not match!'.format(frame_num)) cv2.imwrite('decode_frames_' + str(frame_num) + '.jpg', np.concatenate( (scanner_frame, video_frame, frame_diff), 1)) if __name__ == "__main__": p = argparse.ArgumentParser(description='Extract JPEG frames from videos') p.add_argument('dataset', type=str) extract_frames(p.parse_args().__dict__)
Write out concatenated frame on decode test failure
Write out concatenated frame on decode test failure
Python
apache-2.0
scanner-research/scanner,scanner-research/scanner,scanner-research/scanner,scanner-research/scanner
e2cba02550dfbe8628daf024a2a35c0dffb234e9
python/cli/request.py
python/cli/request.py
import requests import os aport = os.environ.get('MYAPORT') if aport is None: aport = "80" aport = "23456" url1 = 'http://localhost:' + aport + '/' url2 = 'http://localhost:' + aport + '/action/improvesimulateinvest' url3 = 'http://localhost:' + aport + '/action/autosimulateinvest' url4 = 'http://localhost:' + aport + '/action/improveautosimulateinvest' #headers = {'Content-type': 'application/json', 'Accept': 'text/plain'} #headers={'Content-type':'application/json', 'Accept':'application/json'} headers={'Content-Type' : 'application/json;charset=utf-8'} def request1(param, webpath): return requests.post(url1 + webpath, json=param, headers=headers) def request2(market, data): return requests.post(url2 + '/market/' + str(market), json=data, headers=headers) def request3(market, data): return requests.post(url3 + '/market/' + str(market), json=data, headers=headers) def request4(market, data): return requests.post(url4 + '/market/' + str(market), json=data, headers=headers) def request0(data): return requests.post(url, data='', headers=headers) #return requests.post(url, data=json.dumps(data), headers=headers)
import requests import os aport = os.environ.get('MYAPORT') if aport is None: aport = "80" aport = "23456" ahost = os.environ.get('MYAHOST') if ahost is None: ahost = "localhost" url1 = 'http://' + ahost + ':' + aport + '/' #headers = {'Content-type': 'application/json', 'Accept': 'text/plain'} #headers={'Content-type':'application/json', 'Accept':'application/json'} headers={'Content-Type' : 'application/json;charset=utf-8'} def request1(param, webpath): return requests.post(url1 + webpath, json=param, headers=headers) def request0(data): return requests.post(url, data='', headers=headers) #return requests.post(url, data=json.dumps(data), headers=headers)
Handle different environments, for automation (I4).
Handle different environments, for automation (I4).
Python
agpl-3.0
rroart/aether,rroart/aether,rroart/aether,rroart/aether,rroart/aether