commit
stringlengths
40
40
old_file
stringlengths
4
264
new_file
stringlengths
4
264
old_contents
stringlengths
0
3.26k
new_contents
stringlengths
1
4.43k
subject
stringlengths
15
624
message
stringlengths
15
4.7k
lang
stringclasses
3 values
license
stringclasses
13 values
repos
stringlengths
5
91.5k
9ab1530a010e974a376c75da806016185199c545
evelink/__init__.py
evelink/__init__.py
"""EVELink - Python bindings for the EVE API.""" import logging from evelink import account from evelink import api from evelink import char from evelink import constants from evelink import corp from evelink import eve from evelink import map from evelink import server __version__ = "0.6.2" # Implement NullHandler because it was only added in Python 2.7+. class NullHandler(logging.Handler): def emit(self, record): pass # Create a logger, but by default, have it do nothing _log = logging.getLogger('evelink') _log.addHandler(NullHandler()) # Update the version number used in the user-agent api._user_agent = 'evelink v%s' % __version__ __all__ = [ "account", "api", "char", "constants", "corp", "eve", "map", "parsing", "server", ]
"""EVELink - Python bindings for the EVE API.""" import logging from evelink import account from evelink import api from evelink import char from evelink import constants from evelink import corp from evelink import eve from evelink import map from evelink import server __version__ = "0.7.0" # Implement NullHandler because it was only added in Python 2.7+. class NullHandler(logging.Handler): def emit(self, record): pass # Create a logger, but by default, have it do nothing _log = logging.getLogger('evelink') _log.addHandler(NullHandler()) # Update the version number used in the user-agent api._user_agent = 'evelink v%s' % __version__ __all__ = [ "account", "api", "char", "constants", "corp", "eve", "map", "parsing", "server", ]
Update version to 0.7.0 for release
Update version to 0.7.0 for release
Python
mit
FashtimeDotCom/evelink,ayust/evelink,bastianh/evelink,zigdon/evelink
f16994fd3722acba8a60157eed0630a5e2a3d387
macdict/cli.py
macdict/cli.py
from __future__ import absolute_import import sys import argparse from macdict.dictionary import lookup_word, ensure_unicode def parse_args(): parser = argparse.ArgumentParser() parser.add_argument('word') return parser.parse_args() def abort(text): sys.stderr.write(u'%s\n' % text) sys.exit(1) def report(text): sys.stdout.write(u'%s\n' % text) sys.exit(0) def main(): args = parse_args() definition = lookup_word(ensure_unicode(args.word, 'utf-8')) if definition is None: abort(u'Definition not found for "%s"' % args.word) else: report(definition)
from __future__ import absolute_import import sys import argparse from macdict.dictionary import lookup_word, ensure_unicode def parse_args(): parser = argparse.ArgumentParser() parser.add_argument('word') return parser.parse_args() def abort(text): sys.stderr.write(u'%s\n' % text) sys.exit(1) def report(text): sys.stdout.write(u'%s\n' % text) sys.exit(0) def main(): args = parse_args() word = ensure_unicode(args.word, 'utf-8') definition = lookup_word(word) if definition is None: abort(u'Definition not found for "%s"' % word) else: report(definition)
Fix unicode decoding on error messages
Fix unicode decoding on error messages
Python
mit
tonyseek/macdict
554bf6551d0be9d11e046610e4b5772b5beeb9b8
mwdb/schema.py
mwdb/schema.py
from contextlib import contextmanager from sqlalchemy import MetaData, create_engine from sqlalchemy.engine import Engine from sqlalchemy.orm import sessionmaker class Schema(): def __init__(self, engine_or_url, *args, **kwargs): if isinstance(engine_or_url, Engine): self.engine = engine_or_url else: self.engine = create_engine(engine_or_url, *args, **kwargs) self.meta = MetaData(bind=self.engine) self.meta.reflect(views=True) self.Session = sessionmaker(bind=self.engine) def __getattr__(self, table_name): return self.meta.tables[table_name] def execute(self, *args, **kwargs): with self.session() as session: session.execute(*args, **kwargs) @contextmanager def session(self): """Provides a transactional scope around a series of operations.""" session = self.Session() try: yield session session.commit() except: session.rollback() raise finally: session.close()
from contextlib import contextmanager from sqlalchemy import MetaData, create_engine from sqlalchemy.engine import Engine from sqlalchemy.orm import sessionmaker class Schema(): def __init__(self, engine_or_url, *args, **kwargs): if isinstance(engine_or_url, Engine): self.engine = engine_or_url else: self.engine = create_engine(engine_or_url, *args, **kwargs) self.meta = MetaData(bind=self.engine) self.meta.reflect(views=True) self.Session = sessionmaker(bind=self.engine) def __getattr__(self, table_name): return self.meta.tables[table_name] def execute(self, *args, **kwargs): with self.session() as session: return session.execute(*args, **kwargs) @contextmanager def session(self): """Provides a transactional scope around a series of operations.""" session = self.Session() try: yield session session.commit() except: session.rollback() raise finally: session.close()
Return the result of an Execute!
Return the result of an Execute!
Python
mit
mediawiki-utilities/python-mwdb
75171ed80079630d22463685768072ad7323e653
boundary/action_installed.py
boundary/action_installed.py
### ### Copyright 2014-2015 Boundary, Inc. ### ### Licensed under the Apache License, Version 2.0 (the "License"); ### you may not use this file except in compliance with the License. ### You may obtain a copy of the License at ### ### http://www.apache.org/licenses/LICENSE-2.0 ### ### Unless required by applicable law or agreed to in writing, software ### distributed under the License is distributed on an "AS IS" BASIS, ### WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. ### See the License for the specific language governing permissions and ### limitations under the License. ### from api_cli import ApiCli class ActionInstalled (ApiCli): def __init__(self): ApiCli.__init__(self) self.method = "GET" self.path = "v1/actions/installed" def getDescription(self): return "Returns the actions associated with the Boundary account"
# # Copyright 2014-2015 Boundary, Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from api_cli import ApiCli class ActionInstalled (ApiCli): def __init__(self): ApiCli.__init__(self) self.method = "GET" self.path = "v1/actions/installed" def getDescription(self): return "Returns the actions configured within a Boundary account"
Change code to be PEP-8 compliant
Change code to be PEP-8 compliant
Python
apache-2.0
boundary/boundary-api-cli,boundary/boundary-api-cli,jdgwartney/boundary-api-cli,jdgwartney/pulse-api-cli,wcainboundary/boundary-api-cli,wcainboundary/boundary-api-cli,jdgwartney/pulse-api-cli,boundary/pulse-api-cli,jdgwartney/boundary-api-cli,boundary/pulse-api-cli
57bc8b3c40bbafda6f69b23c230ad73750e881ab
hashable/helpers.py
hashable/helpers.py
from .equals_builder import EqualsBuilder from .hash_code_builder import HashCodeBuilder __all__ = [ 'hashable', 'equality_comparable', ] def hashable(cls=None, attributes=None, methods=None): _validate_attributes_and_methods(attributes, methods) def decorator(cls): cls = equality_comparable(cls, attributes, methods) cls.__hash__ = HashCodeBuilder.auto_generate(cls, attributes, methods) return cls return decorator if cls is None else decorator(cls) def equality_comparable(cls=None, attributes=None, methods=None): _validate_attributes_and_methods(attributes, methods) def decorator(cls): cls.__eq__ = EqualsBuilder.auto_generate(cls, attributes, methods) cls.__ne__ = EqualsBuilder.auto_ne_from_eq() return cls return decorator if cls is None else decorator(cls) def _validate_attributes_and_methods(attributes, methods): assert not isinstance(attributes, basestring), 'attributes must be list' assert not isinstance(methods, basestring), 'methods must be list' assert attributes or methods, 'attributes or methods must be NOT empty'
from .equals_builder import EqualsBuilder from .hash_code_builder import HashCodeBuilder __all__ = [ 'hashable', 'equalable', ] def hashable(cls=None, attributes=None, methods=None): _validate_attributes_and_methods(attributes, methods) def decorator(cls): cls = equalable(cls, attributes, methods) cls.__hash__ = HashCodeBuilder.auto_generate(cls, attributes, methods) return cls return decorator if cls is None else decorator(cls) def equalable(cls=None, attributes=None, methods=None): _validate_attributes_and_methods(attributes, methods) def decorator(cls): cls.__eq__ = EqualsBuilder.auto_generate(cls, attributes, methods) cls.__ne__ = EqualsBuilder.auto_ne_from_eq() return cls return decorator if cls is None else decorator(cls) def _validate_attributes_and_methods(attributes, methods): assert not isinstance(attributes, basestring), 'attributes must be list' assert not isinstance(methods, basestring), 'methods must be list' assert attributes or methods, 'attributes or methods must be NOT empty'
Rename decorator equality_comparable to equalable
Rename decorator equality_comparable to equalable
Python
mit
minmax/hashable
4f6e27a6bbc2bbdb19c165f21d47d1491bffd70e
scripts/mc_check_lib_file.py
scripts/mc_check_lib_file.py
#!/usr/bin/env python # -*- mode: python; coding: utf-8 -*- # Copyright 2021 The HERA Collaboration # Licensed under the 2-clause BSD License """ Check that input files are safely in the librarian. This script takes a list of input files and returns the list of those found in the HERA_MC.lib_files table. NOTE: Assumes that lib_files is a faithful list of files uploaded to the Librarian """ import os from hera_mc import mc ap = mc.get_mc_argument_parser() ap.description = """Check that listed files are safely in librarian.""" ap.add_argument("files", type=str, default=None, nargs="*", help="list of files") args = ap.parse_args() db = mc.connect_to_mc_db(args) found_files = [] for pathname in args.files: filename = os.path.basename(pathname) with db.sessionmaker() as session: out = session.get_lib_files(filename) if len(out) > 0: print(pathname) # if we have a file, say so
#!/usr/bin/env python # -*- mode: python; coding: utf-8 -*- # Copyright 2021 The HERA Collaboration # Licensed under the 2-clause BSD License """ Check that input files are safely in the librarian. This script takes a list of input files and returns the list of those found in the HERA_MC.lib_files table. NOTE: Assumes that lib_files is a faithful list of files uploaded to the Librarian """ import os from hera_mc import mc ap = mc.get_mc_argument_parser() ap.description = """Check that listed files are safely in librarian.""" ap.add_argument("files", type=str, default=None, nargs="*", help="list of files") args = ap.parse_args() db = mc.connect_to_mc_db(args) found_files = [] with db.sessionmaker() as session: for pathname in args.files: filename = os.path.basename(pathname) out = session.get_lib_files(filename) if len(out) > 0: print(pathname) # if we have a file, say so
Move sessionmaker outside of loop
Move sessionmaker outside of loop
Python
bsd-2-clause
HERA-Team/hera_mc,HERA-Team/hera_mc
5436068e2a0974a932d59d51dd529af221832735
test/vim_autopep8.py
test/vim_autopep8.py
"""Run autopep8 on the selected buffer in Vim. map <C-I> :pyfile <path_to>/vim_autopep8.py<CR> """ import vim if vim.eval('&syntax') == 'python': encoding = vim.eval('&fileencoding') source = '\n'.join(line.decode(encoding) for line in vim.current.buffer) + '\n' import autopep8 options = autopep8.parse_args(['--range', str(1 + vim.current.range.start), str(1 + vim.current.range.end), '']) formatted = autopep8.fix_code(source, options=options) if source != formatted: if formatted.endswith('\n'): formatted = formatted[:-1] vim.current.buffer[:] = [line.encode(encoding) for line in formatted.splitlines()]
"""Run autopep8 on the selected buffer in Vim. map <C-I> :pyfile <path_to>/vim_autopep8.py<CR> Replace ":pyfile" with ":py3file" if Vim is built with Python 3 support. """ from __future__ import unicode_literals import sys import vim ENCODING = vim.eval('&fileencoding') def encode(text): if sys.version_info[0] >= 3: return text else: return text.encode(ENCODING) def decode(text): if sys.version_info[0] >= 3: return text else: return text.decode(ENCODING) if vim.eval('&syntax') == 'python': source = '\n'.join(decode(line) for line in vim.current.buffer) + '\n' import autopep8 options = autopep8.parse_args(['--range', str(1 + vim.current.range.start), str(1 + vim.current.range.end), '']) formatted = autopep8.fix_code(source, options=options) if source != formatted: if formatted.endswith('\n'): formatted = formatted[:-1] vim.current.buffer[:] = [encode(line) for line in formatted.splitlines()]
Support Python 3 in Vim usage example
Support Python 3 in Vim usage example
Python
mit
vauxoo-dev/autopep8,Vauxoo/autopep8,vauxoo-dev/autopep8,hhatto/autopep8,SG345/autopep8,SG345/autopep8,MeteorAdminz/autopep8,Vauxoo/autopep8,hhatto/autopep8,MeteorAdminz/autopep8
b1402c6ad51af7e76302605e6892684dcb6cd52c
addons/resource/models/res_company.py
addons/resource/models/res_company.py
# -*- coding: utf-8 -*- # Part of Odoo. See LICENSE file for full copyright and licensing details. from odoo import api, fields, models, _ class ResCompany(models.Model): _inherit = 'res.company' resource_calendar_ids = fields.One2many( 'resource.calendar', 'company_id', 'Working Hours') resource_calendar_id = fields.Many2one( 'resource.calendar', 'Default Working Hours', ondelete='restrict') @api.model def _init_data_resource_calendar(self): for company in self.search([('resource_calendar_id', '=', False)]): company.resource_calendar_id = self.env['resource.calendar'].create({'name': _('Standard 40 hours/week')}).id @api.model def create(self, values): if not values.get('resource_calendar_id'): values['resource_calendar_id'] = self.env['resource.calendar'].create({'name': _('Standard 40 hours/week')}).id company = super(ResCompany, self).create(values) # calendar created from form view: no company_id set because record was still not created if not company.resource_calendar_id.company_id: company.resource_calendar_id.company_id = company.id return company
# -*- coding: utf-8 -*- # Part of Odoo. See LICENSE file for full copyright and licensing details. from odoo import api, fields, models, _ class ResCompany(models.Model): _inherit = 'res.company' resource_calendar_ids = fields.One2many( 'resource.calendar', 'company_id', 'Working Hours') resource_calendar_id = fields.Many2one( 'resource.calendar', 'Default Working Hours', ondelete='restrict') @api.model def _init_data_resource_calendar(self): for company in self.search([('resource_calendar_id', '=', False)]): company.resource_calendar_id = self.env['resource.calendar'].create({'name': _('Standard 40 hours/week')}).id @api.model def create(self, values): if not values.get('resource_calendar_id'): values['resource_calendar_id'] = self.env['resource.calendar'].sudo().create({'name': _('Standard 40 hours/week')}).id company = super(ResCompany, self).create(values) # calendar created from form view: no company_id set because record was still not created if not company.resource_calendar_id.company_id: company.resource_calendar_id.company_id = company.id return company
Allow 'Access Rights' users to create companies
[FIX] resource: Allow 'Access Rights' users to create companies Purpose ======= A 'Access Rights' (group_erp_manager) user can create a company A 'Settings' (group_system) user can create a resource.calendar With the resource module, if a resource.calendar is not set on the new company values, a default one is create. As the 'Access Rights' user can create a company, but can't create a resource calendar, this should be done with sudo.
Python
agpl-3.0
ygol/odoo,ygol/odoo,ygol/odoo,ygol/odoo,ygol/odoo,ygol/odoo,ygol/odoo
c105d6f18a5a17b0a47fda5a2df2f8f47352b037
setuptools/command/upload.py
setuptools/command/upload.py
import getpass from distutils.command import upload as orig class upload(orig.upload): """ Override default upload behavior to obtain password in a variety of different ways. """ def finalize_options(self): orig.upload.finalize_options(self) # Attempt to obtain password. Short circuit evaluation at the first # sign of success. self.password = ( self.password or self._load_password_from_keyring() or self._prompt_for_password() ) def _load_password_from_keyring(self): """ Attempt to load password from keyring. Suppress Exceptions. """ try: keyring = __import__('keyring') password = keyring.get_password(self.repository, self.username) except Exception: password = None finally: return password def _prompt_for_password(self): """ Prompt for a password on the tty. Suppress Exceptions. """ password = None try: while not password: password = getpass.getpass() except (Exception, KeyboardInterrupt): password = None finally: return password
import getpass from distutils.command import upload as orig class upload(orig.upload): """ Override default upload behavior to obtain password in a variety of different ways. """ def finalize_options(self): orig.upload.finalize_options(self) # Attempt to obtain password. Short circuit evaluation at the first # sign of success. self.password = ( self.password or self._load_password_from_keyring() or self._prompt_for_password() ) def _load_password_from_keyring(self): """ Attempt to load password from keyring. Suppress Exceptions. """ try: keyring = __import__('keyring') return keyring.get_password(self.repository, self.username) except Exception: pass def _prompt_for_password(self): """ Prompt for a password on the tty. Suppress Exceptions. """ try: return getpass.getpass() except (Exception, KeyboardInterrupt): pass
Simplify logic by eliminating retries in password prompt and returning results directly.
Simplify logic by eliminating retries in password prompt and returning results directly.
Python
mit
pypa/setuptools,pypa/setuptools,pypa/setuptools
7ce46ada7322f2618fd92adf3eb0e8813b118031
changes/api/build_restart.py
changes/api/build_restart.py
from sqlalchemy.orm import joinedload from datetime import datetime from changes.api.base import APIView from changes.api.build_index import execute_build from changes.config import db from changes.constants import Result, Status from changes.models import Build, Job, ItemStat class BuildRestartAPIView(APIView): def post(self, build_id): build = Build.query.options( joinedload('project', innerjoin=True), joinedload('author'), joinedload('source'), ).get(build_id) if build is None: return '', 404 if build.status != Status.finished: return '', 400 # remove any existing job data # TODO(dcramer): this is potentially fairly slow with cascades Job.query.filter( Job.build == build ).delete() ItemStat.query.filter( ItemStat.item_id == build.id ).delete() build.date_started = datetime.utcnow() build.date_modified = build.date_started build.date_finished = None build.duration = None build.status = Status.queued build.result = Result.unknown db.session.add(build) execute_build(build=build) return self.respond(build)
from sqlalchemy.orm import joinedload from datetime import datetime from changes.api.base import APIView from changes.api.build_index import execute_build from changes.config import db from changes.constants import Result, Status from changes.models import Build, Job, ItemStat class BuildRestartAPIView(APIView): def post(self, build_id): build = Build.query.options( joinedload('project', innerjoin=True), joinedload('author'), joinedload('source'), ).get(build_id) if build is None: return '', 404 if build.status != Status.finished: return '', 400 # ItemStat doesnt cascade ItemStat.query.filter( ItemStat.item_id == build.id ).delete() ItemStat.query.filter( ItemStat.item_id.in_(Job.query.filter( Job.build_id == build.id, )), ).delete() # remove any existing job data # TODO(dcramer): this is potentially fairly slow with cascades Job.query.filter( Job.build_id == build.id ).delete() build.date_started = datetime.utcnow() build.date_modified = build.date_started build.date_finished = None build.duration = None build.status = Status.queued build.result = Result.unknown db.session.add(build) execute_build(build=build) return self.respond(build)
Clean up job stats when jobs are removed in build restart
Clean up job stats when jobs are removed in build restart
Python
apache-2.0
dropbox/changes,bowlofstew/changes,bowlofstew/changes,wfxiang08/changes,wfxiang08/changes,wfxiang08/changes,dropbox/changes,wfxiang08/changes,dropbox/changes,dropbox/changes,bowlofstew/changes,bowlofstew/changes
8faf4cd2fa6e155bbe85510ce3ee388bb0e19d3c
src/data/clean_scripts/SG_dengue_malaria_clean.py
src/data/clean_scripts/SG_dengue_malaria_clean.py
import os.path import sys import pandas as pd import logging INPUT_DIRECTORY = '../../../data/raw/disease_SG' INPUT_FILE = "weekly-dengue-malaria.csv" OUTPUT_DIRECTORY = '../../Data/interim/disease_SG' OUTPUT_FILE = "weekly-dengue-malaria.csv" logger = logging.getLogger(__name__) def clean(): input_path = os.path.join(INPUT_DIRECTORY, INPUT_FILE) if not os.path.isfile(input_path): logger.error("Input file is not found %s", os.path.abspath(input_path)) data_frame = pd.read_csv(input_path, names=['week', 'disease', 'number of cases']) print (data_frame) print("end") if __name__ == "__main__": logging.basicConfig(stream=sys.stdout, level=logging.DEBUG) clean()
import os.path import sys import pandas as pd import logging INPUT_DIRECTORY = '../../../data/raw/disease_SG' INPUT_FILE = "weekly-dengue-malaria.csv" OUTPUT_DIRECTORY = '../../../data/interim/disease_SG' OUTPUT_FILE = "weekly-dengue-malaria-cleaned.csv" logger = logging.getLogger(__name__) def clean(): input_path = os.path.join(INPUT_DIRECTORY, INPUT_FILE) if not os.path.isfile(input_path): logger.error("Input file is not found %s", os.path.abspath(input_path)) data_frame = pd.read_csv(input_path, names=['year_week', 'disease', 'number_of_cases']) data_frame['country'] = 'Singapore' year_week = pd.DataFrame(data_frame.year_week.str.split('-').tolist(), columns=['year','week']) data_frame['year'] = year_week['year'] data_frame['week'] = year_week['week'] data_frame.drop('year_week', 1, inplace=True) os.makedirs(OUTPUT_DIRECTORY, exist_ok=True) output_path = os.path.join(OUTPUT_DIRECTORY, OUTPUT_FILE) data_frame.to_csv(output_path, index=False) logger.info('Data clean successfully') if __name__ == "__main__": logging.basicConfig(stream=sys.stdout, level=logging.DEBUG) clean()
Transform data to the target format
Transform data to the target format
Python
mit
DataKind-SG/healthcare_ASEAN
a3213788d0d8591b235359d4b17886ce3f50ab37
tests/test_plugin.py
tests/test_plugin.py
import datajoint.errors as djerr import datajoint.plugin as p import pkg_resources def test_check_pubkey(): base_name = 'datajoint' base_meta = pkg_resources.get_distribution(base_name) pubkey_meta = base_meta.get_metadata('{}.pub'.format(base_name)) with open('./datajoint.pub', "r") as f: assert(f.read() == pubkey_meta) def test_normal_djerror(): try: raise djerr.DataJointError except djerr.DataJointError as e: assert(e.__cause__ is None) def test_verified_djerror(): try: curr_plugins = p.discovered_plugins p.discovered_plugins = dict(test_plugin_module=dict(verified=True, plugon='example')) raise djerr.DataJointError except djerr.DataJointError as e: p.discovered_plugins = curr_plugins assert(e.__cause__ is None) def test_unverified_djerror(): try: curr_plugins = p.discovered_plugins p.discovered_plugins = dict(test_plugin_module=dict(verified=False, plugon='example')) raise djerr.DataJointError("hello") except djerr.DataJointError as e: p.discovered_plugins = curr_plugins assert(isinstance(e.__cause__, djerr.PluginWarning))
import datajoint.errors as djerr import datajoint.plugin as p import pkg_resources from os import path def test_check_pubkey(): base_name = 'datajoint' base_meta = pkg_resources.get_distribution(base_name) pubkey_meta = base_meta.get_metadata('{}.pub'.format(base_name)) with open(path.join(path.abspath( path.dirname(__file__)), '..', 'datajoint.pub'), "r") as f: assert(f.read() == pubkey_meta) def test_normal_djerror(): try: raise djerr.DataJointError except djerr.DataJointError as e: assert(e.__cause__ is None) def test_verified_djerror(): try: curr_plugins = p.discovered_plugins p.discovered_plugins = dict(test_plugin_module=dict(verified=True, plugon='example')) raise djerr.DataJointError except djerr.DataJointError as e: p.discovered_plugins = curr_plugins assert(e.__cause__ is None) def test_unverified_djerror(): try: curr_plugins = p.discovered_plugins p.discovered_plugins = dict(test_plugin_module=dict(verified=False, plugon='example')) raise djerr.DataJointError("hello") except djerr.DataJointError as e: p.discovered_plugins = curr_plugins assert(isinstance(e.__cause__, djerr.PluginWarning))
Make pubkey test more portable.
Make pubkey test more portable.
Python
lgpl-2.1
datajoint/datajoint-python,dimitri-yatsenko/datajoint-python
bc5475bcc3608de75c42d24c5c74e416b41b873f
pages/base.py
pages/base.py
# This Source Code Form is subject to the terms of the Mozilla Public # License, v. 2.0. If a copy of the MPL was not distributed with this file, # You can obtain one at http://mozilla.org/MPL/2.0/. from selenium.webdriver.common.by import By from page import Page class Base(Page): _login_locator = (By.ID, 'login') _logout_locator = (By.ID, 'logout') _notification_locator = (By.CLASS_NAME, 'flash') def click_login(self): self.selenium.find_element(*self._login_locator).click() from pages.login import LoginPage return LoginPage(self.testsetup) def click_logout(self): self.selenium.find_element(*self._logout_locator).click() def login(self, username=None, password=None): login_page = self.click_login() return login_page.login(username, password) def logout(self): self.click_logout() @property def notification(self): return self.selenium.find_element(*self._notification_locator).text
# This Source Code Form is subject to the terms of the Mozilla Public # License, v. 2.0. If a copy of the MPL was not distributed with this file, # You can obtain one at http://mozilla.org/MPL/2.0/. from selenium.webdriver.common.by import By from page import Page class Base(Page): _login_locator = (By.ID, 'login') _logout_locator = (By.ID, 'logout') _notification_locator = (By.CLASS_NAME, 'flash') def click_login(self): self.selenium.find_element(*self._login_locator).click() from pages.login import LoginPage return LoginPage(self.testsetup) def click_logout(self): self.selenium.find_element(*self._logout_locator).click() def login(self, username, password): login_page = self.click_login() return login_page.login(username, password) def logout(self): self.click_logout() @property def notification(self): return self.selenium.find_element(*self._notification_locator).text
Make username and password required arguments
Make username and password required arguments
Python
mpl-2.0
mozilla/mozwebqa-examples,davehunt/mozwebqa-examples,mozilla/mozwebqa-examples,davehunt/mozwebqa-examples
54bce2a224843ec9c1c8b7eb35cdc6bf19d5726b
expensonator/api.py
expensonator/api.py
from tastypie.authorization import Authorization from tastypie.fields import CharField from tastypie.resources import ModelResource from expensonator.models import Expense class ExpenseResource(ModelResource): tags = CharField() def dehydrate_tags(self, bundle): return bundle.obj.tags_as_string() def save(self, bundle, skip_errors=False): bundle = super(ExpenseResource, self).save(bundle, skip_errors) bundle.obj.reset_tags_from_string(bundle.data["tags"]) return bundle class Meta: queryset = Expense.objects.all() excludes = ["created", "updated"] # WARNING: Tastypie docs say that this is VERY INSECURE! # For development only! authorization = Authorization()
from tastypie.authorization import Authorization from tastypie.fields import CharField from tastypie.resources import ModelResource from expensonator.models import Expense class ExpenseResource(ModelResource): tags = CharField() def dehydrate_tags(self, bundle): return bundle.obj.tags_as_string() def save(self, bundle, skip_errors=False): bundle = super(ExpenseResource, self).save(bundle, skip_errors) if "tags" in bundle.data: bundle.obj.reset_tags_from_string(bundle.data["tags"]) return bundle class Meta: queryset = Expense.objects.all() excludes = ["created", "updated"] # WARNING: Tastypie docs say that this is VERY INSECURE! # For development only! authorization = Authorization()
Fix key error when no tags are specified
Fix key error when no tags are specified
Python
mit
matt-haigh/expensonator
f02b6505f190011f06b37619ec4fdf9bda1e804e
cea/interfaces/dashboard/api/utils.py
cea/interfaces/dashboard/api/utils.py
from flask import current_app import cea.inputlocator def deconstruct_parameters(p): params = {'name': p.name, 'type': p.typename, 'value': p.get(), 'help': p.help} try: params['choices'] = p._choices except AttributeError: pass if p.typename == 'WeatherPathParameter': config = current_app.cea_config locator = cea.inputlocator.InputLocator(config.scenario) params['choices'] = {wn: locator.get_weather( wn) for wn in locator.get_weather_names()} return params
from flask import current_app import cea.config import cea.inputlocator def deconstruct_parameters(p): params = {'name': p.name, 'type': p.typename, 'value': p.get(), 'help': p.help} if isinstance(p, cea.config.ChoiceParameter): params['choices'] = p._choices if p.typename == 'WeatherPathParameter': config = current_app.cea_config locator = cea.inputlocator.InputLocator(config.scenario) params['choices'] = {wn: locator.get_weather( wn) for wn in locator.get_weather_names()} elif p.typename == 'DatabasePathParameter': params['choices'] = p._choices return params
Add parameter deconstruction fro DatabasePathParameter
Add parameter deconstruction fro DatabasePathParameter
Python
mit
architecture-building-systems/CEAforArcGIS,architecture-building-systems/CEAforArcGIS
dfdeaf536466cfa8003af4cd5341d1d7127ea6b7
py/_test_py2go.py
py/_test_py2go.py
#!/usr/bin/env python import datetime def return_true(): return True def return_false(): return False def return_int(): return 123 def return_float(): return 1.0 def return_string(): return "ABC" def return_bytearray(): return bytearray('abcdefg') def return_array(): return [1, 2, {"key": 3}] def return_map(): return {"key1": 123, "key2": "str"} def return_nested_map(): return {"key1": {"key2": 123}} def return_none(): return None def return_timestamp(): return datetime.datetime(2015, 4, 1, 14, 27, 0, 500*1000, None)
#!/usr/bin/env python import datetime def return_true(): return True def return_false(): return False def return_int(): return 123 def return_float(): return 1.0 def return_string(): return "ABC" def return_bytearray(): return bytearray('abcdefg') def return_array(): return [1, 2, {"key": 3}] def return_map(): return {"key1": 123, "key2": "str"} def return_nested_map(): return {"key1": {"key2": 123}} def return_none(): return None def return_timestamp(): return datetime.datetime(2015, 4, 1, 14, 27, 0, 500*1000, None)
Update python script for pep8 style
Update python script for pep8 style
Python
mit
sensorbee/py,sensorbee/py
fee11dbff232216726516eea6c8bf7645fdef1a7
pyxif/__init__.py
pyxif/__init__.py
from ._remove import remove from ._load_and_dump import load, dump, ZerothIFD, ExifIFD, GPSIFD from ._transplant import transplant from ._insert import insert try: from ._thumbnail import thumbnail except ImportError: print("'thumbnail' function depends on PIL or Pillow.") VERSION = '0.4.7'
from ._remove import remove from ._load_and_dump import load, dump, ZerothIFD, ExifIFD, GPSIFD from ._transplant import transplant from ._insert import insert try: from ._thumbnail import thumbnail except ImportError: print("'thumbnail' function depends on PIL or Pillow.") VERSION = '0.4.6'
Revert "up version to 0.4.7."
Revert "up version to 0.4.7." This reverts commit 9b1177d4a56070092faa89778911d11c70efdc54.
Python
mit
hMatoba/Piexif
5d2dfa9f40f29ce7ddd23f8aff574c131539ed6c
util/versioncheck.py
util/versioncheck.py
#!/usr/bin/python from subprocess import check_output as co from sys import exit # Actually run bin/mn rather than importing via python path version = 'Mininet ' + co( 'PYTHONPATH=. bin/mn --version', shell=True ) version = version.strip() # Find all Mininet path references lines = co( "egrep -or 'Mininet [0-9\.\+]+\w*' *", shell=True ) error = False for line in lines.split( '\n' ): if line and 'Binary' not in line: fname, fversion = line.split( ':' ) if version != fversion: print( "%s: incorrect version '%s' (should be '%s')" % ( fname, fversion, version ) ) error = True if error: exit( 1 )
#!/usr/bin/python from subprocess import check_output as co from sys import exit # Actually run bin/mn rather than importing via python path version = 'Mininet ' + co( 'PYTHONPATH=. bin/mn --version 2>&1', shell=True ) version = version.strip() # Find all Mininet path references lines = co( "egrep -or 'Mininet [0-9\.\+]+\w*' *", shell=True ) error = False for line in lines.split( '\n' ): if line and 'Binary' not in line: fname, fversion = line.split( ':' ) if version != fversion: print( "%s: incorrect version '%s' (should be '%s')" % ( fname, fversion, version ) ) error = True if error: exit( 1 )
Handle version string sent to stderr
Handle version string sent to stderr An unfortunate side effect of switching from print to output() is that all output() goes to stderr. We should probably carefully consider whether this is the right thing to do.
Python
bsd-3-clause
mininet/mininet,mininet/mininet,mininet/mininet
caf9795cf0f775442bd0c3e06cd550a6e8d0206b
virtool/labels/db.py
virtool/labels/db.py
async def count_samples(db, label_id): return await db.samples.count_documents({"labels": {"$in": [label_id]}})
async def attach_sample_count(db, document, label_id): document.update({"count": await db.samples.count_documents({"labels": {"$in": [label_id]}})})
Rewrite function for sample count
Rewrite function for sample count
Python
mit
virtool/virtool,igboyes/virtool,virtool/virtool,igboyes/virtool
6dc90420dcd7dbfa787bd1e132cf5b304f72bfe7
likes/middleware.py
likes/middleware.py
try: from hashlib import md5 except ImportError: from md5 import md5 from django.http import HttpResponseBadRequest from secretballot.middleware import SecretBallotIpUseragentMiddleware class SecretBallotUserIpUseragentMiddleware(SecretBallotIpUseragentMiddleware): def generate_token(self, request): if request.user.is_authenticated(): return request.user.username else: try: s = ''.join((request.META['REMOTE_ADDR'], request.META['HTTP_USER_AGENT'])) return md5(s).hexdigest() except KeyError: return None
try: from hashlib import md5 except ImportError: from md5 import md5 from django.http import HttpResponseBadRequest from secretballot.middleware import SecretBallotIpUseragentMiddleware class SecretBallotUserIpUseragentMiddleware(SecretBallotIpUseragentMiddleware): def generate_token(self, request): if request.user.is_authenticated(): return request.user.username else: try: s = u''.join((request.META['REMOTE_ADDR'], request.META['HTTP_USER_AGENT'])) return md5(s).hexdigest() except KeyError: return None
Fix hashing for Python 3
Fix hashing for Python 3
Python
bsd-3-clause
Afnarel/django-likes,Afnarel/django-likes,Afnarel/django-likes
453730335b1e8d5d159350e0752faf282378f5e6
newsletter/models.py
newsletter/models.py
from django.db import models from django.utils import timezone from django.utils.datetime_safe import strftime from hashlib import sha256 from markdownx.models import MarkdownxField def generate_unsub_token(email, date): return sha256('{date}:{email}'.format(date=date, email=email).encode()).hexdigest() class Subscription(models.Model): email = models.EmailField() date_subscribed = models.DateTimeField(default=timezone.now) unsubscribe_token = models.CharField(max_length=64, blank=True) def save(self, *args, **kwargs): if not self.pk: self.unsubscribe_token = generate_unsub_token(self.email, self.date_subscribed) super(Subscription, self).save(*args, **kwargs) def __str__(self): return self.email class Mail(models.Model): subject = models.CharField(max_length=120) sender_name = models.CharField(max_length=50, default='UWCS Newsletter') sender_email = models.EmailField(default='[email protected]') text = MarkdownxField() date_created = models.DateTimeField(default=timezone.now) def __str__(self): return '{subject} - {date}'.format(date=strftime(self.date_created, '%Y-%m-%d'), subject=self.subject)
from django.db import models from django.utils import timezone from django.utils.datetime_safe import strftime from hashlib import sha256 from markdownx.models import MarkdownxField def generate_unsub_token(email, date): return sha256('{date}:{email}'.format(date=date, email=email).encode()).hexdigest() class Subscription(models.Model): email = models.EmailField() date_subscribed = models.DateTimeField(default=timezone.now) unsubscribe_token = models.CharField(max_length=64, blank=True) def save(self, *args, **kwargs): if not self.pk: self.unsubscribe_token = generate_unsub_token(self.email, self.date_subscribed) super(Subscription, self).save(*args, **kwargs) def __str__(self): return self.email class Mail(models.Model): subject = models.CharField(max_length=120) sender_name = models.CharField(max_length=50, default='UWCS Newsletter') sender_email = models.EmailField(default='[email protected]') text = MarkdownxField() date_created = models.DateTimeField(default=timezone.now) def __str__(self): return '{subject} - {date}'.format(date=strftime(self.date_created, '%Y-%m-%d'), subject=self.subject)
Change default email to newsletter@uwcs instead of noreply
Change default email to newsletter@uwcs instead of noreply
Python
mit
davidjrichardson/uwcs-zarya,davidjrichardson/uwcs-zarya
51e7cd3bc5a9a56fb53a5b0a8328d0b9d58848dd
modder/utils/desktop_notification.py
modder/utils/desktop_notification.py
# coding: utf-8 import platform if platform.system() == 'Darwin': from Foundation import NSUserNotificationDefaultSoundName import objc NSUserNotification = objc.lookUpClass('NSUserNotification') NSUserNotificationCenter = objc.lookUpClass('NSUserNotificationCenter') def desktop_notify(text, title='Modder', sound=False): notification = NSUserNotification.alloc().init() notification.setTitle_(title.decode('utf-8')) notification.setInformativeText_(text.decode('utf-8')) if sound: notification.setSoundName_(NSUserNotificationDefaultSoundName) center = NSUserNotificationCenter.defaultUserNotificationCenter() center.deliverNotification_(notification) elif platform.system() == 'Windows': def desktop_notify(text, title='Modder', sound=False): pass elif platform.system() == 'Linux': def desktop_notify(text, title='Modder', sound=False): pass
# coding: utf-8 import platform if platform.system() == 'Darwin': from Foundation import NSUserNotificationDefaultSoundName import objc NSUserNotification = objc.lookUpClass('NSUserNotification') NSUserNotificationCenter = objc.lookUpClass('NSUserNotificationCenter') def desktop_notify(text, title=None, sound=False): title = title or 'Modder' notification = NSUserNotification.alloc().init() notification.setTitle_(title.decode('utf-8')) notification.setInformativeText_(text.decode('utf-8')) if sound: notification.setSoundName_(NSUserNotificationDefaultSoundName) center = NSUserNotificationCenter.defaultUserNotificationCenter() center.deliverNotification_(notification) elif platform.system() == 'Windows': def desktop_notify(text, title=None, sound=False): title = title or 'Modder' pass elif platform.system() == 'Linux': def desktop_notify(text, title=None, sound=False): title = title or 'Modder' pass
Fix title for desktop notification
Fix title for desktop notification
Python
mit
JokerQyou/Modder2
925aa2ef91f15511ce7a3c97564f106d57d13623
djangopypi/templatetags/safemarkup.py
djangopypi/templatetags/safemarkup.py
from django import template from django.conf import settings from django.utils.encoding import smart_str, force_unicode from django.utils.safestring import mark_safe register = template.Library() def saferst(value): try: from docutils.core import publish_parts except ImportError: return force_unicode(value) docutils_setttings = getattr(settings, "RESTRUCTUREDTEXT_FILTER_SETTINGS", dict()) try: parts = publish_parts(source=smart_str(value), writer_name="html4css1", settings_overrides=docutils_settings) except: return foce_unicode(value) else: return mark_safe(force_unicode(parts["fragment"])) saferst.is_safe = True register.filter(saferst)
from django import template from django.conf import settings from django.utils.encoding import smart_str, force_unicode from django.utils.safestring import mark_safe register = template.Library() def saferst(value): try: from docutils.core import publish_parts except ImportError: return force_unicode(value) docutils_setttings = getattr(settings, "RESTRUCTUREDTEXT_FILTER_SETTINGS", dict()) try: parts = publish_parts(source=smart_str(value), writer_name="html4css1", settings_overrides=docutils_settings) except: return force_unicode(value) else: return mark_safe(force_unicode(parts["fragment"])) saferst.is_safe = True register.filter(saferst)
Fix typo foce_unicode -> force_unicode
Fix typo foce_unicode -> force_unicode
Python
bsd-3-clause
pitrho/djangopypi2,mattcaldwell/djangopypi,EightMedia/djangopypi,benliles/djangopypi,popen2/djangopypi2,disqus/djangopypi,ask/chishop,pitrho/djangopypi2,hsmade/djangopypi2,popen2/djangopypi2,hsmade/djangopypi2,disqus/djangopypi,EightMedia/djangopypi
02b67810263ac5a39882a1e12a78ba28249dbc0a
webapp/config/settings/development.py
webapp/config/settings/development.py
from .base import * DEBUG = True # TEMPLATES[1]['DEBUG'] = True DATABASES = { 'default': { # 'ENGINE': 'django.db.backends.sqlite3', # 'NAME': os.path.join(BASE_DIR, '..', 'tmp', 'db.sqlite3'), 'ENGINE': 'django.db.backends.postgresql_psycopg2', 'NAME': 'compass_webapp_dev', 'USER': 'compass_webapp', 'PASSWORD': 'password', 'HOST': 'localhost', 'PORT': '5432', } } DEVELOPMENT_APPS = [ 'debug_toolbar', ] INSTALLED_APPS += DEVELOPMENT_APPS
from .base import * DEBUG = True # TEMPLATES[1]['DEBUG'] = True DATABASES = { 'default': { 'ENGINE': 'django.db.backends.postgresql_psycopg2', 'NAME': 'compass_webapp_dev', 'USER': 'compass_webapp', 'PASSWORD': 'password', 'HOST': 'localhost', 'PORT': '5432', } } DEVELOPMENT_APPS = [ 'debug_toolbar', ] INSTALLED_APPS += DEVELOPMENT_APPS
Remove sql comments from settings file
Remove sql comments from settings file
Python
apache-2.0
patrickspencer/compass-python,patrickspencer/compass,patrickspencer/compass-python,patrickspencer/compass-python,patrickspencer/compass-python,patrickspencer/compass-python,patrickspencer/compass,patrickspencer/compass,patrickspencer/compass,patrickspencer/compass
8a7837a8ce7b35c3141374c6a5c99361261fa70a
Cura/avr_isp/chipDB.py
Cura/avr_isp/chipDB.py
avrChipDB = { 'ATMega2560': { 'signature': [0x1E, 0x98, 0x01], 'pageSize': 128, 'pageCount': 1024, }, } def getChipFromDB(sig): for chip in avrChipDB.values(): if chip['signature'] == sig: return chip return False
avrChipDB = { 'ATMega1280': { 'signature': [0x1E, 0x97, 0x03], 'pageSize': 128, 'pageCount': 512, }, 'ATMega2560': { 'signature': [0x1E, 0x98, 0x01], 'pageSize': 128, 'pageCount': 1024, }, } def getChipFromDB(sig): for chip in avrChipDB.values(): if chip['signature'] == sig: return chip return False
Add ATMega1280 chip to programmer chips.
Add ATMega1280 chip to programmer chips.
Python
agpl-3.0
MolarAmbiguity/OctoPrint,EZ3-India/EZ-Remote,JackGavin13/octoprint-test-not-finished,spapadim/OctoPrint,dragondgold/OctoPrint,hudbrog/OctoPrint,CapnBry/OctoPrint,Javierma/OctoPrint-TFG,chriskoz/OctoPrint,javivi001/OctoPrint,shohei/Octoprint,eddieparker/OctoPrint,MolarAmbiguity/OctoPrint,mayoff/OctoPrint,uuv/OctoPrint,C-o-r-E/OctoPrint,Mikk36/OctoPrint,DanLipsitt/OctoPrint,shohei/Octoprint,beeverycreative/BEEweb,alex1818/OctoPrint,EZ3-India/EZ-Remote,alex1818/OctoPrint,shohei/Octoprint,markwal/OctoPrint,beeverycreative/BEEweb,aerickson/OctoPrint,beeverycreative/BEEweb,aerickson/OctoPrint,nicanor-romero/OctoPrint,punkkeks/OctoPrint,d42/octoprint-fork,Javierma/OctoPrint-TFG,3dprintcanalhouse/octoprint2,ErikDeBruijn/OctoPrint,punkkeks/OctoPrint,masterhou/OctoPrint,shaggythesheep/OctoPrint,chriskoz/OctoPrint,madhuni/AstroBox,Catrodigious/OctoPrint-TAM,alephobjects/Cura,javivi001/OctoPrint,uuv/OctoPrint,leductan-nguyen/RaionPi,MoonshineSG/OctoPrint,eliasbakken/OctoPrint,nicanor-romero/OctoPrint,Skeen/OctoPrint,javivi001/OctoPrint,Salandora/OctoPrint,jneves/OctoPrint,hudbrog/OctoPrint,shaggythesheep/OctoPrint,MoonshineSG/OctoPrint,skieast/OctoPrint,abinashk-inf/AstroBox,nickverschoor/OctoPrint,eddieparker/OctoPrint,EZ3-India/EZ-Remote,EZ3-India/EZ-Remote,abinashk-inf/AstroBox,mrbeam/OctoPrint,abinashk-inf/AstroBox,mrbeam/OctoPrint,Voxel8/OctoPrint,sstocker46/OctoPrint,bicephale/OctoPrint,dragondgold/OctoPrint,Jaesin/OctoPrint,mcanes/OctoPrint,ryanneufeld/OctoPrint,Salandora/OctoPrint,CapnBry/OctoPrint,foosel/OctoPrint,nickverschoor/OctoPrint,alephobjects/Cura,mcanes/OctoPrint,markwal/OctoPrint,sstocker46/OctoPrint,Jaesin/OctoPrint,3dprintcanalhouse/octoprint1,skieast/OctoPrint,madhuni/AstroBox,markwal/OctoPrint,Mikk36/OctoPrint,AstroPrint/AstroBox,ymilord/OctoPrint-MrBeam,dansantee/OctoPrint,Jaesin/OctoPrint,punkkeks/OctoPrint,ymilord/OctoPrint-MrBeam,rurkowce/octoprint-fork,foosel/OctoPrint,Salandora/OctoPrint,spapadim/OctoPrint,MoonshineSG/OctoPrint,spapadim/OctoPrint,madhuni/AstroBox,masterhou/OctoPrint,ymilord/OctoPrint-MrBeam,alephobjects/Cura,ryanneufeld/OctoPrint,chriskoz/OctoPrint,hudbrog/OctoPrint,Mikk36/OctoPrint,eddieparker/OctoPrint,leductan-nguyen/RaionPi,JackGavin13/octoprint-test-not-finished,beeverycreative/BEEweb,bicephale/OctoPrint,nicanor-romero/OctoPrint,jneves/OctoPrint,JackGavin13/octoprint-test-not-finished,ErikDeBruijn/OctoPrint,leductan-nguyen/RaionPi,CapnBry/OctoPrint,chriskoz/OctoPrint,ryanneufeld/OctoPrint,3dprintcanalhouse/octoprint1,mrbeam/OctoPrint,senttech/OctoPrint,Javierma/OctoPrint-TFG,dansantee/OctoPrint,Voxel8/OctoPrint,bicephale/OctoPrint,MolarAmbiguity/OctoPrint,MaxOLydian/OctoPrint,eliasbakken/OctoPrint,DanLipsitt/OctoPrint,mayoff/OctoPrint,Skeen/OctoPrint,Jaesin/OctoPrint,rurkowce/octoprint-fork,CapnBry/OctoPrint,AstroPrint/AstroBox,madhuni/AstroBox,uuv/OctoPrint,abinashk-inf/AstroBox,JackGavin13/octoprint-test-not-finished,SeveQ/OctoPrint,sstocker46/OctoPrint,dansantee/OctoPrint,skieast/OctoPrint,mayoff/OctoPrint,C-o-r-E/OctoPrint,eliasbakken/OctoPrint,ryanneufeld/OctoPrint,foosel/OctoPrint,nickverschoor/OctoPrint,bicephale/OctoPrint,SeveQ/OctoPrint,MoonshineSG/OctoPrint,SeveQ/OctoPrint,senttech/OctoPrint,shohei/Octoprint,ymilord/OctoPrint-MrBeam,3dprintcanalhouse/octoprint2,d42/octoprint-fork,mcanes/OctoPrint,Voxel8/OctoPrint,senttech/OctoPrint,ymilord/OctoPrint-MrBeam,leductan-nguyen/RaionPi,Javierma/OctoPrint-TFG,Salandora/OctoPrint,C-o-r-E/OctoPrint,alex1818/OctoPrint,MaxOLydian/OctoPrint,shaggythesheep/OctoPrint,masterhou/OctoPrint,shohei/Octoprint,ErikDeBruijn/OctoPrint,jneves/OctoPrint,Catrodigious/OctoPrint-TAM,foosel/OctoPrint,dragondgold/OctoPrint,senttech/OctoPrint,aerickson/OctoPrint,MaxOLydian/OctoPrint,nickverschoor/OctoPrint,Skeen/OctoPrint,Catrodigious/OctoPrint-TAM,AstroPrint/AstroBox
ef96000b01c50a77b3500fc4071f83f96d7b2458
mrbelvedereci/api/views/cumulusci.py
mrbelvedereci/api/views/cumulusci.py
from django.shortcuts import render from mrbelvedereci.api.serializers.cumulusci import OrgSerializer from mrbelvedereci.api.serializers.cumulusci import ScratchOrgInstanceSerializer from mrbelvedereci.api.serializers.cumulusci import ServiceSerializer from mrbelvedereci.cumulusci.filters import OrgFilter from mrbelvedereci.cumulusci.filters import ScratchOrgInstanceFilter from mrbelvedereci.cumulusci.filters import ServiceFilter from mrbelvedereci.cumulusci.models import Org from mrbelvedereci.cumulusci.models import ScratchOrgInstance from mrbelvedereci.cumulusci.models import Service from rest_framework import viewsets class OrgViewSet(viewsets.ModelViewSet): """ A viewset for viewing and editing Orgs """ serializer_class = OrgSerializer queryset = Org.objects.all() filter_class = OrgFilter class ScratchOrgInstanceViewSet(viewsets.ModelViewSet): """ A viewset for viewing and editing ScratchOrgInstances """ serializer_class = ScratchOrgInstanceSerializer queryset = ScratchOrgInstance.objects.all() filter_class = ScratchOrgInstanceFilter class ServiceViewSet(viewsets.ModelViewSet): """ A viewset for viewing and editing Services """ serializer_class = ServiceSerializer queryset = Service.objects.all() filter_class = ServiceFilter
from django.shortcuts import render from mrbelvedereci.api.serializers.cumulusci import OrgSerializer from mrbelvedereci.api.serializers.cumulusci import ScratchOrgInstanceSerializer from mrbelvedereci.api.serializers.cumulusci import ServiceSerializer from mrbelvedereci.cumulusci.filters import OrgFilter from mrbelvedereci.cumulusci.filters import ScratchOrgInstanceFilter from mrbelvedereci.cumulusci.models import Org from mrbelvedereci.cumulusci.models import ScratchOrgInstance from mrbelvedereci.cumulusci.models import Service from rest_framework import viewsets class OrgViewSet(viewsets.ModelViewSet): """ A viewset for viewing and editing Orgs """ serializer_class = OrgSerializer queryset = Org.objects.all() filter_class = OrgFilter class ScratchOrgInstanceViewSet(viewsets.ModelViewSet): """ A viewset for viewing and editing ScratchOrgInstances """ serializer_class = ScratchOrgInstanceSerializer queryset = ScratchOrgInstance.objects.all() filter_class = ScratchOrgInstanceFilter class ServiceViewSet(viewsets.ModelViewSet): """ A viewset for viewing and editing Services """ serializer_class = ServiceSerializer queryset = Service.objects.all()
Remove ServiceFilter from view since it's not needed. Service only has name and json
Remove ServiceFilter from view since it's not needed. Service only has name and json
Python
bsd-3-clause
SalesforceFoundation/mrbelvedereci,SalesforceFoundation/mrbelvedereci,SalesforceFoundation/mrbelvedereci,SalesforceFoundation/mrbelvedereci
4f0dbf920a6867d8f3e16eb420391c8bcca43c44
onirim/card/_door.py
onirim/card/_door.py
from onirim.card._base import ColorCard class _Door(ColorCard): def drawn(self, agent, content): do_open = agent.ask("if open") if content.can_open(self) else False if do_open: content.discard(self) else: content.limbo(self) def door(color): return _Door(color)
from onirim.card._base import ColorCard from onirim.card._location import LocationKind def _openable(door_card, card): """Check if the door can be opened by another card.""" return card.kind == LocationKind.key and door_card.color == card.color def _may_open(door_card, content): """Check if the door may be opened by agent.""" return any(_openable(door_card, card) for card in content.hand()) class _Door(ColorCard): def drawn(self, agent, content): do_open = agent.ask("if open") if _may_open(self, content) else False if do_open: content.discard(self) else: content.limbo(self) def door(color): """Make a door card.""" return _Door(color)
Implement openable check for door card.
Implement openable check for door card.
Python
mit
cwahbong/onirim-py
24f0402e27ce7e51f370e82aa74c783438875d02
oslo_db/tests/sqlalchemy/__init__.py
oslo_db/tests/sqlalchemy/__init__.py
# Copyright (c) 2014 OpenStack Foundation # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from oslo_db.sqlalchemy import test_base load_tests = test_base.optimize_db_test_loader(__file__)
# Copyright (c) 2014 OpenStack Foundation # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from oslo_db.sqlalchemy import test_fixtures load_tests = test_fixtures.optimize_package_test_loader(__file__)
Remove deprecation warning when loading tests/sqlalchemy
Remove deprecation warning when loading tests/sqlalchemy /home/sam/Work/ironic/.tox/py27/local/lib/python2.7/site-packages/oslo_db/tests/sqlalchemy/__init__.py:20: DeprecationWarning: Function 'oslo_db.sqlalchemy.test_base.optimize_db_test_loader()' has moved to 'oslo_db.sqlalchemy.test_fixtures.optimize_package_test_loader()' Change-Id: I7fb4e776cedb8adcf97c9a43210049c60f796873
Python
apache-2.0
openstack/oslo.db,openstack/oslo.db
db6cb95d5d4261780482b4051f556fcbb2d9f237
rest_api/forms.py
rest_api/forms.py
from django.forms import ModelForm from rest_api.models import Url class UrlForm(ModelForm): class Meta: model = Url
from django.forms import ModelForm from gateway_backend.models import Url class UrlForm(ModelForm): class Meta: model = Url
Remove Url model from admin
Remove Url model from admin
Python
bsd-2-clause
victorpoluceno/shortener_frontend,victorpoluceno/shortener_frontend
3410fba1c8a39156def029eac9c7ff9f779832e6
dev/ci.py
dev/ci.py
# coding: utf-8 from __future__ import unicode_literals, division, absolute_import, print_function import os import site import sys from . import build_root, requires_oscrypto from ._import import _preload deps_dir = os.path.join(build_root, 'modularcrypto-deps') if os.path.exists(deps_dir): site.addsitedir(deps_dir) if sys.version_info[0:2] not in [(2, 6), (3, 2)]: from .lint import run as run_lint else: run_lint = None if sys.version_info[0:2] != (3, 2): from .coverage import run as run_coverage from .coverage import coverage run_tests = None else: from .tests import run as run_tests run_coverage = None def run(): """ Runs the linter and tests :return: A bool - if the linter and tests ran successfully """ _preload(requires_oscrypto, True) if run_lint: print('') lint_result = run_lint() else: lint_result = True if run_coverage: print('\nRunning tests (via coverage.py %s)' % coverage.__version__) sys.stdout.flush() tests_result = run_coverage(ci=True) else: print('\nRunning tests') sys.stdout.flush() tests_result = run_tests(ci=True) sys.stdout.flush() return lint_result and tests_result
# coding: utf-8 from __future__ import unicode_literals, division, absolute_import, print_function import os import site import sys from . import build_root, requires_oscrypto from ._import import _preload deps_dir = os.path.join(build_root, 'modularcrypto-deps') if os.path.exists(deps_dir): site.addsitedir(deps_dir) # In case any of the deps are installed system-wide sys.path.insert(0, deps_dir) if sys.version_info[0:2] not in [(2, 6), (3, 2)]: from .lint import run as run_lint else: run_lint = None if sys.version_info[0:2] != (3, 2): from .coverage import run as run_coverage from .coverage import coverage run_tests = None else: from .tests import run as run_tests run_coverage = None def run(): """ Runs the linter and tests :return: A bool - if the linter and tests ran successfully """ _preload(requires_oscrypto, True) if run_lint: print('') lint_result = run_lint() else: lint_result = True if run_coverage: print('\nRunning tests (via coverage.py %s)' % coverage.__version__) sys.stdout.flush() tests_result = run_coverage(ci=True) else: print('\nRunning tests') sys.stdout.flush() tests_result = run_tests(ci=True) sys.stdout.flush() return lint_result and tests_result
Fix CI to ignore system install of asn1crypto
Fix CI to ignore system install of asn1crypto
Python
mit
wbond/oscrypto
502d99042428175b478e796c067e41995a0ae5bf
picoCTF-web/api/apps/v1/__init__.py
picoCTF-web/api/apps/v1/__init__.py
"""picoCTF API v1 app.""" from flask import Blueprint, jsonify from flask_restplus import Api from api.common import PicoException from .achievements import ns as achievements_ns from .problems import ns as problems_ns from .shell_servers import ns as shell_servers_ns from .exceptions import ns as exceptions_ns from .settings import ns as settings_ns from .bundles import ns as bundles_ns from .submissions import ns as submissions_ns from .feedback import ns as feedback_ns blueprint = Blueprint('v1_api', __name__) api = Api( app=blueprint, title='picoCTF API', version='1.0', ) api.add_namespace(achievements_ns) api.add_namespace(problems_ns) api.add_namespace(shell_servers_ns) api.add_namespace(exceptions_ns) api.add_namespace(settings_ns) api.add_namespace(bundles_ns) api.add_namespace(submissions_ns) api.add_namespace(feedback_ns) @api.errorhandler(PicoException) def handle_pico_exception(e): """Handle exceptions.""" response = jsonify(e.to_dict()) response.status_code = 203 return response
"""picoCTF API v1 app.""" from flask import Blueprint, jsonify from flask_restplus import Api from api.common import PicoException from .achievements import ns as achievements_ns from .problems import ns as problems_ns from .shell_servers import ns as shell_servers_ns from .exceptions import ns as exceptions_ns from .settings import ns as settings_ns from .bundles import ns as bundles_ns from .submissions import ns as submissions_ns from .feedback import ns as feedback_ns blueprint = Blueprint('v1_api', __name__) api = Api( app=blueprint, title='picoCTF API', version='1.0', ) api.add_namespace(achievements_ns) api.add_namespace(problems_ns) api.add_namespace(shell_servers_ns) api.add_namespace(exceptions_ns) api.add_namespace(settings_ns) api.add_namespace(bundles_ns) api.add_namespace(submissions_ns) api.add_namespace(feedback_ns) @api.errorhandler(PicoException) def handle_pico_exception(e): """Handle exceptions.""" response = jsonify(e.to_dict()) response.status_code = e.status_code return response
Fix PicoException response code bug
Fix PicoException response code bug
Python
mit
royragsdale/picoCTF,picoCTF/picoCTF,royragsdale/picoCTF,picoCTF/picoCTF,picoCTF/picoCTF,royragsdale/picoCTF,royragsdale/picoCTF,royragsdale/picoCTF,picoCTF/picoCTF,royragsdale/picoCTF,picoCTF/picoCTF,picoCTF/picoCTF,royragsdale/picoCTF
5d71215645683a059a51407a3768054c9ea77406
pisite/logs/forms.py
pisite/logs/forms.py
from django import forms from logs.models import Log class LineCountForm(forms.Form): linesToFetch = forms.IntegerField(label="Number of lines to show", min_value=0, initial=Log.defaultLinesToShow)
from django import forms from logs.models import Log class LineCountForm(forms.Form): linesToFetch = forms.IntegerField(label="Number of lines to show (0 for all)", min_value=0, initial=Log.defaultLinesToShow)
Add to the label that 0 lines will result in the entire file being downloaded
Add to the label that 0 lines will result in the entire file being downloaded
Python
mit
sizlo/RPiFun,sizlo/RPiFun
94dad4c56a4b6a1968fa15c20b8482fd56774f32
optimize/py/main.py
optimize/py/main.py
from scipy import optimize as o import clean as c def minimize(func, guess): return o.minimize(func, guess) def minimize_scalar(func, options): bracket = options['bracket'] bounds = options['bounds'] method = options['method'] tol = options['tol'] options = options['options'] return o.minimize_scalar(func, bracket=bracket, bounds=bounds, method=method, tol=tol, options=options)
from scipy import optimize as o import numpy as np import clean as c def minimize_scalar(func, options): bracket = options['bracket'] bounds = options['bounds'] method = options['method'] tol = options['tol'] options = options['options'] try: return o.minimize_scalar(func, bracket=bracket, bounds=bounds, method=method, tol=tol, options=options) except Exception as e: return str(e) def nnls(A, b): try: solution, residual = o.nnls(np.array(A), np.array(b)) return { 'solution': solution.tolist(), 'residual': residual } except Exception as e: return str(e)
Add non negative least squares scipy functionality
Add non negative least squares scipy functionality
Python
mit
acjones617/scipy-node,acjones617/scipy-node
a389f20c7f2c8811a5c2f50c43a9ce5c7f3c8387
jobs_backend/vacancies/serializers.py
jobs_backend/vacancies/serializers.py
from rest_framework import serializers from .models import Vacancy class VacancySerializer(serializers.HyperlinkedModelSerializer): """ Common vacancy model serializer """ class Meta: model = Vacancy fields = ( 'id', 'url', 'title', 'description', 'created_on', 'modified_on' ) extra_kwargs = { 'url': {'view_name': 'vacancies:vacancy-detail', 'read_only': True} }
from rest_framework import serializers from .models import Vacancy class VacancySerializer(serializers.ModelSerializer): """ Common vacancy model serializer """ class Meta: model = Vacancy fields = ( 'id', 'url', 'title', 'description', 'created_on', 'modified_on' ) extra_kwargs = { 'url': {'view_name': 'api:vacancies:vacancy-detail', 'read_only': True} }
Fix for correct resolve URL
jobs-010: Fix for correct resolve URL
Python
mit
pyshopml/jobs-backend,pyshopml/jobs-backend
cdd6bc5258a21a1447c6313fad87056163b58a45
product_details/settings_defaults.py
product_details/settings_defaults.py
import logging import os # URL to clone product_details JSON files from. # Include trailing slash. PROD_DETAILS_URL = 'http://svn.mozilla.org/libs/product-details/json/' # Target dir to drop JSON files into (must be writable) PROD_DETAILS_DIR = os.path.join(os.path.dirname(__file__), 'json') # log level. LOG_LEVEL = logging.INFO # name of cache to use PROD_DETAILS_CACHE_NAME = 'default' # django default # how long to cache PROD_DETAILS_CACHE_TIMEOUT = 60 * 60 * 12 # 12 hours # data storage class PROD_DETAILS_STORAGE = 'product_details.storage.PDFileStorage'
import logging import os # URL to clone product_details JSON files from. # Include trailing slash. PROD_DETAILS_URL = 'https://svn.mozilla.org/libs/product-details/json/' # Target dir to drop JSON files into (must be writable) PROD_DETAILS_DIR = os.path.join(os.path.dirname(__file__), 'json') # log level. LOG_LEVEL = logging.INFO # name of cache to use PROD_DETAILS_CACHE_NAME = 'default' # django default # how long to cache PROD_DETAILS_CACHE_TIMEOUT = 60 * 60 * 12 # 12 hours # data storage class PROD_DETAILS_STORAGE = 'product_details.storage.PDFileStorage'
Use HTTPS when fetching JSON files
Use HTTPS when fetching JSON files
Python
bsd-3-clause
mozilla/django-product-details
441a1b85f6ab954ab89f32977e4f00293270aac6
sphinxcontrib/multilatex/__init__.py
sphinxcontrib/multilatex/__init__.py
import directive import builder #=========================================================================== # Node visitor functions def visit_passthrough(self, node): pass def depart_passthrough(self, node): pass passthrough = (visit_passthrough, depart_passthrough) #=========================================================================== # Setup and register extension def setup(app): app.add_node(directive.latex_document, html=passthrough) app.add_directive("latex-document", directive.LatexDocumentDirective) app.add_builder(builder.MultiLatexBuilder) return {"version": "0.0"}
import directive import builder #=========================================================================== # Node visitor functions def visit_passthrough(self, node): pass def depart_passthrough(self, node): pass passthrough = (visit_passthrough, depart_passthrough) #=========================================================================== # Setup and register extension def setup(app): app.add_node(directive.latex_document, latex=passthrough, html=passthrough) app.add_directive("latex-document", directive.LatexDocumentDirective) app.add_builder(builder.MultiLatexBuilder) return {"version": "0.0"}
Set LaTeX builder to skip latex_document nodes
Set LaTeX builder to skip latex_document nodes This stops Sphinx' built-in LaTeX builder from complaining about unknown latex_document node type.
Python
apache-2.0
t4ngo/sphinxcontrib-multilatex,t4ngo/sphinxcontrib-multilatex
84f7fe2d17a82d095ff6cf4f2bbd13a2a8426c2d
go/contacts/urls.py
go/contacts/urls.py
from django.conf.urls.defaults import patterns, url from go.contacts import views urlpatterns = patterns('', url(r'^$', views.index, name='index'), url(r'^groups/$', views.groups, name='groups'), url(r'^groups/(?P<type>[\w ]+)/$', views.groups, name='groups_type'), # TODO: Is the group_name regex sane? url(r'^group/(?P<group_key>[\w ]+)/$', views.group, name='group'), url(r'^people/$', views.people, name='people'), url(r'^people/new/$', views.new_person, name='new_person'), url(r'^people/(?P<person_key>\w+)/$', views.person, name='person'), )
from django.conf.urls.defaults import patterns, url from go.contacts import views urlpatterns = patterns('', url(r'^$', views.index, name='index'), url(r'^groups/$', views.groups, name='groups', kwargs={'type': 'static'}), url(r'^groups/(?P<type>[\w ]+)/$', views.groups, name='groups_type'), # TODO: Is the group_name regex sane? url(r'^group/(?P<group_key>[\w ]+)/$', views.group, name='group'), url(r'^people/$', views.people, name='people'), url(r'^people/new/$', views.new_person, name='new_person'), url(r'^people/(?P<person_key>\w+)/$', views.person, name='person'), )
Change /contacts/groups to display only static groups (instead of all groups)
Change /contacts/groups to display only static groups (instead of all groups)
Python
bsd-3-clause
praekelt/vumi-go,praekelt/vumi-go,praekelt/vumi-go,praekelt/vumi-go
5c11a65af1d51794133895ebe2de92861b0894cf
flask_limiter/errors.py
flask_limiter/errors.py
"""errors and exceptions.""" from distutils.version import LooseVersion from pkg_resources import get_distribution from six import text_type from werkzeug import exceptions werkzeug_exception = None werkzeug_version = get_distribution("werkzeug").version if LooseVersion(werkzeug_version) < LooseVersion("0.9"): # pragma: no cover # sorry, for touching your internals :). import werkzeug._internal werkzeug._internal.HTTP_STATUS_CODES[429] = "Too Many Requests" werkzeug_exception = exceptions.HTTPException else: # Werkzeug 0.9 and up have an existing exception for 429 werkzeug_exception = exceptions.TooManyRequests class RateLimitExceeded(werkzeug_exception): """exception raised when a rate limit is hit. The exception results in ``abort(429)`` being called. """ code = 429 limit = None def __init__(self, limit): self.limit = limit if limit.error_message: description = ( limit.error_message if not callable(limit.error_message) else limit.error_message() ) else: description = text_type(limit.limit) super(RateLimitExceeded, self).__init__(description=description)
"""errors and exceptions.""" from distutils.version import LooseVersion from pkg_resources import get_distribution from six import text_type from werkzeug import exceptions class RateLimitExceeded(exceptions.TooManyRequests): """exception raised when a rate limit is hit. The exception results in ``abort(429)`` being called. """ code = 429 limit = None def __init__(self, limit): self.limit = limit if limit.error_message: description = ( limit.error_message if not callable(limit.error_message) else limit.error_message() ) else: description = text_type(limit.limit) super(RateLimitExceeded, self).__init__(description=description)
Remove backward compatibility hack for exception subclass
Remove backward compatibility hack for exception subclass
Python
mit
alisaifee/flask-limiter,alisaifee/flask-limiter
b3979a46a7bcd71aa9b40892167910fdeed5ad97
frigg/projects/admin.py
frigg/projects/admin.py
from django.contrib import admin from django.template.defaultfilters import pluralize from .forms import EnvironmentVariableForm from .models import EnvironmentVariable, Project class EnvironmentVariableMixin: form = EnvironmentVariableForm @staticmethod def get_readonly_fields(request, obj=None): if obj: return 'key', 'value', 'is_secret' class EnvironmentVariableInline(EnvironmentVariableMixin, admin.TabularInline): model = EnvironmentVariable extra = 0 @admin.register(Project) class ProjectAdmin(admin.ModelAdmin): list_display = ('__str__', 'queue_name', 'approved', 'number_of_members', 'average_time', 'last_build_number', 'can_deploy') list_filter = ['owner', 'queue_name', 'approved', 'can_deploy'] actions = ['sync_members'] inlines = [EnvironmentVariableInline] def sync_members(self, request, queryset): for project in queryset: project.update_members() self.message_user( request, '{} project{} was synced'.format(len(queryset), pluralize(len(queryset))) ) sync_members.short_description = 'Sync members of selected projects' @admin.register(EnvironmentVariable) class EnvironmentVariableAdmin(EnvironmentVariableMixin, admin.ModelAdmin): list_display = ( '__str__', 'is_secret', )
from django.contrib import admin from django.template.defaultfilters import pluralize from .forms import EnvironmentVariableForm from .models import EnvironmentVariable, Project class EnvironmentVariableMixin: form = EnvironmentVariableForm @staticmethod def get_readonly_fields(request, obj=None): if obj: return 'key', 'value', 'is_secret' return tuple() class EnvironmentVariableInline(EnvironmentVariableMixin, admin.TabularInline): model = EnvironmentVariable extra = 0 @admin.register(Project) class ProjectAdmin(admin.ModelAdmin): list_display = ('__str__', 'queue_name', 'approved', 'number_of_members', 'average_time', 'last_build_number', 'can_deploy') list_filter = ['owner', 'queue_name', 'approved', 'can_deploy'] actions = ['sync_members'] inlines = [EnvironmentVariableInline] def sync_members(self, request, queryset): for project in queryset: project.update_members() self.message_user( request, '{} project{} was synced'.format(len(queryset), pluralize(len(queryset))) ) sync_members.short_description = 'Sync members of selected projects' @admin.register(EnvironmentVariable) class EnvironmentVariableAdmin(EnvironmentVariableMixin, admin.ModelAdmin): list_display = ( '__str__', 'is_secret', )
Return empty tuple in get_readonly_fields
fix: Return empty tuple in get_readonly_fields
Python
mit
frigg/frigg-hq,frigg/frigg-hq,frigg/frigg-hq
0d7c0b045c4a2e930fe0d7aa68b96d5a99916a34
scripts/document_path_handlers.py
scripts/document_path_handlers.py
#!/usr/bin/env python from __future__ import print_function, unicode_literals from nikola import nikola n = nikola.Nikola() n.init_plugins() print(""".. title: Path Handlers for Nikola .. slug: path-handlers .. author: The Nikola Team Nikola supports special links with the syntax ``link://kind/name``. Here is the description for all the supported kinds. """) for k in sorted(n.path_handlers.keys()): v = n.path_handlers[k] print(k) print('\n'.join(' '+l.strip() for l in v.__doc__.splitlines())) print()
#!/usr/bin/env python from __future__ import print_function, unicode_literals from nikola import nikola n = nikola.Nikola() n.init_plugins() print(""".. title: Path Handlers for Nikola .. slug: path-handlers .. author: The Nikola Team Nikola supports special links with the syntax ``link://kind/name``. Here is the description for all the supported kinds. .. class:: dl-horizontal """) for k in sorted(n.path_handlers.keys()): v = n.path_handlers[k] print(k) print('\n'.join(' '+l.strip() for l in v.__doc__.splitlines())) print()
Make path handlers list horizontal
Make path handlers list horizontal Signed-off-by: Chris Warrick <[email protected]>
Python
mit
s2hc-johan/nikola,wcmckee/nikola,gwax/nikola,x1101/nikola,okin/nikola,masayuko/nikola,xuhdev/nikola,wcmckee/nikola,gwax/nikola,knowsuchagency/nikola,atiro/nikola,andredias/nikola,gwax/nikola,xuhdev/nikola,atiro/nikola,x1101/nikola,okin/nikola,knowsuchagency/nikola,wcmckee/nikola,okin/nikola,getnikola/nikola,masayuko/nikola,okin/nikola,getnikola/nikola,masayuko/nikola,andredias/nikola,atiro/nikola,xuhdev/nikola,xuhdev/nikola,s2hc-johan/nikola,getnikola/nikola,knowsuchagency/nikola,getnikola/nikola,x1101/nikola,andredias/nikola,s2hc-johan/nikola
c6d50c3feed444f8f450c5c140e8470c6897f2bf
societies/models.py
societies/models.py
# -*- coding: utf-8 -*- from django.db import models from django_countries.fields import CountryField class GuitarSociety(models.Model): """ Represents a single guitar society. .. versionadded:: 0.1 """ #: the name of the society #: ..versionadded:: 0.1 name = models.CharField(max_length=1024) #: the society's url #: ..versionadded:: 0.1 link = models.URLField(max_length=255) #: The country in which the society resides #: .. versionadded:: 0.1 country = CountryField() #: A free form "city" or "region" field used to display where #: exactly the society is within a country #: .. versionadded:: 0.1 region = models.CharField(max_length=512, null=True, default=None, blank=True) def __str__(self): return 'GuitarSociety(name="{}", link="{}")'.format(self.name, self.link)
# -*- coding: utf-8 -*- from django.db import models from django_countries.fields import CountryField class GuitarSociety(models.Model): """ Represents a single guitar society. .. versionadded:: 0.1 """ #: the name of the society #: ..versionadded:: 0.1 name = models.CharField(max_length=1024) #: the society's url #: ..versionadded:: 0.1 link = models.URLField(max_length=255) #: The country in which the society resides #: .. versionadded:: 0.1 country = CountryField() #: A free form "city" or "region" field used to display where #: exactly the society is within a country #: .. versionadded:: 0.1 region = models.CharField(max_length=512, null=True, default=None, blank=True) def __str__(self): return self.name def __repr__(self): return 'GuitarSociety("{}")'.format(self.name)
Make the Guitar Society __str__ Method a bit more Logical
Make the Guitar Society __str__ Method a bit more Logical
Python
bsd-3-clause
chrisguitarguy/GuitarSocieties.org,chrisguitarguy/GuitarSocieties.org
c7a209d2c4455325f1d215ca1c12074b394ae00e
gitdir/host/__init__.py
gitdir/host/__init__.py
import abc import subprocess import gitdir class Host(abc.ABC): @abc.abstractmethod def __iter__(self): raise NotImplementedError() @abc.abstractmethod def __str__(self): raise NotImplementedError() def clone(self, repo_spec): raise NotImplementedError('Host {} does not support cloning'.format(self)) @property def dir(self): return gitdir.GITDIR / str(self) def update(self): for repo_dir in self: subprocess.check_call(['git', 'pull'], cwd=str(repo_dir / 'master')) def all(): for host_dir in gitdir.GITDIR.iterdir(): yield by_name(host_dir.name) def by_name(hostname): if hostname == 'github.com': import gitdir.host.github return gitdir.host.github.GitHub() else: raise ValueError('Unsupported hostname: {}'.format(hostname))
import abc import subprocess import gitdir class Host(abc.ABC): @abc.abstractmethod def __iter__(self): raise NotImplementedError() @abc.abstractmethod def __str__(self): raise NotImplementedError() def clone(self, repo_spec): raise NotImplementedError('Host {} does not support cloning'.format(self)) @property def dir(self): return gitdir.GITDIR / str(self) def update(self): for repo_dir in self: print('[ ** ] updating {}'.format(repo_dir)) subprocess.check_call(['git', 'pull'], cwd=str(repo_dir / 'master')) def all(): for host_dir in gitdir.GITDIR.iterdir(): yield by_name(host_dir.name) def by_name(hostname): if hostname == 'github.com': import gitdir.host.github return gitdir.host.github.GitHub() else: raise ValueError('Unsupported hostname: {}'.format(hostname))
Add status messages to `gitdir update`
Add status messages to `gitdir update`
Python
mit
fenhl/gitdir
051a0ae16f7c387fcab55abff7debb4e0985654e
senlin/db/sqlalchemy/migration.py
senlin/db/sqlalchemy/migration.py
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import os from oslo.db.sqlalchemy import migration as oslo_migration INIT_VERSION = 1 def db_sync(engine, version=None): path = os.path.join(os.path.abspath(os.path.dirname(__file__)), 'migrate_repo') return oslo_migration.db_sync(engine, path, version, init_version=INIT_VERSION) def db_version(engine): path = os.path.join(os.path.abspath(os.path.dirname(__file__)), 'migrate_repo') return oslo_migration.db_version(engine, path, INIT_VERSION) def db_version_control(engine, version=None): path = os.path.join(os.path.abspath(os.path.dirname(__file__)), 'migrate_repo') return oslo_migration.db_version_control(engine, path, version)
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import os from oslo.db.sqlalchemy import migration as oslo_migration INIT_VERSION = 0 def db_sync(engine, version=None): path = os.path.join(os.path.abspath(os.path.dirname(__file__)), 'migrate_repo') return oslo_migration.db_sync(engine, path, version, init_version=INIT_VERSION) def db_version(engine): path = os.path.join(os.path.abspath(os.path.dirname(__file__)), 'migrate_repo') return oslo_migration.db_version(engine, path, INIT_VERSION) def db_version_control(engine, version=None): path = os.path.join(os.path.abspath(os.path.dirname(__file__)), 'migrate_repo') return oslo_migration.db_version_control(engine, path, version)
Make 0 the default version
Make 0 the default version
Python
apache-2.0
tengqm/senlin-container,stackforge/senlin,openstack/senlin,stackforge/senlin,Alzon/senlin,openstack/senlin,tengqm/senlin,Alzon/senlin,openstack/senlin,tengqm/senlin-container,tengqm/senlin
11278ec546cf1c84a6aefff7ed4e5a677203d008
index_addresses.py
index_addresses.py
import csv import re import os from urlparse import urlparse from elasticsearch import Elasticsearch if os.environ.get('BONSAI_URL'): url = urlparse(os.environ['BONSAI_URL']) bonsai_tuple = url.netloc.partition('@') ELASTICSEARCH_HOST = bonsai_tuple[2] ELASTICSEARCH_AUTH = bonsai_tuple[0] es = Elasticsearch([{'host': ELASTICSEARCH_HOST}], http_auth=ELASTICSEARCH_AUTH) else: es = Elasticsearch() with open('data/ParcelCentroids.csv', 'r') as csvfile: print "open file" csv_reader = csv.DictReader(csvfile, fieldnames=[], restkey='undefined-fieldnames', delimiter=',') current_row = 0 for row in csv_reader: current_row += 1 if current_row == 1: csv_reader.fieldnames = row['undefined-fieldnames'] continue address = row if re.match('\d+', address['PVANUM']): es.index(index='addresses', doc_type='address', id=address['PVANUM'], body={'PVANUM': address['PVANUM'], 'NUM1': address['NUM1'], 'NAME': address['NAME'], 'TYPE': address['TYPE'], 'ADDRESS': address['ADDRESS'], 'UNIT': address['UNIT'], 'X': address['X'], 'Y': address['Y']}) csvfile.close()
import sys import csv import re import os from urlparse import urlparse from elasticsearch import Elasticsearch if os.environ.get('BONSAI_URL'): url = urlparse(os.environ['BONSAI_URL']) bonsai_tuple = url.netloc.partition('@') ELASTICSEARCH_HOST = bonsai_tuple[2] ELASTICSEARCH_AUTH = bonsai_tuple[0] es = Elasticsearch([{'host': ELASTICSEARCH_HOST}], http_auth=ELASTICSEARCH_AUTH) else: es = Elasticsearch() files_given = sys.argv for file_name in files_given: if file_name = 'index_addresses.py': continue else: file_path = file_name print 'adding ' + file_path with open(file_path, 'r') as csvfile: print "open file" csv_reader = csv.DictReader(csvfile, fieldnames=[], restkey='undefined-fieldnames', delimiter=',') current_row = 0 for row in csv_reader: current_row += 1 if current_row == 1: csv_reader.fieldnames = row['undefined-fieldnames'] continue address = row es.index(index='addresses', doc_type='address', id=current_row-1, body={'NUMBER': address[' NUMBER'], 'STREET': address[' STREET'], 'ADDRESS': address[' NUMBER'] + ' ' + address[' STREET'], 'X': address['LON'], 'Y': address[' LAT']}) csvfile.close()
Change index to OpenAddresses schema
Change index to OpenAddresses schema
Python
mit
codeforamerica/streetscope,codeforamerica/streetscope
461ad75cdb5b8d1a514ff781fd021b33cfd5aa3d
constants.py
constants.py
from __future__ import ( absolute_import, print_function, ) POSTGRES_MAX_INT = 2147483647 # class statuses STATUS_OPEN = 0 STATUS_FULL = 1 STATUS_CLOSED = 2 STATUS_TENTATIVE = 3 STATUS_CANCELLED = 4 STATUS_STOPPED = 5 # semesters SUMMER_SEM = 0 SEMESTER_1 = 1 SEMESTER_2 = 2 CURRENT_SEM = SEMESTER_1 # contact types CONTACT_TYPE_EMAIL = 0 CONTACT_TYPE_SMS = 1 CONTACT_TYPE_YO = 2 MAX_SEARCH_RESULTS = 5
from __future__ import ( absolute_import, print_function, ) POSTGRES_MAX_INT = 2147483647 # class statuses STATUS_OPEN = 0 STATUS_FULL = 1 STATUS_CLOSED = 2 STATUS_TENTATIVE = 3 STATUS_CANCELLED = 4 STATUS_STOPPED = 5 # semesters SUMMER_SEM = 0 SEMESTER_1 = 1 SEMESTER_2 = 2 CURRENT_SEM = SEMESTER_2 # contact types CONTACT_TYPE_EMAIL = 0 CONTACT_TYPE_SMS = 1 CONTACT_TYPE_YO = 2 MAX_SEARCH_RESULTS = 5
Revert "emails now sent through mailgun, small warning css change"
Revert "emails now sent through mailgun, small warning css change" This reverts commit 12ffeb9562bb9e865fe3ce76266ba3f5c45b815d.
Python
mit
Chybby/Tutorifull,Chybby/Tutorifull,Chybby/Tutorifull
9a3c134de0c1c7d194f3d7e30bd9cea917154cba
gitver/sanity.py
gitver/sanity.py
#!/usr/bin/env python2 # coding=utf-8 """ Implements reused sanity checks """ import os import sys from gitver.termcolors import term, bold from gitver.defines import PRJ_ROOT, CFGDIR, CFGDIRNAME, GITIGNOREFILE def check_project_root(): if len(PRJ_ROOT) == 0: term.err("Couldn't determine your project's root directory, is this " "a valid git repository?") sys.exit(1) def check_config(): # check config directory exists if not os.path.exists(CFGDIR): term.prn("Please run " + bold("gitver init") + " first.") sys.exit(1) def check_gitignore(exit_on_error=True): # check .gitignore for .gitver inclusion try: gifile = os.path.join(GITIGNOREFILE) with open(gifile, 'r') as f: if CFGDIRNAME in f.read(): return except IOError: pass term.warn("Warning: it's highly recommended to EXCLUDE the gitver " "configuration from the repository!") term.prn("Please include the following line in your .gitignore file:") term.prn(" " + CFGDIRNAME) if exit_on_error: sys.exit(1) else: print ""
#!/usr/bin/env python2 # coding=utf-8 """ Implements reused sanity checks """ import os import sys from gitver.termcolors import term, bold from gitver.defines import PRJ_ROOT, CFGDIR, CFGDIRNAME, GITIGNOREFILE def check_project_root(): if len(PRJ_ROOT) == 0: term.err("Couldn't determine your project's root directory, is this " "a valid git repository?") sys.exit(1) def check_config(): # check config directory exists if not os.path.exists(CFGDIR): term.err("Please run " + bold("gitver init") + " first.") sys.exit(1) def check_gitignore(exit_on_error=True): # check .gitignore for .gitver inclusion try: gifile = os.path.join(GITIGNOREFILE) with open(gifile, 'r') as f: if CFGDIRNAME in f.read(): return except IOError: pass term.warn("Warning: it's highly recommended to EXCLUDE the gitver " "configuration from the repository!") term.prn("Please include the following line in your .gitignore file:") term.prn(" " + CFGDIRNAME) if exit_on_error: sys.exit(1) else: print ""
Fix wrong output stream usage
Fix wrong output stream usage
Python
apache-2.0
manuelbua/gitver,manuelbua/gitver,manuelbua/gitver
b2a7b299e38ca2cb0d1a725fcfbf6d5d73fa1dfc
dame/dame.py
dame/dame.py
__author__ = "Richard Lindsley" import sys, os import argparse import sip sip.setapi('QDate', 2) sip.setapi('QDateTime', 2) sip.setapi('QString', 2) sip.setapi('QTextStream', 2) sip.setapi('QTime', 2) sip.setapi('QUrl', 2) sip.setapi('QVariant', 2) from PyQt4 import QtCore,QtGui from . import __version__ from .ui.mainwindow import MainWindow from loadsir import loadsir #class Dame(QtGui.QApplication): # def __init__(self, args): # QtGui.QApplication.__init__(self, args) # self._args = args def main(): parser = argparse.ArgumentParser(description="View SIR file(s)") #parser.add_argument("sir_files", action="store", nargs='*', parser.add_argument("sir_files", action="store", nargs=1, help='List of SIR files to open') parser.add_argument('--version', action='version', version='%(prog)s version {}'.format(__version__)) args = parser.parse_args() qt_app = QtGui.QApplication(sys.argv) frame = MainWindow() frame.show() # TODO: Handle multiple files if len(args.sir_files) > 0: frame.load_sir(args.sir_files[0]) qt_app.exec_() if __name__ == "__main__": main()
__author__ = "Richard Lindsley" import sys, os import argparse import sip sip.setapi('QDate', 2) sip.setapi('QDateTime', 2) sip.setapi('QString', 2) sip.setapi('QTextStream', 2) sip.setapi('QTime', 2) sip.setapi('QUrl', 2) sip.setapi('QVariant', 2) from PyQt4 import QtCore,QtGui from . import __version__ from .ui.mainwindow import MainWindow from loadsir import loadsir #class Dame(QtGui.QApplication): # def __init__(self, args): # QtGui.QApplication.__init__(self, args) # self._args = args def main(): parser = argparse.ArgumentParser(description="View SIR file(s)") parser.add_argument("sir_files", action="store", nargs='*', help='List of SIR files to open') parser.add_argument('--version', action='version', version='%(prog)s version {}'.format(__version__)) args = parser.parse_args() qt_app = QtGui.QApplication(sys.argv) frame = MainWindow() frame.show() # TODO: Handle multiple files if len(args.sir_files) > 0: frame.load_sir(args.sir_files[0]) qt_app.exec_() if __name__ == "__main__": main()
Fix bug if no args were on command line
Fix bug if no args were on command line
Python
mit
richli/dame
932ee2737b822742996f234c90b715771fb876bf
tests/functional/api/view_pdf_test.py
tests/functional/api/view_pdf_test.py
import pytest from tests.conftest import assert_cache_control class TestViewPDFAPI: def test_caching_is_disabled(self, test_app): response = test_app.get("/pdf?url=http://example.com/foo.pdf") assert_cache_control( response.headers, ["max-age=0", "must-revalidate", "no-cache", "no-store"] )
from tests.conftest import assert_cache_control class TestViewPDFAPI: def test_caching_is_disabled(self, test_app): response = test_app.get("/pdf?url=http://example.com/foo.pdf") assert_cache_control( response.headers, ["max-age=0", "must-revalidate", "no-cache", "no-store"] )
Fix lint errors after adding missing __init__ files
Fix lint errors after adding missing __init__ files
Python
bsd-2-clause
hypothesis/via,hypothesis/via,hypothesis/via
50f2cd076aae183376ab14d31594c104ac210738
shivyc.py
shivyc.py
#!/usr/bin/env python3 """Main executable for ShivyC compiler For usage, run "./shivyc.py --help". """ import argparse def get_arguments(): """Set up the argument parser and return an object storing the argument values. return - An object storing argument values, as returned by argparse.parse_args() """ parser = argparse.ArgumentParser(description="Compile C files.") # The C file to compile parser.add_argument("file_name") return parser.parse_args() def main(): """Run the compiler """ arguments = get_arguments() print(arguments) if __name__ == "__main__": main()
#!/usr/bin/env python3 """Main executable for ShivyC compiler For usage, run "./shivyc.py --help". """ import argparse def get_arguments(): """Set up the argument parser and return an object storing the argument values. return - An object storing argument values, as returned by argparse.parse_args() """ parser = argparse.ArgumentParser(description="Compile C files.") # The file name of the C file to compile. The file name gets saved to the # file_name attribute of the returned object, but this parameter appears as # "filename" (no underscore) on the command line. parser.add_argument("file_name", metavar="filename") return parser.parse_args() def main(): """Run the compiler """ arguments = get_arguments() print(arguments) if __name__ == "__main__": main()
Rename file_name argument on command line
Rename file_name argument on command line
Python
mit
ShivamSarodia/ShivyC,ShivamSarodia/ShivyC,ShivamSarodia/ShivyC
d7149d8ea09c897fb954652beeef3bf008448d9e
mopidy/__init__.py
mopidy/__init__.py
import sys if not (2, 6) <= sys.version_info < (3,): sys.exit(u'Mopidy requires Python >= 2.6, < 3') from subprocess import PIPE, Popen VERSION = (0, 4, 0) def get_git_version(): process = Popen(['git', 'describe'], stdout=PIPE, stderr=PIPE) if process.wait() != 0: raise Exception('Execution of "git describe" failed') version = process.stdout.read().strip() if version.startswith('v'): version = version[1:] return version def get_plain_version(): return '.'.join(map(str, VERSION)) def get_version(): try: return get_git_version() except Exception: return get_plain_version() class MopidyException(Exception): def __init__(self, message, *args, **kwargs): super(MopidyException, self).__init__(message, *args, **kwargs) self._message = message @property def message(self): """Reimplement message field that was deprecated in Python 2.6""" return self._message @message.setter def message(self, message): self._message = message class SettingsError(MopidyException): pass class OptionalDependencyError(MopidyException): pass from mopidy import settings as default_settings_module from mopidy.utils.settings import SettingsProxy settings = SettingsProxy(default_settings_module)
import sys if not (2, 6) <= sys.version_info < (3,): sys.exit(u'Mopidy requires Python >= 2.6, < 3') from subprocess import PIPE, Popen VERSION = (0, 4, 0) def get_git_version(): process = Popen(['git', 'describe'], stdout=PIPE, stderr=PIPE) if process.wait() != 0: raise EnvironmentError('Execution of "git describe" failed') version = process.stdout.read().strip() if version.startswith('v'): version = version[1:] return version def get_plain_version(): return '.'.join(map(str, VERSION)) def get_version(): try: return get_git_version() except EnvironmentError: return get_plain_version() class MopidyException(Exception): def __init__(self, message, *args, **kwargs): super(MopidyException, self).__init__(message, *args, **kwargs) self._message = message @property def message(self): """Reimplement message field that was deprecated in Python 2.6""" return self._message @message.setter def message(self, message): self._message = message class SettingsError(MopidyException): pass class OptionalDependencyError(MopidyException): pass from mopidy import settings as default_settings_module from mopidy.utils.settings import SettingsProxy settings = SettingsProxy(default_settings_module)
Raise EnvironmentError instead of Exception to make pylint happy
Raise EnvironmentError instead of Exception to make pylint happy
Python
apache-2.0
pacificIT/mopidy,swak/mopidy,jodal/mopidy,vrs01/mopidy,swak/mopidy,woutervanwijk/mopidy,tkem/mopidy,rawdlite/mopidy,jodal/mopidy,mokieyue/mopidy,rawdlite/mopidy,jmarsik/mopidy,bacontext/mopidy,mokieyue/mopidy,quartz55/mopidy,ZenithDK/mopidy,dbrgn/mopidy,priestd09/mopidy,mopidy/mopidy,quartz55/mopidy,glogiotatidis/mopidy,vrs01/mopidy,hkariti/mopidy,bencevans/mopidy,ZenithDK/mopidy,dbrgn/mopidy,jmarsik/mopidy,mokieyue/mopidy,tkem/mopidy,bencevans/mopidy,pacificIT/mopidy,dbrgn/mopidy,priestd09/mopidy,SuperStarPL/mopidy,hkariti/mopidy,bacontext/mopidy,kingosticks/mopidy,mokieyue/mopidy,bencevans/mopidy,bencevans/mopidy,liamw9534/mopidy,hkariti/mopidy,hkariti/mopidy,jcass77/mopidy,liamw9534/mopidy,kingosticks/mopidy,mopidy/mopidy,adamcik/mopidy,swak/mopidy,tkem/mopidy,ZenithDK/mopidy,ali/mopidy,jcass77/mopidy,diandiankan/mopidy,vrs01/mopidy,jcass77/mopidy,kingosticks/mopidy,jodal/mopidy,jmarsik/mopidy,glogiotatidis/mopidy,diandiankan/mopidy,mopidy/mopidy,SuperStarPL/mopidy,rawdlite/mopidy,glogiotatidis/mopidy,pacificIT/mopidy,pacificIT/mopidy,dbrgn/mopidy,bacontext/mopidy,quartz55/mopidy,jmarsik/mopidy,abarisain/mopidy,diandiankan/mopidy,adamcik/mopidy,adamcik/mopidy,abarisain/mopidy,diandiankan/mopidy,vrs01/mopidy,ali/mopidy,woutervanwijk/mopidy,ali/mopidy,tkem/mopidy,ali/mopidy,glogiotatidis/mopidy,swak/mopidy,SuperStarPL/mopidy,bacontext/mopidy,ZenithDK/mopidy,priestd09/mopidy,SuperStarPL/mopidy,quartz55/mopidy,rawdlite/mopidy
66a9d140feb3a0bd332031853fb1038622fd5c5b
oidc_apis/utils.py
oidc_apis/utils.py
from collections import OrderedDict def combine_uniquely(iterable1, iterable2): """ Combine unique items of two sequences preserving order. :type seq1: Iterable[Any] :type seq2: Iterable[Any] :rtype: list[Any] """ result = OrderedDict.fromkeys(iterable1) for item in iterable2: result[item] = None return list(result.keys()) def after_userlogin_hook(request, user, client): """Marks Django session modified The purpose of this function is to keep the session used by the oidc-provider fresh. This is achieved by pointing 'OIDC_AFTER_USERLOGIN_HOOK' setting to this.""" request.session.modified = True # Return None to continue the login flow return None
from collections import OrderedDict import django from oidc_provider import settings from django.contrib.auth import BACKEND_SESSION_KEY from django.contrib.auth import logout as django_user_logout from users.models import LoginMethod, OidcClientOptions from django.contrib.auth.views import redirect_to_login def combine_uniquely(iterable1, iterable2): """ Combine unique items of two sequences preserving order. :type seq1: Iterable[Any] :type seq2: Iterable[Any] :rtype: list[Any] """ result = OrderedDict.fromkeys(iterable1) for item in iterable2: result[item] = None return list(result.keys()) def after_userlogin_hook(request, user, client): """Marks Django session modified The purpose of this function is to keep the session used by the oidc-provider fresh. This is achieved by pointing 'OIDC_AFTER_USERLOGIN_HOOK' setting to this.""" request.session.modified = True last_login_backend = request.session.get('social_auth_last_login_backend') client_options = OidcClientOptions.objects.get(oidc_client=client) allowed_methods = client_options.login_methods.all() if allowed_methods is None: raise django.core.exceptions.PermissionDenied allowed_providers = set((x.provider_id for x in allowed_methods)) if last_login_backend is not None: active_backend = user.social_auth.filter(provider=last_login_backend) if ((last_login_backend is None and user is not None) or (active_backend.exists() and active_backend.first().provider not in allowed_providers)): django_user_logout(request) next_page = request.get_full_path() return redirect_to_login(next_page, settings.get('OIDC_LOGIN_URL')) # Return None to continue the login flow return None
Implement current session auth method check
Implement current session auth method check
Python
mit
mikkokeskinen/tunnistamo,mikkokeskinen/tunnistamo
23ca8b449a075b4d8ebee19e7756e39f327e9988
dwitter/user/urls.py
dwitter/user/urls.py
from django.conf.urls import url from . import views urlpatterns = [ url(r'^(?P<url_username>\w+)$', views.user_feed, {'page_nr': '1', 'sort': 'new'}, name='user_feed'), url(r'^(?P<url_username>\w+)/(?P<sort>hot|new|top)$', views.user_feed, {'page_nr': '1'}, name='user_sort_feed'), url(r'^(?P<url_username>\w+)/(?P<sort>hot|new|top)/(?P<page_nr>\d+)$', views.user_feed, name='user_feed_page'), ]
from django.conf.urls import url from . import views urlpatterns = [ url(r'^(?P<url_username>[\w.@+-]+)$', views.user_feed, {'page_nr': '1', 'sort': 'new'}, name='user_feed'), url(r'^(?P<url_username>[\w.@+-]+)/(?P<sort>hot|new|top)$', views.user_feed, {'page_nr': '1'}, name='user_sort_feed'), url(r'^(?P<url_username>[\w.@+-]+)/(?P<sort>hot|new|top)/(?P<page_nr>\d+)$', views.user_feed, name='user_feed_page'), ]
Fix url lookup error for usernames certain special characters
Fix url lookup error for usernames certain special characters
Python
apache-2.0
lionleaf/dwitter,lionleaf/dwitter,lionleaf/dwitter
bca736ac15b06263c88d0265339b93b8c2b20d79
test/settings/gyptest-settings.py
test/settings/gyptest-settings.py
#!/usr/bin/env python # Copyright (c) 2011 Google Inc. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. """ Smoke-tests 'settings' blocks. """ import TestGyp test = TestGyp.TestGyp() test.run_gyp('settings.gyp') test.build('test.gyp', test.ALL) test.pass_test()
#!/usr/bin/env python # Copyright (c) 2011 Google Inc. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. """ Smoke-tests 'settings' blocks. """ import TestGyp # 'settings' is only supported for make and scons (and will be removed there as # well eventually). test = TestGyp.TestGyp(formats=['make', 'scons']) test.run_gyp('settings.gyp') test.build('test.gyp', test.ALL) test.pass_test()
Make new settings test not run for xcode generator.
Make new settings test not run for xcode generator. TBR=evan Review URL: http://codereview.chromium.org/7472006
Python
bsd-3-clause
witwall/gyp,witwall/gyp,witwall/gyp,witwall/gyp,witwall/gyp
9ec80ed117ca393a63bf7eb739b4702bfbc0884e
tartpy/eventloop.py
tartpy/eventloop.py
""" Very basic implementation of an event loop ========================================== The eventloop is a singleton to schedule and run events. Exports ------- - ``EventLoop``: the basic eventloop """ import queue import sched import threading import time from .singleton import Singleton class EventLoop(object, metaclass=Singleton): """A generic event loop object.""" def __init__(self): self.scheduler = sched.scheduler() def schedule(self, event): """Schedule an event. An `event` is a thunk. """ self.scheduler.enter(0, 1, event) def stop(self): """Stop the loop.""" pass def run(self, block=False): self.scheduler.run(blocking=block) def run_forever(self, wait=0.05): while True: self.run() time.sleep(wait) def run_in_thread(self): self.thread = threading.Thread(target=self.run_forever, name='event_loop') self.thread.daemon = True self.thread.start()
""" Very basic implementation of an event loop ========================================== The eventloop is a singleton to schedule and run events. Exports ------- - ``EventLoop``: the basic eventloop """ import queue import sched import threading import time from .singleton import Singleton class EventLoop(object, metaclass=Singleton): """A generic event loop object.""" def __init__(self): self.scheduler = sched.scheduler() def schedule(self, event): """Schedule an event. An `event` is a thunk. """ self.scheduler.enter(0, 1, event) def later(self, delay, event): self.scheduler.enter(delay, 1, event) def stop(self): """Stop the loop.""" pass def run(self, block=False): self.scheduler.run(blocking=block) def run_forever(self, wait=0.05): while True: self.run() time.sleep(wait) def run_in_thread(self): self.thread = threading.Thread(target=self.run_forever, name='event_loop') self.thread.daemon = True self.thread.start()
Add function to schedule later
Add function to schedule later
Python
mit
waltermoreira/tartpy
b552d550ca7e4468d95da9a3005e07cbd2ab49d6
tests/test_stock.py
tests/test_stock.py
import cutplanner import unittest class TestStock(unittest.TestCase): def setUp(self): self.stock = cutplanner.Stock(120) def test_cut(self): self.stock.assign_cut(20) self.assertEqual(self.stock.remaining_length, 100) if __name__ == '__main__': unittest.main()
import cutplanner import unittest class TestStock(unittest.TestCase): def setUp(self): self.stock = cutplanner.Stock(120) self.piece = cutplanner.Piece(1, 20) def test_cut(self): self.stock.cut(self.piece) self.assertEqual(self.stock.remaining_length, 100) def test_used_length(self): self.assertEqual(self.stock.used_length, 0) self.stock.cut(self.piece) self.assertEqual(self.stock.used_length, self.piece.length) def test_shrink(self): self.stock.cut(self.piece) print "{0} remaining, {1} used".format(self.stock.remaining_length, self.stock.used_length) new_len = self.stock.shrink(80) self.assertEqual(new_len, 80) if __name__ == '__main__': unittest.main()
Add some initial tests for Stock.
Add some initial tests for Stock.
Python
mit
alanc10n/py-cutplanner
54eb7862d6b17f4e86a380004f6e682452fbebce
git_gutter_change.py
git_gutter_change.py
import sublime import sublime_plugin try: from GitGutter.view_collection import ViewCollection except ImportError: from view_collection import ViewCollection class GitGutterBaseChangeCommand(sublime_plugin.WindowCommand): def run(self): view = self.window.active_view() inserted, modified, deleted = ViewCollection.diff(view) all_changes = sorted(inserted + modified + deleted) row, col = view.rowcol(view.sel()[0].begin()) current_row = row + 1 line = self.jump(all_changes, current_row) self.window.active_view().run_command("goto_line", {"line": line}) class GitGutterNextChangeCommand(GitGutterBaseChangeCommand): def jump(self, all_changes, current_row): return next((change for change in all_changes if change > current_row), current_row) class GitGutterPrevChangeCommand(GitGutterBaseChangeCommand): def jump(self, all_changes, current_row): return next((change for change in reversed(all_changes) if change < current_row), current_row)
import sublime import sublime_plugin try: from GitGutter.view_collection import ViewCollection except ImportError: from view_collection import ViewCollection class GitGutterBaseChangeCommand(sublime_plugin.WindowCommand): def lines_to_blocks(self, lines): blocks = [] last_line = -2 for line in lines: if line > last_line+1: blocks.append(line) last_line = line return blocks def run(self): view = self.window.active_view() inserted, modified, deleted = ViewCollection.diff(view) inserted = self.lines_to_blocks(inserted) modified = self.lines_to_blocks(modified) all_changes = sorted(inserted + modified + deleted) row, col = view.rowcol(view.sel()[0].begin()) current_row = row + 1 line = self.jump(all_changes, current_row) self.window.active_view().run_command("goto_line", {"line": line}) class GitGutterNextChangeCommand(GitGutterBaseChangeCommand): def jump(self, all_changes, current_row): return next((change for change in all_changes if change > current_row), current_row) class GitGutterPrevChangeCommand(GitGutterBaseChangeCommand): def jump(self, all_changes, current_row): return next((change for change in reversed(all_changes) if change < current_row), current_row)
Make lines jumps only jump to blocks over changes
Make lines jumps only jump to blocks over changes Instead of every line in a block of modifications which is tedious
Python
mit
tushortz/GitGutter,biodamasceno/GitGutter,tushortz/GitGutter,akpersad/GitGutter,michaelhogg/GitGutter,natecavanaugh/GitGutter,natecavanaugh/GitGutter,tushortz/GitGutter,michaelhogg/GitGutter,natecavanaugh/GitGutter,biodamasceno/GitGutter,akpersad/GitGutter,akpersad/GitGutter,robfrawley/sublime-git-gutter,natecavanaugh/GitGutter,michaelhogg/GitGutter,robfrawley/sublime-git-gutter,michaelhogg/GitGutter,jisaacks/GitGutter,tushortz/GitGutter,robfrawley/sublime-git-gutter,robfrawley/sublime-git-gutter,biodamasceno/GitGutter,biodamasceno/GitGutter,akpersad/GitGutter
21304ed626998ae4fd359d2e8358bf7563b9020d
harness/summarize.py
harness/summarize.py
#!/usr/bin/env python3 import os import json import math import uncertain TIMINGS_DIR = 'collected' def _mean(values): """The arithmetic mean.""" return sum(values) / len(values) def _mean_err(vals): """The mean and standard error of the mean.""" if len(vals) <= 1: return 0.0 mean = _mean(vals) stdev = math.sqrt(sum((x - mean) ** 2 for x in vals) / (len(vals) - 1)) return mean, stdev / math.sqrt(len(vals)) def summarize_run(data): """Summarize the data from a single run.""" print(data['fn']) all_latencies = [] for msg in data['messages']: # As a sanity check, we can get an average frame latency for the # entire message with: # avg_latency = msg['ms'] / msg['frames'] latencies = msg['latencies'] all_latencies += latencies # TODO Skip the first message as a "warmup" period. umean = uncertain.umean(latencies) mean, err = _mean_err(latencies) print('frame latency:', mean, '+/-', err, 'ms') print('fps:', 1000 / mean) print('frame latency:', umean) print('fps:', 1000.0 / umean) def summarize(): """Sumarrize all the collected data.""" for fn in os.listdir(TIMINGS_DIR): path = os.path.join(TIMINGS_DIR, fn) with open(path) as f: data = json.load(f) summarize_run(data) if __name__ == '__main__': summarize()
#!/usr/bin/env python3 import os import json import uncertain TIMINGS_DIR = 'collected' def summarize_run(data): """Summarize the data from a single run.""" print(data['fn']) all_latencies = [] for msg in data['messages']: # As a sanity check, we can get an average frame latency for the # entire message with: # avg_latency = msg['ms'] / msg['frames'] latencies = msg['latencies'] all_latencies += latencies # TODO Skip the first message as a "warmup" period. mean = uncertain.umean(latencies) print('frame latency:', mean) print('fps:', 1000.0 / mean) def summarize(): """Sumarrize all the collected data.""" for fn in os.listdir(TIMINGS_DIR): path = os.path.join(TIMINGS_DIR, fn) with open(path) as f: data = json.load(f) summarize_run(data) if __name__ == '__main__': summarize()
Remove old uncertainty quantification stuff
Remove old uncertainty quantification stuff
Python
mit
cucapra/braid,cucapra/braid,guoyiteng/braid,guoyiteng/braid,guoyiteng/braid,cucapra/braid,guoyiteng/braid,guoyiteng/braid,cucapra/braid,cucapra/braid,guoyiteng/braid,cucapra/braid
a36fe5002bbf5dfcf27a3251cfed85c341e2156d
cbcollections.py
cbcollections.py
class defaultdict(dict): """Poor man's implementation of defaultdict for Python 2.4 """ def __init__(self, default_factory=None, **kwargs): self.default_factory = default_factory super(defaultdict, self).__init__(**kwargs) def __getitem__(self, key): if self.default_factory is None: return super(defaultdict, self).__getitem__(key) else: try: return super(defaultdict, self).__getitem__(key) except KeyError: return self.default_factory()
class defaultdict(dict): """Poor man's implementation of defaultdict for Python 2.4 """ def __init__(self, default_factory=None, **kwargs): self.default_factory = default_factory super(defaultdict, self).__init__(**kwargs) def __getitem__(self, key): if self.default_factory is None: return super(defaultdict, self).__getitem__(key) else: try: return super(defaultdict, self).__getitem__(key) except KeyError: self[key] = self.default_factory() return self[key]
Save generated value for defaultdict
MB-6867: Save generated value for defaultdict Instead of just returning value, keep it in dict. Change-Id: I2a9862503b71f2234a4a450c48998b5f53a951bc Reviewed-on: http://review.couchbase.org/21602 Tested-by: Bin Cui <[email protected]> Reviewed-by: Pavel Paulau <[email protected]>
Python
apache-2.0
couchbase/couchbase-cli,couchbaselabs/couchbase-cli,membase/membase-cli,membase/membase-cli,couchbase/couchbase-cli,membase/membase-cli,couchbaselabs/couchbase-cli,couchbaselabs/couchbase-cli
3a5e2e34374f92f0412d121fb9552278105f230a
salt/acl/__init__.py
salt/acl/__init__.py
# -*- coding: utf-8 -*- ''' The acl module handles client_acl operations Additional information on client_acl can be found by reading the salt documention: http://docs.saltstack.com/en/latest/ref/clientacl.html ''' # Import python libraries from __future__ import absolute_import import re class ClientACL(object): ''' Represents the client ACL and provides methods to query the ACL for given operations ''' def __init__(self, blacklist): self.blacklist = blacklist def user_is_blacklisted(self, user): ''' Takes a username as a string and returns a boolean. True indicates that the provided user has been blacklisted ''' for blacklisted_user in self.blacklist.get('users', []): if re.match(blacklisted_user, user): return True return False def cmd_is_blacklisted(self, cmd): for blacklisted_module in self.blacklist.get('modules', []): # If this is a regular command, it is a single function if isinstance(cmd, str): funs_to_check = [cmd] # If this is a compound function else: funs_to_check = cmd for fun in funs_to_check: if re.match(blacklisted_module, fun): return True return False
# -*- coding: utf-8 -*- ''' The acl module handles client_acl operations Additional information on client_acl can be found by reading the salt documentation: http://docs.saltstack.com/en/latest/ref/clientacl.html ''' # Import python libraries from __future__ import absolute_import import re class ClientACL(object): ''' Represents the client ACL and provides methods to query the ACL for given operations ''' def __init__(self, blacklist): self.blacklist = blacklist def user_is_blacklisted(self, user): ''' Takes a username as a string and returns a boolean. True indicates that the provided user has been blacklisted ''' for blacklisted_user in self.blacklist.get('users', []): if re.match(blacklisted_user, user): return True return False def cmd_is_blacklisted(self, cmd): for blacklisted_module in self.blacklist.get('modules', []): # If this is a regular command, it is a single function if isinstance(cmd, str): funs_to_check = [cmd] # If this is a compound function else: funs_to_check = cmd for fun in funs_to_check: if re.match(blacklisted_module, fun): return True return False
Fix typo documention -> documentation
Fix typo documention -> documentation
Python
apache-2.0
saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt
7548a1245cc21c92f09302ccaf065bdf6189ef2d
quilt/cli/series.py
quilt/cli/series.py
# vim: fileencoding=utf-8 et sw=4 ts=4 tw=80: # python-quilt - A Python implementation of the quilt patch system # # Copyright (C) 2012 Björn Ricks <[email protected]> # # This library is free software; you can redistribute it and/or # modify it under the terms of the GNU Lesser General Public # License as published by the Free Software Foundation; either # version 2.1 of the License, or (at your option) any later version. # This library is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # Lesser General Public License for more details. # You should have received a copy of the GNU Lesser General Public # License along with this library; if not, write to the Free Software # Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA # 02110-1301 USA from quilt.cli.meta import Command from quilt.db import Series class SeriesCommand(Command): usage = "%prog series" name = "series" def run(self, option, args): series = Series(self.get_patches_dir()) for patch in series.patches(): print patch
# vim: fileencoding=utf-8 et sw=4 ts=4 tw=80: # python-quilt - A Python implementation of the quilt patch system # # Copyright (C) 2012 Björn Ricks <[email protected]> # # This library is free software; you can redistribute it and/or # modify it under the terms of the GNU Lesser General Public # License as published by the Free Software Foundation; either # version 2.1 of the License, or (at your option) any later version. # This library is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # Lesser General Public License for more details. # You should have received a copy of the GNU Lesser General Public # License along with this library; if not, write to the Free Software # Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA # 02110-1301 USA from quilt.cli.meta import Command from quilt.db import Series class SeriesCommand(Command): usage = "%prog series" name = "series" def run(self, option, args): series = Series(self.get_patches_dir()) for patch in series.patches(): print patch
Remove whitespace at end of line
Remove whitespace at end of line
Python
mit
bjoernricks/python-quilt,vadmium/python-quilt
b27a51f19ea3f9d13672a0db51f7d2b05f9539f0
kitten/validation.py
kitten/validation.py
import jsonschema CORE_SCHEMA = { 'type': 'object', 'properties': { 'paradigm': { 'type': 'string', }, 'method': { 'type': 'string', }, }, 'additionalProperties': False, } VALIDATORS = { 'core': CORE_SCHEMA } def validate(request, schema_name): jsonschema.validate(request, VALIDATORS[schema_name])
import jsonschema CORE_SCHEMA = { 'type': 'object', 'properties': { 'paradigm': { 'type': 'string', }, 'method': { 'type': 'string', }, 'address': { 'type': 'string', }, }, 'additionalProperties': False, } VALIDATORS = { 'core': CORE_SCHEMA } def validate(request, schema_name): jsonschema.validate(request, VALIDATORS[schema_name])
Add 'address' field to core schema
Add 'address' field to core schema
Python
mit
thiderman/network-kitten
fb0b956563efbcd22af8300fd4341e3cb277b80a
app/models/user.py
app/models/user.py
from app import db from flask import Flask from datetime import datetime class User(db.Model): id = db.Column(db.Integer, primary_key=True) username = db.Column(db.String(80), unique=True) email = db.Column(db.String(120), unique=True) name = db.Column(db.String(80)) bio = db.Column(db.String(180)) github_id = db.Column(db.Integer, unique=True) github_username = db.Column(db.String(64), unique=True) github_token = db.Column(db.String(300), unique=True) password = db.Column(db.String(300)) created_at = db.Column(db.DateTime) def __init__(self, username, email, password, name=None): self.email = email self.username = username self.password = password if name is None: self.name = username else: self.name = name self.created_at = datetime.now() is_authenticated = True is_anonymous = False is_active = True def get_id(self): return unicode(self.id) def __repr__(self): return '<User %r>' % self.username
from app import db from flask import Flask from datetime import datetime class User(db.Model): id = db.Column(db.Integer, primary_key=True) username = db.Column(db.String(80), unique=True) email = db.Column(db.String(120), unique=True) name = db.Column(db.String(80)) bio = db.Column(db.String(180)) avatar_url = db.Column(db.String(256)) owner = db.Column(db.String(32), nullable=False, default='user') github_id = db.Column(db.Integer, unique=True) github_username = db.Column(db.String(64), unique=True) github_token = db.Column(db.String(300), unique=True) password = db.Column(db.String(300)) created_at = db.Column(db.DateTime) def __init__(self, username, email, password, name=None): self.email = email self.username = username self.password = password if name is None: self.name = username else: self.name = name self.created_at = datetime.now() is_authenticated = True is_anonymous = False is_active = True def get_id(self): return unicode(self.id) def __repr__(self): return '<User %r>' % self.username
Add avatar_url and owner field for User
Add avatar_url and owner field for User
Python
agpl-3.0
lc-soft/GitDigger,lc-soft/GitDigger,lc-soft/GitDigger,lc-soft/GitDigger
f42e62005ea4cc3e71cf10dda8c0bace029014c5
kubespawner/utils.py
kubespawner/utils.py
""" Misc. general utility functions, not tied to Kubespawner directly """ from concurrent.futures import ThreadPoolExecutor from jupyterhub.utils import DT_MIN, DT_MAX, DT_SCALE from tornado import gen, ioloop from traitlets.config import SingletonConfigurable class SingletonExecutor(SingletonConfigurable, ThreadPoolExecutor): """ Simple wrapper to ThreadPoolExecutor that is also a singleton. We want one ThreadPool that is used by all the spawners, rather than one ThreadPool per spawner! """ pass @gen.coroutine def exponential_backoff(func, fail_message, timeout=10, *args, **kwargs): loop = ioloop.IOLoop.current() tic = loop.time() dt = DT_MIN while dt > 0: if func(*args, **kwargs): return else: yield gen.sleep(dt) dt = min(dt * DT_SCALE, DT_MAX, timeout - (loop.time() - tic)) raise TimeoutError(fail_message)
""" Misc. general utility functions, not tied to Kubespawner directly """ from concurrent.futures import ThreadPoolExecutor import random from jupyterhub.utils import DT_MIN, DT_MAX, DT_SCALE from tornado import gen, ioloop from traitlets.config import SingletonConfigurable class SingletonExecutor(SingletonConfigurable, ThreadPoolExecutor): """ Simple wrapper to ThreadPoolExecutor that is also a singleton. We want one ThreadPool that is used by all the spawners, rather than one ThreadPool per spawner! """ pass @gen.coroutine def exponential_backoff(func, fail_message, timeout=10, *args, **kwargs): loop = ioloop.IOLoop.current() start_tic = loop.time() dt = DT_MIN while True: if (loop.time() - start_tic) > timeout: # We time out! break if func(*args, **kwargs): return else: yield gen.sleep(dt) # Add some random jitter to improve performance # This makes sure that we don't overload any single iteration # of the tornado loop with too many things # See https://www.awsarchitectureblog.com/2015/03/backoff.html # for a good example of why and how this helps dt = min(DT_MAX, (1 + random.random()) * (dt * DT_SCALE)) raise TimeoutError(fail_message)
Add random jitter to the exponential backoff function
Add random jitter to the exponential backoff function
Python
bsd-3-clause
yuvipanda/jupyterhub-kubernetes-spawner,jupyterhub/kubespawner
9f6d4d9e82ef575164535a8fb9ea80417458dd6b
website/files/models/dataverse.py
website/files/models/dataverse.py
import requests from framework.auth.core import _get_current_user from website.files.models.base import File, Folder, FileNode, FileVersion __all__ = ('DataverseFile', 'DataverseFolder', 'DataverseFileNode') class DataverseFileNode(FileNode): provider = 'dataverse' class DataverseFolder(DataverseFileNode, Folder): pass class DataverseFile(DataverseFileNode, File): def touch(self, version=None, revision=None, **kwargs): """Note: Dataverse only has psuedo versions, don't save them""" version = revision or version # Use revision or version resp = requests.get(self.generate_waterbutler_url(meta=True, version=version, **kwargs)) if resp.status_code != 200: return None data = resp.json() self.name = data['data']['name'] self.materialized_path = data['data']['materialized'] version = FileVersion(identifier=version) version.update_metadata(data['data'], save=False) user = _get_current_user() if not user or not self.node.can_edit(user=user): try: # Users without edit permission can only see published files if not data['data']['extra']['hasPublishedVersion']: # Blank out name and path for the render # Dont save because there's no reason to persist the change self.name = '' self.materialized_path = '' return (version, '<div class="alert alert-info" role="alert">This file does not exist.</div>') except (KeyError, IndexError): pass
from framework.auth.core import _get_current_user from website.files.models.base import File, Folder, FileNode, FileVersion __all__ = ('DataverseFile', 'DataverseFolder', 'DataverseFileNode') class DataverseFileNode(FileNode): provider = 'dataverse' class DataverseFolder(DataverseFileNode, Folder): pass class DataverseFile(DataverseFileNode, File): version_identifier = 'version' def update(self, revision, data): """Note: Dataverse only has psuedo versions, don't save them""" self.name = data['name'] self.materialized_path = data['materialized'] version = FileVersion(identifier=revision) version.update_metadata(data, save=False) user = _get_current_user() if not user or not self.node.can_edit(user=user): try: # Users without edit permission can only see published files if not data['extra']['hasPublishedVersion']: # Blank out name and path for the render # Dont save because there's no reason to persist the change self.name = '' self.materialized_path = '' return (version, '<div class="alert alert-info" role="alert">This file does not exist.</div>') except (KeyError, IndexError): pass return version
Move override logic into update rather than touch
Move override logic into update rather than touch
Python
apache-2.0
Johnetordoff/osf.io,mluke93/osf.io,SSJohns/osf.io,chrisseto/osf.io,hmoco/osf.io,caseyrygt/osf.io,GageGaskins/osf.io,acshi/osf.io,alexschiller/osf.io,caseyrollins/osf.io,ZobairAlijan/osf.io,wearpants/osf.io,GageGaskins/osf.io,brandonPurvis/osf.io,CenterForOpenScience/osf.io,SSJohns/osf.io,alexschiller/osf.io,adlius/osf.io,samchrisinger/osf.io,sloria/osf.io,hmoco/osf.io,erinspace/osf.io,DanielSBrown/osf.io,samanehsan/osf.io,haoyuchen1992/osf.io,Ghalko/osf.io,crcresearch/osf.io,brandonPurvis/osf.io,CenterForOpenScience/osf.io,danielneis/osf.io,doublebits/osf.io,zamattiac/osf.io,baylee-d/osf.io,Johnetordoff/osf.io,KAsante95/osf.io,wearpants/osf.io,caneruguz/osf.io,petermalcolm/osf.io,pattisdr/osf.io,mattclark/osf.io,brianjgeiger/osf.io,cwisecarver/osf.io,samchrisinger/osf.io,billyhunt/osf.io,asanfilippo7/osf.io,chrisseto/osf.io,Johnetordoff/osf.io,caneruguz/osf.io,rdhyee/osf.io,abought/osf.io,RomanZWang/osf.io,mluke93/osf.io,monikagrabowska/osf.io,TomHeatwole/osf.io,mfraezz/osf.io,cosenal/osf.io,saradbowman/osf.io,felliott/osf.io,aaxelb/osf.io,jnayak1/osf.io,cwisecarver/osf.io,danielneis/osf.io,kch8qx/osf.io,GageGaskins/osf.io,aaxelb/osf.io,aaxelb/osf.io,TomHeatwole/osf.io,crcresearch/osf.io,ZobairAlijan/osf.io,zamattiac/osf.io,caneruguz/osf.io,monikagrabowska/osf.io,brianjgeiger/osf.io,zachjanicki/osf.io,rdhyee/osf.io,samchrisinger/osf.io,kch8qx/osf.io,Johnetordoff/osf.io,chennan47/osf.io,TomBaxter/osf.io,jnayak1/osf.io,felliott/osf.io,binoculars/osf.io,acshi/osf.io,Nesiehr/osf.io,kch8qx/osf.io,Nesiehr/osf.io,Ghalko/osf.io,mfraezz/osf.io,GageGaskins/osf.io,petermalcolm/osf.io,arpitar/osf.io,samanehsan/osf.io,abought/osf.io,laurenrevere/osf.io,emetsger/osf.io,petermalcolm/osf.io,mattclark/osf.io,RomanZWang/osf.io,jnayak1/osf.io,kwierman/osf.io,leb2dg/osf.io,doublebits/osf.io,cslzchen/osf.io,zachjanicki/osf.io,njantrania/osf.io,chennan47/osf.io,kwierman/osf.io,brianjgeiger/osf.io,hmoco/osf.io,GageGaskins/osf.io,zamattiac/osf.io,emetsger/osf.io,caseyrollins/osf.io,binoculars/osf.io,Nesiehr/osf.io,monikagrabowska/osf.io,cslzchen/osf.io,brandonPurvis/osf.io,acshi/osf.io,rdhyee/osf.io,petermalcolm/osf.io,cosenal/osf.io,asanfilippo7/osf.io,adlius/osf.io,mluke93/osf.io,adlius/osf.io,mluo613/osf.io,kch8qx/osf.io,Ghalko/osf.io,haoyuchen1992/osf.io,HalcyonChimera/osf.io,abought/osf.io,pattisdr/osf.io,haoyuchen1992/osf.io,brandonPurvis/osf.io,mluo613/osf.io,saradbowman/osf.io,mattclark/osf.io,jnayak1/osf.io,felliott/osf.io,arpitar/osf.io,caseyrollins/osf.io,amyshi188/osf.io,acshi/osf.io,asanfilippo7/osf.io,emetsger/osf.io,RomanZWang/osf.io,njantrania/osf.io,sloria/osf.io,SSJohns/osf.io,cwisecarver/osf.io,leb2dg/osf.io,leb2dg/osf.io,felliott/osf.io,CenterForOpenScience/osf.io,TomBaxter/osf.io,billyhunt/osf.io,adlius/osf.io,ticklemepierce/osf.io,chrisseto/osf.io,haoyuchen1992/osf.io,erinspace/osf.io,Ghalko/osf.io,acshi/osf.io,doublebits/osf.io,wearpants/osf.io,alexschiller/osf.io,icereval/osf.io,amyshi188/osf.io,doublebits/osf.io,KAsante95/osf.io,alexschiller/osf.io,wearpants/osf.io,samchrisinger/osf.io,njantrania/osf.io,njantrania/osf.io,cwisecarver/osf.io,KAsante95/osf.io,zamattiac/osf.io,kch8qx/osf.io,TomBaxter/osf.io,ZobairAlijan/osf.io,HalcyonChimera/osf.io,KAsante95/osf.io,DanielSBrown/osf.io,aaxelb/osf.io,rdhyee/osf.io,mluke93/osf.io,mfraezz/osf.io,zachjanicki/osf.io,cslzchen/osf.io,hmoco/osf.io,cosenal/osf.io,zachjanicki/osf.io,chennan47/osf.io,doublebits/osf.io,caneruguz/osf.io,monikagrabowska/osf.io,billyhunt/osf.io,KAsante95/osf.io,brandonPurvis/osf.io,brianjgeiger/osf.io,samanehsan/osf.io,binoculars/osf.io,monikagrabowska/osf.io,crcresearch/osf.io,danielneis/osf.io,laurenrevere/osf.io,arpitar/osf.io,kwierman/osf.io,billyhunt/osf.io,kwierman/osf.io,danielneis/osf.io,ticklemepierce/osf.io,baylee-d/osf.io,HalcyonChimera/osf.io,ZobairAlijan/osf.io,cslzchen/osf.io,caseyrygt/osf.io,laurenrevere/osf.io,Nesiehr/osf.io,mluo613/osf.io,TomHeatwole/osf.io,amyshi188/osf.io,caseyrygt/osf.io,mfraezz/osf.io,SSJohns/osf.io,erinspace/osf.io,leb2dg/osf.io,DanielSBrown/osf.io,arpitar/osf.io,baylee-d/osf.io,RomanZWang/osf.io,billyhunt/osf.io,mluo613/osf.io,mluo613/osf.io,amyshi188/osf.io,abought/osf.io,pattisdr/osf.io,ticklemepierce/osf.io,chrisseto/osf.io,emetsger/osf.io,caseyrygt/osf.io,cosenal/osf.io,RomanZWang/osf.io,HalcyonChimera/osf.io,sloria/osf.io,icereval/osf.io,alexschiller/osf.io,samanehsan/osf.io,icereval/osf.io,TomHeatwole/osf.io,ticklemepierce/osf.io,DanielSBrown/osf.io,asanfilippo7/osf.io,CenterForOpenScience/osf.io
06d210cdc811f0051a489f335cc94a604e99a35d
werobot/session/mongodbstorage.py
werobot/session/mongodbstorage.py
# -*- coding: utf-8 -*- from werobot.session import SessionStorage from werobot.utils import json_loads, json_dumps class MongoDBStorage(SessionStorage): """ MongoDBStorage 会把你的 Session 数据储存在一个 MongoDB Collection 中 :: import pymongo import werobot from werobot.session.mongodbstorage import MongoDBStorage collection = pymongo.MongoClient()["wechat"]["session"] session_storage = MongoDBStorage(collection) robot = werobot.WeRoBot(token="token", enable_session=True, session_storage=session_storage) 你需要安装 ``pymongo`` 才能使用 MongoDBStorage 。 :param collection: 一个 MongoDB Collection。 """ def __init__(self, collection): import pymongo assert isinstance(collection, pymongo.collection.Collection) self.collection = collection collection.create_index("wechat_id") def _get_document(self, id): return self.collection.find_one({"wechat_id": id}) def get(self, id): document = self._get_document(id) if document: session_json = document["session"] return json_loads(session_json) return {} def set(self, id, value): document = self._get_document(id) session = json_dumps(value) if document: document["session"] = session self.collection.save(document) else: self.collection.insert({ "wechat_id": id, "session": session }) def delete(self, id): document = self._get_document(id) if document: self.collection.remove(document["_id"])
# -*- coding: utf-8 -*- from werobot.session import SessionStorage from werobot.utils import json_loads, json_dumps class MongoDBStorage(SessionStorage): """ MongoDBStorage 会把你的 Session 数据储存在一个 MongoDB Collection 中 :: import pymongo import werobot from werobot.session.mongodbstorage import MongoDBStorage collection = pymongo.MongoClient()["wechat"]["session"] session_storage = MongoDBStorage(collection) robot = werobot.WeRoBot(token="token", enable_session=True, session_storage=session_storage) 你需要安装 ``pymongo`` 才能使用 MongoDBStorage 。 :param collection: 一个 MongoDB Collection。 """ def __init__(self, collection): self.collection = collection collection.create_index("wechat_id") def _get_document(self, id): return self.collection.find_one({"wechat_id": id}) def get(self, id): document = self._get_document(id) if document: session_json = document["session"] return json_loads(session_json) return {} def set(self, id, value): session = json_dumps(value) self.collection.replace_one({ "wechat_id": id }, { "wechat_id": id, "session": session }, upsert=True) def delete(self, id): self.collection.delete_one({ "wechat_id": id })
Use new pymongo API in MongoDBStorage
Use new pymongo API in MongoDBStorage
Python
mit
whtsky/WeRoBot,whtsky/WeRoBot,adam139/WeRobot,adam139/WeRobot,whtsky/WeRoBot,weberwang/WeRoBot,weberwang/WeRoBot
841ca9cfbdb8faac9d8deb47b65717b5fb7c8eb4
mfh.py
mfh.py
import os import sys import time from multiprocessing import Process, Event import mfhclient import server import update from arguments import parse from settings import HONEYPORT, HIVEPORT def main(): update_event = Event() mfhclient_process = Process( args=(args, update_event,), name="mfhclient_process", target=mfhclient.main, ) server_process = Process( args=(args, update_event,), name="server_process", target=server.main, ) if args.client is not None: mfhclient_process.start() if args.client is not None: server_process.start() if args.updater: trigger_process = Process( args=(update_event,), name="trigger_process", target=update.trigger, ) trigger_process.start() trigger_process.join() while mfhclient_process.is_alive() or server_process.is_alive(): time.sleep(5) else: if args.updater: # update.pull("origin", "master") sys.stdout.flush() os.execl(sys.executable, sys.executable, *sys.argv) if __name__ == '__main__': # Parse arguments args = parse() if args.c: args.client = HONEYPORT if args.s: args.server = HIVEPORT main()
import os import sys import time from multiprocessing import Process, Event import mfhclient import server import update from arguments import parse from settings import HONEYPORT, HIVEPORT def main(): update_event = Event() client = create_process("client", mfhclient.main, args, update_event) serv = create_process("server", server.main, args, update_event) if args.client is not None: client.start() if args.client is not None: serv.start() if args.updater: trigger = create_process("trigger", update.trigger, update_event) trigger.start() trigger.join() while client.is_alive() or serv.is_alive(): time.sleep(5) else: if args.updater: update.pull("origin", "master") sys.stdout.flush() os.execl(sys.executable, sys.executable, *sys.argv) def create_process(name, function, *arguments): process = Process( args=arguments, name="{0}_process".format(name), target=function, ) return process if __name__ == '__main__': # Parse arguments args = parse() if args.c: args.client = HONEYPORT if args.s: args.server = HIVEPORT processes = {} main()
Move all the process creation in a new function
Move all the process creation in a new function This reduces the size of code.
Python
mit
Zloool/manyfaced-honeypot
3aacfd7147836ef95133aa88d558a1d69bbcd0cd
mopidy/exceptions.py
mopidy/exceptions.py
from __future__ import absolute_import, unicode_literals class MopidyException(Exception): def __init__(self, message, *args, **kwargs): super(MopidyException, self).__init__(message, *args, **kwargs) self._message = message @property def message(self): """Reimplement message field that was deprecated in Python 2.6""" return self._message @message.setter # noqa def message(self, message): self._message = message class BackendError(MopidyException): pass class CoreError(MopidyException): def __init(self, message, errno=None): super(CoreError, self).__init(message, errno) self.errno = errno class ExtensionError(MopidyException): pass class FindError(MopidyException): def __init__(self, message, errno=None): super(FindError, self).__init__(message, errno) self.errno = errno class FrontendError(MopidyException): pass class MixerError(MopidyException): pass class ScannerError(MopidyException): pass class TracklistFull(CoreError): def __init(self, message, errno=None): super(TracklistFull, self).__init(message, errno) self.errno = errno class AudioException(MopidyException): pass class ValidationError(ValueError): pass
from __future__ import absolute_import, unicode_literals class MopidyException(Exception): def __init__(self, message, *args, **kwargs): super(MopidyException, self).__init__(message, *args, **kwargs) self._message = message @property def message(self): """Reimplement message field that was deprecated in Python 2.6""" return self._message @message.setter # noqa def message(self, message): self._message = message class BackendError(MopidyException): pass class CoreError(MopidyException): def __init__(self, message, errno=None): super(CoreError, self).__init__(message, errno) self.errno = errno class ExtensionError(MopidyException): pass class FindError(MopidyException): def __init__(self, message, errno=None): super(FindError, self).__init__(message, errno) self.errno = errno class FrontendError(MopidyException): pass class MixerError(MopidyException): pass class ScannerError(MopidyException): pass class TracklistFull(CoreError): def __init__(self, message, errno=None): super(TracklistFull, self).__init__(message, errno) self.errno = errno class AudioException(MopidyException): pass class ValidationError(ValueError): pass
Fix typo in new CoreErrors
exception: Fix typo in new CoreErrors
Python
apache-2.0
mopidy/mopidy,hkariti/mopidy,tkem/mopidy,bacontext/mopidy,swak/mopidy,mokieyue/mopidy,ZenithDK/mopidy,ali/mopidy,mokieyue/mopidy,bencevans/mopidy,jcass77/mopidy,bencevans/mopidy,bacontext/mopidy,diandiankan/mopidy,hkariti/mopidy,dbrgn/mopidy,ZenithDK/mopidy,bacontext/mopidy,mopidy/mopidy,pacificIT/mopidy,SuperStarPL/mopidy,hkariti/mopidy,hkariti/mopidy,mopidy/mopidy,pacificIT/mopidy,kingosticks/mopidy,jodal/mopidy,swak/mopidy,diandiankan/mopidy,jmarsik/mopidy,jmarsik/mopidy,ZenithDK/mopidy,bencevans/mopidy,kingosticks/mopidy,quartz55/mopidy,dbrgn/mopidy,vrs01/mopidy,adamcik/mopidy,quartz55/mopidy,adamcik/mopidy,quartz55/mopidy,adamcik/mopidy,pacificIT/mopidy,swak/mopidy,kingosticks/mopidy,ali/mopidy,vrs01/mopidy,jcass77/mopidy,dbrgn/mopidy,jcass77/mopidy,jodal/mopidy,SuperStarPL/mopidy,jmarsik/mopidy,mokieyue/mopidy,vrs01/mopidy,bacontext/mopidy,ZenithDK/mopidy,tkem/mopidy,jodal/mopidy,jmarsik/mopidy,SuperStarPL/mopidy,quartz55/mopidy,tkem/mopidy,mokieyue/mopidy,SuperStarPL/mopidy,ali/mopidy,bencevans/mopidy,ali/mopidy,diandiankan/mopidy,pacificIT/mopidy,tkem/mopidy,diandiankan/mopidy,swak/mopidy,dbrgn/mopidy,vrs01/mopidy
5f128bbfc61169ac6b5f0e9f4dc6bcd05092382c
requests_cache/serializers/pipeline.py
requests_cache/serializers/pipeline.py
""" .. automodsumm:: requests_cache.serializers.pipeline :classes-only: :nosignatures: """ from typing import Any, List, Union from ..models import CachedResponse class Stage: """Generic class to wrap serialization steps with consistent ``dumps()`` and ``loads()`` methods""" def __init__(self, obj: Any, dumps: str = "dumps", loads: str = "loads"): self.obj = obj self.dumps = getattr(obj, dumps) self.loads = getattr(obj, loads) class SerializerPipeline: """A sequence of steps used to serialize and deserialize response objects""" def __init__(self, steps: List): self.steps = steps self.dump_steps = [step.dumps for step in steps] self.load_steps = [step.loads for step in reversed(steps)] def dumps(self, value) -> Union[str, bytes]: for step in self.dump_steps: value = step(value) return value def loads(self, value) -> CachedResponse: for step in self.load_steps: value = step(value) return value
""" .. automodsumm:: requests_cache.serializers.pipeline :classes-only: :nosignatures: """ from typing import Any, Callable, List, Union from ..models import CachedResponse class Stage: """Generic class to wrap serialization steps with consistent ``dumps()`` and ``loads()`` methods Args: obj: Serializer object or module, if applicable dumps: Serialization function, or name of method on ``obj`` loads: Deserialization function, or name of method on ``obj`` """ def __init__( self, obj: Any = None, dumps: Union[str, Callable] = 'dumps', loads: Union[str, Callable] = 'loads', ): self.obj = obj self.dumps = getattr(obj, dumps) if isinstance(dumps, str) else dumps self.loads = getattr(obj, loads) if isinstance(loads, str) else loads class SerializerPipeline: """A sequence of steps used to serialize and deserialize response objects. This can be initialized with :py:class:`Stage` objects, or any objects with ``dumps()`` and ``loads()`` methods """ def __init__(self, stages: List): self.steps = stages self.dump_steps = [step.dumps for step in stages] self.load_steps = [step.loads for step in reversed(stages)] def dumps(self, value) -> Union[str, bytes]: for step in self.dump_steps: value = step(value) return value def loads(self, value) -> CachedResponse: for step in self.load_steps: value = step(value) return value
Allow Stage objects to take functions instead of object + method names
Allow Stage objects to take functions instead of object + method names
Python
bsd-2-clause
reclosedev/requests-cache
657741f3d4df734afef228e707005dc21d540e34
post-refunds-back.py
post-refunds-back.py
#!/usr/bin/env python -u from __future__ import absolute_import, division, print_function, unicode_literals import csv from gratipay import wireup from gratipay.models.exchange_route import ExchangeRoute from gratipay.models.participant import Participant from gratipay.billing.exchanges import record_exchange db = wireup.db(wireup.env()) inp = csv.reader(open('balanced/refund/refunds.completed.csv')) note = 'refund of advance payment; see https://medium.com/gratipay-blog/charging-in-arrears-18cacf779bee' for ts, id, amount, username, route_id, status_code, content in inp: if status_code != '201': continue amount = '-' + amount[:-2] + '.' + amount[-2:] print('posting {} back for {}'.format(amount, username)) route = ExchangeRoute.from_id(route_id) rp = route.participant participant = Participant.from_id(rp) if type(rp) is long else rp # Such a hack. :( route.set_attributes(participant=participant) record_exchange(db, route, amount, 0, participant, 'pending', note)
#!/usr/bin/env python -u from __future__ import absolute_import, division, print_function, unicode_literals import csv from decimal import Decimal as D from gratipay import wireup from gratipay.models.exchange_route import ExchangeRoute from gratipay.models.participant import Participant from gratipay.billing.exchanges import record_exchange db = wireup.db(wireup.env()) inp = csv.reader(open('refunds.completed.csv')) note = 'refund of advance payment; see https://medium.com/gratipay-blog/18cacf779bee' total = N = 0 for ts, id, amount, username, route_id, success, ref in inp: print('posting {} back for {}'.format(amount, username)) assert success == 'True' total += D(amount) N += 1 amount = D('-' + amount) route = ExchangeRoute.from_id(route_id) # Such a hack. :( rp = route.participant participant = Participant.from_id(rp) if type(rp) is long else rp route.set_attributes(participant=participant) exchange_id = record_exchange(db, route, amount, 0, participant, 'pending', note) db.run("update exchanges set ref=%s where id=%s", (ref, exchange_id)) print('posted {} back for {}'.format(total, N))
Update post-back script for Braintree
Update post-back script for Braintree
Python
mit
gratipay/gratipay.com,gratipay/gratipay.com,gratipay/gratipay.com,gratipay/gratipay.com
9be282d3f2f278ca8fe0dd65d78d03005b6e43cd
url_shortener/forms.py
url_shortener/forms.py
# -*- coding: utf-8 -*- from flask_wtf import Form from flask_wtf.recaptcha import RecaptchaField, Recaptcha from wtforms import StringField, validators from .validation import not_blacklisted_nor_spam class ShortenedURLForm(Form): url = StringField( validators=[ validators.DataRequired(), validators.URL(message="A valid URL is required"), not_blacklisted_nor_spam ] ) recaptcha = RecaptchaField( validators=[ Recaptcha( "Please click on the reCAPTCHA field to prove you are a human" ) ] )
# -*- coding: utf-8 -*- from flask_wtf import Form from flask_wtf.recaptcha import RecaptchaField, Recaptcha from wtforms import StringField, validators from .validation import not_blacklisted_nor_spam class ShortenedURLForm(Form): url = StringField( validators=[ validators.DataRequired(), validators.URL(message='A valid URL is required'), not_blacklisted_nor_spam ] ) recaptcha = RecaptchaField( validators=[ Recaptcha( 'Please click on the reCAPTCHA field to prove you are a human' ) ] )
Replace double quotes with single quotes as string delimiters
Replace double quotes with single quotes as string delimiters This commit replaces double quotes with single quotes as string delimiters to improve consistency.
Python
mit
piotr-rusin/url-shortener,piotr-rusin/url-shortener
022062c409ee06a719b5687ea1feb989c5cad627
app/grandchallenge/pages/sitemaps.py
app/grandchallenge/pages/sitemaps.py
from grandchallenge.core.sitemaps import SubdomainSitemap from grandchallenge.pages.models import Page class PagesSitemap(SubdomainSitemap): priority = 0.8 def items(self): return Page.objects.filter( permission_level=Page.ALL, challenge__hidden=False )
from grandchallenge.core.sitemaps import SubdomainSitemap from grandchallenge.pages.models import Page class PagesSitemap(SubdomainSitemap): priority = 0.8 def items(self): return Page.objects.filter( permission_level=Page.ALL, challenge__hidden=False, hidden=False, )
Remove hidden public pages from sitemap
Remove hidden public pages from sitemap
Python
apache-2.0
comic/comic-django,comic/comic-django,comic/comic-django,comic/comic-django,comic/comic-django
c5239c6bbb40ede4279b33b965c5ded26a78b2ae
app/tests/manual/test_twitter_api.py
app/tests/manual/test_twitter_api.py
# -*- coding: utf-8 -*- """ Twitter API test module. Local test to check that Twitter credentials are valid connect to Twitter API and that the auth functions can be used to do this. s""" from __future__ import absolute_import from unittest import TestCase from lib.twitter_api import authentication class TestAuth(TestCase): def test_generateAppAccessToken(self): auth = authentication._generateAppAccessToken() def test_getTweepyConnection(self): auth = authentication._generateAppAccessToken() api = authentication._getTweepyConnection(auth) def test_getAPIConnection(self): """ Test that App Access token can be used to connect to Twitter API. """ api = authentication.getAPIConnection(userFlow=False) def test_getAppOnlyConnection(self): """ Test App-only token. """ api = authentication.getAppOnlyConnection()
#!/usr/bin/env python # -*- coding: utf-8 -*- """ Twitter API test module. Local test to check that Twitter credentials are valid connect to Twitter API and that the auth functions can be used to do this. """ from __future__ import absolute_import import os import sys import unittest from unittest import TestCase # Allow imports to be done when executing this file directly. sys.path.insert(0, os.path.abspath(os.path.join( os.path.dirname(__file__), os.path.pardir, os.path.pardir) )) from lib.twitter_api import authentication class TestAuth(TestCase): def test_generateAppAccessToken(self): auth = authentication._generateAppAccessToken() def test_getTweepyConnection(self): auth = authentication._generateAppAccessToken() api = authentication._getTweepyConnection(auth) def test_getAPIConnection(self): """ Test that App Access token can be used to connect to Twitter API. """ api = authentication.getAPIConnection(userFlow=False) def test_getAppOnlyConnection(self): """ Test App-only token. """ api = authentication.getAppOnlyConnection() if __name__ == '__main__': unittest.main()
Update Twitter auth test to run directly
test: Update Twitter auth test to run directly
Python
mit
MichaelCurrin/twitterverse,MichaelCurrin/twitterverse
c6862c5f864db4e77dd835f074efdd284667e6fd
util/ldjpp.py
util/ldjpp.py
#! /usr/bin/env python from __future__ import print_function import argparse import json parser = argparse.ArgumentParser(description='Pretty-print LDJSON.') parser.add_argument('--indent', metavar='N', type=int, default=2, dest='indent', help='indentation for pretty-printing') parser.add_argument('--file', metavar='FILE', required=True, dest='file', type=argparse.FileType('r'), help='input LDJSON file') parser.add_argument('--sort', action='store_true', dest='sortkeys', help='sort object keys') args = parser.parse_args() for line in args.file: record = json.loads(line) print(json.dumps(record, indent=args.indent, sort_keys=args.sortkeys))
#! /usr/bin/env python from __future__ import print_function import click import json from collections import OrderedDict def json_loader(sortkeys): def _loader(line): if sortkeys: return json.loads(line) else: # if --no-sortkeys, let's preserve file order return json.JSONDecoder(object_pairs_hook=OrderedDict).decode(line) return _loader @click.command() @click.option('indent', '-i', '--indent', default=2, help='indentation for pretty-printing') @click.option('--sortkeys/--no-sortkeys', default=False, help='sort object keys') @click.argument('infile', type=click.File()) def cli(indent, sortkeys, infile): """Pretty-print LDJSON.""" loader = json_loader(sortkeys) for line in infile: record = loader(line) print(json.dumps(record, indent=indent, sort_keys=sortkeys)) if __name__ == '__main__': cli()
Use click instead of argparse
Use click instead of argparse
Python
mit
mhyfritz/goontools,mhyfritz/goontools,mhyfritz/goontools
fdfa3aae605eaadf099c6d80c86a9406f34fb71c
bluebottle/organizations/urls/api.py
bluebottle/organizations/urls/api.py
from django.conf.urls import url from bluebottle.organizations.views import ( OrganizationList, OrganizationDetail, OrganizationContactList, OrganizationContactDetail ) urlpatterns = [ url(r'^$', OrganizationList.as_view(), name='organization_list'), url(r'^/(?P<pk>\d+)$', OrganizationDetail.as_view(), name='organization_detail'), url(r'^/contacts/$', OrganizationContactList.as_view(), name='organization_contact_list'), url(r'^/contacts/(?P<pk>\d+)$', OrganizationContactDetail.as_view(), name='organization_contact_detail') ]
from django.conf.urls import url from bluebottle.organizations.views import ( OrganizationList, OrganizationDetail, OrganizationContactList, OrganizationContactDetail ) urlpatterns = [ url(r'^$', OrganizationList.as_view(), name='organization_list'), url(r'^/(?P<pk>\d+)$', OrganizationDetail.as_view(), name='organization_detail'), url(r'^/contacts$', OrganizationContactList.as_view(), name='organization_contact_list'), url(r'^/contacts/(?P<pk>\d+)$', OrganizationContactDetail.as_view(), name='organization_contact_detail') ]
Fix organization-contact url having an extra slash
Fix organization-contact url having an extra slash
Python
bsd-3-clause
onepercentclub/bluebottle,onepercentclub/bluebottle,onepercentclub/bluebottle,onepercentclub/bluebottle,onepercentclub/bluebottle
b7decb588f5b6e4d15fb04fa59aa571e5570cbfe
djangae/contrib/contenttypes/apps.py
djangae/contrib/contenttypes/apps.py
from django.apps import AppConfig from django.utils.translation import ugettext_lazy as _ from django.contrib.contenttypes.management import update_contenttypes as django_update_contenttypes from django.db.models.signals import post_migrate from .management import update_contenttypes from .models import SimulatedContentTypeManager class ContentTypesConfig(AppConfig): name = 'djangae.contrib.contenttypes' verbose_name = _("Djangae Content Types") label = "djangae_contenttypes" def ready(self): if django_update_contenttypes != update_contenttypes: post_migrate.disconnect(django_update_contenttypes) from django.db import models from django.contrib.contenttypes import models as django_models if not isinstance(django_models.ContentType.objects, SimulatedContentTypeManager): django_models.ContentType.objects = SimulatedContentTypeManager() django_models.ContentType.objects.auto_created = True # Really force the default manager to use the Simulated one meta = django_models.ContentType._meta meta.local_managers[0] = SimulatedContentTypeManager() meta._expire_cache() # Our generated IDs take up a 64 bit range (signed) but aren't auto # incrementing so update the field to reflect that (for validation) meta.pk.__class__ = models.BigIntegerField
from django.apps import AppConfig from django.utils.translation import ugettext_lazy as _ from django.contrib.contenttypes.management import update_contenttypes as django_update_contenttypes from django.db.models.signals import post_migrate from .management import update_contenttypes from .models import SimulatedContentTypeManager class ContentTypesConfig(AppConfig): name = 'djangae.contrib.contenttypes' verbose_name = _("Djangae Content Types") label = "djangae_contenttypes" def ready(self): if django_update_contenttypes != update_contenttypes: post_migrate.disconnect(django_update_contenttypes) from django.db import models from django.contrib.contenttypes import models as django_models if not isinstance(django_models.ContentType.objects, SimulatedContentTypeManager): django_models.ContentType.objects = SimulatedContentTypeManager(django_models.ContentType) django_models.ContentType.objects.auto_created = True # Really force the default manager to use the Simulated one meta = django_models.ContentType._meta if hasattr(meta, "local_managers"): # Django >= 1.10 meta.local_managers[0] = SimulatedContentTypeManager() else: django_models.ContentType._default_manager = SimulatedContentTypeManager(django_models.ContentType) meta._expire_cache() # Our generated IDs take up a 64 bit range (signed) but aren't auto # incrementing so update the field to reflect that (for validation) meta.pk.__class__ = models.BigIntegerField
Fix up for Django 1.9
Fix up for Django 1.9
Python
bsd-3-clause
grzes/djangae,potatolondon/djangae,grzes/djangae,potatolondon/djangae,grzes/djangae
c02239af435cece9c2664436efbe0b2aeb200a1b
stats/views.py
stats/views.py
from django.contrib.auth.decorators import user_passes_test from django.shortcuts import render from django.db.models import Sum, Count from common.auth import user_is_admin from django.utils.translation import ugettext_lazy as _ from common.models import Purchase, BookType from egielda import settings @user_passes_test(user_is_admin) def index(request): stats = dict() stats['books_sold_value'] = BookType.objects.filter(book__sold=True).annotate(count=Count('book')).aggregate( Sum('price', field='count * price'))['price__sum'] return render(request, 'stats/index.html', {'page_title': _("Statistics"), 'stats': stats, 'currency': getattr(settings, 'CURRENCY', 'USD')}) @user_passes_test(user_is_admin) def books_sold(request): Purchase.objects.all().order_by('-date') return render(request, 'stats/books_sold.html', {'page_title': _("Books sold")})
from django.contrib.auth.decorators import user_passes_test from django.shortcuts import render from django.db.models import Sum, Count from common.auth import user_is_admin from django.utils.translation import ugettext_lazy as _ from common.models import Purchase, BookType from egielda import settings @user_passes_test(user_is_admin) def index(request): stats = dict() stats['books_sold_value'] = BookType.objects.filter(book__sold=True).annotate(count=Count('book')).aggregate( Sum('price', field='count * price'))['price__sum'] or 0 return render(request, 'stats/index.html', {'page_title': _("Statistics"), 'stats': stats, 'currency': getattr(settings, 'CURRENCY', 'USD')}) @user_passes_test(user_is_admin) def books_sold(request): Purchase.objects.all().order_by('-date') return render(request, 'stats/books_sold.html', {'page_title': _("Books sold")})
Fix displaying None in statistics when there's no book sold
Fix displaying None in statistics when there's no book sold
Python
agpl-3.0
m4tx/egielda,m4tx/egielda,m4tx/egielda
dfd3bff4560d1711624b8508795eb3debbaafa40
changes/api/snapshotimage_details.py
changes/api/snapshotimage_details.py
from __future__ import absolute_import from flask.ext.restful import reqparse from changes.api.base import APIView from changes.config import db from changes.models import SnapshotImage, SnapshotStatus class SnapshotImageDetailsAPIView(APIView): parser = reqparse.RequestParser() parser.add_argument('status', choices=SnapshotStatus._member_names_) def get(self, image_id): image = SnapshotImage.query.get(image_id) if image is None: return '', 404 return self.respond(image) def post(self, image_id): image = SnapshotImage.query.get(image_id) if image is None: return '', 404 args = self.parser.parse_args() if args.status: image.status = SnapshotStatus[args.status] db.session.add(image) db.session.flush() if image.status == SnapshotStatus.active: snapshot = image.snapshot inactive_image_query = SnapshotImage.query.filter( SnapshotImage.status != SnapshotStatus.active, SnapshotImage.snapshot_id == snapshot.id, ).exists() if not db.session.query(inactive_image_query).scalar(): snapshot.status = SnapshotStatus.active db.session.add(snapshot) db.session.commit() return self.respond(image)
from __future__ import absolute_import from flask.ext.restful import reqparse from changes.api.base import APIView from changes.config import db from changes.models import SnapshotImage, SnapshotStatus class SnapshotImageDetailsAPIView(APIView): parser = reqparse.RequestParser() parser.add_argument('status', choices=SnapshotStatus._member_names_) def get(self, image_id): image = SnapshotImage.query.get(image_id) if image is None: return '', 404 return self.respond(image) def post(self, image_id): image = SnapshotImage.query.get(image_id) if image is None: return '', 404 args = self.parser.parse_args() if args.status: image.status = SnapshotStatus[args.status] db.session.add(image) db.session.flush() if image.status == SnapshotStatus.active: snapshot = image.snapshot inactive_image_query = SnapshotImage.query.filter( SnapshotImage.status != SnapshotStatus.active, SnapshotImage.snapshot_id == snapshot.id, ).exists() if not db.session.query(inactive_image_query).scalar(): snapshot.status = SnapshotStatus.active db.session.add(snapshot) elif snapshot.status == SnapshotStatus.active: snapshot.status = SnapshotStatus.inactive db.session.add(snapshot) db.session.commit() return self.respond(image)
Mark snapshots as inactive if any are not valid
Mark snapshots as inactive if any are not valid
Python
apache-2.0
dropbox/changes,bowlofstew/changes,wfxiang08/changes,bowlofstew/changes,wfxiang08/changes,wfxiang08/changes,dropbox/changes,bowlofstew/changes,dropbox/changes,dropbox/changes,bowlofstew/changes,wfxiang08/changes
f8b4b1a860b5c0a3ff16dbb8bbf83010bd9a1009
feincms3/plugins/__init__.py
feincms3/plugins/__init__.py
# flake8: noqa from . import html from . import snippet try: from . import external except ImportError: # pragma: no cover pass try: from . import image except ImportError: # pragma: no cover pass try: from . import richtext except ImportError: # pragma: no cover pass try: from . import versatileimage except ImportError: # pragma: no cover pass
# flake8: noqa from . import html from . import snippet try: import requests except ImportError: # pragma: no cover pass else: from . import external try: import imagefield except ImportError: # pragma: no cover pass else: from . import image try: import feincms3.cleanse except ImportError: # pragma: no cover pass else: from . import richtext try: import versatileimagefield except ImportError: # pragma: no cover pass else: from . import versatileimage
Stop hiding local import errors
feincms3.plugins: Stop hiding local import errors
Python
bsd-3-clause
matthiask/feincms3,matthiask/feincms3,matthiask/feincms3
b2eebbdcc14dd47d6ad8bb385966f13ed13890c1
superdesk/coverages.py
superdesk/coverages.py
from superdesk.base_model import BaseModel def init_app(app): CoverageModel(app=app) def rel(resource, embeddable=False): return { 'type': 'objectid', 'data_relation': {'resource': resource, 'field': '_id', 'embeddable': embeddable} } class CoverageModel(BaseModel): endpoint_name = 'coverages' schema = { 'headline': {'type': 'string'}, 'type': {'type': 'string'}, 'ed_note': {'type': 'string'}, 'scheduled': {'type': 'datetime'}, 'delivery': rel('archive'), 'assigned_user': rel('users', True), 'assigned_desk': rel('desks', True), 'planning_item': rel('planning'), }
from superdesk.base_model import BaseModel def init_app(app): CoverageModel(app=app) def rel(resource, embeddable=False): return { 'type': 'objectid', 'data_relation': {'resource': resource, 'field': '_id', 'embeddable': embeddable} } class CoverageModel(BaseModel): endpoint_name = 'coverages' schema = { 'headline': {'type': 'string'}, 'type': {'type': 'string'}, 'ed_note': {'type': 'string'}, 'scheduled': {'type': 'datetime'}, 'delivery': {'type': 'string'}, 'assigned_user': rel('users', True), 'assigned_desk': rel('desks', True), 'planning_item': {'type': 'string'}, }
Fix data relation not working for custom Guids
Fix data relation not working for custom Guids
Python
agpl-3.0
plamut/superdesk,sivakuna-aap/superdesk,mdhaman/superdesk-aap,sivakuna-aap/superdesk,liveblog/superdesk,pavlovicnemanja/superdesk,petrjasek/superdesk,mugurrus/superdesk,ioanpocol/superdesk,pavlovicnemanja/superdesk,Aca-jov/superdesk,akintolga/superdesk,vied12/superdesk,gbbr/superdesk,fritzSF/superdesk,ancafarcas/superdesk,ioanpocol/superdesk-ntb,mdhaman/superdesk-aap,marwoodandrew/superdesk-aap,darconny/superdesk,akintolga/superdesk-aap,amagdas/superdesk,sivakuna-aap/superdesk,thnkloud9/superdesk,fritzSF/superdesk,akintolga/superdesk-aap,ancafarcas/superdesk,akintolga/superdesk,pavlovicnemanja92/superdesk,amagdas/superdesk,vied12/superdesk,verifiedpixel/superdesk,superdesk/superdesk-ntb,Aca-jov/superdesk,superdesk/superdesk,akintolga/superdesk,marwoodandrew/superdesk-aap,hlmnrmr/superdesk,verifiedpixel/superdesk,pavlovicnemanja/superdesk,petrjasek/superdesk-server,pavlovicnemanja/superdesk,liveblog/superdesk,thnkloud9/superdesk,superdesk/superdesk-aap,plamut/superdesk,vied12/superdesk,ioanpocol/superdesk-ntb,plamut/superdesk,darconny/superdesk,sjunaid/superdesk,superdesk/superdesk-aap,amagdas/superdesk,verifiedpixel/superdesk,vied12/superdesk,darconny/superdesk,fritzSF/superdesk,sivakuna-aap/superdesk,amagdas/superdesk,superdesk/superdesk-aap,gbbr/superdesk,marwoodandrew/superdesk-aap,mdhaman/superdesk,petrjasek/superdesk-ntb,sivakuna-aap/superdesk,akintolga/superdesk-aap,akintolga/superdesk,superdesk/superdesk-ntb,fritzSF/superdesk,marwoodandrew/superdesk,marwoodandrew/superdesk,verifiedpixel/superdesk,amagdas/superdesk,marwoodandrew/superdesk-aap,sjunaid/superdesk,petrjasek/superdesk-ntb,sjunaid/superdesk,Aca-jov/superdesk,ioanpocol/superdesk-ntb,petrjasek/superdesk-server,pavlovicnemanja92/superdesk,superdesk/superdesk,ancafarcas/superdesk,superdesk/superdesk,mdhaman/superdesk,fritzSF/superdesk,vied12/superdesk,pavlovicnemanja92/superdesk,petrjasek/superdesk-ntb,hlmnrmr/superdesk,marwoodandrew/superdesk,superdesk/superdesk-ntb,petrjasek/superdesk,hlmnrmr/superdesk,petrjasek/superdesk,mugurrus/superdesk,gbbr/superdesk,pavlovicnemanja92/superdesk,plamut/superdesk,liveblog/superdesk,verifiedpixel/superdesk,petrjasek/superdesk,mdhaman/superdesk,petrjasek/superdesk-ntb,mugurrus/superdesk,mdhaman/superdesk-aap,superdesk/superdesk-aap,superdesk/superdesk-ntb,marwoodandrew/superdesk,pavlovicnemanja92/superdesk,ioanpocol/superdesk,thnkloud9/superdesk,marwoodandrew/superdesk,liveblog/superdesk,mdhaman/superdesk-aap,ioanpocol/superdesk,akintolga/superdesk,plamut/superdesk,liveblog/superdesk,superdesk/superdesk,akintolga/superdesk-aap
4147e6f560889c75abbfd9c8e85ea38ffe408550
suelta/mechanisms/facebook_platform.py
suelta/mechanisms/facebook_platform.py
from suelta.util import bytes from suelta.sasl import Mechanism, register_mechanism try: import urlparse except ImportError: import urllib.parse as urlparse class X_FACEBOOK_PLATFORM(Mechanism): def __init__(self, sasl, name): super(X_FACEBOOK_PLATFORM, self).__init__(sasl, name) self.check_values(['access_token', 'api_key']) def process(self, challenge=None): if challenge is not None: values = {} for kv in challenge.split('&'): key, value = kv.split('=') values[key] = value resp_data = { 'method': values['method'], 'v': '1.0', 'call_id': '1.0', 'nonce': values['nonce'], 'access_token': self.values['access_token'], 'api_key': self.values['api_key'] } resp = '&'.join(['%s=%s' % (k, v) for k, v in resp_data.items()]) return bytes(resp) return bytes('') def okay(self): return True register_mechanism('X-FACEBOOK-PLATFORM', 40, X_FACEBOOK_PLATFORM, use_hashes=False)
from suelta.util import bytes from suelta.sasl import Mechanism, register_mechanism try: import urlparse except ImportError: import urllib.parse as urlparse class X_FACEBOOK_PLATFORM(Mechanism): def __init__(self, sasl, name): super(X_FACEBOOK_PLATFORM, self).__init__(sasl, name) self.check_values(['access_token', 'api_key']) def process(self, challenge=None): if challenge is not None: values = {} for kv in challenge.split(b'&'): key, value = kv.split(b'=') values[key] = value resp_data = { b'method': values[b'method'], b'v': b'1.0', b'call_id': b'1.0', b'nonce': values[b'nonce'], b'access_token': self.values['access_token'], b'api_key': self.values['api_key'] } resp = '&'.join(['%s=%s' % (k, v) for k, v in resp_data.items()]) return bytes(resp) return b'' def okay(self): return True register_mechanism('X-FACEBOOK-PLATFORM', 40, X_FACEBOOK_PLATFORM, use_hashes=False)
Work around Python3's byte semantics.
Work around Python3's byte semantics.
Python
mit
dwd/Suelta
1dbe7acc945a545d3b18ec5025c19b26d1ed110f
test/test_sparql_construct_bindings.py
test/test_sparql_construct_bindings.py
from rdflib import Graph, URIRef, Literal, BNode from rdflib.plugins.sparql import prepareQuery from rdflib.compare import isomorphic import unittest class TestConstructInitBindings(unittest.TestCase): def test_construct_init_bindings(self): """ This is issue https://github.com/RDFLib/rdflib/issues/1001 """ g1 = Graph() q_str = (""" PREFIX : <urn:ns1:> CONSTRUCT { ?uri :prop1 ?val1; :prop2 ?c . } WHERE { bind(uri(concat("urn:ns1:", ?a)) as ?uri) bind(?b as ?val1) } """) q_prepared = prepareQuery(q_str) expected = [ (URIRef('urn:ns1:A'),URIRef('urn:ns1:prop1'), Literal('B')), (URIRef('urn:ns1:A'),URIRef('urn:ns1:prop2'), Literal('C')) ] results = g1.query(q_prepared, initBindings={ 'a': Literal('A'), 'b': Literal('B'), 'c': Literal('C') }) self.assertCountEqual(list(results), expected)
from rdflib import Graph, URIRef, Literal, BNode from rdflib.plugins.sparql import prepareQuery from rdflib.compare import isomorphic import unittest from nose.tools import eq_ class TestConstructInitBindings(unittest.TestCase): def test_construct_init_bindings(self): """ This is issue https://github.com/RDFLib/rdflib/issues/1001 """ g1 = Graph() q_str = (""" PREFIX : <urn:ns1:> CONSTRUCT { ?uri :prop1 ?val1; :prop2 ?c . } WHERE { bind(uri(concat("urn:ns1:", ?a)) as ?uri) bind(?b as ?val1) } """) q_prepared = prepareQuery(q_str) expected = [ (URIRef('urn:ns1:A'),URIRef('urn:ns1:prop1'), Literal('B')), (URIRef('urn:ns1:A'),URIRef('urn:ns1:prop2'), Literal('C')) ] results = g1.query(q_prepared, initBindings={ 'a': Literal('A'), 'b': Literal('B'), 'c': Literal('C') }) eq_(sorted(results, key=lambda x: str(x[1])), expected)
Fix unit tests for python2
Fix unit tests for python2
Python
bsd-3-clause
RDFLib/rdflib,RDFLib/rdflib,RDFLib/rdflib,RDFLib/rdflib
2ebbe2f9f23621d10a70d0817d83da33b002299e
rest_surveys/urls.py
rest_surveys/urls.py
from __future__ import unicode_literals from django.conf import settings from django.conf.urls import include, url from rest_framework_bulk.routes import BulkRouter from rest_surveys.views import ( SurveyViewSet, SurveyResponseViewSet, ) # API # With trailing slash appended: router = BulkRouter() router.register(r'surveys', SurveyViewSet, base_name='survey') router.register(r'survey-responses', SurveyResponseViewSet, base_name='survey-response') slashless_router = BulkRouter(trailing_slash=False) slashless_router.registry = router.registry[:] urlpatterns = [ url(r'^{api_path}'.format(api_path=settings.REST_SURVEYS['API_PATH']), include(router.urls)), url(r'^{api_path}'.format(api_path=settings.REST_SURVEYS['API_PATH']), include(slashless_router.urls)), ]
from __future__ import unicode_literals from django.conf import settings from django.conf.urls import include, url from rest_framework_bulk.routes import BulkRouter from rest_surveys.views import ( SurveyViewSet, SurveyResponseViewSet, ) # API # With trailing slash appended: router = BulkRouter() router.register(r'surveys', SurveyViewSet, base_name='survey') router.register(r'survey-responses', SurveyResponseViewSet, base_name='survey-response') slashless_router = BulkRouter(trailing_slash=False) slashless_router.registry = router.registry[:] urlpatterns = [ url(r'^{api_path}'.format( api_path=settings.REST_SURVEYS.get('API_PATH', 'api/')), include(router.urls)), url(r'^{api_path}'.format( api_path=settings.REST_SURVEYS.get('API_PATH', 'api/')), include(slashless_router.urls)), ]
Set a default api path
Set a default api path
Python
mit
danxshap/django-rest-surveys
1cbd56988478320268838f77e8cc6237d95346fd
test/dunya/conn_test.py
test/dunya/conn_test.py
import unittest from compmusic.dunya.conn import _make_url class ConnTest(unittest.TestCase): def test_make_url(self): params = {"first": "%^grtà"} url = _make_url("path", **params) self.assertEqual(url, 'http://dunya.compmusic.upf.edu/path?first=%25%5Egrt%C3%A0')
#!/usr/local/bin/python # -*- coding: utf-8 -*- import unittest from compmusic.dunya.conn import _make_url class ConnTest(unittest.TestCase): def test_make_url(self): params = {"first": "%^grtà"} url = _make_url("path", **params) self.assertEqual(url, 'http://dunya.compmusic.upf.edu/path?first=%25%5Egrt%C3%A0')
Declare the encoding of conn.py as utf-8
Declare the encoding of conn.py as utf-8
Python
agpl-3.0
MTG/pycompmusic
a7437e657f55cd708baba83421941e67d474daf7
tests/test_utilities.py
tests/test_utilities.py
from __future__ import (absolute_import, division, print_function) from folium.utilities import camelize def test_camelize(): assert camelize('variable_name') == 'variableName' assert camelize('variableName') == 'variableName' assert camelize('name') == 'name' assert camelize('very_long_variable_name') == 'veryLongVariableName'
from __future__ import (absolute_import, division, print_function) from folium.utilities import camelize, deep_copy from folium import Map, FeatureGroup, Marker def test_camelize(): assert camelize('variable_name') == 'variableName' assert camelize('variableName') == 'variableName' assert camelize('name') == 'name' assert camelize('very_long_variable_name') == 'veryLongVariableName' def test_deep_copy(): m = Map() fg = FeatureGroup().add_to(m) Marker(location=(0, 0)).add_to(fg) m_copy = deep_copy(m) def check(item, item_copy): assert type(item) is type(item_copy) assert item._name == item_copy._name for attr in item.__dict__.keys(): if not attr.startswith('_'): assert getattr(item, attr) == getattr(item_copy, attr) assert item is not item_copy assert item._id != item_copy._id for child, child_copy in zip(item._children.values(), item_copy._children.values()): check(child, child_copy) check(m, m_copy)
Add test for deep_copy function
Add test for deep_copy function
Python
mit
python-visualization/folium,ocefpaf/folium,ocefpaf/folium,python-visualization/folium
fe05b5f694671a46dd3391b9cb6561923345c4b7
rpi_gpio_http/app.py
rpi_gpio_http/app.py
from flask import Flask import logging import logging.config import RPi.GPIO as GPIO from .config import config, config_loader from .channel import ChannelFactory app = Flask('rpi_gpio_http') logging.config.dictConfig(config['logger']) logger = logging.getLogger(__name__) logger.info("Config loaded from %s" % config_loader.filename) channels = {} GPIO.setmode(GPIO.BOARD) for ch in config['channels']: if ch['enabled'] != True: continue channel = ChannelFactory.create(ch) if channel: channels[channel.pin] = channel import controllers
from flask import Flask import logging import logging.config import RPi.GPIO as GPIO from .config import config, config_loader from .channel import ChannelFactory app = Flask('rpi_gpio_http') logging.config.dictConfig(config['logger']) logger = logging.getLogger(__name__) logger.info("Config loaded from %s" % config_loader.filename) channels = {} GPIO.setwarnings(False) GPIO.setmode(GPIO.BOARD) for ch in config['channels']: if ch['enabled'] != True: continue channel = ChannelFactory.create(ch) if channel: channels[channel.pin] = channel import controllers
Disable warnings in GPIO lib
Disable warnings in GPIO lib
Python
mit
voidpp/rpi-gpio-http
378f55687131324bb5c43e3b50f9db5fe3b39662
zaqar_ui/__init__.py
zaqar_ui/__init__.py
# Copyright 2015 IBM Corp. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import pbr.version __version__ = pbr.version.VersionInfo( 'neutron_lbaas_dashboard').version_string()
# Copyright 2015 IBM Corp. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import pbr.version __version__ = pbr.version.VersionInfo('zaqar_ui').version_string()
Fix Zaqar-ui with wrong reference pbr version
Fix Zaqar-ui with wrong reference pbr version Change-Id: I84cdb865478a232886ba1059febf56735a0d91ba
Python
apache-2.0
openstack/zaqar-ui,openstack/zaqar-ui,openstack/zaqar-ui,openstack/zaqar-ui
38888d34506b743a06aa93f5dc6c187844774d58
scripts/constants.py
scripts/constants.py
# Copyright 2016 The Kubernetes Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import sys # Kubernetes branch to get the OpenAPI spec from. KUBERNETES_BRANCH = "release-1.5" # Spec version will be set in downloaded spec and all # generated code will refer to it. SPEC_VERSION = "v1.5.0-snapshot" # client version for packaging and releasing. It can # be different than SPEC_VERSION. CLIENT_VERSION = "1.0.0-snapshot" # Name of the release package PACKAGE_NAME = "kubernetes" # If called directly, return the constant value given # its name. Useful in bash scripts. if __name__ == '__main__': if len(sys.argv) != 2: print("Usage: python constant.py CONSTANT_NAME") sys.exit(1) if sys.argv[1] in globals(): print globals()[sys.argv[1]] else: print "Cannot find constant %s" % sys.argv[1] sys.exit(1)
# Copyright 2016 The Kubernetes Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import sys # Kubernetes branch to get the OpenAPI spec from. KUBERNETES_BRANCH = "release-1.5" # Spec version will be set in downloaded spec and all # generated code will refer to it. SPEC_VERSION = "v1.5.0-snapshot" # client version for packaging and releasing. It can # be different than SPEC_VERSION. CLIENT_VERSION = "1.0.0-snapshot" # Name of the release package PACKAGE_NAME = "kubernetes" # If called directly, return the constant value given # its name. Useful in bash scripts. if __name__ == '__main__': if len(sys.argv) != 2: print("Usage: python constant.py CONSTANT_NAME") sys.exit(1) if sys.argv[1] in globals(): print(globals()[sys.argv[1]]) else: print("Cannot find constant %s" % sys.argv[1]) sys.exit(1)
Add missing parentheses to print()
Add missing parentheses to print()
Python
apache-2.0
skuda/client-python,mbohlool/client-python,kubernetes-client/python,djkonro/client-python,sebgoa/client-python,skuda/client-python,mbohlool/client-python,kubernetes-client/python,sebgoa/client-python,djkonro/client-python
d659c685f40de7eb7b2ccd007888177fb158e139
tests/integration/players.py
tests/integration/players.py
#!/usr/bin/env python import urllib.parse import urllib.request def create_player(username, password, email): url = 'https://localhost:3000/players' values = {'username' : username, 'password' : password, 'email' : email } data = urllib.parse.urlencode(values) data = data.encode('utf-8') # data should be bytes req = urllib.request.Request(url, data) response = urllib.request.urlopen(req) the_page = response.read() print("Created user \'{}\' with password \'{}\' and email \'{}\'".format(username, password, email)) if __name__ == '__main__': create_player("chapmang", "password", "[email protected]") create_player("idlee", "deadparrot", "[email protected]") create_player("gilliamt", "lumberjack", "[email protected]") create_player("jonest", "trojanrabbit", "[email protected]") create_player("cleesej", "generaldirection", "[email protected]") create_player("palinm", "fleshwound", "[email protected]")
#!/usr/bin/env python import requests def create_player(username, password, email): url = 'https://localhost:3000/players' values = {'username' : username, 'password' : password, 'email' : email } r = requests.post(url, params=values, verify=False) r.raise_for_status() if (r.status_code == 201): print("Created user \'{}\' with password \'{}\' and email \'{}\'".format(username, password, email)) if __name__ == '__main__': create_player("chapmang", "password", "[email protected]") create_player("idlee", "deadparrot", "[email protected]") create_player("gilliamt", "lumberjack", "[email protected]") create_player("jonest", "trojanrabbit", "[email protected]") create_player("cleesej", "generaldirection", "[email protected]") create_player("palinm", "fleshwound", "[email protected]")
Switch to requests library instead of urllib
Switch to requests library instead of urllib
Python
mit
dropshot/dropshot-server
eeeba609afe732b8e95aa535e70d4cdd2ae1aac7
tests/unit/test_cufflinks.py
tests/unit/test_cufflinks.py
import os import unittest import shutil from bcbio.rnaseq import cufflinks from bcbio.utils import file_exists, safe_makedir from nose.plugins.attrib import attr DATA_DIR = os.path.join(os.path.dirname(__file__), "bcbio-nextgen-test-data", "data") class TestCufflinks(unittest.TestCase): merged_gtf = os.path.join(DATA_DIR, "cufflinks", "merged.gtf") ref_gtf = os.path.join(DATA_DIR, "cufflinks", "ref-transcripts.gtf") out_dir = "cufflinks-test" def setUp(self): safe_makedir(self.out_dir) @attr("unit") def test_cufflinks_clean(self): clean_fn = os.path.join(self.out_dir, "clean.gtf") dirty_fn = os.path.join(self.out_dir, "dirty.gtf") clean, dirty = cufflinks.clean_assembly(self.merged_gtf, clean_fn, dirty_fn) # fixed_fn = os.path.join(self.out_dir, "fixed.gtf") # fixed = cufflinks.fix_cufflinks_attributes(self.ref_gtf, clean, fixed_fn) assert(file_exists(clean)) assert(os.path.exists(dirty)) # assert(file_exists(fixed)) def tearDown(self): shutil.rmtree(self.out_dir)
import os import unittest import shutil from bcbio.rnaseq import cufflinks from bcbio.utils import file_exists, safe_makedir from nose.plugins.attrib import attr DATA_DIR = os.path.join(os.path.dirname(__file__), "bcbio-nextgen-test-data", "data") class TestCufflinks(unittest.TestCase): merged_gtf = os.path.join(DATA_DIR, "cufflinks", "merged.gtf") ref_gtf = os.path.join(DATA_DIR, "cufflinks", "ref-transcripts.gtf") out_dir = "cufflinks-test" def setUp(self): safe_makedir(self.out_dir) @attr("unit") def test_cufflinks_clean(self): clean_fn = os.path.join(self.out_dir, "clean.gtf") dirty_fn = os.path.join(self.out_dir, "dirty.gtf") clean, dirty = cufflinks.clean_assembly(self.merged_gtf, clean_fn, dirty_fn) assert(file_exists(clean)) assert(os.path.exists(dirty)) def tearDown(self): shutil.rmtree(self.out_dir)
Remove some cruft from the cufflinks test.
Remove some cruft from the cufflinks test.
Python
mit
vladsaveliev/bcbio-nextgen,biocyberman/bcbio-nextgen,verdurin/bcbio-nextgen,fw1121/bcbio-nextgen,gifford-lab/bcbio-nextgen,chapmanb/bcbio-nextgen,Cyberbio-Lab/bcbio-nextgen,hjanime/bcbio-nextgen,verdurin/bcbio-nextgen,lbeltrame/bcbio-nextgen,verdurin/bcbio-nextgen,SciLifeLab/bcbio-nextgen,chapmanb/bcbio-nextgen,lpantano/bcbio-nextgen,vladsaveliev/bcbio-nextgen,elkingtonmcb/bcbio-nextgen,mjafin/bcbio-nextgen,brainstorm/bcbio-nextgen,lbeltrame/bcbio-nextgen,guillermo-carrasco/bcbio-nextgen,fw1121/bcbio-nextgen,a113n/bcbio-nextgen,brainstorm/bcbio-nextgen,SciLifeLab/bcbio-nextgen,mjafin/bcbio-nextgen,elkingtonmcb/bcbio-nextgen,mjafin/bcbio-nextgen,lbeltrame/bcbio-nextgen,biocyberman/bcbio-nextgen,Cyberbio-Lab/bcbio-nextgen,chapmanb/bcbio-nextgen,gifford-lab/bcbio-nextgen,lpantano/bcbio-nextgen,lpantano/bcbio-nextgen,elkingtonmcb/bcbio-nextgen,gifford-lab/bcbio-nextgen,fw1121/bcbio-nextgen,vladsaveliev/bcbio-nextgen,guillermo-carrasco/bcbio-nextgen,a113n/bcbio-nextgen,Cyberbio-Lab/bcbio-nextgen,hjanime/bcbio-nextgen,SciLifeLab/bcbio-nextgen,brainstorm/bcbio-nextgen,biocyberman/bcbio-nextgen,hjanime/bcbio-nextgen,a113n/bcbio-nextgen,guillermo-carrasco/bcbio-nextgen
ff09f40b763ac9c968919871d649c47ce6aa7489
main.py
main.py
from BaseHTTPServer import HTTPServer from HTTPHandler import HTTPHandler import socket from threading import currentThread from Cron import Cron from Settings import PORT from Update import check from Event import addEventHandler from event_handlers import * currentThread().name = 'main' check() try: server = HTTPServer(('', PORT), HTTPHandler) except socket.error, (errno, msg): print "Unable to open port %d: %s" % (PORT, msg) exit(1) Cron.start() # addEventHandler(DebugLogger.DebugLogger()) addEventHandler(DBLogger.DBLogger()) addEventHandler(MessageDispatcher.MessageDispatcher()) try: server.serve_forever() except KeyboardInterrupt: pass server.server_close()
from BaseHTTPServer import HTTPServer from HTTPHandler import HTTPHandler import socket from threading import currentThread import signal from Cron import Cron from Settings import PORT from Update import check from Event import addEventHandler from event_handlers import * currentThread().name = 'main' check() try: server = HTTPServer(('', PORT), HTTPHandler) except socket.error, (errno, msg): print "Unable to open port %d: %s" % (PORT, msg) exit(1) Cron.start() # addEventHandler(DebugLogger.DebugLogger()) addEventHandler(DBLogger.DBLogger()) addEventHandler(MessageDispatcher.MessageDispatcher()) # When python is started in the background it ignores SIGINT instead of throwing a KeyboardInterrupt signal.signal(signal.SIGINT, signal.default_int_handler) try: server.serve_forever() except KeyboardInterrupt: pass server.server_close()
Handle SIGINT even if it's ignored by default
Handle SIGINT even if it's ignored by default
Python
mit
mrozekma/Sprint,mrozekma/Sprint,mrozekma/Sprint
ddfd7a3a2a2806045c6f4114c3f7f5a0ca929b7c
main.py
main.py
#!/usr/bin/env python # -*- coding: utf-8 -*- import os import logging from datetime import datetime from update_wrapper import UpdateWrapper if not os.path.isdir("log"): os.mkdir("log") logging.basicConfig( filename="log/{}.log".format(datetime.now().strftime("%Y%m%d%H%M%S%f")), level=logging.DEBUG) logging.captureWarnings(True) wrapper = UpdateWrapper() wrapper.read_config("config.json") wrapper.run()
#!/usr/bin/env python # -*- coding: utf-8 -*- import os import logging from datetime import datetime from update_wrapper import UpdateWrapper if not os.path.isdir("log"): os.mkdir("log") LOG_FILE = datetime.now().strftime("%Y%m%d%H%M%S%f") logging.basicConfig( filename="log/{}.log".format(LOG_FILE), level=logging.DEBUG) logging.captureWarnings(True) wrapper = UpdateWrapper() wrapper.read_config("config.json") wrapper.run()
Move log file to constant
Move log file to constant
Python
mit
stormaaja/csvconverter,stormaaja/csvconverter,stormaaja/csvconverter
c956fbbbc6e4dbd713728c1feda6bce2956a0894
runtime/Python3/src/antlr4/__init__.py
runtime/Python3/src/antlr4/__init__.py
from antlr4.Token import Token from antlr4.InputStream import InputStream from antlr4.FileStream import FileStream from antlr4.BufferedTokenStream import TokenStream from antlr4.CommonTokenStream import CommonTokenStream from antlr4.Lexer import Lexer from antlr4.Parser import Parser from antlr4.dfa.DFA import DFA from antlr4.atn.ATN import ATN from antlr4.atn.ATNDeserializer import ATNDeserializer from antlr4.atn.LexerATNSimulator import LexerATNSimulator from antlr4.atn.ParserATNSimulator import ParserATNSimulator from antlr4.atn.PredictionMode import PredictionMode from antlr4.PredictionContext import PredictionContextCache from antlr4.ParserRuleContext import RuleContext, ParserRuleContext from antlr4.tree.Tree import ParseTreeListener, ParseTreeVisitor, ParseTreeWalker, TerminalNode, ErrorNode, RuleNode from antlr4.error.Errors import RecognitionException, IllegalStateException, NoViableAltException from antlr4.error.ErrorStrategy import BailErrorStrategy from antlr4.error.DiagnosticErrorListener import DiagnosticErrorListener from antlr4.Utils import str_list
from antlr4.Token import Token from antlr4.InputStream import InputStream from antlr4.FileStream import FileStream from antlr4.StdinStream import StdinStream from antlr4.BufferedTokenStream import TokenStream from antlr4.CommonTokenStream import CommonTokenStream from antlr4.Lexer import Lexer from antlr4.Parser import Parser from antlr4.dfa.DFA import DFA from antlr4.atn.ATN import ATN from antlr4.atn.ATNDeserializer import ATNDeserializer from antlr4.atn.LexerATNSimulator import LexerATNSimulator from antlr4.atn.ParserATNSimulator import ParserATNSimulator from antlr4.atn.PredictionMode import PredictionMode from antlr4.PredictionContext import PredictionContextCache from antlr4.ParserRuleContext import RuleContext, ParserRuleContext from antlr4.tree.Tree import ParseTreeListener, ParseTreeVisitor, ParseTreeWalker, TerminalNode, ErrorNode, RuleNode from antlr4.error.Errors import RecognitionException, IllegalStateException, NoViableAltException from antlr4.error.ErrorStrategy import BailErrorStrategy from antlr4.error.DiagnosticErrorListener import DiagnosticErrorListener from antlr4.Utils import str_list
Allow importing StdinStream from antlr4 package
Allow importing StdinStream from antlr4 package
Python
bsd-3-clause
parrt/antlr4,ericvergnaud/antlr4,antlr/antlr4,antlr/antlr4,ericvergnaud/antlr4,parrt/antlr4,ericvergnaud/antlr4,parrt/antlr4,antlr/antlr4,parrt/antlr4,parrt/antlr4,antlr/antlr4,antlr/antlr4,antlr/antlr4,ericvergnaud/antlr4,ericvergnaud/antlr4,ericvergnaud/antlr4,parrt/antlr4,antlr/antlr4,antlr/antlr4,ericvergnaud/antlr4,ericvergnaud/antlr4,antlr/antlr4,ericvergnaud/antlr4,antlr/antlr4,parrt/antlr4,parrt/antlr4,ericvergnaud/antlr4,parrt/antlr4,parrt/antlr4
14c22be85b9c9b3d13cad1130bb8d8d83d69d68a
selenium_testcase/testcases/content.py
selenium_testcase/testcases/content.py
# -*- coding: utf-8 -*- from __future__ import absolute_import from .utils import dom_contains, wait_for class ContentTestMixin: def should_see_immediately(self, text): """ Assert that DOM contains the given text. """ self.assertTrue(dom_contains(self.browser, text)) @wait_for def should_see(self, text): """ Wait for text to appear before testing assertion. """ return self.should_see_immediately(text) def should_not_see(self, text): """ Wait for text to not appear before testing assertion. """ self.assertRaises(AssertionError, self.should_see, text) @wait_for def has_title(self, title): """ Assert that page title matches. """ self.assertEqual(self.browser.title, title) def has_not_title(self, title): """ Assert when page title does not match. """ self.assertRaises(AssertionError, self.has_title, title) @wait_for def title_contains(self, text): """ Assert that page title contains text. """ self.assertIn(text, self.browser.title) def title_does_not_contain(self, text): """ Assert that page title does not contain text. """ self.assertRaises(AssertionError, self.title_contains, text)
# -*- coding: utf-8 -*- from __future__ import absolute_import from .utils import wait_for from selenium.common.exceptions import NoSuchElementException from selenium.webdriver.common.by import By class ContentTestMixin: content_search_list = ( (By.XPATH, '//*[contains(normalize-space(.), "{}") ' 'and not(./*[contains(normalize-space(.), "{}")])]',), ) def should_see_immediately(self, text, **kwargs): """ Assert that DOM contains the given text. """ self.find_element( self.content_search_list, text, text, **kwargs) @wait_for def should_see(self, text): """ Wait for text to appear before testing assertion. """ return self.should_see_immediately(text) def should_not_see(self, text): """ Wait for text to not appear before testing assertion. """ self.assertRaises(NoSuchElementException, self.should_see, text) @wait_for def has_title(self, title): """ Assert that page title matches. """ self.assertEqual(self.browser.title, title) def has_not_title(self, title): """ Assert when page title does not match. """ self.assertRaises(AssertionError, self.has_title, title) @wait_for def title_contains(self, text): """ Assert that page title contains text. """ self.assertIn(text, self.browser.title) def title_does_not_contain(self, text): """ Assert that page title does not contain text. """ self.assertRaises(AssertionError, self.title_contains, text)
Update should_see_immediately to use local find_element method.
Update should_see_immediately to use local find_element method. This commit adds a content_search_list and replaces dom_contains with our local version of find_element. It adds an attribute called content_search_list that can be overridden by the derived TestCase class as necessary for corner cases.
Python
bsd-3-clause
nimbis/django-selenium-testcase,nimbis/django-selenium-testcase
7947d474da8bb086493890d81a6788d76e00b108
numba/cuda/tests/__init__.py
numba/cuda/tests/__init__.py
from numba.testing import SerialSuite from numba.testing import load_testsuite from numba import cuda from os.path import dirname, join def load_tests(loader, tests, pattern): suite = SerialSuite() this_dir = dirname(__file__) suite.addTests(load_testsuite(loader, join(this_dir, 'nocuda'))) suite.addTests(load_testsuite(loader, join(this_dir, 'cudasim'))) if cuda.is_available(): gpus = cuda.list_devices() if gpus and gpus[0].compute_capability >= (2, 0): suite.addTests(load_testsuite(loader, join(this_dir, 'cudadrv'))) suite.addTests(load_testsuite(loader, join(this_dir, 'cudapy'))) else: print("skipped CUDA tests because GPU CC < 2.0") else: print("skipped CUDA tests") return suite
from numba.testing import SerialSuite from numba.testing import load_testsuite from numba import cuda from os.path import dirname, join def load_tests(loader, tests, pattern): suite = SerialSuite() this_dir = dirname(__file__) suite.addTests(load_testsuite(loader, join(this_dir, 'nocuda'))) if cuda.is_available(): suite.addTests(load_testsuite(loader, join(this_dir, 'cudasim'))) gpus = cuda.list_devices() if gpus and gpus[0].compute_capability >= (2, 0): suite.addTests(load_testsuite(loader, join(this_dir, 'cudadrv'))) suite.addTests(load_testsuite(loader, join(this_dir, 'cudapy'))) else: print("skipped CUDA tests because GPU CC < 2.0") else: print("skipped CUDA tests") return suite
Fix tests on machine without CUDA
Fix tests on machine without CUDA
Python
bsd-2-clause
sklam/numba,numba/numba,seibert/numba,IntelLabs/numba,jriehl/numba,stonebig/numba,gmarkall/numba,cpcloud/numba,IntelLabs/numba,gmarkall/numba,jriehl/numba,cpcloud/numba,sklam/numba,cpcloud/numba,numba/numba,stonebig/numba,stefanseefeld/numba,sklam/numba,cpcloud/numba,seibert/numba,sklam/numba,gmarkall/numba,stefanseefeld/numba,jriehl/numba,numba/numba,cpcloud/numba,stefanseefeld/numba,IntelLabs/numba,numba/numba,IntelLabs/numba,stuartarchibald/numba,jriehl/numba,sklam/numba,IntelLabs/numba,numba/numba,stonebig/numba,stuartarchibald/numba,stonebig/numba,jriehl/numba,gmarkall/numba,stefanseefeld/numba,stuartarchibald/numba,stuartarchibald/numba,stonebig/numba,stefanseefeld/numba,seibert/numba,seibert/numba,gmarkall/numba,stuartarchibald/numba,seibert/numba
910d1288adddd0c8dd500c1be5e488502c1ed335
localflavor/nl/forms.py
localflavor/nl/forms.py
# -*- coding: utf-8 -*- """NL-specific Form helpers.""" from __future__ import unicode_literals from django import forms from django.utils import six from .nl_provinces import PROVINCE_CHOICES from .validators import NLBSNFieldValidator, NLZipCodeFieldValidator class NLZipCodeField(forms.CharField): """A Dutch zip code field.""" default_validators = [NLZipCodeFieldValidator()] def clean(self, value): if isinstance(value, six.string_types): value = value.upper().replace(' ', '') if len(value) == 6: value = '%s %s' % (value[:4], value[4:]) return super(NLZipCodeField, self).clean(value) class NLProvinceSelect(forms.Select): """A Select widget that uses a list of provinces of the Netherlands as it's choices.""" def __init__(self, attrs=None): super(NLProvinceSelect, self).__init__(attrs, choices=PROVINCE_CHOICES) class NLBSNFormField(forms.CharField): """ A Dutch social security number (BSN) field. http://nl.wikipedia.org/wiki/Sofinummer .. versionadded:: 1.6 """ default_validators = [NLBSNFieldValidator()] def __init__(self, *args, **kwargs): kwargs['max_length'] = 9 super(NLBSNFormField, self).__init__(*args, **kwargs)
# -*- coding: utf-8 -*- """NL-specific Form helpers.""" from __future__ import unicode_literals from django import forms from django.utils import six from .nl_provinces import PROVINCE_CHOICES from .validators import NLBSNFieldValidator, NLZipCodeFieldValidator class NLZipCodeField(forms.CharField): """A Dutch zip code field.""" default_validators = [NLZipCodeFieldValidator()] def clean(self, value): if isinstance(value, six.string_types): value = value.upper().replace(' ', '') if len(value) == 6: value = '%s %s' % (value[:4], value[4:]) return super(NLZipCodeField, self).clean(value) class NLProvinceSelect(forms.Select): """A Select widget that uses a list of provinces of the Netherlands as it's choices.""" def __init__(self, attrs=None): super(NLProvinceSelect, self).__init__(attrs, choices=PROVINCE_CHOICES) class NLBSNFormField(forms.CharField): """ A Dutch social security number (BSN) field. https://nl.wikipedia.org/wiki/Burgerservicenummer Note that you may only process the BSN if you have a legal basis to do so! .. versionadded:: 1.6 """ default_validators = [NLBSNFieldValidator()] def __init__(self, *args, **kwargs): kwargs['max_length'] = 9 super(NLBSNFormField, self).__init__(*args, **kwargs)
Fix the wikipedia link and include a warning
Fix the wikipedia link and include a warning
Python
bsd-3-clause
django/django-localflavor,rsalmaso/django-localflavor
2e5ec8483930ad328b0a212ccc4b746f73b18c4c
pinax/ratings/tests/tests.py
pinax/ratings/tests/tests.py
from django.test import TestCase from django.contrib.auth.models import User from pinax.ratings.models import Rating from .models import Car class Tests(TestCase): def setUp(self): self.paltman = User.objects.create(username="paltman") self.jtauber = User.objects.create(username="jtauber") self.bronco = Car.objects.create(name="Ford Bronco") self.blazer = Car.objects.create(name="Cheverolet Blazer") self.expedition = Car.objects.create(name="Ford Expedition") def test_rating(self): overall = Rating.update(self.bronco, self.paltman, rating=5) self.assertEquals(overall, 5) overall = Rating.update(self.bronco, self.jtauber, rating=2) self.assertEquals(overall, 3.5)
from decimal import Decimal from django.test import TestCase from django.contrib.auth.models import User from pinax.ratings.models import Rating from .models import Car class Tests(TestCase): def setUp(self): self.paltman = User.objects.create(username="paltman") self.jtauber = User.objects.create(username="jtauber") self.bronco = Car.objects.create(name="Ford Bronco") self.blazer = Car.objects.create(name="Cheverolet Blazer") self.expedition = Car.objects.create(name="Ford Expedition") def test_rating(self): overall = Rating.update(self.bronco, self.paltman, rating=5) self.assertEquals(overall, Decimal("5")) overall = Rating.update(self.bronco, self.jtauber, rating=2) self.assertEquals(overall, Decimal("3.5"))
Use explicit Decimal in test
Use explicit Decimal in test
Python
mit
rizumu/pinax-ratings,pinax/pinax-ratings,arthur-wsw/pinax-ratings,arthur-wsw/pinax-ratings,pinax/pinax-ratings,arthur-wsw/pinax-ratings,pinax/pinax-ratings,rizumu/pinax-ratings,rizumu/pinax-ratings
95fcaffa1dc73ec3c83734587c311b47e79e0d3c
pylamb/bmi_ilamb.py
pylamb/bmi_ilamb.py
#! /usr/bin/env python import sys import subprocess class BmiIlamb(object): _command = 'run_ilamb' _args = None _env = None def __init__(self): self._time = self.get_start_time() @property def args(self): return [self._command] + (self._args or []) def get_component_name(self): return 'ILAMB' def initialize(self, filename): self._args = [filename or 'ILAMB_PARA_SETUP'] def update(self, time): subprocess.check_call(self.args, shell=False, env=self._env) self._time = self.get_end_time() def update_until(self, time): self.update(time) def finalize(self): pass def get_input_var_names(self): return () def get_output_var_names(self): return () def get_start_time(self): return 0.0 def get_end_time(self): return 1.0 def get_current_time(self): return self._time
#! /usr/bin/env python import sys import subprocess class BmiIlamb(object): _command = 'run_ilamb' _args = None _env = None def __init__(self): self._time = self.get_start_time() @property def args(self): return [self._command] + (self._args or []) def get_component_name(self): return 'ILAMB' def initialize(self, filename): self._args = [filename or 'ILAMB_PARA_SETUP'] def update(self): subprocess.check_call(self.args, shell=False, env=self._env) self._time = self.get_end_time() def update_until(self, time): self.update(time) def finalize(self): pass def get_input_var_names(self): return () def get_output_var_names(self): return () def get_start_time(self): return 0.0 def get_end_time(self): return 1.0 def get_current_time(self): return self._time
Update no longer takes an argument
Update no longer takes an argument See the docs: http://bmi-python.readthedocs.io.
Python
mit
permamodel/ILAMB,permamodel/ILAMB,permamodel/ILAMB
b07c26c4d00de2b7dd184e0d173ec9e03ce4b456
qtui/exam_wizard.py
qtui/exam_wizard.py
from PyQt4.QtGui import * from master_page import MasterPage from generate_page import GeneratePage from scan_page import ScanPage from scores_page import ScoresPage from results_page import ResultsPage class ExamWizard(QWizard): def __init__(self, project): super(ExamWizard, self).__init__() self.project = project self.order = None # TODO: Implement order loading here? self.results = None # TODO: Implement result loading here? self.addPage(MasterPage(project, self)) self.addPage(GeneratePage(project, self)) self.addPage(ScanPage(project, self)) self.addPage(ScoresPage(project, self)) self.addPage(ResultsPage(project, self))
from PyQt4.QtGui import * from master_page import MasterPage from generate_page import GeneratePage from scan_page import ScanPage from scores_page import ScoresPage from results_page import ResultsPage class ExamWizard(QWizard): def __init__(self, project): super(ExamWizard, self).__init__() self.project = project self.order = None # TODO: Implement order loading here? self.results = None # TODO: Implement result loading here? self.setOption(QWizard.IndependentPages, False) self.addPage(MasterPage(project, self)) self.addPage(GeneratePage(project, self)) self.addPage(ScanPage(project, self)) # self.addPage(ScoresPage(project, self)) self.addPage(ResultsPage(project, self))
Comment out temporally scoresheet editing page
Comment out temporally scoresheet editing page
Python
mit
matcom/autoexam,matcom/autoexam,matcom/autoexam,matcom/autoexam,matcom/autoexam
41a0fa6412427dadfb33c77da45bc88c576fa67c
rdo/drivers/base.py
rdo/drivers/base.py
from subprocess import call class BaseDriver(object): def __init__(self, config): self.config = config def do(self, cmd): cmd = self.command(cmd) call(cmd) def command(self): raise NotImplementedError()
from subprocess import call class BaseDriver(object): def __init__(self, config): self.config = config def working_dir(self, cmd): command = ' '.join(cmd) working_dir = self.config.get('directory') if working_dir: command = 'cd %s && %s' % (working_dir, command) return command def do(self, cmd): cmd = self.command(cmd) call(cmd) def command(self): raise NotImplementedError()
Add a common function for deriving the working dir.
Add a common function for deriving the working dir.
Python
bsd-3-clause
ionrock/rdo
3940fd8b58b6a21627ef0ff62f7480593e5108eb
remedy/radremedy.py
remedy/radremedy.py
#!/usr/bin/env python """ radremedy.py Main web application file. Contains initial setup of database, API, and other components. Also contains the setup of the routes. """ from flask import Flask, url_for, request, abort from flask.ext.script import Manager from flask.ext.migrate import Migrate, MigrateCommand from rad.models import db, Resource def create_app(config, models=()): from remedyblueprint import remedy, url_for_other_page app = Flask(__name__) app.config.from_object(config) app.register_blueprint(remedy) # searching configurations app.jinja_env.trim_blocks = True # Register the paging helper method with Jinja2 app.jinja_env.globals['url_for_other_page'] = url_for_other_page db.init_app(app) Migrate(app, db, directory=app.config['MIGRATIONS_DIR']) manager = Manager(app) manager.add_command('db', MigrateCommand) # turning API off for now # from api_manager import init_api_manager # api_manager = init_api_manager(app, db) # map(lambda m: api_manager.create_api(m), models) return app, manager if __name__ == '__main__': app, manager = create_app('config.BaseConfig', (Resource, )) with app.app_context(): manager.run()
#!/usr/bin/env python """ radremedy.py Main web application file. Contains initial setup of database, API, and other components. Also contains the setup of the routes. """ from flask import Flask, url_for, request, abort from flask.ext.script import Manager from flask.ext.migrate import Migrate, MigrateCommand from flask.ext.login import current_user from rad.models import db, Resource def create_app(config, models=()): app = Flask(__name__) app.config.from_object(config) from remedyblueprint import remedy, url_for_other_page app.register_blueprint(remedy) from auth.user_auth import auth, login_manager app.register_blueprint(auth) login_manager.init_app(app) # searching configurations app.jinja_env.trim_blocks = True # Register the paging helper method with Jinja2 app.jinja_env.globals['url_for_other_page'] = url_for_other_page app.jinja_env.globals['logged_in'] = lambda : not current_user.is_anonymous() db.init_app(app) Migrate(app, db, directory=app.config['MIGRATIONS_DIR']) manager = Manager(app) manager.add_command('db', MigrateCommand) # turning API off for now # from api_manager import init_api_manager # api_manager = init_api_manager(app, db) # map(lambda m: api_manager.create_api(m), models) return app, manager if __name__ == '__main__': application, manager = create_app('config.BaseConfig', (Resource, )) with application.app_context(): manager.run()
Move around imports and not shadow app
Move around imports and not shadow app
Python
mpl-2.0
radremedy/radremedy,radioprotector/radremedy,radioprotector/radremedy,AllieDeford/radremedy,radremedy/radremedy,radremedy/radremedy,radioprotector/radremedy,radremedy/radremedy,AllieDeford/radremedy,AllieDeford/radremedy,radioprotector/radremedy