commit
stringlengths 40
40
| old_file
stringlengths 4
264
| new_file
stringlengths 4
264
| old_contents
stringlengths 0
3.26k
| new_contents
stringlengths 1
4.43k
| subject
stringlengths 15
624
| message
stringlengths 15
4.7k
| lang
stringclasses 3
values | license
stringclasses 13
values | repos
stringlengths 5
91.5k
|
---|---|---|---|---|---|---|---|---|---|
167ca3f2a91cd20f38b32ab204855a1e86785c67 | st2common/st2common/constants/meta.py | st2common/st2common/constants/meta.py | # Copyright 2020 The StackStorm Authors.
# Copyright 2019 Extreme Networks, Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from __future__ import absolute_import
import yaml
try:
from yaml import CSafeLoader as YamlSafeLoader
except ImportError:
from yaml import SafeLoader as YamlSafeLoader
__all__ = ["ALLOWED_EXTS", "PARSER_FUNCS"]
# NOTE: We utilize CSafeLoader if available since it uses C extensions and is faster.
def yaml_safe_load(stream):
return yaml.load(stream, Loader=YamlSafeLoader)
ALLOWED_EXTS = [".yaml", ".yml"]
PARSER_FUNCS = {".yml": yaml_safe_load, ".yaml": yaml_safe_load}
| # Copyright 2020 The StackStorm Authors.
# Copyright 2019 Extreme Networks, Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from __future__ import absolute_import
import yaml
try:
from yaml import CSafeLoader as YamlSafeLoader
except ImportError:
from yaml import SafeLoader as YamlSafeLoader
__all__ = ["ALLOWED_EXTS", "PARSER_FUNCS"]
# NOTE: We utilize CSafeLoader if available since it uses C extensions and is faster.
#
# SafeLoader / CSafeLoader are both safe to use and don't allow loading arbitrary Python objects.
#
# That's the actual class which is used internally by ``yaml.safe_load()``, but we can't use that
# method directly since we want to use C extension if available (CSafeLoader) for faster parsing.
#
# See pyyaml docs for details https://pyyaml.org/wiki/PyYAMLDocumentation
def yaml_safe_load(stream):
return yaml.load(stream, Loader=YamlSafeLoader)
ALLOWED_EXTS = [".yaml", ".yml"]
PARSER_FUNCS = {".yml": yaml_safe_load, ".yaml": yaml_safe_load}
| Add a comment to custom yaml_safe_load() method. | Add a comment to custom yaml_safe_load() method.
| Python | apache-2.0 | StackStorm/st2,StackStorm/st2,Plexxi/st2,StackStorm/st2,nzlosh/st2,StackStorm/st2,Plexxi/st2,Plexxi/st2,Plexxi/st2,nzlosh/st2,nzlosh/st2,nzlosh/st2 |
0661728da6f38f91d2afa145e0db523a91241e61 | plugins/noah_stuff.py | plugins/noah_stuff.py | from plugin import Plugin
from utils import create_logger
from config import OwnerID
class Jacob_Noah(Plugin):
is_global = True
log = create_logger('noah_stuff')
async def on_message(self, message, pfx):
if message.content.startswith('<@' + self.client.id + '> kish meh'):
await self.client.send_typing(message.channel)
cmd_name = 'Jacob Noah Meme'
self.log.info('User %s [%s] on server %s [%s], used the ' + cmd_name + ' command on #%s channel',
message.author,
message.author.id, message.server.name, message.server.id, message.channel)
if message.author.id == OwnerID:
await self.client.send_message(message.channel, 'Of course <@' + OwnerID + '>, my love, anything for you! Chu~')
else:
await self.client.send_message(message.channel, 'Ew <@' + OwnerID + '>... Would kindly piss off...') | from plugin import Plugin
from utils import create_logger
from config import OwnerID
class Jacob_Noah(Plugin):
is_global = True
log = create_logger('noah_stuff')
async def on_message(self, message, pfx):
if message.content.startswith('<@' + self.client.user.id + '> kish meh'):
await self.client.send_typing(message.channel)
cmd_name = 'Jacob Noah Meme'
self.log.info('User %s [%s] on server %s [%s], used the ' + cmd_name + ' command on #%s channel',
message.author,
message.author.id, message.server.name, message.server.id, message.channel)
if message.author.id == OwnerID:
await self.client.send_message(message.channel, 'Of course <@' + OwnerID + '>, my love, anything for you! Chu~')
else:
await self.client.send_message(message.channel, 'Ew <@' + OwnerID + '>... Would kindly piss off...') | Fix for the noah thing | Fix for the noah thing
| Python | mpl-2.0 | mralext20/apex-sigma,mralext20/apex-sigma |
39b382770ecc1b4c4fc33273fac0467344ccadfc | relay_api/api/main.py | relay_api/api/main.py | from flask import Flask
import relay_api.api.server as backend
server = Flask(__name__)
backend.init_relays()
@server.route("/relay-api/relays/", methods=["GET"])
def get_all_relays():
js = backend.get_all_relays()
return js, 200
@server.route("/relay-api/relays/<relay_name>", methods=["GET"])
def get_relay(relay_name):
js = backend.get_relay(relay_name)
if not js:
return "", 404
return js, 200
| from flask import Flask
import relay_api.api.backend as backend
server = Flask(__name__)
backend.init_relays()
@server.route("/relay-api/relays/", methods=["GET"])
def get_all_relays():
js = backend.get_all_relays()
return js, 200
@server.route("/relay-api/relays/<relay_name>", methods=["GET"])
def get_relay(relay_name):
js = backend.get_relay(relay_name)
if not js:
return "", 404
return js, 200
| Fix issue with backend import | Fix issue with backend import
| Python | mit | pahumadad/raspi-relay-api |
1747fa822c6daf63d0875cf7b3da37423b0932dc | apps/api/paginators.py | apps/api/paginators.py | from rest_framework.pagination import PageNumberPagination
from rest_framework.response import Response
class QuotesResultsSetPagination(PageNumberPagination):
page_size = 10
page_size_query_param = 'page_size'
max_page_size = 10000
def get_paginated_response(self, data):
return Response({
'pages': {
'next': self.page.next_page_number() if self.page.has_next() else None,
'previous': self.page.previous_page_number() if self.page.has_previous() else None
},
'count': self.page.paginator.count,
'results': data
})
| from rest_framework.pagination import PageNumberPagination
from rest_framework.response import Response
class QuotesResultsSetPagination(PageNumberPagination):
page_size = 25
page_size_query_param = 'page_size'
max_page_size = 10000
def get_paginated_response(self, data):
return Response({
'pages': {
'next': self.page.next_page_number() if self.page.has_next() else None,
'previous': self.page.previous_page_number() if self.page.has_previous() else None
},
'count': self.page.paginator.count,
'results': data
})
| Change page size to 25 | Change page size to 25
| Python | bsd-3-clause | lucifurtun/myquotes,lucifurtun/myquotes,lucifurtun/myquotes,lucifurtun/myquotes |
7a281be50ba1fc59281a76470776fa9c8efdfd54 | pijobs/scrolljob.py | pijobs/scrolljob.py | import scrollphat
from pijobs.scrollphatjob import ScrollphatJob
class ScrollJob(ScrollphatJob):
def default_options(self):
opts = {
'brightness': 2,
'interval': 0.1,
'sleep': 1.0,
}
return opts
def init(self):
self.set_brightness()
self.write_message()
def write_message(self):
message = self.parse_message()
scrollphat.write_string(message, 11)
def message(self):
return self.options['message']
def parse_message(self):
message = self.message() + ' '
if self.options['upper'] == True:
message = message.upper()
return message
def run(self):
length = scrollphat.buffer_len()
if self.options['loop'] == True:
counter = 0
while True:
self.scroll()
counter += 1
if counter % length == 0:
time.sleep(self.options['sleep'])
else:
for i in range(length):
self.scroll()
self.sleep()
def scroll(self):
scrollphat.scroll()
self.sleep_interval()
| import scrollphat
from pijobs.scrollphatjob import ScrollphatJob
class ScrollJob(ScrollphatJob):
def default_options(self):
opts = {
'brightness': 2,
'interval': 0.1,
'sleep': 1.0,
}
return opts
def init(self):
self.set_brightness()
self.set_rotate()
self.write_message()
def write_message(self):
message = self.parse_message()
scrollphat.write_string(message, 11)
def message(self):
return self.options['message']
def parse_message(self):
message = self.message() + ' '
if self.options['upper'] == True:
message = message.upper()
return message
def run(self):
length = scrollphat.buffer_len()
if self.options['loop'] == True:
counter = 0
while True:
self.scroll()
counter += 1
if counter % length == 0:
time.sleep(self.options['sleep'])
else:
for i in range(length):
self.scroll()
self.sleep()
def scroll(self):
scrollphat.scroll()
self.sleep_interval()
| Add back rotatation for scroll job. | Add back rotatation for scroll job.
| Python | mit | ollej/piapi,ollej/piapi |
592ffbcd7fbbc29bfd377b5abadb39aa29f1c88d | foyer/tests/conftest.py | foyer/tests/conftest.py | import pytest
@pytest.fixture(scope="session")
def initdir(tmpdir):
tmpdir.chdir()
| import pytest
@pytest.fixture(autouse=True)
def initdir(tmpdir):
tmpdir.chdir()
| Switch from scope="session" to autouse=True | Switch from scope="session" to autouse=True
| Python | mit | iModels/foyer,mosdef-hub/foyer,mosdef-hub/foyer,iModels/foyer |
ea633b9cf062e28525910f5659b6b8f2ddbb74e3 | accelerator/migrations/0099_update_program_model.py | accelerator/migrations/0099_update_program_model.py | # Generated by Django 2.2.28 on 2022-04-20 13:05
from django.db import (
migrations,
models,
)
class Migration(migrations.Migration):
dependencies = [
('accelerator', '0098_update_startup_update_20220408_0441'),
]
operations = [
migrations.AddField(
model_name='program',
name='hubspot_url',
field=models.URLField(blank=True, null=True),
),
migrations.AddField(
model_name='program',
name='program_image',
field=models.ImageField(null=True, upload_to=''),
),
]
| # Generated by Django 2.2.28 on 2022-04-20 13:05
import sorl.thumbnail.fields
from django.db import (
migrations,
models,
)
class Migration(migrations.Migration):
dependencies = [
('accelerator', '0098_update_startup_update_20220408_0441'),
]
operations = [
migrations.AddField(
model_name='program',
name='hubspot_url',
field=models.URLField(blank=True, null=True),
),
migrations.AddField(
model_name='program',
name='program_image',
field=sorl.thumbnail.fields.ImageField(
null=True,
upload_to='program_images'),
),
]
| Fix image field import and migration | [AC-9452] Fix image field import and migration
| Python | mit | masschallenge/django-accelerator,masschallenge/django-accelerator |
5c5e0c7809d8db77327dc259df0547579e515a54 | server-side/Judge.py | server-side/Judge.py | import sys
sys.path.append('../audio2tokens')
from audio2tokens import *
download_dir='/home/michele/Downloads'
from rhyme import get_score
import time
def retrieve_tokens(audiofile):
#time.sleep(3)
audio_path_raw = raw_audio(audiofile.replace(':','-'))
tokens = get_text(audio_path_raw)
print(tokens)
return tokens
def rank(namefile):
#audiofile=os.path.join(download_dir,namefile)
tokens=retrieve_tokens(namefile)
if tokens :
score = get_score({'text':tokens})
else:
score = None
print "No tokens found"
print(score)
return {'score':score, 'rhymes_scheme':6} | import sys
sys.path.append('../audio2tokens')
from audio2tokens import *
download_dir='/home/michele/Downloads'
from rhyme import get_score
import time
def remove_audiofiles(audio_path):
audio_path_16k = audio_path.replace('.wav','_16k.wav')
audio_path_raw = audio_path.replace('.wav','.raw')
os.remove(audio_path)
os.remove(audio_path_16k)
os.remove(audio_path_raw)
def retrieve_tokens(audiofile):
#time.sleep(3)
audio_path_raw = raw_audio(audiofile.replace(':','-'))
tokens = get_text(audio_path_raw)
remove_audiofiles(audio_path_raw)
print(tokens)
return tokens
def rank(namefile):
#audiofile=os.path.join(download_dir,namefile)
tokens=retrieve_tokens(namefile)
if tokens :
score = get_score({'text':tokens})
else:
score = None
print "No tokens found"
print(score)
return {'score':score, 'rhymes_scheme':6} | Add some functions to remove wave files | Add some functions to remove wave files
| Python | mit | hajicj/HAMR_2016,hajicj/HAMR_2016,hajicj/HAMR_2016 |
693dc9d8448740e1a1c4543cc3a91e3769fa7a3e | pySPM/utils/plot.py | pySPM/utils/plot.py | import numpy as np
import matplotlib.pyplot as plt
def plotMask(ax, mask, color, **kargs):
import copy
m = np.ma.masked_array(mask, ~mask)
palette = copy.copy(plt.cm.gray)
palette.set_over(color, 1.0)
ax.imshow(m, cmap=palette, vmin=0, vmax=0.5, **kargs)
def Xdist(ax,left, right, y, color='r', linestyle=':', fmt='.2f', xtransf=lambda x: x, **kargs):
ax.axvline(left,color=color, linestyle=linestyle)
ax.axvline(right,color=color, linestyle=linestyle)
s = "{:"+fmt+"}"+kargs.get('unit','')
ax.annotate(s.format(xtransf(right-left)),(.5*(left+right),y),(0,2),textcoords='offset pixels',va='bottom',ha='center')
ax.annotate("",(left,y),(right,y),arrowprops=dict(arrowstyle=kargs.get('arrowstyle','<->'))) | import numpy as np
import matplotlib.pyplot as plt
def plotMask(ax, mask, color, **kargs):
import copy
m = np.ma.masked_array(mask, ~mask)
palette = copy.copy(plt.cm.gray)
palette.set_over(color, 1.0)
ax.imshow(m, cmap=palette, vmin=0, vmax=0.5, **kargs)
def Xdist(ax,left, right, y, color='r', linestyle=':', fmt='.2f', xtransf=lambda x: x, **kargs):
ax.axvline(left,color=color, linestyle=linestyle)
ax.axvline(right,color=color, linestyle=linestyle)
s = "{:"+fmt+"}"+kargs.get('unit','')
ax.annotate(s.format(xtransf(right-left)),(.5*(left+right),y),(0,2),textcoords='offset pixels',va='bottom',ha='center')
ax.annotate("",(left,y),(right,y),arrowprops=dict(arrowstyle=kargs.get('arrowstyle','<->')))
def DualPlot(ax, col1='C0',col2='C1'):
axb = ax.twinx()
axb.spines['left'].set_color(col1)
axb.spines['right'].set_color(col2)
ax.yaxis.label.set_color(col1)
axb.yaxis.label.set_color(col2)
ax.tick_params(axis='y', colors=col1)
axb.tick_params(axis='y', colors=col2)
return axb | Add helper function to create a DualPlot | Add helper function to create a DualPlot
| Python | apache-2.0 | scholi/pySPM |
2bd551d7fa8da9d7641998a5515fba634d65bc56 | comics/feedback/views.py | comics/feedback/views.py | from django.conf import settings
from django.core.mail import mail_admins
from django.core.urlresolvers import reverse
from django.http import HttpResponseRedirect
from django.shortcuts import render
from comics.feedback.forms import FeedbackForm
def feedback(request):
"""Mail feedback to ADMINS"""
if request.method == 'POST':
form = FeedbackForm(request.POST)
if form.is_valid():
subject = 'Feedback from %s' % settings.COMICS_SITE_TITLE
message = form.cleaned_data['message']
mail_admins(subject, message)
return HttpResponseRedirect(reverse('feedback-thanks'))
else:
form = FeedbackForm()
return render(request, 'feedback/form.html', {'feedback_form': form})
def feedback_thanks(request):
"""Display form submit confirmation page"""
return render(request, 'feedback/thanks.html')
| from django.conf import settings
from django.core.mail import mail_admins
from django.core.urlresolvers import reverse
from django.http import HttpResponseRedirect
from django.shortcuts import render
from comics.feedback.forms import FeedbackForm
def feedback(request):
"""Mail feedback to ADMINS"""
if request.method == 'POST':
form = FeedbackForm(request.POST)
if form.is_valid():
subject = 'Feedback from %s' % settings.COMICS_SITE_TITLE
message = form.cleaned_data['message']
metadata = 'Client IP address: %s\n' % request.META['REMOTE_ADDR']
metadata += 'User agent: %s\n' % request.META['HTTP_USER_AGENT']
if request.user.is_authenticated():
metadata += 'User: %s <%s>\n' % (
request.user.username, request.user.email)
else:
metadata += 'User: anonymous\n'
message = '%s\n\n%s' % (message, metadata)
mail_admins(subject, message)
return HttpResponseRedirect(reverse('feedback-thanks'))
else:
form = FeedbackForm()
return render(request, 'feedback/form.html', {'feedback_form': form})
def feedback_thanks(request):
"""Display form submit confirmation page"""
return render(request, 'feedback/thanks.html')
| Add user information to feedback emails | Add user information to feedback emails
| Python | agpl-3.0 | jodal/comics,datagutten/comics,jodal/comics,datagutten/comics,datagutten/comics,jodal/comics,datagutten/comics,jodal/comics |
c7ab4bc8e0b3dbdd305a7a156ef58dddaa37296c | pystorm/__init__.py | pystorm/__init__.py | from .component import Component, Tuple
from .bolt import BatchingBolt, Bolt, TicklessBatchingBolt
from .spout import Spout
__all__ = [
'BatchingBolt',
'Bolt',
'Component',
'Spout',
'TicklessBatchingBolt',
'Tuple',
]
| '''
pystorm is a production-tested Storm multi-lang implementation for Python
It is mostly intended to be used by other libraries (e.g., streamparse).
'''
from .component import Component, Tuple
from .bolt import BatchingBolt, Bolt, TicklessBatchingBolt
from .spout import Spout
from .version import __version__, VERSION
__all__ = [
'BatchingBolt',
'Bolt',
'Component',
'Spout',
'TicklessBatchingBolt',
'Tuple',
]
| Add VERSION and __version__ directly to pystorm namespace | Add VERSION and __version__ directly to pystorm namespace
| Python | apache-2.0 | pystorm/pystorm |
3a5c7cecbca59567a76be1264d15d45824e69199 | python/src/setup.py | python/src/setup.py | """Setup specs for packaging, distributing, and installing gcs lib."""
import distribute_setup
distribute_setup.use_setuptools()
import setuptools
setuptools.setup(
name="GoogleAppEngineCloudStorageClient",
version="1.9.5.0",
packages=setuptools.find_packages(),
author="Google App Engine",
author_email="[email protected]",
keywords="google app engine cloud storage",
url="https://code.google.com/p/appengine-gcs-client/",
license="Apache License 2.0",
description=("This library is the preferred way of accessing Google "
"Cloud Storage from App Engine. It was designed to "
"replace the Files API. As a result it contains much "
"of the same functionality (streaming reads and writes but "
"not the complete set of GCS APIs). It also provides key "
"stability improvements and a better overall developer "
"experience."),
exclude_package_data={"": ["README"]},
zip_safe=True,
)
| """Setup specs for packaging, distributing, and installing gcs lib."""
import distribute_setup
distribute_setup.use_setuptools()
import setuptools
setuptools.setup(
name="GoogleAppEngineCloudStorageClient",
version="1.9.15.0",
packages=setuptools.find_packages(),
author="Google App Engine",
author_email="[email protected]",
keywords="google app engine cloud storage",
url="https://code.google.com/p/appengine-gcs-client/",
license="Apache License 2.0",
description=("This library is the preferred way of accessing Google "
"Cloud Storage from App Engine. It was designed to "
"replace the Files API. As a result it contains much "
"of the same functionality (streaming reads and writes but "
"not the complete set of GCS APIs). It also provides key "
"stability improvements and a better overall developer "
"experience."),
exclude_package_data={"": ["README"]},
zip_safe=True,
)
| Increment GCS Client PyPi version to 1.9.15.0. - Just doing whatever the most recent GAE version is for the package version. | Increment GCS Client PyPi version to 1.9.15.0.
- Just doing whatever the most recent GAE version is for the package version.
R=tkaitchuck
DELTA=1 (0 added, 0 deleted, 1 changed)
Revision created by MOE tool push_codebase.
MOE_MIGRATION=7186
| Python | apache-2.0 | aozarov/appengine-gcs-client,aozarov/appengine-gcs-client,GoogleCloudPlatform/appengine-gcs-client,GoogleCloudPlatform/appengine-gcs-client,aozarov/appengine-gcs-client,GoogleCloudPlatform/appengine-gcs-client |
eb0aeda225cc7c0aef85559857de4cca35b77efd | ratemyflight/urls.py | ratemyflight/urls.py |
from django.conf.urls.defaults import *
from django.views.generic.simple import direct_to_template
urlpatterns = patterns("ratemyflight.views",
url("^api/airport/list/(?P<south>.*)/(?P<west>.*)/(?P<north>.*)/(?P<east>.*)/$",
"airports_for_boundary", name="airports_for_boundary"),
url("^api/flight/list/(?P<south>.*)/(?P<west>.*)/(?P<north>.*)/(?P<east>.*)/$",
"flights_for_boundary", name="flights_for_boundary"),
url("^$", "home", name="home"),
)
|
from django.conf.urls.defaults import *
from django.views.generic.simple import direct_to_template
urlpatterns = patterns("ratemyflight.views",
url("^api/airport/boundary/(?P<south>.*)/(?P<west>.*)/(?P<north>.*)/(?P<east>.*)/$",
"airports_for_boundary", name="airports_for_boundary"),
url("^api/flight/boundary/(?P<south>.*)/(?P<west>.*)/(?P<north>.*)/(?P<east>.*)/$",
"flights_for_boundary", name="flights_for_boundary"),
url("^api/flight/airline/(?P<iata_code>.*)/$",
"flights_for_airline", name="flights_for_airline"),
url("^api/flight/username/(?P<username>.*)/$",
"flights_for_username", name="flights_for_username"),
url("^api/flight/recent/$", "recent_flights", name="recent_flights"),
url("^$", "home", name="home"),
)
| Clean up URLS for API and point final URLS to views. | Clean up URLS for API and point final URLS to views.
| Python | bsd-2-clause | stephenmcd/ratemyflight,stephenmcd/ratemyflight |
1775782f100f9db9ad101a19887ba95fbc36a6e9 | backend/project_name/celerybeat_schedule.py | backend/project_name/celerybeat_schedule.py | from celery.schedules import crontab
CELERYBEAT_SCHEDULE = {
# Internal tasks
"clearsessions": {"schedule": crontab(hour=3, minute=0), "task": "users.tasks.clearsessions"},
}
| from celery.schedules import crontab # pylint:disable=import-error,no-name-in-module
CELERYBEAT_SCHEDULE = {
# Internal tasks
"clearsessions": {"schedule": crontab(hour=3, minute=0), "task": "users.tasks.clearsessions"},
}
| Disable prospector on celery.schedules import | Disable prospector on celery.schedules import
| Python | mit | vintasoftware/django-react-boilerplate,vintasoftware/django-react-boilerplate,vintasoftware/django-react-boilerplate,vintasoftware/django-react-boilerplate |
691e3581f1602714fba33f6dcb139f32e0507d23 | packages/syft/src/syft/core/node/common/node_table/setup.py | packages/syft/src/syft/core/node/common/node_table/setup.py | # third party
from sqlalchemy import Column
from sqlalchemy import Integer
from sqlalchemy import String
# relative
from . import Base
class SetupConfig(Base):
__tablename__ = "setup"
id = Column(Integer(), primary_key=True, autoincrement=True)
domain_name = Column(String(255), default="")
node_id = Column(String(32), default="")
def __str__(self) -> str:
return f"<Domain Name: {self.domain_name}>"
def create_setup(id: int, domain_name: str, node_id: str) -> SetupConfig:
return SetupConfig(id=id, domain_name=domain_name, node_id=node_id)
| # third party
from sqlalchemy import Column
from sqlalchemy import Integer
from sqlalchemy import String
from sqlalchemy import Boolean
# relative
from . import Base
class SetupConfig(Base):
__tablename__ = "setup"
id = Column(Integer(), primary_key=True, autoincrement=True)
domain_name = Column(String(255), default="")
description = Column(String(255), default="")
contact = Column(String(255), default="")
daa = Column(Boolean(), default=False)
node_id = Column(String(32), default="")
def __str__(self) -> str:
return f"<Domain Name: {self.domain_name}>"
def create_setup(id: int, domain_name: str, node_id: str) -> SetupConfig:
return SetupConfig(id=id, domain_name=domain_name, node_id=node_id)
| ADD description / contact / daa fields | ADD description / contact / daa fields
| Python | apache-2.0 | OpenMined/PySyft,OpenMined/PySyft,OpenMined/PySyft,OpenMined/PySyft |
7fc81e93ea44e52ab7c29087a01eac65a145db09 | qiprofile_rest/server/settings.py | qiprofile_rest/server/settings.py | """This ``settings`` file specifies the Eve configuration."""
import os
# The run environment default is production.
# Modify this by setting the NODE_ENV environment variable.
env = os.getenv('NODE_ENV') or 'production'
# The MongoDB database.
if env == 'production':
MONGO_DBNAME = 'qiprofile'
else:
MONGO_DBNAME = 'qiprofile_test'
# The MongoDB host default is localhost, but can be reset
# by the MONGO_HOST environment variable.
host = os.getenv('MONGO_HOST')
if host:
MONGO_HOST = host
# Even though the domain is defined by the Eve MongoEngine
# adapter, a DOMAIN setting is required by Eve. This setting
# is only used to avoid an Eve complaint about a missing domain.
DOMAIN = {'eve-mongoengine': {}}
| """This ``settings`` file specifies the Eve configuration."""
import os
# The run environment default is production.
# Modify this by setting the NODE_ENV environment variable.
env = os.getenv('NODE_ENV') or 'production'
# The MongoDB database.
if env == 'production':
MONGO_DBNAME = 'qiprofile'
else:
MONGO_DBNAME = 'qiprofile_test'
# The MongoDB host default is localhost, but can be reset
# by the MONGO_HOST environment variable.
host = os.getenv('MONGO_HOST')
if host:
MONGO_HOST = host
# The MongoDB port.
port = os.getenv('MONGO_PORT')
if port:
MONGO_PORT = int(port)
# The MongoDB username.
user = os.getenv('MONGO_USERNAME')
if user:
MONGO_USERNAME = user
# The MongoDB password.
pswd = os.getenv('MONGO_PASSWORD')
if pswd:
MONGO_PASSWORD = pswd
# Even though the domain is defined by the Eve MongoEngine
# adapter, a DOMAIN setting is required by Eve. This setting
# is only used to avoid an Eve complaint about a missing domain.
DOMAIN = {'eve-mongoengine': {}}
| Add Mongo env var overrides. | Add Mongo env var overrides.
| Python | bsd-2-clause | ohsu-qin/qiprofile-rest,ohsu-qin/qirest |
b5e368437a600d78e22a53abe53c0103b20daa24 | _python/main/migrations/0003_auto_20191029_2015.py | _python/main/migrations/0003_auto_20191029_2015.py | # Generated by Django 2.2.6 on 2019-10-29 20:15
from django.db import migrations, models
import main.models
class Migration(migrations.Migration):
dependencies = [
('main', '0002_auto_20191007_1639'),
]
operations = [
migrations.AlterField(
model_name='contentnode',
name='headnote',
field=main.models.SanitizingTextField(blank=True, null=True),
),
migrations.AlterField(
model_name='default',
name='url',
field=models.URLField(max_length=1024),
),
migrations.AlterField(
model_name='textblock',
name='content',
field=main.models.SanitizingCharField(max_length=5242880),
),
]
| # Generated by Django 2.2.6 on 2019-10-29 20:15
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('main', '0002_auto_20191007_1639'),
]
operations = [
migrations.AlterField(
model_name='default',
name='url',
field=models.URLField(max_length=1024),
),
]
| Repair migration, which was a no-op in SQL and was 'faked' anyway. | Repair migration, which was a no-op in SQL and was 'faked' anyway.
| Python | agpl-3.0 | harvard-lil/h2o,harvard-lil/h2o,harvard-lil/h2o,harvard-lil/h2o |
f87b979443bd4a578884bdb327dbd72616d07533 | changes/backends/jenkins/generic_builder.py | changes/backends/jenkins/generic_builder.py | from .builder import JenkinsBuilder
class JenkinsGenericBuilder(JenkinsBuilder):
def __init__(self, *args, **kwargs):
self.script = kwargs.pop('script')
self.cluster = kwargs.pop('cluster')
super(JenkinsGenericBuilder, self).__init__(*args, **kwargs)
def get_job_parameters(self, job, script=None, target_id=None, path=''):
params = super(JenkinsGenericBuilder, self).get_job_parameters(
job, target_id=target_id)
if script is None:
script = self.script
project = job.project
repository = project.repository
vcs = repository.get_vcs()
if vcs:
repo_url = vcs.remote_url
else:
repo_url = repository.url
params.extend([
{'name': 'CHANGES_PID', 'value': project.slug},
{'name': 'REPO_URL', 'value': repo_url},
{'name': 'SCRIPT', 'value': script},
{'name': 'REPO_VCS', 'value': repository.backend.name},
{'name': 'CLUSTER', 'value': self.cluster},
{'name': 'WORK_PATH', 'value': path},
])
return params
| from .builder import JenkinsBuilder
class JenkinsGenericBuilder(JenkinsBuilder):
def __init__(self, *args, **kwargs):
self.script = kwargs.pop('script')
self.cluster = kwargs.pop('cluster')
self.path = kwargs.pop('path', '')
super(JenkinsGenericBuilder, self).__init__(*args, **kwargs)
def get_job_parameters(self, job, script=None, target_id=None, path=None):
params = super(JenkinsGenericBuilder, self).get_job_parameters(
job, target_id=target_id)
if path is None:
path = self.path
if script is None:
script = self.script
project = job.project
repository = project.repository
vcs = repository.get_vcs()
if vcs:
repo_url = vcs.remote_url
else:
repo_url = repository.url
params.extend([
{'name': 'CHANGES_PID', 'value': project.slug},
{'name': 'REPO_URL', 'value': repo_url},
{'name': 'SCRIPT', 'value': script},
{'name': 'REPO_VCS', 'value': repository.backend.name},
{'name': 'CLUSTER', 'value': self.cluster},
{'name': 'WORK_PATH', 'value': path},
])
return params
| Support fixed path in generic builder | Support fixed path in generic builder
| Python | apache-2.0 | wfxiang08/changes,bowlofstew/changes,dropbox/changes,dropbox/changes,dropbox/changes,bowlofstew/changes,wfxiang08/changes,bowlofstew/changes,wfxiang08/changes,dropbox/changes,bowlofstew/changes,wfxiang08/changes |
5cff34d49edfc5697b3e84b84d110c49728d0338 | samples/blaze_sql.py | samples/blaze_sql.py | __author__ = 'mark'
| # -*- coding: utf-8 -*-
"""
This example shows some examples of how to access data in SQL databases using
blaze. It walks through how blaze syntax corresponds to SQL queries.
Select Queries
--------------
"""
from __future__ import absolute_import, division, print_function
import sqlite3 as db
from blaze.io.sql import connect, from_table
from blaze import dshape
def create_sqlite_table():
data = [
(4, "Gilbrecht", 17),
(8, "Bertrand", 48),
(16, "Janssen", 32),
]
conn = db.connect(":memory:")
c = conn.cursor()
c.execute('''create table MyTable
(id INTEGER, name TEXT, age INTEGER)''')
c.executemany("""insert into testtable
values (?, ?, ?)""", data)
conn.commit()
c.close()
return conn
conn = create_sqlite_table()
# Describe the columns. Note: typically you would describe column
# with variables for the column size, e.g. dshape('a, int32')
id = from_table('MyTable', 'id', dshape('3, int32'), conn)
name_col = from_table('MyTable', 'name', dshape('3, int32'), conn)
age_col = from_table('MyTable', 'age', dshape('3, int32'), conn)
table = Table([id, name_col, age_col]) # TODO: Better interface
def select():
"""
SELECT * FROM MyTable WHERE MyTable.id > 5
"""
print(table[table.id > 5])
def select_ordered():
"""
SELECT * FROM MyTable WHERE MyTable.id > 5 ORDER BY MyTable.age
"""
print(index(table, table.id > 5, order=table.age))
def groupby():
"""
SELECT *
FROM MyTable
WHERE MyTable.age > 10 AND MyTable.age < 20
GROUP BY MyTable.age
ORDER BY COUNT(*)
"""
teenagers = index(table, table.age > 10 & table.age < 20)
print(groupby(teenagers, table.age, order=count(table)))
def aggregate():
"""
SELECT AVG(age) FROM MyTable WHERE MyTable.id > 5
"""
print(avg(age_col[id > 5])) | Work on SQL example set | Work on SQL example set
| Python | bsd-3-clause | xlhtc007/blaze,cpcloud/blaze,scls19fr/blaze,ContinuumIO/blaze,aterrel/blaze,FrancescAlted/blaze,mrocklin/blaze,FrancescAlted/blaze,ChinaQuants/blaze,maxalbert/blaze,cpcloud/blaze,LiaoPan/blaze,mwiebe/blaze,aterrel/blaze,cowlicks/blaze,FrancescAlted/blaze,alexmojaki/blaze,maxalbert/blaze,mwiebe/blaze,jdmcbr/blaze,caseyclements/blaze,dwillmer/blaze,nkhuyu/blaze,ContinuumIO/blaze,jcrist/blaze,jdmcbr/blaze,jcrist/blaze,nkhuyu/blaze,mwiebe/blaze,ChinaQuants/blaze,FrancescAlted/blaze,scls19fr/blaze,aterrel/blaze,LiaoPan/blaze,alexmojaki/blaze,mrocklin/blaze,cowlicks/blaze,caseyclements/blaze,dwillmer/blaze,xlhtc007/blaze,mwiebe/blaze |
ba0334459b8318a62014ec945a753fc36fc7d519 | account_verification_flask/models/models.py | account_verification_flask/models/models.py | #from flask.ext.login import UserMixin
from account_verification_flask import db, bcrypt
class User(db.Model):
__tablename__ = "users"
id = db.Column(db.Integer, primary_key = True)
name = db.Column(db.String, nullable = False)
email = db.Column(db.String, nullable = False)
password = db.Column(db.String)
phone_number = db.Column(db.String, nullable = False)
country_code = db.Column(db.String, nullable = False)
phone_number_confirmed = db.Column(db.Boolean, nullable = False, default = False)
authy_user_id = db.Column(db.String, nullable = True)
def __init__(self, name, email, password, phone_number, country_code):
self.name = name
self.email = email
self.password = bcrypt.generate_password_hash(password)
self.phone_number = phone_number
self.country_code = country_code
self.phone_number_confirmed = False
def is_authenticated(self):
return True
def is_active(self):
return True
def is_anonymous(self):
return False
def get_id(self):
return unicode(self.id)
def confirm_phone_number(self):
self.phone_number_confirmed = True
def __unicode__(self):
return self.name
def __repr__(self):
return '<User %r>' % (self.name)
| #from flask.ext.login import UserMixin
from account_verification_flask import db, bcrypt
class User(db.Model):
__tablename__ = "users"
id = db.Column(db.Integer, primary_key = True)
name = db.Column(db.String, nullable = False)
email = db.Column(db.String, nullable = False)
password = db.Column(db.String)
phone_number = db.Column(db.String, nullable = False)
country_code = db.Column(db.String, nullable = False)
phone_number_confirmed = db.Column(db.Boolean, nullable = False, default = False)
authy_user_id = db.Column(db.String, nullable = True)
def __init__(self, name, email, password, phone_number, country_code):
self.name = name
self.email = email
self.password = bcrypt.generate_password_hash(password)
self.phone_number = phone_number
self.country_code = country_code
self.phone_number_confirmed = False
def is_authenticated(self):
return True
def is_active(self):
return True
def is_anonymous(self):
return False
def get_id(self):
return unicode(self.id)
def __unicode__(self):
return self.name
def __repr__(self):
return '<User %r>' % (self.name)
| Remove confirm_phone_number method from model | Remove confirm_phone_number method from model
| Python | mit | TwilioDevEd/account-verification-flask,TwilioDevEd/account-verification-flask,TwilioDevEd/account-verification-flask |
1dc2856368e5e6852b526d86a0c78c5fe10b1550 | myhronet/models.py | myhronet/models.py | # -*- coding: utf-8 -*-
import string
from django.db import models
class Blacklist(models.Model):
domain = models.CharField(max_length=255, unique=True, null=True)
def __unicode__(self):
return self.domain
class URL(models.Model):
hashcode = models.CharField(max_length=10, unique=True,
db_index=True, null=True)
longurl = models.CharField(max_length=1024, unique=True,
db_index=True, null=True)
views = models.IntegerField(default=0)
ip = models.GenericIPAddressField(null=True)
data = models.DateTimeField(auto_now_add=True, null=True)
def save(self, *args, **kwargs):
if URL.objects.count():
last = URL.objects.latest('id').pk + 1
alphabet = string.digits + string.ascii_lowercase
base36 = ''
while last != 0:
last, i = divmod(last, len(alphabet))
base36 = alphabet[i] + base36
self.hashcode = base36
else:
self.hashcode = '1'
return super(URL, self).save(*args, **kwargs)
def short_url(self, request):
return ''.join([
request.scheme,
'://', request.get_host(),
'/', self.hashcode,
])
def __unicode__(self):
return ' - '.join([self.hashcode, self.longurl])
| # -*- coding: utf-8 -*-
import string
from django.db import models
class Blacklist(models.Model):
domain = models.CharField(max_length=255, unique=True, null=True)
def __unicode__(self):
return self.domain
class URL(models.Model):
hashcode = models.CharField(max_length=10, unique=True,
db_index=True, null=True)
longurl = models.CharField(max_length=1024, unique=True,
db_index=True, null=True)
views = models.IntegerField(default=0)
ip = models.GenericIPAddressField(null=True)
data = models.DateTimeField(auto_now_add=True, null=True)
def save(self, *args, **kwargs):
if not self.pk:
if URL.objects.count():
last = URL.objects.latest('id').pk + 1
alphabet = string.digits + string.ascii_lowercase
base36 = ''
while last != 0:
last, i = divmod(last, len(alphabet))
base36 = alphabet[i] + base36
self.hashcode = base36
else:
self.hashcode = '1'
return super(URL, self).save(*args, **kwargs)
def short_url(self, request):
return ''.join([
request.scheme,
'://', request.get_host(),
'/', self.hashcode,
])
def __unicode__(self):
return ' - '.join([self.hashcode, self.longurl])
| Fix hashcode generation for existing URLs | Fix hashcode generation for existing URLs
| Python | mit | myhro/myhronet,myhro/myhronet |
1cad9ab61148173b0f61971805b3e6203da3050d | faker/providers/en_CA/ssn.py | faker/providers/en_CA/ssn.py | # coding=utf-8
from __future__ import unicode_literals
from ..ssn import Provider as SsnProvider
class Provider(SsnProvider):
ssn_formats = ("### ### ###",)
@classmethod
def ssn(cls):
return cls.bothify(cls.random_element(cls.ssn_formats)) | # coding=utf-8
from __future__ import unicode_literals
from ..ssn import Provider as SsnProvider
import random
class Provider(SsnProvider):
#in order to create a valid SIN we need to provide a number that passes a simple modified Luhn Algorithmn checksum
#this function essentially reverses the checksum steps to create a random valid SIN (Social Insurance Number)
@classmethod
def ssn(cls):
#create an array of 8 elements initialized randomly
digits = random.sample(range(10), 8)
# All of the digits must sum to a multiple of 10.
# sum the first 8 and set 9th to the value to get to a multiple of 10
digits.append(10 - (sum(digits) % 10))
#digits is now the digital root of the number we want multiplied by the magic number 121 212 121
#reverse the multiplication which occurred on every other element
for i in range(1, len(digits), 2):
if digits[i] % 2 == 0:
digits[i] = (digits[i] / 2)
else:
digits[i] = (digits[i] + 9) / 2
#build the resulting SIN string
sin = ""
for i in range(0, len(digits), 1):
sin += str(digits[i])
#add a space to make it conform to normal standards in Canada
if i % 3 == 2:
sin += " "
#finally return our random but valid SIN
return sin
| Update Canada SSN/SIN provider to create a valid number | Update Canada SSN/SIN provider to create a valid number
The first revision generated a random number in the correct format.
This commit creates a SIN number that passes the checksum as described
here http://http://en.wikipedia.org/wiki/Social_Insurance_Number
| Python | mit | jaredculp/faker,trtd/faker,xfxf/faker-python,HAYASAKA-Ryosuke/faker,johnraz/faker,joke2k/faker,joke2k/faker,venmo/faker,ericchaves/faker,xfxf/faker-1,GLMeece/faker,danhuss/faker,beetleman/faker,thedrow/faker,meganlkm/faker,yiliaofan/faker,MaryanMorel/faker |
dd1ed907532526a4a70694c46918136ca6d93277 | nqueens/nqueens.py | nqueens/nqueens.py | from nqueens.chessboard import Chessboard
from nqueens.printer import Printer
from nqueens.solver import Solver
board = Chessboard.create(8)
solver = Solver.create(board)
solution = solver.solve()
if solution is not None:
printer = Printer.create(solution)
printer.printBoard()
| #!/usr/bin/env python3
import os
import sys
import getopt
sys.path.insert(0, os.path.abspath(os.path.join(os.path.dirname(__file__), '..')))
from nqueens.chessboard import Chessboard
from nqueens.printer import Printer
from nqueens.solver import Solver
def main():
try:
n = parse_command_line()
except ValueError as e:
print("Error: " + str(e))
print("Usage: nqueens.py <n>")
sys.exit(1)
solution = solve_for(n)
if solution is None:
print("No solution found")
else:
print_solution(solution)
def parse_command_line():
try:
opts, args = getopt.getopt(sys.argv[1:], "", [])
except getopt.GetoptError:
raise ValueError("Could not parse command line")
if len(args) == 0:
raise ValueError("No arguments supplied")
if len(args) > 1:
raise ValueError("Too many arguments supplied")
n = args[0]
if not n.isdigit() or int(n) < 1:
raise ValueError("n must be a positive number")
return int(n)
def solve_for(n):
board = Chessboard.create(n)
solver = Solver.create(board)
return solver.solve()
def print_solution(solution):
printer = Printer.create(solution)
printer.printBoard()
if __name__ == '__main__':
sys.exit(main())
| Add ability to run problems from command line | Add ability to run problems from command line
| Python | mit | stevecshanks/nqueens |
c9cc5585e030951a09687c6a61a489ec51f83446 | cr2/plotter/__init__.py | cr2/plotter/__init__.py | # $Copyright:
# ----------------------------------------------------------------
# This confidential and proprietary software may be used only as
# authorised by a licensing agreement from ARM Limited
# (C) COPYRIGHT 2015 ARM Limited
# ALL RIGHTS RESERVED
# The entire notice above must be reproduced on all authorised
# copies and copies may only be made to the extent permitted
# by a licensing agreement from ARM Limited.
# ----------------------------------------------------------------
# File: __init__.py
# ----------------------------------------------------------------
# $
#
"""Init Module for the Plotter Code"""
import pandas as pd
from LinePlot import LinePlot
| # $Copyright:
# ----------------------------------------------------------------
# This confidential and proprietary software may be used only as
# authorised by a licensing agreement from ARM Limited
# (C) COPYRIGHT 2015 ARM Limited
# ALL RIGHTS RESERVED
# The entire notice above must be reproduced on all authorised
# copies and copies may only be made to the extent permitted
# by a licensing agreement from ARM Limited.
# ----------------------------------------------------------------
# File: __init__.py
# ----------------------------------------------------------------
# $
#
"""Init Module for the Plotter Code"""
import pandas as pd
from LinePlot import LinePlot
import AttrConf
def register_forwarding_arg(arg_name):
"""Allows the user to register args to
be forwarded to matplotlib
"""
if arg_name not in AttrConf.ARGS_TO_FORWARD:
AttrConf.ARGS_TO_FORWARD.append(arg_name)
def unregister_forwarding_arg(arg_name):
"""Unregisters arg_name from being passed to
plotter matplotlib calls
"""
try:
AttrConf.ARGS_TO_FORWARD.remove(arg_name)
except ValueError:
pass
| Enable user specified arg forwarding to matplotlib | plotter: Enable user specified arg forwarding to matplotlib
This change allows the user to register args for forwarding to
matplotlib and also unregister the same.
Change-Id: If53dab43dd4a2f530b3d1faf35582206ac925740
Signed-off-by: Kapileshwar Singh <[email protected]>
| Python | apache-2.0 | JaviMerino/trappy,joelagnel/trappy,bjackman/trappy,derkling/trappy,ARM-software/trappy,sinkap/trappy,JaviMerino/trappy,joelagnel/trappy,ARM-software/trappy,derkling/trappy,bjackman/trappy,sinkap/trappy,ARM-software/trappy,ARM-software/trappy,bjackman/trappy,sinkap/trappy,joelagnel/trappy,sinkap/trappy,JaviMerino/trappy,bjackman/trappy,derkling/trappy,joelagnel/trappy |
38cd50805e080f6613d7e1d5867a84952ec88580 | flask_resty/related.py | flask_resty/related.py | from .exceptions import ApiError
# -----------------------------------------------------------------------------
class Related(object):
def __init__(self, item_class=None, **kwargs):
self._item_class = item_class
self._view_classes = kwargs
def resolve_related(self, data):
for field_name, view_class in self._view_classes.items():
value = data.get(field_name, None)
if value is None:
# If this field were required or non-nullable, the deserializer
# would already have raised an exception.
continue
try:
resolved = self.resolve_field(value, view_class)
except ApiError as e:
pointer = '/data/{}'.format(field_name)
raise e.update({'source': {'pointer': pointer}})
data[field_name] = resolved
if self._item_class:
return self._item_class(**data)
return data
def resolve_field(self, value, view_class):
# marshmallow always uses lists here.
many = isinstance(value, list)
if many and not value:
# As a tiny optimization, there's no need to resolve an empty list.
return value
if isinstance(view_class, Related):
# This is not actually a view class.
resolver = view_class.resolve_related
else:
resolver = view_class().resolve_related_item
if many:
return [resolver(item) for item in value]
return resolver(value)
| from .exceptions import ApiError
# -----------------------------------------------------------------------------
class Related(object):
def __init__(self, item_class=None, **kwargs):
self._item_class = item_class
self._resolvers = kwargs
def resolve_related(self, data):
for field_name, resolver in self._resolvers.items():
value = data.get(field_name, None)
if value is None:
# If this field were required or non-nullable, the deserializer
# would already have raised an exception.
continue
try:
resolved = self.resolve_field(value, resolver)
except ApiError as e:
pointer = '/data/{}'.format(field_name)
raise e.update({'source': {'pointer': pointer}})
data[field_name] = resolved
if self._item_class:
return self._item_class(**data)
return data
def resolve_field(self, value, resolver):
# marshmallow always uses lists here.
many = isinstance(value, list)
if many and not value:
# As a tiny optimization, there's no need to resolve an empty list.
return value
if isinstance(resolver, Related):
resolve_item = resolver.resolve_related
else:
resolve_item = resolver().resolve_related_item
if many:
return [resolve_item(item) for item in value]
return resolve_item(value)
| Fix variable names in Related | Fix variable names in Related
| Python | mit | taion/flask-jsonapiview,4Catalyzer/flask-jsonapiview,4Catalyzer/flask-resty |
8cdbbb52ebe161b0b7fea342e8a3d197ec290ab1 | satnogsclient/settings.py | satnogsclient/settings.py | from os import environ
DEMODULATION_COMMAND = environ.get('SATNOGS_DEMODULATION_COMMAND', None)
ENCODING_COMMAND = environ.get('SATNOGS_ENCODING_COMMAND', None)
DECODING_COMMAND = environ.get('SATNOGS_DECODING_COMMAND', None)
OUTPUT_PATH = environ.get('SATNOGS_OUTPUT_PATH', None)
| from os import environ
DEMODULATION_COMMAND = environ.get('SATNOGS_DEMODULATION_COMMAND', 'rtl_fm')
ENCODING_COMMAND = environ.get('SATNOGS_ENCODING_COMMAND', 'oggenc')
DECODING_COMMAND = environ.get('SATNOGS_DECODING_COMMAND', 'multimon-ng')
OUTPUT_PATH = environ.get('SATNOGS_OUTPUT_PATH', '/tmp')
| Add default commands for encoding/decoding/demodulation. | Add default commands for encoding/decoding/demodulation.
| Python | agpl-3.0 | adamkalis/satnogs-client,adamkalis/satnogs-client,cshields/satnogs-client,cshields/satnogs-client |
3335c8c9ce8419d9d1d1034903687aa02983280d | tests/rules_tests/NoRuleSpecifiedTest.py | tests/rules_tests/NoRuleSpecifiedTest.py | #!/usr/bin/env python
"""
:Author Patrik Valkovic
:Created 23.06.2017 16:39
:Licence GNUv3
Part of grammpy
"""
from unittest import main, TestCase
from grammpy import Rule
class NoRuleSpecifiedTest(TestCase):
pass
if __name__ == '__main__':
main() | #!/usr/bin/env python
"""
:Author Patrik Valkovic
:Created 23.06.2017 16:39
:Licence GNUv3
Part of grammpy
"""
from unittest import main, TestCase
from grammpy import Rule
from grammpy.exceptions import RuleNotDefinedException
class NoRuleSpecifiedTest(TestCase):
def test_noRule(self):
class tmp(Rule):
x = 5
with self.assertRaises(RuleNotDefinedException):
x = tmp.rules
with self.assertRaises(RuleNotDefinedException):
x = tmp.rule
with self.assertRaises(RuleNotDefinedException):
x = tmp.left
with self.assertRaises(RuleNotDefinedException):
x = tmp.right
with self.assertRaises(RuleNotDefinedException):
x = tmp.fromSymbol
with self.assertRaises(RuleNotDefinedException):
x = tmp.toSymbol
if __name__ == '__main__':
main()
| Add tests when no rules is specified | Add tests when no rules is specified
| Python | mit | PatrikValkovic/grammpy |
0b3f6ae3b21cd51b99bcecf17d5ea1275c04abfd | statirator/project_template/project_name/settings.py | statirator/project_template/project_name/settings.py | # Generated by statirator
import os
# directories setup
ROOT_DIR = os.path.abspath(os.path.join(os.path.dirname(__file__), os.pardir))
SOURCE_DIR = os.path.join(ROOT_DIR, '{{ source }}')
BUILD_DIR = os.path.join(ROOT_DIR, '{{ build }}')
# languages setup
LANGUAGE_CODE = '{{default_lang}}'
_ = lambda s:s
LANGUAGES = ({% for code, name in languages %}
('{{code}}', _('{{ name }}')),
{% endfor %})
ROOT_URLCONF = '{{ project_name }}.urls'
TEMPLATE_DIRS = (
os.path.join(ROOT_DIR, 'templates'),
)
INSTALLED_APPS = (
'django.contrib.contenttypes',
'django.contrib.sites',
'django.contrib.staticfiles',
'django_medusa',
)
MEDUSA_RENDERER_CLASS = "django_medusa.renderers.DiskStaticSiteRenderer"
MEDUSA_MULTITHREAD = True
MEDUSA_DEPLOY_DIR = BUILD_DIR
| # Generated by statirator
import os
# directories setup
ROOT_DIR = os.path.abspath(os.path.join(os.path.dirname(__file__), os.pardir))
SOURCE_DIR = os.path.join(ROOT_DIR, '{{ source }}')
BUILD_DIR = os.path.join(ROOT_DIR, '{{ build }}')
# languages setup
LANGUAGE_CODE = '{{default_lang}}'
_ = lambda s:s
LANGUAGES = ({% for code, name in languages %}
('{{code}}', _('{{ name }}')),
{% endfor %})
ROOT_URLCONF = '{{ project_name }}.urls'
TEMPLATE_DIRS = (
os.path.join(ROOT_DIR, 'templates'),
)
INSTALLED_APPS = (
'django.contrib.contenttypes',
'django.contrib.sites',
'django.contrib.staticfiles',
'django_medusa',
'taggit',
'statirator.blog',
)
MEDUSA_RENDERER_CLASS = "django_medusa.renderers.DiskStaticSiteRenderer"
MEDUSA_MULTITHREAD = True
MEDUSA_DEPLOY_DIR = BUILD_DIR
| Add taggit and statirator.blog to INSTALLED_APPS | Add taggit and statirator.blog to INSTALLED_APPS
| Python | mit | MeirKriheli/statirator,MeirKriheli/statirator,MeirKriheli/statirator |
9a8d552b9424c11d6a58fcf38abea6bc2d55b1f0 | tests/install_tests.py | tests/install_tests.py | import sys
import logging
import subprocess
sys.path.insert(0, '..')
from unittest2 import TestCase
class MockLoggingHandler(logging.Handler):
"""Mock logging handler to check for expected logs."""
def __init__(self, *args, **kwargs):
self.reset()
logging.Handler.__init__(self, *args, **kwargs)
def emit(self, record):
self.messages[record.levelname.lower()].append(record.getMessage())
def reset(self):
self.messages = {
'debug': [],
'info': [],
'warning': [],
'error': [],
'critical': [],
}
class TestNode(TestCase):
def test_is_installed(self):
proc = subprocess.Popen(['node', '--version'], stdout=subprocess.PIPE)
output = proc.stdout.read()
self.assertEquals(output.strip(), 'v0.8.11')
| import sys
import logging
import subprocess
sys.path.insert(0, '..')
from unittest2 import TestCase
class MockLoggingHandler(logging.Handler):
"""Mock logging handler to check for expected logs."""
def __init__(self, *args, **kwargs):
self.reset()
logging.Handler.__init__(self, *args, **kwargs)
def emit(self, record):
self.messages[record.levelname.lower()].append(record.getMessage())
def reset(self):
self.messages = {
'debug': [],
'info': [],
'warning': [],
'error': [],
'critical': [],
}
class TestNode(TestCase):
def test_is_installed(self):
proc = subprocess.Popen(['node', '--version'], stdout=subprocess.PIPE)
output = proc.stdout.read()
self.assertEquals(output.strip(), 'v0.10.26')
| Update test to reflect new default version for node | Update test to reflect new default version for node
| Python | bsd-3-clause | elbaschid/virtual-node |
c87f47a3cf4bfd921f7953a9bb14fbc4675ea0e6 | examples/delete_all_policies.py | examples/delete_all_policies.py | #!/usr/bin/env python
#
# Delete all secure policies.
#
import os
import sys
import json
sys.path.insert(0, os.path.join(os.path.dirname(os.path.realpath(sys.argv[0])), '..'))
from sdcclient import SdSecureClient
def usage():
print('usage: %s <sysdig-token>' % sys.argv[0])
print('You can find your token at https://secure.sysdig.com/#/settings/user')
sys.exit(1)
#
# Parse arguments
#
if len(sys.argv) != 2:
usage()
sdc_token = sys.argv[1]
#
# Instantiate the SDC client
#
sdclient = SdSecureClient(sdc_token, 'https://secure.sysdig.com')
# Get a list of policyIds
ok, res = sdclient.list_policies()
policies = []
if not ok:
print(res)
sys.exit(1)
else:
policies = res[1]['policies']
for policy in policies:
print("deleting policy: " + str(policy['id']))
ok, res = sdclient.delete_policy_id(policy['id'])
if not ok:
print(res)
sys.exit(1)
| #!/usr/bin/env python
#
# Delete all secure policies.
#
import os
import sys
import json
sys.path.insert(0, os.path.join(os.path.dirname(os.path.realpath(sys.argv[0])), '..'))
from sdcclient import SdSecureClient
def usage():
print('usage: %s <sysdig-token>' % sys.argv[0])
print('You can find your token at https://secure.sysdig.com/#/settings/user')
sys.exit(1)
#
# Parse arguments
#
if len(sys.argv) != 2:
usage()
sdc_token = sys.argv[1]
#
# Instantiate the SDC client
#
sdclient = SdSecureClient(sdc_token, 'https://secure.sysdig.com')
# Get a list of policyIds
ok, res = sdclient.list_policies()
policies = []
if not ok:
print(res)
sys.exit(1)
else:
policies = res['policies']
for policy in policies:
print("deleting policy: " + str(policy['id']))
ok, res = sdclient.delete_policy_id(policy['id'])
if not ok:
print(res)
sys.exit(1)
| Fix legacy use of action result | Fix legacy use of action result
| Python | mit | draios/python-sdc-client,draios/python-sdc-client |
91a7e4ba30c2c455c58b7069015680b7af511cc4 | tests/test_get_joke.py | tests/test_get_joke.py |
def test_get_joke():
from pyjokes import get_joke
for i in range(10):
assert get_joke()
languages = ['eng', 'de', 'spa']
categories = ['neutral', 'explicit', 'all']
for lang in languages:
for cat in categories:
for i in range(10):
assert get_joke(cat, lang) |
import pytest
from pyjokes import get_joke
from pyjokes.pyjokes import LanguageNotFoundError, CategoryNotFoundError
def test_get_joke():
assert get_joke()
languages = ['en', 'de', 'es']
categories = ['neutral', 'explicit', 'all']
for lang in languages:
assert get_joke(language=lang)
for cat in categories:
assert get_joke(category=cat)
def test_get_joke_raises():
assert pytest.raises(LanguageNotFoundError, get_joke, language='eu')
assert pytest.raises(LanguageNotFoundError, get_joke, language='tr')
assert pytest.raises(CategoryNotFoundError, get_joke, category='123') | Simplify get_joke test, add raise checks | Simplify get_joke test, add raise checks
| Python | bsd-3-clause | borjaayerdi/pyjokes,trojjer/pyjokes,martinohanlon/pyjokes,bennuttall/pyjokes,ElectronicsGeek/pyjokes,pyjokes/pyjokes,gmarkall/pyjokes |
67d067fe499ba2ec78d34083640a4bfe9835d62b | tests/test_sequence.py | tests/test_sequence.py | from unittest import TestCase
from prudent.sequence import Sequence
class SequenceTest(TestCase):
def setUp(self):
self.seq = Sequence([1, 2, 3])
def test_getitem(self):
assert self.seq[0] == 1
self.seq[2]
assert self.seq[2] == 3
def test_len(self):
assert len(self.seq) == 0
self.seq[2]
assert len(self.seq) == 3
def test_iter(self):
for _ in range(2):
assert list(self.seq) == [1, 2, 3]
| from unittest import TestCase
from prudent.sequence import Sequence
class SequenceTest(TestCase):
def setUp(self):
self.seq = Sequence([1, 2, 3])
def test_getitem(self):
assert self.seq[0] == 1
assert self.seq[2] == 3
def test_getitem_raises_indexerror(self):
self.assertRaises(IndexError, lambda: self.seq[3])
def test_len_returns_current_size(self):
assert len(self.seq) == 0
self.seq[2]
assert len(self.seq) == 3
def test_iter_preserves_elems(self):
for _ in range(2):
assert list(self.seq) == [1, 2, 3]
| Test that IndexError is raised when appropriate | Test that IndexError is raised when appropriate
| Python | mit | eugene-eeo/prudent |
987e3b3387124c9eee7b0d69647fe2eeba40b70d | snippets/list_holidays.py | snippets/list_holidays.py | #!/usr/bin/env python
import pandas as pd
from datetime import date
import holidays
def sanitize_holiday_name(name):
new_name = [c for c in name if c.isalpha() or c.isdigit() or c == ' ']
new_name = "".join(new_name).lower().replace(" ", "_")
return new_name
def process_holidays(df):
# Create a date object
user_date = date(df['year'], df['month'], df['day'])
# Get US holidays for this year
holidays_dates = holidays.US(years=df['year'])
for holiday_date, name in holidays_dates.iteritems():
# if 'observed' in name:
# pass
# Compute difference in days
days = (holiday_date - user_date).days
# Clean holiday name
name = sanitize_holiday_name(name)
# Add the computed days to holiday into our DataFrame
df['days_to_' + name] = days
return df
def main():
df = pd.DataFrame()
df['year'] = pd.Series(range(2010, 2015))
df['day'] = pd.Series(range(11, 27, 3))
df['month'] = pd.Series(range(2, 12, 2))
print df.apply(process_holidays, axis=1)
if __name__ == '__main__':
main()
| #!/usr/bin/env python
import pandas as pd
from datetime import date
import holidays
def sanitize_holiday_name(name):
new_name = [c for c in name if c.isalpha() or c.isdigit() or c == ' ']
new_name = "".join(new_name).lower().replace(" ", "_")
return new_name
def process_holidays(df):
# Create a date object
user_date = date(
df['year_account_created'],
df['month_account_created'],
df['day_account_created']
)
# Get US holidays for this year
holidays_dates = holidays.US(years=df['year'])
for holiday_date, name in holidays_dates.iteritems():
# if 'observed' in name:
# pass
# Compute difference in days
days = (holiday_date - user_date).days
# Clean holiday name
name = sanitize_holiday_name(name)
# Add the computed days to holiday into our DataFrame
df['days_to_' + name] = days
return df
def main():
path = '../datasets/processed/'
train_users = pd.read_csv(path + 'processed_train_users.csv')
train_users = train_users.head(500)
test_users = pd.read_csv(path + 'processed_train_users.csv')
test_users = test_users.head(500)
train_users = train_users.apply(process_holidays, axis=1)
print train_users.columns
if __name__ == '__main__':
main()
| Update with real train users | Update with real train users
| Python | mit | davidgasquez/kaggle-airbnb |
4a8aaf3c9e1da5fd10580cc5a3859d801f2c9553 | django_docker/django_docker/urls.py | django_docker/django_docker/urls.py | from django.conf.urls import patterns, include, url
from django.contrib import admin
admin.autodiscover()
urlpatterns = patterns('',
url(r'^admin/', include(admin.site.urls)),
url(r'^$', 'hello_world.views.hello_world', name='hello_world'),
)
| from django.conf.urls import include, url
from hello_world import views as hello_world_views
from django.contrib import admin
admin.autodiscover()
urlpatterns = [
url(r'^admin/', include(admin.site.urls)),
url(r'^$', hello_world_views.hello_world, name='hello_world'),
]
| Update URLs to be compatible with Django 1.10 | Update URLs to be compatible with Django 1.10
| Python | mit | morninj/django-docker,morninj/django-docker,morninj/django-docker |
caeb76cbcb6cdd49138e41f57144573598b722ba | source/clique/__init__.py | source/clique/__init__.py | # :coding: utf-8
# :copyright: Copyright (c) 2013 Martin Pengelly-Phillips
# :license: See LICENSE.txt.
from ._version import __version__
| # :coding: utf-8
# :copyright: Copyright (c) 2013 Martin Pengelly-Phillips
# :license: See LICENSE.txt.
import re
from collections import defaultdict
from ._version import __version__
from .collection import Collection
from .error import CollectionError
#: Pattern for matching an index with optional padding.
DIGITS_PATTERN = '(?P<index>(?P<padding>0*)\d+)'
_DIGITS_REGEX = re.compile(DIGITS_PATTERN)
#: Common patterns that can be passed to :py:func:`~clique.assemble`.
PATTERNS = {
'frames': '\.{0}\.\D+\d?$'.format(DIGITS_PATTERN),
'versions': 'v{0}'.format(DIGITS_PATTERN)
}
def assemble(iterable, patterns=None, minimum_items=2):
'''Assemble items in *iterable* into discreet collections.
*patterns* may be specified as a list of regular expressions to limit
the returned collection possibilities. Use this when interested in
collections that only match specific patterns. Each pattern must contain
the expression from :py:data:`DIGITS_PATTERN` exactly once.
A selection of common expressions are available in :py:data:`PATTERNS`.
.. note::
If a pattern is supplied as a string it will be automatically compiled
to a regular expression for convenience.
When *patterns* is not specified, collections are formed by examining all
possible groupings of the items in *iterable* based around common numerical
components.
*minimum_items* dictates the minimum number of items a collection must have
in order to be included in the result. The default is 2, filtering out
single item collections.
Return list of assembled :py:class:`~clique.collection.Collection`
instances.
'''
collection_map = defaultdict(set)
collections = []
# Compile patterns.
compiled_patterns = []
if patterns is not None:
if not patterns:
return collections
for pattern in patterns:
if isinstance(pattern, basestring):
compiled_patterns.append(re.compile(pattern))
else:
compiled_patterns.append(pattern)
else:
compiled_patterns.append(_DIGITS_REGEX)
# Process iterable.
for item in iterable:
for pattern in compiled_patterns:
for match in pattern.finditer(item):
index = match.group('index')
head = item[:match.start('index')]
tail = item[match.end('index'):]
padding = match.group('padding')
if padding:
padding = len(index)
else:
padding = 0
key = (head, tail, padding)
collection_map[key].add(int(index))
# Form collections, filtering out those that do not have at least
# as many indexes as minimum_items
for (head, tail, padding), indexes in collection_map.items():
if len(indexes) >= minimum_items:
collections.append(
Collection(head, tail, padding, indexes)
)
return collections
| Add top level function to help assemble collections from arbitrary items. | Add top level function to help assemble collections from arbitrary items.
| Python | apache-2.0 | 4degrees/clique |
21b97ceea5b2e667940ddd45682313261eba845b | discode_server/notify.py | discode_server/notify.py | import json
from discode_server import db
from discode_server import fragments
connected = set()
notified = set()
async def feed(request, ws):
global connected
connected.add(ws)
print("Open WebSockets: ", len(connected))
try:
while True:
if not ws.open:
return
async with request.app.config.DB.acquire() as conn:
await conn.execute(f"LISTEN channel")
msg = await conn.connection.notifies.get()
if not ws.open:
return
fingerprint = ws.remote_address, msg.payload
if fingerprint in notified:
continue
notified.add(fingerprint)
paste_id, lineno, comment_id = msg.payload.split(',')
paste = await db.get_paste(conn, int(paste_id))
html = fragments.comment_row(lineno,
paste.comments[int(lineno)])
data = json.dumps({
"html": html,
"lineno": lineno,
"paste_id": paste.id,
"comment_id": comment_id,
})
await ws.send(data)
finally:
connected.remove(ws)
print("Open WebSockets: ", len(connected))
| import asyncio
import json
from discode_server import db
from discode_server import fragments
connected = set()
notified = set()
async def feed(request, ws):
global connected
connected.add(ws)
print("Open WebSockets: ", len(connected))
try:
while True:
if not ws.open:
return
async with request.app.config.DB.acquire() as conn:
await conn.execute(f"LISTEN channel")
try:
msg = await asyncio.wait_for(
conn.connection.notifies.get(), 1)
except asyncio.TimeoutError:
continue
if not ws.open:
return
fingerprint = ws.remote_address, msg.payload
if fingerprint in notified:
continue
notified.add(fingerprint)
paste_id, lineno, comment_id = msg.payload.split(',')
paste = await db.get_paste(conn, int(paste_id))
html = fragments.comment_row(lineno,
paste.comments[int(lineno)])
data = json.dumps({
"html": html,
"lineno": lineno,
"paste_id": paste.id,
"comment_id": comment_id,
})
await ws.send(data)
finally:
connected.remove(ws)
print("Open WebSockets: ", len(connected))
| Add the asyncio wait_for back in | Add the asyncio wait_for back in
| Python | bsd-2-clause | d0ugal/discode-server,d0ugal/discode-server,d0ugal/discode-server |
c3d03629734abfead5ae1eae83d1b6dcec792b45 | iconizer/django_in_iconizer/django_server.py | iconizer/django_in_iconizer/django_server.py | import os
class DjangoServer(object):
default_django_manage_script = "manage.py"
def __init__(self, django_manage_script=None):
super(DjangoServer, self).__init__()
if django_manage_script is None:
self.django_manage_script = self.default_django_manage_script
else:
self.django_manage_script = django_manage_script
def get_task_descriptor(self, task_name, param_list=[]):
task_name_and_param = [self.django_manage_script, task_name]
task_name_and_param.extend(param_list)
return {task_name: task_name_and_param}
# noinspection PyMethodMayBeStatic
def get_cmd_str(self, cmd_name, param_list=[]):
return "python %s %s" % (self.django_manage_script, cmd_name)
def execute_cmd(self, django_cmd):
os.system(self.get_cmd_str(django_cmd))
| import os
class DjangoServer(object):
default_django_manage_script = "manage.py"
def __init__(self, django_manage_script=None):
super(DjangoServer, self).__init__()
if django_manage_script is None:
self.django_manage_script = self.default_django_manage_script
else:
self.django_manage_script = django_manage_script
self.django_manage_script = os.environ.get("MANAGE_PY", self.django_manage_script)
def get_task_descriptor(self, task_name, param_list=[]):
task_name_and_param = [self.django_manage_script, task_name]
task_name_and_param.extend(param_list)
return {task_name: task_name_and_param}
# noinspection PyMethodMayBeStatic
def get_cmd_str(self, cmd_name, param_list=[]):
return "python %s %s" % (self.django_manage_script, cmd_name)
def execute_cmd(self, django_cmd):
os.system(self.get_cmd_str(django_cmd))
| Enable specify command line processor. | Enable specify command line processor.
| Python | bsd-3-clause | weijia/iconizer |
6b57d186173234dbedcfa4758eeb1cb00a90eba7 | src/adhocracy/migration/versions/053_add_votedetail_table.py | src/adhocracy/migration/versions/053_add_votedetail_table.py | from sqlalchemy import MetaData, Column, Table
from sqlalchemy import Integer, ForeignKey
metadata = MetaData()
votedetail_table = Table(
'votedetail', metadata,
Column('instance_id', Integer,
ForeignKey('instance.id', ondelete='CASCADE')),
Column('badge_id', Integer,
ForeignKey('badge.id'), ondelete='CASCADE')),
)
def upgrade(migrate_engine):
meta.bind = migrate_engine
votedetail_table.create()
def downgrade(migrate_engine):
raise NotImplementedError()
| from sqlalchemy import MetaData, Column, Table
from sqlalchemy import Integer, ForeignKey
metadata = MetaData()
votedetail_table = Table(
'votedetail', metadata,
Column('instance_id', Integer,
ForeignKey('instance.id', ondelete='CASCADE')),
Column('badge_id', Integer,
ForeignKey('badge.id', ondelete='CASCADE')),
)
def upgrade(migrate_engine):
meta.bind = migrate_engine
votedetail_table.create()
def downgrade(migrate_engine):
raise NotImplementedError()
| Fix typo in migration script | Fix typo in migration script
| Python | agpl-3.0 | liqd/adhocracy,DanielNeugebauer/adhocracy,alkadis/vcv,alkadis/vcv,liqd/adhocracy,phihag/adhocracy,liqd/adhocracy,DanielNeugebauer/adhocracy,SysTheron/adhocracy,phihag/adhocracy,phihag/adhocracy,alkadis/vcv,alkadis/vcv,DanielNeugebauer/adhocracy,DanielNeugebauer/adhocracy,SysTheron/adhocracy,alkadis/vcv,DanielNeugebauer/adhocracy,liqd/adhocracy,phihag/adhocracy,phihag/adhocracy,SysTheron/adhocracy |
fce890ce9046dd055f219ac880ae9734f334f534 | greenlight/harness/arguments.py | greenlight/harness/arguments.py | # This Source Code Form is subject to the terms of the Mozilla Public
# License, v. 2.0. If a copy of the MPL was not distributed with this
# file, You can obtain one at http://mozilla.org/MPL/2.0/.
from marionette import BaseMarionetteOptions
from greenlight import tests
class ReleaseTestParser(BaseMarionetteOptions):
def parse_args(self, *args, **kwargs):
options, test_files = BaseMarionetteOptions.parse_args(self,
*args, **kwargs)
if not test_files:
test_files = [tests.manifest]
return (options, test_files)
| # This Source Code Form is subject to the terms of the Mozilla Public
# License, v. 2.0. If a copy of the MPL was not distributed with this
# file, You can obtain one at http://mozilla.org/MPL/2.0/.
from marionette import BaseMarionetteOptions
from greenlight import tests
class ReleaseTestParser(BaseMarionetteOptions):
def parse_args(self, *args, **kwargs):
options, test_files = BaseMarionetteOptions.parse_args(self,
*args, **kwargs)
if not any([(k.startswith('log_') and v is not None and '-' in v)
for (k, v) in vars(options).items()]):
options.log_mach = '-'
if not test_files:
test_files = [tests.manifest]
return (options, test_files)
| Make colored terminal output the default log formatter. | Make colored terminal output the default log formatter.
| Python | mpl-2.0 | myrdd/firefox-ui-tests,gbrmachado/firefox-ui-tests,utvar/firefox-ui-tests,chmanchester/firefox-ui-tests,armenzg/firefox-ui-tests,whimboo/firefox-ui-tests,myrdd/firefox-ui-tests,gbrmachado/firefox-ui-tests,sr-murthy/firefox-ui-tests,armenzg/firefox-ui-tests,Motwani/firefox-ui-tests,galgeek/firefox-ui-tests,armenzg/firefox-ui-tests,sr-murthy/firefox-ui-tests,sr-murthy/firefox-ui-tests,galgeek/firefox-ui-tests,myrdd/firefox-ui-tests,utvar/firefox-ui-tests,galgeek/firefox-ui-tests,whimboo/firefox-ui-tests,whimboo/firefox-ui-tests,Motwani/firefox-ui-tests,chmanchester/firefox-ui-tests |
f684692a013e80d0c4d07b1c0eba204bd94d2314 | config.py | config.py |
# Enables detailed tracebacks and an interactive Python console on errors.
# Never use in production!
#DEBUG = True
# Makes the server more performant at sending static files when the
# server is being proxied by a server that supports X-Sendfile.
#USE_X_SENDFILE = True
# Address to listen for clients on
HOST = "0.0.0.0"
# Port to listen on
PORT = 8000
# File to store the JSON server list data in.
FILENAME = "list.json"
# Ammount of time, is seconds, after which servers are removed from the list
# if they haven't updated their listings. Note: By default Minetest servers
# only announce once every 5 minutes, so this should be more than 300.
PURGE_TIME = 350
# List of banned IP addresses.
BANLIST = []
# Creates server entries if a server sends an 'update' and there is no entry yet
# This should only be used to populate the server list after list.json was deleted.
ALLOW_UPDATE_WITHOUT_OLD = False
|
# Enables detailed tracebacks and an interactive Python console on errors.
# Never use in production!
#DEBUG = True
# Makes the server more performant at sending static files when the
# server is being proxied by a server that supports X-Sendfile.
#USE_X_SENDFILE = True
# Address to listen for clients on
HOST = "0.0.0.0"
# Port to listen on
PORT = 8000
# File to store the JSON server list data in.
FILENAME = "list.json"
# Ammount of time, is seconds, after which servers are removed from the list
# if they haven't updated their listings. Note: By default Minetest servers
# only announce once every 5 minutes, so this should be more than 300.
PURGE_TIME = 350
# List of banned IP addresses.
BANLIST = []
# Creates server entries if a server sends an 'update' and there is no entry yet
# This should only be used to populate the server list after list.json was deleted.
# This WILL cause problems such as mapgen, mods and privilege information missing from the list
ALLOW_UPDATE_WITHOUT_OLD = False
| Add note that 'ALLOW_UPDATE_WITHOUT_OLD' causes problems | Add note that 'ALLOW_UPDATE_WITHOUT_OLD' causes problems
| Python | lgpl-2.1 | minetest/master-server,minetest/master-server,minetest/master-server |
63e09b77e3b00a7483249417020fb093f98773a9 | infrastructure/tests/helpers.py | infrastructure/tests/helpers.py | from datetime import datetime
from django.contrib.staticfiles.testing import LiveServerTestCase
from selenium import webdriver
from selenium.webdriver import DesiredCapabilities
from selenium.webdriver.common.by import By
from selenium.webdriver.support import expected_conditions as EC
from selenium.webdriver.support.wait import WebDriverWait
class BaseSeleniumTestCase(LiveServerTestCase):
def setUp(self):
super(BaseSeleniumTestCase, self).setUp()
chrome_options = webdriver.ChromeOptions()
chrome_options.add_argument("headless")
chrome_options.add_argument("--no-sandbox")
d = DesiredCapabilities.CHROME
d["loggingPrefs"] = {"browser": "ALL"}
self.selenium = webdriver.Chrome(
chrome_options=chrome_options, desired_capabilities=d
)
self.selenium.implicitly_wait(10)
self.wait = WebDriverWait(self.selenium, 5)
self.addCleanup(self.selenium.quit)
def wait_until_text_in(self, selector, text):
self.wait.until(
EC.text_to_be_present_in_element((By.CSS_SELECTOR, selector), text)
)
| from datetime import datetime
from django.contrib.staticfiles.testing import LiveServerTestCase
from selenium import webdriver
from selenium.webdriver import DesiredCapabilities
from selenium.webdriver.common.by import By
from selenium.webdriver.support import expected_conditions as EC
from selenium.webdriver.support.wait import WebDriverWait
import logging
logger = logging.Logger(__name__)
class BaseSeleniumTestCase(LiveServerTestCase):
def setUp(self):
super(BaseSeleniumTestCase, self).setUp()
chrome_options = webdriver.ChromeOptions()
chrome_options.add_argument("headless")
chrome_options.add_argument("--no-sandbox")
d = DesiredCapabilities.CHROME
d["loggingPrefs"] = {"browser": "ALL"}
self.selenium = webdriver.Chrome(
chrome_options=chrome_options, desired_capabilities=d
)
self.selenium.implicitly_wait(10)
self.wait = WebDriverWait(self.selenium, 5)
self.addCleanup(self.selenium.quit)
def wait_until_text_in(self, selector, text):
if self.wait.until(EC.text_to_be_present_in_element((By.CSS_SELECTOR, selector), text)):
pass
else:
text_content = self.selenium.find_elements_by_css_selector(selector)[0].text
logger.error("Element contents: %s" % text_content)
| Add logging to selector testing | Add logging to selector testing
| Python | mit | Code4SA/municipal-data,Code4SA/municipal-data,Code4SA/municipal-data,Code4SA/municipal-data |
d07b6483d110eb4c51f7c631f888d1bf30eacc1c | destroyer-runner.py | destroyer-runner.py | #!/usr/bin/python
"""destroyer-runner.py - Run the main application"""
import sys
import subprocess
if __name__ == '__main__':
subprocess.call(['python', './destroyer/destroyer.py'] + [str(arg) for arg in sys.argv[1:]])
| #!/usr/bin/python
"""destroyer-runner.py - Run the main application"""
from destroyer.destroyer import main
if __name__ == '__main__':
main()
| Update with working code to run destroyer | Update with working code to run destroyer
| Python | mit | jaredmichaelsmith/destroyer |
261c294690427b57d111c777cdc4d13b9c84f9d2 | cloudkittydashboard/dashboards/admin/modules/forms.py | cloudkittydashboard/dashboards/admin/modules/forms.py | # Copyright 2017 Objectif Libre
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
import logging
from django.utils.translation import ugettext_lazy as _
from horizon import forms
from cloudkittydashboard.api import cloudkitty as api
LOG = logging.getLogger(__name__)
class EditPriorityForm(forms.SelfHandlingForm):
priority = forms.IntegerField(label=_("Priority"), required=True)
def handle(self, request, data):
ck_client = api.cloudkittyclient(request)
return ck_client.modules.update(
module_id=self.initial["module_id"],
priority=data["priority"]
)
| # Copyright 2017 Objectif Libre
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from django.utils.translation import ugettext_lazy as _
from horizon import forms
from cloudkittydashboard.api import cloudkitty as api
class EditPriorityForm(forms.SelfHandlingForm):
priority = forms.IntegerField(label=_("Priority"), required=True)
def handle(self, request, data):
ck_client = api.cloudkittyclient(request)
return ck_client.modules.update(
module_id=self.initial["module_id"],
priority=data["priority"]
)
| Delete the unused LOG code | Delete the unused LOG code
Change-Id: Ief253cdd226f8c2688429b0ff00785151a99759b
| Python | apache-2.0 | stackforge/cloudkitty-dashboard,openstack/cloudkitty-dashboard,openstack/cloudkitty-dashboard,openstack/cloudkitty-dashboard,stackforge/cloudkitty-dashboard,stackforge/cloudkitty-dashboard |
316323387c508c88595f205182ea2436c271621d | src/highdicom/uid.py | src/highdicom/uid.py | import logging
import pydicom
logger = logging.getLogger(__name__)
class UID(pydicom.uid.UID):
"""Unique DICOM identifier with a highdicom-specific UID prefix."""
def __new__(cls: type) -> str:
prefix = '1.2.826.0.1.3680043.10.511.3.'
identifier = pydicom.uid.generate_uid(prefix=prefix)
return super().__new__(cls, identifier)
| import logging
from typing import Type, TypeVar
import pydicom
logger = logging.getLogger(__name__)
T = TypeVar('T', bound='UID')
class UID(pydicom.uid.UID):
"""Unique DICOM identifier with a highdicom-specific UID prefix."""
def __new__(cls: Type[T]) -> T:
prefix = '1.2.826.0.1.3680043.10.511.3.'
identifier = pydicom.uid.generate_uid(prefix=prefix)
return super().__new__(cls, identifier)
| Fix typing for UID class | Fix typing for UID class
| Python | mit | MGHComputationalPathology/highdicom |
ee53ec51d98802bf0bc55e70c39cc0918f2bb274 | icekit/plugins/blog_post/content_plugins.py | icekit/plugins/blog_post/content_plugins.py | """
Definition of the plugin.
"""
from django.apps import apps
from django.conf import settings
from django.db.models.loading import get_model
from django.utils.translation import ugettext_lazy as _
from fluent_contents.extensions import ContentPlugin, plugin_pool
default_blog_model = 'blog_tools.BlogPost'
icekit_blog_model = getattr(settings, 'ICEKIT_BLOG_MODEL', default_blog_model)
BLOG_MODEL = apps.get_model(*icekit_blog_model.rsplit('.', 1))
if icekit_blog_model != default_blog_model:
@plugin_pool.register
class BlogPostPlugin(ContentPlugin):
model = get_model(getattr(settings, 'ICEKIT_BLOG_CONTENT_ITEM', 'blog_post.PostItem'))
category = _('Blog')
render_template = 'icekit/plugins/post/default.html'
raw_id_fields = ['post', ]
| """
Definition of the plugin.
"""
from django.apps import apps
from django.conf import settings
from django.utils.translation import ugettext_lazy as _
from fluent_contents.extensions import ContentPlugin, plugin_pool
default_blog_model = 'blog_tools.BlogPost'
icekit_blog_model = getattr(settings, 'ICEKIT_BLOG_MODEL', default_blog_model)
BLOG_MODEL = apps.get_model(*icekit_blog_model.rsplit('.', 1))
if icekit_blog_model != default_blog_model:
@plugin_pool.register
class BlogPostPlugin(ContentPlugin):
model = apps.get_model(getattr(settings, 'ICEKIT_BLOG_CONTENT_ITEM', 'blog_post.BlogPostItem'))
category = _('Blog')
render_template = 'icekit/plugins/post/default.html'
raw_id_fields = ['post', ]
| Update Blog model and content item matching | Update Blog model and content item matching
| Python | mit | ic-labs/django-icekit,ic-labs/django-icekit,ic-labs/django-icekit,ic-labs/django-icekit |
c7143cd725fc829c33ad9f9150e5975deb7be93a | irctest/optional_extensions.py | irctest/optional_extensions.py | import unittest
import operator
import itertools
class OptionalExtensionNotSupported(unittest.SkipTest):
def __str__(self):
return 'Unsupported extension: {}'.format(self.args[0])
class OptionalSaslMechanismNotSupported(unittest.SkipTest):
def __str__(self):
return 'Unsupported SASL mechanism: {}'.format(self.args[0])
class OptionalityReportingTextTestRunner(unittest.TextTestRunner):
def run(self, test):
result = super().run(test)
if result.skipped:
print()
print('Some tests were skipped because the following optional'
'specifications/mechanisms are not supported:')
msg_to_tests = itertools.groupby(result.skipped,
key=operator.itemgetter(1))
for (msg, tests) in msg_to_tests:
print('\t{} ({} test(s))'.format(msg, sum(1 for x in tests)))
return result
| import unittest
import operator
import itertools
class NotImplementedByController(unittest.SkipTest):
def __str__(self):
return 'Not implemented by controller: {}'.format(self.args[0])
class OptionalExtensionNotSupported(unittest.SkipTest):
def __str__(self):
return 'Unsupported extension: {}'.format(self.args[0])
class OptionalSaslMechanismNotSupported(unittest.SkipTest):
def __str__(self):
return 'Unsupported SASL mechanism: {}'.format(self.args[0])
class OptionalityReportingTextTestRunner(unittest.TextTestRunner):
def run(self, test):
result = super().run(test)
if result.skipped:
print()
print('Some tests were skipped because the following optional'
'specifications/mechanisms are not supported:')
msg_to_tests = itertools.groupby(result.skipped,
key=operator.itemgetter(1))
for (msg, tests) in sorted(msg_to_tests):
print('\t{} ({} test(s))'.format(msg, sum(1 for x in tests)))
return result
| Add an exception to tell a controller does not implement something. | Add an exception to tell a controller does not implement something.
| Python | mit | ProgVal/irctest |
a0ce4d366681f2f62f232f4f952ac18df07667d4 | ideascube/conf/idb_fra_cultura.py | ideascube/conf/idb_fra_cultura.py | # -*- coding: utf-8 -*-
"""Ideaxbox Cultura, France"""
from .idb import * # noqa
from django.utils.translation import ugettext_lazy as _
IDEASCUBE_NAME = u"Cultura"
IDEASCUBE_PLACE_NAME = _("city")
COUNTRIES_FIRST = ['FR']
TIME_ZONE = None
LANGUAGE_CODE = 'fr'
LOAN_DURATION = 14
MONITORING_ENTRY_EXPORT_FIELDS = ['serial', 'user_id', 'birth_year', 'gender']
USER_FORM_FIELDS = (
(_('Personal informations'), ['serial', 'short_name', 'full_name', 'latin_name', 'birth_year', 'gender']), # noqa
)
HOME_CARDS = HOME_CARDS + [
{
'id': 'cpassorcier',
},
{
'id': 'wikisource',
},
{
'id': 'software',
},
{
'id': 'ted',
},
{
'id': 'ubuntudoc',
},
]
| # -*- coding: utf-8 -*-
"""Ideaxbox Cultura, France"""
from .idb import * # noqa
from django.utils.translation import ugettext_lazy as _
IDEASCUBE_NAME = u"Cultura"
IDEASCUBE_PLACE_NAME = _("city")
COUNTRIES_FIRST = ['FR']
TIME_ZONE = None
LANGUAGE_CODE = 'fr'
LOAN_DURATION = 14
MONITORING_ENTRY_EXPORT_FIELDS = ['serial', 'user_id', 'birth_year', 'gender']
USER_FORM_FIELDS = (
(_('Personal informations'), ['serial', 'short_name', 'full_name', 'latin_name', 'birth_year', 'gender']), # noqa
)
HOME_CARDS = HOME_CARDS + [
{
'id': 'cpassorcier',
},
{
'id': 'wikisource',
},
{
'id': 'ted',
},
{
'id': 'ubuntudoc',
},
]
| Remove "software" card from Cultura conf | Remove "software" card from Cultura conf
| Python | agpl-3.0 | ideascube/ideascube,ideascube/ideascube,ideascube/ideascube,ideascube/ideascube |
a977908efcc176e1e5adbd82843033805953c6cb | tools/reago/format_reago_input_files.py | tools/reago/format_reago_input_files.py | #!/usr/bin/python
# -*- coding: utf-8 -*-
import sys
import os
import argparse
import re
reago_dir = '/tools/rna_manipulation/reago/reago/'
def add_read_pair_num(input_filepath, output_filepath, read_pair_num):
to_add = '.' + str(read_pair_num)
with open(input_filepath,'r') as input_file:
with open(output_filepath,'w') as output_file:
for line in input_file:
if line[0] == '>':
split_line = line.split()
seq_id = split_line[0]
if seq_id.rfind(to_add) != (len(seq_id)-len(to_add)):
split_line[0] = seq_id + to_add
output_file.write(' '.join(split_line) + '\n')
else:
output_file.write(line)
if __name__ == '__main__':
parser = argparse.ArgumentParser()
parser.add_argument('--r1_sequence_file', required=True)
parser.add_argument('--r2_sequence_file', required=True)
args = parser.parse_args()
add_read_pair_num(args.r1_input_sequence_file, args.r1_input_sequence_file, 1)
add_read_pair_num(args.r2_input_sequence_file, args.r2_input_sequence_file, 2) | #!/usr/bin/python
# -*- coding: utf-8 -*-
import sys
import os
import argparse
import re
reago_dir = '/tools/rna_manipulation/reago/reago/'
def add_read_pair_num(input_filepath, output_filepath, read_pair_num):
to_add = '.' + str(read_pair_num)
with open(input_filepath,'r') as input_file:
with open(output_filepath,'w') as output_file:
for line in input_file:
if line[0] == '>':
split_line = line.split()
seq_id = split_line[0]
if seq_id.rfind(to_add) != (len(seq_id)-len(to_add)):
split_line[0] = seq_id + to_add
output_file.write(' '.join(split_line) + '\n')
else:
output_file.write(line)
if __name__ == '__main__':
parser = argparse.ArgumentParser()
parser.add_argument('--r1_sequence_file', required=True)
parser.add_argument('--r2_sequence_file', required=True)
args = parser.parse_args()
add_read_pair_num(args.r1_sequence_file, args.r1_sequence_file, 1)
add_read_pair_num(args.r2_sequence_file, args.r2_sequence_file, 2) | Correct argument name in script to format reago input file | Correct argument name in script to format reago input file
| Python | apache-2.0 | ASaiM/galaxytools,ASaiM/galaxytools |
37c59a8535ac0d68a995af378fcb61f03070e018 | kombu_fernet/serializers/__init__.py | kombu_fernet/serializers/__init__.py | # -*- coding: utf-8 -*-
from __future__ import unicode_literals, absolute_import
import os
from cryptography.fernet import Fernet, MultiFernet
fernet = Fernet(os.environ['KOMBU_FERNET_KEY'])
fallback_fernet = None
try:
fallback_fernet = Fernet(os.environ['OLD_KOMBU_FERNET_KEY'])
except KeyError:
pass
else:
fernet = MultiFernet([fernet, fallback_fernet])
def fernet_encode(func):
def inner(message):
return fernet.encrypt(func(message))
return inner
def fernet_decode(func):
def inner(encoded_message):
if isinstance(encoded_message, unicode):
encoded_message = encoded_message.encode('utf-8')
message = fernet.decrypt(encoded_message)
return func(message)
return inner
| # -*- coding: utf-8 -*-
from __future__ import unicode_literals, absolute_import
import os
from cryptography.fernet import Fernet, MultiFernet
fernet = Fernet(os.environ['KOMBU_FERNET_KEY'])
fallback_fernet = None
try:
fallback_fernet = Fernet(os.environ['KOMBU_FERNET_KEY_PREVIOUS'])
except KeyError:
pass
else:
fernet = MultiFernet([fernet, fallback_fernet])
def fernet_encode(func):
def inner(message):
return fernet.encrypt(func(message))
return inner
def fernet_decode(func):
def inner(encoded_message):
if isinstance(encoded_message, unicode):
encoded_message = encoded_message.encode('utf-8')
message = fernet.decrypt(encoded_message)
return func(message)
return inner
| Make update previous key name | Make update previous key name
| Python | mit | heroku/kombu-fernet-serializers |
a83dd9bfce43a898cf71c9e66192ad589a02b6c8 | macroeco/compare/__init__.py | macroeco/compare/__init__.py | """
=================================
Compare (:mod:`macroeco.compare`)
=================================
This module contains functions that compare the goodness of fit of a
distribution/curve to data or the fit of two distributions/curves to each
other.
Comparison Functions
====================
.. autosummary::
:toctree: generated/
get_AIC
get_AICC
get_AIC_weights
get_nll
get_empirical_cdf
get_sum_of_squares
get_r_squared
get_chi_squared
get_lrt
bin_data
"""
from .compare import *
| """
=================================
Compare (:mod:`macroeco.compare`)
=================================
This module contains functions that compare the goodness of fit of a
distribution/curve to data or the fit of two distributions/curves to each
other.
Comparison Functions
====================
.. autosummary::
:toctree: generated/
AIC
AICC
AIC_weights
nll
empirical_cdf
sum_of_squares
r_squared
chi_squared
lrt
bin_data
"""
from .compare import *
| Remove get_ prefixes from compare docstring | Remove get_ prefixes from compare docstring
| Python | bsd-2-clause | jkitzes/macroeco |
a0972fe1fba507a1e6d17fb18846fb7a13c8cbbe | src/msd/msd_utils.py | src/msd/msd_utils.py | # media-service-demo
#
# Copyright (C) 2012 Intel Corporation. All rights reserved.
#
# This program is free software; you can redistribute it and/or modify it
# under the terms and conditions of the GNU Lesser General Public License,
# version 2.1, as published by the Free Software Foundation.
#
# This program is distributed in the hope it will be useful, but WITHOUT
# ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
# FITNESS FOR A PARTICULAR PURPOSE. See the GNU Lesser General Public License
# for more details.
#
# You should have received a copy of the GNU Lesser General Public License
# along with this program; if not, write to the Free Software Foundation, Inc.,
# 51 Franklin St - Fifth Floor, Boston, MA 02110-1301 USA.
#
# Mark Ryan <[email protected]>
#
import tempfile
import pygtk
pygtk.require('2.0')
import gtk
import urllib2
import os
def image_from_file(url):
tmpfile = tempfile.NamedTemporaryFile(delete=False)
tmpFileName = tmpfile.name
image = None
try:
with tmpfile:
message = urllib2.urlopen(url, None, 1)
tmpfile.write(message.read())
image = gtk.Image()
image.set_from_file(tmpfile.name)
finally:
os.unlink(tmpFileName)
return image
| # media-service-demo
#
# Copyright (C) 2012 Intel Corporation. All rights reserved.
#
# This program is free software; you can redistribute it and/or modify it
# under the terms and conditions of the GNU Lesser General Public License,
# version 2.1, as published by the Free Software Foundation.
#
# This program is distributed in the hope it will be useful, but WITHOUT
# ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
# FITNESS FOR A PARTICULAR PURPOSE. See the GNU Lesser General Public License
# for more details.
#
# You should have received a copy of the GNU Lesser General Public License
# along with this program; if not, write to the Free Software Foundation, Inc.,
# 51 Franklin St - Fifth Floor, Boston, MA 02110-1301 USA.
#
# Mark Ryan <[email protected]>
#
import tempfile
import pygtk
pygtk.require('2.0')
import gtk
import urllib2
import os
def image_from_file(url):
tmpfile = tempfile.NamedTemporaryFile(delete=False)
tmpFileName = tmpfile.name
image = None
try:
with tmpfile:
message = urllib2.urlopen(url, None, 15)
tmpfile.write(message.read())
image = gtk.Image()
image.set_from_file(tmpfile.name)
finally:
os.unlink(tmpFileName)
return image
| Increment timeout when getting images | [FIX] Increment timeout when getting images
Signed-off-by: Ludovic Ferrandis <[email protected]>
| Python | lgpl-2.1 | lferrandis/dleyna-control,01org/dleyna-control,jku/media-service-demo,markdryan/dleyna-control,rmerlino/dleyna-control,jku/dleyna-control |
2015233d252e625419485c269f1f70a7e0edada8 | skmisc/__init__.py | skmisc/__init__.py | from ._version import get_versions
__version__ = get_versions()['version']
del get_versions
__all__ = ['__version__']
# We first need to detect if we're being called as part of the skmisc
# setup procedure itself in a reliable manner.
try:
__SKMISC_SETUP__
except NameError:
__SKMISC_SETUP__ = False
if __SKMISC_SETUP__:
import sys as _sys
_sys.stderr.write('Running from skmisc source directory.\n')
del _sys
else:
from skmisc.__config__ import show as show_config # noqa: F401
# try:
# from skmisc.__config__ import show as show_config # noqa: F401
# except ImportError:
# msg = """Error importing skmisc: you cannot import skmisc while
# being in skmisc source directory; please exit the skmisc source
# tree first, and relaunch your python intepreter."""
# raise ImportError(msg)
__all__.append('show_config')
def test(args=None, plugins=None):
"""
Run tests
"""
# The doctests are not run when called from an installed
# package since the pytest.ini is not included in the
# package.
import os
try:
import pytest
except ImportError:
msg = "To run the tests, you must install pytest"
raise ImportError(msg)
path = os.path.realpath(__file__)
if args is None:
args = [path]
else:
args.append(path)
return pytest.main(args=args, plugins=plugins)
| from ._version import get_versions
__version__ = get_versions()['version']
del get_versions
__all__ = ['__version__']
# We first need to detect if we're being called as part of the skmisc
# setup procedure itself in a reliable manner.
try:
__SKMISC_SETUP__
except NameError:
__SKMISC_SETUP__ = False
if __SKMISC_SETUP__:
import sys as _sys
_sys.stderr.write('Running from skmisc source directory.\n')
del _sys
else:
try:
from skmisc.__config__ import show as show_config # noqa: F401
except ImportError as err:
msg = """Error importing skmisc: you cannot import skmisc while
being in skmisc source directory; please exit the skmisc source
tree first, and relaunch your python intepreter."""
raise ImportError('\n\n'.join([err.message, msg]))
__all__.append('show_config')
def test(args=None, plugins=None):
"""
Run tests
"""
# The doctests are not run when called from an installed
# package since the pytest.ini is not included in the
# package.
import os
try:
import pytest
except ImportError:
msg = "To run the tests, you must install pytest"
raise ImportError(msg)
path = os.path.dirname(os.path.realpath(__file__))
if args is None:
args = [path]
else:
args.append(path)
return pytest.main(args=args, plugins=plugins)
| Fix pytest path to root of package | Fix pytest path to root of package
Instead of the package init file.
| Python | bsd-3-clause | has2k1/onelib,has2k1/onelib,has2k1/onelib |
8187591a0f8255487f4b16b653ba5070bfffe739 | specs/test_diff.py | specs/test_diff.py | '''
This is an example of a python test
that compares a diff function (in this case
a hardcoded one that doesn't work) to the
reference JSON to check compliance.
'''
from nose.tools import eq_
import json
def diff(before, after):
return []
def test_diffs():
test_cases = json.load(open('test_cases.json'))
for test_case in test_cases:
result = diff(test_case['before'], test_case['after'])
eq_(result, test_case['diff'])
| '''
This is an example of a python test
that compares a diff function (in this case
a hardcoded one that doesn't work) to the
reference JSON to check compliance.
'''
from nose.tools import eq_
import json
def diff(before, after):
return []
def test_diffs():
test_cases = json.load(open('test_cases_simple.json'))
for test_case in test_cases:
result = diff(test_case['before'], test_case['after'])
eq_(result, test_case['diff'])
def test_diffs_cells():
test_cases = json.load(open('test_cases_cells.json'))
for test_case in test_cases:
result = diff(test_case['before'], test_case['after'])
eq_(result, test_case['diff'])
| Add code to specs example test | Add code to specs example test
| Python | mit | tarmstrong/nbdiff,tarmstrong/nbdiff,tarmstrong/nbdiff,tarmstrong/nbdiff |
66f6bb4b64ec13412edeef77c6db2f0a2e1932ba | src/sbml_model_checkup.py | src/sbml_model_checkup.py | """Debugging/cleanup checks."""
import itertools
import chen_2009_original_sbml
model = chen_2009_original_sbml.get_model()
# determine truly redundant species -- same name, same compartment
sa = sorted(model.species, key=lambda s: (s.label, s.compartment))
rs = [x
for x in ((n, map(lambda s: s.compartment, it))
for n, it in itertools.groupby(sa, lambda s: s.label))
if len(x[1]) == 2 and x[1][0] == x[1][1]]
# find reactions where product is not a trivial concatenation of reactants
# (e.g. A + B -> A:B)
mismatch_rxns = [r for r in model.reactions
if r.products[0].label != ':'.join([s.label for s in r.reactants])]
| """Debugging/cleanup checks."""
import itertools
import chen_2009_original_sbml
model = chen_2009_original_sbml.get_model()
# determine redundant species (same name)
sa = sorted(model.species, key=lambda s: (s.label, s.compartment))
rs = [x
for x in ((n, map(lambda s: s.compartment, it))
for n, it in itertools.groupby(sa, lambda s: s.label))
if len(x[1]) > 1 ]
# find reactions where label is not reactants -> products
def format_slist(species):
return ' + '.join(s.label.replace('endo|', '') for s in species)
def format_rxn(r):
return ' -> '.join(map(format_slist, [r.reactants, r.products]))
mislabeled_rxns = [(r, fr) for r, fr in
zip(model.reactions, map(format_rxn, model.reactions))
if r.label != fr and not
(' + ->' in r.label and len(r.reactants) == 1)]
print
print "Mismatched reactions:"
print
for r, fr in mislabeled_rxns:
print " label: {}".format(r.label)
print " internal: {}".format(fr)
print
print "Duplicated species:"
for label, compartments in rs:
print " {} @ {}".format(label, ', '.join(compartments))
| Rework model_checkup to print results and make results more useful | Rework model_checkup to print results and make results more useful
| Python | mit | johnbachman/rasmodel,jmuhlich/rasmodel,subkar/rasmodel,johnbachman/ras_model,sorgerlab/rasmodel,jmuhlich/ras_model,jmuhlich/ras_model,jmuhlich/rasmodel,sorgerlab/ras_model,johnbachman/rasmodel,johnbachman/ras_model,sorgerlab/rasmodel,sorgerlab/ras_model |
05151bb3ccd018b37097ddf5288e9984f5b45716 | ci/management/commands/cancel_old_jobs.py | ci/management/commands/cancel_old_jobs.py | from __future__ import unicode_literals, absolute_import
from django.core.management.base import BaseCommand
from ci import models, views, TimeUtils
from datetime import timedelta
class Command(BaseCommand):
help = 'Cancel old Civet jobs. When a specific civet client is no longer running, it can leave jobs lying around that other clients have to ignore.'
def add_arguments(self, parser):
parser.add_argument('--dryrun', default=False, action='store_true', help="Don't make any changes, just report what would have happened")
parser.add_argument('--days', required=True, type=int, help="Cancel jobs older than this many days")
def handle(self, *args, **options):
dryrun = options["dryrun"]
days = options["days"]
d = TimeUtils.get_local_time() - timedelta(days=days)
jobs = models.Job.objects.filter(active=True, ready=True, status=models.JobStatus.NOT_STARTED, created__lt=d)
count = jobs.count()
prefix = ""
if dryrun:
prefix = "DRY RUN: "
for job in jobs.all():
self.stdout.write("%sCancel job %s: %s: %s" % (prefix, job.pk, job, job.created))
if not dryrun:
views.set_job_canceled(job, "Civet client hasn't run this job in too long a time")
job.event.set_complete_if_done()
if count == 0:
self.stdout.write("No jobs to cancel")
| from __future__ import unicode_literals, absolute_import
from django.core.management.base import BaseCommand
from ci import models, views, TimeUtils
from datetime import timedelta
class Command(BaseCommand):
help = 'Cancel old Civet jobs. When a specific civet client is no longer running, it can leave jobs lying around that other clients have to ignore.'
def add_arguments(self, parser):
parser.add_argument('--dryrun', default=False, action='store_true', help="Don't make any changes, just report what would have happened")
parser.add_argument('--days', required=True, type=int, help="Cancel jobs older than this many days")
def handle(self, *args, **options):
dryrun = options["dryrun"]
days = options["days"]
d = TimeUtils.get_local_time() - timedelta(days=days)
jobs = models.Job.objects.filter(active=True, ready=True, status=models.JobStatus.NOT_STARTED, created__lt=d)
count = jobs.count()
prefix = ""
if dryrun:
prefix = "DRY RUN: "
for job in jobs.all():
self.stdout.write("%sCancel job %s: %s: %s" % (prefix, job.pk, job, job.created))
if not dryrun:
views.set_job_canceled(job, "Canceled due to civet client not running this job in too long a time")
job.event.set_complete_if_done()
if count == 0:
self.stdout.write("No jobs to cancel")
| Update cancel old job message | Update cancel old job message
| Python | apache-2.0 | idaholab/civet,brianmoose/civet,idaholab/civet,brianmoose/civet,brianmoose/civet,idaholab/civet,idaholab/civet,brianmoose/civet |
d93f1ff0b226d2e85cefc02afd0d5c44571f70ce | curious/utils.py | curious/utils.py | import time
from . import settings
# for development/debugging
def report_time(f):
def wrap(*args, **kwargs):
t = time.time()
r = f(*args, **kwargs)
if settings.DEBUG:
print '%s.%s: %.4f' % (f.__module__, f.func_name, time.time()-t)
return r
return wrap
| from functools import wraps
import time
from . import settings
# for development/debugging
def report_time(f):
@wraps(f)
def wrap(*args, **kwargs):
t = time.time()
r = f(*args, **kwargs)
if settings.DEBUG:
print '%s.%s: %.4f' % (f.__module__, f.func_name, time.time()-t)
return r
return wrap
| Add functools wraps to report_time decorator | Add functools wraps to report_time decorator
(cherry picked from commit 93d9de9e004896407214c5b67e64cb050bfaa63c)
| Python | mit | ginkgobioworks/curious,benjiec/curious,benjiec/curious,ginkgobioworks/curious,benjiec/curious,ginkgobioworks/curious |
aef51ce5ece86d054f76d86dafca9667f88d3b1a | ccui/testexecution/templatetags/results.py | ccui/testexecution/templatetags/results.py | # Case Conductor is a Test Case Management system.
# Copyright (C) 2011 uTest Inc.
#
# This file is part of Case Conductor.
#
# Case Conductor is free software: you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# Case Conductor is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with Case Conductor. If not, see <http://www.gnu.org/licenses/>.
from django import template
from django.core.urlresolvers import reverse
from ..models import TestCycle, TestRun, TestRunIncludedTestCase
register = template.Library()
@register.filter
def results_detail_url(obj):
if isinstance(obj, TestCycle):
return reverse("results_testruns") + "?testCycle=%s" % obj.id
elif isinstance(obj, TestRun):
return reverse("results_testcases") + "?testRun=%s" % obj.id
elif isinstance(obj, TestRunIncludedTestCase):
return reverse("results_testcase_detail", kwargs={"itc_id": obj.id})
return ""
| # Case Conductor is a Test Case Management system.
# Copyright (C) 2011 uTest Inc.
#
# This file is part of Case Conductor.
#
# Case Conductor is free software: you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# Case Conductor is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with Case Conductor. If not, see <http://www.gnu.org/licenses/>.
from django import template
from django.core.urlresolvers import reverse
from ..models import TestCycle, TestRun, TestRunIncludedTestCase
register = template.Library()
@register.filter
def results_detail_url(obj):
if isinstance(obj, TestCycle):
return reverse("results_testruns") + "?filter-testCycle=%s" % obj.id
elif isinstance(obj, TestRun):
return reverse("results_testcases") + "?filter-testRun=%s" % obj.id
elif isinstance(obj, TestRunIncludedTestCase):
return reverse("results_testcase_detail", kwargs={"itc_id": obj.id})
return ""
| Fix result status chiclet links for new-style filter querystrings. | Fix result status chiclet links for new-style filter querystrings.
| Python | bsd-2-clause | shinglyu/moztrap,shinglyu/moztrap,bobsilverberg/moztrap,mccarrmb/moztrap,mccarrmb/moztrap,mozilla/moztrap,shinglyu/moztrap,bobsilverberg/moztrap,mozilla/moztrap,bobsilverberg/moztrap,mozilla/moztrap,mccarrmb/moztrap,shinglyu/moztrap,mozilla/moztrap,mccarrmb/moztrap,bobsilverberg/moztrap,mccarrmb/moztrap,shinglyu/moztrap,mozilla/moztrap |
538acc8a114c9fda8489dc5fe91fed2314a37c9b | src/sentry/web/forms/invite_organization_member.py | src/sentry/web/forms/invite_organization_member.py | from __future__ import absolute_import
from django import forms
from django.db import transaction, IntegrityError
from sentry.models import (
AuditLogEntry, AuditLogEntryEvent, OrganizationMember,
OrganizationMemberType
)
class InviteOrganizationMemberForm(forms.ModelForm):
class Meta:
fields = ('email',)
model = OrganizationMember
def save(self, actor, organization, ip_address):
om = super(InviteOrganizationMemberForm, self).save(commit=False)
om.organization = organization
om.type = OrganizationMemberType.MEMBER
try:
existing = OrganizationMember.objects.get(
organization=organization,
user__email__iexact=om.email,
)
except OrganizationMember.DoesNotExist:
pass
else:
return existing, False
sid = transaction.savepoint(using='default')
try:
om.save()
except IntegrityError:
transaction.savepoint_rollback(sid, using='default')
return OrganizationMember.objects.get(
email__iexact=om.email,
organization=organization,
), False
transaction.savepoint_commit(sid, using='default')
AuditLogEntry.objects.create(
organization=organization,
actor=actor,
ip_address=ip_address,
target_object=om.id,
event=AuditLogEntryEvent.MEMBER_INVITE,
data=om.get_audit_log_data(),
)
om.send_invite_email()
return om, True
| from __future__ import absolute_import
from django import forms
from django.db import transaction, IntegrityError
from sentry.models import (
AuditLogEntry, AuditLogEntryEvent, OrganizationMember,
OrganizationMemberType
)
class InviteOrganizationMemberForm(forms.ModelForm):
class Meta:
fields = ('email',)
model = OrganizationMember
def save(self, actor, organization, ip_address):
om = super(InviteOrganizationMemberForm, self).save(commit=False)
om.organization = organization
om.type = OrganizationMemberType.MEMBER
try:
existing = OrganizationMember.objects.filter(
organization=organization,
user__email__iexact=om.email,
)[0]
except IndexError:
pass
else:
return existing, False
sid = transaction.savepoint(using='default')
try:
om.save()
except IntegrityError:
transaction.savepoint_rollback(sid, using='default')
return OrganizationMember.objects.get(
email__iexact=om.email,
organization=organization,
), False
transaction.savepoint_commit(sid, using='default')
AuditLogEntry.objects.create(
organization=organization,
actor=actor,
ip_address=ip_address,
target_object=om.id,
event=AuditLogEntryEvent.MEMBER_INVITE,
data=om.get_audit_log_data(),
)
om.send_invite_email()
return om, True
| Handle members with duplicate email addresses | Handle members with duplicate email addresses
| Python | bsd-3-clause | ifduyue/sentry,jean/sentry,gg7/sentry,daevaorn/sentry,Kryz/sentry,songyi199111/sentry,felixbuenemann/sentry,vperron/sentry,looker/sentry,jean/sentry,looker/sentry,hongliang5623/sentry,alexm92/sentry,daevaorn/sentry,pauloschilling/sentry,fuziontech/sentry,TedaLIEz/sentry,mvaled/sentry,beeftornado/sentry,pauloschilling/sentry,BuildingLink/sentry,mitsuhiko/sentry,zenefits/sentry,alexm92/sentry,zenefits/sentry,songyi199111/sentry,wong2/sentry,Kryz/sentry,BuildingLink/sentry,daevaorn/sentry,drcapulet/sentry,ifduyue/sentry,gg7/sentry,JamesMura/sentry,jean/sentry,nicholasserra/sentry,ngonzalvez/sentry,BayanGroup/sentry,TedaLIEz/sentry,JackDanger/sentry,JackDanger/sentry,JamesMura/sentry,BuildingLink/sentry,TedaLIEz/sentry,JamesMura/sentry,boneyao/sentry,ewdurbin/sentry,1tush/sentry,Natim/sentry,korealerts1/sentry,zenefits/sentry,felixbuenemann/sentry,fotinakis/sentry,kevinlondon/sentry,hongliang5623/sentry,gg7/sentry,zenefits/sentry,JTCunning/sentry,kevinlondon/sentry,ngonzalvez/sentry,mvaled/sentry,ifduyue/sentry,mvaled/sentry,JamesMura/sentry,BuildingLink/sentry,gencer/sentry,nicholasserra/sentry,ewdurbin/sentry,zenefits/sentry,ngonzalvez/sentry,JTCunning/sentry,ifduyue/sentry,JamesMura/sentry,korealerts1/sentry,jean/sentry,beeftornado/sentry,fuziontech/sentry,nicholasserra/sentry,kevinlondon/sentry,looker/sentry,imankulov/sentry,hongliang5623/sentry,1tush/sentry,imankulov/sentry,gencer/sentry,wujuguang/sentry,Natim/sentry,kevinastone/sentry,looker/sentry,kevinastone/sentry,mvaled/sentry,drcapulet/sentry,songyi199111/sentry,ifduyue/sentry,jean/sentry,fotinakis/sentry,BayanGroup/sentry,mvaled/sentry,gencer/sentry,mvaled/sentry,boneyao/sentry,imankulov/sentry,fotinakis/sentry,felixbuenemann/sentry,looker/sentry,gencer/sentry,Natim/sentry,alexm92/sentry,JTCunning/sentry,Kryz/sentry,vperron/sentry,wong2/sentry,wong2/sentry,kevinastone/sentry,ewdurbin/sentry,drcapulet/sentry,vperron/sentry,korealerts1/sentry,fuziontech/sentry,BuildingLink/sentry,fotinakis/sentry,1tush/sentry,daevaorn/sentry,BayanGroup/sentry,beeftornado/sentry,mitsuhiko/sentry,gencer/sentry,pauloschilling/sentry,wujuguang/sentry,boneyao/sentry,JackDanger/sentry,wujuguang/sentry |
3c7641c3380acab821dcbf2ae274da4fb8fade96 | students/psbriant/final_project/test_clean_data.py | students/psbriant/final_project/test_clean_data.py | """
Name: Paul Briant
Date: 12/11/16
Class: Introduction to Python
Assignment: Final Project
Description:
Tests for Final Project
"""
import clean_data as cd
import pandas
import io
def get_data():
"""
"""
data = pandas.read_csv("data/Residential_Water_Usage_Zip_Code_on_Top.csv")
return data
def test_clean():
"""
"""
def test_rename_columns():
"""
"""
data = get_data()
data = data.drop(["Date Value"], axis=1)
column_names = list(data.columns.values)
column_list = cd.rename_columns(column_names)
assert column_list[0:5] == ["Date", "90001", "90002", "90003", "90004"]
def test_find_low_water_use():
"""
"""
def test_plot_zipcode():
"""
"""
| """
Name: Paul Briant
Date: 12/11/16
Class: Introduction to Python
Assignment: Final Project
Description:
Tests for Final Project
"""
import clean_data as cd
import pandas
def get_data():
"""
Retrieve data from csv file to test.
"""
data = pandas.read_csv("data/Residential_Water_Usage_Zip_Code_on_Top.csv")
return data
def test_clean():
"""
"""
def test_rename_columns():
"""
Test whether rename_columns successfully renames each column.
"""
data = get_data()
data = data.drop(["Date Value"], axis=1)
column_names = list(data.columns.values)
column_list = cd.rename_columns(column_names)
assert column_list[0:5] == ["Date", "90001", "90002", "90003", "90004"]
def test_find_low_water_use():
"""
"""
def test_plot_zipcode():
"""
"""
| Add docstrings for get_data and test_rename_columns and remove import io statement. | Add docstrings for get_data and test_rename_columns and remove import io statement.
| Python | unlicense | weidnem/IntroPython2016,UWPCE-PythonCert/IntroPython2016,UWPCE-PythonCert/IntroPython2016,weidnem/IntroPython2016,weidnem/IntroPython2016,UWPCE-PythonCert/IntroPython2016 |
17ddcb1b6c293197834b3154830b9521769d76fb | linter.py | linter.py | #
# linter.py
# Linter for SublimeLinter3, a code checking framework for Sublime Text 3
#
# Written by Hardy Jones
# Copyright (c) 2013
#
# License: MIT
#
"""This module exports the Hlint plugin class."""
from SublimeLinter.lint import Linter
class Hlint(Linter):
"""Provides an interface to hlint."""
syntax = ('haskell', 'haskell-sublimehaskell', 'literate haskell')
cmd = 'hlint'
regex = (
r'^.+:(?P<line>\d+):'
'(?P<col>\d+):\s*'
'(?:(?P<error>Error)|(?P<warning>Warning)):\s*'
'(?P<message>.+)$'
)
multiline = True
tempfile_suffix = {
'haskell': 'hs',
'haskell-sublimehaskell': 'hs',
'literate haskell': 'lhs'
}
| #
# linter.py
# Linter for SublimeLinter3, a code checking framework for Sublime Text 3
#
# Written by Hardy Jones
# Copyright (c) 2013
#
# License: MIT
#
"""This module exports the Hlint plugin class."""
from SublimeLinter.lint import Linter
class Hlint(Linter):
"""Provides an interface to hlint."""
defaults = {
'selector': 'source.haskell'
}
cmd = 'hlint'
regex = (
r'^.+:(?P<line>\d+):'
'(?P<col>\d+):\s*'
'(?:(?P<error>Error)|(?P<warning>Warning)):\s*'
'(?P<message>.+)$'
)
multiline = True
tempfile_suffix = 'hs'
| Update to new `defaults` configuration | Update to new `defaults` configuration
| Python | mit | SublimeLinter/SublimeLinter-hlint |
2bab1888b43a9c232b37cc26c37df992ea5df2c5 | project/apps/api/signals.py | project/apps/api/signals.py | from django.db.models.signals import (
post_save,
)
from django.dispatch import receiver
from .models import (
Performance,
Session,
)
@receiver(post_save, sender=Performance)
def performance_post_save(sender, instance=None, created=False, raw=False, **kwargs):
"""Create sentinels."""
if not raw:
if created:
s = 1
while s <= instance.round.num_songs:
song = instance.songs.create(
performance=instance,
num=s,
)
s += 1
judges = instance.round.session.judges.filter(
category__in=[
instance.round.session.judges.model.CATEGORY.music,
instance.round.session.judges.model.CATEGORY.presentation,
instance.round.session.judges.model.CATEGORY.singing,
]
)
for judge in judges:
judge.scores.create(
judge=judge,
song=song,
category=judge.category,
kind=judge.kind,
)
@receiver(post_save, sender=Session)
def session_post_save(sender, instance=None, created=False, raw=False, **kwargs):
"""Create sentinels."""
if not raw:
if created:
i = 1
while i <= instance.num_rounds:
instance.rounds.create(
num=i,
kind=(instance.num_rounds - i) + 1,
)
i += 1
| from django.db.models.signals import (
post_save,
)
from django.dispatch import receiver
from .models import (
Performance,
Session,
)
@receiver(post_save, sender=Session)
def session_post_save(sender, instance=None, created=False, raw=False, **kwargs):
"""Create sentinels."""
if not raw:
if created:
i = 1
while i <= instance.num_rounds:
instance.rounds.create(
num=i,
kind=(instance.num_rounds - i) + 1,
)
i += 1
@receiver(post_save, sender=Performance)
def performance_post_save(sender, instance=None, created=False, raw=False, **kwargs):
"""Create sentinels."""
if not raw:
if created:
s = 1
while s <= instance.round.num_songs:
song = instance.songs.create(
performance=instance,
num=s,
)
s += 1
judges = instance.round.session.judges.filter(
category__in=[
instance.round.session.judges.model.CATEGORY.music,
instance.round.session.judges.model.CATEGORY.presentation,
instance.round.session.judges.model.CATEGORY.singing,
]
)
for judge in judges:
judge.scores.create(
judge=judge,
song=song,
category=judge.category,
kind=judge.kind,
)
| Create sentinel rounds on Session creation | Create sentinel rounds on Session creation
| Python | bsd-2-clause | barberscore/barberscore-api,barberscore/barberscore-api,dbinetti/barberscore-django,dbinetti/barberscore,barberscore/barberscore-api,dbinetti/barberscore-django,dbinetti/barberscore,barberscore/barberscore-api |
0f3f9ac26e968eab788a9ee2e95c34f2fa05ddbf | hooks/settings/settings_gunicorn.py | hooks/settings/settings_gunicorn.py | # -*- coding: utf-8 -*-
"""
Related article: https://sebest.github.io/post/protips-using-gunicorn-inside-a-docker-image/
Parameters you might want to override:
GUNICORN_BIND="0.0.0.0:8005"
"""
import os
workers = 4
bind = "0.0.0.0:8005"
worker_class = "eventlet"
worker_connections = 100
# Overwrite some Gunicorns params by ENV variables
for k, v in os.environ.items():
if k.startswith("GUNICORN_"):
key = k.split('_', 1)[1].lower()
locals()[key] = v
| # -*- coding: utf-8 -*-
"""
Related article: https://sebest.github.io/post/protips-using-gunicorn-inside-a-docker-image/
Parameters you might want to override:
GUNICORN_BIND="0.0.0.0:8005"
"""
import os
workers = 4
bind = "0.0.0.0:8005"
worker_class = "eventlet"
worker_connections = 1000
# Overwrite some Gunicorns params by ENV variables
for k, v in os.environ.items():
if k.startswith("GUNICORN_"):
key = k.split('_', 1)[1].lower()
locals()[key] = v
| Increase Gunicorn worker connections, 100 -> 1000 | Increase Gunicorn worker connections, 100 -> 1000
| Python | mit | business-factory/captain-hook |
a5a5c21dfc8fa03cbf0a5585af3b1e7bd8c875bc | trac/web/__init__.py | trac/web/__init__.py | # With mod_python we'll have to delay importing trac.web.api until
# modpython_frontend.handler() has been called since the
# PYTHON_EGG_CACHE variable is set from there
#
# TODO: Remove this once the Genshi zip_safe issue has been resolved.
import os
from pkg_resources import get_distribution
if not os.path.isdir(get_distribution('genshi').location):
try:
import mod_python.apache
import sys
if 'trac.web.modpython_frontend' in sys.modules:
from trac.web.api import *
except ImportError:
from trac.web.api import *
else:
from trac.web.api import *
| # Workaround for http://bugs.python.org/issue6763 and
# http://bugs.python.org/issue5853 thread issues
import mimetypes
mimetypes.init()
# With mod_python we'll have to delay importing trac.web.api until
# modpython_frontend.handler() has been called since the
# PYTHON_EGG_CACHE variable is set from there
#
# TODO: Remove this once the Genshi zip_safe issue has been resolved.
import os
from pkg_resources import get_distribution
if not os.path.isdir(get_distribution('genshi').location):
try:
import mod_python.apache
import sys
if 'trac.web.modpython_frontend' in sys.modules:
from trac.web.api import *
except ImportError:
from trac.web.api import *
else:
from trac.web.api import *
| Fix race condition during `mimetypes` initialization. | Fix race condition during `mimetypes` initialization.
Initial patch from Steven R. Loomis.
Closes #8629.
| Python | bsd-3-clause | pkdevbox/trac,pkdevbox/trac,pkdevbox/trac,pkdevbox/trac |
375b26fbb6e5ba043a1017e28027241c12374207 | napalm_logs/transport/zeromq.py | napalm_logs/transport/zeromq.py | # -*- coding: utf-8 -*-
'''
ZeroMQ transport for napalm-logs.
'''
from __future__ import absolute_import
from __future__ import unicode_literals
# Import stdlib
import json
# Import third party libs
import zmq
# Import napalm-logs pkgs
from napalm_logs.transport.base import TransportBase
class ZMQTransport(TransportBase):
'''
ZMQ transport class.
'''
def __init__(self, addr, port):
self.addr = addr
self.port = port
def start(self):
self.context = zmq.Context()
self.socket = self.context.socket(zmq.PUB)
self.socket.bind('tcp://{addr}:{port}'.format(
addr=self.addr,
port=self.port)
)
def serialise(self, obj):
return json.dumps(obj)
def publish(self, obj):
self.socket.send(
self.serialise(obj)
)
def tear_down(self):
if hasattr(self, 'socket'):
self.socket.close()
if hasattr(self, 'context'):
self.context.term()
| # -*- coding: utf-8 -*-
'''
ZeroMQ transport for napalm-logs.
'''
from __future__ import absolute_import
from __future__ import unicode_literals
# Import stdlib
import json
import logging
# Import third party libs
import zmq
# Import napalm-logs pkgs
from napalm_logs.exceptions import BindException
from napalm_logs.transport.base import TransportBase
log = logging.getLogger(__name__)
class ZMQTransport(TransportBase):
'''
ZMQ transport class.
'''
def __init__(self, addr, port):
self.addr = addr
self.port = port
def start(self):
self.context = zmq.Context()
self.socket = self.context.socket(zmq.PUB)
try:
self.socket.bind('tcp://{addr}:{port}'.format(
addr=self.addr,
port=self.port)
)
except zmq.error.ZMQError as err:
log.error(err, exc_info=True)
raise BindException(err)
def serialise(self, obj):
return json.dumps(obj)
def publish(self, obj):
self.socket.send(
self.serialise(obj)
)
def tear_down(self):
if hasattr(self, 'socket'):
self.socket.close()
if hasattr(self, 'context'):
self.context.term()
| Raise bind exception and log | Raise bind exception and log
| Python | apache-2.0 | napalm-automation/napalm-logs,napalm-automation/napalm-logs |
c8ce1315caf762f2c0073ab0ebed8ef627be5581 | profile/files/openstack/horizon/overrides.py | profile/files/openstack/horizon/overrides.py | # Disable Floating IPs
from openstack_dashboard.dashboards.project.access_and_security import tabs
from openstack_dashboard.dashboards.project.instances import tables
import horizon
NO = lambda *x: False
tabs.FloatingIPsTab.allowed = NO
tabs.APIAccessTab.allowed = NO
tables.AssociateIP.allowed = NO
tables.SimpleAssociateIP.allowed = NO
tables.SimpleDisassociateIP.allowed = NO
tables.ResizeLink.allowed = NO
project_dashboard = horizon.get_dashboard("project")
# Completely remove panel Network->Routers
routers_panel = project_dashboard.get_panel("routers")
project_dashboard.unregister(routers_panel.__class__)
# Completely remove panel Network->Networks
networks_panel = project_dashboard.get_panel("networks")
project_dashboard.unregister(networks_panel.__class__) # Disable Floating IPs
| # Disable Floating IPs
from openstack_dashboard.dashboards.project.access_and_security import tabs
from openstack_dashboard.dashboards.project.instances import tables
import horizon
NO = lambda *x: False
tabs.FloatingIPsTab.allowed = NO
tabs.APIAccessTab.allowed = NO
tables.AssociateIP.allowed = NO
tables.SimpleAssociateIP.allowed = NO
tables.SimpleDisassociateIP.allowed = NO
tables.ResizeLink.allowed = NO
project_dashboard = horizon.get_dashboard("project")
# Completely remove panel Network->Routers
routers_panel = project_dashboard.get_panel("routers")
project_dashboard.unregister(routers_panel.__class__)
# Completely remove panel Network->Networks
networks_panel = project_dashboard.get_panel("networks")
project_dashboard.unregister(networks_panel.__class__) # Disable Floating IPs
# Remove "Volume Consistency Groups" tab
from openstack_dashboard.dashboards.project.volumes import tabs
tabs.CGroupsTab.allowed = NO
| Remove volume consistency group tab from horizon in mitaka | Remove volume consistency group tab from horizon in mitaka
| Python | apache-2.0 | raykrist/himlar,tanzr/himlar,mikaeld66/himlar,raykrist/himlar,norcams/himlar,eckhart/himlar,mikaeld66/himlar,norcams/himlar,mikaeld66/himlar,raykrist/himlar,tanzr/himlar,TorLdre/himlar,tanzr/himlar,TorLdre/himlar,mikaeld66/himlar,tanzr/himlar,eckhart/himlar,norcams/himlar,raykrist/himlar,eckhart/himlar,TorLdre/himlar,norcams/himlar,TorLdre/himlar,mikaeld66/himlar,tanzr/himlar,raykrist/himlar,norcams/himlar,eckhart/himlar,TorLdre/himlar |
88dc672b8797834b03d67b962dda2de2d40ad4f1 | demos/minimal.py | demos/minimal.py | #!/usr/bin/env python
from gi.repository import GtkClutter
GtkClutter.init([])
from gi.repository import GObject, Gtk, GtkChamplain
GObject.threads_init()
GtkClutter.init([])
window = Gtk.Window(type=Gtk.WindowType.TOPLEVEL)
window.connect("destroy", Gtk.main_quit)
widget = GtkChamplain.Embed()
widget.set_size_request(640, 480)
window.add(widget)
window.show_all()
Gtk.main()
| #!/usr/bin/env python
# To run this example, you need to set the GI_TYPELIB_PATH environment
# variable to point to the gir directory:
#
# export GI_TYPELIB_PATH=$GI_TYPELIB_PATH:/usr/local/lib/girepository-1.0/
from gi.repository import GtkClutter
GtkClutter.init([])
from gi.repository import GObject, Gtk, GtkChamplain
GObject.threads_init()
GtkClutter.init([])
window = Gtk.Window(type=Gtk.WindowType.TOPLEVEL)
window.connect("destroy", Gtk.main_quit)
widget = GtkChamplain.Embed()
widget.set_size_request(640, 480)
window.add(widget)
window.show_all()
Gtk.main()
| Add description how to run the python demo | Add description how to run the python demo
| Python | lgpl-2.1 | Distrotech/libchamplain,PabloCastellano/libchamplain,PabloCastellano/libchamplain,StanciuMarius/Libchamplain-map-wrapping,PabloCastellano/libchamplain,StanciuMarius/Libchamplain-map-wrapping,StanciuMarius/Libchamplain-map-wrapping,PabloCastellano/libchamplain,GNOME/libchamplain,StanciuMarius/Libchamplain-map-wrapping,Distrotech/libchamplain,StanciuMarius/Libchamplain-map-wrapping,Distrotech/libchamplain,Distrotech/libchamplain,Distrotech/libchamplain,GNOME/libchamplain,PabloCastellano/libchamplain |
0cda8aae3c5a8ad4d110c41007279a3364c6f33a | manage.py | manage.py | __author__ = 'zifnab'
from flask_script import Manager, Server
from app import app
manager=Manager(app)
manager.add_command('runserver', Server(host=app.config.get('HOST', '0.0.0.0'), port=app.config.get('PORT', 5000)))
@manager.command
def print_routes():
for rule in app.url_map.iter_rules():
print rule
if __name__ == '__main__':
manager.run() | __author__ = 'zifnab'
from flask_script import Manager, Server
from app import app
from database import Paste
import arrow
manager=Manager(app)
manager.add_command('runserver', Server(host=app.config.get('HOST', '0.0.0.0'), port=app.config.get('PORT', 5000)))
@manager.command
def print_routes():
for rule in app.url_map.iter_rules():
print rule
@manager.command
def remove_expired():
for paste in Paste.objects(expire__lt=arrow.now().datetime, user=None):
print 'delete {0}'.format(paste.name)
paste.delete()
if __name__ == '__main__':
manager.run() | Add way to remove old pastes | Add way to remove old pastes
| Python | mit | zifnab06/zifb.in,zifnab06/zifb.in |
d2f3ff32e6d0a8c03a76f93669bb1f37d28ae124 | parsl/tests/configs/local_threads_globus.py | parsl/tests/configs/local_threads_globus.py | from parsl.config import Config
from parsl.data_provider.scheme import GlobusScheme
from parsl.executors.threads import ThreadPoolExecutor
from parsl.tests.utils import get_rundir
# If you are a developer running tests, make sure to update parsl/tests/configs/user_opts.py
# If you are a user copying-and-pasting this as an example, make sure to either
# 1) create a local `user_opts.py`, or
# 2) delete the user_opts import below and replace all appearances of `user_opts` with the literal value
# (i.e., user_opts['swan']['username'] -> 'your_username')
from .user_opts import user_opts
config = Config(
executors=[
ThreadPoolExecutor(
label='local_threads_globus',
storage_access=GlobusScheme(
endpoint_uuid=user_opts['globus']['endpoint'],
endpoint_path=user_opts['globus']['path']
),
working_dir=user_opts['globus']['path']
)
],
run_dir=get_rundir()
)
| from parsl.config import Config
from parsl.data_provider.scheme import GlobusScheme
from parsl.executors.threads import ThreadPoolExecutor
from parsl.tests.utils import get_rundir
# If you are a developer running tests, make sure to update parsl/tests/configs/user_opts.py
# If you are a user copying-and-pasting this as an example, make sure to either
# 1) create a local `user_opts.py`, or
# 2) delete the user_opts import below and replace all appearances of `user_opts` with the literal value
# (i.e., user_opts['swan']['username'] -> 'your_username')
from .user_opts import user_opts
config = Config(
executors=[
ThreadPoolExecutor(
label='local_threads_globus',
storage_access=[GlobusScheme(
endpoint_uuid=user_opts['globus']['endpoint'],
endpoint_path=user_opts['globus']['path']
)[,
working_dir=user_opts['globus']['path']
)
],
run_dir=get_rundir()
)
| Fix storage_access in the test config | Fix storage_access in the test config
| Python | apache-2.0 | Parsl/parsl,Parsl/parsl,Parsl/parsl,swift-lang/swift-e-lab,Parsl/parsl,swift-lang/swift-e-lab |
3260594268f19dcfe1ea5613f939c892d609b47e | skimage/filters/tests/test_filter_import.py | skimage/filters/tests/test_filter_import.py | from warnings import catch_warnings, simplefilter
def test_filter_import():
with catch_warnings():
simplefilter('ignore')
from skimage import filter as F
assert('sobel' in dir(F))
assert any(['has been renamed' in w
for (w, _, _) in F.__warningregistry__])
| from warnings import catch_warnings, simplefilter
def test_filter_import():
with catch_warnings():
simplefilter('ignore')
from skimage import filter as F
assert('sobel' in dir(F))
assert any(['has been renamed' in w
for (w, _, _) in F.__warningregistry__]), F.__warningregistry__
| Add debug print to failing assert | Add debug print to failing assert
| Python | bsd-3-clause | Hiyorimi/scikit-image,juliusbierk/scikit-image,dpshelio/scikit-image,chriscrosscutler/scikit-image,Britefury/scikit-image,juliusbierk/scikit-image,vighneshbirodkar/scikit-image,jwiggins/scikit-image,WarrenWeckesser/scikits-image,rjeli/scikit-image,Britefury/scikit-image,youprofit/scikit-image,ajaybhat/scikit-image,bennlich/scikit-image,oew1v07/scikit-image,Midafi/scikit-image,Hiyorimi/scikit-image,keflavich/scikit-image,bennlich/scikit-image,michaelpacer/scikit-image,bsipocz/scikit-image,ofgulban/scikit-image,vighneshbirodkar/scikit-image,jwiggins/scikit-image,oew1v07/scikit-image,GaZ3ll3/scikit-image,michaelaye/scikit-image,chriscrosscutler/scikit-image,warmspringwinds/scikit-image,dpshelio/scikit-image,WarrenWeckesser/scikits-image,ajaybhat/scikit-image,blink1073/scikit-image,pratapvardhan/scikit-image,rjeli/scikit-image,Midafi/scikit-image,ofgulban/scikit-image,GaZ3ll3/scikit-image,robintw/scikit-image,emon10005/scikit-image,pratapvardhan/scikit-image,blink1073/scikit-image,ClinicalGraphics/scikit-image,keflavich/scikit-image,rjeli/scikit-image,youprofit/scikit-image,michaelaye/scikit-image,ClinicalGraphics/scikit-image,bsipocz/scikit-image,paalge/scikit-image,paalge/scikit-image,robintw/scikit-image,newville/scikit-image,warmspringwinds/scikit-image,ofgulban/scikit-image,vighneshbirodkar/scikit-image,newville/scikit-image,emon10005/scikit-image,paalge/scikit-image,michaelpacer/scikit-image |
d469a02c697a20d7db36a3a80d44cfca2257a5cd | django/models.py | django/models.py | from django.db import models
from django.utils.translation import ugettext_lazy as _
{%% for model_name, props in all_models.iteritems() %%}
{%% set model_name = model_name|capitalize %%}
class {{{ model_name }}}(models.Model):
{%% for prop, value in props.iteritems() %%}
{{{ prop }}} = {{{ value|model_field }}}{%% endfor %%}
def __unicode__(self):
fields = [{%% for prop, value in props.iteritems() %%}'{{{ prop }}}', {%% endfor %%}]
return unicode('{%% for prop, value in props.iteritems() %%}<{}>, {%% endfor %%}'.format(*fields))
{%% endfor %%}
| from django.db import models
from django.utils.translation import ugettext_lazy as _
{%% for model_name, props in all_models.iteritems() %%}
{%% set model_name = model_name|capitalize %%}
class {{{ model_name }}}(models.Model):
{%% for prop, value in props.iteritems() %%}
{{{ prop }}} = {{{ value|model_field }}}{%% endfor %%}
def __unicode__(self):
fields = [{%% for prop, value in props.iteritems() %%}'{{{ prop }}}', {%% endfor %%}]
return unicode('<{%% for prop, value in props.iteritems() %%}{}.{%% endfor %%}>'.format(*fields))
{%% endfor %%}
| Change string representation for django model | Change string representation for django model
| Python | apache-2.0 | christabor/Skaffold,christabor/Skaffold |
7756236c5e1fa70f1173dbd58b7e57f56214c19f | unitTestUtils/parseXML.py | unitTestUtils/parseXML.py | #!/usr/bin/python
# -*- coding: utf-8 -*-
from __future__ import print_function
from xml.etree.ElementTree import ParseError
import xml.etree.ElementTree as ET
import glob
import sys
def eprint(*args, **kwargs):
print(*args, file=sys.stderr, **kwargs)
def parse():
for infile in glob.glob('*.xml'):
try:
tree = ET.parse(infile)
root = tree.getroot()
if root.findall('.//FatalError'):
eprint("Error detected")
print(infile)
sys.exit(1)
except ParseError:
eprint("The file xml isn't correct. There were some mistakes in the tests ")
sys.exit(1)
def main():
parse()
if __name__ == '__main__':
main()
| #!/usr/bin/python
# -*- coding: utf-8 -*-
from __future__ import print_function
from xml.etree.ElementTree import ParseError
import xml.etree.ElementTree as ET
import glob
import sys
def eprint(*args, **kwargs):
print(*args, file=sys.stderr, **kwargs)
def parse():
for infile in glob.glob('*.xml'):
try:
tree = ET.parse(infile)
root = tree.getroot()
if root.findall('.//FatalError'):
element=root.findall('.//FatalError')[0]
eprint("Error detected")
print(infile)
print(element.text)
sys.exit(1)
except ParseError:
eprint("The file xml isn't correct. There were some mistakes in the tests ")
sys.exit(1)
def main():
parse()
if __name__ == '__main__':
main()
| Add a verbose error reporting on Travis | Add a verbose error reporting on Travis
| Python | apache-2.0 | wkrzemien/j-pet-framework,wkrzemien/j-pet-framework,wkrzemien/j-pet-framework,wkrzemien/j-pet-framework,wkrzemien/j-pet-framework |
0df35e81754f703d1a8164cf0ea5169a53355185 | code/python/knub/thesis/word2vec_gaussian_lda_preprocessing.py | code/python/knub/thesis/word2vec_gaussian_lda_preprocessing.py | import argparse
import logging
import os
from gensim.models import Word2Vec
logging.basicConfig(format='%(asctime)s : %(levelname)s : %(message)s', level=logging.INFO)
if __name__ == "__main__":
parser = argparse.ArgumentParser("Prepare model for Gaussian LDA")
parser.add_argument("--topic_model", type=str)
parser.add_argument("--embedding_model", type=str)
args = parser.parse_args()
word2vec = Word2Vec.load_word2vec_format(args.embedding_model, binary=True)
embedding_name = os.path.basename(args.embedding_model)
with open(args.topic_model + "." + embedding_name + ".gaussian-lda", "w") as output:
with open(args.topic_model + ".restricted.alphabet", "r") as f:
for line in f:
word = line.split("#")[0]
output.write(word + " ")
output.write(" ".join(word2vec[word]))
output.write("\n")
| import argparse
from codecs import open
import logging
import os
from gensim.models import Word2Vec
logging.basicConfig(format='%(asctime)s : %(levelname)s : %(message)s', level=logging.INFO)
if __name__ == "__main__":
parser = argparse.ArgumentParser("Prepare model for Gaussian LDA")
parser.add_argument("--topic-model", type=str)
parser.add_argument("--embedding-model", type=str)
args = parser.parse_args()
word2vec = Word2Vec.load_word2vec_format(args.embedding_model, binary=True)
embedding_name = os.path.basename(args.embedding_model)
with open(args.topic_model + "." + embedding_name + ".gaussian-lda", "w", encoding="utf-8") as output:
with open(args.topic_model + "." + embedding_name + ".restricted.alphabet", "r", encoding="utf-8") as f:
for line in f:
word = line.split("#")[0]
output.write(word + " ")
output.write(" ".join(map(str, word2vec[word])))
output.write("\n")
| Fix parameter parsing in gaussian lda preprocessing. | Fix parameter parsing in gaussian lda preprocessing.
| Python | apache-2.0 | knub/master-thesis,knub/master-thesis,knub/master-thesis,knub/master-thesis |
9ad55a4d532dc98f257206ae82b5d06f4203a4d4 | server/lib/python/cartodb_services/setup.py | server/lib/python/cartodb_services/setup.py | """
CartoDB Services Python Library
See:
https://github.com/CartoDB/geocoder-api
"""
from setuptools import setup, find_packages
setup(
name='cartodb_services',
version='0.6.2',
description='CartoDB Services API Python Library',
url='https://github.com/CartoDB/geocoder-api',
author='Data Services Team - CartoDB',
author_email='[email protected]',
license='MIT',
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Mapping comunity',
'Topic :: Maps :: Mapping Tools',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python :: 2.7',
],
keywords='maps api mapping tools geocoder routing',
packages=find_packages(exclude=['contrib', 'docs', 'tests']),
extras_require={
'dev': ['unittest'],
'test': ['unittest', 'nose', 'mockredispy', 'mock'],
}
)
| """
CartoDB Services Python Library
See:
https://github.com/CartoDB/geocoder-api
"""
from setuptools import setup, find_packages
setup(
name='cartodb_services',
version='0.6.2',
description='CartoDB Services API Python Library',
url='https://github.com/CartoDB/dataservices-api',
author='Data Services Team - CartoDB',
author_email='[email protected]',
license='MIT',
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Mapping comunity',
'Topic :: Maps :: Mapping Tools',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python :: 2.7',
],
keywords='maps api mapping tools geocoder routing',
packages=find_packages(exclude=['contrib', 'docs', 'tests']),
extras_require={
'dev': ['unittest'],
'test': ['unittest', 'nose', 'mockredispy', 'mock'],
}
)
| Update url of pip package | Update url of pip package
| Python | bsd-3-clause | CartoDB/geocoder-api,CartoDB/geocoder-api,CartoDB/dataservices-api,CartoDB/dataservices-api,CartoDB/dataservices-api,CartoDB/geocoder-api,CartoDB/dataservices-api,CartoDB/geocoder-api |
ce0be23f554eb9949a3769da1e4a3d3d51b546f1 | src/server/datab.py | src/server/datab.py | '''
Database module.
Get the database, convert it to the built-in data structure and hold a link
to it. The module should be initialized before any other modules except mailer
and log.
Design: Heranort
'''
'''
Connect to the database.
'''
def connect_to_datab():
pass
'''
Get raw data of the database.
'''
def datab_get_raw_data():
pass
'''
Process the raw data.
'''
def datab_process_data():
pass
'''
Preserve the processed data into somewhere.
'''
def datab_preserve_data():
pass
'''
Check wether the history is modified. If so, emit warning.
'''
def check_health():
pass
| '''
Database module.
Get the database, convert it to the built-in data structure and hold a link
to it. The module should be initialized before any other modules except mailer
and log.
Design: Heranort
'''
import sqlite3, os
'''
Connect to the database.
'''
def connect_to_datab():
path = os.getcwd()
pparent_path = os.path.dirname(os.path.dirname(path)) #get the root dir
# print(pparent_path)
sql = sqlite3.connect(pparent_path + '\data\data.db')
return sql
'''
Get raw data of the database.
'''
def datab_get_raw_data(sql):
cur = sql.cursor()
cur.execute('select * from flight') #fetch the raw data of flight
raw_data_flight = cur.fetchall()
cur.execute('select * from train') #fetch the raw data of train
raw_data_train = cur.fetchall()
cur.execute('select * from highway') #fetch the raw data of highway
raw_data_bus = cur.fetchall()
return (raw_data_flight, raw_data_train, raw_data_bus)
'''
Process the raw data.
'''
def datab_process_data(raw_data_flight, raw_data_train, raw_data_bus):
data_price = [[-1 for i in range(10)] for i in range(10)]
data_instance = [[-1 for i in range(10)] for i in range(10)]
data_time = [[-1 for i in range(10)] for i in range(10)]
for element in raw_data_bus:
pass
'''
Preserve the processed data into somewhere.
'''
def datab_preserve_data():
pass
'''
Check wether the history is modified. If so, emit warning.
'''
def check_health():
pass
if(__name__ == '__main__'):
sql = connect_to_datab()
(raw_data_flight, raw_data_train, raw_data_bus) = datab_get_raw_data(sql)
datab_process_data(raw_data_flight, raw_data_train, raw_data_bus) | Add function of data connection | Add function of data connection
| Python | mit | niwtr/map-walker |
5c6df7c8ddb6ae698ab54b30f211bbcd28549c95 | packages/Python/lldbsuite/test/lang/cpp/template-function/TestTemplateFunctions.py | packages/Python/lldbsuite/test/lang/cpp/template-function/TestTemplateFunctions.py | """
Test that we can call C++ template fucntions.
"""
import lldb
from lldbsuite.test.decorators import *
from lldbsuite.test.lldbtest import *
from lldbsuite.test import lldbutil
class TemplateFunctionsTestCase(TestBase):
mydir = TestBase.compute_mydir(__file__)
def do_test_template_function(self, add_cast):
self.build()
(_, _, thread, _) = lldbutil.run_to_name_breakpoint(self, "main")
frame = thread.GetSelectedFrame()
expr = "foo(42)"
if add_cast:
expr = "(int)" + expr
expr_result = frame.EvaluateExpression(expr)
self.assertTrue(expr_result.IsValid())
self.assertEqual(expr_result.GetValue(), "42")
@skipIfWindows
def test_template_function_with_cast(self):
self.do_test_template_function(True)
@skipIfWindows
@expectedFailureAll(debug_info=["dwarf", "gmodules"])
def test_template_function_without_cast(self):
self.do_test_template_function(False)
| """
Test that we can call C++ template fucntions.
"""
import lldb
from lldbsuite.test.decorators import *
from lldbsuite.test.lldbtest import *
from lldbsuite.test import lldbutil
class TemplateFunctionsTestCase(TestBase):
mydir = TestBase.compute_mydir(__file__)
def do_test_template_function(self, add_cast):
self.build()
(_, _, thread, _) = lldbutil.run_to_name_breakpoint(self, "main")
frame = thread.GetSelectedFrame()
expr = "foo(42)"
if add_cast:
expr = "(int)" + expr
expr_result = frame.EvaluateExpression(expr)
self.assertTrue(expr_result.IsValid())
self.assertEqual(expr_result.GetValue(), "42")
@skipIfWindows
def test_template_function_with_cast(self):
self.do_test_template_function(True)
@skipIfWindows
@expectedFailureAll(debug_info=["dwarf", "gmodules", "dwo"])
def test_template_function_without_cast(self):
self.do_test_template_function(False)
| Add a missing debug info variant. | [TestTemplateFunction] Add a missing debug info variant.
git-svn-id: 4c4cc70b1ef44ba2b7963015e681894188cea27e@359249 91177308-0d34-0410-b5e6-96231b3b80d8
| Python | apache-2.0 | llvm-mirror/lldb,llvm-mirror/lldb,llvm-mirror/lldb,apple/swift-lldb,apple/swift-lldb,apple/swift-lldb,llvm-mirror/lldb,apple/swift-lldb,apple/swift-lldb,apple/swift-lldb,llvm-mirror/lldb |
5d3d47e0fae9ddb9f445972e5186429163aabf40 | statirator/core/management/commands/init.py | statirator/core/management/commands/init.py | import os
from optparse import make_option
from django.core.management.base import BaseCommand
class Command(BaseCommand):
help = "Init the static site project"
args = '[directory]'
option_list = (
make_option(
'--title', '-t', dest='title', default='Default site',
help='Site title [Default: "%defaults"]'),
make_option(
'--domain', '-d', dest='domain', default='example.com',
help='Domain name [Default: "%default"]'),
make_option(
'--languages', '-l', dest='languages', default=['he', 'en'],
action='append', help='Supported languages. [Default: "%default"]')
) + BaseCommand.option_list
def handle(self, directory, **options):
from django.conf.global_settings import LANGUAGES
extra = {
'build': 'build',
'default_lang': options['languages'][0],
'languages': [l for l in LANGUAGES if l[0] in options["languages"]],
'extensions': ('py', ),
'files': (),
'template': os.path.abspath(
os.path.join(
os.path.dirname(__file__),
os.pardir, os.pardir, os.pardir, 'project_template')),
}
extra.update(options)
from django.core.management import call_command
call_command('startproject', 'conf', directory, **extra)
| import os
import logging
from django.core.management.base import BaseCommand
from optparse import make_option
class Command(BaseCommand):
help = "Init the static site project"
args = '[directory]'
option_list = (
make_option(
'--title', '-t', dest='title', default='Default site',
help='Site title [Default: "%defaults"]'),
make_option(
'--domain', '-d', dest='domain', default='example.com',
help='Domain name [Default: "%default"]'),
make_option(
'--languages', '-l', dest='languages', default=['he', 'en'],
action='append', help='Supported languages. [Default: "%default"]')
) + BaseCommand.option_list
def handle(self, directory, **options):
logging.info("Initializing project structure in %s", directory)
os.makedirs(directory)
from django.conf.global_settings import LANGUAGES
extra = {
'build': 'build',
'default_lang': options['languages'][0],
'languages': [l for l in LANGUAGES if l[0] in options["languages"]],
'extensions': ('py', ),
'files': (),
'template': os.path.abspath(
os.path.join(
os.path.dirname(__file__),
os.pardir, os.pardir, os.pardir, 'project_template')),
}
extra.update(options)
from django.core.management import call_command
call_command('startproject', 'conf', directory, **extra)
| Create the directory before calling the startprojcet command | Create the directory before calling the startprojcet command
| Python | mit | MeirKriheli/statirator,MeirKriheli/statirator,MeirKriheli/statirator |
2d584531d043804f3dcf3acf132cb60b463e4c1a | azdweb/markdown_serv.py | azdweb/markdown_serv.py | import os
from flask import request, render_template
from azdweb import app
from azdweb.util import gh_markdown
root_path = os.path.abspath("markdown")
# {filename: (mtime, contents)}
cache = {}
def load(filename):
with open(filename) as file:
return gh_markdown.markdown(file.read())
def load_cached(filename):
mtime = os.path.getmtime(filename)
if filename in cache:
old_mtime, contents = cache[filename]
if mtime != old_mtime:
contents = load(filename)
cache[filename] = (mtime, contents)
else:
contents = load(filename)
cache[filename] = (mtime, contents)
return contents
@app.route("/md/<page>")
def serve_markdown(page):
filename = os.path.join(root_path, "{}.md".format(page))
if not os.path.exists(filename):
return render_template("markdown-404.html", page=page)
return render_template("markdown.html", page=page, content=load_cached(filename)) | import codecs
import os
from flask import render_template
from azdweb import app
from azdweb.util import gh_markdown
root_path = os.path.abspath("markdown")
# {filename: (mtime, contents)}
cache = {}
def load(filename):
with codecs.open(filename, encoding="utf-8") as file:
return gh_markdown.markdown(file.read())
def load_cached(filename):
mtime = os.path.getmtime(filename)
if filename in cache:
old_mtime, contents = cache[filename]
if mtime != old_mtime:
contents = load(filename)
cache[filename] = (mtime, contents)
else:
contents = load(filename)
cache[filename] = (mtime, contents)
return contents
@app.route("/md", defaults={"page": "index"})
@app.route("/md/<path:page>")
def serve_markdown(page):
if "." in page:
return render_template("markdown-404.html", page=page)
if page.endswith("/"):
page += "index"
filename = os.path.join(root_path, "{}.md".format(page))
if not os.path.exists(filename):
return render_template("markdown-404.html", page=page)
sidebar = os.path.join(os.path.dirname(filename), "sidebar.md")
if os.path.exists(sidebar):
sidebar_content = load_cached(sidebar)
else:
sidebar_content = ""
return render_template("markdown.html", title=page, content=load_cached(filename), sidebar=sidebar_content)
@app.route("/sw", defaults={"page": "index"})
@app.route("/sw/<path:page>")
def skywars_alias(page):
return serve_markdown("skywars/{}".format(page))
| Add support for a sidebar, and also add a /sw/ alias for /md/skywars/ | Add support for a sidebar, and also add a /sw/ alias for /md/skywars/
| Python | apache-2.0 | daboross/dabo.guru,daboross/dabo.guru,daboross/dabo.guru,daboross/dabo.guru |
a4c24372ffcbac656a9879cc2fd705d67a875a3e | prime-factors/prime_factors.py | prime-factors/prime_factors.py | # File: prime_factors.py
# Purpose: Compute the prime factors of a given natural number.
# Programmer: Amal Shehu
# Course: Exercism
# Date: Monday 26 September 2016, 12:05 AM
| # File: prime_factors.py
# Purpose: Compute the prime factors of a given natural number.
# Programmer: Amal Shehu
# Course: Exercism
# Date: Monday 26 September 2016, 12:05 AM
def prime(number):
if number <= 1:
return False
else:
if number % 1 == 0 and number % range(2, number)
| Set condition [1 is not a prime] | Set condition [1 is not a prime]
| Python | mit | amalshehu/exercism-python |
14fb663019038b80d42f212e0ad8169cd0d37e84 | neutron_lib/exceptions/address_group.py | neutron_lib/exceptions/address_group.py | # All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from neutron_lib._i18n import _
from neutron_lib import exceptions
class AddressGroupNotFound(exceptions.NotFound):
message = _("Address group %(address_group_id)s could not be found.")
class AddressesNotFound(exceptions.NotFound):
message = _("Addresses %(addresses)s not found in the address group "
"%(address_group_id)s.")
class AddressesAlreadyExist(exceptions.BadRequest):
message = _("Addresses %(addresses)s already exist in the "
"address group %(address_group_id)s.")
| # All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from neutron_lib._i18n import _
from neutron_lib import exceptions
class AddressGroupNotFound(exceptions.NotFound):
message = _("Address group %(address_group_id)s could not be found.")
class AddressGroupInUse(exceptions.InUse):
message = _("Address group %(address_group_id)s is in use on one or more "
"security group rules.")
class AddressesNotFound(exceptions.NotFound):
message = _("Addresses %(addresses)s not found in the address group "
"%(address_group_id)s.")
class AddressesAlreadyExist(exceptions.BadRequest):
message = _("Addresses %(addresses)s already exist in the "
"address group %(address_group_id)s.")
| Add address group in use exception | Add address group in use exception
Related change: https://review.opendev.org/#/c/751110/
Change-Id: I2a9872890ca4d5e59a9e266c1dcacd3488a3265c
| Python | apache-2.0 | openstack/neutron-lib,openstack/neutron-lib,openstack/neutron-lib,openstack/neutron-lib |
c6e10155743eb506c82fb578de180180eb5f1e69 | imagedownloader/stations/tests/units/test_brands.py | imagedownloader/stations/tests/units/test_brands.py | # -*- coding: utf-8 -*-
from stations.models import *
from django.test import TestCase
from datetime import datetime
import pytz
class TestBrands(TestCase):
fixtures = [ 'initial_data.yaml', '*']
def setUp(self):
self.brand = Brand.objects.filter(name = 'Kip&Zonen')[0]
def test_serialization(self):
# check if the __str__ method is defined to return the object name.
self.assertEquals(str(self.brand), self.brand.name.encode("utf-8"))
# check if the __unicode__ method is defined to return the string of bytes as a text.
self.assertEquals(unicode(self.brand), self.brand.name) | # -*- coding: utf-8 -*-
from stations.models import *
from django.test import TestCase
from datetime import datetime
import pytz
class TestBrands(TestCase):
fixtures = [ 'initial_data.yaml', '*']
def setUp(self):
self.brand = Brand.objects.filter(name = 'Kipp&Zonen')[0]
def test_serialization(self):
# check if the __str__ method is defined to return the object name.
self.assertEquals(str(self.brand), self.brand.name.encode("utf-8"))
# check if the __unicode__ method is defined to return the string of bytes as a text.
self.assertEquals(unicode(self.brand), self.brand.name) | Fix a test failure produced by stations/fixture brand change (Kipp&Zonen). | Fix a test failure produced by stations/fixture brand change (Kipp&Zonen).
| Python | mit | gersolar/solar_radiation_model,ahMarrone/solar_radiation_model,scottlittle/solar_radiation_model |
14d51aa701dcc8d1d3f026af947c935abb0eabe3 | examples/rune.py | examples/rune.py | import cassiopeia as cass
from cassiopeia.core import Summoner
def test_cass():
name = "Kalturi"
runes = cass.get_runes()
for rune in runes:
if rune.tier == 3:
print(rune.name)
if __name__ == "__main__":
test_cass()
| import cassiopeia as cass
def print_t3_runes():
for rune in cass.get_runes():
if rune.tier == 3:
print(rune.name)
if __name__ == "__main__":
print_t3_runes()
| Change function name, remove unneeded summoner name | Change function name, remove unneeded summoner name
| Python | mit | robrua/cassiopeia,10se1ucgo/cassiopeia,meraki-analytics/cassiopeia |
d72df78e0dea27ae93bde52e43cec360a963b32c | openprescribing/frontend/management/commands/delete_measure.py | openprescribing/frontend/management/commands/delete_measure.py | from django.conf import settings
from django.core.management import BaseCommand, CommandError
from frontend.models import Measure
class Command(BaseCommand):
def handle(self, measure_id, **options):
if not measure_id.startswith(settings.MEASURE_PREVIEW_PREFIX):
raise CommandError(
f"Not deleting '{measure_id}' because it doesn't look like a preview "
f"measure (it doesn't start with '{settings.MEASURE_PREVIEW_PREFIX}')"
)
try:
measure = Measure.objects.get(id=measure_id)
except Measure.DoesNotExist:
raise CommandError(f"No measure with ID '{measure_id}'")
# The ON DELETE CASCADE configuration ensures that all MeasureValues are deleted
# as well
measure.delete()
self.stdout.write(f"Deleted measure '{measure_id}'")
def add_arguments(self, parser):
parser.add_argument("measure_id")
| from django.conf import settings
from django.core.management import BaseCommand, CommandError
from frontend.models import Measure
from gcutils.bigquery import Client
class Command(BaseCommand):
def handle(self, measure_id, **options):
if not measure_id.startswith(settings.MEASURE_PREVIEW_PREFIX):
raise CommandError(
f"Not deleting '{measure_id}' because it doesn't look like a preview "
f"measure (it doesn't start with '{settings.MEASURE_PREVIEW_PREFIX}')"
)
try:
measure = Measure.objects.get(id=measure_id)
except Measure.DoesNotExist:
raise CommandError(f"No measure with ID '{measure_id}'")
delete_from_bigquery(measure_id)
# The ON DELETE CASCADE configuration ensures that all MeasureValues are deleted
# as well
measure.delete()
self.stdout.write(f"Deleted measure '{measure_id}'")
def add_arguments(self, parser):
parser.add_argument("measure_id")
def delete_from_bigquery(measure_id):
# Dataset name from `import_measures.MeasureCalculation.get_table()`
client = Client("measures")
# Table naming convention from `import_measures.MeasureCalculation.table_name()`
table_suffix = f"_data_{measure_id}"
tables_to_delete = [
table for table in client.list_tables() if table.table_id.endswith(table_suffix)
]
for table in tables_to_delete:
client.delete_table(table.table_id)
| Delete measures from BigQuery as well | Delete measures from BigQuery as well
| Python | mit | ebmdatalab/openprescribing,annapowellsmith/openpresc,ebmdatalab/openprescribing,ebmdatalab/openprescribing,annapowellsmith/openpresc,annapowellsmith/openpresc,ebmdatalab/openprescribing,annapowellsmith/openpresc |
d30eb17fbcc011291418373c1a89e508f8d1c7fe | scs_core/osio/client/topic_client.py | scs_core/osio/client/topic_client.py | """
Created on 9 Nov 2016
@author: Bruno Beloff ([email protected])
"""
from scs_core.data.json import JSONify
from scs_core.data.path_dict import PathDict
# --------------------------------------------------------------------------------------------------------------------
class TopicClient(object):
"""
classdocs
"""
__HOST = "mqtt.opensensors.io" # hard-coded URL
__TIMEOUT = 10.0
# ----------------------------------------------------------------------------------------------------------------
def __init__(self, message_client, auth):
"""
Constructor
"""
self.__message_client = message_client
self.__auth = auth
# ----------------------------------------------------------------------------------------------------------------
def connect(self):
self.__message_client.connect(TopicClient.__HOST,
self.__auth.client_id, self.__auth.user_id, self.__auth.client_password)
def disconnect(self):
self.__message_client.disconnect()
# ----------------------------------------------------------------------------------------------------------------
def publish(self, topic, datum):
datum_jstr = JSONify.dumps(datum)
success = self.__message_client.publish(topic, datum_jstr, TopicClient.__TIMEOUT)
return success
def subscribe(self, topic):
for payload_jstr in self.__message_client.subscribe(topic):
path_dict = PathDict.construct_from_jstr(payload_jstr)
yield (path_dict)
# ----------------------------------------------------------------------------------------------------------------
def __str__(self, *args, **kwargs):
return "TopicClient:{message_client:%s, auth:%s}" % (self.__message_client, self.__auth)
| """
Created on 9 Nov 2016
@author: Bruno Beloff ([email protected])
"""
from scs_core.data.json import JSONify
from scs_core.data.path_dict import PathDict
# --------------------------------------------------------------------------------------------------------------------
class TopicClient(object):
"""
classdocs
"""
__HOST = "mqtt.opensensors.io" # hard-coded URL
__TIMEOUT = 30.0
# ----------------------------------------------------------------------------------------------------------------
def __init__(self, message_client, auth):
"""
Constructor
"""
self.__message_client = message_client
self.__auth = auth
# ----------------------------------------------------------------------------------------------------------------
def connect(self):
self.__message_client.connect(TopicClient.__HOST,
self.__auth.client_id, self.__auth.user_id, self.__auth.client_password)
def disconnect(self):
self.__message_client.disconnect()
# ----------------------------------------------------------------------------------------------------------------
def publish(self, topic, datum):
datum_jstr = JSONify.dumps(datum)
success = self.__message_client.publish(topic, datum_jstr, TopicClient.__TIMEOUT)
return success
def subscribe(self, topic):
for payload_jstr in self.__message_client.subscribe(topic):
path_dict = PathDict.construct_from_jstr(payload_jstr)
yield (path_dict)
# ----------------------------------------------------------------------------------------------------------------
def __str__(self, *args, **kwargs):
return "TopicClient:{message_client:%s, auth:%s}" % (self.__message_client, self.__auth)
| Set abandoned clause in osio_topic_publisher. | Set abandoned clause in osio_topic_publisher.
| Python | mit | south-coast-science/scs_core |
9b6ff8eb88084b69190fed24de92eca31f8509d5 | palindrome-products/palindrome_products.py | palindrome-products/palindrome_products.py | def largest_palindrome():
pass
def smallest_palindrome():
pass
| from collections import defaultdict
def largest_palindrome(max_factor, min_factor=0):
return _palindromes(max_factor, min_factor, max)
def smallest_palindrome(max_factor, min_factor=0):
return _palindromes(max_factor, min_factor, min)
def _palindromes(max_factor, min_factor, minmax):
pals = defaultdict(set)
for i in range(min_factor, max_factor+1):
for j in range(min_factor, max_factor+1):
p = i * j
if is_palindrome(p):
pals[p].add(tuple(sorted([i,j])))
value = minmax(pals)
factors = pals[value]
return (value, factors)
def is_palindrome(n):
return str(n) == str(n)[::-1]
| Add an initial solution that works, but with the wrong output format | Add an initial solution that works, but with the wrong output format
| Python | agpl-3.0 | CubicComet/exercism-python-solutions |
cdcc0aa43025cace37fc1f51379928f0c1a8877d | scripts/staff_public_regs.py | scripts/staff_public_regs.py | # -*- coding: utf-8 -*-
"""Get public registrations for staff members.
python -m scripts.staff_public_regs
"""
from collections import defaultdict
import logging
from modularodm import Q
from website.models import Node, User
from website.app import init_app
logger = logging.getLogger('staff_public_regs')
STAFF_GUIDS = [
'jk5cv', # Jeff
'cdi38', # Brian
'edb8y', # Johanna
'hsey5', # Courtney
'5hdme', # Melissa
]
def main():
init_app(set_backends=True, routes=False)
staff_registrations = defaultdict(list)
users = [User.load(each) for each in STAFF_GUIDS]
for registration in Node.find(Q('is_registration', 'eq', True) & Q('is_public', 'eq', True)):
for user in users:
if registration in user.node__contributed:
staff_registrations[user._id].append(registration)
for uid in staff_registrations:
user = User.load(uid)
user_regs = staff_registrations[uid]
logger.info('{} ({}) on {} Public Registrations:'.format(
user.fullname,
user._id,
len(user_regs))
)
for registration in user_regs:
logger.info('\t{} ({}): {}'.format(registration.title,
registration._id,
registration.absolute_url)
)
if __name__ == '__main__':
main()
| # -*- coding: utf-8 -*-
"""Get public registrations for staff members.
python -m scripts.staff_public_regs
"""
from collections import defaultdict
import logging
from modularodm import Q
from website.models import Node, User
from website.app import init_app
logger = logging.getLogger('staff_public_regs')
STAFF_GUIDS = [
'jk5cv', # Jeff
'cdi38', # Brian
'edb8y', # Johanna
'hsey5', # Courtney
'5hdme', # Melissa
]
def main():
init_app(set_backends=True, routes=False)
staff_registrations = defaultdict(list)
users = [User.load(each) for each in STAFF_GUIDS]
for registration in Node.find(Q('is_registration', 'eq', True) & Q('is_public', 'eq', True)):
for user in users:
if registration in user.contributed:
staff_registrations[user._id].append(registration)
for uid in staff_registrations:
user = User.load(uid)
user_regs = staff_registrations[uid]
logger.info('{} ({}) on {} Public Registrations:'.format(
user.fullname,
user._id,
len(user_regs))
)
for registration in user_regs:
logger.info('\t{} ({}): {}'.format(registration.title,
registration._id,
registration.absolute_url)
)
if __name__ == '__main__':
main()
| Remove backref in main migration | Remove backref in main migration
| Python | apache-2.0 | billyhunt/osf.io,laurenrevere/osf.io,sloria/osf.io,brandonPurvis/osf.io,Ghalko/osf.io,chennan47/osf.io,emetsger/osf.io,hmoco/osf.io,icereval/osf.io,felliott/osf.io,DanielSBrown/osf.io,brianjgeiger/osf.io,DanielSBrown/osf.io,Nesiehr/osf.io,RomanZWang/osf.io,Johnetordoff/osf.io,alexschiller/osf.io,samchrisinger/osf.io,kch8qx/osf.io,caseyrollins/osf.io,acshi/osf.io,binoculars/osf.io,kwierman/osf.io,emetsger/osf.io,kwierman/osf.io,zachjanicki/osf.io,sloria/osf.io,doublebits/osf.io,abought/osf.io,adlius/osf.io,Johnetordoff/osf.io,mluo613/osf.io,HalcyonChimera/osf.io,SSJohns/osf.io,mluo613/osf.io,baylee-d/osf.io,aaxelb/osf.io,wearpants/osf.io,crcresearch/osf.io,brianjgeiger/osf.io,HalcyonChimera/osf.io,brandonPurvis/osf.io,CenterForOpenScience/osf.io,SSJohns/osf.io,erinspace/osf.io,sloria/osf.io,chrisseto/osf.io,hmoco/osf.io,zachjanicki/osf.io,RomanZWang/osf.io,rdhyee/osf.io,aaxelb/osf.io,emetsger/osf.io,caneruguz/osf.io,cslzchen/osf.io,aaxelb/osf.io,KAsante95/osf.io,caneruguz/osf.io,pattisdr/osf.io,brandonPurvis/osf.io,wearpants/osf.io,TomBaxter/osf.io,kwierman/osf.io,zamattiac/osf.io,kch8qx/osf.io,samchrisinger/osf.io,zamattiac/osf.io,brianjgeiger/osf.io,pattisdr/osf.io,cslzchen/osf.io,caneruguz/osf.io,adlius/osf.io,brianjgeiger/osf.io,chennan47/osf.io,cwisecarver/osf.io,rdhyee/osf.io,HalcyonChimera/osf.io,GageGaskins/osf.io,TomBaxter/osf.io,icereval/osf.io,mfraezz/osf.io,DanielSBrown/osf.io,brandonPurvis/osf.io,amyshi188/osf.io,asanfilippo7/osf.io,alexschiller/osf.io,zamattiac/osf.io,acshi/osf.io,GageGaskins/osf.io,pattisdr/osf.io,saradbowman/osf.io,kch8qx/osf.io,binoculars/osf.io,cslzchen/osf.io,erinspace/osf.io,wearpants/osf.io,mfraezz/osf.io,alexschiller/osf.io,chrisseto/osf.io,mluke93/osf.io,caseyrollins/osf.io,crcresearch/osf.io,TomHeatwole/osf.io,hmoco/osf.io,binoculars/osf.io,mattclark/osf.io,abought/osf.io,GageGaskins/osf.io,TomHeatwole/osf.io,cwisecarver/osf.io,baylee-d/osf.io,wearpants/osf.io,KAsante95/osf.io,cslzchen/osf.io,rdhyee/osf.io,zamattiac/osf.io,hmoco/osf.io,zachjanicki/osf.io,caseyrollins/osf.io,billyhunt/osf.io,cwisecarver/osf.io,SSJohns/osf.io,acshi/osf.io,abought/osf.io,samchrisinger/osf.io,chrisseto/osf.io,mattclark/osf.io,CenterForOpenScience/osf.io,emetsger/osf.io,KAsante95/osf.io,mluo613/osf.io,kch8qx/osf.io,jnayak1/osf.io,mfraezz/osf.io,Johnetordoff/osf.io,mluke93/osf.io,leb2dg/osf.io,Ghalko/osf.io,monikagrabowska/osf.io,leb2dg/osf.io,jnayak1/osf.io,Ghalko/osf.io,RomanZWang/osf.io,billyhunt/osf.io,chrisseto/osf.io,adlius/osf.io,acshi/osf.io,zachjanicki/osf.io,jnayak1/osf.io,CenterForOpenScience/osf.io,mfraezz/osf.io,TomHeatwole/osf.io,Nesiehr/osf.io,cwisecarver/osf.io,DanielSBrown/osf.io,monikagrabowska/osf.io,amyshi188/osf.io,leb2dg/osf.io,Nesiehr/osf.io,mluke93/osf.io,asanfilippo7/osf.io,monikagrabowska/osf.io,laurenrevere/osf.io,CenterForOpenScience/osf.io,doublebits/osf.io,billyhunt/osf.io,doublebits/osf.io,aaxelb/osf.io,kch8qx/osf.io,felliott/osf.io,monikagrabowska/osf.io,Johnetordoff/osf.io,erinspace/osf.io,GageGaskins/osf.io,SSJohns/osf.io,laurenrevere/osf.io,samchrisinger/osf.io,amyshi188/osf.io,RomanZWang/osf.io,mluo613/osf.io,mattclark/osf.io,abought/osf.io,asanfilippo7/osf.io,crcresearch/osf.io,alexschiller/osf.io,kwierman/osf.io,doublebits/osf.io,billyhunt/osf.io,TomHeatwole/osf.io,amyshi188/osf.io,felliott/osf.io,baylee-d/osf.io,KAsante95/osf.io,mluo613/osf.io,doublebits/osf.io,RomanZWang/osf.io,saradbowman/osf.io,monikagrabowska/osf.io,HalcyonChimera/osf.io,jnayak1/osf.io,leb2dg/osf.io,GageGaskins/osf.io,Nesiehr/osf.io,KAsante95/osf.io,adlius/osf.io,rdhyee/osf.io,brandonPurvis/osf.io,chennan47/osf.io,mluke93/osf.io,felliott/osf.io,caneruguz/osf.io,asanfilippo7/osf.io,TomBaxter/osf.io,Ghalko/osf.io,alexschiller/osf.io,icereval/osf.io,acshi/osf.io |
d0df78e9f660b138b5f79d6714312740ebcf1648 | fparser/setup.py | fparser/setup.py |
def configuration(parent_package='',top_path=None):
from numpy.distutils.misc_util import Configuration
config = Configuration('fparser',parent_package,top_path)
return config
|
def configuration(parent_package='',top_path=None):
from numpy.distutils.misc_util import Configuration
config = Configuration('fparser',parent_package,top_path)
config.add_data_files('log.config')
return config
| Add log.config to data files to fix installed fparser. | Add log.config to data files to fix installed fparser.
| Python | bsd-3-clause | pemryan/f2py,travellhyne/f2py,pearu/f2py,pemryan/f2py |
fa191537e15dd0729deb94aaa91dbb7fa9295e04 | mathdeck/loadproblem.py | mathdeck/loadproblem.py | # -*- coding: utf-8 -*-
"""
mathdeck.loadproblem
~~~~~~~~~~~~
This module loads a problem file as a module.
:copyright: (c) 2015 by Patrick Spencer.
:license: Apache 2.0, see ../LICENSE for more details.
"""
import os
import sys
# Load problem file as
def load_file_as_module(file):
"""
Load problem file as a module.
:param file: The full path to the problem file
returns a module represented by the problem file
"""
# Create a new module to hold the seed variable so
# the loaded module can reference the seed variable
if sys.version_info[0] == 2:
import imp
problem_module = imp.load_source('prob_module',file)
if sys.version_info[0] == 3:
import importlib.machinery
problem_module = importlib.machinery \
.SourceFileLoader("prob_module",file) \
.load_module()
try:
problem_module.answers
except AttributeError:
raise AttributeError('Problem file has no \'answers\' attribute')
return problem_module
| # -*- coding: utf-8 -*-
"""
mathdeck.loadproblem
~~~~~~~~~~~~~~~~~~~~
This module loads a problem file as a module.
:copyright: (c) 2015 by Patrick Spencer.
:license: Apache 2.0, see ../LICENSE for more details.
"""
import os
import sys
# Load problem file as
def load_file_as_module(file_path):
"""
Load problem file as a module.
:param file: The full path to the problem file
returns a module represented by the problem file
"""
# Create a new module to hold the seed variable so
# the loaded module can reference the seed variable
if sys.version_info[0] == 2:
import imp
problem_module = imp.load_source('prob_mod_pkg',file_path)
if sys.version_info[0] == 3:
import importlib.machinery
problem_module = importlib.machinery \
.SourceFileLoader('prob_mod_pkg',file_path) \
.load_module()
try:
problem_module.answers
except AttributeError:
raise AttributeError('Problem file has no \'answers\' attribute')
return problem_module
| Change package name in loadmodule call | Change package name in loadmodule call
Not much reason to do this. It just happened.
| Python | apache-2.0 | patrickspencer/mathdeck,patrickspencer/mathdeck |
6c39f3504dad1cf918189fd46d9e8529a2fc9586 | Functions/template-python/lambda_function.py | Functions/template-python/lambda_function.py | """Created By: Andrew Ryan DeFilippis"""
print('Lambda cold-start...')
from json import dumps, loads
# Disable 'testing_locally' when deploying to AWS Lambda
testing_locally = True
verbose = True
class CWLogs(object):
def __init__(self, context):
self.context = context
def event(self, message, event_prefix='LOG'):
print('{} RequestId: {}\t{}'.format(
event_prefix,
self.context.aws_request_id,
message
))
def lambda_handler(event, context):
log = CWLogs(context)
if verbose is True:
log.event('Event: {}'.format(dumps(event)))
log.event('Hello World!')
return None
def local_test():
import context
with open('event.json', 'r') as f:
event = loads(f.read())
print('\nFunction Log:\n')
lambda_handler(event, context)
if testing_locally is True:
local_test()
| """Created By: Andrew Ryan DeFilippis"""
print('Lambda cold-start...')
from json import dumps, loads
# Disable 'testing_locally' when deploying to AWS Lambda
testing_locally = True
verbose = True
class CWLogs(object):
"""Define the structure of log events to match all other CloudWatch Log Events logged by AWS Lambda.
"""
def __init__(self, context):
"""Define the instance of the context object.
:param context: Lambda context object
"""
self.context = context
def event(self, message, event_prefix='LOG'):
# type: (any, str) -> None
"""Print an event into the CloudWatch Logs stream for the Function's invocation.
:param message: The information to be logged (required)
:param event_prefix: The prefix that appears before the 'RequestId' (default 'LOG')
:return:
"""
print('{} RequestId: {}\t{}'.format(
event_prefix,
self.context.aws_request_id,
message
))
return None
def lambda_handler(event, context):
"""AWS Lambda executes the 'lambda_handler' function on invocation.
:param event: Ingested JSON event object provided at invocation
:param context: Lambda context object, containing information specific to the invocation and Function
:return: Final response to AWS Lambda, and passed to the invoker if the invocation type is RequestResponse
"""
# Instantiate our CloudWatch logging class
log = CWLogs(context)
if verbose is True:
log.event('Event: {}'.format(dumps(event)))
log.event('Hello World!')
return None
def local_test():
"""Testing on a local development machine (outside of AWS Lambda) is made possible by...
"""
import context
with open('event.json', 'r') as f:
event = loads(f.read())
print('\nFunction Log:\n')
lambda_handler(event, context)
if testing_locally is True:
local_test()
| Add documentation, and modify default return value | Add documentation, and modify default return value
| Python | apache-2.0 | andrewdefilippis/aws-lambda |
e6046cf33af1a4b1f16424740b5093ebd423842e | scipy/spatial/transform/__init__.py | scipy/spatial/transform/__init__.py | """
========================================================
Spatial Transformations (:mod:`scipy.spatial.transform`)
========================================================
.. currentmodule:: scipy.spatial.transform
This package implements various spatial transformations. For now,
only rotations are supported.
Rotations in 3 dimensions
-------------------------
.. autosummary::
:toctree: generated/
Rotation
"""
from __future__ import division, print_function, absolute_import
from .rotation import Rotation
__all__ = ['Rotation']
from scipy._lib._testutils import PytestTester
test = PytestTester(__name__)
del PytestTester
| """
========================================================
Spatial Transformations (:mod:`scipy.spatial.transform`)
========================================================
.. currentmodule:: scipy.spatial.transform
This package implements various spatial transformations. For now,
only rotations are supported.
Rotations in 3 dimensions
-------------------------
.. autosummary::
:toctree: generated/
Rotation
Slerp
"""
from __future__ import division, print_function, absolute_import
from .rotation import Rotation, Slerp
__all__ = ['Rotation', 'Slerp']
from scipy._lib._testutils import PytestTester
test = PytestTester(__name__)
del PytestTester
| Add Slerp class to transform package | BLD: Add Slerp class to transform package
| Python | bsd-3-clause | aarchiba/scipy,rgommers/scipy,mdhaber/scipy,grlee77/scipy,vigna/scipy,perimosocordiae/scipy,gertingold/scipy,anntzer/scipy,ilayn/scipy,anntzer/scipy,jamestwebber/scipy,person142/scipy,grlee77/scipy,gfyoung/scipy,e-q/scipy,perimosocordiae/scipy,matthew-brett/scipy,endolith/scipy,person142/scipy,scipy/scipy,jor-/scipy,ilayn/scipy,matthew-brett/scipy,arokem/scipy,jamestwebber/scipy,endolith/scipy,arokem/scipy,tylerjereddy/scipy,pizzathief/scipy,jor-/scipy,andyfaff/scipy,gfyoung/scipy,perimosocordiae/scipy,grlee77/scipy,gertingold/scipy,aeklant/scipy,person142/scipy,aarchiba/scipy,andyfaff/scipy,vigna/scipy,matthew-brett/scipy,person142/scipy,e-q/scipy,Stefan-Endres/scipy,gfyoung/scipy,zerothi/scipy,endolith/scipy,endolith/scipy,Stefan-Endres/scipy,andyfaff/scipy,matthew-brett/scipy,lhilt/scipy,mdhaber/scipy,arokem/scipy,lhilt/scipy,WarrenWeckesser/scipy,gertingold/scipy,scipy/scipy,anntzer/scipy,perimosocordiae/scipy,Stefan-Endres/scipy,tylerjereddy/scipy,e-q/scipy,jor-/scipy,mdhaber/scipy,gertingold/scipy,Eric89GXL/scipy,lhilt/scipy,jamestwebber/scipy,vigna/scipy,WarrenWeckesser/scipy,lhilt/scipy,rgommers/scipy,pizzathief/scipy,person142/scipy,jor-/scipy,scipy/scipy,ilayn/scipy,WarrenWeckesser/scipy,mdhaber/scipy,mdhaber/scipy,aeklant/scipy,Stefan-Endres/scipy,matthew-brett/scipy,Eric89GXL/scipy,WarrenWeckesser/scipy,Eric89GXL/scipy,perimosocordiae/scipy,pizzathief/scipy,pizzathief/scipy,tylerjereddy/scipy,nmayorov/scipy,Stefan-Endres/scipy,endolith/scipy,zerothi/scipy,andyfaff/scipy,perimosocordiae/scipy,WarrenWeckesser/scipy,Stefan-Endres/scipy,pizzathief/scipy,e-q/scipy,mdhaber/scipy,andyfaff/scipy,aeklant/scipy,jor-/scipy,lhilt/scipy,vigna/scipy,nmayorov/scipy,scipy/scipy,nmayorov/scipy,rgommers/scipy,Eric89GXL/scipy,arokem/scipy,e-q/scipy,grlee77/scipy,Eric89GXL/scipy,zerothi/scipy,aarchiba/scipy,gfyoung/scipy,aarchiba/scipy,anntzer/scipy,tylerjereddy/scipy,scipy/scipy,anntzer/scipy,vigna/scipy,nmayorov/scipy,tylerjereddy/scipy,grlee77/scipy,rgommers/scipy,andyfaff/scipy,zerothi/scipy,WarrenWeckesser/scipy,ilayn/scipy,Eric89GXL/scipy,ilayn/scipy,jamestwebber/scipy,aarchiba/scipy,arokem/scipy,aeklant/scipy,endolith/scipy,zerothi/scipy,aeklant/scipy,zerothi/scipy,anntzer/scipy,scipy/scipy,rgommers/scipy,ilayn/scipy,jamestwebber/scipy,gertingold/scipy,nmayorov/scipy,gfyoung/scipy |
9a543a84bce123329aec985739a44d0e230838d8 | homedisplay/info_internet_connection/management/commands/fetch_internet_information.py | homedisplay/info_internet_connection/management/commands/fetch_internet_information.py | from django.core.management.base import BaseCommand, CommandError
from django.utils.timezone import now
from info_internet_connection.models import Internet
import datetime
import huawei_b593_status
import json
import redis
#{'WIFI': 'off', 'SIG': '5', 'Mode': '4g', 'Roam': 'home', 'SIM': 'normal', 'Connect': 'connected'}
class Command(BaseCommand):
args = ''
help = 'Fetches home 4g router status information'
def handle(self, *args, **options):
r = redis.StrictRedis()
status = huawei_b593_status.HuaweiStatus()
data = status.read()
age_threshold = datetime.timedelta(minutes=2)
try:
latest_data = Internet.objects.latest()
if now() - latest_data.timestamp > age_threshold:
latest_data = Internet()
except Internet.DoesNotExist:
latest_data = Internet()
latest_data.wifi = data["WIFI"] != "off"
latest_data.signal = int(data["SIG"])
latest_data.mode = data["Mode"]
latest_data.sim = data["SIM"]
latest_data.connect_status = data["Connect"]
latest_data.update_timestamp = now()
latest_data.save()
r.publish("home:broadcast:generic", json.dumps({"key": "internet", "value": "updated"}) # TODO: send all information
| from django.core.management.base import BaseCommand, CommandError
from django.utils.timezone import now
from info_internet_connection.models import Internet
import datetime
import huawei_b593_status
import json
import redis
#{'WIFI': 'off', 'SIG': '5', 'Mode': '4g', 'Roam': 'home', 'SIM': 'normal', 'Connect': 'connected'}
class Command(BaseCommand):
args = ''
help = 'Fetches home 4g router status information'
def handle(self, *args, **options):
r = redis.StrictRedis()
status = huawei_b593_status.HuaweiStatus()
data = status.read()
age_threshold = datetime.timedelta(minutes=2)
try:
latest_data = Internet.objects.latest()
if now() - latest_data.timestamp > age_threshold:
latest_data = Internet()
except Internet.DoesNotExist:
latest_data = Internet()
latest_data.wifi = data["WIFI"] != "off"
latest_data.signal = int(data["SIG"])
latest_data.mode = data["Mode"]
latest_data.sim = data["SIM"]
latest_data.connect_status = data["Connect"]
latest_data.update_timestamp = now()
latest_data.save()
r.publish("home:broadcast:generic", json.dumps({"key": "internet", "content": "updated"}) # TODO: send all information
| Fix wrong key in websocket message | Fix wrong key in websocket message
| Python | bsd-3-clause | ojarva/home-info-display,ojarva/home-info-display,ojarva/home-info-display,ojarva/home-info-display |
f8555e70610df47b726e20f312be5beba67f8435 | mopidy/utils/process.py | mopidy/utils/process.py | import logging
import multiprocessing
from multiprocessing.reduction import reduce_connection
import pickle
import sys
from mopidy import SettingsError
logger = logging.getLogger('mopidy.utils.process')
def pickle_connection(connection):
return pickle.dumps(reduce_connection(connection))
def unpickle_connection(pickled_connection):
# From http://stackoverflow.com/questions/1446004
(func, args) = pickle.loads(pickled_connection)
return func(*args)
class BaseProcess(multiprocessing.Process):
def run(self):
logger.debug(u'%s: Starting process', self.name)
try:
self.run_inside_try()
except KeyboardInterrupt:
logger.info(u'%s: Interrupted by user', self.name)
sys.exit(0)
except SettingsError as e:
logger.error(e.message)
sys.exit(1)
except ImportError as e:
logger.error(e)
sys.exit(1)
except Exception as e:
logger.exception(e)
raise e
def run_inside_try(self):
raise NotImplementedError
| import logging
import multiprocessing
import multiprocessing.dummy
from multiprocessing.reduction import reduce_connection
import pickle
import sys
from mopidy import SettingsError
logger = logging.getLogger('mopidy.utils.process')
def pickle_connection(connection):
return pickle.dumps(reduce_connection(connection))
def unpickle_connection(pickled_connection):
# From http://stackoverflow.com/questions/1446004
(func, args) = pickle.loads(pickled_connection)
return func(*args)
class BaseProcess(multiprocessing.Process):
def run(self):
logger.debug(u'%s: Starting process', self.name)
try:
self.run_inside_try()
except KeyboardInterrupt:
logger.info(u'%s: Interrupted by user', self.name)
sys.exit(0)
except SettingsError as e:
logger.error(e.message)
sys.exit(1)
except ImportError as e:
logger.error(e)
sys.exit(1)
except Exception as e:
logger.exception(e)
raise e
def run_inside_try(self):
raise NotImplementedError
class BaseThread(multiprocessing.dummy.Process):
def run(self):
logger.debug(u'%s: Starting process', self.name)
try:
self.run_inside_try()
except KeyboardInterrupt:
logger.info(u'%s: Interrupted by user', self.name)
sys.exit(0)
except SettingsError as e:
logger.error(e.message)
sys.exit(1)
except ImportError as e:
logger.error(e)
sys.exit(1)
except Exception as e:
logger.exception(e)
raise e
def run_inside_try(self):
raise NotImplementedError
| Create BaseThread as copy of BaseProcess but with different superclass | Create BaseThread as copy of BaseProcess but with different superclass
| Python | apache-2.0 | jmarsik/mopidy,hkariti/mopidy,glogiotatidis/mopidy,dbrgn/mopidy,diandiankan/mopidy,mopidy/mopidy,abarisain/mopidy,dbrgn/mopidy,rawdlite/mopidy,tkem/mopidy,mokieyue/mopidy,priestd09/mopidy,ZenithDK/mopidy,ZenithDK/mopidy,adamcik/mopidy,quartz55/mopidy,rawdlite/mopidy,rawdlite/mopidy,woutervanwijk/mopidy,diandiankan/mopidy,glogiotatidis/mopidy,vrs01/mopidy,jmarsik/mopidy,jmarsik/mopidy,jmarsik/mopidy,bencevans/mopidy,bacontext/mopidy,ZenithDK/mopidy,tkem/mopidy,liamw9534/mopidy,ali/mopidy,mokieyue/mopidy,tkem/mopidy,abarisain/mopidy,ali/mopidy,kingosticks/mopidy,liamw9534/mopidy,pacificIT/mopidy,jodal/mopidy,bacontext/mopidy,SuperStarPL/mopidy,diandiankan/mopidy,swak/mopidy,dbrgn/mopidy,mokieyue/mopidy,mokieyue/mopidy,mopidy/mopidy,SuperStarPL/mopidy,tkem/mopidy,rawdlite/mopidy,hkariti/mopidy,vrs01/mopidy,vrs01/mopidy,jodal/mopidy,quartz55/mopidy,swak/mopidy,swak/mopidy,bacontext/mopidy,priestd09/mopidy,hkariti/mopidy,jodal/mopidy,woutervanwijk/mopidy,quartz55/mopidy,jcass77/mopidy,quartz55/mopidy,jcass77/mopidy,pacificIT/mopidy,SuperStarPL/mopidy,pacificIT/mopidy,glogiotatidis/mopidy,bencevans/mopidy,ali/mopidy,hkariti/mopidy,bencevans/mopidy,bacontext/mopidy,priestd09/mopidy,adamcik/mopidy,glogiotatidis/mopidy,SuperStarPL/mopidy,swak/mopidy,kingosticks/mopidy,pacificIT/mopidy,dbrgn/mopidy,diandiankan/mopidy,ZenithDK/mopidy,adamcik/mopidy,bencevans/mopidy,ali/mopidy,vrs01/mopidy,kingosticks/mopidy,jcass77/mopidy,mopidy/mopidy |
f9365594e415e9485dc6a4435cb14b3eaf8b9c58 | situational/settings/heroku.py | situational/settings/heroku.py | import os
from .base import *
DEBUG = False
TEMPLATE_DEBUG = DEBUG
ADMINS = (
('Sym Roe', '[email protected]'),
)
MANAGERS = ADMINS
ALLOWED_HOSTS = ['.herokuapp.com']
INTERNAL_IPS = ()
# Redirect any non-HTTP request to HTTPS
SECURE_SSL_REDIRECT = True
SECURE_PROXY_SSL_HEADER = ('HTTP_X_FORWARDED_PROTO', 'https')
# Only allow sessions when serving the site over HTTPS
SESSION_COOKIE_SECURE = True
# Only send CSRF protection cookies when serving the site over HTTPS
CSRF_COOKIE_SECURE = True
# Use the X-Request=ID HTTP Header as the request ID
LOG_REQUEST_ID_HEADER = "HTTP_X_REQUEST_ID"
import dj_database_url
DATABASES['default'] = dj_database_url.config()
DATABASES['default']['ENGINE'] = 'django_postgrespool'
REDIS_URL = os.environ['REDISCLOUD_URL']
EMAIL_HOST = os.environ['MAILGUN_SMTP_SERVER']
EMAIL_HOST_USER = os.environ['MAILGUN_SMTP_LOGIN']
EMAIL_HOST_PASSWORD = os.environ['MAILGUN_SMTP_PASSWORD']
EMAIL_PORT = os.environ['MAILGUN_SMTP_PORT']
EMAIL_USE_TLS = True
| import os
from .base import *
DEBUG = False
TEMPLATE_DEBUG = DEBUG
ADMINS = (
('Sym Roe', '[email protected]'),
)
MANAGERS = ADMINS
ALLOWED_HOSTS = ['.herokuapp.com']
INTERNAL_IPS = ()
# Redirect any non-HTTP request to HTTPS
SECURE_SSL_REDIRECT = True
SECURE_PROXY_SSL_HEADER = ('HTTP_X_FORWARDED_PROTO', 'https')
# Only allow sessions when serving the site over HTTPS
SESSION_COOKIE_SECURE = True
# Only send CSRF protection cookies when serving the site over HTTPS
CSRF_COOKIE_SECURE = True
# Use the X-Request=ID HTTP Header as the request ID
LOG_REQUEST_ID_HEADER = "HTTP_X_REQUEST_ID"
import dj_database_url
DATABASES['default'] = dj_database_url.config()
DATABASES['default']['ENGINE'] = 'django_postgrespool'
EMAIL_HOST = os.environ['MAILGUN_SMTP_SERVER']
EMAIL_HOST_USER = os.environ['MAILGUN_SMTP_LOGIN']
EMAIL_HOST_PASSWORD = os.environ['MAILGUN_SMTP_PASSWORD']
EMAIL_PORT = os.environ['MAILGUN_SMTP_PORT']
EMAIL_USE_TLS = True
| Remove RedisCloud URL form Heroku | Remove RedisCloud URL form Heroku
| Python | bsd-3-clause | lm-tools/sectors,lm-tools/sectors,lm-tools/sectors,lm-tools/sectors |
6db8a9e779031ae97977a49e4edd11d42fb6389d | samples/04_markdown_parse/epub2markdown.py | samples/04_markdown_parse/epub2markdown.py | #!/usr/bin/env python
import sys
import subprocess
import os
import os.path
from ebooklib import epub
# This is just a basic example which can easily break in real world.
if __name__ == '__main__':
# read epub
book = epub.read_epub(sys.argv[1])
# get base filename from the epub
base_name = os.path.basename(os.path.splitext(sys.argv[1])[0])
for item in book.items:
# convert into markdown if this is html
if isinstance(item, epub.EpubHtml):
proc = subprocess.Popen(['pandoc', '-f', 'html', '-t', 'markdown', '-'],
stdin=subprocess.PIPE,
stdout=subprocess.PIPE
)
content, error = proc.communicate(item.content)
file_name = os.path.splitext(item.file_name)[0]+'.md'
else:
file_name = item.file_name
content = item.content
# create needed directories
dir_name = '%s/%s' % (base_name, os.path.dirname(file_name))
if not os.path.exists(dir_name):
os.makedirs(dir_name)
print '>> ', file_name
# write content to file
f = open('%s/%s' % (base_name, file_name), 'w')
f.write(content)
f.close()
| #!/usr/bin/env python
import os.path
import subprocess
import sys
from ebooklib import epub
# This is just a basic example which can easily break in real world.
if __name__ == '__main__':
# read epub
book = epub.read_epub(sys.argv[1])
# get base filename from the epub
base_name = os.path.basename(os.path.splitext(sys.argv[1])[0])
for item in book.items:
# convert into markdown if this is html
if isinstance(item, epub.EpubHtml):
proc = subprocess.Popen(['pandoc', '-f', 'html', '-t', 'markdown', '-'],
stdin=subprocess.PIPE,
stdout=subprocess.PIPE)
content, error = proc.communicate(item.content)
file_name = os.path.splitext(item.file_name)[0] + '.md'
else:
file_name = item.file_name
content = item.content
# create needed directories
dir_name = '{0}/{1}'.format(base_name, os.path.dirname(file_name))
if not os.path.exists(dir_name):
os.makedirs(dir_name)
print('>> {0}'.format(file_name))
# write content to file
with open('{0}/{1}'.format(base_name, file_name), 'w') as f:
f.write(content)
| Make `samples/04_markdown_parse` Python 2+3 compatible | Make `samples/04_markdown_parse` Python 2+3 compatible
| Python | agpl-3.0 | booktype/ebooklib,aerkalov/ebooklib |
82696dd76351f8d0bb4fcfe9f173ada652947acc | whacked4/whacked4/ui/dialogs/errordialog.py | whacked4/whacked4/ui/dialogs/errordialog.py | #!/usr/bin/env python
#coding=utf8
"""
Error dialog interface.
"""
from whacked4.ui import windows
import wx
class ErrorDialog(windows.ErrorDialogBase):
def __init__(self, parent):
windows.ErrorDialogBase.__init__(self, parent)
wx.EndBusyCursor()
def set_log(self, log_file):
"""
Shows the log file's contents in the report field.
"""
log_file.flush()
log_file.seek(0)
self.Report.ChangeValue(log_file.read())
def copy(self, event):
self.Report.SelectAll()
self.Report.Copy()
self.Report.SetSelection(-1, -1)
def close(self, event):
self.Hide() | #!/usr/bin/env python
#coding=utf8
"""
Error dialog interface.
"""
from whacked4.ui import windows
import wx
class ErrorDialog(windows.ErrorDialogBase):
def __init__(self, parent):
windows.ErrorDialogBase.__init__(self, parent)
if wx.IsBusy() == True:
wx.EndBusyCursor()
def set_log(self, log_file):
"""
Shows the log file's contents in the report field.
"""
log_file.flush()
log_file.seek(0)
self.Report.ChangeValue(log_file.read())
def copy(self, event):
self.Report.SelectAll()
self.Report.Copy()
self.Report.SetSelection(-1, -1)
def close(self, event):
self.Hide() | Fix exceptions not displaying if a busy cursor was set. | Fix exceptions not displaying if a busy cursor was set.
| Python | bsd-2-clause | GitExl/WhackEd4,GitExl/WhackEd4 |
2575946f1b05ac1601a00f8936ab3a701b05bc7e | tests/test_utils.py | tests/test_utils.py | import unittest
from utils import TextLoader
import numpy as np
class TestUtilsMethods(unittest.TestCase):
def setUp(self):
self.data_loader = TextLoader("tests/test_data", batch_size=2, seq_length=5)
def test_init(self):
print (self.data_loader.vocab)
print (self.data_loader.tensor)
print (self.data_loader.vocab_size)
def test_build_vocab(self):
sentences = ["I", "love", "cat", "cat"]
vocab, vocab_inv = self.data_loader.build_vocab(sentences)
print (vocab, vocab_inv)
# Must include I, love, and cat
self.assertItemsEqual(vocab, ["I", "love", "cat"])
self.assertDictEqual(vocab, {'I': 0, 'love': 2, 'cat': 1})
self.assertItemsEqual(vocab_inv, ["I", "love", "cat"])
def test_batch_vocab(self):
print (np.array(self.data_loader.x_batches).shape)
self.assertItemsEqual(self.data_loader.x_batches[0][0][1:],
self.data_loader.y_batches[0][0][:-1])
self.assertItemsEqual(self.data_loader.x_batches[0][1][1:],
self.data_loader.y_batches[0][1][:-1])
if __name__ == '__main__':
unittest.main() | import unittest
from utils import TextLoader
import numpy as np
from collections import Counter
class TestUtilsMethods(unittest.TestCase):
def setUp(self):
self.data_loader = TextLoader("tests/test_data", batch_size=2, seq_length=5)
def test_init(self):
print (self.data_loader.vocab)
print (self.data_loader.tensor)
print (self.data_loader.vocab_size)
def test_build_vocab(self):
sentences = ["I", "love", "cat", "cat"]
vocab, vocab_inv = self.data_loader.build_vocab(sentences)
print (vocab, vocab_inv)
# Must include I, love, and cat
self.assertEqual(Counter(list(vocab)), Counter(list(["I", "love", "cat"])))
self.assertDictEqual(vocab, {'I': 0, 'love': 2, 'cat': 1})
self.assertEqual(Counter(list(vocab_inv)), Counter(list(["I", "love", "cat"])))
def test_batch_vocab(self):
print (np.array(self.data_loader.x_batches).shape)
self.assertEqual(Counter(list(self.data_loader.x_batches[0][0][1:])),
Counter(list(self.data_loader.y_batches[0][0][:-1])))
self.assertEqual(Counter(list(self.data_loader.x_batches[0][1][1:])),
Counter(list(self.data_loader.y_batches[0][1][:-1])))
if __name__ == '__main__':
unittest.main()
| Generalize method names to be compatible with Python 2.7 and 3.4 | Generalize method names to be compatible with Python 2.7 and 3.4
| Python | mit | hunkim/word-rnn-tensorflow,bahmanh/word-rnn-tensorflow |
dbcc7604b7ab554b7bb6000029071bfa8d2a5971 | saleor/payment/gateways/stripe/utils.py | saleor/payment/gateways/stripe/utils.py | from decimal import Decimal
# FIXME: The amount should be dynamically calculated by payment's currency.
# For example, amount will wrong for JPY since JPY does not support cents
def get_amount_for_stripe(amount, currency):
"""Get appropriate amount for stripe.
Stripe is using currency's smallest unit such as cents for USD.
Stripe requires integer instead of decimal.
"""
return int(amount * 100)
def get_amount_from_stripe(amount, currency):
"""Get appropriate amount from stripe.
Stripe is using currency's smallest unit such as cents for USD.
Saleor requires decimal instead of float or integer.
"""
return Decimal(amount / 100.0)
def get_currency_for_stripe(currency):
"""Convert Saleor's currency format to Stripe's currency format.
Stripe's currency is using lowercase while Saleor is using uppercase.
"""
return currency.lower()
def get_currency_from_stripe(currency):
"""Convert Stripe's currency format to Saleor's currency format.
Stripe's currency is using lowercase while Saleor is using uppercase.
"""
return currency.upper()
| from decimal import Decimal
# FIXME: The amount should be dynamically calculated by payment's currency.
# For example, amount will wrong for JPY since JPY does not support cents
def get_amount_for_stripe(amount, currency):
"""Get appropriate amount for stripe.
Stripe is using currency's smallest unit such as cents for USD.
Stripe requires integer instead of decimal.
"""
# Using int(Decimal) directly may yield wrong result
# such as int(Decimal(24.24)) will equal to 2423
return int((amount * 100).to_integral_value())
def get_amount_from_stripe(amount, currency):
"""Get appropriate amount from stripe.
Stripe is using currency's smallest unit such as cents for USD.
Saleor requires decimal instead of float or integer.
"""
# Using Decimal(amount / 100.0) will convert to decimal from float
# which precision may be lost
return Decimal(amount) / Decimal(100)
def get_currency_for_stripe(currency):
"""Convert Saleor's currency format to Stripe's currency format.
Stripe's currency is using lowercase while Saleor is using uppercase.
"""
return currency.lower()
def get_currency_from_stripe(currency):
"""Convert Stripe's currency format to Saleor's currency format.
Stripe's currency is using lowercase while Saleor is using uppercase.
"""
return currency.upper()
| Fix bug where amount for stripe is not correctly converted | Fix bug where amount for stripe is not correctly converted
| Python | bsd-3-clause | UITools/saleor,mociepka/saleor,maferelo/saleor,UITools/saleor,mociepka/saleor,maferelo/saleor,maferelo/saleor,UITools/saleor,UITools/saleor,UITools/saleor,mociepka/saleor |
52fc9bc79343632a034d2dc51645306f4b58210c | tests/services/conftest.py | tests/services/conftest.py | import pytest
from responses import RequestsMock
from netvisor import Netvisor
@pytest.fixture
def netvisor():
kwargs = dict(
sender='Test client',
partner_id='xxx_yyy',
partner_key='E2CEBB1966C7016730C70CA92CBB93DD',
customer_id='xx_yyyy_zz',
customer_key='7767899D6F5FB333784A2520771E5871',
organization_id='1967543-8',
language='EN'
)
return Netvisor(host='http://koulutus.netvisor.fi', **kwargs)
@pytest.yield_fixture(autouse=True)
def responses():
requests_mock = RequestsMock()
requests_mock._start()
yield requests_mock
requests_mock._stop()
requests_mock.reset()
| import pytest
from responses import RequestsMock
from netvisor import Netvisor
@pytest.fixture
def netvisor():
kwargs = dict(
sender='Test client',
partner_id='xxx_yyy',
partner_key='E2CEBB1966C7016730C70CA92CBB93DD',
customer_id='xx_yyyy_zz',
customer_key='7767899D6F5FB333784A2520771E5871',
organization_id='1967543-8',
language='EN'
)
return Netvisor(host='http://koulutus.netvisor.fi', **kwargs)
@pytest.yield_fixture(autouse=True)
def responses():
r = RequestsMock()
with r:
yield r
| Fix tests to work with responses 0.3.0 | Fix tests to work with responses 0.3.0
| Python | mit | fastmonkeys/netvisor.py |
99386fdd222988ca41d3781905cc1b185ca009f1 | plata/__init__.py | plata/__init__.py | VERSION = (1, 0, 0)
__version__ = '.'.join(map(str, VERSION))
import logging
logger = logging.getLogger('plata')
class LazySettings(object):
def _load_settings(self):
from plata import default_settings
from django.conf import settings as django_settings
for key in dir(default_settings):
if not key.startswith(('PLATA', 'CURRENCIES')):
continue
setattr(self, key, getattr(django_settings, key,
getattr(default_settings, key)))
def __getattr__(self, attr):
self._load_settings()
del self.__class__.__getattr__
return self.__dict__[attr]
settings = LazySettings()
shop_instance_cache = None
def register(instance):
logger.debug('Registering shop instance: %s' % instance)
global shop_instance_cache
shop_instance_cache = instance
def shop_instance():
"""
This method ensures that all views and URLs are properly loaded, and
returns the centrally instantiated :class:`plata.shop.views.Shop` object.
"""
if not shop_instance_cache:
# Load default URL patterns to ensure that the shop
# object has been created
from django.core.urlresolvers import get_resolver
get_resolver(None)._populate()
return shop_instance_cache
def product_model():
"""
Return the product model defined by the ``PLATA_SHOP_PRODUCT`` setting.
"""
from django.db.models import loading
return loading.get_model(*settings.PLATA_SHOP_PRODUCT.split('.'))
| VERSION = (1, 1, 0, 'pre')
__version__ = '.'.join(map(str, VERSION))
import logging
logger = logging.getLogger('plata')
class LazySettings(object):
def _load_settings(self):
from plata import default_settings
from django.conf import settings as django_settings
for key in dir(default_settings):
if not key.startswith(('PLATA', 'CURRENCIES')):
continue
setattr(self, key, getattr(django_settings, key,
getattr(default_settings, key)))
def __getattr__(self, attr):
self._load_settings()
del self.__class__.__getattr__
return self.__dict__[attr]
settings = LazySettings()
shop_instance_cache = None
def register(instance):
logger.debug('Registering shop instance: %s' % instance)
global shop_instance_cache
shop_instance_cache = instance
def shop_instance():
"""
This method ensures that all views and URLs are properly loaded, and
returns the centrally instantiated :class:`plata.shop.views.Shop` object.
"""
if not shop_instance_cache:
# Load default URL patterns to ensure that the shop
# object has been created
from django.core.urlresolvers import get_resolver
get_resolver(None)._populate()
return shop_instance_cache
def product_model():
"""
Return the product model defined by the ``PLATA_SHOP_PRODUCT`` setting.
"""
from django.db.models import loading
return loading.get_model(*settings.PLATA_SHOP_PRODUCT.split('.'))
| Change the VERSION to 1.1.0.pre | Change the VERSION to 1.1.0.pre
| Python | bsd-3-clause | armicron/plata,armicron/plata,armicron/plata,stefanklug/plata,allink/plata |
a92118d7ee6acde57ab9853186c43a5c6748e8a6 | tracpro/__init__.py | tracpro/__init__.py | from __future__ import absolute_import
# This will make sure the app is always imported when
# Django starts so that shared_task will use this app.
from .celery import app as celery_app # noqa
__version__ = "1.0.0"
| from __future__ import absolute_import
# This will make sure the app is always imported when
# Django starts so that shared_task will use this app.
from .celery import app as celery_app # noqa
VERSION = (1, 0, 0, "dev")
def get_version(version):
assert len(version) == 4, "Version must be formatted as (major, minor, micro, state)"
major, minor, micro, state = version
assert isinstance(major, int), "Major version must be an integer."
assert isinstance(minor, int), "Minor version must be an integer."
assert isinstance(micro, int), "Micro version must be an integer."
assert state in ('final', 'dev'), "State must be either final or dev."
if state == 'final':
return "{}.{}.{}".format(major, minor, micro)
else:
return "{}.{}.{}.{}".format(major, minor, micro, state)
__version__ = get_version(VERSION)
| Use tuple to represent version | Use tuple to represent version
| Python | bsd-3-clause | rapidpro/tracpro,xkmato/tracpro,xkmato/tracpro,xkmato/tracpro,xkmato/tracpro,rapidpro/tracpro,rapidpro/tracpro |
2b8674528972655937eff797e61fa6819bfc3ba8 | apps/feeds/signals.py | apps/feeds/signals.py | from libs.djpubsubhubbub.signals import updated
def update_handler(sender, update, **kwargs):
"""
Process new content being provided from SuperFeedr
"""
print sender
for entry in update.entries:
print entry
updated.connect(update_handler, dispatch_uid='superfeedr')
| from libs.djpubsubhubbub.signals import updated
from .models import Feed
def update_handler(sender, update, **kwargs):
"""
Process new content being provided from SuperFeedr
"""
print sender.topic
users = []
feeds = Feed.objects.filter(feed_url=sender.topic)
for feed in feeds:
if feed.created_by not in users:
users.append(feed.created_by)
for user in users:
kippt = user.kippt_client()
for entry in update.entries:
title = entry['title']
summary = entry['summary']
link = entry['link']
kippt.addClip(link, user.list_id, title=title, notes=summary)
updated.connect(update_handler, dispatch_uid='superfeedr')
| Handle new content from SuperFeedr to Kippt | Handle new content from SuperFeedr to Kippt | Python | mit | jpadilla/feedleap,jpadilla/feedleap |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.