lang
stringclasses 10
values | seed
stringlengths 5
2.12k
|
---|---|
python | def test_default_return_errors(self):
c = copy(CONFIG_BASE)
del c['return_error_details']
data = get_ldap(c)
self.assertEqual(data['return_error_details'], DEFAULT_RETURN_ERRORS)
"""Not specifying a base_dn should fail"""
def test_missing_base_dn(self):
c = copy(CONFIG_BASE) |
python |
def intersection(lhs: "Data", rhs: "Data", duplicates: bool = False) -> "Data":
"""(Multi)set intersection of two datasets.
This intersection method does not retain duplicates by default.
For multiset behaviour, specify 'duplicates=True'.
Parameters:
lhs: one of the two datasets to intersect ('left hand side')
rhs: one of the two datasets to intersect ('right hand side')
duplicates: if False (default), the returned data do not contain
duplicate entries; if True, duplicates are taken into account.
Both inputs and labels have to match for duplicates. |
python | elif (seconds < 26):
return "tramp "
elif (seconds < 31):
return "slut "
elif (seconds < 36):
return "bitch "
elif (seconds < 41):
return "bimbo "
|
python | ],
ids=["valid_payload", "invalid_payload", "valid_uuid", "invalid_uuid"],
)
def test_validators(
validator: Callable,
expected_output: Union[str, _client.AnyJson],
expected_exception: Exception,
exception_pattern: Optional[str],
):
if expected_exception: |
python | def setUp(self):
self.new_user = User(f_name = 'Grishon', l_name = 'Gikima', email = '<EMAIL>', password_hash = '<PASSWORD>')
self.category = Category(name = 'sports')
def test_user_init(self):
self.assertTrue(self.new_user.f_name is not None)
self.assertTrue(self.new_user.l_name is not None)
self.assertTrue(self.new_user.email is not None)
self.assertTrue(self.new_user.password_hash is not None)
self.assertEqual(self.new_user.f_name, 'Grishon')
self.assertEqual(self.new_user.l_name, 'Gikima') |
python | with open("combined.tex", "w") as combined:
with open("tex/paper.tex", "r") as paper:
for line in paper:
if line.strip()[:7] == "\\input{":
with open("tex/" + line.strip()[7:].replace("}", "").replace("\n", ""), "r") as section:
combined.writelines(section.readlines())
else:
combined.write(line.replace("../figures/", "./figures/"))
|
python |
register_event_type(IncidentSeverityUpdatedEvent)
class IncidentStatusUpdatedEvent(BaseIncidentEvent):
event_type = 'incident_status_updated' |
python | 'interfaces': {
'TenGigabitEthernet1/0/33': {
'duplex': 'a-full',
'speed': 'a-10G',
'status': 'connected',
'type': '100/1000/2.5G/5G/10GBaseTX',
'vlan': '1' |
python | if (a==b) :
print("[%d] Correct" % idx)
else :
print("[%d] Failed" % idx)
check(1, True, ocean.checkOverlap([5],[4],0,1,[4],[3],7,1))
check(2, False, ocean.checkOverlap([5],[4],0,1,[3],[3],7,1))
check(3, True, ocean.checkOverlap([5,3],[4,1],0,1,[3],[10],-3,1))
check(4, True, ocean.checkOverlap([5,3],[4,1],0,1,[2,3],[-40,10],37,1))
check(5, True, ocean.checkOverlap([3],[8],0,1,[4],[5],5,2))
|
python | DfCleaner.clean_column_names_in_place(pd.DataFrame({'seconds': [3600]}, columns=['seconds']))
def test_rename_column_names_in_place_method(self):
DfCleaner.rename_column_names_in_place(pd.DataFrame({'seconds': [3600]}, columns=['seconds']),
{'seconds': 'sec'})
def test_set_columns_to_numeric_method(self): |
python |
syntax = (
'ruby on rails',
'ruby',
'javascript',
'python',
'go',
'html',
'css',
'markdown',
'php',
'haskall', |
python | # Forwards message
with conn.Producer(routing_key="PLUGIN_ERROR") as producer:
producer.publish(newmsg, headers={"source-id": self._queue_name}, priority=message.properties.get("priority"))
# Send ack
message.ack() |
python | target = open("/data2/lbd/kitti/train/" + str(i).zfill(4) + "_td.dat", "wb")
with h5py.File(file,'r') as f:
events = f["events"]
events = rfn.unstructured_to_structured(events, dtype = dtype([('t', '<u8'), ('x', '<f4'), ('y', '<f4'), ('p', '<i4')]))
dat_events_tools.write_event_buffer(target, events) |
python | retval = resource_a.all_names(astype=list)
assert ('grassbuffers' in retval) and isinstance(retval, list)
def test_all_names_query_contains_GRASSBUFFERS_as_series(resource_a):
retval = resource_a.all_names(astype='series') |
python | from wtforms import Form, TextField, TextAreaField, StringField, SubmitField, validators
import easygui
#from identifier_retrieval import retrieve_identifier_semantics
# App config.
DEBUG = True |
python | #
# Note: The ball can only be seen if the height of the rebounding ball
# is strictly greater than the window parameter.
#
# Example:
# 1) h = 3, bounce = 0.66, window = 1.5, result is 3
# 2) h = 3, bounce = 1, window = 1.5, result is -1 (*)
# (*) Condition 2 not fulfilled.
# |
python | ]
)
}
|
python | # limitations under the License.
"""Cornell Movie Dialog Dataset."""
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import collections
import os
import re
from tensor2tensor.data_generators import dialog_abstract
from tensor2tensor.data_generators import text_encoder
from tensor2tensor.utils import registry |
python |
sys.exit()
# soil moisture data (10-minute)
obs = nc.Dataset('observations/cesar_soil_water_lb1_t10_v1.1_200607.nc')
sm_03 = obs.variables['TH03'][tidx]
sm_08 = obs.variables['TH08'][tidx]
sm_20 = obs.variables['TH20'][tidx]
# soil moisture data (daily average from prior day)
# Julian day for July 2, 2006 is 183, so day prior is 182
jdi = 181 |
python | def create_invoice(description, amount):
"""
Call strike's restful api to create a new invoice
"""
cor_id = uuid.uuid4().hex
url = "https://api.strike.me/v1/invoices" |
python | if hasattr(data, 'result') or (
isinstance(data, collections.abc.Iterable) and any([hasattr(item, 'result') for item in data])):
raise exceptions.ValueError(
'Make a Future of type NDArray instead of NDArray of type Future, or call result() first.')
if isinstance(data, (str, bytes)):
data = [data]
length = 1
else:
try:
length = len(data)
except TypeError: |
python | from time import sleep
from cyberpass import login
import cStringIO,StringIO, re
import urllib2, os, urllib
import os
import pycurl
import time
def login1(): |
python | default_method_name = f"get_{field_name}"
default_write_method_name = f"set_{field_name}"
if self.method_name is None:
self.method_name = default_method_name
if self.write_method_name is None:
self.write_method_name = default_write_method_name
super(SerializerMethodField, self).bind(field_name, parent)
def to_representation(self, value):
method = getattr(self.parent, self.method_name)
return method(value)
def to_internal_value(self, data): |
python | parser = get_parser()
options = parser.parse_args()
for pdb_file in options.pdb_files:
FILE = open(pdb_file, 'r')
lines = FILE.readlines()
FILE.close()
|
python | def __init__(self):
self.message = "Override __otp_str__ method to customize output style."
super(NoOtpStrMethod, self).__init__(self.message)
class Case:
"""A base class that TestCases will inherit that, by default it will generate a testCase only once
you can change that by changing the batch_size.
@param
"""
__case_count = 0
batch_size = 1
function: Callable[..., Any] = None
app = None |
python | # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
# THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
# FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
# DEALINGS IN THE SOFTWARE.
#
"""
def on_path(path): |
python | # Test 10: to check the .provides
provider10 = providers.Callable(Cat)
provides10: Optional[Callable[..., Cat]] = provider10.provides
assert provides10 is Cat
|
python |
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/2.2/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: path('', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: path('', Home.as_view(), name='home')
Including another URLconf
1. Import the include() function: from django.urls import include, path |
python |
def _insert_at_index(self, arr, val, index):
"""
The method is only used internally, it inserts in item in an array at a given index and shifts the
values after that index to the right.
:param list arr: list of floats representing the recommendation score.
:param float val: a float representing the recommendation value to be inserted.
:param int index: index where the value will be inserted. |
python | def exclude_pairs(self):
# exclude all pairs with dist >= 4 * tolerance
# dist_allowed is invariant under refine_adjusted_shift:
# exclude all pairs with dist_allowed >= tolerance
# if 0 continuous shifts: dist_allowed == dist:
# exclude all pairs with dist >= tolerance
timer = user_plus_sys_time()
self.add_pair_ext.next_pivot(
self.match_symmetry.continuous_shift_flags,
self.eucl_symop,
self.adjusted_shift,
flex.int(self.singles1),
flex.int(self.singles2)) |
python | parser.add_argument("field_id", help="Use a value of 'all' to delete all the custom fields for the given object")
args = parser.parse_args()
logging.basicConfig(format='%(asctime)s:%(levelname)s:%(message)s', stream=sys.stderr, level=logging.DEBUG)
logging.getLogger("requests").setLevel(logging.DEBUG) |
python | #'sphinx_toolbox.more_autodoc' test required
'sphinx.ext.autodoc',
'sphinx.ext.viewcode',
'sphinx.ext.todo',
]
templates_path = ['_templates']
language = 'en'
exclude_patterns = ['_build', 'Thumbs.db', '.DS_Store']
def skip(app, what, name, obj, would_skip, options):
if name in ( '__init__',): |
python | # -*- coding: utf-8 -*-
"""
Example usage of PyHawk.
You can point this at HAWK's server.js example node server.
Or you can point it at sample_server.py
"""
import requests
|
python | count = 0
foundids = []
tracks = app("iTunes").browser_windows[1].view.tracks
for id, track in zip(tracks.database_ID.get(), tracks.get()):
if id in foundids:
track.delete()
count += 1 |
python | def get_thumbnail_url(self, obj):
return obj.imagefile.thumb.url
def get_file_url(self, obj):
return obj.file.url
return ThisClass
ImageSerializer = basefileserializer_factory( Image ) |
python |
def __del__(self):
llbc.inl.Event_Del(self._cobj)
@property
def evid(self): |
python | from .singletonscanner import SingletonScanner
from .reaper import Reaper
from .reaperstandalone import reaperstandalone
from .prepuller import Prepuller
from .prepullerstandalone import prepullerstandalone
from .parse_args import parse_args
from .primerepocache import prime_repo_cache
__all__ = [ScanRepo, SingletonScanner, Reaper, Prepuller,
standalone, reaperstandalone, prepullerstandalone, parse_args,
prime_repo_cache]
|
python | pred_stack = np.vstack((pred(model, base),
pred(model, r90),
pred(model, r180),
pred(model, r270),
pred(model, r90_VF),
pred(model, r270_VF),
pred(model, VF),
pred(model, HF))) |
python | #!/usr/bin/env python
# -*- coding: utf-8 -*-
# IMPORTS
from sqlalchemy import Column, Integer, String
import modules.globals as sg |
python | Qt.QGraphicsItem.ItemSendsGeometryChanges |
Qt.QGraphicsItem.ItemSendsScenePositionChanges |
Qt.QGraphicsItem.ItemHasNoContents
)
self._size = Qt.QSizeF()
# Children are generally overlay items that should appear over anything else rather than z-fighting
self.setZValue(10)
@property
def size(self):
return self._size
@size.setter
def size(self, v): |
python | acqs.append(a2.toordinal())
image_pairs=[]
for im1 in acqs:
for im2 in acqs:
if im2>im1:
if (im2-im1) in [6,12,18,24,36,48]: |
python | num = int(a[aIndex]) + int(b[bIndex]) + flag
flag = num / 2
num %= 2
s = str(num) + s
aIndex -= 1
bIndex -= 1
while aIndex >= 0:
num = int(a[aIndex]) + flag
flag = num / 2
num %= 2 |
python | def upgrade():
op.alter_column(
"screenshot",
"path",
existing_type=sa.VARCHAR(length=100),
type_=sa.Unicode(length=200),
existing_nullable=False,
) |
python |
'''
file = directory_path + "ing-raw-2.0.csv"
df_ing = ing.df_extract_ing_releve(directory_path + "ing_raw.csv", directory_path + "ing-rules.csv", 0)
df_ing.to_csv(file, index=False)
df_ing_raw = pd.read_csv(
file,
encoding='latin-1',
# error_bad_lines=False,
sep=","
)
view.plot_pie(df_ing_raw, 'category') |
python | Entry point for subclassed commands to add custom arguments.
"""
pass
def get_graph(self, *args, **options):
def not_implemented():
raise NotImplementedError("You must implement {}.get_graph() method.".format(self))
return bonobo.Graph(not_implemented)
def get_services(self): |
python | level=logging.DEBUG)
console = logging.StreamHandler()
console.setLevel(logging.INFO)
return None
def parse_webanno():
grammarian = GrammarAnnotator(import_dir=args.conll)
def get_tsv():
import os
for file in os.listdir(args.webannotsv):
if file.endswith(".tsv"):
yield (os.path.join(args.webannotsv, file))
webanno_parsed = Webanno_Parser(next(get_tsv()))
grammarian.annotate(webanno_parsed)
return None |
python | <gh_stars>0
"""yatube URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/2.2/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: path('', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: path('', Home.as_view(), name='home')
Including another URLconf |
python | """Execution results from Jupyter notebooks."""
|
python |
for element in range(input_length):
total = 0
# Everything to the left of element is bound to be zero additions.
for this_digit in range(element, input_length):
total += input_list[this_digit] * pattern_pos(element_number=element + 1, position=this_digit)
last_digit = int(str(total)[-1:])
output.append(last_digit)
return output
|
python |
class BadStateException(APIException):
status_code = 400
class NothingToDoException(APIException):
status_code = 400
def unzip(filename, extract_dir):
with zipfile.ZipFile(filename) as archive:
for entry in archive.infolist(): |
python | ################################################################
# I: Importing & initialising
import pygame
pygame.init()
# D: Display configuration
# size = (640, 480)
image_size = pygame.image.load("background.jpg").get_size()
screen = pygame.display.set_mode(image_size)
pygame.display.set_caption('Pygame Example') |
python | import argparse
from crypt import crypt
chars = '0123456789abcdef'
parser = argparse.ArgumentParser(description='Generate random unique local IPv6 prefix') |
python |
app = Flask(__name__)
scraper = Scrapper()
@app.route("/")
def main():
return render_template("index.html") |
python | projects = Project.query.filter(Project.name == 'Test Project').all()
assert '422' in resp.status
assert len(projects) == 0
assert resp_data['error'] == 422
assert 'Validation error' in resp_data['message']
|
python | @staticmethod
def list():
from Msgrpc.Handle.handler import Handler
data = Xcache.list_lazyloader()
handlers = Handler.list_handler_config()
context = data_return(200, CODE_MSG.get(200), {"lazyloaders": data, "handlers": handlers})
return context
@staticmethod
def source_code():
filename = "lazyloader.zip"
lazyloader_source_code_path = os.path.join(settings.BASE_DIR, STATIC_STORE_PATH, filename) |
python | for i in range(n - 1, 0, -1):
c_storage[i] = update_c_right(c_storage[i + 1], mps_b[i], mpo_x[i], mps_a[i])
return c_storage
def reduceD_onesite(tensor_a, tensor_x, c_left, c_right):
c_left, numind_x = contract_tensors(c_left, 3, (2, ), tensor_a, 3, (0, ))
c_left, numind_x = contract_tensors(c_left, 4, (1, 3), tensor_x, 4, (0, 3)) |
python | from_city=str_schema(),
from_state=str_schema(),
to_city=str_schema(),
to_state=str_schema(),
quote=null_dict_schema(
id=num_schema(),
transaction_id=num_schema(),
vehicle_request_id=num_schema(),
user_id=num_schema(),
quantity=num_schema(gt=0, multiple_of=1), |
python |
)
urlpatterns += staticfiles_urlpatterns()
urlpatterns += patterns( |
python |
vhdl = ''
for library in self.libraries:
vhdl += library +';\n' |
python | msg = 'Available model checkpoints are:\n'
msg += '\n'.join([f'\t{c}' for c in ckpts])
print(msg)
return
if args.force:
p.cleanup_model_export()
ckpt = args.checkpoint_name or None
try:
return_code = p.export_model(ckpt)
except:
|
python | A views module using views subclassed from BakeView
"""
help = "Enter an app to bake, or no app label to bake all apps"
def add_arguments(self, parser):
parser.add_argument('app', nargs='*', type=str, default=[]) |
python | id = db.Column(db.Integer, primary_key=True)
service = db.Column(db.String(50), index=True)
username = db.Column(db.String(50), index=True)
email = db.Column(db.String(50))
password_enc = db.Column(db.String())
def __init__(self, service, username, email, password, key):
self.service = service
self.username = username
self.email = email
self.password_enc = self.encrypt(password, key)
def encrypt(self, password, key):
aes = AES.new(key, AES.MODE_ECB)
padded = appendPadding(password) |
python | '`test_project_id.dataset_id.concept` as vis_2 on '
'm.visit_source_concept_id = vis_2.concept_id ',
'destination_table_id': 'visit_occurrence', 'write_disposition': 'WRITE_TRUNCATE',
'destination_dataset_id': 'dataset_id'}]
self.chars_to_replace = '[\t\n\\s]+'
self.single_space = ' '
def test_get_fields_dict(self): |
python | def test_RESOURCE_lambda_layer_WITH_deployed_lambda_function_EXPECT_execution_successful():
"""
Test whether the layer provides necessary functionality.
:return: No return.
"""
# Create client for lambda service.
lambda_client = Credentials().boto_session.client('lambda')
# Invoke specific lambda function.
response = lambda_client.invoke( |
python | ]
operations = [
migrations.CreateModel(
name='Message', |
python | long_description_content_type="text/markdown",
license = "MIT Licence",
url = "https://github.com/Bertie97/pyctlib/tree/main/micomputing",
author = "<NAME>",
author_email = "<EMAIL>",
packages = find_packages(),
include_package_data = True, |
python |
def main():
x = np.linspace(0, 10)
y = np.random.normal(x, 0.5)
|
python |
def main():
parser = argparse.ArgumentParser(
description="Get lyrics for currently playing song on Spotify. Either --tab or --cli is required.")
parser.add_argument('-t', '--tab', action='store_true', help='Display lyrics in a browser tab.')
parser.add_argument('-c', '--cli', action='store_true', help='Display lyrics in the command-line.')
args = parser.parse_args()
if args.tab: |
python | )
tab.add_row([experiment,elen,
"%.2f"%(round(oracle_score_before*100,4)), "%.2f"%(round(oracle_score_after*100,4)), "%.2f"%(round((oracle_score_after-oracle_score_before)*100,4)),
"%.2f"%(round(full_acc_before*100,4)), "%.2f"%(round(full_acc_after*100,4)), "%.2f"%(round(( full_acc_after-full_acc_before)*100,4)), |
python | # print(Pc.size(), gt_c.size())
cd1 = CD(P1, gt_1)* 1e3
# print(P1.size(), gt_1.size())
cd2 = CD(P2, gt_2)* 1e3
# print(P2.size(), gt_2.size())
cd3 = CD(P3, gt)* 1e3
# print(P3.size(), gt.size())
loss_all = (cdc + cd1 + cd2 + cd3) * 1e3
losses = (cdc, cd1, cd2, cd3) |
python | game = GameIDSerializer(read_only=True)
pick = TeamIDSerializer(read_only=True)
class Meta:
model = UserPick
fields = ('id', 'team', 'game', 'pick', 'correct')
class UserPickPostSerializer(serializers.ModelSerializer): |
python | from share import views
urlpatterns = [
path('', views.index, name='home'),
path('share', views.share, name='share'),
path('login', views.login, name='login'),
path('register', views.register, name='register'),
path('categories', views.categories, name='categories'),
]
|
python | import time
from geopy.geocoders import Nominatim
geolocator = Nominatim()
def getGPS(place):
#location = geolocator.geocode("Acquafredda di Maratea, Italia")
location = geolocator.geocode(place)
if not location:
return 'NOT FOUND'
#print(str(place).encode("utf-8"))
#print(str(location).encode("utf-8")) |
python | # 1, 2, 3, 5, 8, 13, 21, 34, 55, 89, ...
# By considering the terms in the Fibonacci sequence whose values do not exceed four million,
# find the sum of the even-valued terms.
def solution(limit):
a = 1
b = 2
s = 0 |
python | bin edges, including the rightmost edge, allowing for non-uniform
bin widths.
(default: None, use Freedman-Diaconis rule)
""") |
python | [{'params': model_registrar.get_all_but_name_match('map_encoder').parameters()},
{'params': model_registrar.get_name_match('map_encoder').parameters(), 'lr': 0.0008}],
lr=hyperparams['learning_rate'])
# Set Learning Rate
if hyperparams['learning_rate_style'] == 'const':
lr_scheduler[node_type] = optim.lr_scheduler.ExponentialLR(optimizer[node_type], gamma=1.0)
elif hyperparams['learning_rate_style'] == 'exp':
lr_scheduler[node_type] = optim.lr_scheduler.ExponentialLR(optimizer[node_type],
gamma=hyperparams['learning_decay_rate'])
#################################
# TRAINING # |
python |
def logged_check_call(*args, **kwargs):
try:
subprocess.check_call(*args, **kwargs)
except subprocess.CalledProcessError as e:
logger.error('Command `{}` exited with return value {}\nOutput:\n{}'.format(e.cmd, e.returncode, e.output)) |
python |
def gen_crl():
with _workdir() as workdir:
with open(os.path.join(workdir, 'crlnumber'), 'w') as f:
f.write('{:02x}'.format(Certificate.get_crlnumber())) |
python | # from json import dumps
# api_addr = 'http://127.0.0.1:8000/api/'
# def save_message(data: dict):
# # data = {'message': '你好', 'customData': {'user_id': 7}, 'session_id': 'b9c6bb643e194746b278f0438215daba'}
# tail = "testSaveChat/"
# record = {
# "content": data["message"],
# "from_user_id": data["customData"]["from_user_id"],
# "session_id": data["session_id"],
# "when": time.strftime("%Y-%m-%d %H:%M:%S", time.localtime())
# } |
python | for option in self.options:
if option.name == name or option.alias == name and name:
return option.value
return None
def set(self, name, value):
name = name.upper()
for option in self.options:
if option.name == name or option.alias == name and name:
return option.set(value) |
python | dhi = self.tmy["DHI"]
albedo = 0.16 # Urban environement is 0.14 - 0.22
# POA Beam component
df_poa["E_b_poa"] = dni * np.cos(np.radians(aoi))
# POA Ground component
df_poa["E_g_poa"] = (
ghi * albedo * ((1 - np.cos(np.radians(self.surface_tilt))) / 2)
) |
python |
# template matching
res = cv2.matchTemplate(src, templ, cv2.TM_CCOEFF_NORMED)
res_norm = cv2.normalize(res, None, 0, 255, cv2.NORM_MINMAX, cv2.CV_8U)
_, maxv, _, maxloc = cv2.minMaxLoc(res)
print('maxv', maxv)
print('maxloc', maxloc)
# 매칭 결과를 빨간 사각형으로 표시
th, tw = templ.shape[:2]
dst = cv2.cvtColor(src, cv2.COLOR_GRAY2BGR) |
python | """
PyDMD init
"""
__all__ = ['dmdbase', 'dmd', 'fbdmd', 'mrdmd', 'cdmd', 'hodmd', 'dmdc',
'optdmd', 'hankeldmd']
from .meta import *
from .dmdbase import DMDBase
from .dmd import DMD
from .fbdmd import FbDMD
from .mrdmd import MrDMD
from .cdmd import CDMD
from .hankeldmd import HankelDMD |
python | # will be logged to log.error in the thrown exception
log.status.Print(error)
if 'errorDetail' in log_record:
error_detail = log_record['errorDetail']['message'].strip()
log.status.Print(error_detail) |
python |
def quit_wrapper(webdriver, current_webdriver_list):
for not_closed_webdriver in current_webdriver_list:
not_closed_webdriver.close()
webdriver.quit()
|
python | from rest_framework import serializers
from .models import ItemModel
class ItemModelSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = ItemModel
fields = ['url', 'name', 'description', 'image'] |
python | OPENAI_GPT_CONFIG = ""
convert_openai_checkpoint_to_pytorch(OPENAI_GPT_CHECKPOINT_FOLDER_PATH,
OPENAI_GPT_CONFIG,
PYTORCH_DUMP_OUTPUT)
elif sys.argv[1] == "transfo_xl":
try:
from .convert_transfo_xl_original_tf_checkpoint_to_pytorch import convert_transfo_xl_checkpoint_to_pytorch
except ImportError:
print("transformers can only be used from the commandline to convert TensorFlow models in PyTorch, "
"In that case, it requires TensorFlow to be installed. Please see " |
python | if key not in all_params:
raise TypeError(
"Got an unexpected keyword argument '%s'"
" to method get_repo_mirror_config" % key
)
params[key] = val
del params['kwargs']
# verify the required parameter 'repository' is set
if ('repository' not in params or
params['repository'] is None): |
python | return self.db.lookup_id(n3 + " ")
def count(self, triple):
s, p, o = triple
vs = (s is None) or isinstance(s, rdflib.Variable)
vp = (p is None) or isinstance(p, rdflib.Variable) |
python | Returns
-------
hrf: array of shape(length / tr * oversampling, float),
hrf sampling on the oversampled time grid
"""
dt = tr / oversampling
time_stamps = np.linspace(0, time_length, float(time_length) / dt)
time_stamps -= onset / dt |
python |
def start(self, from_run=False):
self._lock.acquire()
if from_run or self._stopped:
if not self.next_call:
self.next_call = time.time()
self.next_call += self.interval
self._timer = Timer(self.next_call - time.time(), self._run)
self._stopped = False
self._timer.start() |
python | pass
else:
response = requests.get(f'https://api.pubg.com/shards/steam/matches/{matchId}', headers=header)
r = response.json()
participantList = r["included"]
for participant in participantList:
if participant["type"] == "participant":
if participant["attributes"]["stats"]["name"] == pubgUsername:
Stats.objects.create(
matchId=matchId,
kills=participant["attributes"]["stats"]["kills"], |
python | TypeError: 'module' object is not callable
>>>
RESTART: Z:/Coding Classes/Python Text Adventures/wwif/students/simone/kjf dlscvniuekdfs.py
Warning (from warnings module):
File "C:\Users\wwclc\AppData\Local\Programs\Python\Python35\lib\getpass.py", line 101
return fallback_getpass(prompt, stream)
GetPassWarning: Can not control echo on the terminal.
Warning: Password input may be echoed.
Player 1, enter a word
|
python | """ calculates the non-living equations for DOM, POM, and nutrients """
# State variables
o2o = conc[0] # Dissolved oxygen (mg O_2 m^-3)
n3n = conc[2] # Nitrate (mmol N m^-3)
n4n = conc[3] # Ammonium (mmol N m^-3) |
python |
transparent_canvas = Image.new('RGBA', bg.size, color=(0, 0, 0, 0))
transparent_canvas.paste(cutout, offset, cutout)
offset_decal = transparent_canvas
displaced = ImageChops.multiply(offset_decal, source)
product = ImageChops.composite(displaced, source, displaced)
bg.paste(product, ((bg.width-width)//2, (bg.height-height)//2), product)
watermark = Image.open(resources_path+"watermark.png").convert("RGBA")
bg.paste(watermark, (20, bg.height-watermark.height-20), watermark)
final = bg.convert("RGB") |
python |
def test_channel_not_configured():
assert not is_channel_configured(
MagicMock(client_id=123, client_secret="secret", pool_address=None)) |
python | from keras.engine.topology import Layer
from keras import regularizers
from keras import constraints
from keras import activations
from keras import initializations |
python | shutil.copy(file_path, file_name)
except Exception:
raise Exception('Failed to prepare file for upload.')
try:
file_handler = open(file_name, 'rb')
fname = os.path.basename(file_handler.name)
encoded_file_name = base64.b64encode(fname.encode())
file_data = file_handler.read()
encoded_file_data = base64.encodebytes(file_data)
file_handler.seek(0, os.SEEK_END)
file_size = file_handler.tell()
file_handler.close()
finally: |
python | from pydantic import BaseModel
class Auth(BaseModel):
auth_type: str
class NewUser(BaseModel):
email: str
first_name: str
last_name: str |
python | # Define loss functions
def gram_matrix(x):
assert K.ndim(x) == 3
features = K.batch_flatten(x)
gram = K.dot(features, K.transpose(features))
return gram
def region_style_loss(style_image, target_image, style_mask, target_mask):
'''Calculate style loss between style_image and target_image,
for one common region specified by their (boolean) masks
'''
assert 3 == K.ndim(style_image) == K.ndim(target_image) |
Subsets and Splits