code
stringlengths
10
805k
def_use_chains
sequencelengths
0
667
#!/usr/bin/python # -*- coding: utf-8 -*- # #*** <License> ************************************************************# # This module is part of the repository CNDB. # # This module is licensed under the terms of the BSD 3-Clause License # <http://www.c-tanzer.at/license/bsd_3c.html>. # #*** </License> ***********************************************************# from _TFL.pyk import pyk from rsclib.HTML_Parse import tag, Page_Tree from rsclib.autosuper import autosuper from spider.common import Interface, Inet4, Inet6, unroutable from spider.common import WLAN_Config from spider.luci import Version_Mixin class Status (Page_Tree, Version_Mixin) : url = 'cgi-bin/luci/freifunk/status/status' retries = 2 timeout = 10 html_charset = 'utf-8' # force utf-8 encoding wl_names = dict \ ( ssid = 'ssid' , _bsiid = 'bssid' , channel = 'channel' , mode = 'mode' ) def parse (self) : root = self.tree.getroot () self.wlans = [] self.routes = {} for div in root.findall (".//%s" % tag ("div")) : id = div.get ('id') if id == 'cbi-wireless' : wlan_div = div elif id == 'cbi-routes' : route_div = div self.try_get_version (div) for d in self.tbl_iter (wlan_div) : for k, newkey in pyk.iteritems (self.wl_names) : if k in d : d [newkey] = d [k] wl = WLAN_Config (** d) self.wlans.append (wl) for d in self.tbl_iter (route_div) : iface = d.get ('iface') gw = d.get ('gateway') if iface and gw : self.routes [iface] = gw self.set_version (root) # end def parse def tbl_iter (self, div) : tbl = div.find (".//%s" % tag ("table")) assert tbl.get ('class') == 'cbi-section-table' d = {} for tr in tbl : if 'cbi-section-table-row' not in tr.get ('class').split () : continue for input in tr.findall (".//%s" % tag ('input')) : name = input.get ('id').split ('.') [-1] val = input.get ('value') d [name] = val if not d : continue yield d # end def tbl_iter # end class Status class Table_Iter (Page_Tree) : def table_iter (self) : root = self.tree.getroot () for div in root.findall (".//%s" % tag ("div")) : if div.get ('id') == 'maincontent' : break tbl = div.find (".//%s" % tag ("table")) if tbl is None : return for tr in tbl : if tr [0].tag == tag ('th') : continue yield (self.tree.get_text (x) for x in tr) # end def table_iter # end class Table_Iter class OLSR_Connections (Table_Iter) : url = 'cgi-bin/luci/freifunk/olsr/' retries = 2 timeout = 10 html_charset = 'utf-8' # force utf-8 encoding def parse (self) : self.neighbors = {} for l in self.table_iter () : neighbor, ip, lq, nlq, etx = l lq, nlq, etx = (float (x) for x in (lq, nlq, etx)) self.neighbors [neighbor] = [ip, lq, nlq, etx] # end def parse # end class OLSR_Connections class OLSR_Routes (Table_Iter) : url = 'cgi-bin/luci/freifunk/olsr/routes' retries = 2 timeout = 10 html_charset = 'utf-8' # force utf-8 encoding def parse (self) : self.iface_by_gw = {} for l in self.table_iter () : announced, gw, iface, metric, etx = l if gw in self.iface_by_gw : assert iface == self.iface_by_gw [gw] else : self.iface_by_gw [gw] = iface # end def parse # end class OLSR_Routes class OpenWRT (autosuper) : def __init__ (self, site, request) : self.site = site self.request = request if 'interfaces' in self.request or 'ips' in self.request : st = Status (site = site) conn = OLSR_Connections (site = site) route = OLSR_Routes (site = site) self.version = st.version assert len (st.wlans) <= 1 interfaces = {} ips = {} count = 0 for gw, ifname in pyk.iteritems (route.iface_by_gw) : ip, lq, nlq, etx = conn.neighbors [gw] i4 = Inet4 (ip, None, None, iface = ifname) ips [i4] = 1 is_wlan = True if lq == nlq == etx == 1.0 : is_wlan = False if ifname in interfaces : iface = interfaces [ifname] if not iface.is_wlan and is_wlan : iface.is_wlan = True iface.wlan_info = st.wlans [0] else : iface = Interface (count, ifname, None) iface.is_wlan = is_wlan if is_wlan : iface.wlan_info = st.wlans [0] count += 1 interfaces [ifname] = iface if i4 not in iface.inet4 : iface.append_inet4 (i4) wl_if = None for iface in pyk.itervalues (interfaces) : if iface.is_wlan : if wl_if : m = "Duplicate wlan: %s/%s" % (iface.name, wl_if.name) raise ValueError (m) wl_if = iface # check own ip n = 'unknown' i4 = Inet4 (self.request ['ip'], None, None, iface = n) if i4 not in ips : assert n not in interfaces iface = interfaces [n] = Interface (count, n, None) iface.append_inet4 (i4) iface.is_wlan = False if not wl_if and st.wlans : iface.is_wlan = True iface.wlan_info = st.wlans [0] ips [i4] = True self.request ['ips'] = ips self.request ['interfaces'] = interfaces self.request ['version'] = st.version # end def __init__ # end class OpenWRT
[ [ [ 400, 403 ], [ 1447, 1450 ], [ 4508, 4511 ], [ 5479, 5482 ] ], [ [ 438, 441 ], [ 1149, 1152 ], [ 1925, 1928 ], [ 2181, 2184 ], [ 2581, 2584 ], [ 2701, 2704 ], [ 2813, 2816 ] ], [ [ 443, 452 ], [ 672, 681 ], [ 2459, 2468 ] ], [ [ 486, 495 ], [ 3985, 3994 ] ], [ [ 529, 538 ], [ 5099, 5108 ], [ 5970, 5979 ] ], [ [ 540, 545 ], [ 4621, 4626 ], [ 5804, 5809 ] ], [ [ 547, 552 ] ], [ [ 554, 564 ] ], [ [ 598, 609 ], [ 1563, 1574 ] ], [ [ 643, 656 ], [ 683, 696 ] ], [ [ 664, 670 ], [ 4186, 4192 ] ], [ [ 2447, 2457 ], [ 2980, 2990 ], [ 3461, 3471 ] ], [ [ 2962, 2978 ], [ 4237, 4253 ] ], [ [ 3448, 3459 ], [ 4288, 4299 ] ], [ [ 3976, 3983 ] ] ]
# UCF Senior Design 2017-18 # Group 38 from PIL import Image import cv2 import imagehash import math import numpy as np DIFF_THRES = 20 LIMIT = 2 RESIZE = 1000 def calc_hash(img): """ Calculate the wavelet hash of the image img: (ndarray) image file """ # resize image if height > 1000 img = resize(img) return imagehash.whash(Image.fromarray(img)) def compare(hash1, hash2): """ Calculate the difference between two images hash1: (array) first wavelet hash hash2: (array) second wavelet hash """ return hash1 - hash2 def limit(img, std_hash, count): """ Determine whether image should be removed from image dictionary in main.py img: (ndarray) image file std_hash: (array) wavelet hash of comparison standard count: (int) global count of images similar to comparison standard """ # calculate hash for given image cmp_hash = calc_hash(img) # compare to standard diff = compare(std_hash, cmp_hash) # image is similar to standard if diff <= DIFF_THRES: # if there are 3 similar images already, remove image if count >= LIMIT: return 'remove' # non-similar image found else: # update comparison standard return 'update_std' # else continue reading images with same standard return 'continue' def resize(img): """ Resize an image img: (ndarray) RGB color image """ # get dimensions of image width = np.shape(img)[1] height = np.shape(img)[0] # if height of image is greater than 1000, resize it to 1000 if width > RESIZE: # keep resize proportional scale = RESIZE / width resized_img = cv2.resize( img, (RESIZE, math.floor(height / scale)), cv2.INTER_AREA) # return resized image return resized_img # if height of image is less than 1000, return image unresized return img def set_standard(images, filename): """ Set new comparison standard and update information images: (dictionary) dictionary containing all the image data filename: (String) name of the image file """ return filename, calc_hash(images[filename]), 0
[ [ [ 56, 61 ], [ 363, 368 ] ], [ [ 69, 72 ], [ 1745, 1748 ], [ 1812, 1815 ] ], [ [ 80, 89 ], [ 347, 356 ] ], [ [ 97, 101 ], [ 1783, 1787 ] ], [ [ 109, 120 ], [ 1521, 1523 ], [ 1551, 1553 ] ], [ [ 122, 132 ], [ 1073, 1083 ] ], [ [ 138, 143 ], [ 1167, 1172 ] ], [ [ 148, 154 ], [ 1649, 1655 ], [ 1708, 1714 ], [ 1775, 1781 ] ], [ [ 168, 177 ], [ 941, 950 ], [ 2219, 2228 ] ], [ [ 391, 398 ], [ 994, 1001 ] ], [ [ 594, 599 ] ], [ [ 1391, 1397 ], [ 324, 330 ] ], [ [ 1975, 1987 ] ] ]
# Copyright 2018 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from .cli.cli import main # TODO(hongyes): add more commands: # kfp compile (migrate from dsl-compile) # kfp experiment (manage experiments) if __name__ == '__main__': main()
[ [ [ 597, 601 ], [ 750, 754 ] ] ]
import time from PyQt5 import QtGui, QtCore from ui.room_item import Ui_Form from PyQt5.QtWidgets import QWidget class Room_Item(QWidget,Ui_Form): def __init__(self,parent=None,room_data=None): super(Room_Item,self).__init__(parent) self.setupUi(self) self.data = room_data self.setRoomInfo() def setRoomInfo(self): self.room_name.setText('{}({})'.format(self.data['naturalName'], self.data['roomName'])) self.description.setText("<a style='color:#BCBCBC'>{}</a>".format(self.data['description'])) timeStamp = int(self.data['creationDate']) / 1000 timeArray = time.localtime(timeStamp) otherStyleTime = time.strftime("%Y-%m-%d", timeArray) self.create_time.setText("<a style='color:#BCBCBC'>{}</a>".format(otherStyleTime)) members = len(self.data['owners']) + len(self.data['admins']) + len(self.data['members']) memberCounter = "<a style='color:#BCBCBC'>{}/{}</a>".format(members, ('∞' if self.data['maxUsers']==0 else self.data['maxUsers'])) self.member.setText(memberCounter)
[ [ [ 7, 11 ], [ 636, 640 ], [ 687, 691 ] ], [ [ 31, 36 ] ], [ [ 38, 44 ] ], [ [ 71, 78 ], [ 140, 147 ] ], [ [ 107, 114 ], [ 132, 139 ] ], [ [ 122, 131 ], [ 215, 224 ] ] ]
import asyncio import re import sys import traceback import toga from toga import Key from .keys import toga_to_winforms_key from .libs import Threading, WinForms, shcore, user32, win_version from .libs.proactor import WinformsProactorEventLoop from .window import Window class MainWindow(Window): def winforms_FormClosing(self, sender, event): if not self.interface.app._impl._is_exiting: event.Cancel = not self.interface.app.exit() class App: _MAIN_WINDOW_CLASS = MainWindow def __init__(self, interface): self.interface = interface self.interface._impl = self # Winforms app exit is tightly bound to the close of the MainWindow. # The FormClosing message on MainWindow calls app.exit(), which # will then trigger the "on_exit" handler (which might abort the # close). However, if app.exit() succeeds, it will request the # Main Window to close... which calls app.exit(). # So - we have a flag that is only ever sent once a request has been # made to exit the native app. This flag can be used to shortcut any # window-level close handling. self._is_exiting = False self.loop = WinformsProactorEventLoop() asyncio.set_event_loop(self.loop) def create(self): self.native = WinForms.Application self.app_context = WinForms.ApplicationContext() # Check the version of windows and make sure we are setting the DPI mode # with the most up to date API # Windows Versioning Check Sources : https://www.lifewire.com/windows-version-numbers-2625171 # and https://docs.microsoft.com/en-us/windows/release-information/ if win_version.Major >= 6: # Checks for Windows Vista or later # Represents Windows 8.1 up to Windows 10 before Build 1703 which should use # SetProcessDpiAwareness(True) if ((win_version.Major == 6 and win_version.Minor == 3) or (win_version.Major == 10 and win_version.Build < 15063)): shcore.SetProcessDpiAwareness(True) # Represents Windows 10 Build 1703 and beyond which should use # SetProcessDpiAwarenessContext(-2) elif win_version.Major == 10 and win_version.Build >= 15063: user32.SetProcessDpiAwarenessContext(-2) # Any other version of windows should use SetProcessDPIAware() else: user32.SetProcessDPIAware() self.native.EnableVisualStyles() self.native.SetCompatibleTextRenderingDefault(False) self.interface.commands.add( toga.Command( lambda _: self.interface.about(), 'About {}'.format(self.interface.name), group=toga.Group.HELP ), toga.Command(None, 'Preferences', group=toga.Group.FILE), # Quit should always be the last item, in a section on it's own toga.Command( lambda _: self.interface.exit(), 'Exit ' + self.interface.name, shortcut=Key.MOD_1 + 'q', group=toga.Group.FILE, section=sys.maxsize ), toga.Command( lambda _: self.interface.visit_homepage(), 'Visit homepage', enabled=self.interface.home_page is not None, group=toga.Group.HELP ) ) self._create_app_commands() # Call user code to populate the main window self.interface.startup() self.create_menus() self.interface.icon.bind(self.interface.factory) self.interface.main_window._impl.set_app(self) def create_menus(self): self._menu_items = {} self._menu_groups = {} toga.Group.FILE.order = 0 menubar = WinForms.MenuStrip() submenu = None for cmd in self.interface.commands: if cmd == toga.GROUP_BREAK: submenu = None elif cmd == toga.SECTION_BREAK: submenu.DropDownItems.Add('-') else: submenu = self._submenu(cmd.group, menubar) item = WinForms.ToolStripMenuItem(cmd.label) if cmd.action: item.Click += cmd._impl.as_handler() item.Enabled = cmd.enabled if cmd.shortcut is not None: shortcut_keys = toga_to_winforms_key(cmd.shortcut) item.ShortcutKeys = shortcut_keys item.ShowShortcutKeys = True cmd._impl.native.append(item) self._menu_items[item] = cmd submenu.DropDownItems.Add(item) self.interface.main_window._impl.native.Controls.Add(menubar) self.interface.main_window._impl.native.MainMenuStrip = menubar self.interface.main_window.content.refresh() def _submenu(self, group, menubar): try: return self._menu_groups[group] except KeyError: if group is None: submenu = menubar else: parent_menu = self._submenu(group.parent, menubar) submenu = WinForms.ToolStripMenuItem(group.label) # Top level menus are added in a different way to submenus if group.parent is None: parent_menu.Items.Add(submenu) else: parent_menu.DropDownItems.Add(submenu) self._menu_groups[group] = submenu return submenu def _create_app_commands(self): # No extra menus pass def open_document(self, fileURL): '''Add a new document to this app.''' print("STUB: If you want to handle opening documents, implement App.open_document(fileURL)") def winforms_thread_exception(self, sender, winforms_exc): # The PythonException returned by Winforms doesn't give us # easy access to the underlying Python stacktrace; so we # reconstruct it from the string message. # The Python message is helpfully included in square brackets, # as the context for the first line in the .net stack trace. # So, look for the closing bracket and the start of the Python.net # stack trace. Then, reconstruct the line breaks internal to the # remaining string. print("Traceback (most recent call last):") py_exc = winforms_exc.get_Exception() full_stack_trace = py_exc.StackTrace regex = re.compile( r"^\[(?:'(.*?)', )*(?:'(.*?)')\] (?:.*?) Python\.Runtime", re.DOTALL | re.UNICODE ) stacktrace_relevant_lines = regex.findall(full_stack_trace) if len(stacktrace_relevant_lines) == 0: self.print_stack_trace(full_stack_trace) else: for lines in stacktrace_relevant_lines: for line in lines: self.print_stack_trace(line) print(py_exc.Message) @classmethod def print_stack_trace(cls, stack_trace_line): for level in stack_trace_line.split("', '"): for line in level.split("\\n"): if line: print(line) def run_app(self): try: self.create() self.native.ThreadException += self.winforms_thread_exception self.loop.run_forever(self.app_context) except: # NOQA traceback.print_exc() def main_loop(self): thread = Threading.Thread(Threading.ThreadStart(self.run_app)) thread.SetApartmentState(Threading.ApartmentState.STA) thread.Start() thread.Join() def show_about_dialog(self): message_parts = [] if self.interface.name is not None: if self.interface.version is not None: message_parts.append( "{name} v{version}".format( name=self.interface.name, version=self.interface.version, ) ) else: message_parts.append( "{name}".format(name=self.interface.name) ) elif self.interface.version is not None: message_parts.append( "v{version}".format(version=self.interface.version) ) if self.interface.author is not None: message_parts.append( "Author: {author}".format(author=self.interface.author) ) if self.interface.description is not None: message_parts.append( "\n{description}".format( description=self.interface.description ) ) self.interface.main_window.info_dialog( 'About {}'.format(self.interface.name), "\n".join(message_parts) ) def exit(self): self._is_exiting = True self.native.Exit() def set_main_window(self, window): self.app_context.MainForm = window._impl.native def set_on_exit(self, value): pass def current_window(self): self.interface.factory.not_implemented('App.current_window()') def enter_full_screen(self, windows): self.interface.factory.not_implemented('App.enter_full_screen()') def exit_full_screen(self, windows): self.interface.factory.not_implemented('App.exit_full_screen()') def set_cursor(self, value): self.interface.factory.not_implemented('App.set_cursor()') def show_cursor(self): self.interface.factory.not_implemented('App.show_cursor()') def hide_cursor(self): self.interface.factory.not_implemented('App.hide_cursor()') def add_background_task(self, handler): self.loop.call_soon(handler, self) class DocumentApp(App): def _create_app_commands(self): self.interface.commands.add( toga.Command( lambda w: self.open_file, label='Open...', shortcut=Key.MOD_1 + 'o', group=toga.Group.FILE, section=0 ), ) def open_document(self, fileURL): """Open a new document in this app. Args: fileURL (str): The URL/path to the file to add as a document. """ self.interface.factory.not_implemented('DocumentApp.open_document()')
[ [ [ 7, 14 ], [ 1254, 1261 ] ], [ [ 22, 24 ], [ 6600, 6602 ], [ 6697, 6699 ], [ 6709, 6711 ] ], [ [ 32, 35 ], [ 3204, 3207 ] ], [ [ 43, 52 ], [ 7529, 7538 ] ], [ [ 61, 65 ], [ 2658, 2662 ], [ 2800, 2804 ], [ 2843, 2847 ], [ 2883, 2887 ], [ 2989, 2993 ], [ 3163, 3167 ], [ 3243, 3247 ], [ 3434, 3438 ], [ 3836, 3840 ], [ 3990, 3994 ], [ 4063, 4067 ], [ 10015, 10019 ], [ 10168, 10172 ] ], [ [ 83, 86 ], [ 3124, 3127 ], [ 10129, 10132 ] ], [ [ 105, 125 ], [ 4484, 4504 ] ], [ [ 145, 154 ], [ 7594, 7603 ], [ 7611, 7620 ], [ 7681, 7690 ] ], [ [ 156, 164 ], [ 1333, 1341 ], [ 1381, 1389 ], [ 3880, 3888 ], [ 4232, 4240 ], [ 5258, 5266 ] ], [ [ 166, 172 ], [ 2079, 2085 ] ], [ [ 174, 180 ], [ 2327, 2333 ], [ 2477, 2483 ] ], [ [ 182, 193 ], [ 1721, 1732 ], [ 1931, 1942 ], [ 1958, 1969 ], [ 2006, 2017 ], [ 2034, 2045 ], [ 2255, 2266 ], [ 2283, 2294 ] ], [ [ 221, 246 ], [ 1218, 1243 ] ], [ [ 267, 273 ], [ 293, 299 ] ], [ [ 282, 292 ], [ 501, 511 ] ], [ [ 471, 474 ], [ 9924, 9927 ] ], [ [ 9912, 9923 ] ] ]
# -*- coding: utf-8 -*- """ /*************************************************************************** SimplePhotogrammetryRoutePlanner A QGIS plugin A imple photogrammetry route planner. Generated by Plugin Builder: http://g-sherman.github.io/Qgis-Plugin-Builder/ ------------------- begin : 2021-04-24 copyright : (C) 2021 by Xiangyong Luo email : [email protected] git sha : $Format:%H$ ***************************************************************************/ /*************************************************************************** * * * This program is free software; you can redistribute it and/or modify * * it under the terms of the GNU General Public License as published by * * the Free Software Foundation; either version 2 of the License, or * * (at your option) any later version. * * * ***************************************************************************/ This script initializes the plugin, making it known to QGIS. """ __version__ = "0.4.0" # noinspection PyPep8Naming def classFactory(iface): # pylint: disable=invalid-name """Load SimplePhotogrammetryRoutePlanner class from file SimplePhotogrammetryRoutePlanner. :param iface: A QGIS interface instance. :type iface: QgsInterface """ # from .SimplePhotogrammetryRoutePlanner import SimplePhotogrammetryRoutePlanner return SimplePhotogrammetryRoutePlanner(iface)
[ [ [ 1304, 1315 ] ], [ [ 1359, 1371 ] ] ]
# -*- coding: utf-8 -*- """ Created on Tue Jul 24 14:38:20 2018 dimension reduction with VarianceThreshold using sklearn. Feature selector that removes all low-variance features. @author: lenovo """ from sklearn.feature_selection import VarianceThreshold import numpy as np # np.random.seed(1) X = np.random.randn(100, 10) X = np.hstack([X, np.zeros([100, 5])]) # def featureSelection_variance(X, thrd): sel = VarianceThreshold(threshold=thrd) X_selected = sel.fit_transform(X) mask = sel.get_support() return X_selected, mask X = [[0, 2, 0, 3], [0, 1, 4, 3], [0, 1, 1, 3]] selector = VarianceThreshold() selector.fit_transform(X) selector.variances_
[ [ [ 237, 254 ], [ 605, 622 ], [ 416, 433 ] ], [ [ 262, 273 ], [ 276, 278 ], [ 298, 300 ], [ 327, 329 ], [ 341, 343 ] ], [ [ 294, 295 ], [ 338, 339 ] ], [ [ 323, 324 ] ], [ [ 370, 395 ] ], [ [ 547, 548 ], [ 648, 649 ] ], [ [ 594, 602 ], [ 625, 633 ], [ 651, 659 ] ] ]
#!/usr/bin/env python # coding=utf-8 from my_multi_main3 import main import numpy as np import argparse import time parser = argparse.ArgumentParser(description='PyTorch MNIST Example') parser.add_argument('--batch-size', type=int, default=64, metavar='N', help='input batch size for training (default: 64)') parser.add_argument('--test-batch-size', type=int, default=1000, metavar='N', help='input batch size for testing (default: 1000)') parser.add_argument('--epochs', type=int, default=10, metavar='N', help='number of epochs to train (default: 10)') parser.add_argument('--lr', type=float, default=0.01, metavar='LR', help='learning rate (default: 0.01)') parser.add_argument('--momentum', type=float, default=0.5, metavar='M', help='SGD momentum (default: 0.5)') parser.add_argument('--no-cuda', action='store_true', default=False, help='disables CUDA training') parser.add_argument('--seed', type=int, default=1, metavar='S', help='random seed (default: 1)') parser.add_argument('--log-interval', type=int, default=10, metavar='N', help='how many batches to wait before logging training status') parser.add_argument('--save-model', action='store_true', default=False, help='For Saving the current Model') parser.add_argument('--norm-flag', type=bool, default=False, help='Triggering the Layer Normalization flag for attention scores') parser.add_argument('--gamma', type=float, default=None, help='Controlling the sparisty of gfusedmax/sparsemax, the smaller, the more sparse') parser.add_argument('--lam', type=float, default=1.0, help='Lambda: Controlling the smoothness of gfusedmax, the larger, the smoother') parser.add_argument('--max-type', type=str, default='softmax',choices=['softmax','sparsemax','gfusedmax'], help='mapping function in attention') parser.add_argument('--optim-type', type=str, default='SGD',choices=['SGD','Adam'], help='mapping function in attention') parser.add_argument('--head-cnt', type=int, default=2, metavar='S', choices=[1,2,4,5,10], help='Number of heads for attention (default: 1)') args = parser.parse_args() hyperparameter_choices = { 'lr':list(10**np.arange(-4,-1,0.5)), 'norm_flag': [True,False], 'gamma':list(10**np.arange(-1,3,0.5))+[None,], 'lam':list(10**np.arange(-2,2,0.5)), 'max_type':['softmax','sparsemax','gfusedmax'], # 'max_type':['sparsemax'], 'optim_type':['SGD','Adam'], 'head_cnt':[1,2,4,5,10,20] } param_num = 25 record = np.zeros([param_num,len(hyperparameter_choices)+1]) record_name = 'record3_multi_%s.csv'%time.strftime('%Y-%m-%d_%H-%M-%S',time.localtime()) for n in range(param_num): for param_index,(k,v) in enumerate(hyperparameter_choices.items()): print(param_index,k) value_index = np.random.choice(len(v)) if isinstance(v[value_index],str) or isinstance(v[value_index],bool) or v[value_index] is None: record[n,param_index] = value_index else: record[n,param_index] = v[value_index] setattr(args,k,v[value_index]) record[n,-1] = main(args) np.savetxt(record_name, record, delimiter=',')
[ [ [ 65, 69 ], [ 3317, 3321 ] ], [ [ 77, 88 ], [ 2405, 2407 ], [ 2480, 2482 ], [ 2529, 2531 ], [ 2726, 2728 ], [ 3017, 3019 ], [ 3332, 3334 ] ], [ [ 96, 104 ], [ 127, 135 ] ], [ [ 112, 116 ], [ 2815, 2819 ], [ 2849, 2853 ] ], [ [ 118, 124 ], [ 188, 194 ], [ 331, 337 ], [ 482, 488 ], [ 617, 623 ], [ 743, 749 ], [ 871, 877 ], [ 991, 997 ], [ 1108, 1114 ], [ 1265, 1271 ], [ 1394, 1400 ], [ 1544, 1550 ], [ 1707, 1713 ], [ 1863, 1869 ], [ 2028, 2034 ], [ 2170, 2176 ], [ 2339, 2345 ] ], [ [ 2332, 2336 ], [ 3275, 3279 ], [ 3322, 3326 ] ], [ [ 2360, 2382 ], [ 2750, 2772 ], [ 2933, 2955 ] ], [ [ 2702, 2711 ], [ 2736, 2745 ], [ 2882, 2891 ] ], [ [ 2717, 2723 ], [ 3158, 3164 ], [ 3220, 3226 ], [ 3302, 3308 ], [ 3356, 3362 ] ], [ [ 2778, 2789 ], [ 3343, 3354 ] ], [ [ 2871, 2872 ], [ 3165, 3166 ], [ 3227, 3228 ], [ 3309, 3310 ] ], [ [ 2902, 2913 ], [ 2980, 2991 ], [ 3167, 3178 ], [ 3229, 3240 ] ], [ [ 2915, 2916 ], [ 2992, 2993 ], [ 3280, 3281 ] ], [ [ 2917, 2918 ], [ 3038, 3039 ], [ 3064, 3065 ], [ 3098, 3099 ], [ 3122, 3123 ], [ 3244, 3245 ], [ 3282, 3283 ] ], [ [ 3003, 3014 ], [ 3066, 3077 ], [ 3100, 3111 ], [ 3124, 3135 ], [ 3182, 3193 ], [ 3246, 3257 ], [ 3284, 3295 ] ] ]
"""HDF5 related files. This file contains a set of functions that related to read and write HDF5 files. Author: Yuhuang Hu Email : [email protected] """ from __future__ import print_function, absolute_import import h5py from spiker import log logger = log.get_logger("data-hdf5", log.DEBUG) def init_hdf5(file_path, mode="w", cam_type="davis"): """Init HDF5 file object. # Parameters file_path : str absolute path for the HDF5 file. mode : str w : for writing r : for reading cam_type : str davis : for DAVIS camera dvs : for DVS camera # Returns dataset : h5py.File The file object of the given dataset """ if mode == "w": dataset = h5py.File(file_path, mode=mode) dataset.create_group("dvs") dataset.create_group("extra") if cam_type == "davis": dataset.create_group("aps") dataset.create_group("imu") elif mode == "r": dataset = h5py.File(file_path, mode=mode) return dataset
[ [ [ 181, 195 ] ], [ [ 197, 212 ] ], [ [ 221, 225 ], [ 740, 744 ], [ 998, 1002 ] ], [ [ 246, 249 ], [ 260, 263 ], [ 288, 291 ] ], [ [ 251, 257 ] ], [ [ 305, 314 ] ] ]
# automatically generated by the FlatBuffers compiler, do not modify # namespace: flatbuf import flatbuffers class FloatingPoint(object): __slots__ = ['_tab'] @classmethod def GetRootAsFloatingPoint(cls, buf, offset): n = flatbuffers.encode.Get(flatbuffers.packer.uoffset, buf, offset) x = FloatingPoint() x.Init(buf, n + offset) return x # FloatingPoint def Init(self, buf, pos): self._tab = flatbuffers.table.Table(buf, pos) # FloatingPoint def Precision(self): o = flatbuffers.number_types.UOffsetTFlags.py_type(self._tab.Offset(4)) if o != 0: return self._tab.Get(flatbuffers.number_types.Int16Flags, o + self._tab.Pos) return 0 def FloatingPointStart(builder): builder.StartObject(1) def FloatingPointAddPrecision(builder, precision): builder.PrependInt16Slot(0, precision, 0) def FloatingPointEnd(builder): return builder.EndObject()
[ [ [ 99, 110 ], [ 246, 257 ], [ 269, 280 ], [ 458, 469 ], [ 550, 561 ], [ 670, 681 ] ], [ [ 118, 131 ], [ 322, 335 ] ], [ [ 748, 766 ] ], [ [ 804, 829 ] ], [ [ 897, 913 ] ] ]
"""[Scynced Lights] Class attributes are "shared" Instance attributes are not shared. """ def sub(x, y): f class Light: pass a = Light() b = Ligth()
[ [ [ 96, 99 ] ], [ [ 121, 126 ], [ 142, 147 ] ], [ [ 138, 139 ] ], [ [ 150, 151 ] ] ]
#!/usr/bin/env python # -*- encoding: utf-8 -*- # vim: set et sw=4 ts=4 sts=4 ff=unix fenc=utf8: # Author: Binux<[email protected]> # http://binux.me # Created on 2012-11-14 17:09:50 from __future__ import unicode_literals, division, absolute_import import time import logging from collections import deque try: from UserDict import DictMixin except ImportError: from collections import Mapping as DictMixin import six from six import iteritems from six.moves import cPickle class BaseCounter(object): def __init__(self): raise NotImplementedError def event(self, value=1): """Fire a event.""" raise NotImplementedError def value(self, value): """Set counter value.""" raise NotImplementedError @property def avg(self): """Get average value""" raise NotImplementedError @property def sum(self): """Get sum of counter""" raise NotImplementedError def empty(self): """Clear counter""" raise NotImplementedError class TotalCounter(BaseCounter): """Total counter""" def __init__(self): self.cnt = 0 def event(self, value=1): self.cnt += value def value(self, value): self.cnt = value @property def avg(self): return self.cnt @property def sum(self): return self.cnt def empty(self): return self.cnt == 0 class AverageWindowCounter(BaseCounter): """ Record last N(window) value """ def __init__(self, window_size=300): self.window_size = window_size self.values = deque(maxlen=window_size) def event(self, value=1): self.values.append(value) value = event @property def avg(self): return self.sum / len(self.values) @property def sum(self): return sum(self.values) def empty(self): if not self.values: return True class TimebaseAverageWindowCounter(BaseCounter): """ Record last window_size * window_interval seconds values. records will trim evert window_interval seconds """ def __init__(self, window_size=30, window_interval=10): self.max_window_size = window_size self.window_size = 0 self.window_interval = window_interval self.values = deque(maxlen=window_size) self.times = deque(maxlen=window_size) self.cache_value = 0 self.cache_start = None self._first_data_time = None def event(self, value=1): now = time.time() if self._first_data_time is None: self._first_data_time = now if self.cache_start is None: self.cache_value = value self.cache_start = now elif now - self.cache_start > self.window_interval: self.values.append(self.cache_value) self.times.append(self.cache_start) self.on_append(self.cache_value, self.cache_start) self.cache_value = value self.cache_start = now else: self.cache_value += value return self def value(self, value): self.cache_value = value def _trim_window(self): now = time.time() if self.cache_start and now - self.cache_start > self.window_interval: self.values.append(self.cache_value) self.times.append(self.cache_start) self.on_append(self.cache_value, self.cache_start) self.cache_value = 0 self.cache_start = None if self.window_size != self.max_window_size and self._first_data_time is not None: time_passed = now - self._first_data_time self.window_size = min(self.max_window_size, time_passed / self.window_interval) window_limit = now - self.window_size * self.window_interval while self.times and self.times[0] < window_limit: self.times.popleft() self.values.popleft() @property def avg(self): sum = float(self.sum) if not self.window_size: return 0 return sum / self.window_size / self.window_interval @property def sum(self): self._trim_window() return sum(self.values) + self.cache_value def empty(self): self._trim_window() if not self.values and not self.cache_start: return True def on_append(self, value, time): pass class CounterValue(DictMixin): """ A dict like value item for CounterManager. """ def __init__(self, manager, keys): self.manager = manager self._keys = keys def __getitem__(self, key): if key == '__value__': key = self._keys return self.manager.counters[key] else: key = self._keys + (key, ) available_keys = [] for _key in self.manager.counters: if _key[:len(key)] == key: available_keys.append(_key) if len(available_keys) == 0: raise KeyError elif len(available_keys) == 1: if available_keys[0] == key: return self.manager.counters[key] else: return CounterValue(self.manager, key) else: return CounterValue(self.manager, key) def __len__(self): return len(self.keys()) def __iter__(self): return iter(self.keys()) def __contains__(self, key): return key in self.keys() def keys(self): result = set() for key in self.manager.counters: if key[:len(self._keys)] == self._keys: key = key[len(self._keys):] result.add(key[0] if key else '__value__') return result def to_dict(self, get_value=None): """Dump counters as a dict""" result = {} for key, value in iteritems(self): if isinstance(value, BaseCounter): if get_value is not None: value = getattr(value, get_value) result[key] = value else: result[key] = value.to_dict(get_value) return result class CounterManager(DictMixin): """ A dict like counter manager. When using a tuple as event key, say: ('foo', 'bar'), You can visite counter with manager['foo']['bar']. Or get all counters which first element is 'foo' by manager['foo']. It's useful for a group of counters. """ def __init__(self, cls=TimebaseAverageWindowCounter): """init manager with Counter cls""" self.cls = cls self.counters = {} def event(self, key, value=1): """Fire a event of a counter by counter key""" if isinstance(key, six.string_types): key = (key, ) assert isinstance(key, tuple), "event key type error" if key not in self.counters: self.counters[key] = self.cls() self.counters[key].event(value) return self def value(self, key, value=1): """Set value of a counter by counter key""" if isinstance(key, six.string_types): key = (key, ) assert isinstance(key, tuple), "event key type error" if key not in self.counters: self.counters[key] = self.cls() self.counters[key].value(value) return self def trim(self): """Clear not used counters""" for key, value in list(iteritems(self.counters)): if value.empty(): del self.counters[key] def __getitem__(self, key): key = (key, ) available_keys = [] for _key in self.counters: if _key[:len(key)] == key: available_keys.append(_key) if len(available_keys) == 0: raise KeyError elif len(available_keys) == 1: if available_keys[0] == key: return self.counters[key] else: return CounterValue(self, key) else: return CounterValue(self, key) def __iter__(self): return iter(self.keys()) def __len__(self): return len(self.keys()) def keys(self): result = set() for key in self.counters: result.add(key[0] if key else ()) return result def to_dict(self, get_value=None): """Dump counters as a dict""" self.trim() result = {} for key, value in iteritems(self): if isinstance(value, BaseCounter): if get_value is not None: value = getattr(value, get_value) result[key] = value else: result[key] = value.to_dict(get_value) return result def dump(self, filename): """Dump counters to file""" try: with open(filename, 'wb') as fp: cPickle.dump(self.counters, fp) except: logging.error("can't dump counter to file: %s" % filename) return False return True def load(self, filename): """Load counters to file""" try: with open(filename) as fp: self.counters = cPickle.load(fp) except: logging.debug("can't load counter from file: %s" % filename) return False return True
[ [ [ 208, 224 ] ], [ [ 226, 234 ] ], [ [ 236, 251 ] ], [ [ 260, 264 ], [ 2554, 2558 ], [ 3227, 3231 ] ], [ [ 272, 279 ], [ 8972, 8979 ], [ 9272, 9279 ] ], [ [ 304, 309 ], [ 1626, 1631 ], [ 2337, 2342 ], [ 2384, 2389 ] ], [ [ 340, 349 ], [ 4473, 4482 ], [ 6208, 6217 ] ], [ [ 398, 418 ], [ 4473, 4482 ], [ 6208, 6217 ] ], [ [ 427, 430 ], [ 6769, 6772 ], [ 7132, 7135 ] ], [ [ 447, 456 ], [ 5894, 5903 ], [ 7470, 7479 ], [ 8480, 8489 ] ], [ [ 479, 486 ], [ 8912, 8919 ], [ 9227, 9234 ] ], [ [ 495, 506 ], [ 1070, 1081 ], [ 1461, 1472 ], [ 1990, 2001 ], [ 5944, 5955 ], [ 8530, 8541 ] ], [ [ 1057, 1069 ] ], [ [ 1440, 1460 ] ], [ [ 1961, 1989 ], [ 6526, 6554 ] ], [ [ 4460, 4472 ], [ 5228, 5240 ], [ 5293, 5305 ], [ 7995, 8007 ], [ 8052, 8064 ] ], [ [ 6193, 6207 ] ] ]
# coding=utf-8 # Copyright 2021 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Feature extractor class for ViT.""" from typing import List, Optional, Union import numpy as np from PIL import Image from ...feature_extraction_utils import BatchFeature, FeatureExtractionMixin from ...file_utils import TensorType from ...image_utils import IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ImageFeatureExtractionMixin, is_torch_tensor from ...utils import logging logger = logging.get_logger(__name__) class ViTFeatureExtractor(FeatureExtractionMixin, ImageFeatureExtractionMixin): r""" Constructs a ViT feature extractor. This feature extractor inherits from :class:`~transformers.FeatureExtractionMixin` which contains most of the main methods. Users should refer to this superclass for more information regarding those methods. Args: do_resize (:obj:`bool`, `optional`, defaults to :obj:`True`): Whether to resize the input to a certain :obj:`size`. size (:obj:`int` or :obj:`Tuple(int)`, `optional`, defaults to 224): Resize the input to the given size. If a tuple is provided, it should be (width, height). If only an integer is provided, then the input will be resized to (size, size). Only has an effect if :obj:`do_resize` is set to :obj:`True`. resample (:obj:`int`, `optional`, defaults to :obj:`PIL.Image.BILINEAR`): An optional resampling filter. This can be one of :obj:`PIL.Image.NEAREST`, :obj:`PIL.Image.BOX`, :obj:`PIL.Image.BILINEAR`, :obj:`PIL.Image.HAMMING`, :obj:`PIL.Image.BICUBIC` or :obj:`PIL.Image.LANCZOS`. Only has an effect if :obj:`do_resize` is set to :obj:`True`. do_normalize (:obj:`bool`, `optional`, defaults to :obj:`True`): Whether or not to normalize the input with mean and standard deviation. image_mean (:obj:`List[int]`, defaults to :obj:`[0.5, 0.5, 0.5]`): The sequence of means for each channel, to be used when normalizing images. image_std (:obj:`List[int]`, defaults to :obj:`[0.5, 0.5, 0.5]`): The sequence of standard deviations for each channel, to be used when normalizing images. """ model_input_names = ["pixel_values"] def __init__( self, do_resize=True, size=224, resample=Image.BILINEAR, do_normalize=True, image_mean=None, image_std=None, **kwargs ): super().__init__(**kwargs) self.do_resize = do_resize self.size = size self.resample = resample self.do_normalize = do_normalize self.image_mean = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN self.image_std = image_std if image_std is not None else IMAGENET_STANDARD_STD def __call__( self, images: Union[ Image.Image, np.ndarray, "torch.Tensor", List[Image.Image], List[np.ndarray], List["torch.Tensor"] # noqa ], return_tensors: Optional[Union[str, TensorType]] = None, **kwargs ) -> BatchFeature: """ Main method to prepare for the model one or several image(s). .. warning:: NumPy arrays and PyTorch tensors are converted to PIL images when resizing, so the most efficient is to pass PIL images. Args: images (:obj:`PIL.Image.Image`, :obj:`np.ndarray`, :obj:`torch.Tensor`, :obj:`List[PIL.Image.Image]`, :obj:`List[np.ndarray]`, :obj:`List[torch.Tensor]`): The image or batch of images to be prepared. Each image can be a PIL image, NumPy array or PyTorch tensor. In case of a NumPy array/PyTorch tensor, each image should be of shape (C, H, W), where C is a number of channels, H and W are image height and width. return_tensors (:obj:`str` or :class:`~transformers.file_utils.TensorType`, `optional`, defaults to :obj:`'np'`): If set, will return tensors of a particular framework. Acceptable values are: * :obj:`'tf'`: Return TensorFlow :obj:`tf.constant` objects. * :obj:`'pt'`: Return PyTorch :obj:`torch.Tensor` objects. * :obj:`'np'`: Return NumPy :obj:`np.ndarray` objects. * :obj:`'jax'`: Return JAX :obj:`jnp.ndarray` objects. Returns: :class:`~transformers.BatchFeature`: A :class:`~transformers.BatchFeature` with the following fields: - **pixel_values** -- Pixel values to be fed to a model, of shape (batch_size, num_channels, height, width). """ # Input type checking for clearer error valid_images = False # Check that images has a valid type if isinstance(images, (Image.Image, np.ndarray)) or is_torch_tensor(images): valid_images = True elif isinstance(images, (list, tuple)): if len(images) == 0 or isinstance(images[0], (Image.Image, np.ndarray)) or is_torch_tensor(images[0]): valid_images = True if not valid_images: raise ValueError( "Images must of type `PIL.Image.Image`, `np.ndarray` or `torch.Tensor` (single example)," "`List[PIL.Image.Image]`, `List[np.ndarray]` or `List[torch.Tensor]` (batch of examples)." ) is_batched = bool( isinstance(images, (list, tuple)) and (isinstance(images[0], (Image.Image, np.ndarray)) or is_torch_tensor(images[0])) ) if not is_batched: images = [images] # transformations (resizing + normalization) if self.do_resize and self.size is not None: images = [self.resize(image=image, size=self.size, resample=self.resample) for image in images] if self.do_normalize: images = [self.normalize(image=image, mean=self.image_mean, std=self.image_std) for image in images] # return as BatchFeature data = {"pixel_values": images} encoded_inputs = BatchFeature(data=data, tensor_type=return_tensors) return encoded_inputs
[ [ [ 685, 689 ], [ 3487, 3491 ], [ 3506, 3510 ], [ 3524, 3528 ] ], [ [ 691, 699 ], [ 3588, 3596 ] ], [ [ 701, 706 ], [ 3427, 3432 ], [ 3597, 3602 ] ], [ [ 715, 726 ], [ 3459, 3461 ], [ 3511, 3513 ], [ 5368, 5370 ], [ 5560, 5562 ], [ 6054, 6056 ] ], [ [ 743, 748 ], [ 2915, 2920 ], [ 3446, 3451 ], [ 3492, 3497 ], [ 5355, 5360 ], [ 5547, 5552 ], [ 6041, 6046 ] ], [ [ 790, 802 ], [ 3655, 3667 ], [ 6623, 6635 ] ], [ [ 804, 826 ], [ 1080, 1102 ] ], [ [ 853, 863 ], [ 3608, 3618 ] ], [ [ 891, 913 ], [ 3268, 3290 ] ], [ [ 915, 936 ], [ 3356, 3377 ] ], [ [ 938, 965 ], [ 1104, 1131 ] ], [ [ 967, 982 ], [ 5384, 5399 ], [ 5576, 5591 ], [ 6070, 6085 ] ], [ [ 1004, 1011 ], [ 1023, 1030 ] ], [ [ 1014, 1020 ] ], [ [ 1060, 1079 ] ] ]
#!/usr/bin/env python3 UNKNOWN = -1 def read_val(): return int(input()) def read_row(): return list(map(int, input().split())) def read_grid(): return [read_row() for _ in range(read_val())] def make_blank_row(i): return [UNKNOWN] * i def make_blank_grid(n): return [make_blank_row(i) for i in range(1, n + 1)] def compute_max_path_sum(grid): memo = make_blank_grid(len(grid)) def dfs(i, j): if i == len(grid): return 0 if memo[i][j] == UNKNOWN: memo[i][j] = grid[i][j] + max(dfs(i + 1, j), dfs(i + 1, j + 1)) return memo[i][j] return dfs(0, 0) for t in range(read_val()): print(compute_max_path_sum(read_grid()))
[ [ [ 24, 31 ], [ 243, 250 ], [ 514, 521 ] ], [ [ 42, 50 ], [ 676, 684 ], [ 194, 202 ] ], [ [ 83, 91 ], [ 168, 176 ] ], [ [ 143, 152 ], [ 720, 729 ] ], [ [ 212, 226 ], [ 293, 307 ] ], [ [ 261, 276 ], [ 381, 396 ] ], [ [ 342, 362 ], [ 699, 719 ] ], [ [ 665, 666 ] ] ]
import platform # print(platform.system()) operating_system = platform.system().lower() if operating_system == 'darwin': from .blender_utils_macos import get_installed_blender_versions operating_system_name = 'macos' elif operating_system == 'linux': from .blender_utils_linux import get_installed_blender_versions operating_system_name = 'linux' elif operating_system == 'windows': from .blender_utils_windows import get_installed_blender_versions operating_system_name = 'windows' else: raise Exception("Unimplemented for OS {}".format(operating_system)) from .blender_utils_web import get_blender_version_download_links def find_blender(version): # TODO: add fuzzy version matching, ie. '>=2.80', '~2.80', '<2.80', etc. installed_versions = get_installed_blender_versions() if version in installed_versions: return installed_versions[version] else: print("blender version '{}' not found; found {} version(s):".format(version, len(installed_versions))) for v, path in installed_versions.items(): print(" {}: {}".format(v, path)) print("searching web archive...") versions = get_blender_version_download_links(version, operating_system_name) print("found {} download(s) for blender version '{}', platform '{}':".format(len(versions), version, operating_system_name)) for url in versions: print(" {}".format(url)) if __name__ == '__main__': for version, exec_path in get_installed_blender_versions().items(): print("found blender {version}: {path}".format(version=version, path=exec_path)) blender = find_blender('2.80') if blender: print("Found blender: '{}'".format(blender)) else: print("No matching blender version installed :(")
[ [ [ 7, 15 ], [ 64, 72 ] ], [ [ 45, 61 ], [ 93, 109 ], [ 232, 248 ], [ 370, 386 ], [ 568, 584 ] ], [ [ 160, 190 ], [ 1508, 1538 ], [ 785, 815 ] ], [ [ 195, 216 ], [ 1224, 1245 ], [ 1356, 1377 ] ], [ [ 298, 328 ], [ 1508, 1538 ], [ 785, 815 ] ], [ [ 333, 354 ], [ 1224, 1245 ], [ 1356, 1377 ] ], [ [ 440, 470 ], [ 1508, 1538 ], [ 785, 815 ] ], [ [ 475, 496 ], [ 1224, 1245 ], [ 1356, 1377 ] ], [ [ 619, 653 ], [ 1180, 1214 ] ], [ [ 660, 672 ], [ 1708, 1720 ] ], [ [ 1486, 1493 ], [ 1613, 1620 ] ], [ [ 1495, 1504 ], [ 1682, 1691 ] ], [ [ 1698, 1705 ], [ 1736, 1743 ], [ 1788, 1795 ] ] ]
import functools import random from math import cos, pi import cv2 import kornia import numpy as np import torch from kornia.augmentation import ColorJitter from data.util import read_img from PIL import Image from io import BytesIO # Get a rough visualization of the above distribution. (Y-axis is meaningless, just spreads data) from utils.util import opt_get ''' if __name__ == '__main__': import numpy as np import matplotlib.pyplot as plt data = np.asarray([get_rand() for _ in range(5000)]) plt.plot(data, np.random.uniform(size=(5000,)), 'x') plt.show() ''' def kornia_color_jitter_numpy(img, setting): if setting * 255 > 1: # I'm using Kornia's ColorJitter, which requires pytorch arrays in b,c,h,w format. img = torch.from_numpy(img).permute(2,0,1).unsqueeze(0) img = ColorJitter(setting, setting, setting, setting)(img) img = img.squeeze(0).permute(1,2,0).numpy() return img # Performs image corruption on a list of images from a configurable set of corruption # options. class ImageCorruptor: def __init__(self, opt): self.opt = opt self.reset_random() self.blur_scale = opt['corruption_blur_scale'] if 'corruption_blur_scale' in opt.keys() else 1 self.fixed_corruptions = opt['fixed_corruptions'] if 'fixed_corruptions' in opt.keys() else [] self.num_corrupts = opt['num_corrupts_per_image'] if 'num_corrupts_per_image' in opt.keys() else 0 self.cosine_bias = opt_get(opt, ['cosine_bias'], True) if self.num_corrupts == 0: return else: self.random_corruptions = opt['random_corruptions'] if 'random_corruptions' in opt.keys() else [] def reset_random(self): if 'random_seed' in self.opt.keys(): self.rand = random.Random(self.opt['random_seed']) else: self.rand = random.Random() # Feeds a random uniform through a cosine distribution to slightly bias corruptions towards "uncorrupted". # Return is on [0,1] with a bias towards 0. def get_rand(self): r = self.rand.random() if self.cosine_bias: return 1 - cos(r * pi / 2) else: return r def corrupt_images(self, imgs, return_entropy=False): if self.num_corrupts == 0 and not self.fixed_corruptions: if return_entropy: return imgs, [] else: return imgs if self.num_corrupts == 0: augmentations = [] else: augmentations = random.choices(self.random_corruptions, k=self.num_corrupts) # Sources of entropy corrupted_imgs = [] entropy = [] undo_fns = [] applied_augs = augmentations + self.fixed_corruptions for img in imgs: for aug in augmentations: r = self.get_rand() img, undo_fn = self.apply_corruption(img, aug, r, applied_augs) if undo_fn is not None: undo_fns.append(undo_fn) for aug in self.fixed_corruptions: r = self.get_rand() img, undo_fn = self.apply_corruption(img, aug, r, applied_augs) entropy.append(r) if undo_fn is not None: undo_fns.append(undo_fn) # Apply undo_fns after all corruptions are finished, in same order. for ufn in undo_fns: img = ufn(img) corrupted_imgs.append(img) if return_entropy: return corrupted_imgs, entropy else: return corrupted_imgs def apply_corruption(self, img, aug, rand_val, applied_augmentations): undo_fn = None if 'color_quantization' in aug: # Color quantization quant_div = 2 ** (int(rand_val * 10 / 3) + 2) img = img * 255 img = (img // quant_div) * quant_div img = img / 255 elif 'color_jitter' in aug: lo_end = 0 hi_end = .2 setting = rand_val * (hi_end - lo_end) + lo_end img = kornia_color_jitter_numpy(img, setting) elif 'gaussian_blur' in aug: img = cv2.GaussianBlur(img, (0,0), self.blur_scale*rand_val*1.5) elif 'motion_blur' in aug: # Motion blur intensity = self.blur_scale*rand_val * 3 + 1 angle = random.randint(0,360) k = np.zeros((intensity, intensity), dtype=np.float32) k[(intensity - 1) // 2, :] = np.ones(intensity, dtype=np.float32) k = cv2.warpAffine(k, cv2.getRotationMatrix2D((intensity / 2 - 0.5, intensity / 2 - 0.5), angle, 1.0), (intensity, intensity)) k = k * (1.0 / np.sum(k)) img = cv2.filter2D(img, -1, k) elif 'block_noise' in aug: # Large distortion blocks in part of an img, such as is used to mask out a face. pass elif 'lq_resampling' in aug: # Random mode interpolation HR->LR->HR if 'lq_resampling4x' == aug: scale = 4 else: if rand_val < .3: scale = 1 elif rand_val < .7: scale = 2 else: scale = 4 if scale > 1: interpolation_modes = [cv2.INTER_NEAREST, cv2.INTER_CUBIC, cv2.INTER_LINEAR, cv2.INTER_LANCZOS4] mode = random.randint(0,4) % len(interpolation_modes) # Downsample first, then upsample using the random mode. img = cv2.resize(img, dsize=(img.shape[1]//scale, img.shape[0]//scale), interpolation=mode) def lq_resampling_undo_fn(scale, img): return cv2.resize(img, dsize=(img.shape[1]*scale, img.shape[0]*scale), interpolation=cv2.INTER_LINEAR) undo_fn = functools.partial(lq_resampling_undo_fn, scale) elif 'color_shift' in aug: # Color shift pass elif 'interlacing' in aug: # Interlacing distortion pass elif 'chromatic_aberration' in aug: # Chromatic aberration pass elif 'noise' in aug: # Random noise if 'noise-5' == aug: noise_intensity = 5 / 255.0 else: noise_intensity = (rand_val*6) / 255.0 img += np.random.rand(*img.shape) * noise_intensity elif 'jpeg' in aug: if 'noise' not in applied_augmentations and 'noise-5' not in applied_augmentations: if aug == 'jpeg': lo=10 range=20 elif aug == 'jpeg-low': lo=15 range=10 elif aug == 'jpeg-medium': lo=23 range=25 elif aug == 'jpeg-broad': lo=15 range=60 elif aug == 'jpeg-normal': lo=47 range=35 else: raise NotImplementedError("specified jpeg corruption doesn't exist") # JPEG compression qf = (int((1-rand_val)*range) + lo) # Use PIL to perform a mock compression to a data buffer, then swap back to cv2. img = (img * 255).astype(np.uint8) img = Image.fromarray(img) buffer = BytesIO() img.save(buffer, "JPEG", quality=qf, optimize=True) buffer.seek(0) jpeg_img_bytes = np.asarray(bytearray(buffer.read()), dtype="uint8") img = read_img("buffer", jpeg_img_bytes, rgb=True) elif 'saturation' in aug: # Lightening / saturation saturation = rand_val * .3 img = np.clip(img + saturation, a_max=1, a_min=0) elif 'greyscale' in aug: img = np.tile(np.mean(img, axis=2, keepdims=True), [1,1,3]) elif 'none' not in aug: raise NotImplementedError("Augmentation doesn't exist") return img, undo_fn
[ [ [ 7, 16 ], [ 5932, 5941 ] ], [ [ 24, 30 ], [ 1805, 1811 ], [ 1882, 1888 ], [ 2559, 2565 ], [ 4420, 4426 ], [ 5500, 5506 ] ], [ [ 48, 51 ], [ 2165, 2168 ] ], [ [ 53, 55 ], [ 2173, 2175 ] ], [ [ 64, 67 ], [ 4223, 4226 ], [ 4603, 4606 ], [ 4621, 4624 ], [ 4813, 4816 ], [ 5403, 5406 ], [ 5422, 5425 ], [ 5439, 5442 ], [ 5457, 5460 ], [ 5642, 5645 ], [ 5810, 5813 ], [ 5888, 5891 ] ], [ [ 75, 81 ] ], [ [ 89, 100 ], [ 4458, 4460 ], [ 4497, 4499 ], [ 4550, 4552 ], [ 4575, 4577 ], [ 4784, 4786 ], [ 6468, 6470 ], [ 7450, 7452 ], [ 7670, 7672 ], [ 7918, 7920 ], [ 8013, 8015 ], [ 8021, 8023 ] ], [ [ 108, 113 ], [ 768, 773 ] ], [ [ 146, 157 ], [ 832, 843 ] ], [ [ 181, 189 ], [ 7744, 7752 ] ], [ [ 206, 211 ], [ 7482, 7487 ] ], [ [ 227, 234 ], [ 7528, 7535 ] ], [ [ 358, 365 ], [ 1493, 1500 ] ], [ [ 596, 621 ], [ 4128, 4153 ] ], [ [ 1057, 1071 ] ] ]
# This test requires CPython3.5 print(b"%%" % ()) print(b"=%d=" % 1) print(b"=%d=%d=" % (1, 2)) print(b"=%s=" % b"str") print(b"=%r=" % b"str") print("PASS")
[]
# # test_JpegCompression.py # import pytest import albumentations as A from .context import TfDataAugmentation as Tfda from . import test_utils from .test_utils import TestResult @pytest.mark.parametrize( "quality_lower, quality_upper, expected, message", [ # quality_lower (-1, 100, TestResult.Error, "quality_lower < min => Error"), (0, 100, TestResult.OK, "quality_lower == min => OK"), (100, 100, TestResult.OK, "quality_lower == max => OK"), (101, 100, TestResult.Error, "quality_lower >= max => Error"), # quality_upper (0, -1, TestResult.Error, "quality_upper < min => Error"), (0, 0, TestResult.OK, "quality_upper == min => OK"), (0, 100, TestResult.OK, "quality_upper == max => OK"), (0, 101, TestResult.Error, "quality_upper > max => Error"), # Relation (50, 50, TestResult.OK, "quality_lower == quality_upper => OK"), (51, 50, TestResult.Error, "quality_lower > quality_upper => Error"), ]) def test_hue_shift_limit_value( quality_lower, quality_upper, expected, message): try: Tfda.JpegCompression( quality_lower=quality_lower, quality_upper=quality_upper) actual = TestResult.OK except ValueError: actual = TestResult.Error assert expected == actual, message def test_call(): quality_lower = 50 quality_upper = 100 tgt_jpeg = Tfda.JpegCompression( quality_lower=quality_lower, quality_upper=quality_upper, p=1.0) tgt_transform = \ test_utils.make_tgt_transform(tgt_jpeg) image = test_utils.make_test_image() tgt_result = tgt_transform(image=image) actual_image = tgt_result['image'] image_np = image.numpy() quality = float(tgt_jpeg.get_param('quality')) expected_image = A.image_compression( image_np, quality, image_type='.jpg') test_utils.partial_assert_array( expected_image, actual_image, 0.6, "image", eps=0.1)
[ [ [ 38, 44 ], [ 183, 189 ] ], [ [ 52, 71 ], [ 1936, 1937 ] ], [ [ 93, 119 ], [ 1216, 1220 ], [ 1528, 1532 ] ], [ [ 134, 144 ], [ 1669, 1679 ], [ 1721, 1731 ], [ 2008, 2018 ] ], [ [ 169, 179 ], [ 307, 317 ], [ 384, 394 ], [ 458, 468 ], [ 532, 542 ], [ 634, 644 ], [ 709, 719 ], [ 781, 791 ], [ 853, 863 ], [ 950, 960 ], [ 1032, 1042 ], [ 1337, 1347 ], [ 1391, 1401 ] ], [ [ 1113, 1139 ] ], [ [ 1453, 1462 ] ] ]
import os from torch.utils.data import DataLoader from continuum.datasets import CIFAR10, InMemoryDataset from continuum.datasets import MNIST import torchvision from continuum.scenarios import TransformationIncremental import pytest import numpy as np from continuum.transforms.bg_swap import BackgroundSwap DATA_PATH = os.environ.get("CONTINUUM_DATA_PATH") # Uncomment for debugging via image output # import matplotlib.pyplot as plt def test_bg_swap_fast(): """ Fast test for background swap. """ bg_x = np.ones(shape=[2, 5, 5, 3]) * -1 bg_y = np.random.rand(2) fg = np.random.normal(loc=.5, scale=.1, size=[5, 5]) bg = InMemoryDataset(bg_x, bg_y) bg_swap = BackgroundSwap(bg, input_dim=(5, 5), normalize_bg=None) spliced_1_channel = bg_swap(fg)[:, :, 0] assert np.array_equal((spliced_1_channel <= -1), (fg <= .5)) @pytest.mark.slow def test_background_swap_numpy(): """ Test background swap on a single ndarray input. """ mnist = MNIST(DATA_PATH, download=True, train=True) cifar = CIFAR10(DATA_PATH, download=True, train=True) bg_swap = BackgroundSwap(cifar, input_dim=(28, 28)) im = mnist.get_data()[0][0] im = bg_swap(im) # Uncomment for debugging # plt.imshow(im, interpolation='nearest') # plt.show() @pytest.mark.slow def test_background_swap_torch(): """ Test background swap on a single tensor input. """ cifar = CIFAR10(DATA_PATH, download=True, train=True) mnist = torchvision.datasets.MNIST(DATA_PATH, train=True, download=True, transform=torchvision.transforms.Compose([ torchvision.transforms.ToTensor() ])) bg_swap = BackgroundSwap(cifar, input_dim=(28, 28)) im = mnist[0][0] im = bg_swap(im) # Uncomment for debugging # plt.imshow(im.permute(1, 2, 0), interpolation='nearest') # plt.show() @pytest.mark.slow def test_background_tranformation(): """ Example code using TransformationIncremental to create a setting with 3 tasks. """ cifar = CIFAR10(DATA_PATH, train=True) mnist = MNIST(DATA_PATH, download=False, train=True) nb_task = 3 list_trsf = [] for i in range(nb_task): list_trsf.append([torchvision.transforms.ToTensor(), BackgroundSwap(cifar, bg_label=i, input_dim=(28, 28)), torchvision.transforms.ToPILImage()]) scenario = TransformationIncremental(mnist, base_transformations=[torchvision.transforms.ToTensor()], incremental_transformations=list_trsf) folder = "tests/samples/background_trsf/" if not os.path.exists(folder): os.makedirs(folder) for task_id, task_data in enumerate(scenario): task_data.plot(path=folder, title=f"background_{task_id}.jpg", nb_samples=100, shape=[28, 28, 3]) loader = DataLoader(task_data) _, _, _ = next(iter(loader))
[ [ [ 7, 9 ], [ 324, 326 ], [ 2723, 2725 ], [ 2755, 2757 ] ], [ [ 40, 50 ], [ 2949, 2959 ] ], [ [ 82, 89 ], [ 1059, 1066 ], [ 1443, 1450 ], [ 2148, 2155 ] ], [ [ 91, 106 ], [ 658, 673 ] ], [ [ 138, 143 ], [ 1003, 1008 ], [ 2191, 2196 ] ], [ [ 151, 162 ], [ 1502, 1513 ], [ 1616, 1627 ], [ 1692, 1703 ], [ 2326, 2337 ], [ 2442, 2453 ], [ 2550, 2561 ] ], [ [ 195, 220 ], [ 2495, 2520 ] ], [ [ 228, 234 ], [ 872, 878 ], [ 1313, 1319 ], [ 1983, 1989 ] ], [ [ 242, 253 ], [ 529, 531 ], [ 573, 575 ], [ 601, 603 ], [ 815, 817 ] ], [ [ 296, 310 ], [ 701, 715 ], [ 1120, 1134 ], [ 1784, 1798 ], [ 2361, 2375 ] ], [ [ 312, 321 ], [ 1009, 1018 ], [ 1067, 1076 ], [ 1451, 1460 ], [ 1529, 1538 ], [ 2156, 2165 ], [ 2197, 2206 ] ], [ [ 446, 463 ] ], [ [ 893, 919 ] ], [ [ 1334, 1360 ] ], [ [ 2004, 2033 ] ] ]
# ========================================================================================= # Copyright 2015 Community Information Online Consortium (CIOC) and KCL Software Solutions # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ========================================================================================= # std lib import os # jQuery and jQueryUI versions JQUERY_VERSION = "1.6.2" JQUERY_UI_VERSION = "1.8.16" # formatting constants DATE_TEXT_SIZE = 25 TEXT_SIZE = 85 TEXTAREA_COLS = 85 TEXTAREA_ROWS_SHORT = 2 TEXTAREA_ROWS_LONG = 4 TEXTAREA_ROWS_XLONG = 10 MAX_LENGTH_CHECKLIST_NOTES = 255 EMAIL_LENGTH = 60 # application running constants _app_path = None _config_file = None _app_name = None session_lock_dir = None publish_dir = None def update_cache_values(): # called from application init at startup global _app_path, _config_file, _app_name, session_lock_dir, publish_dir if _app_path is None: _app_path = os.path.normpath(os.path.join(os.path.dirname(__file__), '..', '..', '..')) _app_name = os.path.split(_app_path)[1] _config_file = os.path.join(_app_path, '..', '..', 'config', _app_name + '.ini') session_lock_dir = os.path.join(_app_path, 'python', 'session_lock') publish_dir = os.path.join(_app_path, 'python', 'published_files') try: os.makedirs(session_lock_dir) except os.error: pass try: os.makedirs(publish_dir) except os.error: pass
[ [ [ 850, 852 ], [ 1469, 1471 ], [ 1486, 1488 ], [ 1499, 1501 ], [ 1565, 1567 ], [ 1616, 1618 ], [ 1709, 1711 ], [ 1781, 1783 ], [ 1860, 1862 ], [ 1905, 1907 ], [ 1958, 1960 ], [ 1998, 2000 ] ], [ [ 885, 899 ] ], [ [ 910, 927 ] ], [ [ 963, 977 ] ], [ [ 983, 992 ] ], [ [ 998, 1011 ] ], [ [ 1017, 1036 ] ], [ [ 1041, 1059 ] ], [ [ 1064, 1083 ] ], [ [ 1089, 1115 ] ], [ [ 1122, 1134 ] ], [ [ 1173, 1182 ], [ 1430, 1439 ] ], [ [ 1190, 1202 ] ], [ [ 1210, 1219 ] ], [ [ 1227, 1243 ] ], [ [ 1251, 1262 ] ], [ [ 1276, 1295 ] ], [ [ 1457, 1466 ], [ 1579, 1588 ], [ 1629, 1638 ], [ 1722, 1731 ], [ 1794, 1803 ] ], [ [ 1553, 1562 ], [ 1662, 1671 ] ], [ [ 1601, 1613 ] ], [ [ 1690, 1706 ], [ 1872, 1888 ] ], [ [ 1767, 1778 ], [ 1970, 1981 ] ] ]
# coding=utf-8 # -------------------------------------------------------------------------- # Copyright (c) Microsoft Corporation. All rights reserved. # Licensed under the MIT License. See License.txt in the project root for license information. # Code generated by Microsoft (R) AutoRest Code Generator. # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- from typing import Any, TYPE_CHECKING from azure.core.configuration import Configuration from azure.core.pipeline import policies from azure.mgmt.core.policies import ARMHttpLoggingPolicy from .._version import VERSION if TYPE_CHECKING: # pylint: disable=unused-import,ungrouped-imports from azure.core.credentials_async import AsyncTokenCredential class WebSiteManagementClientConfiguration(Configuration): """Configuration for WebSiteManagementClient. Note that all parameters used to create this instance are saved as instance attributes. :param credential: Credential needed for the client to connect to Azure. :type credential: ~azure.core.credentials_async.AsyncTokenCredential :param subscription_id: Your Azure subscription ID. This is a GUID-formatted string (e.g. 00000000-0000-0000-0000-000000000000). :type subscription_id: str """ def __init__( self, credential: "AsyncTokenCredential", subscription_id: str, **kwargs: Any ) -> None: if credential is None: raise ValueError("Parameter 'credential' must not be None.") if subscription_id is None: raise ValueError("Parameter 'subscription_id' must not be None.") super(WebSiteManagementClientConfiguration, self).__init__(**kwargs) self.credential = credential self.subscription_id = subscription_id self.api_version = "2015-08-01" self.credential_scopes = kwargs.pop('credential_scopes', ['https://management.azure.com/.default']) kwargs.setdefault('sdk_moniker', 'mgmt-web/{}'.format(VERSION)) self._configure(**kwargs) def _configure( self, **kwargs: Any ) -> None: self.user_agent_policy = kwargs.get('user_agent_policy') or policies.UserAgentPolicy(**kwargs) self.headers_policy = kwargs.get('headers_policy') or policies.HeadersPolicy(**kwargs) self.proxy_policy = kwargs.get('proxy_policy') or policies.ProxyPolicy(**kwargs) self.logging_policy = kwargs.get('logging_policy') or policies.NetworkTraceLoggingPolicy(**kwargs) self.http_logging_policy = kwargs.get('http_logging_policy') or ARMHttpLoggingPolicy(**kwargs) self.retry_policy = kwargs.get('retry_policy') or policies.AsyncRetryPolicy(**kwargs) self.custom_hook_policy = kwargs.get('custom_hook_policy') or policies.CustomHookPolicy(**kwargs) self.redirect_policy = kwargs.get('redirect_policy') or policies.AsyncRedirectPolicy(**kwargs) self.authentication_policy = kwargs.get('authentication_policy') if self.credential and not self.authentication_policy: self.authentication_policy = policies.AsyncBearerTokenCredentialPolicy(self.credential, *self.credential_scopes, **kwargs)
[ [ [ 487, 490 ], [ 1484, 1487 ], [ 2190, 2193 ] ], [ [ 492, 505 ], [ 693, 706 ] ], [ [ 544, 557 ], [ 873, 886 ] ], [ [ 590, 598 ], [ 2277, 2285 ], [ 2374, 2382 ], [ 2465, 2473 ], [ 2558, 2566 ], [ 2764, 2772 ], [ 2870, 2878 ], [ 2970, 2978 ], [ 3186, 3194 ] ], [ [ 636, 656 ], [ 2675, 2695 ] ], [ [ 681, 688 ], [ 2093, 2100 ] ], [ [ 807, 827 ] ], [ [ 836, 872 ], [ 1735, 1771 ] ] ]
import django.http import unittest.mock from .. import middleware def get_response(req): # dummy get_response, just return an empty response return django.http.HttpResponse() def test_leaves_remote_addr_alone_if_no_real_ip(): remote_addr = object() request = unittest.mock.MagicMock() request.META = {"REMOTE_ADDR": remote_addr} middleware.XRealIPMiddleware(get_response)(request) assert request.META["REMOTE_ADDR"] is remote_addr def test_switches_out_x_real_ip_if_available(): remote_addr = object() x_real_ip = object() request = unittest.mock.MagicMock() request.META = {"REMOTE_ADDR": remote_addr, "HTTP_X_REAL_IP": x_real_ip} middleware.XRealIPMiddleware(get_response)(request) assert request.META["REMOTE_ADDR"] is x_real_ip assert request.META["HTTP_X_REAL_IP"] is x_real_ip
[ [ [ 7, 18 ], [ 160, 166 ] ], [ [ 27, 40 ], [ 281, 289 ], [ 584, 592 ] ], [ [ 57, 67 ], [ 360, 370 ], [ 692, 702 ] ], [ [ 74, 86 ], [ 389, 401 ], [ 721, 733 ] ], [ [ 193, 236 ] ], [ [ 473, 513 ] ] ]
#!/usr/bin/env python import time import RPi.GPIO as GPIO GPIO.setmode(GPIO.BCM) GPIO.setup(21, GPIO.OUT) GPIO.output(21, GPIO.LOW) time.sleep(3.00) GPIO.output(21, GPIO.HIGH) GPIO.cleanup()
[ [ [ 30, 34 ], [ 137, 141 ] ], [ [ 43, 59 ], [ 62, 66 ], [ 75, 79 ], [ 85, 89 ], [ 100, 104 ], [ 110, 114 ], [ 126, 130 ], [ 155, 159 ], [ 171, 175 ], [ 182, 186 ] ] ]
from direct.directnotify.DirectNotifyGlobal import directNotify class Notifier: def __init__(self, name): """ @param name: The name of the notifier. Be sure to add it to your config/Config.prc! @type name: str """ self.notify = directNotify.newCategory(name)
[ [ [ 51, 63 ], [ 274, 286 ] ], [ [ 72, 80 ] ] ]
import numpy as np def train_ml_squarer() -> None: print("Training!") def square() -> int: """Square a number...maybe""" return np.random.randint(1, 100) if __name__ == '__main__': train_ml_squarer()
[ [ [ 7, 18 ], [ 144, 146 ] ], [ [ 25, 41 ], [ 203, 219 ] ], [ [ 82, 88 ] ] ]
""" Platformer Game """ import arcade # Constants SCREEN_WIDTH = 1000 SCREEN_HEIGHT = 650 SCREEN_TITLE = "Platformer" # Constants used to scale our sprites from their original size CHARACTER_SCALING = 1 TILE_SCALING = 0.5 COIN_SCALING = 0.5 SPRITE_PIXEL_SIZE = 128 GRID_PIXEL_SIZE = SPRITE_PIXEL_SIZE * TILE_SCALING # Movement speed of player, in pixels per frame PLAYER_MOVEMENT_SPEED = 10 GRAVITY = 1 PLAYER_JUMP_SPEED = 20 class MyGame(arcade.Window): """ Main application class. """ def __init__(self): # Call the parent class and set up the window super().__init__(SCREEN_WIDTH, SCREEN_HEIGHT, SCREEN_TITLE) # Our TileMap Object self.tile_map = None # Our Scene Object self.scene = None # Separate variable that holds the player sprite self.player_sprite = None # Our physics engine self.physics_engine = None # A Camera that can be used for scrolling the screen self.camera = None # A Camera that can be used to draw GUI elements self.gui_camera = None # Keep track of the score self.score = 0 # Load sounds self.collect_coin_sound = arcade.load_sound(":resources:sounds/coin1.wav") self.jump_sound = arcade.load_sound(":resources:sounds/jump1.wav") arcade.set_background_color(arcade.csscolor.CORNFLOWER_BLUE) def setup(self): """Set up the game here. Call this function to restart the game.""" # Setup the Cameras self.camera = arcade.Camera(self.width, self.height) self.gui_camera = arcade.Camera(self.width, self.height) # Name of map file to load map_name = ":resources:tiled_maps/map.json" # Layer specific options are defined based on Layer names in a dictionary # Doing this will make the SpriteList for the platforms layer # use spatial hashing for detection. layer_options = { "Platforms": { "use_spatial_hash": True, }, } # Read in the tiled map self.tile_map = arcade.load_tilemap(map_name, TILE_SCALING, layer_options) # Initialize Scene with our TileMap, this will automatically add all layers # from the map as SpriteLists in the scene in the proper order. self.scene = arcade.Scene.from_tilemap(self.tile_map) # Keep track of the score self.score = 0 # Set up the player, specifically placing it at these coordinates. image_source = ":resources:images/animated_characters/female_adventurer/femaleAdventurer_idle.png" self.player_sprite = arcade.Sprite(image_source, CHARACTER_SCALING) self.player_sprite.center_x = 128 self.player_sprite.center_y = 128 self.scene.add_sprite("Player", self.player_sprite) # --- Other stuff # Set the background color if self.tile_map.background_color: arcade.set_background_color(self.tile_map.background_color) # Create the 'physics engine' self.physics_engine = arcade.PhysicsEnginePlatformer( self.player_sprite, gravity_constant=GRAVITY, walls=self.scene["Platforms"] ) def on_draw(self): """Render the screen.""" # Clear the screen to the background color arcade.start_render() # Activate the game camera self.camera.use() # Draw our Scene self.scene.draw() # Activate the GUI camera before drawing GUI elements self.gui_camera.use() # Draw our score on the screen, scrolling it with the viewport score_text = f"Score: {self.score}" arcade.draw_text( score_text, 10, 10, arcade.csscolor.WHITE, 18, ) def on_key_press(self, key, modifiers): """Called whenever a key is pressed.""" if key == arcade.key.UP or key == arcade.key.W: if self.physics_engine.can_jump(): self.player_sprite.change_y = PLAYER_JUMP_SPEED arcade.play_sound(self.jump_sound) elif key == arcade.key.LEFT or key == arcade.key.A: self.player_sprite.change_x = -PLAYER_MOVEMENT_SPEED elif key == arcade.key.RIGHT or key == arcade.key.D: self.player_sprite.change_x = PLAYER_MOVEMENT_SPEED def on_key_release(self, key, modifiers): """Called when the user releases a key.""" if key == arcade.key.LEFT or key == arcade.key.A: self.player_sprite.change_x = 0 elif key == arcade.key.RIGHT or key == arcade.key.D: self.player_sprite.change_x = 0 def center_camera_to_player(self): screen_center_x = self.player_sprite.center_x - (self.camera.viewport_width / 2) screen_center_y = self.player_sprite.center_y - ( self.camera.viewport_height / 2 ) if screen_center_x < 0: screen_center_x = 0 if screen_center_y < 0: screen_center_y = 0 player_centered = screen_center_x, screen_center_y self.camera.move_to(player_centered) def on_update(self, delta_time): """Movement and game logic""" # Move the player with the physics engine self.physics_engine.update() # See if we hit any coins coin_hit_list = arcade.check_for_collision_with_list( self.player_sprite, self.scene["Coins"] ) # Loop through each coin we hit (if any) and remove it for coin in coin_hit_list: # Remove the coin coin.remove_from_sprite_lists() # Play a sound arcade.play_sound(self.collect_coin_sound) # Add one to the score self.score += 1 # Position the camera self.center_camera_to_player() def main(): """Main function""" window = MyGame() window.setup() arcade.run() if __name__ == "__main__": main()
[ [ [ 31, 37 ], [ 444, 450 ], [ 1215, 1221 ], [ 1290, 1296 ], [ 1348, 1354 ], [ 1376, 1382 ], [ 1558, 1564 ], [ 1623, 1629 ], [ 2125, 2131 ], [ 2362, 2368 ], [ 2673, 2679 ], [ 2981, 2987 ], [ 3110, 3116 ], [ 3357, 3363 ], [ 3710, 3716 ], [ 3796, 3802 ], [ 3957, 3963 ], [ 3981, 3987 ], [ 4122, 4128 ], [ 4177, 4183 ], [ 4203, 4209 ], [ 4302, 4308 ], [ 4329, 4335 ], [ 4524, 4530 ], [ 4550, 4556 ], [ 4628, 4634 ], [ 4655, 4661 ], [ 5410, 5416 ], [ 5722, 5728 ], [ 5981, 5987 ] ], [ [ 51, 63 ], [ 609, 621 ] ], [ [ 71, 84 ], [ 623, 636 ] ], [ [ 91, 103 ], [ 638, 650 ] ], [ [ 183, 200 ], [ 2701, 2718 ] ], [ [ 205, 217 ], [ 305, 317 ], [ 2155, 2167 ] ], [ [ 224, 236 ] ], [ [ 243, 260 ], [ 285, 302 ] ], [ [ 267, 282 ] ], [ [ 367, 388 ], [ 4260, 4281 ], [ 4385, 4406 ] ], [ [ 394, 401 ], [ 3191, 3198 ] ], [ [ 406, 423 ], [ 4088, 4105 ] ], [ [ 437, 443 ], [ 5949, 5955 ] ], [ [ 5904, 5908 ], [ 6027, 6031 ] ] ]
#!/usr/bin/env python3 ''' lib/ycmd/start.py Server bootstrap logic. Includes a utility class for normalizing parameters and calculating default ones. Also includes a helper to set up the temporary options file. ''' import logging import os import tempfile from ..process import ( FileHandles, Process, ) from ..util.fs import ( default_python_binary_path, save_json_file, ) from ..ycmd.constants import ( YCMD_LOG_SPOOL_OUTPUT, YCMD_LOG_SPOOL_SIZE, YCMD_DEFAULT_SERVER_CHECK_INTERVAL_SECONDS, YCMD_DEFAULT_SERVER_IDLE_SUICIDE_SECONDS, ) from ..ycmd.settings import ( get_default_settings_path, generate_settings_data, ) logger = logging.getLogger('sublime-ycmd.' + __name__) class StartupParameters(object): ''' Startup parameters for a ycmd server instance. Should include all the necessary configuration for creating the ycmd server process. Also calculates defaults for certain parameters. ''' def __init__(self, ycmd_root_directory=None, ycmd_settings_path=None, working_directory=None, python_binary_path=None, server_idle_suicide_seconds=None, server_check_interval_seconds=None): self._ycmd_root_directory = None self._ycmd_settings_path = None self._working_directory = None self._python_binary_path = None self._server_idle_suicide_seconds = None self._server_check_interval_seconds = None # additional attributes, can be set via the properties self._log_level = None self._stdout_log_path = None self._stderr_log_path = None self._keep_logs = None self.ycmd_root_directory = ycmd_root_directory self.ycmd_settings_path = ycmd_settings_path self.working_directory = working_directory self.python_binary_path = python_binary_path self.server_idle_suicide_seconds = server_idle_suicide_seconds self.server_check_interval_seconds = server_check_interval_seconds @property def ycmd_root_directory(self): if self._ycmd_root_directory is None: logger.warning('no ycmd root directory has been set') return self._ycmd_root_directory @ycmd_root_directory.setter def ycmd_root_directory(self, ycmd_root_directory): if ycmd_root_directory is not None and \ not isinstance(ycmd_root_directory, str): raise TypeError(ycmd_root_directory,) self._ycmd_root_directory = ycmd_root_directory @property def ycmd_settings_path(self): if self._ycmd_settings_path is None: if self._ycmd_root_directory is not None: return get_default_settings_path(self._ycmd_root_directory) logger.warning('no ycmd root directory has been set') return self._ycmd_settings_path @ycmd_settings_path.setter def ycmd_settings_path(self, ycmd_settings_path): if ycmd_settings_path is not None and \ not isinstance(ycmd_settings_path, str): raise TypeError(ycmd_settings_path,) self._ycmd_settings_path = ycmd_settings_path @property def working_directory(self): if self._working_directory is None: return os.getcwd() return self._working_directory @working_directory.setter def working_directory(self, working_directory): if working_directory is not None and \ not isinstance(working_directory, str): raise TypeError(working_directory,) self._working_directory = working_directory @property def python_binary_path(self): if self._python_binary_path is None: return default_python_binary_path() return self._python_binary_path @python_binary_path.setter def python_binary_path(self, python_binary_path): if python_binary_path is not None and \ not isinstance(python_binary_path, str): raise TypeError(python_binary_path,) self._python_binary_path = python_binary_path @property def server_idle_suicide_seconds(self): if self._server_idle_suicide_seconds is None: return YCMD_DEFAULT_SERVER_IDLE_SUICIDE_SECONDS return self._server_idle_suicide_seconds @server_idle_suicide_seconds.setter def server_idle_suicide_seconds(self, server_idle_suicide_seconds): if server_idle_suicide_seconds is not None and \ not isinstance(server_idle_suicide_seconds, int): raise TypeError(server_idle_suicide_seconds,) self._server_idle_suicide_seconds = server_idle_suicide_seconds @property def server_check_interval_seconds(self): if self._server_check_interval_seconds is None: return YCMD_DEFAULT_SERVER_CHECK_INTERVAL_SECONDS return self._server_check_interval_seconds @server_check_interval_seconds.setter def server_check_interval_seconds(self, server_check_interval_seconds): if server_check_interval_seconds is not None and \ not isinstance(server_check_interval_seconds, int): raise TypeError(server_check_interval_seconds,) self._server_check_interval_seconds = server_check_interval_seconds @property def log_level(self): return self._log_level @log_level.setter def log_level(self, log_level): if log_level is not None and not isinstance(log_level, str): raise TypeError('log level must be a str: %r' % (log_level)) if log_level is not None and not _is_valid_log_level(log_level): logger.warning('log level unrecognized: %r', log_level) # but fall through and do it anyway self._log_level = log_level @property def stdout_log_path(self): return self._stdout_log_path @stdout_log_path.setter def stdout_log_path(self, stdout_log_path): if stdout_log_path is not None and \ not isinstance(stdout_log_path, str): raise TypeError( 'stdout log path must be a str: %r' % (stdout_log_path) ) self._stdout_log_path = stdout_log_path @property def stderr_log_path(self): return self._stderr_log_path @stderr_log_path.setter def stderr_log_path(self, stderr_log_path): if stderr_log_path is not None and \ not isinstance(stderr_log_path, str): raise TypeError( 'stderr_log_path must be a str: %r' % (stderr_log_path) ) self._stderr_log_path = stderr_log_path @property def keep_logs(self): if self._keep_logs is None: return False return self._keep_logs @keep_logs.setter def keep_logs(self, keep_logs): if keep_logs is not None and not isinstance(keep_logs, bool): raise TypeError('keep-logs must be a bool: %r' % (keep_logs)) self._keep_logs = keep_logs @property def ycmd_module_directory(self): if self._ycmd_root_directory is None: logger.error('no ycmd root directory set') raise AttributeError return os.path.join(self._ycmd_root_directory, 'ycmd') def copy(self): ''' Creates a shallow-copy of the startup parameters. ''' raw_attrs = [ '_ycmd_root_directory', '_ycmd_settings_path', '_working_directory', '_python_binary_path', '_server_idle_suicide_seconds', '_server_check_interval_seconds', '_log_level', '_stdout_log_path', '_stderr_log_path', '_keep_logs', ] result = StartupParameters() for attr in raw_attrs: attr_value = getattr(self, attr) setattr(result, attr, attr_value) return result def __iter__(self): ''' Dictionary-compatible iterator. ''' return iter(( ('ycmd_root_directory', self.ycmd_root_directory), ('ycmd_settings_path', self.ycmd_settings_path), ('working_directory', self.working_directory), ('python_binary_path', self.python_binary_path), ('server_idle_suicide_seconds', self.server_idle_suicide_seconds), ( 'server_check_interval_seconds', self.server_check_interval_seconds, ), ('ycmd_module_directory', self.ycmd_module_directory), ('log_level', self.log_level), ('stdout_log_path', self.stdout_log_path), ('stderr_log_path', self.stderr_log_path), ('keep_logs', self.keep_logs), )) def __str__(self): return ( 'ycmd path, default settings path, ' 'python binary path, working directory: ' '%(ycmd_root_directory)s, %(ycmd_settings_path)s, ' '%(python_binary_path)s, %(working_directory)s' % (dict(self)) ) def __repr__(self): return '%s(%r)' % (StartupParameters, dict(self)) def to_startup_parameters(ycmd_root_directory, ycmd_settings_path=None, working_directory=None, python_binary_path=None, server_idle_suicide_seconds=None, server_check_interval_seconds=None): ''' Internal convenience function. Receives the raw arguments to starting a ycmd server and returns a `StartupParameters` instance from it. If the first argument is already `StartupParameters`, it is returned as-is, and the remaining parameters are ignored. Otherwise, a `StartupParameters` instance is constructed with all the given parameters and returned. ''' if isinstance(ycmd_root_directory, StartupParameters): # great, already in the desired state # check if other params are provided and issue a warning # (they get ignored in that case) if ycmd_settings_path is not None: logger.warning( 'ycmd settings path will be ignored: %s', ycmd_settings_path, ) if working_directory is not None: logger.warning( 'working directory will be ignored: %s', working_directory, ) if python_binary_path is not None: logger.warning( 'python binary path will be ignored: %s', python_binary_path, ) if server_idle_suicide_seconds is not None: logger.warning( 'server idle suicide seconds will be ignored: %s', server_idle_suicide_seconds, ) if server_check_interval_seconds is not None: logger.warning( 'server check interval seconds will be ignored: %s', server_check_interval_seconds, ) return ycmd_root_directory # else, generate them logger.warning('[DEPRECATED] to startup parameters', stack_info=True) logger.debug( 'generating startup parameters with root: %s', ycmd_root_directory, ) return StartupParameters( ycmd_root_directory, ycmd_settings_path=ycmd_settings_path, working_directory=working_directory, python_binary_path=python_binary_path, server_idle_suicide_seconds=server_idle_suicide_seconds, server_check_interval_seconds=server_check_interval_seconds, ) def check_startup_parameters(startup_parameters): ''' Performs quick, non-blocking validation on startup parameters to catch type mismatches or empty configurations. Raises an exception or returns `None`. This is meant to be run on the main thread to catch common startup errors before initializing the server off-thread. It isn't strictly necessary, but produces nicer error messages when the plugin is not configured correctly. NOTE : This does not check the file system for things like missing files, as that can be a blocking operation. ''' if not isinstance(startup_parameters, StartupParameters): raise TypeError( 'startup parameters must be StartupParameters: %r' % (startup_parameters) ) ycmd_root_directory = startup_parameters.ycmd_root_directory if not ycmd_root_directory: raise RuntimeError('no ycmd root directory has been set') ycmd_settings_path = startup_parameters.ycmd_settings_path if not ycmd_settings_path: raise RuntimeError('no ycmd default settings path has been set') logger.debug( 'startup parameters seem to be filled in, ' 'ready to attempt startup: %r', startup_parameters, ) def write_ycmd_settings_file(ycmd_settings_path, ycmd_hmac_secret, out=None): ''' Writes out a ycmd server settings file based on the template file `ycmd_settings_path`. A uniquely-generated `ycmd_hmac_secret` must also be supplied, as it needs to be written into this file. The return value is the path to the settings file, as a `str`. If `out` is omitted, a secure temporary file is created, and the returned path should be passed via the options flag to ycmd. If `out` is provided, it should be a path to an output file (`str`), or a file-like handle (must support `.write`). This is not recommended for use with ycmd, as it may be insecure. ''' ycmd_settings_data = generate_settings_data( ycmd_settings_path, ycmd_hmac_secret, ) out_path = None if out is None: # no point using `with` for this, since we also use `delete=False` temp_file_object = tempfile.NamedTemporaryFile( prefix='ycmd_settings_', suffix='.json', delete=False, ) temp_file_name = temp_file_object.name temp_file_handle = temp_file_object.file # type: io.TextIOWrapper out = temp_file_handle out_path = temp_file_name def flush(): temp_file_handle.flush() def close(): temp_file_object.close() else: raise NotImplementedError('unimplemented: output to specific file') if out_path is None and out is not None: logger.error('failed to get path for output file: %r', out) # fall through and write it out anyway save_json_file(out, ycmd_settings_data) flush() close() logger.debug('successfully wrote file: %s', out_path) return out_path def prepare_ycmd_process(startup_parameters, ycmd_settings_tempfile_path, ycmd_server_hostname, ycmd_server_port): ''' Initializes and returns a `Process` handle, correctly configured to launch a ycmd server process. It does not automatically start it though. The `ycmd_settings_tempfile_path` should be created by (return value of) `write_ycmd_settings_file`. The ycmd server process will read that file on startup and then immediately delete it. The `ycmd_server_hostname` and `ycmd_server_port` must also be provided to instruct the server to listen on the given address. ''' assert isinstance(startup_parameters, StartupParameters), \ 'startup parameters must be StartupParameters: %r' % \ (startup_parameters) assert isinstance(ycmd_settings_tempfile_path, str), \ 'ycmd settings temporary file path must be a str: %r' % \ (ycmd_settings_tempfile_path) # this may throw: check_startup_parameters(startup_parameters) working_directory = startup_parameters.working_directory python_binary_path = startup_parameters.python_binary_path server_idle_suicide_seconds = \ startup_parameters.server_idle_suicide_seconds server_check_interval_seconds = \ startup_parameters.server_check_interval_seconds ycmd_module_directory = startup_parameters.ycmd_module_directory if YCMD_LOG_SPOOL_OUTPUT: stdout_log_spool = \ tempfile.SpooledTemporaryFile(max_size=YCMD_LOG_SPOOL_SIZE) stderr_log_spool = \ tempfile.SpooledTemporaryFile(max_size=YCMD_LOG_SPOOL_SIZE) logger.debug( 'using temporary spools for stdout, stderr: %r, %r', stdout_log_spool, stderr_log_spool, ) stdout_handle = stdout_log_spool stderr_handle = stderr_log_spool else: # explicitly close handles - don't inherit from this process stdout_handle = FileHandles.DEVNULL stderr_handle = FileHandles.DEVNULL ycmd_process_handle = Process() ycmd_process_handle.binary = python_binary_path ycmd_process_handle.args.extend([ ycmd_module_directory, '--host=%s' % (ycmd_server_hostname), '--port=%s' % (ycmd_server_port), '--idle_suicide_seconds=%s' % (server_idle_suicide_seconds), '--check_interval_seconds=%s' % (server_check_interval_seconds), '--options_file=%s' % (ycmd_settings_tempfile_path), ]) ycmd_process_handle.cwd = working_directory ycmd_process_handle.filehandles.stdout = stdout_handle ycmd_process_handle.filehandles.stderr = stderr_handle if startup_parameters.log_level is not None: add_ycmd_debug_args( ycmd_process_handle, log_level=startup_parameters.log_level, stdout_file_name=startup_parameters.stdout_log_path, stderr_file_name=startup_parameters.stderr_log_path, keep_logfiles=startup_parameters.keep_logs, ) return ycmd_process_handle def add_ycmd_debug_args(ycmd_process_handle, log_level='info', stdout_file_name=None, stderr_file_name=None, keep_logfiles=False): ''' Adds startup flags to `ycmd_process_handle` to enable logging output. The `ycmd_process_handle` should be an instance of `Process`. The `log_level` should be one of 'debug', 'info', 'warning', 'error', or 'critical'. Any `str` is accepted, this routine does not actually check it. If `stdout_file_name` and `stderr_file_name` are provided, the server will write log messages to the given files. The bulk of the logs will be on stderr, with only a few startup messages appearing on stdout. If `keep_logfiles` is `True`, then the server won't delete the log files when it exits. Otherwise, the log files will be deleted when it shuts down. ''' if not isinstance(ycmd_process_handle, Process): raise TypeError( 'ycmd process handle must be a Process: %r' % (ycmd_process_handle) ) assert isinstance(ycmd_process_handle, Process) if ycmd_process_handle.alive(): raise ValueError( 'ycmd process is already started, cannot modify it: %r' % (ycmd_process_handle) ) if not _is_valid_log_level(log_level): logger.warning('log level unrecognized: %r', log_level) # but fall through and do it anyway ycmd_debug_args = [ '--log=%s' % (log_level), ] if stdout_file_name and stderr_file_name: ycmd_debug_args.extend([ '--stdout=%s' % (stdout_file_name), '--stderr=%s' % (stderr_file_name), ]) if keep_logfiles: ycmd_debug_args.append( '--keep_logfiles', ) logger.debug('adding ycmd debug args: %r', ycmd_debug_args) ycmd_process_handle.args.extend(ycmd_debug_args) def _is_valid_log_level(log_level): if not isinstance(log_level, str): raise TypeError('log level must be a str: %r' % (log_level)) # these can be found by running `python /path/to/ycmd/ycmd --help` recognized_log_levels = [ 'debug', 'info', 'warning', 'error', 'critical', ] return log_level in recognized_log_levels
[ [ [ 225, 232 ], [ 674, 681 ] ], [ [ 240, 242 ], [ 3305, 3307 ], [ 7231, 7233 ] ], [ [ 250, 258 ], [ 13763, 13771 ], [ 16058, 16066 ], [ 16159, 16167 ] ], [ [ 288, 299 ], [ 16551, 16562 ], [ 16595, 16606 ] ], [ [ 305, 312 ], [ 16642, 16649 ], [ 18549, 18556 ], [ 18717, 18724 ] ], [ [ 344, 370 ], [ 3755, 3781 ] ], [ [ 376, 390 ], [ 14429, 14443 ] ], [ [ 429, 450 ], [ 15994, 16015 ] ], [ [ 456, 475 ], [ 16097, 16116 ], [ 16198, 16217 ] ], [ [ 481, 523 ], [ 4840, 4882 ] ], [ [ 529, 569 ], [ 4249, 4289 ] ], [ [ 607, 632 ], [ 2740, 2765 ] ], [ [ 638, 660 ], [ 13543, 13565 ] ], [ [ 665, 671 ], [ 2172, 2178 ], [ 2805, 2811 ], [ 5674, 5680 ], [ 7140, 7146 ], [ 10141, 10147 ], [ 10303, 10309 ], [ 10464, 10470 ], [ 10636, 10642 ], [ 10844, 10850 ], [ 11057, 11063 ], [ 11131, 11137 ], [ 12690, 12696 ], [ 14317, 14323 ], [ 14499, 14505 ], [ 16228, 16234 ], [ 18954, 18960 ], [ 19422, 19428 ] ], [ [ 728, 745 ], [ 7777, 7794 ], [ 9122, 9139 ], [ 9913, 9930 ], [ 11239, 11256 ], [ 12200, 12217 ], [ 15257, 15274 ] ], [ [ 9159, 9180 ] ], [ [ 11572, 11596 ], [ 15561, 15585 ] ], [ [ 12828, 12852 ] ], [ [ 14579, 14599 ] ], [ [ 17637, 17656 ], [ 17297, 17316 ] ], [ [ 19541, 19560 ], [ 5630, 5649 ], [ 18914, 18933 ] ] ]
#!/usr/bin/env python import serial import sys import struct import pprint import argparse import code pp = pprint.PrettyPrinter() class ConsoleUI: def opStart(self, name): sys.stdout.write(name.ljust(40)) def opProgress(self, progress, total=-1): if (total >= 0): prstr = "0x%04x / 0x%04x" % (progress, total) else: prstr = "0x%04x" % (progress) sys.stdout.write(prstr.ljust(20)) sys.stdout.write('\x08' * 20) sys.stdout.flush() def opEnd(self, result): sys.stdout.write(result.ljust(20)) sys.stdout.write("\n") class XFlash: def __init__(self, serialport): self.serial = serial.Serial(serialport, baudrate=115200) def __del__(self): try: self.serial.close() del self.serial except: pass def cmd(self, cmd, argA=0, argB=0): buffer = struct.pack("<LL", argA, argB) self.serial.write(bytes([cmd])) self.serial.write(buffer) self.serial.flush() def flashPowerOn(self): self.cmd(0x10) def flashShutdown(self): self.cmd(0x11) def update(self): try: self.cmd(0xF0) except: pass def flashInit(self): self.cmd(0x03) buffer = self.serial.read(4) return struct.unpack("<L", buffer)[0] def flashDeInit(self): self.cmd(0x04) def flashStatus(self): self.cmd(0x05) buffer = self.serial.read(2) return struct.unpack("<H", buffer)[0] def flashErase(self, block): self.cmd(0x06, block) # return self.flashStatus() def flashReadBlock(self, block): self.cmd(0x01, block, 528 * 32) # for i in range(0, 32): buffer = self.serial.read(528 * 32) status = self.flashStatus() return (status, buffer) def flashWriteBlock(self, block, buffer): self.cmd(0x02, block, len(buffer)) self.serial.write(buffer) return self.flashStatus() # def calcecc(data): # assert len(data) == 0x210 # val = 0 # for i in range(0x1066): # if not i & 31: # v = ~struct.unpack("<L", data[i/8:i/8+4])[0] # val ^= v & 1 # v >>= 1 # if val & 1: # val ^= 0x6954559 # val >>= 1 # # val = ~val # return data[:-4] + struct.pack("<L", (val << 6) & 0xFFFFFFFF) # # def addecc(data, block = 0, off_8 = "\x00" * 4): # res = "" # while len(data): # d = (data[:0x200] + "\x00" * 0x200)[:0x200] # data = data[0x200:] # # d += struct.pack("<L4B4s4s", block / 32, 0, 0xFF, 0, 0, off_8, "\0\0\0\0") # d = calcecc(d) # block += 1 # res += d # return res def main(argv): parser = argparse.ArgumentParser(description='XBox 360 NAND Flasher') parser.add_argument('port', metavar='port', type=str, help='serial port for comms (e.g. COM5 or /dev/ttyUSB0)') subparsers = parser.add_subparsers(title='Operations', dest='action') parser_read = subparsers.add_parser('read', help='Dumps an image from the NAND') parser_read.add_argument('file', nargs=1, type=argparse.FileType('wb'), help='The file to dump the NAND to') parser_read.add_argument('start', nargs='?', metavar='start', action='store', type=int, default=0, help='The block to start the action from') parser_read.add_argument('end', nargs='?', metavar='end', action='store', type=int, default=0x400, help='The count of blocks to perform the action to') parser_write = subparsers.add_parser('write', help='Writes an image into the NAND') parser_write.add_argument('file', nargs=1, type=argparse.FileType('rb'), help='The image file to write to the NAND') parser_write.add_argument('start', nargs='?', metavar='start', action='store', type=int, default=0, help='The block to start the action from') parser_write.add_argument('end', nargs='?', metavar='end', action='store', type=int, default=0x400, help='The count of blocks to perform the action to') # parser_erase = subparsers.add_parser('erase', help='Erases blocks in the NAND') # parser_erase.add_argument('start', nargs='?', metavar='start', action='store', type=int, default=0, # help='The block to start the action from') # parser_erase.add_argument('end', nargs='?', metavar='end', action='store', type=int, default=0x400, # help='The count of blocks to perform the action to') # # parser_update = subparsers.add_parser('update', # help='Jumps into the bootloader of the NAND Flashing device for updating the firmware') # parser_shutdown = subparsers.add_parser('shutdown', help='Shuts down the attached XBox 360') # parser_poweron = subparsers.add_parser('powerup', help='Powers up the attached XBox 360') arguments = parser.parse_args(argv[1:]) ui = ConsoleUI() xf = XFlash(arguments.port) if arguments.action in ('erase', 'write', 'read'): try: flash_config = xf.flashInit() print("FlashConfig: 0x%08x" % (flash_config)) if flash_config <= 0: raise Exception("FlashConfig invalid!") except Exception as e: print("Error!", e) xf.flashDeInit() return 1 try: if arguments.action == 'erase': # start = 0 # end = (options.flashsize * 1024) / 16 start = arguments.start end = arguments.end ui.opStart('Erase') ui.opProgress(0, end) for b in range(start, end): status = xf.flashErase(b) ui.opProgress(b + 1, end) ui.opEnd('0x%04x blocks OK' % (end)) if arguments.action == 'read': # start = 0 # end = (options.flashsize * 1024) / 16 start = arguments.start end = arguments.end ui.opStart('Read') ui.opProgress(0, end) for b in range(start, end): (status, buffer) = xf.flashReadBlock(b) ui.opProgress(b + 1, end) arguments.file[0].write(buffer) if arguments.action == 'write': # start = 0 # end = (options.flashsize * 1024) / 16 start = arguments.start end = arguments.end blocksize = 528 * 32 ui.opStart('Write') ui.opProgress(0, end) for b in range(start, end): buffer = arguments.file[0].read(blocksize) if len(buffer) < blocksize: buffer += ('\xFF' * (blocksize - len(buffer))) status = xf.flashWriteBlock(b, buffer) ui.opProgress(b + 1, end) # # if arguments.action == 'update': # xf.update() # # if arguments.action == 'powerup': # xf.flashPowerOn() # # if arguments.action == 'shutdown': # xf.flashShutdown() except Exception as e: raise e finally: xf.flashDeInit() return 0 if __name__ == '__main__': sys.exit(main(sys.argv))
[ [ [ 29, 35 ], [ 693, 699 ] ], [ [ 43, 46 ], [ 7333, 7336 ], [ 7347, 7350 ], [ 188, 191 ], [ 416, 419 ], [ 458, 461 ], [ 496, 499 ], [ 553, 556 ], [ 596, 599 ] ], [ [ 54, 60 ], [ 924, 930 ], [ 1361, 1367 ], [ 1547, 1553 ] ], [ [ 68, 74 ], [ 109, 115 ] ], [ [ 82, 90 ], [ 2755, 2763 ], [ 3168, 3176 ], [ 3731, 3739 ] ], [ [ 98, 102 ] ], [ [ 104, 106 ] ], [ [ 140, 149 ], [ 5065, 5074 ] ], [ [ 627, 633 ], [ 5087, 5093 ] ], [ [ 2730, 2734 ], [ 7342, 7346 ] ] ]
# Copyright (C) 2010-2011 Richard Lincoln # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to # deal in the Software without restriction, including without limitation the # rights to use, copy, modify, merge, publish, distribute, sublicense, and/or # sell copies of the Software, and to permit persons to whom the Software is # furnished to do so, subject to the following conditions: # # The above copyright notice and this permission notice shall be included in # all copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING # FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS # IN THE SOFTWARE. from CIM16.IEC61968.Common.ActivityRecord import ActivityRecord class ComplianceEvent(ActivityRecord): """Compliance events are used for reporting regulatory or contract compliance issues and/or variances. These might be created as a consequence of local business processes and associated rules. It is anticipated that this class will be customised extensively to meet local implementation needs. Use inherited 'category' to indicate that, for example, expected performance will not be met or reported as mandated.Compliance events are used for reporting regulatory or contract compliance issues and/or variances. These might be created as a consequence of local business processes and associated rules. It is anticipated that this class will be customised extensively to meet local implementation needs. Use inherited 'category' to indicate that, for example, expected performance will not be met or reported as mandated. """ def __init__(self, deadline='', *args, **kw_args): """Initialises a new 'ComplianceEvent' instance. @param deadline: The deadline for compliance. """ #: The deadline for compliance. self.deadline = deadline super(ComplianceEvent, self).__init__(*args, **kw_args) _attrs = ["deadline"] _attr_types = {"deadline": str} _defaults = {"deadline": ''} _enums = {} _refs = [] _many_refs = []
[ [ [ 1149, 1163 ], [ 1187, 1201 ] ], [ [ 1171, 1186 ], [ 2305, 2320 ] ] ]
import logging from django.db.models.query_utils import Q from django.shortcuts import get_object_or_404 from django.utils.decorators import method_decorator from django_filters.rest_framework import DjangoFilterBackend from drf_yasg import openapi from drf_yasg.openapi import Parameter from drf_yasg.utils import no_body, swagger_auto_schema from notifications.signals import notify from rest_framework import mixins, status, viewsets from rest_framework.decorators import action from rest_framework.decorators import parser_classes as dparser_classes from rest_framework.parsers import FormParser, JSONParser, MultiPartParser from rest_framework.permissions import IsAuthenticated from rest_framework.response import Response from rest_framework_extensions.mixins import DetailSerializerMixin, NestedViewSetMixin from looking_for_group.mixins import AutoPermissionViewSetMixin, ParentObjectAutoPermissionViewSetMixin from . import models, serializers from .signals import player_kicked, player_left logger = logging.getLogger("api") parent_lookup_game__slug = Parameter( name="parent_lookup_game__slug", in_="path", type="string", format=openapi.FORMAT_SLUG, description="Slug of related game object.", ) parent_lookup_session__slug = Parameter( name="parent_lookup_session__slug", in_="path", type="string", format=openapi.FORMAT_SLUG, description="Slug of related session object.", ) parent_lookup_session__game__slug = Parameter( name="parent_lookup_session__game__slug", in_="path", type="string", format=openapi.FORMAT_SLUG, description="Slug of related game object.", ) @method_decorator( name="list", decorator=swagger_auto_schema( operation_summary="List Games", operation_description="Fetch a list of game records. **NOTE**: You will probably want to filter by status at least.", ), ) @method_decorator( name="create", decorator=swagger_auto_schema( operation_summary="Game: Create", operation_description="Create a new game posting.", request_body=serializers.GameDataSerializer, responses={201: serializers.GameDataSerializer}, ), ) @method_decorator( name="retrieve", decorator=swagger_auto_schema( operation_summary="Game: Details", operation_description="Fetch the details for the given game. **NOTE**: If you are not a member of the game, only a subset of the available information will be displayed.", responses={ 200: serializers.GameDataSerializer, 403: "You are not authorized to view this game.", }, ), ) @method_decorator( name="update", decorator=swagger_auto_schema( operation_summary="Game: Update", operation_description="Update the details of this game. (Only available to GM)", request_body=serializers.GameDataSerializer, responses={ 200: serializers.GameDataSerializer, 403: "You are not the GM of this game.", }, ), ) @method_decorator( name="partial_update", decorator=swagger_auto_schema( operation_summary="Game: Update", operation_description="Update the details of this game. (Only available to GM)", request_body=serializers.GameDataSerializer, responses={ 200: serializers.GameDataSerializer, 403: "You are not the GM of this game.", }, ), ) @method_decorator( name="destroy", decorator=swagger_auto_schema( operation_summary="Game: Delete", operation_description="Delete the given game. (Only available to GM.)", request_body=no_body, responses={204: "Game was deleted.", 403: "You are not the GM of this game."}, ), ) @method_decorator( name="leave", decorator=swagger_auto_schema( operation_summary="Game: Leave", operation_description="Leave the current game. (Players only.)", request_body=no_body, reponses={ 204: "You have successfully left the game.", 400: "You are not a member of this game.", 403: "You are the GM and cannot leave.", }, ), ) @method_decorator( name="apply", decorator=swagger_auto_schema( operation_summary="Game: Apply", operation_description="Apply to join this game.", request_body=serializers.GameApplicationSerializer, responses={ 201: serializers.GameApplicationSerializer, 400: "You are already a member of this game.", 403: "You are not permitted to apply to this game either due to your access rights or the game's status.", }, ), ) class GamePostingViewSet( AutoPermissionViewSetMixin, DetailSerializerMixin, NestedViewSetMixin, viewsets.ModelViewSet, ): """ A view set that allows the retrieval and manipulation of posted game data. """ permission_classes = (IsAuthenticated,) parser_classes = [FormParser, MultiPartParser] model = models.GamePosting lookup_field = "slug" lookup_url_kwarg = "slug" serializer_class = serializers.GameDataListSerializer serializer_detail_class = serializers.GameDataSerializer filter_backends = [DjangoFilterBackend] filterset_fields = [ "published_game", "game_system", "published_module", "status", "game_type", "game_mode", ] permission_type_map = { **AutoPermissionViewSetMixin.permission_type_map, "apply": "apply", "leave": "leave", } def get_queryset(self): gamer = self.request.user.gamerprofile friends = gamer.friends.all() communities = [f.id for f in gamer.communities.all()] game_player_ids = [ obj.game.id for obj in models.Player.objects.filter(gamer=gamer).select_related("game") ] q_gm = Q(gm=gamer) q_gm_is_friend = Q(gm__in=friends) & Q(privacy_level="community") q_isplayer = Q(id__in=game_player_ids) q_community = Q(communities__id__in=communities) & Q(privacy_level="community") q_public = Q(privacy_level="public") qs = models.GamePosting.objects.filter( q_gm | q_public | q_gm_is_friend | q_isplayer | q_community ).distinct() return qs def create(self, request, *args, **kwargs): self.serializer_class = serializers.GameDataSerializer return super().create(request, *args, **kwargs) def retrieve(self, request, *args, **kwargs): if not request.user.has_perm("game.is_member", self.get_object()): logger.debug( "User is not a member of game, swtiching serializer to list view mode." ) self.serializer_detail_class = serializers.GameDataListSerializer return super().retrieve(request, *args, **kwargs) @action(methods=["post"], detail=True, parser_classes=[FormParser, JSONParser]) def apply(self, request, *args, **kwargs): obj = self.get_object() logger.debug("Retrieved game object of {}".format(obj)) if request.user.has_perm("game.is_member", obj): return Response( data={"errors": "You are already in this game..."}, status=status.HTTP_400_BAD_REQUEST, ) new_application = serializers.GameApplicationSerializer( data=request.data, context={"request": request} ) if not new_application.is_valid(): return Response( data=new_application.errors, status=status.HTTP_400_BAD_REQUEST ) app = models.GamePostingApplication.objects.create( game=obj, gamer=request.user.gamerprofile, message=new_application.validated_data["message"], status="pending", ) notify.send( request.user.gamerprofile, recipient=obj.gm.user, verb="submitted application", action_object=app, target=obj, ) return Response( data=serializers.GameApplicationSerializer( app, context={"request": request} ).data, status=status.HTTP_201_CREATED, ) @action(methods=["post"], detail=True, parser_classes=[FormParser, JSONParser]) def leave(self, request, *args, **kwargs): obj = self.get_object() if request.user == obj.gm.user: return Response( data={"errors": "The GM cannot leave the game."}, status=status.HTTP_400_BAD_REQUEST, ) player = models.Player.objects.get(gamer=request.user.gamerprofile, game=obj) player_left.send(models.Player, player=player) player.delete() return Response(status=status.HTTP_204_NO_CONTENT) @method_decorator( name="list", decorator=swagger_auto_schema( operation_summary="Game: List Sessions", operation_description="List the sessions for the given game.", manual_parameters=[parent_lookup_game__slug], ), ) @method_decorator( name="retrieve", decorator=swagger_auto_schema( operation_summary="Game Session: Details", operation_description="Get the details for the given session. **NOTE**: If the user is just a player, the GM notes and player details will not be included.", manual_parameters=[parent_lookup_game__slug], responses={ 200: serializers.GameSessionGMSerializer, 403: "You are not a member of this game.", }, ), ) @method_decorator( name="update", decorator=swagger_auto_schema( operation_summary="Game Session: Update", operation_description="Update details of the game session.", manual_parameters=[parent_lookup_game__slug], request_body=serializers.GameSessionGMSerializer, responses={ 200: serializers.GameSessionGMSerializer, 403: "You are not the GM of this game.", }, ), ) @method_decorator( name="partial_update", decorator=swagger_auto_schema( operation_summary="Game Session: Update", operation_description="Update details of the game session.", manual_parameters=[parent_lookup_game__slug], request_body=serializers.GameSessionGMSerializer, responses={ 200: serializers.GameSessionGMSerializer, 403: "You are not the GM of this game.", }, ), ) @method_decorator( name="destroy", decorator=swagger_auto_schema( operation_summary="Game Session: Delete", operation_description="Delete the game session.", manual_parameters=[parent_lookup_game__slug], request_body=serializers.GameSessionGMSerializer, responses={ 204: "Session was deleted.", 403: "You are not the GM of this game.", }, ), ) @method_decorator( name="cancel", decorator=swagger_auto_schema( operation_summary="Game Session: Cancel", operation_description="Cancel the game session.", manual_parameters=[parent_lookup_game__slug], request_body=no_body, responses={ 200: serializers.GameSessionGMSerializer, 400: "This session is already canceled or complete.", 403: "You are not the GM of this game.", }, ), ) @method_decorator( name="uncancel", decorator=swagger_auto_schema( operation_summary="Game Session: Uncancel", operation_description="Uncancel the game session.", manual_parameters=[parent_lookup_game__slug], request_body=no_body, responses={ 200: serializers.GameSessionGMSerializer, 400: "This session is not canceled.", 403: "You are not the GM of this game.", }, ), ) @method_decorator( name="complete", decorator=swagger_auto_schema( operation_summary="Game Session: Mark Complete", operation_description="Mark the game session as complete.", manual_parameters=[parent_lookup_game__slug], request_body=no_body, responses={ 200: serializers.GameSessionGMSerializer, 400: "This session is already canceled or complete.", 403: "You are not the GM of this game.", }, ), ) @method_decorator( name="uncomplete", decorator=swagger_auto_schema( operation_summary="Game Session: Uncomplete", operation_description="Undo the completion status of the session.", manual_parameters=[parent_lookup_game__slug], request_body=no_body, responses={ 200: serializers.GameSessionGMSerializer, 400: "This session isn't marked as complete.", 403: "You are not the GM of this game.", }, ), ) @method_decorator( name="reschedule", decorator=swagger_auto_schema( operation_summary="Game Session: Reschedule", operation_description="Reschedule the game session to another date/time.", manual_parameters=[parent_lookup_game__slug], request_body=serializers.ScheduleSerializer, responses={ 200: serializers.GameSessionGMSerializer, 400: "Your date and time were invalid or the session is already marked as complete or canceled.", 403: "You are not the GM of this game.", }, ), ) @method_decorator( name="addlog", decorator=swagger_auto_schema( operation_summary="Game Session: Add Adventure Log", operation_description="Add an adventure log to this session.", manual_parameters=[parent_lookup_game__slug], request_body=serializers.AdventureLogSerializer, responses={ 201: serializers.AdventureLogSerializer, 400: "This session already has an adventure log. You should update that instead.", 403: "You don't have permission to add an adventure log.", }, ), ) class GameSessionViewSet( ParentObjectAutoPermissionViewSetMixin, NestedViewSetMixin, mixins.ListModelMixin, mixins.RetrieveModelMixin, mixins.UpdateModelMixin, mixins.DestroyModelMixin, viewsets.GenericViewSet, ): """ Views for seeing game session data. """ model = models.GameSession serializer_class = serializers.GameSessionSerializer lookup_field = "slug" lookup_url_kwarg = "slug" parent_dependent_actions = [ "create", "retrieve", "update", "partial_update", "list", "destroy", "reschedule", "cancel", "uncancel", "addlog", "complete", "uncomplete", ] parent_lookup_field = "game" parent_object_model = models.GamePosting parent_object_lookup_field = "slug" parent_object_url_kwarg = "parent_lookup_game__slug" permission_type_map = { **ParentObjectAutoPermissionViewSetMixin.permission_type_map, "addlog": "view", "reschedule": "change", "cancel": "change", "uncancel": "change", "complete": "change", "uncomplete": "change", } permission_type_map["list"] = "view" def get_parent_game(self): return get_object_or_404( models.GamePosting, slug=self.kwargs["parent_lookup_game__slug"] ) def get_queryset(self): return self.model.objects.filter( game__slug=self.kwargs["parent_lookup_game__slug"] ).order_by("-scheduled_time") def dispatch(self, request, *args, **kwargs): if ( request.user.is_authenticated and request.user.gamerprofile == self.get_parent_game().gm ): self.serializer_class = serializers.GameSessionGMSerializer return super().dispatch(request, *args, **kwargs) @action(methods=["post"], detail=True) def reschedule(self, request, *args, **kwargs): date_serializer = serializers.ScheduleSerializer(data=request.data) if not date_serializer.is_valid(): return Response( data=date_serializer.errors, status=status.HTTP_400_BAD_REQUEST ) obj = self.get_object() if obj.status in ["complete", "cancel"]: return Response( data={ "errors": "This session is already marked as {} and cannot be rescheduled.".format( obj.get_status_display() ) }, status=status.HTTP_400_BAD_REQUEST, ) obj.move(date_serializer.validated_data["new_scheduled_time"]) return Response( data=self.serializer_class(obj, context={"request": request}).data, status=status.HTTP_200_OK, ) @action(methods=["post"], detail=True) def complete(self, request, *args, **kwargs): obj = self.get_object() if obj.status in ["complete", "cancel"]: return Response( data={ "errors": "This object is either already completed or canceled and cannot be toggled to complete." }, status=status.HTTP_400_BAD_REQUEST, ) obj.status = "complete" obj.save() return Response( data=self.serializer_class(obj, context={"request": request}).data, status=status.HTTP_200_OK, ) @action(methods=["post"], detail=True) def uncomplete(self, request, *args, **kwargs): obj = self.get_object() if obj.status != "complete": return Response( data={ "errors": "This object is not completed and so completion cannot be undone." }, status=status.HTTP_400_BAD_REQUEST, ) obj.status = "pending" obj.save() return Response( data=self.serializer_class(obj, context={"request": request}).data, status=status.HTTP_200_OK, ) @action(methods=["post"], detail=True) def cancel(self, request, *args, **kwargs): obj = self.get_object() if obj.status in ["complete", "cancel"]: return Response( data={"errors": "This session is already completed or canceled."}, status=status.HTTP_400_BAD_REQUEST, ) obj.cancel() return Response( data=self.serializer_class(obj, context={"request": request}).data, status=status.HTTP_200_OK, ) @action(methods=["post"], detail=True) def uncancel(self, request, *args, **kwargs): obj = self.get_object() if obj.status != "cancel": return Response( data={ "errors": "This session is not canceled and can't be changed this way." }, status=status.HTTP_400_BAD_REQUEST, ) obj.uncancel() return Response( data=self.serializer_class(obj, context={"request": request}).data, status=status.HTTP_200_OK, ) @action(methods=["post"], detail=True) def addlog(self, request, *args, **kwargs): """ Create the adventure log for this session. """ session = self.get_object() if hasattr(session, "adventurelog"): return Response( data={"errors": "This session already has an adventure log."}, status=status.HTTP_400_BAD_REQUEST, ) log_serializer = serializers.AdventureLogSerializer( session=session, data=request.data, context={"request": request} ) if not log_serializer.is_valid(): return Response( data=log_serializer.errors, status=status.HTTP_400_BAD_REQUEST ) new_log = log_serializer.save() return Response( data=serializers.AdventureLogSerializer( new_log, context={"request": request} ).data, status=status.HTTP_201_CREATED, ) @method_decorator( name="retrieve", decorator=swagger_auto_schema( operation_summary="Adventure Log: Details", operation_description="Fetch the details for a given adventure log.", manual_parameters=[ parent_lookup_session__game__slug, parent_lookup_session__slug, ], responses={ 200: serializers.AdventureLogSerializer, 403: "You are not a member of this game.", }, ), ) @method_decorator( name="update", decorator=swagger_auto_schema( operation_summary="Adventure Log: Update", operation_description="Update the details for a given adventure log.", manual_parameters=[ parent_lookup_session__game__slug, parent_lookup_session__slug, ], request_body=serializers.AdventureLogSerializer, responses={ 200: serializers.AdventureLogSerializer, 403: "You don't have permissions to edit this adventure log.", }, ), ) @method_decorator( name="partial_update", decorator=swagger_auto_schema( operation_summary="Adventure Log: Update", operation_description="Update the details for a given adventure log.", manual_parameters=[ parent_lookup_session__game__slug, parent_lookup_session__slug, ], request_body=serializers.AdventureLogSerializer, responses={ 200: serializers.AdventureLogSerializer, 403: "You don't have permissions to edit this adventure log.", }, ), ) @method_decorator( name="destroy", decorator=swagger_auto_schema( operation_summary="Adventure Log: Delete", operation_description="Delete a given adventure log.", manual_parameters=[ parent_lookup_session__game__slug, parent_lookup_session__slug, ], request_body=no_body, responses={ 204: "The adventure log was successfully deleted.", 403: "You don't have permissions to edit this adventure log.", }, ), ) class AdventureLogViewSet( ParentObjectAutoPermissionViewSetMixin, NestedViewSetMixin, mixins.RetrieveModelMixin, mixins.UpdateModelMixin, mixins.DestroyModelMixin, viewsets.GenericViewSet, ): """ Allows the manipulation of view sets. """ model = models.AdventureLog parent_lookup_field = "session__game" parent_object_model = models.GamePosting parent_object_lookup_field = "slug" parent_object_url_kwarg = "parent_lookup_session__game__slug" serializer_class = serializers.AdventureLogSerializer lookup_field = "slug" lookup_url_kwarg = "slug" permission_required = "game.is_member" permission_type_map = {**ParentObjectAutoPermissionViewSetMixin.permission_type_map} permission_type_map["list"] = "add" parent_dependent_actions = [ "create", "retrieve", "update", "partial_update", "destroy", ] def get_queryset(self): return models.AdventureLog.objects.filter( session__slug=self.kwargs["parent_lookup_session__slug"] ) @method_decorator( name="list", decorator=swagger_auto_schema( operation_summary="List Your Game Applications", operation_description="Fetch a list of all your game applications.", ), ) @method_decorator( name="retrieve", decorator=swagger_auto_schema( operation_summary="Your Game Application: Details", operation_description="Fetch the details of your game application.", ), ) @method_decorator( name="update", decorator=swagger_auto_schema( operation_summary="Your Game Application: Update", operation_description="Update the details of your game application.", ), ) @method_decorator( name="partial_update", decorator=swagger_auto_schema( operation_summary="Your Game Application: Update", operation_description="Update the details of your game application.", ), ) @method_decorator( name="destroy", decorator=swagger_auto_schema( operation_summary="Your Game Application: Withdraw", operation_description="Withdraw your game application by deleting the record.", ), ) class GameApplicationViewSet( AutoPermissionViewSetMixin, mixins.ListModelMixin, mixins.RetrieveModelMixin, mixins.UpdateModelMixin, mixins.DestroyModelMixin, viewsets.GenericViewSet, ): """ View for an applicant to review, create, update, and delete their applications to games. """ permission_classes = (IsAuthenticated,) serializer_class = serializers.GameApplicationSerializer filter_backends = [DjangoFilterBackend] filterset_fields = ["status"] lookup_field = "slug" lookup_url_kwarg = "slug" permission_type_map = {**AutoPermissionViewSetMixin.permission_type_map} def get_queryset(self): logger.debug("Fetching gamerprofile from request...") gamer = self.request.user.gamerprofile logger.debug("Fetching game applications for gamer {}".format(gamer)) qs = models.GamePostingApplication.objects.filter( gamer=self.request.user.gamerprofile ).order_by("-modified", "-created", "status") logger.debug( "Retrieved queryset of length {} for gamer {}".format( qs.count(), self.request.user.gamerprofile ) ) return qs @method_decorator( name="list", decorator=swagger_auto_schema( operation_summary="List Applicants for Game", operation_description="List the applicants for the current game. (GM Only)", manual_parameters=[parent_lookup_game__slug], ), ) @method_decorator( name="retrieve", decorator=swagger_auto_schema( operation_summary="Game Applicant: Details", operation_description="Fetch details for a given game application. (GM Only)", manual_parameters=[parent_lookup_game__slug], reponses={ 200: serializers.GameApplicationGMSerializer, 403: "You are not the GM for this game.", }, ), ) @method_decorator( name="approve", decorator=swagger_auto_schema( operation_summary="Game Applicant: Approve", operation_description="Approve the game applicant and add as a player to game.", request_body=no_body, responses={ 201: serializers.PlayerSerializer, 403: "You are not the GM of this game.", }, ), ) @method_decorator( name="reject", decorator=swagger_auto_schema( operation_summary="Game Applicant: Reject", operation_description="Reject the game applicant.", request_body=no_body, responses={ 200: serializers.GameApplicationGMSerializer, 403: "You are not the GM of this game.", }, ), ) class GMGameApplicationViewSet( ParentObjectAutoPermissionViewSetMixin, NestedViewSetMixin, mixins.ListModelMixin, mixins.RetrieveModelMixin, viewsets.GenericViewSet, ): """ View for a GM to review and approve applicants. """ permission_classes = (IsAuthenticated,) serializer_class = serializers.GameApplicationGMSerializer filter_backends = [DjangoFilterBackend] filterset_fields = ["status"] lookup_field = "slug" lookup_url_kwarg = "slug" parent_lookup_field = "game" parent_object_lookup_field = "slug" parent_object_model = models.GamePosting parent_object_url_kwarg = "parent_lookup_game__slug" parent_dependent_actions = ["list", "retrieve", "approve", "reject"] permission_type_map = { **ParentObjectAutoPermissionViewSetMixin.permission_type_map, "approve": "approve", "reject": "approve", } permission_type_map["retrieve"] = "approve" permission_type_map["list"] = "approve" def get_queryset(self): return models.GamePostingApplication.objects.filter( game__slug=self.kwargs["parent_lookup_game__slug"] ).exclude(status="new") def get_parent_game(self): return get_object_or_404( models.GamePosting, slug=self.kwargs["parent_lookup_game__slug"] ) @action(methods=["post"], detail=True) def approve(self, request, *args, **kwargs): """ Approves the game application. """ obj = self.get_object() obj.status = "approve" player = models.Player.objects.create(game=obj.game, gamer=obj.gamer) obj.save() return Response( data=serializers.PlayerSerializer( player, context={"request", request} ).data, status=status.HTTP_201_CREATED, ) @action(methods=["post"], detail=True) def reject(self, request, *args, **kwargs): """ Rejects the game application. """ obj = self.get_object() obj.status = "deny" obj.save() notify.send( obj, recipient=obj.gamer.user, verb="Your player application was not accepted", action_object=obj, target=obj.game, ) return Response( data=serializers.GameApplicationSerializer( obj, context={"request": request} ).data, status=status.HTTP_200_OK, ) @method_decorator( name="list", decorator=swagger_auto_schema( operation_summary="Game: Player List", operation_description="List players for a given game", manual_parameters=[parent_lookup_game__slug], ), ) @method_decorator( name="retrieve", decorator=swagger_auto_schema( operation_summary="Player: Details", operation_description="Details for a player record in a given game.", manual_parameters=[parent_lookup_game__slug], responses={ 200: serializers.PlayerSerializer, 403: "You are not a member of this game.", }, ), ) @method_decorator( name="kick", decorator=swagger_auto_schema( operation_summary="Player: Kick from game", operation_description="Kick the player out of the game.", manual_parameters=[parent_lookup_game__slug], request_body=no_body, responses={ 204: "Player was removed from the game.", 403: "You are not the GM of this game.", }, ), ) class PlayerViewSet( ParentObjectAutoPermissionViewSetMixin, NestedViewSetMixin, mixins.ListModelMixin, mixins.RetrieveModelMixin, viewsets.GenericViewSet, ): """ Provides views for players in a given game. """ permission_classes = (IsAuthenticated,) serializer_class = serializers.PlayerSerializer permission_required = "game.is_member" lookup_field = "slug" lookup_url_kwarg = "slug" parent_lookup_field = "game" parent_object_model = models.GamePosting parent_object_lookup_field = "slug" parent_object_url_kwarg = "parent_lookup_game__slug" parent_dependent_actions = ["list", "retrieve"] permission_type_map = {**ParentObjectAutoPermissionViewSetMixin.permission_type_map} permission_type_map["list"] = "view" def get_parent_game(self): return get_object_or_404( models.GamePosting, slug=self.kwargs["parent_lookup_game__slug"] ) def get_queryset(self): return models.Player.objects.filter(game=self.get_parent_game()) @action(methods=["post"], detail=True) def kick(self, request, *args, **kwargs): obj = self.get_object() player_kicked.send(request.user, player=obj) obj.delete() return Response(status=status.HTTP_204_NO_CONTENT) @method_decorator( name="list", decorator=swagger_auto_schema( operation_summary="Game: List Characters", operation_description="Fetch the list of characters for a given game.", manual_parameters=[parent_lookup_game__slug], ), ) @method_decorator( name="retrieve", decorator=swagger_auto_schema( operation_summary="Game: Character Details", operation_description="Fetch the details of a character for a given game.", manual_parameters=[parent_lookup_game__slug], responses={ 200: serializers.CharacterSerializer, 403: "You are not a member of this game.", }, ), ) @method_decorator( name="update", decorator=swagger_auto_schema( operation_summary="Game: Update Character Details", operation_description="Update the character for the given game.", manual_parameters=[parent_lookup_game__slug], request_body=serializers.CharacterSerializer, responses={ 200: serializers.CharacterSerializer, 403: "You are not the owner of this character or the GM of the game.", }, ), ) @method_decorator( name="partial_update", decorator=swagger_auto_schema( operation_summary="Game: Update Character Details", operation_description="Update the character for the given game.", manual_parameters=[parent_lookup_game__slug], request_body=serializers.CharacterSerializer, responses={ 200: serializers.CharacterSerializer, 403: "You are not the owner of this character or the GM of the game.", }, ), ) @method_decorator( name="deactivate", decorator=swagger_auto_schema( operation_summary="Game: Deactivate Character", operation_description="Mark the character as inactive.", manual_parameters=[parent_lookup_game__slug], request_body=no_body, responses={ 200: serializers.CharacterSerializer, 400: "This character is already inactive.", 403: "You are not the owner of this character or the GM of the game.", }, ), ) @method_decorator( name="reactivate", decorator=swagger_auto_schema( operation_summary="Game: Reactivate Character", operation_description="Mark the character as active.", manual_parameters=[parent_lookup_game__slug], request_body=no_body, responses={ 200: serializers.CharacterSerializer, 400: "This character is already active.", 403: "You are not the owner of this character or the GM of the game.", }, ), ) @method_decorator( name="destroy", decorator=swagger_auto_schema( operation_summary="Game: Delete Character", operation_description="Delete the character.", manual_parameters=[parent_lookup_game__slug], request_body=no_body, responses={ 204: "Character was deleted.", 403: "You are not the owner of this character.", }, ), ) @method_decorator( name="approve", decorator=swagger_auto_schema( operation_summary="Game: Approve Character", operation_description="Mark the character as approved (GM Only).", manual_parameters=[parent_lookup_game__slug], request_body=no_body, responses={ 200: serializers.CharacterSerializer, 400: "This character is already approved.", 403: "You are not the GM of the game.", }, ), ) @method_decorator( name="reject", decorator=swagger_auto_schema( operation_summary="Game: Reject Character", operation_description="Mark the character as rejected (GM Only).", manual_parameters=[parent_lookup_game__slug], request_body=no_body, responses={ 200: serializers.CharacterSerializer, 400: "This character is already rejected.", 403: "You are not the GM of the game.", }, ), ) class CharacterViewSet( ParentObjectAutoPermissionViewSetMixin, NestedViewSetMixin, viewsets.ModelViewSet ): """ Provides views for the characters in a game. """ permission_classes = (IsAuthenticated,) parser_classes = [FormParser, MultiPartParser] parent_object_lookup_field = "slug" parent_object_url_kwarg = "parent_lookup_game__slug" parent_lookup_field = "game" parent_object_model = models.GamePosting parent_dependent_actions = ["create", "list", "retrieve"] serializer_class = serializers.CharacterSerializer lookup_field = "slug" lookup_url_kwarg = "slug" filter_backends = [DjangoFilterBackend] filterset_fields = ["status"] parent_game = None permission_type_map = { **ParentObjectAutoPermissionViewSetMixin.permission_type_map, "approve": "approve", "reject": "approve", "deactivate": "delete", "reactivate": "delete", } permission_type_map["list"] = "gamelist" def get_parent_game(self): if not self.parent_game: self.parent_game = get_object_or_404( models.GamePosting, slug=self.kwargs["parent_lookup_game__slug"] ) return self.parent_game def get_queryset(self): return models.Character.objects.filter(game=self.get_parent_game()) def create(self, request, *args, **kwargs): if request.user.gamerprofile == self.get_parent_game().gm: return Response( data={"errors": "Only a player can create a character."}, status=status.HTTP_403_FORBIDDEN, ) char_ser = serializers.CharacterSerializer( data=request.data, context={"request": request, "game": self.get_parent_game()}, ) if not char_ser.is_valid(): return Response(data=char_ser.errors, status=status.HTTP_400_BAD_REQUEST) char_ser.save() return Response(data=char_ser.data, status=status.HTTP_201_CREATED) @action(methods=["post"], detail=True, parser_classes=[FormParser, JSONParser]) def approve(self, request, *args, **kwargs): """ Approves the proposed character. """ obj = self.get_object() obj.status = "approved" obj.save() return Response( data=self.serializer_class(obj, context={"request": request}).data, status=status.HTTP_200_OK, ) @action(methods=["post"], detail=True, parser_classes=[FormParser, JSONParser]) def reject(self, request, *args, **kwargs): """ Rejects the proposed character. """ obj = self.get_object() obj.status = "rejected" obj.save() return Response( data=self.serializer_class(obj, context={"request": request}).data, status=status.HTTP_200_OK, ) @action(methods=["post"], detail=True, parser_classes=[FormParser, JSONParser]) def deactivate(self, request, *args, **kwargs): """ Make a character inactive. """ obj = self.get_object() obj.status = "inactive" obj.save() return Response( data=self.serializer_class(obj, context={"request": request}).data, status=status.HTTP_200_OK, ) @action(methods=["post"], detail=True, parser_classes=[FormParser, JSONParser]) def reactivate(self, request, *args, **kwargs): """ Reactivate an inactive character. """ obj = self.get_object() obj.status = "pending" obj.save() return Response( data=self.serializer_class(obj, context={"request": request}).data, status=status.HTTP_200_OK, ) @method_decorator( name="list", decorator=swagger_auto_schema( operation_summary="List Your Characters", operation_description="Fetch a list of all of your characters.", ), ) @method_decorator( name="retrieve", decorator=swagger_auto_schema( operation_summary="Your Character: Details", operation_description="Fetch the details of your character.", ), ) @method_decorator( name="update", decorator=swagger_auto_schema( operation_summary="Your Character: Update", operation_description="Update the details of your character.", ), ) @method_decorator( name="partial_update", decorator=swagger_auto_schema( operation_summary="Your Character: Update", operation_description="Update the details of your character.", ), ) @method_decorator( name="destroy", decorator=swagger_auto_schema( operation_summary="Your Character: Delete", operation_description="Delete your character.", request_body=no_body, responses={204: "Character was deleted."}, ), ) @method_decorator( name="deactivate", decorator=swagger_auto_schema( operation_summary="Your Character: Deactivate", operation_description="Mark your character as inactive.", request_body=no_body, responses={ 200: "Character was marked as inactive.", 400: "Character was already inactive.", }, ), ) @method_decorator( name="reactivate", decorator=swagger_auto_schema( operation_summary="Your Character: Reactivate", operation_description="Mark your character as active.", request_body=no_body, responses={ 200: "Character was marked as active.", 400: "Character was already active.", }, ), ) class MyCharacterViewSet( AutoPermissionViewSetMixin, NestedViewSetMixin, mixins.ListModelMixin, mixins.RetrieveModelMixin, mixins.UpdateModelMixin, mixins.DestroyModelMixin, viewsets.GenericViewSet, ): """ Provides a vew so that players can view all their characters in one place. """ serializer_class = serializers.CharacterSerializer permission_classes = (IsAuthenticated,) lookup_field = "slug" lookup_url_kwarg = "slug" filter_backends = [DjangoFilterBackend] filterset_fields = ["status"] permission_type_map = { **AutoPermissionViewSetMixin.permission_type_map, "deactivate": "delete", "reactivate": "delete", } permission_type_map["retrieve"] = "delete" parser_classes = [FormParser, MultiPartParser] def get_queryset(self): return models.Character.objects.filter( player__gamer=self.request.user.gamerprofile ) @action(methods=["post"], detail=True, parser_classes=[FormParser, JSONParser]) def deactivate(self, request, *args, **kwargs): """ Make a character inactive. """ obj = self.get_object() obj.status = "inactive" obj.save() return Response( data=self.serializer_class(obj, context={"request": request}).data, status=status.HTTP_200_OK, ) @action(methods=["post"], detail=True, parser_classes=[FormParser, JSONParser]) def reactivate(self, request, *args, **kwargs): """ Reactivate an inactive character. """ obj = self.get_object() obj.status = "pending" obj.save() return Response( data=self.serializer_class(obj, context={"request": request}).data, status=status.HTTP_200_OK, ) @dparser_classes([FormParser, JSONParser]) def destroy(self, request, *args, **kwargs): self.parser_classes = [FormParser, JSONParser] return super().destroy(request, *args, **kwargs)
[ [ [ 7, 14 ], [ 1015, 1022 ] ], [ [ 57, 58 ], [ 5921, 5922 ], [ 5958, 5959 ], [ 5978, 5979 ], [ 6028, 6029 ], [ 6076, 6077 ], [ 6113, 6114 ], [ 6161, 6162 ] ], [ [ 88, 105 ], [ 15326, 15343 ], [ 28403, 28420 ], [ 31567, 31584 ], [ 37176, 37193 ] ], [ [ 142, 158 ], [ 1647, 1663 ], [ 1893, 1909 ], [ 2187, 2203 ], [ 2636, 2652 ], [ 3035, 3051 ], [ 3442, 3458 ], [ 3764, 3780 ], [ 4184, 4200 ], [ 8882, 8898 ], [ 9136, 9152 ], [ 9631, 9647 ], [ 10082, 10098 ], [ 10541, 10557 ], [ 10969, 10985 ], [ 11447, 11463 ], [ 11915, 11931 ], [ 12412, 12428 ], [ 12909, 12925 ], [ 13487, 13503 ], [ 20194, 20210 ], [ 20674, 20690 ], [ 21229, 21245 ], [ 21792, 21808 ], [ 23404, 23420 ], [ 23618, 23634 ], [ 23839, 23855 ], [ 24058, 24074 ], [ 24285, 24301 ], [ 25724, 25740 ], [ 25997, 26013 ], [ 26417, 26433 ], [ 26803, 26819 ], [ 29667, 29683 ], [ 29911, 29927 ], [ 30305, 30321 ], [ 32033, 32049 ], [ 32298, 32314 ], [ 32709, 32725 ], [ 33197, 33213 ], [ 33693, 33709 ], [ 34204, 34220 ], [ 34711, 34727 ], [ 35120, 35136 ], [ 35604, 35620 ], [ 39844, 39860 ], [ 40047, 40063 ], [ 40254, 40270 ], [ 40459, 40475 ], [ 40672, 40688 ], [ 40944, 40960 ], [ 41319, 41335 ] ], [ [ 201, 220 ], [ 5247, 5266 ], [ 24965, 24984 ], [ 27557, 27576 ], [ 36730, 36749 ], [ 42192, 42211 ] ], [ [ 242, 249 ], [ 1162, 1169 ], [ 1360, 1367 ], [ 1573, 1580 ] ], [ [ 279, 288 ], [ 1068, 1077 ], [ 1263, 1272 ], [ 1470, 1479 ] ], [ [ 316, 323 ], [ 3658, 3665 ], [ 3970, 3977 ], [ 11224, 11231 ], [ 11708, 11715 ], [ 12189, 12196 ], [ 12693, 12700 ], [ 22127, 22134 ], [ 26653, 26660 ], [ 27008, 27015 ], [ 30568, 30575 ], [ 33965, 33972 ], [ 34474, 34481 ], [ 34966, 34973 ], [ 35396, 35403 ], [ 35878, 35885 ], [ 40874, 40881 ], [ 41163, 41170 ], [ 41536, 41543 ] ], [ [ 325, 344 ], [ 1696, 1715 ], [ 1944, 1963 ], [ 2240, 2259 ], [ 2687, 2706 ], [ 3094, 3113 ], [ 3494, 3513 ], [ 3814, 3833 ], [ 4234, 4253 ], [ 8931, 8950 ], [ 9189, 9208 ], [ 9682, 9701 ], [ 10141, 10160 ], [ 10593, 10612 ], [ 11020, 11039 ], [ 11500, 11519 ], [ 11968, 11987 ], [ 12467, 12486 ], [ 12964, 12983 ], [ 13538, 13557 ], [ 20247, 20266 ], [ 20725, 20744 ], [ 21288, 21307 ], [ 21844, 21863 ], [ 23453, 23472 ], [ 23671, 23690 ], [ 23890, 23909 ], [ 24117, 24136 ], [ 24337, 24356 ], [ 25773, 25792 ], [ 26050, 26069 ], [ 26469, 26488 ], [ 26854, 26873 ], [ 29716, 29735 ], [ 29964, 29983 ], [ 30354, 30373 ], [ 32082, 32101 ], [ 32351, 32370 ], [ 32760, 32779 ], [ 33256, 33275 ], [ 33748, 33767 ], [ 34259, 34278 ], [ 34763, 34782 ], [ 35172, 35191 ], [ 35655, 35674 ], [ 39893, 39912 ], [ 40100, 40119 ], [ 40305, 40324 ], [ 40518, 40537 ], [ 40724, 40743 ], [ 40999, 41018 ], [ 41374, 41393 ] ], [ [ 379, 385 ], [ 7891, 7897 ], [ 29265, 29271 ] ], [ [ 413, 419 ], [ 14159, 14165 ], [ 14186, 14192 ], [ 14217, 14223 ], [ 14246, 14252 ], [ 22414, 22420 ], [ 22445, 22451 ], [ 22474, 22480 ], [ 24582, 24588 ], [ 24609, 24615 ], [ 24640, 24646 ], [ 24669, 24675 ], [ 27272, 27278 ], [ 27299, 27305 ], [ 30817, 30823 ], [ 30844, 30850 ], [ 41773, 41779 ], [ 41800, 41806 ], [ 41831, 41837 ], [ 41860, 41866 ] ], [ [ 421, 427 ], [ 7309, 7315 ], [ 7611, 7617 ], [ 8255, 8261 ], [ 8612, 8618 ], [ 8851, 8857 ], [ 16218, 16224 ], [ 16610, 16616 ], [ 16848, 16854 ], [ 17266, 17272 ], [ 17484, 17490 ], [ 17870, 17876 ], [ 18087, 18093 ], [ 18425, 18431 ], [ 18613, 18619 ], [ 18990, 18996 ], [ 19180, 19186 ], [ 19589, 19595 ], [ 19903, 19909 ], [ 20156, 20162 ], [ 28989, 28995 ], [ 29634, 29640 ], [ 32002, 32008 ], [ 37669, 37675 ], [ 37970, 37976 ], [ 38074, 38080 ], [ 38505, 38511 ], [ 38939, 38945 ], [ 39372, 39378 ], [ 39811, 39817 ], [ 43048, 43054 ], [ 43487, 43493 ] ], [ [ 429, 437 ], [ 4801, 4809 ], [ 14276, 14284 ], [ 22504, 22512 ], [ 24699, 24707 ], [ 27330, 27338 ], [ 30875, 30883 ], [ 36173, 36181 ], [ 41890, 41898 ] ], [ [ 476, 482 ], [ 6910, 6916 ], [ 8296, 8302 ], [ 15928, 15934 ], [ 16884, 16890 ], [ 17520, 17526 ], [ 18123, 18129 ], [ 18649, 18655 ], [ 19216, 19222 ], [ 28515, 28521 ], [ 29030, 29036 ], [ 31781, 31787 ], [ 38105, 38111 ], [ 38541, 38547 ], [ 38975, 38981 ], [ 39408, 39414 ], [ 42651, 42657 ], [ 43084, 43090 ] ], [ [ 521, 554 ], [ 43523, 43538 ] ], [ [ 590, 600 ], [ 4989, 4999 ], [ 6964, 6974 ], [ 8350, 8360 ], [ 36330, 36340 ], [ 38159, 38169 ], [ 38595, 38605 ], [ 39029, 39039 ], [ 39462, 39472 ], [ 42472, 42482 ], [ 42705, 42715 ], [ 43138, 43148 ], [ 43540, 43550 ], [ 43645, 43655 ] ], [ [ 602, 612 ], [ 6976, 6986 ], [ 8362, 8372 ], [ 38171, 38181 ], [ 38607, 38617 ], [ 39041, 39051 ], [ 39474, 39484 ], [ 42717, 42727 ], [ 43150, 43160 ], [ 43552, 43562 ], [ 43657, 43667 ] ], [ [ 614, 629 ], [ 5001, 5016 ], [ 36342, 36357 ], [ 42484, 42499 ] ], [ [ 669, 684 ], [ 4949, 4964 ], [ 24863, 24878 ], [ 27453, 27468 ], [ 30994, 31009 ], [ 36290, 36305 ], [ 42095, 42110 ] ], [ [ 721, 729 ], [ 7208, 7216 ], [ 7549, 7557 ], [ 8100, 8108 ], [ 8513, 8521 ], [ 8835, 8843 ], [ 16156, 16164 ], [ 16360, 16368 ], [ 16739, 16747 ], [ 17072, 17080 ], [ 17375, 17383 ], [ 17698, 17706 ], [ 17978, 17986 ], [ 18309, 18317 ], [ 18504, 18512 ], [ 18823, 18831 ], [ 19071, 19079 ], [ 19477, 19485 ], [ 19842, 19850 ], [ 20000, 20008 ], [ 28840, 28848 ], [ 29479, 29487 ], [ 31986, 31994 ], [ 37562, 37570 ], [ 37932, 37940 ], [ 38038, 38046 ], [ 38396, 38404 ], [ 38830, 38838 ], [ 39263, 39271 ], [ 39702, 39710 ], [ 42939, 42947 ], [ 43378, 43386 ] ], [ [ 775, 796 ], [ 4750, 4771 ] ], [ [ 798, 816 ], [ 4777, 4795 ], [ 14135, 14153 ], [ 22390, 22408 ], [ 27248, 27266 ], [ 30793, 30811 ], [ 36153, 36171 ], [ 41749, 41767 ] ], [ [ 855, 881 ], [ 4718, 4744 ], [ 5474, 5500 ], [ 24550, 24576 ], [ 25105, 25131 ], [ 41717, 41743 ], [ 42285, 42311 ] ], [ [ 883, 921 ], [ 14091, 14129 ], [ 14994, 15032 ], [ 22346, 22384 ], [ 23002, 23040 ], [ 27204, 27242 ], [ 27954, 27992 ], [ 30749, 30787 ], [ 31419, 31457 ], [ 36113, 36151 ], [ 36846, 36884 ] ], [ [ 937, 943 ], [ 5030, 5036 ], [ 14373, 14379 ], [ 14840, 14846 ], [ 22603, 22609 ], [ 22691, 22697 ], [ 27767, 27773 ], [ 31222, 31228 ], [ 36515, 36521 ], [ 5831, 5837 ], [ 6200, 6206 ], [ 7667, 7673 ], [ 8672, 8678 ], [ 8766, 8772 ], [ 15357, 15363 ], [ 23286, 23292 ], [ 25382, 25388 ], [ 28215, 28221 ], [ 28434, 28440 ], [ 28745, 28751 ], [ 31598, 31604 ], [ 31717, 31723 ], [ 37211, 37217 ], [ 37366, 37372 ], [ 42545, 42551 ] ], [ [ 945, 956 ], [ 2088, 2099 ], [ 2144, 2155 ], [ 2521, 2532 ], [ 2860, 2871 ], [ 2929, 2940 ], [ 3267, 3278 ], [ 3336, 3347 ], [ 4375, 4386 ], [ 4451, 4462 ], [ 5128, 5139 ], [ 5193, 5204 ], [ 9518, 9529 ], [ 9897, 9908 ], [ 9971, 9982 ], [ 10356, 10367 ], [ 10430, 10441 ], [ 10797, 10808 ], [ 11270, 11281 ], [ 11754, 11765 ], [ 12235, 12246 ], [ 12739, 12750 ], [ 13197, 13208 ], [ 13266, 13277 ], [ 13766, 13777 ], [ 13839, 13850 ], [ 14415, 14426 ], [ 20562, 20573 ], [ 21024, 21035 ], [ 21097, 21108 ], [ 21587, 21598 ], [ 21660, 21671 ], [ 22839, 22850 ], [ 24904, 24915 ], [ 26301, 26312 ], [ 26699, 26710 ], [ 27054, 27065 ], [ 27494, 27505 ], [ 30199, 30210 ], [ 31035, 31046 ], [ 32600, 32611 ], [ 32990, 33001 ], [ 33060, 33071 ], [ 33486, 33497 ], [ 33556, 33567 ], [ 34011, 34022 ], [ 34520, 34531 ], [ 35442, 35453 ], [ 35924, 35935 ], [ 36619, 36630 ], [ 42037, 42048 ], [ 6427, 6438 ], [ 6811, 6822 ], [ 7378, 7389 ], [ 8127, 8138 ], [ 15828, 15839 ], [ 16044, 16055 ], [ 19658, 19669 ], [ 20027, 20038 ], [ 28867, 28878 ], [ 29506, 29517 ], [ 37729, 37740 ] ], [ [ 978, 991 ], [ 31905, 31918 ] ], [ [ 993, 1004 ], [ 8749, 8760 ] ], [ [ 1006, 1012 ], [ 6652, 6658 ], [ 7076, 7082 ], [ 25190, 25196 ], [ 25299, 25305 ], [ 25539, 25545 ] ], [ [ 1041, 1065 ], [ 9099, 9123 ], [ 9454, 9478 ], [ 9849, 9873 ], [ 10308, 10332 ], [ 10749, 10773 ], [ 11176, 11200 ], [ 11660, 11684 ], [ 12141, 12165 ], [ 12645, 12669 ], [ 13149, 13173 ], [ 13718, 13742 ], [ 25960, 25984 ], [ 26238, 26262 ], [ 29874, 29898 ], [ 30135, 30159 ], [ 30520, 30544 ], [ 32261, 32285 ], [ 32536, 32560 ], [ 32942, 32966 ], [ 33438, 33462 ], [ 33917, 33941 ], [ 34426, 34450 ], [ 34918, 34942 ], [ 35348, 35372 ], [ 35830, 35854 ] ], [ [ 1233, 1260 ], [ 20485, 20512 ], [ 20963, 20990 ], [ 21526, 21553 ], [ 22066, 22093 ] ], [ [ 1434, 1467 ], [ 20438, 20471 ], [ 20916, 20949 ], [ 21479, 21512 ], [ 22019, 22052 ] ], [ [ 4694, 4712 ] ], [ [ 14067, 14085 ] ], [ [ 22321, 22340 ] ], [ [ 24522, 24544 ] ], [ [ 27174, 27198 ] ], [ [ 30730, 30743 ] ], [ [ 36091, 36107 ] ], [ [ 41693, 41711 ] ] ]
"""Mypy style test cases for SQLAlchemy stubs and plugin.""" import os import os.path import sys import pytest # type: ignore # no pytest in typeshed from mypy.test.config import test_temp_dir from mypy.test.data import DataDrivenTestCase, DataSuite from mypy.test.helpers import assert_string_arrays_equal from mypy.util import try_find_python2_interpreter from mypy import api this_file_dir = os.path.dirname(os.path.realpath(__file__)) prefix = os.path.dirname(this_file_dir) inipath = os.path.abspath(os.path.join(prefix, 'test')) # Locations of test data files such as test case descriptions (.test). test_data_prefix = os.path.join(prefix, 'test', 'test-data') class SQLDataSuite(DataSuite): files = ['sqlalchemy-basics.test', 'sqlalchemy-sql-elements.test', 'sqlalchemy-sql-sqltypes.test', 'sqlalchemy-sql-selectable.test', 'sqlalchemy-sql-schema.test', 'sqlalchemy-plugin-features.test', 'sqlalchemy-plugin-query.test'] data_prefix = test_data_prefix def run_case(self, testcase: DataDrivenTestCase) -> None: assert testcase.old_cwd is not None, "test was not properly set up" mypy_cmdline = [ '--show-traceback', '--no-silence-site-packages', '--config-file={}/sqlalchemy.ini'.format(inipath), ] py2 = testcase.name.lower().endswith('python2') if py2: if try_find_python2_interpreter() is None: pytest.skip() return mypy_cmdline.append('--py2') else: mypy_cmdline.append('--python-version={}'.format('.'.join(map(str, sys.version_info[:2])))) # Write the program to a file. program_path = os.path.join(test_temp_dir, 'main.py') mypy_cmdline.append(program_path) with open(program_path, 'w') as file: for s in testcase.input: file.write('{}\n'.format(s)) output = [] # Type check the program. out, err, returncode = api.run(mypy_cmdline) # split lines, remove newlines, and remove directory of test case for line in (out + err).splitlines(): if line.startswith(test_temp_dir + os.sep): output.append(line[len(test_temp_dir + os.sep):].rstrip("\r\n").replace('.py', '')) else: output.append(line.rstrip("\r\n")) # Remove temp file. os.remove(program_path) assert_string_arrays_equal(testcase.output, output, 'Invalid output ({}, line {})'.format( testcase.file, testcase.line))
[ [ [ 69, 71 ] ], [ [ 79, 86 ], [ 401, 403 ], [ 417, 419 ], [ 454, 456 ], [ 495, 497 ], [ 511, 513 ], [ 632, 634 ], [ 1842, 1844 ], [ 2325, 2327 ], [ 2389, 2391 ], [ 2627, 2629 ] ], [ [ 94, 97 ], [ 1754, 1757 ] ], [ [ 106, 112 ], [ 1509, 1515 ] ], [ [ 184, 197 ], [ 1855, 1868 ], [ 2309, 2322 ], [ 2373, 2386 ] ], [ [ 225, 243 ], [ 1088, 1106 ] ], [ [ 245, 254 ], [ 695, 704 ] ], [ [ 285, 311 ], [ 2659, 2685 ] ], [ [ 334, 362 ], [ 1453, 1481 ] ], [ [ 380, 383 ], [ 2136, 2139 ] ], [ [ 385, 398 ], [ 470, 483 ] ], [ [ 445, 451 ], [ 524, 530 ], [ 645, 651 ] ], [ [ 485, 492 ], [ 1346, 1353 ] ], [ [ 613, 629 ], [ 1037, 1053 ] ], [ [ 682, 694 ] ] ]
#!/usr/bin/env python # Copyright (C) 2017 The Android Open Source Project # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # Generates the text to paste into the email for announcing a new # release of Gerrit. The text is generated based on a template that # is filled with values either passed to the script or calculated # at runtime. # # The script outputs a plain text file with the announcement text: # # release-announcement-gerrit-X.Y.txt # # and, if GPG is available, the announcement text wrapped with a # signature: # # release-announcement-gerrit-X.Y.txt.asc # # Usage: # # ./tools/release-announcement.py -v 2.14.2 -p 2.14.1 \ # -s "This release fixes several bugs since 2.14.1" # # Parameters: # # --version (-v): The version of Gerrit being released. # # --previous (-p): The previous version of Gerrit. Optional. If # specified, the generated text includes a link to the gitiles # log of commits between the previous and new versions. # # --summary (-s): Short summary of the release. Optional. When # specified, the summary is inserted in the introductory sentence # of the generated text. # # Prerequisites: # # - The Jinja2 python library [1] must be installed. # # - For GPG signing to work, the python-gnupg library [2] must be # installed, and the ~/.gnupg folder must exist. # # - The war file must have been installed to the local Maven repository # using the `./tools/mvn/api.sh war_install` command. # # [1] http://jinja.pocoo.org/ # [2] http://pythonhosted.org/gnupg/ from __future__ import print_function import argparse import hashlib import os import sys from gnupg import GPG from jinja2 import Template class Version: def __init__(self, version): self.version = version parts = version.split('.') if len(parts) > 2: self.major = ".".join(parts[:2]) self.patch = version else: self.major = version self.patch = None def __str__(self): return self.version def _main(): descr = 'Generate Gerrit release announcement email text' parser = argparse.ArgumentParser( description=descr, formatter_class=argparse.ArgumentDefaultsHelpFormatter) parser.add_argument('-v', '--version', dest='version', required=True, help='gerrit version to release') parser.add_argument('-p', '--previous', dest='previous', help='previous gerrit version (optional)') parser.add_argument('-s', '--summary', dest='summary', help='summary of the release content (optional)') options = parser.parse_args() summary = options.summary if summary and not summary.endswith("."): summary = summary + "." data = { "version": Version(options.version), "previous": options.previous, "summary": summary } war = os.path.join( os.path.expanduser("~/.m2/repository/com/google/gerrit/gerrit-war/"), "%(version)s/gerrit-war-%(version)s.war" % data) if not os.path.isfile(war): print("Could not find war file for Gerrit %s in local Maven repository" % data["version"], file=sys.stderr) sys.exit(1) md5 = hashlib.md5() sha1 = hashlib.sha1() sha256 = hashlib.sha256() BUF_SIZE = 65536 # Read data in 64kb chunks with open(war, 'rb') as f: while True: d = f.read(BUF_SIZE) if not d: break md5.update(d) sha1.update(d) sha256.update(d) data["sha1"] = sha1.hexdigest() data["sha256"] = sha256.hexdigest() data["md5"] = md5.hexdigest() template = Template(open("tools/release-announcement-template.txt").read()) output = template.render(data=data) filename = "release-announcement-gerrit-%s.txt" % data["version"] with open(filename, "w") as f: f.write(output) gpghome = os.path.abspath(os.path.expanduser("~/.gnupg")) if not os.path.isdir(gpghome): print("Skipping signing due to missing gnupg home folder") else: try: gpg = GPG(homedir=gpghome) except TypeError: gpg = GPG(gnupghome=gpghome) signed = gpg.sign(output) filename = filename + ".asc" with open(filename, "w") as f: f.write(str(signed)) if __name__ == "__main__": _main()
[ [ [ 2044, 2058 ] ], [ [ 2066, 2074 ], [ 2601, 2609 ], [ 2677, 2685 ] ], [ [ 2082, 2089 ], [ 3763, 3770 ], [ 3788, 3795 ], [ 3816, 3823 ] ], [ [ 2097, 2099 ], [ 3421, 3423 ], [ 3443, 3445 ], [ 3581, 3583 ], [ 4469, 4471 ], [ 4485, 4487 ], [ 4528, 4530 ] ], [ [ 2107, 2110 ], [ 3720, 3723 ], [ 3740, 3743 ] ], [ [ 2129, 2132 ], [ 4660, 4663 ], [ 4725, 4728 ] ], [ [ 2152, 2160 ], [ 4219, 4227 ] ], [ [ 2169, 2176 ], [ 3311, 3318 ] ], [ [ 2517, 2522 ], [ 4924, 4929 ] ] ]
# coding: utf-8 """ An API to insert and retrieve metadata on cloud artifacts. No description provided (generated by Swagger Codegen https://github.com/swagger-api/swagger-codegen) # noqa: E501 OpenAPI spec version: v1alpha1 Generated by: https://github.com/swagger-api/swagger-codegen.git """ import pprint import re # noqa: F401 import six class ApiArtifact(object): """NOTE: This class is auto generated by the swagger code generator program. Do not edit the class manually. """ """ Attributes: swagger_types (dict): The key is attribute name and the value is attribute type. attribute_map (dict): The key is attribute name and the value is json key in definition. """ swagger_types = { 'name': 'str', 'checksum': 'str', 'id': 'str', 'names': 'list[str]' } attribute_map = { 'name': 'name', 'checksum': 'checksum', 'id': 'id', 'names': 'names' } def __init__(self, name=None, checksum=None, id=None, names=None): # noqa: E501 """ApiArtifact - a model defined in Swagger""" # noqa: E501 self._name = None self._checksum = None self._id = None self._names = None self.discriminator = None if name is not None: self.name = name if checksum is not None: self.checksum = checksum if id is not None: self.id = id if names is not None: self.names = names @property def name(self): """Gets the name of this ApiArtifact. # noqa: E501 Name of the artifact. This may be the path to a binary or jar file, or in the case of a container build, the name used to push the container image to Google Container Registry, as presented to `docker push`. This field is deprecated in favor of the plural `names` field; it continues to exist here to allow existing BuildProvenance serialized to json in google.devtools.containeranalysis.v1alpha1.BuildDetails.provenance_bytes to deserialize back into proto. # noqa: E501 :return: The name of this ApiArtifact. # noqa: E501 :rtype: str """ return self._name @name.setter def name(self, name): """Sets the name of this ApiArtifact. Name of the artifact. This may be the path to a binary or jar file, or in the case of a container build, the name used to push the container image to Google Container Registry, as presented to `docker push`. This field is deprecated in favor of the plural `names` field; it continues to exist here to allow existing BuildProvenance serialized to json in google.devtools.containeranalysis.v1alpha1.BuildDetails.provenance_bytes to deserialize back into proto. # noqa: E501 :param name: The name of this ApiArtifact. # noqa: E501 :type: str """ self._name = name @property def checksum(self): """Gets the checksum of this ApiArtifact. # noqa: E501 Hash or checksum value of a binary, or Docker Registry 2.0 digest of a container. # noqa: E501 :return: The checksum of this ApiArtifact. # noqa: E501 :rtype: str """ return self._checksum @checksum.setter def checksum(self, checksum): """Sets the checksum of this ApiArtifact. Hash or checksum value of a binary, or Docker Registry 2.0 digest of a container. # noqa: E501 :param checksum: The checksum of this ApiArtifact. # noqa: E501 :type: str """ self._checksum = checksum @property def id(self): """Gets the id of this ApiArtifact. # noqa: E501 :return: The id of this ApiArtifact. # noqa: E501 :rtype: str """ return self._id @id.setter def id(self, id): """Sets the id of this ApiArtifact. :param id: The id of this ApiArtifact. # noqa: E501 :type: str """ self._id = id @property def names(self): """Gets the names of this ApiArtifact. # noqa: E501 Related artifact names. This may be the path to a binary or jar file, or in the case of a container build, the name used to push the container image to Google Container Registry, as presented to `docker push`. Note that a single Artifact ID can have multiple names, for example if two tags are applied to one image. # noqa: E501 :return: The names of this ApiArtifact. # noqa: E501 :rtype: list[str] """ return self._names @names.setter def names(self, names): """Sets the names of this ApiArtifact. Related artifact names. This may be the path to a binary or jar file, or in the case of a container build, the name used to push the container image to Google Container Registry, as presented to `docker push`. Note that a single Artifact ID can have multiple names, for example if two tags are applied to one image. # noqa: E501 :param names: The names of this ApiArtifact. # noqa: E501 :type: list[str] """ self._names = names def to_dict(self): """Returns the model properties as a dict""" result = {} for attr, _ in six.iteritems(self.swagger_types): value = getattr(self, attr) if isinstance(value, list): result[attr] = list(map( lambda x: x.to_dict() if hasattr(x, "to_dict") else x, value )) elif hasattr(value, "to_dict"): result[attr] = value.to_dict() elif isinstance(value, dict): result[attr] = dict(map( lambda item: (item[0], item[1].to_dict()) if hasattr(item[1], "to_dict") else item, value.items() )) else: result[attr] = value if issubclass(ApiArtifact, dict): for key, value in self.items(): result[key] = value return result def to_str(self): """Returns the string representation of the model""" return pprint.pformat(self.to_dict()) def __repr__(self): """For `print` and `pprint`""" return self.to_str() def __eq__(self, other): """Returns true if both objects are equal""" if not isinstance(other, ApiArtifact): return False return self.__dict__ == other.__dict__ def __ne__(self, other): """Returns true if both objects are not equal""" return not self == other
[ [ [ 328, 334 ], [ 6261, 6267 ] ], [ [ 342, 344 ] ], [ [ 367, 370 ], [ 5335, 5338 ] ], [ [ 379, 390 ], [ 6039, 6050 ], [ 6501, 6512 ] ] ]
# Licensed under a 3-clause BSD style license - see LICENSE.rst # -*- coding: utf-8 -*- from __future__ import absolute_import, division, unicode_literals, print_function import re import os import tempfile import six from .. import environment from ..console import log from .. import util WIN = (os.name == "nt") def _find_conda(): """Find the conda executable robustly across conda versions. Returns ------- conda : str Path to the conda executable. Raises ------ IOError If the executable cannot be found in either the CONDA_EXE environment variable or in the PATH. Notes ----- In POSIX platforms in conda >= 4.4, conda can be set up as a bash function rather than an executable. (This is to enable the syntax ``conda activate env-name``.) In this case, the environment variable ``CONDA_EXE`` contains the path to the conda executable. In other cases, we use standard search for the appropriate name in the PATH. See https://github.com/airspeed-velocity/asv/issues/645 for more details. """ if 'CONDA_EXE' in os.environ: conda = os.environ['CONDA_EXE'] else: conda = util.which('conda') return conda class Conda(environment.Environment): """ Manage an environment using conda. Dependencies are installed using ``conda``. The benchmarked project is installed using ``pip`` (since ``conda`` doesn't have a method to install from an arbitrary ``setup.py``). """ tool_name = "conda" _matches_cache = {} def __init__(self, conf, python, requirements): """ Parameters ---------- conf : Config instance python : str Version of Python. Must be of the form "MAJOR.MINOR". requirements : dict Dictionary mapping a PyPI package name to a version identifier string. """ self._python = python self._requirements = requirements self._conda_channels = conf.conda_channels super(Conda, self).__init__(conf, python, requirements) @classmethod def matches(cls, python): # Calling conda can take a long time, so remember the result if python not in cls._matches_cache: cls._matches_cache[python] = cls._matches(python) return cls._matches_cache[python] @classmethod def _matches(cls, python): if not re.match(r'^[0-9].*$', python): # The python name should be a version number return False try: conda = _find_conda() except IOError: return False else: # This directory never gets created, since we're just # doing a dry run below. All it needs to be is something # that doesn't already exist. path = os.path.join(tempfile.gettempdir(), 'check') # Check that the version number is valid try: util.check_call([ conda, 'create', '--yes', '-p', path, 'python={0}'.format(python), '--dry-run'], display_error=False, dots=False) except util.ProcessError: return False else: return True def _setup(self): try: conda = _find_conda() except IOError as e: raise util.UserError(str(e)) log.info("Creating conda environment for {0}".format(self.name)) # create a temporary environment.yml file # and use that to generate the env for benchmarking env_file = tempfile.NamedTemporaryFile(mode='w', delete=False, suffix=".yml") try: env_file.write('name: {0}\n' 'channels:\n'.format(self.name)) env_file.writelines((' - %s\n' % ch for ch in self._conda_channels)) env_file.write('dependencies:\n' ' - python={0}\n' ' - wheel\n' ' - pip\n'.format(self._python)) # categorize & write dependencies based on pip vs. conda conda_args, pip_args = self._get_requirements(conda) env_file.writelines((' - %s\n' % s for s in conda_args)) if pip_args: # and now specify the packages that are to be installed in # the pip subsection env_file.write(' - pip:\n') env_file.writelines((' - %s\n' % s for s in pip_args)) env_file.close() util.check_output([conda] + ['env', 'create', '-f', env_file.name, '-p', self._path, '--force']) except Exception as exc: if os.path.isfile(env_file.name): with open(env_file.name, 'r') as f: text = f.read() log.info("conda env create failed: in {} with:\n{}".format(self._path, text)) raise finally: os.unlink(env_file.name) def _get_requirements(self, conda): if self._requirements: # retrieve and return all conda / pip dependencies conda_args = [] pip_args = [] for key, val in six.iteritems(self._requirements): if key.startswith('pip+'): if val: pip_args.append("{0}=={1}".format(key[4:], val)) else: pip_args.append(key[4:]) else: if val: conda_args.append("{0}={1}".format(key, val)) else: conda_args.append(key) return conda_args, pip_args else: return [], [] def run(self, args, **kwargs): log.debug("Running '{0}' in {1}".format(' '.join(args), self.name)) return self.run_executable('python', args, **kwargs) def run_executable(self, executable, args, **kwargs): # Conda doesn't guarantee that user site directories are excluded kwargs["env"] = dict(kwargs.pop("env", os.environ), PYTHONNOUSERSITE=str("True")) return super(Conda, self).run_executable(executable, args, **kwargs)
[ [ [ 112, 127 ] ], [ [ 129, 137 ] ], [ [ 139, 155 ] ], [ [ 157, 171 ] ], [ [ 180, 182 ], [ 2443, 2445 ] ], [ [ 190, 192 ], [ 304, 306 ], [ 1117, 1119 ], [ 1145, 1147 ], [ 2865, 2867 ], [ 4879, 4881 ], [ 5139, 5141 ], [ 6263, 6265 ] ], [ [ 200, 208 ], [ 2878, 2886 ], [ 3726, 3734 ] ], [ [ 217, 220 ], [ 5382, 5385 ] ], [ [ 237, 248 ], [ 1246, 1257 ] ], [ [ 271, 274 ], [ 3531, 3534 ], [ 5014, 5017 ], [ 5954, 5957 ] ], [ [ 290, 294 ], [ 1195, 1199 ], [ 2997, 3001 ], [ 3288, 3292 ], [ 3499, 3503 ], [ 4693, 4697 ] ], [ [ 297, 300 ] ], [ [ 327, 338 ], [ 2591, 2602 ], [ 3438, 3449 ] ], [ [ 1240, 1245 ], [ 2063, 2068 ], [ 6356, 6361 ] ] ]
"""Timer class based on the timeit.Timer class, but torch aware.""" import enum import timeit import textwrap from typing import Any, Callable, Dict, List, NoReturn, Optional, Type, Union import numpy as np import torch from torch.utils.benchmark.utils import common, cpp_jit from torch.utils.benchmark.utils._stubs import TimerClass, TimeitModuleType from torch.utils.benchmark.utils.valgrind_wrapper import timer_interface as valgrind_timer_interface __all__ = ["Timer", "timer", "Language"] if torch.has_cuda and torch.cuda.is_available(): def timer() -> float: torch.cuda.synchronize() return timeit.default_timer() else: timer = timeit.default_timer class Language(enum.Enum): PYTHON = 0 CPP = 1 class CPPTimer: def __init__( self, stmt: str, setup: str, timer: Callable[[], float], globals: Dict[str, Any], ) -> None: if timer is not timeit.default_timer: raise NotImplementedError( "PyTorch was built with CUDA and a GPU is present; however " "Timer does not yet support GPU measurements. If your " "code is CPU only, pass `timer=timeit.default_timer` to the " "Timer's constructor to indicate this. (Note that this will " "produce incorrect results if the GPU is in fact used, as " "Timer will not synchronize CUDA.)" ) if globals: raise ValueError("C++ timing does not support globals.") self._stmt: str = textwrap.dedent(stmt) self._setup: str = textwrap.dedent(setup) self._timeit_module: Optional[TimeitModuleType] = None def timeit(self, number: int) -> float: if self._timeit_module is None: self._timeit_module = cpp_jit.compile_timeit_template( self._stmt, self._setup, ) return self._timeit_module.timeit(number) class Timer(object): """Helper class for measuring execution time of PyTorch statements. For a full tutorial on how to use this class, see: https://pytorch.org/tutorials/recipes/recipes/benchmark.html The PyTorch Timer is based on `timeit.Timer` (and in fact uses `timeit.Timer` internally), but with several key differences: 1) Runtime aware: Timer will perform warmups (important as some elements of PyTorch are lazily initialized), set threadpool size so that comparisons are apples-to-apples, and synchronize asynchronous CUDA functions when necessary. 2) Focus on replicates: When measuring code, and particularly complex kernels / models, run-to-run variation is a significant confounding factor. It is expected that all measurements should include replicates to quantify noise and allow median computation, which is more robust than mean. To that effect, this class deviates from the `timeit` API by conceptually merging `timeit.Timer.repeat` and `timeit.Timer.autorange`. (Exact algorithms are discussed in method docstrings.) The `timeit` method is replicated for cases where an adaptive strategy is not desired. 3) Optional metadata: When defining a Timer, one can optionally specify `label`, `sub_label`, `description`, and `env`. (Defined later) These fields are included in the representation of result object and by the `Compare` class to group and display results for comparison. 4) Instruction counts In addition to wall times, Timer can run a statement under Callgrind and report instructions executed. Directly analogous to `timeit.Timer` constructor arguments: `stmt`, `setup`, `timer`, `globals` PyTorch Timer specific constructor arguments: `label`, `sub_label`, `description`, `env`, `num_threads` Args: stmt: Code snippet to be run in a loop and timed. setup: Optional setup code. Used to define variables used in `stmt` timer: Callable which returns the current time. If PyTorch was built without CUDA or there is no GPU present, this defaults to `timeit.default_timer`; otherwise it will synchronize CUDA before measuring the time. globals: A dict which defines the global variables when `stmt` is being executed. This is the other method for providing variables which `stmt` needs. label: String which summarizes `stmt`. For instance, if `stmt` is "torch.nn.functional.relu(torch.add(x, 1, out=out))" one might set label to "ReLU(x + 1)" to improve readability. sub_label: Provide supplemental information to disambiguate measurements with identical stmt or label. For instance, in our example above sub_label might be "float" or "int", so that it is easy to differentiate: "ReLU(x + 1): (float)" "ReLU(x + 1): (int)" when printing Measurements or summarizing using `Compare`. description: String to distinguish measurements with identical label and sub_label. The principal use of `description` is to signal to `Compare` the columns of data. For instance one might set it based on the input size to create a table of the form: :: | n=1 | n=4 | ... ------------- ... ReLU(x + 1): (float) | ... | ... | ... ReLU(x + 1): (int) | ... | ... | ... using `Compare`. It is also included when printing a Measurement. env: This tag indicates that otherwise identical tasks were run in different environments, and are therefore not equivilent, for instance when A/B testing a change to a kernel. `Compare` will treat Measurements with different `env` specification as distinct when merging replicate runs. num_threads: The size of the PyTorch threadpool when executing `stmt`. Single threaded performace is important as both a key inference workload and a good indicator of intrinsic algorithmic efficiency, so the default is set to one. This is in contrast to the default PyTorch threadpool size which tries to utilize all cores. """ _timer_cls: Type[TimerClass] = timeit.Timer def __init__( self, stmt: str = "pass", setup: str = "pass", timer: Callable[[], float] = timer, globals: Optional[Dict[str, Any]] = None, label: Optional[str] = None, sub_label: Optional[str] = None, description: Optional[str] = None, env: Optional[str] = None, num_threads: int = 1, language: Union[Language, str] = Language.PYTHON, ): if not isinstance(stmt, str): raise ValueError("Currently only a `str` stmt is supported.") # We copy `globals` to prevent mutations from leaking. # (For instance, `eval` adds the `__builtins__` key) self._globals = dict(globals or {}) if language in (Language.PYTHON, "py", "python"): # Include `torch` if not specified as a convenience feature. self._globals.setdefault("torch", torch) self._language: Language = Language.PYTHON elif language in (Language.CPP, "cpp", "c++"): assert self._timer_cls is timeit.Timer, "_timer_cls has already been swapped." self._timer_cls = CPPTimer setup = ("" if setup == "pass" else setup) self._language = Language.CPP else: raise ValueError(f"Invalid language `{language}`.") # Convenience adjustment so that multi-line code snippets defined in # functions do not IndentationError (Python) or look odd (C++). The # leading newline removal is for the initial newline that appears when # defining block strings. For instance: # textwrap.dedent(""" # print("This is a stmt") # """) # produces '\nprint("This is a stmt")\n'. # # Stripping this down to 'print("This is a stmt")' doesn't change # what gets executed, but it makes __repr__'s nicer. stmt = textwrap.dedent(stmt) stmt = (stmt[1:] if stmt and stmt[0] == "\n" else stmt).rstrip() setup = textwrap.dedent(setup) setup = (setup[1:] if setup and setup[0] == "\n" else setup).rstrip() self._timer = self._timer_cls( stmt=stmt, setup=setup, timer=timer, globals=valgrind_timer_interface.CopyIfCallgrind.unwrap_all(self._globals), ) self._task_spec = common.TaskSpec( stmt=stmt, setup=setup, label=label, sub_label=sub_label, description=description, env=env, num_threads=num_threads, ) def timeit(self, number: int = 1000000) -> common.Measurement: """Mirrors the semantics of timeit.Timer.timeit(). Execute the main statement (`stmt`) `number` times. https://docs.python.org/3/library/timeit.html#timeit.Timer.timeit """ with common.set_torch_threads(self._task_spec.num_threads): # Warmup self._timer.timeit(number=max(int(number // 100), 1)) return common.Measurement( number_per_run=number, raw_times=[self._timer.timeit(number=number)], task_spec=self._task_spec ) def repeat(self, repeat: int = -1, number: int = -1) -> None: raise NotImplementedError("See `Timer.blocked_autorange.`") def autorange(self, callback: Optional[Callable[[int, float], NoReturn]] = None) -> None: raise NotImplementedError("See `Timer.blocked_autorange.`") def _threaded_measurement_loop( self, number: int, time_hook: Callable[[], float], stop_hook: Callable[[List[float]], bool], min_run_time: float, max_run_time: Optional[float] = None, callback: Optional[Callable[[int, float], NoReturn]] = None ) -> List[float]: total_time = 0.0 can_stop = False times: List[float] = [] with common.set_torch_threads(self._task_spec.num_threads): while (total_time < min_run_time) or (not can_stop): time_spent = time_hook() times.append(time_spent) total_time += time_spent if callback: callback(number, time_spent) can_stop = stop_hook(times) if max_run_time and total_time > max_run_time: break return times def _estimate_block_size(self, min_run_time: float) -> int: with common.set_torch_threads(self._task_spec.num_threads): # Estimate the block size needed for measurement to be negligible # compared to the inner loop. This also serves as a warmup. overhead = np.median([self._timer.timeit(0) for _ in range(5)]) number = 1 while True: time_taken = self._timer.timeit(number) relative_overhead = overhead / time_taken if relative_overhead <= 1e-4 and time_taken >= min_run_time / 1000: break if time_taken > min_run_time: break number *= 10 return number def adaptive_autorange( self, threshold: float = 0.1, *, min_run_time: float = 0.01, max_run_time: float = 10.0, callback: Optional[Callable[[int, float], NoReturn]] = None, ) -> common.Measurement: number = self._estimate_block_size(min_run_time=0.05) def time_hook() -> float: return self._timer.timeit(number) def stop_hook(times: List[float]) -> bool: if len(times) > 3: return common.Measurement( number_per_run=number, raw_times=times, task_spec=self._task_spec ).meets_confidence(threshold=threshold) return False times = self._threaded_measurement_loop( number, time_hook, stop_hook, min_run_time, max_run_time, callback=callback) return common.Measurement( number_per_run=number, raw_times=times, task_spec=self._task_spec ) def blocked_autorange( self, callback: Optional[Callable[[int, float], NoReturn]] = None, min_run_time: float = 0.2, ) -> common.Measurement: """Measure many replicates while keeping timer overhead to a minimum. At a high level, blocked_autorange executes the following pseudo-code:: `setup` total_time = 0 while total_time < min_run_time start = timer() for _ in range(block_size): `stmt` total_time += (timer() - start) Note the variable `block_size` in the inner loop. The choice of block size is important to measurement quality, and must balance two competing objectives: 1) A small block size results in more replicates and generally better statistics. 2) A large block size better amortizes the cost of `timer` invocation, and results in a less biased measurement. This is important because CUDA syncronization time is non-trivial (order single to low double digit microseconds) and would otherwise bias the measurement. blocked_autorange sets block_size by running a warmup period, increasing block size until timer overhead is less than 0.1% of the overall computation. This value is then used for the main measurement loop. Returns: A `Measurement` object that contains measured runtimes and repetition counts, and can be used to compute statistics. (mean, median, etc.) """ number = self._estimate_block_size(min_run_time) def time_hook() -> float: return self._timer.timeit(number) def stop_hook(times: List[float]) -> bool: return True times = self._threaded_measurement_loop( number, time_hook, stop_hook, min_run_time=min_run_time, callback=callback) return common.Measurement( number_per_run=number, raw_times=times, task_spec=self._task_spec ) def collect_callgrind( self, number: int = 100, collect_baseline: bool = True ) -> valgrind_timer_interface.CallgrindStats: """Collect instruction counts using Callgrind. Unlike wall times, instruction counts are deterministic (modulo non-determinism in the program itself and small amounts of jitter from the Python interpreter.) This makes them ideal for detailed performance analysis. This method runs `stmt` in a separate process so that Valgrind can instrument the program. Performance is severely degraded due to the instrumentation, howevever this is ameliorated by the fact that a small number of iterations is generally sufficient to obtain good measurements. In order to to use this method `valgrind`, `callgrind_control`, and `callgrind_annotate` must be installed. Because there is a process boundary between the caller (this process) and the `stmt` execution, `globals` cannot contain arbitrary in-memory data structures. (Unlike timing methods) Instead, globals are restricted to builtins, `nn.Modules`'s, and TorchScripted functions/modules to reduce the surprise factor from serialization and subsequent deserialization. The `GlobalsBridge` class provides more detail on this subject. Take particular care with nn.Modules: they rely on pickle and you may need to add an import to `setup` for them to transfer properly. By default, a profile for an empty statement will be collected and cached to indicate how many instructions are from the Python loop which drives `stmt`. Returns: A `CallgrindStats` object which provides instruction counts and some basic facilities for analyzing and manipulating results. """ if not isinstance(self._task_spec.stmt, str): raise ValueError("`collect_callgrind` currently only supports string `stmt`") # Check that the statement is valid. It doesn't guarantee success, but it's much # simpler and quicker to raise an exception for a faulty `stmt` or `setup` in # the parent process rather than the valgrind subprocess. self._timer.timeit(1) is_python = (self._language == Language.PYTHON) assert is_python or not self._globals return valgrind_timer_interface.wrapper_singleton().collect_callgrind( task_spec=self._task_spec, globals=self._globals, number=number, collect_baseline=collect_baseline and is_python, is_python=is_python)
[ [ [ 75, 79 ], [ 701, 705 ] ], [ [ 87, 93 ], [ 663, 669 ], [ 6580, 6586 ], [ 622, 628 ], [ 937, 943 ], [ 7642, 7648 ] ], [ [ 101, 109 ], [ 1562, 1570 ], [ 1611, 1619 ], [ 8488, 8496 ], [ 8599, 8607 ] ], [ [ 129, 132 ], [ 892, 895 ], [ 6763, 6766 ] ], [ [ 134, 142 ], [ 844, 852 ], [ 6698, 6706 ], [ 9971, 9979 ], [ 10181, 10189 ], [ 10221, 10229 ], [ 10354, 10362 ], [ 11949, 11957 ], [ 12851, 12859 ] ], [ [ 144, 148 ], [ 882, 886 ], [ 6753, 6757 ] ], [ [ 150, 154 ], [ 10404, 10408 ], [ 10231, 10235 ], [ 10482, 10486 ], [ 12193, 12197 ], [ 14603, 14607 ] ], [ [ 156, 164 ], [ 9994, 10002 ], [ 10377, 10385 ], [ 11972, 11980 ], [ 12874, 12882 ] ], [ [ 166, 174 ], [ 1663, 1671 ], [ 6744, 6752 ], [ 6792, 6800 ], [ 6833, 6841 ], [ 6876, 6884 ], [ 6911, 6919 ], [ 9962, 9970 ], [ 10303, 10311 ], [ 10345, 10353 ], [ 11940, 11948 ], [ 12842, 12850 ] ], [ [ 176, 180 ], [ 6561, 6565 ] ], [ [ 182, 187 ], [ 6981, 6986 ] ], [ [ 196, 207 ], [ 11293, 11295 ] ], [ [ 215, 220 ], [ 502, 507 ], [ 521, 526 ], [ 582, 587 ], [ 7486, 7491 ] ], [ [ 261, 267 ], [ 8937, 8943 ], [ 9213, 9219 ], [ 9452, 9458 ], [ 9614, 9620 ], [ 10512, 10518 ], [ 11065, 11071 ], [ 12000, 12006 ], [ 12650, 12656 ], [ 12937, 12943 ], [ 14827, 14833 ], [ 12269, 12275 ] ], [ [ 269, 276 ], [ 1816, 1823 ] ], [ [ 324, 334 ], [ 6566, 6576 ] ], [ [ 336, 352 ], [ 1672, 1688 ] ], [ [ 410, 453 ], [ 8833, 8857 ], [ 15075, 15099 ], [ 17374, 17398 ] ], [ [ 456, 463 ] ], [ [ 556, 561 ], [ 6720, 6725 ] ], [ [ 655, 660 ], [ 6720, 6725 ] ], [ [ 692, 700 ], [ 7004, 7012 ], [ 6987, 6995 ], [ 7333, 7341 ], [ 7532, 7540 ], [ 7521, 7529 ], [ 7575, 7583 ], [ 7818, 7826 ], [ 17296, 17304 ] ], [ [ 748, 756 ], [ 7725, 7733 ] ], [ [ 1979, 1984 ] ] ]
from Ranger.src.Range.Cut import Cut class Range(object): """ Class used to represent a range along some 1-D domain. The range is represented by 2 cutpoints can can be unbounded by specifying an aboveAll or belowAll Cut. """ def __init__(self, lowerCut, upperCut): """ Instantiates a Range Parameters ---------- lowerCut : Cut object Specifies the lower cut for the range upperCut : Cut object Specifies the upper cut for the range Raises ------ ValueError If bound(s) are not Cut objects or lower > upper """ if not all(map(lambda x: isinstance(x, Cut), (lowerCut,upperCut))): raise ValueError("Bounds must be Cut objects") elif lowerCut > upperCut: raise ValueError("Lower bound cannot be greater than upper bound") self.lowerCut = lowerCut self.upperCut = upperCut def __repr__(self): try: return_str = '[' if self.isLowerBoundClosed() else '(' except TypeError: return_str = '(' return_str += (str(self.lowerCut.point) if not self.lowerCut.belowAll \ else '') return_str += ' , ' return_str += (str(self.upperCut.point) if not self.upperCut.aboveAll \ else '') try: return_str += ']' if self.isUpperBoundClosed() else ')' except TypeError: return_str += ')' return return_str def __hash__(self): return (hash(self.lowerCut)*31 + hash(self.upperCut)) def __eq__(self, other): if not isinstance(other, Range): return False else: return ((self.lowerCut == other.lowerCut) and \ (self.upperCut == other.upperCut)) def __ne__(self, other): return not self.__eq__(other) def contains(self, val): """ Returns true if the range contains the value Parameters ---------- val : Comparable object of the appropriate type for the range Value to query whether in the range Raises ------ ValueError If the value type not compatible with cutpoint type Returns ------- True if the range contains the value """ return (self.lowerCut < val and \ self.upperCut > val) def containsAll(self, vals): """ Returns True if the range contains all values in some iterable Parameters ---------- vals : Iterable of comparable object of appropriate type for range Values to query against the range Raises ------ ValueError If there is a value type not compatible with the cutpoint type Returns ------- True if the range contains all values """ for val in vals: if not self.contains(val): return False return True def getDistanceFromPoint(self, val, distFunc = lambda x1, x2: abs(x1-x2)): """ Returns the minimum distance of a Range from a Point, returning 0 if there is an overlap. Note that both upper and lower bounds must be closed for this function to work Parameters ---------- val : comparable, compatible with cutpoint type The value of the point where the distance is desired distFunc : callable Function that calculates the distance between two points in the domain of the Range Raises ------ TypeError If the upper and/or lower bounds of this Range are not closed or if the distFunc not compatible with the type Returns ------- The minimum distance between the Range and the Point. Returns 0 if there is an overlap """ if not all((self.isLowerBoundClosed(), self.isUpperBoundClosed())): raise TypeError("Range is not closed") if self.contains(val): return 0. else: return min(distFunc(self.lowerCut.point, val), distFunc(self.upperCut.point, val)) def getDistanceFromRange(self, other, distFunc = lambda x1,x2: abs(x1-x2)): """ Returns the minimum distance of a Range from another Range, returning 0 if there is any overlap Note that both Ranges must be closed for this function to work Parameters ---------- other : Range, compatible with this Range's domain The Range to compare to distFunc : callable Function that calculates the distance between two points in the domain of the Range Raises ------ TypeError If the upper and/or lower bounds of this Range are not closed or if the distFunc not compatible with the type Returns ------- Minimum distance between the ranges """ if not isinstance(other, Range): raise TypeError("other is not a Range") if not all((self.isLowerBoundClosed(), self.isUpperBoundClosed(), other.isLowerBoundClosed(), other.isUpperBoundClosed())): raise TypeError("Not all Ranges closed") if self.isConnected(other): return 0. else: return min(distFunc(self.lowerCut.point, other.upperCut.point), distFunc(other.lowerCut.point, self.upperCut.point)) def hasLowerBound(self): """ Returns True if the range has a lower endpoint (not unbounded at the lower end) Returns ------- True if the range has a lower endpoint """ return (not self.lowerCut.belowAll) def hasUpperBound(self): """ Returns True if the range has an upper endpoint (not unbounded at the upper end) Returns ------- True if the range has an upper endpoint """ return (not self.upperCut.aboveAll) def lowerEndpoint(self): """ Returns the lower endpoint of the range if it exists. Otherwise raises a TypeError Raises ------ TypeError If the range is unbounded below Returns ------- The lower endpoint of the range """ if self.lowerCut.point is None: raise TypeError("Range unbounded below") else: return self.lowerCut.point def upperEndpoint(self): """ Returns the upper endpoint of the range if it exists. Otherwise raises a TypeError Raises ------ TypeError If the range is unbounded above Returns ------- The upper endpoint of the range """ if self.upperCut.point is None: raise TypeError("Range unbounded above") else: return self.upperCut.point def isLowerBoundClosed(self): """ Returns whether the lower bound is closed (if there is a lower bound) Raises ------ TypeError If the range is unbounded below Returns ------- True if the lower bound is closed """ if self.lowerCut.point is None: raise TypeError("Range unbounded below") else: return self.lowerCut.below def isUpperBoundClosed(self): """ Returns whether the upper bound is closed (if there is an upper bound) Raises ------ TypeError If the range is unbounded above Returns ------- True if the upper bound is closed """ if self.upperCut.point is None: raise TypeError("Range unbounded above") else: return (not self.upperCut.below) def isEmpty(self): """ Returns True if the range is of form [v, v) or (v, v] Returns ------- True if the range is of the form [v,v) or (v,v] """ return self.lowerCut == self.upperCut def encloses(self, other): """ Returns True if the bounds of the other range do not extend outside the bounds of this range Examples: [3,6] encloses [4,5] (3,6) encloses (3,6) [3,6] encloses [4,4] (3,6] does not enclose [3,6] [4,5] does not enclose (3,6) Parameters ---------- other : A Range The range to compare to Raises ------ ValueError If object passed in is not a Range Returns ------- True if the bounds of the other range do not extend outside the bounds of this range """ if not isinstance(other, Range): raise ValueError("Range required") return ((self.lowerCut <= other.lowerCut) and \ (self.upperCut >= other.upperCut)) def isConnected(self, other): """ Returns True if there is a (possibly empty) range that is enclosed by both this range and other Examples: [2,4] and [5,7] are not connected [2,4] and [3,5] are connected [2,4] and [4,6] are connected [3,5] and (5,10) are connected Parameters ---------- other : A range The range to compare to Raises ------ ValueError If object passed in is not a Range Returns ------- True if there is a (possibly empty) range that is enclosed by both this range and other """ if not isinstance(other, Range): raise ValueError("Range required") return ((self.lowerCut <= other.upperCut) and \ (other.lowerCut <= self.upperCut)) def intersection(self, other): """ Returns the maximal range enclosed by both this range and the other range, if such a range exists Examples: Intersection of [1,5] and [3,7] is [3,5] Intersection of [1,5] and [5,7] is [5,5] Parameters ---------- other : A range The range to compare to Raises ------ ValueError If object passed in is not a Range or if there is no intersection Returns ------- The intersection range """ if not isinstance(other, Range): raise ValueError("Range required") if ((self.lowerCut >= other.lowerCut) and \ (self.upperCut <= other.upperCut)): return Range(self.lowerCut, self.upperCut) elif ((self.lowerCut <= other.lowerCut) and \ (self.upperCut >= other.upperCut)): return Range(other.lowerCut, other.upperCut) else: newLower = self.lowerCut if (self.lowerCut >= other.lowerCut) else \ other.lowerCut newUpper = self.upperCut if (self.upperCut <= other.upperCut) else \ other.upperCut return Range(newLower, newUpper) def span(self, other): """ Returns the minimal range that encloses both this range and the other. Note that if the input ranges are not connected, the span can contain values that are not contained within either input range Examples: Span of [1,3] and [5,7] is [1,7] Parameters ---------- other : A range A range to span with Raises ------ ValueError If object passed in is not a Range or if there is no intersection Returns ------- The minimal range enclosing both with and the other range """ if ((self.lowerCut <= other.lowerCut) and \ (self.upperCut >= other.upperCut)): return Range(self.lowerCut, self.upperCut) elif ((self.lowerCut >= other.lowerCut) and \ (self.upperCut <= other.upperCut)): return Range(other.lowerCut, other.upperCut) else: newLower = self.lowerCut if (self.lowerCut <= other.lowerCut) else \ other.lowerCut newUpper = self.upperCut if (self.upperCut >= other.upperCut) else \ other.upperCut return Range(newLower, newUpper) ################## # Static methods # ################## @staticmethod def _validate_cutpoints(*pts): if not all(map(lambda x: (hasattr(x, "__lt__") and \ hasattr(x, "__gt__")) or hasattr(x,'__cmp__'), pts)): raise ValueError("Cutpoint type(s) not comparable") if len(pts) == 2: if not (issubclass(type(pts[0]),type(pts[1])) or \ issubclass(type(pts[1]),type(pts[0]))): raise ValueError("Cutpoints are not compatible") return True @staticmethod def _get_type(*pts): if len(pts) == 1: return type(pts[0]) elif len(pts) == 2: if issubclass(type(pts[0]),type(pts[1])): return type(pts[1]) elif issubclass(type(pts[1]),type(pts[0])): return type(pts[0]) else: raise ValueError("Cutpoints are not compatible") @staticmethod def closed(lower, upper): """ Creates a range including the endpoints (i.e. [lower, upper]) Parameters ---------- lower : comparable, of same type as or subclass of upper type The lower bound upper : comparable, of same type as or subclass of lower type The upper bound Raises ------ ValueError If type(s) are not comparable or compatible Returns ------- A Range object [lower, upper] """ # Ensure cutpoints are of compatible, appropriate types Range._validate_cutpoints(lower, upper) theType = Range._get_type(lower,upper) return Range(Cut.belowValue(lower, theType=theType), Cut.aboveValue(upper, theType=theType)) @staticmethod def closedOpen(lower, upper): """ Creates a range including the lower endpoint (i.e. [lower, upper)) Parameters ---------- lower : comparable, of same type as or subclass of upper type The lower bound upper : comparable, of same type as or subclass of lower type The upper bound Raises ------ ValueError If type(s) are not comparable or compatible Returns ------- A Range object [lower, upper) """ # Ensure cutpoints are of compatible, appropriate types Range._validate_cutpoints(lower, upper) theType = Range._get_type(lower,upper) return Range(Cut.belowValue(lower, theType=theType), Cut.belowValue(upper, theType=theType)) @staticmethod def openClosed(lower, upper): """ Creates a range including the upper (i.e. (lower, upper]) Parameters ---------- lower : comparable, of same type as or subclass of upper type The lower bound upper : comparable, of same type as or subclass of lower type The upper bound Raises ------ ValueError If type(s) are not comparable or compatible Returns ------- A Range object (lower, upper] """ # Ensure cutpoints are of compatible, appropriate types Range._validate_cutpoints(lower, upper) theType = Range._get_type(lower,upper) return Range(Cut.aboveValue(lower, theType=theType), Cut.aboveValue(upper, theType=theType)) @staticmethod def open(lower, upper): """ Creates a range excluding the endpoints (i.e. (lower, upper)) Parameters ---------- lower : comparable, of same type as or subclass of upper type The lower bound upper : comparable, of same type as or subclass of lower type The upper bound Raises ------ ValueError If type(s) are not comparable or compatible or if constructing a range of type (v,v), which is invalid Returns ------- A Range object (lower, upper) """ # Ensure cutpoints are of compatible, appropriate types Range._validate_cutpoints(lower, upper) theType = Range._get_type(lower,upper) if lower == upper: raise TypeError("Range of type (v,v) is not valid") return Range(Cut.aboveValue(lower, theType=theType), Cut.belowValue(upper, theType=theType)) @staticmethod def lessThan(val): """ Makes range including all values less than some value (i.e. (-inf, val)) Parameters ---------- val : comparable The upper bound Raises ------ ValueError If type not comparable Returns ------- A Range object (-inf, val) """ Range._validate_cutpoints(val) theType = Range._get_type(val) return Range(Cut.belowAll(theType=theType), Cut.belowValue(val, theType=theType)) @staticmethod def atMost(val): """ Makes range including all values less than or equal to some value (i.e. (-inf, val]) Parameters ---------- val : comparable The upper bound Raises ------ ValueError If type not comparable Returns ------- A Range object (-inf, val] """ Range._validate_cutpoints(val) theType = Range._get_type(val) return Range(Cut.belowAll(theType=theType), Cut.aboveValue(val, theType=theType)) @staticmethod def greaterThan(val): """ Makes range including all values greater than some value (i.e. (val, inf]) Parameters ---------- val : comparable The lower bound Raises ------ ValueError If type not comparable Returns ------- A Range object (val, inf) """ Range._validate_cutpoints(val) theType = Range._get_type(val) return Range(Cut.aboveValue(val,theType=theType), Cut.aboveAll(theType=theType)) @staticmethod def atLeast(val): """ Makes range including all values greater than or equal to some value (i.e. [val, inf)) Parameters ---------- val : comparable The lower bound Raises ------ ValueError If type not comparable Returns ------- A Range object [val, inf) """ Range._validate_cutpoints(val) theType = Range._get_type(val) return Range(Cut.belowValue(val, theType=theType), Cut.aboveAll(theType=theType))
[ [ [ 33, 36 ], [ 14187, 14190 ], [ 14248, 14251 ], [ 15031, 15034 ], [ 15092, 15095 ], [ 15866, 15869 ], [ 15927, 15930 ], [ 16869, 16872 ], [ 16930, 16933 ], [ 17460, 17463 ], [ 17512, 17515 ], [ 18050, 18053 ], [ 18102, 18105 ], [ 18634, 18637 ], [ 18692, 18695 ], [ 19225, 19228 ], [ 19284, 19287 ], [ 692, 695 ] ], [ [ 44, 49 ], [ 1655, 1660 ], [ 5098, 5103 ], [ 8894, 8899 ], [ 9788, 9793 ], [ 10564, 10569 ], [ 10738, 10743 ], [ 10897, 10902 ], [ 11242, 11247 ], [ 12046, 12051 ], [ 12205, 12210 ], [ 12500, 12505 ], [ 14079, 14084 ], [ 14137, 14142 ], [ 14181, 14186 ], [ 14923, 14928 ], [ 14981, 14986 ], [ 15025, 15030 ], [ 15758, 15763 ], [ 15816, 15821 ], [ 15860, 15865 ], [ 16662, 16667 ], [ 16720, 16725 ], [ 16863, 16868 ], [ 17369, 17374 ], [ 17418, 17423 ], [ 17454, 17459 ], [ 17959, 17964 ], [ 18008, 18013 ], [ 18044, 18049 ], [ 18543, 18548 ], [ 18592, 18597 ], [ 18628, 18633 ], [ 19134, 19139 ], [ 19183, 19188 ], [ 19219, 19224 ] ] ]
from pyrelational.data.data_manager import GenericDataManager
[ [ [ 43, 61 ] ] ]
from selfdrive.car import dbc_dict from cereal import car Ecu = car.CarParams.Ecu class CarControllerParams: ANGLE_DELTA_BP = [0., 5., 15.] ANGLE_DELTA_V = [5., .8, .15] # windup limit ANGLE_DELTA_VU = [5., 3.5, 0.4] # unwind limit LKAS_MAX_TORQUE = 1 # A value of 1 is easy to overpower STEER_THRESHOLD = 1.0 class CAR: XTRAIL = "NISSAN X-TRAIL 2017" LEAF = "NISSAN LEAF 2018" # Leaf with ADAS ECU found behind instrument cluster instead of glovebox # Currently the only known difference between them is the inverted seatbelt signal. LEAF_IC = "NISSAN LEAF 2018 Instrument Cluster" ROGUE = "NISSAN ROGUE 2019" ALTIMA = "NISSAN ALTIMA 2020" FINGERPRINTS = { CAR.XTRAIL: [ { 2: 5, 42: 6, 346: 6, 347: 5, 348: 8, 349: 7, 361: 8, 386: 8, 389: 8, 397: 8, 398: 8, 403: 8, 520: 2, 523: 6, 548: 8, 645: 8, 658: 8, 665: 8, 666: 8, 674: 2, 682: 8, 683: 8, 689: 8, 723: 8, 758: 3, 768: 2, 783: 3, 851: 8, 855: 8, 1041: 8, 1055: 2, 1104: 4, 1105: 6, 1107: 4, 1108: 8, 1111: 4, 1227: 8, 1228: 8, 1247: 4, 1266: 8, 1273: 7, 1342: 1, 1376: 6, 1401: 8, 1474: 2, 1497: 3, 1821: 8, 1823: 8, 1837: 8, 2015: 8, 2016: 8, 2024: 8 }, { 2: 5, 42: 6, 346: 6, 347: 5, 348: 8, 349: 7, 361: 8, 386: 8, 389: 8, 397: 8, 398: 8, 403: 8, 520: 2, 523: 6, 527: 1, 548: 8, 637: 4, 645: 8, 658: 8, 665: 8, 666: 8, 674: 2, 682: 8, 683: 8, 689: 8, 723: 8, 758: 3, 768: 6, 783: 3, 851: 8, 855: 8, 1041: 8, 1055: 2, 1104: 4, 1105: 6, 1107: 4, 1108: 8, 1111: 4, 1227: 8, 1228: 8, 1247: 4, 1266: 8, 1273: 7, 1342: 1, 1376: 6, 1401: 8, 1474: 8, 1497: 3, 1534: 6, 1792: 8, 1821: 8, 1823: 8, 1837: 8, 1872: 8, 1937: 8, 1953: 8, 1968: 8, 2015: 8, 2016: 8, 2024: 8 }, ], CAR.LEAF: [ { 2: 5, 42: 6, 264: 3, 361: 8, 372: 8, 384: 8, 389: 8, 403: 8, 459: 7, 460: 4, 470: 8, 520: 1, 569: 8, 581: 8, 634: 7, 640: 8, 644: 8, 645: 8, 646: 5, 658: 8, 682: 8, 683: 8, 689: 8, 724: 6, 758: 3, 761: 2, 783: 3, 852: 8, 853: 8, 856: 8, 861: 8, 944: 1, 976: 6, 1008: 7, 1011: 7, 1057: 3, 1227: 8, 1228: 8, 1261: 5, 1342: 1, 1354: 8, 1361: 8, 1459: 8, 1477: 8, 1497: 3, 1549: 8, 1573: 6, 1821: 8, 1837: 8, 1856: 8, 1859: 8, 1861: 8, 1864: 8, 1874: 8, 1888: 8, 1891: 8, 1893: 8, 1906: 8, 1947: 8, 1949: 8, 1979: 8, 1981: 8, 2016: 8, 2017: 8, 2021: 8, 643: 5, 1792: 8, 1872: 8, 1937: 8, 1953: 8, 1968: 8, 1988: 8, 2000: 8, 2001: 8, 2004: 8, 2005: 8, 2015: 8 }, # 2020 Leaf SV Plus { 2: 5, 42: 8, 264: 3, 361: 8, 372: 8, 384: 8, 389: 8, 403: 8, 459: 7, 460: 4, 470: 8, 520: 1, 569: 8, 581: 8, 634: 7, 640: 8, 643: 5, 644: 8, 645: 8, 646: 5, 658: 8, 682: 8, 683: 8, 689: 8, 724: 6, 758: 3, 761: 2, 772: 8, 773: 6, 774: 7, 775: 8, 776: 6, 777: 7, 778: 6, 783: 3, 852: 8, 853: 8, 856: 8, 861: 8, 943: 8, 944: 1, 976: 6, 1008: 7, 1009: 8, 1010: 8, 1011: 7, 1012: 8, 1013: 8, 1019: 8, 1020: 8, 1021: 8, 1022: 8, 1057: 3, 1227: 8, 1228: 8, 1261: 5, 1342: 1, 1354: 8, 1361: 8, 1402: 8, 1459: 8, 1477: 8, 1497: 3, 1549: 8, 1573: 6, 1821: 8, 1837: 8 }, ], CAR.LEAF_IC: [ { 2: 5, 42: 6, 264: 3, 282: 8, 361: 8, 372: 8, 384: 8, 389: 8, 403: 8, 459: 7, 460: 4, 470: 8, 520: 1, 569: 8, 581: 8, 634: 7, 640: 8, 643: 5, 644: 8, 645: 8, 646: 5, 658: 8, 682: 8, 683: 8, 689: 8, 756: 5, 758: 3, 761: 2, 783: 3, 830: 2, 852: 8, 853: 8, 856: 8, 861: 8, 943: 8, 944: 1, 1001: 6, 1057: 3, 1227: 8, 1228: 8, 1229: 8, 1342: 1, 1354: 8, 1361: 8, 1459: 8, 1477: 8, 1497: 3, 1514: 6, 1549: 8, 1573: 6, 1792: 8, 1821: 8, 1822: 8, 1837: 8, 1838: 8, 1872: 8, 1937: 8, 1953: 8, 1968: 8, 1988: 8, 2000: 8, 2001: 8, 2004: 8, 2005: 8, 2015: 8, 2016: 8, 2017: 8 }, ], CAR.ROGUE: [ { 2: 5, 42: 6, 346: 6, 347: 5, 348: 8, 349: 7, 361: 8, 386: 8, 389: 8, 397: 8, 398: 8, 403: 8, 520: 2, 523: 6, 548: 8, 634: 7, 643: 5, 645: 8, 658: 8, 665: 8, 666: 8, 674: 2, 682: 8, 683: 8, 689: 8, 723: 8, 758: 3, 772: 8, 773: 6, 774: 7, 775: 8, 776: 6, 777: 7, 778: 6, 783: 3, 851: 8, 855: 8, 1041: 8, 1042: 8, 1055: 2, 1104: 4, 1105: 6, 1107: 4, 1108: 8, 1110: 7, 1111: 7, 1227: 8, 1228: 8, 1247: 4, 1266: 8, 1273: 7, 1342: 1, 1376: 6, 1401: 8, 1474: 2, 1497: 3, 1534: 7, 1792: 8, 1821: 8, 1823: 8, 1837: 8, 1839: 8, 1872: 8, 1937: 8, 1953: 8, 1968: 8, 1988: 8, 2000: 8, 2001: 8, 2004: 8, 2005: 8, 2015: 8, 2016: 8, 2017: 8, 2024: 8, 2025: 8 }, ], CAR.ALTIMA: [ { 2: 5, 42: 6, 346: 6, 347: 5, 348: 8, 349: 7, 361: 8, 386: 8, 389: 8, 397: 8, 398: 8, 403: 8, 438: 8, 451: 8, 517: 8, 520: 2, 522: 8, 523: 6, 539: 8, 541: 7, 542: 8, 543: 8, 544: 8, 545: 8, 546: 8, 547: 8, 548: 8, 570: 8, 576: 8, 577: 8, 582: 8, 583: 8, 584: 8, 586: 8, 587: 8, 588: 8, 589: 8, 590: 8, 591: 8, 592: 8, 600: 8, 601: 8, 610: 8, 611: 8, 612: 8, 614: 8, 615: 8, 616: 8, 617: 8, 622: 8, 623: 8, 634: 7, 638: 8, 645: 8, 648: 5, 654: 6, 658: 8, 659: 8, 660: 8, 661: 8, 665: 8, 666: 8, 674: 2, 675: 8, 676: 8, 682: 8, 683: 8, 684: 8, 685: 8, 686: 8, 687: 8, 689: 8, 690: 8, 703: 8, 708: 7, 709: 7, 711: 7, 712: 7, 713: 7, 714: 8, 715: 8, 716: 8, 717: 7, 718: 7, 719: 7, 720: 7, 723: 8, 726: 7, 727: 7, 728: 7, 735: 8, 746: 8, 748: 6, 749: 6, 750: 8, 758: 3, 772: 8, 773: 6, 774: 7, 775: 8, 776: 6, 777: 7, 778: 6, 779: 7, 781: 7, 782: 7, 783: 3, 851: 8, 855: 5, 1001: 6, 1041: 8, 1042: 8, 1055: 3, 1100: 7, 1104: 4, 1105: 6, 1107: 4, 1108: 8, 1110: 7, 1111: 7, 1144: 7, 1145: 7, 1227: 8, 1228: 8, 1229: 8, 1232: 8, 1247: 4, 1258: 8, 1259: 8, 1266: 8, 1273: 7, 1306: 1, 1314: 8, 1323: 8, 1324: 8, 1342: 1, 1376: 8, 1401: 8, 1454: 8, 1497: 3, 1514: 6, 1526: 8, 1527: 5, 1792: 8, 1821: 8, 1823: 8, 1837: 8, 1872: 8, 1937: 8, 1953: 8, 1968: 8, 1988: 8, 2000: 8, 2001: 8, 2004: 8, 2005: 8, 2015: 8, 2016: 8, 2017: 8, 2024: 8, 2025: 8 }, ] } FW_VERSIONS = { CAR.ALTIMA: { (Ecu.fwdCamera, 0x707, None): [ b'284N86CA1D', ], (Ecu.eps, 0x742, None): [ b'6CA2B\xa9A\x02\x02G8A89P90D6A\x00\x00\x01\x80', ], (Ecu.engine, 0x7e0, None): [ b'237109HE2B', ], (Ecu.gateway, 0x18dad0f1, None): [ b'284U29HE0A', ], }, CAR.LEAF_IC: { (Ecu.fwdCamera, 0x707, None): [ b'5SH1BDB\x04\x18\x00\x00\x00\x00\x00_-?\x04\x91\xf2\x00\x00\x00\x80', b'5SK0ADB\x04\x18\x00\x00\x00\x00\x00_(5\x07\x9aQ\x00\x00\x00\x80', ], (Ecu.esp, 0x740, None): [ b'476605SH1D', b'476605SK2A', ], (Ecu.eps, 0x742, None): [ b'5SH2A\x99A\x05\x02N123F\x15\x81\x00\x00\x00\x00\x00\x00\x00\x80', b'5SK3A\x99A\x05\x02N123F\x15u\x00\x00\x00\x00\x00\x00\x00\x80', ], (Ecu.gateway, 0x18dad0f1, None): [ b'284U25SH3A', b'284U25SK2D', ], }, CAR.XTRAIL: { (Ecu.fwdCamera, 0x707, None): [ b'284N86FR2A', ], (Ecu.esp, 0x740, None): [ b'6FU1BD\x11\x02\x00\x02e\x95e\x80iX#\x01\x00\x00\x00\x00\x00\x80', b'6FU0AD\x11\x02\x00\x02e\x95e\x80iQ#\x01\x00\x00\x00\x00\x00\x80', ], (Ecu.eps, 0x742, None): [ b'6FP2A\x99A\x05\x02N123F\x18\x02\x00\x00\x00\x00\x00\x00\x00\x80', ], (Ecu.combinationMeter, 0x743, None): [ b'6FR2A\x18B\x05\x17\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x80', ], (Ecu.engine, 0x7e0, None): [ b'6FU9B\xa0A\x06\x04\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x80', b'6FR9A\xa0A\x06\x04\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x80', ], (Ecu.gateway, 0x18dad0f1, None): [ b'284U26FR0E', ], }, } DBC = { CAR.XTRAIL: dbc_dict('nissan_x_trail_2017', None), CAR.LEAF: dbc_dict('nissan_leaf_2018', None), CAR.LEAF_IC: dbc_dict('nissan_leaf_2018', None), CAR.ROGUE: dbc_dict('nissan_x_trail_2017', None), CAR.ALTIMA: dbc_dict('nissan_x_trail_2017', None), }
[ [ [ 26, 34 ], [ 7380, 7388 ], [ 7431, 7439 ], [ 7482, 7490 ], [ 7531, 7539 ], [ 7584, 7592 ] ], [ [ 54, 57 ], [ 64, 67 ] ], [ [ 58, 61 ], [ 5700, 5703 ], [ 5764, 5767 ], [ 5857, 5860 ], [ 5918, 5921 ], [ 6007, 6010 ], [ 6201, 6204 ], [ 6280, 6283 ], [ 6462, 6465 ], [ 6571, 6574 ], [ 6635, 6638 ], [ 6820, 6823 ], [ 6931, 6934 ], [ 7070, 7073 ], [ 7288, 7291 ] ], [ [ 90, 109 ] ], [ [ 349, 352 ], [ 709, 712 ], [ 1708, 1711 ], [ 3001, 3004 ], [ 3605, 3608 ], [ 4287, 4290 ], [ 5681, 5684 ], [ 5987, 5990 ], [ 6552, 6555 ], [ 7368, 7371 ], [ 7421, 7424 ], [ 7469, 7472 ], [ 7520, 7523 ], [ 7572, 7575 ] ], [ [ 690, 702 ] ], [ [ 5663, 5674 ] ], [ [ 7358, 7361 ] ] ]
from .core import Core, Settings class Download(Core): host = 'https://artifacts.elastic.co/downloads/beats/elastic-agent/{endpoint}' endpoint = Settings.download_endpoint kwargs = { 'stream': True } def parse_response(self, response): self.__logger.debug('Saving file to download path: {}'.format(Settings.download_path)) with open(Settings.download_path, 'wb+') as f: for chunk in response.raw.stream(1024, decode_content=False): if chunk: f.write(chunk) self.__logger.debug('File saved successfully')
[ [ [ 18, 22 ], [ 50, 54 ] ], [ [ 24, 32 ], [ 156, 164 ], [ 338, 346 ], [ 381, 389 ] ], [ [ 41, 49 ] ] ]
from .single_stage import SingleStageDetector from ..registry import DETECTORS from mmdet.core import bbox2result import torch.nn as nn import torch from .. import builder import numpy as np import cv2 from mmdet.core import bbox2roi, bbox2result, build_assigner, build_sampler @DETECTORS.register_module class CSP(SingleStageDetector): def __init__(self, backbone, neck, bbox_head, refine_roi_extractor=None, refine_head=None, train_cfg=None, test_cfg=None, pretrained=None, detached=True, return_feature_maps=False): super(CSP, self).__init__(backbone, neck, bbox_head, train_cfg, test_cfg, pretrained) if refine_head is not None: self.refine_roi_extractor = builder.build_roi_extractor( refine_roi_extractor) self.refine_head = builder.build_head(refine_head) self.return_feature_maps = return_feature_maps self.train_cfg = train_cfg self.test_cfg = test_cfg self.detached = detached def show_input_debug(self, img, classification_maps, scale_maps, offset_maps): img_numpy = img.cpu().numpy().copy()[0] # img_numpy = np.transpose(img_numpy, [1, 2, 0]) * [58.395, 57.12, 57.375] + [123.675, 116.28, 103.53] img_numpy = np.transpose(img_numpy, [1, 2, 0]) + [102.9801, 115.9465, 122.7717] img_numpy = img_numpy[:, :, ::-1] img_numpy = img_numpy.astype(np.uint8) strides = [8, 16, 32, 64, 128] img_nows = [] for i, stride in enumerate(strides): img_now = img_numpy.copy() # cls_numpy = classification_maps[0][i].cpu().numpy().copy()[0][2] cls_numpy = classification_maps[0][i].cpu().numpy().copy()[0][:80] scale_numpy = scale_maps[0][i].cpu().numpy().copy()[0][0] * stride offset_numpy = offset_maps[0][i].cpu().numpy().copy()[0][:2] cs, ys, xs = cls_numpy.nonzero() print(len(ys)) for c, x, y in zip(cs, xs, ys): cv2.imshow(str(c), classification_maps[0][i].cpu().numpy().copy()[0][80+c]) realx = x realy = y height = scale_numpy[y, x] realy = realy + 0.5 + offset_numpy[0][y, x] realx = realx + 0.5 + offset_numpy[1][y, x] realy = realy * stride realx = realx * stride top_y = int(realy - height/2) top_x = int(realx) down_y = int(realy + height/2) down_x = int(realx) top_left = (int(top_x - height * 0.1), int(top_y)) down_right = (int(down_x + height * 0.1), down_y) cv2.rectangle(img_now, top_left, down_right, (255, 255, 5*int(c)), 2) img_nows.append(img_now) cv2.imshow(str(i) +'img', img_now) cv2.waitKey(0) def show_input_debug_caltech(self, img, classification_maps, scale_maps, offset_maps): for j in range(img.shape[0]): img_numpy = img.cpu().numpy().copy()[j] img_numpy = np.transpose(img_numpy, [1, 2, 0]) * [58.395, 57.12, 57.375] + [123.675, 116.28, 103.53] img_numpy = img_numpy[:, :, ::-1] img_numpy = img_numpy.astype(np.uint8) strides = [4] img_nows = [] for i, stride in enumerate(strides): img_now = img_numpy.copy() cls_numpy = classification_maps[j][i].cpu().numpy().copy()[0][2] ignore_numpy = classification_maps[j][i].cpu().numpy().copy()[0][1] cv2.imshow('ignore', ignore_numpy) scale_numpy = scale_maps[j][i].cpu().numpy().copy()[0][0] * stride offset_numpy = offset_maps[j][i].cpu().numpy().copy()[0][:2] ys, xs = cls_numpy.nonzero() print(len(ys)) for x, y in zip(xs, ys): # cv2.imshow(str(c), classification_maps[j][i].cpu().numpy().copy()[0][c]) realx = x realy = y height = scale_numpy[y, x] realy = realy + 0.5 + offset_numpy[0][y, x] realx = realx + 0.5 + offset_numpy[1][y, x] realy = realy * stride realx = realx * stride top_y = int(realy - height/2) top_x = int(realx) down_y = int(realy + height/2) down_x = int(realx) top_left = (int(top_x - height * 0.1), int(top_y)) down_right = (int(down_x + height * 0.1), down_y) cv2.rectangle(img_now, top_left, down_right, (255, 255, 125), 2) img_nows.append(img_now) cv2.imshow(str(i) +'img', img_now) cv2.waitKey(0) def show_input_debug_head(self, img, classification_maps, scale_maps, offset_maps): for j in range(img.shape[0]): img_numpy = img.cpu().numpy().copy()[j] img_numpy = np.transpose(img_numpy, [1, 2, 0]) * [58.395, 57.12, 57.375] + [123.675, 116.28, 103.53] img_numpy = img_numpy[:, :, ::-1] img_numpy = img_numpy.astype(np.uint8) strides = [4] img_nows = [] for i, stride in enumerate(strides): img_now = img_numpy.copy() cls_numpy = classification_maps[j][i].cpu().numpy().copy()[0][2] ignore_numpy = classification_maps[j][i].cpu().numpy().copy()[0][1] cv2.imshow('ignore', ignore_numpy) scale_numpy = scale_maps[j][i].exp().cpu().numpy().copy()[0][0] * stride offset_numpy = offset_maps[j][i].cpu().numpy().copy()[0][:2] ys, xs = cls_numpy.nonzero() for x, y in zip(xs, ys): # cv2.imshow(str(c), classification_maps[j][i].cpu().numpy().copy()[0][c]) realx = x realy = y height = scale_numpy[y, x] realy = realy + 0.5 + offset_numpy[0][y, x] realx = realx + 0.5 + offset_numpy[1][y, x] realy = realy * stride realx = realx * stride top_y = int(realy) top_x = int(realx) down_y = int(realy + height) down_x = int(realx) top_left = (int(top_x - height * 0.41/2), int(top_y)) down_right = (int(down_x + height * 0.41/2), down_y) cv2.rectangle(img_now, top_left, down_right, (255, 255, 125), 2) img_nows.append(img_now) cv2.imshow(str(i) +'img', img_now) cv2.waitKey(0) def show_mot_input_debug(self, img, classification_maps, scale_maps, offset_maps): for j in range(img.shape[0]): img_numpy = img.cpu().numpy().copy()[j] img_numpy = np.transpose(img_numpy, [1, 2, 0]) * [58.395, 57.12, 57.375] + [123.675, 116.28, 103.53] # img_numpy = np.transpose(img_numpy, [1, 2, 0]) + [102.9801, 115.9465, 122.7717] img_numpy = img_numpy[:, :, ::-1] img_numpy = img_numpy.astype(np.uint8) strides = [4] img_nows = [] for i, stride in enumerate(strides): img_now = img_numpy.copy() # cls_numpy = classification_maps[0][i].cpu().numpy().copy()[0][2] cls_numpy = classification_maps[j][i].cpu().numpy().copy()[0][2] instance_numpy = classification_maps[j][i].cpu().numpy().copy()[0][3] scale_numpy = scale_maps[j][i].cpu().numpy().copy()[0][0] * stride offset_numpy = offset_maps[j][i].cpu().numpy().copy()[0][:2] ys, xs = cls_numpy.nonzero() for x, y in zip(xs, ys): c=0 cv2.imshow(str(c), classification_maps[j][i].cpu().numpy().copy()[0][2]) realx = x realy = y height = scale_numpy[y, x] realy = realy + 0.5 + offset_numpy[0][y, x] realx = realx + 0.5 + offset_numpy[1][y, x] realy = realy * stride realx = realx * stride top_y = int(realy - height/2) top_x = int(realx) down_y = int(realy + height/2) down_x = int(realx) top_left = (int(top_x - height * 0.1), int(top_y)) down_right = (int(down_x + height * 0.1), down_y) cv2.rectangle(img_now, top_left, down_right, (255, 255, 5*int(c)), 2) instance = instance_numpy[y, x] cv2.putText(img_now, str(instance), top_left, cv2.FONT_HERSHEY_COMPLEX, 1, 255) img_nows.append(img_now) cv2.imshow(str(i) +'img', img_now) cv2.waitKey(0) @property def refine(self): return hasattr(self, 'refine_head') and self.refine_head is not None def forward_train(self, img, img_metas, gt_bboxes, gt_labels, gt_bboxes_ignore=None, classification_maps=None, scale_maps=None, offset_maps=None): # for tracking data which batch is produced by dataset instead of data loader if type(img) == list: img=img[0] img_metas=img_metas[0] gt_bboxes=gt_bboxes[0] gt_labels=gt_labels[0] gt_bboxes_ignore = gt_bboxes_ignore[0] classification_maps = classification_maps[0] scale_maps = scale_maps[0] offset_maps = offset_maps[0] losses = dict() x = self.extract_feat(img) # self.show_input_debug(img, classification_maps, scale_maps, offset_maps) # self.show_input_debug_caltech(img, classification_maps, scale_maps, offset_maps) # self.show_mot_input_debug(img, classification_maps, scale_maps, offset_maps) # self.show_input_debug_head(img, classification_maps, scale_maps, offset_maps) outs = self.bbox_head(x) loss_inputs = outs + (gt_bboxes, gt_labels, classification_maps, scale_maps, offset_maps, img_metas, self.train_cfg.csp_head if self.refine else self.train_cfg) losses_bbox = self.bbox_head.loss( *loss_inputs, gt_bboxes_ignore=gt_bboxes_ignore) losses.update(losses_bbox) if self.refine: if self.detached: x = tuple([i.detach() for i in x]) bbox_inputs = outs + (img_metas, self.train_cfg.csp_head, False) bbox_list = self.bbox_head.get_bboxes(*bbox_inputs, no_strides=False) # no_strides to not upscale yet bbox_list = [ bbox2result(det_bboxes, det_labels, self.bbox_head.num_classes)[0] for det_bboxes, det_labels in bbox_list ] bbox_assigner = build_assigner(self.train_cfg.rcnn.assigner) bbox_sampler = build_sampler( self.train_cfg.rcnn.sampler, context=self) num_imgs = img.size(0) if gt_bboxes_ignore is None: gt_bboxes_ignore = [None for _ in range(num_imgs)] sampling_results = [] for i in range(num_imgs): if bbox_list[i].shape[0] == 0 or gt_bboxes[i].shape[0] == 0: continue bbox = torch.tensor(bbox_list[i]).float().cuda() assign_result = bbox_assigner.assign( bbox, gt_bboxes[i], gt_bboxes_ignore[i], gt_labels[i]) sampling_result = bbox_sampler.sample( assign_result, bbox, gt_bboxes[i], gt_labels[i]) sampling_results.append(sampling_result) samp_list = [res.bboxes for res in sampling_results] if len(samp_list) == 0: losses.update(dict(loss_refine_cls=torch.tensor(0).float().cuda(), acc=torch.tensor(0).float().cuda())) return losses rois = bbox2roi(samp_list).float() if self.refine_head.loss_opinion is not None: pred_scores = torch.cat([torch.tensor(bbox[:, 4]).float().cuda() for bbox in bbox_list], dim=0) pred_rois = bbox2roi([torch.tensor(bbox).float().cuda() for bbox in bbox_list]) pred_feats = self.refine_roi_extractor( x, pred_rois) pred_scores_refine = self.refine_head(pred_feats) loss_opinion = self.refine_head.compute_opinion_loss(pred_scores, pred_scores_refine) losses.update(loss_opinion) bbox_feats = self.refine_roi_extractor( x, rois) cls_score = self.refine_head(bbox_feats) bbox_targets = self.refine_head.get_target( sampling_results, gt_bboxes, gt_labels, self.train_cfg.rcnn) loss_refine = self.refine_head.loss(cls_score, *bbox_targets[:2]) losses.update(dict(loss_refine_cls=loss_refine["loss_cls"], distL1=loss_refine["dist"])) return losses def simple_test_accuracy(self, img, img_meta): gts = img_meta[0]["gts"] x = self.extract_feat(img) if self.detached: x = (x[0].detach(),) rois = bbox2roi(gts) if rois.shape[0] == 0: return 0, 0 roi_feats = self.refine_roi_extractor( x, rois) cls_score = self.refine_head.get_scores(roi_feats) return (cls_score > 0.5).float().sum(), rois.size(0) def simple_test(self, img, img_meta, rescale=False, return_id=False): x = self.extract_feat(img) outs = self.bbox_head(x) bbox_inputs = outs + (img_meta, self.test_cfg.csp_head if self.refine else self.test_cfg, False) # TODO://Handle rescalling if self.return_feature_maps: return self.bbox_head.get_bboxes_features(*bbox_inputs) bbox_list = self.bbox_head.get_bboxes(*bbox_inputs, no_strides=False) im_scale = img_meta[0]["scale_factor"] if "id" in img_meta[0]: img_id = img_meta[0]["id"] else: img_id = 0 if self.refine: if self.detached: x = (x[0].detach(),) bbox_list = [ bbox2result(det_bboxes, det_labels, self.bbox_head.num_classes)[0] for det_bboxes, det_labels in bbox_list ] refine_cfg = self.test_cfg.get('rcnn', None) bbox_list = [torch.tensor(bbox).float().cuda() for bbox in bbox_list] rois = bbox2roi(bbox_list) bbox_list = [bbox/im_scale for bbox in bbox_list] if rois.shape[0] == 0: cls_score = None else: roi_feats = self.refine_roi_extractor( x, rois) cls_score = self.refine_head.get_scores(roi_feats) res_buffer = [] if cls_score is not None: if refine_cfg is not None: res_buffer = self.refine_head.suppress_boxes(rois, cls_score, img_meta, cfg=refine_cfg) else: res_buffer = self.refine_head.combine_scores(bbox_list, cls_score) if return_id: return res_buffer, img_id return res_buffer bbox_results = [ bbox2result(det_bboxes, det_labels, self.bbox_head.num_classes) for det_bboxes, det_labels in bbox_list ] if return_id: return bbox_results[0], img_id return bbox_results[0] def foward_features(self, features): bbox_list = self.bbox_head.get_bboxes(*features) bbox_results = [ bbox2result(det_bboxes, det_labels, self.bbox_head.num_classes) for det_bboxes, det_labels in bbox_list ] return bbox_results[0]
[ [ [ 27, 46 ], [ 317, 336 ] ], [ [ 70, 79 ], [ 281, 290 ] ], [ [ 103, 114 ] ], [ [ 122, 136 ] ], [ [ 144, 149 ], [ 11879, 11884 ], [ 12464, 12469 ], [ 12500, 12505 ], [ 12698, 12703 ], [ 12709, 12714 ], [ 12818, 12823 ], [ 15106, 15111 ] ], [ [ 165, 172 ], [ 901, 908 ], [ 999, 1006 ] ], [ [ 180, 191 ], [ 1450, 1452 ], [ 1597, 1599 ], [ 3263, 3265 ], [ 3439, 3441 ], [ 5234, 5236 ], [ 5410, 5412 ], [ 7172, 7174 ], [ 7442, 7444 ] ], [ [ 199, 202 ], [ 2194, 2197 ], [ 2876, 2879 ], [ 2999, 3002 ], [ 3042, 3045 ], [ 3774, 3777 ], [ 4843, 4846 ], [ 4969, 4972 ], [ 5016, 5019 ], [ 5745, 5748 ], [ 6782, 6785 ], [ 6908, 6911 ], [ 6955, 6958 ], [ 8136, 8139 ], [ 8871, 8874 ], [ 9013, 9016 ], [ 9059, 9062 ], [ 9154, 9157 ], [ 9201, 9204 ] ], [ [ 226, 234 ], [ 12582, 12590 ], [ 12808, 12816 ], [ 13882, 13890 ], [ 15182, 15190 ] ], [ [ 236, 247 ], [ 11210, 11221 ], [ 14887, 14898 ], [ 15964, 15975 ], [ 16322, 16333 ] ], [ [ 249, 263 ], [ 11376, 11390 ] ], [ [ 265, 278 ], [ 11448, 11461 ] ], [ [ 313, 316 ], [ 710, 713 ] ] ]
# Simple demo of sending and recieving data with the RFM95 LoRa radio. # Author: Tony DiCola import board import busio import digitalio import adafruit_rfm9x # Define radio parameters. RADIO_FREQ_MHZ = 915.0 # Frequency of the radio in Mhz. Must match your # module! Can be a value like 915.0, 433.0, etc. # Define pins connected to the chip, use these if wiring up the breakout according to the guide: CS = digitalio.DigitalInOut(board.D5) RESET = digitalio.DigitalInOut(board.D6) # Or uncomment and instead use these if using a Feather M0 RFM9x board and the appropriate # CircuitPython build: # CS = digitalio.DigitalInOut(board.RFM9X_CS) # RESET = digitalio.DigitalInOut(board.RFM9X_RST) # Define the onboard LED LED = digitalio.DigitalInOut(board.D13) LED.direction = digitalio.Direction.OUTPUT # Initialize SPI bus. spi = busio.SPI(board.SCK, MOSI=board.MOSI, MISO=board.MISO) # Initialze RFM radio rfm9x = adafruit_rfm9x.RFM9x(spi, CS, RESET, RADIO_FREQ_MHZ) # Note that the radio is configured in LoRa mode so you can't control sync # word, encryption, frequency deviation, or other settings! # You can however adjust the transmit power (in dB). The default is 13 dB but # high power radios like the RFM95 can go up to 23 dB: rfm9x.tx_power = 23 # Send a packet. Note you can only send a packet up to 252 bytes in length. # This is a limitation of the radio packet size, so if you need to send larger # amounts of data you will need to break it into smaller send calls. Each send # call will wait for the previous one to finish before continuing. rfm9x.send(bytes("Hello world!\r\n", "utf-8")) print("Sent Hello World message!") # Wait to receive packets. Note that this library can't receive data at a fast # rate, in fact it can only receive and process one 252 byte packet at a time. # This means you should only use this for low bandwidth scenarios, like sending # and receiving a single message at a time. print("Waiting for packets...") while True: packet = rfm9x.receive() # Optionally change the receive timeout from its default of 0.5 seconds: # packet = rfm9x.receive(timeout=5.0) # If no packet was received during the timeout then None is returned. if packet is None: # Packet has not been received LED.value = False print("Received nothing! Listening again...") else: # Received a packet! LED.value = True # Print out the raw bytes of the packet: print("Received (raw bytes): {0}".format(packet)) # And decode to ASCII text and print it too. Note that you always # receive raw bytes and need to convert to a text format like ASCII # if you intend to do string processing on your data. Make sure the # sending side is sending ASCII data before you try to decode! packet_text = str(packet, "ascii") print("Received (ASCII): {0}".format(packet_text)) # Also read the RSSI (signal strength) of the last received message and # print it. rssi = rfm9x.last_rssi print("Received signal strength: {0} dB".format(rssi))
[ [ [ 100, 105 ], [ 436, 441 ], [ 477, 482 ], [ 752, 757 ], [ 845, 850 ], [ 861, 866 ], [ 878, 883 ] ], [ [ 113, 118 ], [ 835, 840 ] ], [ [ 126, 135 ], [ 413, 422 ], [ 454, 463 ], [ 729, 738 ], [ 779, 788 ] ], [ [ 144, 158 ], [ 921, 935 ] ], [ [ 188, 202 ], [ 958, 972 ] ], [ [ 408, 410 ], [ 947, 949 ] ], [ [ 446, 451 ], [ 951, 956 ] ], [ [ 723, 726 ], [ 763, 766 ], [ 2272, 2275 ], [ 2391, 2394 ] ], [ [ 829, 832 ], [ 942, 945 ] ], [ [ 913, 918 ], [ 1245, 1250 ], [ 1569, 1574 ], [ 1993, 1998 ], [ 3031, 3036 ] ], [ [ 1984, 1990 ], [ 2209, 2215 ], [ 2506, 2512 ], [ 2840, 2846 ] ], [ [ 2822, 2833 ], [ 2902, 2913 ] ], [ [ 3024, 3028 ], [ 3103, 3107 ] ] ]
""" Class to hold clinical outcome model. Predicts probability of good outcome of patient(s) or group(s) of patients. Call `calculate_outcome_for_all(args)` from outside of the object Inputs ====== All inputs take np arrays (for multiple groups of patients). mimic: proportion of patients with stroke mimic ich: proportion of patients with intracerebral haemorrhage (ICH). Or probability of a patient having an ICH, when using for a single patient. nlvo: proportion of patients with non-large vessel occlusions (nLVO). Or probability of a patient having an NLVO, when using for a single patient. lvo: proportion of patients with large vessel occlusions (LVO). Or probability of a patient having a LVO, when using for a single patient. onset_to_needle: minutes from onset to thrombolysis onset_to_ouncture: minutes from onset to thrombectomy nlvo_eligible_for_treatment: proportion of patients with NLVO suitable for treatment with thrombolysis. Or probability of a patient with NVLO being eligible for treatment. lvo_eligible_for_treatment: proportion of patients with LVO suitable for treatment with thrombolysis and/or thrombectomy. Or probability of a patient with LVO being eligible for treatment. Returns ======= Probability of good outcome: The probability of having a good outcome (modified Rankin Scale 0-1) for the patient or group of patients (np array). References for decay of effect of thrombolysis and thrombectomy =============================================================== Decay of effect of thrombolysis without image selection of patients taken from: Emberson, Jonathan, Kennedy R. Lees, Patrick Lyden, Lisa Blackwell, Gregory Albers, Erich Bluhmki, Thomas Brott, et al (2014). “Effect of Treatment Delay, Age, and Stroke Severity on the Effects of Intravenous Thrombolysis with Alteplase for Acute Ischaemic Stroke: A Meta-Analysis of Individual Patient Data from Randomised Trials.” The Lancet 384: 1929–1935. https://doi.org/10.1016/S0140-6736(14)60584-5. * Time to no effect = 6.3hrs Decay of effect of thrombectomy without image selection of patients taken from: Fransen, Puck S. S., Olvert A. Berkhemer, Hester F. Lingsma, Debbie Beumer, Lucie A. van den Berg, Albert J. Yoo, Wouter J. Schonewille, et al. (2016) “Time to Reperfusion and Treatment Effect for Acute Ischemic Stroke: A Randomized Clinical Trial.” JAMA Neurology 73: 190–96. https://doi.org/10.1001/jamaneurol.2015.3886. * Time to no effect = 8hrs """ import numpy as np import pandas as pd class Clinical_outcome: def __init__(self): """Constructor for clinical outcome model """ self.name = "Clinical outcome model" self.thrombectomy_time_no_effect = 8 * 60 self.thrombolysis_time_no_effect = 6.3 * 60 self.maximum_permitted_time_to_thrombectomy = 360 self.maximum_permitted_time_to_thrombolysis = 270 def calculate_outcome_for_all(self, mimic, ich, nlvo, lvo, onset_to_needle, onset_to_puncture, nlvo_eligible_for_treatment, lvo_eligible_for_treatment, prop_thrombolysed_lvo_receiving_thrombectomy): """ Calculates the probability of good outcome for all patients admitted with acute stroke. Based on: Holodinsky JK, Williamson TS, Demchuk AM, et al. Modeling Stroke Patient Transport for All Patients With Suspected Large-Vessel Occlusion. JAMA Neurol. 2018;75(12):1477-1486. doi:10.1001/jamaneurol.2018.2424 Sums outcomes for: 1) mimics 2) ICH 3) non-LVO 4) LVO treated with thrombolysis 5) LVO treated with thrombectomy (if thrombolysis not successful in a drip and ship configuration) arguments --------- np arrays (each row is a given geographic area with different characteristics) mimic: proportion of patients with stroke mimic ich: proportion of patients with ICH nlvo: proportion of patients with non-lvo lvo: proportion of patients with lvo onset_to_needle: minutes from onset to thrombolysis onset_to_ounctureL minutes from onset to thrombectomy nlvo_eligible_for_treatment: proportion of nlvo suitable for treatment lvo_eligible_for_treatment: proportion of lvo suitable for treatment returns ------- probability of good outcome for all (np array) """ # Get outcomes # ------------ outcomes = pd.DataFrame() # Calculate good outcomes for mimics outcomes['mimic'] = self._calculate_outcome_for_stroke_mimics( mimic.shape) # Calculate good outcomes for ich outcomes['ich'] = self._calculate_outcome_for_ICH(mimic.shape) # Calculate good outcomes for nlvo without treatment outcomes['nlvo_base'] = \ np.full(nlvo.shape, 0.4622) # Calculate good outcomes for nlvo with thrombolysis outcomes['nlvo_add_ivt'] = \ self._calculate_thrombolysis_outcome_for_nlvo(onset_to_needle) # Calculate good outcomes for lvo without treatment outcomes['lvo_base'] = \ np.full(nlvo.shape, 0.1328) # Calculate good outcomes for lvo with thrombolysis outcomes['lvo_add_ivt'] = \ self._calculate_thrombolysis_outcome_for_lvo(onset_to_needle) # Calculate good outcomes for lvo with thrombolysis outcomes['lvo_add_et'] = \ self._calculate_thrombectomy_outcome_for_lvo(onset_to_puncture) # Weight outcome results by proportion of patients # ------------------------------------------------ # 'Results' are good outcomes results = pd.DataFrame() # Results for mimic results['mimic'] = outcomes['mimic'] * mimic # Results for ich results['ich'] = outcomes['ich'] * ich # Results for nlvo results['nlvo_base'] = nlvo * outcomes['nlvo_base'] results['nlvo_ivt'] = \ nlvo * outcomes['nlvo_add_ivt'] * nlvo_eligible_for_treatment # Results for lvo results['lvo_base'] = lvo * outcomes['lvo_base'] results['lvo_ivt'] = \ lvo * outcomes['lvo_add_ivt'] * lvo_eligible_for_treatment # Adjust thrombectomy/thrombolysis ratio for LVO # Reduce thrombectomy treatment by LVO responding to IVT lvo_receiving_et = ((lvo * lvo_eligible_for_treatment * prop_thrombolysed_lvo_receiving_thrombectomy) - results['lvo_ivt']) results['lvo_et'] = lvo_receiving_et * outcomes['lvo_add_et'] p_good = results.sum(axis=1).values return p_good @staticmethod def _calculate_outcome_for_ICH(array_shape): """ Calculates the probability of good outcome for patients with intra- cranial haemorrhage (ICH). Sets all values to 0.24 Based on Holodinsky et al. (2018) Drip-and-Ship vs. Mothership: Modelling Stroke Patient Transport for All Suspected Large Vessel Occlusion Patients. JAMA Neuro (in press) arguments --------- array size returns ------- probability of good outcome for ICH (np array) """ # Create an array of required length and set all values to 0.24 p_good = np.zeros(array_shape) p_good[:] = 0.24 return p_good @staticmethod def _calculate_outcome_for_stroke_mimics(array_shape): """ Calculates the probability of good outcome for patients with stroke mimic Sets all values to 1 Based on Holodinsky et al. (2018) Drip-and-Ship vs. Mothership: Modelling Stroke Patient Transport for All Suspected Large Vessel Occlusion Patients. JAMA Neuro (in press) arguments --------- array size returns ------- probability of good outcome for stroke mimiccs (np array) """ # Create an array of required length and set all values to 0.9 p_good = np.zeros(array_shape) p_good[:] = 1 return p_good def _calculate_thrombectomy_outcome_for_lvo(self, onset_to_puncture): """ Calculates the probability of additional good outcome for LVO patients receiving thrombectomy. arguments --------- onset_to_puncture : np array in minutes returns ------- probability of additional good outcome if given thrombectomy (np array) """ p_good_max = 0.5208 p_good_min = 0.1328 # Convert probability to odds odds_good_max = p_good_max / (1 - p_good_max) odds_good_min = p_good_min / (1 - p_good_min) # Calculate fraction of effective time used fraction_max_effect_time_used = \ onset_to_puncture / self.thrombectomy_time_no_effect # Calculate odds of good outcome with treatment odds_good = np.exp(np.log(odds_good_max) - ((np.log(odds_good_max) - np.log(odds_good_min)) * fraction_max_effect_time_used)) # Convert odds to probability prob_good = odds_good / (1 + odds_good) prob_good[prob_good < p_good_min] = p_good_min # Calculate probability of additional good outcome p_good_add = prob_good - p_good_min # Set additional good outcomes to zero if past permitted treatment time mask = onset_to_puncture > self.maximum_permitted_time_to_thrombectomy p_good_add[mask] = 0 # Ensure no negative outcomes mask = p_good_add < 0 p_good_add[mask] = 0 return p_good_add def _calculate_thrombolysis_outcome_for_lvo(self, onset_to_needle): """ Calculates the probability of additional good outcome for LVO patients receiving thrombolysis. Does not include baseline untreated good comes. arguments --------- onset_to_needle : np array in minutes returns ------- probability of additional good outcome if given thrombolysis (np array) """ p_good_max = 0.2441 p_good_min = 0.1328 # Convert probability to odds odds_good_max = p_good_max / (1 - p_good_max) odds_good_min = p_good_min / (1 - p_good_min) # Calculate fraction of effective time used fraction_max_effect_time_used = \ onset_to_needle / self.thrombolysis_time_no_effect # Calculate odds of good outcome with treatment odds_good = np.exp(np.log(odds_good_max) - ((np.log(odds_good_max) - np.log(odds_good_min)) * fraction_max_effect_time_used)) # Convert odds to probability prob_good = odds_good / (1 + odds_good) prob_good[prob_good < p_good_min] = p_good_min # Calculate probability of additional good outcome p_good_add = prob_good - p_good_min # Set additional good outcomes to zero if past permitted treatment time mask = onset_to_needle> self.maximum_permitted_time_to_thrombolysis p_good_add[mask] = 0 # Ensure no negative outcomes mask = p_good_add < 0 p_good_add[mask] = 0 # return outcome and proportion of treated who respond return p_good_add def _calculate_thrombolysis_outcome_for_nlvo(self, onset_to_needle): """ Calculates the probability of good outcome for non-LVO patients receiving thrombolysis. arguments --------- onset_to_needle : np array in minutes returns ------- probability of good outcome if given thrombolysis (np array) """ p_good_max = 0.6444 p_good_min = 0.4622 # Convert probability to odds odds_good_max = p_good_max / (1 - p_good_max) odds_good_min = p_good_min / (1 - p_good_min) # Calculate fraction of effective time used fraction_max_effect_time_used = (onset_to_needle / self.thrombolysis_time_no_effect) # Calculate odds of good outcome with treatment odds_good = np.exp(np.log(odds_good_max) - ((np.log(odds_good_max) - np.log(odds_good_min)) * fraction_max_effect_time_used)) # Convert odds to probability prob_good = odds_good / (1 + odds_good) prob_good[prob_good < p_good_min] = p_good_min # Calculate probability of additional good outcome p_good_add = prob_good - p_good_min mask = onset_to_needle> self.maximum_permitted_time_to_thrombolysis p_good_add[mask] = 0 # Ensure no negative outcomes mask = p_good_add < 0 p_good_add[mask] = 0 # return outcome and proportion of treated who respond return p_good_add
[ [ [ 2483, 2494 ], [ 5194, 5196 ], [ 5514, 5516 ], [ 7799, 7801 ], [ 8543, 8545 ], [ 9489, 9491 ], [ 9496, 9498 ], [ 9535, 9537 ], [ 9559, 9561 ], [ 11185, 11187 ], [ 11192, 11194 ], [ 11231, 11233 ], [ 11255, 11257 ], [ 12857, 12859 ], [ 12864, 12866 ], [ 12903, 12905 ], [ 12927, 12929 ] ], [ [ 2502, 2514 ], [ 4811, 4813 ], [ 6086, 6088 ] ], [ [ 2523, 2539 ] ] ]
# -*- coding: utf-8 -*- from django.conf.urls.defaults import patterns, url from djangopypi.feeds import ReleaseFeed urlpatterns = patterns("djangopypi.views", url(r'^$', "root", name="djangopypi-root"), url(r'^packages/$','packages.index', name='djangopypi-package-index'), url(r'^simple/$','packages.simple_index', name='djangopypi-package-index-simple'), url(r'^search/$','packages.search',name='djangopypi-search'), url(r'^pypi/$', 'root', name='djangopypi-release-index'), url(r'^rss/$', ReleaseFeed(), name='djangopypi-rss'), url(r'^simple/(?P<package>[\w\d_\.\-]+)/$','packages.simple_details', name='djangopypi-package-simple'), url(r'^pypi/(?P<package>[\w\d_\.\-]+)/$','packages.details', name='djangopypi-package'), url(r'^pypi/(?P<package>[\w\d_\.\-]+)/rss/$', ReleaseFeed(), name='djangopypi-package-rss'), url(r'^pypi/(?P<package>[\w\d_\.\-]+)/doap.rdf$','packages.doap', name='djangopypi-package-doap'), url(r'^pypi/(?P<package>[\w\d_\.\-]+)/manage/$','packages.manage', name='djangopypi-package-manage'), url(r'^pypi/(?P<package>[\w\d_\.\-]+)/manage/versions/$','packages.manage_versions', name='djangopypi-package-manage-versions'), url(r'^pypi/(?P<package>[\w\d_\.\-]+)/(?P<version>[\w\d_\.\-]+)/$', 'releases.details',name='djangopypi-release'), url(r'^pypi/(?P<package>[\w\d_\.\-]+)/(?P<version>[\w\d_\.\-]+)/doap.rdf$', 'releases.doap',name='djangopypi-release-doap'), url(r'^pypi/(?P<package>[\w\d_\.\-]+)/(?P<version>[\w\d_\.\-]+)/manage/$', 'releases.manage',name='djangopypi-release-manage'), url(r'^pypi/(?P<package>[\w\d_\.\-]+)/(?P<version>[\w\d_\.\-]+)/metadata/$', 'releases.manage_metadata',name='djangopypi-release-manage-metadata'), url(r'^pypi/(?P<package>[\w\d_\.\-]+)/(?P<version>[\w\d_\.\-]+)/files/$', 'releases.manage_files',name='djangopypi-release-manage-files'), url(r'^pypi/(?P<package>[\w\d_\.\-]+)/(?P<version>[\w\d_\.\-]+)/files/upload/$', 'releases.upload_file',name='djangopypi-release-upload-file'), )
[ [ [ 62, 70 ], [ 132, 140 ] ], [ [ 72, 75 ], [ 165, 168 ], [ 213, 216 ], [ 288, 291 ], [ 375, 378 ], [ 441, 444 ], [ 503, 506 ], [ 566, 569 ], [ 688, 691 ], [ 789, 792 ], [ 898, 901 ], [ 1009, 1012 ], [ 1123, 1126 ], [ 1269, 1272 ], [ 1396, 1399 ], [ 1533, 1536 ], [ 1673, 1676 ], [ 1833, 1836 ], [ 1984, 1987 ] ], [ [ 105, 116 ], [ 518, 529 ], [ 835, 846 ] ], [ [ 118, 129 ] ] ]
import pandas as pd from pandas.compat import StringIO import numpy numpy.set_printoptions(threshold=numpy.nan) def main(): df = pd.read_csv(StringIO(earnings), sep=",", header=None, names=['symbol', 'exchange', 'eps_pct_diff_surp', 'asof_date']) df = df.sort_values(by=['asof_date']) print(df.head()) print(len(df)) df.to_csv('../../data/events/nyse_earnings_surprises_2013.csv', index=False) myString = ', '.join('"{0}"'.format(s) for s in df.symbol.unique()) myString = myString.replace(" ", "") print(myString) #earnings = 'CFN, NYSE, -21.82, 2013-02-09\nNDZ, NYSE, 30.77, 2013-01-29\nAZZ, NYSE, -1.64, 2013-01-10' earnings = 'CFN, NYSE, -21.82, 2013-02-09\n NDZ, NYSE, 30.77, 2013-01-29\n AZZ, NYSE, -1.64, 2013-01-10\n CLC, NYSE, 2.86, 2013-01-17\n CMC, NYSE, 64.71, 2013-01-08\n FC, NYSE, 15.38, 2013-01-04\n FDO, NYSE, -6.76, 2013-01-04\n FUL, NYSE, 14.29, 2013-01-17\n LEN, NYSE, 30.23, 2013-01-16\n LNN, NYSE, 53.33, 2013-01-09\n MKC, NYSE, -3.48, 2013-01-25\n RT, NYSE, 0.00, 2013-01-10\n MSM, NYSE, 1.00, 2013-01-11\n RPM, NYSE, -4.76, 2013-01-09\n SVU, NYSE, -50.00, 2013-01-11\n TISI, NYSE, 10.00, 2013-01-08\n TXI, NYSE, -5.88, 2013-01-10\n UNF, NYSE, 15.79, 2013-01-04\n WOR, NYSE, 12.20, 2013-01-04\n GBX, NYSE, 12.90, 2013-01-10\n SJR, NYSE, 11.11, 2013-01-10\n OMN, NYSE, -50.00, 2013-01-23\n MON, NYSE, 67.57, 2013-01-09\n GPN, NYSE, 6.90, 2013-01-09\n AYI, NYSE, -13.75, 2013-01-09\n STZ, NYSE, 14.55, 2013-01-10\n SNX, NYSE, 11.54, 2013-01-11\n TAL, NYSE, 600.00, 2013-01-23\n IHS, NYSE, 12.35, 2013-01-09\n EDU, NYSE, -150.00, 2013-01-30\n SAR, NYSE, 28.57, 2013-01-15\n ZEP, NYSE, 11.11, 2013-01-08\n MG, NYSE, 0.00, 2013-01-09\n MOS, NYSE, 7.14, 2013-01-04\n ABT, NYSE, 1.33, 2013-01-24\n ABX, NYSE, 1.83, 2013-02-15\n AB, NYSE, 21.21, 2013-02-13\n TAP, NYSE, 7.81, 2013-02-15\n ACO, NYSE, -15.91, 2013-01-26\n ADM, NYSE, -26.83, 2013-02-05\n AEM, NYSE, -13.33, 2013-02-14\n AEP, NYSE, 11.11, 2013-02-16\n AES, NYSE, 6.67, 2013-02-28\n AET, NYSE, -2.08, 2013-02-01\n AFL, NYSE, 0.00, 2013-02-06\n AGCO, NYSE, 1.02, 2013-02-06\n HES, NYSE, -2.44, 2013-01-31\n AIG, NYSE, 322.22, 2013-02-22\n AIN, NYSE, -9.68, 2013-02-07\n AJG, NYSE, 2.63, 2013-01-30\n ALU, NYSE, 0.00, 2013-02-08\n MATX, NYSE, 24.14, 2013-02-08\n ALK, NYSE, -4.11, 2013-01-25\n ALX, NYSE, -11.52, 2013-02-27\n BEAM, NYSE, 0.00, 2013-02-02\n AME, NYSE, 2.08, 2013-01-25\n TWX, NYSE, 6.36, 2013-02-07\n AVD, NYSE, 11.43, 2013-03-01\n AMN, NYSE, 36.36, 2013-02-22\n AN, NYSE, 3.08, 2013-02-01\n AON, NYSE, 1.60, 2013-02-02\n AP, NYSE, 77.78, 2013-02-05\n APA, NYSE, -1.30, 2013-02-15\n APC, NYSE, 30.00, 2013-02-05\n APD, NYSE, 0.78, 2013-01-24\n APH, NYSE, 4.44, 2013-01-18\n ARG, NYSE, -3.70, 2013-01-25\n AAN, NYSE, -4.00, 2013-02-08\n ARW, NYSE, 13.89, 2013-02-08\n ASGN, NYSE, -25.00, 2013-02-15\n ASH, NYSE, -17.65, 2013-01-30\n ASR, NYSE, 56.88, 2013-02-26\n GAS, NYSE, -9.90, 2013-02-07\n ATO, NYSE, -5.13, 2013-02-07\n ATW, NYSE, 17.02, 2013-01-31\n AU, NYSE, -67.44, 2013-02-21\n AVP, NYSE, 37.04, 2013-02-13\n AVT, NYSE, 21.69, 2013-01-25\n AVY, NYSE, 10.20, 2013-01-31\n AXP, NYSE, 0.00, 2013-01-18\n B, NYSE, 7.84, 2013-02-23\n BA, NYSE, 7.56, 2013-01-31\n BAC, NYSE, 50.00, 2013-01-18\n BAX, NYSE, 0.00, 2013-01-25\n BC, NYSE, 122.22, 2013-01-25\n OMX, NYSE, 6.67, 2013-02-21\n BCE, NYSE, -2.99, 2013-02-08\n BCR, NYSE, 1.80, 2013-02-01\n BCS, NYSE, 40.74, 2013-02-13\n BDX, NYSE, 9.76, 2013-02-06\n BEN, NYSE, 1.68, 2013-02-02\n BGG, NYSE, 250.00, 2013-01-25\n BHE, NYSE, 10.00, 2013-02-05\n BHI, NYSE, 1.64, 2013-01-24\n BID, NYSE, 0.92, 2013-03-01\n BIO, NYSE, 15.67, 2013-02-27\n BK, NYSE, 0.00, 2013-01-16\n BKH, NYSE, 9.68, 2013-02-01\n WRB, NYSE, 28.00, 2013-01-29\n BLC, NYSE, 5.71, 2013-02-09\n BLL, NYSE, -3.03, 2013-02-01\n BLX, NYSE, 20.75, 2013-02-08\n BMI, NYSE, -11.36, 2013-02-07\n BMS, NYSE, 4.00, 2013-02-01\n BMY, NYSE, 9.30, 2013-01-25\n BOH, NYSE, 1.12, 2013-01-31\n BXS, NYSE, -25.00, 2013-01-24\n BPL, NYSE, 18.52, 2013-02-09\nBRK.A, NYSE, 175.73, 2013-03-02\n BRO, NYSE, 7.41, 2013-02-02\n BSX, NYSE, 63.64, 2013-01-30\n BT, NYSE, -89.22, 2013-02-02\n MTRN, NYSE, 17.14, 2013-03-01\n CACI, NYSE, 3.66, 2013-01-31\n CAT, NYSE, -13.10, 2013-01-29\n CB, NYSE, 10.00, 2013-01-30\n CBI, NYSE, 9.64, 2013-02-28\n CBM, NYSE, 100.00, 2013-02-07\n CBU, NYSE, -3.70, 2013-01-23\n CBT, NYSE, -28.57, 2013-01-31\n CCC, NYSE, 35.71, 2013-02-22\n CCE, NYSE, 4.65, 2013-02-08\n C, NYSE, -20.69, 2013-01-18\n CCK, NYSE, -7.27, 2013-01-31\n CCU, NYSE, -12.21, 2013-02-01\n CDE, NYSE, -15.15, 2013-02-22\n CDI, NYSE, 8.70, 2013-02-27\n CAH, NYSE, 9.41, 2013-02-06\n CFR, NYSE, 5.38, 2013-01-31\n CHD, NYSE, 0.00, 2013-02-06\n CKP, NYSE, -50.00, 2013-03-06\n CPK, NYSE, 18.60, 2013-03-08\n CI, NYSE, 6.08, 2013-02-08\n CIA, NYSE, -100.00, 2013-03-12\n CKH, NYSE, -93.55, 2013-02-28\n CL, NYSE, 0.71, 2013-02-01\n CLF, NYSE, -25.45, 2013-02-13\n CLH, NYSE, -25.00, 2013-02-21\n CLX, NYSE, 11.11, 2013-02-05\n CMA, NYSE, 7.81, 2013-01-17\n CMO, NYSE, -6.06, 2013-01-31\n CRK, NYSE, -77.42, 2013-02-12\n CMS, NYSE, 4.17, 2013-02-22\n CNA, NYSE, -150.00, 2013-02-12\n CNW, NYSE, -10.34, 2013-02-07\n CHG, NYSE, -4.12, 2013-02-27\n CNL, NYSE, 12.50, 2013-02-20\n COG, NYSE, 14.29, 2013-02-22\n COT, NYSE, -66.67, 2013-02-16\n CP, NYSE, -0.78, 2013-01-30\n CPF, NYSE, 11.54, 2013-02-01\n CQB, NYSE, -17.65, 2013-03-12\n CR, NYSE, -5.15, 2013-01-29\nCRD.B, NYSE, 52.38, 2013-02-14\n CRS, NYSE, 1.64, 2013-02-01\n CSC, NYSE, 22.22, 2013-02-06\n CSL, NYSE, 6.49, 2013-02-09\n CTB, NYSE, 35.29, 2013-02-26\n CTL, NYSE, -1.47, 2013-02-14\n CTS, NYSE, -21.74, 2013-01-29\n CUB, NYSE, -32.86, 2013-02-12\n CMI, NYSE, 14.94, 2013-02-07\n CUZ, NYSE, 40.00, 2013-02-14\n CVC, NYSE, -400.00, 2013-03-01\n CVH, NYSE, 35.82, 2013-02-07\n CW, NYSE, 4.40, 2013-02-21\n CWT, NYSE, 33.33, 2013-02-28\n CX, NYSE, -258.33, 2013-02-08\n CYN, NYSE, -13.00, 2013-01-25\n D, NYSE, 1.47, 2013-02-01\n DBD, NYSE, -8.16, 2013-02-13\n DCO, NYSE, -23.81, 2013-03-05\n DD, NYSE, 22.22, 2013-01-23\n CVA, NYSE, -13.04, 2013-02-07\n DHR, NYSE, 0.00, 2013-01-30\n DIS, NYSE, 2.60, 2013-02-06\n DLX, NYSE, 11.76, 2013-01-25\n DNB, NYSE, -1.24, 2013-02-12\n RRD, NYSE, 16.22, 2013-02-27\n DOV, NYSE, 1.87, 2013-01-25\n DOW, NYSE, -2.94, 2013-02-01\n DRE, NYSE, 0.00, 2013-01-31\n DHI, NYSE, 42.86, 2013-01-30\n UFS, NYSE, -7.09, 2013-02-02\n DTE, NYSE, 0.00, 2013-02-21\n DUK, NYSE, 7.69, 2013-02-14\n DVN, NYSE, 2.63, 2013-02-21\n DV, NYSE, 55.36, 2013-02-07\n EAT, NYSE, 0.00, 2013-01-23\n ECL, NYSE, 0.00, 2013-02-27\n ED, NYSE, -6.85, 2013-02-01\n EDE, NYSE, 27.78, 2013-02-15\n EFX, NYSE, 4.00, 2013-02-07\n EGN, NYSE, -15.58, 2013-01-24\n EGP, NYSE, 0.00, 2013-02-13\n ELY, NYSE, 2.00, 2013-01-31\n EMC, NYSE, 6.98, 2013-01-30\n EMR, NYSE, 0.00, 2013-02-06\n EOG, NYSE, 19.26, 2013-02-14\n EQT, NYSE, 14.29, 2013-01-25\n ESE, NYSE, -44.44, 2013-02-08\n ESV, NYSE, 7.87, 2013-02-21\n ETN, NYSE, -10.87, 2013-02-06\n ETR, NYSE, 21.99, 2013-02-09\n EXAR, NYSE, -14.29, 2013-01-24\n F, NYSE, 19.23, 2013-01-30\n OPY, NYSE, 115.79, 2013-02-02\n CLGX, NYSE, -3.12, 2013-02-22\n FNB, NYSE, 4.55, 2013-01-24\n FCF, NYSE, -18.18, 2013-01-31\n FBP, NYSE, -30.00, 2013-02-06\n FICO, NYSE, 6.94, 2013-01-31\n FLO, NYSE, 12.00, 2013-02-08\n FMC, NYSE, 0.00, 2013-02-07\n FOE, NYSE, -250.00, 2013-03-06\n S, NYSE, 4.35, 2013-02-08\n NEE, NYSE, 9.57, 2013-01-30\n FRT, NYSE, 0.91, 2013-02-13\n FRX, NYSE, -61.54, 2013-01-16\n FUN, NYSE, -433.33, 2013-02-20\n FUR, NYSE, -48.15, 2013-03-08\n GBL, NYSE, -28.72, 2013-02-06\n GVA, NYSE, -29.03, 2013-03-01\n BGC, NYSE, -3.45, 2013-02-26\n GD, NYSE, -26.84, 2013-01-24\n GE, NYSE, 2.33, 2013-01-19\n RHP, NYSE, -50.00, 2013-02-13\n AXLL, NYSE, 95.08, 2013-02-13\n GGG, NYSE, 13.33, 2013-01-29\n GHM, NYSE, -22.22, 2013-02-02\n GIB, NYSE, -4.35, 2013-01-31\n GLT, NYSE, -25.71, 2013-02-08\n GLW, NYSE, 3.03, 2013-01-30\n GSK, NYSE, 8.33, 2013-02-07\n GLF, NYSE, -160.71, 2013-02-26\n GNI, NYSE, -14.44, 2013-01-30\n GPC, NYSE, 0.00, 2013-02-20\n GRA, NYSE, 4.72, 2013-02-07\n GTY, NYSE, -10.34, 2013-03-01\n GWW, NYSE, -7.28, 2013-01-25\n HAE, NYSE, 4.17, 2013-01-31\n HAL, NYSE, 3.28, 2013-01-26\n HAR, NYSE, -32.95, 2013-02-01\n HVT, NYSE, 30.43, 2013-02-26\n HRC, NYSE, 6.82, 2013-01-24\n HCC, NYSE, 43.75, 2013-02-13\n HCN, NYSE, 1.19, 2013-02-26\n HCP, NYSE, 1.41, 2013-02-13\n HOG, NYSE, 0.00, 2013-01-30\n HE, NYSE, 21.88, 2013-02-16\n HL, NYSE, -25.00, 2013-02-26\n HMA, NYSE, -5.00, 2013-02-15\n HMC, NYSE, -29.58, 2013-02-01\n HMN, NYSE, 91.43, 2013-02-06\n HFC, NYSE, -8.97, 2013-02-27\n HOT, NYSE, 7.69, 2013-02-08\n HP, NYSE, 8.53, 2013-02-01\n HLS, NYSE, 40.63, 2013-02-19\n HRS, NYSE, 4.17, 2013-01-30\n HSC, NYSE, -3.23, 2013-02-15\n HSY, NYSE, -1.33, 2013-02-01\n HUBB, NYSE, 0.00, 2013-01-25\n HUM, NYSE, 11.21, 2013-02-05\n HXL, NYSE, -5.26, 2013-01-24\n IBM, NYSE, 2.67, 2013-01-23\n IDA, NYSE, 10.00, 2013-02-22\n IEX, NYSE, 2.99, 2013-02-05\n IFF, NYSE, -1.19, 2013-02-08\n DIN, NYSE, 1.22, 2013-02-28\n INT, NYSE, 0.00, 2013-02-22\n IP, NYSE, 6.15, 2013-01-30\n IPG, NYSE, 3.70, 2013-02-23\n IO, NYSE, 30.77, 2013-02-14\n IR, NYSE, 8.57, 2013-02-02\n IRF, NYSE, 6.38, 2013-01-29\n ITW, NYSE, -1.11, 2013-01-30\n IVC, NYSE, -56.00, 2013-02-09\n JEC, NYSE, 0.00, 2013-01-24\n JNJ, NYSE, 1.71, 2013-01-23\n JNY, NYSE, 75.00, 2013-02-14\n K, NYSE, 3.08, 2013-02-06\n KAMN, NYSE, 0.00, 2013-02-26\n KDN, NYSE, 0.00, 2013-02-22\n KEX, NYSE, 9.30, 2013-01-31\n KEY, NYSE, -4.55, 2013-01-25\n KIM, NYSE, 6.45, 2013-02-06\n KMB, NYSE, 0.74, 2013-01-26\n KEM, NYSE, 53.33, 2013-02-01\n KMT, NYSE, -21.88, 2013-01-25\n KO, NYSE, 2.27, 2013-02-13\n KSU, NYSE, 10.98, 2013-01-23\n LDL, NYSE, -10.53, 2013-02-27\n LDR, NYSE, 10.42, 2013-02-12\n LEE, NYSE, 25.00, 2013-01-23\n LEG, NYSE, 10.34, 2013-02-05\n LLY, NYSE, 8.97, 2013-01-30\n LM, NYSE, 29.63, 2013-02-02\n LNC, NYSE, 3.77, 2013-02-07\n LPX, NYSE, -10.00, 2013-02-09\n LXU, NYSE, 145.00, 2013-03-01\n LTC, NYSE, -1.72, 2013-02-22\n L, NYSE, -37.93, 2013-02-12\n LUK, NYSE, 210.17, 2013-02-26\n LUV, NYSE, 28.57, 2013-01-25\n LUX, NYSE, 4.35, 2013-03-01\n MKL, NYSE, 314.07, 2013-02-05\n MAN, NYSE, 18.18, 2013-01-31\n MTW, NYSE, 12.50, 2013-02-01\n SM, NYSE, 95.65, 2013-02-21\n MAS, NYSE, 500.00, 2013-02-12\n MTZ, NYSE, 2.22, 2013-03-01\n MCD, NYSE, 3.76, 2013-01-24\n MDC, NYSE, 40.48, 2013-02-01\n MDP, NYSE, 1.14, 2013-01-25\n MDR, NYSE, 13.04, 2013-03-01\n MDU, NYSE, 2.56, 2013-02-05\n MED, NYSE, 12.00, 2013-03-08\n CVS, NYSE, 2.73, 2013-02-07\n MFC, NYSE, -12.50, 2013-02-08\n MGA, NYSE, 36.84, 2013-03-02\n MGM, NYSE, 0.00, 2013-02-21\n MLR, NYSE, -11.76, 2013-03-07\n MLI, NYSE, 14.29, 2013-02-06\n MMC, NYSE, 0.00, 2013-02-13\n MMM, NYSE, 0.00, 2013-01-25\n MSA, NYSE, 3.64, 2013-02-14\n MNR, NYSE, 38.46, 2013-02-08\n MO, NYSE, 1.85, 2013-02-01\n MOD, NYSE, -75.00, 2013-02-02\nMOG.A, NYSE, -8.54, 2013-01-26\n MHK, NYSE, 7.45, 2013-02-22\n MSI, NYSE, 7.61, 2013-01-24\n MCY, NYSE, -168.00, 2013-02-05\n MRK, NYSE, 2.47, 2013-02-02\n MRO, NYSE, -19.12, 2013-02-07\n POWR, NYSE, 18.18, 2013-03-08\n MTG, NYSE, -37.87, 2013-03-01\n MTB, NYSE, 2.76, 2013-01-17\n MTX, NYSE, 6.38, 2013-02-01\n MUR, NYSE, 59.23, 2013-01-31\n MYE, NYSE, -7.14, 2013-02-14\n NBL, NYSE, 54.21, 2013-02-08\n NBR, NYSE, 3.45, 2013-02-20\n NE, NYSE, -19.35, 2013-01-24\n NEM, NYSE, 13.27, 2013-02-22\n NFG, NYSE, 6.58, 2013-02-08\n NHI, NYSE, 1.20, 2013-02-15\n NI, NYSE, 0.00, 2013-02-20\n NJR, NYSE, -17.48, 2013-02-08\n THC, NYSE, -24.64, 2013-02-27\n NNN, NYSE, 4.55, 2013-02-08\n NOC, NYSE, 18.39, 2013-01-31\n NPK, NYSE, -11.23, 2013-02-16\n NR, NYSE, 0.00, 2013-02-15\n NSC, NYSE, 9.24, 2013-01-23\n NUE, NYSE, 55.17, 2013-01-30\n NVR, NYSE, 8.22, 2013-01-25\n NWL, NYSE, 2.38, 2013-02-02\n NWN, NYSE, -4.55, 2013-03-02\n NYT, NYSE, 3.23, 2013-02-08\n OCR, NYSE, 1.18, 2013-02-20\n OGE, NYSE, 14.71, 2013-02-28\n OHI, NYSE, 3.57, 2013-02-12\n OI, NYSE, 8.11, 2013-01-31\n OII, NYSE, 2.78, 2013-02-14\n OKE, NYSE, 17.78, 2013-02-26\n OLN, NYSE, 2.94, 2013-01-29\n BRS, NYSE, 32.95, 2013-02-05\n OLP, NYSE, 0.00, 2013-03-15\n OMC, NYSE, 3.67, 2013-02-13\n OMI, NYSE, -12.77, 2013-02-12\n ORB, NYSE, 31.82, 2013-02-15\n ORI, NYSE, -28.57, 2013-01-25\n OSK, NYSE, 93.55, 2013-01-26\n OXY, NYSE, 10.24, 2013-02-01\n PHX, NYSE, -18.75, 2013-02-08\n FCFS, NYSE, 2.20, 2013-01-24\n PBI, NYSE, 7.69, 2013-02-01\n PCG, NYSE, 3.51, 2013-02-22\n PCL, NYSE, 68.97, 2013-01-29\n PCP, NYSE, -3.23, 2013-01-25\n TPC, NYSE, 0.00, 2013-02-22\n PDS, NYSE, 250.00, 2013-02-15\n PEG, NYSE, 5.13, 2013-02-22\n PEI, NYSE, 0.00, 2013-02-26\n PEP, NYSE, 3.81, 2013-02-15\n PFE, NYSE, 6.82, 2013-01-30\n PG, NYSE, 9.91, 2013-01-26\n PGR, NYSE, 0.00, 2013-01-19\n PH, NYSE, 6.25, 2013-01-19\n PHG, NYSE, -4.17, 2013-01-30\n PHM, NYSE, 9.68, 2013-02-01\n PKD, NYSE, -150.00, 2013-02-22\n PKY, NYSE, 17.39, 2013-02-12\n PNC, NYSE, 24.82, 2013-01-18\n PNM, NYSE, 18.18, 2013-03-02\n PNR, NYSE, 6.82, 2013-01-30\n PNW, NYSE, 41.18, 2013-02-23\n POM, NYSE, -5.00, 2013-03-02\n POT, NYSE, -11.86, 2013-02-01\n PPG, NYSE, -0.65, 2013-01-15\n PPL, NYSE, 6.52, 2013-02-15\n PRGO, NYSE, 3.82, 2013-02-02\n PL, NYSE, 11.36, 2013-02-07\n PSB, NYSE, 5.04, 2013-02-20\n CSH, NYSE, 12.61, 2013-01-25\n PWR, NYSE, 36.11, 2013-02-22\n PX, NYSE, 0.00, 2013-01-24\n KWR, NYSE, 26.32, 2013-03-07\n R, NYSE, 6.36, 2013-02-01\n RBC, NYSE, 2.70, 2013-02-05\n RDC, NYSE, 28.57, 2013-03-01\n HTSI, NYSE, -20.69, 2013-02-01\n RES, NYSE, 8.33, 2013-01-24\n RGS, NYSE, -76.92, 2013-02-01\n RGR, NYSE, 36.99, 2013-02-28\n RHI, NYSE, 2.44, 2013-01-30\n RJF, NYSE, 0.00, 2013-01-24\n RLI, NYSE, 102.27, 2013-01-24\n ROG, NYSE, -8.62, 2013-02-20\n ROK, NYSE, -2.38, 2013-01-31\n ROL, NYSE, -5.88, 2013-01-24\n ROP, NYSE, 1.37, 2013-01-29\n RTI, NYSE, 25.00, 2013-02-07\n RTN, NYSE, 23.08, 2013-01-25\n RYL, NYSE, 12.00, 2013-01-30\n BSAC, NYSE, -1.96, 2013-02-05\n T, NYSE, -6.38, 2013-01-25\n SCG, NYSE, 0.00, 2013-02-22\n SCHW, NYSE, 0.00, 2013-01-17\n SCL, NYSE, -5.56, 2013-02-20\n SMG, NYSE, 0.88, 2013-02-07\n SEE, NYSE, 17.24, 2013-02-20\n SF, NYSE, 5.17, 2013-02-26\n SFE, NYSE, -121.74, 2013-03-08\n SHW, NYSE, -0.87, 2013-02-01\n STC, NYSE, 29.27, 2013-02-15\n SJI, NYSE, -6.67, 2013-03-01\n JOE, NYSE, -1000.00, 2013-03-01\n SJW, NYSE, 72.22, 2013-02-20\n SLB, NYSE, 0.00, 2013-01-19\n HSH, NYSE, 29.17, 2013-02-01\n AOS, NYSE, 12.35, 2013-01-25\n SNA, NYSE, 4.38, 2013-02-08\n PII, NYSE, 0.81, 2013-01-30\n SNV, NYSE, 0.00, 2013-01-23\n SO, NYSE, 12.82, 2013-01-31\n SON, NYSE, 3.70, 2013-02-14\n SPA, NYSE, 30.00, 2013-02-06\n TRV, NYSE, 500.00, 2013-01-23\n SR, NYSE, 14.68, 2013-02-06\n NVE, NYSE, 0.00, 2013-02-23\n SCI, NYSE, 10.00, 2013-02-13\n SSP, NYSE, -3.85, 2013-02-27\n STT, NYSE, 11.00, 2013-01-19\n STI, NYSE, 6.56, 2013-01-19\n STJ, NYSE, 2.22, 2013-01-24\n STL, NYSE, 14.29, 2013-01-24\n STR, NYSE, 8.57, 2013-02-21\n STE, NYSE, 3.57, 2013-02-07\n SYK, NYSE, 0.88, 2013-01-24\n SUN, NYSE, -4.88, 2013-03-30\n SUP, NYSE, -61.54, 2013-03-02\n SWK, NYSE, 3.01, 2013-01-25\n SWN, NYSE, 2.33, 2013-02-21\n SWS, NYSE, 0.00, 2013-02-07\n SWX, NYSE, -2.44, 2013-02-27\n SWY, NYSE, 23.68, 2013-02-22\n SXI, NYSE, 1.10, 2013-02-02\n SYY, NYSE, 19.51, 2013-02-05\n TNC, NYSE, 6.90, 2013-02-20\n TCB, NYSE, -16.67, 2013-01-31\n TCO, NYSE, 5.15, 2013-02-14\n TDS, NYSE, -725.00, 2013-02-27\n TDW, NYSE, 38.64, 2013-02-02\n TDY, NYSE, 8.33, 2013-01-25\n TE, NYSE, 0.00, 2013-02-06\n TER, NYSE, 600.00, 2013-01-24\n TEVA, NYSE, -0.75, 2013-02-08\n TEX, NYSE, -51.28, 2013-02-20\n TFX, NYSE, 1.79, 2013-02-22\n TEN, NYSE, -2.94, 2013-02-01\n TKR, NYSE, 25.00, 2013-01-25\n TMK, NYSE, 1.53, 2013-02-05\n TMO, NYSE, 6.25, 2013-02-01\n TOT, NYSE, -1.12, 2013-02-14\n TM, NYSE, -44.72, 2013-02-06\n TR, NYSE, 37.50, 2013-02-14\n TRN, NYSE, 7.14, 2013-02-21\n TRP, NYSE, -15.09, 2013-02-13\n TRR, NYSE, 566.67, 2013-02-07\n TSO, NYSE, -2.90, 2013-02-07\n TSS, NYSE, -3.03, 2013-01-23\n TTI, NYSE, -21.05, 2013-03-01\n TXT, NYSE, -1.75, 2013-01-24\n TYL, NYSE, 10.71, 2013-02-07\n TSN, NYSE, 23.08, 2013-02-02\n UDR, NYSE, 2.94, 2013-02-06\n UFI, NYSE, -42.86, 2013-01-23\n UGI, NYSE, -15.89, 2013-02-01\n UAM, NYSE, 45.45, 2013-02-20\n UHS, NYSE, 9.89, 2013-03-01\n UHT, NYSE, 268.42, 2013-02-28\n UIL, NYSE, -9.68, 2013-02-22\n UNH, NYSE, 0.00, 2013-01-18\n KMPR, NYSE, -250.00, 2013-02-08\n UNM, NYSE, 5.13, 2013-02-06\n UNP, NYSE, 1.39, 2013-01-25\n UNT, NYSE, 2.06, 2013-02-20\n URS, NYSE, -1.04, 2013-02-26\n USG, NYSE, -67.86, 2013-02-07\n MUX, NYSE, -600.00, 2013-03-09\n USM, NYSE, -1100.00, 2013-02-27\n USPH, NYSE, 3.03, 2013-03-08\n UTL, NYSE, 3.13, 2013-01-31\n UTX, NYSE, 26.47, 2013-01-24\n VMI, NYSE, 8.48, 2013-02-13\n VAR, NYSE, 3.49, 2013-01-24\n VFC, NYSE, 1.32, 2013-02-16\n CBS, NYSE, -8.57, 2013-02-15\n VLO, NYSE, 57.98, 2013-01-30\n VMC, NYSE, -81.82, 2013-02-15\n VLY, NYSE, 0.00, 2013-01-31\n VNO, NYSE, 6.09, 2013-02-27\n VSH, NYSE, 37.50, 2013-02-06\n WTS, NYSE, 5.17, 2013-02-20\n WBS, NYSE, 6.12, 2013-01-19\n WEC, NYSE, 4.88, 2013-01-31\n WFC, NYSE, 3.41, 2013-01-14\n WG, NYSE, 57.14, 2013-03-07\n WGL, NYSE, 9.62, 2013-02-07\n WHR, NYSE, 3.15, 2013-02-01\n WMB, NYSE, -3.85, 2013-02-21\n WMK, NYSE, 20.29, 2013-03-06\n WNC, NYSE, 3.23, 2013-02-06\n TEG, NYSE, -5.32, 2013-03-01\n WR, NYSE, 80.00, 2013-03-01\n WRE, NYSE, 2.17, 2013-02-14\n WRI, NYSE, 4.44, 2013-02-15\n WPP, NYSE, -175.00, 2013-02-12\n WSO, NYSE, -12.77, 2013-02-15\n WST, NYSE, 8.93, 2013-02-22\n WWW, NYSE, 200.00, 2013-02-20\n WY, NYSE, 36.84, 2013-01-26\n X, NYSE, 45.33, 2013-01-30\n XL, NYSE, 138.24, 2013-02-08\n XOM, NYSE, 10.00, 2013-02-02\n XRX, NYSE, 7.14, 2013-01-25\n Y, NYSE, 54.64, 2013-02-22\n HRG, NYSE, -50.00, 2013-02-09\n CRY, NYSE, 33.33, 2013-02-15\n CHK, NYSE, 85.71, 2013-02-22\n DDR, NYSE, 0.00, 2013-02-13\n ELS, NYSE, 0.00, 2013-01-29\n ALG, NYSE, 37.93, 2013-03-07\n ETH, NYSE, 5.41, 2013-01-23\n ATR, NYSE, 0.00, 2013-02-08\n GGP, NYSE, 6.90, 2013-02-05\n MSL, NYSE, -10.00, 2013-01-30\n RCL, NYSE, 66.67, 2013-02-05\n CWEI, NYSE, -34.04, 2013-02-22\n HR, NYSE, 0.00, 2013-02-21\n RGA, NYSE, 35.56, 2013-02-01\n RIG, NYSE, 12.35, 2013-03-02\n SKT, NYSE, 2.22, 2013-02-13\n TWI, NYSE, -80.85, 2013-02-26\n BDN, NYSE, 17.86, 2013-02-07\n KGC, NYSE, -4.55, 2013-02-14\n YPF, NYSE, 26.67, 2013-03-13\n CPT, NYSE, 1.04, 2013-02-01\n SGY, NYSE, 67.27, 2013-02-26\n BFS, NYSE, -11.48, 2013-03-08\n BWA, NYSE, 3.57, 2013-02-15\n EQR, NYSE, 0.00, 2013-02-06\n CLP, NYSE, -81.25, 2013-02-08\n KOF, NYSE, -7.78, 2013-02-28\n OKS, NYSE, 3.13, 2013-02-26\n SQM, NYSE, -15.63, 2013-03-06\n BYD, NYSE, -138.46, 2013-03-05\n CBL, NYSE, 8.77, 2013-02-06\n DECK, NYSE, 7.36, 2013-03-01\n IT, NYSE, 6.78, 2013-02-08\n GFI, NYSE, -36.36, 2013-02-15\n HST, NYSE, 8.11, 2013-02-22\n LXP, NYSE, 0.00, 2013-02-22\n OMG, NYSE, -533.33, 2013-02-20\n REG, NYSE, 8.62, 2013-01-31\n TUC, NYSE, -5.56, 2013-03-08\n AF, NYSE, 7.14, 2013-01-24\n BFR, NYSE, 13.33, 2013-02-09\n HHS, NYSE, 26.32, 2013-02-01\n MHO, NYSE, -3.45, 2013-02-01\n NFX, NYSE, -36.36, 2013-02-20\n SPG, NYSE, 13.93, 2013-02-05\n SU, NYSE, -14.20, 2013-02-06\n SUI, NYSE, -2.44, 2013-02-22\n TV, NYSE, 5.13, 2013-02-26\n CGI, NYSE, 0.00, 2013-01-24\n CYT, NYSE, 77.42, 2013-02-01\n EMN, NYSE, 0.00, 2013-02-01\n GRT, NYSE, 0.00, 2013-02-15\n MAA, NYSE, -1.74, 2013-02-07\n PLT, NYSE, 0.00, 2013-01-30\n BZH, NYSE, 24.27, 2013-01-29\n ELX, NYSE, 0.00, 2013-02-01\n AGM, NYSE, -5.41, 2013-03-19\n MLM, NYSE, -13.21, 2013-02-13\n AKS, NYSE, 14.29, 2013-01-30\n ALB, NYSE, 18.18, 2013-01-23\n VRX, NYSE, -4.00, 2013-03-01\n CBR, NYSE, 140.00, 2013-02-22\n MAC, NYSE, 3.45, 2013-02-07\n RKT, NYSE, 5.47, 2013-01-23\n RYN, NYSE, 3.51, 2013-01-25\n ADC, NYSE, 1.96, 2013-02-28\nBRK.B, NYSE, 0.88, 2013-03-02\n EXP, NYSE, 0.00, 2013-02-07\n GGB, NYSE, -66.67, 2013-02-22\n SSD, NYSE, -100.00, 2013-02-08\n ESS, NYSE, 4.02, 2013-02-01\n FR, NYSE, 0.00, 2013-02-21\n HIW, NYSE, 0.00, 2013-02-13\n IMAX, NYSE, 58.33, 2013-02-22\n AIV, NYSE, 4.00, 2013-02-08\n FCH, NYSE, 50.00, 2013-02-20\n ITGR, NYSE, 6.00, 2013-02-26\n GEO, NYSE, 7.32, 2013-02-22\n CLI, NYSE, 4.76, 2013-02-08\n DAR, NYSE, -20.00, 2013-02-28\n RS, NYSE, 9.28, 2013-02-22\n CPE, NYSE, -66.67, 2013-03-15\n KNX, NYSE, 4.76, 2013-01-31\n O, NYSE, 3.70, 2013-02-15\n PKX, NYSE, -15.35, 2013-03-02\n COF, NYSE, -12.35, 2013-01-18\n CYD, NYSE, -23.14, 2013-02-28\n IRS, NYSE, 57.50, 2013-02-20\n MCK, NYSE, -13.50, 2013-02-01\n SWC, NYSE, 116.67, 2013-02-28\n STM, NYSE, -22.22, 2013-01-31\n TEO, NYSE, 28.36, 2013-03-01\n TRK, NYSE, 400.00, 2013-03-07\n GFF, NYSE, 300.00, 2013-01-31\n LMT, NYSE, -0.56, 2013-01-25\n APU, NYSE, -13.89, 2013-02-01\n AGU, NYSE, 6.93, 2013-02-22\n LH, NYSE, -4.35, 2013-02-09\n DDD, NYSE, 0.00, 2013-02-26\n WEX, NYSE, 0.94, 2013-02-07\n AFG, NYSE, 3.08, 2013-02-12\n RMD, NYSE, 3.92, 2013-01-25\n WAB, NYSE, 2.29, 2013-02-20\n CIB, NYSE, 20.39, 2013-03-05\n CAM, NYSE, -1.04, 2013-02-01\n FCX, NYSE, 5.41, 2013-01-23\n RNR, NYSE, 70.27, 2013-02-06\n AVX, NYSE, -20.00, 2013-01-25\n RWT, NYSE, 85.19, 2013-02-22\n AXE, NYSE, 0.76, 2013-01-30\n CLB, NYSE, 3.54, 2013-01-31\n MD, NYSE, 1.54, 2013-02-01\n THG, NYSE, 6.25, 2013-02-07\n BAP, NYSE, 3.72, 2013-02-06\n DO, NYSE, 28.18, 2013-02-06\n RE, NYSE, 175.86, 2013-02-07\n DST, NYSE, 17.82, 2013-02-01\n EL, NYSE, 11.54, 2013-02-06\n ESC, NYSE, -34.88, 2013-03-01\n MIG, NYSE, -100.00, 2013-02-13\n WAT, NYSE, 0.63, 2013-01-23\n EME, NYSE, 11.48, 2013-02-27\n HIG, NYSE, 80.00, 2013-02-05\n ITT, NYSE, 2.63, 2013-02-28\n SPN, NYSE, 4.26, 2013-02-27\n SWM, NYSE, -9.18, 2013-02-07\n SCCO, NYSE, 0.00, 2013-02-02\n RCI, NYSE, 20.55, 2013-02-15\n EIX, NYSE, 66.04, 2013-02-27\n IRM, NYSE, -20.00, 2013-03-01\n REV, NYSE, -19.18, 2013-02-06\n SPH, NYSE, -17.46, 2013-02-08\n CCJ, NYSE, 46.34, 2013-02-09\n PGI, NYSE, -6.67, 2013-02-14\n CRR, NYSE, 2.30, 2013-02-01\n BVN, NYSE, -26.67, 2013-03-01\n FCN, NYSE, 11.67, 2013-03-01\n RPT, NYSE, 8.00, 2013-02-13\n TUP, NYSE, 1.79, 2013-01-30\n ASB, NYSE, 0.00, 2013-01-18\n GWR, NYSE, -2.47, 2013-02-13\n TBI, NYSE, 35.71, 2013-02-07\n FFG, NYSE, 24.00, 2013-02-08\n USNA, NYSE, 4.96, 2013-02-06\n CSV, NYSE, 4.35, 2013-02-26\n LVB, NYSE, 12.77, 2013-03-07\n ALR, NYSE, 6.25, 2013-02-16\n OCN, NYSE, -7.84, 2013-03-01\n PAA, NYSE, 42.03, 2013-02-07\n DNR, NYSE, 24.14, 2013-02-22\n HMY, NYSE, 50.00, 2013-02-05\n TGI, NYSE, 5.80, 2013-01-31\n PAG, NYSE, 7.55, 2013-02-07\n GEL, NYSE, -2.86, 2013-02-15\n IM, NYSE, 23.73, 2013-02-14\n LIN, NYSE, -21.92, 2013-03-01\n NUS, NYSE, 2.11, 2013-02-07\n CNI, NYSE, -0.70, 2013-01-23\n LAD, NYSE, 10.45, 2013-02-21\n NSP, NYSE, 4.44, 2013-02-09\n DEL, NYSE, -29.63, 2013-02-28\n DGX, NYSE, -3.81, 2013-01-24\n KRC, NYSE, 3.23, 2013-01-31\n MTH, NYSE, 50.00, 2013-02-01\n NCR, NYSE, 4.35, 2013-02-08\n OFG, NYSE, -50.00, 2013-02-08\n IVZ, NYSE, -4.26, 2013-02-01\n DX, NYSE, 9.68, 2013-02-21\n FBC, NYSE, 38.27, 2013-02-09\n ALV, NYSE, 9.85, 2013-02-01\n ARE, NYSE, 0.87, 2013-02-08\n BBT, NYSE, 2.86, 2013-01-18\n CGG, NYSE, -59.32, 2013-03-02\n BXP, NYSE, 2.42, 2013-01-30\n MS, NYSE, 73.08, 2013-01-19\n SRT, NYSE, 200.00, 2013-02-28\n HLX, NYSE, 162.86, 2013-02-21\n FLS, NYSE, 0.35, 2013-02-22\n MT, NYSE, -880.00, 2013-02-07\n PXD, NYSE, -2.35, 2013-02-14\n SLG, NYSE, 0.87, 2013-01-31\n NAT, NYSE, 0.00, 2013-02-12\n CSU, NYSE, -22.22, 2013-03-07\n DRQ, NYSE, 2.70, 2013-03-01\n FDP, NYSE, -100.00, 2013-02-20\n NLY, NYSE, 35.29, 2013-02-07\n TLM, NYSE, -300.00, 2013-02-18\n TSM, NYSE, 0.00, 2013-01-18\n YUM, NYSE, 2.47, 2013-02-05\n AMG, NYSE, 4.94, 2013-01-30\n EPR, NYSE, -4.40, 2013-02-27\n FE, NYSE, 1.27, 2013-02-26\n LFL, NYSE, -80.00, 2013-05-01\n MTD, NYSE, 8.44, 2013-02-07\n SID, NYSE, 57.14, 2013-03-29\n IN, NYSE, -18.18, 2013-03-12\n AI, NYSE, 9.91, 2013-02-07\n URI, NYSE, 23.30, 2013-01-24\n INGR, NYSE, 4.26, 2013-02-08\n RAS, NYSE, 153.85, 2013-02-14\n UNS, NYSE, 12.50, 2013-02-27\n ASI, NYSE, -17.95, 2013-03-07\n ANH, NYSE, 7.14, 2013-02-08\n OFC, NYSE, 4.08, 2013-02-09\n GPX, NYSE, 6.67, 2013-02-27\n WAC, NYSE, 11.32, 2013-03-19\n RBA, NYSE, -12.50, 2013-02-27\n WDR, NYSE, 5.17, 2013-01-30\n LHO, NYSE, 4.44, 2013-02-21\n LNT, NYSE, -1.72, 2013-02-15\n LVLT, NYSE, 11.11, 2013-02-13\n MFA, NYSE, 0.00, 2013-03-07\n OME, NYSE, 33.33, 2013-03-06\n EQY, NYSE, 7.14, 2013-02-21\n FII, NYSE, 10.00, 2013-01-25\n FMX, NYSE, 39.60, 2013-02-28\n LLL, NYSE, 6.13, 2013-01-31\n VTR, NYSE, 2.06, 2013-02-16\n WCN, NYSE, -7.69, 2013-02-15\n AVB, NYSE, -0.71, 2013-01-31\n GIL, NYSE, 6.67, 2013-02-07\n HZO, NYSE, 10.00, 2013-01-30\n AWR, NYSE, 43.24, 2013-03-01\n CLS, NYSE, 46.67, 2013-01-23\n EPD, NYSE, 7.58, 2013-02-01\n RSG, NYSE, -13.95, 2013-02-08\n WM, NYSE, -5.00, 2013-02-15\n AKR, NYSE, 3.57, 2013-02-06\n CVG, NYSE, 4.17, 2013-02-08\n RRC, NYSE, 228.57, 2013-02-27\n SAP, NYSE, -2.38, 2013-01-24\n CCI, NYSE, 57.14, 2013-01-24\n PQ, NYSE, -20.00, 2013-03-01\n WFT, NYSE, -94.44, 2013-02-27\n CAA, NYSE, 14.29, 2013-02-01\n ENB, NYSE, -6.67, 2013-02-16\n GMK, NYSE, -8.33, 2013-02-28\n MMR, NYSE, 75.00, 2013-01-19\n PB, NYSE, 1.19, 2013-01-26\n VIV, NYSE, -7.25, 2013-02-26\n AXL, NYSE, -111.76, 2013-02-09\n BP, NYSE, 19.05, 2013-02-06\n ETM, NYSE, 13.04, 2013-02-09\n HT, NYSE, 10.00, 2013-02-21\n BYI, NYSE, 5.26, 2013-02-01\n CEB, NYSE, 4.84, 2013-02-07\n INFY, NYSE, 5.56, 2013-01-12\n JLL, NYSE, -0.38, 2013-01-30\n AZN, NYSE, 24.64, 2013-02-01\n SFG, NYSE, 7.23, 2013-01-30\n TREX, NYSE, 27.78, 2013-02-20\n GS, NYSE, 61.38, 2013-01-17\n SYX, NYSE, -144.44, 2013-03-06\n WCC, NYSE, -2.75, 2013-02-01\n JNPR, NYSE, 26.67, 2013-01-25\n RDN, NYSE, -146.43, 2013-02-12\n RAI, NYSE, 4.11, 2013-02-13\n SKX, NYSE, 172.73, 2013-02-14\n WTM, NYSE, 724.10, 2013-02-06\n NCI, NYSE, 29.17, 2013-02-15\n BLT, NYSE, -21.74, 2013-03-08\n BLK, NYSE, 5.88, 2013-01-18\n CIR, NYSE, 25.45, 2013-03-01\n PKG, NYSE, -1.61, 2013-01-23\n PKI, NYSE, 0.00, 2013-02-01\n UGP, NYSE, 38.10, 2013-02-21\n WWE, NYSE, 0.00, 2013-03-01\n SNN, NYSE, 2.86, 2013-02-08\n UPS, NYSE, -4.35, 2013-02-01\n XOXO, NYSE, 62.50, 2013-03-07\n SLF, NYSE, 36.36, 2013-02-14\n CDR, NYSE, 33.33, 2013-03-08\n RLH, NYSE, -21.43, 2013-03-01\n EW, NYSE, 16.88, 2013-02-05\n MET, NYSE, 5.93, 2013-02-13\n FBR, NYSE, -28.57, 2013-01-31\n VVC, NYSE, 23.81, 2013-02-15\n BAM, NYSE, 148.28, 2013-02-16\n NVS, NYSE, 0.00, 2013-01-24\n VGR, NYSE, -43.75, 2013-02-27\n BHLB, NYSE, 0.00, 2013-01-29\n CRL, NYSE, 6.67, 2013-02-14\n CYH, NYSE, 0.00, 2013-02-22\n MBT, NYSE, 65.71, 2013-03-20\n MTOR, NYSE, -375.00, 2013-01-31\n CNQ, NYSE, -29.55, 2013-03-08\n ERJ, NYSE, -25.27, 2013-03-13\n VZ, NYSE, -28.30, 2013-01-23\n EVC, NYSE, 12.50, 2013-02-28\n PBR, NYSE, 0.00, 2013-02-05\n XEL, NYSE, 3.57, 2013-02-01\n ALE, NYSE, 0.00, 2013-02-16\n HW, NYSE, -20.00, 2013-01-30\n POL, NYSE, 0.00, 2013-01-30\n UMC, NYSE, 0.00, 2013-02-07\n ASX, NYSE, 41.43, 2013-01-31\n COH, NYSE, -4.65, 2013-01-23\n CXW, NYSE, 7.32, 2013-02-14\n DVA, NYSE, 6.33, 2013-02-15\n EXC, NYSE, -1.54, 2013-02-08\n MCO, NYSE, 7.14, 2013-02-09\n BRFS, NYSE, 43.48, 2013-03-06\n TU, NYSE, -1.15, 2013-02-16\n WIT, NYSE, 0.00, 2013-01-18\n ERF, NYSE, 462.50, 2013-02-22\n GG, NYSE, -22.22, 2013-02-15\n HNT, NYSE, -2.70, 2013-01-31\n NXY, NYSE, -23.44, 2013-02-26\n NYCB, NYSE, -3.45, 2013-01-31\n SXT, NYSE, -8.33, 2013-02-08\n CPG, NYSE, -191.67, 2013-03-15\n AMX, NYSE, -40.00, 2013-02-13\n MPX, NYSE, -50.00, 2013-01-24\n OIS, NYSE, -5.82, 2013-02-20\n BH, NYSE, -35.35, 2013-01-26\n MMP, NYSE, 6.15, 2013-02-06\n PES, NYSE, 250.00, 2013-02-14\n ABB, NYSE, -18.75, 2013-02-15\n RDY, NYSE, -27.27, 2013-02-15\n KMR, NYSE, -19.23, 2013-02-22\n GEN, NYSE, -20.00, 2013-02-12\n ADS, NYSE, 2.38, 2013-02-01\n CVI, NYSE, 5.15, 2013-03-13\n FTI, NYSE, 0.00, 2013-02-13\n PRA, NYSE, 10.64, 2013-02-20\n STO, NYSE, 26.47, 2013-02-08\n BEL, NYSE, -266.67, 2013-02-21\n FIS, NYSE, -8.82, 2013-02-13\n COL, NYSE, 4.44, 2013-01-19\n KAI, NYSE, 7.32, 2013-02-27\n FRM, NYSE, 233.33, 2013-03-09\n ABC, NYSE, 0.00, 2013-01-25\n BG, NYSE, -76.15, 2013-02-08\n FRO, NYSE, 106.52, 2013-02-22\n ECA, NYSE, -3.12, 2013-02-15\n CS, NYSE, -54.76, 2013-02-08\n EEP, NYSE, -30.77, 2013-02-14\n CVX, NYSE, -1.65, 2013-02-02\n DB, NYSE, 280.49, 2013-02-01\n GXP, NYSE, 200.00, 2013-03-01\n JHX, NYSE, 371.43, 2013-02-28\n PFG, NYSE, 10.81, 2013-02-01\n PVR, NYSE, -227.78, 2013-02-21\n AAP, NYSE, 17.33, 2013-02-08\n KND, NYSE, 4.55, 2013-02-26\n WTW, NYSE, 9.09, 2013-02-14\n CNC, NYSE, 42.42, 2013-02-06\n PRU, NYSE, -2.87, 2013-02-07\n BCH, NYSE, 12.94, 2013-02-06\n NS, NYSE, -19.35, 2013-02-02\n ITUB, NYSE, -5.00, 2013-02-05\n SXL, NYSE, 20.88, 2013-02-21\n VALE, NYSE, -26.00, 2013-02-28\n TNP, NYSE, -128.57, 2013-04-20\n LCI, NYSE, 233.33, 2013-02-08\n AUO, NYSE, -122.73, 2013-02-07\n GTI, NYSE, 19.05, 2013-02-27\n HNR, NYSE, -127.27, 2013-05-04\n MWE, NYSE, -38.89, 2013-02-28\n NLS, NYSE, 4.55, 2013-03-05\n RGC, NYSE, 40.00, 2013-02-08\n SBS, NYSE, 48.25, 2013-03-22\n JAH, NYSE, 2.40, 2013-02-15\n NPO, NYSE, 110.71, 2013-02-08\n TRI, NYSE, 9.09, 2013-02-14\n CAE, NYSE, 12.50, 2013-02-14\n LF, NYSE, 971.43, 2013-02-07\n SNY, NYSE, 1.30, 2013-02-08\n WHG, NYSE, 15.91, 2013-02-08\n BANC, NYSE, -300.00, 2013-03-02\n GTN, NYSE, 4.35, 2013-02-21\n BAK, NYSE, -150.00, 2013-02-08\n COP, NYSE, 1.42, 2013-01-31\n CNP, NYSE, 40.00, 2013-02-28\n EEQ, NYSE, -18.18, 2013-02-15\n MRH, NYSE, 60.26, 2013-02-08\n NGS, NYSE, 26.09, 2013-03-15\n NRP, NYSE, 34.88, 2013-02-14\n PXP, NYSE, -22.64, 2013-02-22\n XEC, NYSE, 9.26, 2013-02-20\n IAG, NYSE, -11.11, 2013-02-21\n TS, NYSE, -16.44, 2013-02-22\n EGO, NYSE, 6.67, 2013-02-23\n JNS, NYSE, 35.71, 2013-01-25\n PFS, NYSE, 7.41, 2013-02-02\n ENH, NYSE, 21.68, 2013-02-08\n IHG, NYSE, 5.56, 2013-02-20\n CNX, NYSE, 95.45, 2013-02-01\n AMT, NYSE, -17.07, 2013-02-27\n ABG, NYSE, 10.77, 2013-02-20\n LII, NYSE, 0.00, 2013-02-06\n SRE, NYSE, 11.34, 2013-02-27\n AEE, NYSE, -36.36, 2013-02-21\n PLD, NYSE, 0.00, 2013-02-07\n SAH, NYSE, 4.00, 2013-02-21\n GPI, NYSE, -17.50, 2013-02-20\n FIX, NYSE, -11.11, 2013-03-01\n MMS, NYSE, 12.50, 2013-02-08\n SRI, NYSE, -28.57, 2013-03-02\n RTEC, NYSE, 6.25, 2013-02-05\n NOV, NYSE, 3.47, 2013-02-02\n DF, NYSE, 33.33, 2013-02-14\n SAM, NYSE, 1.63, 2013-02-21\n RL, NYSE, 8.60, 2013-02-07\n FLR, NYSE, 132.35, 2013-02-21\n ALL, NYSE, 942.86, 2013-02-07\n ATI, NYSE, 5.88, 2013-01-24\n EE, NYSE, -14.29, 2013-02-20\n AIT, NYSE, 0.00, 2013-02-01\n CHH, NYSE, 9.76, 2013-02-12\n FMS, NYSE, 105.77, 2013-02-27\n BCO, NYSE, -7.69, 2013-02-02\n CBB, NYSE, -125.00, 2013-02-28\n MWW, NYSE, 0.00, 2013-02-08\n PSA, NYSE, 5.68, 2013-02-22\n E, NYSE, 2.83, 2013-02-16\n JPM, NYSE, 15.83, 2013-01-17\n USB, NYSE, 1.35, 2013-01-17\n HON, NYSE, 0.92, 2013-01-26\n ITG, NYSE, 100.00, 2013-02-01\n ARB, NYSE, 6.25, 2013-02-26\n APL, NYSE, 0.00, 2013-02-19\n AVA, NYSE, -42.22, 2013-02-21\n AXS, NYSE, 64.96, 2013-02-05\n CHT, NYSE, 5.26, 2013-01-31\n MOH, NYSE, 145.45, 2013-02-08\n CVD, NYSE, 2.82, 2013-01-25\n AHT, NYSE, 2.63, 2013-02-28\n GPK, NYSE, 12.50, 2013-02-08\n CNO, NYSE, 8.70, 2013-02-12\n AUQ, NYSE, -28.57, 2013-03-26\n JRN, NYSE, 34.62, 2013-03-08\nGRP.U, NYSE, -14.92, 2013-03-06\n NFP, NYSE, 11.43, 2013-02-15\n CRI, NYSE, 2.30, 2013-02-28\n FMD, NYSE, -20.00, 2013-02-08\n FPO, NYSE, 10.34, 2013-02-22\n TRQ, NYSE, -350.00, 2013-03-26\n WLL, NYSE, 9.21, 2013-02-28\n AEL, NYSE, 14.63, 2013-02-21\n AHL, NYSE, 87.60, 2013-02-08\n AUY, NYSE, -3.70, 2013-02-21\n CMP, NYSE, 0.00, 2013-02-07\n KRO, NYSE, -400.00, 2013-03-13\n TPX, NYSE, 9.09, 2013-01-25\n UTI, NYSE, 75.00, 2013-02-01\n PJC, NYSE, 31.34, 2013-01-31\n TRW, NYSE, 14.81, 2013-02-16\n AIZ, NYSE, 122.58, 2013-02-07\n HTH, NYSE, 62.50, 2013-03-16\n ETP, NYSE, 0.00, 2013-02-21\n SMI, NYSE, 500.00, 2013-02-07\n LSE, NYSE, -6.25, 2013-02-16\n BBD, NYSE, -2.63, 2013-01-29\n NRG, NYSE, 124.14, 2013-02-28\n HOS, NYSE, 29.17, 2013-02-07\n ABR, NYSE, 160.00, 2013-02-16\n FHN, NYSE, 0.00, 2013-01-19\n AGO, NYSE, 32.39, 2013-02-28\n HSP, NYSE, 1.85, 2013-02-14\n HNI, NYSE, -6.98, 2013-02-06\n GHL, NYSE, -32.43, 2013-01-24\n XPO, NYSE, -14.00, 2013-02-28\n CVO, NYSE, 23.08, 2013-02-28\n CHE, NYSE, 16.92, 2013-02-19\n GNW, NYSE, 30.77, 2013-02-06\n CBG, NYSE, 12.24, 2013-02-07\n SFL, NYSE, -26.67, 2013-02-26\n NEU, NYSE, -15.57, 2013-01-29\n GOL, NYSE, -109.09, 2013-03-26\n CAB, NYSE, 4.17, 2013-02-15\n LTM, NYSE, 1.82, 2013-02-22\n VVI, NYSE, 10.53, 2013-02-02\n WCG, NYSE, 0.00, 2013-02-14\n HEP, NYSE, -2.63, 2013-02-22\n DPZ, NYSE, 8.47, 2013-03-01\n BDC, NYSE, 9.86, 2013-02-08\n EGY, NYSE, -171.43, 2013-03-15\n LPL, NYSE, 2.63, 2013-02-22\n ENS, NYSE, 12.82, 2013-02-07\n BMR, NYSE, 5.88, 2013-02-06\n ACC, NYSE, 9.26, 2013-02-13\n KRG, NYSE, -9.09, 2013-02-08\n WLK, NYSE, 13.60, 2013-02-20\n EXR, NYSE, 4.65, 2013-02-22\n CNS, NYSE, 16.67, 2013-01-24\n IOC, NYSE, 264.29, 2013-02-28\n STON, NYSE, -233.33, 2013-03-16\n CPL, NYSE, 38.10, 2013-03-13\n TPGI, NYSE, -114.29, 2013-02-14\n SHO, NYSE, -3.33, 2013-02-20\n CUBE, NYSE, 5.00, 2013-02-22\n NRF, NYSE, 170.37, 2013-02-15\n BBW, NYSE, -68.29, 2013-02-15\n DLR, NYSE, 4.31, 2013-02-16\n NWE, NYSE, 2.63, 2013-02-15\n ORA, NYSE, 200.00, 2013-02-28\n NP, NYSE, 5.26, 2013-02-21\n SMA, NYSE, -21.05, 2013-02-22\n BBG, NYSE, 25.00, 2013-02-22\n BXC, NYSE, -163.16, 2013-02-14\n KNL, NYSE, 32.14, 2013-02-06\n LVS, NYSE, -8.47, 2013-01-31\n HLF, NYSE, 0.96, 2013-02-20\n MIC, NYSE, -20.41, 2013-02-21\n PHH, NYSE, -11.54, 2013-02-07\n CE, NYSE, 6.35, 2013-01-29\n EDR, NYSE, 0.00, 2013-02-20\n WTI, NYSE, 8.33, 2013-02-27\n ARC, NYSE, -100.00, 2013-03-01\n PBH, NYSE, 8.82, 2013-02-08\n HUN, NYSE, 0.00, 2013-02-13\n DLB, NYSE, 4.44, 2013-01-30\n DSX, NYSE, -33.33, 2013-03-15\n LAZ, NYSE, 84.85, 2013-02-08\n TGP, NYSE, 1.82, 2013-02-22\n TLP, NYSE, -43.48, 2013-03-13\n DRH, NYSE, 16.00, 2013-03-01\n HTGC, NYSE, 8.70, 2013-03-01\n KFN, NYSE, 5.26, 2013-02-06\n THS, NYSE, 0.00, 2013-02-22\n NSR, NYSE, -12.50, 2013-02-06\n WAL, NYSE, 0.00, 2013-01-25\n SLW, NYSE, 2.04, 2013-03-22\n MPW, NYSE, 0.00, 2013-02-08\nRDS.B, NYSE, 16.00, 2013-02-01\n GNK, NYSE, -24.71, 2013-02-21\n MFB, NYSE, 4.76, 2013-03-07\nRDS.A, NYSE, 9.95, 2013-02-01\n ITC, NYSE, 0.93, 2013-02-28\n FTK, NYSE, -158.82, 2013-03-14\n PIKE, NYSE, 168.00, 2013-02-06\n ALJ, NYSE, 0.00, 2013-03-07\n DRC, NYSE, -4.55, 2013-03-01\n STN, NYSE, 8.06, 2013-02-22\n SSW, NYSE, -6.90, 2013-03-06\n CF, NYSE, 3.41, 2013-02-20\n HPY, NYSE, 0.00, 2013-02-08\n ACCO, NYSE, 0.00, 2013-02-14\n ROC, NYSE, -6.25, 2013-02-20\n WPZ, NYSE, -28.57, 2013-02-20\n LCC, NYSE, 44.44, 2013-01-24\n GLP, NYSE, 58.82, 2013-03-15\n AMP, NYSE, 15.54, 2013-01-31\n DHT, NYSE, 108.33, 2013-01-30\n FNF, NYSE, 17.86, 2013-02-20\n NM, NYSE, 20.00, 2013-02-20\n CCO, NYSE, 25.00, 2013-02-20\n BWP, NYSE, 0.00, 2013-02-12\n ICE, NYSE, 5.14, 2013-02-07\n BKD, NYSE, -57.14, 2013-02-12\n AAV, NYSE, 350.00, 2013-03-28\n BAS, NYSE, -42.11, 2013-02-20\n CPA, NYSE, -9.87, 2013-02-07\n LYV, NYSE, -147.06, 2013-02-27\n WNR, NYSE, 5.84, 2013-03-01\n CMG, NYSE, 0.00, 2013-02-06\n RGP, NYSE, -180.00, 2013-02-21\n KOP, NYSE, 11.86, 2013-02-15\n UAL, NYSE, -7.41, 2013-01-25\n ETE, NYSE, -90.91, 2013-02-21\n RSO, NYSE, -17.65, 2013-03-05\n XCO, NYSE, 6.25, 2013-02-21\n PAC, NYSE, 41.18, 2013-02-28\n NYX, NYSE, 10.26, 2013-02-06\n TDG, NYSE, 51.65, 2013-02-05\n BMA, NYSE, 18.40, 2013-02-15\n THI, NYSE, -2.82, 2013-02-22\n BTE, NYSE, -40.48, 2013-03-08\n CNH, NYSE, 29.58, 2013-02-01\n GLA, NYSE, 67.44, 2013-02-14\n POR, NYSE, -9.52, 2013-02-23\n HIL, NYSE, -100.00, 2013-03-12\n HVB, NYSE, -20.00, 2013-02-01\n KS, NYSE, 0.00, 2013-02-14\n HK, NYSE, 0.00, 2013-03-01\n DCP, NYSE, 59.62, 2013-02-28\n DK, NYSE, 10.10, 2013-03-08\n CODI, NYSE, 14.81, 2013-03-07\n VG, NYSE, 25.00, 2013-02-14\n MA, NYSE, 1.46, 2013-02-01\n MWA, NYSE, -200.00, 2013-02-06\n KOG, NYSE, 14.29, 2013-03-01\n PWE, NYSE, -500.00, 2013-02-15\n PGTI, NYSE, 100.00, 2013-02-21\n AWH, NYSE, 16.23, 2013-02-14\n NSH, NYSE, -65.71, 2013-02-02\n WYN, NYSE, 5.00, 2013-02-07\n WNS, NYSE, 0.00, 2013-01-17\n AYR, NYSE, 36.84, 2013-02-22\n EVR, NYSE, 55.77, 2013-01-31\n HBI, NYSE, 7.00, 2013-02-06\n WU, NYSE, 20.00, 2013-02-13\n OC, NYSE, -31.25, 2013-02-21\n MR, NYSE, 2.08, 2013-02-26\n DAC, NYSE, -21.43, 2013-02-12\n AWI, NYSE, 3.03, 2013-02-20\n SUSS, NYSE, 444.44, 2013-02-28\n DEI, NYSE, 0.00, 2013-02-13\n OB, NYSE, -200.00, 2013-02-06\n SBH, NYSE, -5.88, 2013-02-08\n EBS, NYSE, -4.35, 2013-03-08\n KBR, NYSE, 122.22, 2013-02-21\n AER, NYSE, 30.95, 2013-02-21\n NOA, NYSE, -11.11, 2013-02-06\n SPR, NYSE, -2.27, 2013-02-13\n ANW, NYSE, 0.00, 2013-02-28\n DCT, NYSE, 10.00, 2013-02-08\n SE, NYSE, -3.03, 2013-02-06\n TOO, NYSE, 16.67, 2013-02-22\n TSL, NYSE, -39.77, 2013-02-27\n TWC, NYSE, 1.95, 2013-02-01\n MVO, NYSE, -5.06, 2013-03-15\n CO, NYSE, 40.00, 2013-02-27\n EXK, NYSE, -45.83, 2013-03-13\n EIG, NYSE, -25.00, 2013-02-28\n HF, NYSE, 21.62, 2013-03-07\n CEL, NYSE, 34.78, 2013-03-05\n FIG, NYSE, 53.85, 2013-02-28\n NGLS, NYSE, 0.00, 2013-02-15\n TCAP, NYSE, 3.64, 2013-03-07\n GFA, NYSE, -483.33, 2013-03-12\n BR, NYSE, -5.56, 2013-02-08\n SCR, NYSE, 85.71, 2013-03-08\n CNK, NYSE, -12.82, 2013-02-21\n DAL, NYSE, 0.00, 2013-01-23\n ORN, NYSE, 250.00, 2013-03-01\n ACM, NYSE, 9.09, 2013-02-06\n JMP, NYSE, 62.50, 2013-02-14\n SLH, NYSE, 1.69, 2013-02-08\n CLR, NYSE, 16.85, 2013-02-28\n BGS, NYSE, -17.95, 2013-02-15\n STAR, NYSE, 12.50, 2013-02-27\n YGE, NYSE, -74.07, 2013-03-05\n DFS, NYSE, -9.40, 2013-03-06\n TEL, NYSE, 1.56, 2013-01-24\n BX, NYSE, 25.53, 2013-02-01\n SEP, NYSE, 8.11, 2013-02-06\n BZ, NYSE, -30.00, 2013-02-27\n PPO, NYSE, -28.26, 2013-02-21\n PRO, NYSE, 25.00, 2013-02-13\n WBC, NYSE, 13.68, 2013-02-16\n DHX, NYSE, 7.14, 2013-01-31\n PMC, NYSE, 13.79, 2013-02-08\n HGG, NYSE, 0.00, 2013-02-01\n OWW, NYSE, -14.29, 2013-02-15\n VR, NYSE, 35.58, 2013-02-01\n CXO, NYSE, -5.88, 2013-02-21\n G, NYSE, 4.76, 2013-02-08\n EJ, NYSE, 160.00, 2013-03-13\n WX, NYSE, 32.00, 2013-03-08\n CMLP, NYSE, -50.00, 2013-02-06\n VMW, NYSE, -5.56, 2013-01-29\n CZZ, NYSE, 63.64, 2013-02-08\n CGA, NYSE, -3.23, 2013-02-09\n TDC, NYSE, 5.71, 2013-02-08\n FLY, NYSE, 137.65, 2013-03-08\n DUF, NYSE, 6.25, 2013-02-26\n MAIN, NYSE, 12.00, 2013-03-08\n REN, NYSE, -50.00, 2013-03-08\n TGH, NYSE, 9.57, 2013-02-13\n DFT, NYSE, -5.00, 2013-02-07\n RF, NYSE, 10.00, 2013-01-23\n PZN, NYSE, -22.22, 2013-02-13\n LL, NYSE, 19.05, 2013-02-21\n NMM, NYSE, 0.00, 2013-01-25\n OZM, NYSE, 5.48, 2013-02-08\n ES, NYSE, -5.08, 2013-02-20\n MSCI, NYSE, -1.89, 2013-02-08\n ARR, NYSE, -18.52, 2013-02-23\n KW, NYSE, 275.00, 2013-03-13\n GTS, NYSE, -10.17, 2013-02-07\n FOR, NYSE, 222.22, 2013-02-14\n LRN, NYSE, 4.35, 2013-02-06\n TNK, NYSE, -125.00, 2013-02-22\n N, NYSE, 21.43, 2013-02-01\n DAN, NYSE, 5.56, 2013-02-22\n BIP, NYSE, 12.07, 2013-02-09\n CPN, NYSE, -500.00, 2013-02-14\n SOL, NYSE, 2.70, 2013-03-15\n PM, NYSE, 1.64, 2013-02-08\n HI, NYSE, 7.89, 2013-02-05\n V, NYSE, 2.25, 2013-02-07\n IPI, NYSE, 0.00, 2013-02-14\n AWK, NYSE, -14.29, 2013-02-27\n HTS, NYSE, 37.84, 2013-02-13\n DPS, NYSE, -4.71, 2013-02-14\n CFX, NYSE, 7.69, 2013-02-07\n WES, NYSE, -27.91, 2013-02-28\n SB, NYSE, -10.00, 2013-02-21\n LO, NYSE, 3.95, 2013-02-14\n LPS, NYSE, 10.45, 2013-02-08\n FF, NYSE, -31.82, 2013-03-19\n NNA, NYSE, 150.00, 2013-02-13\n EPB, NYSE, 14.55, 2013-01-17\n JBT, NYSE, 3.23, 2013-03-07\n DL, NYSE, 33.33, 2013-02-27\n RAX, NYSE, -4.55, 2013-02-13\n HCI, NYSE, 67.61, 2013-03-06\n EC, NYSE, -20.47, 2013-02-16\n CLW, NYSE, 10.53, 2013-02-21\n MJN, NYSE, 5.88, 2013-02-01\n EPC, NYSE, 1.85, 2013-02-01\n BPI, NYSE, -3.33, 2013-03-13\n RST, NYSE, 55.56, 2013-03-01\n DGI, NYSE, 92.31, 2013-02-27\n SWI, NYSE, 10.34, 2013-02-05\n CYS, NYSE, -46.15, 2013-02-07\n IVR, NYSE, 20.31, 2013-02-06\n BUD, NYSE, -5.08, 2013-02-28\n PMT, NYSE, -2.35, 2013-02-08\n STWD, NYSE, 15.38, 2013-02-28\n CFN, NYSE, -16.98, 2013-02-09\n SPB, NYSE, 71.43, 2013-02-07\n ARI, NYSE, -10.34, 2013-02-28\n CLNY, NYSE, -13.89, 2013-03-07\n ART, NYSE, 300.00, 2013-02-15\n SEM, NYSE, 12.00, 2013-02-22\n BSBR, NYSE, 578.57, 2013-03-28\n DOLE, NYSE, -6100.00, 2013-03-13\n VSI, NYSE, 0.00, 2013-02-27\n TWO, NYSE, -15.15, 2013-02-07\n CVE, NYSE, -14.29, 2013-02-15\n H, NYSE, 81.82, 2013-02-14\n LEA, NYSE, 7.25, 2013-02-02\n CLD, NYSE, 8.00, 2013-02-14\n AOL, NYSE, 7.50, 2013-02-09\n CHSP, NYSE, 5.13, 2013-02-22\n PEB, NYSE, 0.00, 2013-02-22\n CIT, NYSE, 60.94, 2013-01-30\n KAR, NYSE, -4.55, 2013-02-21\n CIE, NYSE, -66.67, 2013-02-27\n TMH, NYSE, 8.33, 2013-02-06\n KRA, NYSE, -300.00, 2013-02-28\n SYA, NYSE, -29.41, 2013-02-05\n TRNO, NYSE, -162.50, 2013-02-16\n PDM, NYSE, -2.70, 2013-02-08\n GNRC, NYSE, 26.09, 2013-02-15\n ACW, NYSE, -2.17, 2013-03-07\n BALT, NYSE, -11.76, 2013-02-21\n ST, NYSE, 2.17, 2013-01-31\n SEMG, NYSE, 55.56, 2013-03-01\n CALX, NYSE, 20.00, 2013-02-06\n MXL, NYSE, -57.14, 2013-02-06\n STNG, NYSE, -60.00, 2013-02-26\n PRI, NYSE, -1.43, 2013-02-08\n SDRL, NYSE, -93.65, 2013-03-01\n CLDT, NYSE, 0.00, 2013-02-20\n EXL, NYSE, 0.00, 2013-02-28\n LYB, NYSE, -0.88, 2013-02-02\n PNG, NYSE, 7.14, 2013-02-07\n PLOW, NYSE, -25.00, 2013-03-12\n SIX, NYSE, 198.00, 2013-02-21\n NKA, NYSE, 1066.67, 2013-02-01\n RRTS, NYSE, 0.00, 2013-02-07\n JKS, NYSE, -332.48, 2013-04-11\n CODE, NYSE, -13.64, 2013-01-30\n FAF, NYSE, 44.64, 2013-02-22\n QEP, NYSE, 3.13, 2013-02-20\n OAS, NYSE, 6.52, 2013-02-26\n VPG, NYSE, 15.38, 2013-02-13\n HPP, NYSE, 9.52, 2013-03-07\n FN, NYSE, 9.09, 2013-02-05\n ECT, NYSE, 65.85, 2013-03-16\n QUAD, NYSE, -6.67, 2013-03-05\n KKR, NYSE, 54.84, 2013-02-08\n RLD, NYSE, 20.00, 2013-02-07\n AMRC, NYSE, 44.44, 2013-03-19\n GDOT, NYSE, 50.00, 2013-02-01\n AT, NYSE, -160.00, 2013-03-01\n ENV, NYSE, 0.00, 2013-02-15\n IL, NYSE, 200.00, 2013-02-22\n WSR, NYSE, -12.00, 2013-03-13\n SFUN, NYSE, 35.71, 2013-02-09\n COR, NYSE, 5.00, 2013-02-23\n VC, NYSE, 20.62, 2013-03-01\n CCSC, NYSE, -20.00, 2013-03-07\n CCG, NYSE, 0.00, 2013-02-27\n EFC, NYSE, -72.73, 2013-02-14\n TOWR, NYSE, 183.33, 2013-02-16\n CHMT, NYSE, -53.13, 2013-02-26\n HBM, NYSE, 200.00, 2013-02-21\n EXAM, NYSE, 55.56, 2013-02-28\n XUE, NYSE, 7.69, 2013-02-28\n CMRE, NYSE, 6.67, 2013-01-24\n NOAH, NYSE, 20.00, 2013-02-26\n IPHI, NYSE, -40.00, 2013-02-05\n BITA, NYSE, 33.33, 2013-03-08\n BAH, NYSE, 11.11, 2013-01-31\n GM, NYSE, -2.04, 2013-02-15\n TROX, NYSE, -60.00, 2013-02-21\n DANG, NYSE, 20.00, 2013-03-08\n YOKU, NYSE, 9.09, 2013-03-01\n FRC, NYSE, -16.44, 2013-01-17\n RFP, NYSE, 52.38, 2013-02-13\n ISS, NYSE, 15.38, 2013-03-09\n WD, NYSE, -14.29, 2013-03-07\n FLT, NYSE, 10.00, 2013-02-08\n GCAP, NYSE, -325.00, 2013-03-13\n FRF, NYSE, -25.93, 2013-03-29\n SWFT, NYSE, 46.15, 2013-01-24\n AG, NYSE, -10.34, 2013-02-27\n QRE, NYSE, -174.07, 2013-03-07\n AAT, NYSE, 11.76, 2013-02-20\n MCC, NYSE, 5.41, 2013-02-07\n NLSN, NYSE, 3.51, 2013-02-12\n AGRO, NYSE, -71.43, 2013-03-22\n BKU, NYSE, 27.08, 2013-01-30\n INXN, NYSE, -38.89, 2013-02-28\n NPTN, NYSE, 16.67, 2013-02-22\n INN, NYSE, 25.00, 2013-02-27\n KMI, NYSE, -5.88, 2013-01-17\n HCA, NYSE, 9.64, 2013-02-05\n MX, NYSE, 135.21, 2013-01-31\n HII, NYSE, 8.89, 2013-02-28\n QIHU, NYSE, 175.00, 2013-03-06\n APO, NYSE, 119.48, 2013-02-09\n GNC, NYSE, 8.70, 2013-02-15\n SDT, NYSE, 11.48, 2013-03-16\n UAN, NYSE, 16.67, 2013-02-28\n ARCO, NYSE, 5.00, 2013-03-09\n ELLI, NYSE, 36.36, 2013-02-15\n TMS, NYSE, -23.81, 2013-02-15\n SQNS, NYSE, -16.00, 2013-02-08\n STAG, NYSE, 17.24, 2013-02-21\n AL, NYSE, 8.33, 2013-03-01\n TLLP, NYSE, 10.42, 2013-02-12\n RENN, NYSE, 14.29, 2013-03-12\n NQ, NYSE, 800.00, 2013-03-07\n THR, NYSE, -14.29, 2013-02-08\n KOS, NYSE, 125.00, 2013-02-26\n RLJ, NYSE, 4.35, 2013-02-28\n NGL, NYSE, -7.41, 2013-02-16\n FENG, NYSE, 100.00, 2013-03-07\n LNKD, NYSE, 900.00, 2013-02-08\n NMFC, NYSE, 5.88, 2013-03-07\n ACTV, NYSE, 5.26, 2013-02-15\n TAOM, NYSE, 700.00, 2013-03-15\n RATE, NYSE, -60.00, 2013-02-13\n VHS, NYSE, -22.22, 2013-01-31\n MPC, NYSE, 8.13, 2013-01-31\n MITT, NYSE, -1.16, 2013-03-06\n OILT, NYSE, 0.00, 2013-03-07\n SXC, NYSE, 14.71, 2013-02-06\n AMTG, NYSE, -8.57, 2013-03-07\n AMID, NYSE, -2500.00, 2013-04-17\n WAIR, NYSE, -7.41, 2013-01-30\n PER, NYSE, -7.58, 2013-03-02\n PPP, NYSE, -44.44, 2013-02-22\n FNV, NYSE, -8.33, 2013-03-20\n FSM, NYSE, 16.67, 2013-03-21\n FBHS, NYSE, 4.55, 2013-02-01\n XLS, NYSE, 4.44, 2013-03-02\n XYL, NYSE, 2.17, 2013-02-08\n NDRO, NYSE, 4.76, 2013-03-19\n RNF, NYSE, -33.33, 2013-03-20\n VAC, NYSE, 25.53, 2013-02-22\n CHKR, NYSE, -7.25, 2013-03-16\n PACD, NYSE, 14.29, 2013-02-28\n INVN, NYSE, 0.00, 2013-01-24\n DLPH, NYSE, 3.45, 2013-02-06\n MN, NYSE, 0.00, 2013-02-14\n RRMS, NYSE, -25.00, 2013-03-01\n WPX, NYSE, -400.00, 2013-03-01\n LPI, NYSE, 0.00, 2013-03-13\n SN, NYSE, -80.00, 2013-03-07\n KORS, NYSE, 60.00, 2013-02-13\n BCEI, NYSE, -7.89, 2013-03-15\n BOXC, NYSE, 4.78, 2013-01-29\n PVG, NYSE, -25.00, 2013-03-06\n POST, NYSE, 30.43, 2013-02-08\n SLCA, NYSE, 32.26, 2013-02-27\n MTDR, NYSE, -116.67, 2013-03-14\n GWAY, NYSE, -200.00, 2013-02-13\n EPAM, NYSE, -10.81, 2013-02-28\n RNDY, NYSE, 5.56, 2013-03-01\n CPAC, NYSE, -13.33, 2013-02-21\n PRLB, NYSE, 7.69, 2013-02-14\n YELP, NYSE, -50.00, 2013-02-07\n NSM, NYSE, 7.58, 2013-03-08\n ALSN, NYSE, 257.14, 2013-02-20\n DWRE, NYSE, 350.00, 2013-02-15\n VNTV, NYSE, 16.13, 2013-02-21\n ET, NYSE, 34.78, 2013-02-22\n VIPS, NYSE, 1100.00, 2013-02-22\n VCRA, NYSE, -33.33, 2013-02-28\n RM, NYSE, -1.89, 2013-02-28\n BNNY, NYSE, 0.00, 2013-02-12\n MM, NYSE, 200.00, 2013-02-20\n RXN, NYSE, -15.00, 2013-02-12\n GLOG, NYSE, -20.00, 2013-02-28\n PBA, NYSE, 44.44, 2013-03-02\n RPAI, NYSE, 15.79, 2013-02-20\n OAK, NYSE, 63.33, 2013-02-15\n FET, NYSE, -3.45, 2013-02-15\n MRC, NYSE, 17.02, 2013-02-22\n PSX, NYSE, 21.18, 2013-01-31\n TUMI, NYSE, 0.00, 2013-03-21\n ACRE, NYSE, -38.10, 2013-04-02\n EVER, NYSE, 17.24, 2013-01-31\n PDH, NYSE, -13.79, 2013-02-07\n WMC, NYSE, 3.23, 2013-04-03\n WAGE, NYSE, 0.00, 2013-02-21\n HTA, NYSE, 0.00, 2013-02-21\n ALEX, NYSE, 42.86, 2013-02-20\n BKW, NYSE, 53.33, 2013-02-16\n EQM, NYSE, 51.22, 2013-01-25\n NOW, NYSE, 38.46, 2013-01-31\n EGL, NYSE, 18.46, 2013-03-13\n NGVC, NYSE, 25.00, 2013-02-01\n NTI, NYSE, -25.00, 2013-03-14\n AMRE, NYSE, 4.35, 2013-02-20\n GMED, NYSE, 15.79, 2013-02-28\n MANU, NYSE, -46.43, 2013-02-15\n HCLP, NYSE, -28.57, 2013-02-01\n ADT, NYSE, 4.76, 2013-01-31\n TRLA, NYSE, -20.00, 2013-02-13\n SRC, NYSE, 8.82, 2013-02-28\n NBHC, NYSE, -14.29, 2013-01-29\n BSMX, NYSE, -4.17, 2013-02-19\n HY, NYSE, 14.53, 2013-02-20\n SMLP, NYSE, 40.00, 2013-03-14\n DYN, NYSE, -1714.29, 2013-03-15\n LXFR, NYSE, 43.75, 2013-03-12\n LOCK, NYSE, 16.67, 2013-02-21\n JMI, NYSE, 97.78, 2013-03-22\n BERY, NYSE, -40.00, 2013-02-01\n FLTX, NYSE, 0.00, 2013-02-21\n ANFI, NYSE, 30.77, 2013-02-26\n SSTK, NYSE, -100.00, 2013-02-22\n SDLP, NYSE, 90.91, 2013-03-01\n MPLX, NYSE, -25.00, 2013-01-31\n WWAV, NYSE, 5.88, 2013-02-14\n SXE, NYSE, -4121.43, 2013-03-29\n DKL, NYSE, -5.56, 2013-03-06\n RKUS, NYSE, -20.00, 2013-02-13\n WGP, NYSE, 57.14, 2013-02-28\n PBF, NYSE, -92.31, 2013-03-01\n SBY, NYSE, 0.00, 2013-03-01\n RIOM, NYSE, 77.78, 2013-03-29\n BFAM, NYSE, -1186.36, 2013-03-27\n ZTS, NYSE, -79.41, 2013-03-29\n DDC, NYSE, -39.13, 2013-04-04\n ABM, NYSE, 18.18, 2013-03-05\n ANN, NYSE, 0.00, 2013-03-09\n BBY, NYSE, 5.81, 2013-03-02\n BF.B, NYSE, 4.29, 2013-03-07\n BKE, NYSE, 2.40, 2013-03-15\n BNS, NYSE, -3.17, 2013-03-06\n BRC, NYSE, -22.45, 2013-02-22\n CATO, NYSE, -3.57, 2013-03-22\n COO, NYSE, 2.50, 2013-03-08\n CPB, NYSE, 6.06, 2013-02-16\n CFI, NYSE, 10.34, 2013-02-28\n DCI, NYSE, -10.53, 2013-02-26\n DDS, NYSE, -1.03, 2013-02-26\n DE, NYSE, 17.02, 2013-02-14\n DY, NYSE, 50.00, 2013-02-27\n EV, NYSE, -3.85, 2013-02-21\n ENZ, NYSE, -133.33, 2013-03-13\n ESL, NYSE, 13.11, 2013-03-01\nFCE.A, NYSE, 9.09, 2013-03-28\n M, NYSE, 3.54, 2013-02-27\n GCO, NYSE, 1.41, 2013-03-09\n GPS, NYSE, 2.82, 2013-03-01\n HD, NYSE, 4.69, 2013-02-27\n HEI, NYSE, -12.50, 2013-02-21\n HNZ, NYSE, 10.00, 2013-02-28\n HOV, NYSE, -66.67, 2013-03-07\n HRB, NYSE, -633.33, 2013-03-08\n HRL, NYSE, -2.04, 2013-02-22\n HPQ, NYSE, 15.49, 2013-02-22\n JCP, NYSE, -926.32, 2013-02-28\n KR, NYSE, 25.71, 2013-03-08\n KSS, NYSE, 1.84, 2013-03-01\n LB, NYSE, 1.15, 2013-02-28\n LOW, NYSE, 13.04, 2013-02-26\n LZB, NYSE, 16.67, 2013-02-20\n MDT, NYSE, 2.20, 2013-02-20\n MEI, NYSE, 350.00, 2013-03-01\n MPR, NYSE, 0.00, 2013-03-22\n NAV, NYSE, 14.11, 2013-03-08\n JWN, NYSE, 4.48, 2013-02-22\n ODC, NYSE, -35.42, 2013-03-12\n OXM, NYSE, -5.80, 2013-04-03\n PBY, NYSE, -225.00, 2013-04-16\n PLL, NYSE, 8.96, 2013-02-28\n PNY, NYSE, 1.72, 2013-03-07\n PVH, NYSE, 6.67, 2013-03-28\n THO, NYSE, 0.00, 2013-03-08\n TIF, NYSE, 2.19, 2013-03-23\n TJX, NYSE, 1.23, 2013-02-28\n TOL, NYSE, -81.82, 2013-02-21\n TTC, NYSE, 23.26, 2013-02-22\n VAL, NYSE, -9.09, 2013-02-13\n JW.A, NYSE, 13.41, 2013-03-08\n WMT, NYSE, 6.37, 2013-02-22\n WSM, NYSE, 4.69, 2013-03-20\n FL, NYSE, -11.11, 2013-03-09\n CHS, NYSE, 0.00, 2013-03-01\n REX, NYSE, -800.00, 2013-03-29\n BKS, NYSE, -136.00, 2013-03-01\n CAL, NYSE, 75.00, 2013-03-16\n SIG, NYSE, 1.44, 2013-03-29\n ZLC, NYSE, -1.92, 2013-02-22\n AEO, NYSE, 0.00, 2013-03-07\n FGP, NYSE, -10.00, 2013-03-08\n BMO, NYSE, 1.37, 2013-02-27\n RY, NYSE, 0.75, 2013-03-01\n GEF, NYSE, -13.21, 2013-02-28\n MOV, NYSE, 70.83, 2013-03-22\n SKS, NYSE, 13.33, 2013-02-27\n TD, NYSE, 1.55, 2013-03-01\n ANF, NYSE, 14.51, 2013-02-23\n CIEN, NYSE, 116.00, 2013-03-08\n KMG, NYSE, -17.65, 2013-03-09\n IRET, NYSE, -5.88, 2013-03-13\n CM, NYSE, 0.00, 2013-03-01\nHEI.A, NYSE, -18.60, 2013-02-21\n UBA, NYSE, 13.04, 2013-03-07\n KFY, NYSE, 6.90, 2013-03-07\n TGT, NYSE, 12.24, 2013-02-28\n KKD, NYSE, 0.00, 2013-03-15\n NDZ, NYSE, 0.00, 2013-03-06\n MVC, NYSE, -20.00, 2013-03-08\n CBK, NYSE, 52.17, 2013-03-14\n SJM, NYSE, 7.30, 2013-02-16\n BIG, NYSE, 5.03, 2013-03-07\n IDT, NYSE, -7.14, 2013-03-08\n JOY, NYSE, 14.91, 2013-02-28\n SSI, NYSE, -5.93, 2013-03-13\n GME, NYSE, 3.35, 2013-03-29\n DKS, NYSE, -3.74, 2013-03-12\n A, NYSE, -5.97, 2013-02-15\n MTN, NYSE, -3.51, 2013-03-07\n GES, NYSE, 10.47, 2013-03-21\n CRM, NYSE, 66.67, 2013-03-01\n NWY, NYSE, 25.00, 2013-03-22\n PAY, NYSE, 8.11, 2013-03-06\n DSW, NYSE, -4.17, 2013-03-20\n NX, NYSE, -183.33, 2013-03-08\n AGX, NYSE, 15.00, 2013-04-11\n CMD, NYSE, -5.26, 2013-03-08\n DG, NYSE, 7.78, 2013-03-26\n EXPR, NYSE, 1.35, 2013-03-14\n P, NYSE, 0.00, 2013-03-07\n GWRE, NYSE, 181.82, 2013-02-27\n BLOX, NYSE, -20.00, 2013-02-22\n TLYS, NYSE, 6.67, 2013-03-21\n PANW, NYSE, -250.00, 2013-03-01\n WDAY, NYSE, 24.00, 2013-03-08\n RH, NYSE, 4.92, 2013-04-19\n AIR, NYSE, 4.55, 2013-03-20\n ATU, NYSE, -5.41, 2013-03-21\n AZO, NYSE, 0.84, 2013-02-27\n AZZ, NYSE, 2.04, 2013-04-09\n CAG, NYSE, -3.51, 2013-04-04\n CLC, NYSE, 2.17, 2013-03-21\n CMC, NYSE, -80.00, 2013-03-29\n KMX, NYSE, 0.00, 2013-04-11\n FC, NYSE, -27.27, 2013-04-05\n FDO, NYSE, -0.82, 2013-04-11\n FDX, NYSE, -10.87, 2013-03-21\n FUL, NYSE, -3.92, 2013-03-28\n GIS, NYSE, 12.28, 2013-03-21\n KBH, NYSE, 30.43, 2013-03-22\n LEN, NYSE, 100.00, 2013-03-21\n LNN, NYSE, 16.28, 2013-03-28\n LUB, NYSE, -100.00, 2013-03-21\n MKC, NYSE, 1.79, 2013-04-03\n RT, NYSE, 0.00, 2013-04-11\n MSM, NYSE, 0.00, 2013-04-11\n NKE, NYSE, 8.96, 2013-03-22\n ORCL, NYSE, -1.56, 2013-03-21\n PIR, NYSE, 0.00, 2013-04-12\n PKE, NYSE, -21.43, 2013-05-10\n RPM, NYSE, 16.67, 2013-04-05\n SVU, NYSE, -200.00, 2013-04-25\n TXI, NYSE, 25.00, 2013-03-28\n UNF, NYSE, 18.75, 2013-03-28\n WGO, NYSE, 37.50, 2013-03-29\n WOR, NYSE, 6.12, 2013-03-22\n JBL, NYSE, -2.17, 2013-03-21\n GBX, NYSE, 21.62, 2013-04-05\n DRI, NYSE, 0.99, 2013-03-23\n FDS, NYSE, -21.24, 2013-03-20\n SCS, NYSE, 0.00, 2013-03-28\n SJR, NYSE, 5.56, 2013-04-13\n RHT, NYSE, 19.05, 2013-03-28\n OMN, NYSE, -75.00, 2013-04-04\n MON, NYSE, 7.06, 2013-04-04\n GPN, NYSE, -1.14, 2013-04-03\n AYI, NYSE, 0.00, 2013-04-04\n CCL, NYSE, 100.00, 2013-03-16\n CUK, NYSE, 33.33, 2013-03-16\n STZ, NYSE, 4.44, 2013-04-11\n ACN, NYSE, 3.09, 2013-03-29\n SNX, NYSE, 1.15, 2013-03-28\n TAL, NYSE, 50.00, 2013-04-24\n IHS, NYSE, 11.90, 2013-03-22\n EDU, NYSE, 63.64, 2013-04-25\n KED, NYSE, -99.22, 2013-05-02\n CORR, NYSE, -9.09, 2013-05-11\n DFS, NYSE, 18.75, 2013-04-24\n ZEP, NYSE, 54.55, 2013-04-10\n MG, NYSE, -58.82, 2013-04-09\n MOS, NYSE, 5.62, 2013-03-28\n ABT, NYSE, 0.00, 2013-04-18\n ABX, NYSE, 6.98, 2013-04-25\n AB, NYSE, 8.57, 2013-05-02\n ACO, NYSE, -10.64, 2013-04-27\n ADM, NYSE, -5.88, 2013-05-01\n AEM, NYSE, -35.29, 2013-04-26\n AEP, NYSE, 0.00, 2013-04-27\n AES, NYSE, -14.29, 2013-05-10\n AET, NYSE, 8.70, 2013-05-01\n AFL, NYSE, 4.32, 2013-04-25\n AGCO, NYSE, 35.23, 2013-05-01\n HES, NYSE, 24.20, 2013-04-25\n AIG, NYSE, 52.27, 2013-05-03\n AIN, NYSE, 0.00, 2013-05-02\n AJG, NYSE, 33.33, 2013-05-01\n ALU, NYSE, -81.82, 2013-04-27\n MATX, NYSE, 31.25, 2013-05-07\n ALK, NYSE, 15.09, 2013-04-26\n ALX, NYSE, -2.56, 2013-05-07\n BEAM, NYSE, 18.52, 2013-05-03\n AME, NYSE, 3.92, 2013-04-26\n TWX, NYSE, 9.33, 2013-05-02\n AVD, NYSE, 47.50, 2013-05-03\n AMN, NYSE, 33.33, 2013-05-03\n AN, NYSE, 7.94, 2013-04-19\n AON, NYSE, 0.00, 2013-04-27\n APA, NYSE, -9.01, 2013-05-10\n APC, NYSE, 17.39, 2013-05-07\n APD, NYSE, 0.00, 2013-04-24\n APH, NYSE, 1.16, 2013-04-19\n ARG, NYSE, 0.88, 2013-05-03\n AAN, NYSE, -5.63, 2013-04-26\n ARW, NYSE, 3.49, 2013-05-02\n ASGN, NYSE, 94.44, 2013-04-25\n ASH, NYSE, 14.10, 2013-04-25\n ASR, NYSE, -13.25, 2013-04-23\n GAS, NYSE, -2.96, 2013-05-01\n ATO, NYSE, 1.63, 2013-05-02\n ATW, NYSE, 2.40, 2013-05-02\n AU, NYSE, -26.67, 2013-05-14\n AVP, NYSE, 85.71, 2013-05-01\n AVT, NYSE, 3.45, 2013-04-26\n AVY, NYSE, 3.51, 2013-04-25\n AXP, NYSE, 3.60, 2013-04-18\n B, NYSE, -11.11, 2013-04-27\n BA, NYSE, 17.69, 2013-04-25\n BAC, NYSE, -13.04, 2013-04-17\n BAX, NYSE, 0.96, 2013-04-19\n BC, NYSE, 22.58, 2013-04-26\n OMX, NYSE, -52.17, 2013-05-08\n BCE, NYSE, 10.00, 2013-05-10\n BCR, NYSE, 0.00, 2013-04-24\n BDX, NYSE, 6.67, 2013-05-03\n BEN, NYSE, 8.47, 2013-05-01\n BGG, NYSE, -17.59, 2013-04-20\n BHE, NYSE, 10.00, 2013-04-26\n BHI, NYSE, 4.84, 2013-04-20\n BID, NYSE, -175.00, 2013-05-10\n BIO, NYSE, -38.18, 2013-05-08\n BK, NYSE, 9.62, 2013-04-18\n BKH, NYSE, 19.18, 2013-05-03\n WRB, NYSE, 0.00, 2013-04-24\n BLC, NYSE, 6.67, 2013-04-26\n BLL, NYSE, -9.38, 2013-04-26\n BLX, NYSE, -21.82, 2013-04-18\n BMI, NYSE, -58.33, 2013-04-17\n BMS, NYSE, -1.85, 2013-04-26\n BMY, NYSE, 0.00, 2013-04-26\n BOH, NYSE, -6.90, 2013-04-23\n BXS, NYSE, 4.76, 2013-04-23\n BPL, NYSE, 19.44, 2013-05-04\nBRK.A, NYSE, 197.70, 2013-05-04\n BRO, NYSE, 5.13, 2013-04-16\n BSX, NYSE, 0.00, 2013-04-26\n MTRN, NYSE, -2.94, 2013-04-26\n CAI, NYSE, -1.32, 2013-04-25\n CAT, NYSE, -2.24, 2013-04-23\n CB, NYSE, 12.44, 2013-04-23\n CBI, NYSE, 15.49, 2013-05-03\n CBM, NYSE, 85.00, 2013-05-04\n CBU, NYSE, -1.96, 2013-04-24\n CBT, NYSE, -7.25, 2013-05-01\n CCC, NYSE, 20.00, 2013-05-07\n CCE, NYSE, 2.63, 2013-04-26\n C, NYSE, 9.32, 2013-04-16\n CCK, NYSE, 4.17, 2013-04-18\n CDE, NYSE, -74.07, 2013-05-10\n CDI, NYSE, -40.91, 2013-05-03\n CAH, NYSE, 26.32, 2013-05-03\n CFR, NYSE, -4.21, 2013-04-25\n CHD, NYSE, 5.56, 2013-05-03\n CPK, NYSE, 14.93, 2013-05-03\n CI, NYSE, 20.28, 2013-05-03\n CIA, NYSE, 0.00, 2013-05-03\n CKH, NYSE, -156.12, 2013-04-30\n CL, NYSE, 0.00, 2013-04-26\n CLF, NYSE, 87.50, 2013-04-25\n CLH, NYSE, 25.81, 2013-05-02\n CLX, NYSE, -5.66, 2013-05-02\n CMA, NYSE, 4.48, 2013-04-17\n CMO, NYSE, 3.33, 2013-04-25\n CRK, NYSE, -11.36, 2013-04-30\n CMS, NYSE, 15.22, 2013-04-26\n CNA, NYSE, 21.13, 2013-05-01\n CNW, NYSE, -29.63, 2013-05-02\n CHG, NYSE, 19.00, 2013-05-10\n CNL, NYSE, -8.33, 2013-04-30\n COG, NYSE, -20.00, 2013-04-25\n COT, NYSE, -100.00, 2013-05-02\n CP, NYSE, 2.54, 2013-04-25\n CPF, NYSE, 105.00, 2013-04-27\n CQB, NYSE, 28.57, 2013-05-08\n CR, NYSE, -0.95, 2013-04-23\nCRD.B, NYSE, -29.17, 2013-05-09\n CRS, NYSE, -9.21, 2013-04-26\n CSC, NYSE, 32.29, 2013-05-16\n CSL, NYSE, 0.00, 2013-04-25\n CTB, NYSE, 31.82, 2013-05-10\n CTL, NYSE, 10.14, 2013-05-09\n CTS, NYSE, 16.67, 2013-04-24\n CUB, NYSE, 52.24, 2013-05-03\n CMI, NYSE, -22.58, 2013-05-01\n CUZ, NYSE, -8.33, 2013-05-09\n CVC, NYSE, -185.71, 2013-05-10\n CVH, NYSE, 26.58, 2013-05-02\n CW, NYSE, 28.21, 2013-05-02\n CWT, NYSE, -200.00, 2013-05-02\n CX, NYSE, -140.00, 2013-04-27\n CYN, NYSE, -2.17, 2013-04-19\n D, NYSE, -7.78, 2013-04-26\n DBD, NYSE, -125.00, 2013-05-01\n DCO, NYSE, -18.60, 2013-05-07\n DD, NYSE, 1.30, 2013-04-24\n CVA, NYSE, -61.54, 2013-04-18\n DHR, NYSE, -1.32, 2013-04-19\n DIS, NYSE, 2.60, 2013-05-08\n DLX, NYSE, 3.41, 2013-04-26\n DNB, NYSE, 2.26, 2013-05-03\n RRD, NYSE, 12.12, 2013-04-26\n DOV, NYSE, 1.85, 2013-04-18\n DOW, NYSE, 15.00, 2013-04-26\n DRE, NYSE, 0.00, 2013-04-25\n DHI, NYSE, 60.00, 2013-04-27\n UFS, NYSE, -35.37, 2013-04-26\n DTE, NYSE, 30.10, 2013-04-27\n DUK, NYSE, -1.92, 2013-05-04\n DVN, NYSE, 17.86, 2013-05-02\n DV, NYSE, 8.43, 2013-04-24\n EAT, NYSE, 4.35, 2013-04-24\n ECL, NYSE, 3.45, 2013-05-01\n ED, NYSE, 4.85, 2013-05-03\n EDE, NYSE, 11.11, 2013-04-26\n EFX, NYSE, 0.00, 2013-04-25\n EGN, NYSE, -7.32, 2013-04-30\n EGP, NYSE, -1.30, 2013-04-19\n ELP, NYSE, 0.00, 2013-05-17\n ELY, NYSE, 65.00, 2013-04-26\n EMC, NYSE, 3.23, 2013-04-25\n EMR, NYSE, -1.28, 2013-05-08\n EOG, NYSE, 59.29, 2013-05-07\n EQT, NYSE, 26.92, 2013-04-26\n ESE, NYSE, -17.65, 2013-05-08\n ESV, NYSE, 5.43, 2013-04-30\n ETN, NYSE, 6.33, 2013-04-30\n ETR, NYSE, 0.00, 2013-04-26\n EXAR, NYSE, 16.67, 2013-05-01\n F, NYSE, 7.89, 2013-04-25\n CLGX, NYSE, 8.11, 2013-04-25\n FNB, NYSE, -4.76, 2013-04-24\n FCF, NYSE, 0.00, 2013-04-24\n FBP, NYSE, -122.22, 2013-05-04\n FICO, NYSE, -9.38, 2013-04-25\n FLO, NYSE, 6.98, 2013-05-17\n FMC, NYSE, 1.85, 2013-05-01\n FOE, NYSE, 66.67, 2013-04-25\n S, NYSE, 38.24, 2013-04-25\n NEE, NYSE, 10.89, 2013-05-01\n FRT, NYSE, 0.88, 2013-05-02\n FRX, NYSE, 47.06, 2013-04-24\n FSS, NYSE, 20.00, 2013-05-07\n FUN, NYSE, 24.32, 2013-05-09\n FUR, NYSE, 77.78, 2013-05-03\n GBL, NYSE, 17.86, 2013-05-08\n GVA, NYSE, -103.85, 2013-05-10\n BGC, NYSE, -319.23, 2013-05-01\n GD, NYSE, 8.00, 2013-04-25\n GE, NYSE, 11.43, 2013-04-20\n RHP, NYSE, 26.47, 2013-05-08\n AXLL, NYSE, -38.02, 2013-05-08\n GGG, NYSE, 15.07, 2013-04-25\n GHM, NYSE, 28.13, 2013-06-01\n GIB, NYSE, 14.58, 2013-05-01\n GLT, NYSE, 17.65, 2013-05-01\n GLW, NYSE, 15.38, 2013-04-25\n GSK, NYSE, 6.49, 2013-04-26\n GLF, NYSE, 175.00, 2013-04-30\n GNI, NYSE, -14.58, 2013-04-26\n GPC, NYSE, -6.06, 2013-04-20\n GRA, NYSE, 0.00, 2013-04-25\n GTY, NYSE, 0.00, 2013-05-03\n GWW, NYSE, 7.69, 2013-04-17\n HAE, NYSE, 4.35, 2013-05-02\n HAL, NYSE, 17.54, 2013-04-23\n HAR, NYSE, 25.40, 2013-05-03\n HVT, NYSE, 33.33, 2013-05-02\n HRC, NYSE, -2.00, 2013-04-25\n HCC, NYSE, 31.71, 2013-05-01\n HCN, NYSE, 1.11, 2013-05-08\n HCP, NYSE, 2.78, 2013-05-01\n HOG, NYSE, 2.06, 2013-04-26\n HE, NYSE, -12.82, 2013-05-09\n HL, NYSE, -66.67, 2013-05-11\n HMA, NYSE, 0.00, 2013-05-03\n HMC, NYSE, -28.57, 2013-04-27\n HMN, NYSE, 7.84, 2013-04-25\n HFC, NYSE, -7.91, 2013-05-08\n HOT, NYSE, 43.40, 2013-05-01\n HP, NYSE, 5.43, 2013-04-26\n HLS, NYSE, 14.29, 2013-04-26\n HRS, NYSE, 0.00, 2013-05-01\n HSC, NYSE, 50.00, 2013-05-10\n HSY, NYSE, 4.81, 2013-04-26\n HUBB, NYSE, -0.90, 2013-04-19\n HUM, NYSE, 51.12, 2013-05-02\n HXL, NYSE, 4.88, 2013-04-23\n IBM, NYSE, -1.96, 2013-04-19\n IDA, NYSE, 17.54, 2013-05-03\n IEX, NYSE, 4.23, 2013-04-23\n IFF, NYSE, 5.31, 2013-05-08\n DIN, NYSE, 12.87, 2013-05-03\n INT, NYSE, 14.06, 2013-05-01\n IP, NYSE, -12.16, 2013-05-03\n IPG, NYSE, -7.69, 2013-04-20\n IO, NYSE, -85.71, 2013-05-01\n IR, NYSE, 2.44, 2013-04-24\n IRF, NYSE, 27.50, 2013-04-30\n ITW, NYSE, 0.00, 2013-04-24\n JEC, NYSE, -2.44, 2013-04-30\n JNJ, NYSE, 2.13, 2013-04-17\n JNY, NYSE, 0.00, 2013-05-02\n K, NYSE, 0.00, 2013-05-03\n KAMN, NYSE, -2.94, 2013-04-30\n KDN, NYSE, 5.71, 2013-05-10\n KEX, NYSE, 2.15, 2013-04-25\n KEY, NYSE, 5.00, 2013-04-19\n KIM, NYSE, 3.13, 2013-05-01\n KMB, NYSE, 10.45, 2013-04-20\n KEM, NYSE, -133.33, 2013-05-10\n KMT, NYSE, -8.45, 2013-04-26\n KO, NYSE, 2.22, 2013-04-17\n KSU, NYSE, 2.30, 2013-04-20\n LDR, NYSE, -9.52, 2013-05-07\n LEG, NYSE, -13.16, 2013-04-26\n LLY, NYSE, 8.57, 2013-04-25\n LM, NYSE, -13.33, 2013-05-01\n LNC, NYSE, -7.27, 2013-05-02\n LPX, NYSE, 0.00, 2013-05-08\n LXU, NYSE, -110.53, 2013-05-07\n LTC, NYSE, -1.67, 2013-05-01\n L, NYSE, 1.19, 2013-04-30\n LUV, NYSE, 133.33, 2013-04-26\n LUX, NYSE, 7.14, 2013-05-02\n MKL, NYSE, 40.11, 2013-05-01\n MAN, NYSE, 40.00, 2013-04-20\n MTW, NYSE, -35.71, 2013-05-01\n SM, NYSE, 46.43, 2013-05-01\n MAS, NYSE, -7.14, 2013-04-30\n MTZ, NYSE, 12.50, 2013-05-03\n MCD, NYSE, -0.79, 2013-04-20\n MDC, NYSE, 73.08, 2013-05-03\n MDP, NYSE, 4.35, 2013-04-26\n MDR, NYSE, -40.00, 2013-05-09\n MDU, NYSE, 36.36, 2013-05-01\n MED, NYSE, 26.47, 2013-05-09\n CVS, NYSE, 5.06, 2013-05-02\n MFC, NYSE, 18.52, 2013-05-03\n MGA, NYSE, 13.57, 2013-05-11\n MGM, NYSE, 130.00, 2013-05-03\n MMC, NYSE, 4.29, 2013-05-03\n MMM, NYSE, -2.42, 2013-04-26\n MSA, NYSE, -20.31, 2013-04-25\n MNR, NYSE, -7.69, 2013-05-09\n MO, NYSE, 1.89, 2013-04-26\n MOD, NYSE, 5.88, 2013-05-31\nMOG.A, NYSE, -1.23, 2013-04-27\n MHK, NYSE, 3.57, 2013-05-03\n MSI, NYSE, -1.79, 2013-04-25\n MCY, NYSE, 46.81, 2013-04-30\n MRK, NYSE, 8.97, 2013-05-02\n MRO, NYSE, -28.17, 2013-05-08\n POWR, NYSE, 0.00, 2013-05-09\n MTG, NYSE, -60.00, 2013-05-01\n MTB, NYSE, 6.19, 2013-04-16\n MTX, NYSE, 0.00, 2013-04-26\n MUR, NYSE, 11.34, 2013-05-02\n MYE, NYSE, -11.11, 2013-04-25\n NBL, NYSE, 21.31, 2013-04-26\n NBR, NYSE, 13.79, 2013-04-24\n NE, NYSE, 3.51, 2013-04-18\n NEM, NYSE, -8.97, 2013-04-30\n NFG, NYSE, 7.37, 2013-05-03\n NHI, NYSE, 4.94, 2013-05-07\n NI, NYSE, -1.43, 2013-05-01\n NJR, NYSE, 3.16, 2013-05-03\n THC, NYSE, 17.86, 2013-05-01\n NNN, NYSE, 4.35, 2013-05-03\n NOC, NYSE, 12.14, 2013-04-25\n NR, NYSE, 5.88, 2013-04-26\n NSC, NYSE, 3.39, 2013-04-24\n NUE, NYSE, 4.00, 2013-04-19\n NVR, NYSE, -9.64, 2013-04-23\n NWL, NYSE, 9.38, 2013-05-04\n NWN, NYSE, -5.41, 2013-05-03\n NYT, NYSE, -20.00, 2013-04-26\n OCR, NYSE, 4.65, 2013-04-25\n OGE, NYSE, -32.35, 2013-05-03\n OHI, NYSE, 5.08, 2013-05-08\n OI, NYSE, 7.14, 2013-04-24\n OII, NYSE, 16.95, 2013-04-24\n OKE, NYSE, -6.90, 2013-05-01\n OLN, NYSE, 10.64, 2013-04-26\n BRS, NYSE, -1.94, 2013-05-23\n OMC, NYSE, 1.33, 2013-04-19\n OMI, NYSE, 4.76, 2013-04-24\n ORB, NYSE, 43.48, 2013-04-24\n ORI, NYSE, 600.00, 2013-04-26\n OSK, NYSE, 12.94, 2013-05-01\n OXY, NYSE, 7.64, 2013-04-26\n FCFS, NYSE, 0.00, 2013-04-18\n PBI, NYSE, 0.00, 2013-05-01\n PCG, NYSE, -10.00, 2013-05-03\n PCL, NYSE, 9.38, 2013-04-30\n PCP, NYSE, 1.81, 2013-05-10\n TPC, NYSE, 34.78, 2013-05-02\n PDS, NYSE, 14.29, 2013-04-26\n PEG, NYSE, 14.86, 2013-05-01\n PEI, NYSE, 4.76, 2013-04-23\n PEP, NYSE, 8.45, 2013-04-19\n PFE, NYSE, -1.82, 2013-05-01\n PG, NYSE, 3.13, 2013-04-25\n PGR, NYSE, -4.55, 2013-04-11\n PH, NYSE, 0.60, 2013-04-26\n PHM, NYSE, 31.25, 2013-04-26\n PKD, NYSE, 200.00, 2013-05-02\n PKY, NYSE, 15.38, 2013-05-07\n PNC, NYSE, 12.10, 2013-04-18\n PNM, NYSE, -10.00, 2013-05-07\n PNR, NYSE, 3.57, 2013-04-24\n PNW, NYSE, 175.00, 2013-05-04\n POM, NYSE, -4.00, 2013-05-04\n POT, NYSE, 3.28, 2013-04-26\n PPG, NYSE, 1.28, 2013-04-19\n PPL, NYSE, 0.00, 2013-05-03\n PRGO, NYSE, -1.39, 2013-05-08\n PL, NYSE, -4.30, 2013-05-07\n PSB, NYSE, 0.00, 2013-05-07\n WTR, NYSE, 7.41, 2013-05-02\n CSH, NYSE, 8.21, 2013-04-26\n PWR, NYSE, 24.14, 2013-05-03\n PX, NYSE, 0.00, 2013-04-25\n KWR, NYSE, 14.29, 2013-04-30\n R, NYSE, 1.28, 2013-04-24\n RBC, NYSE, -6.09, 2013-05-01\n RDC, NYSE, 5.77, 2013-05-02\n HTSI, NYSE, 11.67, 2013-05-03\n RES, NYSE, -33.33, 2013-04-25\n RGS, NYSE, -90.77, 2013-05-08\n RGR, NYSE, 15.38, 2013-04-30\n RHI, NYSE, -2.44, 2013-04-24\n RJF, NYSE, -9.33, 2013-04-25\n RLI, NYSE, -1.89, 2013-04-18\n ROG, NYSE, 0.00, 2013-05-01\n ROK, NYSE, 2.31, 2013-04-25\n ROL, NYSE, -5.88, 2013-04-25\n ROP, NYSE, 4.10, 2013-04-30\n RTI, NYSE, 20.00, 2013-05-01\n RTN, NYSE, 21.88, 2013-04-26\n RYL, NYSE, 43.33, 2013-04-25\n BSAC, NYSE, -21.74, 2013-04-26\n T, NYSE, 0.00, 2013-04-24\n SCG, NYSE, 7.77, 2013-04-26\n SCHW, NYSE, -6.25, 2013-04-16\n SCL, NYSE, -4.08, 2013-05-01\n SMG, NYSE, -19.60, 2013-05-07\n SEE, NYSE, -5.56, 2013-05-02\n SF, NYSE, 1.75, 2013-05-10\n SFE, NYSE, -46.15, 2013-04-26\n SHW, NYSE, 2.78, 2013-04-19\n SJI, NYSE, -8.43, 2013-05-04\n JOE, NYSE, -200.00, 2013-05-09\n SJW, NYSE, -12.50, 2013-04-25\n SLB, NYSE, 2.02, 2013-04-20\n HSH, NYSE, 9.38, 2013-05-03\n AOS, NYSE, 24.68, 2013-04-24\n SMP, NYSE, 31.25, 2013-05-04\n SNA, NYSE, 4.48, 2013-04-19\n PII, NYSE, 5.94, 2013-04-24\n SNV, NYSE, 0.00, 2013-04-24\n SO, NYSE, -3.92, 2013-04-25\n SON, NYSE, -5.66, 2013-04-19\n SPA, NYSE, -46.15, 2013-05-08\n TRV, NYSE, 14.93, 2013-04-24\n SR, NYSE, -3.36, 2013-05-01\n NVE, NYSE, 12.50, 2013-05-04\n SCI, NYSE, 21.74, 2013-04-25\n SSP, NYSE, 58.33, 2013-05-07\n STT, NYSE, 3.23, 2013-04-20\n STI, NYSE, 3.28, 2013-04-20\n STJ, NYSE, 0.00, 2013-04-18\n STL, NYSE, 7.14, 2013-04-23\n STR, NYSE, -2.38, 2013-05-01\n STE, NYSE, 6.06, 2013-05-08\n SYK, NYSE, 1.98, 2013-04-25\n SUN, NYSE, -7.32, 2013-05-09\n SUP, NYSE, 5.88, 2013-05-04\n SWK, NYSE, 7.29, 2013-04-26\n SWN, NYSE, 7.69, 2013-05-03\n SWX, NYSE, 0.61, 2013-05-04\n SWY, NYSE, -2.78, 2013-04-26\n SYY, NYSE, 16.67, 2013-05-07\n TAC, NYSE, -33.33, 2013-04-24\n TNC, NYSE, -17.14, 2013-04-23\n TCB, NYSE, -15.79, 2013-04-20\n TCO, NYSE, 7.14, 2013-04-26\n TDS, NYSE, 350.00, 2013-05-04\n TDW, NYSE, 55.74, 2013-05-22\n TDY, NYSE, 10.31, 2013-04-25\n TE, NYSE, 11.76, 2013-05-01\n TER, NYSE, 200.00, 2013-04-25\n TEVA, NYSE, 1.82, 2013-05-03\n TEX, NYSE, -17.86, 2013-04-25\n TFX, NYSE, 1.98, 2013-05-01\n TEN, NYSE, 10.77, 2013-04-30\n TKR, NYSE, 0.00, 2013-04-25\n TMK, NYSE, 1.46, 2013-04-24\n TMO, NYSE, 6.20, 2013-04-25\n TOT, NYSE, -2.38, 2013-04-27\n TM, NYSE, 80.67, 2013-05-09\n TR, NYSE, -11.76, 2013-04-25\n TRN, NYSE, 13.75, 2013-05-01\n TRP, NYSE, -8.93, 2013-04-27\n TSO, NYSE, 2.82, 2013-05-02\n TSS, NYSE, -2.94, 2013-04-24\n TTI, NYSE, -40.00, 2013-05-09\n TXT, NYSE, -14.89, 2013-04-18\n TYL, NYSE, 26.09, 2013-04-25\n TSN, NYSE, -21.74, 2013-05-07\n UDR, NYSE, 3.03, 2013-05-01\n UFI, NYSE, -43.75, 2013-04-25\n UAM, NYSE, 17.65, 2013-04-30\n UHS, NYSE, 5.17, 2013-04-25\n UIL, NYSE, 3.06, 2013-05-03\n UIS, NYSE, -145.61, 2013-04-24\n UNH, NYSE, 0.00, 2013-04-19\n KMPR, NYSE, 35.85, 2013-05-03\n UNM, NYSE, 2.56, 2013-05-02\n UNP, NYSE, 3.57, 2013-04-19\n UNT, NYSE, 6.98, 2013-05-08\n URS, NYSE, -14.29, 2013-05-08\n USG, NYSE, -88.89, 2013-04-25\n MUX, NYSE, -300.00, 2013-05-10\n USM, NYSE, 214.29, 2013-05-04\n USPH, NYSE, -3.12, 2013-05-10\n UTL, NYSE, -9.20, 2013-04-24\n UTX, NYSE, -1.54, 2013-04-24\n VMI, NYSE, 15.60, 2013-04-19\n VAR, NYSE, 2.97, 2013-04-25\n CBS, NYSE, 7.35, 2013-05-02\n VLO, NYSE, 16.83, 2013-05-01\n VMC, NYSE, -24.32, 2013-05-03\n VLY, NYSE, -11.11, 2013-04-25\n VNO, NYSE, -38.38, 2013-05-07\n VSH, NYSE, 63.64, 2013-05-01\n WTS, NYSE, -14.04, 2013-05-01\n WBS, NYSE, -2.22, 2013-04-16\n WEC, NYSE, 7.04, 2013-05-01\n WFC, NYSE, 5.75, 2013-04-13\n WG, NYSE, -2400.00, 2013-05-09\n WGL, NYSE, 19.05, 2013-05-02\n WHR, NYSE, 1.03, 2013-04-25\n WMB, NYSE, -8.33, 2013-05-08\n WNC, NYSE, 0.00, 2013-05-01\n TEG, NYSE, 10.69, 2013-05-02\n WR, NYSE, 33.33, 2013-05-09\n WRE, NYSE, -4.35, 2013-04-26\n WRI, NYSE, 4.35, 2013-05-01\n WPP, NYSE, 33.33, 2013-04-30\n WSO, NYSE, 18.18, 2013-04-19\n WST, NYSE, 1.16, 2013-05-03\n WWW, NYSE, 50.00, 2013-04-17\n WY, NYSE, 18.18, 2013-04-27\n X, NYSE, -84.21, 2013-05-01\n XL, NYSE, 38.81, 2013-05-03\n XOM, NYSE, 4.43, 2013-04-26\n XRX, NYSE, 12.50, 2013-04-24\n Y, NYSE, 53.96, 2013-05-07\n HRG, NYSE, 60.00, 2013-05-10\n CRY, NYSE, 28.57, 2013-05-01\n CHK, NYSE, 30.43, 2013-05-02\n DDR, NYSE, 0.00, 2013-05-01\n ELS, NYSE, 0.71, 2013-04-23\n ALG, NYSE, 5.56, 2013-05-02\n ETH, NYSE, -22.22, 2013-04-24\n ATR, NYSE, -3.03, 2013-04-26\n GGP, NYSE, 4.17, 2013-04-30\n MSL, NYSE, 3.70, 2013-05-01\n RCL, NYSE, 84.21, 2013-04-26\n CWEI, NYSE, -61.22, 2013-04-25\n HR, NYSE, 0.00, 2013-05-02\n RGA, NYSE, 2.48, 2013-04-26\n RIG, NYSE, -7.92, 2013-05-09\n SKT, NYSE, 2.44, 2013-05-01\n TWI, NYSE, -16.28, 2013-04-25\n BDN, NYSE, 2.94, 2013-04-25\n KGC, NYSE, 25.00, 2013-05-08\n CPT, NYSE, 2.11, 2013-05-03\n SGY, NYSE, 18.84, 2013-05-07\n BFS, NYSE, -24.49, 2013-05-01\n BWA, NYSE, 6.56, 2013-04-26\n EQR, NYSE, -1.54, 2013-05-01\n CLP, NYSE, 3.03, 2013-04-26\n KOF, NYSE, -16.24, 2013-04-25\n OKS, NYSE, -27.59, 2013-05-01\n SQM, NYSE, -6.45, 2013-05-29\n BYD, NYSE, 114.29, 2013-04-25\n CBL, NYSE, 3.92, 2013-04-30\n DECK, NYSE, 133.33, 2013-04-26\n IT, NYSE, -2.50, 2013-05-03\n HST, NYSE, 21.74, 2013-05-04\n LXP, NYSE, 0.00, 2013-05-03\n REG, NYSE, 3.23, 2013-05-08\n TUC, NYSE, -24.00, 2013-05-03\n AF, NYSE, 7.69, 2013-04-18\n BFR, NYSE, -2.56, 2013-05-11\n HHS, NYSE, 10.00, 2013-04-26\n MHO, NYSE, 28.57, 2013-04-26\n NFX, NYSE, -2.17, 2013-04-24\n SPG, NYSE, 1.99, 2013-04-27\n SU, NYSE, -1.41, 2013-04-30\n SUI, NYSE, 2.20, 2013-04-26\n TV, NYSE, -22.50, 2013-04-26\n CGI, NYSE, -26.92, 2013-04-26\n CYT, NYSE, -12.79, 2013-04-19\n EMN, NYSE, 3.18, 2013-04-26\n GRT, NYSE, 14.29, 2013-04-25\n MAA, NYSE, 5.04, 2013-05-02\n PLT, NYSE, 4.62, 2013-05-08\n BZH, NYSE, 15.38, 2013-05-03\n ELX, NYSE, 114.29, 2013-05-03\n MLM, NYSE, -69.44, 2013-05-01\n AKS, NYSE, 41.67, 2013-04-24\n ALB, NYSE, -7.00, 2013-04-18\n VRX, NYSE, 1.56, 2013-05-03\n CBR, NYSE, 0.00, 2013-05-01\n MAC, NYSE, 8.86, 2013-05-02\n RKT, NYSE, 9.80, 2013-04-24\n RYN, NYSE, 27.42, 2013-04-26\n ADC, NYSE, -2.00, 2013-04-30\nBRK.B, NYSE, 52.31, 2013-05-04\n EXP, NYSE, 5.00, 2013-05-15\n GGB, NYSE, -66.67, 2013-05-08\n SSD, NYSE, -52.38, 2013-04-26\n ESS, NYSE, -0.53, 2013-05-02\n FR, NYSE, -7.69, 2013-04-26\n HIW, NYSE, -2.90, 2013-05-01\n IMAX, NYSE, 0.00, 2013-04-26\n AIV, NYSE, 2.13, 2013-05-03\n FCH, NYSE, 0.00, 2013-05-01\n ITGR, NYSE, 2.33, 2013-04-26\n NOK, NYSE, 33.33, 2013-04-19\n GEO, NYSE, -3.51, 2013-05-09\n CLI, NYSE, 0.00, 2013-04-26\n RS, NYSE, -5.22, 2013-04-26\n CPE, NYSE, 100.00, 2013-05-10\n KNX, NYSE, 0.00, 2013-04-25\n O, NYSE, 1.69, 2013-04-26\n COF, NYSE, 17.79, 2013-04-19\n IRS, NYSE, 10.34, 2013-05-18\n MCK, NYSE, -0.43, 2013-05-08\n SWC, NYSE, 200.00, 2013-04-30\n STM, NYSE, 23.53, 2013-04-23\n TEO, NYSE, 1.30, 2013-04-30\n TRK, NYSE, -400.00, 2013-05-02\n LMT, NYSE, 23.38, 2013-04-24\n APU, NYSE, -35.48, 2013-05-16\n AGU, NYSE, -12.15, 2013-05-10\n LH, NYSE, -1.69, 2013-04-20\n DDD, NYSE, -10.00, 2013-05-01\n AFG, NYSE, 10.84, 2013-05-09\n RMD, NYSE, 3.51, 2013-04-26\n WAB, NYSE, 3.60, 2013-04-25\n CIB, NYSE, 6.78, 2013-05-08\n CAM, NYSE, -5.41, 2013-04-26\n FCX, NYSE, 1.39, 2013-04-19\n RNR, NYSE, 34.25, 2013-05-02\n AVX, NYSE, 7.14, 2013-04-25\n RWT, NYSE, 46.81, 2013-05-03\n AXE, NYSE, -6.62, 2013-04-24\n CLB, NYSE, 6.09, 2013-04-18\n MD, NYSE, 0.92, 2013-05-03\n THG, NYSE, 30.69, 2013-04-30\n BAP, NYSE, -10.94, 2013-05-07\n DO, NYSE, 10.43, 2013-04-26\n RE, NYSE, 36.11, 2013-04-23\n DST, NYSE, -6.60, 2013-04-26\n EL, NYSE, 36.36, 2013-05-03\n ESC, NYSE, -57.14, 2013-05-03\n LXK, NYSE, -7.55, 2013-04-24\n MIG, NYSE, 7.69, 2013-05-01\n WAT, NYSE, -1.83, 2013-04-24\n EME, NYSE, 2.27, 2013-04-26\n HIG, NYSE, 10.84, 2013-04-30\n ITT, NYSE, 9.30, 2013-05-03\n SPN, NYSE, 0.00, 2013-04-26\n SWM, NYSE, 8.60, 2013-05-09\n SCCO, NYSE, -4.84, 2013-04-27\n RCI, NYSE, -1.27, 2013-04-23\n EIX, NYSE, 20.31, 2013-05-01\n IRM, NYSE, 0.00, 2013-05-02\n SPH, NYSE, -4.82, 2013-05-10\n CCJ, NYSE, 0.00, 2013-05-02\n PGI, NYSE, 0.00, 2013-04-19\n CRR, NYSE, -14.61, 2013-04-26\n BVN, NYSE, -40.30, 2013-04-30\n FCN, NYSE, 13.46, 2013-05-10\n RPT, NYSE, 6.90, 2013-04-24\n TUP, NYSE, 4.42, 2013-04-25\n ASB, NYSE, 8.00, 2013-04-19\n GWR, NYSE, -10.11, 2013-05-02\n TBI, NYSE, -50.00, 2013-04-25\n FFG, NYSE, 12.66, 2013-05-03\n USNA, NYSE, 14.29, 2013-04-24\n CSV, NYSE, -3.03, 2013-05-08\n LVB, NYSE, 10.53, 2013-05-09\n ALR, NYSE, 6.25, 2013-05-10\n OCN, NYSE, 0.00, 2013-05-03\n PAA, NYSE, 37.50, 2013-05-07\n DNR, NYSE, 13.79, 2013-05-03\n HMY, NYSE, -119.23, 2013-05-04\n TGI, NYSE, 5.66, 2013-05-02\n PAG, NYSE, 1.61, 2013-04-30\n GEL, NYSE, -17.65, 2013-05-03\n IM, NYSE, 0.00, 2013-04-26\n NUS, NYSE, 13.92, 2013-05-03\n CNI, NYSE, -1.67, 2013-04-23\n LAD, NYSE, 16.67, 2013-04-25\n NSP, NYSE, 0.00, 2013-04-30\n DGX, NYSE, -14.42, 2013-04-18\n KRC, NYSE, 0.00, 2013-05-01\n MTH, NYSE, 32.00, 2013-04-25\n NCR, NYSE, 35.00, 2013-05-01\n OFG, NYSE, 2.78, 2013-04-26\n IVZ, NYSE, 10.64, 2013-05-01\n DX, NYSE, 9.68, 2013-05-02\n FBC, NYSE, -65.98, 2013-04-24\n ALV, NYSE, 1.57, 2013-04-27\n ARE, NYSE, 0.00, 2013-04-30\n BBT, NYSE, 2.99, 2013-04-19\n CGG, NYSE, 6.25, 2013-05-04\n BXP, NYSE, -0.83, 2013-05-01\n CBD, NYSE, -23.73, 2013-05-01\n MS, NYSE, 7.02, 2013-04-19\n SRT, NYSE, -314.29, 2013-05-10\n HLX, NYSE, 38.89, 2013-04-22\n FLS, NYSE, 3.61, 2013-04-25\n MT, NYSE, -400.00, 2013-05-11\n PXD, NYSE, 5.15, 2013-05-02\n SLG, NYSE, 0.83, 2013-04-24\n NAT, NYSE, -16.22, 2013-05-14\n CSU, NYSE, -36.36, 2013-05-07\n DRQ, NYSE, 22.50, 2013-05-04\n FDP, NYSE, -24.47, 2013-05-01\n NLY, NYSE, 30.56, 2013-05-02\n TLM, NYSE, -250.00, 2013-05-02\n TSM, NYSE, 13.04, 2013-04-19\n YUM, NYSE, 12.90, 2013-04-24\n AMG, NYSE, 12.38, 2013-05-01\n EPR, NYSE, -1.05, 2013-05-01\n FE, NYSE, 10.14, 2013-05-08\n LFL, NYSE, 80.00, 2013-05-15\n MTD, NYSE, 2.79, 2013-05-03\n SID, NYSE, -66.67, 2013-05-16\n IN, NYSE, -271.43, 2013-05-04\n CBZ, NYSE, 25.64, 2013-05-03\n URI, NYSE, 11.54, 2013-04-17\n INGR, NYSE, 6.82, 2013-05-03\n RAS, NYSE, 181.82, 2013-05-03\n UNS, NYSE, 35.00, 2013-04-30\n ASI, NYSE, 18.92, 2013-05-09\n ANH, NYSE, 15.38, 2013-04-30\n OFC, NYSE, 17.07, 2013-04-27\n GPX, NYSE, 0.00, 2013-05-03\n WAC, NYSE, 1427.27, 2013-05-10\n RBA, NYSE, -13.33, 2013-05-01\n WDR, NYSE, 1.61, 2013-04-24\n LHO, NYSE, 8.00, 2013-04-18\n LNT, NYSE, 18.03, 2013-05-04\n LVLT, NYSE, 7.14, 2013-04-26\n MFA, NYSE, -4.76, 2013-05-02\n OME, NYSE, 50.00, 2013-05-08\n EQY, NYSE, 6.90, 2013-05-02\n FII, NYSE, -2.38, 2013-04-26\n FMX, NYSE, -37.89, 2013-04-25\n LLL, NYSE, 3.63, 2013-04-26\n VTR, NYSE, 4.04, 2013-04-27\n WCN, NYSE, 20.00, 2013-05-02\n AVB, NYSE, 0.74, 2013-05-01\n GIL, NYSE, 5.36, 2013-05-03\n HZO, NYSE, -92.86, 2013-04-26\n AWR, NYSE, 38.00, 2013-05-11\n CLS, NYSE, 10.00, 2013-04-24\n EPD, NYSE, 16.67, 2013-05-01\n RSG, NYSE, 15.00, 2013-04-26\n WM, NYSE, -2.44, 2013-04-25\n AKR, NYSE, 3.33, 2013-04-24\n CVG, NYSE, 17.39, 2013-05-01\n RRC, NYSE, -38.89, 2013-04-26\n SAP, NYSE, 41.51, 2013-04-20\n CCI, NYSE, 0.00, 2013-04-25\n PQ, NYSE, 100.00, 2013-05-08\n WFT, NYSE, 0.00, 2013-05-03\n CAA, NYSE, 0.00, 2013-05-03\n ENB, NYSE, 13.21, 2013-05-09\n GMK, NYSE, 60.00, 2013-04-25\n MMR, NYSE, 0.00, 2013-05-07\n PB, NYSE, 2.38, 2013-04-25\n VIV, NYSE, -20.00, 2013-05-08\n AXL, NYSE, 53.33, 2013-05-04\n BP, NYSE, 33.33, 2013-05-01\n ETM, NYSE, 0.00, 2013-05-09\n HT, NYSE, 0.00, 2013-05-01\n BYI, NYSE, 10.71, 2013-04-25\n CEB, NYSE, 1.64, 2013-05-02\n INFY, NYSE, 5.41, 2013-04-13\n JLL, NYSE, 56.52, 2013-05-01\n AZN, NYSE, 5.22, 2013-04-26\n SFG, NYSE, 33.75, 2013-04-24\n TREX, NYSE, 14.68, 2013-05-04\n GS, NYSE, 11.43, 2013-04-17\n SYX, NYSE, -157.14, 2013-05-01\n WCC, NYSE, -4.27, 2013-04-19\n JNPR, NYSE, 33.33, 2013-04-24\n RDN, NYSE, 28.57, 2013-05-02\n RAI, NYSE, 4.35, 2013-04-24\n SKX, NYSE, -27.78, 2013-05-16\n WTM, NYSE, 178.02, 2013-04-30\n NCI, NYSE, 12.50, 2013-04-26\n BLT, NYSE, -17.39, 2013-05-08\n QTM, NYSE, -33.33, 2013-05-09\n BLK, NYSE, 1.67, 2013-04-17\n CIR, NYSE, 4.00, 2013-05-03\n MSO, NYSE, 12.50, 2013-05-01\n PKG, NYSE, 10.71, 2013-04-23\n PKI, NYSE, -25.00, 2013-04-26\n WWE, NYSE, -37.50, 2013-05-03\n SNN, NYSE, -2.11, 2013-05-03\n UPS, NYSE, 2.97, 2013-04-26\n XOXO, NYSE, 16.67, 2013-05-10\n SLF, NYSE, 7.25, 2013-05-09\n CDR, NYSE, 9.09, 2013-05-10\n EW, NYSE, -5.26, 2013-04-24\n MET, NYSE, 13.85, 2013-05-01\n FBR, NYSE, -89.47, 2013-04-24\n VVC, NYSE, -7.58, 2013-05-02\n BAM, NYSE, 70.00, 2013-05-10\n NVS, NYSE, 4.00, 2013-04-25\n BHLB, NYSE, -1.82, 2013-04-30\n CRL, NYSE, -2.82, 2013-05-02\n CYH, NYSE, 3.57, 2013-04-30\n MBT, NYSE, -13.04, 2013-06-08\n MTOR, NYSE, 500.00, 2013-05-01\n CNQ, NYSE, -44.19, 2013-05-03\n ERJ, NYSE, -62.79, 2013-04-30\n VZ, NYSE, 3.03, 2013-04-19\n EVC, NYSE, 0.00, 2013-05-03\n PBR, NYSE, 0.00, 2013-04-27\n XEL, NYSE, 11.63, 2013-05-03\n ALE, NYSE, 10.67, 2013-05-09\n HW, NYSE, -30.00, 2013-05-01\n POL, NYSE, 14.81, 2013-05-02\n COH, NYSE, 3.70, 2013-04-24\n CXW, NYSE, 6.38, 2013-05-09\n DVA, NYSE, 3.37, 2013-05-08\n EXC, NYSE, 4.41, 2013-05-02\n MCO, NYSE, 11.49, 2013-05-04\n BRFS, NYSE, 23.53, 2013-04-30\n TU, NYSE, 3.77, 2013-05-10\n WIT, NYSE, 0.00, 2013-04-20\n ERF, NYSE, 100.00, 2013-05-11\n GG, NYSE, -35.00, 2013-05-03\n HNT, NYSE, 34.15, 2013-04-30\n NYCB, NYSE, 3.85, 2013-04-25\n SXT, NYSE, 3.33, 2013-04-19\n CPG, NYSE, -20.00, 2013-05-10\n AMX, NYSE, 16.67, 2013-04-20\n MPX, NYSE, 0.00, 2013-04-25\n OIS, NYSE, -2.70, 2013-04-25\n MMP, NYSE, 4.08, 2013-05-03\n PES, NYSE, 33.33, 2013-05-01\n ABB, NYSE, -12.12, 2013-04-25\n KMR, NYSE, -3.28, 2013-05-02\n GEN, NYSE, -41.18, 2013-05-07\n ADS, NYSE, -2.88, 2013-04-19\n CVI, NYSE, 25.00, 2013-05-03\n FTI, NYSE, -6.52, 2013-04-24\n PRA, NYSE, 27.63, 2013-05-07\n STO, NYSE, -16.46, 2013-05-03\n BEL, NYSE, 41.67, 2013-05-02\n FIS, NYSE, 1.64, 2013-05-01\n COL, NYSE, 0.86, 2013-04-20\n KAI, NYSE, 20.51, 2013-04-30\n ABC, NYSE, -2.25, 2013-04-26\n BG, NYSE, 18.56, 2013-04-26\n FRO, NYSE, 27.08, 2013-05-31\n ECA, NYSE, 150.00, 2013-04-24\n CIG, NYSE, 108.33, 2013-05-17\n EEP, NYSE, 16.67, 2013-05-01\n CVX, NYSE, 3.25, 2013-04-27\n GXP, NYSE, 41.67, 2013-05-10\n JHX, NYSE, -2.78, 2013-05-24\n PFG, NYSE, 5.33, 2013-04-26\n PVR, NYSE, 14.29, 2013-04-26\n AAP, NYSE, 2.48, 2013-05-24\n KND, NYSE, 36.11, 2013-05-02\n WTW, NYSE, 38.10, 2013-05-03\n CNC, NYSE, 5.00, 2013-04-24\n BCH, NYSE, 3.70, 2013-05-09\n NS, NYSE, -86.67, 2013-04-25\n ITUB, NYSE, -4.88, 2013-04-26\n SXL, NYSE, 26.74, 2013-05-09\n VALE, NYSE, 50.00, 2013-04-25\n TNP, NYSE, 150.00, 2013-05-25\n LCI, NYSE, 40.00, 2013-05-09\n GTI, NYSE, 50.00, 2013-04-26\n HNR, NYSE, -26.67, 2013-06-06\n MWE, NYSE, -90.00, 2013-05-09\n NLS, NYSE, 50.00, 2013-05-07\n RGC, NYSE, -7.14, 2013-05-01\n JAH, NYSE, 30.43, 2013-04-25\n NPO, NYSE, -23.29, 2013-05-03\n TRI, NYSE, 22.58, 2013-05-01\n CAE, NYSE, 10.53, 2013-05-17\n LF, NYSE, 28.57, 2013-05-02\n SNY, NYSE, -10.11, 2013-05-03\n BANC, NYSE, 400.00, 2013-05-09\n COP, NYSE, 0.00, 2013-04-26\n CNP, NYSE, -8.11, 2013-05-03\n EEQ, NYSE, -321.43, 2013-05-02\n MRH, NYSE, 32.58, 2013-04-25\n NGS, NYSE, 23.08, 2013-05-10\n NRP, NYSE, 4.88, 2013-05-07\n PXP, NYSE, 17.98, 2013-05-03\n XEC, NYSE, -0.93, 2013-05-08\n IAG, NYSE, 7.14, 2013-05-08\n EGO, NYSE, 0.00, 2013-05-03\n JNS, NYSE, -6.25, 2013-04-24\n PFS, NYSE, 14.81, 2013-04-27\n ENH, NYSE, 74.79, 2013-05-02\n CNX, NYSE, -5.00, 2013-04-26\n AMT, NYSE, -10.42, 2013-05-02\n ABG, NYSE, 13.43, 2013-04-25\n LII, NYSE, 22.22, 2013-04-23\n SRE, NYSE, -4.90, 2013-05-03\n AEE, NYSE, -21.43, 2013-05-03\n PLD, NYSE, 0.00, 2013-04-25\n SAH, NYSE, -2.38, 2013-04-24\n GPI, NYSE, 11.54, 2013-05-03\n FIX, NYSE, 800.00, 2013-05-02\n MMS, NYSE, 1.41, 2013-05-10\n SRI, NYSE, 50.00, 2013-05-10\n RTEC, NYSE, 50.00, 2013-05-03\n NOV, NYSE, -5.84, 2013-04-27\n DF, NYSE, 11.54, 2013-05-10\n SAM, NYSE, -17.74, 2013-05-02\n RL, NYSE, 8.46, 2013-05-24\n FLR, NYSE, 6.25, 2013-05-03\n ALL, NYSE, 2.27, 2013-05-02\n ATI, NYSE, 0.00, 2013-04-25\n EE, NYSE, 72.73, 2013-05-02\n AIT, NYSE, 0.00, 2013-05-03\n CHH, NYSE, -3.70, 2013-04-30\n FMS, NYSE, -17.78, 2013-05-01\n BCO, NYSE, 16.67, 2013-04-26\n CBB, NYSE, 133.33, 2013-05-10\n MWW, NYSE, 14.29, 2013-05-03\n PSA, NYSE, -3.09, 2013-05-10\n E, NYSE, 0.00, 2013-04-25\n JPM, NYSE, 15.22, 2013-04-13\n USB, NYSE, 0.00, 2013-04-17\n HON, NYSE, 6.14, 2013-04-20\n ITG, NYSE, 50.00, 2013-05-03\n ARB, NYSE, -15.49, 2013-05-08\n APL, NYSE, -28.95, 2013-04-30\n AVA, NYSE, 0.00, 2013-05-02\n AXS, NYSE, 85.71, 2013-04-26\n MOH, NYSE, 146.15, 2013-04-26\n CVD, NYSE, 4.17, 2013-05-02\n AHT, NYSE, 2.94, 2013-05-09\n GPK, NYSE, 25.00, 2013-04-26\n CNO, NYSE, 0.00, 2013-04-25\n AUQ, NYSE, -60.00, 2013-05-10\n NFP, NYSE, -5.45, 2013-05-04\n CRI, NYSE, 12.86, 2013-05-10\n FMD, NYSE, 27.27, 2013-04-30\n FPO, NYSE, 3.45, 2013-04-26\n TRQ, NYSE, -25.00, 2013-05-14\n WLL, NYSE, 2.17, 2013-04-25\n AEL, NYSE, 11.36, 2013-05-02\n AHL, NYSE, 0.95, 2013-04-25\n AUY, NYSE, -23.81, 2013-05-01\n CMP, NYSE, 24.32, 2013-04-30\n KRO, NYSE, -800.00, 2013-05-09\n TPX, NYSE, 3.33, 2013-05-03\n UTI, NYSE, -300.00, 2013-05-01\n PJC, NYSE, 9.09, 2013-04-18\n TRW, NYSE, 3.42, 2013-05-01\n AIZ, NYSE, -14.56, 2013-04-25\n HTH, NYSE, 11.43, 2013-05-07\n ETP, NYSE, 33.33, 2013-05-09\n LSE, NYSE, 0.00, 2013-05-09\n BBD, NYSE, 0.00, 2013-04-23\n NRG, NYSE, -37.04, 2013-05-08\n HOS, NYSE, 96.67, 2013-05-02\n ABR, NYSE, 84.62, 2013-05-04\n FHN, NYSE, 0.00, 2013-04-20\n AGO, NYSE, 86.11, 2013-05-10\n HSP, NYSE, 18.18, 2013-05-02\n HNI, NYSE, 250.00, 2013-04-18\n GHL, NYSE, -34.78, 2013-04-18\n XPO, NYSE, -16.44, 2013-05-08\n CVO, NYSE, -200.00, 2013-05-09\n CHE, NYSE, 9.92, 2013-04-19\n GNW, NYSE, 11.11, 2013-05-01\n CBG, NYSE, -5.88, 2013-04-26\n SFL, NYSE, -43.33, 2013-05-31\n NEU, NYSE, 3.28, 2013-04-25\n GOL, NYSE, -1200.00, 2013-05-14\n CAB, NYSE, 18.64, 2013-04-26\n LTM, NYSE, 3.08, 2013-04-26\n VVI, NYSE, 68.00, 2013-04-27\n WCG, NYSE, -8.70, 2013-05-04\n HEP, NYSE, -36.36, 2013-05-01\n DPZ, NYSE, 5.36, 2013-05-01\n BDC, NYSE, 6.33, 2013-05-03\n ENS, NYSE, 2.56, 2013-05-29\n BMR, NYSE, 7.89, 2013-05-02\n ACC, NYSE, -1.54, 2013-04-24\n KRG, NYSE, 27.27, 2013-05-03\n WLK, NYSE, 42.64, 2013-05-07\n EXR, NYSE, 4.55, 2013-04-30\n CNS, NYSE, 7.32, 2013-04-18\n IOC, NYSE, 161.54, 2013-05-14\n STON, NYSE, -150.00, 2013-05-08\n TTM, NYSE, 60.56, 2013-05-30\n CPL, NYSE, 7.69, 2013-05-11\n TPGI, NYSE, -460.00, 2013-05-07\n SHO, NYSE, 0.00, 2013-05-07\n CUBE, NYSE, 0.00, 2013-05-03\n NRF, NYSE, -51.35, 2013-05-04\n DLR, NYSE, -1.69, 2013-04-27\n MTL, NYSE, 100.00, 2013-06-19\n NWE, NYSE, 8.60, 2013-04-26\n ORA, NYSE, 550.00, 2013-05-08\n NP, NYSE, 7.25, 2013-05-09\n SMA, NYSE, -73.33, 2013-05-03\n BBG, NYSE, -2600.00, 2013-05-03\n BXC, NYSE, 35.29, 2013-05-02\n KNL, NYSE, 8.33, 2013-04-19\n LVS, NYSE, 7.58, 2013-05-02\n HLF, NYSE, 18.69, 2013-04-30\n MIC, NYSE, -89.09, 2013-04-30\n PHH, NYSE, -81.13, 2013-05-02\n CE, NYSE, 44.30, 2013-04-19\n EDR, NYSE, 0.00, 2013-04-30\n WTI, NYSE, 34.62, 2013-05-08\n ARC, NYSE, 0.00, 2013-05-08\n PBH, NYSE, 5.88, 2013-05-17\n HUN, NYSE, 18.75, 2013-05-01\n WEX, NYSE, 3.16, 2013-05-02\n DLB, NYSE, 14.29, 2013-04-26\n DSX, NYSE, 66.67, 2013-05-23\n LAZ, NYSE, -17.65, 2013-04-27\n TGP, NYSE, 14.29, 2013-05-10\n TLP, NYSE, 7.69, 2013-05-08\n DRH, NYSE, 55.56, 2013-05-11\n HTGC, NYSE, 8.00, 2013-05-03\n KFN, NYSE, 27.78, 2013-05-02\n THS, NYSE, 5.71, 2013-05-10\n NSR, NYSE, -8.86, 2013-05-03\n WAL, NYSE, 14.29, 2013-04-19\n SLW, NYSE, -9.76, 2013-05-11\n MPW, NYSE, -3.85, 2013-04-27\n GNK, NYSE, -2.75, 2013-05-02\n MFB, NYSE, 28.57, 2013-05-09\nRDS.A, NYSE, 21.74, 2013-05-03\n ITC, NYSE, -3.45, 2013-04-24\n FTK, NYSE, -11.76, 2013-05-10\n PIKE, NYSE, -20.00, 2013-05-07\n ALJ, NYSE, 63.27, 2013-05-09\n DRC, NYSE, 2.38, 2013-04-26\n STN, NYSE, 0.00, 2013-05-10\n SSW, NYSE, -8.70, 2013-04-30\n CF, NYSE, 0.50, 2013-05-09\n HPY, NYSE, 12.50, 2013-05-01\n ROC, NYSE, 1.49, 2013-05-01\n WPZ, NYSE, -57.58, 2013-05-01\n LCC, NYSE, 29.17, 2013-04-24\n GLP, NYSE, -7.27, 2013-05-10\n AMP, NYSE, 1.27, 2013-04-23\n DHT, NYSE, 58.33, 2013-04-30\n FNF, NYSE, 5.00, 2013-05-02\n NM, NYSE, 52.38, 2013-05-22\n CCO, NYSE, -57.14, 2013-05-03\n BWP, NYSE, 5.00, 2013-04-30\n ICE, NYSE, 2.53, 2013-05-02\n BKD, NYSE, 50.00, 2013-05-02\n BAS, NYSE, 12.00, 2013-04-25\n CPA, NYSE, 21.21, 2013-05-14\n LYV, NYSE, 8.33, 2013-05-08\n WNR, NYSE, -6.93, 2013-05-03\n CMG, NYSE, 9.81, 2013-04-19\n RGP, NYSE, -50.00, 2013-05-09\n KOP, NYSE, -16.92, 2013-05-04\n TX, NYSE, 40.43, 2013-05-01\n UAL, NYSE, 10.09, 2013-04-26\n ETE, NYSE, -27.03, 2013-05-09\n RSO, NYSE, -45.00, 2013-05-08\n XCO, NYSE, 62.50, 2013-05-01\n PAC, NYSE, 30.00, 2013-04-26\n NYX, NYSE, 1.79, 2013-05-01\n TDG, NYSE, 0.61, 2013-05-08\n BMA, NYSE, 11.68, 2013-05-09\n THI, NYSE, 1.67, 2013-05-09\n BTE, NYSE, -112.00, 2013-05-10\n CNH, NYSE, 41.49, 2013-05-01\n GLA, NYSE, -82.35, 2013-05-02\n POR, NYSE, 0.00, 2013-05-02\n HIL, NYSE, 50.00, 2013-05-03\n HVB, NYSE, 12.50, 2013-04-24\n KS, NYSE, -9.30, 2013-05-08\n HK, NYSE, -28.57, 2013-05-03\n DCP, NYSE, 3.28, 2013-05-07\n DK, NYSE, 7.56, 2013-05-09\n CODI, NYSE, 0.00, 2013-05-08\n MA, NYSE, 0.65, 2013-05-02\n MWA, NYSE, 150.00, 2013-05-01\n KOG, NYSE, -21.43, 2013-05-03\n PWE, NYSE, -150.00, 2013-05-03\n PGTI, NYSE, 100.00, 2013-05-02\n AWH, NYSE, 8.45, 2013-04-25\n NSH, NYSE, -29.73, 2013-04-25\n WYN, NYSE, 7.58, 2013-04-25\n WNS, NYSE, 15.38, 2013-04-18\n PGH, NYSE, 0.00, 2013-05-02\n AYR, NYSE, 34.48, 2013-05-03\n EVR, NYSE, -24.49, 2013-04-25\n HBI, NYSE, 2.00, 2013-04-24\n WU, NYSE, 12.12, 2013-05-01\n OC, NYSE, 45.00, 2013-04-25\n DAC, NYSE, 44.44, 2013-04-30\n AWI, NYSE, -43.59, 2013-04-30\n SUSS, NYSE, 0.00, 2013-05-09\n DEI, NYSE, 5.71, 2013-05-08\n OB, NYSE, 79.31, 2013-04-30\n SBH, NYSE, -7.69, 2013-05-03\n EBS, NYSE, -144.44, 2013-05-03\n KBR, NYSE, 25.53, 2013-04-26\n AER, NYSE, 23.40, 2013-05-08\n NOA, NYSE, -442.86, 2013-06-11\n SPR, NYSE, 29.79, 2013-05-03\n ANW, NYSE, -7.14, 2013-05-16\n DCT, NYSE, 10.00, 2013-05-03\n SE, NYSE, 6.25, 2013-05-04\n TOO, NYSE, -17.86, 2013-05-10\n TSL, NYSE, -27.78, 2013-05-30\n TWC, NYSE, 2.92, 2013-04-26\n MVO, NYSE, -13.92, 2013-05-09\n CO, NYSE, 150.00, 2013-06-19\n EXK, NYSE, -18.75, 2013-05-07\n EIG, NYSE, 22.22, 2013-05-09\n HF, NYSE, -50.00, 2013-05-02\n FIG, NYSE, 33.33, 2013-05-03\n NGLS, NYSE, -20.00, 2013-05-04\n TCAP, NYSE, -1.75, 2013-05-09\n GFA, NYSE, -211.11, 2013-05-14\n BR, NYSE, 18.18, 2013-05-08\n SCR, NYSE, 12.50, 2013-05-10\n CNK, NYSE, 12.00, 2013-05-08\n DAL, NYSE, 42.86, 2013-04-24\n ORN, NYSE, 42.86, 2013-05-03\n ACM, NYSE, 3.92, 2013-05-08\n SLH, NYSE, 5.00, 2013-05-08\n CLR, NYSE, 2.63, 2013-05-09\n BGS, NYSE, -5.13, 2013-04-19\n STAR, NYSE, 26.42, 2013-05-01\n YGE, NYSE, -40.00, 2013-05-31\n DFS, NYSE, 18.75, 2013-04-24\n TEL, NYSE, 7.04, 2013-04-25\n BX, NYSE, 1.85, 2013-04-19\n SEP, NYSE, 4.65, 2013-05-04\n BZ, NYSE, -77.78, 2013-05-03\n PPO, NYSE, -41.18, 2013-05-09\n PRO, NYSE, 100.00, 2013-05-03\n WBC, NYSE, 7.34, 2013-04-26\n DHX, NYSE, 0.00, 2013-04-24\n PMC, NYSE, 23.53, 2013-05-02\n HGG, NYSE, 3.33, 2013-05-21\n OWW, NYSE, -33.33, 2013-05-10\n VR, NYSE, 35.97, 2013-04-26\n CXO, NYSE, -27.50, 2013-05-02\n G, NYSE, 5.00, 2013-05-02\n EJ, NYSE, 89.47, 2013-05-16\n WX, NYSE, 11.11, 2013-05-14\n CMLP, NYSE, -92.86, 2013-05-08\n VMW, NYSE, 10.87, 2013-04-24\n CZZ, NYSE, -40.00, 2013-06-06\n CGA, NYSE, 6.67, 2013-05-14\n TDC, NYSE, -26.92, 2013-05-03\n FLY, NYSE, 61.73, 2013-05-03\n MAIN, NYSE, 2.04, 2013-05-10\n REN, NYSE, 100.00, 2013-05-07\n TGH, NYSE, -12.90, 2013-05-08\n DFT, NYSE, -5.00, 2013-05-08\n RF, NYSE, 15.00, 2013-04-24\n PZN, NYSE, 0.00, 2013-04-25\n LL, NYSE, 29.55, 2013-04-25\n NMM, NYSE, 0.00, 2013-04-26\n OZM, NYSE, 81.25, 2013-05-03\n ES, NYSE, 12.31, 2013-05-02\n MSCI, NYSE, 5.56, 2013-05-02\n ARR, NYSE, -21.74, 2013-05-03\n KW, NYSE, 62.50, 2013-05-08\n GTS, NYSE, 52.78, 2013-05-02\n FOR, NYSE, 450.00, 2013-05-09\n LRN, NYSE, 34.78, 2013-05-04\n TNK, NYSE, -100.00, 2013-05-10\n N, NYSE, -21.43, 2013-04-26\n DAN, NYSE, -33.33, 2013-04-26\n BIP, NYSE, 0.00, 2013-05-03\n CPN, NYSE, -6.67, 2013-05-03\n SOL, NYSE, -15.38, 2013-05-17\n PM, NYSE, -4.44, 2013-04-19\n V, NYSE, 6.08, 2013-05-02\n IPI, NYSE, 5.26, 2013-05-02\n AWK, NYSE, -5.88, 2013-05-08\n HTS, NYSE, -7.46, 2013-04-23\n DPS, NYSE, 12.77, 2013-04-25\n CFX, NYSE, 8.33, 2013-04-26\n WES, NYSE, -22.50, 2013-05-02\n SB, NYSE, 0.00, 2013-05-16\n LO, NYSE, 4.76, 2013-04-25\n LPS, NYSE, 0.00, 2013-04-25\n FF, NYSE, -6.90, 2013-05-08\n NNA, NYSE, 200.00, 2013-05-03\n EPB, NYSE, 7.41, 2013-04-18\n JBT, NYSE, -17.65, 2013-05-08\n DL, NYSE, -33.33, 2013-05-22\n RAX, NYSE, -5.00, 2013-05-09\n GSL, NYSE, -50.00, 2013-05-10\n HCI, NYSE, 66.06, 2013-05-03\n EC, NYSE, -18.58, 2013-05-04\n CLW, NYSE, -98.08, 2013-04-25\n MJN, NYSE, -1.16, 2013-04-26\n EPC, NYSE, 39.53, 2013-05-02\n BPI, NYSE, 0.00, 2013-05-07\n RST, NYSE, 25.00, 2013-05-09\n DGI, NYSE, 22.22, 2013-05-08\n SWI, NYSE, 6.25, 2013-05-01\n CYS, NYSE, -45.16, 2013-04-18\n IVR, NYSE, 1.59, 2013-05-02\n BUD, NYSE, 50.65, 2013-05-01\n SLD, NYSE, -66.67, 2013-05-15\n PMT, NYSE, 11.11, 2013-04-24\n STWD, NYSE, -20.93, 2013-05-09\n CFN, NYSE, 11.32, 2013-05-10\n SPB, NYSE, 7.32, 2013-05-01\n ARI, NYSE, 33.33, 2013-05-02\n CLNY, NYSE, -26.47, 2013-05-07\n ART, NYSE, -800.00, 2013-05-07\n SEM, NYSE, -11.11, 2013-05-03\n BSBR, NYSE, -71.43, 2013-04-26\n DOLE, NYSE, -50.00, 2013-05-03\n VSI, NYSE, 2.86, 2013-05-08\n TWO, NYSE, -9.38, 2013-05-08\n CVE, NYSE, -6.38, 2013-04-25\n H, NYSE, 12.50, 2013-05-02\n LEA, NYSE, 19.27, 2013-04-26\n SVN, NYSE, -81.82, 2013-05-14\n CLD, NYSE, -59.26, 2013-05-01\n AOL, NYSE, 6.25, 2013-05-09\n CHSP, NYSE, 25.00, 2013-05-08\n PEB, NYSE, 5.88, 2013-04-26\n CIT, NYSE, -8.99, 2013-04-24\n KAR, NYSE, -3.03, 2013-05-02\n CIE, NYSE, -15.38, 2013-05-01\n TMH, NYSE, 0.00, 2013-05-01\n KRA, NYSE, -75.00, 2013-05-02\n SYA, NYSE, 8.82, 2013-04-25\n TRNO, NYSE, -11.11, 2013-05-09\n PDM, NYSE, 0.00, 2013-05-03\n GNRC, NYSE, 23.47, 2013-05-03\n ACW, NYSE, -9.68, 2013-04-24\n BALT, NYSE, -9.52, 2013-05-02\n ST, NYSE, 4.35, 2013-04-24\n SEMG, NYSE, -15.00, 2013-05-09\n CALX, NYSE, 50.00, 2013-04-26\n MXL, NYSE, 33.33, 2013-05-01\n STNG, NYSE, 60.00, 2013-04-30\n PRI, NYSE, -4.35, 2013-05-08\n SDRL, NYSE, 16.95, 2013-05-29\n CLDT, NYSE, 7.50, 2013-05-08\n EXL, NYSE, 5.00, 2013-05-02\n LYB, NYSE, 9.09, 2013-04-27\n PNG, NYSE, 4.35, 2013-05-07\n PLOW, NYSE, 13.33, 2013-05-07\n SIX, NYSE, 19.61, 2013-04-23\n NKA, NYSE, -140.00, 2013-05-10\n RRTS, NYSE, 3.57, 2013-05-02\n JKS, NYSE, 66.27, 2013-06-08\n CODE, NYSE, 7.69, 2013-05-01\n FAF, NYSE, -31.71, 2013-04-26\n QEP, NYSE, -6.67, 2013-05-01\n OAS, NYSE, 31.37, 2013-05-08\n HPP, NYSE, 18.18, 2013-05-07\n FN, NYSE, 3.70, 2013-04-30\n ECT, NYSE, 7.32, 2013-05-11\n QUAD, NYSE, -88.10, 2013-05-08\n KKR, NYSE, 4.76, 2013-04-26\n RLD, NYSE, 70.00, 2013-06-07\n AMRC, NYSE, -200.00, 2013-05-10\n GDOT, NYSE, 9.37, 2013-05-01\n AT, NYSE, 40.00, 2013-05-09\n ENV, NYSE, 0.00, 2013-05-17\n COR, NYSE, 0.00, 2013-04-25\n VC, NYSE, 75.65, 2013-05-10\n CCG, NYSE, 5.88, 2013-05-01\n EFC, NYSE, -32.00, 2013-05-07\n TOWR, NYSE, 255.56, 2013-05-03\n CHMT, NYSE, -21.05, 2013-05-03\n HBM, NYSE, 200.00, 2013-05-02\n EXAM, NYSE, 0.00, 2013-05-09\n XUE, NYSE, -25.00, 2013-05-17\n CMRE, NYSE, 26.09, 2013-04-25\n NOAH, NYSE, 112.50, 2013-05-07\n IPHI, NYSE, 18.18, 2013-05-02\n BITA, NYSE, 0.00, 2013-05-10\n BAH, NYSE, 11.43, 2013-05-23\n GM, NYSE, 19.64, 2013-05-03\n XNY, NYSE, 28.57, 2013-05-20\n TROX, NYSE, -181.25, 2013-05-09\n TRGP, NYSE, 52.38, 2013-05-04\n DANG, NYSE, 21.05, 2013-05-17\n YOKU, NYSE, 0.00, 2013-05-16\n FRC, NYSE, 0.00, 2013-04-16\n RFP, NYSE, 64.29, 2013-05-01\n ISS, NYSE, 50.00, 2013-05-18\n WD, NYSE, -45.65, 2013-05-09\n FLT, NYSE, 10.39, 2013-05-03\n GCAP, NYSE, -15.38, 2013-05-08\n FRF, NYSE, -27.27, 2013-05-14\n SWFT, NYSE, 23.53, 2013-04-23\n AG, NYSE, -8.00, 2013-05-16\n QRE, NYSE, 0.00, 2013-05-09\n AAT, NYSE, 8.57, 2013-05-01\n MCC, NYSE, -2.70, 2013-05-03\n NLSN, NYSE, 9.09, 2013-04-26\n AGRO, NYSE, -100.00, 2013-05-17\n BKU, NYSE, 4.44, 2013-04-25\n INXN, NYSE, -7.14, 2013-05-09\n NPTN, NYSE, 10.00, 2013-05-10\n INN, NYSE, 5.88, 2013-05-07\n KMI, NYSE, -12.50, 2013-04-18\n HCA, NYSE, -4.82, 2013-05-03\n MX, NYSE, 13.04, 2013-05-01\n HII, NYSE, 0.00, 2013-05-09\n QIHU, NYSE, 100.00, 2013-05-20\n APO, NYSE, 56.20, 2013-05-07\n GNC, NYSE, 1.39, 2013-04-27\n SDT, NYSE, 16.07, 2013-05-11\n UAN, NYSE, 4.26, 2013-05-02\n ARCO, NYSE, -142.86, 2013-05-01\n ELLI, NYSE, -16.67, 2013-05-01\n TMS, NYSE, -12.00, 2013-04-26\n SQNS, NYSE, 0.00, 2013-04-26\n STAG, NYSE, 3.13, 2013-05-07\n AL, NYSE, 5.13, 2013-05-10\n TLLP, NYSE, -14.89, 2013-05-07\n RENN, NYSE, 85.71, 2013-05-14\n NQ, NYSE, -16.67, 2013-05-16\n KOS, NYSE, -37.50, 2013-05-10\n RLJ, NYSE, 10.81, 2013-05-09\n NGL, NYSE, -62.86, 2013-06-15\n FENG, NYSE, 60.00, 2013-05-15\n LNKD, NYSE, 340.00, 2013-05-03\n NMFC, NYSE, -2.86, 2013-05-07\n ACTV, NYSE, 32.14, 2013-05-03\n FIO, NYSE, 20.00, 2013-04-25\n TAOM, NYSE, -25.00, 2013-05-24\n RATE, NYSE, 10.00, 2013-05-01\n VHS, NYSE, 8.33, 2013-05-01\n MPC, NYSE, 0.00, 2013-05-01\n MITT, NYSE, -9.64, 2013-05-07\n OILT, NYSE, 17.07, 2013-05-09\n SXC, NYSE, -40.00, 2013-04-26\n AMTG, NYSE, 14.06, 2013-05-07\n AMID, NYSE, -200.00, 2013-05-14\n WAIR, NYSE, 22.22, 2013-04-30\n PER, NYSE, -7.58, 2013-05-11\n PPP, NYSE, 260.00, 2013-05-09\n FSM, NYSE, -28.57, 2013-05-08\n FBHS, NYSE, 41.18, 2013-05-03\n XLS, NYSE, 73.91, 2013-05-04\n XYL, NYSE, -3.57, 2013-05-01\n GNE, NYSE, -550.00, 2013-05-08\n NDRO, NYSE, -8.11, 2013-05-04\n RNF, NYSE, -29.63, 2013-05-10\n VAC, NYSE, 10.20, 2013-04-26\n CHKR, NYSE, -2.90, 2013-05-10\n PACD, NYSE, 250.00, 2013-05-07\n INVN, NYSE, -13.33, 2013-05-03\n DLPH, NYSE, 11.46, 2013-05-02\n MN, NYSE, 0.00, 2013-05-02\n RRMS, NYSE, 51.28, 2013-05-10\n WPX, NYSE, -4.17, 2013-05-03\n LPI, NYSE, -15.38, 2013-05-10\n SN, NYSE, -82.61, 2013-05-08\n KORS, NYSE, 35.14, 2013-05-30\n BCEI, NYSE, -20.93, 2013-05-10\n BOXC, NYSE, 2.56, 2013-04-23\n PVG, NYSE, -25.00, 2013-05-11\n POST, NYSE, -29.63, 2013-05-14\n SLCA, NYSE, -2.78, 2013-05-01\n MTDR, NYSE, 0.00, 2013-05-09\n GWAY, NYSE, -120.00, 2013-05-07\n EPAM, NYSE, -14.71, 2013-05-09\n RNDY, NYSE, -9.52, 2013-05-10\n PRLB, NYSE, 0.00, 2013-04-26\n YELP, NYSE, -40.00, 2013-05-02\n NSM, NYSE, 23.19, 2013-05-08\n ALSN, NYSE, 95.24, 2013-04-30\n DWRE, NYSE, -22.73, 2013-05-08\n VNTV, NYSE, 3.70, 2013-05-07\n ET, NYSE, 0.00, 2013-05-10\n VCRA, NYSE, -160.00, 2013-05-03\n RM, NYSE, -1.82, 2013-05-03\n BNNY, NYSE, 3.57, 2013-06-11\n MM, NYSE, 25.00, 2013-05-09\n RXN, NYSE, 0.00, 2013-05-22\n GLOG, NYSE, -16.67, 2013-05-16\n RPAI, NYSE, 9.52, 2013-05-07\n OAK, NYSE, 39.86, 2013-05-08\n FET, NYSE, 3.03, 2013-04-26\n MRC, NYSE, 4.65, 2013-05-03\n PSX, NYSE, 17.74, 2013-05-02\n TUMI, NYSE, 6.67, 2013-05-09\n ACRE, NYSE, -5.88, 2013-05-16\n EVER, NYSE, 13.79, 2013-04-25\n PDH, NYSE, -13.24, 2013-04-25\n ROYT, NYSE, 10.00, 2013-05-11\n WMC, NYSE, -2.15, 2013-05-16\n WAGE, NYSE, 35.71, 2013-05-10\n HTA, NYSE, 6.67, 2013-05-08\n ALEX, NYSE, -28.57, 2013-05-10\n BKW, NYSE, 0.00, 2013-04-27\n CNCO, NYSE, -88.24, 2013-05-31\n EQM, NYSE, 41.30, 2013-04-26\n NOW, NYSE, 0.00, 2013-04-25\n EGL, NYSE, -11.24, 2013-05-14\n NGVC, NYSE, 7.69, 2013-05-10\n NTI, NYSE, 3.51, 2013-05-14\n AMRE, NYSE, 4.00, 2013-05-08\n GMED, NYSE, 5.00, 2013-05-03\n MANU, NYSE, -25.00, 2013-05-03\n HCLP, NYSE, -23.08, 2013-05-15\n ADT, NYSE, -4.65, 2013-05-02\n TRLA, NYSE, -75.00, 2013-05-01\n SRC, NYSE, 19.44, 2013-05-09\n NBHC, NYSE, -50.00, 2013-04-30\n BSMX, NYSE, 30.43, 2013-04-27\n HY, NYSE, 67.05, 2013-05-02\n SMLP, NYSE, -10.71, 2013-05-14\n DYN, NYSE, -254.55, 2013-05-03\n LXFR, NYSE, 0.00, 2013-05-08\n LOCK, NYSE, 25.00, 2013-05-02\n JMI, NYSE, 224.44, 2013-05-08\n BERY, NYSE, 16.67, 2013-05-03\n FLTX, NYSE, 8.33, 2013-05-09\n ANFI, NYSE, 0.00, 2013-06-11\n SSTK, NYSE, 23.08, 2013-05-09\n RLGY, NYSE, -13.33, 2013-05-02\n SDLP, NYSE, 88.64, 2013-05-29\n MPLX, NYSE, -7.14, 2013-05-01\n WWAV, NYSE, 6.67, 2013-05-10\n SXE, NYSE, -44.44, 2013-05-09\n DKL, NYSE, 31.58, 2013-05-08\n SCM, NYSE, -8.82, 2013-05-10\n RKUS, NYSE, -100.00, 2013-05-07\n ALDW, NYSE, -1.32, 2013-05-08\n WGP, NYSE, 0.00, 2013-05-02\n ABBV, NYSE, 3.03, 2013-04-27\n PBF, NYSE, -54.72, 2013-05-03\n SBY, NYSE, -433.33, 2013-05-14\n RIOM, NYSE, 0.00, 2013-05-15\n USAC, NYSE, -30.00, 2013-05-10\n CVRR, NYSE, -2.56, 2013-05-03\n SXCP, NYSE, -9.76, 2013-04-26\n BFAM, NYSE, 81.82, 2013-05-10\n TPH, NYSE, 200.00, 2013-05-15\n ZTS, NYSE, 5.88, 2013-05-01\n BCC, NYSE, 146.15, 2013-04-23\n AGI, NYSE, 0.00, 2013-04-26\n APAM, NYSE, -11.32, 2013-05-02\n SSNI, NYSE, -1211.77, 2013-05-02\n MODN, NYSE, 0.00, 2013-05-08\n AVIV, NYSE, 150.00, 2013-05-08\n OAKS, NYSE, 509.09, 2013-05-04\n MRIN, NYSE, -7.50, 2013-05-09\n PF, NYSE, 17.24, 2013-05-16\n TMHC, NYSE, -66.67, 2013-05-16\n ARPI, NYSE, -600.00, 2013-06-25\n CSTM, NYSE, -105.08, 2013-06-18\n DDC, NYSE, -80.00, 2013-06-06\n ABM, NYSE, 9.09, 2013-06-04\n ANN, NYSE, 4.76, 2013-06-07\n BBY, NYSE, 28.00, 2013-05-22\n BF.B, NYSE, -2.17, 2013-06-06\n BKE, NYSE, -4.88, 2013-05-24\n NCS, NYSE, -21.74, 2013-06-05\n BNS, NYSE, -0.83, 2013-05-29\n BRC, NYSE, -6.78, 2013-05-17\n CATO, NYSE, 1.94, 2013-05-24\n COO, NYSE, 9.49, 2013-06-07\n CPB, NYSE, 10.71, 2013-05-21\n CFI, NYSE, 10.81, 2013-06-13\n DCI, NYSE, -4.17, 2013-05-18\n DDS, NYSE, 15.38, 2013-05-15\n DE, NYSE, 0.73, 2013-05-16\n DY, NYSE, 0.00, 2013-05-22\n EV, NYSE, 0.00, 2013-05-23\n ESL, NYSE, -11.81, 2013-05-31\n M, NYSE, 3.77, 2013-05-16\n GCO, NYSE, 11.90, 2013-06-01\n GPS, NYSE, 2.90, 2013-05-24\n HD, NYSE, 7.79, 2013-05-22\n HEI, NYSE, 10.00, 2013-05-23\n HOV, NYSE, 120.00, 2013-06-06\n HRB, NYSE, -1.93, 2013-06-13\n HRL, NYSE, 0.00, 2013-05-24\n HPQ, NYSE, 7.41, 2013-05-23\n JCP, NYSE, -12.93, 2013-05-17\n KR, NYSE, 4.55, 2013-06-21\n KSS, NYSE, 15.79, 2013-05-17\n LB, NYSE, 4.35, 2013-05-23\n LOW, NYSE, -3.92, 2013-05-23\n LZB, NYSE, 7.14, 2013-06-19\n MDT, NYSE, 6.80, 2013-05-22\n MEI, NYSE, 60.00, 2013-06-21\n MPR, NYSE, -33.33, 2013-06-07\n NAV, NYSE, -302.75, 2013-06-11\n JWN, NYSE, -3.95, 2013-05-17\n OXM, NYSE, 5.13, 2013-06-12\n PBY, NYSE, -85.71, 2013-06-11\n PLL, NYSE, 1.37, 2013-05-31\n PNY, NYSE, 0.00, 2013-06-08\n PVH, NYSE, 39.42, 2013-06-13\n THO, NYSE, -7.87, 2013-06-07\n TIF, NYSE, 32.08, 2013-05-29\n TJX, NYSE, 0.00, 2013-05-22\n TOL, NYSE, 0.00, 2013-05-23\n TTC, NYSE, 10.92, 2013-05-24\n VAL, NYSE, 2.25, 2013-05-15\n JW.A, NYSE, -16.47, 2013-06-19\n TGT, NYSE, 23.53, 2013-05-23\n WMT, NYSE, -0.87, 2013-05-17\n WSM, NYSE, 11.11, 2013-05-24\n FL, NYSE, 3.41, 2013-05-25\n CHS, NYSE, -11.11, 2013-05-30\n BKS, NYSE, 52.22, 2013-06-26\n CAL, NYSE, 45.45, 2013-05-30\n SIG, NYSE, 0.89, 2013-05-24\n ZLC, NYSE, 1200.00, 2013-05-23\n AEO, NYSE, 5.88, 2013-05-23\n FGP, NYSE, 15.69, 2013-06-07\n BMO, NYSE, -4.73, 2013-05-30\n RY, NYSE, -2.34, 2013-05-31\n GEF, NYSE, 1.45, 2013-06-06\n SKS, NYSE, 0.00, 2013-05-22\n TD, NYSE, 1.09, 2013-05-24\n ANF, NYSE, -80.00, 2013-05-25\n CIEN, NYSE, 20.00, 2013-06-07\n KMG, NYSE, 8.70, 2013-06-11\n IRET, NYSE, 11.76, 2013-07-02\n CM, NYSE, 0.00, 2013-05-31\n UBA, NYSE, 12.00, 2013-06-08\n KFY, NYSE, 3.23, 2013-06-18\n KKD, NYSE, 25.00, 2013-05-31\n MVC, NYSE, -37.50, 2013-06-11\n CBK, NYSE, 150.00, 2013-06-08\n SJM, NYSE, 12.17, 2013-06-07\n BIG, NYSE, 0.00, 2013-05-31\n JOY, NYSE, 11.61, 2013-05-31\n SSI, NYSE, -122.22, 2013-05-18\n GME, NYSE, 15.00, 2013-05-24\n DKS, NYSE, 0.00, 2013-05-22\n A, NYSE, 14.93, 2013-05-15\n MTN, NYSE, -3.62, 2013-06-07\n GES, NYSE, 75.00, 2013-05-31\n CRM, NYSE, -600.00, 2013-05-24\n NWY, NYSE, 128.57, 2013-05-24\n PAY, NYSE, -7.69, 2013-06-06\n DSW, NYSE, 11.11, 2013-05-30\n NX, NYSE, -300.00, 2013-06-08\n DG, NYSE, -1.39, 2013-06-05\n EXPR, NYSE, 5.56, 2013-05-31\n P, NYSE, 0.00, 2013-05-23\n GWRE, NYSE, 44.44, 2013-05-29\n BLOX, NYSE, 100.00, 2013-05-24\n TLYS, NYSE, 14.29, 2013-05-30\n PANW, NYSE, -900.00, 2013-05-31\n WDAY, NYSE, 13.04, 2013-05-23\n RH, NYSE, 50.00, 2013-06-14\n RALY, NYSE, 14.78, 2013-06-07\n AIR, NYSE, 13.64, 2013-07-26\n ATU, NYSE, -1.59, 2013-06-20\n AZO, NYSE, 0.69, 2013-05-22\n AZZ, NYSE, -8.20, 2013-06-29\n CAG, NYSE, 1.69, 2013-06-28\n CLC, NYSE, -1.49, 2013-06-20\n CMC, NYSE, -15.79, 2013-06-28\n FC, NYSE, 18.18, 2013-07-10\n FDO, NYSE, 1.94, 2013-07-11\n FDX, NYSE, 8.67, 2013-06-20\n FUL, NYSE, -5.63, 2013-06-27\n GIS, NYSE, -1.85, 2013-06-27\n KBH, NYSE, 20.00, 2013-06-28\n LEN, NYSE, 30.30, 2013-06-26\n LNN, NYSE, 12.92, 2013-06-27\n MKC, NYSE, 0.00, 2013-06-28\n RT, NYSE, -36.84, 2013-07-25\n MCS, NYSE, -6.25, 2013-07-26\n MSM, NYSE, 9.37, 2013-07-11\n NKE, NYSE, 2.70, 2013-06-28\n ORCL, NYSE, 0.00, 2013-06-21\n PIR, NYSE, 0.00, 2013-06-21\n PKE, NYSE, -13.79, 2013-06-27\n RAD, NYSE, 0.00, 2013-06-21\n RPM, NYSE, 7.46, 2013-07-23\n SVU, NYSE, 250.00, 2013-07-19\n TISI, NYSE, 0.00, 2013-08-07\n TXI, NYSE, 116.00, 2013-07-11\n UNF, NYSE, 2.88, 2013-06-27\n WGO, NYSE, 0.00, 2013-06-28\n WOR, NYSE, -7.46, 2013-06-28\n JBL, NYSE, 4.35, 2013-06-20\n GBX, NYSE, -5.66, 2013-07-03\n DRI, NYSE, -1.94, 2013-06-22\n FDS, NYSE, -1.71, 2013-06-19\n KMX, NYSE, 12.28, 2013-06-22\n SCS, NYSE, 0.00, 2013-06-20\n SJR, NYSE, 16.28, 2013-06-29\n RHT, NYSE, 9.09, 2013-06-20\n OMN, NYSE, 14.29, 2013-06-28\n MON, NYSE, 3.75, 2013-06-27\n GPN, NYSE, -3.92, 2013-07-26\n AYI, NYSE, 7.78, 2013-07-03\n CCL, NYSE, 50.00, 2013-06-26\n CUK, NYSE, 50.00, 2013-06-26\n STZ, NYSE, -7.32, 2013-07-03\n ACN, NYSE, 0.00, 2013-06-28\n SNX, NYSE, 0.00, 2013-06-26\n TAL, NYSE, 66.67, 2013-07-23\n IHS, NYSE, 1.45, 2013-06-21\n EDU, NYSE, 20.00, 2013-07-24\n ZEP, NYSE, -31.71, 2013-07-03\n MG, NYSE, -5.88, 2013-08-08\n MOS, NYSE, -0.88, 2013-07-16\n ABT, NYSE, 4.55, 2013-07-18\n ABX, NYSE, 17.86, 2013-08-02\n AB, NYSE, 7.89, 2013-08-01\n TAP, NYSE, 8.63, 2013-08-07\n ACO, NYSE, 1.79, 2013-07-27\n ADM, NYSE, 9.52, 2013-08-07\n AEM, NYSE, -85.71, 2013-07-25\n AEP, NYSE, -5.19, 2013-07-26\n AES, NYSE, 23.08, 2013-08-09\n AET, NYSE, 9.35, 2013-07-31\n AFL, NYSE, 6.58, 2013-07-31\n AGCO, NYSE, 18.78, 2013-08-01\n AGN, NYSE, 1.01, 2013-07-26\n HES, NYSE, 7.09, 2013-08-01\n AIG, NYSE, 31.76, 2013-08-02\n AIN, NYSE, -23.08, 2013-08-01\n AJG, NYSE, 5.80, 2013-07-31\n ALU, NYSE, 33.33, 2013-07-31\n MATX, NYSE, 6.82, 2013-08-08\n ALK, NYSE, -0.68, 2013-07-26\n BEAM, NYSE, 6.67, 2013-08-09\n AME, NYSE, 0.00, 2013-08-08\n TWX, NYSE, 10.67, 2013-08-08\n AVD, NYSE, -17.14, 2013-08-06\n AMN, NYSE, 20.00, 2013-08-02\n AN, NYSE, -1.35, 2013-07-19\n AON, NYSE, 0.91, 2013-07-27\n APA, NYSE, -0.50, 2013-08-02\n APC, NYSE, 16.67, 2013-07-30\n APD, NYSE, 0.00, 2013-07-24\n APH, NYSE, 1.06, 2013-07-19\n ARG, NYSE, -0.87, 2013-07-26\n AAN, NYSE, 0.00, 2013-07-25\n ARW, NYSE, 8.74, 2013-07-25\n ASGN, NYSE, 14.29, 2013-07-25\n ASH, NYSE, -8.29, 2013-07-26\n ASR, NYSE, 21.90, 2013-07-23\n GAS, NYSE, 51.85, 2013-08-01\n ATO, NYSE, 13.51, 2013-08-07\n ATW, NYSE, 0.74, 2013-08-01\n AVP, NYSE, 11.54, 2013-08-02\n AVT, NYSE, 3.16, 2013-08-08\n AVY, NYSE, 2.90, 2013-07-24\n AXP, NYSE, 4.96, 2013-07-18\n B, NYSE, 0.00, 2013-07-27\n BA, NYSE, 5.70, 2013-07-25\n BAC, NYSE, 28.00, 2013-07-18\n BAX, NYSE, 2.65, 2013-07-19\n BC, NYSE, 13.89, 2013-07-26\n OMX, NYSE, -33.33, 2013-08-07\n BCE, NYSE, -2.67, 2013-08-09\n BCR, NYSE, 2.90, 2013-07-24\n BDX, NYSE, 7.48, 2013-08-02\n BEN, NYSE, 1.18, 2013-07-30\n BGG, NYSE, 15.79, 2013-08-16\n BHE, NYSE, 10.71, 2013-07-26\n BHI, NYSE, -6.15, 2013-07-20\n BID, NYSE, -9.56, 2013-08-07\n BIO, NYSE, 7.14, 2013-08-07\n BK, NYSE, 6.90, 2013-07-18\n BKH, NYSE, -2.38, 2013-08-06\n WRB, NYSE, -2.99, 2013-07-23\n BLC, NYSE, 9.09, 2013-07-31\n BLL, NYSE, 1.19, 2013-07-26\n BLX, NYSE, 5.56, 2013-07-19\n BMI, NYSE, -20.00, 2013-07-19\n BMS, NYSE, 1.67, 2013-07-26\n BMY, NYSE, 0.00, 2013-07-26\n BOH, NYSE, 2.41, 2013-07-23\n BXS, NYSE, 10.00, 2013-07-23\n BPL, NYSE, -8.86, 2013-08-03\nBRK.A, NYSE, 176.30, 2013-08-03\n BRO, NYSE, 2.86, 2013-07-16\n BSX, NYSE, 12.50, 2013-07-26\n BT, NYSE, 6.17, 2013-07-26\n MTRN, NYSE, 7.50, 2013-07-27\n CAI, NYSE, -8.54, 2013-07-31\n CAT, NYSE, -15.20, 2013-07-25\n CB, NYSE, 19.27, 2013-07-24\n CBI, NYSE, 0.00, 2013-07-31\n CBM, NYSE, -64.29, 2013-08-02\n CBU, NYSE, 4.00, 2013-07-24\n CBT, NYSE, -4.35, 2013-08-01\n CCC, NYSE, 14.29, 2013-08-07\n CCE, NYSE, 2.67, 2013-07-26\n C, NYSE, 5.93, 2013-07-16\n CCK, NYSE, 3.23, 2013-07-18\n CCU, NYSE, 25.00, 2013-08-08\n CDE, NYSE, -1100.00, 2013-08-09\n CDI, NYSE, 6.25, 2013-08-02\n CAH, NYSE, 2.60, 2013-08-02\n CFR, NYSE, 0.00, 2013-07-25\n CHD, NYSE, 1.67, 2013-08-03\n CKP, NYSE, -15.38, 2013-08-07\n CPK, NYSE, -7.02, 2013-08-10\n CI, NYSE, 11.95, 2013-08-02\n CKH, NYSE, 51.67, 2013-07-31\n CL, NYSE, 0.00, 2013-07-26\n CLF, NYSE, 85.25, 2013-07-26\n CLH, NYSE, -25.00, 2013-08-08\n CLX, NYSE, 2.99, 2013-08-02\n CMA, NYSE, 8.57, 2013-07-17\n CMO, NYSE, -15.63, 2013-07-25\n CRK, NYSE, -6.67, 2013-07-30\n CMS, NYSE, -14.71, 2013-07-26\n CNA, NYSE, 17.19, 2013-07-31\n CNW, NYSE, 13.56, 2013-08-01\n CNL, NYSE, -6.06, 2013-08-01\n COG, NYSE, 35.48, 2013-07-25\n COT, NYSE, -4.76, 2013-08-02\n CP, NYSE, -4.14, 2013-07-25\n CPF, NYSE, 25.93, 2013-07-26\n CQB, NYSE, 43.48, 2013-08-09\n CR, NYSE, 0.00, 2013-07-23\nCRD.B, NYSE, 42.86, 2013-08-06\n CRS, NYSE, 11.59, 2013-07-31\n CSC, NYSE, 42.19, 2013-08-07\n CSL, NYSE, -14.93, 2013-07-24\n CTB, NYSE, -38.20, 2013-08-09\n CTL, NYSE, 2.99, 2013-08-08\n CTS, NYSE, 33.33, 2013-07-23\n CUB, NYSE, 9.52, 2013-08-02\n CMI, NYSE, 11.11, 2013-07-31\n CUZ, NYSE, 9.09, 2013-07-30\n CVC, NYSE, 80.00, 2013-08-03\n CW, NYSE, 6.06, 2013-08-01\n CWT, NYSE, 0.00, 2013-08-01\n CX, NYSE, 0.00, 2013-07-26\n CYN, NYSE, 8.33, 2013-07-19\n D, NYSE, -4.62, 2013-08-07\n DBD, NYSE, 0.00, 2013-08-15\n DCO, NYSE, 30.77, 2013-08-06\n DD, NYSE, 0.79, 2013-07-24\n CVA, NYSE, 150.00, 2013-07-18\n DHR, NYSE, 2.35, 2013-07-19\n DIS, NYSE, 0.00, 2013-08-07\n DLX, NYSE, 10.34, 2013-07-26\n DNB, NYSE, 2.00, 2013-08-08\n RRD, NYSE, 4.65, 2013-07-30\n DOV, NYSE, 5.43, 2013-07-19\n DOW, NYSE, 1.59, 2013-07-26\n DRE, NYSE, 0.00, 2013-08-01\n DHI, NYSE, 23.53, 2013-07-26\n UFS, NYSE, -25.00, 2013-07-26\n DTE, NYSE, -21.52, 2013-07-27\n DUK, NYSE, -6.45, 2013-08-08\n DVN, NYSE, 28.72, 2013-08-08\n DV, NYSE, 31.71, 2013-08-09\n EAT, NYSE, 4.05, 2013-08-03\n ECL, NYSE, 2.38, 2013-07-31\n ED, NYSE, -5.26, 2013-08-02\n EDE, NYSE, 8.00, 2013-07-26\n EFX, NYSE, 2.22, 2013-07-25\n EGN, NYSE, 8.20, 2013-08-01\n EGP, NYSE, 2.56, 2013-07-19\n ELP, NYSE, 17.65, 2013-08-16\n ELY, NYSE, 20.00, 2013-07-26\n EMC, NYSE, 2.94, 2013-07-25\n EMR, NYSE, -2.02, 2013-08-07\n EOG, NYSE, 19.32, 2013-08-07\n EQT, NYSE, 3.64, 2013-07-26\n ESE, NYSE, -41.07, 2013-08-09\n ESV, NYSE, 3.33, 2013-07-30\n ETN, NYSE, -1.80, 2013-08-03\n ETR, NYSE, 3.06, 2013-07-31\n EXAR, NYSE, 14.29, 2013-07-25\n F, NYSE, 21.62, 2013-07-25\n CLGX, NYSE, 13.64, 2013-07-25\n FNB, NYSE, 0.00, 2013-07-24\n FCF, NYSE, -50.00, 2013-07-25\n FBP, NYSE, -11.11, 2013-07-25\n FICO, NYSE, 6.35, 2013-07-31\n FLO, NYSE, 4.35, 2013-08-14\n FMC, NYSE, 0.00, 2013-07-30\n FOE, NYSE, 27.27, 2013-08-01\n S, NYSE, 6.06, 2013-07-31\n NEE, NYSE, 13.18, 2013-07-31\n FRT, NYSE, 0.88, 2013-08-01\n FRX, NYSE, 300.00, 2013-07-24\n FSS, NYSE, 64.29, 2013-08-10\n FUN, NYSE, 2.41, 2013-08-09\n FUR, NYSE, -48.15, 2013-08-02\n GBL, NYSE, 17.20, 2013-08-07\n GVA, NYSE, -78.13, 2013-08-02\n BGC, NYSE, 23.21, 2013-08-01\n GD, NYSE, 11.73, 2013-07-25\n GE, NYSE, 0.00, 2013-07-20\n RHP, NYSE, -26.85, 2013-08-07\n AXLL, NYSE, 2.59, 2013-08-01\n GGG, NYSE, 9.52, 2013-07-25\n GHM, NYSE, 52.00, 2013-07-26\n GIB, NYSE, 10.71, 2013-08-01\n GLT, NYSE, 20.00, 2013-07-31\n GLW, NYSE, 3.23, 2013-07-31\n GSK, NYSE, -5.88, 2013-07-25\n GLF, NYSE, 25.71, 2013-07-23\n GPC, NYSE, 14.88, 2013-07-19\n GRA, NYSE, 2.75, 2013-07-26\n GTY, NYSE, 36.00, 2013-08-08\n GWW, NYSE, 2.71, 2013-07-18\n HAE, NYSE, 0.00, 2013-07-30\n HAL, NYSE, 1.39, 2013-07-23\n HAR, NYSE, 4.60, 2013-08-07\n HVT, NYSE, 31.25, 2013-08-01\n HRC, NYSE, 0.00, 2013-07-25\n HCC, NYSE, 21.69, 2013-07-31\n HCN, NYSE, 1.09, 2013-08-07\n HCP, NYSE, -2.70, 2013-07-31\n HOG, NYSE, 3.42, 2013-07-26\n HE, NYSE, 7.89, 2013-08-09\n HMA, NYSE, -46.15, 2013-08-10\n HMN, NYSE, 30.00, 2013-07-25\n HFC, NYSE, 0.00, 2013-08-08\n HOT, NYSE, 8.22, 2013-07-26\n HP, NYSE, 6.67, 2013-07-27\n HLS, NYSE, 18.60, 2013-07-26\n HRS, NYSE, 23.68, 2013-07-31\n HSC, NYSE, -11.76, 2013-08-09\n HSY, NYSE, 1.41, 2013-07-26\n HUBB, NYSE, 5.38, 2013-07-19\n HUM, NYSE, 6.91, 2013-08-01\n HXL, NYSE, 2.13, 2013-07-23\n IBM, NYSE, 3.44, 2013-07-18\n IDA, NYSE, 33.82, 2013-08-02\n IEX, NYSE, 2.70, 2013-07-23\n IFF, NYSE, -3.39, 2013-08-07\n DIN, NYSE, 12.09, 2013-07-31\n INT, NYSE, 11.76, 2013-08-01\n IP, NYSE, -5.45, 2013-07-26\n IPG, NYSE, -14.29, 2013-07-20\n IO, NYSE, -100.00, 2013-08-08\n IR, NYSE, 5.56, 2013-07-20\n IRF, NYSE, 81.82, 2013-08-20\n ITW, NYSE, -0.92, 2013-07-24\n JEC, NYSE, -1.19, 2013-07-30\n JNJ, NYSE, 5.71, 2013-07-17\n JNY, NYSE, 116.67, 2013-08-01\n K, NYSE, 3.09, 2013-08-02\n KAMN, NYSE, 13.56, 2013-07-30\n KDN, NYSE, 10.53, 2013-07-26\n KEX, NYSE, 0.94, 2013-07-25\n KEY, NYSE, 5.00, 2013-07-19\n KIM, NYSE, 6.06, 2013-07-30\n KMB, NYSE, 1.44, 2013-07-23\n KEM, NYSE, -95.00, 2013-07-26\n KMT, NYSE, 4.11, 2013-07-26\n KO, NYSE, 0.00, 2013-07-17\n KSU, NYSE, 1.05, 2013-07-20\n LDR, NYSE, -19.64, 2013-08-06\n LEG, NYSE, 0.00, 2013-07-26\n LLY, NYSE, 13.73, 2013-07-25\n LM, NYSE, -1.45, 2013-07-26\n LNC, NYSE, 10.43, 2013-08-01\n LPX, NYSE, 32.26, 2013-08-07\n LXU, NYSE, 29.17, 2013-08-09\n LTC, NYSE, -3.39, 2013-08-09\n L, NYSE, -5.48, 2013-07-30\n LUV, NYSE, -2.56, 2013-07-26\n LUX, NYSE, -1.67, 2013-07-26\n MKL, NYSE, 7.46, 2013-08-08\n MAN, NYSE, 17.98, 2013-07-20\n MTW, NYSE, 25.00, 2013-07-30\n SM, NYSE, 0.00, 2013-07-31\n MAS, NYSE, 21.05, 2013-07-30\n MTZ, NYSE, 2.33, 2013-08-02\n MCD, NYSE, -1.43, 2013-07-23\n MDC, NYSE, 38.18, 2013-07-31\n MDP, NYSE, 5.63, 2013-07-26\n MDR, NYSE, -1966.67, 2013-08-06\n MDU, NYSE, -3.85, 2013-08-01\n MED, NYSE, 2.00, 2013-08-07\n CVS, NYSE, 1.04, 2013-08-07\n MFC, NYSE, -3.12, 2013-08-09\n MGA, NYSE, 11.25, 2013-08-10\n MGM, NYSE, 300.00, 2013-08-07\n MMC, NYSE, 2.94, 2013-08-08\n MMM, NYSE, 0.59, 2013-07-26\n MSA, NYSE, 0.00, 2013-07-25\n MNR, NYSE, -27.78, 2013-08-07\n MO, NYSE, -1.59, 2013-07-24\n MOD, NYSE, 145.45, 2013-08-02\nMOG.A, NYSE, 8.43, 2013-07-27\n MHK, NYSE, 10.84, 2013-08-02\n MSI, NYSE, 11.96, 2013-07-25\n MCY, NYSE, 3.28, 2013-07-30\n MRK, NYSE, 2.44, 2013-07-31\n MRO, NYSE, -5.63, 2013-08-07\n POWR, NYSE, 20.00, 2013-08-08\n MTG, NYSE, 118.75, 2013-07-24\n MTB, NYSE, 26.19, 2013-07-18\n MTX, NYSE, 8.62, 2013-07-26\n MUR, NYSE, 12.90, 2013-08-01\n MYE, NYSE, 19.05, 2013-07-19\n NBL, NYSE, -5.48, 2013-07-26\n NBR, NYSE, -11.11, 2013-07-24\n NE, NYSE, 12.50, 2013-07-18\n NEM, NYSE, -124.39, 2013-07-27\n NFG, NYSE, 6.15, 2013-08-09\n NHI, NYSE, -1.14, 2013-08-07\n NI, NYSE, -4.17, 2013-08-01\n NJR, NYSE, 15.00, 2013-08-08\n THC, NYSE, -4.35, 2013-08-07\n NNN, NYSE, 0.00, 2013-08-02\n NOC, NYSE, 20.59, 2013-07-25\n NR, NYSE, -5.26, 2013-07-26\n NSC, NYSE, -2.67, 2013-07-24\n NUE, NYSE, -10.00, 2013-07-19\n NVR, NYSE, -18.34, 2013-07-23\n NWL, NYSE, 2.04, 2013-07-27\n NWN, NYSE, -11.11, 2013-08-08\n NYT, NYSE, 16.67, 2013-08-02\n OCR, NYSE, 4.65, 2013-07-25\n OGE, NYSE, -2.13, 2013-08-09\n OHI, NYSE, 1.64, 2013-08-01\n OI, NYSE, 2.53, 2013-07-25\n OII, NYSE, 8.33, 2013-07-25\n OKE, NYSE, -225.93, 2013-07-31\n OLN, NYSE, 3.85, 2013-07-26\n BRS, NYSE, 1.01, 2013-08-06\n OMC, NYSE, 0.00, 2013-07-19\n OMI, NYSE, 0.00, 2013-07-30\n ORB, NYSE, 17.39, 2013-07-19\n ORI, NYSE, 1750.00, 2013-07-26\n OSK, NYSE, 53.21, 2013-07-31\n OXY, NYSE, -1.86, 2013-07-31\n FCFS, NYSE, 1.79, 2013-07-18\n PBI, NYSE, 15.56, 2013-07-31\n PCG, NYSE, 9.72, 2013-08-01\n PCL, NYSE, 21.74, 2013-07-30\n PCP, NYSE, -0.69, 2013-07-26\n TPC, NYSE, -11.11, 2013-08-10\n PEG, NYSE, 4.35, 2013-07-31\n PEI, NYSE, 7.69, 2013-07-24\n PEP, NYSE, 10.08, 2013-07-25\n PFE, NYSE, 3.70, 2013-07-31\n PG, NYSE, 2.60, 2013-08-02\n PGR, NYSE, -2.44, 2013-07-12\n PH, NYSE, -8.72, 2013-08-07\n PHM, NYSE, -10.34, 2013-07-26\n PKD, NYSE, 0.00, 2013-08-07\n PKY, NYSE, 0.00, 2013-08-06\n PNC, NYSE, 21.34, 2013-07-18\n PNM, NYSE, 15.15, 2013-08-03\n PNR, NYSE, 2.22, 2013-07-24\n PNW, NYSE, 3.51, 2013-08-03\n POM, NYSE, -8.33, 2013-08-08\n POT, NYSE, -10.98, 2013-07-26\n PPG, NYSE, 4.70, 2013-07-19\n PPL, NYSE, 0.00, 2013-08-02' if __name__ == "__main__": main()
[ [ [ 7, 19 ], [ 135, 137 ] ], [ [ 46, 54 ], [ 147, 155 ] ], [ [ 62, 67 ], [ 68, 73 ], [ 101, 106 ] ], [ [ 118, 122 ], [ 129640, 129644 ] ], [ [ 681, 689 ], [ 156, 164 ] ] ]
# Copyright (C) 2013 eNovance SAS <[email protected]> # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import mock from oslo_config import cfg from oslo_utils import fixture as utils_fixture from oslo_utils import timeutils from oslo_utils import uuidutils from neutron.conf.services import metering_agent as metering_agent_config from neutron.services.metering.agents import metering_agent from neutron.tests import base from neutron.tests import fake_notifier _uuid = uuidutils.generate_uuid TENANT_ID = _uuid() LABEL_ID = _uuid() ROUTERS = [{'status': 'ACTIVE', 'name': 'router1', 'gw_port_id': None, 'admin_state_up': True, 'tenant_id': TENANT_ID, '_metering_labels': [{'rules': [], 'id': LABEL_ID}], 'id': _uuid()}] ROUTERS_WITH_RULE = [{'status': 'ACTIVE', 'name': 'router1', 'gw_port_id': None, 'admin_state_up': True, 'tenant_id': TENANT_ID, '_metering_labels': [{'rule': {}, 'id': LABEL_ID}], 'id': _uuid()}] class TestMeteringOperations(base.BaseTestCase): def setUp(self): super(TestMeteringOperations, self).setUp() metering_agent_config.register_metering_agent_opts() self.noop_driver = ('neutron.services.metering.drivers.noop.' 'noop_driver.NoopMeteringDriver') cfg.CONF.set_override('driver', 'noop') cfg.CONF.set_override('measure_interval', 0) cfg.CONF.set_override('report_interval', 0) self.setup_notification_driver() metering_rpc = ('neutron.services.metering.agents.metering_agent.' 'MeteringPluginRpc._get_sync_data_metering') self.metering_rpc_patch = mock.patch(metering_rpc, return_value=[]) self.metering_rpc_patch.start() self.driver_patch = mock.patch(self.noop_driver, spec=True) self.driver_patch.start() loopingcall_patch = mock.patch( 'oslo_service.loopingcall.FixedIntervalLoopingCall') loopingcall_patch.start() self.agent = metering_agent.MeteringAgent('my agent', cfg.CONF) self.driver = self.agent.metering_driver def test_add_metering_label(self): self.agent.add_metering_label(None, ROUTERS) self.assertEqual(1, self.driver.add_metering_label.call_count) def test_remove_metering_label(self): self.agent.remove_metering_label(None, ROUTERS) self.assertEqual(1, self.driver.remove_metering_label.call_count) def test_update_metering_label_rule(self): self.agent.update_metering_label_rules(None, ROUTERS) self.assertEqual(1, self.driver.update_metering_label_rules.call_count) def test_add_metering_label_rule(self): self.agent.add_metering_label_rule(None, ROUTERS_WITH_RULE) self.assertEqual(1, self.driver.add_metering_label_rule.call_count) def test_remove_metering_label_rule(self): self.agent.remove_metering_label_rule(None, ROUTERS_WITH_RULE) self.assertEqual(1, self.driver.remove_metering_label_rule.call_count) def test_routers_updated(self): self.agent.routers_updated(None, ROUTERS) self.assertEqual(1, self.driver.update_routers.call_count) def test_get_traffic_counters(self): self.agent._get_traffic_counters(None, ROUTERS) self.assertEqual(1, self.driver.get_traffic_counters.call_count) def test_sync_router_namespaces(self): self.agent._sync_router_namespaces(None, ROUTERS) self.assertEqual(1, self.driver.sync_router_namespaces.call_count) def test_notification_report(self): self.agent.routers_updated(None, ROUTERS) self.driver.get_traffic_counters.return_value = {LABEL_ID: {'pkts': 88, 'bytes': 444}} self.agent._metering_loop() self.assertNotEqual(len(fake_notifier.NOTIFICATIONS), 0) for n in fake_notifier.NOTIFICATIONS: if n['event_type'] == 'l3.meter': break self.assertEqual('l3.meter', n['event_type']) payload = n['payload'] self.assertEqual(TENANT_ID, payload['tenant_id']) self.assertEqual(LABEL_ID, payload['label_id']) self.assertEqual(88, payload['pkts']) self.assertEqual(444, payload['bytes']) def test_notification_report_interval(self): measure_interval = 30 report_interval = 600 now = timeutils.utcnow() time_fixture = self.useFixture(utils_fixture.TimeFixture(now)) self.agent.routers_updated(None, ROUTERS) self.driver.get_traffic_counters.return_value = {LABEL_ID: {'pkts': 889, 'bytes': 4440}} cfg.CONF.set_override('measure_interval', measure_interval) cfg.CONF.set_override('report_interval', report_interval) for i in range(report_interval): self.agent._metering_loop() count = 0 if len(fake_notifier.NOTIFICATIONS) > 1: for n in fake_notifier.NOTIFICATIONS: if n['event_type'] == 'l3.meter': # skip the first notification because the time is 0 count += 1 if count > 1: break time_fixture.advance_time_seconds(measure_interval) self.assertEqual('l3.meter', n['event_type']) payload = n['payload'] self.assertEqual(TENANT_ID, payload['tenant_id']) self.assertEqual(LABEL_ID, payload['label_id']) self.assertLess((payload['time'] - report_interval), measure_interval, payload) interval = (payload['last_update'] - payload['first_update']) \ - report_interval self.assertLess(interval, measure_interval, payload) def test_router_deleted(self): label_id = _uuid() self.driver.get_traffic_counters = mock.MagicMock() self.driver.get_traffic_counters.return_value = {label_id: {'pkts': 44, 'bytes': 222}} self.agent._add_metering_info = mock.MagicMock() self.agent.routers_updated(None, ROUTERS) self.agent.router_deleted(None, ROUTERS[0]['id']) self.assertEqual(1, self.agent._add_metering_info.call_count) self.assertEqual(1, self.driver.remove_router.call_count) self.agent._add_metering_info.assert_called_with(label_id, 44, 222) @mock.patch('time.time') def _test_purge_metering_info(self, current_timestamp, is_empty, mock_time): mock_time.return_value = current_timestamp self.agent.metering_infos = {'fake': {'last_update': 1}} self.config(report_interval=1) self.agent._purge_metering_info() self.assertEqual(0 if is_empty else 1, len(self.agent.metering_infos)) self.assertEqual(1, mock_time.call_count) def test_purge_metering_info(self): # 1 < 2 - 1 -> False self._test_purge_metering_info(2, False) def test_purge_metering_info_delete(self): # 1 < 3 - 1 -> False self._test_purge_metering_info(3, True) @mock.patch('time.time') def _test_add_metering_info(self, expected_info, current_timestamp, mock_time): mock_time.return_value = current_timestamp actual_info = self.agent._add_metering_info('fake_label_id', 1, 1) self.assertEqual(1, len(self.agent.metering_infos)) self.assertEqual(expected_info, actual_info) self.assertEqual(expected_info, self.agent.metering_infos['fake_label_id']) self.assertEqual(1, mock_time.call_count) def test_add_metering_info_create(self): expected_info = {'bytes': 1, 'pkts': 1, 'time': 0, 'first_update': 1, 'last_update': 1} self._test_add_metering_info(expected_info, 1) def test_add_metering_info_update(self): expected_info = {'bytes': 1, 'pkts': 1, 'time': 0, 'first_update': 1, 'last_update': 1} self.agent.metering_infos = {'fake_label_id': expected_info} expected_info.update({'bytes': 2, 'pkts': 2, 'time': 1, 'last_update': 2}) self._test_add_metering_info(expected_info, 2) def test_metering_agent_host_value(self): expected_host = 'my agent' self.assertEqual(expected_host, self.agent.host) class TestMeteringDriver(base.BaseTestCase): def setUp(self): super(TestMeteringDriver, self).setUp() metering_agent_config.register_metering_agent_opts() cfg.CONF.set_override('driver', 'noop') self.agent = metering_agent.MeteringAgent('my agent', cfg.CONF) self.driver = mock.Mock() self.agent.metering_driver = self.driver def test_add_metering_label_with_bad_driver_impl(self): del self.driver.add_metering_label with mock.patch.object(metering_agent, 'LOG') as log: self.agent.add_metering_label(None, ROUTERS) log.exception.assert_called_with(mock.ANY, {'driver': 'noop', 'func': 'add_metering_label'}) def test_add_metering_label_runtime_error(self): self.driver.add_metering_label.side_effect = RuntimeError with mock.patch.object(metering_agent, 'LOG') as log: self.agent.add_metering_label(None, ROUTERS) log.exception.assert_called_with(mock.ANY, {'driver': 'noop', 'func': 'add_metering_label'}) def test_init_chain(self): with mock.patch('oslo_service.' 'periodic_task.PeriodicTasks.__init__') as init: metering_agent.MeteringAgent('my agent', cfg.CONF) init.assert_called_once_with(cfg.CONF)
[ [ [ 614, 618 ], [ 7400, 7404 ], [ 8116, 8120 ], [ 2401, 2405 ], [ 2512, 2516 ], [ 2615, 2619 ], [ 6787, 6791 ], [ 7054, 7058 ], [ 9740, 9744 ], [ 9919, 9923 ], [ 10070, 10074 ], [ 10355, 10359 ], [ 10506, 10510 ], [ 10748, 10752 ] ], [ [ 643, 646 ], [ 2035, 2038 ], [ 2083, 2086 ], [ 2136, 2139 ], [ 2789, 2792 ], [ 5565, 5568 ], [ 5633, 5636 ], [ 9605, 9608 ], [ 9708, 9711 ], [ 10901, 10904 ], [ 10948, 10951 ] ], [ [ 670, 694 ], [ 5260, 5273 ] ], [ [ 718, 727 ], [ 5202, 5211 ] ], [ [ 751, 760 ], [ 976, 985 ] ], [ [ 796, 835 ], [ 1841, 1862 ], [ 9543, 9564 ] ], [ [ 881, 895 ], [ 2748, 2762 ], [ 9667, 9681 ], [ 9937, 9951 ], [ 10373, 10387 ], [ 10860, 10874 ] ], [ [ 922, 926 ], [ 1739, 1743 ], [ 9446, 9450 ] ], [ [ 953, 966 ], [ 4635, 4648 ], [ 4685, 4698 ], [ 5815, 5828 ], [ 5874, 5887 ] ], [ [ 968, 973 ], [ 1013, 1018 ], [ 1032, 1037 ], [ 1324, 1329 ], [ 1698, 1703 ], [ 6736, 6741 ] ], [ [ 1001, 1010 ], [ 1196, 1205 ], [ 1541, 1550 ], [ 4894, 4903 ], [ 6317, 6326 ] ], [ [ 1021, 1029 ], [ 1294, 1302 ], [ 1658, 1666 ], [ 4413, 4421 ], [ 4952, 4960 ], [ 5401, 5409 ], [ 6375, 6383 ] ], [ [ 1040, 1047 ], [ 2932, 2939 ], [ 3102, 3109 ], [ 3286, 3293 ], [ 3840, 3847 ], [ 4005, 4012 ], [ 4180, 4187 ], [ 4346, 4353 ], [ 5334, 5341 ], [ 7113, 7120 ], [ 7162, 7169 ], [ 10016, 10023 ], [ 10452, 10459 ] ], [ [ 1335, 1352 ], [ 3469, 3486 ], [ 3664, 3681 ] ], [ [ 1716, 1738 ], [ 1795, 1817 ] ], [ [ 9427, 9445 ], [ 9501, 9519 ] ] ]
from PIL import ImageChops, Image as PILImage from http.client import HTTPConnection from time import sleep from traceback import format_stack, print_exc def Tint(image, color): return ImageChops.blend(image, PILImage.new('RGB', image.size, color), 0.36) def GetStatusCode(host, path="/"): """ This function retreives the status code of a website by requesting HEAD data from the host. This means that it only requests the headers. If the host cannot be reached or something else goes wrong, it returns None instead. """ try: conn = HTTPConnection(host) conn.request("HEAD", path) return conn.getresponse().status except Exception: return None def WaitOK(host, path="/"): while GetStatusCode(host, path) != 200: sleep(5)
[ [ [ 16, 26 ], [ 191, 201 ] ], [ [ 28, 45 ], [ 215, 223 ] ], [ [ 70, 84 ], [ 584, 598 ] ], [ [ 102, 107 ], [ 808, 813 ] ], [ [ 130, 142 ] ], [ [ 144, 153 ] ], [ [ 160, 164 ] ], [ [ 266, 279 ], [ 766, 779 ] ], [ [ 732, 738 ] ] ]
# Copyright 2016 Quantopian, Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from abc import ( ABCMeta, abstractmethod, abstractproperty, ) from numpy import concatenate from lru import LRU from pandas import isnull from pandas.tslib import normalize_date from toolz import sliding_window from six import with_metaclass from zipline.assets import Equity, Future from zipline.assets.continuous_futures import ContinuousFuture from zipline.lib._int64window import AdjustedArrayWindow as Int64Window from zipline.lib._float64window import AdjustedArrayWindow as Float64Window from zipline.lib.adjustment import Float64Multiply, Float64Add from zipline.utils.cache import ExpiringCache from zipline.utils.math_utils import number_of_decimal_places from zipline.utils.memoize import lazyval from zipline.utils.numpy_utils import float64_dtype from zipline.utils.pandas_utils import find_in_sorted_index # Default number of decimal places used for rounding asset prices. DEFAULT_ASSET_PRICE_DECIMALS = 3 class HistoryCompatibleUSEquityAdjustmentReader(object): def __init__(self, adjustment_reader): self._adjustments_reader = adjustment_reader def load_adjustments(self, columns, dts, assets): """ Returns ------- adjustments : list[dict[int -> Adjustment]] A list, where each element corresponds to the `columns`, of mappings from index to adjustment objects to apply at that index. """ out = [None] * len(columns) for i, column in enumerate(columns): adjs = {} for asset in assets: adjs.update(self._get_adjustments_in_range( asset, dts, column)) out[i] = adjs return out def _get_adjustments_in_range(self, asset, dts, field): """ Get the Float64Multiply objects to pass to an AdjustedArrayWindow. For the use of AdjustedArrayWindow in the loader, which looks back from current simulation time back to a window of data the dictionary is structured with: - the key into the dictionary for adjustments is the location of the day from which the window is being viewed. - the start of all multiply objects is always 0 (in each window all adjustments are overlapping) - the end of the multiply object is the location before the calendar location of the adjustment action, making all days before the event adjusted. Parameters ---------- asset : Asset The assets for which to get adjustments. dts : iterable of datetime64-like The dts for which adjustment data is needed. field : str OHLCV field for which to get the adjustments. Returns ------- out : dict[loc -> Float64Multiply] The adjustments as a dict of loc -> Float64Multiply """ sid = int(asset) start = normalize_date(dts[0]) end = normalize_date(dts[-1]) adjs = {} if field != 'volume': mergers = self._adjustments_reader.get_adjustments_for_sid( 'mergers', sid) for m in mergers: dt = m[0] if start < dt <= end: end_loc = dts.searchsorted(dt) adj_loc = end_loc mult = Float64Multiply(0, end_loc - 1, 0, 0, m[1]) try: adjs[adj_loc].append(mult) except KeyError: adjs[adj_loc] = [mult] divs = self._adjustments_reader.get_adjustments_for_sid( 'dividends', sid) for d in divs: dt = d[0] if start < dt <= end: end_loc = dts.searchsorted(dt) adj_loc = end_loc mult = Float64Multiply(0, end_loc - 1, 0, 0, d[1]) try: adjs[adj_loc].append(mult) except KeyError: adjs[adj_loc] = [mult] splits = self._adjustments_reader.get_adjustments_for_sid( 'splits', sid) for s in splits: dt = s[0] if start < dt <= end: if field == 'volume': ratio = 1.0 / s[1] else: ratio = s[1] end_loc = dts.searchsorted(dt) adj_loc = end_loc mult = Float64Multiply(0, end_loc - 1, 0, 0, ratio) try: adjs[adj_loc].append(mult) except KeyError: adjs[adj_loc] = [mult] return adjs class ContinuousFutureAdjustmentReader(object): """ Calculates adjustments for continuous futures, based on the close and open of the contracts on the either side of each roll. """ def __init__(self, trading_calendar, asset_finder, bar_reader, roll_finders, frequency): self._trading_calendar = trading_calendar self._asset_finder = asset_finder self._bar_reader = bar_reader self._roll_finders = roll_finders self._frequency = frequency def load_adjustments(self, columns, dts, assets): """ Returns ------- adjustments : list[dict[int -> Adjustment]] A list, where each element corresponds to the `columns`, of mappings from index to adjustment objects to apply at that index. """ out = [None] * len(columns) for i, column in enumerate(columns): adjs = {} for asset in assets: adjs.update(self._get_adjustments_in_range( asset, dts, column)) out[i] = adjs return out def _make_adjustment(self, adjustment_type, front_close, back_close, end_loc): adj_base = back_close - front_close if adjustment_type == 'mul': adj_value = 1.0 + adj_base / front_close adj_class = Float64Multiply elif adjustment_type == 'add': adj_value = adj_base adj_class = Float64Add return adj_class(0, end_loc, 0, 0, adj_value) def _get_adjustments_in_range(self, cf, dts, field): if field == 'volume' or field == 'sid': return {} if cf.adjustment is None: return {} rf = self._roll_finders[cf.roll_style] partitions = [] rolls = rf.get_rolls(cf.root_symbol, dts[0], dts[-1], cf.offset) tc = self._trading_calendar adjs = {} for front, back in sliding_window(2, rolls): front_sid, roll_dt = front back_sid = back[0] dt = tc.previous_session_label(roll_dt) if self._frequency == 'minute': dt = tc.open_and_close_for_session(dt)[1] roll_dt = tc.open_and_close_for_session(roll_dt)[0] partitions.append((front_sid, back_sid, dt, roll_dt)) for partition in partitions: front_sid, back_sid, dt, roll_dt = partition last_front_dt = self._bar_reader.get_last_traded_dt( self._asset_finder.retrieve_asset(front_sid), dt) last_back_dt = self._bar_reader.get_last_traded_dt( self._asset_finder.retrieve_asset(back_sid), dt) if isnull(last_front_dt) or isnull(last_back_dt): continue front_close = self._bar_reader.get_value( front_sid, last_front_dt, 'close') back_close = self._bar_reader.get_value( back_sid, last_back_dt, 'close') adj_loc = dts.searchsorted(roll_dt) end_loc = adj_loc - 1 adj = self._make_adjustment(cf.adjustment, front_close, back_close, end_loc) try: adjs[adj_loc].append(adj) except KeyError: adjs[adj_loc] = [adj] return adjs class SlidingWindow(object): """ Wrapper around an AdjustedArrayWindow which supports monotonically increasing (by datetime) requests for a sized window of data. Parameters ---------- window : AdjustedArrayWindow Window of pricing data with prefetched values beyond the current simulation dt. cal_start : int Index in the overall calendar at which the window starts. """ def __init__(self, window, size, cal_start, offset): self.window = window self.cal_start = cal_start self.current = next(window) self.offset = offset self.most_recent_ix = self.cal_start + size def get(self, end_ix): """ Returns ------- out : A np.ndarray of the equity pricing up to end_ix after adjustments and rounding have been applied. """ if self.most_recent_ix == end_ix: return self.current target = end_ix - self.cal_start - self.offset + 1 self.current = self.window.seek(target) self.most_recent_ix = end_ix return self.current class HistoryLoader(with_metaclass(ABCMeta)): """ Loader for sliding history windows, with support for adjustments. Parameters ---------- trading_calendar: TradingCalendar Contains the grouping logic needed to assign minutes to periods. reader : DailyBarReader, MinuteBarReader Reader for pricing bars. adjustment_reader : SQLiteAdjustmentReader Reader for adjustment data. """ FIELDS = ('open', 'high', 'low', 'close', 'volume', 'sid') def __init__(self, trading_calendar, reader, equity_adjustment_reader, asset_finder, roll_finders=None, sid_cache_size=1000, prefetch_length=0): self.trading_calendar = trading_calendar self._asset_finder = asset_finder self._reader = reader self._adjustment_readers = {} if equity_adjustment_reader is not None: self._adjustment_readers[Equity] = \ HistoryCompatibleUSEquityAdjustmentReader( equity_adjustment_reader) if roll_finders: self._adjustment_readers[ContinuousFuture] =\ ContinuousFutureAdjustmentReader(trading_calendar, asset_finder, reader, roll_finders, self._frequency) self._window_blocks = { field: ExpiringCache(LRU(sid_cache_size)) for field in self.FIELDS } self._prefetch_length = prefetch_length @abstractproperty def _frequency(self): pass @abstractproperty def _calendar(self): pass @abstractmethod def _array(self, start, end, assets, field): pass def _decimal_places_for_asset(self, asset, reference_date): if isinstance(asset, Future) and asset.tick_size: return number_of_decimal_places(asset.tick_size) elif isinstance(asset, ContinuousFuture): # Tick size should be the same for all contracts of a continuous # future, so arbitrarily get the contract with next upcoming auto # close date. oc = self._asset_finder.get_ordered_contracts(asset.root_symbol) contract_sid = oc.contract_before_auto_close(reference_date.value) if contract_sid is not None: contract = self._asset_finder.retrieve_asset(contract_sid) if contract.tick_size: return number_of_decimal_places(contract.tick_size) return DEFAULT_ASSET_PRICE_DECIMALS def _ensure_sliding_windows(self, assets, dts, field, is_perspective_after): """ Ensure that there is a Float64Multiply window for each asset that can provide data for the given parameters. If the corresponding window for the (assets, len(dts), field) does not exist, then create a new one. If a corresponding window does exist for (assets, len(dts), field), but can not provide data for the current dts range, then create a new one and replace the expired window. Parameters ---------- assets : iterable of Assets The assets in the window dts : iterable of datetime64-like The datetimes for which to fetch data. Makes an assumption that all dts are present and contiguous, in the calendar. field : str The OHLCV field for which to retrieve data. is_perspective_after : bool see: `PricingHistoryLoader.history` Returns ------- out : list of Float64Window with sufficient data so that each asset's window can provide `get` for the index corresponding with the last value in `dts` """ end = dts[-1] size = len(dts) asset_windows = {} needed_assets = [] cal = self._calendar assets = self._asset_finder.retrieve_all(assets) end_ix = find_in_sorted_index(cal, end) for asset in assets: try: window = self._window_blocks[field].get( (asset, size, is_perspective_after), end) except KeyError: needed_assets.append(asset) else: if end_ix < window.most_recent_ix: # Window needs reset. Requested end index occurs before the # end index from the previous history call for this window. # Grab new window instead of rewinding adjustments. needed_assets.append(asset) else: asset_windows[asset] = window if needed_assets: offset = 0 start_ix = find_in_sorted_index(cal, dts[0]) prefetch_end_ix = min(end_ix + self._prefetch_length, len(cal) - 1) prefetch_end = cal[prefetch_end_ix] prefetch_dts = cal[start_ix:prefetch_end_ix + 1] if is_perspective_after: adj_end_ix = min(prefetch_end_ix + 1, len(cal) - 1) adj_dts = cal[start_ix:adj_end_ix + 1] else: adj_dts = prefetch_dts prefetch_len = len(prefetch_dts) array = self._array(prefetch_dts, needed_assets, field) if field == 'sid': window_type = Int64Window else: window_type = Float64Window view_kwargs = {} if field == 'volume': array = array.astype(float64_dtype) for i, asset in enumerate(needed_assets): adj_reader = None try: adj_reader = self._adjustment_readers[type(asset)] except KeyError: adj_reader = None if adj_reader is not None: adjs = adj_reader.load_adjustments( [field], adj_dts, [asset])[0] else: adjs = {} window = window_type( array[:, i].reshape(prefetch_len, 1), view_kwargs, adjs, offset, size, int(is_perspective_after), self._decimal_places_for_asset(asset, dts[-1]), ) sliding_window = SlidingWindow(window, size, start_ix, offset) asset_windows[asset] = sliding_window self._window_blocks[field].set( (asset, size, is_perspective_after), sliding_window, prefetch_end) return [asset_windows[asset] for asset in assets] def history(self, assets, dts, field, is_perspective_after): """ A window of pricing data with adjustments applied assuming that the end of the window is the day before the current simulation time. Parameters ---------- assets : iterable of Assets The assets in the window. dts : iterable of datetime64-like The datetimes for which to fetch data. Makes an assumption that all dts are present and contiguous, in the calendar. field : str The OHLCV field for which to retrieve data. is_perspective_after : bool True, if the window is being viewed immediately after the last dt in the sliding window. False, if the window is viewed on the last dt. This flag is used for handling the case where the last dt in the requested window immediately precedes a corporate action, e.g.: - is_perspective_after is True When the viewpoint is after the last dt in the window, as when a daily history window is accessed from a simulation that uses a minute data frequency, the history call to this loader will not include the current simulation dt. At that point in time, the raw data for the last day in the window will require adjustment, so the most recent adjustment with respect to the simulation time is applied to the last dt in the requested window. An example equity which has a 0.5 split ratio dated for 05-27, with the dts for a history call of 5 bars with a '1d' frequency at 05-27 9:31. Simulation frequency is 'minute'. (In this case this function is called with 4 daily dts, and the calling function is responsible for stitching back on the 'current' dt) | | | | | last dt | <-- viewer is here | | | 05-23 | 05-24 | 05-25 | 05-26 | 05-27 9:31 | | raw | 10.10 | 10.20 | 10.30 | 10.40 | | | adj | 5.05 | 5.10 | 5.15 | 5.25 | | The adjustment is applied to the last dt, 05-26, and all previous dts. - is_perspective_after is False, daily When the viewpoint is the same point in time as the last dt in the window, as when a daily history window is accessed from a simulation that uses a daily data frequency, the history call will include the current dt. At that point in time, the raw data for the last day in the window will be post-adjustment, so no adjustment is applied to the last dt. An example equity which has a 0.5 split ratio dated for 05-27, with the dts for a history call of 5 bars with a '1d' frequency at 05-27 0:00. Simulation frequency is 'daily'. | | | | | | <-- viewer is here | | | | | | | last dt | | | 05-23 | 05-24 | 05-25 | 05-26 | 05-27 | | raw | 10.10 | 10.20 | 10.30 | 10.40 | 5.25 | | adj | 5.05 | 5.10 | 5.15 | 5.20 | 5.25 | Adjustments are applied 05-23 through 05-26 but not to the last dt, 05-27 Returns ------- out : np.ndarray with shape(len(days between start, end), len(assets)) """ block = self._ensure_sliding_windows(assets, dts, field, is_perspective_after) end_ix = self._calendar.searchsorted(dts[-1]) return concatenate( [window.get(end_ix) for window in block], axis=1, ) class DailyHistoryLoader(HistoryLoader): @property def _frequency(self): return 'daily' @property def _calendar(self): return self._reader.sessions def _array(self, dts, assets, field): return self._reader.load_raw_arrays( [field], dts[0], dts[-1], assets, )[0] class MinuteHistoryLoader(HistoryLoader): @property def _frequency(self): return 'minute' @lazyval def _calendar(self): mm = self.trading_calendar.all_minutes start = mm.searchsorted(self._reader.first_trading_day) end = mm.searchsorted(self._reader.last_available_dt, side='right') return mm[start:end] def _array(self, dts, assets, field): return self._reader.load_raw_arrays( [field], dts[0], dts[-1], assets, )[0]
[ [ [ 603, 610 ], [ 10726, 10733 ] ], [ [ 616, 630 ], [ 12478, 12492 ] ], [ [ 636, 652 ], [ 12355, 12371 ], [ 12417, 12433 ] ], [ [ 675, 686 ], [ 21502, 21513 ] ], [ [ 703, 706 ], [ 12233, 12236 ] ], [ [ 726, 732 ], [ 8849, 8855 ], [ 8874, 8880 ] ], [ [ 758, 772 ], [ 3500, 3514 ], [ 3537, 3551 ] ], [ [ 791, 805 ], [ 8003, 8017 ] ], [ [ 823, 837 ], [ 10711, 10725 ] ], [ [ 866, 872 ], [ 11652, 11658 ] ], [ [ 874, 880 ], [ 12649, 12655 ] ], [ [ 927, 943 ], [ 11831, 11847 ], [ 12770, 12786 ] ], [ [ 981, 1015 ], [ 16235, 16246 ] ], [ [ 1055, 1091 ], [ 16295, 16308 ] ], [ [ 1127, 1142 ], [ 3923, 3938 ], [ 4609, 4624 ], [ 5396, 5411 ], [ 7284, 7299 ] ], [ [ 1144, 1154 ], [ 7396, 7406 ] ], [ [ 1187, 1200 ], [ 12219, 12232 ] ], [ [ 1238, 1262 ], [ 12697, 12721 ], [ 13308, 13332 ] ], [ [ 1297, 1304 ], [ 22081, 22088 ] ], [ [ 1343, 1356 ], [ 16410, 16423 ] ], [ [ 1396, 1416 ], [ 14855, 14875 ], [ 15619, 15639 ] ], [ [ 1485, 1513 ], [ 13368, 13396 ] ], [ [ 1526, 1567 ], [ 11680, 11721 ] ], [ [ 5769, 5801 ], [ 11868, 11900 ] ], [ [ 9573, 9586 ], [ 17257, 17270 ] ], [ [ 10697, 10710 ], [ 21626, 21639 ], [ 21994, 22007 ] ], [ [ 21607, 21625 ] ], [ [ 21974, 21993 ] ] ]
import requests API_URL = 'https://secure.techfortesco.com/tescolabsapi/restservice.aspx' class TescoLabsApi(object): def __init__(self, url, developerkey, applicationkey): self.url = url self.developerkey = developerkey self.applicationkey = applicationkey res = requests.get(self.url, params={'command': 'login', 'email': '', 'password': '', 'developerkey': self.developerkey, 'applicationkey': self.applicationkey, }) self.sessionkey = res.json()['SessionKey'] def _command(self, command, **kwargs): params = kwargs params.update({'command': command, 'sessionkey': self.sessionkey}) res = requests.get(self.url, params=params) return res def listproductcategories(self): return self._command('listproductcategories') def listproductsincategory(self, category): return self._command('listproductsincategory', category=category) def listproductoffers(self): return self._command('listproductoffers') def productsearch(self, searchtext, page=1, extendedinfo=False): return self._command('productsearch', searchtext=searchtext, page=page, extendedinfo=extendedinfo)
[ [ [ 7, 15 ], [ 303, 311 ], [ 835, 843 ] ], [ [ 17, 24 ] ], [ [ 99, 111 ] ] ]
# Generated by Django 3.2.12 on 2022-03-21 09:04 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('catalog', '0002_tag'), ] operations = [ migrations.AddField( model_name='item', name='tags', field=models.ManyToManyField(related_name='items', to='catalog.Tag', verbose_name='Теги'), ), ]
[ [ [ 72, 82 ], [ 109, 119 ], [ 221, 231 ] ], [ [ 84, 90 ], [ 316, 322 ] ], [ [ 99, 108 ] ] ]
# # Licensed to the Apache Software Foundation (ASF) under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the # "License"); you may not use this file except in compliance # with the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, # software distributed under the License is distributed on an # "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY # KIND, either express or implied. See the License for the # specific language governing permissions and limitations # under the License. """Base task runner""" import os import subprocess import threading from pwd import getpwnam from tempfile import NamedTemporaryFile from typing import Optional, Union from airflow.configuration import conf from airflow.exceptions import AirflowConfigException from airflow.models.taskinstance import load_error_file from airflow.utils.configuration import tmp_configuration_copy from airflow.utils.log.logging_mixin import LoggingMixin from airflow.utils.net import get_hostname from airflow.utils.platform import getuser PYTHONPATH_VAR = 'PYTHONPATH' class BaseTaskRunner(LoggingMixin): """ Runs Airflow task instances by invoking the `airflow tasks run` command with raw mode enabled in a subprocess. :param local_task_job: The local task job associated with running the associated task instance. :type local_task_job: airflow.jobs.local_task_job.LocalTaskJob """ def __init__(self, local_task_job): # Pass task instance context into log handlers to setup the logger. super().__init__(local_task_job.task_instance) self._task_instance = local_task_job.task_instance popen_prepend = [] if self._task_instance.run_as_user: self.run_as_user = self._task_instance.run_as_user else: try: self.run_as_user = conf.get('core', 'default_impersonation') except AirflowConfigException: self.run_as_user = None # Add sudo commands to change user if we need to. Needed to handle SubDagOperator # case using a SequentialExecutor. self.log.debug("Planning to run as the %s user", self.run_as_user) if self.run_as_user and (self.run_as_user != getuser()): # We want to include any environment variables now, as we won't # want to have to specify them in the sudo call - they would show # up in `ps` that way! And run commands now, as the other user # might not be able to run the cmds to get credentials cfg_path = tmp_configuration_copy(chmod=0o600) # Give ownership of file to user; only they can read and write subprocess.call(['sudo', 'chown', self.run_as_user, cfg_path], close_fds=True) # propagate PYTHONPATH environment variable pythonpath_value = os.environ.get(PYTHONPATH_VAR, '') popen_prepend = ['sudo', '-E', '-H', '-u', self.run_as_user] if pythonpath_value: popen_prepend.append(f'{PYTHONPATH_VAR}={pythonpath_value}') else: # Always provide a copy of the configuration file settings. Since # we are running as the same user, and can pass through environment # variables then we don't need to include those in the config copy # - the runner can read/execute those values as it needs cfg_path = tmp_configuration_copy(chmod=0o600) self._error_file = NamedTemporaryFile(delete=True) if self.run_as_user: try: os.chown(self._error_file.name, getpwnam(self.run_as_user).pw_uid, -1) except KeyError: # No user `run_as_user` found pass self._cfg_path = cfg_path self._command = ( popen_prepend + self._task_instance.command_as_list( raw=True, pickle_id=local_task_job.pickle_id, mark_success=local_task_job.mark_success, job_id=local_task_job.id, pool=local_task_job.pool, cfg_path=cfg_path, ) + ["--error-file", self._error_file.name] ) self.process = None def deserialize_run_error(self) -> Optional[Union[str, Exception]]: """Return task runtime error if its written to provided error file.""" return load_error_file(self._error_file) def _read_task_logs(self, stream): while True: line = stream.readline() if isinstance(line, bytes): line = line.decode('utf-8') if not line: break self.log.info( 'Job %s: Subtask %s %s', self._task_instance.job_id, self._task_instance.task_id, line.rstrip('\n'), ) def run_command(self, run_with=None): """ Run the task command. :param run_with: list of tokens to run the task command with e.g. ``['bash', '-c']`` :type run_with: list :return: the process that was run :rtype: subprocess.Popen """ run_with = run_with or [] full_cmd = run_with + self._command self.log.info("Running on host: %s", get_hostname()) self.log.info('Running: %s', full_cmd) proc = subprocess.Popen( full_cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT, universal_newlines=True, close_fds=True, env=os.environ.copy(), preexec_fn=os.setsid, ) # Start daemon thread to read subprocess logging output log_reader = threading.Thread( target=self._read_task_logs, args=(proc.stdout,), ) log_reader.daemon = True log_reader.start() return proc def start(self): """Start running the task instance in a subprocess.""" raise NotImplementedError() def return_code(self) -> Optional[int]: """ :return: The return code associated with running the task instance or None if the task is not yet done. :rtype: int """ raise NotImplementedError() def terminate(self) -> None: """Force kill the running task instance.""" raise NotImplementedError() def on_finish(self) -> None: """A callback that should be called when this is done running.""" if self._cfg_path and os.path.isfile(self._cfg_path): if self.run_as_user: subprocess.call(['sudo', 'rm', self._cfg_path], close_fds=True) else: os.remove(self._cfg_path) try: self._error_file.close() except FileNotFoundError: # The subprocess has deleted this file before we do # so we ignore pass
[ [ [ 817, 819 ], [ 3132, 3134 ], [ 3853, 3855 ], [ 5847, 5849 ], [ 5889, 5891 ], [ 6808, 6810 ], [ 6987, 6989 ] ], [ [ 827, 837 ], [ 2965, 2975 ], [ 5652, 5662 ], [ 5711, 5721 ], [ 5747, 5757 ], [ 6889, 6899 ] ], [ [ 845, 854 ], [ 5996, 6005 ] ], [ [ 871, 879 ], [ 3885, 3893 ] ], [ [ 901, 919 ], [ 3759, 3777 ] ], [ [ 939, 947 ], [ 4559, 4567 ], [ 6329, 6337 ] ], [ [ 949, 954 ], [ 4568, 4573 ] ], [ [ 990, 994 ], [ 2123, 2127 ] ], [ [ 1026, 1048 ], [ 2184, 2206 ] ], [ [ 1089, 1104 ], [ 4686, 4701 ] ], [ [ 1145, 1167 ], [ 2841, 2863 ], [ 3695, 3717 ] ], [ [ 1212, 1224 ], [ 1365, 1377 ] ], [ [ 1255, 1267 ], [ 5573, 5585 ] ], [ [ 1303, 1310 ], [ 2510, 2517 ] ], [ [ 1312, 1326 ], [ 3147, 3161 ], [ 3314, 3328 ] ], [ [ 1350, 1364 ] ] ]
# ======================================================================================================================================= # VNU-HCM, University of Science # Department Computer Science, Faculty of Information Technology # Authors: Nhut-Nam Le (Tich Phan Suy Rong) # © 2020 import unittest """ Given two strings, return True if either of the strings appears at the very end of the other string, ignoring upper/lower case differences (in other words, the computation should not be "case sensitive"). Note: s.lower() returns the lowercase version of a string. end_other('Hiabc', 'abc') → True end_other('AbC', 'HiaBc') → True end_other('abc', 'abXabc') → True """ def end_other(a, b): a = a.lower() b = b.lower() return (b[(len(b) - len(a)):] == a, a[(len(a) - len(b)):] == b)[len(a) >= len(b)] class TestEndOther(unittest.TestCase): def test_case_00(self): self.assertEqual(end_other('Hiabc', 'abc'), True) def test_case_01(self): self.assertEqual(end_other('AbC', 'HiaBc'), True) def test_case_02(self): self.assertEqual(end_other('abc', 'abXabc'), True) def test_case_03(self): self.assertEqual(end_other('Hiabc', 'abcd'), False) def test_case_04(self): self.assertEqual(end_other('Hiabc', 'bc'), True) def test_case_05(self): self.assertEqual(end_other('Hiabcx', 'bc'), False) def test_case_06(self): self.assertEqual(end_other('abc', 'abc'), True) def test_case_07(self): self.assertEqual(end_other('xyz', '12xyz'), True) def test_case_08(self): self.assertEqual(end_other('yz', '12xz'), False) def test_case_09(self): self.assertEqual(end_other('Z', '12xz'), True) def test_case_10(self): self.assertEqual(end_other('12', '12'), True) def test_case_11(self): self.assertEqual(end_other('abcXYZ', 'abcDEF'), False) def test_case_12(self): self.assertEqual(end_other('ab', 'ab12'), False) def test_case_13(self): self.assertEqual(end_other('ab', '12ab'), True) if __name__ == "__main__": unittest.main()
[ [ [ 298, 306 ], [ 848, 856 ], [ 2114, 2122 ] ], [ [ 688, 697 ], [ 921, 930 ], [ 1008, 1017 ], [ 1095, 1104 ], [ 1183, 1192 ], [ 1272, 1281 ], [ 1358, 1367 ], [ 1446, 1455 ], [ 1531, 1540 ], [ 1618, 1627 ], [ 1704, 1713 ], [ 1788, 1797 ], [ 1871, 1880 ], [ 1963, 1972 ], [ 2049, 2058 ] ], [ [ 835, 847 ] ] ]
# ---------------------------------------------------------------------------- # pyglet # Copyright (c) 2006-2008 Alex Holkner # All rights reserved. # # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions # are met: # # * Redistributions of source code must retain the above copyright # notice, this list of conditions and the following disclaimer. # * Redistributions in binary form must reproduce the above copyright # notice, this list of conditions and the following disclaimer in # the documentation and/or other materials provided with the # distribution. # * Neither the name of pyglet nor the names of its # contributors may be used to endorse or promote products # derived from this software without specific prior written # permission. # # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS # "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT # LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS # FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE # COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, # INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, # BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; # LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER # CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT # LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN # ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE # POSSIBILITY OF SUCH DAMAGE. # ---------------------------------------------------------------------------- """Document formats. :since: pyglet 1.1 """
[]
import os from dotenv import load_dotenv load_dotenv() class Config: SECRET_KEY = os.environ.get('SECRET_KEY') SQLALCHEMY_DATABASE_URI = 'postgresql+psycopg2://fidel:fidel@localhost/blog' UPLOADED_PHOTOS_DEST = 'app/static/photos' QUOTES_URL = 'http://quotes.stormconsultancy.co.uk/random.json' MAIL_SERVER = 'smtp.googlemail.com' MAIL_PORT = 587 MAIL_USE_TLS = True MAIL_USERNAME = os.environ.get("MAIL_USERNAME") MAIL_PASSWORD = os.environ.get("MAIL_PASSWORD") class ProdConfig(Config): SQLALCHEMY_DATABASE_URI =os.environ.get('DATABASE_URL') class DevConfig(Config): #SQLALCHEMY_DATABASE_URI = 'postgresql+psycopg2://fidel:fidel@localhost/blog' DEBUG = True config_options = { 'development':DevConfig, 'production':ProdConfig }
[ [ [ 7, 9 ], [ 87, 89 ], [ 417, 419 ], [ 469, 471 ], [ 561, 563 ] ], [ [ 29, 40 ], [ 41, 52 ] ], [ [ 62, 68 ], [ 523, 529 ], [ 614, 620 ] ], [ [ 512, 522 ], [ 781, 791 ] ], [ [ 604, 613 ], [ 757, 766 ] ], [ [ 724, 738 ] ] ]
_TF_INCLUDE_PATH = "TF_INCLUDE_PATH" _TF_LIB_PATH = "TF_LIB_PATH" def _get_env_var_with_default(repository_ctx, env_var): """Returns evironment variable value.""" if env_var in repository_ctx.os.environ: value = repository_ctx.os.environ[env_var] return value else: fail("Environment variable '%s' was not set." % env_var) def _get_tf_conf(repository_ctx): """Returns structure containing all required information about tensorflow configuration on host platform. """ include_path = _get_env_var_with_default(repository_ctx, _TF_INCLUDE_PATH) lib_path = _get_env_var_with_default(repository_ctx, _TF_LIB_PATH) return struct( include_path = include_path, lib_path = lib_path ) def _tensorflow_autoconf_impl(repository_ctx): """Implementation of tensorflow autoconf. rule.""" tf_conf = _get_tf_conf(repository_ctx) print("Using %s=%s" % (_TF_INCLUDE_PATH, tf_conf.include_path)) print("Using %s=%s" % (_TF_LIB_PATH, tf_conf.lib_path)) repository_ctx.symlink(tf_conf.include_path, 'include') repository_ctx.symlink(tf_conf.lib_path, 'lib') repository_ctx.template('BUILD', Label("//third_party/tensorflow:tensorflow.BUILD")) tensorflow_configure = repository_rule( implementation = _tensorflow_autoconf_impl, environ = [ _TF_INCLUDE_PATH, _TF_LIB_PATH ] )
[ [ [ 0, 16 ], [ 1288, 1304 ], [ 555, 571 ], [ 887, 903 ] ], [ [ 37, 49 ], [ 1310, 1322 ], [ 628, 640 ], [ 953, 965 ] ], [ [ 71, 96 ], [ 513, 538 ], [ 586, 611 ] ], [ [ 347, 359 ], [ 833, 845 ] ], [ [ 725, 750 ], [ 1243, 1268 ] ], [ [ 1184, 1204 ] ] ]
import yaml from os import path from netmiko import ConnectHandler home_dir = path.expanduser("~") filename = path.join(home_dir, ".netmiko.yml") with open(filename) as f: yaml_out = yaml.safe_load(f) cisco3 = yaml_out["cisco3"] net_connect = ConnectHandler(**cisco3) print() print(net_connect.find_prompt()) print()
[ [ [ 7, 11 ], [ 190, 194 ] ], [ [ 27, 31 ], [ 80, 84 ], [ 112, 116 ] ], [ [ 52, 66 ], [ 251, 265 ] ], [ [ 69, 77 ], [ 122, 130 ] ], [ [ 101, 109 ], [ 159, 167 ] ], [ [ 172, 173 ], [ 205, 206 ] ], [ [ 179, 187 ], [ 218, 226 ] ], [ [ 209, 215 ], [ 268, 274 ] ], [ [ 237, 248 ], [ 291, 302 ] ] ]
# -*- coding: iso-8859-1 -*-
[]
# Copyright 1999-2017 Tencent Ltd. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from QcloudApi.modules import base class Trade(base.Base): requestHost = 'trade.api.qcloud.com'
[ [ [ 611, 615 ], [ 630, 634 ] ], [ [ 624, 629 ] ] ]
"""Support for ICS Calendar.""" import copy import logging from datetime import datetime, timedelta from urllib.error import ContentTooShortError, HTTPError, URLError from urllib.request import ( HTTPPasswordMgrWithDefaultRealm, HTTPBasicAuthHandler, HTTPDigestAuthHandler, build_opener, install_opener, urlopen, ) import voluptuous as vol from homeassistant.components.calendar import ( ENTITY_ID_FORMAT, PLATFORM_SCHEMA, CalendarEventDevice, calculate_offset, is_offset_reached, ) from homeassistant.const import CONF_NAME, CONF_PASSWORD, CONF_URL, CONF_USERNAME import homeassistant.helpers.config_validation as cv from homeassistant.helpers.entity import generate_entity_id from homeassistant.util import Throttle from .icalendarparser import ICalendarParser VERSION = "2.0.0" _LOGGER = logging.getLogger(__name__) CONF_DEVICE_ID = "device_id" CONF_CALENDARS = "calendars" CONF_CALENDAR = "calendar" CONF_INCLUDE_ALL_DAY = "includeAllDay" CONF_PARSER = "parser" OFFSET = "!!" PLATFORM_SCHEMA = PLATFORM_SCHEMA.extend( { # pylint: disable=no-value-for-parameter vol.Optional(CONF_CALENDARS, default=[]): vol.All( cv.ensure_list, vol.Schema( [ vol.Schema( { vol.Required(CONF_URL): vol.Url(), vol.Required(CONF_NAME): cv.string, vol.Optional( CONF_INCLUDE_ALL_DAY, default=False ): cv.boolean, vol.Optional(CONF_USERNAME, default=""): cv.string, vol.Optional(CONF_PASSWORD, default=""): cv.string, vol.Optional(CONF_PARSER, default="icalevents"): cv.string, } ) ] ), ) } ) MIN_TIME_BETWEEN_UPDATES = timedelta(minutes=15) # MIN_TIME_BETWEEN_DOWNLOADS is smaller than MIN_TIME_BETWEEN_UPDATES so that # it won't be skipped if an explicit update is called. Eventually, if these are # configurable, we'll let end users worry about if they mean to have it happen # that way. MIN_TIME_BETWEEN_DOWNLOADS = timedelta(minutes=10) def setup_platform(hass, config, add_entities, _=None): """Set up the ICS Calendar platform""" _LOGGER.debug("Setting up ics calendars") calendar_devices = [] for calendar in config.get(CONF_CALENDARS): device_data = { CONF_NAME: calendar.get(CONF_NAME), CONF_URL: calendar.get(CONF_URL), CONF_INCLUDE_ALL_DAY: calendar.get(CONF_INCLUDE_ALL_DAY), CONF_USERNAME: calendar.get(CONF_USERNAME), CONF_PASSWORD: calendar.get(CONF_PASSWORD), CONF_PARSER: calendar.get(CONF_PARSER), } device_id = "{}".format(device_data[CONF_NAME]) entity_id = generate_entity_id(ENTITY_ID_FORMAT, device_id, hass=hass) calendar_devices.append(ICSCalendarEventDevice(entity_id, device_data)) add_entities(calendar_devices) class ICSCalendarEventDevice(CalendarEventDevice): """A device for getting the next Task from an ICS Calendar""" def __init__(self, entity_id, device_data): _LOGGER.debug("Initializing calendar: %s", device_data[CONF_NAME]) self.data = ICSCalendarData(device_data) self.entity_id = entity_id self._event = None self._name = device_data[CONF_NAME] self._offset_reached = False self._last_call = None self._last_event_list = None @property def device_state_attributes(self): """Return the calendar entity's state attributes.""" return {"offset_reached": self._offset_reached} @property def event(self): """Returns the current event for the calendar entity or None""" return self._event @property def name(self): """Returns the name of the calendar entity""" return self._name async def async_get_events(self, hass, start_date, end_date): """Get all events in a specific time frame.""" if ( self._last_event_list is None or self._last_call is None or (datetime.now() - self._last_call) > MIN_TIME_BETWEEN_UPDATES ): self._last_call = datetime.now() self._last_event_list = await self.data.async_get_events( hass, start_date, end_date ) return self._last_event_list def update(self): """Update event data.""" self.data.update() event = copy.deepcopy(self.data.event) if event is None: self._event = event return event = calculate_offset(event, OFFSET) self._offset_reached = is_offset_reached(event) self._event = event class ICSCalendarData: """Calss to use the calendar ICS client object to get next event.""" def __init__(self, device_data): """Set up how we are going to connect to the ICS Calendar""" self.name = device_data[CONF_NAME] self.url = device_data[CONF_URL] self.include_all_day = device_data[CONF_INCLUDE_ALL_DAY] self.parser = ICalendarParser.get_instance(device_data[CONF_PARSER]) self.event = None self._calendar_data = None self._last_download = None if device_data[CONF_USERNAME] != "" and device_data[CONF_PASSWORD] != "": passman = HTTPPasswordMgrWithDefaultRealm() passman.add_password( None, self.url, device_data[CONF_USERNAME], device_data[CONF_PASSWORD] ) basic_auth_handler = HTTPBasicAuthHandler(passman) digest_auth_handler = HTTPDigestAuthHandler(passman) opener = build_opener(digest_auth_handler, basic_auth_handler) install_opener(opener) def _download_calendar(self): if ( self._calendar_data is None or self._last_download is None or (datetime.now() - self._last_download) > MIN_TIME_BETWEEN_DOWNLOADS ): self._last_download = datetime.now() self._calendar_data = None try: with urlopen(self.url) as conn: self._calendar_data = conn.read().decode().replace("\0", "") except HTTPError as http_error: _LOGGER.error(f"{self.name}: Failed to open url: {http_error.reason}") except ContentTooShortError as content_too_short_error: _LOGGER.error( f"{self.name}: Could not download calendar data: {content_too_short_error.reason}" ) except URLError as url_error: _LOGGER.error(f"{self.name}: Failed to open url: {url_error.reason}") except: _LOGGER.error(f"{self.name}: Failed to open url!") return async def async_get_events(self, hass, start_date, end_date): """Get all events in a specific time frame.""" event_list = [] await hass.async_add_job(self._download_calendar) try: events = self.parser.get_event_list( content=self._calendar_data, start=start_date, end=end_date, include_all_day=self.include_all_day, ) event_list = list(map(self.format_dates, events)) except: _LOGGER.error(f"{self.name}: Failed to parse ICS!") event_list = [] return event_list @Throttle(MIN_TIME_BETWEEN_UPDATES) def update(self): """Get the latest data.""" self._download_calendar() try: self.event = self.parser.get_current_event( content=self._calendar_data, include_all_day=self.include_all_day ) self.event["start"] = self.get_hass_date( self.event["start"], self.event["all_day"] ) self.event["end"] = self.get_hass_date( self.event["end"], self.event["all_day"] ) return True except: _LOGGER.error(f"{self.name}: Failed to parse ICS!") return False def format_dates(self, event): event["start"] = self.get_date_formatted(event["start"], event["all_day"]) event["end"] = self.get_date_formatted(event["end"], event["all_day"]) return event def get_date_formatted(self, dt, is_all_day): """Return the formatted date""" # Note that all day events should have a time of 0, and the timezone # must be local. if is_all_day: return dt.strftime("%Y-%m-%d") return dt.strftime("%Y-%m-%dT%H:%M:%S.%f%z") def get_hass_date(self, dt, is_all_day): """Return the wrapped and formatted date""" if is_all_day: return {"date": self.parser.get_date_formatted(dt, is_all_day)} return {"dateTime": self.parser.get_date_formatted(dt, is_all_day)}
[ [ [ 39, 43 ], [ 4662, 4666 ] ], [ [ 51, 58 ], [ 838, 845 ] ], [ [ 80, 88 ], [ 4282, 4290 ], [ 4384, 4392 ], [ 6088, 6096 ], [ 6200, 6208 ] ], [ [ 90, 99 ], [ 1969, 1978 ], [ 2270, 2279 ] ], [ [ 125, 145 ], [ 6550, 6570 ] ], [ [ 147, 156 ], [ 6419, 6428 ] ], [ [ 158, 166 ], [ 6770, 6778 ] ], [ [ 200, 231 ], [ 5534, 5565 ] ], [ [ 237, 257 ], [ 5736, 5756 ] ], [ [ 263, 284 ], [ 5800, 5821 ] ], [ [ 290, 302 ], [ 5852, 5864 ] ], [ [ 308, 322 ], [ 5918, 5932 ] ], [ [ 328, 335 ], [ 6292, 6299 ] ], [ [ 347, 364 ], [ 1135, 1138 ], [ 1177, 1180 ], [ 1226, 1229 ], [ 1276, 1279 ], [ 1342, 1345 ], [ 1405, 1408 ], [ 1469, 1472 ], [ 1622, 1625 ], [ 1702, 1705 ], [ 1782, 1785 ], [ 1366, 1369 ] ], [ [ 417, 433 ], [ 2970, 2986 ] ], [ [ 439, 454 ], [ 1048, 1063 ] ], [ [ 460, 479 ], [ 3157, 3176 ] ], [ [ 485, 501 ], [ 4786, 4802 ] ], [ [ 507, 524 ], [ 4849, 4866 ] ], [ [ 560, 569 ], [ 1418, 1427 ], [ 2549, 2558 ], [ 2573, 2582 ], [ 2919, 2928 ], [ 3357, 3366 ], [ 3513, 3522 ], [ 5139, 5148 ] ], [ [ 571, 584 ], [ 1715, 1728 ], [ 2769, 2782 ], [ 2797, 2810 ], [ 5490, 5503 ], [ 5674, 5687 ] ], [ [ 586, 594 ], [ 1355, 1363 ], [ 2597, 2605 ], [ 2620, 2628 ], [ 5181, 5189 ] ], [ [ 596, 609 ], [ 1635, 1648 ], [ 2713, 2726 ], [ 2741, 2754 ], [ 5453, 5466 ], [ 5646, 5659 ] ], [ [ 617, 662 ], [ 1198, 1200 ], [ 1430, 1432 ], [ 1582, 1584 ], [ 1663, 1665 ], [ 1743, 1745 ], [ 1831, 1833 ] ], [ [ 704, 722 ], [ 2951, 2969 ] ], [ [ 754, 762 ], [ 7627, 7635 ] ], [ [ 792, 807 ], [ 5278, 5293 ] ], [ [ 809, 816 ] ], [ [ 828, 835 ], [ 2397, 2404 ], [ 3302, 3309 ], [ 6460, 6467 ], [ 6615, 6622 ], [ 6809, 6816 ], [ 6915, 6922 ], [ 7514, 7521 ], [ 8220, 8227 ] ], [ [ 867, 881 ] ], [ [ 896, 910 ], [ 1148, 1162 ], [ 2496, 2510 ] ], [ [ 925, 938 ] ], [ [ 952, 972 ], [ 1515, 1535 ], [ 2643, 2663 ], [ 2678, 2698 ], [ 5234, 5254 ] ], [ [ 991, 1002 ], [ 1795, 1806 ], [ 2825, 2836 ], [ 2851, 2862 ], [ 5319, 5330 ] ], [ [ 1015, 1021 ], [ 4810, 4816 ] ], [ [ 1030, 1045 ] ], [ [ 1942, 1966 ], [ 7636, 7660 ], [ 4318, 4342 ] ], [ [ 2241, 2267 ], [ 6128, 6154 ] ], [ [ 2298, 2312 ] ], [ [ 3134, 3156 ], [ 3042, 3064 ] ], [ [ 4910, 4925 ], [ 3389, 3404 ] ] ]
# Copyright 2020 Huawei Technologies Co., Ltd # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================ """SSD dataset""" from __future__ import division import os import json import xml.etree.ElementTree as et import numpy as np import cv2 import mindspore.dataset as de import mindspore.dataset.vision.c_transforms as C from mindspore.mindrecord import FileWriter from .config import config from .box_utils import jaccard_numpy, ssd_bboxes_encode def _rand(a=0., b=1.): """Generate random.""" return np.random.rand() * (b - a) + a def get_imageId_from_fileName(filename, id_iter): """Get imageID from fileName if fileName is int, else return id_iter.""" filename = os.path.splitext(filename)[0] if filename.isdigit(): return int(filename) return id_iter def random_sample_crop(image, boxes): """Random Crop the image and boxes""" height, width, _ = image.shape min_iou = np.random.choice([None, 0.1, 0.3, 0.5, 0.7, 0.9]) if min_iou is None: return image, boxes # max trails (50) for _ in range(50): image_t = image w = _rand(0.3, 1.0) * width h = _rand(0.3, 1.0) * height # aspect ratio constraint b/t .5 & 2 if h / w < 0.5 or h / w > 2: continue left = _rand() * (width - w) top = _rand() * (height - h) rect = np.array([int(top), int(left), int(top + h), int(left + w)]) overlap = jaccard_numpy(boxes, rect) # dropout some boxes drop_mask = overlap > 0 if not drop_mask.any(): continue if overlap[drop_mask].min() < min_iou and overlap[drop_mask].max() > (min_iou + 0.2): continue image_t = image_t[rect[0]:rect[2], rect[1]:rect[3], :] centers = (boxes[:, :2] + boxes[:, 2:4]) / 2.0 m1 = (rect[0] < centers[:, 0]) * (rect[1] < centers[:, 1]) m2 = (rect[2] > centers[:, 0]) * (rect[3] > centers[:, 1]) # mask in that both m1 and m2 are true mask = m1 * m2 * drop_mask # have any valid boxes? try again if not if not mask.any(): continue # take only matching gt boxes boxes_t = boxes[mask, :].copy() boxes_t[:, :2] = np.maximum(boxes_t[:, :2], rect[:2]) boxes_t[:, :2] -= rect[:2] boxes_t[:, 2:4] = np.minimum(boxes_t[:, 2:4], rect[2:4]) boxes_t[:, 2:4] -= rect[:2] return image_t, boxes_t return image, boxes def preprocess_fn(img_id, image, box, is_training): """Preprocess function for dataset.""" cv2.setNumThreads(2) def _infer_data(image, input_shape): img_h, img_w, _ = image.shape input_h, input_w = input_shape image = cv2.resize(image, (input_w, input_h)) # When the channels of image is 1 if len(image.shape) == 2: image = np.expand_dims(image, axis=-1) image = np.concatenate([image, image, image], axis=-1) return img_id, image, np.array((img_h, img_w), np.float32) def _data_aug(image, box, is_training, image_size=(300, 300)): """Data augmentation function.""" ih, iw, _ = image.shape w, h = image_size if not is_training: return _infer_data(image, image_size) # Random crop box = box.astype(np.float32) image, box = random_sample_crop(image, box) ih, iw, _ = image.shape # Resize image image = cv2.resize(image, (w, h)) # Flip image or not flip = _rand() < .5 if flip: image = cv2.flip(image, 1, dst=None) # When the channels of image is 1 if len(image.shape) == 2: image = np.expand_dims(image, axis=-1) image = np.concatenate([image, image, image], axis=-1) box[:, [0, 2]] = box[:, [0, 2]] / ih box[:, [1, 3]] = box[:, [1, 3]] / iw if flip: box[:, [1, 3]] = 1 - box[:, [3, 1]] box, label, num_match = ssd_bboxes_encode(box) return image, box, label, num_match return _data_aug(image, box, is_training, image_size=config.img_shape) def create_voc_label(is_training): """Get image path and annotation from VOC.""" voc_root = config.voc_root cls_map = {name: i for i, name in enumerate(config.classes)} sub_dir = 'train' if is_training else 'eval' voc_dir = os.path.join(voc_root, sub_dir) if not os.path.isdir(voc_dir): raise ValueError(f'Cannot find {sub_dir} dataset path.') image_dir = anno_dir = voc_dir if os.path.isdir(os.path.join(voc_dir, 'Images')): image_dir = os.path.join(voc_dir, 'Images') if os.path.isdir(os.path.join(voc_dir, 'Annotations')): anno_dir = os.path.join(voc_dir, 'Annotations') if not is_training: json_file = os.path.join(config.voc_root, config.voc_json) file_dir = os.path.split(json_file)[0] if not os.path.isdir(file_dir): os.makedirs(file_dir) json_dict = {"images": [], "type": "instances", "annotations": [], "categories": []} bnd_id = 1 image_files_dict = {} image_anno_dict = {} images = [] id_iter = 0 for anno_file in os.listdir(anno_dir): print(anno_file) if not anno_file.endswith('xml'): continue tree = et.parse(os.path.join(anno_dir, anno_file)) root_node = tree.getroot() file_name = root_node.find('filename').text img_id = get_imageId_from_fileName(file_name, id_iter) id_iter += 1 image_path = os.path.join(image_dir, file_name) print(image_path) if not os.path.isfile(image_path): print(f'Cannot find image {file_name} according to annotations.') continue labels = [] for obj in root_node.iter('object'): cls_name = obj.find('name').text if cls_name not in cls_map: print(f'Label "{cls_name}" not in "{config.classes}"') continue bnd_box = obj.find('bndbox') x_min = int(bnd_box.find('xmin').text) - 1 y_min = int(bnd_box.find('ymin').text) - 1 x_max = int(bnd_box.find('xmax').text) - 1 y_max = int(bnd_box.find('ymax').text) - 1 labels.append([y_min, x_min, y_max, x_max, cls_map[cls_name]]) if not is_training: o_width = abs(x_max - x_min) o_height = abs(y_max - y_min) ann = {'area': o_width * o_height, 'iscrowd': 0, 'image_id': \ img_id, 'bbox': [x_min, y_min, o_width, o_height], \ 'category_id': cls_map[cls_name], 'id': bnd_id, \ 'ignore': 0, \ 'segmentation': []} json_dict['annotations'].append(ann) bnd_id = bnd_id + 1 if labels: images.append(img_id) image_files_dict[img_id] = image_path image_anno_dict[img_id] = np.array(labels) if not is_training: size = root_node.find("size") width = int(size.find('width').text) height = int(size.find('height').text) image = {'file_name': file_name, 'height': height, 'width': width, 'id': img_id} json_dict['images'].append(image) if not is_training: for cls_name, cid in cls_map.items(): cat = {'supercategory': 'none', 'id': cid, 'name': cls_name} json_dict['categories'].append(cat) json_fp = open(json_file, 'w') json_str = json.dumps(json_dict) json_fp.write(json_str) json_fp.close() return images, image_files_dict, image_anno_dict def create_coco_label(is_training): """Get image path and annotation from COCO.""" from pycocotools.coco import COCO coco_root = config.coco_root data_type = config.val_data_type if is_training: data_type = config.train_data_type # Classes need to train or test. train_cls = config.classes train_cls_dict = {} for i, cls in enumerate(train_cls): train_cls_dict[cls] = i anno_json = os.path.join(coco_root, config.instances_set.format(data_type)) coco = COCO(anno_json) classs_dict = {} cat_ids = coco.loadCats(coco.getCatIds()) for cat in cat_ids: classs_dict[cat["id"]] = cat["name"] image_ids = coco.getImgIds() images = [] image_path_dict = {} image_anno_dict = {} for img_id in image_ids: image_info = coco.loadImgs(img_id) file_name = image_info[0]["file_name"] anno_ids = coco.getAnnIds(imgIds=img_id, iscrowd=None) anno = coco.loadAnns(anno_ids) image_path = os.path.join(coco_root, data_type, file_name) annos = [] iscrowd = False for label in anno: bbox = label["bbox"] class_name = classs_dict[label["category_id"]] iscrowd = iscrowd or label["iscrowd"] if class_name in train_cls: x_min, x_max = bbox[0], bbox[0] + bbox[2] y_min, y_max = bbox[1], bbox[1] + bbox[3] annos.append(list(map(round, [y_min, x_min, y_max, x_max])) + [train_cls_dict[class_name]]) if not is_training and iscrowd: continue if len(annos) >= 1: images.append(img_id) image_path_dict[img_id] = image_path image_anno_dict[img_id] = np.array(annos) return images, image_path_dict, image_anno_dict def anno_parser(annos_str): """Parse annotation from string to list.""" annos = [] for anno_str in annos_str: anno = list(map(int, anno_str.strip().split(','))) annos.append(anno) return annos def filter_valid_data(image_dir, anno_path): """Filter valid image file, which both in image_dir and anno_path.""" images = [] image_path_dict = {} image_anno_dict = {} if not os.path.isdir(image_dir): raise RuntimeError("Path given is not valid.") if not os.path.isfile(anno_path): raise RuntimeError("Annotation file is not valid.") with open(anno_path, "rb") as f: lines = f.readlines() for img_id, line in enumerate(lines): line_str = line.decode("utf-8").strip() line_split = str(line_str).split(' ') file_name = line_split[0] image_path = os.path.join(image_dir, file_name) if os.path.isfile(image_path): images.append(img_id) image_path_dict[img_id] = image_path image_anno_dict[img_id] = anno_parser(line_split[1:]) return images, image_path_dict, image_anno_dict def voc_data_to_mindrecord(mindrecord_dir, is_training, prefix="ssd.mindrecord", file_num=8): """Create MindRecord file by image_dir and anno_path.""" mindrecord_path = os.path.join(mindrecord_dir, prefix) writer = FileWriter(mindrecord_path, file_num) images, image_path_dict, image_anno_dict = create_voc_label(is_training) ssd_json = { "img_id": {"type": "int32", "shape": [1]}, "image": {"type": "bytes"}, "annotation": {"type": "int32", "shape": [-1, 5]}, } writer.add_schema(ssd_json, "ssd_json") for img_id in images: image_path = image_path_dict[img_id] with open(image_path, 'rb') as f: img = f.read() annos = np.array(image_anno_dict[img_id], dtype=np.int32) img_id = np.array([img_id], dtype=np.int32) row = {"img_id": img_id, "image": img, "annotation": annos} writer.write_raw_data([row]) writer.commit() def data_to_mindrecord_byte_image(dataset="coco", is_training=True, prefix="ssd.mindrecord", file_num=8): """Create MindRecord file.""" mindrecord_dir = config.mindrecord_dir mindrecord_path = os.path.join(mindrecord_dir, prefix) writer = FileWriter(mindrecord_path, file_num) if dataset == "coco": images, image_path_dict, image_anno_dict = create_coco_label(is_training) else: images, image_path_dict, image_anno_dict = filter_valid_data(config.image_dir, config.anno_path) ssd_json = { "img_id": {"type": "int32", "shape": [1]}, "image": {"type": "bytes"}, "annotation": {"type": "int32", "shape": [-1, 5]}, } writer.add_schema(ssd_json, "ssd_json") for img_id in images: image_path = image_path_dict[img_id] with open(image_path, 'rb') as f: img = f.read() annos = np.array(image_anno_dict[img_id], dtype=np.int32) img_id = np.array([img_id], dtype=np.int32) row = {"img_id": img_id, "image": img, "annotation": annos} writer.write_raw_data([row]) writer.commit() def create_ssd_dataset(mindrecord_file, batch_size=32, repeat_num=10, device_num=1, rank=0, is_training=True, num_parallel_workers=4, use_multiprocessing=True): """Create SSD dataset with MindDataset.""" ds = de.MindDataset(mindrecord_file, columns_list=["img_id", "image", "annotation"], num_shards=device_num, shard_id=rank, num_parallel_workers=num_parallel_workers, shuffle=is_training) decode = C.Decode() ds = ds.map(operations=decode, input_columns=["image"]) change_swap_op = C.HWC2CHW() normalize_op = C.Normalize(mean=[0.485 * 255, 0.456 * 255, 0.406 * 255], std=[0.229 * 255, 0.224 * 255, 0.225 * 255]) color_adjust_op = C.RandomColorAdjust(brightness=0.4, contrast=0.4, saturation=0.4) compose_map_func = (lambda img_id, image, annotation: preprocess_fn(img_id, image, annotation, is_training)) if is_training: output_columns = ["image", "box", "label", "num_match"] trans = [color_adjust_op, normalize_op, change_swap_op] else: output_columns = ["img_id", "image", "image_shape"] trans = [normalize_op, change_swap_op] ds = ds.map(operations=compose_map_func, input_columns=["img_id", "image", "annotation"], output_columns=output_columns, column_order=output_columns, python_multiprocessing=use_multiprocessing, num_parallel_workers=num_parallel_workers) ds = ds.map(operations=trans, input_columns=["image"], python_multiprocessing=use_multiprocessing, num_parallel_workers=num_parallel_workers) ds = ds.batch(batch_size, drop_remainder=True) ds = ds.repeat(repeat_num) return ds def create_mindrecord(dataset="coco", prefix="ssd.mindrecord", is_training=True): print("Start create dataset!") # It will generate mindrecord file in config.mindrecord_dir, # and the file name is ssd.mindrecord0, 1, ... file_num. mindrecord_dir = config.mindrecord_dir mindrecord_file = os.path.join(mindrecord_dir, prefix + "0") if not os.path.exists(mindrecord_file): if not os.path.isdir(mindrecord_dir): os.makedirs(mindrecord_dir) if dataset == "coco": if os.path.isdir(config.coco_root): print("Create Mindrecord.") data_to_mindrecord_byte_image("coco", is_training, prefix) print("Create Mindrecord Done, at {}".format(mindrecord_dir)) else: print("coco_root not exits.") elif dataset == "voc": if os.path.isdir(config.voc_root): print("Create Mindrecord.") voc_data_to_mindrecord(mindrecord_dir, is_training, prefix) print("Create Mindrecord Done, at {}".format(mindrecord_dir)) else: print("voc_root not exits.") else: if os.path.isdir(config.image_dir) and os.path.exists(config.anno_path): print("Create Mindrecord.") data_to_mindrecord_byte_image("other", is_training, prefix) print("Create Mindrecord Done, at {}".format(mindrecord_dir)) else: print("image_dir or anno_path not exits.") return mindrecord_file
[ [ [ 710, 718 ] ], [ [ 727, 729 ], [ 1254, 1256 ], [ 4954, 4956 ], [ 4997, 4999 ], [ 5129, 5131 ], [ 5143, 5145 ], [ 5197, 5199 ], [ 5236, 5238 ], [ 5250, 5252 ], [ 5308, 5310 ], [ 5390, 5392 ], [ 5456, 5458 ], [ 5499, 5501 ], [ 5536, 5538 ], [ 5796, 5798 ], [ 5930, 5932 ], [ 6157, 6159 ], [ 6233, 6235 ], [ 8777, 8779 ], [ 9348, 9350 ], [ 10575, 10577 ], [ 10667, 10669 ], [ 11013, 11015 ], [ 11059, 11061 ], [ 11468, 11470 ], [ 12438, 12440 ], [ 15390, 15392 ], [ 15444, 15446 ], [ 15492, 15494 ], [ 15535, 15537 ], [ 15608, 15610 ], [ 15948, 15950 ], [ 16270, 16272 ], [ 16306, 16308 ] ], [ [ 737, 741 ], [ 8202, 8206 ] ], [ [ 749, 776 ], [ 5921, 5923 ] ], [ [ 784, 795 ], [ 1079, 1081 ], [ 1490, 1492 ], [ 1933, 1935 ], [ 2811, 2813 ], [ 2909, 2911 ], [ 7604, 7606 ], [ 10081, 10083 ], [ 12004, 12006 ], [ 12044, 12046 ], [ 12071, 12073 ], [ 12096, 12098 ], [ 13120, 13122 ], [ 13160, 13162 ], [ 13187, 13189 ], [ 13212, 13214 ], [ 3434, 3436 ], [ 3485, 3487 ], [ 3563, 3565 ], [ 3588, 3590 ], [ 3895, 3897 ], [ 4277, 4279 ], [ 4328, 4330 ] ], [ [ 803, 806 ], [ 3142, 3145 ], [ 3299, 3302 ], [ 4031, 4034 ], [ 4151, 4154 ] ], [ [ 815, 838 ], [ 13589, 13591 ] ], [ [ 846, 888 ], [ 13808, 13809 ], [ 13900, 13901 ], [ 13931, 13932 ], [ 14087, 14088 ] ], [ [ 922, 932 ], [ 11518, 11528 ], [ 12488, 12498 ] ], [ [ 953, 959 ], [ 4690, 4696 ], [ 4810, 4816 ], [ 4874, 4880 ], [ 5403, 5409 ], [ 5420, 5426 ], [ 6563, 6569 ], [ 8478, 8484 ], [ 8511, 8517 ], [ 8572, 8578 ], [ 8649, 8655 ], [ 8801, 8807 ], [ 12394, 12400 ], [ 12713, 12719 ], [ 12731, 12737 ], [ 15346, 15352 ], [ 15622, 15628 ], [ 15962, 15968 ], [ 16284, 16290 ], [ 16321, 16327 ] ], [ [ 983, 996 ], [ 2012, 2025 ] ], [ [ 998, 1015 ], [ 4565, 4582 ] ], [ [ 1022, 1027 ], [ 1677, 1682 ], [ 1713, 1718 ], [ 1858, 1863 ], [ 1894, 1899 ], [ 4101, 4106 ] ], [ [ 1116, 1141 ], [ 6069, 6094 ] ], [ [ 1365, 1383 ], [ 3928, 3946 ] ], [ [ 3047, 3060 ], [ 14211, 14224 ] ], [ [ 4714, 4730 ], [ 11603, 11619 ] ], [ [ 8340, 8357 ], [ 12603, 12620 ] ], [ [ 10156, 10167 ], [ 11208, 11219 ] ], [ [ 10383, 10400 ], [ 12695, 12712 ] ], [ [ 11295, 11317 ], [ 16040, 16062 ] ], [ [ 12237, 12266 ], [ 15701, 15730 ], [ 16400, 16429 ] ], [ [ 13353, 13371 ] ], [ [ 15084, 15101 ] ] ]
# -*- coding: utf-8 -*- # PLEASE DO NOT EDIT THIS FILE, IT IS GENERATED AND WILL BE OVERWRITTEN: # https://github.com/ccxt/ccxt/blob/master/CONTRIBUTING.md#how-to-contribute-code from ccxt.base.exchange import Exchange import json from ccxt.base.errors import ExchangeError from ccxt.base.errors import AuthenticationError from ccxt.base.errors import PermissionDenied from ccxt.base.errors import AccountSuspended from ccxt.base.errors import ArgumentsRequired from ccxt.base.errors import BadRequest from ccxt.base.errors import BadSymbol from ccxt.base.errors import InsufficientFunds from ccxt.base.errors import InvalidOrder from ccxt.base.errors import OrderNotFound from ccxt.base.errors import OrderImmediatelyFillable from ccxt.base.errors import NotSupported from ccxt.base.errors import DDoSProtection from ccxt.base.errors import RateLimitExceeded from ccxt.base.errors import ExchangeNotAvailable from ccxt.base.errors import OnMaintenance from ccxt.base.errors import InvalidNonce from ccxt.base.decimal_to_precision import TRUNCATE from ccxt.base.precise import Precise class binance(Exchange): def describe(self): return self.deep_extend(super(binance, self).describe(), { 'id': 'binance', 'name': 'Binance', 'countries': ['JP', 'MT'], # Japan, Malta 'rateLimit': 50, 'certified': True, 'pro': True, # new metainfo interface 'has': { 'cancelAllOrders': True, 'cancelOrder': True, 'CORS': None, 'createOrder': True, 'fetchBalance': True, 'fetchBorrowRate': True, 'fetchBorrowRates': False, 'fetchBidsAsks': True, 'fetchClosedOrders': 'emulated', 'fetchCurrencies': True, 'fetchDepositAddress': True, 'fetchDeposits': True, 'fetchFundingFees': True, 'fetchFundingHistory': True, 'fetchFundingRate': True, 'fetchFundingRateHistory': True, 'fetchFundingRates': True, 'fetchIndexOHLCV': True, 'fetchIsolatedPositions': True, 'fetchMarkets': True, 'fetchMarkOHLCV': True, 'fetchMyTrades': True, 'fetchOHLCV': True, 'fetchOpenOrders': True, 'fetchOrder': True, 'fetchOrderBook': True, 'fetchOrders': True, 'fetchPositions': True, 'fetchPremiumIndexOHLCV': False, 'fetchStatus': True, 'fetchTicker': True, 'fetchTickers': True, 'fetchTime': True, 'fetchTrades': True, 'fetchTradingFee': True, 'fetchTradingFees': True, 'fetchTransactions': False, 'fetchTransfers': True, 'fetchWithdrawals': True, 'setLeverage': True, 'setMarginMode': True, 'setPositionMode': True, 'addMargin': True, 'reduceMargin': True, 'transfer': True, 'withdraw': True, }, 'timeframes': { '1m': '1m', '3m': '3m', '5m': '5m', '15m': '15m', '30m': '30m', '1h': '1h', '2h': '2h', '4h': '4h', '6h': '6h', '8h': '8h', '12h': '12h', '1d': '1d', '3d': '3d', '1w': '1w', '1M': '1M', }, 'urls': { 'logo': 'https://user-images.githubusercontent.com/1294454/29604020-d5483cdc-87ee-11e7-94c7-d1a8d9169293.jpg', 'test': { 'dapiPublic': 'https://testnet.binancefuture.com/dapi/v1', 'dapiPrivate': 'https://testnet.binancefuture.com/dapi/v1', 'fapiPublic': 'https://testnet.binancefuture.com/fapi/v1', 'fapiPrivate': 'https://testnet.binancefuture.com/fapi/v1', 'fapiPrivateV2': 'https://testnet.binancefuture.com/fapi/v2', 'public': 'https://testnet.binance.vision/api/v3', 'private': 'https://testnet.binance.vision/api/v3', 'v1': 'https://testnet.binance.vision/api/v1', }, 'api': { 'wapi': 'https://api.binance.com/wapi/v3', 'sapi': 'https://api.binance.com/sapi/v1', 'dapiPublic': 'https://dapi.binance.com/dapi/v1', 'dapiPrivate': 'https://dapi.binance.com/dapi/v1', 'dapiPrivateV2': 'https://dapi.binance.com/dapi/v2', 'dapiData': 'https://dapi.binance.com/futures/data', 'fapiPublic': 'https://fapi.binance.com/fapi/v1', 'fapiPrivate': 'https://fapi.binance.com/fapi/v1', 'fapiData': 'https://fapi.binance.com/futures/data', 'fapiPrivateV2': 'https://fapi.binance.com/fapi/v2', 'public': 'https://api.binance.com/api/v3', 'private': 'https://api.binance.com/api/v3', 'v1': 'https://api.binance.com/api/v1', }, 'www': 'https://www.binance.com', # 'referral': { # 'url': 'https://www.binance.com/en/register?ref=BLEJC98C', # 'discount': 0.2, # }, 'doc': [ 'https://binance-docs.github.io/apidocs/spot/en', ], 'api_management': 'https://www.binance.com/en/usercenter/settings/api-management', 'fees': 'https://www.binance.com/en/fee/schedule', }, 'depth': 1, 'api': { # the API structure below will need 3-layer apidefs 'sapi': { 'get': { 'accountSnapshot': 1, 'system/status': 1, # these endpoints require self.apiKey 'margin/asset': 1, 'margin/pair': 1, 'margin/allAssets': 1, 'margin/allPairs': 1, 'margin/priceIndex': 1, # these endpoints require self.apiKey + self.secret 'asset/assetDividend': 1, 'asset/dribblet': 1, 'asset/transfer': 1, 'asset/assetDetail': 1, 'asset/tradeFee': 1, 'asset/get-funding-asset': 1, 'margin/loan': 1, 'margin/repay': 1, 'margin/account': 1, 'margin/transfer': 1, 'margin/interestHistory': 1, 'margin/forceLiquidationRec': 1, 'margin/order': 1, 'margin/openOrders': 1, 'margin/allOrders': 1, 'margin/myTrades': 1, 'margin/maxBorrowable': 5, 'margin/maxTransferable': 5, 'margin/isolated/transfer': 1, 'margin/isolated/account': 1, 'margin/isolated/pair': 1, 'margin/isolated/allPairs': 1, 'margin/isolated/accountLimit': 1, 'margin/interestRateHistory': 1, 'margin/orderList': 2, 'margin/allOrderList': 10, 'margin/openOrderList': 3, 'loan/income': 1, 'fiat/orders': 1, 'fiat/payments': 1, 'futures/transfer': 5, 'futures/loan/borrow/history': 1, 'futures/loan/repay/history': 1, 'futures/loan/wallet': 1, 'futures/loan/configs': 1, 'futures/loan/calcAdjustLevel': 1, 'futures/loan/calcMaxAdjustAmount': 1, 'futures/loan/adjustCollateral/history': 1, 'futures/loan/liquidationHistory': 1, # https://binance-docs.github.io/apidocs/spot/en/#withdraw-sapi 'capital/config/getall': 1, # get networks for withdrawing USDT ERC20 vs USDT Omni 'capital/deposit/address': 1, 'capital/deposit/hisrec': 1, 'capital/deposit/subAddress': 1, 'capital/deposit/subHisrec': 1, 'capital/withdraw/history': 1, 'account/status': 1, 'account/apiTradingStatus': 1, 'account/apiRestrictions/ipRestriction': 1, 'bnbBurn': 1, 'sub-account/assets': 1, 'sub-account/futures/account': 1, 'sub-account/futures/accountSummary': 1, 'sub-account/futures/positionRisk': 1, 'sub-account/futures/internalTransfer': 1, 'sub-account/list': 1, 'sub-account/margin/account': 1, 'sub-account/margin/accountSummary': 1, 'sub-account/spotSummary': 5, 'sub-account/status': 1, 'sub-account/sub/transfer/history': 1, 'sub-account/transfer/subUserHistory': 1, 'sub-account/universalTransfer': 1, # lending endpoints 'lending/daily/product/list': 1, 'lending/daily/userLeftQuota': 1, 'lending/daily/userRedemptionQuota': 1, 'lending/daily/token/position': 1, 'lending/union/account': 1, 'lending/union/purchaseRecord': 1, 'lending/union/redemptionRecord': 1, 'lending/union/interestHistory': 1, 'lending/project/list': 1, 'lending/project/position/list': 1, # mining endpoints 'mining/pub/algoList': 1, 'mining/pub/coinList': 1, 'mining/worker/detail': 5, 'mining/worker/list': 5, 'mining/payment/list': 5, 'mining/statistics/user/status': 5, 'mining/statistics/user/list': 5, # liquid swap endpoints 'bswap/pools': 1, 'bswap/liquidity': {'cost': 1, 'noPoolId': 10}, 'bswap/liquidityOps': 2, 'bswap/quote': 2, 'bswap/swap': 1, 'bswap/poolConfigure': 1, 'bswap/addLiquidityPreview': 1, 'bswap/removeLiquidityPreview': 1, # leveraged token endpoints 'blvt/tokenInfo': 1, 'blvt/subscribe/record': 1, 'blvt/redeem/record': 1, 'blvt/userLimit': 1, # broker api 'apiReferral/ifNewUser': 1, 'apiReferral/customization': 1, 'apiReferral/userCustomization': 1, 'apiReferral/rebate/recentRecord': 1, 'apiReferral/rebate/historicalRecord': 1, 'apiReferral/kickback/recentRecord': 1, 'apiReferral/kickback/historicalRecord': 1, # brokerage API 'broker/subAccountApi': 1, 'broker/subAccount': 1, 'broker/subAccountApi/commission/futures': 1, 'broker/subAccountApi/commission/coinFutures': 1, 'broker/info': 1, 'broker/transfer': 1, 'broker/transfer/futures': 1, 'broker/rebate/recentRecord': 1, 'broker/rebate/historicalRecord': 1, 'broker/subAccount/bnbBurn/status': 1, 'broker/subAccount/depositHist': 1, 'broker/subAccount/spotSummary': 1, 'broker/subAccount/marginSummary': 1, 'broker/subAccount/futuresSummary': 1, 'broker/rebate/futures/recentRecord': 1, 'broker/subAccountApi/ipRestriction': 1, 'broker/universalTransfer': 1, # v2 not supported yet # GET /sapi/v2/broker/subAccount/futuresSummary 'account/apiRestrictions': 1, # subaccounts 'managed-subaccount/asset': 1, # c2c / p2p 'c2c/orderMatch/listUserOrderHistory': 1, }, 'post': { 'asset/dust': 1, 'asset/transfer': 1, 'asset/get-funding-asset': 1, 'account/disableFastWithdrawSwitch': 1, 'account/enableFastWithdrawSwitch': 1, 'account/apiRestrictions/ipRestriction': 1, 'account/apiRestrictions/ipRestriction/ipList': 1, 'capital/withdraw/apply': 1, 'margin/transfer': 1, 'margin/loan': 1, 'margin/repay': 1, 'margin/order': 4, 'margin/order/oco': 1, 'margin/isolated/create': 1, 'margin/isolated/transfer': 1, 'margin/isolated/account': 1, 'bnbBurn': 1, 'sub-account/margin/transfer': 1, 'sub-account/margin/enable': 1, # 'sub-account/margin/enable': 1, 'sub-account/futures/enable': 1, 'sub-account/futures/transfer': 1, 'sub-account/futures/internalTransfer': 1, 'sub-account/transfer/subToSub': 1, 'sub-account/transfer/subToMaster': 1, 'sub-account/universalTransfer': 1, 'managed-subaccount/deposit': 1, 'managed-subaccount/withdraw': 1, 'userDataStream': 1, 'userDataStream/isolated': 1, 'futures/transfer': 1, 'futures/loan/borrow': 20, 'futures/loan/repay': 20, 'futures/loan/adjustCollateral': 20, # lending 'lending/customizedFixed/purchase': 1, 'lending/daily/purchase': 1, 'lending/daily/redeem': 1, # liquid swap endpoints 'bswap/liquidityAdd': 2, 'bswap/liquidityRemove': 2, 'bswap/swap': 2, # leveraged token endpoints 'blvt/subscribe': 1, 'blvt/redeem': 1, # brokerage API 'apiReferral/customization': 1, 'apiReferral/userCustomization': 1, 'apiReferral/rebate/historicalRecord': 1, 'apiReferral/kickback/historicalRecord': 1, 'broker/subAccount': 1, 'broker/subAccount/margin': 1, 'broker/subAccount/futures': 1, 'broker/subAccountApi': 1, 'broker/subAccountApi/permission': 1, 'broker/subAccountApi/commission': 1, 'broker/subAccountApi/commission/futures': 1, 'broker/subAccountApi/commission/coinFutures': 1, 'broker/transfer': 1, 'broker/transfer/futures': 1, 'broker/rebate/historicalRecord': 1, 'broker/subAccount/bnbBurn/spot': 1, 'broker/subAccount/bnbBurn/marginInterest': 1, 'broker/subAccount/blvt': 1, 'broker/subAccountApi/ipRestriction': 1, 'broker/subAccountApi/ipRestriction/ipList': 1, 'broker/universalTransfer': 1, 'broker/subAccountApi/permission/universalTransfer': 1, 'broker/subAccountApi/permission/vanillaOptions': 1, }, 'put': { 'userDataStream': 1, 'userDataStream/isolated': 1, }, 'delete': { 'account/apiRestrictions/ipRestriction/ipList': 1, 'margin/openOrders': 1, 'margin/order': 1, 'margin/orderList': 1, 'margin/isolated/account': 1, 'userDataStream': 1, 'userDataStream/isolated': 1, # brokerage API 'broker/subAccountApi': 1, 'broker/subAccountApi/ipRestriction/ipList': 1, }, }, # deprecated 'wapi': { 'post': { 'withdraw': 1, 'sub-account/transfer': 1, }, 'get': { 'depositHistory': 1, 'withdrawHistory': 1, 'depositAddress': 1, 'accountStatus': 1, 'systemStatus': 1, 'apiTradingStatus': 1, 'userAssetDribbletLog': 1, 'tradeFee': 1, 'assetDetail': 1, 'sub-account/list': 1, 'sub-account/transfer/history': 1, 'sub-account/assets': 1, }, }, 'dapiPublic': { 'get': { 'ping': 1, 'time': 1, 'exchangeInfo': 1, 'depth': {'cost': 2, 'byLimit': [[50, 2], [100, 5], [500, 10], [1000, 20]]}, 'trades': 1, 'historicalTrades': 20, 'aggTrades': 20, 'premiumIndex': 10, 'fundingRate': 1, 'klines': {'cost': 1, 'byLimit': [[99, 1], [499, 2], [1000, 5], [10000, 10]]}, 'continuousKlines': {'cost': 1, 'byLimit': [[99, 1], [499, 2], [1000, 5], [10000, 10]]}, 'indexPriceKlines': {'cost': 1, 'byLimit': [[99, 1], [499, 2], [1000, 5], [10000, 10]]}, 'markPriceKlines': {'cost': 1, 'byLimit': [[99, 1], [499, 2], [1000, 5], [10000, 10]]}, 'ticker/24hr': {'cost': 1, 'noSymbol': 40}, 'ticker/price': {'cost': 1, 'noSymbol': 2}, 'ticker/bookTicker': {'cost': 1, 'noSymbol': 2}, 'openInterest': 1, }, }, 'dapiData': { 'get': { 'openInterestHist': 1, 'topLongShortAccountRatio': 1, 'topLongShortPositionRatio': 1, 'globalLongShortAccountRatio': 1, 'takerBuySellVol': 1, 'basis': 1, }, }, 'dapiPrivate': { 'get': { 'positionSide/dual': 30, 'order': 1, 'openOrder': 1, 'openOrders': {'cost': 1, 'noSymbol': 5}, 'allOrders': {'cost': 20, 'noSymbol': 40}, 'balance': 1, 'account': 5, 'positionMargin/history': 1, 'positionRisk': 1, 'userTrades': {'cost': 20, 'noSymbol': 40}, 'income': 20, 'leverageBracket': 1, 'forceOrders': {'cost': 20, 'noSymbol': 50}, 'adlQuantile': 5, }, 'post': { 'positionSide/dual': 1, 'order': 4, 'batchOrders': 5, 'countdownCancelAll': 10, 'leverage': 1, 'marginType': 1, 'positionMargin': 1, 'listenKey': 1, }, 'put': { 'listenKey': 1, }, 'delete': { 'order': 1, 'allOpenOrders': 1, 'batchOrders': 5, 'listenKey': 1, }, }, 'dapiPrivateV2': { 'get': { 'leverageBracket': 1, }, }, 'fapiPublic': { 'get': { 'ping': 1, 'time': 1, 'exchangeInfo': 1, 'depth': {'cost': 2, 'byLimit': [[50, 2], [100, 5], [500, 10], [1000, 20]]}, 'trades': 1, 'historicalTrades': 20, 'aggTrades': 20, 'klines': {'cost': 1, 'byLimit': [[99, 1], [499, 2], [1000, 5], [10000, 10]]}, 'continuousKlines': {'cost': 1, 'byLimit': [[99, 1], [499, 2], [1000, 5], [10000, 10]]}, 'markPriceKlines': {'cost': 1, 'byLimit': [[99, 1], [499, 2], [1000, 5], [10000, 10]]}, 'indexPriceKlines': {'cost': 1, 'byLimit': [[99, 1], [499, 2], [1000, 5], [10000, 10]]}, 'fundingRate': 1, 'premiumIndex': 1, 'ticker/24hr': {'cost': 1, 'noSymbol': 40}, 'ticker/price': {'cost': 1, 'noSymbol': 2}, 'ticker/bookTicker': {'cost': 1, 'noSymbol': 2}, 'openInterest': 1, 'indexInfo': 1, 'apiTradingStatus': {'cost': 1, 'noSymbol': 10}, 'lvtKlines': 1, }, }, 'fapiData': { 'get': { 'openInterestHist': 1, 'topLongShortAccountRatio': 1, 'topLongShortPositionRatio': 1, 'globalLongShortAccountRatio': 1, 'takerlongshortRatio': 1, }, }, 'fapiPrivate': { 'get': { 'forceOrders': {'cost': 20, 'noSymbol': 50}, 'allOrders': 5, 'openOrder': 1, 'openOrders': 1, 'order': 1, 'account': 5, 'balance': 5, 'leverageBracket': 1, 'positionMargin/history': 1, 'positionRisk': 5, 'positionSide/dual': 30, 'userTrades': 5, 'income': 30, 'commissionRate': 20, 'apiTradingStatus': 1, 'multiAssetsMargin': 30, # broker endpoints 'apiReferral/ifNewUser': 1, 'apiReferral/customization': 1, 'apiReferral/userCustomization': 1, 'apiReferral/traderNum': 1, 'apiReferral/overview': 1, 'apiReferral/tradeVol': 1, 'apiReferral/rebateVol': 1, 'apiReferral/traderSummary': 1, 'adlQuantile': 5, }, 'post': { 'batchOrders': 5, 'positionSide/dual': 1, 'positionMargin': 1, 'marginType': 1, 'order': 4, 'leverage': 1, 'listenKey': 1, 'countdownCancelAll': 10, 'multiAssetsMargin': 1, # broker endpoints 'apiReferral/customization': 1, 'apiReferral/userCustomization': 1, }, 'put': { 'listenKey': 1, }, 'delete': { 'batchOrders': 1, 'order': 1, 'allOpenOrders': 1, 'listenKey': 1, }, }, 'fapiPrivateV2': { 'get': { 'account': 1, 'balance': 1, 'positionRisk': 1, }, }, 'public': { 'get': { 'ping': 1, 'time': 1, 'depth': {'cost': 1, 'byLimit': [[100, 1], [500, 5], [1000, 10], [5000, 50]]}, 'trades': 1, 'aggTrades': 1, 'historicalTrades': 5, 'klines': 1, 'ticker/24hr': {'cost': 1, 'noSymbol': 40}, 'ticker/price': {'cost': 1, 'noSymbol': 2}, 'ticker/bookTicker': {'cost': 1, 'noSymbol': 2}, 'exchangeInfo': 10, }, 'put': { 'userDataStream': 1, }, 'post': { 'userDataStream': 1, }, 'delete': { 'userDataStream': 1, }, }, 'private': { 'get': { 'allOrderList': 10, # oco 'openOrderList': 3, # oco 'orderList': 2, # oco 'order': 2, 'openOrders': {'cost': 3, 'noSymbol': 40}, 'allOrders': 10, 'account': 10, 'myTrades': 10, 'rateLimit/order': 20, }, 'post': { 'order/oco': 1, 'order': 4, 'order/test': 1, }, 'delete': { 'openOrders': 1, # added on 2020-04-25 for canceling all open orders per symbol 'orderList': 1, # oco 'order': 1, }, }, }, 'fees': { 'trading': { 'feeSide': 'get', 'tierBased': False, 'percentage': True, 'taker': self.parse_number('0.001'), 'maker': self.parse_number('0.001'), }, 'future': { 'trading': { 'feeSide': 'quote', 'tierBased': True, 'percentage': True, 'taker': self.parse_number('0.000400'), 'maker': self.parse_number('0.000200'), 'tiers': { 'taker': [ [self.parse_number('0'), self.parse_number('0.000400')], [self.parse_number('250'), self.parse_number('0.000400')], [self.parse_number('2500'), self.parse_number('0.000350')], [self.parse_number('7500'), self.parse_number('0.000320')], [self.parse_number('22500'), self.parse_number('0.000300')], [self.parse_number('50000'), self.parse_number('0.000270')], [self.parse_number('100000'), self.parse_number('0.000250')], [self.parse_number('200000'), self.parse_number('0.000220')], [self.parse_number('400000'), self.parse_number('0.000200')], [self.parse_number('750000'), self.parse_number('0.000170')], ], 'maker': [ [self.parse_number('0'), self.parse_number('0.000200')], [self.parse_number('250'), self.parse_number('0.000160')], [self.parse_number('2500'), self.parse_number('0.000140')], [self.parse_number('7500'), self.parse_number('0.000120')], [self.parse_number('22500'), self.parse_number('0.000100')], [self.parse_number('50000'), self.parse_number('0.000080')], [self.parse_number('100000'), self.parse_number('0.000060')], [self.parse_number('200000'), self.parse_number('0.000040')], [self.parse_number('400000'), self.parse_number('0.000020')], [self.parse_number('750000'), self.parse_number('0')], ], }, }, }, 'delivery': { 'trading': { 'feeSide': 'base', 'tierBased': True, 'percentage': True, 'taker': self.parse_number('0.000500'), 'maker': self.parse_number('0.000100'), 'tiers': { 'taker': [ [self.parse_number('0'), self.parse_number('0.000500')], [self.parse_number('250'), self.parse_number('0.000450')], [self.parse_number('2500'), self.parse_number('0.000400')], [self.parse_number('7500'), self.parse_number('0.000300')], [self.parse_number('22500'), self.parse_number('0.000250')], [self.parse_number('50000'), self.parse_number('0.000240')], [self.parse_number('100000'), self.parse_number('0.000240')], [self.parse_number('200000'), self.parse_number('0.000240')], [self.parse_number('400000'), self.parse_number('0.000240')], [self.parse_number('750000'), self.parse_number('0.000240')], ], 'maker': [ [self.parse_number('0'), self.parse_number('0.000100')], [self.parse_number('250'), self.parse_number('0.000080')], [self.parse_number('2500'), self.parse_number('0.000050')], [self.parse_number('7500'), self.parse_number('0.0000030')], [self.parse_number('22500'), self.parse_number('0')], [self.parse_number('50000'), self.parse_number('-0.000050')], [self.parse_number('100000'), self.parse_number('-0.000060')], [self.parse_number('200000'), self.parse_number('-0.000070')], [self.parse_number('400000'), self.parse_number('-0.000080')], [self.parse_number('750000'), self.parse_number('-0.000090')], ], }, }, }, }, 'commonCurrencies': { 'BCC': 'BCC', # kept for backward-compatibility https://github.com/ccxt/ccxt/issues/4848 'YOYO': 'YOYOW', }, # exchange-specific options 'options': { 'fetchCurrencies': True, # self is a private call and it requires API keys # 'fetchTradesMethod': 'publicGetAggTrades', # publicGetTrades, publicGetHistoricalTrades 'defaultTimeInForce': 'GTC', # 'GTC' = Good To Cancel(default), 'IOC' = Immediate Or Cancel 'defaultType': 'spot', # 'spot', 'future', 'margin', 'delivery' 'hasAlreadyAuthenticatedSuccessfully': False, 'warnOnFetchOpenOrdersWithoutSymbol': True, 'fetchPositions': 'positionRisk', # or 'account' 'recvWindow': 5 * 1000, # 5 sec, binance default 'timeDifference': 0, # the difference between system clock and Binance clock 'adjustForTimeDifference': False, # controls the adjustment logic upon instantiation 'newOrderRespType': { 'market': 'FULL', # 'ACK' for order id, 'RESULT' for full order or 'FULL' for order with fills 'limit': 'FULL', # we change it from 'ACK' by default to 'FULL'(returns immediately if limit is not hit) }, 'quoteOrderQty': True, # whether market orders support amounts in quote currency 'broker': { 'spot': 'x-R4BD3S82', 'margin': 'x-R4BD3S82', 'future': 'x-xcKtGhcu', 'delivery': 'x-xcKtGhcu', }, 'accountsByType': { 'main': 'MAIN', 'spot': 'MAIN', 'funding': 'FUNDING', 'margin': 'MARGIN', 'future': 'UMFUTURE', 'delivery': 'CMFUTURE', 'mining': 'MINING', }, 'typesByAccount': { 'MAIN': 'spot', 'FUNDING': 'funding', 'MARGIN': 'margin', 'UMFUTURE': 'future', 'CMFUTURE': 'delivery', 'MINING': 'mining', }, 'networks': { 'ERC20': 'ETH', 'TRC20': 'TRX', 'BEP2': 'BNB', 'BEP20': 'BSC', 'OMNI': 'OMNI', 'EOS': 'EOS', 'SPL': 'SOL', }, 'reverseNetworks': { 'tronscan.org': 'TRC20', 'etherscan.io': 'ERC20', 'bscscan.com': 'BSC', 'explorer.binance.org': 'BEP2', 'bithomp.com': 'XRP', 'bloks.io': 'EOS', 'stellar.expert': 'XLM', 'blockchair.com/bitcoin': 'BTC', 'blockchair.com/bitcoin-cash': 'BCH', 'blockchair.com/ecash': 'XEC', 'explorer.litecoin.net': 'LTC', 'explorer.avax.network': 'AVAX', 'solscan.io': 'SOL', 'polkadot.subscan.io': 'DOT', 'dashboard.internetcomputer.org': 'ICP', 'explorer.chiliz.com': 'CHZ', 'cardanoscan.io': 'ADA', 'mainnet.theoan.com': 'AION', 'algoexplorer.io': 'ALGO', 'explorer.ambrosus.com': 'AMB', 'viewblock.io/zilliqa': 'ZIL', 'viewblock.io/arweave': 'AR', 'explorer.ark.io': 'ARK', 'atomscan.com': 'ATOM', 'www.mintscan.io': 'CTK', 'explorer.bitcoindiamond.org': 'BCD', 'btgexplorer.com': 'BTG', 'bts.ai': 'BTS', 'explorer.celo.org': 'CELO', 'explorer.nervos.org': 'CKB', 'cerebro.cortexlabs.ai': 'CTXC', 'chainz.cryptoid.info': 'VIA', 'explorer.dcrdata.org': 'DCR', 'digiexplorer.info': 'DGB', 'dock.subscan.io': 'DOCK', 'dogechain.info': 'DOGE', 'explorer.elrond.com': 'EGLD', 'blockscout.com': 'ETC', 'explore-fetchhub.fetch.ai': 'FET', 'filfox.info': 'FIL', 'fio.bloks.io': 'FIO', 'explorer.firo.org': 'FIRO', 'neoscan.io': 'NEO', 'ftmscan.com': 'FTM', 'explorer.gochain.io': 'GO', 'block.gxb.io': 'GXS', 'hash-hash.info': 'HBAR', 'www.hiveblockexplorer.com': 'HIVE', 'explorer.helium.com': 'HNT', 'tracker.icon.foundation': 'ICX', 'www.iostabc.com': 'IOST', 'explorer.iota.org': 'IOTA', 'iotexscan.io': 'IOTX', 'irishub.iobscan.io': 'IRIS', 'kava.mintscan.io': 'KAVA', 'scope.klaytn.com': 'KLAY', 'kmdexplorer.io': 'KMD', 'kusama.subscan.io': 'KSM', 'explorer.lto.network': 'LTO', 'polygonscan.com': 'POLYGON', 'explorer.ont.io': 'ONT', 'minaexplorer.com': 'MINA', 'nanolooker.com': 'NANO', 'explorer.nebulas.io': 'NAS', 'explorer.nbs.plus': 'NBS', 'explorer.nebl.io': 'NEBL', 'nulscan.io': 'NULS', 'nxscan.com': 'NXS', 'explorer.harmony.one': 'ONE', 'explorer.poa.network': 'POA', 'qtum.info': 'QTUM', 'explorer.rsk.co': 'RSK', 'www.oasisscan.com': 'ROSE', 'ravencoin.network': 'RVN', 'sc.tokenview.com': 'SC', 'secretnodes.com': 'SCRT', 'explorer.skycoin.com': 'SKY', 'steemscan.com': 'STEEM', 'explorer.stacks.co': 'STX', 'www.thetascan.io': 'THETA', 'scan.tomochain.com': 'TOMO', 'explore.vechain.org': 'VET', 'explorer.vite.net': 'VITE', 'www.wanscan.org': 'WAN', 'wavesexplorer.com': 'WAVES', 'wax.eosx.io': 'WAXP', 'waltonchain.pro': 'WTC', 'chain.nem.ninja': 'XEM', 'verge-blockchain.info': 'XVG', 'explorer.yoyow.org': 'YOYOW', 'explorer.zcha.in': 'ZEC', 'explorer.zensystem.io': 'ZEN', }, 'impliedNetworks': { 'ETH': {'ERC20': 'ETH'}, 'TRX': {'TRC20': 'TRX'}, }, 'legalMoney': { 'MXN': True, 'UGX': True, 'SEK': True, 'CHF': True, 'VND': True, 'AED': True, 'DKK': True, 'KZT': True, 'HUF': True, 'PEN': True, 'PHP': True, 'USD': True, 'TRY': True, 'EUR': True, 'NGN': True, 'PLN': True, 'BRL': True, 'ZAR': True, 'KES': True, 'ARS': True, 'RUB': True, 'AUD': True, 'NOK': True, 'CZK': True, 'GBP': True, 'UAH': True, 'GHS': True, 'HKD': True, 'CAD': True, 'INR': True, 'JPY': True, 'NZD': True, }, }, # https://binance-docs.github.io/apidocs/spot/en/#error-codes-2 'exceptions': { 'exact': { 'System is under maintenance.': OnMaintenance, # {"code":1,"msg":"System is under maintenance."} 'System abnormality': ExchangeError, # {"code":-1000,"msg":"System abnormality"} 'You are not authorized to execute self request.': PermissionDenied, # {"msg":"You are not authorized to execute self request."} 'API key does not exist': AuthenticationError, 'Order would trigger immediately.': OrderImmediatelyFillable, 'Stop price would trigger immediately.': OrderImmediatelyFillable, # {"code":-2010,"msg":"Stop price would trigger immediately."} 'Order would immediately match and take.': OrderImmediatelyFillable, # {"code":-2010,"msg":"Order would immediately match and take."} 'Account has insufficient balance for requested action.': InsufficientFunds, 'Rest API trading is not enabled.': ExchangeNotAvailable, "You don't have permission.": PermissionDenied, # {"msg":"You don't have permission.","success":false} 'Market is closed.': ExchangeNotAvailable, # {"code":-1013,"msg":"Market is closed."} 'Too many requests. Please try again later.': DDoSProtection, # {"msg":"Too many requests. Please try again later.","success":false} '-1000': ExchangeNotAvailable, # {"code":-1000,"msg":"An unknown error occured while processing the request."} '-1001': ExchangeNotAvailable, # 'Internal error; unable to process your request. Please try again.' '-1002': AuthenticationError, # 'You are not authorized to execute self request.' '-1003': RateLimitExceeded, # {"code":-1003,"msg":"Too much request weight used, current limit is 1200 request weight per 1 MINUTE. Please use the websocket for live updates to avoid polling the API."} '-1013': InvalidOrder, # createOrder -> 'invalid quantity'/'invalid price'/MIN_NOTIONAL '-1015': RateLimitExceeded, # 'Too many new orders; current limit is %s orders per %s.' '-1016': ExchangeNotAvailable, # 'This service is no longer available.', '-1020': BadRequest, # 'This operation is not supported.' '-1021': InvalidNonce, # 'your time is ahead of server' '-1022': AuthenticationError, # {"code":-1022,"msg":"Signature for self request is not valid."} '-1100': BadRequest, # createOrder(symbol, 1, asdf) -> 'Illegal characters found in parameter 'price' '-1101': BadRequest, # Too many parameters; expected %s and received %s. '-1102': BadRequest, # Param %s or %s must be sent, but both were empty '-1103': BadRequest, # An unknown parameter was sent. '-1104': BadRequest, # Not all sent parameters were read, read 8 parameters but was sent 9 '-1105': BadRequest, # Parameter %s was empty. '-1106': BadRequest, # Parameter %s sent when not required. '-1111': BadRequest, # Precision is over the maximum defined for self asset. '-1112': InvalidOrder, # No orders on book for symbol. '-1114': BadRequest, # TimeInForce parameter sent when not required. '-1115': BadRequest, # Invalid timeInForce. '-1116': BadRequest, # Invalid orderType. '-1117': BadRequest, # Invalid side. '-1118': BadRequest, # New client order ID was empty. '-1119': BadRequest, # Original client order ID was empty. '-1120': BadRequest, # Invalid interval. '-1121': BadSymbol, # Invalid symbol. '-1125': AuthenticationError, # This listenKey does not exist. '-1127': BadRequest, # More than %s hours between startTime and endTime. '-1128': BadRequest, # {"code":-1128,"msg":"Combination of optional parameters invalid."} '-1130': BadRequest, # Data sent for paramter %s is not valid. '-1131': BadRequest, # recvWindow must be less than 60000 '-2008': AuthenticationError, # {"code":-2008,"msg":"Invalid Api-Key ID."} '-2010': ExchangeError, # generic error code for createOrder -> 'Account has insufficient balance for requested action.', {"code":-2010,"msg":"Rest API trading is not enabled."}, etc... '-2011': OrderNotFound, # cancelOrder(1, 'BTC/USDT') -> 'UNKNOWN_ORDER' '-2013': OrderNotFound, # fetchOrder(1, 'BTC/USDT') -> 'Order does not exist' '-2014': AuthenticationError, # {"code":-2014, "msg": "API-key format invalid."} '-2015': AuthenticationError, # "Invalid API-key, IP, or permissions for action." '-2019': InsufficientFunds, # {"code":-2019,"msg":"Margin is insufficient."} '-3005': InsufficientFunds, # {"code":-3005,"msg":"Transferring out not allowed. Transfer out amount exceeds max amount."} '-3006': InsufficientFunds, # {"code":-3006,"msg":"Your borrow amount has exceed maximum borrow amount."} '-3008': InsufficientFunds, # {"code":-3008,"msg":"Borrow not allowed. Your borrow amount has exceed maximum borrow amount."} '-3010': ExchangeError, # {"code":-3010,"msg":"Repay not allowed. Repay amount exceeds borrow amount."} '-3015': ExchangeError, # {"code":-3015,"msg":"Repay amount exceeds borrow amount."} '-3022': AccountSuspended, # You account's trading is banned. '-4028': BadRequest, # {"code":-4028,"msg":"Leverage 100 is not valid"} '-3020': InsufficientFunds, # {"code":-3020,"msg":"Transfer out amount exceeds max amount."} '-3041': InsufficientFunds, # {"code":-3041,"msg":"Balance is not enough"} '-5013': InsufficientFunds, # Asset transfer failed: insufficient balance" '-11008': InsufficientFunds, # {"code":-11008,"msg":"Exceeding the account's maximum borrowable limit."} '-4051': InsufficientFunds, # {"code":-4051,"msg":"Isolated balance insufficient."} }, 'broad': { 'has no operation privilege': PermissionDenied, 'MAX_POSITION': InvalidOrder, # {"code":-2010,"msg":"Filter failure: MAX_POSITION"} }, }, }) def cost_to_precision(self, symbol, cost): return self.decimal_to_precision(cost, TRUNCATE, self.markets[symbol]['precision']['quote'], self.precisionMode, self.paddingMode) def currency_to_precision(self, currency, fee): # info is available in currencies only if the user has configured his api keys if self.safe_value(self.currencies[currency], 'precision') is not None: return self.decimal_to_precision(fee, TRUNCATE, self.currencies[currency]['precision'], self.precisionMode, self.paddingMode) else: return self.number_to_string(fee) def nonce(self): return self.milliseconds() - self.options['timeDifference'] def fetch_time(self, params={}): defaultType = self.safe_string_2(self.options, 'fetchTime', 'defaultType', 'spot') type = self.safe_string(params, 'type', defaultType) query = self.omit(params, 'type') method = 'publicGetTime' if type == 'future': method = 'fapiPublicGetTime' elif type == 'delivery': method = 'dapiPublicGetTime' response = getattr(self, method)(query) return self.safe_integer(response, 'serverTime') def load_time_difference(self, params={}): serverTime = self.fetch_time(params) after = self.milliseconds() self.options['timeDifference'] = after - serverTime return self.options['timeDifference'] def fetch_currencies(self, params={}): fetchCurrenciesEnabled = self.safe_value(self.options, 'fetchCurrencies') if not fetchCurrenciesEnabled: return None # self endpoint requires authentication # while fetchCurrencies is a public API method by design # therefore we check the keys here # and fallback to generating the currencies from the markets if not self.check_required_credentials(False): return None # sandbox/testnet does not support sapi endpoints apiBackup = self.safe_string(self.urls, 'apiBackup') if apiBackup is not None: return None response = self.sapiGetCapitalConfigGetall(params) result = {} for i in range(0, len(response)): # # { # coin: 'LINK', # depositAllEnable: True, # withdrawAllEnable: True, # name: 'ChainLink', # free: '0.06168', # locked: '0', # freeze: '0', # withdrawing: '0', # ipoing: '0', # ipoable: '0', # storage: '0', # isLegalMoney: False, # trading: True, # networkList: [ # { # network: 'BNB', # coin: 'LINK', # withdrawIntegerMultiple: '0', # isDefault: False, # depositEnable: True, # withdrawEnable: True, # depositDesc: '', # withdrawDesc: '', # specialTips: 'Both a MEMO and an Address are required to successfully deposit your LINK BEP2 tokens to Binance.', # name: 'BEP2', # resetAddressStatus: False, # addressRegex: '^(bnb1)[0-9a-z]{38}$', # memoRegex: '^[0-9A-Za-z\\-_]{1,120}$', # withdrawFee: '0.002', # withdrawMin: '0.01', # withdrawMax: '9999999', # minConfirm: 1, # unLockConfirm: 0 # }, # { # network: 'BSC', # coin: 'LINK', # withdrawIntegerMultiple: '0.00000001', # isDefault: False, # depositEnable: True, # withdrawEnable: True, # depositDesc: '', # withdrawDesc: '', # specialTips: '', # name: 'BEP20(BSC)', # resetAddressStatus: False, # addressRegex: '^(0x)[0-9A-Fa-f]{40}$', # memoRegex: '', # withdrawFee: '0.005', # withdrawMin: '0.01', # withdrawMax: '9999999', # minConfirm: 15, # unLockConfirm: 0 # }, # { # network: 'ETH', # coin: 'LINK', # withdrawIntegerMultiple: '0.00000001', # isDefault: True, # depositEnable: True, # withdrawEnable: True, # depositDesc: '', # withdrawDesc: '', # name: 'ERC20', # resetAddressStatus: False, # addressRegex: '^(0x)[0-9A-Fa-f]{40}$', # memoRegex: '', # withdrawFee: '0.34', # withdrawMin: '0.68', # withdrawMax: '0', # minConfirm: 12, # unLockConfirm: 0 # } # ] # } # entry = response[i] id = self.safe_string(entry, 'coin') name = self.safe_string(entry, 'name') code = self.safe_currency_code(id) precision = None isWithdrawEnabled = True isDepositEnabled = True networkList = self.safe_value(entry, 'networkList', []) fees = {} fee = None for j in range(0, len(networkList)): networkItem = networkList[j] network = self.safe_string(networkItem, 'network') # name = self.safe_string(networkItem, 'name') withdrawFee = self.safe_number(networkItem, 'withdrawFee') depositEnable = self.safe_value(networkItem, 'depositEnable') withdrawEnable = self.safe_value(networkItem, 'withdrawEnable') isDepositEnabled = isDepositEnabled or depositEnable isWithdrawEnabled = isWithdrawEnabled or withdrawEnable fees[network] = withdrawFee isDefault = self.safe_value(networkItem, 'isDefault') if isDefault or fee is None: fee = withdrawFee trading = self.safe_value(entry, 'trading') active = (isWithdrawEnabled and isDepositEnabled and trading) result[code] = { 'id': id, 'name': name, 'code': code, 'precision': precision, 'info': entry, 'active': active, 'networks': networkList, 'fee': fee, 'fees': fees, 'limits': self.limits, } return result def fetch_markets(self, params={}): defaultType = self.safe_string_2(self.options, 'fetchMarkets', 'defaultType', 'spot') type = self.safe_string(params, 'type', defaultType) query = self.omit(params, 'type') if (type != 'spot') and (type != 'future') and (type != 'margin') and (type != 'delivery'): raise ExchangeError(self.id + " does not support '" + type + "' type, set exchange.options['defaultType'] to 'spot', 'margin', 'delivery' or 'future'") # eslint-disable-line quotes method = 'publicGetExchangeInfo' if type == 'future': method = 'fapiPublicGetExchangeInfo' elif type == 'delivery': method = 'dapiPublicGetExchangeInfo' response = getattr(self, method)(query) # # spot / margin # # { # "timezone":"UTC", # "serverTime":1575416692969, # "rateLimits":[ # {"rateLimitType":"REQUEST_WEIGHT","interval":"MINUTE","intervalNum":1,"limit":1200}, # {"rateLimitType":"ORDERS","interval":"SECOND","intervalNum":10,"limit":100}, # {"rateLimitType":"ORDERS","interval":"DAY","intervalNum":1,"limit":200000} # ], # "exchangeFilters":[], # "symbols":[ # { # "symbol":"ETHBTC", # "status":"TRADING", # "baseAsset":"ETH", # "baseAssetPrecision":8, # "quoteAsset":"BTC", # "quotePrecision":8, # "baseCommissionPrecision":8, # "quoteCommissionPrecision":8, # "orderTypes":["LIMIT","LIMIT_MAKER","MARKET","STOP_LOSS_LIMIT","TAKE_PROFIT_LIMIT"], # "icebergAllowed":true, # "ocoAllowed":true, # "quoteOrderQtyMarketAllowed":true, # "isSpotTradingAllowed":true, # "isMarginTradingAllowed":true, # "filters":[ # {"filterType":"PRICE_FILTER","minPrice":"0.00000100","maxPrice":"100000.00000000","tickSize":"0.00000100"}, # {"filterType":"PERCENT_PRICE","multiplierUp":"5","multiplierDown":"0.2","avgPriceMins":5}, # {"filterType":"LOT_SIZE","minQty":"0.00100000","maxQty":"100000.00000000","stepSize":"0.00100000"}, # {"filterType":"MIN_NOTIONAL","minNotional":"0.00010000","applyToMarket":true,"avgPriceMins":5}, # {"filterType":"ICEBERG_PARTS","limit":10}, # {"filterType":"MARKET_LOT_SIZE","minQty":"0.00000000","maxQty":"63100.00000000","stepSize":"0.00000000"}, # {"filterType":"MAX_NUM_ALGO_ORDERS","maxNumAlgoOrders":5} # ] # }, # ], # } # # futures/usdt-margined(fapi) # # { # "timezone":"UTC", # "serverTime":1575417244353, # "rateLimits":[ # {"rateLimitType":"REQUEST_WEIGHT","interval":"MINUTE","intervalNum":1,"limit":1200}, # {"rateLimitType":"ORDERS","interval":"MINUTE","intervalNum":1,"limit":1200} # ], # "exchangeFilters":[], # "symbols":[ # { # "symbol":"BTCUSDT", # "status":"TRADING", # "maintMarginPercent":"2.5000", # "requiredMarginPercent":"5.0000", # "baseAsset":"BTC", # "quoteAsset":"USDT", # "pricePrecision":2, # "quantityPrecision":3, # "baseAssetPrecision":8, # "quotePrecision":8, # "filters":[ # {"minPrice":"0.01","maxPrice":"100000","filterType":"PRICE_FILTER","tickSize":"0.01"}, # {"stepSize":"0.001","filterType":"LOT_SIZE","maxQty":"1000","minQty":"0.001"}, # {"stepSize":"0.001","filterType":"MARKET_LOT_SIZE","maxQty":"1000","minQty":"0.001"}, # {"limit":200,"filterType":"MAX_NUM_ORDERS"}, # {"multiplierDown":"0.8500","multiplierUp":"1.1500","multiplierDecimal":"4","filterType":"PERCENT_PRICE"} # ], # "orderTypes":["LIMIT","MARKET","STOP"], # "timeInForce":["GTC","IOC","FOK","GTX"] # } # ] # } # # delivery/coin-margined(dapi) # # { # "timezone": "UTC", # "serverTime": 1597667052958, # "rateLimits": [ # {"rateLimitType":"REQUEST_WEIGHT","interval":"MINUTE","intervalNum":1,"limit":6000}, # {"rateLimitType":"ORDERS","interval":"MINUTE","intervalNum":1,"limit":6000} # ], # "exchangeFilters": [], # "symbols": [ # { # "symbol": "BTCUSD_200925", # "pair": "BTCUSD", # "contractType": "CURRENT_QUARTER", # "deliveryDate": 1601020800000, # "onboardDate": 1590739200000, # "contractStatus": "TRADING", # "contractSize": 100, # "marginAsset": "BTC", # "maintMarginPercent": "2.5000", # "requiredMarginPercent": "5.0000", # "baseAsset": "BTC", # "quoteAsset": "USD", # "pricePrecision": 1, # "quantityPrecision": 0, # "baseAssetPrecision": 8, # "quotePrecision": 8, # "equalQtyPrecision": 4, # "filters": [ # {"minPrice":"0.1","maxPrice":"100000","filterType":"PRICE_FILTER","tickSize":"0.1"}, # {"stepSize":"1","filterType":"LOT_SIZE","maxQty":"100000","minQty":"1"}, # {"stepSize":"0","filterType":"MARKET_LOT_SIZE","maxQty":"100000","minQty":"1"}, # {"limit":200,"filterType":"MAX_NUM_ORDERS"}, # {"multiplierDown":"0.9500","multiplierUp":"1.0500","multiplierDecimal":"4","filterType":"PERCENT_PRICE"} # ], # "orderTypes": ["LIMIT","MARKET","STOP","STOP_MARKET","TAKE_PROFIT","TAKE_PROFIT_MARKET","TRAILING_STOP_MARKET"], # "timeInForce": ["GTC","IOC","FOK","GTX"] # }, # { # "symbol": "BTCUSD_PERP", # "pair": "BTCUSD", # "contractType": "PERPETUAL", # "deliveryDate": 4133404800000, # "onboardDate": 1596006000000, # "contractStatus": "TRADING", # "contractSize": 100, # "marginAsset": "BTC", # "maintMarginPercent": "2.5000", # "requiredMarginPercent": "5.0000", # "baseAsset": "BTC", # "quoteAsset": "USD", # "pricePrecision": 1, # "quantityPrecision": 0, # "baseAssetPrecision": 8, # "quotePrecision": 8, # "equalQtyPrecision": 4, # "filters": [ # {"minPrice":"0.1","maxPrice":"100000","filterType":"PRICE_FILTER","tickSize":"0.1"}, # {"stepSize":"1","filterType":"LOT_SIZE","maxQty":"100000","minQty":"1"}, # {"stepSize":"1","filterType":"MARKET_LOT_SIZE","maxQty":"100000","minQty":"1"}, # {"limit":200,"filterType":"MAX_NUM_ORDERS"}, # {"multiplierDown":"0.8500","multiplierUp":"1.1500","multiplierDecimal":"4","filterType":"PERCENT_PRICE"} # ], # "orderTypes": ["LIMIT","MARKET","STOP","STOP_MARKET","TAKE_PROFIT","TAKE_PROFIT_MARKET","TRAILING_STOP_MARKET"], # "timeInForce": ["GTC","IOC","FOK","GTX"] # } # ] # } # if self.options['adjustForTimeDifference']: self.load_time_difference() markets = self.safe_value(response, 'symbols', []) result = [] for i in range(0, len(markets)): market = markets[i] spot = (type == 'spot') future = (type == 'future') delivery = (type == 'delivery') id = self.safe_string(market, 'symbol') lowercaseId = self.safe_string_lower(market, 'symbol') baseId = self.safe_string(market, 'baseAsset') quoteId = self.safe_string(market, 'quoteAsset') base = self.safe_currency_code(baseId) quote = self.safe_currency_code(quoteId) contractType = self.safe_string(market, 'contractType') idSymbol = (future or delivery) and (contractType != 'PERPETUAL') symbol = None expiry = None if idSymbol: symbol = id expiry = self.safe_integer(market, 'deliveryDate') else: symbol = base + '/' + quote filters = self.safe_value(market, 'filters', []) filtersByType = self.index_by(filters, 'filterType') precision = { 'base': self.safe_integer(market, 'baseAssetPrecision'), 'quote': self.safe_integer(market, 'quotePrecision'), 'amount': self.safe_integer(market, 'quantityPrecision'), 'price': self.safe_integer(market, 'pricePrecision'), } status = self.safe_string_2(market, 'status', 'contractStatus') active = (status == 'TRADING') margin = self.safe_value(market, 'isMarginTradingAllowed', False) contractSize = None fees = self.fees if future or delivery: contractSize = self.safe_string(market, 'contractSize', '1') fees = self.fees[type] maker = fees['trading']['maker'] taker = fees['trading']['taker'] settleId = self.safe_string(market, 'marginAsset') settle = self.safe_currency_code(settleId) entry = { 'id': id, 'lowercaseId': lowercaseId, 'symbol': symbol, 'base': base, 'quote': quote, 'baseId': baseId, 'quoteId': quoteId, 'info': market, 'spot': spot, 'type': type, 'margin': margin, 'future': future, 'delivery': delivery, 'linear': future, 'inverse': delivery, 'expiry': expiry, 'expiryDatetime': self.iso8601(expiry), 'settleId': settleId, 'settle': settle, 'active': active, 'precision': precision, 'contractSize': contractSize, 'maker': maker, 'taker': taker, 'limits': { 'amount': { 'min': None, 'max': None, }, 'price': { 'min': None, 'max': None, }, 'cost': { 'min': None, 'max': None, }, }, } if 'PRICE_FILTER' in filtersByType: filter = self.safe_value(filtersByType, 'PRICE_FILTER', {}) tickSize = self.safe_string(filter, 'tickSize') entry['precision']['price'] = self.precision_from_string(tickSize) # PRICE_FILTER reports zero values for maxPrice # since they updated filter types in November 2018 # https://github.com/ccxt/ccxt/issues/4286 # therefore limits['price']['max'] doesn't have any meaningful value except None entry['limits']['price'] = { 'min': self.safe_number(filter, 'minPrice'), 'max': self.safe_number(filter, 'maxPrice'), } entry['precision']['price'] = self.precision_from_string(filter['tickSize']) if 'LOT_SIZE' in filtersByType: filter = self.safe_value(filtersByType, 'LOT_SIZE', {}) stepSize = self.safe_string(filter, 'stepSize') entry['precision']['amount'] = self.precision_from_string(stepSize) entry['limits']['amount'] = { 'min': self.safe_number(filter, 'minQty'), 'max': self.safe_number(filter, 'maxQty'), } if 'MARKET_LOT_SIZE' in filtersByType: filter = self.safe_value(filtersByType, 'MARKET_LOT_SIZE', {}) entry['limits']['market'] = { 'min': self.safe_number(filter, 'minQty'), 'max': self.safe_number(filter, 'maxQty'), } if 'MIN_NOTIONAL' in filtersByType: filter = self.safe_value(filtersByType, 'MIN_NOTIONAL', {}) entry['limits']['cost']['min'] = self.safe_number_2(filter, 'minNotional', 'notional') result.append(entry) return result def fetch_balance(self, params={}): self.load_markets() defaultType = self.safe_string_2(self.options, 'fetchBalance', 'defaultType', 'spot') type = self.safe_string(params, 'type', defaultType) method = 'privateGetAccount' if type == 'future': options = self.safe_value(self.options, type, {}) fetchBalanceOptions = self.safe_value(options, 'fetchBalance', {}) method = self.safe_string(fetchBalanceOptions, 'method', 'fapiPrivateV2GetAccount') elif type == 'delivery': options = self.safe_value(self.options, type, {}) fetchBalanceOptions = self.safe_value(options, 'fetchBalance', {}) method = self.safe_string(fetchBalanceOptions, 'method', 'dapiPrivateGetAccount') elif type == 'margin': method = 'sapiGetMarginAccount' elif type == 'savings': method = 'sapiGetLendingUnionAccount' elif type == 'funding': method = 'sapiPostAssetGetFundingAsset' query = self.omit(params, 'type') response = getattr(self, method)(query) # # spot # # { # makerCommission: 10, # takerCommission: 10, # buyerCommission: 0, # sellerCommission: 0, # canTrade: True, # canWithdraw: True, # canDeposit: True, # updateTime: 1575357359602, # accountType: "MARGIN", # balances: [ # {asset: "BTC", free: "0.00219821", locked: "0.00000000" }, # ] # } # # margin # # { # "borrowEnabled":true, # "marginLevel":"999.00000000", # "totalAssetOfBtc":"0.00000000", # "totalLiabilityOfBtc":"0.00000000", # "totalNetAssetOfBtc":"0.00000000", # "tradeEnabled":true, # "transferEnabled":true, # "userAssets":[ # {"asset":"MATIC","borrowed":"0.00000000","free":"0.00000000","interest":"0.00000000","locked":"0.00000000","netAsset":"0.00000000"}, # {"asset":"VET","borrowed":"0.00000000","free":"0.00000000","interest":"0.00000000","locked":"0.00000000","netAsset":"0.00000000"}, # {"asset":"USDT","borrowed":"0.00000000","free":"0.00000000","interest":"0.00000000","locked":"0.00000000","netAsset":"0.00000000"} # ], # } # # futures(fapi) # # fapiPrivateGetAccount # # { # "feeTier":0, # "canTrade":true, # "canDeposit":true, # "canWithdraw":true, # "updateTime":0, # "totalInitialMargin":"0.00000000", # "totalMaintMargin":"0.00000000", # "totalWalletBalance":"4.54000000", # "totalUnrealizedProfit":"0.00000000", # "totalMarginBalance":"4.54000000", # "totalPositionInitialMargin":"0.00000000", # "totalOpenOrderInitialMargin":"0.00000000", # "maxWithdrawAmount":"4.54000000", # "assets":[ # { # "asset":"USDT", # "walletBalance":"4.54000000", # "unrealizedProfit":"0.00000000", # "marginBalance":"4.54000000", # "maintMargin":"0.00000000", # "initialMargin":"0.00000000", # "positionInitialMargin":"0.00000000", # "openOrderInitialMargin":"0.00000000", # "maxWithdrawAmount":"4.54000000" # } # ], # "positions":[ # { # "symbol":"BTCUSDT", # "initialMargin":"0.00000", # "maintMargin":"0.00000", # "unrealizedProfit":"0.00000000", # "positionInitialMargin":"0.00000", # "openOrderInitialMargin":"0.00000" # } # ] # } # # fapiPrivateV2GetAccount # # { # "feeTier":0, # "canTrade":true, # "canDeposit":true, # "canWithdraw":true, # "updateTime":0, # "totalInitialMargin":"0.00000000", # "totalMaintMargin":"0.00000000", # "totalWalletBalance":"0.00000000", # "totalUnrealizedProfit":"0.00000000", # "totalMarginBalance":"0.00000000", # "totalPositionInitialMargin":"0.00000000", # "totalOpenOrderInitialMargin":"0.00000000", # "totalCrossWalletBalance":"0.00000000", # "totalCrossUnPnl":"0.00000000", # "availableBalance":"0.00000000", # "maxWithdrawAmount":"0.00000000", # "assets":[ # { # "asset":"BNB", # "walletBalance":"0.01000000", # "unrealizedProfit":"0.00000000", # "marginBalance":"0.01000000", # "maintMargin":"0.00000000", # "initialMargin":"0.00000000", # "positionInitialMargin":"0.00000000", # "openOrderInitialMargin":"0.00000000", # "maxWithdrawAmount":"0.01000000", # "crossWalletBalance":"0.01000000", # "crossUnPnl":"0.00000000", # "availableBalance":"0.01000000" # } # ], # "positions":[ # { # "symbol":"BTCUSDT", # "initialMargin":"0", # "maintMargin":"0", # "unrealizedProfit":"0.00000000", # "positionInitialMargin":"0", # "openOrderInitialMargin":"0", # "leverage":"20", # "isolated":false, # "entryPrice":"0.00000", # "maxNotional":"5000000", # "positionSide":"BOTH" # }, # ] # } # # fapiPrivateV2GetBalance # # [ # { # "accountAlias":"FzFzXquXXqoC", # "asset":"BNB", # "balance":"0.01000000", # "crossWalletBalance":"0.01000000", # "crossUnPnl":"0.00000000", # "availableBalance":"0.01000000", # "maxWithdrawAmount":"0.01000000" # } # ] # # savings # # { # "totalAmountInBTC": "0.3172", # "totalAmountInUSDT": "10000", # "totalFixedAmountInBTC": "0.3172", # "totalFixedAmountInUSDT": "10000", # "totalFlexibleInBTC": "0", # "totalFlexibleInUSDT": "0", # "positionAmountVos": [ # { # "asset": "USDT", # "amount": "10000", # "amountInBTC": "0.3172", # "amountInUSDT": "10000" # }, # { # "asset": "BUSD", # "amount": "0", # "amountInBTC": "0", # "amountInUSDT": "0" # } # ] # } # # binance pay # # [ # { # "asset": "BUSD", # "free": "1129.83", # "locked": "0", # "freeze": "0", # "withdrawing": "0" # } # ] # result = { 'info': response, } timestamp = None if (type == 'spot') or (type == 'margin'): timestamp = self.safe_integer(response, 'updateTime') balances = self.safe_value_2(response, 'balances', 'userAssets', []) for i in range(0, len(balances)): balance = balances[i] currencyId = self.safe_string(balance, 'asset') code = self.safe_currency_code(currencyId) account = self.account() account['free'] = self.safe_string(balance, 'free') account['used'] = self.safe_string(balance, 'locked') result[code] = account elif type == 'savings': positionAmountVos = self.safe_value(response, 'positionAmountVos') for i in range(0, len(positionAmountVos)): entry = positionAmountVos[i] currencyId = self.safe_string(entry, 'asset') code = self.safe_currency_code(currencyId) account = self.account() usedAndTotal = self.safe_string(entry, 'amount') account['total'] = usedAndTotal account['used'] = usedAndTotal result[code] = account elif type == 'funding': for i in range(0, len(response)): entry = response[i] account = self.account() currencyId = self.safe_string(entry, 'asset') code = self.safe_currency_code(currencyId) account['free'] = self.safe_string(entry, 'free') frozen = self.safe_string(entry, 'freeze') withdrawing = self.safe_string(entry, 'withdrawing') locked = self.safe_string(entry, 'locked') account['used'] = Precise.string_add(frozen, Precise.string_add(locked, withdrawing)) result[code] = account else: balances = response if not isinstance(response, list): balances = self.safe_value(response, 'assets', []) for i in range(0, len(balances)): balance = balances[i] currencyId = self.safe_string(balance, 'asset') code = self.safe_currency_code(currencyId) account = self.account() account['free'] = self.safe_string(balance, 'availableBalance') account['used'] = self.safe_string(balance, 'initialMargin') account['total'] = self.safe_string_2(balance, 'marginBalance', 'balance') result[code] = account result['timestamp'] = timestamp result['datetime'] = self.iso8601(timestamp) return self.parse_balance(result) def fetch_order_book(self, symbol, limit=None, params={}): self.load_markets() market = self.market(symbol) request = { 'symbol': market['id'], } if limit is not None: request['limit'] = limit # default 100, max 5000, see https://github.com/binance-exchange/binance-official-api-docs/blob/master/rest-api.md#order-book method = 'publicGetDepth' if market['linear']: method = 'fapiPublicGetDepth' elif market['inverse']: method = 'dapiPublicGetDepth' response = getattr(self, method)(self.extend(request, params)) # # future # # { # "lastUpdateId":333598053905, # "E":1618631511986, # "T":1618631511964, # "bids":[ # ["2493.56","20.189"], # ["2493.54","1.000"], # ["2493.51","0.005"],["2493.37","0.280"],["2493.31","0.865"],["2493.30","0.514"],["2493.29","2.309"],["2493.25","1.500"],["2493.23","0.012"],["2493.22","7.240"],["2493.21","3.349"],["2493.20","2.030"],["2493.19","58.118"],["2493.18","174.836"],["2493.17","14.436"],["2493.12","2.000"],["2493.09","3.232"],["2493.08","2.010"],["2493.07","2.000"],["2493.06","2.000"],["2493.05","2.684"],["2493.04","2.000"],["2493.03","2.000"],["2493.02","5.000"],["2493.01","2.000"],["2493.00","1.035"],["2492.99","8.546"],["2492.98","4.012"],["2492.96","40.937"],["2492.95","40.595"],["2492.94","21.051"],["2492.92","4.012"],["2492.91","0.200"],["2492.85","2.000"],["2492.83","24.929"],["2492.81","50.000"],["2492.80","0.030"],["2492.76","0.264"],["2492.73","32.098"],["2492.71","32.664"],["2492.70","4.228"],["2492.65","1.230"],["2492.61","5.598"],["2492.60","34.786"],["2492.58","10.393"],["2492.54","4.543"],["2492.50","0.400"],["2492.49","0.600"],["2492.48","4.941"],["2492.45","1.207"],["2492.43","4.878"],["2492.40","4.762"],["2492.39","36.489"],["2492.37","3.000"],["2492.36","4.882"],["2492.33","28.117"],["2492.29","0.490"],["2492.28","76.365"],["2492.27","0.200"],["2492.23","3.804"],["2492.22","1.000"],["2492.19","20.011"],["2492.17","13.500"],["2492.16","4.058"],["2492.14","35.673"],["2492.13","1.915"],["2492.12","76.896"],["2492.10","8.050"],["2492.01","16.615"],["2492.00","10.335"],["2491.95","5.880"],["2491.93","10.000"],["2491.92","3.916"],["2491.90","0.795"],["2491.87","22.000"],["2491.85","1.260"],["2491.84","4.014"],["2491.83","6.668"],["2491.73","0.855"],["2491.72","7.572"],["2491.71","7.000"],["2491.68","3.916"],["2491.66","2.500"],["2491.64","4.945"],["2491.63","2.302"],["2491.62","4.012"],["2491.61","16.170"],["2491.60","0.793"],["2491.59","0.403"],["2491.57","17.445"],["2491.56","88.177"],["2491.53","10.000"],["2491.47","0.013"],["2491.45","0.157"],["2491.44","11.733"],["2491.39","3.593"],["2491.38","3.570"],["2491.36","28.077"],["2491.35","0.808"],["2491.30","0.065"],["2491.29","4.880"],["2491.27","22.000"],["2491.24","9.021"],["2491.23","68.393"],["2491.22","0.050"],["2491.21","1.316"],["2491.20","4.000"],["2491.19","0.108"],["2491.18","0.498"],["2491.17","5.000"],["2491.14","10.000"],["2491.13","0.383"],["2491.12","125.959"],["2491.10","0.870"],["2491.08","10.518"],["2491.05","54.743"],["2491.01","7.980"],["2490.96","3.916"],["2490.95","0.135"],["2490.91","0.140"],["2490.89","8.424"],["2490.88","5.930"],["2490.84","1.208"],["2490.83","2.005"],["2490.82","5.517"],["2490.81","73.707"],["2490.80","1.042"],["2490.79","9.626"],["2490.72","3.916"],["2490.70","0.148"],["2490.69","0.403"],["2490.68","0.012"],["2490.67","21.887"],["2490.66","0.008"],["2490.64","11.500"],["2490.61","0.005"],["2490.58","68.175"],["2490.55","0.218"],["2490.54","14.132"],["2490.53","5.157"],["2490.50","0.018"],["2490.49","9.216"],["2490.48","3.979"],["2490.47","1.884"],["2490.44","0.003"],["2490.36","14.132"],["2490.35","2.008"],["2490.34","0.200"],["2490.33","0.015"],["2490.30","0.065"],["2490.29","5.500"],["2490.28","24.203"],["2490.26","4.373"],["2490.25","0.026"],["2490.24","4.000"],["2490.23","177.628"],["2490.22","14.132"],["2490.21","0.181"],["2490.20","0.645"],["2490.19","9.024"],["2490.18","0.108"],["2490.17","0.085"],["2490.16","0.077"],["2490.14","0.275"],["2490.10","0.080"],["2490.07","0.015"],["2490.04","6.056"],["2490.00","6.796"],["2489.98","0.005"],["2489.97","0.258"],["2489.96","10.084"],["2489.95","1.202"],["2489.91","10.121"],["2489.90","10.084"],["2489.88","0.040"],["2489.87","0.004"],["2489.85","0.003"],["2489.76","3.916"],["2489.73","10.084"],["2489.71","0.272"],["2489.70","12.834"],["2489.67","0.403"],["2489.66","0.362"],["2489.64","0.738"],["2489.63","193.236"],["2489.62","14.152"],["2489.61","0.157"],["2489.59","4.011"],["2489.57","0.015"],["2489.55","0.046"],["2489.52","3.921"],["2489.51","0.005"],["2489.45","80.000"],["2489.44","0.649"],["2489.43","10.088"],["2489.39","0.009"],["2489.37","14.132"],["2489.35","72.262"],["2489.34","10.084"],["2489.33","14.136"],["2489.32","23.953"],["2489.30","0.065"],["2489.28","8.136"],["2489.24","8.022"],["2489.19","14.132"],["2489.18","0.085"],["2489.17","0.108"],["2489.14","10.084"],["2489.13","3.142"],["2489.12","77.827"],["2489.11","10.084"],["2489.10","0.080"],["2489.09","50.024"],["2489.04","3.916"],["2489.03","0.008"],["2489.01","10.084"],["2488.99","0.135"],["2488.98","0.187"],["2488.96","0.324"],["2488.92","0.064"],["2488.85","16.056"],["2488.83","14.132"],["2488.80","3.916"],["2488.79","10.084"],["2488.77","4.414"],["2488.76","0.005"],["2488.75","13.685"],["2488.73","0.020"],["2488.69","0.157"],["2488.60","80.000"],["2488.58","10.164"],["2488.57","0.004"],["2488.56","3.933"],["2488.54","3.311"],["2488.51","12.814"],["2488.50","80.099"],["2488.48","0.684"],["2488.44","0.024"],["2488.42","68.180"],["2488.39","4.412"],["2488.38","26.138"],["2488.34","44.134"],["2488.32","8.014"],["2488.30","0.065"],["2488.29","0.009"],["2488.27","4.513"],["2488.26","4.222"],["2488.25","80.000"],["2488.23","0.007"],["2488.22","0.281"],["2488.19","0.100"],["2488.18","80.100"],["2488.17","80.000"],["2488.16","8.197"],["2488.15","79.184"],["2488.13","0.025"],["2488.11","0.050"],["2488.10","0.080"],["2488.08","3.919"],["2488.04","40.103"],["2488.03","0.120"],["2488.02","0.008"],["2488.01","0.140"],["2488.00","0.406"],["2487.99","0.384"],["2487.98","0.060"],["2487.96","8.010"],["2487.94","0.246"],["2487.93","0.020"],["2487.91","0.136"],["2487.87","0.403"],["2487.84","17.910"],["2487.81","0.005"],["2487.80","0.073"],["2487.74","36.000"],["2487.73","3.225"],["2487.72","0.018"],["2487.71","0.319"],["2487.70","0.006"],["2487.66","0.003"],["2487.64","0.003"],["2487.63","0.008"],["2487.62","0.040"],["2487.60","3.916"],["2487.54","0.805"],["2487.52","0.022"],["2487.51","0.003"],["2487.50","0.051"],["2487.49","6.081"],["2487.47","80.015"],["2487.46","4.735"],["2487.45","30.000"],["2487.41","0.096"],["2487.40","0.078"],["2487.39","0.103"],["2487.37","2.279"],["2487.36","8.152"],["2487.35","2.145"],["2487.32","12.816"],["2487.31","10.023"],["2487.30","0.157"],["2487.27","0.005"],["2487.26","4.010"],["2487.25","0.008"],["2487.24","0.003"],["2487.23","0.014"],["2487.20","0.085"],["2487.17","0.011"],["2487.14","3.217"],["2487.12","3.916"],["2487.11","0.300"],["2487.10","0.088"],["2487.08","10.097"],["2487.07","1.467"],["2487.04","0.600"],["2487.01","18.363"],["2487.00","0.292"],["2486.99","0.014"],["2486.98","0.144"],["2486.97","0.443"],["2486.92","0.005"],["2486.91","0.016"],["2486.89","3.364"],["2486.88","4.166"],["2486.84","24.306"],["2486.83","0.181"],["2486.81","0.015"],["2486.80","0.082"],["2486.79","0.007"],["2486.76","0.011"],["2486.74","0.050"],["2486.73","0.782"],["2486.72","0.004"],["2486.69","0.003"],["2486.68","8.018"],["2486.66","10.004"],["2486.65","40.391"],["2486.64","3.916"],["2486.61","0.489"],["2486.60","0.196"],["2486.57","0.396"],["2486.55","4.015"],["2486.51","3.000"],["2486.50","0.003"],["2486.48","0.005"],["2486.47","0.010"],["2486.45","4.011"],["2486.44","0.602"],["2486.43","0.566"],["2486.42","3.140"],["2486.40","3.958"],["2486.39","0.003"],["2486.34","0.010"],["2486.31","6.281"],["2486.27","0.005"],["2486.26","0.004"],["2486.23","10.088"],["2486.22","0.015"],["2486.17","0.030"],["2486.16","3.916"],["2486.15","0.020"],["2486.13","13.130"],["2486.12","82.414"],["2486.11","0.244"],["2486.10","0.132"],["2486.08","0.720"],["2486.06","0.385"],["2486.01","0.004"],["2486.00","2.359"],["2485.99","154.159"],["2485.98","20.054"],["2485.96","1.000"],["2485.95","0.190"],["2485.92","4.463"],["2485.90","1.557"],["2485.87","0.402"],["2485.85","0.114"],["2485.81","0.900"],["2485.76","4.700"],["2485.75","0.300"],["2485.74","0.196"],["2485.73","4.010"],["2485.72","0.323"],["2485.70","0.263"],["2485.69","0.261"],["2485.68","3.688"],["2485.67","0.005"],["2485.64","1.216"],["2485.63","0.005"],["2485.62","0.015"],["2485.61","0.033"],["2485.60","0.004"],["2485.58","2.012"],["2485.56","0.020"],["2485.54","0.699"],["2485.52","0.003"],["2485.51","1.830"],["2485.48","5.964"],["2485.47","0.015"],["2485.44","7.251"],["2485.43","0.006"],["2485.42","0.644"],["2485.40","8.026"],["2485.38","0.489"],["2485.36","0.014"],["2485.35","0.005"],["2485.31","1.507"],["2485.30","2.107"],["2485.29","0.039"],["2485.28","0.642"],["2485.26","1.990"],["2485.25","4.996"],["2485.23","0.003"],["2485.22","0.277"],["2485.21","0.121"],["2485.20","3.952"],["2485.18","0.006"],["2485.17","0.043"],["2485.15","4.008"],["2485.14","4.434"],["2485.13","1.003"],["2485.05","0.204"],["2485.04","0.254"],["2485.02","5.000"],["2485.01","0.050"],["2485.00","80.821"],["2484.96","3.941"],["2484.95","10.023"],["2484.94","13.935"],["2484.92","0.059"],["2484.90","150.000"],["2484.89","0.004"],["2484.88","150.127"],["2484.87","0.004"],["2484.85","0.100"],["2484.83","0.006"],["2484.82","0.030"],["2484.81","1.246"],["2484.80","0.003"],["2484.79","0.045"],["2484.77","0.003"],["2484.74","0.036"],["2484.72","3.919"],["2484.70","0.134"],["2484.68","1.111"],["2484.66","76.955"],["2484.60","2.580"],["2484.59","31.432"],["2484.58","1.468"],["2484.55","1.153"],["2484.54","0.265"],["2484.53","20.024"],["2484.51","1.047"],["2484.50","0.818"],["2484.49","0.022"],["2484.48","3.887"],["2484.46","0.048"],["2484.45","0.224"],["2484.44","0.174"],["2484.43","223.079"],["2484.42","0.014"],["2484.41","1.115"],["2484.39","26.090"],["2484.38","0.066"],["2484.37","0.121"],["2484.34","0.255"],["2484.33","23.968"],["2484.29","0.085"],["2484.27","1.128"],["2484.26","1.456"],["2484.24","3.916"],["2484.23","28.126"],["2484.22","1.329"],["2484.19","2.015"],["2484.18","0.263"],["2484.15","15.489"],["2484.14","1.135"],["2484.13","0.572"],["2484.12","8.032"],["2484.11","0.021"],["2484.09","0.059"],["2484.08","0.038"],["2484.07","0.147"],["2484.05","24.156"],["2484.04","0.008"],["2484.01","1.184"],["2484.00","4.641"],["2483.99","0.006"],["2483.97","0.294"],["2483.96","0.424"],["2483.94","3.660"],["2483.93","2.067"],["2483.92","0.008"],["2483.89","0.141"],["2483.88","1.089"], # ["2483.87","110.000"],["2483.85","4.018"],["2483.81","150.077"],["2483.80","0.003"],["2483.77","0.020"] # ], # "asks":[ # ["2493.57","0.877"], # ["2493.62","0.063"], # ["2493.71","12.054"], # ] # } timestamp = self.safe_integer(response, 'T') orderbook = self.parse_order_book(response, symbol, timestamp) orderbook['nonce'] = self.safe_integer(response, 'lastUpdateId') return orderbook def parse_ticker(self, ticker, market=None): # # { # symbol: 'ETHBTC', # priceChange: '0.00068700', # priceChangePercent: '2.075', # weightedAvgPrice: '0.03342681', # prevClosePrice: '0.03310300', # lastPrice: '0.03378900', # lastQty: '0.07700000', # bidPrice: '0.03378900', # bidQty: '7.16800000', # askPrice: '0.03379000', # askQty: '24.00000000', # openPrice: '0.03310200', # highPrice: '0.03388900', # lowPrice: '0.03306900', # volume: '205478.41000000', # quoteVolume: '6868.48826294', # openTime: 1601469986932, # closeTime: 1601556386932, # firstId: 196098772, # lastId: 196186315, # count: 87544 # } # # coinm # { # baseVolume: '214549.95171161', # closeTime: '1621965286847', # count: '1283779', # firstId: '152560106', # highPrice: '39938.3', # lastId: '153843955', # lastPrice: '37993.4', # lastQty: '1', # lowPrice: '36457.2', # openPrice: '37783.4', # openTime: '1621878840000', # pair: 'BTCUSD', # priceChange: '210.0', # priceChangePercent: '0.556', # symbol: 'BTCUSD_PERP', # volume: '81990451', # weightedAvgPrice: '38215.08713747' # } # timestamp = self.safe_integer(ticker, 'closeTime') marketId = self.safe_string(ticker, 'symbol') symbol = self.safe_symbol(marketId, market) last = self.safe_number(ticker, 'lastPrice') isCoinm = ('baseVolume' in ticker) baseVolume = None quoteVolume = None if isCoinm: baseVolume = self.safe_number(ticker, 'baseVolume') quoteVolume = self.safe_number(ticker, 'volume') else: baseVolume = self.safe_number(ticker, 'volume') quoteVolume = self.safe_number(ticker, 'quoteVolume') return self.safe_ticker({ 'symbol': symbol, 'timestamp': timestamp, 'datetime': self.iso8601(timestamp), 'high': self.safe_number(ticker, 'highPrice'), 'low': self.safe_number(ticker, 'lowPrice'), 'bid': self.safe_number(ticker, 'bidPrice'), 'bidVolume': self.safe_number(ticker, 'bidQty'), 'ask': self.safe_number(ticker, 'askPrice'), 'askVolume': self.safe_number(ticker, 'askQty'), 'vwap': self.safe_number(ticker, 'weightedAvgPrice'), 'open': self.safe_number(ticker, 'openPrice'), 'close': last, 'last': last, 'previousClose': self.safe_number(ticker, 'prevClosePrice'), # previous day close 'change': self.safe_number(ticker, 'priceChange'), 'percentage': self.safe_number(ticker, 'priceChangePercent'), 'average': None, 'baseVolume': baseVolume, 'quoteVolume': quoteVolume, 'info': ticker, }, market) def fetch_status(self, params={}): response = self.sapiGetSystemStatus(params) status = self.safe_string(response, 'status') if status is not None: status = 'ok' if (status == '0') else 'maintenance' self.status = self.extend(self.status, { 'status': status, 'updated': self.milliseconds(), }) return self.status def fetch_ticker(self, symbol, params={}): self.load_markets() market = self.market(symbol) request = { 'symbol': market['id'], } method = 'publicGetTicker24hr' if market['linear']: method = 'fapiPublicGetTicker24hr' elif market['inverse']: method = 'dapiPublicGetTicker24hr' response = getattr(self, method)(self.extend(request, params)) if isinstance(response, list): firstTicker = self.safe_value(response, 0, {}) return self.parse_ticker(firstTicker, market) return self.parse_ticker(response, market) def fetch_bids_asks(self, symbols=None, params={}): self.load_markets() defaultType = self.safe_string_2(self.options, 'fetchBidsAsks', 'defaultType', 'spot') type = self.safe_string(params, 'type', defaultType) query = self.omit(params, 'type') method = None if type == 'future': method = 'fapiPublicGetTickerBookTicker' elif type == 'delivery': method = 'dapiPublicGetTickerBookTicker' else: method = 'publicGetTickerBookTicker' response = getattr(self, method)(query) return self.parse_tickers(response, symbols) def fetch_tickers(self, symbols=None, params={}): self.load_markets() defaultType = self.safe_string_2(self.options, 'fetchTickers', 'defaultType', 'spot') type = self.safe_string(params, 'type', defaultType) query = self.omit(params, 'type') defaultMethod = None if type == 'future': defaultMethod = 'fapiPublicGetTicker24hr' elif type == 'delivery': defaultMethod = 'dapiPublicGetTicker24hr' else: defaultMethod = 'publicGetTicker24hr' method = self.safe_string(self.options, 'fetchTickersMethod', defaultMethod) response = getattr(self, method)(query) return self.parse_tickers(response, symbols) def parse_ohlcv(self, ohlcv, market=None): # when api method = publicGetKlines or fapiPublicGetKlines or dapiPublicGetKlines # [ # 1591478520000, # open time # "0.02501300", # open # "0.02501800", # high # "0.02500000", # low # "0.02500000", # close # "22.19000000", # volume # 1591478579999, # close time # "0.55490906", # quote asset volume # 40, # number of trades # "10.92900000", # taker buy base asset volume # "0.27336462", # taker buy quote asset volume # "0" # ignore # ] # # when api method = fapiPublicGetMarkPriceKlines or fapiPublicGetIndexPriceKlines # [ # [ # 1591256460000, # Open time # "9653.29201333", # Open # "9654.56401333", # High # "9653.07367333", # Low # "9653.07367333", # Close(or latest price) # "0", # Ignore # 1591256519999, # Close time # "0", # Ignore # 60, # Number of bisic data # "0", # Ignore # "0", # Ignore # "0" # Ignore # ] # ] # return [ self.safe_integer(ohlcv, 0), self.safe_number(ohlcv, 1), self.safe_number(ohlcv, 2), self.safe_number(ohlcv, 3), self.safe_number(ohlcv, 4), self.safe_number(ohlcv, 5), ] def fetch_ohlcv(self, symbol, timeframe='1m', since=None, limit=None, params={}): self.load_markets() market = self.market(symbol) # binance docs say that the default limit 500, max 1500 for futures, max 1000 for spot markets # the reality is that the time range wider than 500 candles won't work right defaultLimit = 500 maxLimit = 1500 price = self.safe_string(params, 'price') params = self.omit(params, 'price') limit = defaultLimit if (limit is None) else min(limit, maxLimit) request = { 'interval': self.timeframes[timeframe], 'limit': limit, } if price == 'index': request['pair'] = market['id'] # Index price takes self argument instead of symbol else: request['symbol'] = market['id'] # duration = self.parse_timeframe(timeframe) if since is not None: request['startTime'] = since # # It didn't work before without the endTime # https://github.com/ccxt/ccxt/issues/8454 # # if since > 0: # endTime = self.sum(since, limit * duration * 1000 - 1) # now = self.milliseconds() # request['endTime'] = min(now, endTime) # } method = 'publicGetKlines' if price == 'mark': if market['inverse']: method = 'dapiPublicGetMarkPriceKlines' else: method = 'fapiPublicGetMarkPriceKlines' elif price == 'index': if market['inverse']: method = 'dapiPublicGetIndexPriceKlines' else: method = 'fapiPublicGetIndexPriceKlines' elif market['linear']: method = 'fapiPublicGetKlines' elif market['inverse']: method = 'dapiPublicGetKlines' response = getattr(self, method)(self.extend(request, params)) # # [ # [1591478520000,"0.02501300","0.02501800","0.02500000","0.02500000","22.19000000",1591478579999,"0.55490906",40,"10.92900000","0.27336462","0"], # [1591478580000,"0.02499600","0.02500900","0.02499400","0.02500300","21.34700000",1591478639999,"0.53370468",24,"7.53800000","0.18850725","0"], # [1591478640000,"0.02500800","0.02501100","0.02500300","0.02500800","154.14200000",1591478699999,"3.85405839",97,"5.32300000","0.13312641","0"], # ] # return self.parse_ohlcvs(response, market, timeframe, since, limit) def fetch_mark_ohlcv(self, symbol, timeframe='1m', since=None, limit=None, params={}): request = { 'price': 'mark', } return self.fetch_ohlcv(symbol, timeframe, since, limit, self.extend(request, params)) def fetch_index_ohlcv(self, symbol, timeframe='1m', since=None, limit=None, params={}): request = { 'price': 'index', } return self.fetch_ohlcv(symbol, timeframe, since, limit, self.extend(request, params)) def parse_trade(self, trade, market=None): if 'isDustTrade' in trade: return self.parse_dust_trade(trade, market) # # aggregate trades # https://github.com/binance-exchange/binance-official-api-docs/blob/master/rest-api.md#compressedaggregate-trades-list # # { # "a": 26129, # Aggregate tradeId # "p": "0.01633102", # Price # "q": "4.70443515", # Quantity # "f": 27781, # First tradeId # "l": 27781, # Last tradeId # "T": 1498793709153, # Timestamp # "m": True, # Was the buyer the maker? # "M": True # Was the trade the best price match? # } # # recent public trades and old public trades # https://github.com/binance-exchange/binance-official-api-docs/blob/master/rest-api.md#recent-trades-list # https://github.com/binance-exchange/binance-official-api-docs/blob/master/rest-api.md#old-trade-lookup-market_data # # { # "id": 28457, # "price": "4.00000100", # "qty": "12.00000000", # "time": 1499865549590, # "isBuyerMaker": True, # "isBestMatch": True # } # # private trades # https://github.com/binance-exchange/binance-official-api-docs/blob/master/rest-api.md#account-trade-list-user_data # # { # "symbol": "BNBBTC", # "id": 28457, # "orderId": 100234, # "price": "4.00000100", # "qty": "12.00000000", # "commission": "10.10000000", # "commissionAsset": "BNB", # "time": 1499865549590, # "isBuyer": True, # "isMaker": False, # "isBestMatch": True # } # # futures trades # https://binance-docs.github.io/apidocs/futures/en/#account-trade-list-user_data # # { # "accountId": 20, # "buyer": False, # "commission": "-0.07819010", # "commissionAsset": "USDT", # "counterPartyId": 653, # "id": 698759, # "maker": False, # "orderId": 25851813, # "price": "7819.01", # "qty": "0.002", # "quoteQty": "0.01563", # "realizedPnl": "-0.91539999", # "side": "SELL", # "symbol": "BTCUSDT", # "time": 1569514978020 # } # { # "symbol": "BTCUSDT", # "id": 477128891, # "orderId": 13809777875, # "side": "SELL", # "price": "38479.55", # "qty": "0.001", # "realizedPnl": "-0.00009534", # "marginAsset": "USDT", # "quoteQty": "38.47955", # "commission": "-0.00076959", # "commissionAsset": "USDT", # "time": 1612733566708, # "positionSide": "BOTH", # "maker": True, # "buyer": False # } # # {respType: FULL} # # { # "price": "4000.00000000", # "qty": "1.00000000", # "commission": "4.00000000", # "commissionAsset": "USDT", # "tradeId": "1234", # } # timestamp = self.safe_integer_2(trade, 'T', 'time') price = self.safe_string_2(trade, 'p', 'price') amount = self.safe_string_2(trade, 'q', 'qty') cost = self.safe_string_2(trade, 'quoteQty', 'baseQty') # inverse futures marketId = self.safe_string(trade, 'symbol') symbol = self.safe_symbol(marketId, market) id = self.safe_string_2(trade, 't', 'a') id = self.safe_string_2(trade, 'id', 'tradeId', id) side = None orderId = self.safe_string(trade, 'orderId') if 'm' in trade: side = 'sell' if trade['m'] else 'buy' # self is reversed intentionally elif 'isBuyerMaker' in trade: side = 'sell' if trade['isBuyerMaker'] else 'buy' elif 'side' in trade: side = self.safe_string_lower(trade, 'side') else: if 'isBuyer' in trade: side = 'buy' if trade['isBuyer'] else 'sell' # self is a True side fee = None if 'commission' in trade: fee = { 'cost': self.safe_string(trade, 'commission'), 'currency': self.safe_currency_code(self.safe_string(trade, 'commissionAsset')), } takerOrMaker = None if 'isMaker' in trade: takerOrMaker = 'maker' if trade['isMaker'] else 'taker' if 'maker' in trade: takerOrMaker = 'maker' if trade['maker'] else 'taker' return self.safe_trade({ 'info': trade, 'timestamp': timestamp, 'datetime': self.iso8601(timestamp), 'symbol': symbol, 'id': id, 'order': orderId, 'type': None, 'side': side, 'takerOrMaker': takerOrMaker, 'price': price, 'amount': amount, 'cost': cost, 'fee': fee, }, market) def fetch_trades(self, symbol, since=None, limit=None, params={}): self.load_markets() market = self.market(symbol) request = { 'symbol': market['id'], # 'fromId': 123, # ID to get aggregate trades from INCLUSIVE. # 'startTime': 456, # Timestamp in ms to get aggregate trades from INCLUSIVE. # 'endTime': 789, # Timestamp in ms to get aggregate trades until INCLUSIVE. # 'limit': 500, # default = 500, maximum = 1000 } defaultType = self.safe_string_2(self.options, 'fetchTrades', 'defaultType', 'spot') type = self.safe_string(params, 'type', defaultType) query = self.omit(params, 'type') defaultMethod = None if type == 'future': defaultMethod = 'fapiPublicGetAggTrades' elif type == 'delivery': defaultMethod = 'dapiPublicGetAggTrades' else: defaultMethod = 'publicGetAggTrades' method = self.safe_string(self.options, 'fetchTradesMethod', defaultMethod) if method == 'publicGetAggTrades': if since is not None: request['startTime'] = since # https://github.com/ccxt/ccxt/issues/6400 # https://github.com/binance-exchange/binance-official-api-docs/blob/master/rest-api.md#compressedaggregate-trades-list request['endTime'] = self.sum(since, 3600000) if type == 'future': method = 'fapiPublicGetAggTrades' elif type == 'delivery': method = 'dapiPublicGetAggTrades' elif method == 'publicGetHistoricalTrades': if type == 'future': method = 'fapiPublicGetHistoricalTrades' elif type == 'delivery': method = 'dapiPublicGetHistoricalTrades' if limit is not None: request['limit'] = limit # default = 500, maximum = 1000 # # Caveats: # - default limit(500) applies only if no other parameters set, trades up # to the maximum limit may be returned to satisfy other parameters # - if both limit and time window is set and time window contains more # trades than the limit then the last trades from the window are returned # - 'tradeId' accepted and returned by self method is "aggregate" trade id # which is different from actual trade id # - setting both fromId and time window results in error response = getattr(self, method)(self.extend(request, query)) # # aggregate trades # # [ # { # "a": 26129, # Aggregate tradeId # "p": "0.01633102", # Price # "q": "4.70443515", # Quantity # "f": 27781, # First tradeId # "l": 27781, # Last tradeId # "T": 1498793709153, # Timestamp # "m": True, # Was the buyer the maker? # "M": True # Was the trade the best price match? # } # ] # # recent public trades and historical public trades # # [ # { # "id": 28457, # "price": "4.00000100", # "qty": "12.00000000", # "time": 1499865549590, # "isBuyerMaker": True, # "isBestMatch": True # } # ] # return self.parse_trades(response, market, since, limit) def parse_order_status(self, status): statuses = { 'NEW': 'open', 'PARTIALLY_FILLED': 'open', 'FILLED': 'closed', 'CANCELED': 'canceled', 'PENDING_CANCEL': 'canceling', # currently unused 'REJECTED': 'rejected', 'EXPIRED': 'expired', } return self.safe_string(statuses, status, status) def parse_order(self, order, market=None): # # spot # # { # "symbol": "LTCBTC", # "orderId": 1, # "clientOrderId": "myOrder1", # "price": "0.1", # "origQty": "1.0", # "executedQty": "0.0", # "cummulativeQuoteQty": "0.0", # "status": "NEW", # "timeInForce": "GTC", # "type": "LIMIT", # "side": "BUY", # "stopPrice": "0.0", # "icebergQty": "0.0", # "time": 1499827319559, # "updateTime": 1499827319559, # "isWorking": True # } # # futures # # { # "symbol": "BTCUSDT", # "orderId": 1, # "clientOrderId": "myOrder1", # "price": "0.1", # "origQty": "1.0", # "executedQty": "1.0", # "cumQuote": "10.0", # "status": "NEW", # "timeInForce": "GTC", # "type": "LIMIT", # "side": "BUY", # "stopPrice": "0.0", # "updateTime": 1499827319559 # } # # createOrder with {"newOrderRespType": "FULL"} # # { # "symbol": "BTCUSDT", # "orderId": 5403233939, # "orderListId": -1, # "clientOrderId": "x-R4BD3S825e669e75b6c14f69a2c43e", # "transactTime": 1617151923742, # "price": "0.00000000", # "origQty": "0.00050000", # "executedQty": "0.00050000", # "cummulativeQuoteQty": "29.47081500", # "status": "FILLED", # "timeInForce": "GTC", # "type": "MARKET", # "side": "BUY", # "fills": [ # { # "price": "58941.63000000", # "qty": "0.00050000", # "commission": "0.00007050", # "commissionAsset": "BNB", # "tradeId": 737466631 # } # ] # } # # delivery # # { # "orderId": "18742727411", # "symbol": "ETHUSD_PERP", # "pair": "ETHUSD", # "status": "FILLED", # "clientOrderId": "x-xcKtGhcu3e2d1503fdd543b3b02419", # "price": "0", # "avgPrice": "4522.14", # "origQty": "1", # "executedQty": "1", # "cumBase": "0.00221134", # "timeInForce": "GTC", # "type": "MARKET", # "reduceOnly": False, # "closePosition": False, # "side": "SELL", # "positionSide": "BOTH", # "stopPrice": "0", # "workingType": "CONTRACT_PRICE", # "priceProtect": False, # "origType": "MARKET", # "time": "1636061952660", # "updateTime": "1636061952660" # } # status = self.parse_order_status(self.safe_string(order, 'status')) marketId = self.safe_string(order, 'symbol') symbol = self.safe_symbol(marketId, market) filled = self.safe_string(order, 'executedQty', '0') timestamp = None lastTradeTimestamp = None if 'time' in order: timestamp = self.safe_integer(order, 'time') elif 'transactTime' in order: timestamp = self.safe_integer(order, 'transactTime') elif 'updateTime' in order: if status == 'open': if Precise.string_gt(filled, '0'): lastTradeTimestamp = self.safe_integer(order, 'updateTime') else: timestamp = self.safe_integer(order, 'updateTime') average = self.safe_string(order, 'avgPrice') price = self.safe_string(order, 'price') amount = self.safe_string(order, 'origQty') # - Spot/Margin market: cummulativeQuoteQty # - Futures market: cumQuote. # Note self is not the actual cost, since Binance futures uses leverage to calculate margins. cost = self.safe_string_2(order, 'cummulativeQuoteQty', 'cumQuote') cost = self.safe_string(order, 'cumBase', cost) id = self.safe_string(order, 'orderId') type = self.safe_string_lower(order, 'type') side = self.safe_string_lower(order, 'side') fills = self.safe_value(order, 'fills', []) clientOrderId = self.safe_string(order, 'clientOrderId') timeInForce = self.safe_string(order, 'timeInForce') postOnly = (type == 'limit_maker') or (timeInForce == 'GTX') if type == 'limit_maker': type = 'limit' stopPriceString = self.safe_string(order, 'stopPrice') stopPrice = self.parse_number(self.omit_zero(stopPriceString)) return self.safe_order2({ 'info': order, 'id': id, 'clientOrderId': clientOrderId, 'timestamp': timestamp, 'datetime': self.iso8601(timestamp), 'lastTradeTimestamp': lastTradeTimestamp, 'symbol': symbol, 'type': type, 'timeInForce': timeInForce, 'postOnly': postOnly, 'side': side, 'price': price, 'stopPrice': stopPrice, 'amount': amount, 'cost': cost, 'average': average, 'filled': filled, 'remaining': None, 'status': status, 'fee': None, 'trades': fills, }, market) def create_reduce_only_order(self, symbol, type, side, amount, price=None, params={}): request = { 'reduceOnly': True, } return self.create_order(symbol, type, side, amount, price, self.extend(request, params)) def create_order(self, symbol, type, side, amount, price=None, params={}): self.load_markets() market = self.market(symbol) defaultType = self.safe_string_2(self.options, 'createOrder', 'defaultType', 'spot') orderType = self.safe_string(params, 'type', defaultType) clientOrderId = self.safe_string_2(params, 'newClientOrderId', 'clientOrderId') postOnly = self.safe_value(params, 'postOnly', False) params = self.omit(params, ['type', 'newClientOrderId', 'clientOrderId', 'postOnly']) reduceOnly = self.safe_value(params, 'reduceOnly') if reduceOnly is not None: if (orderType != 'future') and (orderType != 'delivery'): raise InvalidOrder(self.id + ' createOrder() does not support reduceOnly for ' + orderType + ' orders, reduceOnly orders are supported for futures and perpetuals only') method = 'privatePostOrder' if orderType == 'future': method = 'fapiPrivatePostOrder' elif orderType == 'delivery': method = 'dapiPrivatePostOrder' elif orderType == 'margin': method = 'sapiPostMarginOrder' # the next 5 lines are added to support for testing orders if market['spot']: test = self.safe_value(params, 'test', False) if test: method += 'Test' params = self.omit(params, 'test') # only supported for spot/margin api(all margin markets are spot markets) if postOnly: type = 'LIMIT_MAKER' uppercaseType = type.upper() validOrderTypes = self.safe_value(market['info'], 'orderTypes') if not self.in_array(uppercaseType, validOrderTypes): raise InvalidOrder(self.id + ' ' + type + ' is not a valid order type in market ' + symbol) request = { 'symbol': market['id'], 'type': uppercaseType, 'side': side.upper(), } if clientOrderId is None: broker = self.safe_value(self.options, 'broker') if broker is not None: brokerId = self.safe_string(broker, orderType) if brokerId is not None: request['newClientOrderId'] = brokerId + self.uuid22() else: request['newClientOrderId'] = clientOrderId if (orderType == 'spot') or (orderType == 'margin'): request['newOrderRespType'] = self.safe_value(self.options['newOrderRespType'], type, 'RESULT') # 'ACK' for order id, 'RESULT' for full order or 'FULL' for order with fills else: # delivery and future request['newOrderRespType'] = 'RESULT' # "ACK", "RESULT", default "ACK" # additional required fields depending on the order type timeInForceIsRequired = False priceIsRequired = False stopPriceIsRequired = False quantityIsRequired = False # # spot/margin # # LIMIT timeInForce, quantity, price # MARKET quantity or quoteOrderQty # STOP_LOSS quantity, stopPrice # STOP_LOSS_LIMIT timeInForce, quantity, price, stopPrice # TAKE_PROFIT quantity, stopPrice # TAKE_PROFIT_LIMIT timeInForce, quantity, price, stopPrice # LIMIT_MAKER quantity, price # # futures # # LIMIT timeInForce, quantity, price # MARKET quantity # STOP/TAKE_PROFIT quantity, price, stopPrice # STOP_MARKET stopPrice # TAKE_PROFIT_MARKET stopPrice # TRAILING_STOP_MARKET callbackRate # if uppercaseType == 'MARKET': quoteOrderQty = self.safe_value(self.options, 'quoteOrderQty', False) if quoteOrderQty: quoteOrderQty = self.safe_number(params, 'quoteOrderQty') precision = market['precision']['price'] if quoteOrderQty is not None: request['quoteOrderQty'] = self.decimal_to_precision(quoteOrderQty, TRUNCATE, precision, self.precisionMode) params = self.omit(params, 'quoteOrderQty') elif price is not None: request['quoteOrderQty'] = self.decimal_to_precision(amount * price, TRUNCATE, precision, self.precisionMode) else: quantityIsRequired = True else: quantityIsRequired = True elif uppercaseType == 'LIMIT': priceIsRequired = True timeInForceIsRequired = True quantityIsRequired = True elif (uppercaseType == 'STOP_LOSS') or (uppercaseType == 'TAKE_PROFIT'): stopPriceIsRequired = True quantityIsRequired = True if market['linear'] or market['inverse']: priceIsRequired = True elif (uppercaseType == 'STOP_LOSS_LIMIT') or (uppercaseType == 'TAKE_PROFIT_LIMIT'): quantityIsRequired = True stopPriceIsRequired = True priceIsRequired = True timeInForceIsRequired = True elif uppercaseType == 'LIMIT_MAKER': priceIsRequired = True quantityIsRequired = True elif uppercaseType == 'STOP': quantityIsRequired = True stopPriceIsRequired = True priceIsRequired = True elif (uppercaseType == 'STOP_MARKET') or (uppercaseType == 'TAKE_PROFIT_MARKET'): closePosition = self.safe_value(params, 'closePosition') if closePosition is None: quantityIsRequired = True stopPriceIsRequired = True elif uppercaseType == 'TRAILING_STOP_MARKET': quantityIsRequired = True callbackRate = self.safe_number(params, 'callbackRate') if callbackRate is None: raise InvalidOrder(self.id + ' createOrder() requires a callbackRate extra param for a ' + type + ' order') if quantityIsRequired: request['quantity'] = self.amount_to_precision(symbol, amount) if priceIsRequired: if price is None: raise InvalidOrder(self.id + ' createOrder() requires a price argument for a ' + type + ' order') request['price'] = self.price_to_precision(symbol, price) if timeInForceIsRequired: request['timeInForce'] = self.options['defaultTimeInForce'] # 'GTC' = Good To Cancel(default), 'IOC' = Immediate Or Cancel if stopPriceIsRequired: stopPrice = self.safe_number(params, 'stopPrice') if stopPrice is None: raise InvalidOrder(self.id + ' createOrder() requires a stopPrice extra param for a ' + type + ' order') else: params = self.omit(params, 'stopPrice') request['stopPrice'] = self.price_to_precision(symbol, stopPrice) response = getattr(self, method)(self.extend(request, params)) return self.parse_order(response, market) def fetch_order(self, id, symbol=None, params={}): if symbol is None: raise ArgumentsRequired(self.id + ' fetchOrder() requires a symbol argument') self.load_markets() market = self.market(symbol) defaultType = self.safe_string_2(self.options, 'fetchOrder', 'defaultType', 'spot') type = self.safe_string(params, 'type', defaultType) method = 'privateGetOrder' if type == 'future': method = 'fapiPrivateGetOrder' elif type == 'delivery': method = 'dapiPrivateGetOrder' elif type == 'margin': method = 'sapiGetMarginOrder' request = { 'symbol': market['id'], } clientOrderId = self.safe_value_2(params, 'origClientOrderId', 'clientOrderId') if clientOrderId is not None: request['origClientOrderId'] = clientOrderId else: request['orderId'] = id query = self.omit(params, ['type', 'clientOrderId', 'origClientOrderId']) response = getattr(self, method)(self.extend(request, query)) return self.parse_order(response, market) def fetch_orders(self, symbol=None, since=None, limit=None, params={}): if symbol is None: raise ArgumentsRequired(self.id + ' fetchOrders() requires a symbol argument') self.load_markets() market = self.market(symbol) defaultType = self.safe_string_2(self.options, 'fetchOrders', 'defaultType', 'spot') type = self.safe_string(params, 'type', defaultType) method = 'privateGetAllOrders' if type == 'future': method = 'fapiPrivateGetAllOrders' elif type == 'delivery': method = 'dapiPrivateGetAllOrders' elif type == 'margin': method = 'sapiGetMarginAllOrders' request = { 'symbol': market['id'], } if since is not None: request['startTime'] = since if limit is not None: request['limit'] = limit query = self.omit(params, 'type') response = getattr(self, method)(self.extend(request, query)) # # spot # # [ # { # "symbol": "LTCBTC", # "orderId": 1, # "clientOrderId": "myOrder1", # "price": "0.1", # "origQty": "1.0", # "executedQty": "0.0", # "cummulativeQuoteQty": "0.0", # "status": "NEW", # "timeInForce": "GTC", # "type": "LIMIT", # "side": "BUY", # "stopPrice": "0.0", # "icebergQty": "0.0", # "time": 1499827319559, # "updateTime": 1499827319559, # "isWorking": True # } # ] # # futures # # [ # { # "symbol": "BTCUSDT", # "orderId": 1, # "clientOrderId": "myOrder1", # "price": "0.1", # "origQty": "1.0", # "executedQty": "1.0", # "cumQuote": "10.0", # "status": "NEW", # "timeInForce": "GTC", # "type": "LIMIT", # "side": "BUY", # "stopPrice": "0.0", # "updateTime": 1499827319559 # } # ] # return self.parse_orders(response, market, since, limit) def fetch_open_orders(self, symbol=None, since=None, limit=None, params={}): self.load_markets() market = None query = None type = None request = {} if symbol is not None: market = self.market(symbol) request['symbol'] = market['id'] defaultType = self.safe_string_2(self.options, 'fetchOpenOrders', 'defaultType', 'spot') type = self.safe_string(params, 'type', defaultType) query = self.omit(params, 'type') elif self.options['warnOnFetchOpenOrdersWithoutSymbol']: symbols = self.symbols numSymbols = len(symbols) fetchOpenOrdersRateLimit = int(numSymbols / 2) raise ExchangeError(self.id + ' fetchOpenOrders WARNING: fetching open orders without specifying a symbol is rate-limited to one call per ' + str(fetchOpenOrdersRateLimit) + ' seconds. Do not call self method frequently to avoid ban. Set ' + self.id + '.options["warnOnFetchOpenOrdersWithoutSymbol"] = False to suppress self warning message.') else: defaultType = self.safe_string_2(self.options, 'fetchOpenOrders', 'defaultType', 'spot') type = self.safe_string(params, 'type', defaultType) query = self.omit(params, 'type') method = 'privateGetOpenOrders' if type == 'future': method = 'fapiPrivateGetOpenOrders' elif type == 'delivery': method = 'dapiPrivateGetOpenOrders' elif type == 'margin': method = 'sapiGetMarginOpenOrders' response = getattr(self, method)(self.extend(request, query)) return self.parse_orders(response, market, since, limit) def fetch_closed_orders(self, symbol=None, since=None, limit=None, params={}): orders = self.fetch_orders(symbol, since, limit, params) return self.filter_by(orders, 'status', 'closed') def cancel_order(self, id, symbol=None, params={}): if symbol is None: raise ArgumentsRequired(self.id + ' cancelOrder() requires a symbol argument') self.load_markets() market = self.market(symbol) defaultType = self.safe_string_2(self.options, 'fetchOpenOrders', 'defaultType', 'spot') type = self.safe_string(params, 'type', defaultType) # https://github.com/ccxt/ccxt/issues/6507 origClientOrderId = self.safe_value_2(params, 'origClientOrderId', 'clientOrderId') request = { 'symbol': market['id'], # 'orderId': id, # 'origClientOrderId': id, } if origClientOrderId is None: request['orderId'] = id else: request['origClientOrderId'] = origClientOrderId method = 'privateDeleteOrder' if type == 'future': method = 'fapiPrivateDeleteOrder' elif type == 'delivery': method = 'dapiPrivateDeleteOrder' elif type == 'margin': method = 'sapiDeleteMarginOrder' query = self.omit(params, ['type', 'origClientOrderId', 'clientOrderId']) response = getattr(self, method)(self.extend(request, query)) return self.parse_order(response, market) def cancel_all_orders(self, symbol=None, params={}): if symbol is None: raise ArgumentsRequired(self.id + ' cancelAllOrders() requires a symbol argument') self.load_markets() market = self.market(symbol) request = { 'symbol': market['id'], } defaultType = self.safe_string_2(self.options, 'cancelAllOrders', 'defaultType', 'spot') type = self.safe_string(params, 'type', defaultType) query = self.omit(params, 'type') method = 'privateDeleteOpenOrders' if type == 'margin': method = 'sapiDeleteMarginOpenOrders' elif type == 'future': method = 'fapiPrivateDeleteAllOpenOrders' elif type == 'delivery': method = 'dapiPrivateDeleteAllOpenOrders' response = getattr(self, method)(self.extend(request, query)) if isinstance(response, list): return self.parse_orders(response, market) else: return response def fetch_my_trades(self, symbol=None, since=None, limit=None, params={}): if symbol is None: raise ArgumentsRequired(self.id + ' fetchMyTrades() requires a symbol argument') self.load_markets() market = self.market(symbol) defaultType = self.safe_string_2(self.options, 'fetchMyTrades', 'defaultType', 'spot') type = self.safe_string(params, 'type', defaultType) params = self.omit(params, 'type') method = None if type == 'spot': method = 'privateGetMyTrades' elif type == 'margin': method = 'sapiGetMarginMyTrades' elif type == 'future': method = 'fapiPrivateGetUserTrades' elif type == 'delivery': method = 'dapiPrivateGetUserTrades' request = { 'symbol': market['id'], } if since is not None: request['startTime'] = since if limit is not None: request['limit'] = limit response = getattr(self, method)(self.extend(request, params)) # # spot trade # # [ # { # "symbol": "BNBBTC", # "id": 28457, # "orderId": 100234, # "price": "4.00000100", # "qty": "12.00000000", # "commission": "10.10000000", # "commissionAsset": "BNB", # "time": 1499865549590, # "isBuyer": True, # "isMaker": False, # "isBestMatch": True, # } # ] # # futures trade # # [ # { # "accountId": 20, # "buyer": False, # "commission": "-0.07819010", # "commissionAsset": "USDT", # "counterPartyId": 653, # "id": 698759, # "maker": False, # "orderId": 25851813, # "price": "7819.01", # "qty": "0.002", # "quoteQty": "0.01563", # "realizedPnl": "-0.91539999", # "side": "SELL", # "symbol": "BTCUSDT", # "time": 1569514978020 # } # ] # return self.parse_trades(response, market, since, limit) def fetch_my_dust_trades(self, symbol=None, since=None, limit=None, params={}): # # Binance provides an opportunity to trade insignificant(i.e. non-tradable and non-withdrawable) # token leftovers(of any asset) into `BNB` coin which in turn can be used to pay trading fees with it. # The corresponding trades history is called the `Dust Log` and can be requested via the following end-point: # https://github.com/binance-exchange/binance-official-api-docs/blob/master/wapi-api.md#dustlog-user_data # self.load_markets() request = {} if since is not None: request['startTime'] = since request['endTime'] = self.sum(since, 7776000000) response = self.sapiGetAssetDribblet(self.extend(request, params)) # { # "total": "4", # "userAssetDribblets": [ # { # "operateTime": "1627575731000", # "totalServiceChargeAmount": "0.00001453", # "totalTransferedAmount": "0.00072693", # "transId": "70899815863", # "userAssetDribbletDetails": [ # { # "fromAsset": "LTC", # "amount": "0.000006", # "transferedAmount": "0.00000267", # "serviceChargeAmount": "0.00000005", # "operateTime": "1627575731000", # "transId": "70899815863" # }, # { # "fromAsset": "GBP", # "amount": "0.15949157", # "transferedAmount": "0.00072426", # "serviceChargeAmount": "0.00001448", # "operateTime": "1627575731000", # "transId": "70899815863" # } # ] # }, # ] # } results = self.safe_value(response, 'userAssetDribblets', []) rows = self.safe_integer(response, 'total', 0) data = [] for i in range(0, rows): logs = self.safe_value(results[i], 'userAssetDribbletDetails', []) for j in range(0, len(logs)): logs[j]['isDustTrade'] = True data.append(logs[j]) trades = self.parse_trades(data, None, since, limit) return self.filter_by_since_limit(trades, since, limit) def parse_dust_trade(self, trade, market=None): # # { # "fromAsset": "USDT", # "amount": "0.009669", # "transferedAmount": "0.00002992", # "serviceChargeAmount": "0.00000059", # "operateTime": "1628076010000", # "transId": "71416578712", # "isDustTrade": True # } # orderId = self.safe_string(trade, 'transId') timestamp = self.safe_integer(trade, 'operateTime') currencyId = self.safe_string(trade, 'fromAsset') tradedCurrency = self.safe_currency_code(currencyId) bnb = self.currency('BNB') earnedCurrency = bnb['code'] applicantSymbol = earnedCurrency + '/' + tradedCurrency tradedCurrencyIsQuote = False if applicantSymbol in self.markets: tradedCurrencyIsQuote = True feeCostString = self.safe_string(trade, 'serviceChargeAmount') fee = { 'currency': earnedCurrency, 'cost': self.parse_number(feeCostString), } symbol = None amountString = None costString = None side = None if tradedCurrencyIsQuote: symbol = applicantSymbol amountString = self.safe_string(trade, 'transferedAmount') costString = self.safe_string(trade, 'amount') side = 'buy' else: symbol = tradedCurrency + '/' + earnedCurrency amountString = self.safe_string(trade, 'amount') costString = self.safe_string(trade, 'transferedAmount') side = 'sell' priceString = None if costString is not None: if amountString: priceString = Precise.string_div(costString, amountString) id = None amount = self.parse_number(amountString) price = self.parse_number(priceString) cost = self.parse_number(costString) type = None takerOrMaker = None return { 'id': id, 'timestamp': timestamp, 'datetime': self.iso8601(timestamp), 'symbol': symbol, 'order': orderId, 'type': type, 'takerOrMaker': takerOrMaker, 'side': side, 'amount': amount, 'price': price, 'cost': cost, 'fee': fee, 'info': trade, } def fetch_deposits(self, code=None, since=None, limit=None, params={}): self.load_markets() currency = None response = None request = {} legalMoney = self.safe_value(self.options, 'legalMoney', {}) if code in legalMoney: if code is not None: currency = self.currency(code) request['transactionType'] = 0 if since is not None: request['beginTime'] = since raw = self.sapiGetFiatOrders(self.extend(request, params)) response = self.safe_value(raw, 'data') # { # "code": "000000", # "message": "success", # "data": [ # { # "orderNo": "25ced37075c1470ba8939d0df2316e23", # "fiatCurrency": "EUR", # "indicatedAmount": "15.00", # "amount": "15.00", # "totalFee": "0.00", # "method": "card", # "status": "Failed", # "createTime": 1627501026000, # "updateTime": 1627501027000 # } # ], # "total": 1, # "success": True # } else: if code is not None: currency = self.currency(code) request['coin'] = currency['id'] if since is not None: request['startTime'] = since # max 3 months range https://github.com/ccxt/ccxt/issues/6495 request['endTime'] = self.sum(since, 7776000000) if limit is not None: request['limit'] = limit response = self.sapiGetCapitalDepositHisrec(self.extend(request, params)) # [ # { # "amount": "0.01844487", # "coin": "BCH", # "network": "BCH", # "status": 1, # "address": "1NYxAJhW2281HK1KtJeaENBqHeygA88FzR", # "addressTag": "", # "txId": "bafc5902504d6504a00b7d0306a41154cbf1d1b767ab70f3bc226327362588af", # "insertTime": 1610784980000, # "transferType": 0, # "confirmTimes": "2/2" # }, # { # "amount": "4500", # "coin": "USDT", # "network": "BSC", # "status": 1, # "address": "0xc9c923c87347ca0f3451d6d308ce84f691b9f501", # "addressTag": "", # "txId": "Internal transfer 51376627901", # "insertTime": 1618394381000, # "transferType": 1, # "confirmTimes": "1/15" # } # ] return self.parse_transactions(response, currency, since, limit) def fetch_withdrawals(self, code=None, since=None, limit=None, params={}): self.load_markets() legalMoney = self.safe_value(self.options, 'legalMoney', {}) request = {} response = None currency = None if code in legalMoney: if code is not None: currency = self.currency(code) request['transactionType'] = 1 if since is not None: request['beginTime'] = since raw = self.sapiGetFiatOrders(self.extend(request, params)) response = self.safe_value(raw, 'data') # { # "code": "000000", # "message": "success", # "data": [ # { # "orderNo": "CJW706452266115170304", # "fiatCurrency": "GBP", # "indicatedAmount": "10001.50", # "amount": "100.00", # "totalFee": "1.50", # "method": "bank transfer", # "status": "Successful", # "createTime": 1620037745000, # "updateTime": 1620038480000 # }, # { # "orderNo": "CJW706287492781891584", # "fiatCurrency": "GBP", # "indicatedAmount": "10001.50", # "amount": "100.00", # "totalFee": "1.50", # "method": "bank transfer", # "status": "Successful", # "createTime": 1619998460000, # "updateTime": 1619998823000 # } # ], # "total": 39, # "success": True # } else: if code is not None: currency = self.currency(code) request['coin'] = currency['id'] if since is not None: request['startTime'] = since # max 3 months range https://github.com/ccxt/ccxt/issues/6495 request['endTime'] = self.sum(since, 7776000000) if limit is not None: request['limit'] = limit response = self.sapiGetCapitalWithdrawHistory(self.extend(request, params)) # [ # { # "id": "69e53ad305124b96b43668ceab158a18", # "amount": "28.75", # "transactionFee": "0.25", # "coin": "XRP", # "status": 6, # "address": "r3T75fuLjX51mmfb5Sk1kMNuhBgBPJsjza", # "addressTag": "101286922", # "txId": "19A5B24ED0B697E4F0E9CD09FCB007170A605BC93C9280B9E6379C5E6EF0F65A", # "applyTime": "2021-04-15 12:09:16", # "network": "XRP", # "transferType": 0 # }, # { # "id": "9a67628b16ba4988ae20d329333f16bc", # "amount": "20", # "transactionFee": "20", # "coin": "USDT", # "status": 6, # "address": "0x0AB991497116f7F5532a4c2f4f7B1784488628e1", # "txId": "0x77fbf2cf2c85b552f0fd31fd2e56dc95c08adae031d96f3717d8b17e1aea3e46", # "applyTime": "2021-04-15 12:06:53", # "network": "ETH", # "transferType": 0 # }, # { # "id": "a7cdc0afbfa44a48bd225c9ece958fe2", # "amount": "51", # "transactionFee": "1", # "coin": "USDT", # "status": 6, # "address": "TYDmtuWL8bsyjvcauUTerpfYyVhFtBjqyo", # "txId": "168a75112bce6ceb4823c66726ad47620ad332e69fe92d9cb8ceb76023f9a028", # "applyTime": "2021-04-13 12:46:59", # "network": "TRX", # "transferType": 0 # } # ] return self.parse_transactions(response, currency, since, limit) def parse_transaction_status_by_type(self, status, type=None): statusesByType = { 'deposit': { '0': 'pending', '1': 'ok', # Fiat # Processing, Failed, Successful, Finished, Refunding, Refunded, Refund Failed, Order Partial credit Stopped 'Processing': 'pending', 'Failed': 'failed', 'Successful': 'ok', 'Refunding': 'canceled', 'Refunded': 'canceled', 'Refund Failed': 'failed', }, 'withdrawal': { '0': 'pending', # Email Sent '1': 'canceled', # Cancelled(different from 1 = ok in deposits) '2': 'pending', # Awaiting Approval '3': 'failed', # Rejected '4': 'pending', # Processing '5': 'failed', # Failure '6': 'ok', # Completed # Fiat # Processing, Failed, Successful, Finished, Refunding, Refunded, Refund Failed, Order Partial credit Stopped 'Processing': 'pending', 'Failed': 'failed', 'Successful': 'ok', 'Refunding': 'canceled', 'Refunded': 'canceled', 'Refund Failed': 'failed', }, } statuses = self.safe_value(statusesByType, type, {}) return self.safe_string(statuses, status, status) def parse_transaction(self, transaction, currency=None): # # fetchDeposits # # { # "amount": "4500", # "coin": "USDT", # "network": "BSC", # "status": 1, # "address": "0xc9c923c87347ca0f3451d6d308ce84f691b9f501", # "addressTag": "", # "txId": "Internal transfer 51376627901", # "insertTime": 1618394381000, # "transferType": 1, # "confirmTimes": "1/15" # } # # fetchWithdrawals # # { # "id": "69e53ad305124b96b43668ceab158a18", # "amount": "28.75", # "transactionFee": "0.25", # "coin": "XRP", # "status": 6, # "address": "r3T75fuLjX51mmfb5Sk1kMNuhBgBPJsjza", # "addressTag": "101286922", # "txId": "19A5B24ED0B697E4F0E9CD09FCB007170A605BC93C9280B9E6379C5E6EF0F65A", # "applyTime": "2021-04-15 12:09:16", # "network": "XRP", # "transferType": 0 # } # # fiat transaction # withdraw # { # "orderNo": "CJW684897551397171200", # "fiatCurrency": "GBP", # "indicatedAmount": "29.99", # "amount": "28.49", # "totalFee": "1.50", # "method": "bank transfer", # "status": "Successful", # "createTime": 1614898701000, # "updateTime": 1614898820000 # } # # deposit # { # "orderNo": "25ced37075c1470ba8939d0df2316e23", # "fiatCurrency": "EUR", # "indicatedAmount": "15.00", # "amount": "15.00", # "totalFee": "0.00", # "method": "card", # "status": "Failed", # "createTime": "1627501026000", # "updateTime": "1627501027000" # } # id = self.safe_string_2(transaction, 'id', 'orderNo') address = self.safe_string(transaction, 'address') tag = self.safe_string(transaction, 'addressTag') # set but unused if tag is not None: if len(tag) < 1: tag = None txid = self.safe_string(transaction, 'txId') if (txid is not None) and (txid.find('Internal transfer ') >= 0): txid = txid[18:] currencyId = self.safe_string_2(transaction, 'coin', 'fiatCurrency') code = self.safe_currency_code(currencyId, currency) timestamp = None insertTime = self.safe_integer_2(transaction, 'insertTime', 'createTime') applyTime = self.parse8601(self.safe_string(transaction, 'applyTime')) type = self.safe_string(transaction, 'type') if type is None: if (insertTime is not None) and (applyTime is None): type = 'deposit' timestamp = insertTime elif (insertTime is None) and (applyTime is not None): type = 'withdrawal' timestamp = applyTime status = self.parse_transaction_status_by_type(self.safe_string(transaction, 'status'), type) amount = self.safe_number(transaction, 'amount') feeCost = self.safe_number_2(transaction, 'transactionFee', 'totalFee') fee = None if feeCost is not None: fee = {'currency': code, 'cost': feeCost} updated = self.safe_integer_2(transaction, 'successTime', 'updateTime') internal = self.safe_integer(transaction, 'transferType', False) internal = True if internal else False return { 'info': transaction, 'id': id, 'txid': txid, 'timestamp': timestamp, 'datetime': self.iso8601(timestamp), 'address': address, 'addressTo': address, 'addressFrom': None, 'tag': tag, 'tagTo': tag, 'tagFrom': None, 'type': type, 'amount': amount, 'currency': code, 'status': status, 'updated': updated, 'internal': internal, 'fee': fee, } def parse_transfer_status(self, status): statuses = { 'CONFIRMED': 'ok', } return self.safe_string(statuses, status, status) def parse_transfer(self, transfer, currency=None): # # transfer # # { # "tranId":13526853623 # } # # fetchTransfers # # { # timestamp: 1614640878000, # asset: 'USDT', # amount: '25', # type: 'MAIN_UMFUTURE', # status: 'CONFIRMED', # tranId: 43000126248 # } # id = self.safe_string(transfer, 'tranId') currencyId = self.safe_string(transfer, 'asset') code = self.safe_currency_code(currencyId, currency) amount = self.safe_number(transfer, 'amount') type = self.safe_string(transfer, 'type') fromAccount = None toAccount = None typesByAccount = self.safe_value(self.options, 'typesByAccount', {}) if type is not None: parts = type.split('_') fromAccount = self.safe_value(parts, 0) toAccount = self.safe_value(parts, 1) fromAccount = self.safe_string(typesByAccount, fromAccount, fromAccount) toAccount = self.safe_string(typesByAccount, toAccount, toAccount) timestamp = self.safe_integer(transfer, 'timestamp') status = self.parse_transfer_status(self.safe_string(transfer, 'status')) return { 'info': transfer, 'id': id, 'timestamp': timestamp, 'datetime': self.iso8601(timestamp), 'currency': code, 'amount': amount, 'fromAccount': fromAccount, 'toAccount': toAccount, 'status': status, } def parse_income(self, income, market=None): # # { # "symbol": "ETHUSDT", # "incomeType": "FUNDING_FEE", # "income": "0.00134317", # "asset": "USDT", # "time": "1621584000000", # "info": "FUNDING_FEE", # "tranId": "4480321991774044580", # "tradeId": "" # } # marketId = self.safe_string(income, 'symbol') symbol = self.safe_symbol(marketId, market) amount = self.safe_number(income, 'income') currencyId = self.safe_string(income, 'asset') code = self.safe_currency_code(currencyId) id = self.safe_string(income, 'tranId') timestamp = self.safe_integer(income, 'time') return { 'info': income, 'symbol': symbol, 'code': code, 'timestamp': timestamp, 'datetime': self.iso8601(timestamp), 'id': id, 'amount': amount, } def parse_incomes(self, incomes, market=None, since=None, limit=None): result = [] for i in range(0, len(incomes)): entry = incomes[i] parsed = self.parse_income(entry, market) result.append(parsed) sorted = self.sort_by(result, 'timestamp') return self.filter_by_since_limit(sorted, since, limit) def transfer(self, code, amount, fromAccount, toAccount, params={}): self.load_markets() currency = self.currency(code) type = self.safe_string(params, 'type') if type is None: accountsByType = self.safe_value(self.options, 'accountsByType', {}) fromAccount = fromAccount.lower() toAccount = toAccount.lower() fromId = self.safe_string(accountsByType, fromAccount) toId = self.safe_string(accountsByType, toAccount) if fromId is None: keys = list(accountsByType.keys()) raise ExchangeError(self.id + ' fromAccount must be one of ' + ', '.join(keys)) if toId is None: keys = list(accountsByType.keys()) raise ExchangeError(self.id + ' toAccount must be one of ' + ', '.join(keys)) type = fromId + '_' + toId request = { 'asset': currency['id'], 'amount': self.currency_to_precision(code, amount), 'type': type, } response = self.sapiPostAssetTransfer(self.extend(request, params)) # # { # "tranId":13526853623 # } # transfer = self.parse_transfer(response, currency) return self.extend(transfer, { 'amount': amount, 'currency': code, 'fromAccount': fromAccount, 'toAccount': toAccount, }) def fetch_transfers(self, code=None, since=None, limit=None, params={}): self.load_markets() currency = None if code is not None: currency = self.currency(code) defaultType = self.safe_string_2(self.options, 'fetchTransfers', 'defaultType', 'spot') fromAccount = self.safe_string(params, 'fromAccount', defaultType) defaultTo = 'spot' if (fromAccount == 'future') else 'future' toAccount = self.safe_string(params, 'toAccount', defaultTo) type = self.safe_string(params, 'type') accountsByType = self.safe_value(self.options, 'accountsByType', {}) fromId = self.safe_string(accountsByType, fromAccount) toId = self.safe_string(accountsByType, toAccount) if type is None: if fromId is None: keys = list(accountsByType.keys()) raise ExchangeError(self.id + ' fromAccount parameter must be one of ' + ', '.join(keys)) if toId is None: keys = list(accountsByType.keys()) raise ExchangeError(self.id + ' toAccount parameter must be one of ' + ', '.join(keys)) type = fromId + '_' + toId request = { 'type': type, } if since is not None: request['startTime'] = since if limit is not None: request['size'] = limit response = self.sapiGetAssetTransfer(self.extend(request, params)) # # { # total: 3, # rows: [ # { # timestamp: 1614640878000, # asset: 'USDT', # amount: '25', # type: 'MAIN_UMFUTURE', # status: 'CONFIRMED', # tranId: 43000126248 # }, # ] # } # rows = self.safe_value(response, 'rows', []) return self.parse_transfers(rows, currency, since, limit) def fetch_deposit_address(self, code, params={}): self.load_markets() currency = self.currency(code) request = { 'coin': currency['id'], # 'network': 'ETH', # 'BSC', 'XMR', you can get network and isDefault in networkList in the response of sapiGetCapitalConfigDetail } networks = self.safe_value(self.options, 'networks', {}) network = self.safe_string_upper(params, 'network') # self line allows the user to specify either ERC20 or ETH network = self.safe_string(networks, network, network) # handle ERC20>ETH alias if network is not None: request['network'] = network params = self.omit(params, 'network') # has support for the 'network' parameter # https://binance-docs.github.io/apidocs/spot/en/#deposit-address-supporting-network-user_data response = self.sapiGetCapitalDepositAddress(self.extend(request, params)) # # { # currency: 'XRP', # address: 'rEb8TK3gBgk5auZkwc6sHnwrGVJH8DuaLh', # tag: '108618262', # info: { # coin: 'XRP', # address: 'rEb8TK3gBgk5auZkwc6sHnwrGVJH8DuaLh', # tag: '108618262', # url: 'https://bithomp.com/explorer/rEb8TK3gBgk5auZkwc6sHnwrGVJH8DuaLh' # } # } # address = self.safe_string(response, 'address') url = self.safe_string(response, 'url') impliedNetwork = None if url is not None: reverseNetworks = self.safe_value(self.options, 'reverseNetworks', {}) parts = url.split('/') topLevel = self.safe_string(parts, 2) if (topLevel == 'blockchair.com') or (topLevel == 'viewblock.io'): subLevel = self.safe_string(parts, 3) if subLevel is not None: topLevel = topLevel + '/' + subLevel impliedNetwork = self.safe_string(reverseNetworks, topLevel) impliedNetworks = self.safe_value(self.options, 'impliedNetworks', { 'ETH': {'ERC20': 'ETH'}, 'TRX': {'TRC20': 'TRX'}, }) if code in impliedNetworks: conversion = self.safe_value(impliedNetworks, code, {}) impliedNetwork = self.safe_string(conversion, impliedNetwork, impliedNetwork) tag = self.safe_string(response, 'tag', '') if len(tag) == 0: tag = None self.check_address(address) return { 'currency': code, 'address': address, 'tag': tag, 'network': impliedNetwork, 'info': response, } def fetch_funding_fees(self, codes=None, params={}): self.load_markets() response = self.sapiGetCapitalConfigGetall(params) # # [ # { # coin: 'BAT', # depositAllEnable: True, # withdrawAllEnable: True, # name: 'Basic Attention Token', # free: '0', # locked: '0', # freeze: '0', # withdrawing: '0', # ipoing: '0', # ipoable: '0', # storage: '0', # isLegalMoney: False, # trading: True, # networkList: [ # { # network: 'BNB', # coin: 'BAT', # withdrawIntegerMultiple: '0.00000001', # isDefault: False, # depositEnable: True, # withdrawEnable: True, # depositDesc: '', # withdrawDesc: '', # specialTips: 'The name of self asset is Basic Attention Token(BAT). Both a MEMO and an Address are required to successfully deposit your BEP2 tokens to Binance.', # name: 'BEP2', # resetAddressStatus: False, # addressRegex: '^(bnb1)[0-9a-z]{38}$', # memoRegex: '^[0-9A-Za-z\\-_]{1,120}$', # withdrawFee: '0.27', # withdrawMin: '0.54', # withdrawMax: '10000000000', # minConfirm: '1', # unLockConfirm: '0' # }, # { # network: 'BSC', # coin: 'BAT', # withdrawIntegerMultiple: '0.00000001', # isDefault: False, # depositEnable: True, # withdrawEnable: True, # depositDesc: '', # withdrawDesc: '', # specialTips: 'The name of self asset is Basic Attention Token. Please ensure you are depositing Basic Attention Token(BAT) tokens under the contract address ending in 9766e.', # name: 'BEP20(BSC)', # resetAddressStatus: False, # addressRegex: '^(0x)[0-9A-Fa-f]{40}$', # memoRegex: '', # withdrawFee: '0.27', # withdrawMin: '0.54', # withdrawMax: '10000000000', # minConfirm: '15', # unLockConfirm: '0' # }, # { # network: 'ETH', # coin: 'BAT', # withdrawIntegerMultiple: '0.00000001', # isDefault: True, # depositEnable: True, # withdrawEnable: True, # depositDesc: '', # withdrawDesc: '', # specialTips: 'The name of self asset is Basic Attention Token. Please ensure you are depositing Basic Attention Token(BAT) tokens under the contract address ending in 887ef.', # name: 'ERC20', # resetAddressStatus: False, # addressRegex: '^(0x)[0-9A-Fa-f]{40}$', # memoRegex: '', # withdrawFee: '27', # withdrawMin: '54', # withdrawMax: '10000000000', # minConfirm: '12', # unLockConfirm: '0' # } # ] # } # ] # withdrawFees = {} for i in range(0, len(response)): entry = response[i] currencyId = self.safe_string(entry, 'coin') code = self.safe_currency_code(currencyId) networkList = self.safe_value(entry, 'networkList') withdrawFees[code] = {} for j in range(0, len(networkList)): networkEntry = networkList[j] networkId = self.safe_string(networkEntry, 'network') networkCode = self.safe_currency_code(networkId) fee = self.safe_number(networkEntry, 'withdrawFee') withdrawFees[code][networkCode] = fee return { 'withdraw': withdrawFees, 'deposit': {}, 'info': response, } def withdraw(self, code, amount, address, tag=None, params={}): tag, params = self.handle_withdraw_tag_and_params(tag, params) self.check_address(address) self.load_markets() currency = self.currency(code) request = { 'coin': currency['id'], 'address': address, 'amount': amount, # https://binance-docs.github.io/apidocs/spot/en/#withdraw-sapi # issue sapiGetCapitalConfigGetall() to get networks for withdrawing USDT ERC20 vs USDT Omni # 'network': 'ETH', # 'BTC', 'TRX', etc, optional } if tag is not None: request['addressTag'] = tag networks = self.safe_value(self.options, 'networks', {}) network = self.safe_string_upper(params, 'network') # self line allows the user to specify either ERC20 or ETH network = self.safe_string(networks, network, network) # handle ERC20>ETH alias if network is not None: request['network'] = network params = self.omit(params, 'network') response = self.sapiPostCapitalWithdrawApply(self.extend(request, params)) # {id: '9a67628b16ba4988ae20d329333f16bc'} return { 'info': response, 'id': self.safe_string(response, 'id'), } def parse_trading_fee(self, fee, market=None): # # { # "symbol": "ADABNB", # "makerCommission": 0.001, # "takerCommission": 0.001 # } # marketId = self.safe_string(fee, 'symbol') symbol = self.safe_symbol(marketId) return { 'info': fee, 'symbol': symbol, 'maker': self.safe_number(fee, 'makerCommission'), 'taker': self.safe_number(fee, 'takerCommission'), } def fetch_trading_fee(self, symbol, params={}): self.load_markets() market = self.market(symbol) request = { 'symbol': market['id'], } response = self.sapiGetAssetTradeFee(self.extend(request, params)) # # [ # { # "symbol": "BTCUSDT", # "makerCommission": "0.001", # "takerCommission": "0.001" # } # ] # first = self.safe_value(response, 0, {}) return self.parse_trading_fee(first) def fetch_trading_fees(self, params={}): self.load_markets() method = None defaultType = self.safe_string_2(self.options, 'fetchFundingRates', 'defaultType', 'future') type = self.safe_string(params, 'type', defaultType) query = self.omit(params, 'type') if (type == 'spot') or (type == 'margin'): method = 'sapiGetAssetTradeFee' elif type == 'future': method = 'fapiPrivateGetAccount' elif type == 'delivery': method = 'dapiPrivateGetAccount' response = getattr(self, method)(query) # # sapi / spot # # [ # { # "symbol": "ZRXBNB", # "makerCommission": "0.001", # "takerCommission": "0.001" # }, # { # "symbol": "ZRXBTC", # "makerCommission": "0.001", # "takerCommission": "0.001" # }, # ] # # fapi / future / linear # # { # "feeTier": 0, # account commisssion tier # "canTrade": True, # if can trade # "canDeposit": True, # if can transfer in asset # "canWithdraw": True, # if can transfer out asset # "updateTime": 0, # "totalInitialMargin": "0.00000000", # total initial margin required with current mark price(useless with isolated positions), only for USDT asset # "totalMaintMargin": "0.00000000", # total maintenance margin required, only for USDT asset # "totalWalletBalance": "23.72469206", # total wallet balance, only for USDT asset # "totalUnrealizedProfit": "0.00000000", # total unrealized profit, only for USDT asset # "totalMarginBalance": "23.72469206", # total margin balance, only for USDT asset # "totalPositionInitialMargin": "0.00000000", # initial margin required for positions with current mark price, only for USDT asset # "totalOpenOrderInitialMargin": "0.00000000", # initial margin required for open orders with current mark price, only for USDT asset # "totalCrossWalletBalance": "23.72469206", # crossed wallet balance, only for USDT asset # "totalCrossUnPnl": "0.00000000", # unrealized profit of crossed positions, only for USDT asset # "availableBalance": "23.72469206", # available balance, only for USDT asset # "maxWithdrawAmount": "23.72469206" # maximum amount for transfer out, only for USDT asset # ... # } # # dapi / delivery / inverse # # { # "canDeposit": True, # "canTrade": True, # "canWithdraw": True, # "feeTier": 2, # "updateTime": 0 # } # if (type == 'spot') or (type == 'margin'): # # [ # { # "symbol": "ZRXBNB", # "makerCommission": "0.001", # "takerCommission": "0.001" # }, # { # "symbol": "ZRXBTC", # "makerCommission": "0.001", # "takerCommission": "0.001" # }, # ] # result = {} for i in range(0, len(response)): fee = self.parse_trading_fee(response[i]) symbol = fee['symbol'] result[symbol] = fee return result elif type == 'future': # # { # "feeTier": 0, # account commisssion tier # "canTrade": True, # if can trade # "canDeposit": True, # if can transfer in asset # "canWithdraw": True, # if can transfer out asset # "updateTime": 0, # "totalInitialMargin": "0.00000000", # total initial margin required with current mark price(useless with isolated positions), only for USDT asset # "totalMaintMargin": "0.00000000", # total maintenance margin required, only for USDT asset # "totalWalletBalance": "23.72469206", # total wallet balance, only for USDT asset # "totalUnrealizedProfit": "0.00000000", # total unrealized profit, only for USDT asset # "totalMarginBalance": "23.72469206", # total margin balance, only for USDT asset # "totalPositionInitialMargin": "0.00000000", # initial margin required for positions with current mark price, only for USDT asset # "totalOpenOrderInitialMargin": "0.00000000", # initial margin required for open orders with current mark price, only for USDT asset # "totalCrossWalletBalance": "23.72469206", # crossed wallet balance, only for USDT asset # "totalCrossUnPnl": "0.00000000", # unrealized profit of crossed positions, only for USDT asset # "availableBalance": "23.72469206", # available balance, only for USDT asset # "maxWithdrawAmount": "23.72469206" # maximum amount for transfer out, only for USDT asset # ... # } # symbols = list(self.markets.keys()) result = {} feeTier = self.safe_integer(response, 'feeTier') feeTiers = self.fees[type]['trading']['tiers'] maker = feeTiers['maker'][feeTier][1] taker = feeTiers['taker'][feeTier][1] for i in range(0, len(symbols)): symbol = symbols[i] result[symbol] = { 'info': { 'feeTier': feeTier, }, 'symbol': symbol, 'maker': maker, 'taker': taker, } return result elif type == 'delivery': # # { # "canDeposit": True, # "canTrade": True, # "canWithdraw": True, # "feeTier": 2, # "updateTime": 0 # } # symbols = list(self.markets.keys()) result = {} feeTier = self.safe_integer(response, 'feeTier') feeTiers = self.fees[type]['trading']['tiers'] maker = feeTiers['maker'][feeTier][1] taker = feeTiers['taker'][feeTier][1] for i in range(0, len(symbols)): symbol = symbols[i] result[symbol] = { 'info': { 'feeTier': feeTier, }, 'symbol': symbol, 'maker': maker, 'taker': taker, } return result def futures_transfer(self, code, amount, type, params={}): if (type < 1) or (type > 4): raise ArgumentsRequired(self.id + ' type must be between 1 and 4') self.load_markets() currency = self.currency(code) request = { 'asset': currency['id'], 'amount': amount, 'type': type, } response = self.sapiPostFuturesTransfer(self.extend(request, params)) # # { # "tranId": 100000001 # } # return self.parse_transfer(response, currency) def fetch_funding_rate(self, symbol, params={}): self.load_markets() market = self.market(symbol) request = { 'symbol': market['id'], } method = None if market['linear']: method = 'fapiPublicGetPremiumIndex' elif market['inverse']: method = 'dapiPublicGetPremiumIndex' else: raise NotSupported(self.id + ' fetchFundingRate() supports linear and inverse contracts only') response = getattr(self, method)(self.extend(request, params)) if market['inverse']: response = response[0] # # { # "symbol": "BTCUSDT", # "markPrice": "45802.81129892", # "indexPrice": "45745.47701915", # "estimatedSettlePrice": "45133.91753671", # "lastFundingRate": "0.00063521", # "interestRate": "0.00010000", # "nextFundingTime": "1621267200000", # "time": "1621252344001" # } # return self.parse_funding_rate(response, market) def fetch_funding_rate_history(self, symbol=None, since=None, limit=None, params={}): # # Gets a history of funding rates with their timestamps # (param) symbol: Future currency pair(e.g. "BTC/USDT") # (param) limit: maximum number of data points returned # (param) since: Unix timestamp in miliseconds for the time of the earliest requested funding rate # (param) params: Object containing more params for the request # - until: Unix timestamp in miliseconds for the time of the earliest requested funding rate # return: [{symbol, fundingRate, timestamp}] # self.load_markets() request = {} method = None defaultType = self.safe_string_2(self.options, 'fetchFundingRateHistory', 'defaultType', 'future') type = self.safe_string(params, 'type', defaultType) params = self.omit(params, 'type') if type == 'future': method = 'fapiPublicGetFundingRate' elif type == 'delivery': method = 'dapiPublicGetFundingRate' if symbol is not None: market = self.market(symbol) request['symbol'] = market['id'] if market['linear']: method = 'fapiPublicGetFundingRate' elif market['inverse']: method = 'dapiPublicGetFundingRate' if method is None: raise NotSupported(self.id + ' fetchFundingRateHistory() not supported for ' + type + ' markets') if since is not None: request['startTime'] = since till = self.safe_integer(params, 'till') # unified in milliseconds endTime = self.safe_string(params, 'endTime', till) # exchange-specific in milliseconds params = self.omit(params, ['endTime', 'till']) if endTime is not None: request['endTime'] = endTime if limit is not None: request['limit'] = limit response = getattr(self, method)(self.extend(request, params)) # # { # "symbol": "BTCUSDT", # "fundingRate": "0.00063521", # "fundingTime": "1621267200000", # } # rates = [] for i in range(0, len(response)): entry = response[i] timestamp = self.safe_integer(entry, 'fundingTime') rates.append({ 'info': entry, 'symbol': self.safe_symbol(self.safe_string(entry, 'symbol')), 'fundingRate': self.safe_number(entry, 'fundingRate'), 'timestamp': timestamp, 'datetime': self.iso8601(timestamp), }) sorted = self.sort_by(rates, 'timestamp') return self.filter_by_symbol_since_limit(sorted, symbol, since, limit) def fetch_funding_rates(self, symbols=None, params={}): self.load_markets() method = None defaultType = self.safe_string_2(self.options, 'fetchFundingRates', 'defaultType', 'future') type = self.safe_string(params, 'type', defaultType) query = self.omit(params, 'type') if type == 'future': method = 'fapiPublicGetPremiumIndex' elif type == 'delivery': method = 'dapiPublicGetPremiumIndex' else: raise NotSupported(self.id + ' fetchFundingRates() supports linear and inverse contracts only') response = getattr(self, method)(query) result = [] for i in range(0, len(response)): entry = response[i] parsed = self.parse_funding_rate(entry) result.append(parsed) return self.filter_by_array(result, 'symbol', symbols) def parse_funding_rate(self, premiumIndex, market=None): # ensure it matches with https://www.binance.com/en/futures/funding-history/0 # # { # "symbol": "BTCUSDT", # "markPrice": "45802.81129892", # "indexPrice": "45745.47701915", # "estimatedSettlePrice": "45133.91753671", # "lastFundingRate": "0.00063521", # "interestRate": "0.00010000", # "nextFundingTime": "1621267200000", # "time": "1621252344001" # } # timestamp = self.safe_integer(premiumIndex, 'time') marketId = self.safe_string(premiumIndex, 'symbol') symbol = self.safe_symbol(marketId, market) markPrice = self.safe_number(premiumIndex, 'markPrice') indexPrice = self.safe_number(premiumIndex, 'indexPrice') interestRate = self.safe_number(premiumIndex, 'interestRate') estimatedSettlePrice = self.safe_number(premiumIndex, 'estimatedSettlePrice') nextFundingRate = self.safe_number(premiumIndex, 'lastFundingRate') nextFundingTime = self.safe_integer(premiumIndex, 'nextFundingTime') previousFundingTime = nextFundingTime - (8 * 3600000) return { 'info': premiumIndex, 'symbol': symbol, 'markPrice': markPrice, 'indexPrice': indexPrice, 'interestRate': interestRate, 'estimatedSettlePrice': estimatedSettlePrice, 'timestamp': timestamp, 'datetime': self.iso8601(timestamp), 'previousFundingRate': None, 'nextFundingRate': nextFundingRate, 'previousFundingTimestamp': previousFundingTime, # subtract 8 hours 'nextFundingTimestamp': nextFundingTime, 'previousFundingDatetime': self.iso8601(previousFundingTime), 'nextFundingDatetime': self.iso8601(nextFundingTime), } def parse_account_positions(self, account): positions = self.safe_value(account, 'positions') assets = self.safe_value(account, 'assets') balances = {} for i in range(0, len(assets)): entry = assets[i] currencyId = self.safe_string(entry, 'asset') code = self.safe_currency_code(currencyId) crossWalletBalance = self.safe_string(entry, 'crossWalletBalance') crossUnPnl = self.safe_string(entry, 'crossUnPnl') balances[code] = { 'crossMargin': Precise.string_add(crossWalletBalance, crossUnPnl), 'crossWalletBalance': crossWalletBalance, } result = [] for i in range(0, len(positions)): position = positions[i] marketId = self.safe_string(position, 'symbol') market = self.safe_market(marketId) code = market['quote'] if (self.options['defaultType'] == 'future') else market['base'] # sometimes not all the codes are correctly returned... if code in balances: parsed = self.parse_account_position(self.extend(position, { 'crossMargin': balances[code]['crossMargin'], 'crossWalletBalance': balances[code]['crossWalletBalance'], }), market) result.append(parsed) return result def parse_account_position(self, position, market=None): # # usdm # { # "symbol": "BTCBUSD", # "initialMargin": "0", # "maintMargin": "0", # "unrealizedProfit": "0.00000000", # "positionInitialMargin": "0", # "openOrderInitialMargin": "0", # "leverage": "20", # "isolated": False, # "entryPrice": "0.0000", # "maxNotional": "100000", # "positionSide": "BOTH", # "positionAmt": "0.000", # "notional": "0", # "isolatedWallet": "0", # "updateTime": "0", # "crossMargin": "100.93634809", # } # # coinm # { # "symbol": "BTCUSD_210625", # "initialMargin": "0.00024393", # "maintMargin": "0.00002439", # "unrealizedProfit": "-0.00000163", # "positionInitialMargin": "0.00024393", # "openOrderInitialMargin": "0", # "leverage": "10", # "isolated": False, # "positionSide": "BOTH", # "entryPrice": "41021.20000069", # "maxQty": "100", # "notionalValue": "0.00243939", # "isolatedWallet": "0", # "crossMargin": "0.314" # "crossWalletBalance": "34", # } # marketId = self.safe_string(position, 'symbol') market = self.safe_market(marketId, market) symbol = market['symbol'] leverageString = self.safe_string(position, 'leverage') leverage = int(leverageString) initialMarginString = self.safe_string(position, 'initialMargin') initialMargin = self.parse_number(initialMarginString) initialMarginPercentageString = Precise.string_div('1', leverageString, 8) rational = (1000 % leverage) == 0 if not rational: initialMarginPercentageString = Precise.string_div(Precise.string_add(initialMarginPercentageString, '1e-8'), '1', 8) usdm = ('notional' in position) maintenanceMarginString = self.safe_string(position, 'maintMargin') maintenanceMargin = self.parse_number(maintenanceMarginString) entryPriceString = self.safe_string(position, 'entryPrice') entryPrice = self.parse_number(entryPriceString) notionalString = self.safe_string_2(position, 'notional', 'notionalValue') notionalStringAbs = Precise.string_abs(notionalString) notionalFloat = float(notionalString) notionalFloatAbs = float(notionalStringAbs) notional = self.parse_number(Precise.string_abs(notionalString)) contractsString = self.safe_string(position, 'positionAmt') contractsStringAbs = Precise.string_abs(contractsString) if contractsString is None: entryNotional = Precise.string_mul(Precise.string_mul(leverageString, initialMarginString), entryPriceString) contractsString = Precise.string_div(entryNotional, market['contractSize']) contractsStringAbs = Precise.string_div(Precise.string_add(contractsString, '0.5'), '1', 0) contracts = self.parse_number(contractsStringAbs) leverageBrackets = self.safe_value(self.options, 'leverageBrackets', {}) leverageBracket = self.safe_value(leverageBrackets, symbol, []) maintenanceMarginPercentageString = None for i in range(0, len(leverageBracket)): bracket = leverageBracket[i] if notionalFloatAbs < bracket[0]: break maintenanceMarginPercentageString = bracket[1] maintenanceMarginPercentage = self.parse_number(maintenanceMarginPercentageString) unrealizedPnlString = self.safe_string(position, 'unrealizedProfit') unrealizedPnl = self.parse_number(unrealizedPnlString) timestamp = self.safe_integer(position, 'updateTime') if timestamp == 0: timestamp = None isolated = self.safe_value(position, 'isolated') marginType = None collateralString = None walletBalance = None if isolated: marginType = 'isolated' walletBalance = self.safe_string(position, 'isolatedWallet') collateralString = Precise.string_add(walletBalance, unrealizedPnlString) else: marginType = 'cross' walletBalance = self.safe_string(position, 'crossWalletBalance') collateralString = self.safe_string(position, 'crossMargin') collateral = self.parse_number(collateralString) marginRatio = None side = None percentage = None liquidationPriceStringRaw = None liquidationPrice = None if notionalFloat == 0.0: entryPrice = None else: side = 'short' if (notionalFloat < 0) else 'long' marginRatio = self.parse_number(Precise.string_div(Precise.string_add(Precise.string_div(maintenanceMarginString, collateralString), '5e-5'), '1', 4)) percentage = self.parse_number(Precise.string_mul(Precise.string_div(unrealizedPnlString, initialMarginString, 4), '100')) if usdm: # calculate liquidation price # # liquidationPrice = (walletBalance / (contracts * (±1 + mmp))) + (±entryPrice / (±1 + mmp)) # # mmp = maintenanceMarginPercentage # where ± is negative for long and positive for short # TODO: calculate liquidation price for coinm contracts onePlusMaintenanceMarginPercentageString = None entryPriceSignString = entryPriceString if side == 'short': onePlusMaintenanceMarginPercentageString = Precise.string_add('1', maintenanceMarginPercentageString) else: onePlusMaintenanceMarginPercentageString = Precise.string_add('-1', maintenanceMarginPercentageString) entryPriceSignString = Precise.string_mul('-1', entryPriceSignString) leftSide = Precise.string_div(walletBalance, Precise.string_mul(contractsStringAbs, onePlusMaintenanceMarginPercentageString)) rightSide = Precise.string_div(entryPriceSignString, onePlusMaintenanceMarginPercentageString) liquidationPriceStringRaw = Precise.string_add(leftSide, rightSide) else: # calculate liquidation price # # liquidationPrice = (contracts * contractSize(±1 - mmp)) / (±1/entryPrice * contracts * contractSize - walletBalance) # onePlusMaintenanceMarginPercentageString = None entryPriceSignString = entryPriceString if side == 'short': onePlusMaintenanceMarginPercentageString = Precise.string_sub('1', maintenanceMarginPercentageString) else: onePlusMaintenanceMarginPercentageString = Precise.string_sub('-1', maintenanceMarginPercentageString) entryPriceSignString = Precise.string_mul('-1', entryPriceSignString) size = Precise.string_mul(contractsStringAbs, market['contractSize']) leftSide = Precise.string_mul(size, onePlusMaintenanceMarginPercentageString) rightSide = Precise.string_sub(Precise.string_mul(Precise.string_div('1', entryPriceSignString), size), walletBalance) liquidationPriceStringRaw = Precise.string_div(leftSide, rightSide) pricePrecision = market['precision']['price'] pricePrecisionPlusOne = pricePrecision + 1 pricePrecisionPlusOneString = str(pricePrecisionPlusOne) # round half up rounder = Precise('5e-' + pricePrecisionPlusOneString) rounderString = str(rounder) liquidationPriceRoundedString = Precise.string_add(rounderString, liquidationPriceStringRaw) truncatedLiquidationPrice = Precise.string_div(liquidationPriceRoundedString, '1', pricePrecision) if truncatedLiquidationPrice[0] == '-': # user cannot be liquidated # since he has more collateral than the size of the position truncatedLiquidationPrice = None liquidationPrice = self.parse_number(truncatedLiquidationPrice) positionSide = self.safe_string(position, 'positionSide') hedged = positionSide != 'BOTH' return { 'info': position, 'symbol': symbol, 'timestamp': timestamp, 'datetime': self.iso8601(timestamp), 'initialMargin': initialMargin, 'initialMarginPercentage': self.parse_number(initialMarginPercentageString), 'maintenanceMargin': maintenanceMargin, 'maintenanceMarginPercentage': maintenanceMarginPercentage, 'entryPrice': entryPrice, 'notional': notional, 'leverage': self.parse_number(leverageString), 'unrealizedPnl': unrealizedPnl, 'contracts': contracts, 'contractSize': self.parse_number(market['contractSize']), 'marginRatio': marginRatio, 'liquidationPrice': liquidationPrice, 'markPrice': None, 'collateral': collateral, 'marginType': marginType, 'side': side, 'hedged': hedged, 'percentage': percentage, } def parse_position_risk(self, position, market=None): # # usdm # { # "symbol": "BTCUSDT", # "positionAmt": "0.001", # "entryPrice": "43578.07000", # "markPrice": "43532.30000000", # "unRealizedProfit": "-0.04577000", # "liquidationPrice": "21841.24993976", # "leverage": "2", # "maxNotionalValue": "300000000", # "marginType": "isolated", # "isolatedMargin": "21.77841506", # "isAutoAddMargin": "false", # "positionSide": "BOTH", # "notional": "43.53230000", # "isolatedWallet": "21.82418506", # "updateTime": "1621358023886" # } # # coinm # { # "symbol": "BTCUSD_PERP", # "positionAmt": "2", # "entryPrice": "37643.10000021", # "markPrice": "38103.05510455", # "unRealizedProfit": "0.00006413", # "liquidationPrice": "25119.97445760", # "leverage": "2", # "maxQty": "1500", # "marginType": "isolated", # "isolatedMargin": "0.00274471", # "isAutoAddMargin": "false", # "positionSide": "BOTH", # "notionalValue": "0.00524892", # "isolatedWallet": "0.00268058" # } # marketId = self.safe_string(position, 'symbol') market = self.safe_market(marketId, market) symbol = market['symbol'] leverageBrackets = self.safe_value(self.options, 'leverageBrackets', {}) leverageBracket = self.safe_value(leverageBrackets, symbol, []) notionalString = self.safe_string_2(position, 'notional', 'notionalValue') notionalStringAbs = Precise.string_abs(notionalString) notionalFloatAbs = float(notionalStringAbs) notionalFloat = float(notionalString) maintenanceMarginPercentageString = None for i in range(0, len(leverageBracket)): bracket = leverageBracket[i] if notionalFloatAbs < bracket[0]: break maintenanceMarginPercentageString = bracket[1] notional = self.parse_number(notionalStringAbs) contractsAbs = Precise.string_abs(self.safe_string(position, 'positionAmt')) contracts = self.parse_number(contractsAbs) unrealizedPnlString = self.safe_string(position, 'unRealizedProfit') unrealizedPnl = self.parse_number(unrealizedPnlString) leverageString = self.safe_string(position, 'leverage') leverage = int(leverageString) liquidationPriceString = self.omit_zero(self.safe_string(position, 'liquidationPrice')) liquidationPrice = self.parse_number(liquidationPriceString) collateralString = None marginType = self.safe_string(position, 'marginType') side = None if notionalFloat > 0: side = 'long' elif notionalFloat < 0: side = 'short' entryPriceString = self.safe_string(position, 'entryPrice') entryPrice = self.parse_number(entryPriceString) if marginType == 'cross': # calculate collateral if market['linear']: # walletBalance = (liquidationPrice * (±1 + mmp) ± entryPrice) * contracts onePlusMaintenanceMarginPercentageString = None entryPriceSignString = entryPriceString if side == 'short': onePlusMaintenanceMarginPercentageString = Precise.string_add('1', maintenanceMarginPercentageString) entryPriceSignString = Precise.string_mul('-1', entryPriceSignString) else: onePlusMaintenanceMarginPercentageString = Precise.string_add('-1', maintenanceMarginPercentageString) inner = Precise.string_mul(liquidationPriceString, onePlusMaintenanceMarginPercentageString) leftSide = Precise.string_add(inner, entryPriceSignString) collateralString = Precise.string_div(Precise.string_mul(leftSide, contractsAbs), '1', market['precision']['quote']) else: # walletBalance = (contracts * contractSize) * (±1/entryPrice - (±1 - mmp) / liquidationPrice) onePlusMaintenanceMarginPercentageString = None entryPriceSignString = entryPriceString if side == 'short': onePlusMaintenanceMarginPercentageString = Precise.string_sub('1', maintenanceMarginPercentageString) else: onePlusMaintenanceMarginPercentageString = Precise.string_sub('-1', maintenanceMarginPercentageString) entryPriceSignString = Precise.string_mul('-1', entryPriceSignString) leftSide = Precise.string_mul(contractsAbs, market['contractSize']) rightSide = Precise.string_sub(Precise.string_div('1', entryPriceSignString), Precise.string_div(onePlusMaintenanceMarginPercentageString, liquidationPriceString)) collateralString = Precise.string_div(Precise.string_mul(leftSide, rightSide), '1', market['precision']['base']) else: collateralString = self.safe_string(position, 'isolatedMargin') collateralString = '0' if (collateralString is None) else collateralString collateralFloat = float(collateralString) collateral = self.parse_number(collateralString) markPrice = self.parse_number(self.omit_zero(self.safe_string(position, 'markPrice'))) timestamp = self.safe_integer(position, 'updateTime') if timestamp == 0: timestamp = None maintenanceMarginPercentage = self.parse_number(maintenanceMarginPercentageString) maintenanceMarginString = Precise.string_mul(maintenanceMarginPercentageString, notionalStringAbs) maintenanceMargin = self.parse_number(maintenanceMarginString) initialMarginPercentageString = Precise.string_div('1', leverageString, 8) rational = (1000 % leverage) == 0 if not rational: initialMarginPercentageString = Precise.string_add(initialMarginPercentageString, '1e-8') initialMarginString = Precise.string_div(Precise.string_mul(notionalStringAbs, initialMarginPercentageString), '1', 8) initialMargin = self.parse_number(initialMarginString) marginRatio = None percentage = None if collateralFloat != 0.0: marginRatio = self.parse_number(Precise.string_div(Precise.string_add(Precise.string_div(maintenanceMarginString, collateralString), '5e-5'), '1', 4)) percentage = self.parse_number(Precise.string_mul(Precise.string_div(unrealizedPnlString, initialMarginString, 4), '100')) positionSide = self.safe_string(position, 'positionSide') hedged = positionSide != 'BOTH' return { 'info': position, 'symbol': symbol, 'contracts': contracts, 'contractSize': self.parse_number(market['contractSize']), 'unrealizedPnl': unrealizedPnl, 'leverage': self.parse_number(leverageString), 'liquidationPrice': liquidationPrice, 'collateral': collateral, 'notional': notional, 'markPrice': markPrice, 'entryPrice': entryPrice, 'timestamp': timestamp, 'initialMargin': initialMargin, 'initialMarginPercentage': self.parse_number(initialMarginPercentageString), 'maintenanceMargin': maintenanceMargin, 'maintenanceMarginPercentage': maintenanceMarginPercentage, 'marginRatio': marginRatio, 'datetime': self.iso8601(timestamp), 'marginType': marginType, 'side': side, 'hedged': hedged, 'percentage': percentage, } def load_leverage_brackets(self, reload=False, params={}): self.load_markets() # by default cache the leverage bracket # it contains useful stuff like the maintenance margin and initial margin for positions leverageBrackets = self.safe_value(self.options, 'leverageBrackets') if (leverageBrackets is None) or (reload): method = None defaultType = self.safe_string(self.options, 'defaultType', 'future') type = self.safe_string(params, 'type', defaultType) query = self.omit(params, 'type') if type == 'future': method = 'fapiPrivateGetLeverageBracket' elif type == 'delivery': method = 'dapiPrivateV2GetLeverageBracket' else: raise NotSupported(self.id + ' loadLeverageBrackets() supports linear and inverse contracts only') response = getattr(self, method)(query) self.options['leverageBrackets'] = {} for i in range(0, len(response)): entry = response[i] marketId = self.safe_string(entry, 'symbol') symbol = self.safe_symbol(marketId) brackets = self.safe_value(entry, 'brackets') result = [] for j in range(0, len(brackets)): bracket = brackets[j] # we use floats here internally on purpose floorValue = self.safe_float_2(bracket, 'notionalFloor', 'qtyFloor') maintenanceMarginPercentage = self.safe_string(bracket, 'maintMarginRatio') result.append([floorValue, maintenanceMarginPercentage]) self.options['leverageBrackets'][symbol] = result return self.options['leverageBrackets'] def fetch_positions(self, symbols=None, params={}): defaultMethod = self.safe_string(self.options, 'fetchPositions', 'positionRisk') if defaultMethod == 'positionRisk': return self.fetch_positions_risk(symbols, params) elif defaultMethod == 'account': return self.fetch_account_positions(symbols, params) else: raise NotSupported(self.id + '.options["fetchPositions"] = "' + defaultMethod + '" is invalid, please choose between "account" and "positionRisk"') def fetch_account_positions(self, symbols=None, params={}): if symbols is not None: if not isinstance(symbols, list): raise ArgumentsRequired(self.id + ' fetchPositions requires an array argument for symbols') self.load_markets() self.load_leverage_brackets() method = None defaultType = self.safe_string(self.options, 'defaultType', 'future') type = self.safe_string(params, 'type', defaultType) query = self.omit(params, 'type') if type == 'future': method = 'fapiPrivateGetAccount' elif type == 'delivery': method = 'dapiPrivateGetAccount' else: raise NotSupported(self.id + ' fetchPositions() supports linear and inverse contracts only') account = getattr(self, method)(query) result = self.parse_account_positions(account) return self.filter_by_array(result, 'symbol', symbols, False) def fetch_positions_risk(self, symbols=None, params={}): if symbols is not None: if not isinstance(symbols, list): raise ArgumentsRequired(self.id + ' fetchPositions requires an array argument for symbols') self.load_markets() self.load_leverage_brackets() request = {} method = None defaultType = 'future' defaultType = self.safe_string(self.options, 'defaultType', defaultType) type = self.safe_string(params, 'type', defaultType) params = self.omit(params, 'type') if (type == 'future') or (type == 'linear'): method = 'fapiPrivateGetPositionRisk' elif (type == 'delivery') or (type == 'inverse'): method = 'dapiPrivateGetPositionRisk' else: raise NotSupported(self.id + ' fetchIsolatedPositions() supports linear and inverse contracts only') response = getattr(self, method)(self.extend(request, params)) result = [] for i in range(0, len(response)): parsed = self.parse_position_risk(response[i]) result.append(parsed) return self.filter_by_array(result, 'symbol', symbols, False) def fetch_funding_history(self, symbol=None, since=None, limit=None, params={}): self.load_markets() market = None method = None defaultType = 'future' request = { 'incomeType': 'FUNDING_FEE', # "TRANSFER","WELCOME_BONUS", "REALIZED_PNL","FUNDING_FEE", "COMMISSION" and "INSURANCE_CLEAR" } if symbol is not None: market = self.market(symbol) request['symbol'] = market['id'] if market['linear']: defaultType = 'future' elif market['inverse']: defaultType = 'delivery' else: raise NotSupported(self.id + ' fetchFundingHistory() supports linear and inverse contracts only') if since is not None: request['startTime'] = since if limit is not None: request['limit'] = limit defaultType = self.safe_string_2(self.options, 'fetchFundingHistory', 'defaultType', defaultType) type = self.safe_string(params, 'type', defaultType) params = self.omit(params, 'type') if (type == 'future') or (type == 'linear'): method = 'fapiPrivateGetIncome' elif (type == 'delivery') or (type == 'inverse'): method = 'dapiPrivateGetIncome' else: raise NotSupported(self.id + ' fetchFundingHistory() supports linear and inverse contracts only') response = getattr(self, method)(self.extend(request, params)) return self.parse_incomes(response, market, since, limit) def set_leverage(self, leverage, symbol=None, params={}): if symbol is None: raise ArgumentsRequired(self.id + ' setLeverage() requires a symbol argument') # WARNING: THIS WILL INCREASE LIQUIDATION PRICE FOR OPEN ISOLATED LONG POSITIONS # AND DECREASE LIQUIDATION PRICE FOR OPEN ISOLATED SHORT POSITIONS if (leverage < 1) or (leverage > 125): raise BadRequest(self.id + ' leverage should be between 1 and 125') self.load_markets() market = self.market(symbol) method = None if market['linear']: method = 'fapiPrivatePostLeverage' elif market['inverse']: method = 'dapiPrivatePostLeverage' else: raise NotSupported(self.id + ' setLeverage() supports linear and inverse contracts only') request = { 'symbol': market['id'], 'leverage': leverage, } return getattr(self, method)(self.extend(request, params)) def set_margin_mode(self, marginType, symbol=None, params={}): # # {"code": -4048 , "msg": "Margin type cannot be changed if there exists position."} # # or # # {"code": 200, "msg": "success"} # marginType = marginType.upper() if (marginType != 'ISOLATED') and (marginType != 'CROSSED'): raise BadRequest(self.id + ' marginType must be either isolated or crossed') self.load_markets() market = self.market(symbol) method = None if market['linear']: method = 'fapiPrivatePostMarginType' elif market['inverse']: method = 'dapiPrivatePostMarginType' else: raise NotSupported(self.id + ' setMarginMode() supports linear and inverse contracts only') request = { 'symbol': market['id'], 'marginType': marginType, } return getattr(self, method)(self.extend(request, params)) def set_position_mode(self, hedged, symbol=None, params={}): defaultType = self.safe_string(self.options, 'defaultType', 'future') type = self.safe_string(params, 'type', defaultType) params = self.omit(params, ['type']) dualSidePosition = None if hedged: dualSidePosition = 'true' else: dualSidePosition = 'false' request = { 'dualSidePosition': dualSidePosition, } method = None if type == 'delivery': method = 'dapiPrivatePostPositionSideDual' else: # default to future method = 'fapiPrivatePostPositionSideDual' # # { # "code": 200, # "msg": "success" # } # return getattr(self, method)(self.extend(request, params)) def sign(self, path, api='public', method='GET', params={}, headers=None, body=None): if not (api in self.urls['api']): raise NotSupported(self.id + ' does not have a testnet/sandbox URL for ' + api + ' endpoints') url = self.urls['api'][api] url += '/' + path if api == 'wapi': url += '.html' if path == 'historicalTrades': if self.apiKey: headers = { 'X-MBX-APIKEY': self.apiKey, } else: raise AuthenticationError(self.id + ' historicalTrades endpoint requires `apiKey` credential') userDataStream = (path == 'userDataStream') or (path == 'listenKey') if userDataStream: if self.apiKey: # v1 special case for userDataStream headers = { 'X-MBX-APIKEY': self.apiKey, 'Content-Type': 'application/x-www-form-urlencoded', } if method != 'GET': body = self.urlencode(params) else: raise AuthenticationError(self.id + ' userDataStream endpoint requires `apiKey` credential') elif (api == 'private') or (api == 'sapi') or (api == 'wapi' and path != 'systemStatus') or (api == 'dapiPrivate') or (api == 'dapiPrivateV2') or (api == 'fapiPrivate') or (api == 'fapiPrivateV2'): self.check_required_credentials() query = None recvWindow = self.safe_integer(self.options, 'recvWindow', 5000) if (api == 'sapi') and (path == 'asset/dust'): query = self.urlencode_with_array_repeat(self.extend({ 'timestamp': self.nonce(), 'recvWindow': recvWindow, }, params)) elif (path == 'batchOrders') or (path.find('sub-account') >= 0): query = self.rawencode(self.extend({ 'timestamp': self.nonce(), 'recvWindow': recvWindow, }, params)) else: query = self.urlencode(self.extend({ 'timestamp': self.nonce(), 'recvWindow': recvWindow, }, params)) signature = self.hmac(self.encode(query), self.encode(self.secret)) query += '&' + 'signature=' + signature headers = { 'X-MBX-APIKEY': self.apiKey, } if (method == 'GET') or (method == 'DELETE') or (api == 'wapi'): url += '?' + query else: body = query headers['Content-Type'] = 'application/x-www-form-urlencoded' else: if params: url += '?' + self.urlencode(params) return {'url': url, 'method': method, 'body': body, 'headers': headers} def handle_errors(self, code, reason, url, method, headers, body, response, requestHeaders, requestBody): if (code == 418) or (code == 429): raise DDoSProtection(self.id + ' ' + str(code) + ' ' + reason + ' ' + body) # error response in a form: {"code": -1013, "msg": "Invalid quantity."} # following block cointains legacy checks against message patterns in "msg" property # will switch "code" checks eventually, when we know all of them if code >= 400: if body.find('Price * QTY is zero or less') >= 0: raise InvalidOrder(self.id + ' order cost = amount * price is zero or less ' + body) if body.find('LOT_SIZE') >= 0: raise InvalidOrder(self.id + ' order amount should be evenly divisible by lot size ' + body) if body.find('PRICE_FILTER') >= 0: raise InvalidOrder(self.id + ' order price is invalid, i.e. exceeds allowed price precision, exceeds min price or max price limits or is invalid float value in general, use self.price_to_precision(symbol, amount) ' + body) if response is None: return # fallback to default error handler # check success value for wapi endpoints # response in format {'msg': 'The coin does not exist.', 'success': True/false} success = self.safe_value(response, 'success', True) if not success: message = self.safe_string(response, 'msg') parsedMessage = None if message is not None: try: parsedMessage = json.loads(message) except Exception as e: # do nothing parsedMessage = None if parsedMessage is not None: response = parsedMessage message = self.safe_string(response, 'msg') if message is not None: self.throw_exactly_matched_exception(self.exceptions['exact'], message, self.id + ' ' + message) self.throw_broadly_matched_exception(self.exceptions['broad'], message, self.id + ' ' + message) # checks against error codes error = self.safe_string(response, 'code') if error is not None: # https://github.com/ccxt/ccxt/issues/6501 # https://github.com/ccxt/ccxt/issues/7742 if (error == '200') or Precise.string_equals(error, '0'): return # a workaround for {"code":-2015,"msg":"Invalid API-key, IP, or permissions for action."} # despite that their message is very confusing, it is raised by Binance # on a temporary ban, the API key is valid, but disabled for a while if (error == '-2015') and self.options['hasAlreadyAuthenticatedSuccessfully']: raise DDoSProtection(self.id + ' temporary banned: ' + body) feedback = self.id + ' ' + body self.throw_exactly_matched_exception(self.exceptions['exact'], error, feedback) raise ExchangeError(feedback) if not success: raise ExchangeError(self.id + ' ' + body) def calculate_rate_limiter_cost(self, api, method, path, params, config={}, context={}): if ('noSymbol' in config) and not ('symbol' in params): return config['noSymbol'] elif ('noPoolId' in config) and not ('poolId' in params): return config['noPoolId'] elif ('byLimit' in config) and ('limit' in params): limit = params['limit'] byLimit = config['byLimit'] for i in range(0, len(byLimit)): entry = byLimit[i] if limit <= entry[0]: return entry[1] return self.safe_integer(config, 'cost', 1) def request(self, path, api='public', method='GET', params={}, headers=None, body=None, config={}, context={}): response = self.fetch2(path, api, method, params, headers, body, config, context) # a workaround for {"code":-2015,"msg":"Invalid API-key, IP, or permissions for action."} if (api == 'private') or (api == 'wapi'): self.options['hasAlreadyAuthenticatedSuccessfully'] = True return response def modify_margin_helper(self, symbol, amount, addOrReduce, params={}): # used to modify isolated positions defaultType = self.safe_string(self.options, 'defaultType', 'future') if defaultType == 'spot': defaultType = 'future' type = self.safe_string(params, 'type', defaultType) if (type == 'margin') or (type == 'spot'): raise NotSupported(self.id + ' add / reduce margin only supported with type future or delivery') self.load_markets() market = self.market(symbol) request = { 'type': addOrReduce, 'symbol': market['id'], 'amount': amount, } method = None code = None if type == 'future': method = 'fapiPrivatePostPositionMargin' code = market['quote'] else: method = 'dapiPrivatePostPositionMargin' code = market['base'] response = getattr(self, method)(self.extend(request, params)) # # { # "code": 200, # "msg": "Successfully modify position margin.", # "amount": 0.001, # "type": 1 # } # rawType = self.safe_integer(response, 'type') resultType = 'add' if (rawType == 1) else 'reduce' resultAmount = self.safe_number(response, 'amount') errorCode = self.safe_string(response, 'code') status = 'ok' if (errorCode == '200') else 'failed' return { 'info': response, 'type': resultType, 'amount': resultAmount, 'code': code, 'symbol': market['symbol'], 'status': status, } def reduce_margin(self, symbol, amount, params={}): return self.modify_margin_helper(symbol, amount, 2, params) def add_margin(self, symbol, amount, params={}): return self.modify_margin_helper(symbol, amount, 1, params) def fetch_borrow_rate(self, code, params={}): self.load_markets() currency = self.currency(code) request = { 'asset': currency['id'], # 'vipLevel': self.safe_integer(params, 'vipLevel'), } response = self.sapiGetMarginInterestRateHistory(self.extend(request, params)) # # [ # { # "asset": "USDT", # "timestamp": 1638230400000, # "dailyInterestRate": "0.0006", # "vipLevel": 0 # }, # ... # ] # rate = self.safe_value(response, 0) timestamp = self.safe_number(rate, 'timestamp') return { 'currency': code, 'rate': self.safe_number(rate, 'dailyInterestRate'), 'period': 86400000, 'timestamp': timestamp, 'datetime': self.iso8601(timestamp), 'info': response, }
[ [ [ 212, 220 ], [ 1103, 1111 ] ], [ [ 228, 232 ], [ 218575, 218579 ] ], [ [ 262, 275 ], [ 42414, 42427 ], [ 46753, 46766 ], [ 47873, 47886 ], [ 47998, 48011 ], [ 56974, 56987 ], [ 131345, 131358 ], [ 159206, 159219 ], [ 159382, 159395 ], [ 160950, 160963 ], [ 161136, 161149 ], [ 220011, 220024 ], [ 220077, 220090 ] ], [ [ 305, 324 ], [ 42670, 42689 ], [ 43922, 43941 ], [ 44716, 44735 ], [ 46205, 46224 ], [ 46657, 46676 ], [ 47152, 47171 ], [ 47254, 47273 ], [ 214649, 214668 ], [ 215217, 215236 ] ], [ [ 354, 370 ], [ 42545, 42561 ], [ 43304, 43320 ], [ 48884, 48900 ] ], [ [ 400, 416 ], [ 48104, 48120 ] ], [ [ 446, 463 ], [ 127036, 127053 ], [ 128205, 128222 ], [ 132629, 132646 ], [ 133924, 133941 ], [ 134957, 134974 ], [ 178911, 178928 ], [ 207677, 207694 ], [ 208637, 208654 ], [ 211352, 211369 ] ], [ [ 493, 503 ], [ 44560, 44570 ], [ 44833, 44843 ], [ 44956, 44966 ], [ 45050, 45060 ], [ 45143, 45153 ], [ 45218, 45228 ], [ 45330, 45340 ], [ 45398, 45408 ], [ 45479, 45489 ], [ 45653, 45663 ], [ 45743, 45753 ], [ 45808, 45818 ], [ 45871, 45881 ], [ 45929, 45939 ], [ 46004, 46014 ], [ 46084, 46094 ], [ 46289, 46299 ], [ 46383, 46393 ], [ 46494, 46504 ], [ 46578, 46588 ], [ 48187, 48197 ], [ 211654, 211664 ], [ 212624, 212634 ] ], [ [ 533, 542 ], [ 46146, 46155 ] ], [ [ 572, 589 ], [ 43157, 43174 ], [ 47357, 47374 ], [ 47455, 47472 ], [ 47599, 47616 ], [ 47726, 47743 ], [ 48280, 48297 ], [ 48394, 48411 ], [ 48490, 48507 ], [ 48587, 48604 ], [ 48712, 48729 ] ], [ [ 619, 631 ], [ 44248, 44260 ], [ 45577, 45589 ], [ 48938, 48950 ], [ 120508, 120520 ], [ 121536, 121548 ], [ 125789, 125801 ], [ 126077, 126089 ], [ 126559, 126571 ], [ 217569, 217581 ], [ 217713, 217725 ], [ 217869, 217881 ] ], [ [ 661, 674 ], [ 46960, 46973 ], [ 47053, 47066 ] ], [ [ 704, 728 ], [ 42747, 42771 ], [ 42834, 42858 ], [ 42987, 43011 ] ], [ [ 758, 770 ], [ 179777, 179789 ], [ 181929, 181941 ], [ 183829, 183841 ], [ 205969, 205981 ], [ 207370, 207382 ], [ 208216, 208228 ], [ 209291, 209303 ], [ 210344, 210356 ], [ 211015, 211027 ], [ 211990, 212002 ], [ 212973, 212985 ], [ 214243, 214255 ], [ 221603, 221615 ] ], [ [ 800, 814 ], [ 43551, 43565 ], [ 217145, 217159 ], [ 219802, 219816 ] ], [ [ 844, 861 ], [ 44025, 44042 ], [ 44357, 44374 ] ], [ [ 891, 911 ], [ 43232, 43252 ], [ 43419, 43439 ], [ 43668, 43688 ], [ 43800, 43820 ], [ 44466, 44486 ] ], [ [ 941, 954 ], [ 42306, 42319 ] ], [ [ 984, 996 ], [ 44639, 44651 ] ], [ [ 1040, 1048 ], [ 49147, 49155 ], [ 49509, 49517 ], [ 123971, 123979 ], [ 124205, 124213 ] ], [ [ 1079, 1086 ], [ 80969, 80976 ], [ 80996, 81003 ], [ 117502, 117509 ], [ 141572, 141579 ], [ 186720, 186727 ], [ 189456, 189463 ], [ 189610, 189617 ], [ 189629, 189636 ], [ 190119, 190126 ], [ 190289, 190296 ], [ 190422, 190429 ], [ 190522, 190529 ], [ 190541, 190548 ], [ 190646, 190653 ], [ 190737, 190744 ], [ 190756, 190763 ], [ 191939, 191946 ], [ 192577, 192584 ], [ 192596, 192603 ], [ 192615, 192622 ], [ 192739, 192746 ], [ 192758, 192765 ], [ 193456, 193463 ], [ 193600, 193607 ], [ 193703, 193710 ], [ 193777, 193784 ], [ 193811, 193818 ], [ 193921, 193928 ], [ 194048, 194055 ], [ 194542, 194549 ], [ 194686, 194693 ], [ 194789, 194796 ], [ 194859, 194866 ], [ 194949, 194956 ], [ 195044, 195051 ], [ 195063, 195070 ], [ 195082, 195089 ], [ 195195, 195202 ], [ 195467, 195474 ], [ 195597, 195604 ], [ 195698, 195705 ], [ 199045, 199052 ], [ 199523, 199530 ], [ 200811, 200818 ], [ 200913, 200920 ], [ 201045, 201052 ], [ 201129, 201136 ], [ 201241, 201248 ], [ 201324, 201331 ], [ 201343, 201350 ], [ 201770, 201777 ], [ 201914, 201921 ], [ 202017, 202024 ], [ 202091, 202098 ], [ 202176, 202183 ], [ 202195, 202202 ], [ 202242, 202249 ], [ 202363, 202370 ], [ 202382, 202389 ], [ 203075, 203082 ], [ 203259, 203266 ], [ 203413, 203420 ], [ 203501, 203508 ], [ 203520, 203527 ], [ 203793, 203800 ], [ 203812, 203819 ], [ 203831, 203838 ], [ 203955, 203962 ], [ 203974, 203981 ], [ 219364, 219371 ] ], [ [ 1095, 1102 ], [ 1177, 1184 ] ] ]
from arm.logicnode.arm_nodes import * class SetTransformNode(ArmLogicTreeNode): """Use to set the transform of an object.""" bl_idname = 'LNSetTransformNode' bl_label = 'Set Object Transform' arm_version = 1 def init(self, context): super(SetTransformNode, self).init(context) self.add_input('ArmNodeSocketAction', 'In') self.add_input('ArmNodeSocketObject', 'Object') self.add_input('NodeSocketShader', 'Transform') self.add_output('ArmNodeSocketAction', 'Out') add_node(SetTransformNode, category=PKG_AS_CATEGORY)
[ [ [ 36, 37 ], [ 62, 78 ], [ 526, 534 ], [ 562, 577 ] ], [ [ 45, 61 ], [ 535, 551 ], [ 269, 285 ] ] ]
# SPDX-FileCopyrightText: 2017 Scott Shawcroft, written for Adafruit Industries # SPDX-FileCopyrightText: Copyright (c) 2021 Jose David M. for circuitpython # # SPDX-License-Identifier: MIT """A setuptools based setup module. See: https://packaging.python.org/en/latest/distributing.html https://github.com/pypa/sampleproject """ from setuptools import setup, find_packages # To use a consistent encoding from codecs import open from os import path here = path.abspath(path.dirname(__file__)) # Get the long description from the README file with open(path.join(here, "README.rst"), encoding="utf-8") as f: long_description = f.read() setup( # Community Bundle Information name="circuitpython-displayio-cartesian", use_scm_version=True, setup_requires=["setuptools_scm"], description="A cartesian plane widget for displaying graphical information.", long_description=long_description, long_description_content_type="text/x-rst", # The project's main homepage. url="https://github.com/circuitpython/CircuitPython_Org_DisplayIO_Cartesian.git", # Author details author="Jose David M.", author_email="", install_requires=[ "Adafruit-Blinka", "adafruit-circuitpython-display-text", "adafruit-circuitpython-displayio-layout", ], # Choose your license license="MIT", # See https://pypi.python.org/pypi?%3Aaction=list_classifiers classifiers=[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "Topic :: Software Development :: Libraries", "Topic :: System :: Hardware", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", ], # What does your project relate to? keywords="adafruit blinka circuitpython micropython displayio_cartesian displayio widget " "graphics gui graph chart graphic", # You can just specify the packages manually here if your project is # simple. Or you can use find_packages(). # TODO: IF LIBRARY FILES ARE A PACKAGE FOLDER, # CHANGE `py_modules=['...']` TO `packages=['...']` py_modules=["displayio_cartesian"], )
[ [ [ 356, 361 ], [ 645, 650 ] ], [ [ 363, 376 ] ], [ [ 428, 432 ], [ 552, 556 ] ], [ [ 448, 452 ], [ 461, 465 ], [ 474, 478 ], [ 557, 561 ] ], [ [ 454, 458 ], [ 567, 571 ] ], [ [ 609, 610 ], [ 635, 636 ] ], [ [ 616, 632 ], [ 901, 917 ] ] ]
from ssmpfwd.helpers import verify_plugin_version, verbose_debug_quiet, time_decorator from unittest.mock import MagicMock, patch import unittest class TestVerifyPluginVersion(unittest.TestCase): @patch("ssmpfwd.helpers.subprocess") def test_verify_plugin_version_success(self, mock_subprocess): result = mock_subprocess.run() result.stdout = b"9.8.3" self.assertTrue(verify_plugin_version("9.8.3")) @patch("ssmpfwd.helpers.subprocess") def test_verify_plugin_version_fail(self, mock_subprocess): with self.assertLogs("ssmpfwd.helpers", level="INFO") as cm: result = mock_subprocess.run() result.stdout = b"1.8.1" self.assertFalse(verify_plugin_version("9.2.3")) self.assertEqual(cm.output[0], "ERROR:ssmpfwd.helpers:session-manager-plugin version 1.8.1 is installed, 9.2.3 is required") class TestVerboseDebugQuiet(unittest.TestCase): import logging def setUp(self): @verbose_debug_quiet def test_func(): pass self.vdq = test_func self.vdq() def test_quiet(self): option_name = "quiet" self.assertTrue(any([p.name == option_name for p in self.vdq.__click_params__]), msg=f"Can not find {option_name} in option parameters") def test_debug(self): flag_value = self.logging.DEBUG self.assertTrue(any([p.flag_value == flag_value for p in self.vdq.__click_params__]), msg=f"Can not find {flag_value} in option flag values") def test_verbose(self): flag_value = self.logging.INFO self.assertTrue(any([p.flag_value == flag_value for p in self.vdq.__click_params__]), msg=f"Can not find {flag_value} in option flag values") def test_default_loglevel(self): flag_value = self.logging.WARN self.assertTrue(any([p.flag_value == flag_value for p in self.vdq.__click_params__]), msg=f"Can not find {flag_value} in option flag values") class TestTimeDecorator(unittest.TestCase): from time import sleep def setUp(self): @time_decorator def test_func(): self.sleep(0.5) self.time_decorated_method = test_func def test_time_decorartor(self): with self.assertLogs("ssmpfwd.helpers", level="INFO") as cm: self.time_decorated_method() self.assertEqual(cm.output[0], "INFO:ssmpfwd.helpers:[*] starting test_func")
[ [ [ 28, 49 ], [ 412, 433 ], [ 735, 756 ] ], [ [ 51, 70 ], [ 1012, 1031 ] ], [ [ 72, 86 ], [ 2119, 2133 ] ], [ [ 114, 123 ] ], [ [ 125, 130 ], [ 209, 214 ], [ 452, 457 ] ], [ [ 139, 147 ], [ 183, 191 ], [ 938, 946 ], [ 2037, 2045 ] ], [ [ 159, 182 ] ], [ [ 916, 937 ] ], [ [ 2019, 2036 ] ] ]
import numpy as np import matplotlib.pyplot as plt import pint # Use the same registry from main import ureg ureg.setup_matplotlib(True) from uncertainties import ufloat, umath, unumpy import pandas as pd from scipy.signal import find_peaks from scipy.integrate import simpson from scipy.optimize import curve_fit plt.rcParams['text.usetex'] = True amp = 700*ureg.mV R=ufloat(0.82, 0.82*0.1)*ureg.ohm df = pd.read_csv("./ESRB.csv") # The I0_modulation signal is horrible, the system was too noisy, so instead: # # I0_modulation = (unumpy.uarray( # df['V_modulation_raw'].values, # df['V_modulation_err'].values # )*ureg.mV/R).to('ampere') # # we regnerate it, assuming it should be linear, just as V_DC is. I0_modulation = (unumpy.uarray(np.linspace( df['V_modulation_raw'].min(), df['V_modulation_raw'].max(), len(df) ), df['V_modulation_err'].mean())*ureg.mV/R).to('ampere') ptp_Y = unumpy.uarray( df['ptp_Y_raw'].values*df['phase_sign'].values, df['ptp_Y_err'].values )*ureg.mV ptp_X_modulation = ufloat(3.09, 0.01)*ureg.mV fig, ax = plt.subplots() I0_modulation_err = np.array([val.m.s for val in I0_modulation]) I0_modulation_raw = np.array([val.m.n for val in I0_modulation]) ptp_ratio = ptp_Y/ptp_X_modulation absorption_deriviative = ptp_ratio/max(ptp_ratio) absorption_deriviative_raw = np.array([val.m.n for val in absorption_deriviative]) absorption_deriviative_err = np.array([val.m.s for val in absorption_deriviative]) ax.errorbar( I0_modulation_raw*ureg.ampere, absorption_deriviative_raw, # Dimensionless fmt='.', yerr=absorption_deriviative_err, # TODO: Mention in report that error is too big to be drafted #xerr=I_modulation_err, # TODO: Is this the correct label? label='Absorption Deriviative' ) def lorentzian_dif_fit(I, I0, gamma, amplitude): return amplitude*(-2*(gamma**2)*(I - I0))/ \ (gamma**2 + (I - I0)**2)**2 def lorentzian_fit(I, I0, gamma, amplitude): return amplitude*gamma**2/\ (gamma**2 + (I - I0)**2)**2 ##### By MATLAB: # Goodness of fit: # SSE: 0.197 # R-square: 0.9845 # Adjusted R-square: 0.9838 # RMSE: 0.06769 # I0 gamma amplitude matlab_p0 = [0.5479, 0.03847, 0.05554] matlab_bounds=((0.547, 0.03672, 0.05304), (0.5488, 0.04021, 0.05805)) I_rf = ufloat(matlab_p0[0], abs(matlab_bounds[0][0] - matlab_p0[0]))*ureg.ampere I_hwhm = ufloat(matlab_p0[1], abs(matlab_bounds[0][1] - matlab_p0[1]))*ureg.ampere from main import g_times_bohr # TODO: Take this value from Itamar & Tomer H_RF = ufloat(34.914, 0.009)*ureg.gauss k = H_RF/I_rf # Converts current I To frequency f using all of the constants def I2f(I): return (I*k*g_times_bohr/ureg.planck_constant).to('megahertz') f0_modulation = I2f(I0_modulation) f_rf = I2f(I_rf) f_hwhm = I2f(I_hwhm) T2 = (1/f_hwhm).to('nanosecond') ##### A failing Python fit attempt - I consider it as a failure because it hits ##### the bounds :/ # popt, pcov = curve_fit( # lorentzian_dif_fit, absorption_deriviative_raw, I0_modulation_raw, # p0=matlab_p0, bounds=matlab_bounds # ) # lorentzian_dif_fit_points = lorentzian_dif_fit(I0_modulation_raw, *popt) # ax.plot( # I0_modulation_raw*ureg.ampere, # lorentzian_dif_fit_points, # label="Python fit" # ) I0_modulation_seq = np.linspace( I0_modulation.min().m.n, I0_modulation.max().m.n, len(I0_modulation)*100 ) ax.plot( I0_modulation_seq*ureg.ampere, lorentzian_dif_fit(I0_modulation_seq, I_rf.m.n, I_hwhm.m.n, matlab_p0[2]), label="Matlab fit" ) ax.set_yticks([]) axt = ax.twiny() axt.grid(linestyle='--') axt.set_yticks([]) f0_modulation_seq = np.linspace( f0_modulation.min().m.n, f0_modulation.max().m.n, len(f0_modulation)*100 ) def lorentzian_wrapper(f0): # From some reason this need to be amplified by a factor of 800 so it will # look good. return lorentzian_fit(f0, f_rf.m.n, f_hwhm.m.n, matlab_p0[2]*800) axt.plot( f0_modulation_seq*ureg.megahertz, lorentzian_wrapper(f0_modulation_seq), label = "Lorenzian fit", color='green' ) axt.set_xticks( [(f_rf - f_hwhm).m.n, f_rf.m.n, (f_rf + f_hwhm).m.n], ['', '$f_{rf}$', ''] ) axt.set_xlabel('') axt.arrow( length_includes_head = True, x = (f_rf - f_hwhm).m.n*ureg.megahertz, y = lorentzian_wrapper((f_rf - f_hwhm).m.n), dx = 2*f_hwhm.m.n*ureg.megahertz, dy = 0, head_length = f_hwhm.m.n/10, head_width = matlab_p0[2], label="Full Width Half Max", ) axt.arrow( length_includes_head = True, x = (f_rf + f_hwhm).m.n*ureg.megahertz, y = lorentzian_wrapper((f_rf + f_hwhm).m.n), dx = -2*f_hwhm.m.n*ureg.megahertz, head_length = f_hwhm.m.n/10, head_width = matlab_p0[2], dy = 0, ) axt.text( 0.5, 0.63, # (f_hwhm.m.n/10), # lorentzian_wrapper((f0 - f_hwhm).m.n)*2, "FWHM", transform=ax.transAxes, # fontsize=00 ) ax.legend(loc='upper right') # axt.legend(loc='upper left') plt.show() fig.savefig("ESRB.pgf") fig.savefig("ESRB.png") # TODO: Integrate numerically / or fit to a laurenzian's differentiation # TODO: Scale the x axis to frequency and find the width of the laurenzian in # frequency scale
[ [ [ 7, 18 ], [ 752, 754 ], [ 1108, 1110 ], [ 1173, 1175 ], [ 1332, 1334 ], [ 1415, 1417 ], [ 3335, 3337 ], [ 3682, 3684 ] ], [ [ 26, 50 ], [ 314, 317 ], [ 1073, 1076 ], [ 4991, 4994 ] ], [ [ 58, 62 ] ], [ [ 104, 108 ], [ 109, 113 ], [ 360, 364 ], [ 393, 397 ], [ 879, 883 ], [ 1008, 1012 ], [ 1054, 1058 ], [ 1504, 1508 ], [ 2401, 2405 ], [ 2484, 2488 ], [ 2600, 2604 ], [ 3466, 3470 ], [ 4008, 4012 ], [ 4304, 4308 ], [ 4391, 4395 ], [ 4590, 4594 ], [ 4678, 4682 ], [ 2729, 2733 ] ], [ [ 163, 169 ], [ 370, 376 ], [ 1035, 1041 ], [ 2339, 2345 ], [ 2422, 2428 ], [ 2578, 2584 ] ], [ [ 171, 176 ] ], [ [ 178, 184 ], [ 738, 744 ], [ 912, 918 ] ], [ [ 192, 204 ], [ 408, 410 ] ], [ [ 230, 240 ] ], [ [ 269, 276 ] ], [ [ 304, 313 ] ], [ [ 350, 353 ] ], [ [ 368, 369 ], [ 887, 888 ] ], [ [ 403, 405 ], [ 769, 771 ], [ 803, 805 ], [ 841, 843 ], [ 848, 850 ], [ 931, 933 ], [ 954, 956 ], [ 983, 985 ] ], [ [ 721, 734 ], [ 1137, 1150 ], [ 1202, 1215 ], [ 2788, 2801 ], [ 3352, 3365 ], [ 3381, 3394 ], [ 3414, 3427 ] ], [ [ 904, 909 ], [ 1230, 1235 ] ], [ [ 1016, 1032 ], [ 1236, 1252 ] ], [ [ 1063, 1066 ], [ 5002, 5005 ], [ 5026, 5029 ] ], [ [ 1068, 1070 ], [ 1469, 1471 ], [ 3435, 3437 ], [ 3583, 3585 ], [ 3607, 3609 ], [ 4895, 4897 ], [ 4930, 4932 ] ], [ [ 1088, 1105 ] ], [ [ 1153, 1170 ], [ 1486, 1503 ] ], [ [ 1218, 1227 ], [ 1278, 1287 ], [ 1292, 1301 ] ], [ [ 1253, 1275 ], [ 1361, 1383 ], [ 1444, 1466 ] ], [ [ 1303, 1329 ], [ 1521, 1547 ] ], [ [ 1386, 1412 ], [ 1587, 1613 ] ], [ [ 1790, 1808 ], [ 3483, 3501 ] ], [ [ 1928, 1942 ], [ 3917, 3931 ] ], [ [ 2201, 2210 ], [ 2346, 2355 ], [ 2386, 2395 ], [ 2429, 2438 ], [ 2469, 2478 ], [ 3543, 3552 ], [ 4469, 4478 ], [ 4744, 4753 ], [ 3958, 3967 ] ], [ [ 2243, 2256 ], [ 2364, 2377 ], [ 2447, 2460 ] ], [ [ 2332, 2336 ], [ 2620, 2624 ], [ 2814, 2818 ], [ 3521, 3525 ] ], [ [ 2413, 2419 ], [ 2833, 2839 ], [ 3531, 3537 ] ], [ [ 2514, 2526 ], [ 2716, 2728 ] ], [ [ 2571, 2575 ], [ 2615, 2619 ] ], [ [ 2611, 2612 ], [ 2714, 2715 ] ], [ [ 2692, 2695 ], [ 2784, 2787 ], [ 2810, 2813 ], [ 2829, 2832 ] ], [ [ 2768, 2781 ], [ 3699, 3712 ], [ 3728, 3741 ], [ 3761, 3774 ] ], [ [ 2803, 2807 ], [ 4134, 4138 ], [ 4154, 4158 ], [ 4165, 4169 ], [ 4285, 4289 ], [ 4348, 4352 ], [ 4571, 4575 ], [ 4634, 4638 ], [ 3936, 3940 ] ], [ [ 2820, 2826 ], [ 2849, 2855 ], [ 4141, 4147 ], [ 4172, 4178 ], [ 4292, 4298 ], [ 4355, 4361 ], [ 4380, 4386 ], [ 4437, 4443 ], [ 4578, 4584 ], [ 4641, 4647 ], [ 4667, 4673 ], [ 4712, 4718 ], [ 3946, 3952 ] ], [ [ 2841, 2843 ] ], [ [ 3315, 3332 ], [ 3448, 3465 ], [ 3502, 3519 ] ], [ [ 3601, 3604 ], [ 3618, 3621 ], [ 3643, 3646 ], [ 3976, 3979 ], [ 4112, 4115 ], [ 4213, 4216 ], [ 4232, 4235 ], [ 4518, 4521 ], [ 4772, 4775 ] ], [ [ 3662, 3679 ], [ 3990, 4007 ], [ 4047, 4064 ] ], [ [ 3786, 3804 ], [ 4028, 4046 ], [ 4328, 4346 ], [ 4614, 4632 ] ] ]
from flask import Flask, request, redirect from twilio.twiml.messaging_response import MessagingResponse from get_secrets import * def main(): resp = MessagingResponse() resp.message ("You have reached the DogBot. Thanks for contacting us :)") return str(resp) if __name__ == "__main__": main()
[ [ [ 18, 23 ] ], [ [ 25, 32 ] ], [ [ 34, 42 ] ], [ [ 87, 104 ], [ 155, 172 ] ], [ [ 129, 130 ] ], [ [ 136, 140 ], [ 308, 312 ] ] ]
from __future__ import annotations from enum import IntEnum class Algorithm(IntEnum): """ https://developers.yubico.com/YubiHSM2/Concepts/Algorithms.html """ RSA_PKCS1_SHA1 = 1 RSA_PKCS1_SHA256 = 2 RSA_PKCS1_SHA384 = 3 RSA_PKCS1_SHA512 = 4 RSA_PSS_SHA1 = 5 RSA_PSS_SHA256 = 6 RSA_PSS_SHA384 = 7 RSA_PSS_SHA512 = 8 RSA_2048 = 9 RSA_3072 = 10 RSA_4096 = 11 EC_P256 = 12 EC_P384 = 13 EC_P521 = 14 EC_K256 = 15 EC_BP256 = 16 EC_BP384 = 17 EC_BP512 = 18 HMAC_SHA1 = 19 HMAC_SHA256 = 20 HMAC_SHA384 = 21 HMAC_SHA512 = 22 ECDSA_SHA1 = 23 EC_ECDH = 24 RSA_OAEP_SHA1 = 25 RSA_OAEP_SHA256 = 26 RSA_OAEP_SHA384 = 27 RSA_OAEP_SHA512 = 28 AES128_CCM_WRAP = 29 Opaque_Data = 30 Opaque_X509_Certificate = 31 MGF1_SHA1 = 32 MGF1_SHA256 = 33 MGF1_SHA384 = 34 MGF1_SHA512 = 35 SSH_Template = 36 Yubico_OTP_AES128 = 37 Yubico_AES_Authentication = 38 Yubico_OTP_AES192 = 39 Yubico_OTP_AES256 = 40 AES192_CCM_WRAP = 41 AES256_CCM_WRAP = 42 ECDSA_SHA256 = 43 ECDSA_SHA384 = 44 ECDSA_SHA512 = 45 ED25519 = 46 EC_P224 = 47 class Capability(IntEnum): """ https://developers.yubico.com/YubiHSM2/Concepts/Capability.html """ GetOpaque = 0 PutOpaque = 1 PutAuthenticationKey = 2 PutAsymmetricKey = 3 GenerateAsymmetricKey = 4 SignPkcs = 5 SignPss = 6 SignEcdsa = 7 SignEddsa = 8 DecryptPkcs = 9 DecryptOaep = 10 DeriveEcdh = 11 ExportWrapped = 12 ImportWrapped = 13 PutWrapKey = 14 GenerateWrapKey = 15 ExportableUnderWrap = 16 SetOption = 17 GetOption = 18 GetPseudoRandom = 19 PutMacKey = 20 GenerateHmacKey = 21 SignHmac = 22 VerifyHmac = 23 GetLogEntries = 24 SignSshCertificate = 25 GetTemplate = 26 PutTemplate = 27 ResetDevice = 28 DecryptOtp = 29 CreateOtpAead = 30 RandomizeOtpAead = 31 RewrapFromOtpAeadKey = 32 RewrapToOtpAeadKey = 33 SignAttestationCertificate = 34 PutOtpAeadKey = 35 GenerateOtpAeadKey = 36 WrapData = 37 UnwrapData = 38 DeleteOpaque = 39 DeleteAuthenticationKey = 40 DeleteAsymmetricKey = 41 DeleteWrapKey = 42 DeleteHmacKey = 43 DeleteTemplate = 44 DeleteOtpAeadKey = 45 ChangeAuthenticationKey = 46 class Command(IntEnum): """ https://developers.yubico.com/YubiHSM2/Commands/ """ Echo = 0x01 CreateSession = 0x03 AuthenticateSession = 0x04 SessionMessage = 0x05 GetDeviceInfo = 0x06 ResetDevice = 0x08 CloseSession = 0x40 GetStorageInfo = 0x41 PutOpaque = 0x42 GetOpaque = 0x43 PutAuthenticationKey = 0x44 PutAsymmetricKey = 0x45 GenerateAsymmetricKey = 0x46 SignPkcs1 = 0x47 ListObjects = 0x48 DecryptPkcs1 = 0x49 ExportWrapped = 0x4A ImportWrapped = 0x4B PutWrapKey = 0x4C GetLogEntries = 0x4D GetObjectInfo = 0x4E SetOption = 0x4F GetOption = 0x50 GetPseudoRandom = 0x51 PutHmacKey = 0x52 SignHmac = 0x53 GetPublicKey = 0x54 SignPss = 0x55 SignEcdsa = 0x56 DeriveEcdh = 0x57 DeleteObject = 0x58 DecryptOaep = 0x59 GenerateHmacKey = 0x5A GenerateWrapKey = 0x5B VerifyHmac = 0x5C SignSshCertificate = 0x5D PutTemplate = 0x5E GetTemplate = 0x5F DecryptOtp = 0x60 CreateOtpAead = 0x61 RandomizeOtpAead = 0x62 RewrapOtpAead = 0x63 SignAttestationCertificate = 0x64 PutOtpAeadKey = 0x65 GenerateOtpAeadKey = 0x66 SetLogIndex = 0x67 WrapData = 0x68 UnwrapData = 0x69 SignEddsa = 0x6A BlinkDevice = 0x6B ChangeAuthenticationKey = 0x6C Error = 0x7F class Error(IntEnum): """ https://developers.yubico.com/YubiHSM2/Concepts/Errors.html """ OK = 0x00 INVALID_COMMAND = 0x01 INVALID_DATA = 0x02 INVALID_SESSION = 0x03 AUTHENTICATION_FAILED = 0x04 SESSIONS_FULL = 0x05 SESSION_FAILED = 0x06 STORAGE_FAILED = 0x07 WRONG_LENGTH = 0x08 INSUFFICIENT_PERMISSIONS = 0x09 LOG_FULL = 0x0A OBJECT_NOT_FOUND = 0x0B INVALID_ID = 0x0C SSH_CA_CONSTRAINT_VIOLATION = 0x0E INVALID_OTP = 0x0F DEMO_MODE = 0x10 OBJECT_EXISTS = 0x11 class ObjectType(IntEnum): """ https://developers.yubico.com/YubiHSM2/Concepts/Object.html """ Opaque = 0x01 AuthenticationKey = 0x02 AsymmetricKey = 0x03 WrapKey = 0x04 HmacKey = 0x05 Template = 0x06 OtpAeadKey = 0x07 class Option(IntEnum): """ https://developers.yubico.com/YubiHSM2/Concepts/Options.html """ ForceAudit = 0x01 CommandAudit = 0x03
[ [ [ 23, 34 ] ], [ [ 53, 60 ], [ 79, 86 ], [ 1222, 1229 ], [ 2427, 2434 ], [ 3792, 3799 ], [ 4342, 4349 ], [ 4600, 4607 ] ], [ [ 69, 78 ] ], [ [ 1211, 1221 ] ], [ [ 2419, 2426 ] ], [ [ 3786, 3791 ] ], [ [ 4331, 4341 ] ], [ [ 4593, 4599 ] ] ]
import unittest.mock from functools import partial import bokeh.core.properties as bp import param import pytest from bokeh.document import Document from bokeh.io.doc import patch_curdoc from bokeh.models import Div from panel.layout import Tabs, WidgetBox from panel.reactive import Reactive, ReactiveHTML from panel.viewable import Viewable from panel.widgets import ( Checkbox, IntInput, StaticText, TextInput, ) def test_reactive_default_title(): doc = ReactiveHTML().server_doc() assert doc.title == 'Panel Application' def test_reactive_servable_title(): doc = Document() session_context = unittest.mock.Mock() with patch_curdoc(doc): doc._session_context = lambda: session_context ReactiveHTML().servable(title='A') ReactiveHTML().servable(title='B') assert doc.title == 'B' def test_link(): "Link two Reactive objects" class ReactiveLink(Reactive): a = param.Parameter() obj = ReactiveLink() obj2 = ReactiveLink() obj.link(obj2, a='a') obj.a = 1 assert obj.a == 1 assert obj2.a == 1 def test_param_rename(): "Test that Reactive renames params and properties" class ReactiveRename(Reactive): a = param.Parameter() _rename = {'a': 'b'} obj = ReactiveRename() params = obj._process_property_change({'b': 1}) assert params == {'a': 1} properties = obj._process_param_change({'a': 1}) assert properties == {'b': 1} def test_link_properties_nb(document, comm): class ReactiveLink(Reactive): text = param.String(default='A') obj = ReactiveLink() div = Div() # Link property and check bokeh js property callback is defined obj._link_props(div, ['text'], document, div, comm) assert 'text' in div._callbacks # Assert callback is set up correctly cb = div._callbacks['text'][0] assert isinstance(cb, partial) assert cb.args == (document, div.ref['id'], comm, None) assert cb.func == obj._comm_change def test_link_properties_server(document): class ReactiveLink(Reactive): text = param.String(default='A') obj = ReactiveLink() div = Div() # Link property and check bokeh callback is defined obj._link_props(div, ['text'], document, div) assert 'text' in div._callbacks # Assert callback is set up correctly cb = div._callbacks['text'][0] assert isinstance(cb, partial) assert cb.args == (document, div.ref['id'], None) assert cb.func == obj._server_change def test_text_input_controls(): text_input = TextInput() controls = text_input.controls() assert isinstance(controls, Tabs) assert len(controls) == 2 wb1, wb2 = controls assert isinstance(wb1, WidgetBox) assert len(wb1) == 6 name, disabled, *(ws) = wb1 assert isinstance(name, StaticText) assert isinstance(disabled, Checkbox) not_checked = [] for w in ws: if w.name == 'Value': assert isinstance(w, TextInput) text_input.value = "New value" assert w.value == "New value" elif w.name == 'Value input': assert isinstance(w, TextInput) elif w.name == 'Placeholder': assert isinstance(w, TextInput) text_input.placeholder = "Test placeholder..." assert w.value == "Test placeholder..." elif w.name == 'Max length': assert isinstance(w, IntInput) else: not_checked.append(w) assert not not_checked assert isinstance(wb2, WidgetBox) assert len(wb2) == len(list(Viewable.param)) + 1 def test_text_input_controls_explicit(): text_input = TextInput() controls = text_input.controls(['placeholder', 'disabled']) assert isinstance(controls, WidgetBox) assert len(controls) == 3 name, disabled, placeholder = controls assert isinstance(name, StaticText) assert isinstance(disabled, Checkbox) assert isinstance(placeholder, TextInput) text_input.disabled = True assert disabled.value text_input.placeholder = "Test placeholder..." assert placeholder.value == "Test placeholder..." def test_reactive_html_basic(): class Test(ReactiveHTML): int = param.Integer(default=3, doc='An integer') float = param.Number(default=3.14, doc='A float') _template = '<div id="div" width=${int}></div>' data_model = Test._data_model assert data_model.__name__ == 'Test1' properties = data_model.properties() assert 'int' in properties assert 'float' in properties int_prop = data_model.lookup('int') assert isinstance(int_prop.property, bp.Int) assert int_prop.class_default(data_model) == 3 float_prop = data_model.lookup('float') assert isinstance(float_prop.property, bp.Float) assert float_prop.class_default(data_model) == 3.14 assert Test._node_callbacks == {} test = Test() root = test.get_root() assert test._attrs == {'div': [('width', ['int'], '{int}')]} assert root.callbacks == {} assert root.events == {} def test_reactive_html_no_id_param_error(): with pytest.raises(ValueError) as excinfo: class Test(ReactiveHTML): width = param.Number(default=200) _template = '<div width=${width}></div>' assert "Found <div> node with the `width` attribute referencing the `width` parameter." in str(excinfo.value) def test_reactive_html_no_id_method_error(): with pytest.raises(ValueError) as excinfo: class Test(ReactiveHTML): _template = '<div onclick=${_onclick}></div>' def _onclick(self): pass assert "Found <div> node with the `onclick` callback referencing the `_onclick` method." in str(excinfo.value) def test_reactive_html_dom_events(): class TestDOMEvents(ReactiveHTML): int = param.Integer(default=3, doc='An integer') float = param.Number(default=3.14, doc='A float') _template = '<div id="div" width=${int}></div>' _dom_events = {'div': ['change']} data_model = TestDOMEvents._data_model assert data_model.__name__ == 'TestDOMEvents1' properties = data_model.properties() assert 'int' in properties assert 'float' in properties int_prop = data_model.lookup('int') assert isinstance(int_prop.property, bp.Int) assert int_prop.class_default(data_model) == 3 float_prop = data_model.lookup('float') assert isinstance(float_prop.property, bp.Float) assert float_prop.class_default(data_model) == 3.14 assert TestDOMEvents._node_callbacks == {} test = TestDOMEvents() root = test.get_root() assert test._attrs == {'div': [('width', ['int'], '{int}')]} assert root.callbacks == {} assert root.events == {'div': {'change': True}} def test_reactive_html_inline(): class TestInline(ReactiveHTML): int = param.Integer(default=3, doc='An integer') _template = '<div id="div" onchange=${_div_change} width=${int}></div>' def _div_change(self, event): pass data_model = TestInline._data_model assert data_model.__name__ == 'TestInline1' properties = data_model.properties() assert 'int' in properties int_prop = data_model.lookup('int') assert isinstance(int_prop.property, bp.Int) assert int_prop.class_default(data_model) == 3 assert TestInline._node_callbacks == {'div': [('onchange', '_div_change')]} assert TestInline._inline_callbacks == [('div', 'onchange', '_div_change')] test = TestInline() root = test.get_root() assert test._attrs == { 'div': [ ('onchange', [], '{_div_change}'), ('width', ['int'], '{int}') ] } assert root.callbacks == {'div': [('onchange', '_div_change')]} assert root.events == {} test.on_event('div', 'click', print) assert root.events == {'div': {'click': False}} def test_reactive_html_children(): class TestChildren(ReactiveHTML): children = param.List(default=[]) _template = '<div id="div">${children}</div>' assert TestChildren._node_callbacks == {} assert TestChildren._inline_callbacks == [] assert TestChildren._parser.children == {'div': 'children'} widget = TextInput() test = TestChildren(children=[widget]) root = test.get_root() assert test._attrs == {} assert root.children == {'div': [widget._models[root.ref['id']][0]]} assert len(widget._models) == 1 assert test._panes == {'children': [widget]} widget_new = TextInput() test.children = [widget_new] assert len(widget._models) == 0 assert root.children == {'div': [widget_new._models[root.ref['id']][0]]} assert test._panes == {'children': [widget_new]} test._cleanup(root) assert len(test._models) == 0 assert len(widget_new._models) == 0 def test_reactive_html_templated_children(): class TestTemplatedChildren(ReactiveHTML): children = param.List(default=[]) _template = """ <select id="select"> {% for option in children %} <option id="option-{{ loop.index0 }}">${children[{{ loop.index0 }}]}</option> {% endfor %} </div> """ assert TestTemplatedChildren._node_callbacks == {} assert TestTemplatedChildren._inline_callbacks == [] assert TestTemplatedChildren._parser.children == {'option': 'children'} widget = TextInput() test = TestTemplatedChildren(children=[widget]) root = test.get_root() assert test._attrs == {} assert root.looped == ['option'] assert root.children == {'option': [widget._models[root.ref['id']][0]]} assert test._panes == {'children': [widget]} widget_new = TextInput() test.children = [widget_new] assert len(widget._models) == 0 assert root.children == {'option': [widget_new._models[root.ref['id']][0]]} assert test._panes == {'children': [widget_new]} def test_reactive_html_templated_dict_children(): class TestTemplatedChildren(ReactiveHTML): children = param.Dict(default={}) _template = """ <select id="select"> {% for key, option in children.items() %} <option id="option-{{ loop.index0 }}">${children[{{ key }}]}</option> {% endfor %} </div> """ assert TestTemplatedChildren._node_callbacks == {} assert TestTemplatedChildren._inline_callbacks == [] assert TestTemplatedChildren._parser.children == {'option': 'children'} widget = TextInput() test = TestTemplatedChildren(children={'test': widget}) root = test.get_root() assert test._attrs == {} assert root.looped == ['option'] assert root.children == {'option': [widget._models[root.ref['id']][0]]} assert test._panes == {'children': [widget]} widget_model = widget._models[root.ref['id']][0] widget_new = TextInput() test.children = {'test': widget_new, 'test2': widget} assert len(widget._models) == 1 assert root.children == { 'option': [ widget_new._models[root.ref['id']][0], widget_model ] } assert test._panes == {'children': [widget_new, widget]} def test_reactive_html_templated_children_add_loop_id(): class TestTemplatedChildren(ReactiveHTML): children = param.List(default=[]) _template = """ <select id="select"> {%- for option in children %} <option id="option">${children[{{ loop.index0 }}]}</option> {%- endfor %} </select> """ assert TestTemplatedChildren._node_callbacks == {} assert TestTemplatedChildren._inline_callbacks == [] assert TestTemplatedChildren._parser.children == {'option': 'children'} test = TestTemplatedChildren(children=['A', 'B', 'C']) assert test._get_template()[0] == """ <select id="select-${id}"> <option id="option-0-${id}"></option> <option id="option-1-${id}"></option> <option id="option-2-${id}"></option> </select> """ model = test.get_root() assert test._attrs == {} assert model.looped == ['option'] def test_reactive_html_templated_children_add_loop_id_and_for_loop_var(): class TestTemplatedChildren(ReactiveHTML): children = param.List(default=[]) _template = """ <select id="select"> {%- for option in children %} <option id="option">${option}</option> {%- endfor %} </select> """ assert TestTemplatedChildren._node_callbacks == {} assert TestTemplatedChildren._inline_callbacks == [] assert TestTemplatedChildren._parser.children == {'option': 'children'} test = TestTemplatedChildren(children=['A', 'B', 'C']) assert test._get_template()[0] == """ <select id="select-${id}"> <option id="option-0-${id}"></option> <option id="option-1-${id}"></option> <option id="option-2-${id}"></option> </select> """ model = test.get_root() assert test._attrs == {} assert model.looped == ['option'] @pytest.mark.parametrize('operator', ['', '+', '-', '*', '\\', '%', '**', '>>', '<<', '>>>', '&', '^', '&&', '||', '??']) @pytest.mark.parametrize('sep', [' ', '']) def test_reactive_html_scripts_linked_properties_assignment_operator(operator, sep): class TestScripts(ReactiveHTML): clicks = param.Integer() _template = "<div id='test'></div>" _scripts = {'render': f'test.onclick = () => {{ data.clicks{sep}{operator}= 1 }}'} assert TestScripts()._linked_properties() == ['clicks']
[ [ [ 7, 20 ], [ 626, 634 ] ], [ [ 44, 51 ], [ 1911, 1918 ], [ 2430, 2437 ] ], [ [ 60, 87 ], [ 4679, 4681 ], [ 4826, 4828 ], [ 6379, 6381 ], [ 6526, 6528 ], [ 7354, 7356 ] ], [ [ 95, 100 ], [ 945, 950 ], [ 1233, 1238 ], [ 1578, 1583 ], [ 2115, 2120 ], [ 4255, 4260 ], [ 4315, 4320 ], [ 5250, 5255 ], [ 5894, 5899 ], [ 5954, 5959 ], [ 6930, 6935 ], [ 8061, 8066 ], [ 9024, 9029 ], [ 10109, 10114 ], [ 11362, 11367 ], [ 12340, 12345 ], [ 13461, 13466 ] ], [ [ 108, 114 ], [ 13155, 13161 ], [ 13278, 13284 ], [ 5158, 5164 ], [ 5501, 5507 ] ], [ [ 143, 151 ], [ 592, 600 ] ], [ [ 177, 189 ], [ 657, 669 ] ], [ [ 215, 218 ], [ 1640, 1643 ], [ 2177, 2180 ] ], [ [ 245, 249 ], [ 2668, 2672 ] ], [ [ 251, 260 ], [ 2755, 2764 ], [ 3563, 3572 ], [ 3798, 3807 ] ], [ [ 288, 296 ], [ 921, 929 ], [ 1209, 1217 ], [ 1551, 1559 ], [ 2088, 2096 ] ], [ [ 298, 310 ], [ 471, 483 ], [ 739, 751 ], [ 782, 794 ], [ 4225, 4237 ], [ 5215, 5227 ], [ 5558, 5570 ], [ 5864, 5876 ], [ 6900, 6912 ], [ 8026, 8038 ], [ 8989, 9001 ], [ 10074, 10086 ], [ 11327, 11339 ], [ 12305, 12317 ], [ 13428, 13440 ] ], [ [ 338, 346 ], [ 3606, 3614 ] ], [ [ 379, 387 ], [ 2896, 2904 ], [ 3955, 3963 ] ], [ [ 389, 397 ], [ 3449, 3457 ] ], [ [ 399, 409 ], [ 2852, 2862 ], [ 3911, 3921 ] ], [ [ 411, 420 ], [ 2585, 2594 ], [ 3008, 3017 ], [ 3175, 3184 ], [ 3257, 3266 ], [ 3688, 3697 ], [ 4000, 4009 ], [ 8312, 8321 ], [ 8599, 8608 ], [ 9475, 9484 ], [ 9775, 9784 ], [ 10565, 10574 ], [ 10926, 10935 ] ], [ [ 430, 457 ] ], [ [ 550, 578 ] ], [ [ 852, 861 ] ], [ [ 1107, 1124 ] ], [ [ 1486, 1509 ] ], [ [ 2025, 2052 ] ], [ [ 2540, 2564 ] ], [ [ 3634, 3667 ] ], [ [ 4181, 4205 ] ], [ [ 5108, 5144 ] ], [ [ 5450, 5487 ] ], [ [ 5806, 5835 ] ], [ [ 6850, 6875 ] ], [ [ 7971, 7998 ] ], [ [ 8915, 8952 ] ], [ [ 9995, 10037 ] ], [ [ 11241, 11290 ] ], [ [ 12202, 12268 ] ], [ [ 13324, 13388 ] ] ]
# Copyright © 2019 Province of British Columbia # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """This module holds data for ppr queue event tracking.""" from __future__ import annotations from mhr_api.models import utils as model_utils from mhr_api.utils.base import BaseEnum from .db import db class EventTracking(db.Model): # pylint: disable=too-many-instance-attributes """This class manages all of the event tracking information.""" class EventTrackingTypes(BaseEnum): """Render an Enum of the event tracking types.""" SEARCH_REPORT = 'SEARCH_REPORT' API_NOTIFICATION = 'API_NOTIFICATION' EMAIL = 'EMAIL' SURFACE_MAIL = 'SURFACE_MAIL' EMAIL_REPORT = 'EMAIL_REPORT' REGISTRATION_REPORT = 'REGISTRATION_REPORT' __tablename__ = 'event_tracking' id = db.Column('id', db.Integer, db.Sequence('event_tracking_id_seq'), primary_key=True) key_id = db.Column('key_id', db.Integer, nullable=False, index=True) event_ts = db.Column('event_ts', db.DateTime, nullable=False, index=True) event_tracking_type = db.Column('event_tracking_type', db.String(20), db.ForeignKey('event_tracking_types.event_tracking_type'), nullable=False, index=True) status = db.Column('status', db.Integer, nullable=True) message = db.Column('message', db.String(2000), nullable=True) email_id = db.Column('email_address', db.String(250), nullable=True) # Relationships - SerialType tracking_type = db.relationship('EventTrackingType', foreign_keys=[event_tracking_type], back_populates='event_tracking', cascade='all, delete', uselist=False) def save(self): """Save the object to the database immediately.""" db.session.add(self) db.session.commit() @property def json(self) -> dict: """Return the event tracking record as a json object.""" event_tracking = { 'eventTrackingId': self.id, 'keyId': self.key_id, 'type': self.event_tracking_type, 'createDateTime': model_utils.format_ts(self.event_ts) } if self.status: event_tracking['status'] = self.status if self.message: event_tracking['message'] = self.message if self.email_id: event_tracking['emailAddress'] = self.email_id return event_tracking @classmethod def find_by_id(cls, event_id: int): """Return a tracking object by ID.""" if event_id: return cls.query.get(event_id) return None @classmethod def find_by_key_id(cls, key_id: int): """Return a list of event tracking objects by key id.""" event_tracking = None if key_id: event_tracking = cls.query.filter(EventTracking.key_id == key_id) \ .order_by(EventTracking.id).all() return event_tracking @classmethod def find_by_key_id_type(cls, key_id: int, event_tracking_type: str, extra_key: str = None): """Return a list of event tracking objects by key id and event tracking type.""" event_tracking = None if key_id and event_tracking_type: event_tracking = cls.query.filter(EventTracking.key_id == key_id, EventTracking.event_tracking_type == event_tracking_type) \ .order_by(EventTracking.id).all() if event_tracking is not None and extra_key: events = [] for event in event_tracking: if event.message and event.message.find(extra_key) > 0: events.append(event) return events return event_tracking @staticmethod def create(key_id: int, event_type: str, status: int = None, message: str = None): """Create an EventTracking record.""" event_tracking = EventTracking(key_id=key_id, event_tracking_type=event_type, status=status, message=message) event_tracking.event_ts = model_utils.now_ts() event_tracking.save() return event_tracking
[ [ [ 676, 687 ] ], [ [ 716, 736 ], [ 2656, 2667 ], [ 4666, 4677 ] ], [ [ 768, 776 ], [ 976, 984 ] ], [ [ 794, 796 ], [ 819, 821 ], [ 1332, 1334 ], [ 1348, 1350 ], [ 1360, 1362 ], [ 1429, 1431 ], [ 1449, 1451 ], [ 1504, 1506 ], [ 1526, 1528 ], [ 1593, 1595 ], [ 1626, 1628 ], [ 1677, 1679 ], [ 1813, 1815 ], [ 1833, 1835 ], [ 1874, 1876 ], [ 1895, 1897 ], [ 1942, 1944 ], [ 1969, 1971 ], [ 2054, 2056 ], [ 2322, 2324 ], [ 2351, 2353 ] ], [ [ 805, 818 ], [ 3381, 3394 ], [ 3463, 3476 ], [ 3840, 3853 ], [ 3918, 3931 ], [ 4026, 4039 ], [ 4539, 4552 ] ] ]
import unittest from hydrus.core import HydrusConstants as HC from hydrus.core import HydrusData from hydrus.core import HydrusSerialisable from hydrus.client import ClientApplicationCommand as CAC from hydrus.client import ClientConstants as CC from hydrus.client import ClientData from hydrus.client import ClientDefaults from hydrus.client import ClientDuplicates from hydrus.client import ClientSearch from hydrus.client.gui import ClientGUIShortcuts from hydrus.client.importing import ClientImportOptions from hydrus.client.importing import ClientImportSubscriptions from hydrus.client.importing import ClientImportSubscriptionQuery from hydrus.client.media import ClientMedia from hydrus.client.media import ClientMediaManagers from hydrus.client.media import ClientMediaResult from hydrus.client.metadata import ClientTags from hydrus.test import TestController as TC class TestSerialisables( unittest.TestCase ): def _dump_and_load_and_test( self, obj, test_func ): serialisable_tuple = obj.GetSerialisableTuple() self.assertIsInstance( serialisable_tuple, tuple ) if isinstance( obj, HydrusSerialisable.SerialisableBaseNamed ): ( serialisable_type, name, version, serialisable_info ) = serialisable_tuple elif isinstance( obj, HydrusSerialisable.SerialisableBase ): ( serialisable_type, version, serialisable_info ) = serialisable_tuple self.assertEqual( serialisable_type, obj.SERIALISABLE_TYPE ) self.assertEqual( version, obj.SERIALISABLE_VERSION ) dupe_obj = HydrusSerialisable.CreateFromSerialisableTuple( serialisable_tuple ) self.assertIsNot( obj, dupe_obj ) test_func( obj, dupe_obj ) # json_string = obj.DumpToString() self.assertIsInstance( json_string, str ) dupe_obj = HydrusSerialisable.CreateFromString( json_string ) self.assertIsNot( obj, dupe_obj ) test_func( obj, dupe_obj ) # network_bytes = obj.DumpToNetworkBytes() self.assertIsInstance( network_bytes, bytes ) dupe_obj = HydrusSerialisable.CreateFromNetworkBytes( network_bytes ) self.assertIsNot( obj, dupe_obj ) test_func( obj, dupe_obj ) def test_basics( self ): def test( obj, dupe_obj ): self.assertEqual( len( list(obj.items()) ), len( list(dupe_obj.items()) ) ) for ( key, value ) in list(obj.items()): self.assertEqual( value, dupe_obj[ key ] ) # d = HydrusSerialisable.SerialisableDictionary() d[ 1 ] = 2 d[ 3 ] = 'test1' d[ 'test2' ] = 4 d[ 'test3' ] = 5 d[ 6 ] = HydrusSerialisable.SerialisableDictionary( { i : 'test' + str( i ) for i in range( 20 ) } ) d[ ClientSearch.Predicate( ClientSearch.PREDICATE_TYPE_TAG, 'test pred 1' ) ] = 56 d[ ClientSearch.Predicate( ClientSearch.PREDICATE_TYPE_TAG, 'test pred 2' ) ] = HydrusSerialisable.SerialisableList( [ ClientSearch.Predicate( ClientSearch.PREDICATE_TYPE_TAG, 'test' + str( i ) ) for i in range( 10 ) ] ) self.assertEqual( len( list(d.keys()) ), 7 ) for ( key, value ) in list(d.items()): self.assertEqual( d[ key ], value ) self._dump_and_load_and_test( d, test ) # db = HydrusSerialisable.SerialisableBytesDictionary() db[ HydrusData.GenerateKey() ] = HydrusData.GenerateKey() db[ HydrusData.GenerateKey() ] = [ HydrusData.GenerateKey() for i in range( 10 ) ] db[ 1 ] = HydrusData.GenerateKey() db[ 2 ] = [ HydrusData.GenerateKey() for i in range( 10 ) ] self.assertEqual( len( list(db.keys()) ), 4 ) for ( key, value ) in list(db.items()): self.assertEqual( db[ key ], value ) self._dump_and_load_and_test( db, test ) def test_SERIALISABLE_TYPE_APPLICATION_COMMAND( self ): def test( obj, dupe_obj ): self.assertEqual( obj.GetCommandType(), dupe_obj.GetCommandType() ) self.assertEqual( obj.GetData(), dupe_obj.GetData() ) acs = [] acs.append( ( CAC.ApplicationCommand( CAC.APPLICATION_COMMAND_TYPE_SIMPLE, CAC.SIMPLE_ARCHIVE_FILE ), 'archive file' ) ) acs.append( ( CAC.ApplicationCommand( CAC.APPLICATION_COMMAND_TYPE_CONTENT, ( HydrusData.GenerateKey(), HC.CONTENT_TYPE_MAPPINGS, HC.CONTENT_UPDATE_FLIP, 'test' ) ), 'flip on/off mappings "test" for unknown service!' ) ) acs.append( ( CAC.ApplicationCommand( CAC.APPLICATION_COMMAND_TYPE_CONTENT, ( CC.DEFAULT_LOCAL_TAG_SERVICE_KEY, HC.CONTENT_TYPE_MAPPINGS, HC.CONTENT_UPDATE_FLIP, 'test' ) ), 'flip on/off mappings "test" for my tags' ) ) acs.append( ( CAC.ApplicationCommand( CAC.APPLICATION_COMMAND_TYPE_CONTENT, ( HydrusData.GenerateKey(), HC.CONTENT_TYPE_RATINGS, HC.CONTENT_UPDATE_SET, 0.4 ) ), 'set ratings uncertain rating, "0.4" for unknown service!' ) ) for ( ac, s ) in acs: self._dump_and_load_and_test( ac, test ) self.assertEqual( ac.ToString(), s ) def test_SERIALISABLE_TYPE_DUPLICATE_ACTION_OPTIONS( self ): def test( obj, dupe_obj ): self.assertEqual( obj.ToTuple(), dupe_obj.ToTuple() ) duplicate_action_options_delete_and_move = ClientDuplicates.DuplicateActionOptions( [ ( CC.DEFAULT_LOCAL_TAG_SERVICE_KEY, HC.CONTENT_MERGE_ACTION_MOVE, ClientTags.TagFilter() ) ], [ ( TC.LOCAL_RATING_LIKE_SERVICE_KEY, HC.CONTENT_MERGE_ACTION_MOVE ), ( TC.LOCAL_RATING_NUMERICAL_SERVICE_KEY, HC.CONTENT_MERGE_ACTION_MOVE ) ] ) duplicate_action_options_copy = ClientDuplicates.DuplicateActionOptions( [ ( CC.DEFAULT_LOCAL_TAG_SERVICE_KEY, HC.CONTENT_MERGE_ACTION_COPY, ClientTags.TagFilter() ) ], [ ( TC.LOCAL_RATING_LIKE_SERVICE_KEY, HC.CONTENT_MERGE_ACTION_COPY ), ( TC.LOCAL_RATING_NUMERICAL_SERVICE_KEY, HC.CONTENT_MERGE_ACTION_COPY ) ] ) duplicate_action_options_merge = ClientDuplicates.DuplicateActionOptions( [ ( CC.DEFAULT_LOCAL_TAG_SERVICE_KEY, HC.CONTENT_MERGE_ACTION_TWO_WAY_MERGE, ClientTags.TagFilter() ) ], [ ( TC.LOCAL_RATING_LIKE_SERVICE_KEY, HC.CONTENT_MERGE_ACTION_TWO_WAY_MERGE ), ( TC.LOCAL_RATING_NUMERICAL_SERVICE_KEY, HC.CONTENT_MERGE_ACTION_TWO_WAY_MERGE ) ] ) inbox = True size = 40960 mime = HC.IMAGE_JPEG width = 640 height = 480 duration = None num_frames = None has_audio = False num_words = None local_locations_manager = ClientMediaManagers.LocationsManager( { CC.LOCAL_FILE_SERVICE_KEY, CC.COMBINED_LOCAL_FILE_SERVICE_KEY }, set(), set(), set(), inbox ) trash_locations_manager = ClientMediaManagers.LocationsManager( { CC.TRASH_SERVICE_KEY, CC.COMBINED_LOCAL_FILE_SERVICE_KEY }, set(), set(), set(), inbox ) deleted_locations_manager = ClientMediaManagers.LocationsManager( set(), { CC.COMBINED_LOCAL_FILE_SERVICE_KEY }, set(), set(), inbox ) # duplicate to generate proper dicts one_tags_manager = ClientMediaManagers.TagsManager( { CC.DEFAULT_LOCAL_TAG_SERVICE_KEY : { HC.CONTENT_STATUS_CURRENT : { 'one' } } }, { CC.DEFAULT_LOCAL_TAG_SERVICE_KEY : { HC.CONTENT_STATUS_CURRENT : { 'one' } } } ).Duplicate() two_tags_manager = ClientMediaManagers.TagsManager( { CC.DEFAULT_LOCAL_TAG_SERVICE_KEY : { HC.CONTENT_STATUS_CURRENT : { 'two' } } }, { CC.DEFAULT_LOCAL_TAG_SERVICE_KEY : { HC.CONTENT_STATUS_CURRENT : { 'two' } } } ).Duplicate() substantial_tags_manager = ClientMediaManagers.TagsManager( { CC.DEFAULT_LOCAL_TAG_SERVICE_KEY : { HC.CONTENT_STATUS_CURRENT : { 'test tag', 'series:namespaced test tag' } } }, { CC.DEFAULT_LOCAL_TAG_SERVICE_KEY : { HC.CONTENT_STATUS_CURRENT : { 'test tag', 'series:namespaced test tag' } } } ).Duplicate() empty_tags_manager = ClientMediaManagers.TagsManager( {}, {} ).Duplicate() one_ratings_manager = ClientMediaManagers.RatingsManager( { TC.LOCAL_RATING_LIKE_SERVICE_KEY : 1.0, TC.LOCAL_RATING_NUMERICAL_SERVICE_KEY : 0.8 } ) two_ratings_manager = ClientMediaManagers.RatingsManager( { TC.LOCAL_RATING_LIKE_SERVICE_KEY : 0.0, TC.LOCAL_RATING_NUMERICAL_SERVICE_KEY : 0.6 } ) substantial_ratings_manager = ClientMediaManagers.RatingsManager( { TC.LOCAL_RATING_LIKE_SERVICE_KEY : 1.0, TC.LOCAL_RATING_NUMERICAL_SERVICE_KEY : 0.8 } ) empty_ratings_manager = ClientMediaManagers.RatingsManager( {} ) notes_manager = ClientMediaManagers.NotesManager( {} ) file_viewing_stats_manager = ClientMediaManagers.FileViewingStatsManager.STATICGenerateEmptyManager() # local_hash_has_values = HydrusData.GenerateKey() file_info_manager = ClientMediaManagers.FileInfoManager( 1, local_hash_has_values, size, mime, width, height, duration, num_frames, has_audio, num_words ) media_result = ClientMediaResult.MediaResult( file_info_manager, substantial_tags_manager, local_locations_manager, substantial_ratings_manager, notes_manager, file_viewing_stats_manager ) local_media_has_values = ClientMedia.MediaSingleton( media_result ) # other_local_hash_has_values = HydrusData.GenerateKey() file_info_manager = ClientMediaManagers.FileInfoManager( 2, other_local_hash_has_values, size, mime, width, height, duration, num_frames, has_audio, num_words ) media_result = ClientMediaResult.MediaResult( file_info_manager, substantial_tags_manager, local_locations_manager, substantial_ratings_manager, notes_manager, file_viewing_stats_manager ) other_local_media_has_values = ClientMedia.MediaSingleton( media_result ) # local_hash_empty = HydrusData.GenerateKey() file_info_manager = ClientMediaManagers.FileInfoManager( 3, local_hash_empty, size, mime, width, height, duration, num_frames, has_audio, num_words ) media_result = ClientMediaResult.MediaResult( file_info_manager, empty_tags_manager, local_locations_manager, empty_ratings_manager, notes_manager, file_viewing_stats_manager ) local_media_empty = ClientMedia.MediaSingleton( media_result ) # trashed_hash_empty = HydrusData.GenerateKey() file_info_manager = ClientMediaManagers.FileInfoManager( 4, trashed_hash_empty, size, mime, width, height, duration, num_frames, has_audio, num_words ) media_result = ClientMediaResult.MediaResult( file_info_manager, empty_tags_manager, trash_locations_manager, empty_ratings_manager, notes_manager, file_viewing_stats_manager ) trashed_media_empty = ClientMedia.MediaSingleton( media_result ) # deleted_hash_empty = HydrusData.GenerateKey() file_info_manager = ClientMediaManagers.FileInfoManager( 5, deleted_hash_empty, size, mime, width, height, duration, num_frames, has_audio, num_words ) media_result = ClientMediaResult.MediaResult( file_info_manager, empty_tags_manager, deleted_locations_manager, empty_ratings_manager, notes_manager, file_viewing_stats_manager ) deleted_media_empty = ClientMedia.MediaSingleton( media_result ) # one_hash = HydrusData.GenerateKey() file_info_manager = ClientMediaManagers.FileInfoManager( 6, one_hash, size, mime, width, height, duration, num_frames, has_audio, num_words ) media_result = ClientMediaResult.MediaResult( file_info_manager, one_tags_manager, local_locations_manager, one_ratings_manager, notes_manager, file_viewing_stats_manager ) one_media = ClientMedia.MediaSingleton( media_result ) # two_hash = HydrusData.GenerateKey() file_info_manager = ClientMediaManagers.FileInfoManager( 7, two_hash, size, mime, width, height, duration, num_frames, has_audio, num_words ) media_result = ClientMediaResult.MediaResult( file_info_manager, two_tags_manager, local_locations_manager, two_ratings_manager, notes_manager, file_viewing_stats_manager ) two_media = ClientMedia.MediaSingleton( media_result ) # self._dump_and_load_and_test( duplicate_action_options_delete_and_move, test ) self._dump_and_load_and_test( duplicate_action_options_copy, test ) self._dump_and_load_and_test( duplicate_action_options_merge, test ) # def assertSCUEqual( one, two ): self.assertEqual( TC.ConvertServiceKeysToContentUpdatesToComparable( one ), TC.ConvertServiceKeysToContentUpdatesToComparable( two ) ) file_deletion_reason = 'test delete' # result = duplicate_action_options_delete_and_move.ProcessPairIntoContentUpdates( local_media_has_values, local_media_empty, delete_second = True, file_deletion_reason = file_deletion_reason ) scu = {} scu[ CC.LOCAL_FILE_SERVICE_KEY ] = [ HydrusData.ContentUpdate( HC.CONTENT_TYPE_FILES, HC.CONTENT_UPDATE_DELETE, { local_hash_empty }, reason = file_deletion_reason ) ] assertSCUEqual( result, scu ) # result = duplicate_action_options_delete_and_move.ProcessPairIntoContentUpdates( local_media_has_values, trashed_media_empty, delete_second = True, file_deletion_reason = file_deletion_reason ) scu = {} scu[ CC.TRASH_SERVICE_KEY ] = [ HydrusData.ContentUpdate( HC.CONTENT_TYPE_FILES, HC.CONTENT_UPDATE_DELETE, { trashed_hash_empty }, reason = file_deletion_reason ) ] assertSCUEqual( result, scu ) # result = duplicate_action_options_delete_and_move.ProcessPairIntoContentUpdates( local_media_has_values, deleted_media_empty, delete_second = True, file_deletion_reason = file_deletion_reason ) self.assertEqual( result, {} ) # result = duplicate_action_options_delete_and_move.ProcessPairIntoContentUpdates( local_media_has_values, other_local_media_has_values, delete_second = True, file_deletion_reason = file_deletion_reason ) scu = {} scu[ CC.DEFAULT_LOCAL_TAG_SERVICE_KEY ] = [ HydrusData.ContentUpdate( HC.CONTENT_TYPE_MAPPINGS, HC.CONTENT_UPDATE_DELETE, ( 'test tag', { other_local_hash_has_values } ) ), HydrusData.ContentUpdate( HC.CONTENT_TYPE_MAPPINGS, HC.CONTENT_UPDATE_DELETE, ( 'series:namespaced test tag', { other_local_hash_has_values } ) ) ] scu[ TC.LOCAL_RATING_LIKE_SERVICE_KEY ] = [ HydrusData.ContentUpdate( HC.CONTENT_TYPE_RATINGS, HC.CONTENT_UPDATE_ADD, ( None, { other_local_hash_has_values } ) ) ] scu[ TC.LOCAL_RATING_NUMERICAL_SERVICE_KEY ] = [ HydrusData.ContentUpdate( HC.CONTENT_TYPE_RATINGS, HC.CONTENT_UPDATE_ADD, ( None, { other_local_hash_has_values } ) ) ] scu[ CC.LOCAL_FILE_SERVICE_KEY ] = [ HydrusData.ContentUpdate( HC.CONTENT_TYPE_FILES, HC.CONTENT_UPDATE_DELETE, { other_local_hash_has_values }, reason = file_deletion_reason ) ] assertSCUEqual( result, scu ) # result = duplicate_action_options_delete_and_move.ProcessPairIntoContentUpdates( local_media_empty, other_local_media_has_values, delete_second = True, file_deletion_reason = file_deletion_reason ) scu = {} scu[ CC.DEFAULT_LOCAL_TAG_SERVICE_KEY ] = [ HydrusData.ContentUpdate( HC.CONTENT_TYPE_MAPPINGS, HC.CONTENT_UPDATE_ADD, ( 'test tag', { local_hash_empty } ) ), HydrusData.ContentUpdate( HC.CONTENT_TYPE_MAPPINGS, HC.CONTENT_UPDATE_ADD, ( 'series:namespaced test tag', { local_hash_empty } ) ), HydrusData.ContentUpdate( HC.CONTENT_TYPE_MAPPINGS, HC.CONTENT_UPDATE_DELETE, ( 'test tag', { other_local_hash_has_values } ) ), HydrusData.ContentUpdate( HC.CONTENT_TYPE_MAPPINGS, HC.CONTENT_UPDATE_DELETE, ( 'series:namespaced test tag', { other_local_hash_has_values } ) ) ] scu[ TC.LOCAL_RATING_LIKE_SERVICE_KEY ] = [ HydrusData.ContentUpdate( HC.CONTENT_TYPE_RATINGS, HC.CONTENT_UPDATE_ADD, ( 1.0, { local_hash_empty } ) ), HydrusData.ContentUpdate( HC.CONTENT_TYPE_RATINGS, HC.CONTENT_UPDATE_ADD, ( None, { other_local_hash_has_values } ) ) ] scu[ TC.LOCAL_RATING_NUMERICAL_SERVICE_KEY ] = [ HydrusData.ContentUpdate( HC.CONTENT_TYPE_RATINGS, HC.CONTENT_UPDATE_ADD, ( 0.8, { local_hash_empty } ) ), HydrusData.ContentUpdate( HC.CONTENT_TYPE_RATINGS, HC.CONTENT_UPDATE_ADD, ( None, { other_local_hash_has_values } ) ) ] scu[ CC.LOCAL_FILE_SERVICE_KEY ] = [ HydrusData.ContentUpdate( HC.CONTENT_TYPE_FILES, HC.CONTENT_UPDATE_DELETE, { other_local_hash_has_values }, reason = file_deletion_reason ) ] assertSCUEqual( result, scu ) # # result = duplicate_action_options_copy.ProcessPairIntoContentUpdates( local_media_has_values, local_media_empty, file_deletion_reason = file_deletion_reason ) self.assertEqual( result, {} ) # result = duplicate_action_options_copy.ProcessPairIntoContentUpdates( local_media_empty, other_local_media_has_values, file_deletion_reason = file_deletion_reason ) scu = {} scu[ CC.DEFAULT_LOCAL_TAG_SERVICE_KEY ] = [ HydrusData.ContentUpdate( HC.CONTENT_TYPE_MAPPINGS, HC.CONTENT_UPDATE_ADD, ( 'test tag', { local_hash_empty } ) ), HydrusData.ContentUpdate( HC.CONTENT_TYPE_MAPPINGS, HC.CONTENT_UPDATE_ADD, ( 'series:namespaced test tag', { local_hash_empty } ) ) ] scu[ TC.LOCAL_RATING_LIKE_SERVICE_KEY ] = [ HydrusData.ContentUpdate( HC.CONTENT_TYPE_RATINGS, HC.CONTENT_UPDATE_ADD, ( 1.0, { local_hash_empty } ) ) ] scu[ TC.LOCAL_RATING_NUMERICAL_SERVICE_KEY ] = [ HydrusData.ContentUpdate( HC.CONTENT_TYPE_RATINGS, HC.CONTENT_UPDATE_ADD, ( 0.8, { local_hash_empty } ) ) ] assertSCUEqual( result, scu ) # # result = duplicate_action_options_merge.ProcessPairIntoContentUpdates( local_media_has_values, local_media_empty, file_deletion_reason = file_deletion_reason ) scu = {} scu[ CC.DEFAULT_LOCAL_TAG_SERVICE_KEY ] = [ HydrusData.ContentUpdate( HC.CONTENT_TYPE_MAPPINGS, HC.CONTENT_UPDATE_ADD, ( 'test tag', { local_hash_empty } ) ), HydrusData.ContentUpdate( HC.CONTENT_TYPE_MAPPINGS, HC.CONTENT_UPDATE_ADD, ( 'series:namespaced test tag', { local_hash_empty } ) ) ] scu[ TC.LOCAL_RATING_LIKE_SERVICE_KEY ] = [ HydrusData.ContentUpdate( HC.CONTENT_TYPE_RATINGS, HC.CONTENT_UPDATE_ADD, ( 1.0, { local_hash_empty } ) ) ] scu[ TC.LOCAL_RATING_NUMERICAL_SERVICE_KEY ] = [ HydrusData.ContentUpdate( HC.CONTENT_TYPE_RATINGS, HC.CONTENT_UPDATE_ADD, ( 0.8, { local_hash_empty } ) ) ] assertSCUEqual( result, scu ) # result = duplicate_action_options_merge.ProcessPairIntoContentUpdates( local_media_empty, other_local_media_has_values, file_deletion_reason = file_deletion_reason ) scu = {} scu[ CC.DEFAULT_LOCAL_TAG_SERVICE_KEY ] = [ HydrusData.ContentUpdate( HC.CONTENT_TYPE_MAPPINGS, HC.CONTENT_UPDATE_ADD, ( 'test tag', { local_hash_empty } ) ), HydrusData.ContentUpdate( HC.CONTENT_TYPE_MAPPINGS, HC.CONTENT_UPDATE_ADD, ( 'series:namespaced test tag', { local_hash_empty } ) ) ] scu[ TC.LOCAL_RATING_LIKE_SERVICE_KEY ] = [ HydrusData.ContentUpdate( HC.CONTENT_TYPE_RATINGS, HC.CONTENT_UPDATE_ADD, ( 1.0, { local_hash_empty } ) ) ] scu[ TC.LOCAL_RATING_NUMERICAL_SERVICE_KEY ] = [ HydrusData.ContentUpdate( HC.CONTENT_TYPE_RATINGS, HC.CONTENT_UPDATE_ADD, ( 0.8, { local_hash_empty } ) ) ] assertSCUEqual( result, scu ) # result = duplicate_action_options_merge.ProcessPairIntoContentUpdates( one_media, two_media, file_deletion_reason = file_deletion_reason ) scu = {} scu[ CC.DEFAULT_LOCAL_TAG_SERVICE_KEY ] = [ HydrusData.ContentUpdate( HC.CONTENT_TYPE_MAPPINGS, HC.CONTENT_UPDATE_ADD, ( 'one', { two_hash } ) ), HydrusData.ContentUpdate( HC.CONTENT_TYPE_MAPPINGS, HC.CONTENT_UPDATE_ADD, ( 'two', { one_hash } ) ) ] scu[ TC.LOCAL_RATING_LIKE_SERVICE_KEY ] = [ HydrusData.ContentUpdate( HC.CONTENT_TYPE_RATINGS, HC.CONTENT_UPDATE_ADD, ( 1.0, { two_hash } ) ) ] scu[ TC.LOCAL_RATING_NUMERICAL_SERVICE_KEY ] = [ HydrusData.ContentUpdate( HC.CONTENT_TYPE_RATINGS, HC.CONTENT_UPDATE_ADD, ( 0.8, { two_hash } ) ) ] assertSCUEqual( result, scu ) def test_SERIALISABLE_TYPE_SHORTCUT( self ): def test( obj, dupe_obj ): self.assertEqual( dupe_obj.__hash__(), ( dupe_obj.shortcut_type, dupe_obj.shortcut_key, dupe_obj.shortcut_press_type, tuple( dupe_obj.modifiers ) ).__hash__() ) self.assertEqual( obj, dupe_obj ) shortcuts = [] shortcuts.append( ( ClientGUIShortcuts.Shortcut(), 'f7' ) ) shortcuts.append( ( ClientGUIShortcuts.Shortcut( ClientGUIShortcuts.SHORTCUT_TYPE_KEYBOARD_SPECIAL, ClientGUIShortcuts.SHORTCUT_KEY_SPECIAL_SPACE, ClientGUIShortcuts.SHORTCUT_PRESS_TYPE_PRESS, [] ), 'space' ) ) shortcuts.append( ( ClientGUIShortcuts.Shortcut( ClientGUIShortcuts.SHORTCUT_TYPE_KEYBOARD_CHARACTER, ord( 'a' ), ClientGUIShortcuts.SHORTCUT_PRESS_TYPE_PRESS, [ ClientGUIShortcuts.SHORTCUT_MODIFIER_CTRL ] ), 'ctrl+a' ) ) shortcuts.append( ( ClientGUIShortcuts.Shortcut( ClientGUIShortcuts.SHORTCUT_TYPE_KEYBOARD_CHARACTER, ord( 'A' ), ClientGUIShortcuts.SHORTCUT_PRESS_TYPE_PRESS, [ ClientGUIShortcuts.SHORTCUT_MODIFIER_CTRL ] ), 'ctrl+a' ) ) shortcuts.append( ( ClientGUIShortcuts.Shortcut( ClientGUIShortcuts.SHORTCUT_TYPE_KEYBOARD_SPECIAL, ClientGUIShortcuts.SHORTCUT_KEY_SPECIAL_HOME, ClientGUIShortcuts.SHORTCUT_PRESS_TYPE_PRESS, [ ClientGUIShortcuts.SHORTCUT_MODIFIER_ALT, ClientGUIShortcuts.SHORTCUT_MODIFIER_CTRL ] ), 'ctrl+alt+home' ) ) shortcuts.append( ( ClientGUIShortcuts.Shortcut( ClientGUIShortcuts.SHORTCUT_TYPE_MOUSE, ClientGUIShortcuts.SHORTCUT_MOUSE_LEFT, ClientGUIShortcuts.SHORTCUT_PRESS_TYPE_PRESS, [] ), 'left-click' ) ) shortcuts.append( ( ClientGUIShortcuts.Shortcut( ClientGUIShortcuts.SHORTCUT_TYPE_MOUSE, ClientGUIShortcuts.SHORTCUT_MOUSE_MIDDLE, ClientGUIShortcuts.SHORTCUT_PRESS_TYPE_PRESS, [ ClientGUIShortcuts.SHORTCUT_MODIFIER_CTRL ] ), 'ctrl+middle-click' ) ) shortcuts.append( ( ClientGUIShortcuts.Shortcut( ClientGUIShortcuts.SHORTCUT_TYPE_MOUSE, ClientGUIShortcuts.SHORTCUT_MOUSE_SCROLL_DOWN, ClientGUIShortcuts.SHORTCUT_PRESS_TYPE_PRESS, [ ClientGUIShortcuts.SHORTCUT_MODIFIER_ALT, ClientGUIShortcuts.SHORTCUT_MODIFIER_SHIFT ] ), 'alt+shift+scroll down' ) ) for ( shortcut, s ) in shortcuts: self._dump_and_load_and_test( shortcut, test ) self.assertEqual( shortcut.ToString(), s ) def test_SERIALISABLE_TYPE_SHORTCUT_SET( self ): def test( obj, dupe_obj ): for ( shortcut, command ) in obj: self.assertEqual( dupe_obj.GetCommand( shortcut ).GetData(), command.GetData() ) default_shortcuts = ClientDefaults.GetDefaultShortcuts() for shortcuts in default_shortcuts: self._dump_and_load_and_test( shortcuts, test ) command_1 = CAC.ApplicationCommand( CAC.APPLICATION_COMMAND_TYPE_SIMPLE, CAC.SIMPLE_ARCHIVE_FILE ) command_2 = CAC.ApplicationCommand( CAC.APPLICATION_COMMAND_TYPE_CONTENT, ( HydrusData.GenerateKey(), HC.CONTENT_TYPE_MAPPINGS, HC.CONTENT_UPDATE_FLIP, 'test' ) ) command_3 = CAC.ApplicationCommand( CAC.APPLICATION_COMMAND_TYPE_CONTENT, ( CC.DEFAULT_LOCAL_TAG_SERVICE_KEY, HC.CONTENT_TYPE_MAPPINGS, HC.CONTENT_UPDATE_FLIP, 'test' ) ) k_shortcut_1 = ClientGUIShortcuts.Shortcut( ClientGUIShortcuts.SHORTCUT_TYPE_KEYBOARD_SPECIAL, ClientGUIShortcuts.SHORTCUT_KEY_SPECIAL_SPACE, ClientGUIShortcuts.SHORTCUT_PRESS_TYPE_PRESS, [] ) k_shortcut_2 = ClientGUIShortcuts.Shortcut( ClientGUIShortcuts.SHORTCUT_TYPE_KEYBOARD_CHARACTER, ord( 'a' ), ClientGUIShortcuts.SHORTCUT_PRESS_TYPE_PRESS, [ ClientGUIShortcuts.SHORTCUT_MODIFIER_CTRL ] ) k_shortcut_3 = ClientGUIShortcuts.Shortcut( ClientGUIShortcuts.SHORTCUT_TYPE_KEYBOARD_CHARACTER, ord( 'A' ), ClientGUIShortcuts.SHORTCUT_PRESS_TYPE_PRESS, [ ClientGUIShortcuts.SHORTCUT_MODIFIER_CTRL ] ) k_shortcut_4 = ClientGUIShortcuts.Shortcut( ClientGUIShortcuts.SHORTCUT_TYPE_KEYBOARD_SPECIAL, ClientGUIShortcuts.SHORTCUT_KEY_SPECIAL_HOME, ClientGUIShortcuts.SHORTCUT_PRESS_TYPE_PRESS, [ ClientGUIShortcuts.SHORTCUT_MODIFIER_ALT, ClientGUIShortcuts.SHORTCUT_MODIFIER_CTRL ] ) m_shortcut_1 = ClientGUIShortcuts.Shortcut( ClientGUIShortcuts.SHORTCUT_TYPE_MOUSE, ClientGUIShortcuts.SHORTCUT_MOUSE_LEFT, ClientGUIShortcuts.SHORTCUT_PRESS_TYPE_PRESS, [] ) m_shortcut_2 = ClientGUIShortcuts.Shortcut( ClientGUIShortcuts.SHORTCUT_TYPE_MOUSE, ClientGUIShortcuts.SHORTCUT_MOUSE_MIDDLE, ClientGUIShortcuts.SHORTCUT_PRESS_TYPE_PRESS, [ ClientGUIShortcuts.SHORTCUT_MODIFIER_CTRL ] ) m_shortcut_3 = ClientGUIShortcuts.Shortcut( ClientGUIShortcuts.SHORTCUT_TYPE_MOUSE, ClientGUIShortcuts.SHORTCUT_MOUSE_SCROLL_DOWN, ClientGUIShortcuts.SHORTCUT_PRESS_TYPE_PRESS, [ ClientGUIShortcuts.SHORTCUT_MODIFIER_ALT, ClientGUIShortcuts.SHORTCUT_MODIFIER_SHIFT ] ) shortcut_set = ClientGUIShortcuts.ShortcutSet( 'test' ) shortcut_set.SetCommand( k_shortcut_1, command_1 ) shortcut_set.SetCommand( k_shortcut_2, command_2 ) shortcut_set.SetCommand( k_shortcut_3, command_2 ) shortcut_set.SetCommand( k_shortcut_4, command_3 ) shortcut_set.SetCommand( m_shortcut_1, command_1 ) shortcut_set.SetCommand( m_shortcut_2, command_2 ) shortcut_set.SetCommand( m_shortcut_3, command_3 ) self._dump_and_load_and_test( shortcut_set, test ) self.assertEqual( shortcut_set.GetCommand( k_shortcut_1 ).GetData(), command_1.GetData() ) shortcut_set.SetCommand( k_shortcut_1, command_3 ) self.assertEqual( shortcut_set.GetCommand( k_shortcut_1 ).GetData(), command_3.GetData() ) def test_SERIALISABLE_TYPE_SUBSCRIPTION( self ): def test( obj, dupe_obj ): self.assertEqual( obj.GetName(), dupe_obj.GetName() ) self.assertEqual( obj._gug_key_and_name, dupe_obj._gug_key_and_name ) self.assertEqual( len( obj._query_headers ), len( dupe_obj._query_headers ) ) self.assertEqual( obj._initial_file_limit, dupe_obj._initial_file_limit ) self.assertEqual( obj._periodic_file_limit, dupe_obj._periodic_file_limit ) self.assertEqual( obj._paused, dupe_obj._paused ) self.assertEqual( obj._file_import_options.GetSerialisableTuple(), dupe_obj._file_import_options.GetSerialisableTuple() ) self.assertEqual( obj._tag_import_options.GetSerialisableTuple(), dupe_obj._tag_import_options.GetSerialisableTuple() ) self.assertEqual( obj._no_work_until, dupe_obj._no_work_until ) sub = ClientImportSubscriptions.Subscription( 'test sub' ) self._dump_and_load_and_test( sub, test ) gug_key_and_name = ( HydrusData.GenerateKey(), 'muh test gug' ) query_headers = [] q = ClientImportSubscriptionQuery.SubscriptionQueryHeader() q.SetQueryText( 'test query' ) query_headers.append( q ) q = ClientImportSubscriptionQuery.SubscriptionQueryHeader() q.SetQueryText( 'test query 2' ) query_headers.append( q ) checker_options = ClientImportOptions.CheckerOptions() initial_file_limit = 100 periodic_file_limit = 50 paused = False file_import_options = ClientImportOptions.FileImportOptions() service_tag_import_options = ClientImportOptions.ServiceTagImportOptions( get_tags = False, additional_tags = { 'test additional tag', 'and another' } ) tag_import_options = ClientImportOptions.TagImportOptions( service_keys_to_service_tag_import_options = { HydrusData.GenerateKey() : service_tag_import_options } ) no_work_until = HydrusData.GetNow() - 86400 * 20 sub.SetTuple( gug_key_and_name, checker_options, initial_file_limit, periodic_file_limit, paused, file_import_options, tag_import_options, no_work_until ) sub.SetQueryHeaders( query_headers ) self.assertEqual( sub.GetGUGKeyAndName(), gug_key_and_name ) self.assertEqual( sub.GetTagImportOptions(), tag_import_options ) self.assertEqual( sub.GetQueryHeaders(), query_headers ) self.assertEqual( sub._paused, False ) sub.PauseResume() self.assertEqual( sub._paused, True ) sub.PauseResume() self.assertEqual( sub._paused, False ) self._dump_and_load_and_test( sub, test ) def test_SERIALISABLE_TYPE_TAG_FILTER( self ): def test( obj, dupe_obj ): self.assertEqual( obj._tag_slices_to_rules, dupe_obj._tag_slices_to_rules ) tags = set() tags.add( 'title:test title' ) tags.add( 'series:neon genesis evangelion' ) tags.add( 'series:kill la kill' ) tags.add( 'smile' ) tags.add( 'blue eyes' ) # tag_filter = ClientTags.TagFilter() self._dump_and_load_and_test( tag_filter, test ) self.assertEqual( tag_filter.Filter( tags ), { 'smile', 'blue eyes', 'title:test title', 'series:neon genesis evangelion', 'series:kill la kill' } ) # tag_filter = ClientTags.TagFilter() tag_filter.SetRule( '', CC.FILTER_BLACKLIST ) tag_filter.SetRule( ':', CC.FILTER_BLACKLIST ) self._dump_and_load_and_test( tag_filter, test ) self.assertEqual( tag_filter.Filter( tags ), set() ) # tag_filter = ClientTags.TagFilter() tag_filter.SetRule( '', CC.FILTER_BLACKLIST ) tag_filter.SetRule( ':', CC.FILTER_BLACKLIST ) tag_filter.SetRule( 'series:', CC.FILTER_WHITELIST ) self._dump_and_load_and_test( tag_filter, test ) self.assertEqual( tag_filter.Filter( tags ), { 'series:neon genesis evangelion', 'series:kill la kill' } ) # tag_filter = ClientTags.TagFilter() tag_filter.SetRule( '', CC.FILTER_BLACKLIST ) tag_filter.SetRule( ':', CC.FILTER_BLACKLIST ) tag_filter.SetRule( 'series:kill la kill', CC.FILTER_WHITELIST ) self._dump_and_load_and_test( tag_filter, test ) self.assertEqual( tag_filter.Filter( tags ), { 'series:kill la kill' } ) # tag_filter = ClientTags.TagFilter() tag_filter.SetRule( '', CC.FILTER_BLACKLIST ) tag_filter.SetRule( ':', CC.FILTER_BLACKLIST ) tag_filter.SetRule( 'smile', CC.FILTER_WHITELIST ) self._dump_and_load_and_test( tag_filter, test ) self.assertEqual( tag_filter.Filter( tags ), { 'smile' } ) # tag_filter = ClientTags.TagFilter() tag_filter.SetRule( ':', CC.FILTER_BLACKLIST ) self._dump_and_load_and_test( tag_filter, test ) self.assertEqual( tag_filter.Filter( tags ), { 'smile', 'blue eyes' } ) # tag_filter = ClientTags.TagFilter() tag_filter.SetRule( ':', CC.FILTER_BLACKLIST ) tag_filter.SetRule( 'series:', CC.FILTER_WHITELIST ) self._dump_and_load_and_test( tag_filter, test ) self.assertEqual( tag_filter.Filter( tags ), { 'smile', 'blue eyes', 'series:neon genesis evangelion', 'series:kill la kill' } ) # tag_filter = ClientTags.TagFilter() tag_filter.SetRule( ':', CC.FILTER_BLACKLIST ) tag_filter.SetRule( 'series:kill la kill', CC.FILTER_WHITELIST ) self._dump_and_load_and_test( tag_filter, test ) self.assertEqual( tag_filter.Filter( tags ), { 'smile', 'blue eyes', 'series:kill la kill' } ) # tag_filter = ClientTags.TagFilter() tag_filter.SetRule( 'series:', CC.FILTER_BLACKLIST ) self._dump_and_load_and_test( tag_filter, test ) self.assertEqual( tag_filter.Filter( tags ), { 'smile', 'blue eyes', 'title:test title' } ) # tag_filter = ClientTags.TagFilter() tag_filter.SetRule( 'series:', CC.FILTER_BLACKLIST ) tag_filter.SetRule( 'series:neon genesis evangelion', CC.FILTER_WHITELIST ) self._dump_and_load_and_test( tag_filter, test ) self.assertEqual( tag_filter.Filter( tags ), { 'smile', 'blue eyes', 'title:test title', 'series:neon genesis evangelion' } ) # tag_filter = ClientTags.TagFilter() tag_filter.SetRule( '', CC.FILTER_BLACKLIST ) self._dump_and_load_and_test( tag_filter, test ) self.assertEqual( tag_filter.Filter( tags ), { 'title:test title', 'series:neon genesis evangelion', 'series:kill la kill' } ) # tag_filter = ClientTags.TagFilter() tag_filter.SetRule( '', CC.FILTER_BLACKLIST ) tag_filter.SetRule( 'blue eyes', CC.FILTER_WHITELIST ) self._dump_and_load_and_test( tag_filter, test ) self.assertEqual( tag_filter.Filter( tags ), { 'title:test title', 'series:neon genesis evangelion', 'series:kill la kill', 'blue eyes' } ) # blacklist namespace test blacklist_tags = { 'nintendo', 'studio:nintendo' } # tag_filter = ClientTags.TagFilter() tag_filter.SetRule( 'nintendo', CC.FILTER_BLACKLIST ) self._dump_and_load_and_test( tag_filter, test ) self.assertEqual( tag_filter.Filter( blacklist_tags ), { 'studio:nintendo' } ) # tag_filter = ClientTags.TagFilter() tag_filter.SetRule( 'nintendo', CC.FILTER_BLACKLIST ) self._dump_and_load_and_test( tag_filter, test ) self.assertEqual( tag_filter.Filter( blacklist_tags, apply_unnamespaced_rules_to_namespaced_tags = True ), set() ) # tag_filter = ClientTags.TagFilter() tag_filter.SetRule( 'nintendo', CC.FILTER_BLACKLIST ) tag_filter.SetRule( 'studio:nintendo', CC.FILTER_WHITELIST ) self._dump_and_load_and_test( tag_filter, test ) self.assertEqual( tag_filter.Filter( blacklist_tags, apply_unnamespaced_rules_to_namespaced_tags = True ), { 'studio:nintendo' } )
[ [ [ 7, 15 ], [ 905, 913 ] ], [ [ 41, 62 ], [ 4902, 4904 ], [ 4928, 4930 ], [ 5139, 5141 ], [ 5165, 5167 ], [ 5359, 5361 ], [ 5384, 5386 ], [ 6013, 6015 ], [ 6109, 6111 ], [ 6182, 6184 ], [ 6336, 6338 ], [ 6432, 6434 ], [ 6505, 6507 ], [ 6660, 6662 ], [ 6765, 6767 ], [ 6847, 6849 ], [ 6957, 6959 ], [ 7758, 7760 ], [ 7840, 7842 ], [ 7995, 7997 ], [ 8077, 8079 ], [ 8240, 8242 ], [ 8357, 8359 ], [ 13854, 13856 ], [ 13877, 13879 ], [ 14337, 14339 ], [ 14360, 14362 ], [ 15121, 15123 ], [ 15147, 15149 ], [ 15250, 15252 ], [ 15276, 15278 ], [ 15450, 15452 ], [ 15475, 15477 ], [ 15627, 15629 ], [ 15652, 15654 ], [ 15792, 15794 ], [ 15815, 15817 ], [ 16302, 16304 ], [ 16328, 16330 ], [ 16417, 16419 ], [ 16443, 16445 ], [ 16550, 16552 ], [ 16576, 16578 ], [ 16679, 16681 ], [ 16705, 16707 ], [ 16879, 16881 ], [ 16904, 16906 ], [ 16986, 16988 ], [ 17011, 17013 ], [ 17163, 17165 ], [ 17188, 17190 ], [ 17270, 17272 ], [ 17295, 17297 ], [ 17435, 17437 ], [ 17458, 17460 ], [ 18165, 18167 ], [ 18191, 18193 ], [ 18280, 18282 ], [ 18306, 18308 ], [ 18466, 18468 ], [ 18491, 18493 ], [ 18631, 18633 ], [ 18656, 18658 ], [ 19079, 19081 ], [ 19105, 19107 ], [ 19194, 19196 ], [ 19220, 19222 ], [ 19380, 19382 ], [ 19405, 19407 ], [ 19545, 19547 ], [ 19570, 19572 ], [ 19989, 19991 ], [ 20015, 20017 ], [ 20104, 20106 ], [ 20130, 20132 ], [ 20290, 20292 ], [ 20315, 20317 ], [ 20455, 20457 ], [ 20480, 20482 ], [ 20872, 20874 ], [ 20898, 20900 ], [ 20974, 20976 ], [ 21000, 21002 ], [ 21129, 21131 ], [ 21154, 21156 ], [ 21286, 21288 ], [ 21311, 21313 ], [ 24620, 24622 ], [ 24646, 24648 ], [ 24799, 24801 ], [ 24825, 24827 ] ], [ [ 87, 97 ], [ 3843, 3853 ], [ 3814, 3824 ], [ 3911, 3921 ], [ 3880, 3890 ], [ 3977, 3987 ], [ 4022, 4032 ], [ 4876, 4886 ], [ 5333, 5343 ], [ 9340, 9350 ], [ 9894, 9904 ], [ 10449, 10459 ], [ 10972, 10982 ], [ 11499, 11509 ], [ 12018, 12028 ], [ 12511, 12521 ], [ 13828, 13838 ], [ 14311, 14321 ], [ 15095, 15105 ], [ 15224, 15234 ], [ 15424, 15434 ], [ 15601, 15611 ], [ 15766, 15776 ], [ 16276, 16286 ], [ 16391, 16401 ], [ 16524, 16534 ], [ 16653, 16663 ], [ 16853, 16863 ], [ 16960, 16970 ], [ 17137, 17147 ], [ 17244, 17254 ], [ 17409, 17419 ], [ 18139, 18149 ], [ 18254, 18264 ], [ 18440, 18450 ], [ 18605, 18615 ], [ 19053, 19063 ], [ 19168, 19178 ], [ 19354, 19364 ], [ 19519, 19529 ], [ 19963, 19973 ], [ 20078, 20088 ], [ 20264, 20274 ], [ 20429, 20439 ], [ 20846, 20856 ], [ 20948, 20958 ], [ 21103, 21113 ], [ 21260, 21270 ], [ 24594, 24604 ], [ 28494, 28504 ], [ 29408, 29418 ], [ 29499, 29509 ] ], [ [ 122, 140 ], [ 1158, 1176 ], [ 1347, 1365 ], [ 1663, 1681 ], [ 1983, 2001 ], [ 2297, 2315 ], [ 2851, 2869 ], [ 3033, 3051 ], [ 3313, 3331 ], [ 3744, 3762 ] ], [ [ 168, 199 ], [ 4683, 4686 ], [ 4707, 4710 ], [ 4744, 4747 ], [ 4812, 4815 ], [ 4836, 4839 ], [ 5041, 5044 ], [ 5065, 5068 ], [ 5269, 5272 ], [ 5293, 5296 ], [ 24423, 24426 ], [ 24447, 24450 ], [ 24484, 24487 ], [ 24530, 24533 ], [ 24554, 24557 ], [ 24701, 24704 ], [ 24725, 24728 ] ], [ [ 226, 247 ], [ 5105, 5107 ], [ 5979, 5981 ], [ 6302, 6304 ], [ 6626, 6628 ], [ 7196, 7198 ], [ 7223, 7225 ], [ 7364, 7366 ], [ 7386, 7388 ], [ 7536, 7538 ], [ 7721, 7723 ], [ 7803, 7805 ], [ 7958, 7960 ], [ 8040, 8042 ], [ 8203, 8205 ], [ 8320, 8322 ], [ 13796, 13798 ], [ 14284, 14286 ], [ 15056, 15058 ], [ 15734, 15736 ], [ 16237, 16239 ], [ 17377, 17379 ], [ 18100, 18102 ], [ 19014, 19016 ], [ 19924, 19926 ], [ 20807, 20809 ], [ 24765, 24767 ], [ 31108, 31110 ], [ 31163, 31165 ], [ 31434, 31436 ], [ 31489, 31491 ], [ 31550, 31552 ], [ 31875, 31877 ], [ 31930, 31932 ], [ 32003, 32005 ], [ 32294, 32296 ], [ 32349, 32351 ], [ 32408, 32410 ], [ 32686, 32688 ], [ 32977, 32979 ], [ 33038, 33040 ], [ 33386, 33388 ], [ 33459, 33461 ], [ 33779, 33781 ], [ 34096, 34098 ], [ 34180, 34182 ], [ 34524, 34526 ], [ 34869, 34871 ], [ 34932, 34934 ], [ 35410, 35412 ], [ 35715, 35717 ], [ 36056, 36058 ], [ 36125, 36127 ] ], [ [ 274, 284 ] ], [ [ 311, 325 ], [ 24218, 24232 ] ], [ [ 352, 368 ], [ 5934, 5950 ], [ 6257, 6273 ], [ 6581, 6597 ] ], [ [ 395, 407 ], [ 3136, 3148 ], [ 3160, 3172 ], [ 3352, 3364 ], [ 3376, 3388 ], [ 3236, 3248 ], [ 3260, 3272 ] ], [ [ 438, 456 ], [ 21841, 21859 ], [ 21918, 21936 ], [ 21947, 21965 ], [ 21998, 22016 ], [ 22045, 22063 ], [ 22137, 22155 ], [ 22166, 22184 ], [ 22231, 22249 ], [ 22279, 22297 ], [ 22367, 22385 ], [ 22396, 22414 ], [ 22461, 22479 ], [ 22509, 22527 ], [ 22597, 22615 ], [ 22626, 22644 ], [ 22677, 22695 ], [ 22723, 22741 ], [ 22771, 22789 ], [ 22813, 22831 ], [ 22917, 22935 ], [ 22946, 22964 ], [ 22986, 23004 ], [ 23026, 23044 ], [ 23123, 23141 ], [ 23152, 23170 ], [ 23192, 23210 ], [ 23234, 23252 ], [ 23282, 23300 ], [ 23381, 23399 ], [ 23410, 23428 ], [ 23450, 23468 ], [ 23497, 23515 ], [ 23545, 23563 ], [ 23587, 23605 ], [ 24892, 24910 ], [ 24921, 24939 ], [ 24972, 24990 ], [ 25019, 25037 ], [ 25093, 25111 ], [ 25122, 25140 ], [ 25187, 25205 ], [ 25235, 25253 ], [ 25304, 25322 ], [ 25333, 25351 ], [ 25398, 25416 ], [ 25446, 25464 ], [ 25515, 25533 ], [ 25544, 25562 ], [ 25595, 25613 ], [ 25641, 25659 ], [ 25689, 25707 ], [ 25731, 25749 ], [ 25809, 25827 ], [ 25838, 25856 ], [ 25878, 25896 ], [ 25918, 25936 ], [ 25992, 26010 ], [ 26021, 26039 ], [ 26061, 26079 ], [ 26103, 26121 ], [ 26151, 26169 ], [ 26220, 26238 ], [ 26249, 26267 ], [ 26289, 26307 ], [ 26336, 26354 ], [ 26384, 26402 ], [ 26426, 26444 ], [ 26505, 26523 ] ], [ [ 493, 512 ], [ 28910, 28929 ], [ 29075, 29094 ], [ 29161, 29180 ], [ 29323, 29342 ] ], [ [ 549, 574 ], [ 28344, 28369 ] ], [ [ 611, 640 ], [ 28594, 28623 ], [ 28744, 28773 ] ], [ [ 673, 684 ], [ 9785, 9796 ], [ 10351, 10362 ], [ 10872, 10883 ], [ 11399, 11410 ], [ 11928, 11939 ], [ 12421, 12432 ], [ 12914, 12925 ] ], [ [ 717, 736 ], [ 7156, 7175 ], [ 7324, 7343 ], [ 7489, 7508 ], [ 7686, 7705 ], [ 7923, 7942 ], [ 8168, 8187 ], [ 8477, 8496 ], [ 8570, 8589 ], [ 8726, 8745 ], [ 8890, 8909 ], [ 9048, 9067 ], [ 9122, 9141 ], [ 9207, 9226 ], [ 9402, 9421 ], [ 9956, 9975 ], [ 10511, 10530 ], [ 11034, 11053 ], [ 11561, 11580 ], [ 12080, 12099 ], [ 12573, 12592 ] ], [ [ 769, 786 ], [ 9569, 9586 ], [ 10129, 10146 ], [ 10673, 10690 ], [ 11198, 11215 ], [ 11725, 11742 ], [ 12234, 12251 ], [ 12727, 12744 ] ], [ [ 822, 832 ], [ 6043, 6053 ], [ 6366, 6376 ], [ 6699, 6709 ], [ 30740, 30750 ], [ 31044, 31054 ], [ 31370, 31380 ], [ 31811, 31821 ], [ 32230, 32240 ], [ 32621, 32631 ], [ 32912, 32922 ], [ 33321, 33331 ], [ 33708, 33718 ], [ 34025, 34035 ], [ 34460, 34470 ], [ 34805, 34815 ], [ 35338, 35348 ], [ 35643, 35653 ], [ 35984, 35994 ] ], [ [ 858, 878 ], [ 6075, 6077 ], [ 6143, 6145 ], [ 6398, 6400 ], [ 6466, 6468 ], [ 6731, 6733 ], [ 6808, 6810 ], [ 8608, 8610 ], [ 8648, 8650 ], [ 8764, 8766 ], [ 8804, 8806 ], [ 8928, 8930 ], [ 8968, 8970 ], [ 15385, 15387 ], [ 15557, 15559 ], [ 16814, 16816 ], [ 17093, 17095 ], [ 18401, 18403 ], [ 18561, 18563 ], [ 19315, 19317 ], [ 19475, 19477 ], [ 20225, 20227 ], [ 20385, 20387 ], [ 21064, 21066 ], [ 21216, 21218 ], [ 13336, 13338 ], [ 13394, 13396 ] ], [ [ 886, 903 ] ] ]
import requests import json import time import random from . import conf, data, lang from inukit.timestamp import natural_date, natural_time, timestamp_now def is_same_day(ts1, ts2) -> bool: def d(ts): return natural_date(ts, '%Y-%m-%d') return d(ts1) == d(ts2) def handle_morning(qq): last_morning = data.get(qq, 'last_morning') last_night = data.get(qq, 'last_night') now = timestamp_now() if last_morning > last_night: msg = lang.no_sleep else: msg = lang.morning_success % ( natural_time(now - last_night) ) data.set(qq, 'last_morning', now) return msg def handle_night(qq): last_morning = data.get(qq, 'last_morning') last_night = data.get(qq, 'last_night') now = timestamp_now() if last_night > last_morning: msg = lang.no_getup else: data.set(qq, 'last_night', now) msg = lang.night_success % ( natural_time(now - last_morning) ) return msg def gen_sign_info(): rp = random.randint(1,100) return { "rp": rp } def handle_sign(qq): last_sign = data.get(qq, 'last_sign') now = timestamp_now() msg = '' if is_same_day(last_sign, now): info = data.get(qq, 'last_sign_info') msg = lang.already_sign else: msg = lang.sign_success info = gen_sign_info() data.set(qq, 'last_sign', now) data.set(qq, 'last_sign_info', info) msg += lang.sign % ( natural_date(last_sign), info['rp'] ) return msg
[ [ [ 7, 15 ] ], [ [ 23, 27 ] ], [ [ 35, 39 ] ], [ [ 47, 53 ], [ 1038, 1044 ] ], [ [ 68, 72 ] ], [ [ 74, 78 ], [ 323, 327 ], [ 369, 373 ], [ 594, 598 ], [ 685, 689 ], [ 731, 735 ], [ 864, 868 ], [ 1134, 1138 ], [ 1250, 1254 ], [ 1394, 1398 ], [ 1433, 1437 ] ], [ [ 80, 84 ], [ 470, 474 ], [ 508, 512 ], [ 832, 836 ], [ 910, 914 ], [ 1295, 1299 ], [ 1337, 1341 ], [ 1481, 1485 ] ], [ [ 114, 126 ], [ 1503, 1515 ], [ 222, 234 ] ], [ [ 128, 140 ], [ 545, 557 ], [ 945, 957 ] ], [ [ 142, 155 ], [ 406, 419 ], [ 768, 781 ], [ 1170, 1183 ] ], [ [ 161, 172 ], [ 1206, 1217 ] ], [ [ 284, 298 ] ], [ [ 648, 660 ] ], [ [ 1012, 1025 ], [ 1370, 1383 ] ], [ [ 1101, 1112 ] ] ]
""" Django settings for app project. Generated by 'django-admin startproject' using Django 2.1.15. For more information on this file, see https://docs.djangoproject.com/en/2.1/topics/settings/ For the full list of settings and their values, see https://docs.djangoproject.com/en/2.1/ref/settings/ """ import os # Build paths inside the project like this: os.path.join(BASE_DIR, ...) BASE_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) # Quick-start development settings - unsuitable for production # See https://docs.djangoproject.com/en/2.1/howto/deployment/checklist/ # SECURITY WARNING: keep the secret key used in production secret! SECRET_KEY = 'lxb!(o00)qtw0p+6q_vs$01&wtsw(m*s!ol0_6^v*flo^!&ek&' # SECURITY WARNING: don't run with debug turned on in production! DEBUG = True ALLOWED_HOSTS = [] # Application definition INSTALLED_APPS = [ 'django.contrib.admin', 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.sessions', 'django.contrib.messages', 'django.contrib.staticfiles', 'rest_framework', 'rest_framework.authtoken', 'core', 'user', 'recipe', ] MIDDLEWARE = [ 'django.middleware.security.SecurityMiddleware', 'django.contrib.sessions.middleware.SessionMiddleware', 'django.middleware.common.CommonMiddleware', 'django.middleware.csrf.CsrfViewMiddleware', 'django.contrib.auth.middleware.AuthenticationMiddleware', 'django.contrib.messages.middleware.MessageMiddleware', 'django.middleware.clickjacking.XFrameOptionsMiddleware', ] ROOT_URLCONF = 'app.urls' TEMPLATES = [ { 'BACKEND': 'django.template.backends.django.DjangoTemplates', 'DIRS': [], 'APP_DIRS': True, 'OPTIONS': { 'context_processors': [ 'django.template.context_processors.debug', 'django.template.context_processors.request', 'django.contrib.auth.context_processors.auth', 'django.contrib.messages.context_processors.messages', ], }, }, ] WSGI_APPLICATION = 'app.wsgi.application' # Database # https://docs.djangoproject.com/en/2.1/ref/settings/#databases DATABASES = { 'default': { 'ENGINE': 'django.db.backends.postgresql', 'HOST': os.environ.get('DB_HOST'), 'NAME': os.environ.get('DB_NAME'), 'USER': os.environ.get('DB_USER'), 'PASSWORD': os.environ.get('DB_PASS'), } } # Password validation # https://docs.djangoproject.com/en/2.1/ref/settings/#auth-password-validators AUTH_PASSWORD_VALIDATORS = [ { 'NAME': 'django.contrib.auth.password_validation.UserAttributeSimilarityValidator', }, { 'NAME': 'django.contrib.auth.password_validation.MinimumLengthValidator', }, { 'NAME': 'django.contrib.auth.password_validation.CommonPasswordValidator', }, { 'NAME': 'django.contrib.auth.password_validation.NumericPasswordValidator', }, ] # Internationalization # https://docs.djangoproject.com/en/2.1/topics/i18n/ LANGUAGE_CODE = 'en-us' TIME_ZONE = 'UTC' USE_I18N = True USE_L10N = True USE_TZ = True # Static files (CSS, JavaScript, Images) # https://docs.djangoproject.com/en/2.1/howto/static-files/ STATIC_URL = '/static/' MEDIA_URL = '/media/' MEDIA_ROOT = '/vol/web/media' STATIC_ROOT = '/vol/web/static' AUTH_USER_MODEL = 'core.User'
[ [ [ 312, 314 ], [ 399, 401 ], [ 415, 417 ], [ 431, 433 ], [ 2302, 2304 ], [ 2345, 2347 ], [ 2388, 2390 ], [ 2435, 2437 ] ], [ [ 388, 396 ] ], [ [ 664, 674 ] ], [ [ 797, 802 ] ], [ [ 811, 824 ] ], [ [ 858, 872 ] ], [ [ 1158, 1168 ] ], [ [ 1572, 1584 ] ], [ [ 1599, 1608 ] ], [ [ 2084, 2100 ] ], [ [ 2204, 2213 ] ], [ [ 2574, 2598 ] ], [ [ 3077, 3090 ] ], [ [ 3102, 3111 ] ], [ [ 3121, 3129 ] ], [ [ 3138, 3146 ] ], [ [ 3155, 3161 ] ], [ [ 3273, 3283 ] ], [ [ 3297, 3306 ] ], [ [ 3320, 3330 ] ], [ [ 3350, 3361 ] ], [ [ 3383, 3398 ] ] ]
#!/usr/bin/env python3 '''Test config updates ''' # ------------------------------------------------------------------------------ # Imports # ------------------------------------------------------------------------------ import subprocess import os import json import time import datetime import requests import pytest # ------------------------------------------------------------------------------ # Constants # ------------------------------------------------------------------------------ G_TEST_HOST = 'http://127.0.0.1:12345' # ------------------------------------------------------------------------------ # run_command # ------------------------------------------------------------------------------ def run_command(command): p = subprocess.Popen(command, shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE) stdout, stderr = p.communicate() return (p.returncode, stdout, stderr) # ------------------------------------------------------------------------------ # setup scopez server in action mode # ------------------------------------------------------------------------------ @pytest.fixture() def setup_scopez_server_action(): # ------------------------------------------------------ # setup # ------------------------------------------------------ l_file_path = os.path.dirname(os.path.abspath(__file__)) l_geoip2city_path = os.path.realpath(os.path.join(l_file_path, '../../data/waf/db/GeoLite2-City.mmdb')) l_geoip2ISP_path = os.path.realpath(os.path.join(l_file_path, '../../data/waf/db/GeoLite2-ASN.mmdb')) l_conf_dir = os.path.realpath(os.path.join(l_file_path, '../../data/waf/conf')) l_ruleset_path = os.path.realpath(os.path.join(l_file_path, '../../data/waf/ruleset')) l_scopez_dir = os.path.realpath(os.path.join(l_file_path, '../../data/waf/conf/scopes')) l_an_list = os.path.realpath(os.path.join(l_file_path, '../../data/an/an-scopes.json')) l_scopez_server_path = os.path.abspath(os.path.join(l_file_path, '../../../build/util/scopez_server/scopez_server')) l_bot_challenge = os.path.realpath(os.path.join(l_file_path, '../../data/bot/bot-challenges.json')) l_subproc = subprocess.Popen([l_scopez_server_path, '-d', l_conf_dir, '-S', l_scopez_dir, '-l', l_an_list, '-r', l_ruleset_path, '-g', l_geoip2city_path, '-i', l_geoip2ISP_path, '-c', l_bot_challenge, '-a' ]) print('cmd: {}'.format(' '.join([l_scopez_server_path, '-d', l_conf_dir, '-S', l_scopez_dir, '-l', l_an_list, '-r', l_ruleset_path, '-g', l_geoip2city_path, '-i', l_geoip2ISP_path, '-c', l_bot_challenge, '-a']))) # '-b']))) time.sleep(1) # ------------------------------------------------------ # yield... # ------------------------------------------------------ yield setup_scopez_server_action # ------------------------------------------------------ # tear down # ------------------------------------------------------ _, _, _ = run_command('kill -9 %d'%(l_subproc.pid)) time.sleep(0.5) def test_acl_config_update(setup_scopez_server_action): ''' update acl config 0050-ZrLf2KkQ - remove gizoogle from user agent black list and test if request returns 200 ''' # ------------------------------------------------------ # test an 0050 with user-agent acl 'gizoogle' in the # request # ------------------------------------------------------ l_uri = G_TEST_HOST l_headers = {'host': 'monkeez.com', 'user-agent': 'gizoogle', 'waf-scopes-id': '0050'} l_r = requests.get(l_uri, headers=l_headers) assert l_r.status_code == 403 assert l_r.text == 'This is acl custom response\n' #------------------------------------------------------- # load acl config and remove gizoogle from blacklist # ------------------------------------------------------ l_conf = {} l_file_path = os.path.dirname(os.path.abspath(__file__)) l_acl_conf_path = os.path.realpath(os.path.join(l_file_path, '../../data/waf/conf/acl/0050-ZrLf2KkQ.acl.json')) try: with open(l_acl_conf_path) as l_f: l_conf = json.load(l_f) except Exception as l_e: print('error opening config file: %s. Reason: %s error: %s, doc: %s' % ( l_acl_conf_path, type(l_e), l_e, l_e.__doc__)) assert False l_conf['user_agent']['blacklist'] = [] l_conf['last_modified_date'] = datetime.datetime.utcnow().strftime('%Y-%m-%dT%H:%M:%S.%fZ') # ------------------------------------------------------ # post/update acl conf # ------------------------------------------------------ l_url = '%s/update_acl'%(G_TEST_HOST) l_headers = {'Content-Type': 'application/json', 'waf-scopes-id': '0050'} l_r = requests.post(l_url, headers=l_headers, data=json.dumps(l_conf)) assert l_r.status_code == 200 # ------------------------------------------------------ # blacklist should have been updated and should get 200 #------------------------------------------------------- l_uri = G_TEST_HOST l_headers = {'host': 'monkeez.com', 'user-agent': 'gizoogle', 'waf-scopes-id': '0050'} l_r = requests.get(l_uri, headers=l_headers) assert l_r.status_code == 200 def test_rules_config_update(setup_scopez_server_action): ''' update rules config 0050-ZrLf3KKq.rules.json - change user agent to Donkeez from Monkeez ''' # ------------------------------------------------------ # test an 0050 with user-agent 'Monkeez' in the # request # ------------------------------------------------------ l_uri = G_TEST_HOST l_headers = {'host': 'monkeez.com', 'user-agent': 'monkeez', 'waf-scopes-id': '0050'} l_r = requests.get(l_uri, headers=l_headers) assert l_r.status_code == 403 assert l_r.text == 'This is rules custom response\n' #------------------------------------------------------- # load rules config and changes monkeez to donkeez in # custom rules # ------------------------------------------------------ l_conf = {} l_file_path = os.path.dirname(os.path.abspath(__file__)) l_rules_conf_path = os.path.realpath(os.path.join(l_file_path, '../../data/waf/conf/rules/0050-ZrLf3KkQ.rules.json')) try: with open(l_rules_conf_path) as l_f: l_conf = json.load(l_f) except Exception as l_e: print('error opening config file: %s. Reason: %s error: %s, doc: %s' % ( l_file_path, type(l_e), l_e, l_e.__doc__)) assert False l_conf['directive'][1]['sec_rule']['operator']['value'] = 'donkeez' l_conf['last_modified_date'] = datetime.datetime.utcnow().strftime('%Y-%m-%dT%H:%M:%S.%fZ') # ------------------------------------------------------ # post/update rules conf # ------------------------------------------------------ l_url = '%s/update_rules'%(G_TEST_HOST) l_headers = {'Content-Type': 'application/json', 'waf-scopes-id': '0050'} l_r = requests.post(l_url, headers=l_headers, data=json.dumps(l_conf)) assert l_r.status_code == 200 # ------------------------------------------------------ # test again with user-agent 'Monkeez' in the # request. It should pass # ------------------------------------------------------ l_uri = G_TEST_HOST l_headers = {'host': 'monkeez.com', 'user-agent': 'monkeez', 'waf-scopes-id': '0050'} l_r = requests.get(l_uri, headers=l_headers) assert l_r.status_code == 200 # ------------------------------------------------------ # test with user-agent 'donkeez' in the # request. should be blocked # ------------------------------------------------------ l_uri = G_TEST_HOST l_headers = {'host': 'monkeez.com', 'user-agent': 'donkeez', 'waf-scopes-id': '0050'} l_r = requests.get(l_uri, headers=l_headers) assert l_r.status_code == 403 assert l_r.text == 'This is rules custom response\n' def test_profile_config_update(setup_scopez_server_action): ''' update profile config 0050-YrLf3KkQ.wafprof.json - change ignore_query_args to test from ignore ''' # ------------------------------------------------------ # test an 0050 with sql injection # ------------------------------------------------------ l_uri = G_TEST_HOST+'/profile.html?a=%27select%20*%20from%20testing%27' l_headers = {'host': 'monkeez.com', 'waf-scopes-id': '0050'} l_r = requests.get(l_uri, headers=l_headers) assert l_r.status_code == 403 assert l_r.text == 'This is profile custom response\n' # ------------------------------------------------------ # test an 0050 with sql injection and query_args "ignore" # ------------------------------------------------------ l_uri = G_TEST_HOST+'/profile.html?ignore=%27select%20*%20from%20testing%27' l_headers = {'host': 'monkeez.com', 'waf-scopes-id': '0050'} l_r = requests.get(l_uri, headers=l_headers) assert l_r.status_code == 200 #------------------------------------------------------- # load profile config and change "ignore_query_args" # to "test" # ------------------------------------------------------ l_conf = {} l_file_path = os.path.dirname(os.path.abspath(__file__)) l_profile_conf_path = os.path.realpath(os.path.join(l_file_path, '../../data/waf/conf/profile/0050-YrLf3KkQ.wafprof.json')) try: with open(l_profile_conf_path) as l_f: l_conf = json.load(l_f) except Exception as l_e: print('error opening config file: %s. Reason: %s error: %s, doc: %s' % ( l_profile_conf_path, type(l_e), l_e, l_e.__doc__)) assert False l_conf["general_settings"]["ignore_query_args"] = ["test"] l_conf['last_modified_date'] = datetime.datetime.utcnow().strftime('%Y-%m-%dT%H:%M:%S.%fZ') # ------------------------------------------------------ # post/update profile conf # ------------------------------------------------------ l_url = '%s/update_profile'%(G_TEST_HOST) l_headers = {'Content-Type': 'application/json', 'waf-scopes-id': '0050'} l_r = requests.post(l_url, headers=l_headers, data=json.dumps(l_conf)) assert l_r.status_code == 200 # ------------------------------------------------------ # test an 0050 with sql injection and query_args "ignore" # should get 403 # ------------------------------------------------------ l_uri = G_TEST_HOST+'/profile.html?ignore=%27select%20*%20from%20testing%27' l_headers = {'host': 'monkeez.com', 'waf-scopes-id': '0050'} l_r = requests.get(l_uri, headers=l_headers) assert l_r.status_code == 403 assert l_r.text == 'This is profile custom response\n' # ------------------------------------------------------ # test an 0050 with sql injection and query_args "test" # sql injection should be ignored and get 200 # ------------------------------------------------------ l_uri = G_TEST_HOST+'/profile.html?test=%27select%20*%20from%20testing%27' l_headers = {'host': 'monkeez.com', 'waf-scopes-id': '0050'} l_r = requests.get(l_uri, headers=l_headers) assert l_r.status_code == 200 def test_limit_config_update(setup_scopez_server_action): # ------------------------------------------------------ # Make 3 request in 2 sec for 3rd and # 4th scope. Third request should get rate limited # ------------------------------------------------------ l_uri = G_TEST_HOST+'/test.html' l_headers = {'host': 'limit.com', 'waf-scopes-id': '0050'} for _ in range(2): l_r = requests.get(l_uri, headers=l_headers) assert l_r.status_code == 200 l_r = requests.get(l_uri, headers=l_headers) assert l_r.status_code == 403 assert l_r.text == 'This is ddos custom response\n' l_uri = G_TEST_HOST+'/test.html' l_headers = {'host': 'test.limit.com', 'waf-scopes-id': '0050'} for _ in range(2): l_r = requests.get(l_uri, headers=l_headers) assert l_r.status_code == 200 l_r = requests.get(l_uri, headers=l_headers) assert l_r.status_code == 403 assert l_r.text == 'custom response for limits from limit_id_2\n' # ------------------------------------------------------ # sleep for 2 seconds. Enforcements should expire # ------------------------------------------------------ time.sleep(2) #------------------------------------------------------- # load limit config and change duration_sec to 3 # ------------------------------------------------------ l_conf = {} l_file_path = os.path.dirname(os.path.abspath(__file__)) l_limit_conf_path = os.path.realpath(os.path.join(l_file_path, '../../data/waf/conf/limit/0050-MjMhNXMR.limit.json')) try: with open(l_limit_conf_path) as l_f: l_conf = json.load(l_f) except Exception as l_e: print('error opening config file: %s. Reason: %s error: %s, doc: %s' % ( l_limit_conf_path, type(l_e), l_e, l_e.__doc__)) assert False l_conf["num"] = 3 l_conf['last_modified_date'] = datetime.datetime.utcnow().strftime('%Y-%m-%dT%H:%M:%S.%fZ') #------------------------------------------------------- # POST conf # ------------------------------------------------------ l_url = '%s/update_limit'%(G_TEST_HOST) l_headers = {'Content-Type': 'application/json', 'waf-scopes-id': '0050'} l_r = requests.post(l_url, headers=l_headers, data=json.dumps(l_conf)) assert l_r.status_code == 200 # ------------------------------------------------------ # Make 4 request in 2 sec. fourth request should get # rate limited. Third request shouldn't be blocked # because of the update # ------------------------------------------------------ l_uri = G_TEST_HOST+'/test.html' l_headers = {'host': 'limit.com', 'waf-scopes-id': '0050'} for _ in range(3): l_r = requests.get(l_uri, headers=l_headers) assert l_r.status_code == 200 l_r = requests.get(l_uri, headers=l_headers) assert l_r.status_code == 403 assert l_r.text == 'This is ddos custom response\n' # ------------------------------------------------------ # Make 4 request in 2 sec for fourth scope. # verify if 4th scope was also updated # ------------------------------------------------------ l_uri = G_TEST_HOST+'/test.html' l_headers = {'host': 'test.limit.com', 'waf-scopes-id': '0050'} for _ in range(3): l_r = requests.get(l_uri, headers=l_headers) assert l_r.status_code == 200 l_r = requests.get(l_uri, headers=l_headers) assert l_r.status_code == 403 assert l_r.text == 'custom response for limits from limit_id_2\n' def test_scopes_update(setup_scopez_server_action): #------------------------------------------------------- # check second scope for AN 0051 working correctly # ------------------------------------------------------ l_uri = G_TEST_HOST+'/path.html' l_headers = {'host': 'www.regexhost.com', 'waf-scopes-id':'0051', 'User-Agent': 'bananas'} l_r = requests.get(l_uri, headers=l_headers) assert l_r.status_code == 403 assert l_r.text == 'This is from RX scope\n' #------------------------------------------------------- # change the 'path' value for scope and update. # check if update was successful # ------------------------------------------------------ l_conf = {} l_file_path = os.path.dirname(os.path.abspath(__file__)) l_scopes_conf_path = os.path.realpath(os.path.join(l_file_path, '../../data/waf/conf/scopes/0051.scopes.json')) try: with open(l_scopes_conf_path) as l_f: l_conf = json.load(l_f) except Exception as l_e: print('error opening config file: %s. Reason: %s error: %s, doc: %s' % ( l_scopes_conf_path, type(l_e), l_e, l_e.__doc__)) assert False l_conf['scopes'][1]['path']['value'] = ".*/test.html" l_conf['last_modified_date'] = datetime.datetime.utcnow().strftime('%Y-%m-%dT%H:%M:%S.%fZ') #------------------------------------------------------- # POST conf # ------------------------------------------------------ l_url = '%s/update_scopes'%(G_TEST_HOST) l_headers = {'Content-Type': 'application/json'} l_r = requests.post(l_url, headers=l_headers, data=json.dumps(l_conf)) assert l_r.status_code == 200 #------------------------------------------------------- # make a request with same path '/path.html', # should match GLOB scope # ------------------------------------------------------ l_uri = G_TEST_HOST+'/path.html' l_headers = {'host': 'www.regexhost.com', 'waf-scopes-id':'0051', 'User-Agent': 'bananas'} l_r = requests.get(l_uri, headers=l_headers) assert l_r.status_code == 403 assert l_r.text == 'This is from GLOB scope\n' #------------------------------------------------------- # make a request with updated path '/test.html', # should get 403 with custom response # ------------------------------------------------------ l_uri = G_TEST_HOST+'/test.html' l_headers = {'host': 'www.regexhost.com', 'waf-scopes-id':'0051', 'User-Agent': 'bananas'} l_r = requests.get(l_uri, headers=l_headers) assert l_r.status_code == 403 assert l_r.text == 'This is from RX scope\n' def test_scopes_linkage_update(setup_scopez_server_action): """ Test linkage update. Update rules config in second scope (0050-scopes.json) to 0050-0gG8osWJ.rules.json from 0050-ZrLf3KkQ.rules.json check if update worked """ #------------------------------------------------------- # check second scope for AN 0050 working correctly # ------------------------------------------------------ l_uri = G_TEST_HOST+'/path.html' l_headers = {'host': 'test.com', 'waf-scopes-id':'0050', 'User-Agent': 'monkeez'} l_r = requests.get(l_uri, headers=l_headers) assert l_r.status_code == 403 assert l_r.text == 'This is rules custom response\n' #------------------------------------------------------- # change the 'rules_prod_id' value for second scope # and update. # check if update was successful # ------------------------------------------------------ l_conf = {} l_file_path = os.path.dirname(os.path.abspath(__file__)) l_scopes_conf_path = os.path.realpath(os.path.join(l_file_path, '../../data/waf/conf/scopes/0050.scopes.json')) try: with open(l_scopes_conf_path) as l_f: l_conf = json.load(l_f) except Exception as l_e: print('error opening config file: %s. Reason: %s error: %s, doc: %s' % ( l_scopes_conf_path, type(l_e), l_e, l_e.__doc__)) assert False l_conf['scopes'][1]['rules_prod_id'] = "0gG8osWJ" l_conf['last_modified_date'] = datetime.datetime.utcnow().strftime('%Y-%m-%dT%H:%M:%S.%fZ') #------------------------------------------------------- # POST conf # ------------------------------------------------------ l_url = '%s/update_scopes'%(G_TEST_HOST) l_headers = {'Content-Type': 'application/json'} l_r = requests.post(l_url, headers=l_headers, data=json.dumps(l_conf)) assert l_r.status_code == 200 #------------------------------------------------------- # make the same request. should get 200 # ------------------------------------------------------ l_uri = G_TEST_HOST+'/path.html' l_headers = {'host': 'test.com', 'waf-scopes-id':'0050', 'User-Agent': 'monkeez'} l_r = requests.get(l_uri, headers=l_headers) assert l_r.status_code == 200 #assert l_r.text == 'This is from GLOB scope\n' #------------------------------------------------------- # make a request with user-agent bananas # ------------------------------------------------------ l_uri = G_TEST_HOST+'/path.html' l_headers = {'host': 'test.com', 'waf-scopes-id':'0050', 'User-Agent': 'bananas'} l_r = requests.get(l_uri, headers=l_headers) assert l_r.status_code == 403 assert l_r.text == 'This is rules custom response\n' # ------------------------------------------------------------------------------ # test /update_bots endpoint # ------------------------------------------------------------------------------ def test_update_bots_endpoint(setup_scopez_server_action): l_url = G_TEST_HOST + '/update_bots' l_file_path = os.path.dirname(os.path.abspath(__file__)) l_test_file = os.path.realpath(os.path.join(l_file_path, '../../data/waf/conf/bots/0052-wHyMHxV7.bots.json')) l_test_payload = '' # ------------------------------------------------------ # check setup # ------------------------------------------------------ assert os.path.exists(l_test_file), 'test file not found!' # ------------------------------------------------------ # slurp test file # ------------------------------------------------------ with open(l_test_file) as l_tf: l_test_payload = l_tf.read() # ------------------------------------------------------ # check setup # ------------------------------------------------------ assert l_test_payload, 'payload is empty!' l_json_payload = json.loads(l_test_payload) # ------------------------------------------------------ # Check that challenge works # ------------------------------------------------------ l_uri = G_TEST_HOST+'/test.html' l_headers = {'host': 'mybot.com', 'user-agent': 'bot-testing', 'waf-scopes-id': '0052'} l_r = requests.get(l_uri, headers=l_headers) assert l_r.status_code == 401 # ------------------------------------------------------ # Update the bot config # ------------------------------------------------------ l_json_payload['directive'][0]['sec_rule']['operator']['value'] = 'chowdah' # ------------------------------------------------------ # update the timestamp, else it will silently do nothing and return 200 # ref: scopes.cc:load_bots (compare time) # ------------------------------------------------------ l_json_payload['last_modified_date'] = datetime.datetime.now().strftime('%Y-%m-%dT%H:%M:%SZ') l_result = requests.post(l_url, timeout=3, json=l_json_payload) assert l_result.status_code == 200 assert l_result.json()['status'] == 'success' # ------------------------------------------------------ # Expect 200 # ------------------------------------------------------ l_uri = G_TEST_HOST+'/test.html' l_headers = {'host': 'mybot.com', 'user-agent': 'bot-testing', 'waf-scopes-id': '0052'} l_r = requests.get(l_uri, headers=l_headers) assert l_r.status_code == 200,\ "expecting 200, got {resp_code} since user-agent changed to chowdah".format(resp_code=l_r.status_code) # ------------------------------------------------------ # Expect 401 due to new UA # ------------------------------------------------------ l_uri = G_TEST_HOST+'/test.html' l_headers = {'host': 'mybot.com', 'user-agent': 'chowdah', 'waf-scopes-id': '0052'} l_r = requests.get(l_uri, headers=l_headers) assert l_r.status_code == 401,\ "expecting 401, got {resp_code} since user-agent changed to chowdah".format(resp_code=l_r.status_code) # ------------------------------------------------------ # check negative test - missing customer_id field # ------------------------------------------------------ l_cust_id = l_json_payload.pop('customer_id') l_n2_result = requests.post(l_url, json=l_json_payload) assert l_n2_result.status_code == 500,\ 'expected 500 since customer_id {} is removed'.format(l_cust_id)
[ [ [ 229, 239 ], [ 743, 753 ], [ 788, 798 ], [ 812, 822 ], [ 2169, 2179 ] ], [ [ 247, 249 ], [ 1311, 1313 ], [ 1327, 1329 ], [ 1378, 1380 ], [ 1395, 1397 ], [ 1485, 1487 ], [ 1502, 1504 ], [ 1585, 1587 ], [ 1602, 1604 ], [ 1673, 1675 ], [ 1690, 1692 ], [ 1762, 1764 ], [ 1779, 1781 ], [ 1852, 1854 ], [ 1869, 1871 ], [ 1955, 1957 ], [ 1971, 1973 ], [ 2071, 2073 ], [ 2088, 2090 ], [ 4496, 4498 ], [ 4512, 4514 ], [ 4561, 4563 ], [ 4578, 4580 ], [ 6814, 6816 ], [ 6830, 6832 ], [ 6881, 6883 ], [ 6898, 6900 ], [ 10093, 10095 ], [ 10109, 10111 ], [ 10162, 10164 ], [ 10179, 10181 ], [ 13588, 13590 ], [ 13604, 13606 ], [ 13655, 13657 ], [ 13672, 13674 ], [ 16598, 16600 ], [ 16614, 16616 ], [ 16666, 16668 ], [ 16683, 16685 ], [ 19599, 19601 ], [ 19615, 19617 ], [ 19667, 19669 ], [ 19684, 19686 ], [ 21818, 21820 ], [ 21834, 21836 ], [ 21879, 21881 ], [ 21896, 21898 ], [ 22198, 22200 ] ], [ [ 257, 261 ], [ 4728, 4732 ], [ 5462, 5466 ], [ 7054, 7058 ], [ 7817, 7821 ], [ 10341, 10345 ], [ 11107, 11111 ], [ 13828, 13832 ], [ 14534, 14538 ], [ 16833, 16837 ], [ 17536, 17540 ], [ 19834, 19838 ], [ 20533, 20537 ], [ 22675, 22679 ] ], [ [ 269, 273 ], [ 3210, 3214 ], [ 3596, 3600 ], [ 13365, 13369 ] ], [ [ 281, 289 ], [ 5012, 5020 ], [ 7363, 7371 ], [ 10649, 10657 ], [ 14093, 14101 ], [ 17136, 17144 ], [ 20133, 20141 ], [ 23620, 23628 ] ], [ [ 297, 305 ], [ 4155, 4163 ], [ 5369, 5377 ], [ 5857, 5865 ], [ 6450, 6458 ], [ 7724, 7732 ], [ 8232, 8240 ], [ 8663, 8671 ], [ 9302, 9310 ], [ 9791, 9799 ], [ 11014, 11022 ], [ 11539, 11547 ], [ 12074, 12082 ], [ 12579, 12587 ], [ 12666, 12674 ], [ 12955, 12963 ], [ 13042, 13050 ], [ 14441, 14449 ], [ 15004, 15012 ], [ 15091, 15099 ], [ 15592, 15600 ], [ 15679, 15687 ], [ 16229, 16237 ], [ 17443, 17451 ], [ 17968, 17976 ], [ 18485, 18493 ], [ 19198, 19206 ], [ 20440, 20448 ], [ 20920, 20928 ], [ 21379, 21387 ], [ 23030, 23038 ], [ 23690, 23698 ], [ 24144, 24152 ], [ 24652, 24660 ], [ 25082, 25090 ] ], [ [ 313, 319 ], [ 1108, 1114 ] ], [ [ 494, 505 ], [ 4008, 4019 ], [ 5251, 5262 ], [ 5710, 5721 ], [ 6304, 6315 ], [ 7606, 7617 ], [ 8086, 8097 ], [ 8517, 8528 ], [ 9146, 9157 ], [ 9630, 9641 ], [ 10896, 10907 ], [ 11378, 11389 ], [ 11915, 11926 ], [ 12437, 12448 ], [ 12808, 12819 ], [ 14323, 14334 ], [ 14862, 14873 ], [ 15445, 15456 ], [ 16065, 16076 ], [ 17367, 17378 ], [ 17804, 17815 ], [ 18321, 18332 ], [ 19043, 19054 ], [ 20364, 20375 ], [ 20765, 20776 ], [ 21224, 21235 ], [ 21771, 21782 ], [ 22869, 22880 ], [ 23983, 23994 ], [ 24495, 24506 ] ], [ [ 713, 724 ], [ 3550, 3561 ] ], [ [ 1129, 1155 ], [ 3371, 3397 ] ], [ [ 3617, 3639 ] ], [ [ 5935, 5959 ] ], [ [ 8798, 8824 ] ], [ [ 12152, 12176 ] ], [ [ 15827, 15845 ] ], [ [ 18612, 18638 ] ], [ [ 21704, 21729 ] ] ]
#!/usr/bin/env python # -*- coding: utf-8 -*- """The setup script.""" from setuptools import find_packages, setup test_requirements = [ "black>=19.10b0", "flake8>=3.8.3", "flake8-debugger>=3.2.1", ] dev_requirements = [ *test_requirements, "wheel>=0.34.2", ] requirements = [ "cdp-backend[pipeline]==3.0.2", "cdp-scrapers[king_county]>=0.3.2", ] extra_requirements = { "test": test_requirements, "dev": dev_requirements, "all": [ *requirements, *dev_requirements, ], } setup( author="JacksonMaxfield", classifiers=[ "Development Status :: 2 - Pre-Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Programming Language :: Python :: 3.9", ], description="Package containing the gather functions for Example.", install_requires=requirements, license="MIT license", long_description_content_type="text/markdown", include_package_data=True, keywords="civic technology, open government", name="cdp-king_county-backend", packages=find_packages(exclude=["tests", "*.tests", "*.tests.*"]), python_requires=">=3.9", tests_require=test_requirements, extras_require=extra_requirements, url="https://github.com/CouncilDataProject/king-county", version="1.0.0", zip_safe=False, )
[ [ [ 95, 108 ], [ 1141, 1154 ] ], [ [ 110, 115 ], [ 536, 541 ] ], [ [ 117, 134 ], [ 241, 258 ], [ 415, 432 ], [ 1246, 1263 ] ], [ [ 215, 231 ], [ 445, 461 ], [ 508, 524 ] ], [ [ 284, 296 ], [ 485, 497 ], [ 919, 931 ] ], [ [ 380, 398 ], [ 1284, 1302 ] ] ]
from scrapy.spider import BaseSpider from scrapy.http import Request from scrapy.selector import XmlXPathSelector from openrecipes.spiders.elanaspantry_spider import ElanaspantryMixin class ElanaspantryfeedSpider(BaseSpider, ElanaspantryMixin): name = "elanaspantry.feed" allowed_domains = [ "www.elanaspantry.com", "feeds.feedburner.com", "feedproxy.google.com", ] start_urls = [ "http://feeds.feedburner.com/elanaspantry", ] def parse(self, response): xxs = XmlXPathSelector(response) links = xxs.select("//item/*[local-name()='origLink']/text()").extract() return [Request(x, callback=self.parse_item) for x in links]
[ [ [ 26, 36 ], [ 215, 225 ] ], [ [ 61, 68 ], [ 652, 659 ] ], [ [ 97, 113 ], [ 527, 543 ] ], [ [ 166, 183 ], [ 227, 244 ] ], [ [ 192, 214 ] ] ]
#!/usr/bin/env python #-*- encoding: UTF-8 -*- ############################################### # Todos los derechos reservados a: # # CreceLibre Consultores en Tecnologías Ltda. # # # # ©Milton Inostroza Aguilera # # [email protected] # # 2009 # ############################################### from django.db import models from AlyMoly.mantenedor.models import Producto, Promocion, Trabajador class Turno(models.Model): """ estado: 1 --> abierto 2 --> cerrado """ fecha_apertura_sistema = models.DateTimeField() fecha_cierre_sistema = models.DateTimeField(null=True, blank=True) estado = models.IntegerField(default=1, blank=True) trabajador = models.ForeignKey(Trabajador, blank=True) monto_apertura_caja = models.IntegerField(default=0) monto_cierre_calculado = models.IntegerField(default=0, blank=True) monto_afecto = models.IntegerField(default=0, blank=True) monto_exento = models.IntegerField(default=0, blank=True) def monto_cierre_informado(self): return self.boletadeposito.total def estado_turno(self): if self.estado == 1: return "Abierto" else: return "Cerrado" def save(self, force_insert=False, force_update=False): """ Al guardar un turno abierto se verifica que el trabajador ya no cuente con un turno abierto anteriormente. """ if self.estado == 1 and len(Turno.objects.exclude(id=self.id).filter(trabajador__id=self.trabajador.id).filter(estado=1)) > 0: raise Exception(u"Usted ya cuenta con un turno abierto.") super(Turno, self).save(force_insert, force_update) class BoletaDeposito(models.Model): turno = models.OneToOneField(Turno, blank=True) veintemil = models.PositiveIntegerField(default=0, blank=True) diezmil = models.PositiveIntegerField(default=0, blank=True) cincomil = models.PositiveIntegerField(default=0, blank=True) dosmil = models.PositiveIntegerField(default=0, blank=True) mil = models.PositiveIntegerField(default=0, blank=True) quinientos = models.PositiveIntegerField(default=0, blank=True) cien = models.PositiveIntegerField(default=0, blank=True) cincuenta = models.PositiveIntegerField(default=0, blank=True) diez = models.PositiveIntegerField(default=0, blank=True) tarjetas = models.PositiveIntegerField(default=0, blank=True) otros = models.PositiveIntegerField(default=0, blank=True) total = models.PositiveIntegerField(default=0, blank=True) class Venta(models.Model): """ medio_pago: 1 --> efectivo 2 --> otro """ fecha_venta = models.DateTimeField() folio_boleta = models.PositiveIntegerField(null=True, blank=True) monto_total = models.PositiveIntegerField() monto_afecto = models.PositiveIntegerField() monto_exento = models.PositiveIntegerField() cantidad_productos = models.PositiveIntegerField() medio_pago = models.PositiveIntegerField() monto_pago = models.PositiveIntegerField(null=True) turno = models.ForeignKey('Turno') def __unicode__(self): return u"%s-%s" % (self.id, self.folio_boleta) class LineaDetalle(models.Model): cantidad = models.IntegerField() precio_venta = models.IntegerField() precio_venta_total = models.IntegerField() producto = models.ForeignKey(Producto, null=True, blank=True) promocion = models.ForeignKey(Promocion, null=True, blank=True) venta = models.ForeignKey('Venta')
[ [ [ 454, 460 ], [ 545, 551 ], [ 673, 679 ], [ 723, 729 ], [ 780, 786 ], [ 840, 846 ], [ 908, 914 ], [ 968, 974 ], [ 1030, 1036 ], [ 1092, 1098 ], [ 1849, 1855 ], [ 1876, 1882 ], [ 1932, 1938 ], [ 1997, 2003 ], [ 2063, 2069 ], [ 2127, 2133 ], [ 2188, 2194 ], [ 2256, 2262 ], [ 2318, 2324 ], [ 2385, 2391 ], [ 2447, 2453 ], [ 2513, 2519 ], [ 2576, 2582 ], [ 2639, 2645 ], [ 2704, 2710 ], [ 2823, 2829 ], [ 2865, 2871 ], [ 2934, 2940 ], [ 2983, 2989 ], [ 3032, 3038 ], [ 3087, 3093 ], [ 3134, 3140 ], [ 3181, 3187 ], [ 3232, 3238 ], [ 3363, 3369 ], [ 3393, 3399 ], [ 3434, 3440 ], [ 3481, 3487 ], [ 3518, 3524 ], [ 3585, 3591 ], [ 3649, 3655 ] ], [ [ 499, 507 ], [ 3536, 3544 ] ], [ [ 509, 518 ], [ 3603, 3612 ] ], [ [ 520, 530 ], [ 858, 868 ] ], [ [ 539, 544 ], [ 1897, 1902 ], [ 1597, 1602 ], [ 1780, 1785 ] ], [ [ 1834, 1848 ] ], [ [ 2698, 2703 ] ], [ [ 3350, 3362 ] ] ]
"""Train (basic) densely-connected oracle.""" import os import time import multiprocessing as mp import pandas as pd import torch from torch import optim from torch.utils.data import DataLoader, Subset, TensorDataset, WeightedRandomSampler from profit.dataset.splitters import split_method_dict from profit.models.torch import SequenceOracle from profit.utils.data_utils.tokenizers import AminoAcidTokenizer from profit.utils.training_utils.torch import losses as L from profit.utils.training_utils.torch.callbacks import ModelCheckpoint from profit.utils.training_utils.torch.callbacks import EarlyStopping from examples.gb1.data import load_dataset timestep = time.strftime("%Y-%b-%d-%H:%M:%S", time.gmtime()) device = torch.device("cuda" if torch.cuda.is_available() else "cpu") tensor = torch.cuda.FloatTensor if torch.cuda.is_available() else torch.Tensor splits = ["train", "valid"] # Preprocess + load the dataset dataset = load_dataset("lstm", "primary", labels="Fitness", num_data=-1, filetype="mdb", as_numpy=False, vocab="aa20") # Stratify train/val/test sets s.t. the target labels are equally represented in # each subset. Each subset will have the same ratio of low/mid/high variants in # each batch as the full dataset. See: https://discuss.pytorch.org/t/29907/2 _dataset = dataset[:]["arr_0"] _labels = dataset[:]["arr_1"].view(-1) # # Remove samples below a certain threshold # high_idx = torch.where(_labels > _labels.mean()) # dataset = Subset(dataset, sorted(high_idx)) # _dataset = _dataset[high_idx] # _labels = _labels[high_idx] # Compute sample weights (each sample should get its own weight) def sampler(labels: torch.Tensor, nbins: int = 10, stratify: bool = False) -> WeightedRandomSampler: discretize = pd.qcut if stratify else pd.cut bin_labels = torch.LongTensor(discretize(labels.tolist(), nbins, labels=False, duplicates="drop")) class_sample_count = torch.LongTensor( [(bin_labels == t).sum() for t in torch.arange(nbins)]) weight = 1. / class_sample_count.float() sample_weights = torch.zeros_like(labels) for t in torch.unique(bin_labels): sample_weights[bin_labels == t] = weight[t] return WeightedRandomSampler(sample_weights, len(sample_weights)) # Compute sample weights and add to original dataset weights = sampler(_labels, nbins=10, stratify=False).weights.type(torch.float) dataset = TensorDataset(*dataset[:].values(), weights) # Create subset indicies subset_idx = split_method_dict["stratified"]().train_valid_test_split( dataset=_dataset, labels=_labels.tolist(), frac_train=0.9, frac_valid=0.1, frac_test=0.0, return_idxs=True, n_bins=10) stratified = {split: Subset(dataset, sorted(idx)) for split, idx in zip(splits, subset_idx)} # Create stratified sampler (only needed for training) train_sampler = sampler(stratified["train"][:][1].view(-1), stratify=True) # Initialize model tokenizer = AminoAcidTokenizer("aa20") vocab_size = tokenizer.vocab_size seqlen = stratified["train"][0][0].size(0) model = SequenceOracle(seqlen, vocab_size, hidden_size=50, out_size=2) # Initialize callbacks # NOTE: Must set model (within save_clbk) to ensure weights get saved stop_clbk = EarlyStopping(patience=5, verbose=1) save_clbk = ModelCheckpoint(os.path.join("bin/3gb1/oracle", timestep), monitor="val_loss", verbose=1, save_weights_only=True) save_clbk.set_model(model) # Initialize callbacks optimizer = optim.AdamW(model.parameters(), lr=1e-3) epochs = 50 for epoch in range(1, epochs+1): for split in splits: summed_loss = 0 data_loader = DataLoader( dataset=stratified[split], batch_size=32, sampler=train_sampler if split == "train" else None, num_workers=mp.cpu_count(), pin_memory=torch.cuda.is_available() ) # Enable/disable dropout model.train() if split == "train" else model.eval() for it, batch in enumerate(data_loader): data = batch[0].long().to(device) target = batch[1].to(device) sample_weight = batch[2].to(device) # One-hot encode (see: https://discuss.pytorch.org/t/507/34) batch_size, seqlen = data.size() onehot = torch.zeros(batch_size, seqlen, vocab_size) onehot.scatter_(2, torch.unsqueeze(data, 2), 1) # Forward pass pred = model(onehot) # Loss calculation nll_loss = L.gaussian_nll_loss(pred, target, reduction="none") # Reweight nll_loss w/ sample weights nll_loss = (nll_loss * sample_weight).sum() summed_loss += nll_loss.item() loss = nll_loss / batch_size # Compute gradients and update params/weights if split == "train": optimizer.zero_grad() loss.backward() optimizer.step() # Bookkeeping (batch) if it % 5 == 0 or it+1 == len(data_loader): print("{} Batch {:04d}/{:d} ({:.2f}%)\tLoss: {:.4f}".format( split.upper(), it+1, len(data_loader), 100. * ((it+1)/len(data_loader)), loss.item())) # Bookkeeping (epoch) avg_loss = summed_loss / len(data_loader.dataset) print("{} Epoch {}/{}, Average NLL loss: {:.4f}".format( split.upper(), epoch, epochs, avg_loss)) # Stop training (based off val loss) and save (top k) ckpts if split == "valid": save_clbk.on_epoch_end(epoch, logs={"val_loss": avg_loss}) should_stop = stop_clbk.on_epoch_end(epoch, logs={"val_loss": avg_loss}) if should_stop: break else: continue break
[ [ [ 54, 56 ], [ 3358, 3360 ] ], [ [ 64, 68 ], [ 669, 673 ], [ 704, 708 ] ], [ [ 76, 97 ], [ 3928, 3930 ] ], [ [ 106, 118 ], [ 1791, 1793 ], [ 1816, 1818 ] ], [ [ 127, 132 ], [ 728, 733 ], [ 751, 756 ], [ 824, 829 ], [ 798, 803 ], [ 855, 860 ], [ 2450, 2455 ], [ 3967, 3972 ], [ 4421, 4426 ], [ 4496, 4501 ], [ 1669, 1674 ], [ 1840, 1845 ], [ 1996, 2001 ], [ 2056, 2061 ], [ 2144, 2149 ], [ 2182, 2187 ] ], [ [ 151, 156 ], [ 3603, 3608 ] ], [ [ 186, 196 ], [ 3761, 3771 ] ], [ [ 198, 204 ], [ 2763, 2769 ] ], [ [ 206, 219 ], [ 2473, 2486 ] ], [ [ 221, 242 ], [ 1751, 1772 ], [ 2271, 2292 ] ], [ [ 281, 298 ], [ 2557, 2574 ] ], [ [ 331, 345 ], [ 3124, 3138 ] ], [ [ 393, 411 ], [ 3012, 3030 ] ], [ [ 458, 469 ], [ 4640, 4641 ] ], [ [ 526, 541 ], [ 3342, 3357 ] ], [ [ 598, 611 ], [ 3293, 3306 ] ], [ [ 643, 655 ], [ 939, 951 ] ], [ [ 658, 666 ], [ 3390, 3398 ] ], [ [ 719, 725 ], [ 4185, 4191 ], [ 4226, 4232 ], [ 4274, 4280 ] ], [ [ 789, 795 ] ], [ [ 868, 874 ], [ 2828, 2834 ], [ 3707, 3713 ] ], [ [ 929, 936 ], [ 1320, 1327 ], [ 1350, 1357 ], [ 2488, 2495 ] ], [ [ 1309, 1317 ], [ 2627, 2635 ] ], [ [ 1340, 1347 ], [ 2402, 2409 ], [ 2644, 2651 ] ], [ [ 1653, 1660 ], [ 2394, 2401 ], [ 2921, 2928 ] ], [ [ 2384, 2391 ], [ 2509, 2516 ] ], [ [ 2463, 2470 ], [ 2770, 2777 ] ], [ [ 2544, 2554 ], [ 2836, 2846 ] ], [ [ 2742, 2752 ], [ 2929, 2939 ], [ 3082, 3092 ], [ 3793, 3803 ] ], [ [ 2905, 2918 ], [ 3859, 3872 ] ], [ [ 3000, 3009 ], [ 3052, 3061 ] ], [ [ 3039, 3049 ], [ 3147, 3157 ], [ 4453, 4463 ] ], [ [ 3073, 3079 ], [ 3139, 3145 ] ], [ [ 3116, 3121 ], [ 3560, 3565 ], [ 3615, 3620 ], [ 4045, 4050 ], [ 4084, 4089 ], [ 4572, 4577 ] ], [ [ 3281, 3290 ], [ 5773, 5782 ] ], [ [ 3330, 3339 ], [ 3540, 3549 ], [ 5688, 5697 ] ], [ [ 3591, 3600 ], [ 4989, 4998 ], [ 5059, 5068 ] ], [ [ 3645, 3651 ], [ 3679, 3685 ], [ 5559, 5565 ] ], [ [ 3661, 3666 ], [ 5552, 5557 ], [ 5711, 5716 ], [ 5796, 5801 ] ], [ [ 3698, 3703 ], [ 3804, 3809 ], [ 3876, 3881 ], [ 4062, 4067 ], [ 4955, 4960 ], [ 5264, 5269 ], [ 5537, 5542 ], [ 5658, 5663 ] ], [ [ 3723, 3734 ], [ 4810, 4821 ], [ 5421, 5432 ] ], [ [ 3747, 3758 ], [ 4133, 4144 ], [ 5153, 5164 ], [ 5289, 5300 ], [ 5342, 5353 ], [ 5439, 5450 ] ], [ [ 4110, 4112 ], [ 5126, 5128 ], [ 5141, 5143 ], [ 5279, 5281 ], [ 5332, 5334 ] ], [ [ 4114, 4119 ], [ 4166, 4171 ], [ 4214, 4219 ], [ 4262, 4267 ] ], [ [ 4159, 4163 ], [ 4388, 4392 ], [ 4512, 4516 ] ], [ [ 4205, 4211 ], [ 4666, 4672 ] ], [ [ 4246, 4259 ], [ 4777, 4790 ] ], [ [ 4367, 4377 ], [ 4433, 4443 ], [ 4871, 4881 ] ], [ [ 4379, 4385 ], [ 4445, 4451 ] ], [ [ 4412, 4418 ], [ 4477, 4483 ], [ 4578, 4584 ] ], [ [ 4565, 4569 ], [ 4660, 4664 ] ], [ [ 4629, 4637 ], [ 4766, 4774 ] ], [ [ 4754, 4762 ], [ 4825, 4833 ], [ 4860, 4868 ] ], [ [ 4853, 4857 ], [ 5027, 5031 ], [ 5357, 5361 ] ], [ [ 5410, 5418 ], [ 5567, 5575 ], [ 5736, 5744 ], [ 5821, 5829 ] ], [ [ 5759, 5770 ], [ 5847, 5858 ] ] ]
# -*- coding: utf-8 -*- # Copyright 2020 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import os import mock import grpc from grpc.experimental import aio import math import pytest from proto.marshal.rules.dates import DurationRule, TimestampRule from google.api_core import client_options from google.api_core import exceptions as core_exceptions from google.api_core import future from google.api_core import gapic_v1 from google.api_core import grpc_helpers from google.api_core import grpc_helpers_async from google.api_core import operation_async # type: ignore from google.api_core import operations_v1 from google.api_core import path_template from google.auth import credentials as ga_credentials from google.auth.exceptions import MutualTLSChannelError from google.cloud.deploy_v1.services.cloud_deploy import CloudDeployAsyncClient from google.cloud.deploy_v1.services.cloud_deploy import CloudDeployClient from google.cloud.deploy_v1.services.cloud_deploy import pagers from google.cloud.deploy_v1.services.cloud_deploy import transports from google.cloud.deploy_v1.types import cloud_deploy from google.longrunning import operations_pb2 from google.oauth2 import service_account from google.protobuf import field_mask_pb2 # type: ignore from google.protobuf import timestamp_pb2 # type: ignore import google.auth def client_cert_source_callback(): return b"cert bytes", b"key bytes" # If default endpoint is localhost, then default mtls endpoint will be the same. # This method modifies the default endpoint so the client can produce a different # mtls endpoint for endpoint testing purposes. def modify_default_endpoint(client): return ( "foo.googleapis.com" if ("localhost" in client.DEFAULT_ENDPOINT) else client.DEFAULT_ENDPOINT ) def test__get_default_mtls_endpoint(): api_endpoint = "example.googleapis.com" api_mtls_endpoint = "example.mtls.googleapis.com" sandbox_endpoint = "example.sandbox.googleapis.com" sandbox_mtls_endpoint = "example.mtls.sandbox.googleapis.com" non_googleapi = "api.example.com" assert CloudDeployClient._get_default_mtls_endpoint(None) is None assert ( CloudDeployClient._get_default_mtls_endpoint(api_endpoint) == api_mtls_endpoint ) assert ( CloudDeployClient._get_default_mtls_endpoint(api_mtls_endpoint) == api_mtls_endpoint ) assert ( CloudDeployClient._get_default_mtls_endpoint(sandbox_endpoint) == sandbox_mtls_endpoint ) assert ( CloudDeployClient._get_default_mtls_endpoint(sandbox_mtls_endpoint) == sandbox_mtls_endpoint ) assert CloudDeployClient._get_default_mtls_endpoint(non_googleapi) == non_googleapi @pytest.mark.parametrize("client_class", [CloudDeployClient, CloudDeployAsyncClient,]) def test_cloud_deploy_client_from_service_account_info(client_class): creds = ga_credentials.AnonymousCredentials() with mock.patch.object( service_account.Credentials, "from_service_account_info" ) as factory: factory.return_value = creds info = {"valid": True} client = client_class.from_service_account_info(info) assert client.transport._credentials == creds assert isinstance(client, client_class) assert client.transport._host == "clouddeploy.googleapis.com:443" @pytest.mark.parametrize( "transport_class,transport_name", [ (transports.CloudDeployGrpcTransport, "grpc"), (transports.CloudDeployGrpcAsyncIOTransport, "grpc_asyncio"), ], ) def test_cloud_deploy_client_service_account_always_use_jwt( transport_class, transport_name ): with mock.patch.object( service_account.Credentials, "with_always_use_jwt_access", create=True ) as use_jwt: creds = service_account.Credentials(None, None, None) transport = transport_class(credentials=creds, always_use_jwt_access=True) use_jwt.assert_called_once_with(True) with mock.patch.object( service_account.Credentials, "with_always_use_jwt_access", create=True ) as use_jwt: creds = service_account.Credentials(None, None, None) transport = transport_class(credentials=creds, always_use_jwt_access=False) use_jwt.assert_not_called() @pytest.mark.parametrize("client_class", [CloudDeployClient, CloudDeployAsyncClient,]) def test_cloud_deploy_client_from_service_account_file(client_class): creds = ga_credentials.AnonymousCredentials() with mock.patch.object( service_account.Credentials, "from_service_account_file" ) as factory: factory.return_value = creds client = client_class.from_service_account_file("dummy/file/path.json") assert client.transport._credentials == creds assert isinstance(client, client_class) client = client_class.from_service_account_json("dummy/file/path.json") assert client.transport._credentials == creds assert isinstance(client, client_class) assert client.transport._host == "clouddeploy.googleapis.com:443" def test_cloud_deploy_client_get_transport_class(): transport = CloudDeployClient.get_transport_class() available_transports = [ transports.CloudDeployGrpcTransport, ] assert transport in available_transports transport = CloudDeployClient.get_transport_class("grpc") assert transport == transports.CloudDeployGrpcTransport @pytest.mark.parametrize( "client_class,transport_class,transport_name", [ (CloudDeployClient, transports.CloudDeployGrpcTransport, "grpc"), ( CloudDeployAsyncClient, transports.CloudDeployGrpcAsyncIOTransport, "grpc_asyncio", ), ], ) @mock.patch.object( CloudDeployClient, "DEFAULT_ENDPOINT", modify_default_endpoint(CloudDeployClient) ) @mock.patch.object( CloudDeployAsyncClient, "DEFAULT_ENDPOINT", modify_default_endpoint(CloudDeployAsyncClient), ) def test_cloud_deploy_client_client_options( client_class, transport_class, transport_name ): # Check that if channel is provided we won't create a new one. with mock.patch.object(CloudDeployClient, "get_transport_class") as gtc: transport = transport_class(credentials=ga_credentials.AnonymousCredentials()) client = client_class(transport=transport) gtc.assert_not_called() # Check that if channel is provided via str we will create a new one. with mock.patch.object(CloudDeployClient, "get_transport_class") as gtc: client = client_class(transport=transport_name) gtc.assert_called() # Check the case api_endpoint is provided. options = client_options.ClientOptions(api_endpoint="squid.clam.whelk") with mock.patch.object(transport_class, "__init__") as patched: patched.return_value = None client = client_class(transport=transport_name, client_options=options) patched.assert_called_once_with( credentials=None, credentials_file=None, host="squid.clam.whelk", scopes=None, client_cert_source_for_mtls=None, quota_project_id=None, client_info=transports.base.DEFAULT_CLIENT_INFO, always_use_jwt_access=True, ) # Check the case api_endpoint is not provided and GOOGLE_API_USE_MTLS_ENDPOINT is # "never". with mock.patch.dict(os.environ, {"GOOGLE_API_USE_MTLS_ENDPOINT": "never"}): with mock.patch.object(transport_class, "__init__") as patched: patched.return_value = None client = client_class(transport=transport_name) patched.assert_called_once_with( credentials=None, credentials_file=None, host=client.DEFAULT_ENDPOINT, scopes=None, client_cert_source_for_mtls=None, quota_project_id=None, client_info=transports.base.DEFAULT_CLIENT_INFO, always_use_jwt_access=True, ) # Check the case api_endpoint is not provided and GOOGLE_API_USE_MTLS_ENDPOINT is # "always". with mock.patch.dict(os.environ, {"GOOGLE_API_USE_MTLS_ENDPOINT": "always"}): with mock.patch.object(transport_class, "__init__") as patched: patched.return_value = None client = client_class(transport=transport_name) patched.assert_called_once_with( credentials=None, credentials_file=None, host=client.DEFAULT_MTLS_ENDPOINT, scopes=None, client_cert_source_for_mtls=None, quota_project_id=None, client_info=transports.base.DEFAULT_CLIENT_INFO, always_use_jwt_access=True, ) # Check the case api_endpoint is not provided and GOOGLE_API_USE_MTLS_ENDPOINT has # unsupported value. with mock.patch.dict(os.environ, {"GOOGLE_API_USE_MTLS_ENDPOINT": "Unsupported"}): with pytest.raises(MutualTLSChannelError): client = client_class(transport=transport_name) # Check the case GOOGLE_API_USE_CLIENT_CERTIFICATE has unsupported value. with mock.patch.dict( os.environ, {"GOOGLE_API_USE_CLIENT_CERTIFICATE": "Unsupported"} ): with pytest.raises(ValueError): client = client_class(transport=transport_name) # Check the case quota_project_id is provided options = client_options.ClientOptions(quota_project_id="octopus") with mock.patch.object(transport_class, "__init__") as patched: patched.return_value = None client = client_class(client_options=options, transport=transport_name) patched.assert_called_once_with( credentials=None, credentials_file=None, host=client.DEFAULT_ENDPOINT, scopes=None, client_cert_source_for_mtls=None, quota_project_id="octopus", client_info=transports.base.DEFAULT_CLIENT_INFO, always_use_jwt_access=True, ) @pytest.mark.parametrize( "client_class,transport_class,transport_name,use_client_cert_env", [ (CloudDeployClient, transports.CloudDeployGrpcTransport, "grpc", "true"), ( CloudDeployAsyncClient, transports.CloudDeployGrpcAsyncIOTransport, "grpc_asyncio", "true", ), (CloudDeployClient, transports.CloudDeployGrpcTransport, "grpc", "false"), ( CloudDeployAsyncClient, transports.CloudDeployGrpcAsyncIOTransport, "grpc_asyncio", "false", ), ], ) @mock.patch.object( CloudDeployClient, "DEFAULT_ENDPOINT", modify_default_endpoint(CloudDeployClient) ) @mock.patch.object( CloudDeployAsyncClient, "DEFAULT_ENDPOINT", modify_default_endpoint(CloudDeployAsyncClient), ) @mock.patch.dict(os.environ, {"GOOGLE_API_USE_MTLS_ENDPOINT": "auto"}) def test_cloud_deploy_client_mtls_env_auto( client_class, transport_class, transport_name, use_client_cert_env ): # This tests the endpoint autoswitch behavior. Endpoint is autoswitched to the default # mtls endpoint, if GOOGLE_API_USE_CLIENT_CERTIFICATE is "true" and client cert exists. # Check the case client_cert_source is provided. Whether client cert is used depends on # GOOGLE_API_USE_CLIENT_CERTIFICATE value. with mock.patch.dict( os.environ, {"GOOGLE_API_USE_CLIENT_CERTIFICATE": use_client_cert_env} ): options = client_options.ClientOptions( client_cert_source=client_cert_source_callback ) with mock.patch.object(transport_class, "__init__") as patched: patched.return_value = None client = client_class(client_options=options, transport=transport_name) if use_client_cert_env == "false": expected_client_cert_source = None expected_host = client.DEFAULT_ENDPOINT else: expected_client_cert_source = client_cert_source_callback expected_host = client.DEFAULT_MTLS_ENDPOINT patched.assert_called_once_with( credentials=None, credentials_file=None, host=expected_host, scopes=None, client_cert_source_for_mtls=expected_client_cert_source, quota_project_id=None, client_info=transports.base.DEFAULT_CLIENT_INFO, always_use_jwt_access=True, ) # Check the case ADC client cert is provided. Whether client cert is used depends on # GOOGLE_API_USE_CLIENT_CERTIFICATE value. with mock.patch.dict( os.environ, {"GOOGLE_API_USE_CLIENT_CERTIFICATE": use_client_cert_env} ): with mock.patch.object(transport_class, "__init__") as patched: with mock.patch( "google.auth.transport.mtls.has_default_client_cert_source", return_value=True, ): with mock.patch( "google.auth.transport.mtls.default_client_cert_source", return_value=client_cert_source_callback, ): if use_client_cert_env == "false": expected_host = client.DEFAULT_ENDPOINT expected_client_cert_source = None else: expected_host = client.DEFAULT_MTLS_ENDPOINT expected_client_cert_source = client_cert_source_callback patched.return_value = None client = client_class(transport=transport_name) patched.assert_called_once_with( credentials=None, credentials_file=None, host=expected_host, scopes=None, client_cert_source_for_mtls=expected_client_cert_source, quota_project_id=None, client_info=transports.base.DEFAULT_CLIENT_INFO, always_use_jwt_access=True, ) # Check the case client_cert_source and ADC client cert are not provided. with mock.patch.dict( os.environ, {"GOOGLE_API_USE_CLIENT_CERTIFICATE": use_client_cert_env} ): with mock.patch.object(transport_class, "__init__") as patched: with mock.patch( "google.auth.transport.mtls.has_default_client_cert_source", return_value=False, ): patched.return_value = None client = client_class(transport=transport_name) patched.assert_called_once_with( credentials=None, credentials_file=None, host=client.DEFAULT_ENDPOINT, scopes=None, client_cert_source_for_mtls=None, quota_project_id=None, client_info=transports.base.DEFAULT_CLIENT_INFO, always_use_jwt_access=True, ) @pytest.mark.parametrize("client_class", [CloudDeployClient, CloudDeployAsyncClient]) @mock.patch.object( CloudDeployClient, "DEFAULT_ENDPOINT", modify_default_endpoint(CloudDeployClient) ) @mock.patch.object( CloudDeployAsyncClient, "DEFAULT_ENDPOINT", modify_default_endpoint(CloudDeployAsyncClient), ) def test_cloud_deploy_client_get_mtls_endpoint_and_cert_source(client_class): mock_client_cert_source = mock.Mock() # Test the case GOOGLE_API_USE_CLIENT_CERTIFICATE is "true". with mock.patch.dict(os.environ, {"GOOGLE_API_USE_CLIENT_CERTIFICATE": "true"}): mock_api_endpoint = "foo" options = client_options.ClientOptions( client_cert_source=mock_client_cert_source, api_endpoint=mock_api_endpoint ) api_endpoint, cert_source = client_class.get_mtls_endpoint_and_cert_source( options ) assert api_endpoint == mock_api_endpoint assert cert_source == mock_client_cert_source # Test the case GOOGLE_API_USE_CLIENT_CERTIFICATE is "false". with mock.patch.dict(os.environ, {"GOOGLE_API_USE_CLIENT_CERTIFICATE": "false"}): mock_client_cert_source = mock.Mock() mock_api_endpoint = "foo" options = client_options.ClientOptions( client_cert_source=mock_client_cert_source, api_endpoint=mock_api_endpoint ) api_endpoint, cert_source = client_class.get_mtls_endpoint_and_cert_source( options ) assert api_endpoint == mock_api_endpoint assert cert_source is None # Test the case GOOGLE_API_USE_MTLS_ENDPOINT is "never". with mock.patch.dict(os.environ, {"GOOGLE_API_USE_MTLS_ENDPOINT": "never"}): api_endpoint, cert_source = client_class.get_mtls_endpoint_and_cert_source() assert api_endpoint == client_class.DEFAULT_ENDPOINT assert cert_source is None # Test the case GOOGLE_API_USE_MTLS_ENDPOINT is "always". with mock.patch.dict(os.environ, {"GOOGLE_API_USE_MTLS_ENDPOINT": "always"}): api_endpoint, cert_source = client_class.get_mtls_endpoint_and_cert_source() assert api_endpoint == client_class.DEFAULT_MTLS_ENDPOINT assert cert_source is None # Test the case GOOGLE_API_USE_MTLS_ENDPOINT is "auto" and default cert doesn't exist. with mock.patch.dict(os.environ, {"GOOGLE_API_USE_CLIENT_CERTIFICATE": "true"}): with mock.patch( "google.auth.transport.mtls.has_default_client_cert_source", return_value=False, ): api_endpoint, cert_source = client_class.get_mtls_endpoint_and_cert_source() assert api_endpoint == client_class.DEFAULT_ENDPOINT assert cert_source is None # Test the case GOOGLE_API_USE_MTLS_ENDPOINT is "auto" and default cert exists. with mock.patch.dict(os.environ, {"GOOGLE_API_USE_CLIENT_CERTIFICATE": "true"}): with mock.patch( "google.auth.transport.mtls.has_default_client_cert_source", return_value=True, ): with mock.patch( "google.auth.transport.mtls.default_client_cert_source", return_value=mock_client_cert_source, ): ( api_endpoint, cert_source, ) = client_class.get_mtls_endpoint_and_cert_source() assert api_endpoint == client_class.DEFAULT_MTLS_ENDPOINT assert cert_source == mock_client_cert_source @pytest.mark.parametrize( "client_class,transport_class,transport_name", [ (CloudDeployClient, transports.CloudDeployGrpcTransport, "grpc"), ( CloudDeployAsyncClient, transports.CloudDeployGrpcAsyncIOTransport, "grpc_asyncio", ), ], ) def test_cloud_deploy_client_client_options_scopes( client_class, transport_class, transport_name ): # Check the case scopes are provided. options = client_options.ClientOptions(scopes=["1", "2"],) with mock.patch.object(transport_class, "__init__") as patched: patched.return_value = None client = client_class(client_options=options, transport=transport_name) patched.assert_called_once_with( credentials=None, credentials_file=None, host=client.DEFAULT_ENDPOINT, scopes=["1", "2"], client_cert_source_for_mtls=None, quota_project_id=None, client_info=transports.base.DEFAULT_CLIENT_INFO, always_use_jwt_access=True, ) @pytest.mark.parametrize( "client_class,transport_class,transport_name", [ (CloudDeployClient, transports.CloudDeployGrpcTransport, "grpc"), ( CloudDeployAsyncClient, transports.CloudDeployGrpcAsyncIOTransport, "grpc_asyncio", ), ], ) def test_cloud_deploy_client_client_options_credentials_file( client_class, transport_class, transport_name ): # Check the case credentials file is provided. options = client_options.ClientOptions(credentials_file="credentials.json") with mock.patch.object(transport_class, "__init__") as patched: patched.return_value = None client = client_class(client_options=options, transport=transport_name) patched.assert_called_once_with( credentials=None, credentials_file="credentials.json", host=client.DEFAULT_ENDPOINT, scopes=None, client_cert_source_for_mtls=None, quota_project_id=None, client_info=transports.base.DEFAULT_CLIENT_INFO, always_use_jwt_access=True, ) def test_cloud_deploy_client_client_options_from_dict(): with mock.patch( "google.cloud.deploy_v1.services.cloud_deploy.transports.CloudDeployGrpcTransport.__init__" ) as grpc_transport: grpc_transport.return_value = None client = CloudDeployClient(client_options={"api_endpoint": "squid.clam.whelk"}) grpc_transport.assert_called_once_with( credentials=None, credentials_file=None, host="squid.clam.whelk", scopes=None, client_cert_source_for_mtls=None, quota_project_id=None, client_info=transports.base.DEFAULT_CLIENT_INFO, always_use_jwt_access=True, ) @pytest.mark.parametrize( "request_type", [cloud_deploy.ListDeliveryPipelinesRequest, dict,] ) def test_list_delivery_pipelines(request_type, transport: str = "grpc"): client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, # and we are mocking out the actual API, so just send an empty request. request = request_type() # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.list_delivery_pipelines), "__call__" ) as call: # Designate an appropriate return value for the call. call.return_value = cloud_deploy.ListDeliveryPipelinesResponse( next_page_token="next_page_token_value", unreachable=["unreachable_value"], ) response = client.list_delivery_pipelines(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.ListDeliveryPipelinesRequest() # Establish that the response is the type that we expect. assert isinstance(response, pagers.ListDeliveryPipelinesPager) assert response.next_page_token == "next_page_token_value" assert response.unreachable == ["unreachable_value"] def test_list_delivery_pipelines_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.list_delivery_pipelines), "__call__" ) as call: client.list_delivery_pipelines() call.assert_called() _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.ListDeliveryPipelinesRequest() @pytest.mark.asyncio async def test_list_delivery_pipelines_async( transport: str = "grpc_asyncio", request_type=cloud_deploy.ListDeliveryPipelinesRequest, ): client = CloudDeployAsyncClient( credentials=ga_credentials.AnonymousCredentials(), transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, # and we are mocking out the actual API, so just send an empty request. request = request_type() # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.list_delivery_pipelines), "__call__" ) as call: # Designate an appropriate return value for the call. call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( cloud_deploy.ListDeliveryPipelinesResponse( next_page_token="next_page_token_value", unreachable=["unreachable_value"], ) ) response = await client.list_delivery_pipelines(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.ListDeliveryPipelinesRequest() # Establish that the response is the type that we expect. assert isinstance(response, pagers.ListDeliveryPipelinesAsyncPager) assert response.next_page_token == "next_page_token_value" assert response.unreachable == ["unreachable_value"] @pytest.mark.asyncio async def test_list_delivery_pipelines_async_from_dict(): await test_list_delivery_pipelines_async(request_type=dict) def test_list_delivery_pipelines_field_headers(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cloud_deploy.ListDeliveryPipelinesRequest() request.parent = "parent/value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.list_delivery_pipelines), "__call__" ) as call: call.return_value = cloud_deploy.ListDeliveryPipelinesResponse() client.list_delivery_pipelines(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] assert args[0] == request # Establish that the field header was sent. _, _, kw = call.mock_calls[0] assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] @pytest.mark.asyncio async def test_list_delivery_pipelines_field_headers_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cloud_deploy.ListDeliveryPipelinesRequest() request.parent = "parent/value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.list_delivery_pipelines), "__call__" ) as call: call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( cloud_deploy.ListDeliveryPipelinesResponse() ) await client.list_delivery_pipelines(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] assert args[0] == request # Establish that the field header was sent. _, _, kw = call.mock_calls[0] assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] def test_list_delivery_pipelines_flattened(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.list_delivery_pipelines), "__call__" ) as call: # Designate an appropriate return value for the call. call.return_value = cloud_deploy.ListDeliveryPipelinesResponse() # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. client.list_delivery_pipelines(parent="parent_value",) # Establish that the underlying call was made with the expected # request object values. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] arg = args[0].parent mock_val = "parent_value" assert arg == mock_val def test_list_delivery_pipelines_flattened_error(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Attempting to call a method with both a request object and flattened # fields is an error. with pytest.raises(ValueError): client.list_delivery_pipelines( cloud_deploy.ListDeliveryPipelinesRequest(), parent="parent_value", ) @pytest.mark.asyncio async def test_list_delivery_pipelines_flattened_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.list_delivery_pipelines), "__call__" ) as call: # Designate an appropriate return value for the call. call.return_value = cloud_deploy.ListDeliveryPipelinesResponse() call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( cloud_deploy.ListDeliveryPipelinesResponse() ) # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. response = await client.list_delivery_pipelines(parent="parent_value",) # Establish that the underlying call was made with the expected # request object values. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] arg = args[0].parent mock_val = "parent_value" assert arg == mock_val @pytest.mark.asyncio async def test_list_delivery_pipelines_flattened_error_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Attempting to call a method with both a request object and flattened # fields is an error. with pytest.raises(ValueError): await client.list_delivery_pipelines( cloud_deploy.ListDeliveryPipelinesRequest(), parent="parent_value", ) def test_list_delivery_pipelines_pager(transport_name: str = "grpc"): client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials, transport=transport_name, ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.list_delivery_pipelines), "__call__" ) as call: # Set the response to a series of pages. call.side_effect = ( cloud_deploy.ListDeliveryPipelinesResponse( delivery_pipelines=[ cloud_deploy.DeliveryPipeline(), cloud_deploy.DeliveryPipeline(), cloud_deploy.DeliveryPipeline(), ], next_page_token="abc", ), cloud_deploy.ListDeliveryPipelinesResponse( delivery_pipelines=[], next_page_token="def", ), cloud_deploy.ListDeliveryPipelinesResponse( delivery_pipelines=[cloud_deploy.DeliveryPipeline(),], next_page_token="ghi", ), cloud_deploy.ListDeliveryPipelinesResponse( delivery_pipelines=[ cloud_deploy.DeliveryPipeline(), cloud_deploy.DeliveryPipeline(), ], ), RuntimeError, ) metadata = () metadata = tuple(metadata) + ( gapic_v1.routing_header.to_grpc_metadata((("parent", ""),)), ) pager = client.list_delivery_pipelines(request={}) assert pager._metadata == metadata results = [i for i in pager] assert len(results) == 6 assert all(isinstance(i, cloud_deploy.DeliveryPipeline) for i in results) def test_list_delivery_pipelines_pages(transport_name: str = "grpc"): client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials, transport=transport_name, ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.list_delivery_pipelines), "__call__" ) as call: # Set the response to a series of pages. call.side_effect = ( cloud_deploy.ListDeliveryPipelinesResponse( delivery_pipelines=[ cloud_deploy.DeliveryPipeline(), cloud_deploy.DeliveryPipeline(), cloud_deploy.DeliveryPipeline(), ], next_page_token="abc", ), cloud_deploy.ListDeliveryPipelinesResponse( delivery_pipelines=[], next_page_token="def", ), cloud_deploy.ListDeliveryPipelinesResponse( delivery_pipelines=[cloud_deploy.DeliveryPipeline(),], next_page_token="ghi", ), cloud_deploy.ListDeliveryPipelinesResponse( delivery_pipelines=[ cloud_deploy.DeliveryPipeline(), cloud_deploy.DeliveryPipeline(), ], ), RuntimeError, ) pages = list(client.list_delivery_pipelines(request={}).pages) for page_, token in zip(pages, ["abc", "def", "ghi", ""]): assert page_.raw_page.next_page_token == token @pytest.mark.asyncio async def test_list_delivery_pipelines_async_pager(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials,) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.list_delivery_pipelines), "__call__", new_callable=mock.AsyncMock, ) as call: # Set the response to a series of pages. call.side_effect = ( cloud_deploy.ListDeliveryPipelinesResponse( delivery_pipelines=[ cloud_deploy.DeliveryPipeline(), cloud_deploy.DeliveryPipeline(), cloud_deploy.DeliveryPipeline(), ], next_page_token="abc", ), cloud_deploy.ListDeliveryPipelinesResponse( delivery_pipelines=[], next_page_token="def", ), cloud_deploy.ListDeliveryPipelinesResponse( delivery_pipelines=[cloud_deploy.DeliveryPipeline(),], next_page_token="ghi", ), cloud_deploy.ListDeliveryPipelinesResponse( delivery_pipelines=[ cloud_deploy.DeliveryPipeline(), cloud_deploy.DeliveryPipeline(), ], ), RuntimeError, ) async_pager = await client.list_delivery_pipelines(request={},) assert async_pager.next_page_token == "abc" responses = [] async for response in async_pager: responses.append(response) assert len(responses) == 6 assert all(isinstance(i, cloud_deploy.DeliveryPipeline) for i in responses) @pytest.mark.asyncio async def test_list_delivery_pipelines_async_pages(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials,) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.list_delivery_pipelines), "__call__", new_callable=mock.AsyncMock, ) as call: # Set the response to a series of pages. call.side_effect = ( cloud_deploy.ListDeliveryPipelinesResponse( delivery_pipelines=[ cloud_deploy.DeliveryPipeline(), cloud_deploy.DeliveryPipeline(), cloud_deploy.DeliveryPipeline(), ], next_page_token="abc", ), cloud_deploy.ListDeliveryPipelinesResponse( delivery_pipelines=[], next_page_token="def", ), cloud_deploy.ListDeliveryPipelinesResponse( delivery_pipelines=[cloud_deploy.DeliveryPipeline(),], next_page_token="ghi", ), cloud_deploy.ListDeliveryPipelinesResponse( delivery_pipelines=[ cloud_deploy.DeliveryPipeline(), cloud_deploy.DeliveryPipeline(), ], ), RuntimeError, ) pages = [] async for page_ in (await client.list_delivery_pipelines(request={})).pages: pages.append(page_) for page_, token in zip(pages, ["abc", "def", "ghi", ""]): assert page_.raw_page.next_page_token == token @pytest.mark.parametrize( "request_type", [cloud_deploy.GetDeliveryPipelineRequest, dict,] ) def test_get_delivery_pipeline(request_type, transport: str = "grpc"): client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, # and we are mocking out the actual API, so just send an empty request. request = request_type() # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.get_delivery_pipeline), "__call__" ) as call: # Designate an appropriate return value for the call. call.return_value = cloud_deploy.DeliveryPipeline( name="name_value", uid="uid_value", description="description_value", etag="etag_value", serial_pipeline=cloud_deploy.SerialPipeline( stages=[cloud_deploy.Stage(target_id="target_id_value")] ), ) response = client.get_delivery_pipeline(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.GetDeliveryPipelineRequest() # Establish that the response is the type that we expect. assert isinstance(response, cloud_deploy.DeliveryPipeline) assert response.name == "name_value" assert response.uid == "uid_value" assert response.description == "description_value" assert response.etag == "etag_value" def test_get_delivery_pipeline_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.get_delivery_pipeline), "__call__" ) as call: client.get_delivery_pipeline() call.assert_called() _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.GetDeliveryPipelineRequest() @pytest.mark.asyncio async def test_get_delivery_pipeline_async( transport: str = "grpc_asyncio", request_type=cloud_deploy.GetDeliveryPipelineRequest, ): client = CloudDeployAsyncClient( credentials=ga_credentials.AnonymousCredentials(), transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, # and we are mocking out the actual API, so just send an empty request. request = request_type() # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.get_delivery_pipeline), "__call__" ) as call: # Designate an appropriate return value for the call. call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( cloud_deploy.DeliveryPipeline( name="name_value", uid="uid_value", description="description_value", etag="etag_value", ) ) response = await client.get_delivery_pipeline(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.GetDeliveryPipelineRequest() # Establish that the response is the type that we expect. assert isinstance(response, cloud_deploy.DeliveryPipeline) assert response.name == "name_value" assert response.uid == "uid_value" assert response.description == "description_value" assert response.etag == "etag_value" @pytest.mark.asyncio async def test_get_delivery_pipeline_async_from_dict(): await test_get_delivery_pipeline_async(request_type=dict) def test_get_delivery_pipeline_field_headers(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cloud_deploy.GetDeliveryPipelineRequest() request.name = "name/value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.get_delivery_pipeline), "__call__" ) as call: call.return_value = cloud_deploy.DeliveryPipeline() client.get_delivery_pipeline(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] assert args[0] == request # Establish that the field header was sent. _, _, kw = call.mock_calls[0] assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] @pytest.mark.asyncio async def test_get_delivery_pipeline_field_headers_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cloud_deploy.GetDeliveryPipelineRequest() request.name = "name/value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.get_delivery_pipeline), "__call__" ) as call: call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( cloud_deploy.DeliveryPipeline() ) await client.get_delivery_pipeline(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] assert args[0] == request # Establish that the field header was sent. _, _, kw = call.mock_calls[0] assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] def test_get_delivery_pipeline_flattened(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.get_delivery_pipeline), "__call__" ) as call: # Designate an appropriate return value for the call. call.return_value = cloud_deploy.DeliveryPipeline() # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. client.get_delivery_pipeline(name="name_value",) # Establish that the underlying call was made with the expected # request object values. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] arg = args[0].name mock_val = "name_value" assert arg == mock_val def test_get_delivery_pipeline_flattened_error(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Attempting to call a method with both a request object and flattened # fields is an error. with pytest.raises(ValueError): client.get_delivery_pipeline( cloud_deploy.GetDeliveryPipelineRequest(), name="name_value", ) @pytest.mark.asyncio async def test_get_delivery_pipeline_flattened_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.get_delivery_pipeline), "__call__" ) as call: # Designate an appropriate return value for the call. call.return_value = cloud_deploy.DeliveryPipeline() call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( cloud_deploy.DeliveryPipeline() ) # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. response = await client.get_delivery_pipeline(name="name_value",) # Establish that the underlying call was made with the expected # request object values. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] arg = args[0].name mock_val = "name_value" assert arg == mock_val @pytest.mark.asyncio async def test_get_delivery_pipeline_flattened_error_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Attempting to call a method with both a request object and flattened # fields is an error. with pytest.raises(ValueError): await client.get_delivery_pipeline( cloud_deploy.GetDeliveryPipelineRequest(), name="name_value", ) @pytest.mark.parametrize( "request_type", [cloud_deploy.CreateDeliveryPipelineRequest, dict,] ) def test_create_delivery_pipeline(request_type, transport: str = "grpc"): client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, # and we are mocking out the actual API, so just send an empty request. request = request_type() # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.create_delivery_pipeline), "__call__" ) as call: # Designate an appropriate return value for the call. call.return_value = operations_pb2.Operation(name="operations/spam") response = client.create_delivery_pipeline(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.CreateDeliveryPipelineRequest() # Establish that the response is the type that we expect. assert isinstance(response, future.Future) def test_create_delivery_pipeline_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.create_delivery_pipeline), "__call__" ) as call: client.create_delivery_pipeline() call.assert_called() _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.CreateDeliveryPipelineRequest() @pytest.mark.asyncio async def test_create_delivery_pipeline_async( transport: str = "grpc_asyncio", request_type=cloud_deploy.CreateDeliveryPipelineRequest, ): client = CloudDeployAsyncClient( credentials=ga_credentials.AnonymousCredentials(), transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, # and we are mocking out the actual API, so just send an empty request. request = request_type() # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.create_delivery_pipeline), "__call__" ) as call: # Designate an appropriate return value for the call. call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( operations_pb2.Operation(name="operations/spam") ) response = await client.create_delivery_pipeline(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.CreateDeliveryPipelineRequest() # Establish that the response is the type that we expect. assert isinstance(response, future.Future) @pytest.mark.asyncio async def test_create_delivery_pipeline_async_from_dict(): await test_create_delivery_pipeline_async(request_type=dict) def test_create_delivery_pipeline_field_headers(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cloud_deploy.CreateDeliveryPipelineRequest() request.parent = "parent/value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.create_delivery_pipeline), "__call__" ) as call: call.return_value = operations_pb2.Operation(name="operations/op") client.create_delivery_pipeline(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] assert args[0] == request # Establish that the field header was sent. _, _, kw = call.mock_calls[0] assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] @pytest.mark.asyncio async def test_create_delivery_pipeline_field_headers_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cloud_deploy.CreateDeliveryPipelineRequest() request.parent = "parent/value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.create_delivery_pipeline), "__call__" ) as call: call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( operations_pb2.Operation(name="operations/op") ) await client.create_delivery_pipeline(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] assert args[0] == request # Establish that the field header was sent. _, _, kw = call.mock_calls[0] assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] def test_create_delivery_pipeline_flattened(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.create_delivery_pipeline), "__call__" ) as call: # Designate an appropriate return value for the call. call.return_value = operations_pb2.Operation(name="operations/op") # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. client.create_delivery_pipeline( parent="parent_value", delivery_pipeline=cloud_deploy.DeliveryPipeline(name="name_value"), delivery_pipeline_id="delivery_pipeline_id_value", ) # Establish that the underlying call was made with the expected # request object values. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] arg = args[0].parent mock_val = "parent_value" assert arg == mock_val arg = args[0].delivery_pipeline mock_val = cloud_deploy.DeliveryPipeline(name="name_value") assert arg == mock_val arg = args[0].delivery_pipeline_id mock_val = "delivery_pipeline_id_value" assert arg == mock_val def test_create_delivery_pipeline_flattened_error(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Attempting to call a method with both a request object and flattened # fields is an error. with pytest.raises(ValueError): client.create_delivery_pipeline( cloud_deploy.CreateDeliveryPipelineRequest(), parent="parent_value", delivery_pipeline=cloud_deploy.DeliveryPipeline(name="name_value"), delivery_pipeline_id="delivery_pipeline_id_value", ) @pytest.mark.asyncio async def test_create_delivery_pipeline_flattened_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.create_delivery_pipeline), "__call__" ) as call: # Designate an appropriate return value for the call. call.return_value = operations_pb2.Operation(name="operations/op") call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( operations_pb2.Operation(name="operations/spam") ) # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. response = await client.create_delivery_pipeline( parent="parent_value", delivery_pipeline=cloud_deploy.DeliveryPipeline(name="name_value"), delivery_pipeline_id="delivery_pipeline_id_value", ) # Establish that the underlying call was made with the expected # request object values. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] arg = args[0].parent mock_val = "parent_value" assert arg == mock_val arg = args[0].delivery_pipeline mock_val = cloud_deploy.DeliveryPipeline(name="name_value") assert arg == mock_val arg = args[0].delivery_pipeline_id mock_val = "delivery_pipeline_id_value" assert arg == mock_val @pytest.mark.asyncio async def test_create_delivery_pipeline_flattened_error_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Attempting to call a method with both a request object and flattened # fields is an error. with pytest.raises(ValueError): await client.create_delivery_pipeline( cloud_deploy.CreateDeliveryPipelineRequest(), parent="parent_value", delivery_pipeline=cloud_deploy.DeliveryPipeline(name="name_value"), delivery_pipeline_id="delivery_pipeline_id_value", ) @pytest.mark.parametrize( "request_type", [cloud_deploy.UpdateDeliveryPipelineRequest, dict,] ) def test_update_delivery_pipeline(request_type, transport: str = "grpc"): client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, # and we are mocking out the actual API, so just send an empty request. request = request_type() # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.update_delivery_pipeline), "__call__" ) as call: # Designate an appropriate return value for the call. call.return_value = operations_pb2.Operation(name="operations/spam") response = client.update_delivery_pipeline(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.UpdateDeliveryPipelineRequest() # Establish that the response is the type that we expect. assert isinstance(response, future.Future) def test_update_delivery_pipeline_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.update_delivery_pipeline), "__call__" ) as call: client.update_delivery_pipeline() call.assert_called() _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.UpdateDeliveryPipelineRequest() @pytest.mark.asyncio async def test_update_delivery_pipeline_async( transport: str = "grpc_asyncio", request_type=cloud_deploy.UpdateDeliveryPipelineRequest, ): client = CloudDeployAsyncClient( credentials=ga_credentials.AnonymousCredentials(), transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, # and we are mocking out the actual API, so just send an empty request. request = request_type() # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.update_delivery_pipeline), "__call__" ) as call: # Designate an appropriate return value for the call. call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( operations_pb2.Operation(name="operations/spam") ) response = await client.update_delivery_pipeline(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.UpdateDeliveryPipelineRequest() # Establish that the response is the type that we expect. assert isinstance(response, future.Future) @pytest.mark.asyncio async def test_update_delivery_pipeline_async_from_dict(): await test_update_delivery_pipeline_async(request_type=dict) def test_update_delivery_pipeline_field_headers(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cloud_deploy.UpdateDeliveryPipelineRequest() request.delivery_pipeline.name = "delivery_pipeline.name/value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.update_delivery_pipeline), "__call__" ) as call: call.return_value = operations_pb2.Operation(name="operations/op") client.update_delivery_pipeline(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] assert args[0] == request # Establish that the field header was sent. _, _, kw = call.mock_calls[0] assert ( "x-goog-request-params", "delivery_pipeline.name=delivery_pipeline.name/value", ) in kw["metadata"] @pytest.mark.asyncio async def test_update_delivery_pipeline_field_headers_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cloud_deploy.UpdateDeliveryPipelineRequest() request.delivery_pipeline.name = "delivery_pipeline.name/value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.update_delivery_pipeline), "__call__" ) as call: call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( operations_pb2.Operation(name="operations/op") ) await client.update_delivery_pipeline(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] assert args[0] == request # Establish that the field header was sent. _, _, kw = call.mock_calls[0] assert ( "x-goog-request-params", "delivery_pipeline.name=delivery_pipeline.name/value", ) in kw["metadata"] def test_update_delivery_pipeline_flattened(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.update_delivery_pipeline), "__call__" ) as call: # Designate an appropriate return value for the call. call.return_value = operations_pb2.Operation(name="operations/op") # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. client.update_delivery_pipeline( delivery_pipeline=cloud_deploy.DeliveryPipeline(name="name_value"), update_mask=field_mask_pb2.FieldMask(paths=["paths_value"]), ) # Establish that the underlying call was made with the expected # request object values. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] arg = args[0].delivery_pipeline mock_val = cloud_deploy.DeliveryPipeline(name="name_value") assert arg == mock_val arg = args[0].update_mask mock_val = field_mask_pb2.FieldMask(paths=["paths_value"]) assert arg == mock_val def test_update_delivery_pipeline_flattened_error(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Attempting to call a method with both a request object and flattened # fields is an error. with pytest.raises(ValueError): client.update_delivery_pipeline( cloud_deploy.UpdateDeliveryPipelineRequest(), delivery_pipeline=cloud_deploy.DeliveryPipeline(name="name_value"), update_mask=field_mask_pb2.FieldMask(paths=["paths_value"]), ) @pytest.mark.asyncio async def test_update_delivery_pipeline_flattened_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.update_delivery_pipeline), "__call__" ) as call: # Designate an appropriate return value for the call. call.return_value = operations_pb2.Operation(name="operations/op") call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( operations_pb2.Operation(name="operations/spam") ) # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. response = await client.update_delivery_pipeline( delivery_pipeline=cloud_deploy.DeliveryPipeline(name="name_value"), update_mask=field_mask_pb2.FieldMask(paths=["paths_value"]), ) # Establish that the underlying call was made with the expected # request object values. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] arg = args[0].delivery_pipeline mock_val = cloud_deploy.DeliveryPipeline(name="name_value") assert arg == mock_val arg = args[0].update_mask mock_val = field_mask_pb2.FieldMask(paths=["paths_value"]) assert arg == mock_val @pytest.mark.asyncio async def test_update_delivery_pipeline_flattened_error_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Attempting to call a method with both a request object and flattened # fields is an error. with pytest.raises(ValueError): await client.update_delivery_pipeline( cloud_deploy.UpdateDeliveryPipelineRequest(), delivery_pipeline=cloud_deploy.DeliveryPipeline(name="name_value"), update_mask=field_mask_pb2.FieldMask(paths=["paths_value"]), ) @pytest.mark.parametrize( "request_type", [cloud_deploy.DeleteDeliveryPipelineRequest, dict,] ) def test_delete_delivery_pipeline(request_type, transport: str = "grpc"): client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, # and we are mocking out the actual API, so just send an empty request. request = request_type() # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.delete_delivery_pipeline), "__call__" ) as call: # Designate an appropriate return value for the call. call.return_value = operations_pb2.Operation(name="operations/spam") response = client.delete_delivery_pipeline(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.DeleteDeliveryPipelineRequest() # Establish that the response is the type that we expect. assert isinstance(response, future.Future) def test_delete_delivery_pipeline_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.delete_delivery_pipeline), "__call__" ) as call: client.delete_delivery_pipeline() call.assert_called() _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.DeleteDeliveryPipelineRequest() @pytest.mark.asyncio async def test_delete_delivery_pipeline_async( transport: str = "grpc_asyncio", request_type=cloud_deploy.DeleteDeliveryPipelineRequest, ): client = CloudDeployAsyncClient( credentials=ga_credentials.AnonymousCredentials(), transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, # and we are mocking out the actual API, so just send an empty request. request = request_type() # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.delete_delivery_pipeline), "__call__" ) as call: # Designate an appropriate return value for the call. call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( operations_pb2.Operation(name="operations/spam") ) response = await client.delete_delivery_pipeline(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.DeleteDeliveryPipelineRequest() # Establish that the response is the type that we expect. assert isinstance(response, future.Future) @pytest.mark.asyncio async def test_delete_delivery_pipeline_async_from_dict(): await test_delete_delivery_pipeline_async(request_type=dict) def test_delete_delivery_pipeline_field_headers(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cloud_deploy.DeleteDeliveryPipelineRequest() request.name = "name/value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.delete_delivery_pipeline), "__call__" ) as call: call.return_value = operations_pb2.Operation(name="operations/op") client.delete_delivery_pipeline(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] assert args[0] == request # Establish that the field header was sent. _, _, kw = call.mock_calls[0] assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] @pytest.mark.asyncio async def test_delete_delivery_pipeline_field_headers_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cloud_deploy.DeleteDeliveryPipelineRequest() request.name = "name/value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.delete_delivery_pipeline), "__call__" ) as call: call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( operations_pb2.Operation(name="operations/op") ) await client.delete_delivery_pipeline(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] assert args[0] == request # Establish that the field header was sent. _, _, kw = call.mock_calls[0] assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] def test_delete_delivery_pipeline_flattened(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.delete_delivery_pipeline), "__call__" ) as call: # Designate an appropriate return value for the call. call.return_value = operations_pb2.Operation(name="operations/op") # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. client.delete_delivery_pipeline(name="name_value",) # Establish that the underlying call was made with the expected # request object values. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] arg = args[0].name mock_val = "name_value" assert arg == mock_val def test_delete_delivery_pipeline_flattened_error(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Attempting to call a method with both a request object and flattened # fields is an error. with pytest.raises(ValueError): client.delete_delivery_pipeline( cloud_deploy.DeleteDeliveryPipelineRequest(), name="name_value", ) @pytest.mark.asyncio async def test_delete_delivery_pipeline_flattened_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.delete_delivery_pipeline), "__call__" ) as call: # Designate an appropriate return value for the call. call.return_value = operations_pb2.Operation(name="operations/op") call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( operations_pb2.Operation(name="operations/spam") ) # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. response = await client.delete_delivery_pipeline(name="name_value",) # Establish that the underlying call was made with the expected # request object values. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] arg = args[0].name mock_val = "name_value" assert arg == mock_val @pytest.mark.asyncio async def test_delete_delivery_pipeline_flattened_error_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Attempting to call a method with both a request object and flattened # fields is an error. with pytest.raises(ValueError): await client.delete_delivery_pipeline( cloud_deploy.DeleteDeliveryPipelineRequest(), name="name_value", ) @pytest.mark.parametrize("request_type", [cloud_deploy.ListTargetsRequest, dict,]) def test_list_targets(request_type, transport: str = "grpc"): client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, # and we are mocking out the actual API, so just send an empty request. request = request_type() # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_targets), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = cloud_deploy.ListTargetsResponse( next_page_token="next_page_token_value", unreachable=["unreachable_value"], ) response = client.list_targets(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.ListTargetsRequest() # Establish that the response is the type that we expect. assert isinstance(response, pagers.ListTargetsPager) assert response.next_page_token == "next_page_token_value" assert response.unreachable == ["unreachable_value"] def test_list_targets_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_targets), "__call__") as call: client.list_targets() call.assert_called() _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.ListTargetsRequest() @pytest.mark.asyncio async def test_list_targets_async( transport: str = "grpc_asyncio", request_type=cloud_deploy.ListTargetsRequest ): client = CloudDeployAsyncClient( credentials=ga_credentials.AnonymousCredentials(), transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, # and we are mocking out the actual API, so just send an empty request. request = request_type() # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_targets), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( cloud_deploy.ListTargetsResponse( next_page_token="next_page_token_value", unreachable=["unreachable_value"], ) ) response = await client.list_targets(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.ListTargetsRequest() # Establish that the response is the type that we expect. assert isinstance(response, pagers.ListTargetsAsyncPager) assert response.next_page_token == "next_page_token_value" assert response.unreachable == ["unreachable_value"] @pytest.mark.asyncio async def test_list_targets_async_from_dict(): await test_list_targets_async(request_type=dict) def test_list_targets_field_headers(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cloud_deploy.ListTargetsRequest() request.parent = "parent/value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_targets), "__call__") as call: call.return_value = cloud_deploy.ListTargetsResponse() client.list_targets(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] assert args[0] == request # Establish that the field header was sent. _, _, kw = call.mock_calls[0] assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] @pytest.mark.asyncio async def test_list_targets_field_headers_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cloud_deploy.ListTargetsRequest() request.parent = "parent/value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_targets), "__call__") as call: call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( cloud_deploy.ListTargetsResponse() ) await client.list_targets(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] assert args[0] == request # Establish that the field header was sent. _, _, kw = call.mock_calls[0] assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] def test_list_targets_flattened(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_targets), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = cloud_deploy.ListTargetsResponse() # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. client.list_targets(parent="parent_value",) # Establish that the underlying call was made with the expected # request object values. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] arg = args[0].parent mock_val = "parent_value" assert arg == mock_val def test_list_targets_flattened_error(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Attempting to call a method with both a request object and flattened # fields is an error. with pytest.raises(ValueError): client.list_targets( cloud_deploy.ListTargetsRequest(), parent="parent_value", ) @pytest.mark.asyncio async def test_list_targets_flattened_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_targets), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = cloud_deploy.ListTargetsResponse() call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( cloud_deploy.ListTargetsResponse() ) # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. response = await client.list_targets(parent="parent_value",) # Establish that the underlying call was made with the expected # request object values. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] arg = args[0].parent mock_val = "parent_value" assert arg == mock_val @pytest.mark.asyncio async def test_list_targets_flattened_error_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Attempting to call a method with both a request object and flattened # fields is an error. with pytest.raises(ValueError): await client.list_targets( cloud_deploy.ListTargetsRequest(), parent="parent_value", ) def test_list_targets_pager(transport_name: str = "grpc"): client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials, transport=transport_name, ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_targets), "__call__") as call: # Set the response to a series of pages. call.side_effect = ( cloud_deploy.ListTargetsResponse( targets=[ cloud_deploy.Target(), cloud_deploy.Target(), cloud_deploy.Target(), ], next_page_token="abc", ), cloud_deploy.ListTargetsResponse(targets=[], next_page_token="def",), cloud_deploy.ListTargetsResponse( targets=[cloud_deploy.Target(),], next_page_token="ghi", ), cloud_deploy.ListTargetsResponse( targets=[cloud_deploy.Target(), cloud_deploy.Target(),], ), RuntimeError, ) metadata = () metadata = tuple(metadata) + ( gapic_v1.routing_header.to_grpc_metadata((("parent", ""),)), ) pager = client.list_targets(request={}) assert pager._metadata == metadata results = [i for i in pager] assert len(results) == 6 assert all(isinstance(i, cloud_deploy.Target) for i in results) def test_list_targets_pages(transport_name: str = "grpc"): client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials, transport=transport_name, ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_targets), "__call__") as call: # Set the response to a series of pages. call.side_effect = ( cloud_deploy.ListTargetsResponse( targets=[ cloud_deploy.Target(), cloud_deploy.Target(), cloud_deploy.Target(), ], next_page_token="abc", ), cloud_deploy.ListTargetsResponse(targets=[], next_page_token="def",), cloud_deploy.ListTargetsResponse( targets=[cloud_deploy.Target(),], next_page_token="ghi", ), cloud_deploy.ListTargetsResponse( targets=[cloud_deploy.Target(), cloud_deploy.Target(),], ), RuntimeError, ) pages = list(client.list_targets(request={}).pages) for page_, token in zip(pages, ["abc", "def", "ghi", ""]): assert page_.raw_page.next_page_token == token @pytest.mark.asyncio async def test_list_targets_async_pager(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials,) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.list_targets), "__call__", new_callable=mock.AsyncMock ) as call: # Set the response to a series of pages. call.side_effect = ( cloud_deploy.ListTargetsResponse( targets=[ cloud_deploy.Target(), cloud_deploy.Target(), cloud_deploy.Target(), ], next_page_token="abc", ), cloud_deploy.ListTargetsResponse(targets=[], next_page_token="def",), cloud_deploy.ListTargetsResponse( targets=[cloud_deploy.Target(),], next_page_token="ghi", ), cloud_deploy.ListTargetsResponse( targets=[cloud_deploy.Target(), cloud_deploy.Target(),], ), RuntimeError, ) async_pager = await client.list_targets(request={},) assert async_pager.next_page_token == "abc" responses = [] async for response in async_pager: responses.append(response) assert len(responses) == 6 assert all(isinstance(i, cloud_deploy.Target) for i in responses) @pytest.mark.asyncio async def test_list_targets_async_pages(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials,) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.list_targets), "__call__", new_callable=mock.AsyncMock ) as call: # Set the response to a series of pages. call.side_effect = ( cloud_deploy.ListTargetsResponse( targets=[ cloud_deploy.Target(), cloud_deploy.Target(), cloud_deploy.Target(), ], next_page_token="abc", ), cloud_deploy.ListTargetsResponse(targets=[], next_page_token="def",), cloud_deploy.ListTargetsResponse( targets=[cloud_deploy.Target(),], next_page_token="ghi", ), cloud_deploy.ListTargetsResponse( targets=[cloud_deploy.Target(), cloud_deploy.Target(),], ), RuntimeError, ) pages = [] async for page_ in (await client.list_targets(request={})).pages: pages.append(page_) for page_, token in zip(pages, ["abc", "def", "ghi", ""]): assert page_.raw_page.next_page_token == token @pytest.mark.parametrize("request_type", [cloud_deploy.GetTargetRequest, dict,]) def test_get_target(request_type, transport: str = "grpc"): client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, # and we are mocking out the actual API, so just send an empty request. request = request_type() # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_target), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = cloud_deploy.Target( name="name_value", target_id="target_id_value", uid="uid_value", description="description_value", require_approval=True, etag="etag_value", gke=cloud_deploy.GkeCluster(cluster="cluster_value"), ) response = client.get_target(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.GetTargetRequest() # Establish that the response is the type that we expect. assert isinstance(response, cloud_deploy.Target) assert response.name == "name_value" assert response.target_id == "target_id_value" assert response.uid == "uid_value" assert response.description == "description_value" assert response.require_approval is True assert response.etag == "etag_value" def test_get_target_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_target), "__call__") as call: client.get_target() call.assert_called() _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.GetTargetRequest() @pytest.mark.asyncio async def test_get_target_async( transport: str = "grpc_asyncio", request_type=cloud_deploy.GetTargetRequest ): client = CloudDeployAsyncClient( credentials=ga_credentials.AnonymousCredentials(), transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, # and we are mocking out the actual API, so just send an empty request. request = request_type() # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_target), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( cloud_deploy.Target( name="name_value", target_id="target_id_value", uid="uid_value", description="description_value", require_approval=True, etag="etag_value", ) ) response = await client.get_target(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.GetTargetRequest() # Establish that the response is the type that we expect. assert isinstance(response, cloud_deploy.Target) assert response.name == "name_value" assert response.target_id == "target_id_value" assert response.uid == "uid_value" assert response.description == "description_value" assert response.require_approval is True assert response.etag == "etag_value" @pytest.mark.asyncio async def test_get_target_async_from_dict(): await test_get_target_async(request_type=dict) def test_get_target_field_headers(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cloud_deploy.GetTargetRequest() request.name = "name/value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_target), "__call__") as call: call.return_value = cloud_deploy.Target() client.get_target(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] assert args[0] == request # Establish that the field header was sent. _, _, kw = call.mock_calls[0] assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] @pytest.mark.asyncio async def test_get_target_field_headers_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cloud_deploy.GetTargetRequest() request.name = "name/value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_target), "__call__") as call: call.return_value = grpc_helpers_async.FakeUnaryUnaryCall(cloud_deploy.Target()) await client.get_target(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] assert args[0] == request # Establish that the field header was sent. _, _, kw = call.mock_calls[0] assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] def test_get_target_flattened(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_target), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = cloud_deploy.Target() # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. client.get_target(name="name_value",) # Establish that the underlying call was made with the expected # request object values. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] arg = args[0].name mock_val = "name_value" assert arg == mock_val def test_get_target_flattened_error(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Attempting to call a method with both a request object and flattened # fields is an error. with pytest.raises(ValueError): client.get_target( cloud_deploy.GetTargetRequest(), name="name_value", ) @pytest.mark.asyncio async def test_get_target_flattened_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_target), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = cloud_deploy.Target() call.return_value = grpc_helpers_async.FakeUnaryUnaryCall(cloud_deploy.Target()) # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. response = await client.get_target(name="name_value",) # Establish that the underlying call was made with the expected # request object values. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] arg = args[0].name mock_val = "name_value" assert arg == mock_val @pytest.mark.asyncio async def test_get_target_flattened_error_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Attempting to call a method with both a request object and flattened # fields is an error. with pytest.raises(ValueError): await client.get_target( cloud_deploy.GetTargetRequest(), name="name_value", ) @pytest.mark.parametrize("request_type", [cloud_deploy.CreateTargetRequest, dict,]) def test_create_target(request_type, transport: str = "grpc"): client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, # and we are mocking out the actual API, so just send an empty request. request = request_type() # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.create_target), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = operations_pb2.Operation(name="operations/spam") response = client.create_target(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.CreateTargetRequest() # Establish that the response is the type that we expect. assert isinstance(response, future.Future) def test_create_target_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.create_target), "__call__") as call: client.create_target() call.assert_called() _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.CreateTargetRequest() @pytest.mark.asyncio async def test_create_target_async( transport: str = "grpc_asyncio", request_type=cloud_deploy.CreateTargetRequest ): client = CloudDeployAsyncClient( credentials=ga_credentials.AnonymousCredentials(), transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, # and we are mocking out the actual API, so just send an empty request. request = request_type() # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.create_target), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( operations_pb2.Operation(name="operations/spam") ) response = await client.create_target(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.CreateTargetRequest() # Establish that the response is the type that we expect. assert isinstance(response, future.Future) @pytest.mark.asyncio async def test_create_target_async_from_dict(): await test_create_target_async(request_type=dict) def test_create_target_field_headers(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cloud_deploy.CreateTargetRequest() request.parent = "parent/value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.create_target), "__call__") as call: call.return_value = operations_pb2.Operation(name="operations/op") client.create_target(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] assert args[0] == request # Establish that the field header was sent. _, _, kw = call.mock_calls[0] assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] @pytest.mark.asyncio async def test_create_target_field_headers_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cloud_deploy.CreateTargetRequest() request.parent = "parent/value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.create_target), "__call__") as call: call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( operations_pb2.Operation(name="operations/op") ) await client.create_target(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] assert args[0] == request # Establish that the field header was sent. _, _, kw = call.mock_calls[0] assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] def test_create_target_flattened(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.create_target), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = operations_pb2.Operation(name="operations/op") # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. client.create_target( parent="parent_value", target=cloud_deploy.Target(name="name_value"), target_id="target_id_value", ) # Establish that the underlying call was made with the expected # request object values. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] arg = args[0].parent mock_val = "parent_value" assert arg == mock_val arg = args[0].target mock_val = cloud_deploy.Target(name="name_value") assert arg == mock_val arg = args[0].target_id mock_val = "target_id_value" assert arg == mock_val def test_create_target_flattened_error(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Attempting to call a method with both a request object and flattened # fields is an error. with pytest.raises(ValueError): client.create_target( cloud_deploy.CreateTargetRequest(), parent="parent_value", target=cloud_deploy.Target(name="name_value"), target_id="target_id_value", ) @pytest.mark.asyncio async def test_create_target_flattened_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.create_target), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = operations_pb2.Operation(name="operations/op") call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( operations_pb2.Operation(name="operations/spam") ) # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. response = await client.create_target( parent="parent_value", target=cloud_deploy.Target(name="name_value"), target_id="target_id_value", ) # Establish that the underlying call was made with the expected # request object values. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] arg = args[0].parent mock_val = "parent_value" assert arg == mock_val arg = args[0].target mock_val = cloud_deploy.Target(name="name_value") assert arg == mock_val arg = args[0].target_id mock_val = "target_id_value" assert arg == mock_val @pytest.mark.asyncio async def test_create_target_flattened_error_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Attempting to call a method with both a request object and flattened # fields is an error. with pytest.raises(ValueError): await client.create_target( cloud_deploy.CreateTargetRequest(), parent="parent_value", target=cloud_deploy.Target(name="name_value"), target_id="target_id_value", ) @pytest.mark.parametrize("request_type", [cloud_deploy.UpdateTargetRequest, dict,]) def test_update_target(request_type, transport: str = "grpc"): client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, # and we are mocking out the actual API, so just send an empty request. request = request_type() # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.update_target), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = operations_pb2.Operation(name="operations/spam") response = client.update_target(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.UpdateTargetRequest() # Establish that the response is the type that we expect. assert isinstance(response, future.Future) def test_update_target_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.update_target), "__call__") as call: client.update_target() call.assert_called() _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.UpdateTargetRequest() @pytest.mark.asyncio async def test_update_target_async( transport: str = "grpc_asyncio", request_type=cloud_deploy.UpdateTargetRequest ): client = CloudDeployAsyncClient( credentials=ga_credentials.AnonymousCredentials(), transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, # and we are mocking out the actual API, so just send an empty request. request = request_type() # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.update_target), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( operations_pb2.Operation(name="operations/spam") ) response = await client.update_target(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.UpdateTargetRequest() # Establish that the response is the type that we expect. assert isinstance(response, future.Future) @pytest.mark.asyncio async def test_update_target_async_from_dict(): await test_update_target_async(request_type=dict) def test_update_target_field_headers(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cloud_deploy.UpdateTargetRequest() request.target.name = "target.name/value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.update_target), "__call__") as call: call.return_value = operations_pb2.Operation(name="operations/op") client.update_target(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] assert args[0] == request # Establish that the field header was sent. _, _, kw = call.mock_calls[0] assert ("x-goog-request-params", "target.name=target.name/value",) in kw["metadata"] @pytest.mark.asyncio async def test_update_target_field_headers_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cloud_deploy.UpdateTargetRequest() request.target.name = "target.name/value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.update_target), "__call__") as call: call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( operations_pb2.Operation(name="operations/op") ) await client.update_target(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] assert args[0] == request # Establish that the field header was sent. _, _, kw = call.mock_calls[0] assert ("x-goog-request-params", "target.name=target.name/value",) in kw["metadata"] def test_update_target_flattened(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.update_target), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = operations_pb2.Operation(name="operations/op") # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. client.update_target( target=cloud_deploy.Target(name="name_value"), update_mask=field_mask_pb2.FieldMask(paths=["paths_value"]), ) # Establish that the underlying call was made with the expected # request object values. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] arg = args[0].target mock_val = cloud_deploy.Target(name="name_value") assert arg == mock_val arg = args[0].update_mask mock_val = field_mask_pb2.FieldMask(paths=["paths_value"]) assert arg == mock_val def test_update_target_flattened_error(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Attempting to call a method with both a request object and flattened # fields is an error. with pytest.raises(ValueError): client.update_target( cloud_deploy.UpdateTargetRequest(), target=cloud_deploy.Target(name="name_value"), update_mask=field_mask_pb2.FieldMask(paths=["paths_value"]), ) @pytest.mark.asyncio async def test_update_target_flattened_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.update_target), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = operations_pb2.Operation(name="operations/op") call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( operations_pb2.Operation(name="operations/spam") ) # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. response = await client.update_target( target=cloud_deploy.Target(name="name_value"), update_mask=field_mask_pb2.FieldMask(paths=["paths_value"]), ) # Establish that the underlying call was made with the expected # request object values. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] arg = args[0].target mock_val = cloud_deploy.Target(name="name_value") assert arg == mock_val arg = args[0].update_mask mock_val = field_mask_pb2.FieldMask(paths=["paths_value"]) assert arg == mock_val @pytest.mark.asyncio async def test_update_target_flattened_error_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Attempting to call a method with both a request object and flattened # fields is an error. with pytest.raises(ValueError): await client.update_target( cloud_deploy.UpdateTargetRequest(), target=cloud_deploy.Target(name="name_value"), update_mask=field_mask_pb2.FieldMask(paths=["paths_value"]), ) @pytest.mark.parametrize("request_type", [cloud_deploy.DeleteTargetRequest, dict,]) def test_delete_target(request_type, transport: str = "grpc"): client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, # and we are mocking out the actual API, so just send an empty request. request = request_type() # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.delete_target), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = operations_pb2.Operation(name="operations/spam") response = client.delete_target(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.DeleteTargetRequest() # Establish that the response is the type that we expect. assert isinstance(response, future.Future) def test_delete_target_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.delete_target), "__call__") as call: client.delete_target() call.assert_called() _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.DeleteTargetRequest() @pytest.mark.asyncio async def test_delete_target_async( transport: str = "grpc_asyncio", request_type=cloud_deploy.DeleteTargetRequest ): client = CloudDeployAsyncClient( credentials=ga_credentials.AnonymousCredentials(), transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, # and we are mocking out the actual API, so just send an empty request. request = request_type() # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.delete_target), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( operations_pb2.Operation(name="operations/spam") ) response = await client.delete_target(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.DeleteTargetRequest() # Establish that the response is the type that we expect. assert isinstance(response, future.Future) @pytest.mark.asyncio async def test_delete_target_async_from_dict(): await test_delete_target_async(request_type=dict) def test_delete_target_field_headers(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cloud_deploy.DeleteTargetRequest() request.name = "name/value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.delete_target), "__call__") as call: call.return_value = operations_pb2.Operation(name="operations/op") client.delete_target(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] assert args[0] == request # Establish that the field header was sent. _, _, kw = call.mock_calls[0] assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] @pytest.mark.asyncio async def test_delete_target_field_headers_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cloud_deploy.DeleteTargetRequest() request.name = "name/value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.delete_target), "__call__") as call: call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( operations_pb2.Operation(name="operations/op") ) await client.delete_target(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] assert args[0] == request # Establish that the field header was sent. _, _, kw = call.mock_calls[0] assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] def test_delete_target_flattened(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.delete_target), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = operations_pb2.Operation(name="operations/op") # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. client.delete_target(name="name_value",) # Establish that the underlying call was made with the expected # request object values. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] arg = args[0].name mock_val = "name_value" assert arg == mock_val def test_delete_target_flattened_error(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Attempting to call a method with both a request object and flattened # fields is an error. with pytest.raises(ValueError): client.delete_target( cloud_deploy.DeleteTargetRequest(), name="name_value", ) @pytest.mark.asyncio async def test_delete_target_flattened_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.delete_target), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = operations_pb2.Operation(name="operations/op") call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( operations_pb2.Operation(name="operations/spam") ) # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. response = await client.delete_target(name="name_value",) # Establish that the underlying call was made with the expected # request object values. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] arg = args[0].name mock_val = "name_value" assert arg == mock_val @pytest.mark.asyncio async def test_delete_target_flattened_error_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Attempting to call a method with both a request object and flattened # fields is an error. with pytest.raises(ValueError): await client.delete_target( cloud_deploy.DeleteTargetRequest(), name="name_value", ) @pytest.mark.parametrize("request_type", [cloud_deploy.ListReleasesRequest, dict,]) def test_list_releases(request_type, transport: str = "grpc"): client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, # and we are mocking out the actual API, so just send an empty request. request = request_type() # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_releases), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = cloud_deploy.ListReleasesResponse( next_page_token="next_page_token_value", unreachable=["unreachable_value"], ) response = client.list_releases(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.ListReleasesRequest() # Establish that the response is the type that we expect. assert isinstance(response, pagers.ListReleasesPager) assert response.next_page_token == "next_page_token_value" assert response.unreachable == ["unreachable_value"] def test_list_releases_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_releases), "__call__") as call: client.list_releases() call.assert_called() _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.ListReleasesRequest() @pytest.mark.asyncio async def test_list_releases_async( transport: str = "grpc_asyncio", request_type=cloud_deploy.ListReleasesRequest ): client = CloudDeployAsyncClient( credentials=ga_credentials.AnonymousCredentials(), transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, # and we are mocking out the actual API, so just send an empty request. request = request_type() # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_releases), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( cloud_deploy.ListReleasesResponse( next_page_token="next_page_token_value", unreachable=["unreachable_value"], ) ) response = await client.list_releases(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.ListReleasesRequest() # Establish that the response is the type that we expect. assert isinstance(response, pagers.ListReleasesAsyncPager) assert response.next_page_token == "next_page_token_value" assert response.unreachable == ["unreachable_value"] @pytest.mark.asyncio async def test_list_releases_async_from_dict(): await test_list_releases_async(request_type=dict) def test_list_releases_field_headers(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cloud_deploy.ListReleasesRequest() request.parent = "parent/value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_releases), "__call__") as call: call.return_value = cloud_deploy.ListReleasesResponse() client.list_releases(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] assert args[0] == request # Establish that the field header was sent. _, _, kw = call.mock_calls[0] assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] @pytest.mark.asyncio async def test_list_releases_field_headers_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cloud_deploy.ListReleasesRequest() request.parent = "parent/value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_releases), "__call__") as call: call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( cloud_deploy.ListReleasesResponse() ) await client.list_releases(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] assert args[0] == request # Establish that the field header was sent. _, _, kw = call.mock_calls[0] assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] def test_list_releases_flattened(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_releases), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = cloud_deploy.ListReleasesResponse() # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. client.list_releases(parent="parent_value",) # Establish that the underlying call was made with the expected # request object values. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] arg = args[0].parent mock_val = "parent_value" assert arg == mock_val def test_list_releases_flattened_error(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Attempting to call a method with both a request object and flattened # fields is an error. with pytest.raises(ValueError): client.list_releases( cloud_deploy.ListReleasesRequest(), parent="parent_value", ) @pytest.mark.asyncio async def test_list_releases_flattened_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_releases), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = cloud_deploy.ListReleasesResponse() call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( cloud_deploy.ListReleasesResponse() ) # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. response = await client.list_releases(parent="parent_value",) # Establish that the underlying call was made with the expected # request object values. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] arg = args[0].parent mock_val = "parent_value" assert arg == mock_val @pytest.mark.asyncio async def test_list_releases_flattened_error_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Attempting to call a method with both a request object and flattened # fields is an error. with pytest.raises(ValueError): await client.list_releases( cloud_deploy.ListReleasesRequest(), parent="parent_value", ) def test_list_releases_pager(transport_name: str = "grpc"): client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials, transport=transport_name, ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_releases), "__call__") as call: # Set the response to a series of pages. call.side_effect = ( cloud_deploy.ListReleasesResponse( releases=[ cloud_deploy.Release(), cloud_deploy.Release(), cloud_deploy.Release(), ], next_page_token="abc", ), cloud_deploy.ListReleasesResponse(releases=[], next_page_token="def",), cloud_deploy.ListReleasesResponse( releases=[cloud_deploy.Release(),], next_page_token="ghi", ), cloud_deploy.ListReleasesResponse( releases=[cloud_deploy.Release(), cloud_deploy.Release(),], ), RuntimeError, ) metadata = () metadata = tuple(metadata) + ( gapic_v1.routing_header.to_grpc_metadata((("parent", ""),)), ) pager = client.list_releases(request={}) assert pager._metadata == metadata results = [i for i in pager] assert len(results) == 6 assert all(isinstance(i, cloud_deploy.Release) for i in results) def test_list_releases_pages(transport_name: str = "grpc"): client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials, transport=transport_name, ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_releases), "__call__") as call: # Set the response to a series of pages. call.side_effect = ( cloud_deploy.ListReleasesResponse( releases=[ cloud_deploy.Release(), cloud_deploy.Release(), cloud_deploy.Release(), ], next_page_token="abc", ), cloud_deploy.ListReleasesResponse(releases=[], next_page_token="def",), cloud_deploy.ListReleasesResponse( releases=[cloud_deploy.Release(),], next_page_token="ghi", ), cloud_deploy.ListReleasesResponse( releases=[cloud_deploy.Release(), cloud_deploy.Release(),], ), RuntimeError, ) pages = list(client.list_releases(request={}).pages) for page_, token in zip(pages, ["abc", "def", "ghi", ""]): assert page_.raw_page.next_page_token == token @pytest.mark.asyncio async def test_list_releases_async_pager(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials,) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.list_releases), "__call__", new_callable=mock.AsyncMock ) as call: # Set the response to a series of pages. call.side_effect = ( cloud_deploy.ListReleasesResponse( releases=[ cloud_deploy.Release(), cloud_deploy.Release(), cloud_deploy.Release(), ], next_page_token="abc", ), cloud_deploy.ListReleasesResponse(releases=[], next_page_token="def",), cloud_deploy.ListReleasesResponse( releases=[cloud_deploy.Release(),], next_page_token="ghi", ), cloud_deploy.ListReleasesResponse( releases=[cloud_deploy.Release(), cloud_deploy.Release(),], ), RuntimeError, ) async_pager = await client.list_releases(request={},) assert async_pager.next_page_token == "abc" responses = [] async for response in async_pager: responses.append(response) assert len(responses) == 6 assert all(isinstance(i, cloud_deploy.Release) for i in responses) @pytest.mark.asyncio async def test_list_releases_async_pages(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials,) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.list_releases), "__call__", new_callable=mock.AsyncMock ) as call: # Set the response to a series of pages. call.side_effect = ( cloud_deploy.ListReleasesResponse( releases=[ cloud_deploy.Release(), cloud_deploy.Release(), cloud_deploy.Release(), ], next_page_token="abc", ), cloud_deploy.ListReleasesResponse(releases=[], next_page_token="def",), cloud_deploy.ListReleasesResponse( releases=[cloud_deploy.Release(),], next_page_token="ghi", ), cloud_deploy.ListReleasesResponse( releases=[cloud_deploy.Release(), cloud_deploy.Release(),], ), RuntimeError, ) pages = [] async for page_ in (await client.list_releases(request={})).pages: pages.append(page_) for page_, token in zip(pages, ["abc", "def", "ghi", ""]): assert page_.raw_page.next_page_token == token @pytest.mark.parametrize("request_type", [cloud_deploy.GetReleaseRequest, dict,]) def test_get_release(request_type, transport: str = "grpc"): client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, # and we are mocking out the actual API, so just send an empty request. request = request_type() # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_release), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = cloud_deploy.Release( name="name_value", uid="uid_value", description="description_value", skaffold_config_uri="skaffold_config_uri_value", skaffold_config_path="skaffold_config_path_value", render_state=cloud_deploy.Release.RenderState.SUCCEEDED, etag="etag_value", skaffold_version="skaffold_version_value", ) response = client.get_release(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.GetReleaseRequest() # Establish that the response is the type that we expect. assert isinstance(response, cloud_deploy.Release) assert response.name == "name_value" assert response.uid == "uid_value" assert response.description == "description_value" assert response.skaffold_config_uri == "skaffold_config_uri_value" assert response.skaffold_config_path == "skaffold_config_path_value" assert response.render_state == cloud_deploy.Release.RenderState.SUCCEEDED assert response.etag == "etag_value" assert response.skaffold_version == "skaffold_version_value" def test_get_release_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_release), "__call__") as call: client.get_release() call.assert_called() _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.GetReleaseRequest() @pytest.mark.asyncio async def test_get_release_async( transport: str = "grpc_asyncio", request_type=cloud_deploy.GetReleaseRequest ): client = CloudDeployAsyncClient( credentials=ga_credentials.AnonymousCredentials(), transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, # and we are mocking out the actual API, so just send an empty request. request = request_type() # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_release), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( cloud_deploy.Release( name="name_value", uid="uid_value", description="description_value", skaffold_config_uri="skaffold_config_uri_value", skaffold_config_path="skaffold_config_path_value", render_state=cloud_deploy.Release.RenderState.SUCCEEDED, etag="etag_value", skaffold_version="skaffold_version_value", ) ) response = await client.get_release(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.GetReleaseRequest() # Establish that the response is the type that we expect. assert isinstance(response, cloud_deploy.Release) assert response.name == "name_value" assert response.uid == "uid_value" assert response.description == "description_value" assert response.skaffold_config_uri == "skaffold_config_uri_value" assert response.skaffold_config_path == "skaffold_config_path_value" assert response.render_state == cloud_deploy.Release.RenderState.SUCCEEDED assert response.etag == "etag_value" assert response.skaffold_version == "skaffold_version_value" @pytest.mark.asyncio async def test_get_release_async_from_dict(): await test_get_release_async(request_type=dict) def test_get_release_field_headers(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cloud_deploy.GetReleaseRequest() request.name = "name/value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_release), "__call__") as call: call.return_value = cloud_deploy.Release() client.get_release(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] assert args[0] == request # Establish that the field header was sent. _, _, kw = call.mock_calls[0] assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] @pytest.mark.asyncio async def test_get_release_field_headers_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cloud_deploy.GetReleaseRequest() request.name = "name/value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_release), "__call__") as call: call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( cloud_deploy.Release() ) await client.get_release(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] assert args[0] == request # Establish that the field header was sent. _, _, kw = call.mock_calls[0] assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] def test_get_release_flattened(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_release), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = cloud_deploy.Release() # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. client.get_release(name="name_value",) # Establish that the underlying call was made with the expected # request object values. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] arg = args[0].name mock_val = "name_value" assert arg == mock_val def test_get_release_flattened_error(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Attempting to call a method with both a request object and flattened # fields is an error. with pytest.raises(ValueError): client.get_release( cloud_deploy.GetReleaseRequest(), name="name_value", ) @pytest.mark.asyncio async def test_get_release_flattened_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_release), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = cloud_deploy.Release() call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( cloud_deploy.Release() ) # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. response = await client.get_release(name="name_value",) # Establish that the underlying call was made with the expected # request object values. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] arg = args[0].name mock_val = "name_value" assert arg == mock_val @pytest.mark.asyncio async def test_get_release_flattened_error_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Attempting to call a method with both a request object and flattened # fields is an error. with pytest.raises(ValueError): await client.get_release( cloud_deploy.GetReleaseRequest(), name="name_value", ) @pytest.mark.parametrize("request_type", [cloud_deploy.CreateReleaseRequest, dict,]) def test_create_release(request_type, transport: str = "grpc"): client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, # and we are mocking out the actual API, so just send an empty request. request = request_type() # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.create_release), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = operations_pb2.Operation(name="operations/spam") response = client.create_release(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.CreateReleaseRequest() # Establish that the response is the type that we expect. assert isinstance(response, future.Future) def test_create_release_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.create_release), "__call__") as call: client.create_release() call.assert_called() _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.CreateReleaseRequest() @pytest.mark.asyncio async def test_create_release_async( transport: str = "grpc_asyncio", request_type=cloud_deploy.CreateReleaseRequest ): client = CloudDeployAsyncClient( credentials=ga_credentials.AnonymousCredentials(), transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, # and we are mocking out the actual API, so just send an empty request. request = request_type() # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.create_release), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( operations_pb2.Operation(name="operations/spam") ) response = await client.create_release(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.CreateReleaseRequest() # Establish that the response is the type that we expect. assert isinstance(response, future.Future) @pytest.mark.asyncio async def test_create_release_async_from_dict(): await test_create_release_async(request_type=dict) def test_create_release_field_headers(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cloud_deploy.CreateReleaseRequest() request.parent = "parent/value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.create_release), "__call__") as call: call.return_value = operations_pb2.Operation(name="operations/op") client.create_release(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] assert args[0] == request # Establish that the field header was sent. _, _, kw = call.mock_calls[0] assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] @pytest.mark.asyncio async def test_create_release_field_headers_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cloud_deploy.CreateReleaseRequest() request.parent = "parent/value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.create_release), "__call__") as call: call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( operations_pb2.Operation(name="operations/op") ) await client.create_release(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] assert args[0] == request # Establish that the field header was sent. _, _, kw = call.mock_calls[0] assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] def test_create_release_flattened(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.create_release), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = operations_pb2.Operation(name="operations/op") # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. client.create_release( parent="parent_value", release=cloud_deploy.Release(name="name_value"), release_id="release_id_value", ) # Establish that the underlying call was made with the expected # request object values. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] arg = args[0].parent mock_val = "parent_value" assert arg == mock_val arg = args[0].release mock_val = cloud_deploy.Release(name="name_value") assert arg == mock_val arg = args[0].release_id mock_val = "release_id_value" assert arg == mock_val def test_create_release_flattened_error(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Attempting to call a method with both a request object and flattened # fields is an error. with pytest.raises(ValueError): client.create_release( cloud_deploy.CreateReleaseRequest(), parent="parent_value", release=cloud_deploy.Release(name="name_value"), release_id="release_id_value", ) @pytest.mark.asyncio async def test_create_release_flattened_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.create_release), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = operations_pb2.Operation(name="operations/op") call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( operations_pb2.Operation(name="operations/spam") ) # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. response = await client.create_release( parent="parent_value", release=cloud_deploy.Release(name="name_value"), release_id="release_id_value", ) # Establish that the underlying call was made with the expected # request object values. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] arg = args[0].parent mock_val = "parent_value" assert arg == mock_val arg = args[0].release mock_val = cloud_deploy.Release(name="name_value") assert arg == mock_val arg = args[0].release_id mock_val = "release_id_value" assert arg == mock_val @pytest.mark.asyncio async def test_create_release_flattened_error_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Attempting to call a method with both a request object and flattened # fields is an error. with pytest.raises(ValueError): await client.create_release( cloud_deploy.CreateReleaseRequest(), parent="parent_value", release=cloud_deploy.Release(name="name_value"), release_id="release_id_value", ) @pytest.mark.parametrize("request_type", [cloud_deploy.ApproveRolloutRequest, dict,]) def test_approve_rollout(request_type, transport: str = "grpc"): client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, # and we are mocking out the actual API, so just send an empty request. request = request_type() # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.approve_rollout), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = cloud_deploy.ApproveRolloutResponse() response = client.approve_rollout(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.ApproveRolloutRequest() # Establish that the response is the type that we expect. assert isinstance(response, cloud_deploy.ApproveRolloutResponse) def test_approve_rollout_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.approve_rollout), "__call__") as call: client.approve_rollout() call.assert_called() _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.ApproveRolloutRequest() @pytest.mark.asyncio async def test_approve_rollout_async( transport: str = "grpc_asyncio", request_type=cloud_deploy.ApproveRolloutRequest ): client = CloudDeployAsyncClient( credentials=ga_credentials.AnonymousCredentials(), transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, # and we are mocking out the actual API, so just send an empty request. request = request_type() # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.approve_rollout), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( cloud_deploy.ApproveRolloutResponse() ) response = await client.approve_rollout(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.ApproveRolloutRequest() # Establish that the response is the type that we expect. assert isinstance(response, cloud_deploy.ApproveRolloutResponse) @pytest.mark.asyncio async def test_approve_rollout_async_from_dict(): await test_approve_rollout_async(request_type=dict) def test_approve_rollout_field_headers(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cloud_deploy.ApproveRolloutRequest() request.name = "name/value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.approve_rollout), "__call__") as call: call.return_value = cloud_deploy.ApproveRolloutResponse() client.approve_rollout(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] assert args[0] == request # Establish that the field header was sent. _, _, kw = call.mock_calls[0] assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] @pytest.mark.asyncio async def test_approve_rollout_field_headers_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cloud_deploy.ApproveRolloutRequest() request.name = "name/value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.approve_rollout), "__call__") as call: call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( cloud_deploy.ApproveRolloutResponse() ) await client.approve_rollout(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] assert args[0] == request # Establish that the field header was sent. _, _, kw = call.mock_calls[0] assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] def test_approve_rollout_flattened(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.approve_rollout), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = cloud_deploy.ApproveRolloutResponse() # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. client.approve_rollout(name="name_value",) # Establish that the underlying call was made with the expected # request object values. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] arg = args[0].name mock_val = "name_value" assert arg == mock_val def test_approve_rollout_flattened_error(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Attempting to call a method with both a request object and flattened # fields is an error. with pytest.raises(ValueError): client.approve_rollout( cloud_deploy.ApproveRolloutRequest(), name="name_value", ) @pytest.mark.asyncio async def test_approve_rollout_flattened_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.approve_rollout), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = cloud_deploy.ApproveRolloutResponse() call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( cloud_deploy.ApproveRolloutResponse() ) # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. response = await client.approve_rollout(name="name_value",) # Establish that the underlying call was made with the expected # request object values. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] arg = args[0].name mock_val = "name_value" assert arg == mock_val @pytest.mark.asyncio async def test_approve_rollout_flattened_error_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Attempting to call a method with both a request object and flattened # fields is an error. with pytest.raises(ValueError): await client.approve_rollout( cloud_deploy.ApproveRolloutRequest(), name="name_value", ) @pytest.mark.parametrize("request_type", [cloud_deploy.ListRolloutsRequest, dict,]) def test_list_rollouts(request_type, transport: str = "grpc"): client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, # and we are mocking out the actual API, so just send an empty request. request = request_type() # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_rollouts), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = cloud_deploy.ListRolloutsResponse( next_page_token="next_page_token_value", unreachable=["unreachable_value"], ) response = client.list_rollouts(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.ListRolloutsRequest() # Establish that the response is the type that we expect. assert isinstance(response, pagers.ListRolloutsPager) assert response.next_page_token == "next_page_token_value" assert response.unreachable == ["unreachable_value"] def test_list_rollouts_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_rollouts), "__call__") as call: client.list_rollouts() call.assert_called() _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.ListRolloutsRequest() @pytest.mark.asyncio async def test_list_rollouts_async( transport: str = "grpc_asyncio", request_type=cloud_deploy.ListRolloutsRequest ): client = CloudDeployAsyncClient( credentials=ga_credentials.AnonymousCredentials(), transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, # and we are mocking out the actual API, so just send an empty request. request = request_type() # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_rollouts), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( cloud_deploy.ListRolloutsResponse( next_page_token="next_page_token_value", unreachable=["unreachable_value"], ) ) response = await client.list_rollouts(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.ListRolloutsRequest() # Establish that the response is the type that we expect. assert isinstance(response, pagers.ListRolloutsAsyncPager) assert response.next_page_token == "next_page_token_value" assert response.unreachable == ["unreachable_value"] @pytest.mark.asyncio async def test_list_rollouts_async_from_dict(): await test_list_rollouts_async(request_type=dict) def test_list_rollouts_field_headers(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cloud_deploy.ListRolloutsRequest() request.parent = "parent/value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_rollouts), "__call__") as call: call.return_value = cloud_deploy.ListRolloutsResponse() client.list_rollouts(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] assert args[0] == request # Establish that the field header was sent. _, _, kw = call.mock_calls[0] assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] @pytest.mark.asyncio async def test_list_rollouts_field_headers_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cloud_deploy.ListRolloutsRequest() request.parent = "parent/value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_rollouts), "__call__") as call: call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( cloud_deploy.ListRolloutsResponse() ) await client.list_rollouts(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] assert args[0] == request # Establish that the field header was sent. _, _, kw = call.mock_calls[0] assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] def test_list_rollouts_flattened(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_rollouts), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = cloud_deploy.ListRolloutsResponse() # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. client.list_rollouts(parent="parent_value",) # Establish that the underlying call was made with the expected # request object values. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] arg = args[0].parent mock_val = "parent_value" assert arg == mock_val def test_list_rollouts_flattened_error(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Attempting to call a method with both a request object and flattened # fields is an error. with pytest.raises(ValueError): client.list_rollouts( cloud_deploy.ListRolloutsRequest(), parent="parent_value", ) @pytest.mark.asyncio async def test_list_rollouts_flattened_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_rollouts), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = cloud_deploy.ListRolloutsResponse() call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( cloud_deploy.ListRolloutsResponse() ) # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. response = await client.list_rollouts(parent="parent_value",) # Establish that the underlying call was made with the expected # request object values. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] arg = args[0].parent mock_val = "parent_value" assert arg == mock_val @pytest.mark.asyncio async def test_list_rollouts_flattened_error_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Attempting to call a method with both a request object and flattened # fields is an error. with pytest.raises(ValueError): await client.list_rollouts( cloud_deploy.ListRolloutsRequest(), parent="parent_value", ) def test_list_rollouts_pager(transport_name: str = "grpc"): client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials, transport=transport_name, ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_rollouts), "__call__") as call: # Set the response to a series of pages. call.side_effect = ( cloud_deploy.ListRolloutsResponse( rollouts=[ cloud_deploy.Rollout(), cloud_deploy.Rollout(), cloud_deploy.Rollout(), ], next_page_token="abc", ), cloud_deploy.ListRolloutsResponse(rollouts=[], next_page_token="def",), cloud_deploy.ListRolloutsResponse( rollouts=[cloud_deploy.Rollout(),], next_page_token="ghi", ), cloud_deploy.ListRolloutsResponse( rollouts=[cloud_deploy.Rollout(), cloud_deploy.Rollout(),], ), RuntimeError, ) metadata = () metadata = tuple(metadata) + ( gapic_v1.routing_header.to_grpc_metadata((("parent", ""),)), ) pager = client.list_rollouts(request={}) assert pager._metadata == metadata results = [i for i in pager] assert len(results) == 6 assert all(isinstance(i, cloud_deploy.Rollout) for i in results) def test_list_rollouts_pages(transport_name: str = "grpc"): client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials, transport=transport_name, ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.list_rollouts), "__call__") as call: # Set the response to a series of pages. call.side_effect = ( cloud_deploy.ListRolloutsResponse( rollouts=[ cloud_deploy.Rollout(), cloud_deploy.Rollout(), cloud_deploy.Rollout(), ], next_page_token="abc", ), cloud_deploy.ListRolloutsResponse(rollouts=[], next_page_token="def",), cloud_deploy.ListRolloutsResponse( rollouts=[cloud_deploy.Rollout(),], next_page_token="ghi", ), cloud_deploy.ListRolloutsResponse( rollouts=[cloud_deploy.Rollout(), cloud_deploy.Rollout(),], ), RuntimeError, ) pages = list(client.list_rollouts(request={}).pages) for page_, token in zip(pages, ["abc", "def", "ghi", ""]): assert page_.raw_page.next_page_token == token @pytest.mark.asyncio async def test_list_rollouts_async_pager(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials,) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.list_rollouts), "__call__", new_callable=mock.AsyncMock ) as call: # Set the response to a series of pages. call.side_effect = ( cloud_deploy.ListRolloutsResponse( rollouts=[ cloud_deploy.Rollout(), cloud_deploy.Rollout(), cloud_deploy.Rollout(), ], next_page_token="abc", ), cloud_deploy.ListRolloutsResponse(rollouts=[], next_page_token="def",), cloud_deploy.ListRolloutsResponse( rollouts=[cloud_deploy.Rollout(),], next_page_token="ghi", ), cloud_deploy.ListRolloutsResponse( rollouts=[cloud_deploy.Rollout(), cloud_deploy.Rollout(),], ), RuntimeError, ) async_pager = await client.list_rollouts(request={},) assert async_pager.next_page_token == "abc" responses = [] async for response in async_pager: responses.append(response) assert len(responses) == 6 assert all(isinstance(i, cloud_deploy.Rollout) for i in responses) @pytest.mark.asyncio async def test_list_rollouts_async_pages(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials,) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object( type(client.transport.list_rollouts), "__call__", new_callable=mock.AsyncMock ) as call: # Set the response to a series of pages. call.side_effect = ( cloud_deploy.ListRolloutsResponse( rollouts=[ cloud_deploy.Rollout(), cloud_deploy.Rollout(), cloud_deploy.Rollout(), ], next_page_token="abc", ), cloud_deploy.ListRolloutsResponse(rollouts=[], next_page_token="def",), cloud_deploy.ListRolloutsResponse( rollouts=[cloud_deploy.Rollout(),], next_page_token="ghi", ), cloud_deploy.ListRolloutsResponse( rollouts=[cloud_deploy.Rollout(), cloud_deploy.Rollout(),], ), RuntimeError, ) pages = [] async for page_ in (await client.list_rollouts(request={})).pages: pages.append(page_) for page_, token in zip(pages, ["abc", "def", "ghi", ""]): assert page_.raw_page.next_page_token == token @pytest.mark.parametrize("request_type", [cloud_deploy.GetRolloutRequest, dict,]) def test_get_rollout(request_type, transport: str = "grpc"): client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, # and we are mocking out the actual API, so just send an empty request. request = request_type() # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_rollout), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = cloud_deploy.Rollout( name="name_value", uid="uid_value", description="description_value", target_id="target_id_value", approval_state=cloud_deploy.Rollout.ApprovalState.NEEDS_APPROVAL, state=cloud_deploy.Rollout.State.SUCCEEDED, failure_reason="failure_reason_value", deploying_build="deploying_build_value", etag="etag_value", ) response = client.get_rollout(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.GetRolloutRequest() # Establish that the response is the type that we expect. assert isinstance(response, cloud_deploy.Rollout) assert response.name == "name_value" assert response.uid == "uid_value" assert response.description == "description_value" assert response.target_id == "target_id_value" assert response.approval_state == cloud_deploy.Rollout.ApprovalState.NEEDS_APPROVAL assert response.state == cloud_deploy.Rollout.State.SUCCEEDED assert response.failure_reason == "failure_reason_value" assert response.deploying_build == "deploying_build_value" assert response.etag == "etag_value" def test_get_rollout_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_rollout), "__call__") as call: client.get_rollout() call.assert_called() _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.GetRolloutRequest() @pytest.mark.asyncio async def test_get_rollout_async( transport: str = "grpc_asyncio", request_type=cloud_deploy.GetRolloutRequest ): client = CloudDeployAsyncClient( credentials=ga_credentials.AnonymousCredentials(), transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, # and we are mocking out the actual API, so just send an empty request. request = request_type() # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_rollout), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( cloud_deploy.Rollout( name="name_value", uid="uid_value", description="description_value", target_id="target_id_value", approval_state=cloud_deploy.Rollout.ApprovalState.NEEDS_APPROVAL, state=cloud_deploy.Rollout.State.SUCCEEDED, failure_reason="failure_reason_value", deploying_build="deploying_build_value", etag="etag_value", ) ) response = await client.get_rollout(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.GetRolloutRequest() # Establish that the response is the type that we expect. assert isinstance(response, cloud_deploy.Rollout) assert response.name == "name_value" assert response.uid == "uid_value" assert response.description == "description_value" assert response.target_id == "target_id_value" assert response.approval_state == cloud_deploy.Rollout.ApprovalState.NEEDS_APPROVAL assert response.state == cloud_deploy.Rollout.State.SUCCEEDED assert response.failure_reason == "failure_reason_value" assert response.deploying_build == "deploying_build_value" assert response.etag == "etag_value" @pytest.mark.asyncio async def test_get_rollout_async_from_dict(): await test_get_rollout_async(request_type=dict) def test_get_rollout_field_headers(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cloud_deploy.GetRolloutRequest() request.name = "name/value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_rollout), "__call__") as call: call.return_value = cloud_deploy.Rollout() client.get_rollout(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] assert args[0] == request # Establish that the field header was sent. _, _, kw = call.mock_calls[0] assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] @pytest.mark.asyncio async def test_get_rollout_field_headers_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cloud_deploy.GetRolloutRequest() request.name = "name/value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_rollout), "__call__") as call: call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( cloud_deploy.Rollout() ) await client.get_rollout(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] assert args[0] == request # Establish that the field header was sent. _, _, kw = call.mock_calls[0] assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] def test_get_rollout_flattened(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_rollout), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = cloud_deploy.Rollout() # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. client.get_rollout(name="name_value",) # Establish that the underlying call was made with the expected # request object values. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] arg = args[0].name mock_val = "name_value" assert arg == mock_val def test_get_rollout_flattened_error(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Attempting to call a method with both a request object and flattened # fields is an error. with pytest.raises(ValueError): client.get_rollout( cloud_deploy.GetRolloutRequest(), name="name_value", ) @pytest.mark.asyncio async def test_get_rollout_flattened_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_rollout), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = cloud_deploy.Rollout() call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( cloud_deploy.Rollout() ) # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. response = await client.get_rollout(name="name_value",) # Establish that the underlying call was made with the expected # request object values. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] arg = args[0].name mock_val = "name_value" assert arg == mock_val @pytest.mark.asyncio async def test_get_rollout_flattened_error_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Attempting to call a method with both a request object and flattened # fields is an error. with pytest.raises(ValueError): await client.get_rollout( cloud_deploy.GetRolloutRequest(), name="name_value", ) @pytest.mark.parametrize("request_type", [cloud_deploy.CreateRolloutRequest, dict,]) def test_create_rollout(request_type, transport: str = "grpc"): client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, # and we are mocking out the actual API, so just send an empty request. request = request_type() # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.create_rollout), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = operations_pb2.Operation(name="operations/spam") response = client.create_rollout(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.CreateRolloutRequest() # Establish that the response is the type that we expect. assert isinstance(response, future.Future) def test_create_rollout_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.create_rollout), "__call__") as call: client.create_rollout() call.assert_called() _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.CreateRolloutRequest() @pytest.mark.asyncio async def test_create_rollout_async( transport: str = "grpc_asyncio", request_type=cloud_deploy.CreateRolloutRequest ): client = CloudDeployAsyncClient( credentials=ga_credentials.AnonymousCredentials(), transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, # and we are mocking out the actual API, so just send an empty request. request = request_type() # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.create_rollout), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( operations_pb2.Operation(name="operations/spam") ) response = await client.create_rollout(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.CreateRolloutRequest() # Establish that the response is the type that we expect. assert isinstance(response, future.Future) @pytest.mark.asyncio async def test_create_rollout_async_from_dict(): await test_create_rollout_async(request_type=dict) def test_create_rollout_field_headers(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cloud_deploy.CreateRolloutRequest() request.parent = "parent/value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.create_rollout), "__call__") as call: call.return_value = operations_pb2.Operation(name="operations/op") client.create_rollout(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] assert args[0] == request # Establish that the field header was sent. _, _, kw = call.mock_calls[0] assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] @pytest.mark.asyncio async def test_create_rollout_field_headers_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cloud_deploy.CreateRolloutRequest() request.parent = "parent/value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.create_rollout), "__call__") as call: call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( operations_pb2.Operation(name="operations/op") ) await client.create_rollout(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] assert args[0] == request # Establish that the field header was sent. _, _, kw = call.mock_calls[0] assert ("x-goog-request-params", "parent=parent/value",) in kw["metadata"] def test_create_rollout_flattened(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.create_rollout), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = operations_pb2.Operation(name="operations/op") # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. client.create_rollout( parent="parent_value", rollout=cloud_deploy.Rollout(name="name_value"), rollout_id="rollout_id_value", ) # Establish that the underlying call was made with the expected # request object values. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] arg = args[0].parent mock_val = "parent_value" assert arg == mock_val arg = args[0].rollout mock_val = cloud_deploy.Rollout(name="name_value") assert arg == mock_val arg = args[0].rollout_id mock_val = "rollout_id_value" assert arg == mock_val def test_create_rollout_flattened_error(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Attempting to call a method with both a request object and flattened # fields is an error. with pytest.raises(ValueError): client.create_rollout( cloud_deploy.CreateRolloutRequest(), parent="parent_value", rollout=cloud_deploy.Rollout(name="name_value"), rollout_id="rollout_id_value", ) @pytest.mark.asyncio async def test_create_rollout_flattened_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.create_rollout), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = operations_pb2.Operation(name="operations/op") call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( operations_pb2.Operation(name="operations/spam") ) # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. response = await client.create_rollout( parent="parent_value", rollout=cloud_deploy.Rollout(name="name_value"), rollout_id="rollout_id_value", ) # Establish that the underlying call was made with the expected # request object values. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] arg = args[0].parent mock_val = "parent_value" assert arg == mock_val arg = args[0].rollout mock_val = cloud_deploy.Rollout(name="name_value") assert arg == mock_val arg = args[0].rollout_id mock_val = "rollout_id_value" assert arg == mock_val @pytest.mark.asyncio async def test_create_rollout_flattened_error_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Attempting to call a method with both a request object and flattened # fields is an error. with pytest.raises(ValueError): await client.create_rollout( cloud_deploy.CreateRolloutRequest(), parent="parent_value", rollout=cloud_deploy.Rollout(name="name_value"), rollout_id="rollout_id_value", ) @pytest.mark.parametrize("request_type", [cloud_deploy.GetConfigRequest, dict,]) def test_get_config(request_type, transport: str = "grpc"): client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, # and we are mocking out the actual API, so just send an empty request. request = request_type() # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_config), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = cloud_deploy.Config( name="name_value", default_skaffold_version="default_skaffold_version_value", ) response = client.get_config(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.GetConfigRequest() # Establish that the response is the type that we expect. assert isinstance(response, cloud_deploy.Config) assert response.name == "name_value" assert response.default_skaffold_version == "default_skaffold_version_value" def test_get_config_empty_call(): # This test is a coverage failsafe to make sure that totally empty calls, # i.e. request == None and no flattened fields passed, work. client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport="grpc", ) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_config), "__call__") as call: client.get_config() call.assert_called() _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.GetConfigRequest() @pytest.mark.asyncio async def test_get_config_async( transport: str = "grpc_asyncio", request_type=cloud_deploy.GetConfigRequest ): client = CloudDeployAsyncClient( credentials=ga_credentials.AnonymousCredentials(), transport=transport, ) # Everything is optional in proto3 as far as the runtime is concerned, # and we are mocking out the actual API, so just send an empty request. request = request_type() # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_config), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = grpc_helpers_async.FakeUnaryUnaryCall( cloud_deploy.Config( name="name_value", default_skaffold_version="default_skaffold_version_value", ) ) response = await client.get_config(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] assert args[0] == cloud_deploy.GetConfigRequest() # Establish that the response is the type that we expect. assert isinstance(response, cloud_deploy.Config) assert response.name == "name_value" assert response.default_skaffold_version == "default_skaffold_version_value" @pytest.mark.asyncio async def test_get_config_async_from_dict(): await test_get_config_async(request_type=dict) def test_get_config_field_headers(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cloud_deploy.GetConfigRequest() request.name = "name/value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_config), "__call__") as call: call.return_value = cloud_deploy.Config() client.get_config(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] assert args[0] == request # Establish that the field header was sent. _, _, kw = call.mock_calls[0] assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] @pytest.mark.asyncio async def test_get_config_field_headers_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Any value that is part of the HTTP/1.1 URI should be sent as # a field header. Set these to a non-empty value. request = cloud_deploy.GetConfigRequest() request.name = "name/value" # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_config), "__call__") as call: call.return_value = grpc_helpers_async.FakeUnaryUnaryCall(cloud_deploy.Config()) await client.get_config(request) # Establish that the underlying gRPC stub method was called. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] assert args[0] == request # Establish that the field header was sent. _, _, kw = call.mock_calls[0] assert ("x-goog-request-params", "name=name/value",) in kw["metadata"] def test_get_config_flattened(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_config), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = cloud_deploy.Config() # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. client.get_config(name="name_value",) # Establish that the underlying call was made with the expected # request object values. assert len(call.mock_calls) == 1 _, args, _ = call.mock_calls[0] arg = args[0].name mock_val = "name_value" assert arg == mock_val def test_get_config_flattened_error(): client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) # Attempting to call a method with both a request object and flattened # fields is an error. with pytest.raises(ValueError): client.get_config( cloud_deploy.GetConfigRequest(), name="name_value", ) @pytest.mark.asyncio async def test_get_config_flattened_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Mock the actual call within the gRPC stub, and fake the request. with mock.patch.object(type(client.transport.get_config), "__call__") as call: # Designate an appropriate return value for the call. call.return_value = cloud_deploy.Config() call.return_value = grpc_helpers_async.FakeUnaryUnaryCall(cloud_deploy.Config()) # Call the method with a truthy value for each flattened field, # using the keyword arguments to the method. response = await client.get_config(name="name_value",) # Establish that the underlying call was made with the expected # request object values. assert len(call.mock_calls) _, args, _ = call.mock_calls[0] arg = args[0].name mock_val = "name_value" assert arg == mock_val @pytest.mark.asyncio async def test_get_config_flattened_error_async(): client = CloudDeployAsyncClient(credentials=ga_credentials.AnonymousCredentials(),) # Attempting to call a method with both a request object and flattened # fields is an error. with pytest.raises(ValueError): await client.get_config( cloud_deploy.GetConfigRequest(), name="name_value", ) def test_credentials_transport_error(): # It is an error to provide credentials and a transport instance. transport = transports.CloudDeployGrpcTransport( credentials=ga_credentials.AnonymousCredentials(), ) with pytest.raises(ValueError): client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport=transport, ) # It is an error to provide a credentials file and a transport instance. transport = transports.CloudDeployGrpcTransport( credentials=ga_credentials.AnonymousCredentials(), ) with pytest.raises(ValueError): client = CloudDeployClient( client_options={"credentials_file": "credentials.json"}, transport=transport, ) # It is an error to provide an api_key and a transport instance. transport = transports.CloudDeployGrpcTransport( credentials=ga_credentials.AnonymousCredentials(), ) options = client_options.ClientOptions() options.api_key = "api_key" with pytest.raises(ValueError): client = CloudDeployClient(client_options=options, transport=transport,) # It is an error to provide an api_key and a credential. options = mock.Mock() options.api_key = "api_key" with pytest.raises(ValueError): client = CloudDeployClient( client_options=options, credentials=ga_credentials.AnonymousCredentials() ) # It is an error to provide scopes and a transport instance. transport = transports.CloudDeployGrpcTransport( credentials=ga_credentials.AnonymousCredentials(), ) with pytest.raises(ValueError): client = CloudDeployClient( client_options={"scopes": ["1", "2"]}, transport=transport, ) def test_transport_instance(): # A client may be instantiated with a custom transport instance. transport = transports.CloudDeployGrpcTransport( credentials=ga_credentials.AnonymousCredentials(), ) client = CloudDeployClient(transport=transport) assert client.transport is transport def test_transport_get_channel(): # A client may be instantiated with a custom transport instance. transport = transports.CloudDeployGrpcTransport( credentials=ga_credentials.AnonymousCredentials(), ) channel = transport.grpc_channel assert channel transport = transports.CloudDeployGrpcAsyncIOTransport( credentials=ga_credentials.AnonymousCredentials(), ) channel = transport.grpc_channel assert channel @pytest.mark.parametrize( "transport_class", [transports.CloudDeployGrpcTransport, transports.CloudDeployGrpcAsyncIOTransport,], ) def test_transport_adc(transport_class): # Test default credentials are used if not provided. with mock.patch.object(google.auth, "default") as adc: adc.return_value = (ga_credentials.AnonymousCredentials(), None) transport_class() adc.assert_called_once() def test_transport_grpc_default(): # A client should use the gRPC transport by default. client = CloudDeployClient(credentials=ga_credentials.AnonymousCredentials(),) assert isinstance(client.transport, transports.CloudDeployGrpcTransport,) def test_cloud_deploy_base_transport_error(): # Passing both a credentials object and credentials_file should raise an error with pytest.raises(core_exceptions.DuplicateCredentialArgs): transport = transports.CloudDeployTransport( credentials=ga_credentials.AnonymousCredentials(), credentials_file="credentials.json", ) def test_cloud_deploy_base_transport(): # Instantiate the base transport. with mock.patch( "google.cloud.deploy_v1.services.cloud_deploy.transports.CloudDeployTransport.__init__" ) as Transport: Transport.return_value = None transport = transports.CloudDeployTransport( credentials=ga_credentials.AnonymousCredentials(), ) # Every method on the transport should just blindly # raise NotImplementedError. methods = ( "list_delivery_pipelines", "get_delivery_pipeline", "create_delivery_pipeline", "update_delivery_pipeline", "delete_delivery_pipeline", "list_targets", "get_target", "create_target", "update_target", "delete_target", "list_releases", "get_release", "create_release", "approve_rollout", "list_rollouts", "get_rollout", "create_rollout", "get_config", ) for method in methods: with pytest.raises(NotImplementedError): getattr(transport, method)(request=object()) with pytest.raises(NotImplementedError): transport.close() # Additionally, the LRO client (a property) should # also raise NotImplementedError with pytest.raises(NotImplementedError): transport.operations_client def test_cloud_deploy_base_transport_with_credentials_file(): # Instantiate the base transport with a credentials file with mock.patch.object( google.auth, "load_credentials_from_file", autospec=True ) as load_creds, mock.patch( "google.cloud.deploy_v1.services.cloud_deploy.transports.CloudDeployTransport._prep_wrapped_messages" ) as Transport: Transport.return_value = None load_creds.return_value = (ga_credentials.AnonymousCredentials(), None) transport = transports.CloudDeployTransport( credentials_file="credentials.json", quota_project_id="octopus", ) load_creds.assert_called_once_with( "credentials.json", scopes=None, default_scopes=("https://www.googleapis.com/auth/cloud-platform",), quota_project_id="octopus", ) def test_cloud_deploy_base_transport_with_adc(): # Test the default credentials are used if credentials and credentials_file are None. with mock.patch.object(google.auth, "default", autospec=True) as adc, mock.patch( "google.cloud.deploy_v1.services.cloud_deploy.transports.CloudDeployTransport._prep_wrapped_messages" ) as Transport: Transport.return_value = None adc.return_value = (ga_credentials.AnonymousCredentials(), None) transport = transports.CloudDeployTransport() adc.assert_called_once() def test_cloud_deploy_auth_adc(): # If no credentials are provided, we should use ADC credentials. with mock.patch.object(google.auth, "default", autospec=True) as adc: adc.return_value = (ga_credentials.AnonymousCredentials(), None) CloudDeployClient() adc.assert_called_once_with( scopes=None, default_scopes=("https://www.googleapis.com/auth/cloud-platform",), quota_project_id=None, ) @pytest.mark.parametrize( "transport_class", [transports.CloudDeployGrpcTransport, transports.CloudDeployGrpcAsyncIOTransport,], ) def test_cloud_deploy_transport_auth_adc(transport_class): # If credentials and host are not provided, the transport class should use # ADC credentials. with mock.patch.object(google.auth, "default", autospec=True) as adc: adc.return_value = (ga_credentials.AnonymousCredentials(), None) transport_class(quota_project_id="octopus", scopes=["1", "2"]) adc.assert_called_once_with( scopes=["1", "2"], default_scopes=("https://www.googleapis.com/auth/cloud-platform",), quota_project_id="octopus", ) @pytest.mark.parametrize( "transport_class,grpc_helpers", [ (transports.CloudDeployGrpcTransport, grpc_helpers), (transports.CloudDeployGrpcAsyncIOTransport, grpc_helpers_async), ], ) def test_cloud_deploy_transport_create_channel(transport_class, grpc_helpers): # If credentials and host are not provided, the transport class should use # ADC credentials. with mock.patch.object( google.auth, "default", autospec=True ) as adc, mock.patch.object( grpc_helpers, "create_channel", autospec=True ) as create_channel: creds = ga_credentials.AnonymousCredentials() adc.return_value = (creds, None) transport_class(quota_project_id="octopus", scopes=["1", "2"]) create_channel.assert_called_with( "clouddeploy.googleapis.com:443", credentials=creds, credentials_file=None, quota_project_id="octopus", default_scopes=("https://www.googleapis.com/auth/cloud-platform",), scopes=["1", "2"], default_host="clouddeploy.googleapis.com", ssl_credentials=None, options=[ ("grpc.max_send_message_length", -1), ("grpc.max_receive_message_length", -1), ], ) @pytest.mark.parametrize( "transport_class", [transports.CloudDeployGrpcTransport, transports.CloudDeployGrpcAsyncIOTransport], ) def test_cloud_deploy_grpc_transport_client_cert_source_for_mtls(transport_class): cred = ga_credentials.AnonymousCredentials() # Check ssl_channel_credentials is used if provided. with mock.patch.object(transport_class, "create_channel") as mock_create_channel: mock_ssl_channel_creds = mock.Mock() transport_class( host="squid.clam.whelk", credentials=cred, ssl_channel_credentials=mock_ssl_channel_creds, ) mock_create_channel.assert_called_once_with( "squid.clam.whelk:443", credentials=cred, credentials_file=None, scopes=None, ssl_credentials=mock_ssl_channel_creds, quota_project_id=None, options=[ ("grpc.max_send_message_length", -1), ("grpc.max_receive_message_length", -1), ], ) # Check if ssl_channel_credentials is not provided, then client_cert_source_for_mtls # is used. with mock.patch.object(transport_class, "create_channel", return_value=mock.Mock()): with mock.patch("grpc.ssl_channel_credentials") as mock_ssl_cred: transport_class( credentials=cred, client_cert_source_for_mtls=client_cert_source_callback, ) expected_cert, expected_key = client_cert_source_callback() mock_ssl_cred.assert_called_once_with( certificate_chain=expected_cert, private_key=expected_key ) def test_cloud_deploy_host_no_port(): client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), client_options=client_options.ClientOptions( api_endpoint="clouddeploy.googleapis.com" ), ) assert client.transport._host == "clouddeploy.googleapis.com:443" def test_cloud_deploy_host_with_port(): client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), client_options=client_options.ClientOptions( api_endpoint="clouddeploy.googleapis.com:8000" ), ) assert client.transport._host == "clouddeploy.googleapis.com:8000" def test_cloud_deploy_grpc_transport_channel(): channel = grpc.secure_channel("http://localhost/", grpc.local_channel_credentials()) # Check that channel is used if provided. transport = transports.CloudDeployGrpcTransport( host="squid.clam.whelk", channel=channel, ) assert transport.grpc_channel == channel assert transport._host == "squid.clam.whelk:443" assert transport._ssl_channel_credentials == None def test_cloud_deploy_grpc_asyncio_transport_channel(): channel = aio.secure_channel("http://localhost/", grpc.local_channel_credentials()) # Check that channel is used if provided. transport = transports.CloudDeployGrpcAsyncIOTransport( host="squid.clam.whelk", channel=channel, ) assert transport.grpc_channel == channel assert transport._host == "squid.clam.whelk:443" assert transport._ssl_channel_credentials == None # Remove this test when deprecated arguments (api_mtls_endpoint, client_cert_source) are # removed from grpc/grpc_asyncio transport constructor. @pytest.mark.parametrize( "transport_class", [transports.CloudDeployGrpcTransport, transports.CloudDeployGrpcAsyncIOTransport], ) def test_cloud_deploy_transport_channel_mtls_with_client_cert_source(transport_class): with mock.patch( "grpc.ssl_channel_credentials", autospec=True ) as grpc_ssl_channel_cred: with mock.patch.object( transport_class, "create_channel" ) as grpc_create_channel: mock_ssl_cred = mock.Mock() grpc_ssl_channel_cred.return_value = mock_ssl_cred mock_grpc_channel = mock.Mock() grpc_create_channel.return_value = mock_grpc_channel cred = ga_credentials.AnonymousCredentials() with pytest.warns(DeprecationWarning): with mock.patch.object(google.auth, "default") as adc: adc.return_value = (cred, None) transport = transport_class( host="squid.clam.whelk", api_mtls_endpoint="mtls.squid.clam.whelk", client_cert_source=client_cert_source_callback, ) adc.assert_called_once() grpc_ssl_channel_cred.assert_called_once_with( certificate_chain=b"cert bytes", private_key=b"key bytes" ) grpc_create_channel.assert_called_once_with( "mtls.squid.clam.whelk:443", credentials=cred, credentials_file=None, scopes=None, ssl_credentials=mock_ssl_cred, quota_project_id=None, options=[ ("grpc.max_send_message_length", -1), ("grpc.max_receive_message_length", -1), ], ) assert transport.grpc_channel == mock_grpc_channel assert transport._ssl_channel_credentials == mock_ssl_cred # Remove this test when deprecated arguments (api_mtls_endpoint, client_cert_source) are # removed from grpc/grpc_asyncio transport constructor. @pytest.mark.parametrize( "transport_class", [transports.CloudDeployGrpcTransport, transports.CloudDeployGrpcAsyncIOTransport], ) def test_cloud_deploy_transport_channel_mtls_with_adc(transport_class): mock_ssl_cred = mock.Mock() with mock.patch.multiple( "google.auth.transport.grpc.SslCredentials", __init__=mock.Mock(return_value=None), ssl_credentials=mock.PropertyMock(return_value=mock_ssl_cred), ): with mock.patch.object( transport_class, "create_channel" ) as grpc_create_channel: mock_grpc_channel = mock.Mock() grpc_create_channel.return_value = mock_grpc_channel mock_cred = mock.Mock() with pytest.warns(DeprecationWarning): transport = transport_class( host="squid.clam.whelk", credentials=mock_cred, api_mtls_endpoint="mtls.squid.clam.whelk", client_cert_source=None, ) grpc_create_channel.assert_called_once_with( "mtls.squid.clam.whelk:443", credentials=mock_cred, credentials_file=None, scopes=None, ssl_credentials=mock_ssl_cred, quota_project_id=None, options=[ ("grpc.max_send_message_length", -1), ("grpc.max_receive_message_length", -1), ], ) assert transport.grpc_channel == mock_grpc_channel def test_cloud_deploy_grpc_lro_client(): client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport="grpc", ) transport = client.transport # Ensure that we have a api-core operations client. assert isinstance(transport.operations_client, operations_v1.OperationsClient,) # Ensure that subsequent calls to the property send the exact same object. assert transport.operations_client is transport.operations_client def test_cloud_deploy_grpc_lro_async_client(): client = CloudDeployAsyncClient( credentials=ga_credentials.AnonymousCredentials(), transport="grpc_asyncio", ) transport = client.transport # Ensure that we have a api-core operations client. assert isinstance(transport.operations_client, operations_v1.OperationsAsyncClient,) # Ensure that subsequent calls to the property send the exact same object. assert transport.operations_client is transport.operations_client def test_build_path(): project = "squid" location = "clam" build = "whelk" expected = "projects/{project}/locations/{location}/builds/{build}".format( project=project, location=location, build=build, ) actual = CloudDeployClient.build_path(project, location, build) assert expected == actual def test_parse_build_path(): expected = { "project": "octopus", "location": "oyster", "build": "nudibranch", } path = CloudDeployClient.build_path(**expected) # Check that the path construction is reversible. actual = CloudDeployClient.parse_build_path(path) assert expected == actual def test_cluster_path(): project = "cuttlefish" location = "mussel" cluster = "winkle" expected = "projects/{project}/locations/{location}/clusters/{cluster}".format( project=project, location=location, cluster=cluster, ) actual = CloudDeployClient.cluster_path(project, location, cluster) assert expected == actual def test_parse_cluster_path(): expected = { "project": "nautilus", "location": "scallop", "cluster": "abalone", } path = CloudDeployClient.cluster_path(**expected) # Check that the path construction is reversible. actual = CloudDeployClient.parse_cluster_path(path) assert expected == actual def test_config_path(): project = "squid" location = "clam" expected = "projects/{project}/locations/{location}/config".format( project=project, location=location, ) actual = CloudDeployClient.config_path(project, location) assert expected == actual def test_parse_config_path(): expected = { "project": "whelk", "location": "octopus", } path = CloudDeployClient.config_path(**expected) # Check that the path construction is reversible. actual = CloudDeployClient.parse_config_path(path) assert expected == actual def test_delivery_pipeline_path(): project = "oyster" location = "nudibranch" delivery_pipeline = "cuttlefish" expected = "projects/{project}/locations/{location}/deliveryPipelines/{delivery_pipeline}".format( project=project, location=location, delivery_pipeline=delivery_pipeline, ) actual = CloudDeployClient.delivery_pipeline_path( project, location, delivery_pipeline ) assert expected == actual def test_parse_delivery_pipeline_path(): expected = { "project": "mussel", "location": "winkle", "delivery_pipeline": "nautilus", } path = CloudDeployClient.delivery_pipeline_path(**expected) # Check that the path construction is reversible. actual = CloudDeployClient.parse_delivery_pipeline_path(path) assert expected == actual def test_release_path(): project = "scallop" location = "abalone" delivery_pipeline = "squid" release = "clam" expected = "projects/{project}/locations/{location}/deliveryPipelines/{delivery_pipeline}/releases/{release}".format( project=project, location=location, delivery_pipeline=delivery_pipeline, release=release, ) actual = CloudDeployClient.release_path( project, location, delivery_pipeline, release ) assert expected == actual def test_parse_release_path(): expected = { "project": "whelk", "location": "octopus", "delivery_pipeline": "oyster", "release": "nudibranch", } path = CloudDeployClient.release_path(**expected) # Check that the path construction is reversible. actual = CloudDeployClient.parse_release_path(path) assert expected == actual def test_rollout_path(): project = "cuttlefish" location = "mussel" delivery_pipeline = "winkle" release = "nautilus" rollout = "scallop" expected = "projects/{project}/locations/{location}/deliveryPipelines/{delivery_pipeline}/releases/{release}/rollouts/{rollout}".format( project=project, location=location, delivery_pipeline=delivery_pipeline, release=release, rollout=rollout, ) actual = CloudDeployClient.rollout_path( project, location, delivery_pipeline, release, rollout ) assert expected == actual def test_parse_rollout_path(): expected = { "project": "abalone", "location": "squid", "delivery_pipeline": "clam", "release": "whelk", "rollout": "octopus", } path = CloudDeployClient.rollout_path(**expected) # Check that the path construction is reversible. actual = CloudDeployClient.parse_rollout_path(path) assert expected == actual def test_target_path(): project = "oyster" location = "nudibranch" target = "cuttlefish" expected = "projects/{project}/locations/{location}/targets/{target}".format( project=project, location=location, target=target, ) actual = CloudDeployClient.target_path(project, location, target) assert expected == actual def test_parse_target_path(): expected = { "project": "mussel", "location": "winkle", "target": "nautilus", } path = CloudDeployClient.target_path(**expected) # Check that the path construction is reversible. actual = CloudDeployClient.parse_target_path(path) assert expected == actual def test_worker_pool_path(): project = "scallop" location = "abalone" worker_pool = "squid" expected = "projects/{project}/locations/{location}/workerPools/{worker_pool}".format( project=project, location=location, worker_pool=worker_pool, ) actual = CloudDeployClient.worker_pool_path(project, location, worker_pool) assert expected == actual def test_parse_worker_pool_path(): expected = { "project": "clam", "location": "whelk", "worker_pool": "octopus", } path = CloudDeployClient.worker_pool_path(**expected) # Check that the path construction is reversible. actual = CloudDeployClient.parse_worker_pool_path(path) assert expected == actual def test_common_billing_account_path(): billing_account = "oyster" expected = "billingAccounts/{billing_account}".format( billing_account=billing_account, ) actual = CloudDeployClient.common_billing_account_path(billing_account) assert expected == actual def test_parse_common_billing_account_path(): expected = { "billing_account": "nudibranch", } path = CloudDeployClient.common_billing_account_path(**expected) # Check that the path construction is reversible. actual = CloudDeployClient.parse_common_billing_account_path(path) assert expected == actual def test_common_folder_path(): folder = "cuttlefish" expected = "folders/{folder}".format(folder=folder,) actual = CloudDeployClient.common_folder_path(folder) assert expected == actual def test_parse_common_folder_path(): expected = { "folder": "mussel", } path = CloudDeployClient.common_folder_path(**expected) # Check that the path construction is reversible. actual = CloudDeployClient.parse_common_folder_path(path) assert expected == actual def test_common_organization_path(): organization = "winkle" expected = "organizations/{organization}".format(organization=organization,) actual = CloudDeployClient.common_organization_path(organization) assert expected == actual def test_parse_common_organization_path(): expected = { "organization": "nautilus", } path = CloudDeployClient.common_organization_path(**expected) # Check that the path construction is reversible. actual = CloudDeployClient.parse_common_organization_path(path) assert expected == actual def test_common_project_path(): project = "scallop" expected = "projects/{project}".format(project=project,) actual = CloudDeployClient.common_project_path(project) assert expected == actual def test_parse_common_project_path(): expected = { "project": "abalone", } path = CloudDeployClient.common_project_path(**expected) # Check that the path construction is reversible. actual = CloudDeployClient.parse_common_project_path(path) assert expected == actual def test_common_location_path(): project = "squid" location = "clam" expected = "projects/{project}/locations/{location}".format( project=project, location=location, ) actual = CloudDeployClient.common_location_path(project, location) assert expected == actual def test_parse_common_location_path(): expected = { "project": "whelk", "location": "octopus", } path = CloudDeployClient.common_location_path(**expected) # Check that the path construction is reversible. actual = CloudDeployClient.parse_common_location_path(path) assert expected == actual def test_client_with_default_client_info(): client_info = gapic_v1.client_info.ClientInfo() with mock.patch.object( transports.CloudDeployTransport, "_prep_wrapped_messages" ) as prep: client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), client_info=client_info, ) prep.assert_called_once_with(client_info) with mock.patch.object( transports.CloudDeployTransport, "_prep_wrapped_messages" ) as prep: transport_class = CloudDeployClient.get_transport_class() transport = transport_class( credentials=ga_credentials.AnonymousCredentials(), client_info=client_info, ) prep.assert_called_once_with(client_info) @pytest.mark.asyncio async def test_transport_close_async(): client = CloudDeployAsyncClient( credentials=ga_credentials.AnonymousCredentials(), transport="grpc_asyncio", ) with mock.patch.object( type(getattr(client.transport, "grpc_channel")), "close" ) as close: async with client: close.assert_not_called() close.assert_called_once() def test_transport_close(): transports = { "grpc": "_grpc_channel", } for transport, close_name in transports.items(): client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport=transport ) with mock.patch.object( type(getattr(client.transport, close_name)), "close" ) as close: with client: close.assert_not_called() close.assert_called_once() def test_client_ctx(): transports = [ "grpc", ] for transport in transports: client = CloudDeployClient( credentials=ga_credentials.AnonymousCredentials(), transport=transport ) # Test client calls underlying transport. with mock.patch.object(type(client.transport), "close") as close: close.assert_not_called() with client: pass close.assert_called() @pytest.mark.parametrize( "client_class,transport_class", [ (CloudDeployClient, transports.CloudDeployGrpcTransport), (CloudDeployAsyncClient, transports.CloudDeployGrpcAsyncIOTransport), ], ) def test_api_key_credentials(client_class, transport_class): with mock.patch.object( google.auth._default, "get_api_key_credentials", create=True ) as get_api_key_credentials: mock_cred = mock.Mock() get_api_key_credentials.return_value = mock_cred options = client_options.ClientOptions() options.api_key = "api_key" with mock.patch.object(transport_class, "__init__") as patched: patched.return_value = None client = client_class(client_options=options) patched.assert_called_once_with( credentials=mock_cred, credentials_file=None, host=client.DEFAULT_ENDPOINT, scopes=None, client_cert_source_for_mtls=None, quota_project_id=None, client_info=transports.base.DEFAULT_CLIENT_INFO, always_use_jwt_access=True, )
[ [ [ 607, 609 ], [ 11467, 11469 ], [ 7941, 7943 ], [ 8702, 8704 ], [ 9479, 9481 ], [ 9765, 9767 ], [ 11996, 11998 ], [ 13285, 13287 ], [ 14865, 14867 ], [ 16259, 16261 ], [ 16807, 16809 ], [ 17378, 17380 ], [ 17703, 17705 ], [ 18063, 18065 ], [ 18567, 18569 ] ], [ [ 617, 621 ], [ 6264, 6268 ], [ 6372, 6376 ], [ 11216, 11220 ], [ 11324, 11328 ], [ 11451, 11455 ], [ 15814, 15818 ], [ 15922, 15926 ], [ 3460, 3464 ], [ 4184, 4188 ], [ 4501, 4505 ], [ 5017, 5021 ], [ 6672, 6676 ], [ 6994, 6998 ], [ 7279, 7283 ], [ 7925, 7929 ], [ 8010, 8014 ], [ 8686, 8690 ], [ 8772, 8776 ], [ 9463, 9467 ], [ 9740, 9744 ], [ 10068, 10072 ], [ 11971, 11975 ], [ 12204, 12208 ], [ 13260, 13264 ], [ 13376, 13380 ], [ 13452, 13456 ], [ 13612, 13616 ], [ 14840, 14844 ], [ 14956, 14960 ], [ 15032, 15036 ], [ 16156, 16160 ], [ 16243, 16247 ], [ 16791, 16795 ], [ 16902, 16906 ], [ 17362, 17366 ], [ 17687, 17691 ], [ 18047, 18051 ], [ 18136, 18140 ], [ 18551, 18555 ], [ 18640, 18644 ], [ 18784, 18788 ], [ 19756, 19760 ], [ 20866, 20870 ], [ 21488, 21492 ], [ 22677, 22681 ], [ 23928, 23932 ], [ 24763, 24767 ], [ 26295, 26299 ], [ 27335, 27339 ], [ 28178, 28182 ], [ 29538, 29542 ], [ 31113, 31117 ], [ 32875, 32879 ], [ 34403, 34407 ], [ 34519, 34523 ], [ 36129, 36133 ], [ 36245, 36249 ], [ 38074, 38078 ], [ 39549, 39553 ], [ 40374, 40378 ], [ 41966, 41970 ], [ 42977, 42981 ], [ 43797, 43801 ], [ 45120, 45124 ], [ 46929, 46933 ], [ 47951, 47955 ], [ 48791, 48795 ], [ 50068, 50072 ], [ 51114, 51118 ], [ 51962, 51966 ], [ 53911, 53915 ], [ 56360, 56364 ], [ 57382, 57386 ], [ 58222, 58226 ], [ 59531, 59535 ], [ 60663, 60667 ], [ 61565, 61569 ], [ 63380, 63384 ], [ 65695, 65699 ], [ 66717, 66721 ], [ 67557, 67561 ], [ 68830, 68834 ], [ 69868, 69872 ], [ 70712, 70716 ], [ 72068, 72072 ], [ 73895, 73899 ], [ 75069, 75073 ], [ 75832, 75836 ], [ 77255, 77259 ], [ 78228, 78232 ], [ 79014, 79018 ], [ 80285, 80289 ], [ 81761, 81765 ], [ 83218, 83222 ], [ 84451, 84455 ], [ 84540, 84544 ], [ 85869, 85873 ], [ 85958, 85962 ], [ 87500, 87504 ], [ 88991, 88995 ], [ 89744, 89748 ], [ 91407, 91411 ], [ 92351, 92355 ], [ 93092, 93096 ], [ 94326, 94330 ], [ 95998, 96002 ], [ 96963, 96967 ], [ 97731, 97735 ], [ 98919, 98923 ], [ 99908, 99912 ], [ 100709, 100713 ], [ 102450, 102454 ], [ 104675, 104679 ], [ 105640, 105644 ], [ 106408, 106412 ], [ 107606, 107610 ], [ 108615, 108619 ], [ 109426, 109430 ], [ 111099, 111103 ], [ 113256, 113260 ], [ 114221, 114225 ], [ 114989, 114993 ], [ 116173, 116177 ], [ 117154, 117158 ], [ 117951, 117955 ], [ 119228, 119232 ], [ 120989, 120993 ], [ 122169, 122173 ], [ 122937, 122941 ], [ 124369, 124373 ], [ 125347, 125351 ], [ 126137, 126141 ], [ 127415, 127419 ], [ 128899, 128903 ], [ 130374, 130378 ], [ 131624, 131628 ], [ 131714, 131718 ], [ 133060, 133064 ], [ 133150, 133154 ], [ 134709, 134713 ], [ 136504, 136508 ], [ 137262, 137266 ], [ 139306, 139310 ], [ 140255, 140259 ], [ 141022, 141026 ], [ 142263, 142267 ], [ 143966, 143970 ], [ 144935, 144939 ], [ 145708, 145712 ], [ 146903, 146907 ], [ 147896, 147900 ], [ 148700, 148704 ], [ 150459, 150463 ], [ 152705, 152709 ], [ 153689, 153693 ], [ 154467, 154471 ], [ 155676, 155680 ], [ 156656, 156660 ], [ 157450, 157454 ], [ 158730, 158734 ], [ 160481, 160485 ], [ 161661, 161665 ], [ 162429, 162433 ], [ 163861, 163865 ], [ 164839, 164843 ], [ 165629, 165633 ], [ 166907, 166911 ], [ 168391, 168395 ], [ 169866, 169870 ], [ 171116, 171120 ], [ 171206, 171210 ], [ 172552, 172556 ], [ 172642, 172646 ], [ 174201, 174205 ], [ 176068, 176072 ], [ 176826, 176830 ], [ 178946, 178950 ], [ 179895, 179899 ], [ 180662, 180666 ], [ 181903, 181907 ], [ 183606, 183610 ], [ 184575, 184579 ], [ 185348, 185352 ], [ 186543, 186547 ], [ 187536, 187540 ], [ 188340, 188344 ], [ 190099, 190103 ], [ 192335, 192339 ], [ 193500, 193504 ], [ 194253, 194257 ], [ 195640, 195644 ], [ 196584, 196588 ], [ 197325, 197329 ], [ 198559, 198563 ], [ 200936, 200940 ], [ 202508, 202512 ], [ 203405, 203409 ], [ 204816, 204820 ], [ 204921, 204925 ], [ 205702, 205706 ], [ 205767, 205771 ], [ 206221, 206225 ], [ 206885, 206889 ], [ 207696, 207700 ], [ 207775, 207779 ], [ 208932, 208936 ], [ 209042, 209046 ], [ 209754, 209758 ], [ 209820, 209824 ], [ 209847, 209851 ], [ 212216, 212220 ], [ 212327, 212331 ], [ 212454, 212458 ], [ 212562, 212566 ], [ 212769, 212773 ], [ 214302, 214306 ], [ 214323, 214327 ], [ 214414, 214418 ], [ 214468, 214472 ], [ 214535, 214539 ], [ 214666, 214670 ], [ 214767, 214771 ], [ 225664, 225668 ], [ 225958, 225962 ], [ 226509, 226513 ], [ 226993, 226997 ], [ 227494, 227498 ], [ 227966, 227970 ], [ 228108, 228112 ], [ 228275, 228279 ] ], [ [ 630, 634 ], [ 210991, 210995 ], [ 211032, 211036 ], [ 211486, 211490 ] ], [ [ 665, 668 ], [ 211446, 211449 ] ], [ [ 676, 680 ] ], [ [ 688, 694 ], [ 3245, 3251 ], [ 3872, 3878 ], [ 4802, 4808 ], [ 5957, 5963 ], [ 10616, 10622 ], [ 15728, 15734 ], [ 19231, 19237 ], [ 20305, 20311 ], [ 22126, 22132 ], [ 24212, 24218 ], [ 25720, 25726 ], [ 26867, 26873 ], [ 29291, 29297 ], [ 30393, 30399 ], [ 34162, 34168 ], [ 35888, 35894 ], [ 37527, 37533 ], [ 39827, 39833 ], [ 41403, 41409 ], [ 42517, 42523 ], [ 44875, 44881 ], [ 45937, 45943 ], [ 46376, 46382 ], [ 48238, 48244 ], [ 49489, 49495 ], [ 50644, 50650 ], [ 53663, 53669 ], [ 55200, 55206 ], [ 55807, 55813 ], [ 57669, 57675 ], [ 58920, 58926 ], [ 60161, 60167 ], [ 63132, 63138 ], [ 64560, 64566 ], [ 65142, 65148 ], [ 67004, 67010 ], [ 68255, 68261 ], [ 69402, 69408 ], [ 71820, 71826 ], [ 72923, 72929 ], [ 73371, 73377 ], [ 75307, 75313 ], [ 76723, 76729 ], [ 77781, 77787 ], [ 80049, 80055 ], [ 81084, 81090 ], [ 84221, 84227 ], [ 85639, 85645 ], [ 86980, 86986 ], [ 89223, 89229 ], [ 90887, 90893 ], [ 91912, 91918 ], [ 94092, 94098 ], [ 95065, 95071 ], [ 95472, 95478 ], [ 97204, 97210 ], [ 98383, 98389 ], [ 99459, 99465 ], [ 102213, 102219 ], [ 103617, 103623 ], [ 104149, 104155 ], [ 105881, 105887 ], [ 107060, 107066 ], [ 108156, 108162 ], [ 110862, 110868 ], [ 112201, 112207 ], [ 112730, 112736 ], [ 114462, 114468 ], [ 115641, 115647 ], [ 116709, 116715 ], [ 118991, 118997 ], [ 120047, 120053 ], [ 120463, 120469 ], [ 122410, 122416 ], [ 123833, 123839 ], [ 124898, 124904 ], [ 127178, 127184 ], [ 128218, 128224 ], [ 131393, 131399 ], [ 132829, 132835 ], [ 134187, 134193 ], [ 136739, 136745 ], [ 138782, 138788 ], [ 139814, 139820 ], [ 142028, 142034 ], [ 143028, 143034 ], [ 143438, 143444 ], [ 145179, 145185 ], [ 146363, 146369 ], [ 147445, 147451 ], [ 150221, 150227 ], [ 151636, 151642 ], [ 152175, 152181 ], [ 153936, 153942 ], [ 155136, 155142 ], [ 156207, 156213 ], [ 158491, 158497 ], [ 159533, 159539 ], [ 159955, 159961 ], [ 161902, 161908 ], [ 163325, 163331 ], [ 164390, 164396 ], [ 166670, 166676 ], [ 167710, 167716 ], [ 170885, 170891 ], [ 172321, 172327 ], [ 173679, 173685 ], [ 176303, 176309 ], [ 178422, 178428 ], [ 179454, 179460 ], [ 181668, 181674 ], [ 182668, 182674 ], [ 183078, 183084 ], [ 184819, 184825 ], [ 186003, 186009 ], [ 187085, 187091 ], [ 189861, 189867 ], [ 191276, 191282 ], [ 191815, 191821 ], [ 193732, 193738 ], [ 195120, 195126 ], [ 196145, 196151 ], [ 198325, 198331 ], [ 199298, 199304 ], [ 202263, 202269 ], [ 206577, 206583 ], [ 207295, 207301 ], [ 208596, 208602 ], [ 211983, 211989 ], [ 214073, 214079 ], [ 226312, 226318 ], [ 227676, 227682 ], [ 9554, 9560 ], [ 9850, 9856 ], [ 29131, 29137 ], [ 30676, 30682 ], [ 44723, 44729 ], [ 46218, 46224 ], [ 53346, 53352 ], [ 55484, 55490 ], [ 62840, 62846 ], [ 64844, 64850 ], [ 71662, 71668 ], [ 73207, 73213 ], [ 79910, 79916 ], [ 81356, 81362 ], [ 93961, 93967 ], [ 95335, 95341 ], [ 101960, 101966 ], [ 103890, 103896 ], [ 110612, 110618 ], [ 112474, 112480 ], [ 118854, 118860 ], [ 120320, 120326 ], [ 127037, 127043 ], [ 128491, 128497 ], [ 141895, 141901 ], [ 143299, 143305 ], [ 149962, 149968 ], [ 151910, 151916 ], [ 158350, 158356 ], [ 159808, 159814 ], [ 166529, 166535 ], [ 167983, 167989 ], [ 181535, 181541 ], [ 182939, 182945 ], [ 189602, 189608 ], [ 191550, 191556 ], [ 198194, 198200 ], [ 199568, 199574 ], [ 199941, 199947 ], [ 200303, 200309 ], [ 200752, 200758 ], [ 200989, 200995 ], [ 201341, 201347 ], [ 203085, 203091 ], [ 204343, 204349 ], [ 204446, 204452 ], [ 204610, 204616 ], [ 212714, 212720 ], [ 214797, 214803 ] ], [ [ 733, 745 ] ], [ [ 747, 760 ] ], [ [ 791, 805 ], [ 7208, 7222 ], [ 10002, 10016 ], [ 12092, 12106 ], [ 16371, 16385 ], [ 16966, 16980 ], [ 19698, 19712 ], [ 20791, 20805 ], [ 200680, 200694 ], [ 210423, 210437 ], [ 210750, 210764 ], [ 228195, 228209 ] ], [ [ 834, 863 ], [ 203099, 203114 ] ], [ [ 892, 898 ], [ 47547, 47553 ], [ 49471, 49477 ], [ 56978, 56984 ], [ 58902, 58908 ], [ 66313, 66319 ], [ 68237, 68243 ], [ 96570, 96576 ], [ 98365, 98371 ], [ 105247, 105253 ], [ 107042, 107048 ], [ 113828, 113834 ], [ 115623, 115629 ], [ 144541, 144547 ], [ 146345, 146351 ], [ 184181, 184187 ], [ 185985, 185991 ] ], [ [ 927, 935 ], [ 32274, 32282 ], [ 82649, 82657 ], [ 129802, 129810 ], [ 169294, 169302 ], [ 225620, 225628 ] ], [ [ 964, 976 ], [ 207408, 207420 ] ], [ [ 1005, 1023 ], [ 207476, 207494 ], [ 24954, 24972 ], [ 27464, 27482 ], [ 29803, 29821 ], [ 40563, 40581 ], [ 43104, 43122 ], [ 45370, 45388 ], [ 48983, 49001 ], [ 51244, 51262 ], [ 54179, 54197 ], [ 58414, 58432 ], [ 60793, 60811 ], [ 63648, 63666 ], [ 67749, 67767 ], [ 69998, 70016 ], [ 72336, 72354 ], [ 75998, 76016 ], [ 78332, 78350 ], [ 80515, 80533 ], [ 89908, 89926 ], [ 92453, 92471 ], [ 94541, 94559 ], [ 97898, 97916 ], [ 100013, 100031 ], [ 102693, 102711 ], [ 106575, 106593 ], [ 108720, 108738 ], [ 111342, 111360 ], [ 115156, 115174 ], [ 117259, 117277 ], [ 119471, 119489 ], [ 123104, 123122 ], [ 125452, 125470 ], [ 127647, 127665 ], [ 137427, 137445 ], [ 140358, 140376 ], [ 142480, 142498 ], [ 145876, 145894 ], [ 148002, 148020 ], [ 150703, 150721 ], [ 154636, 154654 ], [ 156763, 156781 ], [ 158966, 158984 ], [ 162596, 162614 ], [ 164944, 164962 ], [ 167139, 167157 ], [ 176991, 177009 ], [ 179998, 180016 ], [ 182120, 182138 ], [ 185516, 185534 ], [ 187642, 187660 ], [ 190343, 190361 ], [ 194417, 194435 ], [ 196686, 196704 ], [ 198774, 198792 ] ], [ [ 1052, 1067 ] ], [ [ 1112, 1125 ], [ 215926, 215939 ], [ 216427, 216440 ] ], [ [ 1154, 1167 ] ], [ [ 1192, 1221 ], [ 3413, 3427 ], [ 4970, 4984 ], [ 6788, 6802 ], [ 22349, 22363 ], [ 23784, 23798 ], [ 24435, 24449 ], [ 25957, 25971 ], [ 26997, 27011 ], [ 28057, 28071 ], [ 28980, 28994 ], [ 29417, 29431 ], [ 30525, 30539 ], [ 30963, 30977 ], [ 32725, 32739 ], [ 34284, 34298 ], [ 36010, 36024 ], [ 37746, 37760 ], [ 39405, 39419 ], [ 40046, 40060 ], [ 41634, 41648 ], [ 42645, 42659 ], [ 43676, 43690 ], [ 44572, 44586 ], [ 44999, 45013 ], [ 46067, 46081 ], [ 46601, 46615 ], [ 47807, 47821 ], [ 48463, 48477 ], [ 49729, 49743 ], [ 50775, 50789 ], [ 51841, 51855 ], [ 53195, 53209 ], [ 53790, 53804 ], [ 55333, 55347 ], [ 56032, 56046 ], [ 57238, 57252 ], [ 57894, 57908 ], [ 59160, 59174 ], [ 60292, 60306 ], [ 61444, 61458 ], [ 62689, 62703 ], [ 63259, 63273 ], [ 64693, 64707 ], [ 65367, 65381 ], [ 66573, 66587 ], [ 67229, 67243 ], [ 68495, 68509 ], [ 69533, 69547 ], [ 70591, 70605 ], [ 71511, 71525 ], [ 71947, 71961 ], [ 73056, 73070 ], [ 73567, 73581 ], [ 74925, 74939 ], [ 75504, 75518 ], [ 76927, 76941 ], [ 77900, 77914 ], [ 78893, 78907 ], [ 79759, 79773 ], [ 80164, 80178 ], [ 81205, 81219 ], [ 81611, 81625 ], [ 83068, 83082 ], [ 84332, 84346 ], [ 85750, 85764 ], [ 87172, 87186 ], [ 88847, 88861 ], [ 89416, 89430 ], [ 91085, 91099 ], [ 92029, 92043 ], [ 92971, 92985 ], [ 93810, 93824 ], [ 94205, 94219 ], [ 95184, 95198 ], [ 95670, 95684 ], [ 96819, 96833 ], [ 97403, 97417 ], [ 98590, 98604 ], [ 99579, 99593 ], [ 100588, 100602 ], [ 101809, 101823 ], [ 102329, 102343 ], [ 103739, 103753 ], [ 104347, 104361 ], [ 105496, 105510 ], [ 106080, 106094 ], [ 107267, 107281 ], [ 108276, 108290 ], [ 109305, 109319 ], [ 110461, 110475 ], [ 110978, 110992 ], [ 112323, 112337 ], [ 112928, 112942 ], [ 114077, 114091 ], [ 114661, 114675 ], [ 115848, 115862 ], [ 116829, 116843 ], [ 117830, 117844 ], [ 118703, 118717 ], [ 119107, 119121 ], [ 120169, 120183 ], [ 120661, 120675 ], [ 122025, 122039 ], [ 122609, 122623 ], [ 124040, 124054 ], [ 125018, 125032 ], [ 126016, 126030 ], [ 126886, 126900 ], [ 127294, 127308 ], [ 128340, 128354 ], [ 128749, 128763 ], [ 130224, 130238 ], [ 131505, 131519 ], [ 132941, 132955 ], [ 134381, 134395 ], [ 136360, 136374 ], [ 136934, 136948 ], [ 138983, 138997 ], [ 139932, 139946 ], [ 140901, 140915 ], [ 141744, 141758 ], [ 142142, 142156 ], [ 143148, 143162 ], [ 143638, 143652 ], [ 144791, 144805 ], [ 145380, 145394 ], [ 146573, 146587 ], [ 147566, 147580 ], [ 148579, 148593 ], [ 149811, 149825 ], [ 150338, 150352 ], [ 151759, 151773 ], [ 152377, 152391 ], [ 153545, 153559 ], [ 154139, 154153 ], [ 155349, 155363 ], [ 156329, 156343 ], [ 157329, 157343 ], [ 158199, 158213 ], [ 158609, 158623 ], [ 159657, 159671 ], [ 160153, 160167 ], [ 161517, 161531 ], [ 162101, 162115 ], [ 163532, 163546 ], [ 164510, 164524 ], [ 165508, 165522 ], [ 166378, 166392 ], [ 166786, 166800 ], [ 167832, 167846 ], [ 168241, 168255 ], [ 169716, 169730 ], [ 170997, 171011 ], [ 172433, 172447 ], [ 173873, 173887 ], [ 175924, 175938 ], [ 176498, 176512 ], [ 178623, 178637 ], [ 179572, 179586 ], [ 180541, 180555 ], [ 181384, 181398 ], [ 181782, 181796 ], [ 182788, 182802 ], [ 183278, 183292 ], [ 184431, 184445 ], [ 185020, 185034 ], [ 186213, 186227 ], [ 187206, 187220 ], [ 188219, 188233 ], [ 189451, 189465 ], [ 189978, 189992 ], [ 191399, 191413 ], [ 192007, 192021 ], [ 193356, 193370 ], [ 193925, 193939 ], [ 195318, 195332 ], [ 196262, 196276 ], [ 197204, 197218 ], [ 198043, 198057 ], [ 198438, 198452 ], [ 199417, 199431 ], [ 199887, 199901 ], [ 200028, 200042 ], [ 200249, 200263 ], [ 200621, 200635 ], [ 201100, 201114 ], [ 201287, 201301 ], [ 201661, 201675 ], [ 201977, 201991 ], [ 202159, 202173 ], [ 202586, 202600 ], [ 202827, 202841 ], [ 203218, 203232 ], [ 203648, 203662 ], [ 205136, 205150 ], [ 205975, 205989 ], [ 206314, 206328 ], [ 206978, 206992 ], [ 207889, 207903 ], [ 208827, 208841 ], [ 210361, 210375 ], [ 210688, 210702 ], [ 212659, 212673 ], [ 215722, 215736 ], [ 216215, 216229 ], [ 225824, 225838 ], [ 226185, 226199 ], [ 226429, 226443 ], [ 226911, 226925 ], [ 227362, 227376 ] ], [ [ 1257, 1278 ], [ 9568, 9589 ] ], [ [ 1336, 1358 ], [ 3305, 3327 ], [ 4862, 4884 ], [ 6135, 6157 ], [ 6395, 6417 ], [ 6471, 6493 ], [ 10822, 10844 ], [ 11066, 11088 ], [ 11347, 11369 ], [ 11423, 11445 ], [ 15788, 15810 ], [ 15945, 15967 ], [ 16021, 16043 ], [ 19409, 19431 ], [ 20483, 20505 ], [ 227818, 227840 ], [ 24391, 24413 ], [ 26962, 26984 ], [ 29382, 29404 ], [ 30490, 30512 ], [ 34249, 34271 ], [ 35975, 35997 ], [ 40002, 40024 ], [ 42610, 42632 ], [ 44964, 44986 ], [ 46032, 46054 ], [ 48419, 48441 ], [ 50740, 50762 ], [ 53755, 53777 ], [ 55298, 55320 ], [ 57850, 57872 ], [ 60257, 60279 ], [ 63224, 63246 ], [ 64658, 64680 ], [ 67185, 67207 ], [ 69498, 69520 ], [ 71912, 71934 ], [ 73021, 73043 ], [ 75460, 75482 ], [ 77865, 77887 ], [ 80129, 80151 ], [ 81170, 81192 ], [ 84297, 84319 ], [ 85715, 85737 ], [ 89372, 89394 ], [ 91994, 92016 ], [ 94170, 94192 ], [ 95149, 95171 ], [ 97359, 97381 ], [ 99544, 99566 ], [ 102294, 102316 ], [ 103704, 103726 ], [ 106036, 106058 ], [ 108241, 108263 ], [ 110943, 110965 ], [ 112288, 112310 ], [ 114617, 114639 ], [ 116794, 116816 ], [ 119072, 119094 ], [ 120134, 120156 ], [ 122565, 122587 ], [ 124983, 125005 ], [ 127259, 127281 ], [ 128305, 128327 ], [ 131470, 131492 ], [ 132906, 132928 ], [ 136890, 136912 ], [ 139897, 139919 ], [ 142107, 142129 ], [ 143113, 143135 ], [ 145336, 145358 ], [ 147531, 147553 ], [ 150303, 150325 ], [ 151724, 151746 ], [ 154095, 154117 ], [ 156294, 156316 ], [ 158574, 158596 ], [ 159622, 159644 ], [ 162057, 162079 ], [ 164475, 164497 ], [ 166751, 166773 ], [ 167797, 167819 ], [ 170962, 170984 ], [ 172398, 172420 ], [ 176454, 176476 ], [ 179537, 179559 ], [ 181747, 181769 ], [ 182753, 182775 ], [ 184976, 184998 ], [ 187171, 187193 ], [ 189943, 189965 ], [ 191364, 191386 ], [ 193881, 193903 ], [ 196227, 196249 ], [ 198403, 198425 ], [ 199382, 199404 ], [ 216171, 216193 ], [ 226385, 226407 ] ], [ [ 1416, 1433 ], [ 3286, 3303 ], [ 4843, 4860 ], [ 6048, 6065 ], [ 6287, 6304 ], [ 6350, 6367 ], [ 10727, 10744 ], [ 10970, 10987 ], [ 11239, 11256 ], [ 11302, 11319 ], [ 15769, 15786 ], [ 15837, 15854 ], [ 15900, 15917 ], [ 19322, 19339 ], [ 20396, 20413 ], [ 227752, 227769 ], [ 2617, 2634 ], [ 2697, 2714 ], [ 2804, 2821 ], [ 2924, 2941 ], [ 3047, 3064 ], [ 3165, 3182 ], [ 5666, 5683 ], [ 5848, 5865 ], [ 6690, 6707 ], [ 7012, 7029 ], [ 21685, 21702 ], [ 22310, 22327 ], [ 23745, 23762 ], [ 25927, 25944 ], [ 28027, 28044 ], [ 28950, 28967 ], [ 30924, 30941 ], [ 32686, 32703 ], [ 37707, 37724 ], [ 39366, 39383 ], [ 41604, 41621 ], [ 43646, 43663 ], [ 44542, 44559 ], [ 46562, 46579 ], [ 47768, 47785 ], [ 49699, 49716 ], [ 51811, 51828 ], [ 53165, 53182 ], [ 55993, 56010 ], [ 57199, 57216 ], [ 59130, 59147 ], [ 61414, 61431 ], [ 62659, 62676 ], [ 65328, 65345 ], [ 66534, 66551 ], [ 68465, 68482 ], [ 70561, 70578 ], [ 71481, 71498 ], [ 73528, 73545 ], [ 74886, 74903 ], [ 76897, 76914 ], [ 78863, 78880 ], [ 79729, 79746 ], [ 81572, 81589 ], [ 83029, 83046 ], [ 87133, 87150 ], [ 88808, 88825 ], [ 91055, 91072 ], [ 92941, 92958 ], [ 93780, 93797 ], [ 95631, 95648 ], [ 96780, 96797 ], [ 98560, 98577 ], [ 100558, 100575 ], [ 101779, 101796 ], [ 104308, 104325 ], [ 105457, 105474 ], [ 107237, 107254 ], [ 109275, 109292 ], [ 110431, 110448 ], [ 112889, 112906 ], [ 114038, 114055 ], [ 115818, 115835 ], [ 117800, 117817 ], [ 118673, 118690 ], [ 120622, 120639 ], [ 121986, 122003 ], [ 124010, 124027 ], [ 125986, 126003 ], [ 126856, 126873 ], [ 128710, 128727 ], [ 130185, 130202 ], [ 134342, 134359 ], [ 136321, 136338 ], [ 138953, 138970 ], [ 140871, 140888 ], [ 141714, 141731 ], [ 143599, 143616 ], [ 144752, 144769 ], [ 146543, 146560 ], [ 148549, 148566 ], [ 149781, 149798 ], [ 152338, 152355 ], [ 153506, 153523 ], [ 155319, 155336 ], [ 157299, 157316 ], [ 158169, 158186 ], [ 160114, 160131 ], [ 161478, 161495 ], [ 163502, 163519 ], [ 165478, 165495 ], [ 166348, 166365 ], [ 168202, 168219 ], [ 169677, 169694 ], [ 173834, 173851 ], [ 175885, 175902 ], [ 178593, 178610 ], [ 180511, 180528 ], [ 181354, 181371 ], [ 183239, 183256 ], [ 184392, 184409 ], [ 186183, 186200 ], [ 188189, 188206 ], [ 189421, 189438 ], [ 191968, 191985 ], [ 193317, 193334 ], [ 195288, 195305 ], [ 197174, 197191 ], [ 198013, 198030 ], [ 199985, 200002 ], [ 200347, 200364 ], [ 200796, 200813 ], [ 201033, 201050 ], [ 201385, 201402 ], [ 201719, 201736 ], [ 202797, 202814 ], [ 206367, 206384 ], [ 210322, 210339 ], [ 210649, 210666 ], [ 215683, 215700 ], [ 216860, 216877 ], [ 217101, 217118 ], [ 217210, 217227 ], [ 217546, 217563 ], [ 217794, 217811 ], [ 217905, 217922 ], [ 218183, 218200 ], [ 218387, 218404 ], [ 218497, 218514 ], [ 218897, 218914 ], [ 219197, 219214 ], [ 219318, 219335 ], [ 219793, 219810 ], [ 220113, 220130 ], [ 220224, 220241 ], [ 220764, 220781 ], [ 221116, 221133 ], [ 221227, 221244 ], [ 221563, 221580 ], [ 221805, 221822 ], [ 221915, 221932 ], [ 222272, 222289 ], [ 222530, 222547 ], [ 222645, 222662 ], [ 222914, 222931 ], [ 223130, 223147 ], [ 223256, 223273 ], [ 223473, 223490 ], [ 223649, 223666 ], [ 223766, 223783 ], [ 224006, 224023 ], [ 224208, 224225 ], [ 224331, 224348 ], [ 224548, 224565 ], [ 224729, 224746 ], [ 224847, 224864 ], [ 225134, 225151 ], [ 225356, 225373 ], [ 225475, 225492 ], [ 225781, 225798 ], [ 226084, 226101 ], [ 226868, 226885 ], [ 227319, 227336 ] ], [ [ 1491, 1497 ], [ 23385, 23391 ], [ 25557, 25563 ], [ 74547, 74553 ], [ 76570, 76576 ], [ 121645, 121651 ], [ 123679, 123685 ], [ 161137, 161143 ], [ 163171, 163177 ] ], [ [ 1555, 1565 ], [ 3950, 3960 ], [ 4005, 4015 ], [ 6067, 6077 ], [ 6171, 6181 ], [ 10746, 10756 ], [ 10858, 10868 ], [ 10989, 10999 ], [ 11102, 11112 ], [ 19341, 19351 ], [ 19445, 19455 ], [ 20415, 20425 ], [ 20519, 20529 ], [ 202316, 202326 ], [ 202353, 202363 ], [ 206630, 206640 ], [ 206667, 206677 ], [ 207371, 207381 ], [ 207432, 207442 ], [ 208649, 208659 ], [ 208686, 208696 ], [ 212036, 212046 ], [ 212073, 212083 ], [ 214126, 214136 ], [ 214163, 214173 ], [ 227771, 227781 ], [ 227842, 227852 ], [ 5743, 5753 ], [ 5918, 5928 ], [ 7727, 7737 ], [ 8479, 8489 ], [ 9246, 9256 ], [ 10526, 10536 ], [ 13019, 13029 ], [ 14641, 14651 ], [ 15622, 15632 ], [ 20215, 20225 ], [ 21333, 21343 ], [ 22036, 22046 ], [ 199830, 199840 ], [ 200192, 200202 ], [ 200564, 200574 ], [ 201230, 201240 ], [ 201604, 201614 ], [ 201920, 201930 ], [ 202095, 202105 ], [ 202907, 202917 ], [ 203161, 203171 ], [ 203591, 203601 ], [ 205201, 205211 ], [ 206040, 206050 ], [ 211129, 211139 ], [ 211583, 211593 ], [ 225691, 225701 ], [ 225985, 225995 ], [ 228747, 228757 ] ], [ [ 1607, 1619 ], [ 22172, 22184 ], [ 24332, 24344 ], [ 37573, 37585 ], [ 39945, 39957 ], [ 46422, 46434 ], [ 48359, 48371 ], [ 55853, 55865 ], [ 57790, 57802 ], [ 65188, 65200 ], [ 67125, 67137 ], [ 73412, 73424 ], [ 75412, 75424 ], [ 87021, 87033 ], [ 89326, 89338 ], [ 95513, 95525 ], [ 97310, 97322 ], [ 104190, 104202 ], [ 105987, 105999 ], [ 112771, 112783 ], [ 114568, 114580 ], [ 120504, 120516 ], [ 122516, 122528 ], [ 134228, 134240 ], [ 136843, 136855 ], [ 143479, 143491 ], [ 145286, 145298 ], [ 152216, 152228 ], [ 154044, 154056 ], [ 159996, 160008 ], [ 162008, 162020 ], [ 173720, 173732 ], [ 176407, 176419 ], [ 183119, 183131 ], [ 184926, 184938 ], [ 191856, 191868 ], [ 193835, 193847 ], [ 22868, 22880 ], [ 23246, 23258 ], [ 24165, 24177 ], [ 25005, 25017 ], [ 25418, 25430 ], [ 26133, 26145 ], [ 26424, 26436 ], [ 27173, 27185 ], [ 27515, 27527 ], [ 28369, 28381 ], [ 29210, 29222 ], [ 29729, 29741 ], [ 29854, 29866 ], [ 30761, 30773 ], [ 31304, 31316 ], [ 31405, 31417 ], [ 31458, 31470 ], [ 31511, 31523 ], [ 31629, 31641 ], [ 31762, 31774 ], [ 31842, 31854 ], [ 31943, 31955 ], [ 32044, 32056 ], [ 32097, 32109 ], [ 32552, 32564 ], [ 33066, 33078 ], [ 33167, 33179 ], [ 33220, 33232 ], [ 33273, 33285 ], [ 33391, 33403 ], [ 33524, 33536 ], [ 33604, 33616 ], [ 33705, 33717 ], [ 33806, 33818 ], [ 33859, 33871 ], [ 34640, 34652 ], [ 34741, 34753 ], [ 34794, 34806 ], [ 34847, 34859 ], [ 34965, 34977 ], [ 35098, 35110 ], [ 35178, 35190 ], [ 35279, 35291 ], [ 35380, 35392 ], [ 35433, 35445 ], [ 35834, 35846 ], [ 36366, 36378 ], [ 36467, 36479 ], [ 36520, 36532 ], [ 36573, 36585 ], [ 36691, 36703 ], [ 36824, 36836 ], [ 36904, 36916 ], [ 37005, 37017 ], [ 37106, 37118 ], [ 37159, 37171 ], [ 38263, 38275 ], [ 38458, 38470 ], [ 38511, 38523 ], [ 38819, 38831 ], [ 38956, 38968 ], [ 39782, 39794 ], [ 40614, 40626 ], [ 41056, 41068 ], [ 41193, 41205 ], [ 41810, 41822 ], [ 42093, 42105 ], [ 42821, 42833 ], [ 43155, 43167 ], [ 43986, 43998 ], [ 44800, 44812 ], [ 45309, 45321 ], [ 45421, 45433 ], [ 46301, 46313 ], [ 47407, 47419 ], [ 48190, 48202 ], [ 49331, 49343 ], [ 49905, 49917 ], [ 50951, 50963 ], [ 52432, 52444 ], [ 52895, 52907 ], [ 53426, 53438 ], [ 53537, 53549 ], [ 54537, 54549 ], [ 54995, 55007 ], [ 55570, 55582 ], [ 55681, 55693 ], [ 56838, 56850 ], [ 57621, 57633 ], [ 58762, 58774 ], [ 59336, 59348 ], [ 60468, 60480 ], [ 62000, 62012 ], [ 62379, 62391 ], [ 62920, 62932 ], [ 62996, 63008 ], [ 63971, 63983 ], [ 64345, 64357 ], [ 64930, 64942 ], [ 65006, 65018 ], [ 66173, 66185 ], [ 66956, 66968 ], [ 68097, 68109 ], [ 68671, 68683 ], [ 69709, 69721 ], [ 71742, 71754 ], [ 73293, 73305 ], [ 74061, 74073 ], [ 74418, 74430 ], [ 75270, 75282 ], [ 76049, 76061 ], [ 76441, 76453 ], [ 77103, 77115 ], [ 77359, 77371 ], [ 78076, 78088 ], [ 78383, 78395 ], [ 79180, 79192 ], [ 79978, 79990 ], [ 80451, 80463 ], [ 80566, 80578 ], [ 81430, 81442 ], [ 81927, 81939 ], [ 82007, 82019 ], [ 82050, 82062 ], [ 82093, 82105 ], [ 82201, 82213 ], [ 82283, 82295 ], [ 82342, 82354 ], [ 82417, 82429 ], [ 82476, 82488 ], [ 82499, 82511 ], [ 82916, 82928 ], [ 83384, 83396 ], [ 83464, 83476 ], [ 83507, 83519 ], [ 83550, 83562 ], [ 83658, 83670 ], [ 83740, 83752 ], [ 83799, 83811 ], [ 83874, 83886 ], [ 83933, 83945 ], [ 83956, 83968 ], [ 84660, 84672 ], [ 84740, 84752 ], [ 84783, 84795 ], [ 84826, 84838 ], [ 84934, 84946 ], [ 85016, 85028 ], [ 85075, 85087 ], [ 85150, 85162 ], [ 85209, 85221 ], [ 85232, 85244 ], [ 85595, 85607 ], [ 86078, 86090 ], [ 86158, 86170 ], [ 86201, 86213 ], [ 86244, 86256 ], [ 86352, 86364 ], [ 86434, 86446 ], [ 86493, 86505 ], [ 86568, 86580 ], [ 86627, 86639 ], [ 86650, 86662 ], [ 87664, 87676 ], [ 87913, 87925 ], [ 88196, 88208 ], [ 88323, 88335 ], [ 89188, 89200 ], [ 89959, 89971 ], [ 90464, 90476 ], [ 90591, 90603 ], [ 91261, 91273 ], [ 91509, 91521 ], [ 92205, 92217 ], [ 92491, 92503 ], [ 93256, 93268 ], [ 94027, 94039 ], [ 94490, 94502 ], [ 94579, 94591 ], [ 95407, 95419 ], [ 96440, 96452 ], [ 97166, 97178 ], [ 98235, 98247 ], [ 98766, 98778 ], [ 99755, 99767 ], [ 101132, 101144 ], [ 101552, 101564 ], [ 102029, 102041 ], [ 102119, 102131 ], [ 103029, 103041 ], [ 103444, 103456 ], [ 103965, 103977 ], [ 104055, 104067 ], [ 105117, 105129 ], [ 105843, 105855 ], [ 106912, 106924 ], [ 107443, 107455 ], [ 108452, 108464 ], [ 109814, 109826 ], [ 110172, 110184 ], [ 110681, 110693 ], [ 110736, 110748 ], [ 111643, 111655 ], [ 111996, 112008 ], [ 112549, 112561 ], [ 112604, 112616 ], [ 113698, 113710 ], [ 114424, 114436 ], [ 115493, 115505 ], [ 116024, 116036 ], [ 117005, 117017 ], [ 118923, 118935 ], [ 120395, 120407 ], [ 121156, 121168 ], [ 121515, 121527 ], [ 122372, 122384 ], [ 123155, 123167 ], [ 123549, 123561 ], [ 124216, 124228 ], [ 124474, 124486 ], [ 125194, 125206 ], [ 125503, 125515 ], [ 126304, 126316 ], [ 127106, 127118 ], [ 127582, 127594 ], [ 127698, 127710 ], [ 128566, 128578 ], [ 129066, 129078 ], [ 129148, 129160 ], [ 129192, 129204 ], [ 129236, 129248 ], [ 129345, 129357 ], [ 129429, 129441 ], [ 129490, 129502 ], [ 129566, 129578 ], [ 129627, 129639 ], [ 129651, 129663 ], [ 130070, 130082 ], [ 130541, 130553 ], [ 130623, 130635 ], [ 130667, 130679 ], [ 130711, 130723 ], [ 130820, 130832 ], [ 130904, 130916 ], [ 130965, 130977 ], [ 131041, 131053 ], [ 131102, 131114 ], [ 131126, 131138 ], [ 131834, 131846 ], [ 131916, 131928 ], [ 131960, 131972 ], [ 132004, 132016 ], [ 132113, 132125 ], [ 132197, 132209 ], [ 132258, 132270 ], [ 132334, 132346 ], [ 132395, 132407 ], [ 132419, 132431 ], [ 132784, 132796 ], [ 133270, 133282 ], [ 133352, 133364 ], [ 133396, 133408 ], [ 133440, 133452 ], [ 133549, 133561 ], [ 133633, 133645 ], [ 133694, 133706 ], [ 133770, 133782 ], [ 133831, 133843 ], [ 133855, 133867 ], [ 134874, 134886 ], [ 135150, 135162 ], [ 135514, 135526 ], [ 135642, 135654 ], [ 135979, 135991 ], [ 136703, 136715 ], [ 137478, 137490 ], [ 137778, 137790 ], [ 138165, 138177 ], [ 138293, 138305 ], [ 138630, 138642 ], [ 139159, 139171 ], [ 139409, 139421 ], [ 140108, 140120 ], [ 140409, 140421 ], [ 141187, 141199 ], [ 141962, 141974 ], [ 142428, 142440 ], [ 142531, 142543 ], [ 143372, 143384 ], [ 144410, 144422 ], [ 145140, 145152 ], [ 146214, 146226 ], [ 146749, 146761 ], [ 147742, 147754 ], [ 149126, 149138 ], [ 149550, 149562 ], [ 150032, 150044 ], [ 150124, 150136 ], [ 151041, 151053 ], [ 151460, 151472 ], [ 151986, 151998 ], [ 152078, 152090 ], [ 152874, 152886 ], [ 153140, 153152 ], [ 153272, 153284 ], [ 153896, 153908 ], [ 154687, 154699 ], [ 154964, 154976 ], [ 155096, 155108 ], [ 155525, 155537 ], [ 155783, 155795 ], [ 156505, 156517 ], [ 156814, 156826 ], [ 157619, 157631 ], [ 158421, 158433 ], [ 158899, 158911 ], [ 159017, 159029 ], [ 159885, 159897 ], [ 160648, 160660 ], [ 161007, 161019 ], [ 161864, 161876 ], [ 162647, 162659 ], [ 163041, 163053 ], [ 163708, 163720 ], [ 163966, 163978 ], [ 164686, 164698 ], [ 164995, 165007 ], [ 165796, 165808 ], [ 166598, 166610 ], [ 167074, 167086 ], [ 167190, 167202 ], [ 168058, 168070 ], [ 168558, 168570 ], [ 168640, 168652 ], [ 168684, 168696 ], [ 168728, 168740 ], [ 168837, 168849 ], [ 168921, 168933 ], [ 168982, 168994 ], [ 169058, 169070 ], [ 169119, 169131 ], [ 169143, 169155 ], [ 169562, 169574 ], [ 170033, 170045 ], [ 170115, 170127 ], [ 170159, 170171 ], [ 170203, 170215 ], [ 170312, 170324 ], [ 170396, 170408 ], [ 170457, 170469 ], [ 170533, 170545 ], [ 170594, 170606 ], [ 170618, 170630 ], [ 171326, 171338 ], [ 171408, 171420 ], [ 171452, 171464 ], [ 171496, 171508 ], [ 171605, 171617 ], [ 171689, 171701 ], [ 171750, 171762 ], [ 171826, 171838 ], [ 171887, 171899 ], [ 171911, 171923 ], [ 172276, 172288 ], [ 172762, 172774 ], [ 172844, 172856 ], [ 172888, 172900 ], [ 172932, 172944 ], [ 173041, 173053 ], [ 173125, 173137 ], [ 173186, 173198 ], [ 173262, 173274 ], [ 173323, 173335 ], [ 173347, 173359 ], [ 174366, 174378 ], [ 174561, 174573 ], [ 174630, 174642 ], [ 175037, 175049 ], [ 175165, 175177 ], [ 175411, 175423 ], [ 175490, 175502 ], [ 176267, 176279 ], [ 177042, 177054 ], [ 177257, 177269 ], [ 177330, 177342 ], [ 177764, 177776 ], [ 177892, 177904 ], [ 178138, 178150 ], [ 178217, 178229 ], [ 178799, 178811 ], [ 179049, 179061 ], [ 179748, 179760 ], [ 180049, 180061 ], [ 180827, 180839 ], [ 181602, 181614 ], [ 182068, 182080 ], [ 182171, 182183 ], [ 183012, 183024 ], [ 184050, 184062 ], [ 184780, 184792 ], [ 185854, 185866 ], [ 186389, 186401 ], [ 187382, 187394 ], [ 188766, 188778 ], [ 189190, 189202 ], [ 189672, 189684 ], [ 189764, 189776 ], [ 190681, 190693 ], [ 191100, 191112 ], [ 191626, 191638 ], [ 191718, 191730 ], [ 192499, 192511 ], [ 192855, 192867 ], [ 192982, 192994 ], [ 193697, 193709 ], [ 194468, 194480 ], [ 194847, 194859 ], [ 194974, 194986 ], [ 195494, 195506 ], [ 195742, 195754 ], [ 196438, 196450 ], [ 196724, 196736 ], [ 197489, 197501 ], [ 198260, 198272 ], [ 198723, 198735 ], [ 198812, 198824 ], [ 199640, 199652 ] ], [ [ 1651, 1665 ], [ 47121, 47135 ], [ 49034, 49048 ], [ 50198, 50212 ], [ 51295, 51309 ], [ 52154, 52168 ], [ 54103, 54117 ], [ 54230, 54244 ], [ 56552, 56566 ], [ 58465, 58479 ], [ 59661, 59675 ], [ 60844, 60858 ], [ 61757, 61771 ], [ 63572, 63586 ], [ 63699, 63713 ], [ 65887, 65901 ], [ 67800, 67814 ], [ 68960, 68974 ], [ 70049, 70063 ], [ 70904, 70918 ], [ 72260, 72274 ], [ 72387, 72401 ], [ 96165, 96179 ], [ 97949, 97963 ], [ 99024, 99038 ], [ 100064, 100078 ], [ 100876, 100890 ], [ 102617, 102631 ], [ 102744, 102758 ], [ 104842, 104856 ], [ 106626, 106640 ], [ 107711, 107725 ], [ 108771, 108785 ], [ 109593, 109607 ], [ 111266, 111280 ], [ 111393, 111407 ], [ 113423, 113437 ], [ 115207, 115221 ], [ 116278, 116292 ], [ 117310, 117324 ], [ 118118, 118132 ], [ 119395, 119409 ], [ 119522, 119536 ], [ 144134, 144148 ], [ 145927, 145941 ], [ 147009, 147023 ], [ 148053, 148067 ], [ 148868, 148882 ], [ 150627, 150641 ], [ 150754, 150768 ], [ 183774, 183788 ], [ 185567, 185581 ], [ 186649, 186663 ], [ 187693, 187707 ], [ 188508, 188522 ], [ 190267, 190281 ], [ 190394, 190408 ] ], [ [ 1692, 1707 ], [ 3487, 3502 ], [ 4211, 4226 ], [ 4316, 4331 ], [ 4528, 4543 ], [ 4633, 4648 ], [ 5044, 5059 ] ], [ [ 1736, 1750 ], [ 62074, 62088 ], [ 62512, 62526 ], [ 63070, 63084 ], [ 64045, 64059 ], [ 64478, 64492 ], [ 65080, 65094 ], [ 109878, 109892 ], [ 110295, 110309 ], [ 110800, 110814 ], [ 111707, 111721 ], [ 112119, 112133 ], [ 112668, 112682 ] ], [ [ 1795, 1808 ] ], [ [ 1832, 1843 ], [ 202526, 202532 ], [ 204843, 204849 ], [ 205720, 205726 ], [ 206239, 206245 ], [ 206903, 206909 ], [ 207723, 207729 ], [ 212787, 212793 ], [ 227993, 227999 ] ], [ [ 1850, 1877 ], [ 12153, 12180 ], [ 12606, 12633 ], [ 13734, 13761 ], [ 14109, 14136 ], [ 210015, 210042 ], [ 210100, 210127 ], [ 213079, 213106 ] ], [ [ 2136, 2159 ], [ 6326, 6349 ], [ 6447, 6470 ], [ 11278, 11301 ], [ 11399, 11422 ], [ 15876, 15899 ], [ 15997, 16020 ] ], [ [ 2312, 2343 ] ], [ [ 3335, 3385 ] ], [ [ 4079, 4134 ] ], [ [ 4892, 4942 ] ], [ [ 5602, 5646 ] ], [ [ 6502, 6541 ] ], [ [ 11525, 11563 ] ], [ [ 16052, 16110 ] ], [ [ 19541, 19587 ] ], [ [ 20615, 20671 ] ], [ [ 21426, 21475 ] ], [ [ 22228, 22256 ] ], [ [ 23546, 23585 ] ], [ [ 24232, 25716 ], [ 25808, 25842 ] ], [ [ 25740, 25861 ] ], [ [ 25868, 25910 ] ], [ [ 26887, 27965 ] ], [ [ 27972, 28010 ] ], [ [ 28889, 28933 ] ], [ [ 29311, 30389 ] ], [ [ 30413, 30838 ] ], [ [ 30845, 30879 ] ], [ [ 32607, 32641 ] ], [ [ 34182, 35884 ] ], [ [ 35908, 37523 ] ], [ [ 37627, 37653 ] ], [ [ 39169, 39206 ] ], [ [ 39847, 41399 ], [ 41489, 41521 ] ], [ [ 41423, 41540 ] ], [ [ 41547, 41587 ] ], [ [ 42537, 43586 ] ], [ [ 43593, 43629 ] ], [ [ 44483, 44525 ] ], [ [ 44895, 45933 ] ], [ [ 45957, 46372 ] ], [ [ 46479, 46508 ] ], [ [ 47568, 47608 ] ], [ [ 48258, 49485 ], [ 49578, 49613 ] ], [ [ 49509, 49632 ] ], [ [ 49639, 49682 ] ], [ [ 50664, 51748 ] ], [ [ 51755, 51794 ] ], [ [ 53103, 53148 ] ], [ [ 53683, 55196 ] ], [ [ 55220, 55803 ] ], [ [ 55910, 55939 ] ], [ [ 56999, 57039 ] ], [ [ 57689, 58916 ], [ 59009, 59044 ] ], [ [ 58940, 59063 ] ], [ [ 59070, 59113 ] ], [ [ 60181, 61351 ] ], [ [ 61358, 61397 ] ], [ [ 62597, 62642 ] ], [ [ 63152, 64556 ] ], [ [ 64580, 65138 ] ], [ [ 65245, 65274 ] ], [ [ 66334, 66374 ] ], [ [ 67024, 68251 ], [ 68344, 68379 ] ], [ [ 68275, 68398 ] ], [ [ 68405, 68448 ] ], [ [ 69422, 70498 ] ], [ [ 70505, 70544 ] ], [ [ 71419, 71464 ] ], [ [ 71840, 72919 ] ], [ [ 72943, 73367 ] ], [ [ 73457, 73474 ] ], [ [ 74698, 74726 ] ], [ [ 75327, 76719 ], [ 76800, 76823 ] ], [ [ 76743, 76842 ] ], [ [ 76849, 76880 ] ], [ [ 77801, 78812 ] ], [ [ 78819, 78846 ] ], [ [ 79679, 79712 ] ], [ [ 80069, 81080 ] ], [ [ 81104, 81497 ] ], [ [ 81504, 81527 ] ], [ [ 82961, 82984 ] ], [ [ 84241, 85635 ] ], [ [ 85659, 86976 ] ], [ [ 87064, 87079 ] ], [ [ 88622, 88648 ] ], [ [ 89243, 90883 ], [ 90962, 90983 ] ], [ [ 90907, 91002 ] ], [ [ 91009, 91038 ] ], [ [ 91932, 92892 ] ], [ [ 92899, 92924 ] ], [ [ 93732, 93763 ] ], [ [ 94112, 95061 ] ], [ [ 95085, 95468 ] ], [ [ 95559, 95577 ] ], [ [ 96591, 96620 ] ], [ [ 97224, 98379 ], [ 98461, 98485 ] ], [ [ 98403, 98504 ] ], [ [ 98511, 98543 ] ], [ [ 99479, 100506 ] ], [ [ 100513, 100541 ] ], [ [ 101728, 101762 ] ], [ [ 102233, 103613 ] ], [ [ 103637, 104145 ] ], [ [ 104236, 104254 ] ], [ [ 105268, 105297 ] ], [ [ 105901, 107056 ], [ 107138, 107162 ] ], [ [ 107080, 107181 ] ], [ [ 107188, 107220 ] ], [ [ 108176, 109223 ] ], [ [ 109230, 109258 ] ], [ [ 110380, 110414 ] ], [ [ 110882, 112197 ] ], [ [ 112221, 112726 ] ], [ [ 112817, 112835 ] ], [ [ 113849, 113878 ] ], [ [ 114482, 115637 ], [ 115719, 115743 ] ], [ [ 115661, 115762 ] ], [ [ 115769, 115801 ] ], [ [ 116729, 117748 ] ], [ [ 117755, 117783 ] ], [ [ 118622, 118656 ] ], [ [ 119011, 120043 ] ], [ [ 120067, 120459 ] ], [ [ 120550, 120568 ] ], [ [ 121797, 121826 ] ], [ [ 122430, 123829 ], [ 123911, 123935 ] ], [ [ 123853, 123954 ] ], [ [ 123961, 123993 ] ], [ [ 124918, 125934 ] ], [ [ 125941, 125969 ] ], [ [ 126805, 126839 ] ], [ [ 127198, 128214 ] ], [ [ 128238, 128634 ] ], [ [ 128641, 128665 ] ], [ [ 130116, 130140 ] ], [ [ 131413, 132825 ] ], [ [ 132849, 134183 ] ], [ [ 134272, 134288 ] ], [ [ 136134, 136161 ] ], [ [ 136759, 138778 ], [ 138858, 138880 ] ], [ [ 138802, 138899 ] ], [ [ 138906, 138936 ] ], [ [ 139834, 140821 ] ], [ [ 140828, 140854 ] ], [ [ 141665, 141697 ] ], [ [ 142048, 143024 ] ], [ [ 143048, 143434 ] ], [ [ 143526, 143545 ] ], [ [ 144562, 144592 ] ], [ [ 145199, 146359 ], [ 146442, 146467 ] ], [ [ 146383, 146486 ] ], [ [ 146493, 146526 ] ], [ [ 147465, 148496 ] ], [ [ 148503, 148532 ] ], [ [ 149729, 149764 ] ], [ [ 150241, 151632 ] ], [ [ 151656, 152171 ] ], [ [ 152264, 152284 ] ], [ [ 153315, 153346 ] ], [ [ 153956, 155132 ], [ 155216, 155242 ] ], [ [ 155156, 155261 ] ], [ [ 155268, 155302 ] ], [ [ 156227, 157245 ] ], [ [ 157252, 157282 ] ], [ [ 158116, 158152 ] ], [ [ 158511, 159529 ] ], [ [ 159553, 159951 ] ], [ [ 160042, 160060 ] ], [ [ 161289, 161318 ] ], [ [ 161922, 163321 ], [ 163403, 163427 ] ], [ [ 163345, 163446 ] ], [ [ 163453, 163485 ] ], [ [ 164410, 165426 ] ], [ [ 165433, 165461 ] ], [ [ 166297, 166331 ] ], [ [ 166690, 167706 ] ], [ [ 167730, 168126 ] ], [ [ 168133, 168157 ] ], [ [ 169608, 169632 ] ], [ [ 170905, 172317 ] ], [ [ 172341, 173675 ] ], [ [ 173764, 173780 ] ], [ [ 175698, 175725 ] ], [ [ 176323, 178418 ], [ 178498, 178520 ] ], [ [ 178442, 178539 ] ], [ [ 178546, 178576 ] ], [ [ 179474, 180461 ] ], [ [ 180468, 180494 ] ], [ [ 181305, 181337 ] ], [ [ 181688, 182664 ] ], [ [ 182688, 183074 ] ], [ [ 183166, 183185 ] ], [ [ 184202, 184232 ] ], [ [ 184839, 185999 ], [ 186082, 186107 ] ], [ [ 186023, 186126 ] ], [ [ 186133, 186166 ] ], [ [ 187105, 188136 ] ], [ [ 188143, 188172 ] ], [ [ 189369, 189404 ] ], [ [ 189881, 191272 ] ], [ [ 191296, 191811 ] ], [ [ 191899, 191914 ] ], [ [ 193131, 193157 ] ], [ [ 193752, 195116 ], [ 195195, 195216 ] ], [ [ 195140, 195235 ] ], [ [ 195242, 195271 ] ], [ [ 196165, 197125 ] ], [ [ 197132, 197157 ] ], [ [ 197965, 197996 ] ], [ [ 198345, 199294 ] ], [ [ 199318, 199701 ] ], [ [ 199708, 199740 ] ], [ [ 201492, 201515 ] ], [ [ 201805, 201831 ] ], [ [ 202405, 202423 ] ], [ [ 202696, 202723 ] ], [ [ 202951, 202989 ] ], [ [ 203322, 203354 ] ], [ [ 204688, 204742 ] ], [ [ 205558, 205599 ] ], [ [ 206113, 206139 ] ], [ [ 206719, 206755 ] ], [ [ 207510, 207552 ] ], [ [ 208737, 208797 ] ], [ [ 210275, 210305 ] ], [ [ 210600, 210632 ] ], [ [ 210933, 210973 ] ], [ [ 211380, 211428 ] ], [ [ 212124, 212188 ] ], [ [ 214214, 214263 ] ], [ [ 215633, 215666 ] ], [ [ 216115, 216154 ] ], [ [ 216621, 216636 ] ], [ [ 216951, 216972 ] ], [ [ 217287, 217304 ] ], [ [ 217641, 217664 ] ], [ [ 217984, 218000 ] ], [ [ 218268, 218290 ] ], [ [ 218575, 218602 ] ], [ [ 219026, 219059 ] ], [ [ 219407, 219424 ] ], [ [ 219921, 219944 ] ], [ [ 220303, 220320 ] ], [ [ 220901, 220924 ] ], [ [ 221306, 221322 ] ], [ [ 221656, 221678 ] ], [ [ 221993, 222014 ] ], [ [ 222375, 222402 ] ], [ [ 222728, 222760 ] ], [ [ 223013, 223051 ] ], [ [ 223350, 223373 ] ], [ [ 223554, 223583 ] ], [ [ 223851, 223880 ] ], [ [ 224099, 224134 ] ], [ [ 224422, 224446 ] ], [ [ 224631, 224661 ] ], [ [ 224933, 224958 ] ], [ [ 225228, 225259 ] ], [ [ 225562, 225598 ] ], [ [ 226332, 226708 ] ], [ [ 226715, 226735 ] ], [ [ 227209, 227224 ] ], [ [ 227900, 227924 ] ] ]
import numpy as np from skimage.morphology import max_tree, area_closing, area_opening from skimage.morphology import max_tree_local_maxima, diameter_opening from skimage.morphology import diameter_closing from skimage.util import invert from skimage._shared.testing import assert_array_equal, TestCase eps = 1e-12 def _full_type_test(img, param, expected, func, param_scale=False, **keywords): # images as they are out = func(img, param, **keywords) assert_array_equal(out, expected) # unsigned int for dt in [np.uint32, np.uint64]: img_cast = img.astype(dt) out = func(img_cast, param, **keywords) exp_cast = expected.astype(dt) assert_array_equal(out, exp_cast) # float data_float = img.astype(np.float64) data_float = data_float / 255.0 expected_float = expected.astype(np.float64) expected_float = expected_float / 255.0 if param_scale: param_cast = param / 255.0 else: param_cast = param for dt in [np.float32, np.float64]: data_cast = data_float.astype(dt) out = func(data_cast, param_cast, **keywords) exp_cast = expected_float.astype(dt) error_img = 255.0 * exp_cast - 255.0 * out error = (error_img >= 1.0).sum() assert error < eps # signed images img_signed = img.astype(np.int16) img_signed = img_signed - 128 exp_signed = expected.astype(np.int16) exp_signed = exp_signed - 128 for dt in [np.int8, np.int16, np.int32, np.int64]: img_s = img_signed.astype(dt) out = func(img_s, param, **keywords) exp_s = exp_signed.astype(dt) assert_array_equal(out, exp_s) class TestMaxtree(TestCase): def test_max_tree(self): "Test for max tree" img_type = np.uint8 img = np.array([[10, 8, 8, 9], [7, 7, 9, 9], [8, 7, 10, 10], [9, 9, 10, 10]], dtype=img_type) P_exp = np.array([[1, 4, 1, 1], [4, 4, 3, 3], [1, 4, 3, 10], [3, 3, 10, 10]], dtype=np.int64) S_exp = np.array([4, 5, 9, 1, 2, 8, 3, 6, 7, 12, 13, 0, 10, 11, 14, 15], dtype=np.int64) for img_type in [np.uint8, np.uint16, np.uint32, np.uint64]: img = img.astype(img_type) P, S = max_tree(img, connectivity=2) assert_array_equal(P, P_exp) assert_array_equal(S, S_exp) for img_type in [np.int8, np.int16, np.int32, np.int64]: img = img.astype(img_type) img_shifted = img - 9 P, S = max_tree(img_shifted, connectivity=2) assert_array_equal(P, P_exp) assert_array_equal(S, S_exp) img_float = img.astype(float) img_float = (img_float - 8) / 2.0 for img_type in [np.float32, np.float64]: img_float = img_float.astype(img_type) P, S = max_tree(img_float, connectivity=2) assert_array_equal(P, P_exp) assert_array_equal(S, S_exp) return def test_area_closing(self): "Test for Area Closing (2 thresholds, all types)" # original image img = np.array( [[240, 240, 240, 240, 240, 240, 240, 240, 240, 240, 240, 240], [240, 200, 200, 240, 200, 240, 200, 200, 240, 240, 200, 240], [240, 200, 40, 240, 240, 240, 240, 240, 240, 240, 40, 240], [240, 240, 240, 240, 100, 240, 100, 100, 240, 240, 200, 240], [240, 240, 240, 240, 240, 240, 240, 240, 240, 240, 240, 240], [200, 200, 200, 200, 200, 200, 200, 240, 200, 200, 255, 255], [200, 255, 200, 200, 200, 255, 200, 240, 255, 255, 255, 40], [200, 200, 200, 100, 200, 200, 200, 240, 255, 255, 255, 255], [200, 200, 200, 100, 200, 200, 200, 240, 200, 200, 255, 255], [200, 200, 200, 200, 200, 40, 200, 240, 240, 100, 255, 255], [200, 40, 255, 255, 255, 40, 200, 255, 200, 200, 255, 255], [200, 200, 200, 200, 200, 200, 200, 255, 255, 255, 255, 255]], dtype=np.uint8) # expected area closing with area 2 expected_2 = np.array( [[240, 240, 240, 240, 240, 240, 240, 240, 240, 240, 240, 240], [240, 200, 200, 240, 240, 240, 200, 200, 240, 240, 200, 240], [240, 200, 200, 240, 240, 240, 240, 240, 240, 240, 200, 240], [240, 240, 240, 240, 240, 240, 100, 100, 240, 240, 200, 240], [240, 240, 240, 240, 240, 240, 240, 240, 240, 240, 240, 240], [200, 200, 200, 200, 200, 200, 200, 240, 200, 200, 255, 255], [200, 255, 200, 200, 200, 255, 200, 240, 255, 255, 255, 255], [200, 200, 200, 100, 200, 200, 200, 240, 255, 255, 255, 255], [200, 200, 200, 100, 200, 200, 200, 240, 200, 200, 255, 255], [200, 200, 200, 200, 200, 40, 200, 240, 240, 200, 255, 255], [200, 200, 255, 255, 255, 40, 200, 255, 200, 200, 255, 255], [200, 200, 200, 200, 200, 200, 200, 255, 255, 255, 255, 255]], dtype=np.uint8) # expected diameter closing with diameter 4 expected_4 = np.array( [[240, 240, 240, 240, 240, 240, 240, 240, 240, 240, 240, 240], [240, 200, 200, 240, 240, 240, 240, 240, 240, 240, 240, 240], [240, 200, 200, 240, 240, 240, 240, 240, 240, 240, 240, 240], [240, 240, 240, 240, 240, 240, 240, 240, 240, 240, 240, 240], [240, 240, 240, 240, 240, 240, 240, 240, 240, 240, 240, 240], [200, 200, 200, 200, 200, 200, 200, 240, 240, 240, 255, 255], [200, 255, 200, 200, 200, 255, 200, 240, 255, 255, 255, 255], [200, 200, 200, 200, 200, 200, 200, 240, 255, 255, 255, 255], [200, 200, 200, 200, 200, 200, 200, 240, 200, 200, 255, 255], [200, 200, 200, 200, 200, 200, 200, 240, 240, 200, 255, 255], [200, 200, 255, 255, 255, 200, 200, 255, 200, 200, 255, 255], [200, 200, 200, 200, 200, 200, 200, 255, 255, 255, 255, 255]], dtype=np.uint8) # _full_type_test makes a test with many image types. _full_type_test(img, 2, expected_2, area_closing, connectivity=2) _full_type_test(img, 4, expected_4, area_closing, connectivity=2) P, S = max_tree(invert(img), connectivity=2) _full_type_test(img, 4, expected_4, area_closing, parent=P, tree_traverser=S) def test_area_opening(self): "Test for Area Opening (2 thresholds, all types)" # original image img = np.array([[15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15], [15, 55, 55, 15, 55, 15, 55, 55, 15, 15, 55, 15], [15, 55, 215, 15, 15, 15, 15, 15, 15, 15, 215, 15], [15, 15, 15, 15, 155, 15, 155, 155, 15, 15, 55, 15], [15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15], [55, 55, 55, 55, 55, 55, 55, 15, 55, 55, 0, 0], [55, 0, 55, 55, 55, 0, 55, 15, 0, 0, 0, 215], [55, 55, 55, 155, 55, 55, 55, 15, 0, 0, 0, 0], [55, 55, 55, 155, 55, 55, 55, 15, 55, 55, 0, 0], [55, 55, 55, 55, 55, 215, 55, 15, 15, 155, 0, 0], [55, 215, 0, 0, 0, 215, 55, 0, 55, 55, 0, 0], [55, 55, 55, 55, 55, 55, 55, 0, 0, 0, 0, 0]], dtype=np.uint8) # expected area closing with area 2 expected_2 = np.array([[15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15], [15, 55, 55, 15, 15, 15, 55, 55, 15, 15, 55, 15], [15, 55, 55, 15, 15, 15, 15, 15, 15, 15, 55, 15], [15, 15, 15, 15, 15, 15, 155, 155, 15, 15, 55, 15], [15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15], [55, 55, 55, 55, 55, 55, 55, 15, 55, 55, 0, 0], [55, 0, 55, 55, 55, 0, 55, 15, 0, 0, 0, 0], [55, 55, 55, 155, 55, 55, 55, 15, 0, 0, 0, 0], [55, 55, 55, 155, 55, 55, 55, 15, 55, 55, 0, 0], [55, 55, 55, 55, 55, 215, 55, 15, 15, 55, 0, 0], [55, 55, 0, 0, 0, 215, 55, 0, 55, 55, 0, 0], [55, 55, 55, 55, 55, 55, 55, 0, 0, 0, 0, 0]], dtype=np.uint8) # expected diameter closing with diameter 4 expected_4 = np.array([[15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15], [15, 55, 55, 15, 15, 15, 15, 15, 15, 15, 15, 15], [15, 55, 55, 15, 15, 15, 15, 15, 15, 15, 15, 15], [15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15], [15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15], [55, 55, 55, 55, 55, 55, 55, 15, 15, 15, 0, 0], [55, 0, 55, 55, 55, 0, 55, 15, 0, 0, 0, 0], [55, 55, 55, 55, 55, 55, 55, 15, 0, 0, 0, 0], [55, 55, 55, 55, 55, 55, 55, 15, 55, 55, 0, 0], [55, 55, 55, 55, 55, 55, 55, 15, 15, 55, 0, 0], [55, 55, 0, 0, 0, 55, 55, 0, 55, 55, 0, 0], [55, 55, 55, 55, 55, 55, 55, 0, 0, 0, 0, 0]], dtype=np.uint8) # _full_type_test makes a test with many image types. _full_type_test(img, 2, expected_2, area_opening, connectivity=2) _full_type_test(img, 4, expected_4, area_opening, connectivity=2) P, S = max_tree(img, connectivity=2) _full_type_test(img, 4, expected_4, area_opening, parent=P, tree_traverser=S) def test_diameter_closing(self): "Test for Diameter Opening (2 thresholds, all types)" img = np.array([[97, 95, 93, 92, 91, 90, 90, 90, 91, 92, 93, 95], [95, 93, 91, 89, 88, 88, 88, 88, 88, 89, 91, 93], [93, 63, 63, 63, 63, 86, 86, 86, 87, 43, 43, 91], [92, 89, 88, 86, 85, 85, 84, 85, 85, 43, 43, 89], [91, 88, 87, 85, 84, 84, 83, 84, 84, 85, 87, 88], [90, 88, 86, 85, 84, 83, 83, 83, 84, 85, 86, 88], [90, 88, 86, 84, 83, 83, 82, 83, 83, 84, 86, 88], [90, 88, 86, 85, 84, 83, 83, 83, 84, 85, 86, 88], [91, 88, 87, 85, 84, 84, 83, 84, 84, 85, 87, 88], [92, 89, 23, 23, 85, 85, 84, 85, 85, 3, 3, 89], [93, 91, 23, 23, 87, 86, 86, 86, 87, 88, 3, 91], [95, 93, 91, 89, 88, 88, 88, 88, 88, 89, 91, 93]], dtype=np.uint8) ex2 = np.array([[97, 95, 93, 92, 91, 90, 90, 90, 91, 92, 93, 95], [95, 93, 91, 89, 88, 88, 88, 88, 88, 89, 91, 93], [93, 63, 63, 63, 63, 86, 86, 86, 87, 43, 43, 91], [92, 89, 88, 86, 85, 85, 84, 85, 85, 43, 43, 89], [91, 88, 87, 85, 84, 84, 83, 84, 84, 85, 87, 88], [90, 88, 86, 85, 84, 83, 83, 83, 84, 85, 86, 88], [90, 88, 86, 84, 83, 83, 83, 83, 83, 84, 86, 88], [90, 88, 86, 85, 84, 83, 83, 83, 84, 85, 86, 88], [91, 88, 87, 85, 84, 84, 83, 84, 84, 85, 87, 88], [92, 89, 23, 23, 85, 85, 84, 85, 85, 3, 3, 89], [93, 91, 23, 23, 87, 86, 86, 86, 87, 88, 3, 91], [95, 93, 91, 89, 88, 88, 88, 88, 88, 89, 91, 93]], dtype=np.uint8) ex4 = np.array([[97, 95, 93, 92, 91, 90, 90, 90, 91, 92, 93, 95], [95, 93, 91, 89, 88, 88, 88, 88, 88, 89, 91, 93], [93, 63, 63, 63, 63, 86, 86, 86, 87, 84, 84, 91], [92, 89, 88, 86, 85, 85, 84, 85, 85, 84, 84, 89], [91, 88, 87, 85, 84, 84, 83, 84, 84, 85, 87, 88], [90, 88, 86, 85, 84, 83, 83, 83, 84, 85, 86, 88], [90, 88, 86, 84, 83, 83, 83, 83, 83, 84, 86, 88], [90, 88, 86, 85, 84, 83, 83, 83, 84, 85, 86, 88], [91, 88, 87, 85, 84, 84, 83, 84, 84, 85, 87, 88], [92, 89, 84, 84, 85, 85, 84, 85, 85, 84, 84, 89], [93, 91, 84, 84, 87, 86, 86, 86, 87, 88, 84, 91], [95, 93, 91, 89, 88, 88, 88, 88, 88, 89, 91, 93]], dtype=np.uint8) # _full_type_test makes a test with many image types. _full_type_test(img, 2, ex2, diameter_closing, connectivity=2) _full_type_test(img, 4, ex4, diameter_closing, connectivity=2) P, S = max_tree(invert(img), connectivity=2) _full_type_test(img, 4, ex4, diameter_opening, parent=P, tree_traverser=S) def test_diameter_opening(self): "Test for Diameter Opening (2 thresholds, all types)" img = np.array([[5, 7, 9, 11, 12, 12, 12, 12, 12, 11, 9, 7], [7, 10, 11, 13, 14, 14, 15, 14, 14, 13, 11, 10], [9, 40, 40, 40, 40, 16, 16, 16, 16, 60, 60, 11], [11, 13, 15, 16, 17, 18, 18, 18, 17, 60, 60, 13], [12, 14, 16, 17, 18, 19, 19, 19, 18, 17, 16, 14], [12, 14, 16, 18, 19, 19, 19, 19, 19, 18, 16, 14], [12, 15, 16, 18, 19, 19, 20, 19, 19, 18, 16, 15], [12, 14, 16, 18, 19, 19, 19, 19, 19, 18, 16, 14], [12, 14, 16, 17, 18, 19, 19, 19, 18, 17, 16, 14], [11, 13, 80, 80, 17, 18, 18, 18, 17, 100, 100, 13], [9, 11, 80, 80, 16, 16, 16, 16, 16, 15, 100, 11], [7, 10, 11, 13, 14, 14, 15, 14, 14, 13, 11, 10]]) ex2 = np.array([[5, 7, 9, 11, 12, 12, 12, 12, 12, 11, 9, 7], [7, 10, 11, 13, 14, 14, 15, 14, 14, 13, 11, 10], [9, 40, 40, 40, 40, 16, 16, 16, 16, 60, 60, 11], [11, 13, 15, 16, 17, 18, 18, 18, 17, 60, 60, 13], [12, 14, 16, 17, 18, 19, 19, 19, 18, 17, 16, 14], [12, 14, 16, 18, 19, 19, 19, 19, 19, 18, 16, 14], [12, 15, 16, 18, 19, 19, 19, 19, 19, 18, 16, 15], [12, 14, 16, 18, 19, 19, 19, 19, 19, 18, 16, 14], [12, 14, 16, 17, 18, 19, 19, 19, 18, 17, 16, 14], [11, 13, 80, 80, 17, 18, 18, 18, 17, 100, 100, 13], [9, 11, 80, 80, 16, 16, 16, 16, 16, 15, 100, 11], [7, 10, 11, 13, 14, 14, 15, 14, 14, 13, 11, 10]]) ex4 = np.array([[5, 7, 9, 11, 12, 12, 12, 12, 12, 11, 9, 7], [7, 10, 11, 13, 14, 14, 15, 14, 14, 13, 11, 10], [9, 40, 40, 40, 40, 16, 16, 16, 16, 18, 18, 11], [11, 13, 15, 16, 17, 18, 18, 18, 17, 18, 18, 13], [12, 14, 16, 17, 18, 19, 19, 19, 18, 17, 16, 14], [12, 14, 16, 18, 19, 19, 19, 19, 19, 18, 16, 14], [12, 15, 16, 18, 19, 19, 19, 19, 19, 18, 16, 15], [12, 14, 16, 18, 19, 19, 19, 19, 19, 18, 16, 14], [12, 14, 16, 17, 18, 19, 19, 19, 18, 17, 16, 14], [11, 13, 18, 18, 17, 18, 18, 18, 17, 18, 18, 13], [9, 11, 18, 18, 16, 16, 16, 16, 16, 15, 18, 11], [7, 10, 11, 13, 14, 14, 15, 14, 14, 13, 11, 10]]) # _full_type_test makes a test with many image types. _full_type_test(img, 2, ex2, diameter_opening, connectivity=2) _full_type_test(img, 4, ex4, diameter_opening, connectivity=2) P, S = max_tree(img, connectivity=2) _full_type_test(img, 4, ex4, diameter_opening, parent=P, tree_traverser=S) def test_local_maxima(self): "local maxima for various data types" data = np.array([[10, 11, 13, 14, 14, 15, 14, 14, 13, 11], [11, 13, 15, 16, 16, 16, 16, 16, 15, 13], [13, 15, 40, 40, 18, 18, 18, 60, 60, 15], [14, 16, 40, 40, 19, 19, 19, 60, 60, 16], [14, 16, 18, 19, 19, 19, 19, 19, 18, 16], [15, 16, 18, 19, 19, 20, 19, 19, 18, 16], [14, 16, 18, 19, 19, 19, 19, 19, 18, 16], [14, 16, 80, 80, 19, 19, 19, 100, 100, 16], [13, 15, 80, 80, 18, 18, 18, 100, 100, 15], [11, 13, 15, 16, 16, 16, 16, 16, 15, 13]], dtype=np.uint8) expected_result = np.array([[0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 1, 0, 0, 0, 1, 1, 0], [0, 0, 1, 1, 0, 0, 0, 1, 1, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 1, 0, 0, 0, 1, 1, 0], [0, 0, 1, 1, 0, 0, 0, 1, 1, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], dtype=np.uint64) for dtype in [np.uint8, np.uint64, np.int8, np.int64]: test_data = data.astype(dtype) out = max_tree_local_maxima(test_data, connectivity=1) out_bin = out > 0 assert_array_equal(expected_result, out_bin) assert out.dtype == expected_result.dtype assert np.max(out) == 5 P, S = max_tree(test_data) out = max_tree_local_maxima(test_data, parent=P, tree_traverser=S) assert_array_equal(expected_result, out_bin) assert out.dtype == expected_result.dtype assert np.max(out) == 5 def test_extrema_float(self): "specific tests for float type" data = np.array([[0.10, 0.11, 0.13, 0.14, 0.14, 0.15, 0.14, 0.14, 0.13, 0.11], [0.11, 0.13, 0.15, 0.16, 0.16, 0.16, 0.16, 0.16, 0.15, 0.13], [0.13, 0.15, 0.40, 0.40, 0.18, 0.18, 0.18, 0.60, 0.60, 0.15], [0.14, 0.16, 0.40, 0.40, 0.19, 0.19, 0.19, 0.60, 0.60, 0.16], [0.14, 0.16, 0.18, 0.19, 0.19, 0.19, 0.19, 0.19, 0.18, 0.16], [0.15, 0.182, 0.18, 0.19, 0.204, 0.20, 0.19, 0.19, 0.18, 0.16], [0.14, 0.16, 0.18, 0.19, 0.19, 0.19, 0.19, 0.19, 0.18, 0.16], [0.14, 0.16, 0.80, 0.80, 0.19, 0.19, 0.19, 4.0, 1.0, 0.16], [0.13, 0.15, 0.80, 0.80, 0.18, 0.18, 0.18, 1.0, 1.0, 0.15], [0.11, 0.13, 0.15, 0.16, 0.16, 0.16, 0.16, 0.16, 0.15, 0.13]], dtype=np.float32) expected_result = np.array([[0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 1, 0, 0, 0, 1, 1, 0], [0, 0, 1, 1, 0, 0, 0, 1, 1, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 1, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 1, 0, 0, 0, 1, 0, 0], [0, 0, 1, 1, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], dtype=np.uint8) # test for local maxima out = max_tree_local_maxima(data, connectivity=1) out_bin = out > 0 assert_array_equal(expected_result, out_bin) assert np.max(out) == 6 def test_3d(self): """tests the detection of maxima in 3D.""" img = np.zeros((8, 8, 8), dtype=np.uint8) local_maxima = np.zeros((8, 8, 8), dtype=np.uint64) # first maximum: only one pixel img[1, 1:3, 1:3] = 100 img[2, 2, 2] = 200 img[3, 1:3, 1:3] = 100 local_maxima[2, 2, 2] = 1 # second maximum: three pixels in z-direction img[5:8, 1, 1] = 200 local_maxima[5:8, 1, 1] = 1 # third: two maxima in 0 and 3. img[0, 5:8, 5:8] = 200 img[1, 6, 6] = 100 img[2, 5:7, 5:7] = 200 img[0:3, 5:8, 5:8] += 50 local_maxima[0, 5:8, 5:8] = 1 local_maxima[2, 5:7, 5:7] = 1 # four : one maximum in the corner of the square img[6:8, 6:8, 6:8] = 200 img[7, 7, 7] = 255 local_maxima[7, 7, 7] = 1 out = max_tree_local_maxima(img) out_bin = out > 0 assert_array_equal(local_maxima, out_bin) assert np.max(out) == 5
[ [ [ 7, 18 ], [ 557, 559 ], [ 568, 570 ], [ 784, 786 ], [ 869, 871 ], [ 1032, 1034 ], [ 1044, 1046 ], [ 1366, 1368 ], [ 1443, 1445 ], [ 1502, 1504 ], [ 1511, 1513 ], [ 1521, 1523 ], [ 1531, 1533 ], [ 1810, 1812 ], [ 1833, 1835 ], [ 2010, 2012 ], [ 2164, 2166 ], [ 2191, 2193 ], [ 2313, 2315 ], [ 2349, 2351 ], [ 2359, 2361 ], [ 2370, 2372 ], [ 2381, 2383 ], [ 2589, 2591 ], [ 2598, 2600 ], [ 2608, 2610 ], [ 2618, 2620 ], [ 2947, 2949 ], [ 2959, 2961 ], [ 3308, 3310 ], [ 4231, 4233 ], [ 4307, 4309 ], [ 5234, 5236 ], [ 5318, 5320 ], [ 6247, 6249 ], [ 6764, 6766 ], [ 7654, 7656 ], [ 7730, 7732 ], [ 9081, 9083 ], [ 9165, 9167 ], [ 10510, 10512 ], [ 11001, 11003 ], [ 11902, 11904 ], [ 11927, 11929 ], [ 12828, 12830 ], [ 12853, 12855 ], [ 13757, 13759 ], [ 14247, 14249 ], [ 15131, 15133 ], [ 16015, 16017 ], [ 17334, 17336 ], [ 18024, 18026 ], [ 18060, 18062 ], [ 18756, 18758 ], [ 18789, 18791 ], [ 18799, 18801 ], [ 18810, 18812 ], [ 18819, 18821 ], [ 19101, 19103 ], [ 19449, 19451 ], [ 19556, 19558 ], [ 20700, 20702 ], [ 20739, 20741 ], [ 21435, 21437 ], [ 21630, 21632 ], [ 21736, 21738 ], [ 21762, 21764 ], [ 21795, 21797 ], [ 21821, 21823 ], [ 22640, 22642 ] ], [ [ 50, 58 ], [ 2451, 2459 ], [ 2721, 2729 ], [ 3042, 3050 ], [ 6484, 6492 ], [ 10747, 10755 ], [ 13988, 13996 ], [ 17102, 17110 ], [ 19138, 19146 ] ], [ [ 60, 72 ], [ 6364, 6376 ], [ 6438, 6450 ], [ 6566, 6578 ] ], [ [ 74, 86 ], [ 10627, 10639 ], [ 10701, 10713 ], [ 10821, 10833 ] ], [ [ 118, 139 ], [ 18892, 18913 ], [ 19176, 19197 ], [ 21492, 21513 ], [ 22522, 22543 ] ], [ [ 141, 157 ], [ 14063, 14079 ], [ 16981, 16997 ], [ 17052, 17068 ], [ 17169, 17185 ] ], [ [ 189, 205 ], [ 13867, 13883 ], [ 13938, 13954 ] ], [ [ 231, 237 ], [ 6493, 6499 ], [ 13997, 14003 ] ], [ [ 275, 293 ], [ 488, 506 ], [ 709, 727 ], [ 1671, 1689 ], [ 2493, 2511 ], [ 2534, 2552 ], [ 2771, 2789 ], [ 2812, 2830 ], [ 3090, 3108 ], [ 3131, 3149 ], [ 18983, 19001 ], [ 19330, 19348 ], [ 21570, 21588 ], [ 22583, 22601 ] ], [ [ 295, 303 ], [ 1722, 1730 ] ], [ [ 305, 308 ], [ 1313, 1316 ] ], [ [ 323, 338 ], [ 6328, 6343 ], [ 6402, 6417 ], [ 6530, 6545 ], [ 10591, 10606 ], [ 10665, 10680 ], [ 10785, 10800 ], [ 13838, 13853 ], [ 13909, 13924 ], [ 14034, 14049 ], [ 16952, 16967 ], [ 17023, 17038 ], [ 17140, 17155 ] ], [ [ 1710, 1721 ] ] ]
import logging import os from pathlib import Path from typing import Any, Callable, Optional from torch.utils.data import Dataset from torchvision import transforms from PIL import Image import cv2 import numpy as np class URISC(Dataset): def __init__( self, dir: str, mode: str = 'train', transform: Optional[Callable] = None, data_rank: str = 'simple', ): super(URISC, self).__init__() self.dir = dir self.mode = mode self.transform = transform self.data_rank = data_rank if data_rank == 'simple': self.transform_normalize = transforms.Normalize(mean=0.520, std=0.185) elif data_rank == 'complex': self.transform_normalize = transforms.Normalize(mean=0.518, std=0.190) self.transform_totensor = transforms.ToTensor() self.ids = [os.path.join(dir, data_rank, mode, filename) for filename in os.listdir(os.path.join(dir, data_rank, mode))] if not self.ids: raise RuntimeError(f'No input file found in {os.path.join(dir, data_rank, mode)}, make sure you put your images there') logging.info(f'Creating dataset with {len(self.ids)} examples') def __len__(self): return len(self.ids) def __getitem__(self, idx): image = cv2.imread(self.ids[idx]) # print(image.shape) if self.mode == 'test': if self.transform is not None: image = self.transform(image=image) return image.float().contiguous(), self.ids[idx] mask_path = self.ids[idx].replace(self.mode, "label/"+self.mode) mask = cv2.imread(mask_path, cv2.IMREAD_GRAYSCALE) # print(mask) if self.transform is not None: transformed = self.transform(image=image, mask=mask) transformed_image = transformed['image'] transformed_mask = transformed['mask'] else: transformed_image = image transformed_mask = mask transformed_image = self.transform_totensor(transformed_image) transformed_image = self.transform_normalize(transformed_image) transformed_mask = self.transform_totensor(transformed_mask) # transformed_image = np.transpose(transformed_image, (2, 0, 1)) # transformed_mask = np.expand_dims(transformed_mask, axis=0) return transformed_image, transformed_mask
[ [ [ 7, 14 ], [ 1158, 1165 ] ], [ [ 22, 24 ], [ 945, 947 ], [ 956, 958 ], [ 884, 886 ], [ 1075, 1077 ] ], [ [ 45, 49 ] ], [ [ 69, 72 ] ], [ [ 74, 82 ], [ 352, 360 ] ], [ [ 84, 92 ], [ 343, 351 ] ], [ [ 123, 130 ], [ 233, 240 ] ], [ [ 155, 165 ], [ 643, 653 ], [ 763, 773 ], [ 841, 851 ] ], [ [ 182, 187 ] ], [ [ 195, 198 ], [ 1324, 1327 ], [ 1673, 1676 ], [ 1695, 1698 ] ], [ [ 207, 218 ] ], [ [ 227, 232 ], [ 427, 432 ] ] ]
# Copyright (C) 2021 NVIDIA CORPORATION & AFFILIATES. All rights reserved. # # This work is made available under the Nvidia Source Code License-NC. # To view a copy of this license, check out LICENSE.md import torch.nn as nn class FeatureMatchingLoss(nn.Module): r"""Compute feature matching loss""" def __init__(self, criterion='l1'): super(FeatureMatchingLoss, self).__init__() if criterion == 'l1': self.criterion = nn.L1Loss() elif criterion == 'l2' or criterion == 'mse': self.criterion = nn.MSELoss() else: raise ValueError('Criterion %s is not recognized' % criterion) def forward(self, fake_features, real_features): r"""Return the target vector for the binary cross entropy loss computation. Args: fake_features (list of lists): Discriminator features of fake images. real_features (list of lists): Discriminator features of real images. Returns: (tensor): Loss value. """ num_d = len(fake_features) dis_weight = 1.0 / num_d loss = fake_features[0][0].new_tensor(0) for i in range(num_d): for j in range(len(fake_features[i])): tmp_loss = self.criterion(fake_features[i][j], real_features[i][j].detach()) loss += dis_weight * tmp_loss return loss
[ [ [ 211, 225 ], [ 254, 256 ], [ 458, 460 ], [ 553, 555 ] ], [ [ 234, 253 ], [ 361, 380 ] ] ]
# Copyright 2017 Datera # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import time import uuid from eventlet.green import threading from oslo_config import cfg from oslo_log import log as logging import six from cinder import exception from cinder.i18n import _ from cinder import utils from cinder.volume import configuration from cinder.volume.drivers.san import san import cinder.volume.drivers.datera.datera_api2 as api2 import cinder.volume.drivers.datera.datera_api21 as api21 import cinder.volume.drivers.datera.datera_common as datc LOG = logging.getLogger(__name__) d_opts = [ cfg.StrOpt('datera_api_port', default='7717', help='Datera API port.'), cfg.StrOpt('datera_api_version', default='2', deprecated_for_removal=True, help='Datera API version.'), cfg.IntOpt('datera_503_timeout', default='120', help='Timeout for HTTP 503 retry messages'), cfg.IntOpt('datera_503_interval', default='5', help='Interval between 503 retries'), cfg.BoolOpt('datera_debug', default=False, help="True to set function arg and return logging"), cfg.BoolOpt('datera_debug_replica_count_override', default=False, help="ONLY FOR DEBUG/TESTING PURPOSES\n" "True to set replica_count to 1"), cfg.StrOpt('datera_tenant_id', default=None, help="If set to 'Map' --> OpenStack project ID will be mapped " "implicitly to Datera tenant ID\n" "If set to 'None' --> Datera tenant ID will not be used " "during volume provisioning\n" "If set to anything else --> Datera tenant ID will be the " "provided value"), cfg.BoolOpt('datera_disable_profiler', default=False, help="Set to True to disable profiling in the Datera driver"), ] CONF = cfg.CONF CONF.import_opt('driver_use_ssl', 'cinder.volume.driver') CONF.register_opts(d_opts, group=configuration.SHARED_CONF_GROUP) @six.add_metaclass(utils.TraceWrapperWithABCMetaclass) class DateraDriver(san.SanISCSIDriver, api2.DateraApi, api21.DateraApi): """The OpenStack Datera Driver Version history: * 1.0 - Initial driver * 1.1 - Look for lun-0 instead of lun-1. * 2.0 - Update For Datera API v2 * 2.1 - Multipath, ACL and reorg * 2.2 - Capabilites List, Extended Volume-Type Support Naming convention change, Volume Manage/Unmanage support * 2.3 - Templates, Tenants, Snapshot Polling, 2.1 Api Version Support, Restructure * 2.3.1 - Scalability bugfixes * 2.3.2 - Volume Placement, ACL multi-attach bugfix * 2.4.0 - Fast Retype Support """ VERSION = '2.4.0' CI_WIKI_NAME = "datera-ci" HEADER_DATA = {'Datera-Driver': 'OpenStack-Cinder-{}'.format(VERSION)} # TODO(jsbryant) Remove driver in the 'U' release if CI is not fixed. SUPPORTED = False def __init__(self, *args, **kwargs): super(DateraDriver, self).__init__(*args, **kwargs) self.configuration.append_config_values(d_opts) self.username = self.configuration.san_login self.password = self.configuration.san_password self.cluster_stats = {} self.datera_api_token = None self.interval = self.configuration.datera_503_interval self.retry_attempts = (self.configuration.datera_503_timeout / self.interval) self.driver_prefix = str(uuid.uuid4())[:4] self.datera_debug = self.configuration.datera_debug self.datera_api_versions = [] if self.datera_debug: utils.setup_tracing(['method']) self.tenant_id = self.configuration.datera_tenant_id if self.tenant_id and self.tenant_id.lower() == 'none': self.tenant_id = None self.api_check = time.time() self.api_cache = [] self.api_timeout = 0 self.do_profile = not self.configuration.datera_disable_profiler self.thread_local = threading.local() backend_name = self.configuration.safe_get( 'volume_backend_name') self.backend_name = backend_name or 'Datera' datc.register_driver(self) def do_setup(self, context): # If we can't authenticate through the old and new method, just fail # now. if not all([self.username, self.password]): msg = _("san_login and/or san_password is not set for Datera " "driver in the cinder.conf. Set this information and " "start the cinder-volume service again.") LOG.error(msg) raise exception.InvalidInput(msg) self.login() self._create_tenant() # ================= # ================= # = Create Volume = # ================= @datc._api_lookup def create_volume(self, volume): """Create a logical volume.""" pass # ================= # = Extend Volume = # ================= @datc._api_lookup def extend_volume(self, volume, new_size): pass # ================= # ================= # = Cloned Volume = # ================= @datc._api_lookup def create_cloned_volume(self, volume, src_vref): pass # ================= # = Delete Volume = # ================= @datc._api_lookup def delete_volume(self, volume): pass # ================= # = Ensure Export = # ================= @datc._api_lookup def ensure_export(self, context, volume, connector=None): """Gets the associated account, retrieves CHAP info and updates.""" # ========================= # = Initialize Connection = # ========================= @datc._api_lookup def initialize_connection(self, volume, connector): pass # ================= # = Create Export = # ================= @datc._api_lookup def create_export(self, context, volume, connector): pass # ================= # = Detach Volume = # ================= @datc._api_lookup def detach_volume(self, context, volume, attachment=None): pass # =================== # = Create Snapshot = # =================== @datc._api_lookup def create_snapshot(self, snapshot): pass # =================== # = Delete Snapshot = # =================== @datc._api_lookup def delete_snapshot(self, snapshot): pass # ======================== # = Volume From Snapshot = # ======================== @datc._api_lookup def create_volume_from_snapshot(self, volume, snapshot): pass # ========== # = Retype = # ========== @datc._api_lookup def retype(self, ctxt, volume, new_type, diff, host): """Convert the volume to be of the new type. Returns a boolean indicating whether the retype occurred. :param ctxt: Context :param volume: A dictionary describing the volume to migrate :param new_type: A dictionary describing the volume type to convert to :param diff: A dictionary with the difference between the two types :param host: A dictionary describing the host to migrate to, where host['host'] is its name, and host['capabilities'] is a dictionary of its reported capabilities (Not Used). """ pass # ========== # = Manage = # ========== @datc._api_lookup def manage_existing(self, volume, existing_ref): """Manage an existing volume on the Datera backend The existing_ref must be either the current name or Datera UUID of an app_instance on the Datera backend in a colon separated list with the storage instance name and volume name. This means only single storage instances and single volumes are supported for managing by cinder. Eg. (existing_ref['source-name'] == tenant:app_inst_name:storage_inst_name:vol_name) if using Datera 2.1 API or (existing_ref['source-name'] == app_inst_name:storage_inst_name:vol_name) if using 2.0 API :param volume: Cinder volume to manage :param existing_ref: Driver-specific information used to identify a volume """ pass # =================== # = Manage Get Size = # =================== @datc._api_lookup def manage_existing_get_size(self, volume, existing_ref): """Get the size of an unmanaged volume on the Datera backend The existing_ref must be either the current name or Datera UUID of an app_instance on the Datera backend in a colon separated list with the storage instance name and volume name. This means only single storage instances and single volumes are supported for managing by cinder. Eg. existing_ref == app_inst_name:storage_inst_name:vol_name :param volume: Cinder volume to manage :param existing_ref: Driver-specific information used to identify a volume on the Datera backend """ pass # ========================= # = Get Manageable Volume = # ========================= @datc._api_lookup def get_manageable_volumes(self, cinder_volumes, marker, limit, offset, sort_keys, sort_dirs): """List volumes on the backend available for management by Cinder. Returns a list of dictionaries, each specifying a volume in the host, with the following keys: - reference (dictionary): The reference for a volume, which can be passed to 'manage_existing'. - size (int): The size of the volume according to the storage backend, rounded up to the nearest GB. - safe_to_manage (boolean): Whether or not this volume is safe to manage according to the storage backend. For example, is the volume in use or invalid for any reason. - reason_not_safe (string): If safe_to_manage is False, the reason why. - cinder_id (string): If already managed, provide the Cinder ID. - extra_info (string): Any extra information to return to the user :param cinder_volumes: A list of volumes in this host that Cinder currently manages, used to determine if a volume is manageable or not. :param marker: The last item of the previous page; we return the next results after this value (after sorting) :param limit: Maximum number of items to return :param offset: Number of items to skip after marker :param sort_keys: List of keys to sort results by (valid keys are 'identifier' and 'size') :param sort_dirs: List of directions to sort by, corresponding to sort_keys (valid directions are 'asc' and 'desc') """ pass # ============ # = Unmanage = # ============ @datc._api_lookup def unmanage(self, volume): """Unmanage a currently managed volume in Cinder :param volume: Cinder volume to unmanage """ pass # ================ # = Volume Stats = # ================ @datc._api_lookup def get_volume_stats(self, refresh=False): """Get volume stats. If 'refresh' is True, run update first. The name is a bit misleading as the majority of the data here is cluster data. """ pass # ========= # = Login = # ========= @datc._api_lookup def login(self): pass # ======= # = QoS = # ======= def _update_qos(self, resource, policies): url = datc.URL_TEMPLATES['vol_inst']( policies['default_storage_name'], policies['default_volume_name']) + '/performance_policy' url = url.format(datc._get_name(resource['id'])) type_id = resource.get('volume_type_id', None) if type_id is not None: # Filter for just QOS policies in result. All of their keys # should end with "max" fpolicies = {k: int(v) for k, v in policies.items() if k.endswith("max")} # Filter all 0 values from being passed fpolicies = dict(filter(lambda _v: _v[1] > 0, fpolicies.items())) if fpolicies: self._issue_api_request(url, 'post', body=fpolicies, api_version='2') def _get_lunid(self): return 0 # ============================ # = Volume-Types/Extra-Specs = # ============================ def _init_vendor_properties(self): """Create a dictionary of vendor unique properties. This method creates a dictionary of vendor unique properties and returns both created dictionary and vendor name. Returned vendor name is used to check for name of vendor unique properties. - Vendor name shouldn't include colon(:) because of the separator and it is automatically replaced by underscore(_). ex. abc:d -> abc_d - Vendor prefix is equal to vendor name. ex. abcd - Vendor unique properties must start with vendor prefix + ':'. ex. abcd:maxIOPS Each backend driver needs to override this method to expose its own properties using _set_property() like this: self._set_property( properties, "vendorPrefix:specific_property", "Title of property", _("Description of property"), "type") : return dictionary of vendor unique properties : return vendor name prefix: DF --> Datera Fabric """ properties = {} self._set_property( properties, "DF:placement_mode", "Datera Volume Placement", _("'single_flash' for single-flash-replica placement, " "'all_flash' for all-flash-replica placement, " "'hybrid' for hybrid placement"), "string", default="hybrid") self._set_property( properties, "DF:round_robin", "Datera Round Robin Portals", _("True to round robin the provided portals for a target"), "boolean", default=False) if self.configuration.get('datera_debug_replica_count_override'): replica_count = 1 else: replica_count = 3 self._set_property( properties, "DF:replica_count", "Datera Volume Replica Count", _("Specifies number of replicas for each volume. Can only be " "increased once volume is created"), "integer", minimum=1, default=replica_count) self._set_property( properties, "DF:acl_allow_all", "Datera ACL Allow All", _("True to set acl 'allow_all' on volumes created. Cannot be " "changed on volume once set"), "boolean", default=False) self._set_property( properties, "DF:ip_pool", "Datera IP Pool", _("Specifies IP pool to use for volume"), "string", default="default") self._set_property( properties, "DF:template", "Datera Template", _("Specifies Template to use for volume provisioning"), "string", default="") # ###### QoS Settings ###### # self._set_property( properties, "DF:read_bandwidth_max", "Datera QoS Max Bandwidth Read", _("Max read bandwidth setting for volume qos, " "use 0 for unlimited"), "integer", minimum=0, default=0) self._set_property( properties, "DF:default_storage_name", "Datera Default Storage Instance Name", _("The name to use for storage instances created"), "string", default="storage-1") self._set_property( properties, "DF:default_volume_name", "Datera Default Volume Name", _("The name to use for volumes created"), "string", default="volume-1") self._set_property( properties, "DF:write_bandwidth_max", "Datera QoS Max Bandwidth Write", _("Max write bandwidth setting for volume qos, " "use 0 for unlimited"), "integer", minimum=0, default=0) self._set_property( properties, "DF:total_bandwidth_max", "Datera QoS Max Bandwidth Total", _("Max total bandwidth setting for volume qos, " "use 0 for unlimited"), "integer", minimum=0, default=0) self._set_property( properties, "DF:read_iops_max", "Datera QoS Max iops Read", _("Max read iops setting for volume qos, " "use 0 for unlimited"), "integer", minimum=0, default=0) self._set_property( properties, "DF:write_iops_max", "Datera QoS Max IOPS Write", _("Max write iops setting for volume qos, " "use 0 for unlimited"), "integer", minimum=0, default=0) self._set_property( properties, "DF:total_iops_max", "Datera QoS Max IOPS Total", _("Max total iops setting for volume qos, " "use 0 for unlimited"), "integer", minimum=0, default=0) # ###### End QoS Settings ###### # return properties, 'DF'
[ [ [ 629, 633 ], [ 4627, 4631 ] ], [ [ 641, 645 ], [ 4252, 4256 ] ], [ [ 674, 683 ], [ 4797, 4806 ] ], [ [ 708, 711 ], [ 1147, 1150 ], [ 1253, 1256 ], [ 1406, 1409 ], [ 1533, 1536 ], [ 1652, 1655 ], [ 1784, 1787 ], [ 1983, 1986 ], [ 2429, 2432 ], [ 2589, 2592 ] ], [ [ 733, 747 ], [ 1103, 1110 ] ], [ [ 755, 758 ], [ 2725, 2728 ] ], [ [ 779, 788 ], [ 5427, 5436 ] ], [ [ 813, 814 ], [ 5188, 5189 ], [ 14952, 14953 ], [ 15307, 15308 ], [ 15705, 15706 ], [ 16033, 16034 ], [ 16313, 16314 ], [ 16531, 16532 ], [ 16819, 16820 ], [ 17130, 17131 ], [ 17382, 17383 ], [ 17627, 17628 ], [ 17932, 17933 ], [ 18225, 18226 ], [ 18514, 18515 ], [ 18804, 18805 ] ], [ [ 834, 839 ], [ 2743, 2748 ], [ 4411, 4416 ] ], [ [ 866, 879 ], [ 2689, 2702 ] ], [ [ 918, 921 ], [ 2798, 2801 ] ], [ [ 930, 978 ], [ 2818, 2822 ] ], [ [ 986, 1036 ], [ 2834, 2839 ] ], [ [ 1044, 1094 ], [ 5611, 5615 ], [ 5796, 5800 ], [ 5977, 5981 ], [ 6140, 6144 ], [ 6286, 6290 ], [ 6544, 6548 ], [ 6709, 6713 ], [ 6875, 6879 ], [ 7053, 7057 ], [ 7209, 7213 ], [ 7380, 7384 ], [ 7529, 7533 ], [ 8286, 8290 ], [ 9293, 9297 ], [ 10155, 10159 ], [ 11990, 11994 ], [ 12253, 12257 ], [ 12578, 12582 ], [ 4965, 4969 ], [ 12734, 12738 ], [ 12906, 12910 ] ], [ [ 1097, 1100 ], [ 5394, 5397 ] ], [ [ 1132, 1138 ], [ 2675, 2681 ], [ 3853, 3859 ] ], [ [ 2582, 2586 ], [ 2598, 2602 ], [ 2656, 2660 ] ], [ [ 2785, 2797 ], [ 3759, 3771 ] ] ]
#!/usr/bin/env python """ Author: Alexander David Leech Date: 30/09/2015 Rev: 2 Lang: Python 2.7 Deps: Pyserial, Pymodbus, logging """ import time # For sleep functionality import logging # For detailed error output from pymodbus.client.sync import ModbusSerialClient \ as ModbusClient # Import MODBUS support class comSettings = { "method" : 'rtu', "port" : 'COM3', "stopbits" : 1, "bytesize" : 8, "parity" : 'N', "baudrate" : 9600, "timeout" : 1 } logging.basicConfig() # Setup error logging log = logging.getLogger() # Start logging client = ModbusClient(**comSettings) # Setup connection object client.connect() # Open the MODBUS connection while(True): client.write_register(3,1000,unit=0x01) # Write valve to 100% time.sleep(4) # Sleep 4 seconds client.write_register(3,0,unit=0x01) # Write valve to 0% time.sleep(4) # Sleep 4 seconds client.close() # Close the connection
[ [ [ 150, 154 ], [ 1167, 1171 ], [ 1317, 1321 ] ], [ [ 231, 238 ], [ 753, 760 ], [ 837, 844 ] ], [ [ 340, 376 ], [ 913, 925 ] ], [ [ 447, 458 ], [ 928, 939 ] ], [ [ 831, 834 ] ], [ [ 904, 910 ], [ 986, 992 ], [ 1089, 1095 ], [ 1241, 1247 ], [ 1388, 1394 ] ] ]
import os import subprocess from unittest import mock import pytest from pre_commit.constants import VERSION as PRE_COMMIT_VERSION import testing.git from all_repos import autofix_lib from all_repos import clone from all_repos import git from all_repos.config import load_config @pytest.mark.parametrize( ('cli_repos', 'expected'), ( (None, ['found_repo']), ([], []), (['cli_repo'], ['cli_repo']), ), ) def test_filter_repos(file_config, cli_repos, expected): ret = autofix_lib.filter_repos( file_config, cli_repos, lambda _: ['found_repo'], ) assert ret == expected def test_assert_importable_is_importable(): autofix_lib.assert_importable('pre_commit', install='pre-commit') def test_assert_importable_not_importable(): with pytest.raises(SystemExit) as excinfo: autofix_lib.assert_importable('watmodule', install='wat') msg, = excinfo.value.args assert msg == ( 'This tool requires the `watmodule` module to be installed.\n' 'Try installing it via `pip install wat`.' ) def test_require_version_new_enough(): autofix_lib.require_version_gte('pre-commit', '0.17.0') def test_require_version_not_new_enough(): with pytest.raises(SystemExit) as excinfo: autofix_lib.require_version_gte('pre-commit', '999') msg, = excinfo.value.args assert msg == ( f'This tool requires the `pre-commit` package is at least version ' f'999. The currently installed version is {PRE_COMMIT_VERSION}.\n\n' f'Try `pip install --upgrade pre-commit`' ) def test_run(capfd): autofix_lib.run('echo', 'h"i') out, _ = capfd.readouterr() assert out == ( '$ echo \'h"i\'\n' 'h"i\n' ) def test_cwd(tmpdir): orig = os.getcwd() with autofix_lib.cwd(tmpdir): assert os.getcwd() == tmpdir assert os.getcwd() == orig def test_repo_context_success(file_config_files, capsys): expected_rev = testing.git.revparse(file_config_files.dir1) with autofix_lib.repo_context( str(file_config_files.output_dir.join('repo1')), use_color=False, ): assert testing.git.revparse('.') == expected_rev assert git.remote('.') == file_config_files.dir1 out, err = capsys.readouterr() assert err == '' assert 'Errored' not in out def test_repo_context_errors(file_config_files, capsys): with autofix_lib.repo_context( str(file_config_files.output_dir.join('repo1')), use_color=False, ): assert False out, err = capsys.readouterr() assert 'Errored' in out assert 'assert False' in err def test_interactive_control_c(mock_input, capfd): mock_input.set_side_effect(KeyboardInterrupt) with pytest.raises(SystemExit): autofix_lib._interactive_check(use_color=False) out, _ = capfd.readouterr() assert out == ( '***Looks good [y,n,s,q,?]? ^C\n' 'Goodbye!\n' ) def test_interactive_eof(mock_input, capfd): mock_input.set_side_effect(EOFError) with pytest.raises(SystemExit): autofix_lib._interactive_check(use_color=False) out, _ = capfd.readouterr() assert out == ( '***Looks good [y,n,s,q,?]? ^D\n' 'Goodbye!\n' ) def test_interactive_quit(mock_input, capfd): mock_input.set_side_effect('q') with pytest.raises(SystemExit): autofix_lib._interactive_check(use_color=False) out, _ = capfd.readouterr() assert out == ( '***Looks good [y,n,s,q,?]? <<q\n' 'Goodbye!\n' ) def test_interactive_yes(mock_input, capfd): mock_input.set_side_effect('y') assert autofix_lib._interactive_check(use_color=False) is True out, _ = capfd.readouterr() assert out == '***Looks good [y,n,s,q,?]? <<y\n' def test_interactive_no(mock_input, capfd): mock_input.set_side_effect('n') assert autofix_lib._interactive_check(use_color=False) is False out, _ = capfd.readouterr() assert out == '***Looks good [y,n,s,q,?]? <<n\n' def test_interactive_shell(mock_input, capfd): mock_input.set_side_effect('s', 'n') with mock.patch.dict(os.environ, {'SHELL': 'echo'}): assert autofix_lib._interactive_check(use_color=False) is False out, _ = capfd.readouterr() assert out == ( '***Looks good [y,n,s,q,?]? <<s\n' 'Opening an interactive shell, type `exit` to continue.\n' 'Any modifications will be committed.\n' # A newline from echo '\n' '***Looks good [y,n,s,q,?]? <<n\n' ) def test_interactive_help(mock_input, capfd): mock_input.set_side_effect('?', 'n') assert autofix_lib._interactive_check(use_color=False) is False out, _ = capfd.readouterr() assert out == ( '***Looks good [y,n,s,q,?]? <<?\n' 'y (yes): yes it looks good, commit and continue.\n' 'n (no): no, do not commit this repository.\n' 's (shell): open an interactive shell in the repo.\n' 'q (quit, ^C): early exit from the autofixer.\n' '? (help): show this help message.\n' '***Looks good [y,n,s,q,?]? <<n\n' ) def test_interactive_garbage(mock_input, capfd): mock_input.set_side_effect('garbage', 'n') assert autofix_lib._interactive_check(use_color=False) is False out, _ = capfd.readouterr() assert out == ( '***Looks good [y,n,s,q,?]? <<garbage\n' 'Unexpected input: garbage\n' 'y (yes): yes it looks good, commit and continue.\n' 'n (no): no, do not commit this repository.\n' 's (shell): open an interactive shell in the repo.\n' 'q (quit, ^C): early exit from the autofixer.\n' '? (help): show this help message.\n' '***Looks good [y,n,s,q,?]? <<n\n' ) def lower_case_f(): f_contents = open('f').read() with open('f', 'w') as f: f.write(f_contents.lower()) def failing_check_fix(): raise AssertionError('nope!') def test_fix_dry_run_no_change(file_config_files, capfd): autofix_lib.fix( ( str(file_config_files.output_dir.join('repo1')), str(file_config_files.output_dir.join('repo2')), ), apply_fix=lower_case_f, config=load_config(file_config_files.cfg), commit=autofix_lib.Commit('message!', 'test-branch', None), autofix_settings=autofix_lib.AutofixSettings( jobs=1, color=False, limit=None, dry_run=True, interactive=False, ), ) out, err = capfd.readouterr() assert err == '' assert 'Errored' not in out # Showed the diff of what would have happened assert '-OHAI\n+ohai\n' in out assert '-OHELLO\n+ohello\n' in out # Didn't actually perform any changes assert file_config_files.dir1.join('f').read() == 'OHAI\n' assert file_config_files.dir2.join('f').read() == 'OHELLO\n' def test_fix_with_limit(file_config_files, capfd): autofix_lib.fix( ( str(file_config_files.output_dir.join('repo1')), str(file_config_files.output_dir.join('repo2')), ), apply_fix=lower_case_f, config=load_config(file_config_files.cfg), commit=autofix_lib.Commit('message!', 'test-branch', None), autofix_settings=autofix_lib.AutofixSettings( jobs=1, color=False, limit=1, dry_run=True, interactive=False, ), ) out, err = capfd.readouterr() assert err == '' assert 'Errored' not in out # Should still see the diff from the first repository assert '-OHAI\n+ohai\n' in out assert '-OHELLO\n+ohello\n' not in out def test_fix_interactive(file_config_files, capfd, mock_input): mock_input.set_side_effect('y', 'n') autofix_lib.fix( ( str(file_config_files.output_dir.join('repo1')), str(file_config_files.output_dir.join('repo2')), ), apply_fix=lower_case_f, config=load_config(file_config_files.cfg), commit=autofix_lib.Commit('message!', 'test-branch', None), autofix_settings=autofix_lib.AutofixSettings( jobs=1, color=False, limit=None, dry_run=False, interactive=True, ), ) assert file_config_files.dir1.join('f').read() == 'ohai\n' assert file_config_files.dir2.join('f').read() == 'OHELLO\n' def test_autofix_makes_commits(file_config_files, capfd): autofix_lib.fix( ( str(file_config_files.output_dir.join('repo1')), str(file_config_files.output_dir.join('repo2')), ), apply_fix=lower_case_f, config=load_config(file_config_files.cfg), commit=autofix_lib.Commit('message!', 'test-branch', 'A B <[email protected]>'), autofix_settings=autofix_lib.AutofixSettings( jobs=1, color=False, limit=None, dry_run=False, interactive=False, ), ) out, err = capfd.readouterr() assert err == '' assert 'Errored' not in out assert file_config_files.dir1.join('f').read() == 'ohai\n' assert file_config_files.dir2.join('f').read() == 'ohello\n' # The branch name should be what we specified last_commit_msg = subprocess.check_output(( 'git', '-C', file_config_files.dir1, 'log', '--format=%s', '--first-parent', '-1', )).decode() assert last_commit_msg == "Merge branch 'all-repos_autofix_test-branch'\n" # We should see a commit from the autofix change we made commit = subprocess.check_output(( 'git', '-C', file_config_files.dir1, 'log', '--patch', '--grep', 'message!', '--format=%an %ae\n%B', )).decode() assert commit.startswith( 'A B [email protected]\n' 'message!\n' '\n' 'Committed via https://github.com/asottile/all-repos\n', ) assert commit.endswith('-OHAI\n+ohai\n') def test_fix_failing_check_no_changes(file_config_files, capfd): autofix_lib.fix( ( str(file_config_files.output_dir.join('repo1')), str(file_config_files.output_dir.join('repo2')), ), apply_fix=lower_case_f, check_fix=failing_check_fix, config=load_config(file_config_files.cfg), commit=autofix_lib.Commit('message!', 'test-branch', None), autofix_settings=autofix_lib.AutofixSettings( jobs=1, color=False, limit=None, dry_run=False, interactive=False, ), ) out, err = capfd.readouterr() assert 'nope!' in err assert out.count('Errored') == 2 # An error while checking should not allow the changes assert file_config_files.dir1.join('f').read() == 'OHAI\n' assert file_config_files.dir2.join('f').read() == 'OHELLO\n' def test_noop_does_not_commit(file_config_files): rev_before1 = testing.git.revparse(file_config_files.dir1) rev_before2 = testing.git.revparse(file_config_files.dir2) autofix_lib.fix( ( str(file_config_files.output_dir.join('repo1')), str(file_config_files.output_dir.join('repo2')), ), apply_fix=lambda: None, config=load_config(file_config_files.cfg), commit=autofix_lib.Commit('message!', 'test-branch', None), autofix_settings=autofix_lib.AutofixSettings( jobs=1, color=False, limit=None, dry_run=False, interactive=False, ), ) rev_after1 = testing.git.revparse(file_config_files.dir1) rev_after2 = testing.git.revparse(file_config_files.dir2) assert (rev_before1, rev_before2) == (rev_after1, rev_after2) def test_fix_non_default_branch(file_config_non_default): clone.main(('--config-filename', str(file_config_non_default.cfg))) autofix_lib.fix( ( str(file_config_non_default.output_dir.join('repo1')), ), apply_fix=lower_case_f, config=load_config(file_config_non_default.cfg), commit=autofix_lib.Commit('message!', 'test-branch', 'A B <[email protected]>'), autofix_settings=autofix_lib.AutofixSettings( jobs=1, color=False, limit=None, dry_run=False, interactive=False, ), ) assert file_config_non_default.dir1.join('f').read() == 'ohai\n'
[ [ [ 7, 9 ], [ 1789, 1791 ], [ 1850, 1852 ], [ 1883, 1885 ], [ 4145, 4147 ] ], [ [ 17, 27 ], [ 9121, 9131 ], [ 9416, 9426 ] ], [ [ 49, 53 ], [ 4129, 4133 ] ], [ [ 62, 68 ], [ 284, 290 ], [ 799, 805 ], [ 1236, 1242 ], [ 2757, 2763 ], [ 3058, 3064 ], [ 3355, 3361 ] ], [ [ 102, 131 ], [ 1513, 1531 ] ], [ [ 140, 151 ], [ 1982, 1989 ], [ 2162, 2169 ], [ 10702, 10709 ], [ 10765, 10772 ], [ 11292, 11299 ], [ 11354, 11361 ] ], [ [ 174, 185 ], [ 510, 521 ], [ 677, 688 ], [ 845, 856 ], [ 1126, 1137 ], [ 1282, 1293 ], [ 1622, 1633 ], [ 1810, 1821 ], [ 2036, 2047 ], [ 2417, 2428 ], [ 2792, 2803 ], [ 3093, 3104 ], [ 3390, 3401 ], [ 3654, 3665 ], [ 3888, 3899 ], [ 4192, 4203 ], [ 4652, 4663 ], [ 5243, 5254 ], [ 6016, 6027 ], [ 6274, 6285 ], [ 6352, 6363 ], [ 6916, 6927 ], [ 7174, 7185 ], [ 7252, 7263 ], [ 7708, 7719 ], [ 7966, 7977 ], [ 8044, 8055 ], [ 8361, 8372 ], [ 8619, 8630 ], [ 8706, 8717 ], [ 9848, 9859 ], [ 10143, 10154 ], [ 10221, 10232 ], [ 10814, 10825 ], [ 11072, 11083 ], [ 11150, 11161 ], [ 11602, 11613 ], [ 11811, 11822 ], [ 11898, 11909 ] ], [ [ 208, 213 ], [ 11529, 11534 ] ], [ [ 236, 239 ], [ 2219, 2222 ] ], [ [ 269, 280 ], [ 6223, 6234 ], [ 7123, 7134 ], [ 7915, 7926 ], [ 8568, 8579 ], [ 10092, 10103 ], [ 11021, 11032 ], [ 11754, 11765 ] ], [ [ 447, 464 ] ], [ [ 633, 669 ] ], [ [ 749, 786 ] ], [ [ 1087, 1118 ] ], [ [ 1188, 1223 ] ], [ [ 1601, 1609 ] ], [ [ 1760, 1768 ] ], [ [ 1909, 1934 ] ], [ [ 2355, 2379 ] ], [ [ 2651, 2677 ] ], [ [ 2967, 2987 ] ], [ [ 3268, 3289 ] ], [ [ 3566, 3586 ] ], [ [ 3801, 3820 ] ], [ [ 4036, 4058 ] ], [ [ 4558, 4579 ] ], [ [ 5140, 5164 ] ], [ [ 5775, 5787 ], [ 6194, 6206 ], [ 7094, 7106 ], [ 7886, 7898 ], [ 8539, 8551 ], [ 10026, 10038 ], [ 11725, 11737 ] ], [ [ 5897, 5914 ], [ 10058, 10075 ] ], [ [ 5958, 5984 ] ], [ [ 6865, 6884 ] ], [ [ 7603, 7623 ] ], [ [ 8303, 8329 ] ], [ [ 9783, 9816 ] ], [ [ 10638, 10663 ] ], [ [ 11471, 11498 ] ] ]
#!/usr/bin/env python # -*- coding: utf-8 -*- """ Generate random usernames in """ import random from .names import names as default_names class NameGenerator(object): def __init__(self, names=None): self.names = names or default_names def __call__(self): return self.names.pop(random.randrange(len(self.names))) def __iter__(self): while self.names: yield self()
[ [ [ 97, 103 ], [ 315, 321 ] ], [ [ 124, 146 ], [ 245, 258 ] ], [ [ 155, 168 ] ] ]
from django.utils import timezone from rest_framework.authtoken.models import Token class AuthTokenHandler: """ Handles variations in auth token """ @staticmethod def expired_token(auth_token): """ Checks expiry of auth token """ utc_now = timezone.now() expired = auth_token.created < utc_now - \ timezone.timedelta(hours=24) return expired @staticmethod def create_auth_token(user): """ Creates an auth token for a user """ token, created = Token.objects.get_or_create(user=user) if not created: token.created = timezone.now() token.save() return token
[ [ [ 25, 33 ], [ 294, 302 ], [ 372, 380 ], [ 657, 665 ] ], [ [ 78, 83 ], [ 566, 571 ] ], [ [ 92, 108 ] ] ]
from .useful_functions import get_ngrams, words_to_ngrams_list, remove_hook_words, remove_words from .transformers import phrases_transform, phrases2lower, phrases_without_excess_symbols from .tokenizers import text2sentences, split_by_words, sentence_split from .stemlem_operators import create_stemmer_lemmer, create_stemmer, create_lemmatizer from .pipeline import StemLemPipeline from .simplifiers import sum_phrases, wordlist2set from .stopwords import stopwords from .metrics import Levenstein
[ [ [ 32, 42 ] ], [ [ 44, 64 ] ], [ [ 66, 83 ] ], [ [ 85, 97 ] ], [ [ 125, 142 ] ], [ [ 144, 157 ] ], [ [ 159, 189 ] ], [ [ 215, 229 ] ], [ [ 231, 245 ] ], [ [ 247, 261 ] ], [ [ 294, 315 ] ], [ [ 317, 331 ] ], [ [ 333, 350 ] ], [ [ 374, 389 ] ], [ [ 416, 427 ] ], [ [ 429, 441 ] ], [ [ 466, 475 ] ], [ [ 498, 508 ] ] ]
# internal imports import dependency_checker import dependency_installer import dependency_updater import logger from rendering import VortexWindow # external imports import pyglet import sys # check if python version is too old. If it is, exit. if sys.version_info < (3, 6): # if python version is less than 3.6 logger.critical( "Vortex", "Python version is too old. Please use python 3.6 or higher.") sys.exit(1) # check all deps and update them if needed if not dependency_checker.check_deps(): # if any deps are missing dependency_installer.install_deps() # install them if not dependency_checker.check_deps(): # if any deps are still missing # warn user and exit logger.warn( "Vortex", "Dependencies are not installed. Please install them manually.") sys.exit(1) else: dependency_updater.update_deps() # update deps window = VortexWindow() # create the window pyglet.app.run() # run the app
[ [ [ 26, 44 ], [ 485, 503 ], [ 612, 630 ] ], [ [ 52, 72 ], [ 549, 569 ] ], [ [ 80, 98 ], [ 845, 863 ] ], [ [ 106, 112 ], [ 320, 326 ], [ 715, 721 ] ], [ [ 135, 147 ], [ 903, 915 ] ], [ [ 175, 181 ], [ 939, 945 ] ], [ [ 189, 192 ], [ 251, 254 ], [ 422, 425 ], [ 823, 826 ] ], [ [ 894, 900 ] ] ]
""" Intergation of the pytorch_transformers openai and gpt2 modules. Note that these objects are only to be used to load pretrained models. The pytorch-transformers library wasn't designed to train these models from scratch. """ import pytorch_transformers as pt from flambe.nlp.transformers.utils import TransformerTextField, TransformerEmbedder class GPTTextField(TransformerTextField): """Integrate the pytorch_transformers OpenAIGPTTokenizer. Currently available aliases: . `openai-gpt` """ _cls = pt.OpenAIGPTTokenizer class GPTEmbedder(TransformerEmbedder): """Integrate the pytorch_transformers OpenAIGPTmodel. Currently available aliases: . `openai-gpt` """ _cls = pt.OpenAIGPTModel class GPT2TextField(TransformerTextField): """Integrate the pytorch_transformers GPT2Tokenizer. Currently available aliases: . `gpt2` . `gpt2-medium` . `gpt2-large` """ _cls = pt.GPT2Tokenizer class GPT2Embedder(TransformerEmbedder): """Integrate the pytorch_transformers GPT2Model. Currently available aliases: . `gpt2` . `gpt2-medium` . `gpt2-large` """ _cls = pt.GPT2Model
[ [ [ 239, 265 ], [ 535, 537 ], [ 735, 737 ], [ 974, 976 ], [ 1206, 1208 ] ], [ [ 309, 329 ], [ 372, 392 ], [ 775, 795 ] ], [ [ 331, 350 ], [ 577, 596 ], [ 1012, 1031 ] ], [ [ 359, 371 ] ], [ [ 565, 576 ] ], [ [ 761, 774 ] ], [ [ 999, 1011 ] ] ]
# # Copyright 2018 Analytics Zoo Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from tensorflow.keras.models import Model from tensorflow.keras.layers import Input, LSTM, Dense import tensorflow.keras as keras from zoo.automl.model.abstract import BaseModel from zoo.automl.common.util import * from zoo.automl.common.metrics import Evaluator class LSTMSeq2Seq(BaseModel): def __init__(self, check_optional_config=True, future_seq_len=2): """ Constructor of LSTM Seq2Seq model """ self.model = None self.past_seq_len = None self.future_seq_len = future_seq_len self.feature_num = None self.target_col_num = None self.metric = None self.latent_dim = None self.batch_size = None self.check_optional_config = check_optional_config def _build_train(self, mc=False, **config): """ build LSTM Seq2Seq model :param config: :return: """ super()._check_config(**config) self.metric = config.get('metric', 'mean_squared_error') self.latent_dim = config.get('latent_dim', 128) self.dropout = config.get('dropout', 0.2) self.lr = config.get('lr', 0.001) # for restore in continuous training self.batch_size = config.get('batch_size', 64) training = True if mc else None # Define an input sequence and process it. self.encoder_inputs = Input(shape=(None, self.feature_num), name="encoder_inputs") encoder = LSTM(units=self.latent_dim, dropout=self.dropout, return_state=True, name="encoder_lstm") encoder_outputs, state_h, state_c = encoder(self.encoder_inputs, training=training) # We discard `encoder_outputs` and only keep the states. self.encoder_states = [state_h, state_c] # Set up the decoder, using `encoder_states` as initial state. self.decoder_inputs = Input(shape=(None, self.target_col_num), name="decoder_inputs") # We set up our decoder to return full output sequences, # and to return internal states as well. We don't use the # return states in the training model, but we will use them in inference. self.decoder_lstm = LSTM(self.latent_dim, dropout=self.dropout, return_sequences=True, return_state=True, name="decoder_lstm") decoder_outputs, _, _ = self.decoder_lstm(self.decoder_inputs, training=training, initial_state=self.encoder_states) self.decoder_dense = Dense(self.target_col_num, name="decoder_dense") decoder_outputs = self.decoder_dense(decoder_outputs) # Define the model that will turn # `encoder_input_data` & `decoder_input_data` into `decoder_target_data` self.model = Model([self.encoder_inputs, self.decoder_inputs], decoder_outputs) self.model.compile(loss='mse', metrics=[self.metric], optimizer=keras.optimizers.RMSprop(lr=self.lr)) return self.model def _restore_model(self): self.encoder_inputs = self.model.input[0] # input_1 encoder_outputs, state_h_enc, state_c_enc = self.model.layers[2].output # lstm_1 self.encoder_states = [state_h_enc, state_c_enc] self.decoder_inputs = self.model.input[1] # input_2 self.decoder_lstm = self.model.layers[3] self.decoder_dense = self.model.layers[4] def _build_inference(self, mc=False): training = True if mc else None # from our previous model - mapping encoder sequence to state vectors encoder_model = Model(self.encoder_inputs, self.encoder_states) # A modified version of the decoding stage that takes in predicted target inputs # and encoded state vectors, returning predicted target outputs and decoder state vectors. # We need to hang onto these state vectors to run the next step of the inference loop. decoder_state_input_h = Input(shape=(self.latent_dim,)) decoder_state_input_c = Input(shape=(self.latent_dim,)) decoder_states_inputs = [decoder_state_input_h, decoder_state_input_c] decoder_outputs, state_h, state_c = self.decoder_lstm(self.decoder_inputs, training=training, initial_state=decoder_states_inputs) decoder_states = [state_h, state_c] decoder_outputs = self.decoder_dense(decoder_outputs) decoder_model = Model([self.decoder_inputs] + decoder_states_inputs, [decoder_outputs] + decoder_states) return encoder_model, decoder_model def _decode_sequence(self, input_seq, mc=False): encoder_model, decoder_model = self._build_inference(mc=mc) # Encode the input as state vectors. states_value = encoder_model.predict(input_seq) # Generate empty target sequence of length 1. target_seq = np.zeros((len(input_seq), 1, self.target_col_num)) # Populate the first target sequence with end of encoding series value target_seq[:, 0] = input_seq[:, -1, :self.target_col_num] # Sampling loop for a batch of sequences - we will fill decoded_seq with predictions # (to simplify, here we assume a batch of size 1). decoded_seq = np.zeros((len(input_seq), self.future_seq_len, self.target_col_num)) for i in range(self.future_seq_len): output, h, c = decoder_model.predict([target_seq] + states_value) decoded_seq[:, i] = output[:, 0] # Update the target sequence (of length 1). target_seq = np.zeros((len(input_seq), 1, self.target_col_num)) target_seq[:, 0] = output[:, 0] # Update states states_value = [h, c] return decoded_seq def _get_decoder_inputs(self, x, y): """ lagged target series for teacher forcing decoder_input data is one timestamp ahead of y :param x: 3-d array in format of (sample_num, past_sequence_len, feature_num) :param y: 3-d array in format of (sample_num, future_sequence_len, target_col_num) Need to expand dimension if y is a 2-d array with one target col :return: 3-d array of decoder inputs """ decoder_input_data = np.zeros(y.shape) decoder_input_data[1:, ] = y[:-1, ] decoder_input_data[0, 0] = x[-1, -1, :self.target_col_num] decoder_input_data[0, 1:] = y[0, :-1] return decoder_input_data def _get_len(self, x, y): self.past_seq_len = x.shape[1] self.feature_num = x.shape[2] # self.future_seq_len = y.shape[1] self.target_col_num = y.shape[2] def _expand_y(self, y): """ expand dims for y. :param y: :return: """ while len(y.shape) < 3: y = np.expand_dims(y, axis=2) return y def _pre_processing(self, x, y, validation_data): """ pre_process input data. 1. expand dims for y and val_y 2. get decoder inputs for train data 3. get decoder inputs for validation data :param x: train_x :param y: train_y :param validation_data: :return: network input """ y = self._expand_y(y) self._get_len(x, y) decoder_input_data = self._get_decoder_inputs(x, y) if validation_data is not None: val_x, val_y = validation_data val_y = self._expand_y(val_y) val_decoder_input = self._get_decoder_inputs(val_x, val_y) validation_data = ([val_x, val_decoder_input], val_y) return x, y, decoder_input_data, validation_data def fit_eval(self, data, validation_data=None, mc=False, verbose=0, **config): """ fit for one iteration :param data: could be a tuple with numpy ndarray with form (x, y) x: 3-d array in format (no. of samples, past sequence length, 2+feature length), in the last dimension, the 1st col is the time index (data type needs to be numpy datetime type, e.g. "datetime64"), the 2nd col is the target value (data type should be numeric) y: 2-d numpy array in format (no. of samples, future sequence length) if future sequence length > 1, or 1-d numpy array in format (no. of samples, ) if future sequence length = 1 :param validation_data: tuple in format (x_test,y_test), data used for validation. If this is specified, validation result will be the optimization target for automl. Otherwise, train metric will be the optimization target. :param config: optimization hyper parameters :return: the resulting metric """ x, y = data[0], data[1] x, y, decoder_input_data, validation_data = self._pre_processing(x, y, validation_data) # if model is not initialized, __build the model if self.model is None: self._build_train(mc=mc, **config) # batch_size = config.get('batch_size', 64) # lr = self.lr # name = "seq2seq-batch_size-{}-epochs-{}-lr-{}-time-{}"\ # .format(batch_size, epochs, lr, time()) # tensorboard = TensorBoard(log_dir="logs/" + name) hist = self.model.fit([x, decoder_input_data], y, validation_data=validation_data, batch_size=self.batch_size, epochs=config.get("epochs", 10), verbose=verbose, # callbacks=[tensorboard] ) # print(hist.history) if validation_data is None: # get train metrics # results = self.model.evaluate(x, y) result = hist.history.get(self.metric)[-1] else: result = hist.history.get('val_' + str(self.metric))[-1] return result def evaluate(self, x, y, metric=['mse']): """ Evaluate on x, y :param x: input :param y: target :param metric: a list of metrics in string format :return: a list of metric evaluation results """ y_pred = self.predict(x) # y = np.squeeze(y, axis=2) if self.target_col_num == 1: return [Evaluator.evaluate(m, y, y_pred) for m in metric] else: return [np.array([Evaluator.evaluate(m, y[:, i, :], y_pred[:, i, :]) for i in range(self.future_seq_len)]) for m in metric] def predict(self, x, mc=False): """ Prediction on x. :param x: input :return: predicted y (expected dimension = 2) """ y_pred = self._decode_sequence(x, mc=mc) if self.target_col_num == 1: y_pred = np.squeeze(y_pred, axis=2) return y_pred def predict_with_uncertainty(self, x, n_iter=100): result = np.array([self.predict(x, mc=True) for i in range(n_iter)]) prediction = result.mean(axis=0) uncertainty = result.var(axis=0) return prediction, uncertainty def save(self, model_path, config_path): """ save model to file. :param model_path: the model file path to be saved to. :param config_path: the config file path to be saved to. :return: """ self.model.save(model_path) config_to_save = {"past_seq_len": self.past_seq_len, "feature_num": self.feature_num, "future_seq_len": self.future_seq_len, "target_col_num": self.target_col_num, "metric": self.metric, "latent_dim": self.latent_dim, "batch_size": self.batch_size} save_config(config_path, config_to_save) def restore(self, model_path, **config): """ restore model from file :param model_path: the model file :param config: the trial config :return: the restored model """ self.past_seq_len = config["past_seq_len"] self.feature_num = config["feature_num"] self.future_seq_len = config["future_seq_len"] self.target_col_num = config["target_col_num"] self.metric = config["metric"] self.latent_dim = config["latent_dim"] self.batch_size = config["batch_size"] self.model = keras.models.load_model(model_path) self._restore_model() # self.model.load_weights(file_path) def _get_required_parameters(self): return { # 'input_shape_x', # 'input_shape_y', # 'out_units' } def _get_optional_parameters(self): return { 'past_seq_len' 'latent_dim' 'dropout', 'metric', 'lr', 'epochs', 'batch_size' }
[ [ [ 627, 632 ], [ 3565, 3570 ], [ 4408, 4413 ], [ 5342, 5347 ] ], [ [ 669, 674 ], [ 1964, 1969 ], [ 2510, 2515 ], [ 4772, 4777 ], [ 4836, 4841 ] ], [ [ 676, 680 ], [ 2043, 2047 ], [ 2815, 2819 ] ], [ [ 682, 687 ], [ 3309, 3314 ] ], [ [ 695, 720 ], [ 3758, 3763 ], [ 13381, 13386 ] ], [ [ 760, 769 ], [ 875, 884 ] ], [ [ 805, 806 ], [ 5804, 5806 ], [ 6177, 6179 ], [ 6498, 6500 ], [ 7188, 7190 ], [ 7753, 7755 ], [ 11312, 11314 ], [ 11749, 11751 ], [ 11871, 11873 ], [ 12754, 12765 ] ], [ [ 845, 854 ], [ 11228, 11237 ], [ 11322, 11331 ] ], [ [ 863, 874 ] ] ]
from src.layers.LayerHelper import * from settings import LayerSettings as layerSettings import tensorflow as tf import os CUDA_VISIBLE_DEVICES=0 os.environ["CUDA_VISIBLE_DEVICES"] = "0" # set gpu number def LSTM(name_, inputTensor_, numberOfOutputs_, isTraining_, dropoutProb_=None): with tf.name_scope(name_): cell = tf.nn.rnn_cell.LSTMCell(num_units=numberOfOutputs_, use_peepholes=True, initializer=layerSettings.LSTM_INITIALIZER, forget_bias=1.0, state_is_tuple=True, activation=tf.nn.tanh, name=name_+"_cell") if dropoutProb_ != None: dropoutProbTensor = tf.cond(isTraining_, lambda: 0.5, lambda: 1.0) cell = tf.nn.rnn_cell.DropoutWrapper(cell, input_keep_prob=dropoutProbTensor, output_keep_prob=dropoutProbTensor) statePlaceHolder = tf.nn.rnn_cell.LSTMStateTuple( tf.placeholder(layerSettings.FLOAT_TYPE, [None, numberOfOutputs_]), tf.placeholder(layerSettings.FLOAT_TYPE, [None, numberOfOutputs_]) ) outputTensor, stateTensor = tf.nn.dynamic_rnn( cell=cell, initial_state=statePlaceHolder, inputs=inputTensor_) # Add Regularization Loss for eachVariable in tf.trainable_variables(): if name_ in eachVariable.name: if ('bias' not in eachVariable.name)and(layerSettings.REGULARIZER_WEIGHTS_DECAY != None): regularizationLoss = L2_Regularizer(eachVariable) tf.losses.add_loss(regularizationLoss, loss_collection=tf.GraphKeys.REGULARIZATION_LOSSES) return outputTensor, stateTensor, statePlaceHolder
[ [ [ 35, 36 ], [ 1362, 1376 ] ], [ [ 58, 88 ], [ 421, 434 ], [ 869, 882 ], [ 947, 960 ], [ 1286, 1299 ] ], [ [ 96, 112 ], [ 292, 294 ], [ 323, 325 ], [ 523, 525 ], [ 613, 615 ], [ 670, 672 ], [ 823, 825 ], [ 854, 856 ], [ 932, 934 ], [ 1032, 1034 ], [ 1182, 1184 ], [ 1396, 1398 ], [ 1451, 1453 ] ], [ [ 120, 122 ], [ 146, 148 ] ], [ [ 123, 143 ] ], [ [ 209, 213 ] ] ]
"""LCM type definitions This file automatically generated by lcm. DO NOT MODIFY BY HAND!!!! """ import cStringIO as StringIO import struct class request_t(object): __slots__ = ["utime"] def __init__(self): self.utime = 0 def encode(self): buf = StringIO.StringIO() buf.write(request_t._get_packed_fingerprint()) self._encode_one(buf) return buf.getvalue() def _encode_one(self, buf): buf.write(struct.pack(">q", self.utime)) def decode(data): if hasattr(data, 'read'): buf = data else: buf = StringIO.StringIO(data) if buf.read(8) != request_t._get_packed_fingerprint(): raise ValueError("Decode error") return request_t._decode_one(buf) decode = staticmethod(decode) def _decode_one(buf): self = request_t() self.utime = struct.unpack(">q", buf.read(8))[0] return self _decode_one = staticmethod(_decode_one) _hash = None def _get_hash_recursive(parents): if request_t in parents: return 0 tmphash = (0xa686a0e0f882d897) & 0xffffffffffffffff tmphash = (((tmphash<<1)&0xffffffffffffffff) + (tmphash>>63)) & 0xffffffffffffffff return tmphash _get_hash_recursive = staticmethod(_get_hash_recursive) _packed_fingerprint = None def _get_packed_fingerprint(): if request_t._packed_fingerprint is None: request_t._packed_fingerprint = struct.pack(">Q", request_t._get_hash_recursive([])) return request_t._packed_fingerprint _get_packed_fingerprint = staticmethod(_get_packed_fingerprint)
[ [ [ 104, 125 ], [ 277, 285 ], [ 606, 614 ] ], [ [ 133, 139 ], [ 463, 469 ], [ 889, 895 ], [ 1484, 1490 ] ], [ [ 147, 156 ], [ 315, 324 ], [ 656, 665 ], [ 753, 762 ], [ 856, 865 ], [ 1056, 1065 ], [ 1401, 1410 ], [ 1502, 1511 ], [ 1452, 1461 ], [ 1552, 1561 ] ] ]
from conans import ConanFile, CMake import os channel = os.getenv("CONAN_CHANNEL", "testing") username = os.getenv("CONAN_USERNAME", "memsharded") class EasyLoggingTestConan(ConanFile): settings = "os", "compiler", "build_type", "arch" requires = "easyloggingpp/9.94.1@%s/%s" % (username, channel) generators = "cmake" def build(self): cmake = CMake(self.settings) self.run('cmake "%s" %s' % (self.conanfile_directory, cmake.command_line)) self.run("cmake --build . %s" % cmake.build_config) def imports(self): self.copy("*.cc") def test(self): os.chdir("bin") self.run(".%sexample" % os.sep)
[ [ [ 19, 28 ], [ 176, 185 ] ], [ [ 30, 35 ], [ 371, 376 ] ], [ [ 43, 45 ], [ 57, 59 ], [ 106, 108 ], [ 614, 616 ], [ 662, 664 ] ], [ [ 47, 54 ], [ 299, 306 ] ], [ [ 95, 103 ], [ 289, 297 ] ], [ [ 155, 175 ] ] ]
# Author: Guilherme Aldeia # Contact: [email protected] # Version: 1.0.1 # Last modified: 06-07-2021 by Guilherme Aldeia """Interaction Transformation expression's **Inspector** Sub-module containing three classes to help inspect and explain the results obtained with the itea. - ``ITExpr_explainer``: Implementations of feature importances methods specific to the Interaction-Transformation representation, and several visualization tools to help interpret the final expression; - ``ITExpr_inspector``: Based on a more statistical approach, this class implements methods to measure the quality of the final expression by calculating information between individual terms; - ``ITExpr_texifier``: Creation of latex representations of the final expression and its derivatives. In cases where the final expression is simple enough, the analysis of the expression can provide useful insights. All the modules are designed to work with `ITExpr`s. After the evolutionary process is performed (by calling `fit()` on the `ITEA_classifier` or `ITEA_regressor`), the best final expression can be accessed by `itea.bestsol_`, and those classes are specialized in different ways of inspecting the final model. Additionally, there is one class designed to work with the ´`itea``, instead of ``ITExpr`` expressions. The class ``ITEA_summarizer`` implements a method to automatically create a pdf file containing information generated with all the inspection classes, in an attempt to automate the task of generating an interpretability report. """ from itea.inspection._ITExpr_explainer import ITExpr_explainer from itea.inspection._ITExpr_inspector import ITExpr_inspector from itea.inspection._ITExpr_texifier import ITExpr_texifier from itea.inspection._ITEA_summarizer import ITEA_summarizer import jax # Must be used at startup. We'll perform lightweight usage with jax jax.config.update('jax_platform_name', 'cpu') __all__ = [ 'ITExpr_explainer', 'ITExpr_inspector', 'ITExpr_texifier', 'ITEA_summarizer' ]
[ [ [ 1650, 1666 ] ], [ [ 1714, 1730 ] ], [ [ 1778, 1793 ] ], [ [ 1841, 1856 ] ], [ [ 1867, 1870 ], [ 1946, 1949 ] ], [ [ 1997, 2004 ] ] ]
# write your first unittest! import unittest from ovos_plugin_manager.skills import find_skill_plugins class TestPlugin(unittest.TestCase): @classmethod def setUpClass(self): self.skill_id = "ovos-skill-timer.OpenVoiceOS" def test_find_plugin(self): plugins = find_skill_plugins() self.assertIn(self.skill_id, list(plugins))
[ [ [ 36, 44 ], [ 122, 130 ] ], [ [ 84, 102 ], [ 291, 309 ] ], [ [ 111, 121 ] ] ]
from keras.preprocessing.image import ImageDataGenerator from keras.models import Sequential from keras.layers import Convolution2D, MaxPooling2D from keras.layers import Activation, Dropout, Flatten, Dense, Lambda, ELU from keras.optimizers import Adam from sklearn.model_selection import train_test_split from keras.models import model_from_json from sklearn.preprocessing import normalize import cv2 import numpy as np import glob import json from keras.layers import merge from keras.layers.core import Lambda from keras.models import Model import tensorflow as tf def make_parallel(model, gpu_count): def get_slice(data, idx, parts): shape = tf.shape(data) size = tf.concat(0, [shape[:1] // parts, shape[1:]]) stride = tf.concat(0, [shape[:1] // parts, shape[1:] * 0]) start = stride * idx return tf.slice(data, start, size) outputs_all = [] for i in range(len(model.outputs)): outputs_all.append([]) # Place a copy of the model on each GPU, each getting a slice of the batch for i in range(gpu_count): with tf.device('/gpu:%d' % i): with tf.name_scope('tower_%d' % i) as scope: inputs = [] # Slice each input into a piece for processing on this GPU for x in model.inputs: input_shape = tuple(x.get_shape().as_list())[1:] slice_n = Lambda(get_slice, output_shape=input_shape, arguments={'idx': i, 'parts': gpu_count})(x) inputs.append(slice_n) outputs = model(inputs) if not isinstance(outputs, list): outputs = [outputs] # Save all the outputs for merging back together later for l in range(len(outputs)): outputs_all[l].append(outputs[l]) # merge outputs on CPU with tf.device('/cpu:0'): merged = [] for outputs in outputs_all: merged.append(merge(outputs, mode='concat', concat_axis=0)) return Model(input=model.inputs, output=merged) class CNNClassifier: def __init__(self): self.classifier = None def get_model(self, parallel=False): model = Sequential() #model.add(Lambda(lambda x: x / 127.5 - 1., input_shape=(64, 64, 3))) model.add(Convolution2D(8, 8, 8, subsample=(4, 4), border_mode="same", activation='elu', name='Conv1')) model.add(Convolution2D(16, 5, 5, subsample=(2, 2), border_mode="same", activation='elu', name='Conv2')) model.add(Convolution2D(32, 5, 5, subsample=(2, 2), border_mode="same", activation='elu', name='Conv3')) model.add(Flatten()) model.add(ELU()) model.add(Dense(1024, activation='elu')) model.add(Dropout(.5)) model.add(ELU()) model.add(Dense(512, activation='elu')) model.add(Dropout(.5)) model.add(Dense(1, name='output')) model.add(Activation('sigmoid')) if parallel: model = make_parallel(model, 2) #model.compile(optimizer='sgd', loss='binary_crossentropy', metrics=['accuracy']) self.model = model return model def _model(self): img_width, img_height = 64, 64 model = Sequential() model.add(Convolution2D(8, 3, 3, input_shape=(img_width, img_height, 3))) model.add(Activation('elu')) model.add(MaxPooling2D(pool_size=(2, 2))) #model.add(Convolution2D(16, 3, 3)) #model.add(Activation('elu')) #model.add(MaxPooling2D(pool_size=(2, 2))) #model.add(Convolution2D(32, 3, 3)) #model.add(Activation('elu')) #model.add(MaxPooling2D(pool_size=(2, 2))) model.add(Flatten()) model.add(Dense(512)) model.add(Dropout(0.5)) model.add(Dense(1, activation='sigmoid')) #model = make_parallel(model, 2) self.model = model def compile(self): self.model.compile(loss='binary_crossentropy', optimizer='rmsprop', class_mode='binary', metrics=['accuracy']) def save(self): model_json = self.model.to_json() with open("./model.json", "w") as json_file: json.dump(model_json, json_file) self.model.save_weights("./model.h5") print("Saved model to disk") def load(self): with open('./model.json', 'r') as jfile: self.model = model_from_json(json.load(jfile)) self.compile() self.model.load_weights('./model.h5') def get_list(self): vehicles = np.array(glob.glob('training_data/vehicles/*/*')) y_vehicles = np.zeros(vehicles.shape) + 1 non_vehicles = np.array(glob.glob('training_data/non-vehicles/*/*')) y_non_vehicles = np.zeros(non_vehicles.shape) X_data = np.concatenate((vehicles, non_vehicles)) Y_data = np.concatenate((y_vehicles, y_non_vehicles)) return X_data, Y_data def predict(self, image): #img = np.copy(image) #img = cv2.resize(img, (64, 64)) x = image[None, :, :, :] result = self.model.predict(x, 1) return result def train(self, file_list, labels, test_size=0.2, nb_epoch=30, batch_size=128): X_train, X_test, Y_train, Y_test = train_test_split(file_list, labels, test_size=test_size, random_state=100) test_images = build_images(X_test) train_images = build_images(X_train) train_datagen = ImageDataGenerator( rescale=1. / 255, shear_range=0.05, zoom_range=0.05, width_shift_range=0.1, height_shift_range=0.1, rotation_range=5, horizontal_flip=True) test_datagen = ImageDataGenerator(rescale=1. / 255) train_generator = train_datagen.flow(train_images, Y_train, batch_size) test_generator = test_datagen.flow(test_images, Y_test, batch_size) nb_train_samples = (batch_size-1)*100 nb_validation_samples = (batch_size-1)*20 #self.get_model(parallel=False) self._model() self.compile() self.model.fit_generator( train_generator, samples_per_epoch=nb_train_samples, nb_epoch=nb_epoch, show_accuracy=True, validation_data=test_generator, nb_val_samples=nb_validation_samples) def build_images(x): images = np.zeros((len(x), 64, 64, 3)) for idx, img_fname in enumerate(x): im = cv2.imread(img_fname) im = cv2.cvtColor(im, cv2.COLOR_BGR2RGB) im = cv2.resize(im, (64, 64), interpolation=cv2.INTER_AREA) images[idx] = im return images def do_all(nb_epoch=30, batch_size=256): clf = CNNClassifier() x, y = clf.get_list() clf.train(x, y, nb_epoch=nb_epoch, batch_size=batch_size) clf.save()
[ [ [ 38, 56 ], [ 5497, 5515 ], [ 5764, 5782 ] ], [ [ 82, 92 ], [ 2234, 2244 ], [ 3266, 3276 ] ], [ [ 118, 131 ], [ 2343, 2356 ], [ 2455, 2468 ], [ 2568, 2581 ], [ 3297, 3310 ] ], [ [ 133, 145 ], [ 3416, 3428 ] ], [ [ 171, 181 ], [ 2962, 2972 ], [ 3379, 3389 ] ], [ [ 183, 190 ], [ 2784, 2791 ], [ 2888, 2895 ], [ 3794, 3801 ] ], [ [ 192, 199 ], [ 2681, 2688 ], [ 3735, 3742 ] ], [ [ 201, 206 ], [ 2735, 2740 ], [ 2840, 2845 ], [ 2919, 2924 ], [ 3764, 3769 ], [ 3826, 3831 ] ], [ [ 208, 214 ] ], [ [ 216, 219 ], [ 2710, 2713 ], [ 2815, 2818 ] ], [ [ 249, 253 ] ], [ [ 290, 306 ], [ 5308, 5324 ] ], [ [ 332, 347 ], [ 4452, 4467 ] ], [ [ 382, 391 ] ], [ [ 399, 402 ], [ 6515, 6518 ], [ 6550, 6553 ], [ 6567, 6570 ], [ 6599, 6602 ], [ 6638, 6641 ] ], [ [ 410, 421 ], [ 4600, 4602 ], [ 4671, 4673 ], [ 4723, 4725 ], [ 4802, 4804 ], [ 4848, 4850 ], [ 4906, 4908 ], [ 6432, 6434 ] ], [ [ 429, 433 ], [ 4609, 4613 ], [ 4732, 4736 ] ], [ [ 441, 445 ], [ 4241, 4245 ], [ 4468, 4472 ] ], [ [ 471, 476 ], [ 1995, 2000 ] ], [ [ 507, 513 ], [ 1419, 1425 ] ], [ [ 539, 544 ], [ 2057, 2062 ] ], [ [ 553, 569 ], [ 1094, 1096 ], [ 1137, 1139 ], [ 1892, 1894 ], [ 662, 664 ], [ 692, 694 ], [ 755, 757 ], [ 849, 851 ] ], [ [ 576, 589 ], [ 3026, 3039 ] ], [ [ 2106, 2119 ], [ 6749, 6762 ] ], [ [ 6402, 6414 ], [ 5406, 5418 ], [ 5450, 5462 ] ], [ [ 6702, 6708 ] ] ]
import pytest from brownie import interface def test_uniswap_add_two_tokens( admin, alice, chain, bank, werc20, ufactory, urouter, simple_oracle, oracle, celo, cusd, ceur, UniswapV2SpellV1, UniswapV2Oracle, core_oracle ): spell = UniswapV2SpellV1.deploy(bank, werc20, urouter, celo, {'from': admin}) cusd.mint(admin, 10000000 * 10**6, {'from': admin}) ceur.mint(admin, 10000000 * 10**6, {'from': admin}) cusd.approve(urouter, 2**256-1, {'from': admin}) ceur.approve(urouter, 2**256-1, {'from': admin}) urouter.addLiquidity( cusd, ceur, 1000000 * 10**6, 1000000 * 10**6, 0, 0, admin, chain.time() + 60, {'from': admin}, ) lp = ufactory.getPair(cusd, ceur) print('admin lp bal', interface.IERC20(lp).balanceOf(admin)) uniswap_lp_oracle = UniswapV2Oracle.deploy(core_oracle, {'from': admin}) print('ceur Px', simple_oracle.getCELOPx(ceur)) print('cusd Px', simple_oracle.getCELOPx(cusd)) core_oracle.setRoute([cusd, ceur, lp], [simple_oracle, simple_oracle, uniswap_lp_oracle]) print('lp Px', uniswap_lp_oracle.getCELOPx(lp)) oracle.setTokenFactors( [cusd, ceur, lp], [ [10000, 10000, 10000], [10000, 10000, 10000], [10000, 10000, 10000], ], {'from': admin}, ) cusd.mint(alice, 10000000 * 10**6, {'from': admin}) ceur.mint(alice, 10000000 * 10**6, {'from': admin}) cusd.approve(bank, 2**256-1, {'from': alice}) ceur.approve(bank, 2**256-1, {'from': alice}) spell.getAndApprovePair(cusd, ceur, {'from': admin}) lp = ufactory.getPair(cusd, ceur) spell.setWhitelistLPTokens([lp], [True], {'from': admin}) bank.setWhitelistSpells([spell], [True], {'from': admin}) bank.setWhitelistTokens([cusd, ceur], [True, True], {'from': admin}) tx = bank.execute( 0, spell, spell.addLiquidityWERC20.encode_input( ceur, # token 0 cusd, # token 1 [ 40000 * 10**6, # 40000 ceur 50000 * 10**6, # 50000 cusd 0, 1000 * 10**6, # 1000 ceur 200 * 10**6, # 200 cusd 0, # borrow LP tokens 0, # min ceur 0, # min cusd ], ), {'from': alice} ) position_id = tx.return_value print('tx gas used', tx.gas_used) print('bank collateral size', bank.getPositionInfo(position_id)) print('bank collateral value', bank.getCollateralCELOValue(position_id)) print('bank borrow value', bank.getBorrowCELOValue(position_id)) print('bank ceur', bank.getBankInfo(ceur)) print('bank cusd', bank.getBankInfo(cusd)) print('ceur Px', simple_oracle.getCELOPx(ceur)) print('cusd Px', simple_oracle.getCELOPx(cusd)) print('lp Px', uniswap_lp_oracle.getCELOPx(lp))
[ [ [ 7, 13 ] ], [ [ 34, 43 ], [ 792, 801 ] ], [ [ 50, 77 ] ] ]
""" This module patches a few core functions to add compression capabilities, since gevent-websocket does not appear to be maintained anymore. """ from socket import error from zlib import ( decompressobj, MAX_WBITS, Z_FULL_FLUSH, ) from geventwebsocket.exceptions import ( ProtocolError, WebSocketError, ) from geventwebsocket.websocket import ( MSG_SOCKET_DEAD, Header, WebSocket, ) DECOMPRESSOR = decompressobj(-MAX_WBITS) def _encode_bytes(text): if isinstance(text, str): return text if not isinstance(text, unicode): text = unicode(text or '') return text.encode('utf-8') def make_compressed_frame(message, compressor): """ Make a compressed websocket frame from a message and compressor. Generates header and a compressed message which can then be used on any websocket connection where `no_context_takeover` has been negotiated. This prevents the need to re-compress a broadcast-style message for every websocket connection. `compressor` is a zlib compressor object. """ binary = not isinstance(message, (str, unicode)) opcode = WebSocket.OPCODE_BINARY if binary else WebSocket.OPCODE_TEXT if binary: message = str(message) else: message = _encode_bytes(message) message = compressor.compress(message) # We use Z_FULL_FLUSH (rather than Z_SYNC_FLUSH) here when # server_no_context_takeover has been passed, to reset the context at # the end of every frame. Patches to the actual gevent-websocket # library should probably be able to support both. message += compressor.flush(Z_FULL_FLUSH) # See https://tools.ietf.org/html/rfc7692#page-19 if message.endswith('\x00\x00\xff\xff'): message = message[:-4] # Generate header. The RSV0 bit indicates the payload is compressed. flags = Header.RSV0_MASK header = Header.encode_header( fin=True, opcode=opcode, mask='', length=len(message), flags=flags) return header + message def send_raw_frame(websocket, raw_message): """ `raw_message` includes both the header and the encoded message. """ try: websocket.raw_write(raw_message) except error: websocket.current_app.on_close(MSG_SOCKET_DEAD) raise WebSocketError(MSG_SOCKET_DEAD) def read_frame(websocket): # Patched `read_frame` method that supports decompression header = Header.decode_header(websocket.stream) # Start patched lines compressed = header.flags & header.RSV0_MASK if compressed: header.flags &= ~header.RSV0_MASK # End patched lines if header.flags: raise ProtocolError if not header.length: return header, '' try: payload = websocket.raw_read(header.length) except error: payload = '' except Exception: # Start patched lines raise WebSocketError('Could not read payload') # End patched lines if len(payload) != header.length: raise WebSocketError('Unexpected EOF reading frame payload') if header.mask: payload = header.unmask_payload(payload) # Start patched lines if compressed: payload = ''.join(( DECOMPRESSOR.decompress(payload), DECOMPRESSOR.decompress('\0\0\xff\xff'), DECOMPRESSOR.flush(), )) # End patched lines return header, payload
[ [ [ 166, 171 ], [ 2222, 2227 ], [ 2812, 2817 ] ], [ [ 195, 208 ], [ 435, 448 ] ], [ [ 214, 223 ], [ 450, 459 ] ], [ [ 229, 241 ], [ 1643, 1655 ] ], [ [ 291, 304 ], [ 2672, 2685 ] ], [ [ 310, 324 ], [ 2299, 2313 ], [ 2907, 2921 ], [ 3029, 3043 ] ], [ [ 372, 387 ], [ 2268, 2283 ], [ 2314, 2329 ] ], [ [ 393, 399 ], [ 1874, 1880 ], [ 1904, 1910 ], [ 2436, 2442 ] ], [ [ 405, 414 ], [ 1148, 1157 ], [ 1187, 1196 ] ], [ [ 420, 432 ], [ 3240, 3252 ], [ 3286, 3298 ], [ 3339, 3351 ] ], [ [ 467, 480 ], [ 1283, 1296 ] ], [ [ 651, 672 ] ], [ [ 2037, 2051 ] ], [ [ 2337, 2347 ] ] ]
import os import json Environ = os._Environ def is_on_cloudfoundry(env: Environ=os.environ) -> bool: return 'VCAP_SERVICES' in env def load_cups_from_vcap_services(name: str, env: Environ=os.environ) -> None: ''' Detects if VCAP_SERVICES exists in the environment; if so, parses it and imports all the credentials from the given custom user-provided service (CUPS) as strings into the environment. For more details on CUPS, see: https://docs.cloudfoundry.org/devguide/services/user-provided.html ''' if not is_on_cloudfoundry(env): return vcap = json.loads(env['VCAP_SERVICES']) for entry in vcap.get('user-provided', []): if entry['name'] == name: for key, value in entry['credentials'].items(): env[key] = value def load_database_url_from_vcap_services(name: str, service: str, env: Environ=os.environ) -> str: """ Sets os.environ[DATABASE_URL] from a service entry in VCAP_SERVICES. """ if not is_on_cloudfoundry(env): return # FIXME: this'll break if there are multiple databases. Not an issue right # now, but could be in the future. Keep an eye on it. vcap = json.loads(env['VCAP_SERVICES']) env['DATABASE_URL'] = vcap[service][0]["credentials"]["uri"]
[ [ [ 7, 9 ], [ 33, 35 ], [ 83, 85 ], [ 197, 199 ], [ 932, 934 ] ], [ [ 17, 21 ], [ 601, 605 ], [ 1241, 1245 ] ], [ [ 23, 30 ], [ 75, 82 ], [ 189, 196 ], [ 924, 931 ] ], [ [ 51, 69 ], [ 549, 567 ], [ 1052, 1070 ] ], [ [ 144, 172 ] ], [ [ 816, 852 ] ] ]
from pymongo import MongoClient from pymongo import ReadPreference from datetime import datetime, timedelta class Mongo(MongoClient): def __init__(self, username, password, host, db='tags', collection='tweets_pipeline_v2'): uri = f"mongodb://{username}:{password}@{host}/{db}" super(Mongo, self).__init__(host=uri, authSource=db, authMechanism='SCRAM-SHA-256', port=27017, replicaset="rs0", read_preference=ReadPreference.SECONDARY, ) self.database = self.get_default_database() self.collection = collection def pipelined(self, count=True): query = {"status": "pipelined"} if count: return self.database[self.collection].count_documents(query) return self.database[self.collection].find(query) def feed(self, count=True): query = {"status": "graphicone_feed"} if count: return self.database[self.collection].count_documents(query) return self.database[self.collection].find(query) def search(self, count=True): query = {"status": "graphicone_search"} if count: return self.database[self.collection].count_documents(query) return self.database[self.collection].find(query) def left_for_analysts(self, count=True): query = {"in_app": {"$exists": False}, "status": "graphicone_feed"} if count: return self.database[self.collection].count_documents(query) return self.database[self.collection].find(query) def removed_validators(self, count=True): query = {"validator_username": {"$exists": True}, "status": "deleted"} if count: return self.database[self.collection].count_documents(query) return self.database[self.collection].find(query) def removed_analysts(self, count=True): query = {"status": "deleted_from_analytics"} if count: return self.database[self.collection].count_documents(query) return self.database[self.collection].find(query) # if __name__ == "__main__": # _username = "login" # _password = "passwd" # mongodb_host = "host address" # # mongo_client = Mongo(_username, _password, mongodb_host) # print(mongo_client.pipelined()) # print(mongo_client.search()) # print(mongo_client.feed()) # print(mongo_client.left_for_analysts()) # print(mongo_client.removed_validators()) # print(mongo_client.removed_analysts())
[ [ [ 20, 31 ], [ 122, 133 ] ], [ [ 52, 66 ], [ 610, 624 ] ], [ [ 88, 96 ] ], [ [ 98, 107 ] ], [ [ 116, 121 ], [ 306, 311 ] ] ]
# Data Preprocessing Template # Importing the libraries import numpy as np import matplotlib.pyplot as plt import pandas as pd # Importing the dataset dataset = pd.read_csv('50_Startups.csv') X = dataset.iloc[:, :-1].values y = dataset.iloc[:, 4].values #encoding independent variable state #from sklearn.preprocessing import LabelEncoder, OneHotEncoder #labelencoder_X = LabelEncoder() #X[:, 3] = labelencoder_X.fit_transform(X[:, 3]) #onehotencoder = OneHotEncoder(categorical_features = [3]) #X = onehotencoder.fit_transform(X).toarray() from sklearn.preprocessing import LabelEncoder,OneHotEncoder from sklearn.compose import ColumnTransformer ct = ColumnTransformer([("State", OneHotEncoder(), [3])], remainder = 'passthrough') X= ct.fit_transform(X) #avoiding the dummy variable trap X=X[:,1:] # Splitting the dataset into the Training set and Test set from sklearn.model_selection import train_test_split X_train, X_test, y_train, y_test = train_test_split(X, y, test_size = 0.2, random_state = 0) # Feature Scaling """from sklearn.preprocessing import StandardScaler sc_X = StandardScaler() X_train = sc_X.fit_transform(X_train) X_test = sc_X.transform(X_test) sc_y = StandardScaler() y_train = sc_y.fit_transform(y_train)""" #fitting multiple linear regression to the training set from sklearn.linear_model import LinearRegression regressor=LinearRegression() regressor.fit(X_train,y_train) #Predicting the test set results y_pred=regressor.predict(X_test) #Building the optimal model using backward elimination import statsmodels.api as sm X=np.append(arr=np.ones((50,1)).astype(int),values=X,axis=1) #X_opt=X[:,[0,1,2,3,4,5]] X_opt = np.array(X[:, [0, 1, 2, 3, 4, 5]], dtype=float) regressor_OLS=sm.OLS(endog=y,exog=X_opt).fit() regressor_OLS.summary() X_opt = np.array(X[:, [0, 1, 3, 4, 5]], dtype=float) regressor_OLS=sm.OLS(endog=y,exog=X_opt).fit() regressor_OLS.summary() X_opt = np.array(X[:, [0, 3, 4, 5]], dtype=float) regressor_OLS=sm.OLS(endog=y,exog=X_opt).fit() regressor_OLS.summary() X_opt = np.array(X[:, [0, 3, 5]], dtype=float) regressor_OLS=sm.OLS(endog=y,exog=X_opt).fit() regressor_OLS.summary() X_opt = np.array(X[:, [0, 3]], dtype=float) regressor_OLS=sm.OLS(endog=y,exog=X_opt).fit() regressor_OLS.summary()
[ [ [ 64, 75 ], [ 1565, 1567 ], [ 1579, 1581 ], [ 1659, 1661 ], [ 1788, 1790 ], [ 1913, 1915 ], [ 2036, 2038 ], [ 2156, 2158 ] ], [ [ 83, 107 ] ], [ [ 115, 127 ], [ 163, 165 ] ], [ [ 153, 160 ], [ 198, 205 ], [ 230, 237 ] ], [ [ 194, 195 ], [ 758, 759 ] ], [ [ 226, 227 ], [ 974, 975 ], [ 1734, 1735 ], [ 1860, 1861 ], [ 1982, 1983 ], [ 2102, 2103 ], [ 2219, 2220 ] ], [ [ 580, 592 ] ], [ [ 593, 606 ], [ 687, 700 ] ], [ [ 635, 652 ], [ 658, 675 ] ], [ [ 653, 655 ], [ 741, 743 ] ], [ [ 738, 739 ], [ 798, 799 ] ], [ [ 796, 797 ], [ 971, 972 ], [ 1614, 1615 ] ], [ [ 902, 918 ], [ 954, 970 ] ], [ [ 919, 926 ], [ 1393, 1400 ] ], [ [ 928, 934 ], [ 1470, 1476 ] ], [ [ 936, 943 ], [ 1401, 1408 ] ], [ [ 945, 951 ] ], [ [ 1333, 1349 ], [ 1360, 1376 ] ], [ [ 1350, 1359 ], [ 1379, 1388 ], [ 1452, 1461 ] ], [ [ 1445, 1451 ] ], [ [ 1541, 1562 ], [ 1721, 1723 ], [ 1847, 1849 ], [ 1969, 1971 ], [ 2089, 2091 ], [ 2206, 2208 ] ], [ [ 1563, 1564 ], [ 1668, 1669 ], [ 1797, 1798 ], [ 1922, 1923 ], [ 2045, 2046 ], [ 2165, 2166 ] ], [ [ 1651, 1656 ], [ 1741, 1746 ] ], [ [ 1707, 1720 ], [ 1754, 1767 ] ], [ [ 1780, 1785 ], [ 1867, 1872 ] ], [ [ 1833, 1846 ], [ 1880, 1893 ] ], [ [ 1905, 1910 ], [ 1989, 1994 ] ], [ [ 1955, 1968 ], [ 2002, 2015 ] ], [ [ 2028, 2033 ], [ 2109, 2114 ] ], [ [ 2075, 2088 ], [ 2122, 2135 ] ], [ [ 2148, 2153 ], [ 2226, 2231 ] ], [ [ 2192, 2205 ], [ 2239, 2252 ] ] ]