lines
sequencelengths
1
383
raw_lines
sequencelengths
1
383
label
sequencelengths
1
383
type
sequencelengths
1
383
[ "def FUNC_77(self, VAR_148, VAR_149='', VAR_150=0, VAR_90=None):...\n", "\"\"\"docstring\"\"\"\n", "def FUNC_84(self=self, VAR_148=VAR_148, VAR_149=VAR_149, VAR_150=VAR_150):...\n", "return self.has_permission(VAR_148, VAR_149, VAR_150)\n" ]
[ "def requires_permission(self, name, table_name='', record_id=0, otherwise=None...\n", "\"\"\"docstring\"\"\"\n", "def has_permission(self=self, name=name, table_name=table_name, record_id=...\n", "return self.has_permission(name, table_name, record_id)\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "FunctionDef'", "Return'" ]
[ "def FUNC_16(self, VAR_5, VAR_10):...\n", "\"\"\"docstring\"\"\"\n", "VAR_34 = tempfile.mkdtemp(dir=FLAGS.image_decryption_dir)\n", "VAR_35 = VAR_10['properties']['image_location']\n", "VAR_36 = VAR_35.split('/')[0]\n", "VAR_37 = VAR_35[len(VAR_36) + 1:]\n", "VAR_13 = self._conn(VAR_5).get_bucket(VAR_36)\n", "VAR_27 = VAR_13.get_key(VAR_37)\n", "VAR_16 = VAR_27.get_contents_as_string()\n", "VAR_16, VAR_9, VAR_7 = self._s3_parse_manifest(VAR_5, VAR_10, VAR_16)\n", "def FUNC_21():...\n", "\"\"\"docstring\"\"\"\n", "VAR_47 = {'image_location': VAR_35, 'image_path': VAR_34}\n", "VAR_10['properties']['image_state'] = 'downloading'\n", "self.service.update(VAR_5, VAR_7, VAR_10)\n", "VAR_49 = []\n", "VAR_0.exception(_('Failed to download %(image_location)s to %(image_path)s'\n ), VAR_47)\n", "VAR_10['properties']['image_state'] = 'decrypting'\n", "VAR_50 = VAR_16.find('image').getiterator('filename')\n", "VAR_10['properties']['image_state'] = 'failed_download'\n", "self.service.update(VAR_5, VAR_7, VAR_10)\n", "for fn_element in VAR_50:\n", "self.service.update(VAR_5, VAR_7, VAR_10)\n", "VAR_52 = VAR_16.find('image/ec2_encrypted_key').text\n", "VAR_0.exception(_('Failed to decrypt %(image_location)s to %(image_path)s'),\n VAR_47)\n", "VAR_10['properties']['image_state'] = 'untarring'\n", "VAR_57 = self._download_file(VAR_13, fn_element.text, VAR_34)\n", "VAR_51 = os.path.join(VAR_34, 'image.encrypted')\n", "return\n", "VAR_18 = binascii.a2b_hex(VAR_52)\n", "VAR_10['properties']['image_state'] = 'failed_decrypt'\n", "self.service.update(VAR_5, VAR_7, VAR_10)\n", "VAR_49.append(VAR_57)\n", "for VAR_14 in VAR_49:\n", "VAR_53 = VAR_16.find('image/ec2_encrypted_iv').text\n", "self.service.update(VAR_5, VAR_7, VAR_10)\n", "VAR_56 = self._untarzip_image(VAR_34, VAR_55)\n", "VAR_0.exception(_('Failed to untar %(image_location)s to %(image_path)s'),\n VAR_47)\n", "VAR_10['properties']['image_state'] = 'uploading'\n", "shutil.copyfileobj(VAR_57, combined)\n", "VAR_19 = binascii.a2b_hex(VAR_53)\n", "return\n", "VAR_10['properties']['image_state'] = 'failed_untar'\n", "self.service.update(VAR_5, VAR_7, VAR_10)\n", "VAR_54 = crypto.key_path(VAR_5.project_id)\n", "self.service.update(VAR_5, VAR_7, VAR_10)\n", "self.service.update(VAR_5, VAR_7, VAR_10, VAR_42)\n", "VAR_0.exception(_('Failed to upload %(image_location)s to %(image_path)s'),\n VAR_47)\n", "VAR_10['properties']['image_state'] = 'available'\n", "VAR_55 = os.path.join(VAR_34, 'image.tar.gz')\n", "return\n", "VAR_10['properties']['image_state'] = 'failed_upload'\n", "VAR_10['status'] = 'active'\n", "self._decrypt_image(VAR_51, VAR_18, VAR_19, VAR_54, VAR_55)\n", "self.service.update(VAR_5, VAR_7, VAR_10)\n", "self.service.update(VAR_5, VAR_7, VAR_10)\n", "return\n", "shutil.rmtree(VAR_34)\n", "eventlet.spawn_n(FUNC_21)\n", "return VAR_9\n" ]
[ "def _s3_create(self, context, metadata):...\n", "\"\"\"docstring\"\"\"\n", "image_path = tempfile.mkdtemp(dir=FLAGS.image_decryption_dir)\n", "image_location = metadata['properties']['image_location']\n", "bucket_name = image_location.split('/')[0]\n", "manifest_path = image_location[len(bucket_name) + 1:]\n", "bucket = self._conn(context).get_bucket(bucket_name)\n", "key = bucket.get_key(manifest_path)\n", "manifest = key.get_contents_as_string()\n", "manifest, image, image_uuid = self._s3_parse_manifest(context, metadata,\n manifest)\n", "def delayed_create():...\n", "\"\"\"docstring\"\"\"\n", "log_vars = {'image_location': image_location, 'image_path': image_path}\n", "metadata['properties']['image_state'] = 'downloading'\n", "self.service.update(context, image_uuid, metadata)\n", "parts = []\n", "LOG.exception(_('Failed to download %(image_location)s to %(image_path)s'),\n log_vars)\n", "metadata['properties']['image_state'] = 'decrypting'\n", "elements = manifest.find('image').getiterator('filename')\n", "metadata['properties']['image_state'] = 'failed_download'\n", "self.service.update(context, image_uuid, metadata)\n", "for fn_element in elements:\n", "self.service.update(context, image_uuid, metadata)\n", "hex_key = manifest.find('image/ec2_encrypted_key').text\n", "LOG.exception(_('Failed to decrypt %(image_location)s to %(image_path)s'),\n log_vars)\n", "metadata['properties']['image_state'] = 'untarring'\n", "part = self._download_file(bucket, fn_element.text, image_path)\n", "enc_filename = os.path.join(image_path, 'image.encrypted')\n", "return\n", "encrypted_key = binascii.a2b_hex(hex_key)\n", "metadata['properties']['image_state'] = 'failed_decrypt'\n", "self.service.update(context, image_uuid, metadata)\n", "parts.append(part)\n", "for filename in parts:\n", "hex_iv = manifest.find('image/ec2_encrypted_iv').text\n", "self.service.update(context, image_uuid, metadata)\n", "unz_filename = self._untarzip_image(image_path, dec_filename)\n", "LOG.exception(_('Failed to untar %(image_location)s to %(image_path)s'),\n log_vars)\n", "metadata['properties']['image_state'] = 'uploading'\n", "shutil.copyfileobj(part, combined)\n", "encrypted_iv = binascii.a2b_hex(hex_iv)\n", "return\n", "metadata['properties']['image_state'] = 'failed_untar'\n", "self.service.update(context, image_uuid, metadata)\n", "cloud_pk = crypto.key_path(context.project_id)\n", "self.service.update(context, image_uuid, metadata)\n", "self.service.update(context, image_uuid, metadata, image_file)\n", "LOG.exception(_('Failed to upload %(image_location)s to %(image_path)s'),\n log_vars)\n", "metadata['properties']['image_state'] = 'available'\n", "dec_filename = os.path.join(image_path, 'image.tar.gz')\n", "return\n", "metadata['properties']['image_state'] = 'failed_upload'\n", "metadata['status'] = 'active'\n", "self._decrypt_image(enc_filename, encrypted_key, encrypted_iv, cloud_pk,\n dec_filename)\n", "self.service.update(context, image_uuid, metadata)\n", "self.service.update(context, image_uuid, metadata)\n", "return\n", "shutil.rmtree(image_path)\n", "eventlet.spawn_n(delayed_create)\n", "return image\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "FunctionDef'", "Docstring", "Assign'", "Assign'", "Expr'", "Assign'", "Expr'", "Assign'", "Assign'", "Assign'", "Expr'", "For", "Expr'", "Assign'", "Expr'", "Assign'", "Assign'", "Assign'", "Return'", "Assign'", "Assign'", "Expr'", "Expr'", "For", "Assign'", "Expr'", "Assign'", "Expr'", "Assign'", "Expr'", "Assign'", "Return'", "Assign'", "Expr'", "Assign'", "Expr'", "Expr'", "Expr'", "Assign'", "Assign'", "Return'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'", "Return'", "Expr'", "Expr'", "Return'" ]
[ "@staticmethod...\n", "VAR_11 = VAR_5.add_argument_group('registration')\n", "VAR_11.add_argument('--enable-registration', action='store_true', default=\n None, help='Enable registration for new users.')\n" ]
[ "@staticmethod...\n", "reg_group = parser.add_argument_group('registration')\n", "reg_group.add_argument('--enable-registration', action='store_true',\n default=None, help='Enable registration for new users.')\n" ]
[ 0, 0, 0 ]
[ "Condition", "Assign'", "Expr'" ]
[ "@VAR_0.route('/shutdown', methods=['GET'])...\n", "if VAR_0.config['WEB_CLI'] is True:\n", "flash(\n 'WEB CLI Mode is on, cannot self shutdown server. Consider doing manually.'\n , 'warning')\n", "VAR_98 = FUNC_58('/shutdown', 'get')\n", "return jsonify({'success': False, 'message': str(err)}), 400\n", "return 'GUI Shutdown Successful.'\n", "return redirect(url_for('dashboard'))\n" ]
[ "@gui.route('/shutdown', methods=['GET'])...\n", "if gui.config['WEB_CLI'] is True:\n", "flash(\n 'WEB CLI Mode is on, cannot self shutdown server. Consider doing manually.'\n , 'warning')\n", "response_info = query_internal_api('/shutdown', 'get')\n", "return jsonify({'success': False, 'message': str(err)}), 400\n", "return 'GUI Shutdown Successful.'\n", "return redirect(url_for('dashboard'))\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Condition", "Expr'", "Assign'", "Return'", "Return'", "Return'" ]
[ "def FUNC_0(VAR_1):...\n", "if not VAR_1:\n", "VAR_3 = None\n", "VAR_3 = reverse_with_params(viewname=lookup_view['viewname'], VAR_23=\n lookup_view.get('args', []), query_string=lookup_view.get(\n 'query_string', None))\n", "if VAR_3 is None:\n", "resolve(VAR_1)\n", "VAR_0.error(\"Reverse for '%s' not found.\" % VAR_1)\n", "return VAR_3\n", "VAR_3 = VAR_1\n" ]
[ "def parse_url(lookup_view):...\n", "if not lookup_view:\n", "url = None\n", "url = reverse_with_params(viewname=lookup_view['viewname'], args=\n lookup_view.get('args', []), query_string=lookup_view.get(\n 'query_string', None))\n", "if url is None:\n", "resolve(lookup_view)\n", "logger.error(\"Reverse for '%s' not found.\" % lookup_view)\n", "return url\n", "url = lookup_view\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Assign'", "Assign'", "Condition", "Expr'", "Expr'", "Return'", "Assign'" ]
[ "def FUNC_98(VAR_91=None, VAR_158=False, VAR_159=True, VAR_160=None, VAR_161...\n", "\"\"\"docstring\"\"\"\n", "from frappe.utils.logger import get_logger\n", "return get_logger(VAR_91=module, VAR_158=with_more_info, VAR_159=allow_site,\n VAR_160=filter, VAR_161=max_size, VAR_162=file_count)\n" ]
[ "def logger(module=None, with_more_info=False, allow_site=True, filter=None,...\n", "\"\"\"docstring\"\"\"\n", "from frappe.utils.logger import get_logger\n", "return get_logger(module=module, with_more_info=with_more_info, allow_site=\n allow_site, filter=filter, max_size=max_size, file_count=file_count)\n" ]
[ 0, 0, 0, 0 ]
[ "Condition", "Docstring", "ImportFrom'", "Return'" ]
[ "def FUNC_24(self, VAR_12):...\n", "VAR_14 = None\n", "if VAR_12.get('sort_by'):\n", "VAR_26 = CLASS_0._format(VAR_12.get('sort_by').split('.')) + ' ' + VAR_12.get(\n 'sort_order')\n", "if VAR_12.get('order_by'):\n", "if VAR_12.get('sort_by_next'):\n", "VAR_26 = VAR_12.get('order_by')\n", "VAR_26 = CLASS_0._format([self.ref_doctype, 'modified']) + ' desc'\n", "VAR_26 += ', ' + CLASS_0._format(VAR_12.get('sort_by_next').split('.')\n ) + ' ' + VAR_12.get('sort_order_next')\n", "VAR_27 = None\n", "if VAR_12.get('group_by'):\n", "VAR_14 = frappe._dict(VAR_12['group_by'])\n", "return VAR_26, VAR_27, VAR_14\n", "VAR_27 = VAR_14['group_by']\n", "VAR_26 = '_aggregate_column desc'\n" ]
[ "def get_standard_report_order_by(self, params):...\n", "group_by_args = None\n", "if params.get('sort_by'):\n", "order_by = Report._format(params.get('sort_by').split('.')) + ' ' + params.get(\n 'sort_order')\n", "if params.get('order_by'):\n", "if params.get('sort_by_next'):\n", "order_by = params.get('order_by')\n", "order_by = Report._format([self.ref_doctype, 'modified']) + ' desc'\n", "order_by += ', ' + Report._format(params.get('sort_by_next').split('.')\n ) + ' ' + params.get('sort_order_next')\n", "group_by = None\n", "if params.get('group_by'):\n", "group_by_args = frappe._dict(params['group_by'])\n", "return order_by, group_by, group_by_args\n", "group_by = group_by_args['group_by']\n", "order_by = '_aggregate_column desc'\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Condition", "Assign'", "Condition", "Condition", "Assign'", "Assign'", "AugAssign'", "Assign'", "Condition", "Assign'", "Return'", "Assign'", "Assign'" ]
[ "@CLASS_4('bindings')...\n", "\"\"\"docstring\"\"\"\n", "VAR_26 = {}\n", "VAR_27 = config.val.bindings.default\n", "VAR_28 = set(VAR_27.keys()).union(config.val.bindings.commands)\n", "VAR_28.remove('normal')\n", "VAR_28 = ['normal'] + sorted(list(VAR_28))\n", "for VAR_41 in VAR_28:\n", "VAR_26[VAR_41] = config.key_instance.get_bindings_for(VAR_41)\n", "VAR_20 = jinja.render('bindings.html', title='Bindings', VAR_26=bindings)\n", "return 'text/html', VAR_20\n" ]
[ "@add_handler('bindings')...\n", "\"\"\"docstring\"\"\"\n", "bindings = {}\n", "defaults = config.val.bindings.default\n", "modes = set(defaults.keys()).union(config.val.bindings.commands)\n", "modes.remove('normal')\n", "modes = ['normal'] + sorted(list(modes))\n", "for mode in modes:\n", "bindings[mode] = config.key_instance.get_bindings_for(mode)\n", "html = jinja.render('bindings.html', title='Bindings', bindings=bindings)\n", "return 'text/html', html\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "For", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_2():...\n", "" ]
[ "def get_local_ip_address():...\n", "" ]
[ 0, 0 ]
[ "FunctionDef'", "Condition" ]
[ "def FUNC_18(self):...\n", "\"\"\"docstring\"\"\"\n", "self.reactor.advance(100)\n", "VAR_34 = 'server2'\n", "VAR_36 = signedjson.key.generate_signing_key('ver1')\n", "VAR_37 = signedjson.key.get_verify_key(VAR_36)\n", "VAR_38 = 'ed25519:ver1'\n", "VAR_39 = 200 * 1000\n", "def FUNC_29():...\n", "return self.build_perspectives_response(VAR_34, VAR_36, VAR_39)\n" ]
[ "def test_invalid_perspectives_responses(self):...\n", "\"\"\"docstring\"\"\"\n", "self.reactor.advance(100)\n", "SERVER_NAME = 'server2'\n", "testkey = signedjson.key.generate_signing_key('ver1')\n", "testverifykey = signedjson.key.get_verify_key(testkey)\n", "testverifykey_id = 'ed25519:ver1'\n", "VALID_UNTIL_TS = 200 * 1000\n", "def build_response():...\n", "return self.build_perspectives_response(SERVER_NAME, testkey, VALID_UNTIL_TS)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Expr'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "FunctionDef'", "Return'" ]
[ "def FUNC_7(self, VAR_14=None):...\n", "\"\"\"docstring\"\"\"\n", "if VAR_14 is not None:\n", "return VAR_14\n", "if self.app is not None:\n", "return self.app\n", "VAR_18 = _app_ctx_stack.top\n", "if VAR_18 is not None:\n", "return VAR_18.app\n" ]
[ "def get_app(self, reference_app=None):...\n", "\"\"\"docstring\"\"\"\n", "if reference_app is not None:\n", "return reference_app\n", "if self.app is not None:\n", "return self.app\n", "ctx = _app_ctx_stack.top\n", "if ctx is not None:\n", "return ctx.app\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Return'", "Condition", "Return'", "Assign'", "Condition", "Return'" ]
[ "import glob\n", "import os\n", "import sys\n", "from collections import defaultdict\n", "from functools import partial as curry\n", "from . import biblio, boilerplate, caniuse, conditional, config, constants, datablocks, dfns, extensions, fingerprinting, h, headings, highlight, idl, includes, inlineTags, lint, markdown, mdnspeclinks, metadata, shorthands, wpt\n", "from .func import Functor\n", "from .h import *\n", "from .InputSource import FileInputSource, InputSource\n", "from .messages import *\n", "from .refs import ReferenceManager\n", "from .unsortedJunk import *\n", "def __init__(self, VAR_6, VAR_7=False, VAR_8=None, VAR_9=False, VAR_10=None,...\n", "self.valid = False\n", "self.lineNumbers = VAR_9\n", "if VAR_9:\n", "constants.dryRun = True\n", "if VAR_6 is None:\n", "VAR_6 = FUNC_0()\n", "if VAR_6 is None:\n", "die('string')\n", "self.inputSource = InputSource(VAR_6)\n", "return\n", "self.transitiveDependencies = set()\n", "self.debug = VAR_7\n", "self.token = VAR_8\n", "self.testing = VAR_11\n", "if VAR_10 is None:\n", "self.dataFile = config.defaultRequester\n", "self.dataFile = VAR_10\n", "self.md = None\n", "self.mdBaseline = None\n", "self.mdDocument = None\n", "self.mdCommandLine = None\n", "self.mdDefaults = None\n", "self.mdOverridingDefaults = None\n", "self.lines = []\n", "self.document = None\n", "self.html = None\n", "self.head = None\n", "self.body = None\n", "self.fillContainers = None\n", "self.valid = self.initializeState()\n", "def FUNC_1(self):...\n", "self.normativeRefs = {}\n", "self.informativeRefs = {}\n", "self.refs = ReferenceManager(VAR_10=self.dataFile, VAR_11=self.testing)\n", "self.externalRefsUsed = defaultdict(lambda : defaultdict(dict))\n", "self.md = None\n", "self.mdBaseline = metadata.MetadataManager()\n", "self.mdDocument = None\n", "self.mdCommandLine = metadata.MetadataManager()\n", "self.mdDefaults = None\n", "self.mdOverridingDefaults = None\n", "self.biblios = {}\n", "self.typeExpansions = {}\n", "self.macros = defaultdict(lambda x: '???')\n", "self.canIUse = {}\n", "self.mdnSpecLinks = {}\n", "self.widl = idl.getParser()\n", "self.testSuites = json.loads(self.dataFile.fetch('test-suites.json', str=True))\n", "self.languages = json.loads(self.dataFile.fetch('languages.json', str=True))\n", "self.extraStyles = defaultdict(str)\n", "self.extraStyles['style-colors'] = VAR_0\n", "self.extraStyles['style-darkmode'] = VAR_1\n", "self.extraStyles['style-md-lists'] = VAR_2\n", "self.extraStyles['style-autolinks'] = VAR_3\n", "self.extraStyles['style-selflinks'] = VAR_4\n", "self.extraStyles['style-counters'] = VAR_5\n", "self.extraScripts = defaultdict(str)\n", "VAR_31 = self.inputSource.read()\n", "die(\"Couldn't find the input file at the specified location '{0}'.\", self.\n inputSource)\n", "return True\n", "self.lines = VAR_31.lines\n", "return False\n", "if VAR_31.date is not None:\n", "self.mdBaseline.addParsedData('Date', VAR_31.date)\n" ]
[ "import glob\n", "import os\n", "import sys\n", "from collections import defaultdict\n", "from functools import partial as curry\n", "from . import biblio, boilerplate, caniuse, conditional, config, constants, datablocks, dfns, extensions, fingerprinting, h, headings, highlight, idl, includes, inlineTags, lint, markdown, mdnspeclinks, metadata, shorthands, wpt\n", "from .func import Functor\n", "from .h import *\n", "from .InputSource import FileInputSource, InputSource\n", "from .messages import *\n", "from .refs import ReferenceManager\n", "from .unsortedJunk import *\n", "def __init__(self, inputFilename, debug=False, token=None, lineNumbers=...\n", "self.valid = False\n", "self.lineNumbers = lineNumbers\n", "if lineNumbers:\n", "constants.dryRun = True\n", "if inputFilename is None:\n", "inputFilename = findImplicitInputFile()\n", "if inputFilename is None:\n", "die(\"\"\"No input file specified, and no *.bs or *.src.html files found in current directory.\nPlease specify an input file, or use - to pipe from STDIN.\"\"\"\n )\n", "self.inputSource = InputSource(inputFilename)\n", "return\n", "self.transitiveDependencies = set()\n", "self.debug = debug\n", "self.token = token\n", "self.testing = testing\n", "if fileRequester is None:\n", "self.dataFile = config.defaultRequester\n", "self.dataFile = fileRequester\n", "self.md = None\n", "self.mdBaseline = None\n", "self.mdDocument = None\n", "self.mdCommandLine = None\n", "self.mdDefaults = None\n", "self.mdOverridingDefaults = None\n", "self.lines = []\n", "self.document = None\n", "self.html = None\n", "self.head = None\n", "self.body = None\n", "self.fillContainers = None\n", "self.valid = self.initializeState()\n", "def initializeState(self):...\n", "self.normativeRefs = {}\n", "self.informativeRefs = {}\n", "self.refs = ReferenceManager(fileRequester=self.dataFile, testing=self.testing)\n", "self.externalRefsUsed = defaultdict(lambda : defaultdict(dict))\n", "self.md = None\n", "self.mdBaseline = metadata.MetadataManager()\n", "self.mdDocument = None\n", "self.mdCommandLine = metadata.MetadataManager()\n", "self.mdDefaults = None\n", "self.mdOverridingDefaults = None\n", "self.biblios = {}\n", "self.typeExpansions = {}\n", "self.macros = defaultdict(lambda x: '???')\n", "self.canIUse = {}\n", "self.mdnSpecLinks = {}\n", "self.widl = idl.getParser()\n", "self.testSuites = json.loads(self.dataFile.fetch('test-suites.json', str=True))\n", "self.languages = json.loads(self.dataFile.fetch('languages.json', str=True))\n", "self.extraStyles = defaultdict(str)\n", "self.extraStyles['style-colors'] = styleColors\n", "self.extraStyles['style-darkmode'] = styleDarkMode\n", "self.extraStyles['style-md-lists'] = styleMdLists\n", "self.extraStyles['style-autolinks'] = styleAutolinks\n", "self.extraStyles['style-selflinks'] = styleSelflinks\n", "self.extraStyles['style-counters'] = styleCounters\n", "self.extraScripts = defaultdict(str)\n", "inputContent = self.inputSource.read()\n", "die(\"Couldn't find the input file at the specified location '{0}'.\", self.\n inputSource)\n", "return True\n", "self.lines = inputContent.lines\n", "return False\n", "if inputContent.date is not None:\n", "self.mdBaseline.addParsedData('Date', inputContent.date)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Import'", "Import'", "Import'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Condition", "Assign'", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Condition", "Expr'", "Assign'", "Return'", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Return'", "Assign'", "Return'", "Condition", "Expr'" ]
[ "async def FUNC_39(self, VAR_28, VAR_45, VAR_1, VAR_32):...\n", "VAR_144 = VAR_188.ThirdPartyInvite, VAR_1.content['third_party_invite'][\n 'signed']['token']\n", "VAR_145 = None\n", "VAR_94 = await VAR_32.get_prev_state_ids()\n", "VAR_146 = VAR_94.get(VAR_144)\n", "if VAR_146:\n", "VAR_145 = await self.store.get_event(VAR_146, allow_none=True)\n", "if VAR_145:\n", "VAR_182 = VAR_145.content.get('display_name')\n", "VAR_0.info('Could not find invite event for third_party_invite: %r', VAR_45)\n", "VAR_45['content']['third_party_invite']['display_name'] = VAR_182\n", "VAR_93 = self.event_builder_factory.new(VAR_28, VAR_45)\n", "EventValidator().validate_builder(VAR_93)\n", "VAR_1, VAR_32 = await self.event_creation_handler.create_new_client_event(\n VAR_93=builder)\n", "EventValidator().validate_new(VAR_1, self.config)\n", "return VAR_1, VAR_32\n" ]
[ "async def add_display_name_to_third_party_invite(self, room_version,...\n", "key = EventTypes.ThirdPartyInvite, event.content['third_party_invite']['signed'\n ]['token']\n", "original_invite = None\n", "prev_state_ids = await context.get_prev_state_ids()\n", "original_invite_id = prev_state_ids.get(key)\n", "if original_invite_id:\n", "original_invite = await self.store.get_event(original_invite_id, allow_none\n =True)\n", "if original_invite:\n", "display_name = original_invite.content.get('display_name')\n", "logger.info('Could not find invite event for third_party_invite: %r',\n event_dict)\n", "event_dict['content']['third_party_invite']['display_name'] = display_name\n", "builder = self.event_builder_factory.new(room_version, event_dict)\n", "EventValidator().validate_builder(builder)\n", "event, context = await self.event_creation_handler.create_new_client_event(\n builder=builder)\n", "EventValidator().validate_new(event, self.config)\n", "return event, context\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "AsyncFunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Expr'", "Assign'", "Assign'", "Expr'", "Assign'", "Expr'", "Return'" ]
[ "def FUNC_6(VAR_4, VAR_6):...\n", "\"\"\"docstring\"\"\"\n", "VAR_6.start(VAR_11=['--temp-basedir', '--loglines=0'] + _base_args(request.\n config))\n", "VAR_6.open_path('qute://log')\n", "assert VAR_6.get_content() == 'Log output was disabled.'\n" ]
[ "def test_no_loglines(request, quteproc_new):...\n", "\"\"\"docstring\"\"\"\n", "quteproc_new.start(args=['--temp-basedir', '--loglines=0'] + _base_args(\n request.config))\n", "quteproc_new.open_path('qute://log')\n", "assert quteproc_new.get_content() == 'Log output was disabled.'\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Expr'", "Expr'", "Assert'" ]
[ "def FUNC_13():...\n", "yield '<?xml version=\"1.0\" encoding=\"UTF-8\" ?>\\n'\n", "if VAR_1:\n", "yield '<?xml-stylesheet type=\"text/xsl\" href=\"/scopelist.xsl\" ?>\\n'\n", "yield '<objectlist count=\"{:d}\">\\n'.format(VAR_39)\n", "VAR_40 = 0\n", "VAR_41 = 0\n", "if VAR_26:\n", "VAR_40 = int(VAR_25 * VAR_5)\n", "VAR_42 = []\n", "VAR_44 = cycle(VAR_17)\n", "def FUNC_15():...\n", "random.seed(VAR_24)\n", "return list(map(lambda x: x + VAR_5 * VAR_41, sorted(random.sample(list(\n range(VAR_5)), VAR_40))))\n" ]
[ "def generate():...\n", "yield '<?xml version=\"1.0\" encoding=\"UTF-8\" ?>\\n'\n", "if STYLE:\n", "yield '<?xml-stylesheet type=\"text/xsl\" href=\"/scopelist.xsl\" ?>\\n'\n", "yield '<objectlist count=\"{:d}\">\\n'.format(total_entries)\n", "mix_per_iteration = 0\n", "iteration_count = 0\n", "if make_cocktail:\n", "mix_per_iteration = int(percentage * ITEMS_PER_ITERATION)\n", "mix_indices = []\n", "pool = cycle(mixer_list)\n", "def generate_mix_indices():...\n", "random.seed(seed)\n", "return list(map(lambda x: x + ITEMS_PER_ITERATION * iteration_count, sorted\n (random.sample(list(range(ITEMS_PER_ITERATION)), mix_per_iteration))))\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Condition", "Expr'", "Expr'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "FunctionDef'", "Expr'", "Return'" ]
[ "def FUNC_121(*VAR_79, **VAR_42):...\n", "if VAR_13.read_from_replica:\n", "FUNC_6()\n", "VAR_225 = VAR_129(*VAR_79, **get_newargs(fn, kwargs))\n", "if VAR_1 and hasattr(VAR_1, 'primary_db'):\n", "return VAR_225\n", "VAR_1.db.close()\n", "VAR_1.db = VAR_1.primary_db\n" ]
[ "def wrapper_fn(*args, **kwargs):...\n", "if conf.read_from_replica:\n", "connect_replica()\n", "retval = fn(*args, **get_newargs(fn, kwargs))\n", "if local and hasattr(local, 'primary_db'):\n", "return retval\n", "local.db.close()\n", "local.db = local.primary_db\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Expr'", "Assign'", "Condition", "Return'", "Expr'", "Assign'" ]
[ "def FUNC_68():...\n", "VAR_54 = []\n", "if VAR_136['grabData'] and VAR_136['ctx'] == 'imgs':\n", "VAR_284 = min(VAR_136['start'], len(VAR_283) - 1)\n", "return [VAR_30.simpleMarshal(VAR_121=xtra, parents=opts['parents']) for\n VAR_30 in VAR_283]\n", "if VAR_136['limit'] == 0:\n", "VAR_317 = len(VAR_283)\n", "VAR_317 = min(len(VAR_283), VAR_284 + VAR_136['limit'])\n", "for VAR_212 in range(VAR_284, VAR_317):\n", "VAR_318 = VAR_283[VAR_212]\n", "return VAR_54\n", "VAR_54.append(FUNC_23(VAR_2, VAR_19, VAR_6=e.id, VAR_79=opts['key'], VAR_8=\n conn, VAR_31=True))\n", "VAR_1.debug('(iid %i) ignoring Attribute Error: %s' % (VAR_318.id, str(VAR_30))\n )\n", "VAR_1.debug('(iid %i) ignoring Server Error: %s' % (VAR_318.id, str(VAR_30)))\n" ]
[ "def marshal():...\n", "rv = []\n", "if opts['grabData'] and opts['ctx'] == 'imgs':\n", "bottom = min(opts['start'], len(sr) - 1)\n", "return [x.simpleMarshal(xtra=xtra, parents=opts['parents']) for x in sr]\n", "if opts['limit'] == 0:\n", "top = len(sr)\n", "top = min(len(sr), bottom + opts['limit'])\n", "for i in range(bottom, top):\n", "e = sr[i]\n", "return rv\n", "rv.append(imageData_json(request, server_id, iid=e.id, key=opts['key'],\n conn=conn, _internal=True))\n", "logger.debug('(iid %i) ignoring Attribute Error: %s' % (e.id, str(x)))\n", "logger.debug('(iid %i) ignoring Server Error: %s' % (e.id, str(x)))\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Condition", "Assign'", "Return'", "Condition", "Assign'", "Assign'", "For", "Assign'", "Return'", "Expr'", "Expr'", "Expr'" ]
[ "@defer.inlineCallbacks...\n", "VAR_14 = self.mock_handler.set_displayname\n", "VAR_14.side_effect = AuthError(400, 'message')\n", "VAR_12, VAR_13 = yield self.mock_resource.trigger('PUT', \n '/profile/%s/displayname' % '@4567:test', b'{\"displayname\": \"Frank Jr.\"}')\n", "self.assertTrue(400 <= VAR_12 < 499, msg='code %d is in the 4xx range' % code)\n" ]
[ "@defer.inlineCallbacks...\n", "mocked_set = self.mock_handler.set_displayname\n", "mocked_set.side_effect = AuthError(400, 'message')\n", "code, response = yield self.mock_resource.trigger('PUT', \n '/profile/%s/displayname' % '@4567:test', b'{\"displayname\": \"Frank Jr.\"}')\n", "self.assertTrue(400 <= code < 499, msg='code %d is in the 4xx range' % code)\n" ]
[ 0, 0, 0, 0, 0 ]
[ "Condition", "Assign'", "Assign'", "Assign'", "Expr'" ]
[ "@VAR_8.route('/login', methods=['GET', 'POST'])...\n", "if request.method == 'GET':\n", "VAR_78 = FUNC_7()\n", "if request.method == 'POST':\n", "return render_template('frontend/index.html', VAR_78=config)\n", "VAR_25 = request.form.get('username')\n", "VAR_91 = request.form.get('password')\n", "if not callable(VAR_8.auth) and VAR_25 in VAR_8.auth and VAR_8.auth[VAR_25\n", "login_user(CLASS_0(VAR_25))\n", "return abort(401)\n", "return redirect('/')\n" ]
[ "@app.route('/login', methods=['GET', 'POST'])...\n", "if request.method == 'GET':\n", "config = get_config()\n", "if request.method == 'POST':\n", "return render_template('frontend/index.html', config=config)\n", "username = request.form.get('username')\n", "password = request.form.get('password')\n", "if not callable(app.auth) and username in app.auth and app.auth[username\n", "login_user(User(username))\n", "return abort(401)\n", "return redirect('/')\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Condition", "Assign'", "Condition", "Return'", "Assign'", "Assign'", "Condition", "Expr'", "Return'", "Return'" ]
[ "def FUNC_39(self):...\n", "VAR_53 = {'membership': 'join', 'displayname': 'other test user'}\n", "VAR_54 = json.dumps(VAR_53)\n", "VAR_22, VAR_23 = self.make_request('PUT', \n '/_matrix/client/r0/rooms/%s/state/m.room.member/%s' % (self.room_id,\n self.user_id), VAR_54, VAR_16=self.tok)\n", "self.assertEqual(VAR_23.code, 200, VAR_23.result)\n", "VAR_55 = VAR_23.json_body['event_id']\n", "VAR_22, VAR_23 = self.make_request('GET', \n '/_matrix/client/r0/rooms/%s/event/%s' % (self.room_id, VAR_55), VAR_16\n =self.tok)\n", "self.assertEqual(VAR_23.code, 200, VAR_23.result)\n", "VAR_56 = VAR_23.json_body['content']['displayname']\n", "self.assertEqual(VAR_56, self.displayname, VAR_23.result)\n" ]
[ "def test_per_room_profile_forbidden(self):...\n", "data = {'membership': 'join', 'displayname': 'other test user'}\n", "request_data = json.dumps(data)\n", "request, channel = self.make_request('PUT', \n '/_matrix/client/r0/rooms/%s/state/m.room.member/%s' % (self.room_id,\n self.user_id), request_data, access_token=self.tok)\n", "self.assertEqual(channel.code, 200, channel.result)\n", "event_id = channel.json_body['event_id']\n", "request, channel = self.make_request('GET', \n '/_matrix/client/r0/rooms/%s/event/%s' % (self.room_id, event_id),\n access_token=self.tok)\n", "self.assertEqual(channel.code, 200, channel.result)\n", "res_displayname = channel.json_body['content']['displayname']\n", "self.assertEqual(res_displayname, self.displayname, channel.result)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "Expr'", "Assign'", "Expr'" ]
[ "async def FUNC_35(self, VAR_1: EventBase, VAR_32: EventContext, VAR_3:...\n", "\"\"\"docstring\"\"\"\n", "if VAR_1.is_state():\n", "VAR_179 = VAR_1.type, VAR_1.state_key\n", "VAR_179 = None\n", "VAR_125 = {k: a.event_id for k, a in VAR_3.items() if k != VAR_179}\n", "VAR_126 = await VAR_32.get_current_state_ids()\n", "VAR_126 = dict(VAR_126)\n", "VAR_126.update(VAR_125)\n", "VAR_94 = await VAR_32.get_prev_state_ids()\n", "VAR_94 = dict(VAR_94)\n", "VAR_94.update({k: a.event_id for k, a in VAR_3.items()})\n", "VAR_127 = VAR_32.state_group\n", "VAR_128 = await self.state_store.store_state_group(VAR_1.event_id, VAR_1.\n room_id, VAR_127=prev_group, delta_ids=state_updates, VAR_126=\n current_state_ids)\n", "return EventContext.with_state(VAR_128=state_group,\n state_group_before_event=context.state_group_before_event, VAR_126=\n current_state_ids, VAR_94=prev_state_ids, VAR_127=prev_group, delta_ids\n =state_updates)\n" ]
[ "async def _update_context_for_auth_events(self, event: EventBase, context:...\n", "\"\"\"docstring\"\"\"\n", "if event.is_state():\n", "event_key = event.type, event.state_key\n", "event_key = None\n", "state_updates = {k: a.event_id for k, a in auth_events.items() if k !=\n event_key}\n", "current_state_ids = await context.get_current_state_ids()\n", "current_state_ids = dict(current_state_ids)\n", "current_state_ids.update(state_updates)\n", "prev_state_ids = await context.get_prev_state_ids()\n", "prev_state_ids = dict(prev_state_ids)\n", "prev_state_ids.update({k: a.event_id for k, a in auth_events.items()})\n", "prev_group = context.state_group\n", "state_group = await self.state_store.store_state_group(event.event_id,\n event.room_id, prev_group=prev_group, delta_ids=state_updates,\n current_state_ids=current_state_ids)\n", "return EventContext.with_state(state_group=state_group,\n state_group_before_event=context.state_group_before_event,\n current_state_ids=current_state_ids, prev_state_ids=prev_state_ids,\n prev_group=prev_group, delta_ids=state_updates)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_17(self, VAR_44):...\n", "VAR_13 = FUNC_1(VAR_44.group(1))\n", "if VAR_13 in self.def_footnotes:\n", "return\n", "self.def_footnotes[VAR_13] = 0\n", "self.tokens.append({'type': 'footnote_start', 'key': VAR_13})\n", "VAR_14 = VAR_44.group(2)\n", "if '\\n' in VAR_14:\n", "VAR_97 = VAR_14.split('\\n')\n", "self.parse(VAR_14, self.footnote_rules)\n", "VAR_98 = None\n", "self.tokens.append({'type': 'footnote_end', 'key': VAR_13})\n", "for VAR_85 in VAR_97[1:]:\n", "VAR_94 = len(VAR_85) - len(VAR_85.lstrip())\n", "VAR_99 = [VAR_97[0]]\n", "if VAR_94 and (not VAR_98 or VAR_94 < VAR_98):\n", "for VAR_85 in VAR_97[1:]:\n", "VAR_98 = VAR_94\n", "VAR_99.append(VAR_85[VAR_98:])\n", "VAR_14 = '\\n'.join(VAR_99)\n" ]
[ "def parse_def_footnotes(self, m):...\n", "key = _keyify(m.group(1))\n", "if key in self.def_footnotes:\n", "return\n", "self.def_footnotes[key] = 0\n", "self.tokens.append({'type': 'footnote_start', 'key': key})\n", "text = m.group(2)\n", "if '\\n' in text:\n", "lines = text.split('\\n')\n", "self.parse(text, self.footnote_rules)\n", "whitespace = None\n", "self.tokens.append({'type': 'footnote_end', 'key': key})\n", "for line in lines[1:]:\n", "space = len(line) - len(line.lstrip())\n", "newlines = [lines[0]]\n", "if space and (not whitespace or space < whitespace):\n", "for line in lines[1:]:\n", "whitespace = space\n", "newlines.append(line[whitespace:])\n", "text = '\\n'.join(newlines)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Condition", "Return'", "Assign'", "Expr'", "Assign'", "Condition", "Assign'", "Expr'", "Assign'", "Expr'", "For", "Assign'", "Assign'", "Condition", "For", "Assign'", "Expr'", "Assign'" ]
[ "@defer.inlineCallbacks...\n", "self.mock_federation = Mock()\n", "self.mock_registry = Mock()\n", "self.query_handlers = {}\n", "def FUNC_10(VAR_0, VAR_1):...\n", "self.query_handlers[VAR_0] = VAR_1\n", "self.mock_registry.register_query_handler = FUNC_10\n", "VAR_2 = yield setup_test_homeserver(self.addCleanup, http_client=None,\n resource_for_federation=Mock(), federation_client=self.mock_federation,\n federation_server=Mock(), federation_registry=self.mock_registry)\n", "self.store = VAR_2.get_datastore()\n", "self.frank = UserID.from_string('@1234ABCD:test')\n", "self.bob = UserID.from_string('@4567:test')\n", "self.alice = UserID.from_string('@alice:remote')\n", "yield defer.ensureDeferred(self.store.create_profile(self.frank.localpart))\n", "self.handler = VAR_2.get_profile_handler()\n", "self.hs = VAR_2\n" ]
[ "@defer.inlineCallbacks...\n", "self.mock_federation = Mock()\n", "self.mock_registry = Mock()\n", "self.query_handlers = {}\n", "def register_query_handler(query_type, handler):...\n", "self.query_handlers[query_type] = handler\n", "self.mock_registry.register_query_handler = register_query_handler\n", "hs = yield setup_test_homeserver(self.addCleanup, http_client=None,\n resource_for_federation=Mock(), federation_client=self.mock_federation,\n federation_server=Mock(), federation_registry=self.mock_registry)\n", "self.store = hs.get_datastore()\n", "self.frank = UserID.from_string('@1234ABCD:test')\n", "self.bob = UserID.from_string('@4567:test')\n", "self.alice = UserID.from_string('@alice:remote')\n", "yield defer.ensureDeferred(self.store.create_profile(self.frank.localpart))\n", "self.handler = hs.get_profile_handler()\n", "self.hs = hs\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 4, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Assign'", "Assign'", "Assign'", "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'" ]
[ "@VAR_0.route('/<path:path>', methods=['HEAD'])...\n", "" ]
[ "@app.route('/<path:path>', methods=['HEAD'])...\n", "" ]
[ 0, 0 ]
[ "Condition", "Condition" ]
[ "@log_function...\n", "\"\"\"docstring\"\"\"\n", "if VAR_33:\n", "VAR_2 = FUNC_1('/groups/%s/summary/categories/%s/rooms/%s', VAR_30, VAR_33,\n VAR_6)\n", "VAR_2 = FUNC_1('/groups/%s/summary/rooms/%s', VAR_30, VAR_6)\n", "return self.client.delete_json(VAR_5=destination, VAR_2=path, VAR_3={\n 'requester_user_id': user_id}, VAR_15=True)\n" ]
[ "@log_function...\n", "\"\"\"docstring\"\"\"\n", "if category_id:\n", "path = _create_v1_path('/groups/%s/summary/categories/%s/rooms/%s',\n group_id, category_id, room_id)\n", "path = _create_v1_path('/groups/%s/summary/rooms/%s', group_id, room_id)\n", "return self.client.delete_json(destination=destination, path=path, args={\n 'requester_user_id': user_id}, ignore_backoff=True)\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Condition", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_3(self, VAR_0, VAR_1):...\n", "VAR_1.AWS_LOCATION = ''\n", "VAR_7 = SimpleUploadedFile('uploaded_file.txt', b'uploaded')\n", "VAR_8 = VAR_0.post('/', data={'file': uploaded_file})\n", "S3FileMiddleware(lambda x: None)(VAR_8)\n", "assert VAR_8.FILES.getlist('file')\n", "assert VAR_8.FILES.get('file').read() == b'uploaded'\n", "storage.save('tmp/s3file/s3_file.txt', ContentFile(b's3file'))\n", "VAR_8 = VAR_0.post('/', data={'file': 'tmp/s3file/s3_file.txt', 's3file':\n 'file'})\n", "S3FileMiddleware(lambda x: None)(VAR_8)\n", "assert VAR_8.FILES.getlist('file')\n", "assert VAR_8.FILES.get('file').read() == b's3file'\n" ]
[ "def test_process_request__no_location(self, rf, settings):...\n", "settings.AWS_LOCATION = ''\n", "uploaded_file = SimpleUploadedFile('uploaded_file.txt', b'uploaded')\n", "request = rf.post('/', data={'file': uploaded_file})\n", "S3FileMiddleware(lambda x: None)(request)\n", "assert request.FILES.getlist('file')\n", "assert request.FILES.get('file').read() == b'uploaded'\n", "storage.save('tmp/s3file/s3_file.txt', ContentFile(b's3file'))\n", "request = rf.post('/', data={'file': 'tmp/s3file/s3_file.txt', 's3file':\n 'file'})\n", "S3FileMiddleware(lambda x: None)(request)\n", "assert request.FILES.getlist('file')\n", "assert request.FILES.get('file').read() == b's3file'\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Expr'", "Assert'", "Assert'", "Expr'", "Assign'", "Expr'", "Assert'", "Assert'" ]
[ "def FUNC_2(VAR_3):...\n", "VAR_13 = VAR_25.local.request.method\n", "if VAR_13 not in VAR_25.allowed_http_methods_for_whitelisted_func[VAR_3]:\n", "VAR_25.throw(_('Not permitted'), VAR_25.PermissionError)\n" ]
[ "def is_valid_http_method(method):...\n", "http_method = frappe.local.request.method\n", "if http_method not in frappe.allowed_http_methods_for_whitelisted_func[method]:\n", "frappe.throw(_('Not permitted'), frappe.PermissionError)\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "For", "Expr'" ]
[ "@VAR_2.route('/read/<int:book_id>/<book_format>')...\n", "VAR_95 = calibre_db.get_filtered_book(VAR_5)\n", "if not VAR_95:\n", "flash(_(\n u'Oops! Selected book title is unavailable. File does not exist or is not accessible'\n ), category='error')\n", "VAR_99 = None\n", "VAR_3.debug(\n u'Oops! Selected book title is unavailable. File does not exist or is not accessible'\n )\n", "if VAR_87.is_authenticated:\n", "return redirect(url_for('web.index'))\n", "VAR_99 = ub.session.query(ub.Bookmark).filter(and_(ub.Bookmark.user_id ==\n int(VAR_87.id), ub.Bookmark.book_id == VAR_5, ub.Bookmark.format ==\n VAR_6.upper())).first()\n", "if VAR_6.lower() == 'epub':\n", "VAR_3.debug(u'Start epub reader for %d', VAR_5)\n", "if VAR_6.lower() == 'pdf':\n", "return render_title_template('read.html', bookid=book_id, VAR_149=book.\n title, VAR_99=bookmark)\n", "VAR_3.debug(u'Start pdf reader for %d', VAR_5)\n", "if VAR_6.lower() == 'txt':\n", "return render_title_template('readpdf.html', pdffile=book_id, VAR_149=book.\n title)\n", "VAR_3.debug(u'Start txt reader for %d', VAR_5)\n", "if VAR_6.lower() == 'djvu':\n", "return render_title_template('readtxt.html', txtfile=book_id, VAR_149=book.\n title)\n", "VAR_3.debug(u'Start djvu reader for %d', VAR_5)\n", "for fileExt in constants.EXTENSIONS_AUDIO:\n", "return render_title_template('readdjvu.html', djvufile=book_id, VAR_149=\n book.title)\n", "if VAR_6.lower() == fileExt:\n", "for fileExt in ['cbr', 'cbt', 'cbz']:\n", "VAR_63 = calibre_db.get_filtered_book(VAR_5)\n", "if VAR_6.lower() == fileExt:\n", "VAR_3.debug(\n u'Oops! Selected book title is unavailable. File does not exist or is not accessible'\n )\n", "VAR_3.debug(u'Start mp3 listening for %d', VAR_5)\n", "VAR_148 = str(VAR_5)\n", "flash(_(\n u'Oops! Selected book title is unavailable. File does not exist or is not accessible'\n ), category='error')\n", "return render_title_template('listenmp3.html', mp3file=book_id, audioformat\n =book_format.lower(), entry=entries, VAR_99=bookmark)\n", "VAR_149 = VAR_95.title\n", "return redirect(url_for('web.index'))\n", "if len(VAR_95.series):\n", "VAR_149 = VAR_149 + ' - ' + VAR_95.series[0].name\n", "VAR_3.debug(u'Start comic reader for %d', VAR_5)\n", "if VAR_95.series_index:\n", "return render_title_template('readcbr.html', comicfile=all_name, VAR_149=\n title, extension=fileExt)\n", "VAR_149 = VAR_149 + ' #' + '{0:.2f}'.format(VAR_95.series_index).rstrip('0'\n ).rstrip('.')\n" ]
[ "@web.route('/read/<int:book_id>/<book_format>')...\n", "book = calibre_db.get_filtered_book(book_id)\n", "if not book:\n", "flash(_(\n u'Oops! Selected book title is unavailable. File does not exist or is not accessible'\n ), category='error')\n", "bookmark = None\n", "log.debug(\n u'Oops! Selected book title is unavailable. File does not exist or is not accessible'\n )\n", "if current_user.is_authenticated:\n", "return redirect(url_for('web.index'))\n", "bookmark = ub.session.query(ub.Bookmark).filter(and_(ub.Bookmark.user_id ==\n int(current_user.id), ub.Bookmark.book_id == book_id, ub.Bookmark.\n format == book_format.upper())).first()\n", "if book_format.lower() == 'epub':\n", "log.debug(u'Start epub reader for %d', book_id)\n", "if book_format.lower() == 'pdf':\n", "return render_title_template('read.html', bookid=book_id, title=book.title,\n bookmark=bookmark)\n", "log.debug(u'Start pdf reader for %d', book_id)\n", "if book_format.lower() == 'txt':\n", "return render_title_template('readpdf.html', pdffile=book_id, title=book.title)\n", "log.debug(u'Start txt reader for %d', book_id)\n", "if book_format.lower() == 'djvu':\n", "return render_title_template('readtxt.html', txtfile=book_id, title=book.title)\n", "log.debug(u'Start djvu reader for %d', book_id)\n", "for fileExt in constants.EXTENSIONS_AUDIO:\n", "return render_title_template('readdjvu.html', djvufile=book_id, title=book.\n title)\n", "if book_format.lower() == fileExt:\n", "for fileExt in ['cbr', 'cbt', 'cbz']:\n", "entries = calibre_db.get_filtered_book(book_id)\n", "if book_format.lower() == fileExt:\n", "log.debug(\n u'Oops! Selected book title is unavailable. File does not exist or is not accessible'\n )\n", "log.debug(u'Start mp3 listening for %d', book_id)\n", "all_name = str(book_id)\n", "flash(_(\n u'Oops! Selected book title is unavailable. File does not exist or is not accessible'\n ), category='error')\n", "return render_title_template('listenmp3.html', mp3file=book_id, audioformat\n =book_format.lower(), entry=entries, bookmark=bookmark)\n", "title = book.title\n", "return redirect(url_for('web.index'))\n", "if len(book.series):\n", "title = title + ' - ' + book.series[0].name\n", "log.debug(u'Start comic reader for %d', book_id)\n", "if book.series_index:\n", "return render_title_template('readcbr.html', comicfile=all_name, title=\n title, extension=fileExt)\n", "title = title + ' #' + '{0:.2f}'.format(book.series_index).rstrip('0').rstrip(\n '.')\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "For", "Assign'", "Condition", "Expr'", "Assign'", "Expr'", "Condition", "Return'", "Assign'", "Condition", "Expr'", "Condition", "Return'", "Expr'", "Condition", "Return'", "Expr'", "Condition", "Return'", "Expr'", "For", "Return'", "Condition", "For", "Assign'", "Condition", "Expr'", "Expr'", "Assign'", "Expr'", "Return'", "Assign'", "Return'", "Condition", "Assign'", "Expr'", "Condition", "Return'", "Assign'" ]
[ "def FUNC_47(self):...\n", "self.model.updateLibrary()\n", "return 'success'\n" ]
[ "def api_updatedb(self):...\n", "self.model.updateLibrary()\n", "return 'success'\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Return'" ]
[ "def FUNC_73(VAR_38, VAR_40):...\n", "\"\"\"docstring\"\"\"\n", "VAR_41 = CertificateWhitelist.objects.get(VAR_19=student, VAR_10=course_key)\n", "VAR_44 = GeneratedCertificate.objects.get(VAR_19=student, VAR_10=course_key)\n", "VAR_41.delete()\n", "VAR_44.invalidate()\n", "VAR_0.info(\n u'Certificate invalidated for %s in course %s when removed from certificate exception list'\n , VAR_40.username, VAR_38)\n" ]
[ "def remove_certificate_exception(course_key, student):...\n", "\"\"\"docstring\"\"\"\n", "certificate_exception = CertificateWhitelist.objects.get(user=student,\n course_id=course_key)\n", "generated_certificate = GeneratedCertificate.objects.get(user=student,\n course_id=course_key)\n", "certificate_exception.delete()\n", "generated_certificate.invalidate()\n", "log.info(\n u'Certificate invalidated for %s in course %s when removed from certificate exception list'\n , student.username, course_key)\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'" ]
[ "def FUNC_1(VAR_2, VAR_3):...\n", "VAR_5 = VAR_2.get_view(VAR_3)\n", "if hasattr(VAR_5, 'as_view'):\n", "VAR_5 = VAR_5.as_view()\n", "if VAR_5 and not callable(VAR_5):\n", "return VAR_5\n" ]
[ "def _get_view_by_name(theme, view_name):...\n", "view = theme.get_view(view_name)\n", "if hasattr(view, 'as_view'):\n", "view = view.as_view()\n", "if view and not callable(view):\n", "return view\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Condition", "Assign'", "Condition", "Return'" ]
[ "@FUNC_0...\n", "from synapse.handlers.oidc_handler import OidcHandler\n", "return OidcHandler(self)\n" ]
[ "@cache_in_self...\n", "from synapse.handlers.oidc_handler import OidcHandler\n", "return OidcHandler(self)\n" ]
[ 0, 0, 0 ]
[ "Condition", "ImportFrom'", "Return'" ]
[ "def FUNC_54(self, VAR_112=VAR_3, VAR_134=VAR_3, VAR_135=VAR_3, VAR_136=VAR_3):...\n", "\"\"\"docstring\"\"\"\n", "VAR_254 = self.table_user()\n", "if 'username' not in VAR_254.fields:\n", "VAR_56 = VAR_263.request\n", "VAR_244 = VAR_263.response\n", "VAR_139 = VAR_263.session\n", "VAR_232 = (self.settings.retrieve_username_captcha or self.settings.\n retrieve_username_captcha is not False and self.settings.captcha)\n", "if not self.settings.mailer:\n", "VAR_244.flash = self.messages.function_disabled\n", "if VAR_112 is VAR_3:\n", "return ''\n", "VAR_112 = self.get_vars_next() or self.settings.retrieve_username_next\n", "if VAR_134 is VAR_3:\n", "VAR_134 = self.settings.retrieve_username_onvalidation\n", "if VAR_135 is VAR_3:\n", "VAR_135 = self.settings.retrieve_username_onaccept\n", "if VAR_136 is VAR_3:\n", "VAR_136 = self.messages['retrieve_username_log']\n", "VAR_271 = VAR_254.email.requires\n", "VAR_254.email.requires = [IS_IN_DB(self.db, VAR_254.email, VAR_59=self.\n messages.invalid_email)]\n", "VAR_7 = SQLFORM(VAR_254, VAR_133=['email'], hidden=dict(_next=next), showid\n =self.settings.showid, submit_button=self.messages.submit_button,\n delete_label=self.messages.delete_label, VAR_273=self.settings.\n formstyle, separator=self.settings.label_separator)\n", "if VAR_232:\n", "FUNC_5(VAR_7, VAR_232.label, VAR_232, VAR_232.comment, self.settings.\n formstyle, 'captcha__row')\n", "if VAR_7.accepts(VAR_56, VAR_139 if self.csrf_prevention else None, VAR_171\n", "VAR_373 = VAR_254._db(VAR_254.email == VAR_7.vars.email).select()\n", "VAR_254.email.requires = VAR_271\n", "if not VAR_373:\n", "return VAR_7\n", "VAR_263.session.flash = self.messages.invalid_email\n", "VAR_124 = ', '.join(u.username for u in VAR_373)\n", "redirect(self.url(VAR_11=request.args))\n", "self.settings.mailer.send(VAR_35=form.vars.email, VAR_36=self.messages.\n retrieve_username_subject, VAR_37=self.messages.retrieve_username %\n dict(username=username))\n", "VAR_139.flash = self.messages.email_sent\n", "for VAR_141 in VAR_373:\n", "self.log_event(VAR_136, VAR_141)\n", "VAR_26(VAR_135, VAR_7)\n", "if not VAR_112:\n", "VAR_112 = self.url(VAR_11=request.args)\n", "VAR_112 = FUNC_4(VAR_112, VAR_7)\n", "redirect(VAR_112)\n" ]
[ "def retrieve_username(self, next=DEFAULT, onvalidation=DEFAULT, onaccept=...\n", "\"\"\"docstring\"\"\"\n", "table_user = self.table_user()\n", "if 'username' not in table_user.fields:\n", "request = current.request\n", "response = current.response\n", "session = current.session\n", "captcha = (self.settings.retrieve_username_captcha or self.settings.\n retrieve_username_captcha is not False and self.settings.captcha)\n", "if not self.settings.mailer:\n", "response.flash = self.messages.function_disabled\n", "if next is DEFAULT:\n", "return ''\n", "next = self.get_vars_next() or self.settings.retrieve_username_next\n", "if onvalidation is DEFAULT:\n", "onvalidation = self.settings.retrieve_username_onvalidation\n", "if onaccept is DEFAULT:\n", "onaccept = self.settings.retrieve_username_onaccept\n", "if log is DEFAULT:\n", "log = self.messages['retrieve_username_log']\n", "old_requires = table_user.email.requires\n", "table_user.email.requires = [IS_IN_DB(self.db, table_user.email,\n error_message=self.messages.invalid_email)]\n", "form = SQLFORM(table_user, fields=['email'], hidden=dict(_next=next),\n showid=self.settings.showid, submit_button=self.messages.submit_button,\n delete_label=self.messages.delete_label, formstyle=self.settings.\n formstyle, separator=self.settings.label_separator)\n", "if captcha:\n", "addrow(form, captcha.label, captcha, captcha.comment, self.settings.\n formstyle, 'captcha__row')\n", "if form.accepts(request, session if self.csrf_prevention else None,\n", "users = table_user._db(table_user.email == form.vars.email).select()\n", "table_user.email.requires = old_requires\n", "if not users:\n", "return form\n", "current.session.flash = self.messages.invalid_email\n", "username = ', '.join(u.username for u in users)\n", "redirect(self.url(args=request.args))\n", "self.settings.mailer.send(to=form.vars.email, subject=self.messages.\n retrieve_username_subject, message=self.messages.retrieve_username %\n dict(username=username))\n", "session.flash = self.messages.email_sent\n", "for user in users:\n", "self.log_event(log, user)\n", "callback(onaccept, form)\n", "if not next:\n", "next = self.url(args=request.args)\n", "next = replace_id(next, form)\n", "redirect(next)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Condition", "Return'", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Expr'", "Condition", "Assign'", "Assign'", "Condition", "Return'", "Assign'", "Assign'", "Expr'", "Expr'", "Assign'", "For", "Expr'", "Expr'", "Condition", "Assign'", "Assign'", "Expr'" ]
[ "@app.route('/images/<filename>')...\n", "if VAR_2 and data.valid_image_filename(VAR_2):\n", "VAR_34 = data.image_exists(VAR_2)\n", "return 'Invalid file request', 413\n", "if VAR_34:\n", "return send_file(VAR_34)\n", "return 'Image not found', 404\n" ]
[ "@app.route('/images/<filename>')...\n", "if filename and data.valid_image_filename(filename):\n", "image_path = data.image_exists(filename)\n", "return 'Invalid file request', 413\n", "if image_path:\n", "return send_file(image_path)\n", "return 'Image not found', 404\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Condition", "Assign'", "Return'", "Condition", "Return'", "Return'" ]
[ "def FUNC_1(VAR_18: Text=DEFAULT_MODELS_PATH) ->TempDirectoryPath:...\n", "\"\"\"docstring\"\"\"\n", "VAR_18 = FUNC_0(VAR_18)\n", "VAR_65 = os.path.relpath(VAR_18)\n", "VAR_65 = VAR_18\n", "VAR_0.info(f'Loading model {VAR_65}...')\n", "return FUNC_3(VAR_18)\n" ]
[ "def get_model(model_path: Text=DEFAULT_MODELS_PATH) ->TempDirectoryPath:...\n", "\"\"\"docstring\"\"\"\n", "model_path = get_local_model(model_path)\n", "model_relative_path = os.path.relpath(model_path)\n", "model_relative_path = model_path\n", "logger.info(f'Loading model {model_relative_path}...')\n", "return unpack_model(model_path)\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Assign'", "Expr'", "Return'" ]
[ "def FUNC_15(self, VAR_2):...\n", "VAR_2.config.from_mapping({'MY_KEY': 'some.endpoint'})\n", "assert url_for('MY_KEY')\n", "VAR_2.add_url_rule('/some-endpoint', endpoint='some.endpoint')\n", "assert url_for('MY_KEY') == '/some-endpoint'\n" ]
[ "def test_it_works_with_config_keys_returning_endpoints(self, app):...\n", "app.config.from_mapping({'MY_KEY': 'some.endpoint'})\n", "assert url_for('MY_KEY')\n", "app.add_url_rule('/some-endpoint', endpoint='some.endpoint')\n", "assert url_for('MY_KEY') == '/some-endpoint'\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Assert'", "Expr'", "Assert'" ]
[ "@app.context_processor...\n", "VAR_3 = data.get_items(load_content=False)\n", "VAR_4 = require('archivy')[0].version\n", "VAR_5 = sep\n", "if VAR_5 == '\\\\':\n", "VAR_5 += '\\\\'\n", "return dict(VAR_3=dataobjs, VAR_5=SEP, VAR_4=version)\n" ]
[ "@app.context_processor...\n", "dataobjs = data.get_items(load_content=False)\n", "version = require('archivy')[0].version\n", "SEP = sep\n", "if SEP == '\\\\':\n", "SEP += '\\\\'\n", "return dict(dataobjs=dataobjs, SEP=SEP, version=version)\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Assign'", "Assign'", "Assign'", "Condition", "AugAssign'", "Return'" ]
[ "\"\"\"``chameleon.tales`` expressions.\"\"\"\n", "from ast import NodeTransformer\n", "from ast import parse\n", "from chameleon.astutil import Static\n", "from chameleon.astutil import Symbol\n", "from chameleon.codegen import template\n", "from chameleon.tales import NotExpr\n", "from chameleon.tales import StringExpr\n", "from AccessControl.ZopeGuards import guarded_apply\n", "from AccessControl.ZopeGuards import guarded_getattr\n", "from AccessControl.ZopeGuards import guarded_getitem\n", "from AccessControl.ZopeGuards import guarded_iter\n", "from AccessControl.ZopeGuards import protected_inplacevar\n", "from OFS.interfaces import ITraversable\n", "from RestrictedPython import RestrictingNodeTransformer\n", "from RestrictedPython.Utilities import utility_builtins\n", "from z3c.pt import expressions\n", "from zExceptions import NotFound\n", "from zExceptions import Unauthorized\n", "from zope.interface import implementer\n", "from zope.tales.tales import ExpressionEngine\n", "from zope.traversing.adapters import traversePathElement\n", "from zope.traversing.interfaces import TraversalError\n", "from .Expressions import render\n", "from .interfaces import IZopeAwareEngine\n", "VAR_0 = object()\n", "VAR_1 = (AttributeError, LookupError, NameError, TypeError, ValueError,\n NotFound, Unauthorized, TraversalError)\n", "def FUNC_0(VAR_2):...\n", "return Static(template('obj', VAR_2=Symbol(obj), mode='eval'))\n" ]
[ "\"\"\"``chameleon.tales`` expressions.\"\"\"\n", "from ast import NodeTransformer\n", "from ast import parse\n", "from chameleon.astutil import Static\n", "from chameleon.astutil import Symbol\n", "from chameleon.codegen import template\n", "from chameleon.tales import NotExpr\n", "from chameleon.tales import StringExpr\n", "from AccessControl.ZopeGuards import guarded_apply\n", "from AccessControl.ZopeGuards import guarded_getattr\n", "from AccessControl.ZopeGuards import guarded_getitem\n", "from AccessControl.ZopeGuards import guarded_iter\n", "from AccessControl.ZopeGuards import protected_inplacevar\n", "from OFS.interfaces import ITraversable\n", "from RestrictedPython import RestrictingNodeTransformer\n", "from RestrictedPython.Utilities import utility_builtins\n", "from z3c.pt import expressions\n", "from zExceptions import NotFound\n", "from zExceptions import Unauthorized\n", "from zope.interface import implementer\n", "from zope.tales.tales import ExpressionEngine\n", "from zope.traversing.adapters import traversePathElement\n", "from zope.traversing.interfaces import TraversalError\n", "from .Expressions import render\n", "from .interfaces import IZopeAwareEngine\n", "_marker = object()\n", "zope2_exceptions = (AttributeError, LookupError, NameError, TypeError,\n ValueError, NotFound, Unauthorized, TraversalError)\n", "def static(obj):...\n", "return Static(template('obj', obj=Symbol(obj), mode='eval'))\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Expr'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Assign'", "Assign'", "FunctionDef'", "Return'" ]
[ "def FUNC_2(VAR_4, VAR_5):...\n", "\"\"\"docstring\"\"\"\n", "if VAR_5 not in VAR_4.signature_def:\n", "return VAR_4.signature_def[VAR_5].inputs\n" ]
[ "def _get_inputs_tensor_info_from_meta_graph_def(meta_graph_def,...\n", "\"\"\"docstring\"\"\"\n", "if signature_def_key not in meta_graph_def.signature_def:\n", "return meta_graph_def.signature_def[signature_def_key].inputs\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Return'" ]
[ "def FUNC_50(self):...\n", "\"\"\"docstring\"\"\"\n", "self.load_doc_before_save()\n", "self.reset_seen()\n", "if self._action in ('save', 'submit'):\n", "self.run_method('before_validate')\n", "if self.flags.ignore_validate:\n", "return\n", "if self._action == 'save':\n", "self.run_method('validate')\n", "if self._action == 'submit':\n", "self.run_method('before_save')\n", "self.run_method('validate')\n", "if self._action == 'cancel':\n", "self.set_title_field()\n", "self.run_method('before_submit')\n", "self.run_method('before_cancel')\n", "if self._action == 'update_after_submit':\n", "self.run_method('before_update_after_submit')\n" ]
[ "def run_before_save_methods(self):...\n", "\"\"\"docstring\"\"\"\n", "self.load_doc_before_save()\n", "self.reset_seen()\n", "if self._action in ('save', 'submit'):\n", "self.run_method('before_validate')\n", "if self.flags.ignore_validate:\n", "return\n", "if self._action == 'save':\n", "self.run_method('validate')\n", "if self._action == 'submit':\n", "self.run_method('before_save')\n", "self.run_method('validate')\n", "if self._action == 'cancel':\n", "self.set_title_field()\n", "self.run_method('before_submit')\n", "self.run_method('before_cancel')\n", "if self._action == 'update_after_submit':\n", "self.run_method('before_update_after_submit')\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Expr'", "Expr'", "Condition", "Expr'", "Condition", "Return'", "Condition", "Expr'", "Condition", "Expr'", "Expr'", "Condition", "Expr'", "Expr'", "Expr'", "Condition", "Expr'" ]
[ "def FUNC_18(self):...\n", "VAR_5 = {'rooms': ['!secretbase:unknown']}\n", "VAR_6 = FUNC_0(sender='@foo:bar', type='m.room.message', room_id=\n '!anothersecretbase:unknown')\n", "self.assertFalse(Filter(VAR_5).check(VAR_6))\n" ]
[ "def test_definition_rooms_works_with_unknowns(self):...\n", "definition = {'rooms': ['!secretbase:unknown']}\n", "event = MockEvent(sender='@foo:bar', type='m.room.message', room_id=\n '!anothersecretbase:unknown')\n", "self.assertFalse(Filter(definition).check(event))\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Expr'" ]
[ "def FUNC_31(self):...\n", "self.parser = saved_model_cli.create_parser()\n", "VAR_9 = test.test_src_dir_path(VAR_0)\n", "VAR_44 = np.array([[1], [2]])\n", "VAR_45 = np.zeros((6, 3))\n", "VAR_31 = os.path.join(test.get_temp_dir(),\n 'testRunCommandOutOverwrite_inputs.npz')\n", "np.savez(VAR_31, VAR_26=x, VAR_27=x_notused)\n", "VAR_46 = os.path.join(test.get_temp_dir(), 'y.npy')\n", "open(VAR_46, 'a').close()\n", "VAR_10 = self.parser.parse_args(['run', '--dir', VAR_9, '--tag_set',\n 'serve', '--signature_def', 'serving_default', '--inputs', 'x=' +\n VAR_31 + '[x0]', '--outdir', test.get_temp_dir()])\n", "saved_model_cli.run(VAR_10)\n" ]
[ "def testRunCommandOutputFileExistError(self):...\n", "self.parser = saved_model_cli.create_parser()\n", "base_path = test.test_src_dir_path(SAVED_MODEL_PATH)\n", "x = np.array([[1], [2]])\n", "x_notused = np.zeros((6, 3))\n", "input_path = os.path.join(test.get_temp_dir(),\n 'testRunCommandOutOverwrite_inputs.npz')\n", "np.savez(input_path, x0=x, x1=x_notused)\n", "output_file = os.path.join(test.get_temp_dir(), 'y.npy')\n", "open(output_file, 'a').close()\n", "args = self.parser.parse_args(['run', '--dir', base_path, '--tag_set',\n 'serve', '--signature_def', 'serving_default', '--inputs', 'x=' +\n input_path + '[x0]', '--outdir', test.get_temp_dir()])\n", "saved_model_cli.run(args)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "Expr'", "Assign'", "Expr'" ]
[ "@VAR_0.before_request...\n", "g.conn = Connections(VAR_0.config)\n" ]
[ "@app.before_request...\n", "g.conn = Connections(app.config)\n" ]
[ 0, 0 ]
[ "For", "Assign'" ]
[ "def FUNC_3(self, VAR_13: IProtocolFactory) ->defer.Deferred:...\n", "\"\"\"docstring\"\"\"\n", "return run_in_background(self._do_connect, VAR_13)\n" ]
[ "def connect(self, protocol_factory: IProtocolFactory) ->defer.Deferred:...\n", "\"\"\"docstring\"\"\"\n", "return run_in_background(self._do_connect, protocol_factory)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Return'" ]
[ "def FUNC_8(self):...\n", "super().setUp()\n", "self.server_factory = ReplicationStreamProtocolFactory(self.hs)\n", "self.streamer = self.hs.get_replication_streamer()\n", "self._redis_server = CLASS_6()\n", "VAR_43 = self.hs.get_datastore()\n", "self.database_pool = VAR_43.db_pool\n", "self.reactor.lookups['testserv'] = '1.2.3.4'\n", "self.reactor.lookups['localhost'] = '127.0.0.1'\n", "self._hs_to_site = {self.hs: self.site}\n", "if self.hs.config.redis.redis_enabled:\n", "self.reactor.add_tcp_client_callback('localhost', 6379, self.\n connect_any_redis_attempts)\n", "self.reactor.add_tcp_client_callback('1.2.3.4', 8765, lambda : self.\n _handle_http_replication_attempt(self.hs, 8765))\n", "self.hs.get_tcp_replication().start_replication(self.hs)\n" ]
[ "def setUp(self):...\n", "super().setUp()\n", "self.server_factory = ReplicationStreamProtocolFactory(self.hs)\n", "self.streamer = self.hs.get_replication_streamer()\n", "self._redis_server = FakeRedisPubSubServer()\n", "store = self.hs.get_datastore()\n", "self.database_pool = store.db_pool\n", "self.reactor.lookups['testserv'] = '1.2.3.4'\n", "self.reactor.lookups['localhost'] = '127.0.0.1'\n", "self._hs_to_site = {self.hs: self.site}\n", "if self.hs.config.redis.redis_enabled:\n", "self.reactor.add_tcp_client_callback('localhost', 6379, self.\n connect_any_redis_attempts)\n", "self.reactor.add_tcp_client_callback('1.2.3.4', 8765, lambda : self.\n _handle_http_replication_attempt(self.hs, 8765))\n", "self.hs.get_tcp_replication().start_replication(self.hs)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Expr'", "Expr'", "Expr'" ]
[ "def FUNC_2(self, VAR_1):...\n", "from openapi_python_client.parser.properties import Property\n", "VAR_5 = VAR_1.MagicMock()\n", "VAR_1.patch('openapi_python_client.utils.snake_case')\n", "VAR_4 = Property(VAR_5=name, VAR_26=True, default=None)\n", "assert VAR_4.get_imports(VAR_9='') == set()\n", "VAR_4.required = False\n", "assert VAR_4.get_imports(VAR_9='') == {'from typing import Optional'}\n" ]
[ "def test_get_imports(self, mocker):...\n", "from openapi_python_client.parser.properties import Property\n", "name = mocker.MagicMock()\n", "mocker.patch('openapi_python_client.utils.snake_case')\n", "p = Property(name=name, required=True, default=None)\n", "assert p.get_imports(prefix='') == set()\n", "p.required = False\n", "assert p.get_imports(prefix='') == {'from typing import Optional'}\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "ImportFrom'", "Assign'", "Expr'", "Assign'", "Assert'", "Assign'", "Assert'" ]
[ "\"\"\" A FastAPI app used to create an OpenAPI document for end-to-end testing \"\"\"\n", "import json\n", "from datetime import date, datetime\n", "from enum import Enum\n", "from pathlib import Path\n", "from typing import Any, Dict, List, Union\n", "from fastapi import APIRouter, FastAPI, File, Header, Query, UploadFile\n", "from pydantic import BaseModel\n", "VAR_0 = FastAPI(title='My Test API', description=\n 'An API for testing openapi-python-client')\n", "@VAR_0.get('/ping', response_model=bool)...\n", "\"\"\"docstring\"\"\"\n", "return True\n" ]
[ "\"\"\" A FastAPI app used to create an OpenAPI document for end-to-end testing \"\"\"\n", "import json\n", "from datetime import date, datetime\n", "from enum import Enum\n", "from pathlib import Path\n", "from typing import Any, Dict, List, Union\n", "from fastapi import APIRouter, FastAPI, File, Header, Query, UploadFile\n", "from pydantic import BaseModel\n", "app = FastAPI(title='My Test API', description=\n 'An API for testing openapi-python-client')\n", "@app.get('/ping', response_model=bool)...\n", "\"\"\"docstring\"\"\"\n", "return True\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Expr'", "Import'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Assign'", "Condition", "Docstring", "Return'" ]
[ "def FUNC_14(VAR_35: Fingerprint, VAR_36: Optional[Text], VAR_37: Text,...\n", "\"\"\"docstring\"\"\"\n", "VAR_61 = CLASS_1()\n", "if VAR_36 is None or not os.path.exists(VAR_36):\n", "return VAR_61\n", "VAR_40 = FUNC_10(unpacked)\n", "VAR_0.error(\n f\"\"\"Failed to get the fingerprint. Error: {e}.\nProceeding with running default retrain...\"\"\"\n )\n", "VAR_67, VAR_68 = FUNC_4(unpacked)\n", "return VAR_61\n", "VAR_61 = CLASS_1(VAR_41=did_section_fingerprint_change(last_fingerprint,\n new_fingerprint, SECTION_CORE), VAR_42=did_section_fingerprint_change(\n last_fingerprint, new_fingerprint, SECTION_NLU), VAR_48=\n did_section_fingerprint_change(last_fingerprint, new_fingerprint,\n SECTION_NLG), VAR_39=force_training)\n", "if VAR_38 and VAR_61.should_retrain_nlu():\n", "VAR_61.core = True\n", "VAR_69 = False\n", "if not VAR_61.should_retrain_core():\n", "VAR_70 = os.path.join(VAR_37, DEFAULT_CORE_SUBDIRECTORY_NAME)\n", "if not VAR_61.should_retrain_nlg() and VAR_69:\n", "VAR_69 = not FUNC_13(VAR_67, VAR_70)\n", "VAR_61.nlg = True\n", "if not VAR_61.should_retrain_nlu():\n", "VAR_61.core = VAR_69\n", "VAR_70 = os.path.join(VAR_37, 'nlu')\n", "return VAR_61\n", "VAR_61.nlu = not FUNC_13(VAR_68, VAR_70)\n" ]
[ "def should_retrain(new_fingerprint: Fingerprint, old_model: Optional[Text],...\n", "\"\"\"docstring\"\"\"\n", "fingerprint_comparison = FingerprintComparisonResult()\n", "if old_model is None or not os.path.exists(old_model):\n", "return fingerprint_comparison\n", "last_fingerprint = fingerprint_from_path(unpacked)\n", "logger.error(\n f\"\"\"Failed to get the fingerprint. Error: {e}.\nProceeding with running default retrain...\"\"\"\n )\n", "old_core, old_nlu = get_model_subdirectories(unpacked)\n", "return fingerprint_comparison\n", "fingerprint_comparison = FingerprintComparisonResult(core=\n did_section_fingerprint_change(last_fingerprint, new_fingerprint,\n SECTION_CORE), nlu=did_section_fingerprint_change(last_fingerprint,\n new_fingerprint, SECTION_NLU), nlg=did_section_fingerprint_change(\n last_fingerprint, new_fingerprint, SECTION_NLG), force_training=\n force_training)\n", "if has_e2e_examples and fingerprint_comparison.should_retrain_nlu():\n", "fingerprint_comparison.core = True\n", "core_merge_failed = False\n", "if not fingerprint_comparison.should_retrain_core():\n", "target_path = os.path.join(train_path, DEFAULT_CORE_SUBDIRECTORY_NAME)\n", "if not fingerprint_comparison.should_retrain_nlg() and core_merge_failed:\n", "core_merge_failed = not move_model(old_core, target_path)\n", "fingerprint_comparison.nlg = True\n", "if not fingerprint_comparison.should_retrain_nlu():\n", "fingerprint_comparison.core = core_merge_failed\n", "target_path = os.path.join(train_path, 'nlu')\n", "return fingerprint_comparison\n", "fingerprint_comparison.nlu = not move_model(old_nlu, target_path)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Condition", "Return'", "Assign'", "Expr'", "Assign'", "Return'", "Assign'", "Condition", "Assign'", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Return'", "Assign'" ]
[ "def FUNC_18(VAR_14, VAR_31):...\n", "return FUNC_19(FUNC_23(VAR_14, VAR_31))\n" ]
[ "def _compute_date_for_i18n(locale, domain):...\n", "return _compute_date(_get_all_translationfiles(locale, domain))\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_1(self, VAR_1):...\n", "\"\"\"docstring\"\"\"\n", "VAR_2 = VAR_1.requestUrl()\n", "if VAR_2.scheme() in ['chrome-error', 'chrome-extension']:\n", "VAR_1.fail(QWebEngineUrlRequestJob.UrlInvalid)\n", "assert VAR_1.requestMethod() == b'GET'\n", "return\n", "assert VAR_2.scheme() == 'qute'\n", "log.misc.debug('Got request for {}'.format(VAR_2.toDisplayString()))\n", "VAR_3, VAR_4 = qutescheme.data_for_url(VAR_2)\n", "log.misc.debug('No handler found for {}'.format(VAR_2.toDisplayString()))\n", "log.misc.debug('Returning {} data'.format(VAR_3))\n", "VAR_1.fail(QWebEngineUrlRequestJob.UrlNotFound)\n", "VAR_5 = QBuffer(parent=self)\n", "log.misc.exception('OSError while handling qute://* URL')\n", "VAR_5.open(QIODevice.WriteOnly)\n", "VAR_1.fail(QWebEngineUrlRequestJob.UrlNotFound)\n", "VAR_5.write(VAR_4)\n", "log.misc.exception('Error while handling qute://* URL')\n", "VAR_5.seek(0)\n", "VAR_1.fail(QWebEngineUrlRequestJob.RequestFailed)\n", "VAR_5.close()\n", "qtutils.ensure_valid(e.url)\n", "VAR_1.reply(VAR_3.encode('ascii'), VAR_5)\n", "VAR_1.redirect(e.url)\n" ]
[ "def requestStarted(self, job):...\n", "\"\"\"docstring\"\"\"\n", "url = job.requestUrl()\n", "if url.scheme() in ['chrome-error', 'chrome-extension']:\n", "job.fail(QWebEngineUrlRequestJob.UrlInvalid)\n", "assert job.requestMethod() == b'GET'\n", "return\n", "assert url.scheme() == 'qute'\n", "log.misc.debug('Got request for {}'.format(url.toDisplayString()))\n", "mimetype, data = qutescheme.data_for_url(url)\n", "log.misc.debug('No handler found for {}'.format(url.toDisplayString()))\n", "log.misc.debug('Returning {} data'.format(mimetype))\n", "job.fail(QWebEngineUrlRequestJob.UrlNotFound)\n", "buf = QBuffer(parent=self)\n", "log.misc.exception('OSError while handling qute://* URL')\n", "buf.open(QIODevice.WriteOnly)\n", "job.fail(QWebEngineUrlRequestJob.UrlNotFound)\n", "buf.write(data)\n", "log.misc.exception('Error while handling qute://* URL')\n", "buf.seek(0)\n", "job.fail(QWebEngineUrlRequestJob.RequestFailed)\n", "buf.close()\n", "qtutils.ensure_valid(e.url)\n", "job.reply(mimetype.encode('ascii'), buf)\n", "job.redirect(e.url)\n" ]
[ 0, 0, 0, 0, 0, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Condition", "Expr'", "Assert'", "Return'", "Assert'", "Expr'", "Assign'", "Expr'", "Expr'", "Expr'", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'" ]
[ "def FUNC_28(self, VAR_27):...\n", "return [entry.to_dict() for entry in self.playlistdb.loadPlaylist(VAR_27=\n playlistid, VAR_15=self.getUserId())]\n" ]
[ "def api_loadplaylist(self, playlistid):...\n", "return [entry.to_dict() for entry in self.playlistdb.loadPlaylist(\n playlistid=playlistid, userid=self.getUserId())]\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_54(VAR_21, VAR_31, VAR_32):...\n", "for serie in VAR_31:\n", "VAR_21 = VAR_21.filter(db.Books.series.any(db.Series.id == serie))\n", "for serie in VAR_32:\n", "VAR_21 = VAR_21.filter(not_(db.Books.series.any(db.Series.id == serie)))\n", "return VAR_21\n" ]
[ "def adv_search_serie(q, include_series_inputs, exclude_series_inputs):...\n", "for serie in include_series_inputs:\n", "q = q.filter(db.Books.series.any(db.Series.id == serie))\n", "for serie in exclude_series_inputs:\n", "q = q.filter(not_(db.Books.series.any(db.Series.id == serie)))\n", "return q\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "For", "Assign'", "For", "Assign'", "Return'" ]
[ "def FUNC_17(VAR_0, VAR_1: FlaskClient):...\n", "VAR_8 = VAR_1.post('/folders/create', data={'parent_dir': ''},\n follow_redirects=True)\n", "assert VAR_8.status_code == 200\n", "assert request.path == '/'\n", "assert b'Could not create folder.' in VAR_8.data\n" ]
[ "def test_creating_without_dirname_fails(test_app, client: FlaskClient):...\n", "resp = client.post('/folders/create', data={'parent_dir': ''},\n follow_redirects=True)\n", "assert resp.status_code == 200\n", "assert request.path == '/'\n", "assert b'Could not create folder.' in resp.data\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assert'", "Assert'", "Assert'" ]
[ "def FUNC_22(self):...\n", "\"\"\"docstring\"\"\"\n", "self._producer.resumeProducing()\n", "VAR_0.exception('Failed to call resumeProducing')\n", "self._consumer.unregisterProducer()\n", "self.stopProducing()\n" ]
[ "def _run_once(self):...\n", "\"\"\"docstring\"\"\"\n", "self._producer.resumeProducing()\n", "logger.exception('Failed to call resumeProducing')\n", "self._consumer.unregisterProducer()\n", "self.stopProducing()\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Expr'", "Expr'", "Expr'", "Expr'" ]
[ "def __call__(self, VAR_17=None, VAR_18={}, VAR_19=None, **VAR_20):...\n", "return 'dummy'\n" ]
[ "def __call__(self, client=None, REQUEST={}, RESPONSE=None, **kw):...\n", "return 'dummy'\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_0(VAR_7: IPAddress, VAR_8: Optional[IPSet], VAR_9: IPSet) ->bool:...\n", "\"\"\"docstring\"\"\"\n", "if VAR_7 in VAR_9:\n", "if VAR_8 is None or VAR_7 not in VAR_8:\n", "return False\n", "return True\n" ]
[ "def check_against_blacklist(ip_address: IPAddress, ip_whitelist: Optional[...\n", "\"\"\"docstring\"\"\"\n", "if ip_address in ip_blacklist:\n", "if ip_whitelist is None or ip_address not in ip_whitelist:\n", "return False\n", "return True\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Condition", "Return'", "Return'" ]
[ "def FUNC_14(self):...\n", "self.assert_expected(self.folder.t, 'CheckNothing.html')\n" ]
[ "def testReplaceWithNothing(self):...\n", "self.assert_expected(self.folder.t, 'CheckNothing.html')\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Expr'" ]
[ "def FUNC_36(self):...\n", "self.login()\n", "VAR_3 = self.client.post('/password_change/custom/named/', {'old_password':\n 'password', 'new_password1': 'password1', 'new_password2': 'password1'})\n", "self.assertEqual(VAR_3.status_code, 302)\n", "self.assertURLEqual(VAR_3.url, '/password_reset/')\n" ]
[ "def test_password_change_redirect_custom_named(self):...\n", "self.login()\n", "response = self.client.post('/password_change/custom/named/', {\n 'old_password': 'password', 'new_password1': 'password1',\n 'new_password2': 'password1'})\n", "self.assertEqual(response.status_code, 302)\n", "self.assertURLEqual(response.url, '/password_reset/')\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Assign'", "Expr'", "Expr'" ]
[ "@app.route('/dataobj/delete/<int:dataobj_id>', methods=['POST'])...\n", "" ]
[ "@app.route('/dataobj/delete/<int:dataobj_id>', methods=['POST'])...\n", "" ]
[ 0, 0 ]
[ "Condition", "Condition" ]
[ "def FUNC_21(self):...\n", "from zope.tales.expressions import LazyWrapper\n", "VAR_5 = self._makeContext()\n", "VAR_8 = VAR_5.evaluate('lazy: b')\n", "self.assertIsInstance(VAR_8, LazyWrapper)\n" ]
[ "def test_lazy_expression_returns_wrapper(self):...\n", "from zope.tales.expressions import LazyWrapper\n", "ec = self._makeContext()\n", "lazy = ec.evaluate('lazy: b')\n", "self.assertIsInstance(lazy, LazyWrapper)\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "ImportFrom'", "Assign'", "Assign'", "Expr'" ]
[ "def FUNC_3(VAR_0):...\n", "VAR_9 = Category.objects.visible().parents().ordered()\n", "VAR_10 = Topic.objects.visible().global_().with_bookmarks(VAR_4=request.user\n ).order_by('-is_globally_pinned', '-last_active').select_related('category'\n )\n", "VAR_10 = yt_paginate(VAR_10, per_page=config.topics_per_page, page_number=\n request.GET.get('page', 1))\n", "return render(VAR_0=request, template_name='spirit/topic/active.html',\n context={'categories': categories, 'topics': topics})\n" ]
[ "def index_active(request):...\n", "categories = Category.objects.visible().parents().ordered()\n", "topics = Topic.objects.visible().global_().with_bookmarks(user=request.user\n ).order_by('-is_globally_pinned', '-last_active').select_related('category'\n )\n", "topics = yt_paginate(topics, per_page=config.topics_per_page, page_number=\n request.GET.get('page', 1))\n", "return render(request=request, template_name='spirit/topic/active.html',\n context={'categories': categories, 'topics': topics})\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_50(VAR_19=None, VAR_20=False):...\n", "VAR_86 = {}\n", "if VAR_19:\n", "VAR_0.error(VAR_19)\n", "VAR_86['result'] = [{'type': 'success', 'message': _(\n u'Calibre-Web configuration updated')}]\n", "config.load()\n", "VAR_86['reboot'] = VAR_20\n", "VAR_86['result'] = [{'type': 'danger', 'message': VAR_19}]\n", "VAR_86['config_upload'] = config.config_upload_formats\n", "return Response(json.dumps(VAR_86), mimetype='application/json')\n" ]
[ "def _configuration_result(error_flash=None, reboot=False):...\n", "resp = {}\n", "if error_flash:\n", "log.error(error_flash)\n", "resp['result'] = [{'type': 'success', 'message': _(\n u'Calibre-Web configuration updated')}]\n", "config.load()\n", "resp['reboot'] = reboot\n", "resp['result'] = [{'type': 'danger', 'message': error_flash}]\n", "resp['config_upload'] = config.config_upload_formats\n", "return Response(json.dumps(resp), mimetype='application/json')\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Condition", "Expr'", "Assign'", "Expr'", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "def __init__(self, VAR_24=None, VAR_25=None, **VAR_7):...\n", "VAR_7['label'] = ''\n", "super().__init__(widget=PluralTextarea, **kwargs)\n" ]
[ "def __init__(self, max_length=None, min_length=None, **kwargs):...\n", "kwargs['label'] = ''\n", "super().__init__(widget=PluralTextarea, **kwargs)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Expr'" ]
[ "@FUNC_0...\n", "return Filtering(self)\n" ]
[ "@cache_in_self...\n", "return Filtering(self)\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "def FUNC_7(self):...\n", "while not self.get_success(self.store.db_pool.updates.\n", "self.get_success(self.store.db_pool.updates.do_next_background_update(100),\n by=0.1)\n", "VAR_13 = UserID('alice', 'test')\n", "VAR_14 = create_requester(VAR_13)\n", "self.get_success(self.room_creator.create_room(VAR_14, {}))\n", "self.get_success(self.store.db_pool.simple_insert(table=\n 'background_updates', values={'update_name':\n 'current_state_events_membership', 'progress_json': '{}', 'depends_on':\n None}))\n", "self.store.db_pool.updates._all_done = False\n", "while not self.get_success(self.store.db_pool.updates.\n", "self.get_success(self.store.db_pool.updates.do_next_background_update(100),\n by=0.1)\n" ]
[ "def test_can_rerun_update(self):...\n", "while not self.get_success(self.store.db_pool.updates.\n", "self.get_success(self.store.db_pool.updates.do_next_background_update(100),\n by=0.1)\n", "user = UserID('alice', 'test')\n", "requester = create_requester(user)\n", "self.get_success(self.room_creator.create_room(requester, {}))\n", "self.get_success(self.store.db_pool.simple_insert(table=\n 'background_updates', values={'update_name':\n 'current_state_events_membership', 'progress_json': '{}', 'depends_on':\n None}))\n", "self.store.db_pool.updates._all_done = False\n", "while not self.get_success(self.store.db_pool.updates.\n", "self.get_success(self.store.db_pool.updates.do_next_background_update(100),\n by=0.1)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Expr'", "Assign'", "Assign'", "Expr'", "Expr'", "Assign'", "Condition", "Expr'" ]
[ "def FUNC_14(self, VAR_0, VAR_1, VAR_2):...\n", "self.presence_handler = VAR_2.get_presence_handler()\n", "self.clock = VAR_2.get_clock()\n" ]
[ "def prepare(self, reactor, clock, hs):...\n", "self.presence_handler = hs.get_presence_handler()\n", "self.clock = hs.get_clock()\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'" ]
[ "def FUNC_4(self, VAR_24, VAR_25=None):...\n", "self.encoding = VAR_25\n", "VAR_27 = False\n", "self.errors = []\n", "if VAR_25 and self.inject_meta_charset:\n", "from ..filters.inject_meta_charset import Filter\n", "if self.strip_whitespace:\n", "VAR_24 = Filter(VAR_24, VAR_25)\n", "from ..filters.whitespace import Filter\n", "if self.sanitize:\n", "VAR_24 = Filter(VAR_24)\n", "from ..filters.sanitizer import Filter\n", "if self.omit_optional_tags:\n", "VAR_24 = Filter(VAR_24)\n", "from ..filters.optionaltags import Filter\n", "if self.alphabetical_attributes:\n", "VAR_24 = Filter(VAR_24)\n", "from ..filters.alphabeticalattributes import Filter\n", "for token in VAR_24:\n", "VAR_24 = Filter(VAR_24)\n", "VAR_31 = token['type']\n", "if VAR_31 == 'Doctype':\n", "VAR_35 = '<!DOCTYPE %s' % token['name']\n", "if VAR_31 in ('Characters', 'SpaceCharacters'):\n", "if token['publicId']:\n", "if VAR_31 == 'SpaceCharacters' or VAR_27:\n", "if VAR_31 in ('StartTag', 'EmptyTag'):\n", "VAR_35 += ' PUBLIC \"%s\"' % token['publicId']\n", "if token['systemId']:\n", "if VAR_27 and token['data'].find('</') >= 0:\n", "yield self.encode(escape(token['data']))\n", "VAR_37 = token['name']\n", "if VAR_31 == 'EndTag':\n", "if token['systemId']:\n", "VAR_35 += ' SYSTEM'\n", "self.serializeError('Unexpected </ in CDATA')\n", "yield self.encode(token['data'])\n", "yield self.encodeStrict('<%s' % VAR_37)\n", "VAR_37 = token['name']\n", "if VAR_31 == 'Comment':\n", "if token['systemId'].find('\"') >= 0:\n", "VAR_35 += '>'\n", "if VAR_37 in rcdataElements and not self.escape_rcdata:\n", "if VAR_37 in rcdataElements:\n", "VAR_26 = token['data']\n", "if VAR_31 == 'Entity':\n", "if token['systemId'].find(\"'\") >= 0:\n", "VAR_8 = '\"'\n", "yield self.encodeStrict(VAR_35)\n", "VAR_27 = True\n", "if VAR_27:\n", "VAR_27 = False\n", "if VAR_27:\n", "if VAR_26.find('--') >= 0:\n", "VAR_37 = token['name']\n", "self.serializeError(token['data'])\n", "self.serializeError(\n 'System identifer contains both single and double quote characters')\n", "VAR_8 = \"'\"\n", "VAR_35 += ' %s%s%s' % (VAR_8, token['systemId'], VAR_8)\n", "for (attr_namespace, attr_name), attr_value in token['data'].items():\n", "self.serializeError('Unexpected child element of a CDATA element')\n", "yield self.encodeStrict('</%s>' % VAR_37)\n", "self.serializeError('Unexpected child element of a CDATA element')\n", "self.serializeError('Comment contains --')\n", "yield self.encodeStrict('<!--%s-->' % token['data'])\n", "VAR_40 = VAR_37 + ';'\n", "VAR_38 = attr_name\n", "if VAR_37 in voidElements and self.use_trailing_solidus:\n", "if VAR_40 not in entities:\n", "VAR_32 = attr_value\n", "if self.space_before_trailing_solidus:\n", "yield self.encode('>')\n", "self.serializeError('Entity %s not recognized' % VAR_37)\n", "if self.resolve_entities and VAR_40 not in xmlEntities:\n", "yield self.encodeStrict(' ')\n", "yield self.encodeStrict(' /')\n", "yield self.encodeStrict('/')\n", "VAR_26 = entities[VAR_40]\n", "VAR_26 = '&%s;' % VAR_37\n", "yield self.encodeStrict(VAR_38)\n", "yield self.encodeStrict(VAR_26)\n", "if not self.minimize_boolean_attributes or VAR_38 not in booleanAttributes.get(\n", "yield self.encodeStrict('=')\n", "if self.quote_attr_values:\n", "VAR_39 = True\n", "VAR_39 = len(VAR_32) == 0 or VAR_1.search(VAR_32)\n", "VAR_32 = VAR_32.replace('&', '&amp;')\n", "if self.escape_lt_in_attrs:\n", "VAR_32 = VAR_32.replace('<', '&lt;')\n", "if VAR_39:\n", "VAR_8 = self.quote_char\n", "yield self.encode(VAR_32)\n", "if self.use_best_quote_char:\n", "if \"'\" in VAR_32 and '\"' not in VAR_32:\n", "if VAR_8 == \"'\":\n", "VAR_8 = '\"'\n", "if '\"' in VAR_32 and \"'\" not in VAR_32:\n", "VAR_32 = VAR_32.replace(\"'\", '&#39;')\n", "VAR_32 = VAR_32.replace('\"', '&quot;')\n", "VAR_8 = \"'\"\n", "yield self.encodeStrict(VAR_8)\n", "yield self.encode(VAR_32)\n", "yield self.encodeStrict(VAR_8)\n" ]
[ "def serialize(self, treewalker, encoding=None):...\n", "self.encoding = encoding\n", "in_cdata = False\n", "self.errors = []\n", "if encoding and self.inject_meta_charset:\n", "from ..filters.inject_meta_charset import Filter\n", "if self.strip_whitespace:\n", "treewalker = Filter(treewalker, encoding)\n", "from ..filters.whitespace import Filter\n", "if self.sanitize:\n", "treewalker = Filter(treewalker)\n", "from ..filters.sanitizer import Filter\n", "if self.omit_optional_tags:\n", "treewalker = Filter(treewalker)\n", "from ..filters.optionaltags import Filter\n", "if self.alphabetical_attributes:\n", "treewalker = Filter(treewalker)\n", "from ..filters.alphabeticalattributes import Filter\n", "for token in treewalker:\n", "treewalker = Filter(treewalker)\n", "type = token['type']\n", "if type == 'Doctype':\n", "doctype = '<!DOCTYPE %s' % token['name']\n", "if type in ('Characters', 'SpaceCharacters'):\n", "if token['publicId']:\n", "if type == 'SpaceCharacters' or in_cdata:\n", "if type in ('StartTag', 'EmptyTag'):\n", "doctype += ' PUBLIC \"%s\"' % token['publicId']\n", "if token['systemId']:\n", "if in_cdata and token['data'].find('</') >= 0:\n", "yield self.encode(escape(token['data']))\n", "name = token['name']\n", "if type == 'EndTag':\n", "if token['systemId']:\n", "doctype += ' SYSTEM'\n", "self.serializeError('Unexpected </ in CDATA')\n", "yield self.encode(token['data'])\n", "yield self.encodeStrict('<%s' % name)\n", "name = token['name']\n", "if type == 'Comment':\n", "if token['systemId'].find('\"') >= 0:\n", "doctype += '>'\n", "if name in rcdataElements and not self.escape_rcdata:\n", "if name in rcdataElements:\n", "data = token['data']\n", "if type == 'Entity':\n", "if token['systemId'].find(\"'\") >= 0:\n", "quote_char = '\"'\n", "yield self.encodeStrict(doctype)\n", "in_cdata = True\n", "if in_cdata:\n", "in_cdata = False\n", "if in_cdata:\n", "if data.find('--') >= 0:\n", "name = token['name']\n", "self.serializeError(token['data'])\n", "self.serializeError(\n 'System identifer contains both single and double quote characters')\n", "quote_char = \"'\"\n", "doctype += ' %s%s%s' % (quote_char, token['systemId'], quote_char)\n", "for (attr_namespace, attr_name), attr_value in token['data'].items():\n", "self.serializeError('Unexpected child element of a CDATA element')\n", "yield self.encodeStrict('</%s>' % name)\n", "self.serializeError('Unexpected child element of a CDATA element')\n", "self.serializeError('Comment contains --')\n", "yield self.encodeStrict('<!--%s-->' % token['data'])\n", "key = name + ';'\n", "k = attr_name\n", "if name in voidElements and self.use_trailing_solidus:\n", "if key not in entities:\n", "v = attr_value\n", "if self.space_before_trailing_solidus:\n", "yield self.encode('>')\n", "self.serializeError('Entity %s not recognized' % name)\n", "if self.resolve_entities and key not in xmlEntities:\n", "yield self.encodeStrict(' ')\n", "yield self.encodeStrict(' /')\n", "yield self.encodeStrict('/')\n", "data = entities[key]\n", "data = '&%s;' % name\n", "yield self.encodeStrict(k)\n", "yield self.encodeStrict(data)\n", "if not self.minimize_boolean_attributes or k not in booleanAttributes.get(name,\n", "yield self.encodeStrict('=')\n", "if self.quote_attr_values:\n", "quote_attr = True\n", "quote_attr = len(v) == 0 or quoteAttributeSpec.search(v)\n", "v = v.replace('&', '&amp;')\n", "if self.escape_lt_in_attrs:\n", "v = v.replace('<', '&lt;')\n", "if quote_attr:\n", "quote_char = self.quote_char\n", "yield self.encode(v)\n", "if self.use_best_quote_char:\n", "if \"'\" in v and '\"' not in v:\n", "if quote_char == \"'\":\n", "quote_char = '\"'\n", "if '\"' in v and \"'\" not in v:\n", "v = v.replace(\"'\", '&#39;')\n", "v = v.replace('\"', '&quot;')\n", "quote_char = \"'\"\n", "yield self.encodeStrict(quote_char)\n", "yield self.encode(v)\n", "yield self.encodeStrict(quote_char)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2, 0, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Condition", "ImportFrom'", "Condition", "Assign'", "ImportFrom'", "Condition", "Assign'", "ImportFrom'", "Condition", "Assign'", "ImportFrom'", "Condition", "Assign'", "ImportFrom'", "For", "Assign'", "Assign'", "Condition", "Assign'", "Condition", "Condition", "Condition", "Condition", "AugAssign'", "Condition", "Condition", "Expr'", "Assign'", "Condition", "Condition", "AugAssign'", "Expr'", "Expr'", "Expr'", "Assign'", "Condition", "Condition", "AugAssign'", "Condition", "Condition", "Assign'", "Condition", "Condition", "Assign'", "Expr'", "Assign'", "Condition", "Assign'", "Condition", "Condition", "Assign'", "Expr'", "Expr'", "Assign'", "AugAssign'", "For", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Assign'", "Assign'", "Condition", "Condition", "Assign'", "For", "Expr'", "Expr'", "Condition", "Expr'", "Expr'", "Expr'", "Assign'", "Assign'", "Expr'", "Expr'", "Condition", "Expr'", "Condition", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Expr'", "Condition", "Condition", "Condition", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'" ]
[ "def FUNC_19(self, VAR_0):...\n", "from openapi_python_client import Project\n", "VAR_8 = Project(VAR_3=mocker.MagicMock(title='My Test API'))\n", "VAR_8.package_dir = VAR_0.MagicMock()\n", "VAR_8.package_dir.is_dir.return_value = False\n", "VAR_8._build_models = VAR_0.MagicMock()\n", "VAR_8.update()\n", "VAR_8.package_dir.is_dir.assert_called_once()\n", "VAR_8._build_models.assert_not_called()\n" ]
[ "def test_update_missing_dir(self, mocker):...\n", "from openapi_python_client import Project\n", "project = Project(openapi=mocker.MagicMock(title='My Test API'))\n", "project.package_dir = mocker.MagicMock()\n", "project.package_dir.is_dir.return_value = False\n", "project._build_models = mocker.MagicMock()\n", "project.update()\n", "project.package_dir.is_dir.assert_called_once()\n", "project._build_models.assert_not_called()\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "ImportFrom'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'" ]
[ "@FUNC_0...\n", "if self.config.worker_app:\n", "return DeviceWorkerHandler(self)\n", "return DeviceHandler(self)\n" ]
[ "@cache_in_self...\n", "if self.config.worker_app:\n", "return DeviceWorkerHandler(self)\n", "return DeviceHandler(self)\n" ]
[ 0, 0, 0, 0 ]
[ "Condition", "Condition", "Return'", "Return'" ]
[ "def FUNC_61(VAR_118, *VAR_119):...\n", "\"\"\"docstring\"\"\"\n", "if not 'public' in VAR_119:\n", "VAR_119 = [FUNC_56(part) for part in VAR_119]\n", "return os.path.join(os.path.dirname(FUNC_55(FUNC_56(VAR_118)).__file__), *\n VAR_119)\n" ]
[ "def get_pymodule_path(modulename, *joins):...\n", "\"\"\"docstring\"\"\"\n", "if not 'public' in joins:\n", "joins = [scrub(part) for part in joins]\n", "return os.path.join(os.path.dirname(get_module(scrub(modulename)).__file__),\n *joins)\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Assign'", "Return'" ]
[ "def FUNC_63(self, VAR_34, VAR_33=None):...\n", "\"\"\"docstring\"\"\"\n", "if not (isinstance(self.get(VAR_34), list) and len(self.get(VAR_34)) > 0):\n", "VAR_93 = self.meta.get_label(VAR_34)\n", "frappe.throw(_('Table {0} cannot be empty').format(VAR_93), VAR_33 or\n frappe.EmptyTableError)\n" ]
[ "def validate_table_has_rows(self, parentfield, raise_exception=None):...\n", "\"\"\"docstring\"\"\"\n", "if not (isinstance(self.get(parentfield), list) and len(self.get(\n", "label = self.meta.get_label(parentfield)\n", "frappe.throw(_('Table {0} cannot be empty').format(label), raise_exception or\n frappe.EmptyTableError)\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Assign'", "Expr'" ]
[ "async def FUNC_6(VAR_6):...\n", "" ]
[ "async def middleware_handler(request):...\n", "" ]
[ 0, 0 ]
[ "AsyncFunctionDef'", "Condition" ]
[ "def FUNC_9(self):...\n", "return 'dataverse-' + escapism.escape(self.identifier, escape_char='-').lower()\n" ]
[ "def get_build_slug(self):...\n", "return 'dataverse-' + escapism.escape(self.identifier, escape_char='-').lower()\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_32(self):...\n", "if self.meta.istable:\n", "VAR_85 = frappe.get_meta(self.parenttype).permissions\n", "VAR_85 = self.meta.permissions\n", "return VAR_85\n" ]
[ "def get_permissions(self):...\n", "if self.meta.istable:\n", "permissions = frappe.get_meta(self.parenttype).permissions\n", "permissions = self.meta.permissions\n", "return permissions\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_18(self):...\n", "VAR_26 = np.array([[1], [2]])\n", "VAR_31 = os.path.join(test.get_temp_dir(), 'input.npz')\n", "np.savez(VAR_31, VAR_53=x0)\n", "VAR_27 = np.ones([2, 10])\n", "VAR_20 = 'x0=' + VAR_31 + '[a]'\n", "VAR_21 = 'x1=np.ones([2,10])'\n", "VAR_30 = saved_model_cli.load_inputs_from_input_arg_string(VAR_20, VAR_21, '')\n", "self.assertTrue(np.all(VAR_30['x0'] == VAR_26))\n", "self.assertTrue(np.all(VAR_30['x1'] == VAR_27))\n" ]
[ "def testInputParserBoth(self):...\n", "x0 = np.array([[1], [2]])\n", "input_path = os.path.join(test.get_temp_dir(), 'input.npz')\n", "np.savez(input_path, a=x0)\n", "x1 = np.ones([2, 10])\n", "input_str = 'x0=' + input_path + '[a]'\n", "input_expr_str = 'x1=np.ones([2,10])'\n", "feed_dict = saved_model_cli.load_inputs_from_input_arg_string(input_str,\n input_expr_str, '')\n", "self.assertTrue(np.all(feed_dict['x0'] == x0))\n", "self.assertTrue(np.all(feed_dict['x1'] == x1))\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'" ]
[ "def FUNC_28():...\n", "\"\"\"docstring\"\"\"\n", "VAR_3 = request.vars.app or ''\n", "VAR_56 = apath('%(app)s' % {'app': VAR_3}, VAR_122=request)\n", "VAR_65 = ['models', 'controllers', 'modules', 'private']\n", "def FUNC_38(VAR_3, VAR_13, VAR_14='.*\\\\.py$'):...\n", "VAR_45 = sorted(listdir(apath('%(app)s/%(dir)s/' % {'app': VAR_3, 'dir':\n VAR_13}, VAR_122=request), VAR_14))\n", "VAR_45 = [x.replace(os.path.sep, '/') for x in VAR_45 if not x.endswith('.bak')\n ]\n", "return VAR_45\n" ]
[ "def todolist():...\n", "\"\"\"docstring\"\"\"\n", "app = request.vars.app or ''\n", "app_path = apath('%(app)s' % {'app': app}, r=request)\n", "dirs = ['models', 'controllers', 'modules', 'private']\n", "def listfiles(app, dir, regexp='.*\\\\.py$'):...\n", "files = sorted(listdir(apath('%(app)s/%(dir)s/' % {'app': app, 'dir': dir},\n r=request), regexp))\n", "files = [x.replace(os.path.sep, '/') for x in files if not x.endswith('.bak')]\n", "return files\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Assign'", "FunctionDef'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_7(self):...\n", "self._test_thumbnail('crop', self.test_image.expected_cropped, self.\n test_image.expected_found)\n" ]
[ "def test_thumbnail_crop(self):...\n", "self._test_thumbnail('crop', self.test_image.expected_cropped, self.\n test_image.expected_found)\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Expr'" ]
[ "def FUNC_52(self, VAR_112=VAR_3, VAR_134=VAR_3, VAR_135=VAR_3, VAR_136=VAR_3):...\n", "\"\"\"docstring\"\"\"\n", "VAR_254 = self.table_user()\n", "VAR_56 = VAR_263.request\n", "VAR_244 = VAR_263.response\n", "VAR_139 = VAR_263.session\n", "if self.is_logged_in():\n", "redirect(self.settings.logged_url, client_side=self.settings.client_side)\n", "if VAR_112 is VAR_3:\n", "VAR_112 = self.get_vars_next() or self.settings.register_next\n", "if VAR_134 is VAR_3:\n", "VAR_134 = self.settings.register_onvalidation\n", "if VAR_135 is VAR_3:\n", "VAR_135 = self.settings.register_onaccept\n", "if VAR_136 is VAR_3:\n", "VAR_136 = self.messages['register_log']\n", "VAR_254 = self.table_user()\n", "if self.settings.login_userfield:\n", "VAR_124 = self.settings.login_userfield\n", "if 'username' in VAR_254.fields:\n", "VAR_272 = IS_NOT_IN_DB(self.db, VAR_254[VAR_124])\n", "VAR_124 = 'username'\n", "VAR_124 = 'email'\n", "if not VAR_254[VAR_124].requires:\n", "VAR_254[VAR_124].requires = VAR_272\n", "if isinstance(VAR_254[VAR_124].requires, (list, tuple)):\n", "VAR_261 = self.settings.password_field\n", "if not any([isinstance(validator, IS_NOT_IN_DB) for validator in VAR_254[\n", "if not isinstance(VAR_254[VAR_124].requires, IS_NOT_IN_DB):\n", "VAR_273 = self.settings.formstyle\n", "if isinstance(VAR_254[VAR_124].requires, list):\n", "VAR_254[VAR_124].requires = [VAR_254[VAR_124].requires, VAR_272]\n", "VAR_254[VAR_261].requires[-1].min_length = self.settings.password_min_length\n", "if self.settings.register_verify_password:\n", "VAR_254[VAR_124].requires.append(VAR_272)\n", "VAR_254[VAR_124].requires += VAR_272,\n", "if self.settings.register_fields is None:\n", "VAR_360 = []\n", "self.settings.register_fields = [VAR_10.name for VAR_10 in VAR_254 if \n VAR_10.writable and not VAR_10.compute]\n", "VAR_360 = [VAR_1('password_two', 'password', VAR_279=IS_EQUAL_TO(request.\n post_vars.get(passfield, None), error_message=self.messages.\n mismatched_password), VAR_60=current.T('Confirm Password'))]\n", "VAR_7 = SQLFORM(VAR_254, VAR_133=self.settings.register_fields, hidden=dict\n (_next=next), showid=self.settings.showid, submit_button=self.messages.\n register_button, delete_label=self.messages.delete_label, VAR_273=\n formstyle, separator=self.settings.label_separator, VAR_360=extra_fields)\n", "VAR_346 = self.settings.register_fields.index(VAR_261)\n", "VAR_232 = self.settings.register_captcha or self.settings.captcha\n", "self.settings.register_fields.insert(VAR_346 + 1, 'password_two')\n", "if VAR_232:\n", "FUNC_5(VAR_7, VAR_232.label, VAR_232, VAR_232.comment, self.settings.\n formstyle, 'captcha__row')\n", "if self.settings.pre_registration_div:\n", "FUNC_5(VAR_7, '', DIV(*self.settings.pre_registration_div, VAR_16='pre-reg'\n ), '', VAR_273, '')\n", "VAR_199 = web2py_uuid()\n", "if self.settings.registration_requires_approval:\n", "VAR_199 = 'pending-' + VAR_199\n", "VAR_254.registration_key.default = VAR_199\n", "if VAR_7.accepts(VAR_56, VAR_139 if self.csrf_prevention else None, VAR_171\n", "VAR_372 = self.messages.group_description % VAR_7.vars\n", "return VAR_7\n", "if self.settings.create_user_groups:\n", "VAR_147 = self.add_group(self.settings.create_user_groups % VAR_7.vars, VAR_372\n )\n", "if self.settings.everybody_group_id:\n", "self.add_membership(VAR_147, VAR_7.vars.id)\n", "self.add_membership(self.settings.everybody_group_id, VAR_7.vars.id)\n", "if self.settings.registration_requires_verification:\n", "VAR_276 = self.url(self.settings.function, VAR_11=('verify_email', key),\n VAR_97=True)\n", "if self.settings.registration_requires_approval and not self.settings.registration_requires_verification:\n", "VAR_27 = dict(VAR_7.vars)\n", "VAR_254[VAR_7.vars.id] = dict(registration_key='pending')\n", "if not self.settings.registration_requires_verification or self.settings.login_after_registration:\n", "VAR_27.update(dict(VAR_199=key, VAR_276=link, VAR_124=form.vars[username],\n firstname=form.vars['firstname'], lastname=form.vars['lastname']))\n", "VAR_139.flash = self.messages.registration_pending\n", "if not self.settings.registration_requires_verification:\n", "self.log_event(VAR_136, VAR_7.vars)\n", "if not (self.settings.mailer and self.settings.mailer.send(VAR_35=form.vars\n", "VAR_254[VAR_7.vars.id] = dict(registration_key='')\n", "VAR_139.flash = self.messages.registration_successful\n", "VAR_26(VAR_135, VAR_7)\n", "self.db.rollback()\n", "VAR_139.flash = self.messages.email_sent\n", "VAR_141 = VAR_254(**{username: form.vars[username]})\n", "if not VAR_112:\n", "VAR_244.flash = self.messages.unable_send_email\n", "self.login_user(VAR_141)\n", "VAR_112 = self.url(VAR_11=request.args)\n", "VAR_112 = FUNC_4(VAR_112, VAR_7)\n", "return VAR_7\n", "VAR_139.flash = self.messages.logged_in\n", "redirect(VAR_112, client_side=self.settings.client_side)\n" ]
[ "def register(self, next=DEFAULT, onvalidation=DEFAULT, onaccept=DEFAULT,...\n", "\"\"\"docstring\"\"\"\n", "table_user = self.table_user()\n", "request = current.request\n", "response = current.response\n", "session = current.session\n", "if self.is_logged_in():\n", "redirect(self.settings.logged_url, client_side=self.settings.client_side)\n", "if next is DEFAULT:\n", "next = self.get_vars_next() or self.settings.register_next\n", "if onvalidation is DEFAULT:\n", "onvalidation = self.settings.register_onvalidation\n", "if onaccept is DEFAULT:\n", "onaccept = self.settings.register_onaccept\n", "if log is DEFAULT:\n", "log = self.messages['register_log']\n", "table_user = self.table_user()\n", "if self.settings.login_userfield:\n", "username = self.settings.login_userfield\n", "if 'username' in table_user.fields:\n", "unique_validator = IS_NOT_IN_DB(self.db, table_user[username])\n", "username = 'username'\n", "username = 'email'\n", "if not table_user[username].requires:\n", "table_user[username].requires = unique_validator\n", "if isinstance(table_user[username].requires, (list, tuple)):\n", "passfield = self.settings.password_field\n", "if not any([isinstance(validator, IS_NOT_IN_DB) for validator in table_user\n", "if not isinstance(table_user[username].requires, IS_NOT_IN_DB):\n", "formstyle = self.settings.formstyle\n", "if isinstance(table_user[username].requires, list):\n", "table_user[username].requires = [table_user[username].requires,\n unique_validator]\n", "table_user[passfield].requires[-1\n ].min_length = self.settings.password_min_length\n", "if self.settings.register_verify_password:\n", "table_user[username].requires.append(unique_validator)\n", "table_user[username].requires += unique_validator,\n", "if self.settings.register_fields is None:\n", "extra_fields = []\n", "self.settings.register_fields = [f.name for f in table_user if f.writable and\n not f.compute]\n", "extra_fields = [Field('password_two', 'password', requires=IS_EQUAL_TO(\n request.post_vars.get(passfield, None), error_message=self.messages.\n mismatched_password), label=current.T('Confirm Password'))]\n", "form = SQLFORM(table_user, fields=self.settings.register_fields, hidden=\n dict(_next=next), showid=self.settings.showid, submit_button=self.\n messages.register_button, delete_label=self.messages.delete_label,\n formstyle=formstyle, separator=self.settings.label_separator,\n extra_fields=extra_fields)\n", "k = self.settings.register_fields.index(passfield)\n", "captcha = self.settings.register_captcha or self.settings.captcha\n", "self.settings.register_fields.insert(k + 1, 'password_two')\n", "if captcha:\n", "addrow(form, captcha.label, captcha, captcha.comment, self.settings.\n formstyle, 'captcha__row')\n", "if self.settings.pre_registration_div:\n", "addrow(form, '', DIV(*self.settings.pre_registration_div, _id='pre-reg'),\n '', formstyle, '')\n", "key = web2py_uuid()\n", "if self.settings.registration_requires_approval:\n", "key = 'pending-' + key\n", "table_user.registration_key.default = key\n", "if form.accepts(request, session if self.csrf_prevention else None,\n", "description = self.messages.group_description % form.vars\n", "return form\n", "if self.settings.create_user_groups:\n", "group_id = self.add_group(self.settings.create_user_groups % form.vars,\n description)\n", "if self.settings.everybody_group_id:\n", "self.add_membership(group_id, form.vars.id)\n", "self.add_membership(self.settings.everybody_group_id, form.vars.id)\n", "if self.settings.registration_requires_verification:\n", "link = self.url(self.settings.function, args=('verify_email', key), scheme=True\n )\n", "if self.settings.registration_requires_approval and not self.settings.registration_requires_verification:\n", "d = dict(form.vars)\n", "table_user[form.vars.id] = dict(registration_key='pending')\n", "if not self.settings.registration_requires_verification or self.settings.login_after_registration:\n", "d.update(dict(key=key, link=link, username=form.vars[username], firstname=\n form.vars['firstname'], lastname=form.vars['lastname']))\n", "session.flash = self.messages.registration_pending\n", "if not self.settings.registration_requires_verification:\n", "self.log_event(log, form.vars)\n", "if not (self.settings.mailer and self.settings.mailer.send(to=form.vars.\n", "table_user[form.vars.id] = dict(registration_key='')\n", "session.flash = self.messages.registration_successful\n", "callback(onaccept, form)\n", "self.db.rollback()\n", "session.flash = self.messages.email_sent\n", "user = table_user(**{username: form.vars[username]})\n", "if not next:\n", "response.flash = self.messages.unable_send_email\n", "self.login_user(user)\n", "next = self.url(args=request.args)\n", "next = replace_id(next, form)\n", "return form\n", "session.flash = self.messages.logged_in\n", "redirect(next, client_side=self.settings.client_side)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Expr'", "Condition", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "For", "Condition", "Assign'", "Condition", "Assign'", "Assign'", "Condition", "Expr'", "AugAssign'", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Condition", "Expr'", "Condition", "Expr'", "Assign'", "Condition", "Assign'", "Assign'", "Condition", "Assign'", "Return'", "Condition", "Assign'", "Condition", "Expr'", "Expr'", "Condition", "Assign'", "Condition", "Assign'", "Assign'", "Condition", "Expr'", "Assign'", "Condition", "Expr'", "For", "Assign'", "Assign'", "Expr'", "Expr'", "Assign'", "Assign'", "Condition", "Assign'", "Expr'", "Assign'", "Assign'", "Return'", "Assign'", "Expr'" ]
[ "def FUNC_66(self, VAR_18: str, VAR_17: int=200):...\n", "VAR_68 = '/_matrix/client/r0/directory/room/' + VAR_18\n", "VAR_53 = {'room_id': self.room_id}\n", "VAR_54 = json.dumps(VAR_53)\n", "VAR_22, VAR_23 = self.make_request('PUT', VAR_68, VAR_54, VAR_16=self.\n room_owner_tok)\n", "self.assertEqual(VAR_23.code, VAR_17, VAR_23.result)\n" ]
[ "def _set_alias_via_directory(self, alias: str, expected_code: int=200):...\n", "url = '/_matrix/client/r0/directory/room/' + alias\n", "data = {'room_id': self.room_id}\n", "request_data = json.dumps(data)\n", "request, channel = self.make_request('PUT', url, request_data, access_token\n =self.room_owner_tok)\n", "self.assertEqual(channel.code, expected_code, channel.result)\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'" ]
[ "def FUNC_46(VAR_62, VAR_106=None, VAR_107='creation desc'):...\n", "\"\"\"docstring\"\"\"\n", "VAR_174 = FUNC_83(VAR_62, VAR_106=filters, limit_page_length=1, VAR_107=\n order_by, pluck='name')\n", "if VAR_174:\n", "return FUNC_45(VAR_62, VAR_174[0])\n" ]
[ "def get_last_doc(doctype, filters=None, order_by='creation desc'):...\n", "\"\"\"docstring\"\"\"\n", "d = get_all(doctype, filters=filters, limit_page_length=1, order_by=\n order_by, pluck='name')\n", "if d:\n", "return get_doc(doctype, d[0])\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Condition", "Return'" ]
[ "def FUNC_21(self):...\n", "\"\"\"docstring\"\"\"\n", "self.stop()\n", "self._producer.stopProducing()\n" ]
[ "def stopProducing(self):...\n", "\"\"\"docstring\"\"\"\n", "self.stop()\n", "self._producer.stopProducing()\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Expr'", "Expr'" ]
[ "def __getattr__(self, VAR_4):...\n", "\"\"\"docstring\"\"\"\n", "print(f\"No member '{VAR_4}' contained in InputSource.\")\n", "return ''\n" ]
[ "def __getattr__(self, name):...\n", "\"\"\"docstring\"\"\"\n", "print(f\"No member '{name}' contained in InputSource.\")\n", "return ''\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Expr'", "Return'" ]
[ "def FUNC_25(self, VAR_23):...\n", "VAR_53.session['playlist'] = VAR_23\n" ]
[ "def api_rememberplaylist(self, playlist):...\n", "cherrypy.session['playlist'] = playlist\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Assign'" ]
[ "def FUNC_50(VAR_21, VAR_24, VAR_25):...\n", "if VAR_24:\n", "VAR_24 = int(VAR_24) * 2\n", "if VAR_25:\n", "VAR_21 = VAR_21.filter(db.Books.ratings.any(db.Ratings.rating <= VAR_24))\n", "VAR_25 = int(VAR_25) * 2\n", "return VAR_21\n", "VAR_21 = VAR_21.filter(db.Books.ratings.any(db.Ratings.rating >= VAR_25))\n" ]
[ "def adv_search_ratings(q, rating_high, rating_low):...\n", "if rating_high:\n", "rating_high = int(rating_high) * 2\n", "if rating_low:\n", "q = q.filter(db.Books.ratings.any(db.Ratings.rating <= rating_high))\n", "rating_low = int(rating_low) * 2\n", "return q\n", "q = q.filter(db.Books.ratings.any(db.Ratings.rating >= rating_low))\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Assign'", "Condition", "Assign'", "Assign'", "Return'", "Assign'" ]
[ "\"\"\"string\"\"\"\n", "from binascii import hexlify\n", "from datetime import datetime\n", "from os import urandom\n", "from flask import g, Blueprint, url_for, abort, request\n", "from flask_login import login_user, current_user, login_required\n", "from flask_babel import gettext as _\n", "from . import logger, config, calibre_db, db, helper, ub, lm\n", "from .render_template import render_title_template\n", "from functools import wraps\n", "VAR_0 = logger.create()\n", "def FUNC_0(VAR_1):...\n", "@VAR_1.url_value_preprocessor...\n", "g.auth_token = VAR_7.pop('auth_token')\n", "def FUNC_1(VAR_2):...\n", "VAR_8.blueprint_login_views[VAR_2.name] = None\n", "def FUNC_2():...\n", "if 'auth_token' in g:\n", "return g.get('auth_token')\n", "return None\n" ]
[ "\"\"\"This module is used to control authentication/authorization of Kobo sync requests.\nThis module also includes research notes into the auth protocol used by Kobo devices.\n\nLog-in:\nWhen first booting a Kobo device the user must sign into a Kobo (or affiliate) account.\nUpon successful sign-in, the user is redirected to\n https://auth.kobobooks.com/CrossDomainSignIn?id=<some id>\nwhich serves the following response:\n <script type='text/javascript'>\n location.href='kobo://UserAuthenticated?userId=<redacted>&userKey<redacted>&email=<redacted>&returnUrl=https%3a%2f%2fwww.kobo.com';\n </script>\nAnd triggers the insertion of a userKey into the device's User table.\n\nTogether, the device's DeviceId and UserKey act as an *irrevocable* authentication\ntoken to most (if not all) Kobo APIs. In fact, in most cases only the UserKey is\nrequired to authorize the API call.\n\nChanging Kobo password *does not* invalidate user keys! This is apparently a known\nissue for a few years now https://www.mobileread.com/forums/showpost.php?p=3476851&postcount=13\n(although this poster hypothesised that Kobo could blacklist a DeviceId, many endpoints\nwill still grant access given the userkey.)\n\nOfficial Kobo Store Api authorization:\n* For most of the endpoints we care about (sync, metadata, tags, etc), the userKey is\npassed in the x-kobo-userkey header, and is sufficient to authorize the API call.\n* Some endpoints (e.g: AnnotationService) instead make use of Bearer tokens pass through\nan authorization header. To get a BearerToken, the device makes a POST request to the\nv1/auth/device endpoint with the secret UserKey and the device's DeviceId.\n* The book download endpoint passes an auth token as a URL param instead of a header.\n\nOur implementation:\nWe pretty much ignore all of the above. To authenticate the user, we generate a random\nand unique token that they append to the CalibreWeb Url when setting up the api_store\nsetting on the device.\nThus, every request from the device to the api_store will hit CalibreWeb with the\nauth_token in the url (e.g: https://mylibrary.com/<auth_token>/v1/library/sync).\nIn addition, once authenticated we also set the login cookie on the response that will\nbe sent back for the duration of the session to authorize subsequent API calls (in\nparticular calls to non-Kobo specific endpoints such as the CalibreWeb book download).\n\"\"\"\n", "from binascii import hexlify\n", "from datetime import datetime\n", "from os import urandom\n", "from flask import g, Blueprint, url_for, abort, request\n", "from flask_login import login_user, current_user, login_required\n", "from flask_babel import gettext as _\n", "from . import logger, config, calibre_db, db, helper, ub, lm\n", "from .render_template import render_title_template\n", "from functools import wraps\n", "log = logger.create()\n", "def register_url_value_preprocessor(kobo):...\n", "@kobo.url_value_preprocessor...\n", "g.auth_token = values.pop('auth_token')\n", "def disable_failed_auth_redirect_for_blueprint(bp):...\n", "lm.blueprint_login_views[bp.name] = None\n", "def get_auth_token():...\n", "if 'auth_token' in g:\n", "return g.get('auth_token')\n", "return None\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Expr'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Assign'", "FunctionDef'", "Condition", "Assign'", "FunctionDef'", "Assign'", "FunctionDef'", "Condition", "Return'", "Return'" ]
[ "def FUNC_104():...\n", "from frappe.core.doctype.domain_settings.domain_settings import get_active_domains\n", "return FUNC_104()\n" ]
[ "def get_active_domains():...\n", "from frappe.core.doctype.domain_settings.domain_settings import get_active_domains\n", "return get_active_domains()\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "ImportFrom'", "Return'" ]
[ "@login_required()...\n", "\"\"\"docstring\"\"\"\n", "VAR_154 = FUNC_42(VAR_2, VAR_5)\n", "VAR_114 = FUNC_43(VAR_2)\n", "VAR_115 = {'selected': VAR_114, 'images': VAR_154['image'], 'datasets':\n VAR_154['dataset'], 'projects': VAR_154['project'], 'screens': VAR_154[\n 'screen'], 'plates': VAR_154['plate'], 'acquisitions': VAR_154[\n 'acquisition'], 'wells': VAR_154['well']}\n", "for obs in VAR_154.values():\n", "if len(obs) > 0:\n", "VAR_155 = sum([len(VAR_114[types]) for types in VAR_114])\n", "VAR_5.SERVICE_OPTS.setOmeroGroup(obs[0].getDetails().group.id.val)\n", "if VAR_155 == 0:\n", "VAR_104 = None\n", "if VAR_155 == 1:\n", "for VAR_433 in VAR_114:\n", "if VAR_104 is not None:\n", "if len(VAR_114[VAR_433]) > 0:\n", "if VAR_28 in ('dataset', 'project', 'image', 'screen', 'plate',\n", "VAR_284 = VAR_104.getFilesByObject()\n", "VAR_104 = BaseContainer(VAR_5)\n", "VAR_28 = VAR_433[:-1]\n", "if VAR_28 == 'tagset':\n", "VAR_115['files'] = VAR_284\n", "for VAR_215, VAR_143 in VAR_154.items():\n", "VAR_29 = VAR_114[VAR_433][0]\n", "VAR_28 = 'tag'\n", "VAR_94 = {}\n", "if VAR_2.method == 'POST':\n", "if len(VAR_143) > 0:\n", "if VAR_28 is not None and int(VAR_29) > 0:\n", "VAR_285 = FilesAnnotationForm(VAR_115=initial, VAR_158=request.POST.copy())\n", "VAR_285 = FilesAnnotationForm(VAR_115=initial)\n", "VAR_284 = VAR_104.getFilesByObject(VAR_12=dtype, parent_ids=[o.getId() for\n o in objs])\n", "VAR_94[VAR_345(VAR_28)] = int(VAR_29)\n", "VAR_104 = BaseContainer(VAR_5, **kw)\n", "return handlerInternalError(VAR_2, x)\n", "if VAR_285.is_valid():\n", "VAR_53 = {'form_file': VAR_285}\n", "VAR_284 = VAR_285.cleaned_data['files']\n", "return HttpResponse(VAR_285.errors)\n", "VAR_54 = 'webclient/annotations/files_form.html'\n", "VAR_353 = []\n", "VAR_53['template'] = VAR_54\n", "if VAR_284 is not None and len(VAR_284) > 0:\n", "return VAR_53\n", "VAR_353 = VAR_104.createAnnotationsLinks('file', VAR_284, VAR_154)\n", "VAR_234 = 'annotation_file' in VAR_2.FILES and VAR_2.FILES['annotation_file'\n ] or None\n", "if VAR_234 is not None and VAR_234 != '':\n", "VAR_399 = VAR_104.createFileAnnotations(VAR_234, VAR_154)\n", "return JsonResponse({'fileIds': VAR_353})\n", "VAR_353.append(VAR_399)\n" ]
[ "@login_required()...\n", "\"\"\"docstring\"\"\"\n", "oids = getObjects(request, conn)\n", "selected = getIds(request)\n", "initial = {'selected': selected, 'images': oids['image'], 'datasets': oids[\n 'dataset'], 'projects': oids['project'], 'screens': oids['screen'],\n 'plates': oids['plate'], 'acquisitions': oids['acquisition'], 'wells':\n oids['well']}\n", "for obs in oids.values():\n", "if len(obs) > 0:\n", "obj_count = sum([len(selected[types]) for types in selected])\n", "conn.SERVICE_OPTS.setOmeroGroup(obs[0].getDetails().group.id.val)\n", "if obj_count == 0:\n", "manager = None\n", "if obj_count == 1:\n", "for t in selected:\n", "if manager is not None:\n", "if len(selected[t]) > 0:\n", "if o_type in ('dataset', 'project', 'image', 'screen', 'plate',\n", "files = manager.getFilesByObject()\n", "manager = BaseContainer(conn)\n", "o_type = t[:-1]\n", "if o_type == 'tagset':\n", "initial['files'] = files\n", "for dtype, objs in oids.items():\n", "o_id = selected[t][0]\n", "o_type = 'tag'\n", "kw = {}\n", "if request.method == 'POST':\n", "if len(objs) > 0:\n", "if o_type is not None and int(o_id) > 0:\n", "form_file = FilesAnnotationForm(initial=initial, data=request.POST.copy())\n", "form_file = FilesAnnotationForm(initial=initial)\n", "files = manager.getFilesByObject(parent_type=dtype, parent_ids=[o.getId() for\n o in objs])\n", "kw[str(o_type)] = int(o_id)\n", "manager = BaseContainer(conn, **kw)\n", "return handlerInternalError(request, x)\n", "if form_file.is_valid():\n", "context = {'form_file': form_file}\n", "files = form_file.cleaned_data['files']\n", "return HttpResponse(form_file.errors)\n", "template = 'webclient/annotations/files_form.html'\n", "added_files = []\n", "context['template'] = template\n", "if files is not None and len(files) > 0:\n", "return context\n", "added_files = manager.createAnnotationsLinks('file', files, oids)\n", "fileupload = 'annotation_file' in request.FILES and request.FILES[\n 'annotation_file'] or None\n", "if fileupload is not None and fileupload != '':\n", "newFileId = manager.createFileAnnotations(fileupload, oids)\n", "return JsonResponse({'fileIds': added_files})\n", "added_files.append(newFileId)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Assign'", "Assign'", "For", "Condition", "Assign'", "Expr'", "Condition", "Assign'", "Condition", "For", "Condition", "Condition", "Condition", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "For", "Assign'", "Assign'", "Assign'", "Condition", "Condition", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Return'", "Condition", "Assign'", "Assign'", "Return'", "Assign'", "Assign'", "Assign'", "Condition", "Return'", "Assign'", "Assign'", "Condition", "Assign'", "Return'", "Expr'" ]
[ "def FUNC_35():...\n", "return FUNC_0(VAR_6=request.url_root, VAR_7=additional_unless)\n" ]
[ "def unless():...\n", "return _preemptive_unless(base_url=request.url_root, additional_unless=\n additional_unless)\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "@tenacity.retry(reraise=True, stop=tenacity.stop_after_attempt(3), wait=...\n", "VAR_6 = requests.get(self.sourceName, timeout=10)\n", "if VAR_6.status_code == 404:\n", "VAR_6.raise_for_status()\n", "return VAR_6\n" ]
[ "@tenacity.retry(reraise=True, stop=tenacity.stop_after_attempt(3), wait=...\n", "response = requests.get(self.sourceName, timeout=10)\n", "if response.status_code == 404:\n", "response.raise_for_status()\n", "return response\n" ]
[ 0, 0, 0, 0, 0 ]
[ "Condition", "Assign'", "Condition", "Expr'", "Return'" ]
[ "def FUNC_3(*, VAR_0: Optional[str], VAR_1: Optional[Path]) ->Union[Dict[str,...\n", "VAR_35: bytes\n", "if VAR_0 is not None and VAR_1 is not None:\n", "return GeneratorError(header='Provide URL or Path, not both.')\n", "if VAR_0 is not None:\n", "if VAR_1 is not None:\n", "VAR_34 = httpx.get(VAR_0)\n", "return GeneratorError(header='Could not get OpenAPI document from provided URL'\n )\n", "return yaml.safe_load(VAR_35)\n", "return GeneratorError(header='Invalid YAML from provided source')\n", "VAR_35 = VAR_1.read_bytes()\n", "return GeneratorError(header='No URL or Path provided')\n", "VAR_35 = VAR_34.content\n" ]
[ "def _get_document(*, url: Optional[str], path: Optional[Path]) ->Union[Dict...\n", "yaml_bytes: bytes\n", "if url is not None and path is not None:\n", "return GeneratorError(header='Provide URL or Path, not both.')\n", "if url is not None:\n", "if path is not None:\n", "response = httpx.get(url)\n", "return GeneratorError(header='Could not get OpenAPI document from provided URL'\n )\n", "return yaml.safe_load(yaml_bytes)\n", "return GeneratorError(header='Invalid YAML from provided source')\n", "yaml_bytes = path.read_bytes()\n", "return GeneratorError(header='No URL or Path provided')\n", "yaml_bytes = response.content\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "AnnAssign'", "Condition", "Return'", "Condition", "Condition", "Assign'", "Return'", "Return'", "Return'", "Assign'", "Return'", "Assign'" ]
[ "def FUNC_12(self) ->Databases:...\n", "if not self.datastores:\n", "return self.datastores\n" ]
[ "def get_datastores(self) ->Databases:...\n", "if not self.datastores:\n", "return self.datastores\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Return'" ]
[ "def FUNC_84(*VAR_79, **VAR_42):...\n", "\"\"\"docstring\"\"\"\n", "return VAR_12.get_value(*VAR_79, **kwargs)\n" ]
[ "def get_value(*args, **kwargs):...\n", "\"\"\"docstring\"\"\"\n", "return db.get_value(*args, **kwargs)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Return'" ]
[ "def FUNC_2(self):...\n", "from Products.PageTemplates.Expressions import createZopeEngine\n", "return createZopeEngine()\n" ]
[ "def _makeEngine(self):...\n", "from Products.PageTemplates.Expressions import createZopeEngine\n", "return createZopeEngine()\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "ImportFrom'", "Return'" ]
[ "def FUNC_30(self):...\n", "VAR_22 = models.Band.objects.create(name='Linkin Park')\n", "VAR_22.album_set.create(name='Hybrid Theory', cover_art=\n 'albums\\\\hybrid_theory.jpg')\n", "VAR_29 = models.Album._meta.get_field('band').rel\n", "VAR_27 = widgets.ForeignKeyRawIdWidget(VAR_29, widget_admin_site)\n", "self.assertHTMLEqual(conditional_escape(VAR_27.render('test', VAR_22.pk,\n attrs={})), 'string' % dict(VAR_0(), bandpk=band.pk))\n" ]
[ "def test_render(self):...\n", "band = models.Band.objects.create(name='Linkin Park')\n", "band.album_set.create(name='Hybrid Theory', cover_art=\n 'albums\\\\hybrid_theory.jpg')\n", "rel = models.Album._meta.get_field('band').rel\n", "w = widgets.ForeignKeyRawIdWidget(rel, widget_admin_site)\n", "self.assertHTMLEqual(conditional_escape(w.render('test', band.pk, attrs={})\n ), \n '<input type=\"text\" name=\"test\" value=\"%(bandpk)s\" class=\"vForeignKeyRawIdAdminField\" /><a href=\"/widget_admin/admin_widgets/band/?t=id\" class=\"related-lookup\" id=\"lookup_id_test\" onclick=\"return showRelatedObjectLookupPopup(this);\"> <img src=\"%(ADMIN_STATIC_PREFIX)simg/selector-search.gif\" width=\"16\" height=\"16\" alt=\"Lookup\" /></a>&nbsp;<strong>Linkin Park</strong>'\n % dict(admin_static_prefix(), bandpk=band.pk))\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Expr'", "Assign'", "Assign'", "Expr'" ]
[ "def FUNC_31(VAR_10=None, VAR_62=None):...\n", "\"\"\"docstring\"\"\"\n", "import frappe.cache_manager\n", "if VAR_62:\n", "frappe.cache_manager.clear_doctype_cache(VAR_62)\n", "if VAR_10:\n", "FUNC_38()\n", "frappe.cache_manager.clear_user_cache(VAR_10)\n", "from frappe import translate\n", "VAR_1.role_permissions = {}\n", "frappe.cache_manager.clear_user_cache()\n", "frappe.cache_manager.clear_domain_cache()\n", "translate.clear_cache()\n", "FUNC_38()\n", "VAR_1.cache = {}\n", "VAR_1.new_doc_templates = {}\n", "for VAR_129 in FUNC_66('clear_cache'):\n", "FUNC_72(VAR_129)()\n" ]
[ "def clear_cache(user=None, doctype=None):...\n", "\"\"\"docstring\"\"\"\n", "import frappe.cache_manager\n", "if doctype:\n", "frappe.cache_manager.clear_doctype_cache(doctype)\n", "if user:\n", "reset_metadata_version()\n", "frappe.cache_manager.clear_user_cache(user)\n", "from frappe import translate\n", "local.role_permissions = {}\n", "frappe.cache_manager.clear_user_cache()\n", "frappe.cache_manager.clear_domain_cache()\n", "translate.clear_cache()\n", "reset_metadata_version()\n", "local.cache = {}\n", "local.new_doc_templates = {}\n", "for fn in get_hooks('clear_cache'):\n", "get_attr(fn)()\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Import'", "Condition", "Expr'", "Condition", "Expr'", "Expr'", "ImportFrom'", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'", "Assign'", "Assign'", "For", "Expr'" ]
[ "def FUNC_3(self, VAR_16: Text) ->None:...\n", "\"\"\"docstring\"\"\"\n", "VAR_26 = self.bucket.blob(VAR_16)\n", "VAR_26.download_to_filename(VAR_16)\n" ]
[ "def _retrieve_tar(self, target_filename: Text) ->None:...\n", "\"\"\"docstring\"\"\"\n", "blob = self.bucket.blob(target_filename)\n", "blob.download_to_filename(target_filename)\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Expr'" ]
[ "def FUNC_10(self, *VAR_0, **VAR_1):...\n", "\"\"\"docstring\"\"\"\n", "return self._save(*VAR_0, **kwargs)\n" ]
[ "def save(self, *args, **kwargs):...\n", "\"\"\"docstring\"\"\"\n", "return self._save(*args, **kwargs)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Return'" ]
[ "def FUNC_9(self, VAR_20: Text) ->None:...\n", "from azure.core.exceptions import ResourceExistsError\n", "self.blob_service.create_container(VAR_20)\n" ]
[ "def _ensure_container_exists(self, container_name: Text) ->None:...\n", "from azure.core.exceptions import ResourceExistsError\n", "self.blob_service.create_container(container_name)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "ImportFrom'", "Expr'" ]
[ "def FUNC_12(VAR_21):...\n", "\"\"\"docstring\"\"\"\n", "VAR_21 = force_bytes(VAR_21)\n", "return base64.urlsafe_b64decode(VAR_21.ljust(len(VAR_21) + len(VAR_21) % 4,\n b'='))\n" ]
[ "def urlsafe_base64_decode(s):...\n", "\"\"\"docstring\"\"\"\n", "s = force_bytes(s)\n", "return base64.urlsafe_b64decode(s.ljust(len(s) + len(s) % 4, b'='))\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Return'" ]
[ "def FUNC_49(self):...\n", "\"\"\"docstring\"\"\"\n", "VAR_55 = self._send_labelled_messages_in_room()\n", "VAR_22, VAR_23 = self.make_request('GET', '/rooms/%s/context/%s?filter=%s' %\n (self.room_id, VAR_55, json.dumps(self.FILTER_NOT_LABELS)), VAR_16=self.tok\n )\n", "self.assertEqual(VAR_23.code, 200, VAR_23.result)\n", "VAR_58 = VAR_23.json_body['events_before']\n", "self.assertEqual(len(VAR_58), 1, [event['content'] for event in VAR_58])\n", "self.assertEqual(VAR_58[0]['content']['body'], 'without label', VAR_58[0])\n", "VAR_59 = VAR_23.json_body['events_after']\n", "self.assertEqual(len(VAR_59), 2, [event['content'] for event in VAR_59])\n", "self.assertEqual(VAR_59[0]['content']['body'], 'with wrong label', VAR_59[0])\n", "self.assertEqual(VAR_59[1]['content']['body'], 'with two wrong labels',\n VAR_59[1])\n" ]
[ "def test_context_filter_not_labels(self):...\n", "\"\"\"docstring\"\"\"\n", "event_id = self._send_labelled_messages_in_room()\n", "request, channel = self.make_request('GET', \n '/rooms/%s/context/%s?filter=%s' % (self.room_id, event_id, json.dumps(\n self.FILTER_NOT_LABELS)), access_token=self.tok)\n", "self.assertEqual(channel.code, 200, channel.result)\n", "events_before = channel.json_body['events_before']\n", "self.assertEqual(len(events_before), 1, [event['content'] for event in\n events_before])\n", "self.assertEqual(events_before[0]['content']['body'], 'without label',\n events_before[0])\n", "events_after = channel.json_body['events_after']\n", "self.assertEqual(len(events_after), 2, [event['content'] for event in\n events_after])\n", "self.assertEqual(events_after[0]['content']['body'], 'with wrong label',\n events_after[0])\n", "self.assertEqual(events_after[1]['content']['body'],\n 'with two wrong labels', events_after[1])\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Expr'", "Assign'", "Expr'", "Expr'", "Assign'", "Expr'", "Expr'", "Expr'" ]