lines
sequencelengths
1
383
raw_lines
sequencelengths
1
383
label
sequencelengths
1
383
type
sequencelengths
1
383
[ "def FUNC_173(VAR_445):...\n", "return A(VAR_445, _href=URL(args='_search', vars=dict(q=t)))\n" ]
[ "def link(t):...\n", "return A(t, _href=URL(args='_search', vars=dict(q=t)))\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_16(self, *VAR_40, **VAR_7):...\n", "if not self.date:\n", "self.date = timezone.now()\n", "return super(CLASS_16, self).save(*VAR_40, **kwargs)\n" ]
[ "def save(self, *args, **kwargs):...\n", "if not self.date:\n", "self.date = timezone.now()\n", "return super(IgnoreEmail, self).save(*args, **kwargs)\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Assign'", "Return'" ]
[ "def FUNC_4(self):...\n", "\"\"\"docstring\"\"\"\n", "return reverse(settings.LOGIN_VIEW)\n" ]
[ "def get_login_url(self):...\n", "\"\"\"docstring\"\"\"\n", "return reverse(settings.LOGIN_VIEW)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Return'" ]
[ "def FUNC_0(self, VAR_2, VAR_3):...\n", "VAR_4 = self.setup_test_homeserver('server', http_client=None)\n", "return VAR_4\n" ]
[ "def make_homeserver(self, reactor, clock):...\n", "hs = self.setup_test_homeserver('server', http_client=None)\n", "return hs\n" ]
[ 0, 4, 0 ]
[ "FunctionDef'", "Assign'", "Return'" ]
[ "def FUNC_25(self, VAR_19, VAR_38):...\n", "if VAR_19 in ('first', 'last'):\n", "VAR_46 = getattr(self, VAR_19)\n", "return getattr(self, VAR_19)\n", "VAR_19 = VAR_38[:]\n", "if not VAR_19:\n", "VAR_19 = None\n", "VAR_38[:] = []\n", "return VAR_46(VAR_19)\n" ]
[ "def traverse(self, name, furtherPath):...\n", "if name in ('first', 'last'):\n", "method = getattr(self, name)\n", "return getattr(self, name)\n", "name = furtherPath[:]\n", "if not name:\n", "name = None\n", "furtherPath[:] = []\n", "return method(name)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Assign'", "Return'", "Assign'", "Condition", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_8(VAR_0):...\n", "from openapi_python_client import GeneratorData, Project\n", "from openapi_python_client.parser.openapi import EndpointCollection, Schemas\n", "VAR_3 = VAR_0.MagicMock(autospec=GeneratorData, title='My Test API',\n endpoint_collections_by_tag={'default': mocker.MagicMock(autospec=\n EndpointCollection, parse_errors=[1]), 'other': mocker.MagicMock(\n autospec=EndpointCollection, parse_errors=[2])}, schemas=mocker.\n MagicMock(autospec=Schemas, errors=[3]))\n", "VAR_8 = Project(VAR_3=openapi)\n", "assert VAR_8._get_errors() == [1, 2, 3]\n" ]
[ "def test__get_errors(mocker):...\n", "from openapi_python_client import GeneratorData, Project\n", "from openapi_python_client.parser.openapi import EndpointCollection, Schemas\n", "openapi = mocker.MagicMock(autospec=GeneratorData, title='My Test API',\n endpoint_collections_by_tag={'default': mocker.MagicMock(autospec=\n EndpointCollection, parse_errors=[1]), 'other': mocker.MagicMock(\n autospec=EndpointCollection, parse_errors=[2])}, schemas=mocker.\n MagicMock(autospec=Schemas, errors=[3]))\n", "project = Project(openapi=openapi)\n", "assert project._get_errors() == [1, 2, 3]\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "ImportFrom'", "ImportFrom'", "Assign'", "Assign'", "Assert'" ]
[ "@defer.inlineCallbacks...\n", "\"\"\"docstring\"\"\"\n", "VAR_12 = urllib.parse.urlparse(VAR_8)\n", "assert VAR_12.hostname\n", "VAR_15 = None\n", "if VAR_12.scheme == b'matrix' and not FUNC_0(VAR_12.hostname\n", "VAR_20 = yield defer.ensureDeferred(self._well_known_resolver.\n get_well_known(VAR_12.hostname))\n", "if VAR_15:\n", "VAR_15 = VAR_20.delegated_server\n", "VAR_8 = urllib.parse.urlunparse((VAR_12.scheme, VAR_15, VAR_12.path, VAR_12\n .params, VAR_12.query, VAR_12.fragment))\n", "if VAR_9 is None:\n", "VAR_12 = urllib.parse.urlparse(VAR_8)\n", "VAR_9 = Headers()\n", "VAR_9 = VAR_9.copy()\n", "if not VAR_9.hasHeader(b'host'):\n", "VAR_9.addRawHeader(b'host', VAR_12.netloc)\n", "if not VAR_9.hasHeader(b'user-agent'):\n", "VAR_9.addRawHeader(b'user-agent', self.user_agent)\n", "VAR_16 = yield make_deferred_yieldable(self._agent.request(VAR_7, VAR_8,\n VAR_9, VAR_10))\n", "return VAR_16\n" ]
[ "@defer.inlineCallbacks...\n", "\"\"\"docstring\"\"\"\n", "parsed_uri = urllib.parse.urlparse(uri)\n", "assert parsed_uri.hostname\n", "delegated_server = None\n", "if parsed_uri.scheme == b'matrix' and not _is_ip_literal(parsed_uri.hostname\n", "well_known_result = yield defer.ensureDeferred(self._well_known_resolver.\n get_well_known(parsed_uri.hostname))\n", "if delegated_server:\n", "delegated_server = well_known_result.delegated_server\n", "uri = urllib.parse.urlunparse((parsed_uri.scheme, delegated_server,\n parsed_uri.path, parsed_uri.params, parsed_uri.query, parsed_uri.fragment))\n", "if headers is None:\n", "parsed_uri = urllib.parse.urlparse(uri)\n", "headers = Headers()\n", "headers = headers.copy()\n", "if not headers.hasHeader(b'host'):\n", "headers.addRawHeader(b'host', parsed_uri.netloc)\n", "if not headers.hasHeader(b'user-agent'):\n", "headers.addRawHeader(b'user-agent', self.user_agent)\n", "res = yield make_deferred_yieldable(self._agent.request(method, uri,\n headers, bodyProducer))\n", "return res\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Assert'", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Condition", "Expr'", "Condition", "Expr'", "Assign'", "Return'" ]
[ "@ensure_csrf_cookie...\n", "\"\"\"docstring\"\"\"\n", "VAR_10 = SlashSeparatedCourseKey.from_deprecated_string(VAR_10)\n", "VAR_135 = ReportStore.from_config(config_name='GRADES_DOWNLOAD')\n", "VAR_63 = {'downloads': [dict(VAR_16=name, VAR_147=url, link=HTML(\n '<a href=\"{}\">{}</a>').format(HTML(url), Text(name))) for VAR_16,\n VAR_147 in VAR_135.links_for(VAR_10)]}\n", "return JsonResponse(VAR_63)\n" ]
[ "@ensure_csrf_cookie...\n", "\"\"\"docstring\"\"\"\n", "course_id = SlashSeparatedCourseKey.from_deprecated_string(course_id)\n", "report_store = ReportStore.from_config(config_name='GRADES_DOWNLOAD')\n", "response_payload = {'downloads': [dict(name=name, url=url, link=HTML(\n '<a href=\"{}\">{}</a>').format(HTML(url), Text(name))) for name, url in\n report_store.links_for(course_id)]}\n", "return JsonResponse(response_payload)\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_6():...\n", "from frappe.database import get_db\n", "VAR_10 = VAR_1.conf.db_name\n", "VAR_151 = VAR_1.conf.db_password\n", "if VAR_1.conf.different_credentials_for_replica:\n", "VAR_10 = VAR_1.conf.replica_db_name\n", "VAR_1.replica_db = get_db(host=local.conf.replica_host, VAR_10=user,\n VAR_151=password)\n", "VAR_151 = VAR_1.conf.replica_db_password\n", "VAR_1.primary_db = VAR_1.db\n", "VAR_1.db = VAR_1.replica_db\n" ]
[ "def connect_replica():...\n", "from frappe.database import get_db\n", "user = local.conf.db_name\n", "password = local.conf.db_password\n", "if local.conf.different_credentials_for_replica:\n", "user = local.conf.replica_db_name\n", "local.replica_db = get_db(host=local.conf.replica_host, user=user, password\n =password)\n", "password = local.conf.replica_db_password\n", "local.primary_db = local.db\n", "local.db = local.replica_db\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "ImportFrom'", "Assign'", "Assign'", "For", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'" ]
[ "async def FUNC_5(self):...\n", "if hasattr(self, 'resolved_ref'):\n", "return self.resolved_ref\n", "self.sha1_validate(self.unresolved_ref)\n", "VAR_64 = ['git', 'ls-remote', self.repo, self.unresolved_ref]\n", "self.resolved_ref = self.unresolved_ref\n", "return self.resolved_ref\n", "VAR_65 = subprocess.run(VAR_64, universal_newlines=True, stdout=subprocess.\n PIPE, stderr=subprocess.PIPE)\n", "if VAR_65.returncode:\n", "if not VAR_65.stdout:\n", "return None\n", "VAR_66 = VAR_65.stdout.split(None, 1)[0]\n", "self.sha1_validate(VAR_66)\n", "self.resolved_ref = VAR_66\n" ]
[ "async def get_resolved_ref(self):...\n", "if hasattr(self, 'resolved_ref'):\n", "return self.resolved_ref\n", "self.sha1_validate(self.unresolved_ref)\n", "command = ['git', 'ls-remote', self.repo, self.unresolved_ref]\n", "self.resolved_ref = self.unresolved_ref\n", "return self.resolved_ref\n", "result = subprocess.run(command, universal_newlines=True, stdout=subprocess\n .PIPE, stderr=subprocess.PIPE)\n", "if result.returncode:\n", "if not result.stdout:\n", "return None\n", "resolved_ref = result.stdout.split(None, 1)[0]\n", "self.sha1_validate(resolved_ref)\n", "self.resolved_ref = resolved_ref\n" ]
[ 0, 0, 0, 0, 5, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "AsyncFunctionDef'", "Condition", "Return'", "Expr'", "Assign'", "Assign'", "Return'", "Assign'", "Condition", "Condition", "Return'", "Assign'", "Expr'", "Assign'" ]
[ "@parameterized.named_parameters(('VariablesToFeedNone', '', 'func2', None),...\n", "if not test.is_built_with_xla():\n", "self.skipTest('Skipping test because XLA is not compiled in.')\n", "VAR_13 = os.path.join(test.get_temp_dir(), 'dummy_model')\n", "VAR_14 = self.AOTCompileDummyModel()\n", "VAR_6 = getattr(VAR_14, VAR_6)\n", "self.evaluate(VAR_14.var.initializer)\n", "self.evaluate(VAR_14.write_var.initializer)\n", "save.save(VAR_14, VAR_13, signatures={'func': func})\n", "self.parser = saved_model_cli.create_parser()\n", "VAR_50 = os.path.join(test.get_temp_dir(), 'aot_compile_cpu_dir/out')\n", "VAR_10 = ['aot_compile_cpu', '--dir', VAR_13, '--tag_set', 'serve',\n '--signature_def_key', 'func', '--output_prefix', VAR_50,\n '--variables_to_feed', VAR_5, '--cpp_class', 'Generated']\n", "if VAR_7:\n", "VAR_10.extend(['--target_triple', VAR_7])\n", "VAR_10 = self.parser.parse_args(VAR_10)\n", "saved_model_cli.aot_compile_cpu(VAR_10)\n", "self.assertRegex(str(captured_warn.call_args),\n \"Signature input key 'y'.*has been pruned while freezing the graph.\")\n", "self.assertTrue(file_io.file_exists('{}.o'.format(VAR_50)))\n", "self.assertTrue(file_io.file_exists('{}.h'.format(VAR_50)))\n", "self.assertTrue(file_io.file_exists('{}_metadata.o'.format(VAR_50)))\n", "self.assertTrue(file_io.file_exists('{}_makefile.inc'.format(VAR_50)))\n", "VAR_51 = file_io.read_file_to_string('{}.h'.format(VAR_50))\n", "self.assertIn('class Generated', VAR_51)\n", "self.assertIn('arg_feed_x_data', VAR_51)\n", "self.assertIn('result_fetch_res_data', VAR_51)\n", "self.assertNotIn('arg_feed_y_data', VAR_51)\n", "if VAR_5:\n", "self.assertIn('set_var_param_my_var_data(const float', VAR_51)\n", "if VAR_6 == VAR_14.func_write:\n", "self.assertNotIn('set_var_param_my_var_data(float', VAR_51)\n", "self.assertIn('set_var_param_write_var_data(float', VAR_51)\n", "VAR_52 = file_io.read_file_to_string('{}_makefile.inc'.format(VAR_50))\n", "self.assertNotIn('set_var_param_write_var_data(const float', VAR_51)\n", "self.assertIn('-D_GLIBCXX_USE_CXX11_ABI=', VAR_52)\n" ]
[ "@parameterized.named_parameters(('VariablesToFeedNone', '', 'func2', None),...\n", "if not test.is_built_with_xla():\n", "self.skipTest('Skipping test because XLA is not compiled in.')\n", "saved_model_dir = os.path.join(test.get_temp_dir(), 'dummy_model')\n", "dummy_model = self.AOTCompileDummyModel()\n", "func = getattr(dummy_model, func)\n", "self.evaluate(dummy_model.var.initializer)\n", "self.evaluate(dummy_model.write_var.initializer)\n", "save.save(dummy_model, saved_model_dir, signatures={'func': func})\n", "self.parser = saved_model_cli.create_parser()\n", "output_prefix = os.path.join(test.get_temp_dir(), 'aot_compile_cpu_dir/out')\n", "args = ['aot_compile_cpu', '--dir', saved_model_dir, '--tag_set', 'serve',\n '--signature_def_key', 'func', '--output_prefix', output_prefix,\n '--variables_to_feed', variables_to_feed, '--cpp_class', 'Generated']\n", "if target_triple:\n", "args.extend(['--target_triple', target_triple])\n", "args = self.parser.parse_args(args)\n", "saved_model_cli.aot_compile_cpu(args)\n", "self.assertRegex(str(captured_warn.call_args),\n \"Signature input key 'y'.*has been pruned while freezing the graph.\")\n", "self.assertTrue(file_io.file_exists('{}.o'.format(output_prefix)))\n", "self.assertTrue(file_io.file_exists('{}.h'.format(output_prefix)))\n", "self.assertTrue(file_io.file_exists('{}_metadata.o'.format(output_prefix)))\n", "self.assertTrue(file_io.file_exists('{}_makefile.inc'.format(output_prefix)))\n", "header_contents = file_io.read_file_to_string('{}.h'.format(output_prefix))\n", "self.assertIn('class Generated', header_contents)\n", "self.assertIn('arg_feed_x_data', header_contents)\n", "self.assertIn('result_fetch_res_data', header_contents)\n", "self.assertNotIn('arg_feed_y_data', header_contents)\n", "if variables_to_feed:\n", "self.assertIn('set_var_param_my_var_data(const float', header_contents)\n", "if func == dummy_model.func_write:\n", "self.assertNotIn('set_var_param_my_var_data(float', header_contents)\n", "self.assertIn('set_var_param_write_var_data(float', header_contents)\n", "makefile_contents = file_io.read_file_to_string('{}_makefile.inc'.format(\n output_prefix))\n", "self.assertNotIn('set_var_param_write_var_data(const float', header_contents)\n", "self.assertIn('-D_GLIBCXX_USE_CXX11_ABI=', makefile_contents)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Condition", "Expr'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'", "Assign'", "Assign'", "Assign'", "Condition", "Expr'", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'", "Condition", "Expr'", "Condition", "Expr'", "Expr'", "Assign'", "Expr'", "Expr'" ]
[ "def FUNC_64(self, VAR_2):...\n", "" ]
[ "def write(self, data):...\n", "" ]
[ 0, 0 ]
[ "FunctionDef'", "Condition" ]
[ "async def FUNC_6(self):...\n", "return 'fake/repo/1a2b3c4d5e6f'\n" ]
[ "async def get_resolved_spec(self):...\n", "return 'fake/repo/1a2b3c4d5e6f'\n" ]
[ 0, 0 ]
[ "AsyncFunctionDef'", "Return'" ]
[ "def FUNC_19(self, VAR_2):...\n", "def FUNC_30(self, VAR_8):...\n", "VAR_2.add_url_rule('/sites/<int:id>', endpoint='site_resource.get')\n", "url_for('delete', VAR_8=1, _cls=SiteResource)\n" ]
[ "def test_it_falls_through_if_class_endpoint_not_found(self, app):...\n", "def get(self, id):...\n", "app.add_url_rule('/sites/<int:id>', endpoint='site_resource.get')\n", "url_for('delete', id=1, _cls=SiteResource)\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "FunctionDef'", "Expr'", "Expr'" ]
[ "def FUNC_15(self, VAR_0, VAR_8, VAR_5, VAR_7):...\n", "VAR_0.get(VAR_8 + self.url)\n", "VAR_16 = VAR_0.find_element(By.XPATH, \"//input[@name='file']\")\n", "VAR_16.send_keys(VAR_5)\n", "assert VAR_16.get_attribute('name') == 'file'\n", "VAR_17 = VAR_0.find_element(By.XPATH, \"//input[@name='save']\")\n", "VAR_17.click()\n", "assert 'save' in VAR_0.page_source\n", "VAR_0.get(VAR_8 + self.url)\n", "VAR_16 = VAR_0.find_element(By.XPATH, \"//input[@name='file']\")\n", "VAR_16.send_keys(VAR_5)\n", "assert VAR_16.get_attribute('name') == 'file'\n", "VAR_17 = VAR_0.find_element(By.XPATH, \"//button[@name='save_continue']\")\n", "VAR_17.click()\n", "VAR_12 = json.loads(VAR_0.find_elements(By.CSS_SELECTOR, 'pre')[0].text)\n", "assert VAR_12['POST']['progress'] == '1'\n" ]
[ "def test_progress(self, driver, live_server, upload_file, freeze):...\n", "driver.get(live_server + self.url)\n", "file_input = driver.find_element(By.XPATH, \"//input[@name='file']\")\n", "file_input.send_keys(upload_file)\n", "assert file_input.get_attribute('name') == 'file'\n", "save_button = driver.find_element(By.XPATH, \"//input[@name='save']\")\n", "save_button.click()\n", "assert 'save' in driver.page_source\n", "driver.get(live_server + self.url)\n", "file_input = driver.find_element(By.XPATH, \"//input[@name='file']\")\n", "file_input.send_keys(upload_file)\n", "assert file_input.get_attribute('name') == 'file'\n", "save_button = driver.find_element(By.XPATH, \"//button[@name='save_continue']\")\n", "save_button.click()\n", "response = json.loads(driver.find_elements(By.CSS_SELECTOR, 'pre')[0].text)\n", "assert response['POST']['progress'] == '1'\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Assign'", "Expr'", "Assert'", "Assign'", "Expr'", "Assert'", "Expr'", "Assign'", "Expr'", "Assert'", "Assign'", "Expr'", "Assign'", "Assert'" ]
[ "@VAR_0.simple_tag...\n", "\"\"\"docstring\"\"\"\n", "VAR_52 = reverse('rest_framework:login')\n", "return 'log in'\n", "VAR_41 = \"<a href='{href}?next={next}'>log in</a>\"\n", "VAR_41 = format_html(VAR_41, href=login_url, next=escape(request.path))\n", "return mark_safe(VAR_41)\n" ]
[ "@register.simple_tag...\n", "\"\"\"docstring\"\"\"\n", "login_url = reverse('rest_framework:login')\n", "return 'log in'\n", "snippet = \"<a href='{href}?next={next}'>log in</a>\"\n", "snippet = format_html(snippet, href=login_url, next=escape(request.path))\n", "return mark_safe(snippet)\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Return'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_155():...\n", "def FUNC_172(VAR_354):...\n", "return VAR_354 + ' ' + VAR_354.replace('icon', 'glyphicon')\n" ]
[ "def bootstrap3():...\n", "def rename(icon):...\n", "return icon + ' ' + icon.replace('icon', 'glyphicon')\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "FunctionDef'", "Return'" ]
[ "def FUNC_15(self):...\n", "return getattr(self, '_doc_before_save', None)\n" ]
[ "def get_doc_before_save(self):...\n", "return getattr(self, '_doc_before_save', None)\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_13(self):...\n", "\"\"\"docstring\"\"\"\n", "for VAR_19 in self.meta.get_table_fields():\n", "self.update_child_table(VAR_19.fieldname, VAR_19)\n" ]
[ "def update_children(self):...\n", "\"\"\"docstring\"\"\"\n", "for df in self.meta.get_table_fields():\n", "self.update_child_table(df.fieldname, df)\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "For", "Expr'" ]
[ "def FUNC_20(self, VAR_19):...\n", "if os.path.exists(VAR_19):\n", "return f.read()\n" ]
[ "def albumartcache_load(self, imgb64path):...\n", "if os.path.exists(imgb64path):\n", "return f.read()\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Return'" ]
[ "def FUNC_0():...\n", "\"\"\"docstring\"\"\"\n", "if os.path.isfile('index.bs'):\n", "return 'index.bs'\n", "if os.path.isfile('Overview.bs'):\n", "return 'Overview.bs'\n", "VAR_20 = glob.glob('*.bs')\n", "if VAR_20:\n", "return VAR_20[0]\n", "VAR_21 = glob.glob('*.src.html')\n", "if VAR_21:\n", "return VAR_21[0]\n", "return None\n" ]
[ "def findImplicitInputFile():...\n", "\"\"\"docstring\"\"\"\n", "if os.path.isfile('index.bs'):\n", "return 'index.bs'\n", "if os.path.isfile('Overview.bs'):\n", "return 'Overview.bs'\n", "allBs = glob.glob('*.bs')\n", "if allBs:\n", "return allBs[0]\n", "allHtml = glob.glob('*.src.html')\n", "if allHtml:\n", "return allHtml[0]\n", "return None\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Return'", "Condition", "Return'", "Assign'", "Condition", "Return'", "Assign'", "Condition", "Return'", "Return'" ]
[ "def FUNC_6(self):...\n", "VAR_7 = self.helper.create_room_as(self.u_alice, tok=self.t_alice)\n", "VAR_8 = self.get_success(event_injection.inject_member_event(self.hs, VAR_7,\n self.u_bob, Membership.JOIN))\n", "VAR_9, VAR_10 = self.get_success(event_injection.create_event(self.hs,\n room_id=room, sender=self.u_alice, prev_event_ids=[bob_event.event_id],\n type='m.test.1', content={}))\n", "VAR_11 = self.get_success(self.store.get_joined_users_from_context(VAR_9,\n VAR_10))\n", "self.assertEqual(VAR_11.keys(), {self.u_alice, self.u_bob})\n", "VAR_12 = self.get_success(event_injection.inject_event(self.hs, room_id=\n room, sender=self.u_bob, prev_event_ids=[bob_event.event_id], type=\n 'm.test.2', state_key=self.u_bob, content={}))\n", "VAR_9, VAR_10 = self.get_success(event_injection.create_event(self.hs,\n room_id=room, sender=self.u_alice, prev_event_ids=[non_member_event.\n event_id], type='m.test.3', content={}))\n", "VAR_11 = self.get_success(self.store.get_joined_users_from_context(VAR_9,\n VAR_10))\n", "self.assertEqual(VAR_11.keys(), {self.u_alice, self.u_bob})\n" ]
[ "def test_get_joined_users_from_context(self):...\n", "room = self.helper.create_room_as(self.u_alice, tok=self.t_alice)\n", "bob_event = self.get_success(event_injection.inject_member_event(self.hs,\n room, self.u_bob, Membership.JOIN))\n", "event, context = self.get_success(event_injection.create_event(self.hs,\n room_id=room, sender=self.u_alice, prev_event_ids=[bob_event.event_id],\n type='m.test.1', content={}))\n", "users = self.get_success(self.store.get_joined_users_from_context(event,\n context))\n", "self.assertEqual(users.keys(), {self.u_alice, self.u_bob})\n", "non_member_event = self.get_success(event_injection.inject_event(self.hs,\n room_id=room, sender=self.u_bob, prev_event_ids=[bob_event.event_id],\n type='m.test.2', state_key=self.u_bob, content={}))\n", "event, context = self.get_success(event_injection.create_event(self.hs,\n room_id=room, sender=self.u_alice, prev_event_ids=[non_member_event.\n event_id], type='m.test.3', content={}))\n", "users = self.get_success(self.store.get_joined_users_from_context(event,\n context))\n", "self.assertEqual(users.keys(), {self.u_alice, self.u_bob})\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "Assign'", "Expr'" ]
[ "@CLASS_4('bookmarks')...\n", "\"\"\"docstring\"\"\"\n", "VAR_18 = sorted(objreg.get('bookmark-manager').marks.items(), key=lambda x:\n x[1])\n", "VAR_19 = sorted(objreg.get('quickmark-manager').marks.items(), key=lambda x:\n x[0])\n", "VAR_20 = jinja.render('bookmarks.html', title='Bookmarks', VAR_18=bookmarks,\n VAR_19=quickmarks)\n", "return 'text/html', VAR_20\n" ]
[ "@add_handler('bookmarks')...\n", "\"\"\"docstring\"\"\"\n", "bookmarks = sorted(objreg.get('bookmark-manager').marks.items(), key=lambda\n x: x[1])\n", "quickmarks = sorted(objreg.get('quickmark-manager').marks.items(), key=lambda\n x: x[0])\n", "html = jinja.render('bookmarks.html', title='Bookmarks', bookmarks=\n bookmarks, quickmarks=quickmarks)\n", "return 'text/html', html\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "async def FUNC_5(VAR_7=None, VAR_8=False):...\n", "return {'user': UserID.from_string(self.auth_user_id), 'token_id': 1,\n 'is_guest': False}\n" ]
[ "async def get_user_by_access_token(token=None, allow_guest=False):...\n", "return {'user': UserID.from_string(self.auth_user_id), 'token_id': 1,\n 'is_guest': False}\n" ]
[ 0, 0 ]
[ "AsyncFunctionDef'", "Return'" ]
[ "async def FUNC_27(self, VAR_5: str, VAR_11: str, VAR_34: Iterable[...\n", "\"\"\"docstring\"\"\"\n", "async def FUNC_49(VAR_102: _NewEventInfo):...\n", "VAR_1 = VAR_102.event\n", "VAR_168 = await self._prep_event(VAR_5, VAR_1, VAR_2=ev_info.state, VAR_3=\n ev_info.auth_events, VAR_33=backfilled)\n", "return VAR_168\n" ]
[ "async def _handle_new_events(self, origin: str, room_id: str, event_infos:...\n", "\"\"\"docstring\"\"\"\n", "async def prep(ev_info: _NewEventInfo):...\n", "event = ev_info.event\n", "res = await self._prep_event(origin, event, state=ev_info.state,\n auth_events=ev_info.auth_events, backfilled=backfilled)\n", "return res\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "AsyncFunctionDef'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_2(VAR_0, VAR_1):...\n", "return FUNC_0(VAR_0=request, VAR_1=pk, VAR_2='is_removed', VAR_3=False,\n VAR_5=_('The topic has been undeleted'))\n" ]
[ "def undelete(request, pk):...\n", "return _moderate(request=request, pk=pk, field_name='is_removed', to_value=\n False, message=_('The topic has been undeleted'))\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_0(VAR_0: Optional[str], VAR_1: Optional[Path]) ->Union[CLASS_0,...\n", "VAR_2 = FUNC_3(VAR_0=url, VAR_1=path)\n", "if isinstance(VAR_2, GeneratorError):\n", "return VAR_2\n", "VAR_3 = GeneratorData.from_dict(VAR_2)\n", "if isinstance(VAR_3, GeneratorError):\n", "return VAR_3\n", "return CLASS_0(VAR_3=openapi)\n" ]
[ "def _get_project_for_url_or_path(url: Optional[str], path: Optional[Path]...\n", "data_dict = _get_document(url=url, path=path)\n", "if isinstance(data_dict, GeneratorError):\n", "return data_dict\n", "openapi = GeneratorData.from_dict(data_dict)\n", "if isinstance(openapi, GeneratorError):\n", "return openapi\n", "return Project(openapi=openapi)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Assign'", "Condition", "Return'", "Assign'", "Condition", "Return'", "Return'" ]
[ "def FUNC_17(self, VAR_19, VAR_40):...\n", "return self\n" ]
[ "def creatorForNetloc(self, hostname, port):...\n", "return self\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "import pathlib\n", "import httpcore\n", "import jinja2\n", "import pytest\n", "import yaml\n", "from openapi_python_client import GeneratorError\n", "def FUNC_0(VAR_0):...\n", "VAR_1 = VAR_0.MagicMock()\n", "VAR_2 = VAR_0.patch('openapi_python_client._get_document', return_value=\n data_dict)\n", "VAR_3 = VAR_0.MagicMock()\n", "VAR_4 = VAR_0.patch('openapi_python_client.parser.GeneratorData.from_dict',\n return_value=openapi)\n", "VAR_5 = VAR_0.patch('openapi_python_client.Project')\n", "VAR_6 = VAR_0.MagicMock()\n", "VAR_7 = VAR_0.MagicMock()\n", "from openapi_python_client import _get_project_for_url_or_path\n", "VAR_8 = VAR_10(VAR_6=url, VAR_7=path)\n", "VAR_2.assert_called_once_with(VAR_6=url, VAR_7=path)\n", "VAR_4.assert_called_once_with(VAR_1)\n", "VAR_5.assert_called_once_with(VAR_3=openapi)\n", "assert VAR_8 == VAR_5()\n", "def FUNC_1(VAR_0):...\n", "VAR_1 = VAR_0.MagicMock()\n", "VAR_2 = VAR_0.patch('openapi_python_client._get_document', return_value=\n data_dict)\n", "VAR_9 = GeneratorError()\n", "VAR_4 = VAR_0.patch('openapi_python_client.parser.GeneratorData.from_dict',\n return_value=error)\n", "VAR_5 = VAR_0.patch('openapi_python_client.Project')\n", "VAR_6 = VAR_0.MagicMock()\n", "VAR_7 = VAR_0.MagicMock()\n", "from openapi_python_client import _get_project_for_url_or_path\n", "VAR_8 = VAR_10(VAR_6=url, VAR_7=path)\n", "VAR_2.assert_called_once_with(VAR_6=url, VAR_7=path)\n", "VAR_4.assert_called_once_with(VAR_1)\n", "VAR_5.assert_not_called()\n", "assert VAR_8 == VAR_9\n", "def FUNC_2(VAR_0):...\n", "VAR_9 = GeneratorError()\n", "VAR_2 = VAR_0.patch('openapi_python_client._get_document', return_value=error)\n", "VAR_4 = VAR_0.patch('openapi_python_client.parser.GeneratorData.from_dict')\n", "VAR_6 = VAR_0.MagicMock()\n", "VAR_7 = VAR_0.MagicMock()\n", "from openapi_python_client import _get_project_for_url_or_path\n", "VAR_8 = VAR_10(VAR_6=url, VAR_7=path)\n", "VAR_2.assert_called_once_with(VAR_6=url, VAR_7=path)\n", "VAR_4.assert_not_called()\n", "assert VAR_8 == VAR_9\n", "def FUNC_3(VAR_0):...\n", "VAR_8 = VAR_0.MagicMock()\n", "VAR_10 = VAR_0.patch('openapi_python_client._get_project_for_url_or_path',\n return_value=project)\n", "VAR_6 = VAR_0.MagicMock()\n", "VAR_7 = VAR_0.MagicMock()\n", "from openapi_python_client import create_new_client\n", "VAR_11 = create_new_client(VAR_6=url, VAR_7=path)\n", "VAR_10.assert_called_once_with(VAR_6=url, VAR_7=path)\n", "VAR_8.build.assert_called_once()\n", "assert VAR_11 == VAR_8.build.return_value\n", "def FUNC_4(VAR_0):...\n", "VAR_9 = GeneratorError()\n", "VAR_10 = VAR_0.patch('openapi_python_client._get_project_for_url_or_path',\n return_value=error)\n", "VAR_6 = VAR_0.MagicMock()\n", "VAR_7 = VAR_0.MagicMock()\n", "from openapi_python_client import create_new_client\n", "VAR_11 = create_new_client(VAR_6=url, VAR_7=path)\n", "VAR_10.assert_called_once_with(VAR_6=url, VAR_7=path)\n", "assert VAR_11 == [VAR_9]\n", "def FUNC_5(VAR_0):...\n", "VAR_8 = VAR_0.MagicMock()\n", "VAR_10 = VAR_0.patch('openapi_python_client._get_project_for_url_or_path',\n return_value=project)\n", "VAR_6 = VAR_0.MagicMock()\n", "VAR_7 = VAR_0.MagicMock()\n", "from openapi_python_client import update_existing_client\n", "VAR_11 = update_existing_client(VAR_6=url, VAR_7=path)\n", "VAR_10.assert_called_once_with(VAR_6=url, VAR_7=path)\n", "VAR_8.update.assert_called_once()\n", "assert VAR_11 == VAR_8.update.return_value\n", "def FUNC_6(VAR_0):...\n", "VAR_9 = GeneratorError()\n", "VAR_10 = VAR_0.patch('openapi_python_client._get_project_for_url_or_path',\n return_value=error)\n", "VAR_6 = VAR_0.MagicMock()\n", "VAR_7 = VAR_0.MagicMock()\n", "from openapi_python_client import update_existing_client\n", "VAR_11 = update_existing_client(VAR_6=url, VAR_7=path)\n", "VAR_10.assert_called_once_with(VAR_6=url, VAR_7=path)\n", "assert VAR_11 == [VAR_9]\n", "def FUNC_9(self, VAR_0):...\n", "VAR_13 = VAR_0.patch('httpx.get')\n", "VAR_14 = VAR_0.patch('openapi_python_client.Path')\n", "VAR_15 = VAR_0.patch('yaml.safe_load')\n", "from openapi_python_client import _get_document\n", "VAR_11 = VAR_2(VAR_6=None, VAR_7=None)\n", "assert VAR_11 == GeneratorError(header='No URL or Path provided')\n", "VAR_13.assert_not_called()\n", "VAR_14.assert_not_called()\n", "VAR_15.assert_not_called()\n", "def FUNC_10(self, VAR_0):...\n", "VAR_13 = VAR_0.patch('httpx.get')\n", "VAR_14 = VAR_0.patch('openapi_python_client.Path')\n", "VAR_15 = VAR_0.patch('yaml.safe_load')\n", "from openapi_python_client import _get_document\n", "VAR_11 = VAR_2(VAR_6=mocker.MagicMock(), VAR_7=mocker.MagicMock())\n", "assert VAR_11 == GeneratorError(header='Provide URL or Path, not both.')\n", "VAR_13.assert_not_called()\n", "VAR_14.assert_not_called()\n", "VAR_15.assert_not_called()\n", "def FUNC_11(self, VAR_0):...\n", "VAR_13 = VAR_0.patch('httpx.get', side_effect=httpcore.NetworkError)\n", "VAR_14 = VAR_0.patch('openapi_python_client.Path')\n", "VAR_15 = VAR_0.patch('yaml.safe_load')\n", "from openapi_python_client import _get_document\n", "VAR_6 = VAR_0.MagicMock()\n", "VAR_11 = VAR_2(VAR_6=url, VAR_7=None)\n", "assert VAR_11 == GeneratorError(header=\n 'Could not get OpenAPI document from provided URL')\n", "VAR_13.assert_called_once_with(VAR_6)\n", "VAR_14.assert_not_called()\n", "VAR_15.assert_not_called()\n", "def FUNC_12(self, VAR_0):...\n", "VAR_13 = VAR_0.patch('httpx.get')\n", "VAR_14 = VAR_0.patch('openapi_python_client.Path')\n", "VAR_15 = VAR_0.patch('yaml.safe_load')\n", "from openapi_python_client import _get_document\n", "VAR_6 = VAR_0.MagicMock()\n", "VAR_2(VAR_6=url, VAR_7=None)\n", "VAR_13.assert_called_once_with(VAR_6)\n", "VAR_14.assert_not_called()\n", "VAR_15.assert_called_once_with(VAR_13().content)\n", "def FUNC_13(self, VAR_0):...\n", "VAR_13 = VAR_0.patch('httpx.get')\n", "VAR_15 = VAR_0.patch('yaml.safe_load')\n", "from openapi_python_client import _get_document\n", "VAR_7 = VAR_0.MagicMock()\n", "VAR_2(VAR_6=None, VAR_7=path)\n", "VAR_13.assert_not_called()\n", "VAR_7.read_bytes.assert_called_once()\n", "VAR_15.assert_called_once_with(VAR_7.read_bytes())\n", "def FUNC_14(self, VAR_0):...\n", "VAR_13 = VAR_0.patch('httpx.get')\n", "VAR_15 = VAR_0.patch('yaml.safe_load', side_effect=yaml.YAMLError)\n", "from openapi_python_client import _get_document\n", "VAR_7 = VAR_0.MagicMock()\n", "VAR_11 = VAR_2(VAR_6=None, VAR_7=path)\n", "VAR_13.assert_not_called()\n", "VAR_7.read_bytes.assert_called_once()\n", "VAR_15.assert_called_once_with(VAR_7.read_bytes())\n", "assert VAR_11 == GeneratorError(header='Invalid YAML from provided source')\n", "def test___init__(self, VAR_0):...\n", "VAR_3 = VAR_0.MagicMock(title='My Test API')\n", "from openapi_python_client import Project\n", "VAR_8 = Project(VAR_3=openapi)\n", "assert VAR_8.openapi == VAR_3\n", "assert VAR_8.project_name == 'my-test-api-client'\n", "assert VAR_8.package_name == 'my_test_api_client'\n", "assert VAR_8.package_description == 'A client library for accessing My Test API'\n", "def FUNC_15(self, VAR_0):...\n", "VAR_3 = VAR_0.MagicMock(title='My Test API')\n", "from openapi_python_client import Project\n", "Project.project_name_override = 'my-special-project-name'\n", "VAR_8 = Project(VAR_3=openapi)\n", "assert VAR_8.project_name == 'my-special-project-name'\n", "assert VAR_8.package_name == 'my_special_project_name'\n", "Project.package_name_override = 'my_special_package_name'\n", "VAR_8 = Project(VAR_3=openapi)\n", "assert VAR_8.project_name == 'my-special-project-name'\n", "assert VAR_8.package_name == 'my_special_package_name'\n", "def FUNC_16(self, VAR_0):...\n", "from openapi_python_client import Project\n", "VAR_8 = Project(VAR_3=mocker.MagicMock(title='My Test API'))\n", "VAR_8.project_dir = VAR_0.MagicMock()\n", "VAR_8.package_dir = VAR_0.MagicMock()\n", "VAR_8._build_metadata = VAR_0.MagicMock()\n", "VAR_8._build_models = VAR_0.MagicMock()\n", "VAR_8._build_api = VAR_0.MagicMock()\n", "VAR_8._create_package = VAR_0.MagicMock()\n", "VAR_8._reformat = VAR_0.MagicMock()\n", "VAR_8._get_errors = VAR_0.MagicMock()\n", "VAR_11 = VAR_8.build()\n", "VAR_8.project_dir.mkdir.assert_called_once()\n", "VAR_8._create_package.assert_called_once()\n", "VAR_8._build_metadata.assert_called_once()\n", "VAR_8._build_models.assert_called_once()\n", "VAR_8._build_api.assert_called_once()\n", "VAR_8._reformat.assert_called_once()\n", "VAR_8._get_errors.assert_called_once()\n", "assert VAR_11 == VAR_8._get_errors.return_value\n", "def FUNC_17(self, VAR_0):...\n", "from openapi_python_client import Project\n", "VAR_8 = Project(VAR_3=mocker.MagicMock(title='My Test API'))\n", "VAR_8.project_dir = VAR_0.MagicMock()\n", "VAR_8.project_dir.mkdir.side_effect = FileExistsError\n", "VAR_11 = VAR_8.build()\n", "VAR_8.project_dir.mkdir.assert_called_once()\n", "assert VAR_11 == [GeneratorError(detail=\n 'Directory already exists. Delete it or use the update command.')]\n", "def FUNC_18(self, VAR_0):...\n", "from openapi_python_client import Project, shutil\n", "VAR_16 = VAR_0.patch.object(shutil, 'rmtree')\n", "VAR_8 = Project(VAR_3=mocker.MagicMock(title='My Test API'))\n", "VAR_8.package_dir = VAR_0.MagicMock()\n", "VAR_8._build_metadata = VAR_0.MagicMock()\n", "VAR_8._build_models = VAR_0.MagicMock()\n", "VAR_8._build_api = VAR_0.MagicMock()\n", "VAR_8._create_package = VAR_0.MagicMock()\n", "VAR_8._reformat = VAR_0.MagicMock()\n", "VAR_8._get_errors = VAR_0.MagicMock()\n", "VAR_11 = VAR_8.update()\n", "VAR_16.assert_called_once_with(VAR_8.package_dir)\n", "VAR_8._create_package.assert_called_once()\n", "VAR_8._build_models.assert_called_once()\n", "VAR_8._build_api.assert_called_once()\n", "VAR_8._reformat.assert_called_once()\n", "VAR_8._get_errors.assert_called_once()\n", "assert VAR_11 == VAR_8._get_errors.return_value\n", "def FUNC_19(self, VAR_0):...\n", "from openapi_python_client import Project\n", "VAR_8 = Project(VAR_3=mocker.MagicMock(title='My Test API'))\n", "VAR_8.package_dir = VAR_0.MagicMock()\n", "VAR_8.package_dir.is_dir.return_value = False\n", "VAR_8._build_models = VAR_0.MagicMock()\n", "VAR_8.update()\n", "VAR_8.package_dir.is_dir.assert_called_once()\n", "VAR_8._build_models.assert_not_called()\n", "def FUNC_20(self, VAR_0):...\n", "from openapi_python_client import Project\n", "VAR_8 = Project(VAR_3=mocker.MagicMock(title='My Test API'))\n", "VAR_8.package_dir = VAR_0.MagicMock()\n", "VAR_17 = VAR_0.MagicMock()\n", "VAR_8.env = VAR_0.MagicMock()\n", "VAR_8.env.get_template.return_value = VAR_17\n", "VAR_18 = VAR_0.MagicMock(autospec=pathlib.Path)\n", "VAR_19 = VAR_0.MagicMock(autospec=pathlib.Path)\n", "VAR_20 = {'__init__.py': VAR_18, 'py.typed': VAR_19}\n", "VAR_8.package_dir.__truediv__.side_effect = lambda VAR_40: VAR_20[VAR_40]\n", "VAR_8._create_package()\n", "VAR_8.package_dir.mkdir.assert_called_once()\n", "VAR_8.env.get_template.assert_called_once_with('package_init.pyi')\n", "VAR_17.render.assert_called_once_with(description=project.package_description)\n", "VAR_18.write_text.assert_called_once_with(VAR_17.render())\n", "VAR_19.write_text.assert_called_once_with('# Marker file for PEP 561')\n", "def FUNC_21(self, VAR_0):...\n", "from openapi_python_client import Project\n", "VAR_8 = Project(VAR_3=mocker.MagicMock(title='My Test API'))\n", "VAR_8.project_dir = VAR_0.MagicMock()\n", "VAR_21 = VAR_0.MagicMock(autospec=pathlib.Path)\n", "VAR_22 = VAR_0.MagicMock(autospec=pathlib.Path)\n", "VAR_23 = VAR_0.MagicMock(autospec=pathlib.Path)\n", "VAR_20 = {'pyproject.toml': VAR_21, 'README.md': VAR_22, '.gitignore': VAR_23}\n", "VAR_8.project_dir.__truediv__.side_effect = lambda VAR_40: VAR_20[VAR_40]\n", "VAR_24 = VAR_0.MagicMock(autospec=jinja2.Template)\n", "VAR_25 = VAR_0.MagicMock(autospec=jinja2.Template)\n", "VAR_26 = VAR_0.MagicMock(autospec=jinja2.Template)\n", "VAR_8.env = VAR_0.MagicMock(autospec=jinja2.Environment)\n", "VAR_27 = {'pyproject.toml': VAR_24, 'README.md': VAR_25, '.gitignore': VAR_26}\n", "VAR_8.env.get_template.side_effect = lambda VAR_40: VAR_27[VAR_40]\n", "VAR_8._build_metadata()\n", "VAR_8.env.get_template.assert_has_calls([VAR_0.call('pyproject.toml'),\n VAR_0.call('README.md'), VAR_0.call('.gitignore')])\n", "VAR_24.render.assert_called_once_with(project_name=project.project_name,\n package_name=project.package_name, version=project.version, description\n =project.package_description)\n", "VAR_21.write_text.assert_called_once_with(VAR_24.render())\n", "VAR_25.render.assert_called_once_with(description=project.\n package_description, project_name=project.project_name, package_name=\n project.package_name)\n", "VAR_22.write_text.assert_called_once_with(VAR_25.render())\n", "VAR_26.render.assert_called_once()\n", "VAR_23.write_text.assert_called_once_with(VAR_26.render())\n", "def FUNC_22(self, VAR_0):...\n", "from openapi_python_client import GeneratorData, Project\n", "VAR_3 = VAR_0.MagicMock(autospec=GeneratorData, title='My Test API')\n", "VAR_28 = VAR_0.MagicMock()\n", "VAR_29 = VAR_0.MagicMock()\n", "VAR_3.schemas.models = {'1': VAR_28, '2': VAR_29}\n", "VAR_30 = VAR_0.MagicMock()\n", "VAR_31 = VAR_0.MagicMock()\n", "VAR_3.enums = {'1': VAR_30, '2': VAR_31}\n", "VAR_8 = Project(VAR_3=openapi)\n", "VAR_8.package_dir = VAR_0.MagicMock()\n", "VAR_32 = VAR_0.MagicMock()\n", "VAR_33 = VAR_0.MagicMock()\n", "VAR_34 = VAR_0.MagicMock()\n", "VAR_35 = VAR_0.MagicMock()\n", "VAR_36 = VAR_0.MagicMock()\n", "VAR_37 = VAR_0.MagicMock()\n", "VAR_38 = VAR_0.MagicMock()\n", "VAR_39 = {'__init__.py': VAR_32, 'types.py': VAR_33,\n f'{VAR_28.reference.module_name}.py': VAR_35,\n f'{VAR_29.reference.module_name}.py': VAR_36,\n f'{VAR_30.reference.module_name}.py': VAR_37,\n f'{VAR_31.reference.module_name}.py': VAR_38}\n", "def FUNC_24(VAR_40):...\n", "return VAR_39[VAR_40]\n" ]
[ "import pathlib\n", "import httpcore\n", "import jinja2\n", "import pytest\n", "import yaml\n", "from openapi_python_client import GeneratorError\n", "def test__get_project_for_url_or_path(mocker):...\n", "data_dict = mocker.MagicMock()\n", "_get_document = mocker.patch('openapi_python_client._get_document',\n return_value=data_dict)\n", "openapi = mocker.MagicMock()\n", "from_dict = mocker.patch('openapi_python_client.parser.GeneratorData.from_dict'\n , return_value=openapi)\n", "_Project = mocker.patch('openapi_python_client.Project')\n", "url = mocker.MagicMock()\n", "path = mocker.MagicMock()\n", "from openapi_python_client import _get_project_for_url_or_path\n", "project = _get_project_for_url_or_path(url=url, path=path)\n", "_get_document.assert_called_once_with(url=url, path=path)\n", "from_dict.assert_called_once_with(data_dict)\n", "_Project.assert_called_once_with(openapi=openapi)\n", "assert project == _Project()\n", "def test__get_project_for_url_or_path_generator_error(mocker):...\n", "data_dict = mocker.MagicMock()\n", "_get_document = mocker.patch('openapi_python_client._get_document',\n return_value=data_dict)\n", "error = GeneratorError()\n", "from_dict = mocker.patch('openapi_python_client.parser.GeneratorData.from_dict'\n , return_value=error)\n", "_Project = mocker.patch('openapi_python_client.Project')\n", "url = mocker.MagicMock()\n", "path = mocker.MagicMock()\n", "from openapi_python_client import _get_project_for_url_or_path\n", "project = _get_project_for_url_or_path(url=url, path=path)\n", "_get_document.assert_called_once_with(url=url, path=path)\n", "from_dict.assert_called_once_with(data_dict)\n", "_Project.assert_not_called()\n", "assert project == error\n", "def test__get_project_for_url_or_path_document_error(mocker):...\n", "error = GeneratorError()\n", "_get_document = mocker.patch('openapi_python_client._get_document',\n return_value=error)\n", "from_dict = mocker.patch('openapi_python_client.parser.GeneratorData.from_dict'\n )\n", "url = mocker.MagicMock()\n", "path = mocker.MagicMock()\n", "from openapi_python_client import _get_project_for_url_or_path\n", "project = _get_project_for_url_or_path(url=url, path=path)\n", "_get_document.assert_called_once_with(url=url, path=path)\n", "from_dict.assert_not_called()\n", "assert project == error\n", "def test_create_new_client(mocker):...\n", "project = mocker.MagicMock()\n", "_get_project_for_url_or_path = mocker.patch(\n 'openapi_python_client._get_project_for_url_or_path', return_value=project)\n", "url = mocker.MagicMock()\n", "path = mocker.MagicMock()\n", "from openapi_python_client import create_new_client\n", "result = create_new_client(url=url, path=path)\n", "_get_project_for_url_or_path.assert_called_once_with(url=url, path=path)\n", "project.build.assert_called_once()\n", "assert result == project.build.return_value\n", "def test_create_new_client_project_error(mocker):...\n", "error = GeneratorError()\n", "_get_project_for_url_or_path = mocker.patch(\n 'openapi_python_client._get_project_for_url_or_path', return_value=error)\n", "url = mocker.MagicMock()\n", "path = mocker.MagicMock()\n", "from openapi_python_client import create_new_client\n", "result = create_new_client(url=url, path=path)\n", "_get_project_for_url_or_path.assert_called_once_with(url=url, path=path)\n", "assert result == [error]\n", "def test_update_existing_client(mocker):...\n", "project = mocker.MagicMock()\n", "_get_project_for_url_or_path = mocker.patch(\n 'openapi_python_client._get_project_for_url_or_path', return_value=project)\n", "url = mocker.MagicMock()\n", "path = mocker.MagicMock()\n", "from openapi_python_client import update_existing_client\n", "result = update_existing_client(url=url, path=path)\n", "_get_project_for_url_or_path.assert_called_once_with(url=url, path=path)\n", "project.update.assert_called_once()\n", "assert result == project.update.return_value\n", "def test_update_existing_client_project_error(mocker):...\n", "error = GeneratorError()\n", "_get_project_for_url_or_path = mocker.patch(\n 'openapi_python_client._get_project_for_url_or_path', return_value=error)\n", "url = mocker.MagicMock()\n", "path = mocker.MagicMock()\n", "from openapi_python_client import update_existing_client\n", "result = update_existing_client(url=url, path=path)\n", "_get_project_for_url_or_path.assert_called_once_with(url=url, path=path)\n", "assert result == [error]\n", "def test__get_document_no_url_or_path(self, mocker):...\n", "get = mocker.patch('httpx.get')\n", "Path = mocker.patch('openapi_python_client.Path')\n", "loads = mocker.patch('yaml.safe_load')\n", "from openapi_python_client import _get_document\n", "result = _get_document(url=None, path=None)\n", "assert result == GeneratorError(header='No URL or Path provided')\n", "get.assert_not_called()\n", "Path.assert_not_called()\n", "loads.assert_not_called()\n", "def test__get_document_url_and_path(self, mocker):...\n", "get = mocker.patch('httpx.get')\n", "Path = mocker.patch('openapi_python_client.Path')\n", "loads = mocker.patch('yaml.safe_load')\n", "from openapi_python_client import _get_document\n", "result = _get_document(url=mocker.MagicMock(), path=mocker.MagicMock())\n", "assert result == GeneratorError(header='Provide URL or Path, not both.')\n", "get.assert_not_called()\n", "Path.assert_not_called()\n", "loads.assert_not_called()\n", "def test__get_document_bad_url(self, mocker):...\n", "get = mocker.patch('httpx.get', side_effect=httpcore.NetworkError)\n", "Path = mocker.patch('openapi_python_client.Path')\n", "loads = mocker.patch('yaml.safe_load')\n", "from openapi_python_client import _get_document\n", "url = mocker.MagicMock()\n", "result = _get_document(url=url, path=None)\n", "assert result == GeneratorError(header=\n 'Could not get OpenAPI document from provided URL')\n", "get.assert_called_once_with(url)\n", "Path.assert_not_called()\n", "loads.assert_not_called()\n", "def test__get_document_url_no_path(self, mocker):...\n", "get = mocker.patch('httpx.get')\n", "Path = mocker.patch('openapi_python_client.Path')\n", "loads = mocker.patch('yaml.safe_load')\n", "from openapi_python_client import _get_document\n", "url = mocker.MagicMock()\n", "_get_document(url=url, path=None)\n", "get.assert_called_once_with(url)\n", "Path.assert_not_called()\n", "loads.assert_called_once_with(get().content)\n", "def test__get_document_path_no_url(self, mocker):...\n", "get = mocker.patch('httpx.get')\n", "loads = mocker.patch('yaml.safe_load')\n", "from openapi_python_client import _get_document\n", "path = mocker.MagicMock()\n", "_get_document(url=None, path=path)\n", "get.assert_not_called()\n", "path.read_bytes.assert_called_once()\n", "loads.assert_called_once_with(path.read_bytes())\n", "def test__get_document_bad_yaml(self, mocker):...\n", "get = mocker.patch('httpx.get')\n", "loads = mocker.patch('yaml.safe_load', side_effect=yaml.YAMLError)\n", "from openapi_python_client import _get_document\n", "path = mocker.MagicMock()\n", "result = _get_document(url=None, path=path)\n", "get.assert_not_called()\n", "path.read_bytes.assert_called_once()\n", "loads.assert_called_once_with(path.read_bytes())\n", "assert result == GeneratorError(header='Invalid YAML from provided source')\n", "def test___init__(self, mocker):...\n", "openapi = mocker.MagicMock(title='My Test API')\n", "from openapi_python_client import Project\n", "project = Project(openapi=openapi)\n", "assert project.openapi == openapi\n", "assert project.project_name == 'my-test-api-client'\n", "assert project.package_name == 'my_test_api_client'\n", "assert project.package_description == 'A client library for accessing My Test API'\n", "def test_project_and_package_name_overrides(self, mocker):...\n", "openapi = mocker.MagicMock(title='My Test API')\n", "from openapi_python_client import Project\n", "Project.project_name_override = 'my-special-project-name'\n", "project = Project(openapi=openapi)\n", "assert project.project_name == 'my-special-project-name'\n", "assert project.package_name == 'my_special_project_name'\n", "Project.package_name_override = 'my_special_package_name'\n", "project = Project(openapi=openapi)\n", "assert project.project_name == 'my-special-project-name'\n", "assert project.package_name == 'my_special_package_name'\n", "def test_build(self, mocker):...\n", "from openapi_python_client import Project\n", "project = Project(openapi=mocker.MagicMock(title='My Test API'))\n", "project.project_dir = mocker.MagicMock()\n", "project.package_dir = mocker.MagicMock()\n", "project._build_metadata = mocker.MagicMock()\n", "project._build_models = mocker.MagicMock()\n", "project._build_api = mocker.MagicMock()\n", "project._create_package = mocker.MagicMock()\n", "project._reformat = mocker.MagicMock()\n", "project._get_errors = mocker.MagicMock()\n", "result = project.build()\n", "project.project_dir.mkdir.assert_called_once()\n", "project._create_package.assert_called_once()\n", "project._build_metadata.assert_called_once()\n", "project._build_models.assert_called_once()\n", "project._build_api.assert_called_once()\n", "project._reformat.assert_called_once()\n", "project._get_errors.assert_called_once()\n", "assert result == project._get_errors.return_value\n", "def test_build_file_exists(self, mocker):...\n", "from openapi_python_client import Project\n", "project = Project(openapi=mocker.MagicMock(title='My Test API'))\n", "project.project_dir = mocker.MagicMock()\n", "project.project_dir.mkdir.side_effect = FileExistsError\n", "result = project.build()\n", "project.project_dir.mkdir.assert_called_once()\n", "assert result == [GeneratorError(detail=\n 'Directory already exists. Delete it or use the update command.')]\n", "def test_update(self, mocker):...\n", "from openapi_python_client import Project, shutil\n", "rmtree = mocker.patch.object(shutil, 'rmtree')\n", "project = Project(openapi=mocker.MagicMock(title='My Test API'))\n", "project.package_dir = mocker.MagicMock()\n", "project._build_metadata = mocker.MagicMock()\n", "project._build_models = mocker.MagicMock()\n", "project._build_api = mocker.MagicMock()\n", "project._create_package = mocker.MagicMock()\n", "project._reformat = mocker.MagicMock()\n", "project._get_errors = mocker.MagicMock()\n", "result = project.update()\n", "rmtree.assert_called_once_with(project.package_dir)\n", "project._create_package.assert_called_once()\n", "project._build_models.assert_called_once()\n", "project._build_api.assert_called_once()\n", "project._reformat.assert_called_once()\n", "project._get_errors.assert_called_once()\n", "assert result == project._get_errors.return_value\n", "def test_update_missing_dir(self, mocker):...\n", "from openapi_python_client import Project\n", "project = Project(openapi=mocker.MagicMock(title='My Test API'))\n", "project.package_dir = mocker.MagicMock()\n", "project.package_dir.is_dir.return_value = False\n", "project._build_models = mocker.MagicMock()\n", "project.update()\n", "project.package_dir.is_dir.assert_called_once()\n", "project._build_models.assert_not_called()\n", "def test__create_package(self, mocker):...\n", "from openapi_python_client import Project\n", "project = Project(openapi=mocker.MagicMock(title='My Test API'))\n", "project.package_dir = mocker.MagicMock()\n", "package_init_template = mocker.MagicMock()\n", "project.env = mocker.MagicMock()\n", "project.env.get_template.return_value = package_init_template\n", "package_init_path = mocker.MagicMock(autospec=pathlib.Path)\n", "pytyped_path = mocker.MagicMock(autospec=pathlib.Path)\n", "paths = {'__init__.py': package_init_path, 'py.typed': pytyped_path}\n", "project.package_dir.__truediv__.side_effect = lambda x: paths[x]\n", "project._create_package()\n", "project.package_dir.mkdir.assert_called_once()\n", "project.env.get_template.assert_called_once_with('package_init.pyi')\n", "package_init_template.render.assert_called_once_with(description=project.\n package_description)\n", "package_init_path.write_text.assert_called_once_with(package_init_template.\n render())\n", "pytyped_path.write_text.assert_called_once_with('# Marker file for PEP 561')\n", "def test__build_metadata(self, mocker):...\n", "from openapi_python_client import Project\n", "project = Project(openapi=mocker.MagicMock(title='My Test API'))\n", "project.project_dir = mocker.MagicMock()\n", "pyproject_path = mocker.MagicMock(autospec=pathlib.Path)\n", "readme_path = mocker.MagicMock(autospec=pathlib.Path)\n", "git_ignore_path = mocker.MagicMock(autospec=pathlib.Path)\n", "paths = {'pyproject.toml': pyproject_path, 'README.md': readme_path,\n '.gitignore': git_ignore_path}\n", "project.project_dir.__truediv__.side_effect = lambda x: paths[x]\n", "pyproject_template = mocker.MagicMock(autospec=jinja2.Template)\n", "readme_template = mocker.MagicMock(autospec=jinja2.Template)\n", "git_ignore_template = mocker.MagicMock(autospec=jinja2.Template)\n", "project.env = mocker.MagicMock(autospec=jinja2.Environment)\n", "templates = {'pyproject.toml': pyproject_template, 'README.md':\n readme_template, '.gitignore': git_ignore_template}\n", "project.env.get_template.side_effect = lambda x: templates[x]\n", "project._build_metadata()\n", "project.env.get_template.assert_has_calls([mocker.call('pyproject.toml'),\n mocker.call('README.md'), mocker.call('.gitignore')])\n", "pyproject_template.render.assert_called_once_with(project_name=project.\n project_name, package_name=project.package_name, version=project.\n version, description=project.package_description)\n", "pyproject_path.write_text.assert_called_once_with(pyproject_template.render())\n", "readme_template.render.assert_called_once_with(description=project.\n package_description, project_name=project.project_name, package_name=\n project.package_name)\n", "readme_path.write_text.assert_called_once_with(readme_template.render())\n", "git_ignore_template.render.assert_called_once()\n", "git_ignore_path.write_text.assert_called_once_with(git_ignore_template.render()\n )\n", "def test__build_models(self, mocker):...\n", "from openapi_python_client import GeneratorData, Project\n", "openapi = mocker.MagicMock(autospec=GeneratorData, title='My Test API')\n", "model_1 = mocker.MagicMock()\n", "model_2 = mocker.MagicMock()\n", "openapi.schemas.models = {'1': model_1, '2': model_2}\n", "enum_1 = mocker.MagicMock()\n", "enum_2 = mocker.MagicMock()\n", "openapi.enums = {'1': enum_1, '2': enum_2}\n", "project = Project(openapi=openapi)\n", "project.package_dir = mocker.MagicMock()\n", "models_init = mocker.MagicMock()\n", "types_py = mocker.MagicMock()\n", "models_dir = mocker.MagicMock()\n", "model_1_module_path = mocker.MagicMock()\n", "model_2_module_path = mocker.MagicMock()\n", "enum_1_module_path = mocker.MagicMock()\n", "enum_2_module_path = mocker.MagicMock()\n", "module_paths = {'__init__.py': models_init, 'types.py': types_py,\n f'{model_1.reference.module_name}.py': model_1_module_path,\n f'{model_2.reference.module_name}.py': model_2_module_path,\n f'{enum_1.reference.module_name}.py': enum_1_module_path,\n f'{enum_2.reference.module_name}.py': enum_2_module_path}\n", "def models_dir_get(x):...\n", "return module_paths[x]\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Import'", "Import'", "Import'", "Import'", "Import'", "ImportFrom'", "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "ImportFrom'", "Assign'", "Expr'", "Expr'", "Expr'", "Assert'", "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "ImportFrom'", "Assign'", "Expr'", "Expr'", "Expr'", "Assert'", "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "ImportFrom'", "Assign'", "Expr'", "Expr'", "Assert'", "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "ImportFrom'", "Assign'", "Expr'", "Expr'", "Assert'", "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "ImportFrom'", "Assign'", "Expr'", "Assert'", "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "ImportFrom'", "Assign'", "Expr'", "Expr'", "Assert'", "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "ImportFrom'", "Assign'", "Expr'", "Assert'", "FunctionDef'", "Assign'", "Assign'", "Assign'", "ImportFrom'", "Assign'", "Assert'", "Expr'", "Expr'", "Expr'", "FunctionDef'", "Assign'", "Assign'", "Assign'", "ImportFrom'", "Assign'", "Assert'", "Expr'", "Expr'", "Expr'", "FunctionDef'", "Assign'", "Assign'", "Assign'", "ImportFrom'", "Assign'", "Assign'", "Assert'", "Expr'", "Expr'", "Expr'", "FunctionDef'", "Assign'", "Assign'", "Assign'", "ImportFrom'", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'", "FunctionDef'", "Assign'", "Assign'", "ImportFrom'", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'", "FunctionDef'", "Assign'", "Assign'", "ImportFrom'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'", "Assert'", "FunctionDef'", "Assign'", "ImportFrom'", "Assign'", "Assert'", "Assert'", "Assert'", "Assert'", "FunctionDef'", "Assign'", "ImportFrom'", "Assign'", "Assign'", "Assert'", "Assert'", "Assign'", "Assign'", "Assert'", "Assert'", "FunctionDef'", "ImportFrom'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Assert'", "FunctionDef'", "ImportFrom'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Assert'", "FunctionDef'", "ImportFrom'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Assert'", "FunctionDef'", "ImportFrom'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'", "FunctionDef'", "ImportFrom'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "FunctionDef'", "ImportFrom'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "FunctionDef'", "ImportFrom'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "FunctionDef'", "Return'" ]
[ "def FUNC_3(self, VAR_15: 'ComponentRegistry'):...\n", "self._registry = VAR_15\n" ]
[ "def initialize(self, registry: 'ComponentRegistry'):...\n", "self._registry = registry\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Assign'" ]
[ "def __init__(self, VAR_3=None, VAR_7=None):...\n", "VAR_28 = {'class': 'vTimeField', 'size': '8'}\n", "if VAR_3 is not None:\n", "VAR_28.update(VAR_3)\n", "super(CLASS_2, self).__init__(VAR_3=final_attrs, VAR_7=format)\n" ]
[ "def __init__(self, attrs=None, format=None):...\n", "final_attrs = {'class': 'vTimeField', 'size': '8'}\n", "if attrs is not None:\n", "final_attrs.update(attrs)\n", "super(AdminTimeWidget, self).__init__(attrs=final_attrs, format=format)\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Condition", "Expr'", "Expr'" ]
[ "@login_required()...\n", "VAR_71 = VAR_2.GET\n", "VAR_46 = FUNC_1(VAR_2, 'image')\n", "VAR_45 = FUNC_1(VAR_2, 'dataset')\n", "VAR_44 = FUNC_1(VAR_2, 'project')\n", "VAR_47 = FUNC_1(VAR_2, 'screen')\n", "VAR_48 = FUNC_1(VAR_2, 'plate')\n", "VAR_85 = FUNC_1(VAR_2, 'acquisition')\n", "VAR_86 = FUNC_1(VAR_2, 'well')\n", "VAR_87 = FUNC_0(VAR_2, 'page', 1)\n", "VAR_88 = FUNC_0(VAR_2, 'limit', VAR_1)\n", "VAR_89 = VAR_71.get('type', None)\n", "VAR_90 = VAR_71.get('ns', None)\n", "VAR_91, VAR_92 = tree.marshal_annotations(VAR_5, VAR_44=project_ids, VAR_45\n =dataset_ids, VAR_46=image_ids, VAR_47=screen_ids, VAR_48=plate_ids,\n VAR_85=run_ids, VAR_86=well_ids, VAR_89=ann_type, VAR_90=ns, VAR_87=\n page, VAR_88=limit)\n", "return JsonResponse({'annotations': VAR_91, 'experimenters': VAR_92})\n" ]
[ "@login_required()...\n", "r = request.GET\n", "image_ids = get_list(request, 'image')\n", "dataset_ids = get_list(request, 'dataset')\n", "project_ids = get_list(request, 'project')\n", "screen_ids = get_list(request, 'screen')\n", "plate_ids = get_list(request, 'plate')\n", "run_ids = get_list(request, 'acquisition')\n", "well_ids = get_list(request, 'well')\n", "page = get_long_or_default(request, 'page', 1)\n", "limit = get_long_or_default(request, 'limit', ANNOTATIONS_LIMIT)\n", "ann_type = r.get('type', None)\n", "ns = r.get('ns', None)\n", "anns, exps = tree.marshal_annotations(conn, project_ids=project_ids,\n dataset_ids=dataset_ids, image_ids=image_ids, screen_ids=screen_ids,\n plate_ids=plate_ids, run_ids=run_ids, well_ids=well_ids, ann_type=\n ann_type, ns=ns, page=page, limit=limit)\n", "return JsonResponse({'annotations': anns, 'experimenters': exps})\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_6(self) ->None:...\n", "if streamlit.server.routes.allow_cross_origin_requests():\n", "self.set_header('Access-Control-Allow-Origin', '*')\n" ]
[ "def set_default_headers(self) ->None:...\n", "if streamlit.server.routes.allow_cross_origin_requests():\n", "self.set_header('Access-Control-Allow-Origin', '*')\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Expr'" ]
[ "@override_settings(LOGIN_URL='https:///login/')...\n", "VAR_25 = urlquote('http://testserver/login_required/')\n", "VAR_6 = 'https:///login/?next=%s' % VAR_25\n", "self.assertLoginURLEquals(VAR_6)\n" ]
[ "@override_settings(LOGIN_URL='https:///login/')...\n", "quoted_next = urlquote('http://testserver/login_required/')\n", "expected = 'https:///login/?next=%s' % quoted_next\n", "self.assertLoginURLEquals(expected)\n" ]
[ 0, 0, 0, 0 ]
[ "Condition", "Assign'", "Assign'", "Expr'" ]
[ "def FUNC_5(self, VAR_24, VAR_25=None):...\n", "if VAR_25:\n", "return b''.join(list(self.serialize(VAR_24, VAR_25)))\n", "return ''.join(list(self.serialize(VAR_24)))\n" ]
[ "def render(self, treewalker, encoding=None):...\n", "if encoding:\n", "return b''.join(list(self.serialize(treewalker, encoding)))\n", "return ''.join(list(self.serialize(treewalker)))\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Return'", "Return'" ]
[ "def FUNC_7(self, VAR_16, VAR_17, VAR_5):...\n", "VAR_6 = VAR_16.get(VAR_5)\n", "if VAR_6:\n", "return VAR_6.split(',')\n" ]
[ "def value_from_datadict(self, data, files, name):...\n", "value = data.get(name)\n", "if value:\n", "return value.split(',')\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Condition", "Return'" ]
[ "def FUNC_59(self):...\n", "\"\"\"docstring\"\"\"\n", "self.login()\n", "VAR_3 = self.client.get('/logout/custom_query/?follow=/somewhere/')\n", "self.assertEqual(VAR_3.status_code, 302)\n", "self.assertURLEqual(VAR_3.url, '/somewhere/')\n", "self.confirm_logged_out()\n" ]
[ "def test_logout_with_custom_redirect_argument(self):...\n", "\"\"\"docstring\"\"\"\n", "self.login()\n", "response = self.client.get('/logout/custom_query/?follow=/somewhere/')\n", "self.assertEqual(response.status_code, 302)\n", "self.assertURLEqual(response.url, '/somewhere/')\n", "self.confirm_logged_out()\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Expr'", "Assign'", "Expr'", "Expr'", "Expr'" ]
[ "def FUNC_141(self, VAR_212):...\n", "if not self.can_read(VAR_212):\n", "VAR_407 = (VAR_212.body or '').replace('\\r', '')\n", "return ''\n", "VAR_408 = [VAR_446 for VAR_446 in VAR_407.split('\\n\\n') if not VAR_446.\n startswith('#') and VAR_446.strip()]\n", "if VAR_408:\n", "return VAR_408[0]\n" ]
[ "def first_paragraph(self, page):...\n", "if not self.can_read(page):\n", "mm = (page.body or '').replace('\\r', '')\n", "return ''\n", "ps = [p for p in mm.split('\\n\\n') if not p.startswith('#') and p.strip()]\n", "if ps:\n", "return ps[0]\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Assign'", "Return'", "Assign'", "Condition", "Return'" ]
[ "@staticmethod...\n", "VAR_1 = AnEnum(VAR_0['an_enum_value'])\n", "def FUNC_2(VAR_4: Dict[str, Any]) ->Union[datetime, date]:...\n", "VAR_5: Union[datetime, date]\n", "VAR_5 = datetime.fromisoformat(VAR_0['aCamelDateTime'])\n", "VAR_5 = date.fromisoformat(VAR_0['aCamelDateTime'])\n", "return VAR_5\n", "return VAR_5\n" ]
[ "@staticmethod...\n", "an_enum_value = AnEnum(d['an_enum_value'])\n", "def _parse_a_camel_date_time(data: Dict[str, Any]) ->Union[datetime, date]:...\n", "a_camel_date_time: Union[datetime, date]\n", "a_camel_date_time = datetime.fromisoformat(d['aCamelDateTime'])\n", "a_camel_date_time = date.fromisoformat(d['aCamelDateTime'])\n", "return a_camel_date_time\n", "return a_camel_date_time\n" ]
[ 0, 0, 0, 5, 5, 5, 0, 0 ]
[ "Condition", "Assign'", "FunctionDef'", "AnnAssign'", "Assign'", "Assign'", "Return'", "Return'" ]
[ "def FUNC_8():...\n", "VAR_6 = Token(TokenType.TEXT,\n \"unicorn 'tests.templatetags.test_unicorn_render.FakeComponentModel' model_id=model.id\"\n )\n", "VAR_7 = unicorn(None, VAR_6)\n", "VAR_8 = {'model': {'pk': 123}}\n", "VAR_9 = VAR_7.render(VAR_8)\n", "assert '==123==' in VAR_9\n" ]
[ "def test_unicorn_render_id_use_pk():...\n", "token = Token(TokenType.TEXT,\n \"unicorn 'tests.templatetags.test_unicorn_render.FakeComponentModel' model_id=model.id\"\n )\n", "unicorn_node = unicorn(None, token)\n", "context = {'model': {'pk': 123}}\n", "actual = unicorn_node.render(context)\n", "assert '==123==' in actual\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Assert'" ]
[ "def FUNC_70(self):...\n", "VAR_76 = getattr(self, '_liked_by', None)\n", "if VAR_76:\n", "return json.loads(VAR_76)\n", "return []\n" ]
[ "def get_liked_by(self):...\n", "liked_by = getattr(self, '_liked_by', None)\n", "if liked_by:\n", "return json.loads(liked_by)\n", "return []\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Condition", "Return'", "Return'" ]
[ "def FUNC_6(self):...\n", "VAR_5 = self._makeContext()\n", "self.assertEqual(VAR_5.evaluate('d/one'), 1)\n" ]
[ "def test_evaluate_simple_path_dict_key_int_value(self):...\n", "ec = self._makeContext()\n", "self.assertEqual(ec.evaluate('d/one'), 1)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Expr'" ]
[ "async def FUNC_6(self):...\n", "if not hasattr(self, 'resolved_ref'):\n", "self.resolved_ref = await self.get_resolved_ref()\n", "return f'{self.quoted_namespace}/{self.resolved_ref}'\n" ]
[ "async def get_resolved_spec(self):...\n", "if not hasattr(self, 'resolved_ref'):\n", "self.resolved_ref = await self.get_resolved_ref()\n", "return f'{self.quoted_namespace}/{self.resolved_ref}'\n" ]
[ 0, 0, 0, 0 ]
[ "AsyncFunctionDef'", "Condition", "Assign'", "Return'" ]
[ "def FUNC_4(self):...\n", "assert controller_name(CLASS_11) == 'more_than_one_word'\n" ]
[ "def test_it_works_with_more_than_one_word(self):...\n", "assert controller_name(MoreThanOneWordController) == 'more_than_one_word'\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Assert'" ]
[ "def FUNC_11(self, VAR_11, VAR_9, VAR_12, VAR_10, **VAR_4):...\n", "self.layout.move_cell_to_position(VAR_11, VAR_9, VAR_12, VAR_10)\n", "self.save_layout()\n" ]
[ "def dispatch_move_cell_to_position(self, from_x, from_y, to_x, to_y, **kwargs):...\n", "self.layout.move_cell_to_position(from_x, from_y, to_x, to_y)\n", "self.save_layout()\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Expr'" ]
[ "def FUNC_8(self, VAR_6):...\n", "VAR_13 = UploadForm(data={'file-clear': '1'}, instance=filemodel)\n", "assert VAR_13.is_valid()\n", "assert not VAR_13.cleaned_data['file']\n" ]
[ "def test_clear(self, filemodel):...\n", "form = UploadForm(data={'file-clear': '1'}, instance=filemodel)\n", "assert form.is_valid()\n", "assert not form.cleaned_data['file']\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assert'", "Assert'" ]
[ "def FUNC_3(self, VAR_13):...\n", "return VAR_13\n" ]
[ "def to_python(self, value):...\n", "return value\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_121(*VAR_79, **VAR_42):...\n", "if VAR_13.read_from_replica:\n", "FUNC_6()\n", "VAR_225 = VAR_129(*VAR_79, **get_newargs(fn, kwargs))\n", "if VAR_1 and hasattr(VAR_1, 'primary_db'):\n", "return VAR_225\n", "VAR_1.db.close()\n", "VAR_1.db = VAR_1.primary_db\n" ]
[ "def wrapper_fn(*args, **kwargs):...\n", "if conf.read_from_replica:\n", "connect_replica()\n", "retval = fn(*args, **get_newargs(fn, kwargs))\n", "if local and hasattr(local, 'primary_db'):\n", "return retval\n", "local.db.close()\n", "local.db = local.primary_db\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Expr'", "Assign'", "Condition", "Return'", "Expr'", "Assign'" ]
[ "@login_required()...\n", "\"\"\"docstring\"\"\"\n", "VAR_96 = []\n", "VAR_97 = set()\n", "VAR_98 = []\n", "VAR_68 = {}\n", "VAR_77 = {}\n", "for VAR_215 in ('Project', 'Dataset', 'Image', 'Screen', 'Plate'):\n", "VAR_154 = VAR_2.GET.get(VAR_215, None)\n", "VAR_96 = list(set(VAR_96))\n", "if VAR_154 is not None:\n", "if len(VAR_96) == 0:\n", "for o in VAR_5.getObjects(VAR_215, VAR_154.split(',')):\n", "VAR_96 = [VAR_5.getUserId()]\n", "for VAR_103 in VAR_5.getObjects('Experimenter', VAR_96, opts={\n", "VAR_96.append(o.getDetails().owner.id.val)\n", "VAR_271 = []\n", "VAR_99 = set.intersection(*VAR_98)\n", "VAR_97.add(o.getDetails().group.id.val)\n", "VAR_77[VAR_103.id] = VAR_103.getFullName()\n", "VAR_100 = VAR_5.getAdminService().getSecurityRoles().userGroupId\n", "for VAR_101 in VAR_103.copyGroupExperimenterMap():\n", "if VAR_100 in VAR_99:\n", "VAR_68[VAR_101.parent.id.val] = VAR_101.parent\n", "VAR_98.append(set(VAR_271))\n", "VAR_99.remove(VAR_100)\n", "if len(VAR_97) == 1:\n", "VAR_271.append(VAR_101.parent.id.val)\n", "VAR_272 = VAR_97.pop()\n", "def FUNC_83(VAR_101):...\n", "if VAR_272 in VAR_99:\n", "VAR_256 = VAR_101.getDetails().permissions\n", "VAR_99.remove(VAR_272)\n", "return {'write': VAR_256.isGroupWrite(), 'annotate': VAR_256.\n isGroupAnnotate(), 'read': VAR_256.isGroupRead()}\n" ]
[ "@login_required()...\n", "\"\"\"docstring\"\"\"\n", "ownerIds = []\n", "currentGroups = set()\n", "groupSets = []\n", "groups = {}\n", "owners = {}\n", "for dtype in ('Project', 'Dataset', 'Image', 'Screen', 'Plate'):\n", "oids = request.GET.get(dtype, None)\n", "ownerIds = list(set(ownerIds))\n", "if oids is not None:\n", "if len(ownerIds) == 0:\n", "for o in conn.getObjects(dtype, oids.split(',')):\n", "ownerIds = [conn.getUserId()]\n", "for owner in conn.getObjects('Experimenter', ownerIds, opts={\n", "ownerIds.append(o.getDetails().owner.id.val)\n", "gids = []\n", "targetGroupIds = set.intersection(*groupSets)\n", "currentGroups.add(o.getDetails().group.id.val)\n", "owners[owner.id] = owner.getFullName()\n", "userGroupId = conn.getAdminService().getSecurityRoles().userGroupId\n", "for group in owner.copyGroupExperimenterMap():\n", "if userGroupId in targetGroupIds:\n", "groups[group.parent.id.val] = group.parent\n", "groupSets.append(set(gids))\n", "targetGroupIds.remove(userGroupId)\n", "if len(currentGroups) == 1:\n", "gids.append(group.parent.id.val)\n", "curr_grp = currentGroups.pop()\n", "def getPerms(group):...\n", "if curr_grp in targetGroupIds:\n", "p = group.getDetails().permissions\n", "targetGroupIds.remove(curr_grp)\n", "return {'write': p.isGroupWrite(), 'annotate': p.isGroupAnnotate(), 'read':\n p.isGroupRead()}\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "For", "Assign'", "Assign'", "Condition", "Condition", "For", "Assign'", "For", "Expr'", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "For", "Condition", "Assign'", "Expr'", "Expr'", "Condition", "Expr'", "Assign'", "FunctionDef'", "Condition", "Assign'", "Expr'", "Return'" ]
[ "def __init__(self, VAR_5, VAR_4, *VAR_6, **VAR_7):...\n", "if VAR_7['initial'] is None:\n", "VAR_7['initial'] = {}\n", "VAR_7['initial']['terminology'] = True\n", "super().__init__(VAR_5, VAR_4, *VAR_6, **kwargs)\n" ]
[ "def __init__(self, translation, user, *args, **kwargs):...\n", "if kwargs['initial'] is None:\n", "kwargs['initial'] = {}\n", "kwargs['initial']['terminology'] = True\n", "super().__init__(translation, user, *args, **kwargs)\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Assign'", "Assign'", "Expr'" ]
[ "async def FUNC_4(self, VAR_5: str, VAR_1: EventBase, VAR_2: Optional[...\n", "\"\"\"docstring\"\"\"\n", "VAR_11 = VAR_1.room_id\n", "VAR_12 = VAR_1.event_id\n", "VAR_0.debug('[%s %s] Processing event: %s', VAR_11, VAR_12, VAR_1)\n", "await self._handle_new_event(VAR_5, VAR_1, VAR_2=state)\n", "if VAR_1.type == VAR_188.Encrypted:\n", "VAR_155 = VAR_1.content.get('device_id')\n", "VAR_156 = VAR_1.content.get('sender_key')\n", "VAR_157 = await self.store.get_cached_devices_for_user(VAR_1.sender)\n", "VAR_158 = False\n", "VAR_159 = None\n", "if VAR_155 is not None:\n", "VAR_159 = VAR_157.get(VAR_155)\n", "if VAR_156 is not None:\n", "if VAR_159 is None:\n", "VAR_186 = []\n", "if VAR_158:\n", "VAR_0.info('Received event from remote device not in our cache: %s %s',\n VAR_1.sender, VAR_155)\n", "if VAR_159:\n", "run_as_background_process('resync_device_due_to_pdu', self._resync_device,\n VAR_1.sender)\n", "VAR_158 = True\n", "VAR_195 = VAR_159.get('keys', {}).get('keys', {})\n", "if VAR_155:\n", "if VAR_1.content.get('algorithm'\n", "VAR_186 = [VAR_144 for VAR_159 in VAR_157.values() for VAR_144 in VAR_159.\n get('keys', {}).get('keys', {}).values()]\n", "if VAR_156 not in VAR_186:\n", "VAR_199 = 'curve25519:%s' % (VAR_155,)\n", "VAR_186 = VAR_195.values()\n", "VAR_0.info(\n 'Received event from remote device with unexpected sender key: %s %s: %s',\n VAR_1.sender, VAR_155 or '<no device_id>', VAR_156)\n", "VAR_186 = [VAR_195.get(VAR_199)]\n", "VAR_158 = True\n" ]
[ "async def _process_received_pdu(self, origin: str, event: EventBase, state:...\n", "\"\"\"docstring\"\"\"\n", "room_id = event.room_id\n", "event_id = event.event_id\n", "logger.debug('[%s %s] Processing event: %s', room_id, event_id, event)\n", "await self._handle_new_event(origin, event, state=state)\n", "if event.type == EventTypes.Encrypted:\n", "device_id = event.content.get('device_id')\n", "sender_key = event.content.get('sender_key')\n", "cached_devices = await self.store.get_cached_devices_for_user(event.sender)\n", "resync = False\n", "device = None\n", "if device_id is not None:\n", "device = cached_devices.get(device_id)\n", "if sender_key is not None:\n", "if device is None:\n", "current_keys = []\n", "if resync:\n", "logger.info('Received event from remote device not in our cache: %s %s',\n event.sender, device_id)\n", "if device:\n", "run_as_background_process('resync_device_due_to_pdu', self._resync_device,\n event.sender)\n", "resync = True\n", "keys = device.get('keys', {}).get('keys', {})\n", "if device_id:\n", "if event.content.get('algorithm'\n", "current_keys = [key for device in cached_devices.values() for key in device\n .get('keys', {}).get('keys', {}).values()]\n", "if sender_key not in current_keys:\n", "key_name = 'curve25519:%s' % (device_id,)\n", "current_keys = keys.values()\n", "logger.info(\n 'Received event from remote device with unexpected sender key: %s %s: %s',\n event.sender, device_id or '<no device_id>', sender_key)\n", "current_keys = [keys.get(key_name)]\n", "resync = True\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Assign'", "Expr'", "Expr'", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Condition", "Condition", "Assign'", "Condition", "Expr'", "Condition", "Expr'", "Assign'", "Assign'", "Condition", "Condition", "Assign'", "Condition", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'" ]
[ "@ensure_csrf_cookie...\n", "\"\"\"docstring\"\"\"\n", "VAR_10 = SlashSeparatedCourseKey.from_deprecated_string(VAR_10)\n", "VAR_120 = CourseRegistrationCode.objects.filter(VAR_10=course_id).order_by(\n 'invoice_item__invoice__company_name')\n", "VAR_88 = VAR_9.POST['active_company_name']\n", "if VAR_88:\n", "VAR_120 = VAR_120.filter(invoice_item__invoice__company_name=company_name)\n", "VAR_121 = RegistrationCodeRedemption.objects.select_related('registration_code'\n , 'registration_code__invoice_item__invoice').filter(\n registration_code__course_id=course_id)\n", "if VAR_121.exists():\n", "VAR_184 = [VAR_82.registration_code.code for VAR_82 in VAR_121]\n", "return FUNC_34('Active_Registration_Codes.csv', VAR_120)\n", "VAR_120 = VAR_120.exclude(code__in=redeemed_registration_codes)\n" ]
[ "@ensure_csrf_cookie...\n", "\"\"\"docstring\"\"\"\n", "course_id = SlashSeparatedCourseKey.from_deprecated_string(course_id)\n", "registration_codes_list = CourseRegistrationCode.objects.filter(course_id=\n course_id).order_by('invoice_item__invoice__company_name')\n", "company_name = request.POST['active_company_name']\n", "if company_name:\n", "registration_codes_list = registration_codes_list.filter(\n invoice_item__invoice__company_name=company_name)\n", "code_redemption_set = RegistrationCodeRedemption.objects.select_related(\n 'registration_code', 'registration_code__invoice_item__invoice').filter(\n registration_code__course_id=course_id)\n", "if code_redemption_set.exists():\n", "redeemed_registration_codes = [code.registration_code.code for code in\n code_redemption_set]\n", "return registration_codes_csv('Active_Registration_Codes.csv',\n registration_codes_list)\n", "registration_codes_list = registration_codes_list.exclude(code__in=\n redeemed_registration_codes)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Condition", "Assign'", "Return'", "Assign'" ]
[ "def FUNC_8(self):...\n", "self._record_users()\n", "self.get_success(self.handler.delete_device(VAR_0, 'abc'))\n", "self.get_failure(self.handler.get_device(VAR_0, 'abc'), synapse.api.errors.\n NotFoundError)\n" ]
[ "def test_delete_device(self):...\n", "self._record_users()\n", "self.get_success(self.handler.delete_device(user1, 'abc'))\n", "self.get_failure(self.handler.get_device(user1, 'abc'), synapse.api.errors.\n NotFoundError)\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Expr'", "Expr'" ]
[ "def FUNC_39(self):...\n", "\"\"\"docstring\"\"\"\n", "signal.signal(signal.SIGINT, self._handle_sigint)\n" ]
[ "def _restore_sigint_handler(self):...\n", "\"\"\"docstring\"\"\"\n", "signal.signal(signal.SIGINT, self._handle_sigint)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Expr'" ]
[ "@app.route('/folders/delete', methods=['POST'])...\n", "VAR_9 = forms.DeleteFolderForm()\n", "if VAR_9.validate_on_submit():\n", "if data.delete_dir(VAR_9.dir_name.data):\n", "flash('Could not delete folder.', 'error')\n", "flash('Folder successfully deleted.', 'success')\n", "flash('Folder not found.', 'error')\n", "return redirect(request.referrer or '/')\n", "return redirect('/')\n", "return redirect(request.referrer or '/', 404)\n" ]
[ "@app.route('/folders/delete', methods=['POST'])...\n", "form = forms.DeleteFolderForm()\n", "if form.validate_on_submit():\n", "if data.delete_dir(form.dir_name.data):\n", "flash('Could not delete folder.', 'error')\n", "flash('Folder successfully deleted.', 'success')\n", "flash('Folder not found.', 'error')\n", "return redirect(request.referrer or '/')\n", "return redirect('/')\n", "return redirect(request.referrer or '/', 404)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Assign'", "Condition", "Condition", "Expr'", "Expr'", "Expr'", "Return'", "Return'", "Return'" ]
[ "def FUNC_4(VAR_2, VAR_3, VAR_5, VAR_6=0):...\n", "\"\"\"docstring\"\"\"\n", "VAR_4 = saved_model_utils.get_meta_graph_def(VAR_2, VAR_3)\n", "VAR_24 = FUNC_2(VAR_4, VAR_5)\n", "VAR_25 = FUNC_3(VAR_4, VAR_5)\n", "VAR_26 = ' ' * VAR_6\n", "def FUNC_30(VAR_27):...\n", "print(VAR_26 + VAR_27)\n", "FUNC_30('The given SavedModel SignatureDef contains the following input(s):')\n", "for VAR_68, input_tensor in sorted(VAR_24.items()):\n", "FUNC_30(\" inputs['%s'] tensor_info:\" % VAR_68)\n", "FUNC_30('The given SavedModel SignatureDef contains the following output(s):')\n", "FUNC_7(input_tensor, VAR_6 + 1)\n", "for output_key, output_tensor in sorted(VAR_25.items()):\n", "FUNC_30(\" outputs['%s'] tensor_info:\" % output_key)\n", "FUNC_30('Method name is: %s' % VAR_4.signature_def[VAR_5].method_name)\n", "FUNC_7(output_tensor, VAR_6 + 1)\n" ]
[ "def _show_inputs_outputs(saved_model_dir, tag_set, signature_def_key, indent=0...\n", "\"\"\"docstring\"\"\"\n", "meta_graph_def = saved_model_utils.get_meta_graph_def(saved_model_dir, tag_set)\n", "inputs_tensor_info = _get_inputs_tensor_info_from_meta_graph_def(meta_graph_def\n , signature_def_key)\n", "outputs_tensor_info = _get_outputs_tensor_info_from_meta_graph_def(\n meta_graph_def, signature_def_key)\n", "indent_str = ' ' * indent\n", "def in_print(s):...\n", "print(indent_str + s)\n", "in_print('The given SavedModel SignatureDef contains the following input(s):')\n", "for input_key, input_tensor in sorted(inputs_tensor_info.items()):\n", "in_print(\" inputs['%s'] tensor_info:\" % input_key)\n", "in_print('The given SavedModel SignatureDef contains the following output(s):')\n", "_print_tensor_info(input_tensor, indent + 1)\n", "for output_key, output_tensor in sorted(outputs_tensor_info.items()):\n", "in_print(\" outputs['%s'] tensor_info:\" % output_key)\n", "in_print('Method name is: %s' % meta_graph_def.signature_def[\n signature_def_key].method_name)\n", "_print_tensor_info(output_tensor, indent + 1)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Assign'", "Assign'", "FunctionDef'", "Expr'", "Expr'", "For", "Expr'", "Expr'", "Expr'", "For", "Expr'", "Expr'", "Expr'" ]
[ "@wraps(VAR_2)...\n", "if current_user.role_upload():\n", "return VAR_2(*VAR_48, **kwargs)\n", "abort(403)\n" ]
[ "@wraps(f)...\n", "if current_user.role_upload():\n", "return f(*args, **kwargs)\n", "abort(403)\n" ]
[ 0, 0, 0, 0 ]
[ "Condition", "Condition", "Return'", "Expr'" ]
[ "def __str__(self):...\n", "if self.shared:\n", "return '%s (*)' % self.title\n", "return '%s' % self.title\n" ]
[ "def __str__(self):...\n", "if self.shared:\n", "return '%s (*)' % self.title\n", "return '%s' % self.title\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Return'", "Return'" ]
[ "\"\"\"string\"\"\"\n", "import copy\n", "import os\n", "import datetime\n", "import Ice\n", "from Ice import Exception as IceException\n", "import logging\n", "import traceback\n", "import json\n", "import re\n", "import sys\n", "import warnings\n", "from past.builtins import unicode\n", "from future.utils import bytes_to_native_str\n", "from time import time\n", "from omeroweb.version import omeroweb_buildyear as build_year\n", "from omeroweb.version import omeroweb_version as omero_version\n", "import omero\n", "import omero.scripts\n", "from omero.rtypes import wrap, unwrap, rlong, rlist\n", "from omero.gateway.utils import toBoolean\n", "from django.conf import settings\n", "from django.template import loader as template_loader\n", "from django.http import Http404, HttpResponse, HttpResponseRedirect, JsonResponse, HttpResponseForbidden\n", "from django.http import HttpResponseServerError, HttpResponseBadRequest\n", "from django.utils.http import urlencode\n", "from django.core.urlresolvers import reverse, NoReverseMatch\n", "from django.utils.encoding import smart_str\n", "from django.views.decorators.cache import never_cache\n", "from django.views.decorators.http import require_POST\n", "from django.shortcuts import render\n", "from omeroweb.webclient.webclient_utils import _formatReport, _purgeCallback\n", "from .forms import GlobalSearchForm, ContainerForm\n", "from .forms import ShareForm, BasketShareForm\n", "from .forms import ContainerNameForm, ContainerDescriptionForm\n", "from .forms import CommentAnnotationForm, TagsAnnotationForm\n", "from .forms import MetadataFilterForm, MetadataDetectorForm\n", "from .forms import MetadataChannelForm, MetadataEnvironmentForm\n", "from .forms import MetadataObjectiveForm, MetadataObjectiveSettingsForm\n", "from .forms import MetadataStageLabelForm, MetadataLightSourceForm\n", "from .forms import MetadataDichroicForm, MetadataMicroscopeForm\n", "from .forms import FilesAnnotationForm, WellIndexForm, NewTagsAnnotationFormSet\n", "from .controller.container import BaseContainer\n", "from .controller.history import BaseCalendar\n", "from .controller.search import BaseSearch\n", "from .controller.share import BaseShare\n", "from omeroweb.webadmin.forms import LoginForm\n", "from omeroweb.webgateway import views as webgateway_views\n", "from omeroweb.webgateway.marshal import graphResponseMarshal\n", "from omeroweb.webgateway.util import get_longs as webgateway_get_longs\n", "from omeroweb.feedback.views import handlerInternalError\n", "from omeroweb.webclient.decorators import login_required\n", "from omeroweb.webclient.decorators import render_response\n", "from omeroweb.webclient.show import Show, IncorrectMenuError, paths_to_object, paths_to_tag\n", "from omeroweb.decorators import ConnCleaningHttpResponse, parse_url, TableClosingHttpResponse\n", "from omeroweb.webgateway.util import getIntOrDefault\n", "from omero.model import AnnotationAnnotationLinkI, DatasetI, DatasetImageLinkI, ExperimenterI, ImageI, OriginalFileI, PlateI, ProjectI, ProjectDatasetLinkI, ScreenI, ScreenPlateLinkI, TagAnnotationI\n", "from omero import ApiUsageException, ServerError, CmdError\n", "from omeroweb.webgateway.views import LoginView\n", "from . import tree\n", "import long\n", "VAR_241 = int\n", "VAR_0 = logging.getLogger(__name__)\n", "VAR_0.info(\"INIT '%s'\" % os.getpid())\n", "VAR_1 = settings.PAGE * 100\n", "def FUNC_0(VAR_2, VAR_3, VAR_4):...\n", "\"\"\"docstring\"\"\"\n", "VAR_51 = None\n", "VAR_52 = VAR_2.GET.get(VAR_3, VAR_4)\n", "if VAR_52 is not None:\n", "VAR_51 = VAR_241(VAR_52)\n", "return VAR_51\n" ]
[ "\"\"\" A view functions is simply a Python function that takes a Web request and\nreturns a Web response. This response can be the HTML contents of a Web page,\nor a redirect, or the 404 and 500 error, or an XML document, or an image...\nor anything.\"\"\"\n", "import copy\n", "import os\n", "import datetime\n", "import Ice\n", "from Ice import Exception as IceException\n", "import logging\n", "import traceback\n", "import json\n", "import re\n", "import sys\n", "import warnings\n", "from past.builtins import unicode\n", "from future.utils import bytes_to_native_str\n", "from time import time\n", "from omeroweb.version import omeroweb_buildyear as build_year\n", "from omeroweb.version import omeroweb_version as omero_version\n", "import omero\n", "import omero.scripts\n", "from omero.rtypes import wrap, unwrap, rlong, rlist\n", "from omero.gateway.utils import toBoolean\n", "from django.conf import settings\n", "from django.template import loader as template_loader\n", "from django.http import Http404, HttpResponse, HttpResponseRedirect, JsonResponse, HttpResponseForbidden\n", "from django.http import HttpResponseServerError, HttpResponseBadRequest\n", "from django.utils.http import urlencode\n", "from django.core.urlresolvers import reverse, NoReverseMatch\n", "from django.utils.encoding import smart_str\n", "from django.views.decorators.cache import never_cache\n", "from django.views.decorators.http import require_POST\n", "from django.shortcuts import render\n", "from omeroweb.webclient.webclient_utils import _formatReport, _purgeCallback\n", "from .forms import GlobalSearchForm, ContainerForm\n", "from .forms import ShareForm, BasketShareForm\n", "from .forms import ContainerNameForm, ContainerDescriptionForm\n", "from .forms import CommentAnnotationForm, TagsAnnotationForm\n", "from .forms import MetadataFilterForm, MetadataDetectorForm\n", "from .forms import MetadataChannelForm, MetadataEnvironmentForm\n", "from .forms import MetadataObjectiveForm, MetadataObjectiveSettingsForm\n", "from .forms import MetadataStageLabelForm, MetadataLightSourceForm\n", "from .forms import MetadataDichroicForm, MetadataMicroscopeForm\n", "from .forms import FilesAnnotationForm, WellIndexForm, NewTagsAnnotationFormSet\n", "from .controller.container import BaseContainer\n", "from .controller.history import BaseCalendar\n", "from .controller.search import BaseSearch\n", "from .controller.share import BaseShare\n", "from omeroweb.webadmin.forms import LoginForm\n", "from omeroweb.webgateway import views as webgateway_views\n", "from omeroweb.webgateway.marshal import graphResponseMarshal\n", "from omeroweb.webgateway.util import get_longs as webgateway_get_longs\n", "from omeroweb.feedback.views import handlerInternalError\n", "from omeroweb.webclient.decorators import login_required\n", "from omeroweb.webclient.decorators import render_response\n", "from omeroweb.webclient.show import Show, IncorrectMenuError, paths_to_object, paths_to_tag\n", "from omeroweb.decorators import ConnCleaningHttpResponse, parse_url, TableClosingHttpResponse\n", "from omeroweb.webgateway.util import getIntOrDefault\n", "from omero.model import AnnotationAnnotationLinkI, DatasetI, DatasetImageLinkI, ExperimenterI, ImageI, OriginalFileI, PlateI, ProjectI, ProjectDatasetLinkI, ScreenI, ScreenPlateLinkI, TagAnnotationI\n", "from omero import ApiUsageException, ServerError, CmdError\n", "from omeroweb.webgateway.views import LoginView\n", "from . import tree\n", "import long\n", "long = int\n", "logger = logging.getLogger(__name__)\n", "logger.info(\"INIT '%s'\" % os.getpid())\n", "ANNOTATIONS_LIMIT = settings.PAGE * 100\n", "def get_long_or_default(request, name, default):...\n", "\"\"\"docstring\"\"\"\n", "val = None\n", "val_raw = request.GET.get(name, default)\n", "if val_raw is not None:\n", "val = long(val_raw)\n", "return val\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Expr'", "Import'", "Import'", "Import'", "Import'", "ImportFrom'", "Import'", "Import'", "Import'", "Import'", "Import'", "Import'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Import'", "Import'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Import'", "Assign'", "Assign'", "Expr'", "Assign'", "FunctionDef'", "Docstring", "Assign'", "Assign'", "Condition", "Assign'", "Return'" ]
[ "async def FUNC_8(self, VAR_16):...\n", "\"\"\"docstring\"\"\"\n", "async def FUNC_15(VAR_20):...\n", "VAR_58 = await self.get_server_verify_key_v2_indirect(VAR_16, VAR_20)\n", "VAR_0.warning('Key lookup failed from %r: %s', VAR_20.server_name, e)\n", "return {}\n", "return VAR_58\n", "VAR_0.exception('Unable to get key from %r: %s %s', VAR_20.server_name,\n type(e).__name__, str(e))\n" ]
[ "async def get_keys(self, keys_to_fetch):...\n", "\"\"\"docstring\"\"\"\n", "async def get_key(key_server):...\n", "result = await self.get_server_verify_key_v2_indirect(keys_to_fetch, key_server\n )\n", "logger.warning('Key lookup failed from %r: %s', key_server.server_name, e)\n", "return {}\n", "return result\n", "logger.exception('Unable to get key from %r: %s %s', key_server.server_name,\n type(e).__name__, str(e))\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "AsyncFunctionDef'", "Docstring", "AsyncFunctionDef'", "Assign'", "Expr'", "Return'", "Return'", "Expr'" ]
[ "def FUNC_34(VAR_32, VAR_33, VAR_34=None):...\n", "\"\"\"docstring\"\"\"\n", "VAR_68 = ['code', 'redeem_code_url', 'course_id', 'company_name',\n 'created_by', 'redeemed_by', 'invoice_id', 'purchaser',\n 'customer_reference_number', 'internal_reference', 'is_valid']\n", "VAR_85 = instructor_analytics.basic.course_registration_features(VAR_68,\n VAR_33, VAR_34)\n", "VAR_86, VAR_81 = instructor_analytics.csvs.format_dictlist(VAR_85, VAR_68)\n", "return instructor_analytics.csvs.create_csv_response(VAR_32, VAR_86, VAR_81)\n" ]
[ "def registration_codes_csv(file_name, codes_list, csv_type=None):...\n", "\"\"\"docstring\"\"\"\n", "query_features = ['code', 'redeem_code_url', 'course_id', 'company_name',\n 'created_by', 'redeemed_by', 'invoice_id', 'purchaser',\n 'customer_reference_number', 'internal_reference', 'is_valid']\n", "registration_codes = instructor_analytics.basic.course_registration_features(\n query_features, codes_list, csv_type)\n", "header, data_rows = instructor_analytics.csvs.format_dictlist(\n registration_codes, query_features)\n", "return instructor_analytics.csvs.create_csv_response(file_name, header,\n data_rows)\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_3(VAR_7, VAR_8):...\n", "" ]
[ "def check_user_session(user_id, session_key):...\n", "" ]
[ 0, 0 ]
[ "FunctionDef'", "Condition" ]
[ "@VAR_2.route('/register', methods=['GET', 'POST'])...\n", "if not config.config_public_reg:\n", "abort(404)\n", "if VAR_87 is not None and VAR_87.is_authenticated:\n", "return redirect(url_for('web.index'))\n", "if not config.get_mail_server_configured():\n", "flash(_(\n u'E-Mail server is not configured, please contact your administrator!'),\n category='error')\n", "if request.method == 'POST':\n", "return render_title_template('register.html', VAR_149=_('Register'), VAR_9=\n 'register')\n", "VAR_55 = request.form.to_dict()\n", "if VAR_0['oauth']:\n", "VAR_115 = VAR_55['email'].strip(\n ) if config.config_register_email else VAR_55.get('name')\n", "register_user_with_oauth()\n", "return render_title_template('register.html', config=config, VAR_149=_(\n 'Register'), VAR_9='register')\n", "if not VAR_115 or not VAR_55.get('email'):\n", "flash(_(u'Please fill out all fields!'), category='error')\n", "VAR_115 = check_username(VAR_115)\n", "flash(str(ex), category='error')\n", "VAR_116 = ub.User()\n", "return render_title_template('register.html', VAR_149=_('Register'), VAR_9=\n 'register')\n", "VAR_136 = check_email(VAR_55['email'])\n", "return render_title_template('register.html', VAR_149=_('Register'), VAR_9=\n 'register')\n", "if check_valid_domain(VAR_136):\n", "VAR_116.name = VAR_115\n", "flash(_(u'Your e-mail is not allowed to register'), category='error')\n", "VAR_116.email = VAR_136\n", "VAR_3.warning('Registering failed for user \"%s\" e-mail address: %s',\n VAR_115, VAR_55['email'])\n", "VAR_137 = generate_random_password()\n", "return render_title_template('register.html', VAR_149=_('Register'), VAR_9=\n 'register')\n", "VAR_116.password = generate_password_hash(VAR_137)\n", "VAR_116.role = config.config_default_role\n", "VAR_116.sidebar_view = config.config_default_show\n", "ub.session.add(VAR_116)\n", "ub.session.rollback()\n", "flash(_(u'Confirmation e-mail was send to your e-mail account.'), category=\n 'success')\n", "ub.session.commit()\n", "flash(_(u'An unknown error occurred. Please try again later.'), category=\n 'error')\n", "return redirect(url_for('web.login'))\n", "if VAR_0['oauth']:\n", "return render_title_template('register.html', VAR_149=_('Register'), VAR_9=\n 'register')\n", "register_user_with_oauth(VAR_116)\n", "send_registration_mail(VAR_55['email'].strip(), VAR_115, VAR_137)\n" ]
[ "@web.route('/register', methods=['GET', 'POST'])...\n", "if not config.config_public_reg:\n", "abort(404)\n", "if current_user is not None and current_user.is_authenticated:\n", "return redirect(url_for('web.index'))\n", "if not config.get_mail_server_configured():\n", "flash(_(\n u'E-Mail server is not configured, please contact your administrator!'),\n category='error')\n", "if request.method == 'POST':\n", "return render_title_template('register.html', title=_('Register'), page=\n 'register')\n", "to_save = request.form.to_dict()\n", "if feature_support['oauth']:\n", "nickname = to_save['email'].strip(\n ) if config.config_register_email else to_save.get('name')\n", "register_user_with_oauth()\n", "return render_title_template('register.html', config=config, title=_(\n 'Register'), page='register')\n", "if not nickname or not to_save.get('email'):\n", "flash(_(u'Please fill out all fields!'), category='error')\n", "nickname = check_username(nickname)\n", "flash(str(ex), category='error')\n", "content = ub.User()\n", "return render_title_template('register.html', title=_('Register'), page=\n 'register')\n", "email = check_email(to_save['email'])\n", "return render_title_template('register.html', title=_('Register'), page=\n 'register')\n", "if check_valid_domain(email):\n", "content.name = nickname\n", "flash(_(u'Your e-mail is not allowed to register'), category='error')\n", "content.email = email\n", "log.warning('Registering failed for user \"%s\" e-mail address: %s', nickname,\n to_save['email'])\n", "password = generate_random_password()\n", "return render_title_template('register.html', title=_('Register'), page=\n 'register')\n", "content.password = generate_password_hash(password)\n", "content.role = config.config_default_role\n", "content.sidebar_view = config.config_default_show\n", "ub.session.add(content)\n", "ub.session.rollback()\n", "flash(_(u'Confirmation e-mail was send to your e-mail account.'), category=\n 'success')\n", "ub.session.commit()\n", "flash(_(u'An unknown error occurred. Please try again later.'), category=\n 'error')\n", "return redirect(url_for('web.login'))\n", "if feature_support['oauth']:\n", "return render_title_template('register.html', title=_('Register'), page=\n 'register')\n", "register_user_with_oauth(content)\n", "send_registration_mail(to_save['email'].strip(), nickname, password)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Condition", "Expr'", "Condition", "Return'", "Condition", "Expr'", "Condition", "Return'", "Assign'", "Condition", "Assign'", "Expr'", "Return'", "Condition", "Expr'", "Assign'", "Expr'", "Assign'", "Return'", "Assign'", "Return'", "Condition", "Assign'", "Expr'", "Assign'", "Expr'", "Assign'", "Return'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Return'", "Condition", "Return'", "Expr'", "Expr'" ]
[ "def FUNC_6(self, VAR_1):...\n", "from openapi_python_client.parser import properties\n", "properties._existing_enums = VAR_1.MagicMock()\n", "assert properties.EnumProperty.get_all_enums() == properties._existing_enums\n", "properties._existing_enums = {}\n" ]
[ "def test_get_all_enums(self, mocker):...\n", "from openapi_python_client.parser import properties\n", "properties._existing_enums = mocker.MagicMock()\n", "assert properties.EnumProperty.get_all_enums() == properties._existing_enums\n", "properties._existing_enums = {}\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "ImportFrom'", "Assign'", "Assert'", "Assign'" ]
[ "def FUNC_114():...\n", "if VAR_32:\n", "if VAR_19.rollback_on_exception:\n", "VAR_12.rollback()\n", "import inspect\n", "if inspect.isclass(VAR_32) and issubclass(VAR_32, Exception):\n" ]
[ "def _raise_exception():...\n", "if raise_exception:\n", "if flags.rollback_on_exception:\n", "db.rollback()\n", "import inspect\n", "if inspect.isclass(raise_exception) and issubclass(raise_exception, Exception):\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Condition", "Expr'", "Import'", "Condition" ]
[ "def __init__(self, VAR_9, *VAR_6, **VAR_7):...\n", "super().__init__(*VAR_6, **kwargs)\n", "VAR_118 = Language.objects.filter(translation__component=component).exclude(pk\n =component.source_language_id)\n", "self.fields['lang'].choices = VAR_118.as_choices()\n" ]
[ "def __init__(self, component, *args, **kwargs):...\n", "super().__init__(*args, **kwargs)\n", "languages = Language.objects.filter(translation__component=component).exclude(\n pk=component.source_language_id)\n", "self.fields['lang'].choices = languages.as_choices()\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Assign'", "Assign'" ]
[ "@VAR_4.route('/quiz')...\n", "return render_template('quiz/quiz_container.html')\n" ]
[ "@app.route('/quiz')...\n", "return render_template('quiz/quiz_container.html')\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "@login_required(isAdmin=True)...\n", "\"\"\"docstring\"\"\"\n", "if VAR_2.method != 'POST':\n", "return {'template': 'webclient/scripts/upload_script.html'}\n", "VAR_236 = VAR_2.POST.get('script_path')\n", "VAR_237 = VAR_2.FILES['script_file']\n", "VAR_237.seek(0)\n", "VAR_238 = VAR_237.read().decode('utf-8')\n", "if not VAR_236.endswith('/'):\n", "VAR_236 = VAR_236 + '/'\n", "VAR_236 = VAR_236 + VAR_237.name\n", "VAR_202 = VAR_5.getScriptService()\n", "VAR_239 = VAR_202.getScriptID(VAR_236)\n", "if VAR_239 > 0:\n", "VAR_382 = VAR_345(ex)\n", "return {'Message': VAR_382, 'script_id': VAR_239}\n", "VAR_173 = OriginalFileI(VAR_239, False)\n", "VAR_239 = VAR_202.uploadOfficialScript(VAR_236, VAR_238)\n", "VAR_202.editScript(VAR_173, VAR_238)\n", "VAR_382 = 'Script Uploaded: %s' % VAR_237.name\n", "VAR_382 = 'Script Replaced: %s' % VAR_237.name\n" ]
[ "@login_required(isAdmin=True)...\n", "\"\"\"docstring\"\"\"\n", "if request.method != 'POST':\n", "return {'template': 'webclient/scripts/upload_script.html'}\n", "script_path = request.POST.get('script_path')\n", "script_file = request.FILES['script_file']\n", "script_file.seek(0)\n", "script_text = script_file.read().decode('utf-8')\n", "if not script_path.endswith('/'):\n", "script_path = script_path + '/'\n", "script_path = script_path + script_file.name\n", "scriptService = conn.getScriptService()\n", "script_id = scriptService.getScriptID(script_path)\n", "if script_id > 0:\n", "message = str(ex)\n", "return {'Message': message, 'script_id': script_id}\n", "orig_file = OriginalFileI(script_id, False)\n", "script_id = scriptService.uploadOfficialScript(script_path, script_text)\n", "scriptService.editScript(orig_file, script_text)\n", "message = 'Script Uploaded: %s' % script_file.name\n", "message = 'Script Replaced: %s' % script_file.name\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Condition", "Return'", "Assign'", "Assign'", "Expr'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Return'", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'" ]
[ "\"\"\"A cleanup tool for HTML.\n\nRemoves unwanted tags and content. See the `Cleaner` class for\ndetails.\n\"\"\"\n", "from __future__ import absolute_import\n", "import copy\n", "import re\n", "import sys\n", "from urlparse import urlsplit\n", "from urllib.parse import urlsplit, unquote_plus\n", "from lxml import etree\n", "from urllib import unquote_plus\n", "from lxml.html import defs\n", "from lxml.html import fromstring, XHTML_NAMESPACE\n", "from lxml.html import xhtml_to_html, _transform_result\n", "VAR_70\n", "VAR_70 = chr\n", "VAR_71\n", "VAR_71 = str\n", "VAR_72\n", "VAR_72 = str, bytes\n", "__all__ = ['clean_html', 'clean', 'Cleaner', 'autolink', 'autolink_html',\n 'word_break', 'word_break_html']\n", "VAR_0 = re.compile('expression\\\\s*\\\\(.*?\\\\)', re.S | re.I).sub\n", "VAR_1 = re.compile('@\\\\s*import', re.I).sub\n", "VAR_2 = re.compile('</?[a-zA-Z]+|\\\\son[a-zA-Z]+\\\\s*=', *((re.ASCII,) if sys\n .version_info[0] >= 3 else ())).search\n", "VAR_3 = re.compile('^data:image/.+;base64', re.I).search\n", "VAR_4 = re.compile(\n '(?:javascript|jscript|livescript|vbscript|data|about|mocha):', re.I\n ).search\n", "def FUNC_0(VAR_5):...\n", "if VAR_3(VAR_5):\n", "return None\n", "return VAR_4(VAR_5)\n" ]
[ "\"\"\"A cleanup tool for HTML.\n\nRemoves unwanted tags and content. See the `Cleaner` class for\ndetails.\n\"\"\"\n", "from __future__ import absolute_import\n", "import copy\n", "import re\n", "import sys\n", "from urlparse import urlsplit\n", "from urllib.parse import urlsplit, unquote_plus\n", "from lxml import etree\n", "from urllib import unquote_plus\n", "from lxml.html import defs\n", "from lxml.html import fromstring, XHTML_NAMESPACE\n", "from lxml.html import xhtml_to_html, _transform_result\n", "unichr\n", "unichr = chr\n", "unicode\n", "unicode = str\n", "basestring\n", "basestring = str, bytes\n", "__all__ = ['clean_html', 'clean', 'Cleaner', 'autolink', 'autolink_html',\n 'word_break', 'word_break_html']\n", "_replace_css_javascript = re.compile('expression\\\\s*\\\\(.*?\\\\)', re.S | re.I\n ).sub\n", "_replace_css_import = re.compile('@\\\\s*import', re.I).sub\n", "_looks_like_tag_content = re.compile('</?[a-zA-Z]+|\\\\son[a-zA-Z]+\\\\s*=', *(\n (re.ASCII,) if sys.version_info[0] >= 3 else ())).search\n", "_is_image_dataurl = re.compile('^data:image/.+;base64', re.I).search\n", "_is_possibly_malicious_scheme = re.compile(\n '(?:javascript|jscript|livescript|vbscript|data|about|mocha):', re.I\n ).search\n", "def _is_javascript_scheme(s):...\n", "if _is_image_dataurl(s):\n", "return None\n", "return _is_possibly_malicious_scheme(s)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2, 2, 0, 2, 2, 2 ]
[ "Expr'", "ImportFrom'", "Import'", "Import'", "Import'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Expr'", "Assign'", "Expr'", "Assign'", "Expr'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "FunctionDef'", "Condition", "Return'", "Return'" ]
[ "def FUNC_97(VAR_62):...\n", "def FUNC_120():...\n", "VAR_214 = VAR_1.db.get_value('DocType', VAR_62, 'module')\n", "return VAR_1.module_app[FUNC_56(VAR_214)]\n" ]
[ "def get_doctype_app(doctype):...\n", "def _get_doctype_app():...\n", "doctype_module = local.db.get_value('DocType', doctype, 'module')\n", "return local.module_app[scrub(doctype_module)]\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "FunctionDef'", "Assign'", "Return'" ]
[ "def FUNC_9(self):...\n", "def __call__(self):...\n", "VAR_6 = CLASS_5()\n", "VAR_5 = self._makeContext(VAR_0={'dummy': dummy})\n", "self.assertIs(VAR_5.evaluate('dummy'), VAR_6)\n" ]
[ "def test_evaluate_with_unimplemented_call(self):...\n", "def __call__(self):...\n", "dummy = Dummy()\n", "ec = self._makeContext(bindings={'dummy': dummy})\n", "self.assertIs(ec.evaluate('dummy'), dummy)\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "FunctionDef'", "Assign'", "Assign'", "Expr'" ]
[ "async def FUNC_36(self, VAR_40: Iterable[EventBase], VAR_41: Iterable[...\n", "\"\"\"docstring\"\"\"\n", "VAR_0.debug('construct_auth_difference Start!')\n", "def FUNC_50(VAR_21):...\n", "return VAR_21.depth, VAR_21.event_id\n" ]
[ "async def construct_auth_difference(self, local_auth: Iterable[EventBase],...\n", "\"\"\"docstring\"\"\"\n", "logger.debug('construct_auth_difference Start!')\n", "def sort_fun(ev):...\n", "return ev.depth, ev.event_id\n" ]
[ 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Expr'", "FunctionDef'", "Return'" ]
[ "@CLASS_4('log')...\n", "\"\"\"docstring\"\"\"\n", "if log.ram_handler is None:\n", "VAR_38 = None\n", "VAR_37 = QUrlQuery(VAR_3).queryItemValue('level')\n", "VAR_20 = jinja.render('log.html', title='log', content=html_log)\n", "if not VAR_37:\n", "return 'text/html', VAR_20\n", "VAR_37 = 'vdebug'\n", "VAR_38 = log.ram_handler.dump_log(VAR_20=True, VAR_37=level)\n" ]
[ "@add_handler('log')...\n", "\"\"\"docstring\"\"\"\n", "if log.ram_handler is None:\n", "html_log = None\n", "level = QUrlQuery(url).queryItemValue('level')\n", "html = jinja.render('log.html', title='log', content=html_log)\n", "if not level:\n", "return 'text/html', html\n", "level = 'vdebug'\n", "html_log = log.ram_handler.dump_log(html=True, level=level)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Condition", "Assign'", "Assign'", "Assign'", "Condition", "Return'", "Assign'", "Assign'" ]
[ "def FUNC_84(*VAR_79, **VAR_42):...\n", "\"\"\"docstring\"\"\"\n", "return VAR_12.get_value(*VAR_79, **kwargs)\n" ]
[ "def get_value(*args, **kwargs):...\n", "\"\"\"docstring\"\"\"\n", "return db.get_value(*args, **kwargs)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Return'" ]
[ "def FUNC_33(VAR_30):...\n", "print(VAR_29 + VAR_30)\n" ]
[ "def in_print(s):...\n", "print(indent_str + s)\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Expr'" ]
[ "def FUNC_29(self):...\n", "self.called = True\n" ]
[ "def to_html(self):...\n", "self.called = True\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Assign'" ]
[ "def FUNC_2(VAR_6, VAR_7):...\n", "\"\"\"docstring\"\"\"\n", "if hasattr(VAR_6, '__render_with_namespace__'):\n", "VAR_6 = ZRPythonExpr.call_with_ns(VAR_6.__render_with_namespace__, VAR_7)\n", "VAR_40 = aq_base(VAR_6)\n", "return VAR_6\n", "VAR_40 = removeAllProxies(VAR_40)\n", "if callable(VAR_40):\n", "if getattr(VAR_40, 'isDocTemp', 0):\n", "VAR_6 = ZRPythonExpr.call_with_ns(VAR_6, VAR_7, 2)\n", "VAR_6 = VAR_6()\n" ]
[ "def render(ob, ns):...\n", "\"\"\"docstring\"\"\"\n", "if hasattr(ob, '__render_with_namespace__'):\n", "ob = ZRPythonExpr.call_with_ns(ob.__render_with_namespace__, ns)\n", "base = aq_base(ob)\n", "return ob\n", "base = removeAllProxies(base)\n", "if callable(base):\n", "if getattr(base, 'isDocTemp', 0):\n", "ob = ZRPythonExpr.call_with_ns(ob, ns, 2)\n", "ob = ob()\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Assign'", "Assign'", "Return'", "Assign'", "Condition", "Condition", "Assign'", "Assign'" ]
[ "def FUNC_1(self, VAR_2, VAR_3, VAR_4):...\n", "self.room_id = self.helper.create_room_as(self.user_id)\n" ]
[ "def prepare(self, reactor, clock, hs):...\n", "self.room_id = self.helper.create_room_as(self.user_id)\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Assign'" ]
[ "def FUNC_2(VAR_4, VAR_5, VAR_6):...\n", "\"\"\"docstring\"\"\"\n", "for VAR_32, node in enumerate(VAR_4):\n", "if VAR_32 % 2 == 0:\n", "return VAR_4\n", "VAR_4[VAR_32:VAR_32 + 1] = FUNC_3(VAR_5, node, VAR_6)\n" ]
[ "def processTextNodes(nodes, regex, replacer):...\n", "\"\"\"docstring\"\"\"\n", "for i, node in enumerate(nodes):\n", "if i % 2 == 0:\n", "return nodes\n", "nodes[i:i + 1] = reSubObject(regex, node, replacer)\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "For", "Condition", "Return'", "Assign'" ]
[ "def FUNC_2(VAR_5: EscapableEntity) ->str:...\n", "\"\"\"docstring\"\"\"\n", "if isinstance(VAR_5, HTML):\n", "return VAR_5.__html__()\n", "VAR_5 = FUNC_0(VAR_5)\n", "VAR_5 = VAR_1.sub('<\\\\1\\\\2>', VAR_5)\n", "for a_href in VAR_3.finditer(VAR_5):\n", "VAR_5 = VAR_5.replace(a_href.group(0), u'<a href=%s>' % VAR_2.sub(u'\"',\n a_href.group(1)))\n", "return VAR_5.replace(u'&amp;nbsp;', u'&nbsp;')\n" ]
[ "def escape_text(text: EscapableEntity) ->str:...\n", "\"\"\"docstring\"\"\"\n", "if isinstance(text, HTML):\n", "return text.__html__()\n", "text = escape_attribute(text)\n", "text = _UNESCAPER_TEXT.sub('<\\\\1\\\\2>', text)\n", "for a_href in _A_HREF.finditer(text):\n", "text = text.replace(a_href.group(0), u'<a href=%s>' % _QUOTE.sub(u'\"',\n a_href.group(1)))\n", "return text.replace(u'&amp;nbsp;', u'&nbsp;')\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 2, 2 ]
[ "FunctionDef'", "Docstring", "Condition", "Return'", "Assign'", "Assign'", "For", "Assign'", "Return'" ]
[ "def FUNC_9(VAR_10, VAR_18=False):...\n", "\"\"\"docstring\"\"\"\n", "def FUNC_7(self):...\n", "self.distribution.run_command('jsversion')\n", "VAR_60 = self.distribution.get_command_obj('jsdeps')\n", "VAR_61 = self.distribution.get_command_obj('js')\n", "VAR_62 = self.distribution.get_command_obj('css')\n", "VAR_60.force = VAR_61.force = VAR_18\n", "VAR_37 = [VAR_60.bower_dir]\n", "VAR_37.extend(VAR_61.targets)\n", "VAR_37.extend(VAR_62.targets)\n", "VAR_63 = [t for t in VAR_37 if not os.path.exists(t)]\n", "if not VAR_3 and not VAR_63:\n", "VAR_10.run(self)\n", "self.distribution.run_command('js')\n", "VAR_63 = [t for t in VAR_37 if not os.path.exists(t)]\n", "VAR_63 = [t for t in VAR_37 if not os.path.exists(t)]\n", "return\n", "self.distribution.run_command('css')\n", "if VAR_18 or VAR_63:\n", "if VAR_63:\n", "self.distribution.run_command('backendtranslations')\n", "VAR_67 = os.path.commonprefix([VAR_2 + os.sep] + VAR_63)\n", "log.warn('rebuilding js and css failed (not a problem)')\n", "VAR_67 = os.path.commonprefix([VAR_2 + os.sep] + VAR_63)\n", "VAR_10.run(self)\n", "VAR_63 = [m[len(VAR_67):] for m in VAR_63]\n", "log.warn(str(e))\n", "VAR_63 = [m[len(VAR_67):] for m in VAR_63]\n", "return CLASS_5\n", "log.warn(\n 'rebuilding js and css failed. The following required files are missing: %s'\n % VAR_63)\n" ]
[ "def css_js_prerelease(command, strict=False):...\n", "\"\"\"docstring\"\"\"\n", "def run(self):...\n", "self.distribution.run_command('jsversion')\n", "jsdeps = self.distribution.get_command_obj('jsdeps')\n", "js = self.distribution.get_command_obj('js')\n", "css = self.distribution.get_command_obj('css')\n", "jsdeps.force = js.force = strict\n", "targets = [jsdeps.bower_dir]\n", "targets.extend(js.targets)\n", "targets.extend(css.targets)\n", "missing = [t for t in targets if not os.path.exists(t)]\n", "if not is_repo and not missing:\n", "command.run(self)\n", "self.distribution.run_command('js')\n", "missing = [t for t in targets if not os.path.exists(t)]\n", "missing = [t for t in targets if not os.path.exists(t)]\n", "return\n", "self.distribution.run_command('css')\n", "if strict or missing:\n", "if missing:\n", "self.distribution.run_command('backendtranslations')\n", "prefix = os.path.commonprefix([repo_root + os.sep] + missing)\n", "log.warn('rebuilding js and css failed (not a problem)')\n", "prefix = os.path.commonprefix([repo_root + os.sep] + missing)\n", "command.run(self)\n", "missing = [m[len(prefix):] for m in missing]\n", "log.warn(str(e))\n", "missing = [m[len(prefix):] for m in missing]\n", "return DecoratedCommand\n", "log.warn(\n 'rebuilding js and css failed. The following required files are missing: %s'\n % missing)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "FunctionDef'", "Expr'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'", "Assign'", "Condition", "Expr'", "Expr'", "Assign'", "Assign'", "Return'", "Expr'", "Condition", "Condition", "Expr'", "Assign'", "Expr'", "Assign'", "Expr'", "Assign'", "Expr'", "Assign'", "Return'", "Expr'" ]
[ "def FUNC_1(self):...\n", "VAR_3 = VAR_0.get('/', {'q': '查询'})\n", "VAR_4 = add_query_param(VAR_3, 'format', 'json')\n", "self.assertIn('q=%E6%9F%A5%E8%AF%A2', VAR_4)\n", "self.assertIn('format=json', VAR_4)\n" ]
[ "def test_add_query_param_with_non_latin_character(self):...\n", "request = factory.get('/', {'q': '查询'})\n", "json_url = add_query_param(request, 'format', 'json')\n", "self.assertIn('q=%E6%9F%A5%E8%AF%A2', json_url)\n", "self.assertIn('format=json', json_url)\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Expr'", "Expr'" ]
[ "def FUNC_12(VAR_4):...\n", "VAR_21 = [m.upper() for m in VAR_13]\n", "for r, m, v in itertools.product(VAR_12, VAR_21, [VAR_4.__name__]):\n", "self.before_acl.append((r, m, v, VAR_3))\n", "return VAR_4\n" ]
[ "def decorator(view_func):...\n", "_methods = [m.upper() for m in methods]\n", "for r, m, v in itertools.product(roles, _methods, [view_func.__name__]):\n", "self.before_acl.append((r, m, v, with_children))\n", "return view_func\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "For", "Expr'", "Return'" ]
[ "def FUNC_3(VAR_17):...\n", "\"\"\"docstring\"\"\"\n", "VAR_22 = VAR_17.lower().strip('\\x00\\x1a \\n\\r\\t')\n", "for scheme in VAR_11:\n", "if re.sub('[^A-Za-z0-9\\\\/:]+', '', VAR_22).startswith(scheme):\n", "return VAR_19(VAR_17, VAR_15=True, VAR_16=False)\n", "return ''\n" ]
[ "def escape_link(url):...\n", "\"\"\"docstring\"\"\"\n", "lower_url = url.lower().strip('\\x00\\x1a \\n\\r\\t')\n", "for scheme in _scheme_blacklist:\n", "if re.sub('[^A-Za-z0-9\\\\/:]+', '', lower_url).startswith(scheme):\n", "return escape(url, quote=True, smart_amp=False)\n", "return ''\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "For", "Condition", "Return'", "Return'" ]
[ "def FUNC_1(self):...\n", "self.client.get('/logout/')\n" ]
[ "def logout(self):...\n", "self.client.get('/logout/')\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Expr'" ]
[ "@wrapt.decorator...\n", "VAR_18.update(dict(zip(VAR_15.__code__.co_varnames, VAR_17)))\n", "FUNC_0(VAR_18['searchfield'])\n", "VAR_18['start'] = cint(VAR_18['start'])\n", "VAR_18['page_len'] = cint(VAR_18['page_len'])\n", "if VAR_18['doctype'] and not VAR_20.db.exists('DocType', VAR_18['doctype']):\n", "return []\n", "return VAR_15(**kwargs)\n" ]
[ "@wrapt.decorator...\n", "kwargs.update(dict(zip(fn.__code__.co_varnames, args)))\n", "sanitize_searchfield(kwargs['searchfield'])\n", "kwargs['start'] = cint(kwargs['start'])\n", "kwargs['page_len'] = cint(kwargs['page_len'])\n", "if kwargs['doctype'] and not frappe.db.exists('DocType', kwargs['doctype']):\n", "return []\n", "return fn(**kwargs)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 2 ]
[ "Condition", "Expr'", "Expr'", "Assign'", "Assign'", "Condition", "Return'", "Return'" ]
[ "def FUNC_10(self) ->None:...\n", "VAR_16 = self.package_dir / 'models'\n", "VAR_16.mkdir()\n", "VAR_17 = VAR_16 / '__init__.py'\n", "VAR_18 = []\n", "VAR_19 = self.env.get_template('types.py')\n", "VAR_20 = VAR_16 / 'types.py'\n", "VAR_20.write_text(VAR_19.render())\n", "VAR_21 = self.env.get_template('model.pyi')\n", "for model in self.openapi.schemas.models.values():\n", "VAR_36 = VAR_16 / f'{model.reference.module_name}.py'\n", "VAR_22 = self.env.get_template('enum.pyi')\n", "VAR_36.write_text(VAR_21.render(model=model))\n", "for enum in self.openapi.enums.values():\n", "VAR_18.append(import_string_from_reference(model.reference))\n", "VAR_36 = VAR_16 / f'{enum.reference.module_name}.py'\n", "VAR_23 = self.env.get_template('models_init.pyi')\n", "VAR_36.write_text(VAR_22.render(enum=enum))\n", "VAR_17.write_text(VAR_23.render(VAR_18=imports))\n", "VAR_18.append(import_string_from_reference(enum.reference))\n" ]
[ "def _build_models(self) ->None:...\n", "models_dir = self.package_dir / 'models'\n", "models_dir.mkdir()\n", "models_init = models_dir / '__init__.py'\n", "imports = []\n", "types_template = self.env.get_template('types.py')\n", "types_path = models_dir / 'types.py'\n", "types_path.write_text(types_template.render())\n", "model_template = self.env.get_template('model.pyi')\n", "for model in self.openapi.schemas.models.values():\n", "module_path = models_dir / f'{model.reference.module_name}.py'\n", "enum_template = self.env.get_template('enum.pyi')\n", "module_path.write_text(model_template.render(model=model))\n", "for enum in self.openapi.enums.values():\n", "imports.append(import_string_from_reference(model.reference))\n", "module_path = models_dir / f'{enum.reference.module_name}.py'\n", "models_init_template = self.env.get_template('models_init.pyi')\n", "module_path.write_text(enum_template.render(enum=enum))\n", "models_init.write_text(models_init_template.render(imports=imports))\n", "imports.append(import_string_from_reference(enum.reference))\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Expr'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "For", "Assign'", "Assign'", "Expr'", "For", "Expr'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'" ]
[ "def __init__(self, VAR_58, *VAR_6, **VAR_7):...\n", "self.project = VAR_58\n", "super().__init__(*VAR_6, **kwargs)\n", "self.fields['token'].queryset = VAR_58.projecttoken_set.all()\n" ]
[ "def __init__(self, project, *args, **kwargs):...\n", "self.project = project\n", "super().__init__(*args, **kwargs)\n", "self.fields['token'].queryset = project.projecttoken_set.all()\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Expr'", "Assign'" ]
[ "@VAR_20.whitelist()...\n", "VAR_9 = cint(VAR_9)\n", "if isinstance(VAR_5, string_types):\n", "VAR_5 = json.loads(VAR_5)\n", "if VAR_1:\n", "FUNC_0(VAR_1)\n", "if not VAR_1:\n", "VAR_1 = 'name'\n", "VAR_21 = VAR_20.get_hooks().standard_queries or {}\n", "if VAR_4 and VAR_4.split()[0].lower() != 'select':\n", "if not VAR_4 and VAR_2 in VAR_21:\n", "is_whitelisted(VAR_20.get_attr(VAR_4))\n", "if VAR_20.local.conf.developer_mode:\n", "FUNC_2(VAR_2, VAR_3, VAR_21[VAR_2][0], VAR_1, VAR_9, VAR_6, VAR_5)\n", "VAR_12 = VAR_20.get_meta(VAR_2)\n", "VAR_20.response['values'] = VAR_20.call(VAR_4, VAR_2, VAR_3, VAR_1, VAR_9,\n VAR_6, VAR_5, VAR_11=as_dict)\n", "VAR_20.respond_as_web_page(title='Invalid Method', html='Method not found',\n indicator_color='red', http_status_code=404)\n", "return\n", "if VAR_4:\n", "VAR_20.throw(_('This query style is discontinued'))\n", "if isinstance(VAR_5, dict):\n", "VAR_34 = VAR_5.items()\n", "if VAR_5 == None:\n", "VAR_5 = []\n", "VAR_5 = []\n", "VAR_26 = []\n", "for f in VAR_34:\n", "if VAR_3:\n", "if isinstance(f[1], (list, tuple)):\n", "VAR_35 = ['name']\n", "if VAR_12.get('fields', {'fieldname': 'enabled', 'fieldtype': 'Check'}):\n", "VAR_5.append([VAR_2, f[0], f[1][0], f[1][1]])\n", "VAR_5.append([VAR_2, f[0], '=', f[1]])\n", "if VAR_12.title_field:\n", "VAR_5.append([VAR_2, 'enabled', '=', 1])\n", "if VAR_12.get('fields', {'fieldname': 'disabled', 'fieldtype': 'Check'}):\n", "VAR_35.append(VAR_12.title_field)\n", "if VAR_12.search_fields:\n", "VAR_5.append([VAR_2, 'disabled', '!=', 1])\n", "VAR_27 = FUNC_3(VAR_12, VAR_1 or 'name')\n", "VAR_35.extend(VAR_12.get_search_fields())\n", "for f in VAR_35:\n", "if VAR_10:\n", "VAR_36 = VAR_12.get_field(f.strip())\n", "VAR_27 = list(set(VAR_27 + json.loads(VAR_10)))\n", "VAR_28 = [('`tab%s`.`%s`' % (VAR_12.name, f.strip())) for f in VAR_27]\n", "if VAR_2 not in VAR_0 and (f == 'name' or VAR_36 and VAR_36.fieldtype in [\n", "VAR_28.append('locate({_txt}, `tab{doctype}`.`name`) as `_relevance`'.\n format(_txt=frappe.db.escape((txt or '').replace('%', '').replace('@',\n '')), VAR_2=doctype))\n", "VAR_26.append([VAR_2, f.strip(), 'like', '%{0}%'.format(VAR_3)])\n", "from frappe.model.db_query import get_order_by\n", "VAR_29 = get_order_by(VAR_2, VAR_12)\n", "VAR_30 = '_relevance, {0}, `tab{1}`.idx desc'.format(VAR_29, VAR_2)\n", "VAR_31 = 'select' if VAR_20.only_has_select_perm(VAR_2) else 'read'\n", "VAR_32 = True if VAR_2 == 'DocType' else cint(VAR_8) and has_permission(VAR_2,\n VAR_31=ptype)\n", "if VAR_2 in VAR_0:\n", "VAR_6 = None\n", "VAR_33 = VAR_20.get_list(VAR_2, VAR_5=filters, VAR_27=formatted_fields,\n VAR_26=or_filters, limit_start=start, limit_page_length=page_length,\n VAR_30=order_by, VAR_32=ignore_permissions, VAR_7=reference_doctype,\n as_list=not as_dict, strict=False)\n", "if VAR_2 in VAR_0:\n", "VAR_33 = tuple([v for v in list(VAR_33) if re.search(re.escape(VAR_3) +\n '.*', _(v.name) if VAR_11 else _(v[0]), re.IGNORECASE)])\n", "if VAR_11:\n", "for r in VAR_33:\n", "VAR_20.response['values'] = [r[:-1] for r in VAR_33]\n", "r.pop('_relevance')\n", "VAR_20.response['values'] = VAR_33\n" ]
[ "@frappe.whitelist()...\n", "start = cint(start)\n", "if isinstance(filters, string_types):\n", "filters = json.loads(filters)\n", "if searchfield:\n", "sanitize_searchfield(searchfield)\n", "if not searchfield:\n", "searchfield = 'name'\n", "standard_queries = frappe.get_hooks().standard_queries or {}\n", "if query and query.split()[0].lower() != 'select':\n", "if not query and doctype in standard_queries:\n", "is_whitelisted(frappe.get_attr(query))\n", "if frappe.local.conf.developer_mode:\n", "search_widget(doctype, txt, standard_queries[doctype][0], searchfield,\n start, page_length, filters)\n", "meta = frappe.get_meta(doctype)\n", "frappe.response['values'] = frappe.call(query, doctype, txt, searchfield,\n start, page_length, filters, as_dict=as_dict)\n", "frappe.respond_as_web_page(title='Invalid Method', html='Method not found',\n indicator_color='red', http_status_code=404)\n", "return\n", "if query:\n", "frappe.throw(_('This query style is discontinued'))\n", "if isinstance(filters, dict):\n", "filters_items = filters.items()\n", "if filters == None:\n", "filters = []\n", "filters = []\n", "or_filters = []\n", "for f in filters_items:\n", "if txt:\n", "if isinstance(f[1], (list, tuple)):\n", "search_fields = ['name']\n", "if meta.get('fields', {'fieldname': 'enabled', 'fieldtype': 'Check'}):\n", "filters.append([doctype, f[0], f[1][0], f[1][1]])\n", "filters.append([doctype, f[0], '=', f[1]])\n", "if meta.title_field:\n", "filters.append([doctype, 'enabled', '=', 1])\n", "if meta.get('fields', {'fieldname': 'disabled', 'fieldtype': 'Check'}):\n", "search_fields.append(meta.title_field)\n", "if meta.search_fields:\n", "filters.append([doctype, 'disabled', '!=', 1])\n", "fields = get_std_fields_list(meta, searchfield or 'name')\n", "search_fields.extend(meta.get_search_fields())\n", "for f in search_fields:\n", "if filter_fields:\n", "fmeta = meta.get_field(f.strip())\n", "fields = list(set(fields + json.loads(filter_fields)))\n", "formatted_fields = [('`tab%s`.`%s`' % (meta.name, f.strip())) for f in fields]\n", "if doctype not in UNTRANSLATED_DOCTYPES and (f == 'name' or fmeta and fmeta\n", "formatted_fields.append('locate({_txt}, `tab{doctype}`.`name`) as `_relevance`'\n .format(_txt=frappe.db.escape((txt or '').replace('%', '').replace('@',\n '')), doctype=doctype))\n", "or_filters.append([doctype, f.strip(), 'like', '%{0}%'.format(txt)])\n", "from frappe.model.db_query import get_order_by\n", "order_by_based_on_meta = get_order_by(doctype, meta)\n", "order_by = '_relevance, {0}, `tab{1}`.idx desc'.format(order_by_based_on_meta,\n doctype)\n", "ptype = 'select' if frappe.only_has_select_perm(doctype) else 'read'\n", "ignore_permissions = True if doctype == 'DocType' else cint(\n ignore_user_permissions) and has_permission(doctype, ptype=ptype)\n", "if doctype in UNTRANSLATED_DOCTYPES:\n", "page_length = None\n", "values = frappe.get_list(doctype, filters=filters, fields=formatted_fields,\n or_filters=or_filters, limit_start=start, limit_page_length=page_length,\n order_by=order_by, ignore_permissions=ignore_permissions,\n reference_doctype=reference_doctype, as_list=not as_dict, strict=False)\n", "if doctype in UNTRANSLATED_DOCTYPES:\n", "values = tuple([v for v in list(values) if re.search(re.escape(txt) + '.*',\n _(v.name) if as_dict else _(v[0]), re.IGNORECASE)])\n", "if as_dict:\n", "for r in values:\n", "frappe.response['values'] = [r[:-1] for r in values]\n", "r.pop('_relevance')\n", "frappe.response['values'] = values\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Assign'", "Condition", "Assign'", "Condition", "Expr'", "Condition", "Assign'", "Assign'", "Condition", "Condition", "Expr'", "Condition", "Expr'", "Assign'", "Assign'", "Expr'", "Return'", "Condition", "Expr'", "Condition", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "For", "Condition", "Condition", "Assign'", "Condition", "Expr'", "Expr'", "Condition", "Expr'", "Condition", "Expr'", "Condition", "Expr'", "Assign'", "Expr'", "For", "Condition", "Assign'", "Assign'", "Assign'", "Condition", "Expr'", "Expr'", "ImportFrom'", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Condition", "Assign'", "Condition", "For", "Assign'", "Expr'", "Assign'" ]
[ "def FUNC_2(VAR_0, VAR_1: FlaskClient):...\n", "VAR_5 = VAR_1.get('/bookmarks/new')\n", "assert VAR_5.status_code == 200\n" ]
[ "def test_get_new_bookmark(test_app, client: FlaskClient):...\n", "response = client.get('/bookmarks/new')\n", "assert response.status_code == 200\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assert'" ]
[ "async def FUNC_30(self, VAR_1: EventBase, VAR_2: Optional[Iterable[...\n", "\"\"\"docstring\"\"\"\n", "if VAR_33 or VAR_1.internal_metadata.is_outlier():\n", "return\n", "VAR_109 = await self.store.get_latest_event_ids_in_room(VAR_1.room_id)\n", "VAR_110 = set(VAR_109)\n", "VAR_111 = set(VAR_1.prev_event_ids())\n", "if VAR_110 == VAR_111:\n", "return\n", "VAR_28 = await self.store.get_room_version_id(VAR_1.room_id)\n", "VAR_90 = KNOWN_ROOM_VERSIONS[VAR_28]\n", "if VAR_2 is not None:\n", "VAR_175 = await self.state_store.get_state_groups(VAR_1.room_id, VAR_110)\n", "VAR_126 = await self.state_handler.get_current_state_ids(VAR_1.room_id,\n latest_event_ids=extrem_ids)\n", "VAR_176 = list(VAR_175.values())\n", "VAR_0.debug('Doing soft-fail check for %s: state %s', VAR_1.event_id, VAR_126)\n", "VAR_176.append(VAR_2)\n", "VAR_112 = auth_types_for_event(VAR_1)\n", "VAR_177 = await self.state_handler.resolve_events(VAR_28, VAR_176, VAR_1)\n", "VAR_113 = [VAR_170 for k, VAR_170 in VAR_126.items() if k in VAR_112]\n", "VAR_126 = {k: VAR_170.event_id for k, VAR_170 in VAR_177.items()}\n", "VAR_114 = await self.store.get_events(VAR_113)\n", "VAR_115 = {(VAR_170.type, VAR_170.state_key): VAR_170 for VAR_170 in\n VAR_114.values()}\n", "event_auth.check(VAR_90, VAR_1, VAR_3=current_auth_events)\n", "VAR_0.warning('Soft-failing %r because %s', VAR_1, VAR_170)\n", "VAR_1.internal_metadata.soft_failed = True\n" ]
[ "async def _check_for_soft_fail(self, event: EventBase, state: Optional[...\n", "\"\"\"docstring\"\"\"\n", "if backfilled or event.internal_metadata.is_outlier():\n", "return\n", "extrem_ids_list = await self.store.get_latest_event_ids_in_room(event.room_id)\n", "extrem_ids = set(extrem_ids_list)\n", "prev_event_ids = set(event.prev_event_ids())\n", "if extrem_ids == prev_event_ids:\n", "return\n", "room_version = await self.store.get_room_version_id(event.room_id)\n", "room_version_obj = KNOWN_ROOM_VERSIONS[room_version]\n", "if state is not None:\n", "state_sets_d = await self.state_store.get_state_groups(event.room_id,\n extrem_ids)\n", "current_state_ids = await self.state_handler.get_current_state_ids(event.\n room_id, latest_event_ids=extrem_ids)\n", "state_sets = list(state_sets_d.values())\n", "logger.debug('Doing soft-fail check for %s: state %s', event.event_id,\n current_state_ids)\n", "state_sets.append(state)\n", "auth_types = auth_types_for_event(event)\n", "current_states = await self.state_handler.resolve_events(room_version,\n state_sets, event)\n", "current_state_ids_list = [e for k, e in current_state_ids.items() if k in\n auth_types]\n", "current_state_ids = {k: e.event_id for k, e in current_states.items()}\n", "auth_events_map = await self.store.get_events(current_state_ids_list)\n", "current_auth_events = {(e.type, e.state_key): e for e in auth_events_map.\n values()}\n", "event_auth.check(room_version_obj, event, auth_events=current_auth_events)\n", "logger.warning('Soft-failing %r because %s', event, e)\n", "event.internal_metadata.soft_failed = True\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Condition", "Return'", "Assign'", "Assign'", "Assign'", "Condition", "Return'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'", "Assign'" ]
[ "\"\"\"Tests for SavedModelCLI tool.\"\"\"\n", "import contextlib\n", "import os\n", "import pickle\n", "import platform\n", "import shutil\n", "import sys\n", "from absl.testing import parameterized\n", "import numpy as np\n", "from six import StringIO\n", "from tensorflow.core.example import example_pb2\n", "from tensorflow.core.framework import types_pb2\n", "from tensorflow.core.protobuf import meta_graph_pb2\n", "from tensorflow.python.debug.wrappers import local_cli_wrapper\n", "from tensorflow.python.eager import def_function\n", "from tensorflow.python.framework import constant_op\n", "from tensorflow.python.framework import dtypes\n", "from tensorflow.python.framework import tensor_spec\n", "from tensorflow.python.lib.io import file_io\n", "from tensorflow.python.ops import variables\n", "from tensorflow.python.platform import test\n", "from tensorflow.python.platform import tf_logging as logging\n", "from tensorflow.python.saved_model import save\n", "from tensorflow.python.tools import saved_model_cli\n", "from tensorflow.python.training.tracking import tracking\n", "VAR_0 = 'cc/saved_model/testdata/half_plus_two/00000123'\n", "@contextlib.contextmanager...\n", "VAR_1, VAR_2 = StringIO(), StringIO()\n", "VAR_3, VAR_4 = VAR_8.stdout, VAR_8.stderr\n", "VAR_8.stdout, VAR_8.stderr = VAR_1, VAR_2\n", "VAR_8.stdout, VAR_8.stderr = VAR_3, VAR_4\n", "def FUNC_1(self):...\n", "yield VAR_8.stdout, VAR_8.stderr\n", "super(CLASS_0, self).setUp()\n", "if platform.system() == 'Windows':\n", "self.skipTest('Skipping failing tests on Windows.')\n", "def FUNC_2(self):...\n", "VAR_9 = test.test_src_dir_path(VAR_0)\n", "self.parser = saved_model_cli.create_parser()\n", "VAR_10 = self.parser.parse_args(['show', '--dir', VAR_9, '--all'])\n", "saved_model_cli.show(VAR_10)\n", "VAR_11 = out.getvalue().strip()\n", "VAR_12 = 'string'\n", "self.maxDiff = None\n", "self.assertMultiLineEqual(VAR_11, VAR_12)\n", "self.assertEqual(err.getvalue().strip(), '')\n", "def FUNC_3(self):...\n", "\"\"\"Model with callable polymorphic functions specified.\"\"\"\n", "@def_function.function...\n", "if VAR_55:\n", "return VAR_53 + VAR_54\n", "return VAR_53 * VAR_54\n" ]
[ "\"\"\"Tests for SavedModelCLI tool.\"\"\"\n", "import contextlib\n", "import os\n", "import pickle\n", "import platform\n", "import shutil\n", "import sys\n", "from absl.testing import parameterized\n", "import numpy as np\n", "from six import StringIO\n", "from tensorflow.core.example import example_pb2\n", "from tensorflow.core.framework import types_pb2\n", "from tensorflow.core.protobuf import meta_graph_pb2\n", "from tensorflow.python.debug.wrappers import local_cli_wrapper\n", "from tensorflow.python.eager import def_function\n", "from tensorflow.python.framework import constant_op\n", "from tensorflow.python.framework import dtypes\n", "from tensorflow.python.framework import tensor_spec\n", "from tensorflow.python.lib.io import file_io\n", "from tensorflow.python.ops import variables\n", "from tensorflow.python.platform import test\n", "from tensorflow.python.platform import tf_logging as logging\n", "from tensorflow.python.saved_model import save\n", "from tensorflow.python.tools import saved_model_cli\n", "from tensorflow.python.training.tracking import tracking\n", "SAVED_MODEL_PATH = 'cc/saved_model/testdata/half_plus_two/00000123'\n", "@contextlib.contextmanager...\n", "new_out, new_err = StringIO(), StringIO()\n", "old_out, old_err = sys.stdout, sys.stderr\n", "sys.stdout, sys.stderr = new_out, new_err\n", "sys.stdout, sys.stderr = old_out, old_err\n", "def setUp(self):...\n", "yield sys.stdout, sys.stderr\n", "super(SavedModelCLITestCase, self).setUp()\n", "if platform.system() == 'Windows':\n", "self.skipTest('Skipping failing tests on Windows.')\n", "def testShowCommandAll(self):...\n", "base_path = test.test_src_dir_path(SAVED_MODEL_PATH)\n", "self.parser = saved_model_cli.create_parser()\n", "args = self.parser.parse_args(['show', '--dir', base_path, '--all'])\n", "saved_model_cli.show(args)\n", "output = out.getvalue().strip()\n", "exp_out = \"\"\"MetaGraphDef with tag-set: 'serve' contains the following SignatureDefs:\n\nsignature_def['classify_x2_to_y3']:\n The given SavedModel SignatureDef contains the following input(s):\n inputs['inputs'] tensor_info:\n dtype: DT_FLOAT\n shape: (-1, 1)\n name: x2:0\n The given SavedModel SignatureDef contains the following output(s):\n outputs['scores'] tensor_info:\n dtype: DT_FLOAT\n shape: (-1, 1)\n name: y3:0\n Method name is: tensorflow/serving/classify\n\nsignature_def['classify_x_to_y']:\n The given SavedModel SignatureDef contains the following input(s):\n inputs['inputs'] tensor_info:\n dtype: DT_STRING\n shape: unknown_rank\n name: tf_example:0\n The given SavedModel SignatureDef contains the following output(s):\n outputs['scores'] tensor_info:\n dtype: DT_FLOAT\n shape: (-1, 1)\n name: y:0\n Method name is: tensorflow/serving/classify\n\nsignature_def['regress_x2_to_y3']:\n The given SavedModel SignatureDef contains the following input(s):\n inputs['inputs'] tensor_info:\n dtype: DT_FLOAT\n shape: (-1, 1)\n name: x2:0\n The given SavedModel SignatureDef contains the following output(s):\n outputs['outputs'] tensor_info:\n dtype: DT_FLOAT\n shape: (-1, 1)\n name: y3:0\n Method name is: tensorflow/serving/regress\n\nsignature_def['regress_x_to_y']:\n The given SavedModel SignatureDef contains the following input(s):\n inputs['inputs'] tensor_info:\n dtype: DT_STRING\n shape: unknown_rank\n name: tf_example:0\n The given SavedModel SignatureDef contains the following output(s):\n outputs['outputs'] tensor_info:\n dtype: DT_FLOAT\n shape: (-1, 1)\n name: y:0\n Method name is: tensorflow/serving/regress\n\nsignature_def['regress_x_to_y2']:\n The given SavedModel SignatureDef contains the following input(s):\n inputs['inputs'] tensor_info:\n dtype: DT_STRING\n shape: unknown_rank\n name: tf_example:0\n The given SavedModel SignatureDef contains the following output(s):\n outputs['outputs'] tensor_info:\n dtype: DT_FLOAT\n shape: (-1, 1)\n name: y2:0\n Method name is: tensorflow/serving/regress\n\nsignature_def['serving_default']:\n The given SavedModel SignatureDef contains the following input(s):\n inputs['x'] tensor_info:\n dtype: DT_FLOAT\n shape: (-1, 1)\n name: x:0\n The given SavedModel SignatureDef contains the following output(s):\n outputs['y'] tensor_info:\n dtype: DT_FLOAT\n shape: (-1, 1)\n name: y:0\n Method name is: tensorflow/serving/predict\"\"\"\n", "self.maxDiff = None\n", "self.assertMultiLineEqual(output, exp_out)\n", "self.assertEqual(err.getvalue().strip(), '')\n", "def testShowAllWithFunctions(self):...\n", "\"\"\"Model with callable polymorphic functions specified.\"\"\"\n", "@def_function.function...\n", "if c:\n", "return a + b\n", "return a * b\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Expr'", "Import'", "Import'", "Import'", "Import'", "Import'", "Import'", "ImportFrom'", "Import'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "FunctionDef'", "Expr'", "Expr'", "For", "Expr'", "FunctionDef'", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'", "FunctionDef'", "Expr'", "Condition", "Condition", "Return'", "Return'" ]
[ "@FUNC_0...\n", "return EventCreationHandler(self)\n" ]
[ "@cache_in_self...\n", "return EventCreationHandler(self)\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "def FUNC_25(self):...\n", "VAR_5, VAR_15 = self._test_confirm_start()\n", "VAR_15 = VAR_15.replace('/reset/', '/reset/custom/')\n", "VAR_3 = self.client.post(VAR_15, {'new_password1': 'anewpassword',\n 'new_password2': 'anewpassword'})\n", "self.assertEqual(VAR_3.status_code, 302)\n", "self.assertURLEqual(VAR_3.url, '/custom/')\n" ]
[ "def test_confirm_redirect_custom(self):...\n", "url, path = self._test_confirm_start()\n", "path = path.replace('/reset/', '/reset/custom/')\n", "response = self.client.post(path, {'new_password1': 'anewpassword',\n 'new_password2': 'anewpassword'})\n", "self.assertEqual(response.status_code, 302)\n", "self.assertURLEqual(response.url, '/custom/')\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'" ]
[ "def FUNC_114():...\n", "if VAR_32:\n", "if VAR_19.rollback_on_exception:\n", "VAR_12.rollback()\n", "import inspect\n", "if inspect.isclass(VAR_32) and issubclass(VAR_32, Exception):\n" ]
[ "def _raise_exception():...\n", "if raise_exception:\n", "if flags.rollback_on_exception:\n", "db.rollback()\n", "import inspect\n", "if inspect.isclass(raise_exception) and issubclass(raise_exception, Exception):\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Condition", "Expr'", "Import'", "Condition" ]
[ "def FUNC_43(self):...\n", "if self.user and self.user.email is not None:\n", "return self.user.email\n", "return self.email\n" ]
[ "def _email_address(self):...\n", "if self.user and self.user.email is not None:\n", "return self.user.email\n", "return self.email\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Return'", "Return'" ]
[ "def __init__(self, VAR_2: IReactorCore, VAR_3: Optional[...\n", "self._reactor = VAR_2\n", "self._parsed_uri = VAR_12\n", "if VAR_3 is None:\n", "self._tls_options = None\n", "self._tls_options = VAR_3.get_options(self._parsed_uri.host)\n", "self._srv_resolver = VAR_11\n" ]
[ "def __init__(self, reactor: IReactorCore, tls_client_options_factory:...\n", "self._reactor = reactor\n", "self._parsed_uri = parsed_uri\n", "if tls_client_options_factory is None:\n", "self._tls_options = None\n", "self._tls_options = tls_client_options_factory.get_options(self._parsed_uri\n .host)\n", "self._srv_resolver = srv_resolver\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'" ]
[ "def FUNC_29():...\n", "VAR_58 = Config(os.path.join(request.folder, 'settings.cfg'), section=\n 'editor_sessions', default_values={})\n", "VAR_57 = VAR_58.read()\n", "if request.vars.session_name and request.vars.files:\n", "VAR_135 = request.vars.session_name\n", "return VAR_43.render('default/editor_sessions.html', {'editor_sessions':\n VAR_57})\n", "VAR_45 = request.vars.files\n", "VAR_57.update({VAR_135: ','.join(VAR_45)})\n", "if VAR_58.save(VAR_57.items()):\n", "VAR_43.headers['web2py-component-flash'] = T('Session saved correctly')\n", "VAR_43.headers['web2py-component-flash'] = T('Session saved on session only')\n" ]
[ "def editor_sessions():...\n", "config = Config(os.path.join(request.folder, 'settings.cfg'), section=\n 'editor_sessions', default_values={})\n", "preferences = config.read()\n", "if request.vars.session_name and request.vars.files:\n", "session_name = request.vars.session_name\n", "return response.render('default/editor_sessions.html', {'editor_sessions':\n preferences})\n", "files = request.vars.files\n", "preferences.update({session_name: ','.join(files)})\n", "if config.save(preferences.items()):\n", "response.headers['web2py-component-flash'] = T('Session saved correctly')\n", "response.headers['web2py-component-flash'] = T('Session saved on session only')\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Condition", "Assign'", "Return'", "Assign'", "Expr'", "Condition", "Assign'", "Assign'" ]
[ "def FUNC_14(self):...\n", "self.assertFormfield(models.Band, 'members', widgets.ManyToManyRawIdWidget,\n raw_id_fields=['members'])\n" ]
[ "def testRawIDManyTOMany(self):...\n", "self.assertFormfield(models.Band, 'members', widgets.ManyToManyRawIdWidget,\n raw_id_fields=['members'])\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Expr'" ]
[ "def FUNC_1(self):...\n", "if 'ping' not in current_app.config.get('ACTIVE_PROVIDERS'):\n", "return 'Ping is not enabled in the config. See the ACTIVE_PROVIDERS section.', 404\n", "VAR_6 = ('clientId,{client_id},redirectUri,{redirectUri},return_to,{return_to}'\n .format(VAR_8=current_app.config.get('PING_CLIENT_ID'), redirectUri=\n current_app.config.get('PING_REDIRECT_URI'), VAR_10=current_app.config.\n get('WEB_PATH')))\n", "self.reqparse.add_argument('code', type=str, required=True)\n", "self.reqparse.add_argument('state', type=str, required=False, default=\n default_state)\n", "VAR_7 = self.reqparse.parse_args()\n", "VAR_8 = VAR_7['state'].split(',')[1]\n", "VAR_9 = VAR_7['state'].split(',')[3]\n", "VAR_10 = VAR_7['state'].split(',')[5]\n", "if not validate_redirect_url(VAR_10):\n", "VAR_10 = current_app.config.get('WEB_PATH')\n", "VAR_11 = {'client_id': VAR_8, 'grant_type': 'authorization_code', 'scope':\n 'openid email profile address', 'redirect_uri': VAR_9, 'code': VAR_7[\n 'code']}\n", "VAR_12 = current_app.config.get('PING_ACCESS_TOKEN_URL')\n", "VAR_13 = current_app.config.get('PING_USER_API_URL')\n", "VAR_14 = base64.b64encode(bytes('{0}:{1}'.format(VAR_8, current_app.config.\n get('PING_SECRET'))))\n", "VAR_15 = {'Authorization': 'Basic {0}'.format(VAR_14.decode('utf-8'))}\n", "VAR_16 = requests.post(VAR_12, VAR_15=headers, VAR_11=params)\n", "VAR_17 = VAR_16.json()['id_token']\n", "VAR_18 = VAR_16.json()['access_token']\n", "VAR_19 = fetch_token_header_payload(VAR_17)[0]\n", "VAR_20 = current_app.config.get('PING_JWKS_URL')\n", "VAR_16 = requests.get(VAR_20)\n", "for key in VAR_16.json()['keys']:\n", "if key['kid'] == VAR_19['kid']:\n", "current_app.logger.debug(VAR_17)\n", "return dict(message='Token is invalid'), 403\n", "VAR_21 = dict(VAR_18=access_token, schema='profile')\n", "VAR_34 = get_rsa_public_key(key['n'], key['e'])\n", "current_app.logger.debug(VAR_34)\n", "VAR_16 = requests.get(VAR_13, VAR_11=user_params)\n", "VAR_35 = VAR_19['alg']\n", "current_app.logger.debug(VAR_35)\n", "VAR_22 = VAR_16.json()\n", "jwt.decode(VAR_17, VAR_34.decode('utf-8'), algorithms=[algo], audience=\n client_id)\n", "VAR_23 = User.query.filter(User.email == VAR_22['email']).first()\n", "if not VAR_23:\n", "VAR_23 = User(VAR_36=profile['email'], active=True, role='View')\n", "identity_changed.send(current_app._get_current_object(), identity=Identity(\n user.id))\n", "db.session.add(VAR_23)\n", "login_user(VAR_23)\n", "db.session.commit()\n", "return redirect(VAR_10, code=302)\n", "db.session.refresh(VAR_23)\n" ]
[ "def post(self):...\n", "if 'ping' not in current_app.config.get('ACTIVE_PROVIDERS'):\n", "return 'Ping is not enabled in the config. See the ACTIVE_PROVIDERS section.', 404\n", "default_state = (\n 'clientId,{client_id},redirectUri,{redirectUri},return_to,{return_to}'.\n format(client_id=current_app.config.get('PING_CLIENT_ID'), redirectUri=\n current_app.config.get('PING_REDIRECT_URI'), return_to=current_app.\n config.get('WEB_PATH')))\n", "self.reqparse.add_argument('code', type=str, required=True)\n", "self.reqparse.add_argument('state', type=str, required=False, default=\n default_state)\n", "args = self.reqparse.parse_args()\n", "client_id = args['state'].split(',')[1]\n", "redirect_uri = args['state'].split(',')[3]\n", "return_to = args['state'].split(',')[5]\n", "if not validate_redirect_url(return_to):\n", "return_to = current_app.config.get('WEB_PATH')\n", "params = {'client_id': client_id, 'grant_type': 'authorization_code',\n 'scope': 'openid email profile address', 'redirect_uri': redirect_uri,\n 'code': args['code']}\n", "access_token_url = current_app.config.get('PING_ACCESS_TOKEN_URL')\n", "user_api_url = current_app.config.get('PING_USER_API_URL')\n", "basic = base64.b64encode(bytes('{0}:{1}'.format(client_id, current_app.\n config.get('PING_SECRET'))))\n", "headers = {'Authorization': 'Basic {0}'.format(basic.decode('utf-8'))}\n", "r = requests.post(access_token_url, headers=headers, params=params)\n", "id_token = r.json()['id_token']\n", "access_token = r.json()['access_token']\n", "header_data = fetch_token_header_payload(id_token)[0]\n", "jwks_url = current_app.config.get('PING_JWKS_URL')\n", "r = requests.get(jwks_url)\n", "for key in r.json()['keys']:\n", "if key['kid'] == header_data['kid']:\n", "current_app.logger.debug(id_token)\n", "return dict(message='Token is invalid'), 403\n", "user_params = dict(access_token=access_token, schema='profile')\n", "secret = get_rsa_public_key(key['n'], key['e'])\n", "current_app.logger.debug(secret)\n", "r = requests.get(user_api_url, params=user_params)\n", "algo = header_data['alg']\n", "current_app.logger.debug(algo)\n", "profile = r.json()\n", "jwt.decode(id_token, secret.decode('utf-8'), algorithms=[algo], audience=\n client_id)\n", "user = User.query.filter(User.email == profile['email']).first()\n", "if not user:\n", "user = User(email=profile['email'], active=True, role='View')\n", "identity_changed.send(current_app._get_current_object(), identity=Identity(\n user.id))\n", "db.session.add(user)\n", "login_user(user)\n", "db.session.commit()\n", "return redirect(return_to, code=302)\n", "db.session.refresh(user)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Return'", "Assign'", "Expr'", "Expr'", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "For", "Condition", "Expr'", "Return'", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "Expr'", "Assign'", "Expr'", "Assign'", "Condition", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'", "Return'", "Expr'" ]