lines
sequencelengths 1
383
| raw_lines
sequencelengths 1
383
| label
sequencelengths 1
383
| type
sequencelengths 1
383
|
---|---|---|---|
[
"def FUNC_4(VAR_1: Any) ->str:...\n",
"assert isinstance(VAR_1, type), repr(VAR_1)\n",
"if VAR_1.__module__.startswith('pytorch_lightning.'):\n",
"return f'pl.{VAR_1.__name__}'\n",
"return f'{VAR_1.__module__}.{VAR_1.__qualname__}'\n"
] | [
"def _get_abbrev_qualified_cls_name(cls: Any) ->str:...\n",
"assert isinstance(cls, type), repr(cls)\n",
"if cls.__module__.startswith('pytorch_lightning.'):\n",
"return f'pl.{cls.__name__}'\n",
"return f'{cls.__module__}.{cls.__qualname__}'\n"
] | [
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assert'",
"Condition",
"Return'",
"Return'"
] |
[
"@VAR_2.route('/ajax/simulatedbchange', methods=['POST'])...\n",
"VAR_84, VAR_85 = FUNC_47()\n",
"return Response(json.dumps({'change': VAR_84, 'valid': VAR_85}), mimetype=\n 'application/json')\n"
] | [
"@admi.route('/ajax/simulatedbchange', methods=['POST'])...\n",
"db_change, db_valid = _db_simulate_change()\n",
"return Response(json.dumps({'change': db_change, 'valid': db_valid}),\n mimetype='application/json')\n"
] | [
0,
0,
0
] | [
"Condition",
"Assign'",
"Return'"
] |
[
"def FUNC_47(self, VAR_12):...\n",
"VAR_22, VAR_23 = self.make_request('GET',\n '/_matrix/client/r0/rooms/{}/state/m.room.member/{}'.format(self.\n room_id, self.second_user_id), VAR_16=self.creator_tok)\n",
"self.assertEqual(VAR_23.code, 200, VAR_23.result)\n",
"VAR_57 = VAR_23.json_body\n",
"self.assertEqual(VAR_57.get('reason'), VAR_12, VAR_23.result)\n"
] | [
"def _check_for_reason(self, reason):...\n",
"request, channel = self.make_request('GET',\n '/_matrix/client/r0/rooms/{}/state/m.room.member/{}'.format(self.\n room_id, self.second_user_id), access_token=self.creator_tok)\n",
"self.assertEqual(channel.code, 200, channel.result)\n",
"event_content = channel.json_body\n",
"self.assertEqual(event_content.get('reason'), reason, channel.result)\n"
] | [
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Expr'",
"Assign'",
"Expr'"
] |
[
"def FUNC_1(VAR_6: list[FavaLedger]) ->None:...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_38: dict[str, FavaLedger] = {}\n",
"for VAR_5 in VAR_6:\n",
"VAR_36 = FUNC_0(VAR_5)\n",
"VAR_1.config['LEDGERS'] = VAR_38\n",
"VAR_37 = next_key(VAR_36, VAR_38)\n",
"VAR_38[VAR_37] = VAR_5\n"
] | [
"def update_ledger_slugs(ledgers: list[FavaLedger]) ->None:...\n",
"\"\"\"docstring\"\"\"\n",
"ledgers_by_slug: dict[str, FavaLedger] = {}\n",
"for ledger in ledgers:\n",
"slug = ledger_slug(ledger)\n",
"app.config['LEDGERS'] = ledgers_by_slug\n",
"unique_key = next_key(slug, ledgers_by_slug)\n",
"ledgers_by_slug[unique_key] = ledger\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"AnnAssign'",
"For",
"Assign'",
"Assign'",
"Assign'",
"Assign'"
] |
[
"def FUNC_118(VAR_120=None):...\n",
"VAR_184 = {}\n",
"for VAR_219 in ([VAR_120] if VAR_120 else FUNC_64(VAR_122=True)):\n",
"VAR_219 = 'frappe' if VAR_219 == 'webnotes' else VAR_219\n",
"return VAR_184\n",
"VAR_226 = FUNC_55(VAR_219 + '.hooks')\n",
"if VAR_1.flags.in_install_app:\n",
"for VAR_46 in dir(VAR_226):\n",
"print('Could not find app \"{0}\"'.format(VAR_120))\n",
"if not VAR_46.startswith('_'):\n",
"if not VAR_16:\n",
"FUNC_67(VAR_184, VAR_46, getattr(VAR_226, VAR_46))\n",
"sys.exit(1)\n"
] | [
"def load_app_hooks(app_name=None):...\n",
"hooks = {}\n",
"for app in ([app_name] if app_name else get_installed_apps(sort=True)):\n",
"app = 'frappe' if app == 'webnotes' else app\n",
"return hooks\n",
"app_hooks = get_module(app + '.hooks')\n",
"if local.flags.in_install_app:\n",
"for key in dir(app_hooks):\n",
"print('Could not find app \"{0}\"'.format(app_name))\n",
"if not key.startswith('_'):\n",
"if not request:\n",
"append_hook(hooks, key, getattr(app_hooks, key))\n",
"sys.exit(1)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"For",
"Assign'",
"Return'",
"Assign'",
"Condition",
"For",
"Expr'",
"Condition",
"Condition",
"Expr'",
"Expr'"
] |
[
"def FUNC_51(self):...\n",
"\"\"\"docstring\"\"\"\n",
"self._doc_before_save = None\n",
"if not self.is_new():\n",
"self._doc_before_save = frappe.get_doc(self.doctype, self.name)\n",
"self._doc_before_save = None\n",
"frappe.clear_last_message()\n"
] | [
"def load_doc_before_save(self):...\n",
"\"\"\"docstring\"\"\"\n",
"self._doc_before_save = None\n",
"if not self.is_new():\n",
"self._doc_before_save = frappe.get_doc(self.doctype, self.name)\n",
"self._doc_before_save = None\n",
"frappe.clear_last_message()\n"
] | [
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Expr'"
] |
[
"def FUNC_1(VAR_0, VAR_1: FlaskClient):...\n",
"VAR_0.config['THEME_CONF']['use_custom_css'] = True\n",
"VAR_6 = 'custom.css'\n",
"VAR_7 = \"\"\"\n body {\n color: red\n }\n \"\"\"\n",
"os.mkdir(f\"{VAR_0.config['USER_DIR']}/css/\")\n",
"f.write(VAR_7)\n",
"VAR_0.config['THEME_CONF']['custom_css_file'] = VAR_6\n",
"VAR_8 = VAR_1.get('/static/custom.css')\n",
"assert VAR_7.encode('utf-8') in VAR_8.data\n",
"VAR_0.config['THEME_CONF']['use_custom_css'] = False\n"
] | [
"def test_get_custom_css(test_app, client: FlaskClient):...\n",
"test_app.config['THEME_CONF']['use_custom_css'] = True\n",
"css_file = 'custom.css'\n",
"css_contents = \"\"\"\n body {\n color: red\n }\n \"\"\"\n",
"os.mkdir(f\"{test_app.config['USER_DIR']}/css/\")\n",
"f.write(css_contents)\n",
"test_app.config['THEME_CONF']['custom_css_file'] = css_file\n",
"resp = client.get('/static/custom.css')\n",
"assert css_contents.encode('utf-8') in resp.data\n",
"test_app.config['THEME_CONF']['use_custom_css'] = False\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Assign'",
"Assign'",
"Assert'",
"Assign'"
] |
[
"def FUNC_33(VAR_30, VAR_35='\\n'):...\n",
"print(VAR_29 + VAR_30, VAR_35=end)\n"
] | [
"def in_print(s, end='\\n'):...\n",
"print(indent_str + s, end=end)\n"
] | [
0,
0
] | [
"FunctionDef'",
"Expr'"
] |
[
"def __init__(self, VAR_26: str, VAR_27: Callable[..., Any], *, VAR_29:...\n",
"if isinstance(VAR_13, enum.IntEnum):\n",
"VAR_13 = int(VAR_13)\n",
"self.path = VAR_26\n",
"self.endpoint = VAR_27\n",
"self.name = get_name(VAR_27) if VAR_28 is None else VAR_28\n",
"self.path_regex, self.path_format, self.param_convertors = compile_path(VAR_26)\n",
"if VAR_37 is None:\n",
"VAR_37 = ['GET']\n",
"self.methods: Set[str] = set([method.upper() for method in VAR_37])\n",
"self.unique_id = generate_operation_id_for_path(VAR_28=self.name, VAR_26=\n self.path_format, method=list(methods)[0])\n",
"self.response_model = VAR_29\n",
"if self.response_model:\n",
"assert VAR_13 not in STATUS_CODES_WITH_NO_BODY, f'Status code {VAR_13} must not have a response body'\n",
"self.response_field = None\n",
"VAR_73 = 'Response_' + self.unique_id\n",
"self.secure_cloned_response_field = None\n",
"self.response_field = create_response_field(VAR_28=response_name, type_=\n self.response_model)\n",
"self.status_code = VAR_13\n",
"self.secure_cloned_response_field: Optional[ModelField] = create_cloned_field(\n self.response_field)\n",
"self.tags = VAR_30 or []\n",
"if VAR_31:\n",
"self.dependencies = list(VAR_31)\n",
"self.dependencies = []\n",
"self.summary = VAR_32\n",
"self.description = VAR_33 or inspect.cleandoc(self.endpoint.__doc__ or '')\n",
"self.description = self.description.split('\\x0c')[0]\n",
"self.response_description = VAR_34\n",
"self.responses = VAR_35 or {}\n",
"VAR_59 = {}\n",
"for VAR_75, VAR_72 in self.responses.items():\n",
"assert isinstance(VAR_72, dict), 'An additional response must be a dict'\n",
"if VAR_59:\n",
"VAR_74 = VAR_72.get('model')\n",
"self.response_fields: Dict[Union[int, str], ModelField] = VAR_59\n",
"self.response_fields = {}\n",
"if VAR_74:\n",
"self.deprecated = VAR_36\n",
"assert VAR_75 not in STATUS_CODES_WITH_NO_BODY, f'Status code {VAR_75} must not have a response body'\n",
"self.operation_id = VAR_38\n",
"VAR_73 = f'Response_{VAR_75}_{self.unique_id}'\n",
"self.response_model_include = VAR_16\n",
"VAR_15 = create_response_field(VAR_28=response_name, type_=model)\n",
"self.response_model_exclude = VAR_17\n",
"VAR_59[VAR_75] = VAR_15\n",
"self.response_model_by_alias = VAR_18\n",
"self.response_model_exclude_unset = VAR_19\n",
"self.response_model_exclude_defaults = VAR_20\n",
"self.response_model_exclude_none = VAR_21\n",
"self.include_in_schema = VAR_39\n",
"self.response_class = VAR_14\n",
"assert callable(VAR_27), 'An endpoint must be a callable'\n",
"self.dependant = get_dependant(VAR_26=self.path_format, call=self.endpoint)\n",
"for depends in self.dependencies[::-1]:\n",
"self.dependant.dependencies.insert(0, get_parameterless_sub_dependant(\n depends=depends, VAR_26=self.path_format))\n",
"self.body_field = get_body_field(VAR_10=self.dependant, VAR_28=self.unique_id)\n",
"self.dependency_overrides_provider = VAR_22\n",
"self.callbacks = VAR_40\n",
"self.app = request_response(self.get_route_handler())\n"
] | [
"def __init__(self, path: str, endpoint: Callable[..., Any], *,...\n",
"if isinstance(status_code, enum.IntEnum):\n",
"status_code = int(status_code)\n",
"self.path = path\n",
"self.endpoint = endpoint\n",
"self.name = get_name(endpoint) if name is None else name\n",
"self.path_regex, self.path_format, self.param_convertors = compile_path(path)\n",
"if methods is None:\n",
"methods = ['GET']\n",
"self.methods: Set[str] = set([method.upper() for method in methods])\n",
"self.unique_id = generate_operation_id_for_path(name=self.name, path=self.\n path_format, method=list(methods)[0])\n",
"self.response_model = response_model\n",
"if self.response_model:\n",
"assert status_code not in STATUS_CODES_WITH_NO_BODY, f'Status code {status_code} must not have a response body'\n",
"self.response_field = None\n",
"response_name = 'Response_' + self.unique_id\n",
"self.secure_cloned_response_field = None\n",
"self.response_field = create_response_field(name=response_name, type_=self.\n response_model)\n",
"self.status_code = status_code\n",
"self.secure_cloned_response_field: Optional[ModelField] = create_cloned_field(\n self.response_field)\n",
"self.tags = tags or []\n",
"if dependencies:\n",
"self.dependencies = list(dependencies)\n",
"self.dependencies = []\n",
"self.summary = summary\n",
"self.description = description or inspect.cleandoc(self.endpoint.__doc__ or '')\n",
"self.description = self.description.split('\\x0c')[0]\n",
"self.response_description = response_description\n",
"self.responses = responses or {}\n",
"response_fields = {}\n",
"for additional_status_code, response in self.responses.items():\n",
"assert isinstance(response, dict), 'An additional response must be a dict'\n",
"if response_fields:\n",
"model = response.get('model')\n",
"self.response_fields: Dict[Union[int, str], ModelField] = response_fields\n",
"self.response_fields = {}\n",
"if model:\n",
"self.deprecated = deprecated\n",
"assert additional_status_code not in STATUS_CODES_WITH_NO_BODY, f'Status code {additional_status_code} must not have a response body'\n",
"self.operation_id = operation_id\n",
"response_name = f'Response_{additional_status_code}_{self.unique_id}'\n",
"self.response_model_include = response_model_include\n",
"response_field = create_response_field(name=response_name, type_=model)\n",
"self.response_model_exclude = response_model_exclude\n",
"response_fields[additional_status_code] = response_field\n",
"self.response_model_by_alias = response_model_by_alias\n",
"self.response_model_exclude_unset = response_model_exclude_unset\n",
"self.response_model_exclude_defaults = response_model_exclude_defaults\n",
"self.response_model_exclude_none = response_model_exclude_none\n",
"self.include_in_schema = include_in_schema\n",
"self.response_class = response_class\n",
"assert callable(endpoint), 'An endpoint must be a callable'\n",
"self.dependant = get_dependant(path=self.path_format, call=self.endpoint)\n",
"for depends in self.dependencies[::-1]:\n",
"self.dependant.dependencies.insert(0, get_parameterless_sub_dependant(\n depends=depends, path=self.path_format))\n",
"self.body_field = get_body_field(dependant=self.dependant, name=self.unique_id)\n",
"self.dependency_overrides_provider = dependency_overrides_provider\n",
"self.callbacks = callbacks\n",
"self.app = request_response(self.get_route_handler())\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"AnnAssign'",
"Assign'",
"Assign'",
"Condition",
"Assert'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"AnnAssign'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"For",
"Assert'",
"Condition",
"Assign'",
"AnnAssign'",
"Assign'",
"Condition",
"Assign'",
"Assert'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assert'",
"Assign'",
"For",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Assign'"
] |
[
"@defer.inlineCallbacks...\n",
"VAR_18 = yield FUNC_7(VAR_11)\n",
"for member in VAR_18:\n",
"if VAR_14 is not None and member == VAR_14:\n",
"if VAR_6.is_mine(member):\n",
"if VAR_12 is not None:\n",
"if VAR_13 is not None:\n",
"VAR_12.add(member)\n",
"VAR_13.add(member.domain)\n"
] | [
"@defer.inlineCallbacks...\n",
"members = yield get_room_members(room_id)\n",
"for member in members:\n",
"if ignore_user is not None and member == ignore_user:\n",
"if hs.is_mine(member):\n",
"if localusers is not None:\n",
"if remotedomains is not None:\n",
"localusers.add(member)\n",
"remotedomains.add(member.domain)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Assign'",
"For",
"Condition",
"Condition",
"Condition",
"Condition",
"Expr'",
"Expr'"
] |
[
"def FUNC_3(self):...\n",
"VAR_10 = self.get_success(self.handler.check_device_registered(VAR_5=\n '@boris:foo', VAR_6='fco', initial_device_display_name='display name'))\n",
"self.assertEqual(VAR_10, 'fco')\n",
"VAR_11 = self.get_success(self.handler.store.get_device('@boris:foo', 'fco'))\n",
"self.assertEqual(VAR_11['display_name'], 'display name')\n"
] | [
"def test_device_is_created_if_doesnt_exist(self):...\n",
"res = self.get_success(self.handler.check_device_registered(user_id=\n '@boris:foo', device_id='fco', initial_device_display_name='display name'))\n",
"self.assertEqual(res, 'fco')\n",
"dev = self.get_success(self.handler.store.get_device('@boris:foo', 'fco'))\n",
"self.assertEqual(dev['display_name'], 'display name')\n"
] | [
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Expr'",
"Assign'",
"Expr'"
] |
[
"@VAR_2.route('/ajax/listbooks')...\n",
"VAR_79 = int(request.args.get('offset') or 0)\n",
"VAR_19 = int(request.args.get('limit') or config.config_books_per_page)\n",
"VAR_80 = request.args.get('search')\n",
"VAR_7 = request.args.get('sort', 'id')\n",
"VAR_10 = request.args.get('order', '').lower()\n",
"VAR_81 = None\n",
"VAR_76 = tuple()\n",
"if VAR_7 == 'state':\n",
"VAR_81 = json.loads(request.args.get('state', '[]'))\n",
"if VAR_7 == 'tags':\n",
"VAR_82 = VAR_83 = calibre_db.session.query(db.Books).filter(calibre_db.\n common_filters(False)).count()\n",
"VAR_10 = [db.Tags.name.asc()] if VAR_10 == 'asc' else [db.Tags.name.desc()]\n",
"if VAR_7 == 'series':\n",
"if VAR_81 is not None:\n",
"VAR_76 = db.books_tags_link, db.Books.id == db.books_tags_link.c.book, db.Tags\n",
"VAR_10 = [db.Series.name.asc()] if VAR_10 == 'asc' else [db.Series.name.desc()]\n",
"if VAR_7 == 'publishers':\n",
"if VAR_80:\n",
"if VAR_80:\n",
"VAR_76 = (db.books_series_link, db.Books.id == db.books_series_link.c.book,\n db.Series)\n",
"VAR_10 = [db.Publishers.name.asc()] if VAR_10 == 'asc' else [db.Publishers.\n name.desc()]\n",
"if VAR_7 == 'authors':\n",
"VAR_127 = calibre_db.search_query(VAR_80).all()\n",
"VAR_127 = calibre_db.session.query(db.Books).filter(calibre_db.common_filters()\n ).all()\n",
"VAR_63, VAR_83, VAR_64 = calibre_db.get_search_results(VAR_80, VAR_79, [\n VAR_10, ''], VAR_19, *VAR_76)\n",
"VAR_63, VAR_64, VAR_64 = calibre_db.fill_indexpage(int(VAR_79) / int(VAR_19\n ) + 1, VAR_19, db.Books, True, VAR_10, *VAR_76)\n",
"VAR_76 = (db.books_publishers_link, db.Books.id == db.books_publishers_link\n .c.book, db.Publishers)\n",
"VAR_10 = [db.Authors.name.asc(), db.Series.name, db.Books.series_index\n ] if VAR_10 == 'asc' else [db.Authors.name.desc(), db.Series.name.desc(\n ), db.Books.series_index.desc()]\n",
"if VAR_7 == 'languages':\n",
"VAR_83 = len(VAR_127)\n",
"VAR_63 = calibre_db.get_checkbox_sorted(VAR_127, VAR_81, VAR_79, VAR_19, VAR_10\n )\n",
"for entry in VAR_63:\n",
"VAR_76 = (db.books_authors_link, db.Books.id == db.books_authors_link.c.\n book, db.Authors, db.books_series_link, db.Books.id == db.\n books_series_link.c.book, db.Series)\n",
"VAR_10 = [db.Languages.lang_code.asc()] if VAR_10 == 'asc' else [db.\n Languages.lang_code.desc()]\n",
"if VAR_10 and VAR_7 in ['sort', 'title', 'authors_sort', 'series_index']:\n",
"for FUNC_31 in range(0, len(entry.languages)):\n",
"VAR_84 = {'totalNotFiltered': VAR_82, 'total': VAR_83, 'rows': VAR_63}\n",
"VAR_76 = (db.books_languages_link, db.Books.id == db.books_languages_link.c\n .book, db.Languages)\n",
"VAR_10 = [text(VAR_7 + ' ' + VAR_10)]\n",
"if not VAR_81:\n",
"entry.languages[FUNC_31].language_name = isoLanguages.get_language_name(\n get_locale(), entry.languages[FUNC_31].lang_code)\n",
"VAR_85 = json.dumps(VAR_84, cls=db.AlchemyEncoder)\n",
"VAR_10 = [db.Books.timestamp.desc()]\n",
"VAR_86 = make_response(VAR_85)\n",
"VAR_86.headers['Content-Type'] = 'application/json; charset=utf-8'\n",
"return VAR_86\n"
] | [
"@web.route('/ajax/listbooks')...\n",
"off = int(request.args.get('offset') or 0)\n",
"limit = int(request.args.get('limit') or config.config_books_per_page)\n",
"search = request.args.get('search')\n",
"sort = request.args.get('sort', 'id')\n",
"order = request.args.get('order', '').lower()\n",
"state = None\n",
"join = tuple()\n",
"if sort == 'state':\n",
"state = json.loads(request.args.get('state', '[]'))\n",
"if sort == 'tags':\n",
"total_count = filtered_count = calibre_db.session.query(db.Books).filter(\n calibre_db.common_filters(False)).count()\n",
"order = [db.Tags.name.asc()] if order == 'asc' else [db.Tags.name.desc()]\n",
"if sort == 'series':\n",
"if state is not None:\n",
"join = db.books_tags_link, db.Books.id == db.books_tags_link.c.book, db.Tags\n",
"order = [db.Series.name.asc()] if order == 'asc' else [db.Series.name.desc()]\n",
"if sort == 'publishers':\n",
"if search:\n",
"if search:\n",
"join = (db.books_series_link, db.Books.id == db.books_series_link.c.book,\n db.Series)\n",
"order = [db.Publishers.name.asc()] if order == 'asc' else [db.Publishers.\n name.desc()]\n",
"if sort == 'authors':\n",
"books = calibre_db.search_query(search).all()\n",
"books = calibre_db.session.query(db.Books).filter(calibre_db.common_filters()\n ).all()\n",
"entries, filtered_count, __ = calibre_db.get_search_results(search, off, [\n order, ''], limit, *join)\n",
"entries, __, __ = calibre_db.fill_indexpage(int(off) / int(limit) + 1,\n limit, db.Books, True, order, *join)\n",
"join = (db.books_publishers_link, db.Books.id == db.books_publishers_link.c\n .book, db.Publishers)\n",
"order = [db.Authors.name.asc(), db.Series.name, db.Books.series_index\n ] if order == 'asc' else [db.Authors.name.desc(), db.Series.name.desc(),\n db.Books.series_index.desc()]\n",
"if sort == 'languages':\n",
"filtered_count = len(books)\n",
"entries = calibre_db.get_checkbox_sorted(books, state, off, limit, order)\n",
"for entry in entries:\n",
"join = (db.books_authors_link, db.Books.id == db.books_authors_link.c.book,\n db.Authors, db.books_series_link, db.Books.id == db.books_series_link.c\n .book, db.Series)\n",
"order = [db.Languages.lang_code.asc()] if order == 'asc' else [db.Languages\n .lang_code.desc()]\n",
"if order and sort in ['sort', 'title', 'authors_sort', 'series_index']:\n",
"for index in range(0, len(entry.languages)):\n",
"table_entries = {'totalNotFiltered': total_count, 'total': filtered_count,\n 'rows': entries}\n",
"join = (db.books_languages_link, db.Books.id == db.books_languages_link.c.\n book, db.Languages)\n",
"order = [text(sort + ' ' + order)]\n",
"if not state:\n",
"entry.languages[index].language_name = isoLanguages.get_language_name(\n get_locale(), entry.languages[index].lang_code)\n",
"js_list = json.dumps(table_entries, cls=db.AlchemyEncoder)\n",
"order = [db.Books.timestamp.desc()]\n",
"response = make_response(js_list)\n",
"response.headers['Content-Type'] = 'application/json; charset=utf-8'\n",
"return response\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Condition",
"Condition",
"Assign'",
"Assign'",
"Condition",
"Condition",
"Condition",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"For",
"Assign'",
"Assign'",
"Condition",
"For",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Return'"
] |
[
"def __init__(self, VAR_3=None):...\n",
"VAR_28 = {'class': 'vCommaSeparatedIntegerField'}\n",
"if VAR_3 is not None:\n",
"VAR_28.update(VAR_3)\n",
"super(CLASS_16, self).__init__(VAR_3=final_attrs)\n"
] | [
"def __init__(self, attrs=None):...\n",
"final_attrs = {'class': 'vCommaSeparatedIntegerField'}\n",
"if attrs is not None:\n",
"final_attrs.update(attrs)\n",
"super(AdminCommaSeparatedIntegerFieldWidget, self).__init__(attrs=final_attrs)\n"
] | [
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Condition",
"Expr'",
"Expr'"
] |
[
"def FUNC_32(VAR_87):...\n",
"VAR_86.update(to_bytes(VAR_87, encoding='utf-8', errors='replace'))\n"
] | [
"def hash_update(value):...\n",
"hash.update(to_bytes(value, encoding='utf-8', errors='replace'))\n"
] | [
0,
0
] | [
"FunctionDef'",
"Expr'"
] |
[
"def FUNC_9(self):...\n",
"if not self.get('roles') and self.is_standard == 'No':\n",
"VAR_5 = frappe.get_meta(self.ref_doctype)\n",
"if not VAR_5.istable:\n",
"VAR_35 = [{'role': d.role} for d in VAR_5.permissions if d.permlevel == 0]\n",
"self.set('roles', VAR_35)\n"
] | [
"def set_doctype_roles(self):...\n",
"if not self.get('roles') and self.is_standard == 'No':\n",
"meta = frappe.get_meta(self.ref_doctype)\n",
"if not meta.istable:\n",
"roles = [{'role': d.role} for d in meta.permissions if d.permlevel == 0]\n",
"self.set('roles', roles)\n"
] | [
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Expr'"
] |
[
"def FUNC_10(VAR_0, VAR_1):...\n",
"self.query_handlers[VAR_0] = VAR_1\n"
] | [
"def register_query_handler(query_type, handler):...\n",
"self.query_handlers[query_type] = handler\n"
] | [
0,
0
] | [
"FunctionDef'",
"Assign'"
] |
[
"def FUNC_26(VAR_9, VAR_13, VAR_10):...\n",
""
] | [
"def render_language_books(page, name, order):...\n",
""
] | [
0,
0
] | [
"FunctionDef'",
"Condition"
] |
[
"def __unicode__(self):...\n",
"return f\"<Text '{self.offendingText}' shows up in both lt and local-lt>\"\n"
] | [
"def __unicode__(self):...\n",
"return f\"<Text '{self.offendingText}' shows up in both lt and local-lt>\"\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
[
"def FUNC_0(self, VAR_2, VAR_3):...\n",
"VAR_4 = self.setup_test_homeserver('server', http_client=None)\n",
"self.handler = VAR_4.get_device_handler()\n",
"self.registration = VAR_4.get_registration_handler()\n",
"self.auth = VAR_4.get_auth()\n",
"self.store = VAR_4.get_datastore()\n",
"return VAR_4\n"
] | [
"def make_homeserver(self, reactor, clock):...\n",
"hs = self.setup_test_homeserver('server', http_client=None)\n",
"self.handler = hs.get_device_handler()\n",
"self.registration = hs.get_registration_handler()\n",
"self.auth = hs.get_auth()\n",
"self.store = hs.get_datastore()\n",
"return hs\n"
] | [
0,
4,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Return'"
] |
[
"@VAR_1.route('/ajax/delete/<int:book_id>')...\n",
"return Response(FUNC_13(VAR_15, '', True), mimetype='application/json')\n"
] | [
"@editbook.route('/ajax/delete/<int:book_id>')...\n",
"return Response(delete_book(book_id, '', True), mimetype='application/json')\n"
] | [
0,
0
] | [
"Condition",
"Return'"
] |
[
"def FUNC_0(self):...\n",
"VAR_5 = super().default_config()\n",
"VAR_5['send_federation'] = False\n",
"return VAR_5\n"
] | [
"def default_config(self):...\n",
"conf = super().default_config()\n",
"conf['send_federation'] = False\n",
"return conf\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Return'"
] |
[
"async def FUNC_17(VAR_46: 'TrainingDataImporter', VAR_21: Union[Path, Text]...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_18 = Path(VAR_21) / DEFAULT_CORE_SUBDIRECTORY_NAME\n",
"VAR_52 = await VAR_46.get_domain()\n",
"VAR_52.persist(VAR_18 / DEFAULT_DOMAIN_PATH)\n"
] | [
"async def update_model_with_new_domain(importer: 'TrainingDataImporter',...\n",
"\"\"\"docstring\"\"\"\n",
"model_path = Path(unpacked_model_path) / DEFAULT_CORE_SUBDIRECTORY_NAME\n",
"domain = await importer.get_domain()\n",
"domain.persist(model_path / DEFAULT_DOMAIN_PATH)\n"
] | [
0,
0,
0,
0,
0
] | [
"Condition",
"Docstring",
"Assign'",
"Assign'",
"Expr'"
] |
[
"def FUNC_11(self):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_6 = self._make_request(b'GET', b'/res/')\n",
"self.assertEqual(VAR_6.result['code'], b'200')\n",
"self.assertEqual(VAR_6.result['body'], b'/res/')\n"
] | [
"def test_known_request(self):...\n",
"\"\"\"docstring\"\"\"\n",
"channel = self._make_request(b'GET', b'/res/')\n",
"self.assertEqual(channel.result['code'], b'200')\n",
"self.assertEqual(channel.result['body'], b'/res/')\n"
] | [
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Expr'",
"Expr'"
] |
[
"def __str__(self):...\n",
"return '%s' % self.title\n"
] | [
"def __str__(self):...\n",
"return '%s' % self.title\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
[
"@wraps(VAR_10)...\n",
"VAR_1 = self.__class__\n",
"if VAR_2:\n",
"VAR_42 = [arg[0] for arg in FUNC_3(VAR_1)]\n",
"VAR_39 = vars(FUNC_2(VAR_1))\n",
"VAR_3.update(dict(zip(VAR_42, VAR_2)))\n",
"VAR_3 = dict(list(VAR_39.items()) + list(VAR_3.items()))\n",
"return VAR_10(self, **kwargs)\n"
] | [
"@wraps(fn)...\n",
"cls = self.__class__\n",
"if args:\n",
"cls_arg_names = [arg[0] for arg in get_init_arguments_and_types(cls)]\n",
"env_variables = vars(parse_env_variables(cls))\n",
"kwargs.update(dict(zip(cls_arg_names, args)))\n",
"kwargs = dict(list(env_variables.items()) + list(kwargs.items()))\n",
"return fn(self, **kwargs)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Return'"
] |
[
"def FUNC_7(self):...\n",
"return self.spec\n"
] | [
"def get_repo_url(self):...\n",
"return self.spec\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
[
"from __future__ import unicode_literals\n",
"from django import template\n",
"from djblets.gravatars import get_gravatar_url, get_gravatar_url_for_email\n",
"from djblets.util.decorators import basictag\n",
"VAR_0 = template.Library()\n",
"@VAR_0.tag...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_5 = get_gravatar_url(VAR_1['request'], VAR_2, VAR_3)\n",
"if VAR_5:\n",
"return '<img src=\"%s\" width=\"%s\" height=\"%s\" alt=\"%s\" class=\"gravatar\"/>' % (\n VAR_5, VAR_3, VAR_3, VAR_2.get_full_name() or VAR_2.username)\n",
"return ''\n"
] | [
"from __future__ import unicode_literals\n",
"from django import template\n",
"from djblets.gravatars import get_gravatar_url, get_gravatar_url_for_email\n",
"from djblets.util.decorators import basictag\n",
"register = template.Library()\n",
"@register.tag...\n",
"\"\"\"docstring\"\"\"\n",
"url = get_gravatar_url(context['request'], user, size)\n",
"if url:\n",
"return '<img src=\"%s\" width=\"%s\" height=\"%s\" alt=\"%s\" class=\"gravatar\"/>' % (\n url, size, size, user.get_full_name() or user.username)\n",
"return ''\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
2,
0
] | [
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"Assign'",
"Condition",
"Docstring",
"Assign'",
"Condition",
"Return'",
"Return'"
] |
[
"def FUNC_140(self):...\n",
"\"\"\"docstring\"\"\"\n",
"if not self.wiki_menu_items and self.settings.controller and self.settings.function:\n",
"self.wiki_menu_items = self.menu(self.settings.controller, self.settings.\n function)\n",
"VAR_263.response.menu += self.wiki_menu_items\n"
] | [
"def automenu(self):...\n",
"\"\"\"docstring\"\"\"\n",
"if not self.wiki_menu_items and self.settings.controller and self.settings.function:\n",
"self.wiki_menu_items = self.menu(self.settings.controller, self.settings.\n function)\n",
"current.response.menu += self.wiki_menu_items\n"
] | [
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Condition",
"Assign'",
"AugAssign'"
] |
[
"@app.route('/plugin_assets/<string:name>/<path:filename>')...\n",
"return redirect(url_for('plugin.' + VAR_23 + '.static', VAR_32=filename))\n"
] | [
"@app.route('/plugin_assets/<string:name>/<path:filename>')...\n",
"return redirect(url_for('plugin.' + name + '.static', filename=filename))\n"
] | [
0,
0
] | [
"Condition",
"Return'"
] |
[
"def FUNC_1(self, VAR_3, VAR_4, VAR_11):...\n",
"self.user_id = self.register_user('user', 'password')\n",
"self.tok = self.login('user', 'password')\n",
"self.room_id = self.helper.create_room_as(self.user_id, VAR_52=self.tok,\n is_public=False)\n",
"self.other_user_id = self.register_user('user2', 'password')\n",
"self.other_tok = self.login('user2', 'password')\n",
"self.helper.invite(self.room_id, self.user_id, self.other_user_id, VAR_52=\n self.tok)\n",
"self.helper.join(self.room_id, self.other_user_id, VAR_52=self.other_tok)\n"
] | [
"def prepare(self, reactor, clock, homeserver):...\n",
"self.user_id = self.register_user('user', 'password')\n",
"self.tok = self.login('user', 'password')\n",
"self.room_id = self.helper.create_room_as(self.user_id, tok=self.tok,\n is_public=False)\n",
"self.other_user_id = self.register_user('user2', 'password')\n",
"self.other_tok = self.login('user2', 'password')\n",
"self.helper.invite(self.room_id, self.user_id, self.other_user_id, tok=self.tok\n )\n",
"self.helper.join(self.room_id, self.other_user_id, tok=self.other_tok)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Expr'"
] |
[
"@app.after_request...\n",
"VAR_1.headers['Content-Security-Policy'\n ] = \"default-src 'self' 'unsafe-inline' 'unsafe-eval';\"\n",
"if request.endpoint == 'editbook.edit_book':\n",
"VAR_1.headers['Content-Security-Policy'] += 'img-src * data:'\n",
"VAR_1.headers['X-Content-Type-Options'] = 'nosniff'\n",
"VAR_1.headers['X-Frame-Options'] = 'SAMEORIGIN'\n",
"VAR_1.headers['X-XSS-Protection'] = '1; mode=block'\n",
"VAR_1.headers['Strict-Transport-Security'\n ] = 'max-age=31536000; includeSubDomains'\n",
"return VAR_1\n"
] | [
"@app.after_request...\n",
"resp.headers['Content-Security-Policy'\n ] = \"default-src 'self' 'unsafe-inline' 'unsafe-eval';\"\n",
"if request.endpoint == 'editbook.edit_book':\n",
"resp.headers['Content-Security-Policy'] += 'img-src * data:'\n",
"resp.headers['X-Content-Type-Options'] = 'nosniff'\n",
"resp.headers['X-Frame-Options'] = 'SAMEORIGIN'\n",
"resp.headers['X-XSS-Protection'] = '1; mode=block'\n",
"resp.headers['Strict-Transport-Security'\n ] = 'max-age=31536000; includeSubDomains'\n",
"return resp\n"
] | [
0,
3,
0,
3,
0,
0,
0,
0,
0
] | [
"Condition",
"Assign'",
"Condition",
"AugAssign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Return'"
] |
[
"def FUNC_23(self):...\n",
"VAR_5 = self._makeContext()\n",
"self.assertEqual(VAR_5.evaluate('path: '), None)\n"
] | [
"def test_empty_path_expression_explicit_with_trailing_whitespace(self):...\n",
"ec = self._makeContext()\n",
"self.assertEqual(ec.evaluate('path: '), None)\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Expr'"
] |
[
"def FUNC_1(VAR_1, VAR_2=False):...\n",
"\"\"\"docstring\"\"\"\n",
"for hook in VAR_25.get_hooks('override_whitelisted_methods', {}).get(VAR_1, []\n",
"VAR_1 = hook\n",
"if run_server_script_api(VAR_1):\n",
"return None\n",
"VAR_3 = FUNC_11(VAR_1)\n",
"if VAR_25.local.conf.developer_mode:\n",
"if VAR_2:\n",
"VAR_25.respond_as_web_page(title='Invalid Method', html='Method not found',\n indicator_color='red', http_status_code=404)\n",
"return\n",
"VAR_3 = VAR_3.queue\n",
"FUNC_3(VAR_3)\n",
"FUNC_2(VAR_3)\n",
"return VAR_25.call(VAR_3, **frappe.form_dict)\n"
] | [
"def execute_cmd(cmd, from_async=False):...\n",
"\"\"\"docstring\"\"\"\n",
"for hook in frappe.get_hooks('override_whitelisted_methods', {}).get(cmd, []):\n",
"cmd = hook\n",
"if run_server_script_api(cmd):\n",
"return None\n",
"method = get_attr(cmd)\n",
"if frappe.local.conf.developer_mode:\n",
"if from_async:\n",
"frappe.respond_as_web_page(title='Invalid Method', html='Method not found',\n indicator_color='red', http_status_code=404)\n",
"return\n",
"method = method.queue\n",
"is_whitelisted(method)\n",
"is_valid_http_method(method)\n",
"return frappe.call(method, **frappe.form_dict)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
2,
2,
0
] | [
"FunctionDef'",
"Docstring",
"For",
"Assign'",
"Condition",
"Return'",
"Assign'",
"Condition",
"Condition",
"Expr'",
"Return'",
"Assign'",
"Expr'",
"Expr'",
"Return'"
] |
[
"def FUNC_19(self):...\n",
"\"\"\"docstring\"\"\"\n",
"if not self.assigned_to:\n",
"return _('Unassigned')\n",
"if self.assigned_to.get_full_name():\n",
"return self.assigned_to.get_full_name()\n",
"return self.assigned_to.get_username()\n"
] | [
"def _get_assigned_to(self):...\n",
"\"\"\"docstring\"\"\"\n",
"if not self.assigned_to:\n",
"return _('Unassigned')\n",
"if self.assigned_to.get_full_name():\n",
"return self.assigned_to.get_full_name()\n",
"return self.assigned_to.get_username()\n"
] | [
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Condition",
"Return'",
"Condition",
"Return'",
"Return'"
] |
[
"def FUNC_9(VAR_18, VAR_19, VAR_20, VAR_21):...\n",
"VAR_69 = settings().getBoolean(['server', 'firstRun'])\n",
"VAR_70 = {}\n",
"for VAR_131 in LOCALES:\n",
"VAR_41 = [permission.as_dict() for permission in Permissions.all()]\n",
"VAR_70[VAR_131.language] = {'language': VAR_131.language, 'display':\n VAR_131.display_name, 'english': VAR_131.english_name}\n",
"VAR_0.exception('Error while collecting available locales')\n",
"VAR_71 = list(sorted(full_extension_tree().keys()))\n",
"VAR_72 = list(map(lambda ext: f'.{ext}', get_all_extensions()))\n",
"VAR_43 = {'debug': debug, 'firstRun': VAR_69, 'version': {'number': VERSION,\n 'display': DISPLAY_VERSION, 'branch': BRANCH}, 'python_version':\n get_python_version_string(), 'templates': VAR_18, 'pluginNames': VAR_19,\n 'locales': VAR_70, 'permissions': VAR_41, 'supportedFiletypes': VAR_71,\n 'supportedExtensions': VAR_72}\n",
"VAR_43.update(VAR_20)\n",
"return VAR_43\n"
] | [
"def _get_render_kwargs(templates, plugin_names, plugin_vars, now):...\n",
"first_run = settings().getBoolean(['server', 'firstRun'])\n",
"locales = {}\n",
"for loc in LOCALES:\n",
"permissions = [permission.as_dict() for permission in Permissions.all()]\n",
"locales[loc.language] = {'language': loc.language, 'display': loc.\n display_name, 'english': loc.english_name}\n",
"_logger.exception('Error while collecting available locales')\n",
"filetypes = list(sorted(full_extension_tree().keys()))\n",
"extensions = list(map(lambda ext: f'.{ext}', get_all_extensions()))\n",
"render_kwargs = {'debug': debug, 'firstRun': first_run, 'version': {\n 'number': VERSION, 'display': DISPLAY_VERSION, 'branch': BRANCH},\n 'python_version': get_python_version_string(), 'templates': templates,\n 'pluginNames': plugin_names, 'locales': locales, 'permissions':\n permissions, 'supportedFiletypes': filetypes, 'supportedExtensions':\n extensions}\n",
"render_kwargs.update(plugin_vars)\n",
"return render_kwargs\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"For",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Return'"
] |
[
"def FUNC_49():...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_3 = FUNC_5()\n",
"update_all_languages(apath(VAR_3, VAR_122=request))\n",
"session.flash = T('Language files (static strings) updated')\n",
"redirect(URL('design', VAR_98=app, VAR_157='languages'))\n"
] | [
"def update_languages():...\n",
"\"\"\"docstring\"\"\"\n",
"app = get_app()\n",
"update_all_languages(apath(app, r=request))\n",
"session.flash = T('Language files (static strings) updated')\n",
"redirect(URL('design', args=app, anchor='languages'))\n"
] | [
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Expr'",
"Assign'",
"Expr'"
] |
[
"def FUNC_22(self):...\n",
"assert join('/', '/') == '/'\n",
"assert join(None, '/') == '/'\n",
"assert join('/', None) == '/'\n"
] | [
"def test_it_doesnt_eat_single_slash(self):...\n",
"assert join('/', '/') == '/'\n",
"assert join(None, '/') == '/'\n",
"assert join('/', None) == '/'\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"Assert'",
"Assert'",
"Assert'"
] |
[
"def FUNC_0(VAR_0):...\n",
"VAR_1 = os.path.dirname(os.path.abspath(__file__))\n",
"return open(os.path.join(VAR_1, VAR_0)).read()\n"
] | [
"def _read(fname):...\n",
"here = os.path.dirname(os.path.abspath(__file__))\n",
"return open(os.path.join(here, fname)).read()\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Return'"
] |
[
"def FUNC_15(self):...\n",
"VAR_26 = np.array([[1], [2]])\n",
"VAR_31 = os.path.join(test.get_temp_dir(), 'input.npz')\n",
"np.savez(VAR_31, VAR_53=x0)\n",
"VAR_20 = 'x=' + VAR_31 + '[a];y=' + VAR_31\n",
"VAR_30 = saved_model_cli.load_inputs_from_input_arg_string(VAR_20, '', '')\n",
"self.assertTrue(np.all(VAR_30['x'] == VAR_26))\n",
"self.assertTrue(np.all(VAR_30['y'] == VAR_26))\n"
] | [
"def testInputParserNPZ(self):...\n",
"x0 = np.array([[1], [2]])\n",
"input_path = os.path.join(test.get_temp_dir(), 'input.npz')\n",
"np.savez(input_path, a=x0)\n",
"input_str = 'x=' + input_path + '[a];y=' + input_path\n",
"feed_dict = saved_model_cli.load_inputs_from_input_arg_string(input_str, '', ''\n )\n",
"self.assertTrue(np.all(feed_dict['x'] == x0))\n",
"self.assertTrue(np.all(feed_dict['y'] == x0))\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Assign'",
"Expr'",
"Expr'"
] |
[
"def FUNC_26(VAR_37, VAR_38):...\n",
"VAR_76 = FUNC_25(VAR_38 + '/v1/tunnel-request' if VAR_38 is not None else VAR_3\n )\n",
"if VAR_76 and VAR_76.code == 200:\n",
"VAR_97 = json.loads(VAR_76.read().decode('utf-8'))[0]\n",
"return create_tunnel(VAR_97, VAR_2, VAR_37)\n"
] | [
"def setup_tunnel(local_server_port, endpoint):...\n",
"response = url_request(endpoint + '/v1/tunnel-request' if endpoint is not\n None else GRADIO_API_SERVER)\n",
"if response and response.code == 200:\n",
"payload = json.loads(response.read().decode('utf-8'))[0]\n",
"return create_tunnel(payload, LOCALHOST_NAME, local_server_port)\n"
] | [
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Condition",
"Assign'",
"Return'"
] |
[
"def FUNC_5(self, VAR_24):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_79 = []\n",
"VAR_79.append(('/deprecatedwidgets/(.*)', CLASS_0))\n",
"VAR_79.extend(FUNC_1('tree.handlers'))\n",
"VAR_79.extend([('/login', VAR_24['login_handler_class'])])\n",
"VAR_79.extend([('/logout', VAR_24['logout_handler_class'])])\n",
"VAR_79.extend(FUNC_1('files.handlers'))\n",
"VAR_79.extend(FUNC_1('notebook.handlers'))\n",
"VAR_79.extend(FUNC_1('nbconvert.handlers'))\n",
"VAR_79.extend(FUNC_1('kernelspecs.handlers'))\n",
"VAR_79.extend(FUNC_1('edit.handlers'))\n",
"VAR_79.extend(FUNC_1('services.api.handlers'))\n",
"VAR_79.extend(FUNC_1('services.config.handlers'))\n",
"VAR_79.extend(FUNC_1('services.kernels.handlers'))\n",
"VAR_79.extend(FUNC_1('services.contents.handlers'))\n",
"VAR_79.extend(FUNC_1('services.sessions.handlers'))\n",
"VAR_79.extend(FUNC_1('services.nbconvert.handlers'))\n",
"VAR_79.extend(FUNC_1('services.kernelspecs.handlers'))\n",
"VAR_79.extend(FUNC_1('services.security.handlers'))\n",
"import ipywidgets\n",
"app_log.warn('ipywidgets package not installed. Widgets are unavailable.')\n",
"VAR_79.append(('/nbextensions/(.*)', FileFindHandler, {'path': VAR_24[\n 'nbextensions_path'], 'no_cache_paths': ['/']}))\n",
"VAR_79.append(('/nbextensions/widgets/(.*)', FileFindHandler, {'path':\n ipywidgets.find_static_assets(), 'no_cache_paths': ['/']}))\n",
"VAR_79.append(('/custom/(.*)', FileFindHandler, {'path': VAR_24[\n 'static_custom_path'], 'no_cache_paths': ['/']}))\n",
"VAR_79.extend(FUNC_1('base.handlers'))\n",
"VAR_79.append(('/?', web.RedirectHandler, {'url': VAR_24['default_url'],\n 'permanent': False}))\n",
"VAR_85 = []\n",
"for handler in VAR_79:\n",
"VAR_98 = url_path_join(VAR_24['base_url'], handler[0])\n",
"VAR_85.append(('(.*)', Template404))\n",
"VAR_99 = tuple([VAR_98] + list(handler[1:]))\n",
"return VAR_85\n",
"VAR_85.append(VAR_99)\n"
] | [
"def init_handlers(self, settings):...\n",
"\"\"\"docstring\"\"\"\n",
"handlers = []\n",
"handlers.append(('/deprecatedwidgets/(.*)', DeprecationHandler))\n",
"handlers.extend(load_handlers('tree.handlers'))\n",
"handlers.extend([('/login', settings['login_handler_class'])])\n",
"handlers.extend([('/logout', settings['logout_handler_class'])])\n",
"handlers.extend(load_handlers('files.handlers'))\n",
"handlers.extend(load_handlers('notebook.handlers'))\n",
"handlers.extend(load_handlers('nbconvert.handlers'))\n",
"handlers.extend(load_handlers('kernelspecs.handlers'))\n",
"handlers.extend(load_handlers('edit.handlers'))\n",
"handlers.extend(load_handlers('services.api.handlers'))\n",
"handlers.extend(load_handlers('services.config.handlers'))\n",
"handlers.extend(load_handlers('services.kernels.handlers'))\n",
"handlers.extend(load_handlers('services.contents.handlers'))\n",
"handlers.extend(load_handlers('services.sessions.handlers'))\n",
"handlers.extend(load_handlers('services.nbconvert.handlers'))\n",
"handlers.extend(load_handlers('services.kernelspecs.handlers'))\n",
"handlers.extend(load_handlers('services.security.handlers'))\n",
"import ipywidgets\n",
"app_log.warn('ipywidgets package not installed. Widgets are unavailable.')\n",
"handlers.append(('/nbextensions/(.*)', FileFindHandler, {'path': settings[\n 'nbextensions_path'], 'no_cache_paths': ['/']}))\n",
"handlers.append(('/nbextensions/widgets/(.*)', FileFindHandler, {'path':\n ipywidgets.find_static_assets(), 'no_cache_paths': ['/']}))\n",
"handlers.append(('/custom/(.*)', FileFindHandler, {'path': settings[\n 'static_custom_path'], 'no_cache_paths': ['/']}))\n",
"handlers.extend(load_handlers('base.handlers'))\n",
"handlers.append(('/?', web.RedirectHandler, {'url': settings['default_url'],\n 'permanent': False}))\n",
"new_handlers = []\n",
"for handler in handlers:\n",
"pattern = url_path_join(settings['base_url'], handler[0])\n",
"new_handlers.append(('(.*)', Template404))\n",
"new_handler = tuple([pattern] + list(handler[1:]))\n",
"return new_handlers\n",
"new_handlers.append(new_handler)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Import'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Assign'",
"For",
"Assign'",
"Expr'",
"Assign'",
"Return'",
"Expr'"
] |
[
"@app.after_request...\n",
"VAR_1.headers['Content-Security-Policy'] = (\"default-src 'self'\" + ''.join(\n [(' ' + host) for host in config.config_trustedhosts.strip().split(',')\n ]) +\n \" 'unsafe-inline' 'unsafe-eval'; font-src 'self' data:; img-src 'self' data:\"\n )\n",
"if request.endpoint == 'editbook.edit_book' or config.config_use_google_drive:\n",
"VAR_1.headers['Content-Security-Policy'] += ' *'\n",
"if request.endpoint == 'web.read_book':\n",
"VAR_1.headers['X-Content-Type-Options'] = 'nosniff'\n",
"VAR_1.headers['Content-Security-Policy'\n ] += \" blob:;style-src-elem 'self' blob: 'unsafe-inline';\"\n",
"VAR_1.headers['X-Frame-Options'] = 'SAMEORIGIN'\n",
"VAR_1.headers['X-XSS-Protection'] = '1; mode=block'\n",
"VAR_1.headers['Strict-Transport-Security'] = 'max-age=31536000;'\n",
"return VAR_1\n"
] | [
"@app.after_request...\n",
"resp.headers['Content-Security-Policy'] = (\"default-src 'self'\" + ''.join([\n (' ' + host) for host in config.config_trustedhosts.strip().split(',')]\n ) +\n \" 'unsafe-inline' 'unsafe-eval'; font-src 'self' data:; img-src 'self' data:\"\n )\n",
"if request.endpoint == 'editbook.edit_book' or config.config_use_google_drive:\n",
"resp.headers['Content-Security-Policy'] += ' *'\n",
"if request.endpoint == 'web.read_book':\n",
"resp.headers['X-Content-Type-Options'] = 'nosniff'\n",
"resp.headers['Content-Security-Policy'\n ] += \" blob:;style-src-elem 'self' blob: 'unsafe-inline';\"\n",
"resp.headers['X-Frame-Options'] = 'SAMEORIGIN'\n",
"resp.headers['X-XSS-Protection'] = '1; mode=block'\n",
"resp.headers['Strict-Transport-Security'] = 'max-age=31536000;'\n",
"return resp\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Assign'",
"Condition",
"AugAssign'",
"Condition",
"Assign'",
"AugAssign'",
"Assign'",
"Assign'",
"Assign'",
"Return'"
] |
[
"@VAR_0.route('/plugins')...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_110 = VAR_8\n",
"flash(str(err), 'danger')\n",
"return render_template('plugins.html', VAR_110=plugins_info)\n",
"return redirect(url_for('dashboard'))\n"
] | [
"@gui.route('/plugins')...\n",
"\"\"\"docstring\"\"\"\n",
"plugins_info = plugins\n",
"flash(str(err), 'danger')\n",
"return render_template('plugins.html', plugins_info=plugins_info)\n",
"return redirect(url_for('dashboard'))\n"
] | [
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Docstring",
"Assign'",
"Expr'",
"Return'",
"Return'"
] |
[
"def FUNC_121(*VAR_79, **VAR_42):...\n",
"if VAR_13.read_from_replica:\n",
"FUNC_6()\n",
"VAR_225 = VAR_129(*VAR_79, **get_newargs(fn, kwargs))\n",
"if VAR_1 and hasattr(VAR_1, 'primary_db'):\n",
"return VAR_225\n",
"VAR_1.db.close()\n",
"VAR_1.db = VAR_1.primary_db\n"
] | [
"def wrapper_fn(*args, **kwargs):...\n",
"if conf.read_from_replica:\n",
"connect_replica()\n",
"retval = fn(*args, **get_newargs(fn, kwargs))\n",
"if local and hasattr(local, 'primary_db'):\n",
"return retval\n",
"local.db.close()\n",
"local.db = local.primary_db\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Condition",
"Expr'",
"Assign'",
"Condition",
"Return'",
"Expr'",
"Assign'"
] |
[
"def FUNC_2(self):...\n",
"VAR_9 = test.test_src_dir_path(VAR_0)\n",
"self.parser = saved_model_cli.create_parser()\n",
"VAR_10 = self.parser.parse_args(['show', '--dir', VAR_9, '--all'])\n",
"saved_model_cli.show(VAR_10)\n",
"VAR_11 = out.getvalue().strip()\n",
"VAR_12 = 'string'\n",
"self.maxDiff = None\n",
"self.assertMultiLineEqual(VAR_11, VAR_12)\n",
"self.assertEqual(err.getvalue().strip(), '')\n"
] | [
"def testShowCommandAll(self):...\n",
"base_path = test.test_src_dir_path(SAVED_MODEL_PATH)\n",
"self.parser = saved_model_cli.create_parser()\n",
"args = self.parser.parse_args(['show', '--dir', base_path, '--all'])\n",
"saved_model_cli.show(args)\n",
"output = out.getvalue().strip()\n",
"exp_out = \"\"\"MetaGraphDef with tag-set: 'serve' contains the following SignatureDefs:\n\nsignature_def['classify_x2_to_y3']:\n The given SavedModel SignatureDef contains the following input(s):\n inputs['inputs'] tensor_info:\n dtype: DT_FLOAT\n shape: (-1, 1)\n name: x2:0\n The given SavedModel SignatureDef contains the following output(s):\n outputs['scores'] tensor_info:\n dtype: DT_FLOAT\n shape: (-1, 1)\n name: y3:0\n Method name is: tensorflow/serving/classify\n\nsignature_def['classify_x_to_y']:\n The given SavedModel SignatureDef contains the following input(s):\n inputs['inputs'] tensor_info:\n dtype: DT_STRING\n shape: unknown_rank\n name: tf_example:0\n The given SavedModel SignatureDef contains the following output(s):\n outputs['scores'] tensor_info:\n dtype: DT_FLOAT\n shape: (-1, 1)\n name: y:0\n Method name is: tensorflow/serving/classify\n\nsignature_def['regress_x2_to_y3']:\n The given SavedModel SignatureDef contains the following input(s):\n inputs['inputs'] tensor_info:\n dtype: DT_FLOAT\n shape: (-1, 1)\n name: x2:0\n The given SavedModel SignatureDef contains the following output(s):\n outputs['outputs'] tensor_info:\n dtype: DT_FLOAT\n shape: (-1, 1)\n name: y3:0\n Method name is: tensorflow/serving/regress\n\nsignature_def['regress_x_to_y']:\n The given SavedModel SignatureDef contains the following input(s):\n inputs['inputs'] tensor_info:\n dtype: DT_STRING\n shape: unknown_rank\n name: tf_example:0\n The given SavedModel SignatureDef contains the following output(s):\n outputs['outputs'] tensor_info:\n dtype: DT_FLOAT\n shape: (-1, 1)\n name: y:0\n Method name is: tensorflow/serving/regress\n\nsignature_def['regress_x_to_y2']:\n The given SavedModel SignatureDef contains the following input(s):\n inputs['inputs'] tensor_info:\n dtype: DT_STRING\n shape: unknown_rank\n name: tf_example:0\n The given SavedModel SignatureDef contains the following output(s):\n outputs['outputs'] tensor_info:\n dtype: DT_FLOAT\n shape: (-1, 1)\n name: y2:0\n Method name is: tensorflow/serving/regress\n\nsignature_def['serving_default']:\n The given SavedModel SignatureDef contains the following input(s):\n inputs['x'] tensor_info:\n dtype: DT_FLOAT\n shape: (-1, 1)\n name: x:0\n The given SavedModel SignatureDef contains the following output(s):\n outputs['y'] tensor_info:\n dtype: DT_FLOAT\n shape: (-1, 1)\n name: y:0\n Method name is: tensorflow/serving/predict\"\"\"\n",
"self.maxDiff = None\n",
"self.assertMultiLineEqual(output, exp_out)\n",
"self.assertEqual(err.getvalue().strip(), '')\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Expr'"
] |
[
"def FUNC_18(VAR_5):...\n",
"if VAR_5 != '0':\n",
"if not calibre_db.session.query(db.Custom_Columns).filter(db.Custom_Columns\n",
"return True\n",
"return False\n"
] | [
"def check_valid_read_column(column):...\n",
"if column != '0':\n",
"if not calibre_db.session.query(db.Custom_Columns).filter(db.Custom_Columns\n",
"return True\n",
"return False\n"
] | [
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Condition",
"Condition",
"Return'",
"Return'"
] |
[
"def FUNC_27(VAR_10, VAR_11):...\n",
"VAR_69 = VAR_11()\n",
"if VAR_69 == ['']:\n",
"VAR_69 = []\n",
"if not VAR_10['add_element'] in VAR_69:\n",
"VAR_69 += [VAR_10['add_element']]\n",
"return ','.join(VAR_69)\n"
] | [
"def restriction_addition(element, list_func):...\n",
"elementlist = list_func()\n",
"if elementlist == ['']:\n",
"elementlist = []\n",
"if not element['add_element'] in elementlist:\n",
"elementlist += [element['add_element']]\n",
"return ','.join(elementlist)\n"
] | [
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Condition",
"Assign'",
"Condition",
"AugAssign'",
"Return'"
] |
[
"@defer.inlineCallbacks...\n",
"VAR_4 = defer.ensureDeferred(self.handler.set_displayname(self.frank,\n synapse.types.create_requester(self.bob), 'Frank Jr.'))\n",
"yield self.assertFailure(VAR_4, AuthError)\n"
] | [
"@defer.inlineCallbacks...\n",
"d = defer.ensureDeferred(self.handler.set_displayname(self.frank, synapse.\n types.create_requester(self.bob), 'Frank Jr.'))\n",
"yield self.assertFailure(d, AuthError)\n"
] | [
0,
0,
0
] | [
"Condition",
"Assign'",
"Expr'"
] |
[
"@property...\n",
"return ''.join(self.rawLines)\n"
] | [
"@property...\n",
"return ''.join(self.rawLines)\n"
] | [
0,
0
] | [
"Condition",
"Return'"
] |
[
"def __repr__(self):...\n",
"return '<User %r>' % self.name\n"
] | [
"def __repr__(self):...\n",
"return '<User %r>' % self.name\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
[
"def FUNC_35(self):...\n",
"self.login()\n",
"VAR_3 = self.client.post('/password_change/custom/', {'old_password':\n 'password', 'new_password1': 'password1', 'new_password2': 'password1'})\n",
"self.assertEqual(VAR_3.status_code, 302)\n",
"self.assertURLEqual(VAR_3.url, '/custom/')\n"
] | [
"def test_password_change_redirect_custom(self):...\n",
"self.login()\n",
"response = self.client.post('/password_change/custom/', {'old_password':\n 'password', 'new_password1': 'password1', 'new_password2': 'password1'})\n",
"self.assertEqual(response.status_code, 302)\n",
"self.assertURLEqual(response.url, '/custom/')\n"
] | [
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Expr'",
"Assign'",
"Expr'",
"Expr'"
] |
[
"@log_function...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_2 = FUNC_1('/groups/%s/roles/%s', VAR_30, VAR_34)\n",
"return self.client.get_json(VAR_5=destination, VAR_2=path, VAR_3={\n 'requester_user_id': requester_user_id}, VAR_15=True)\n"
] | [
"@log_function...\n",
"\"\"\"docstring\"\"\"\n",
"path = _create_v1_path('/groups/%s/roles/%s', group_id, role_id)\n",
"return self.client.get_json(destination=destination, path=path, args={\n 'requester_user_id': requester_user_id}, ignore_backoff=True)\n"
] | [
0,
0,
0,
0
] | [
"Condition",
"Docstring",
"Assign'",
"Return'"
] |
[
"def FUNC_74(self, VAR_44=None):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_77 = self.get_signature()\n",
"if file_lock.lock_exists(VAR_77):\n",
"VAR_96 = True\n",
"file_lock.create_lock(VAR_77)\n",
"if VAR_44:\n",
"for i in range(VAR_44):\n",
"if VAR_96:\n",
"time.sleep(1)\n",
"if not file_lock.lock_exists(VAR_77):\n",
"VAR_96 = False\n"
] | [
"def lock(self, timeout=None):...\n",
"\"\"\"docstring\"\"\"\n",
"signature = self.get_signature()\n",
"if file_lock.lock_exists(signature):\n",
"lock_exists = True\n",
"file_lock.create_lock(signature)\n",
"if timeout:\n",
"for i in range(timeout):\n",
"if lock_exists:\n",
"time.sleep(1)\n",
"if not file_lock.lock_exists(signature):\n",
"lock_exists = False\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Condition",
"Assign'",
"Expr'",
"Condition",
"For",
"Condition",
"Expr'",
"Condition",
"Assign'"
] |
[
"@defer.inlineCallbacks...\n",
"VAR_50 = yield VAR_4.verify_json_for_server(*VAR_2)\n",
"return VAR_50\n"
] | [
"@defer.inlineCallbacks...\n",
"rv1 = yield kr.verify_json_for_server(*args)\n",
"return rv1\n"
] | [
0,
0,
0
] | [
"Condition",
"Assign'",
"Return'"
] |
[
"def FUNC_25():...\n",
"return ''\n"
] | [
"def keepalive():...\n",
"return ''\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
[
"def FUNC_24(self, VAR_47: str) ->str:...\n",
"\"\"\"docstring\"\"\"\n",
"if not self.instances:\n",
"return 'master'\n",
"if len(self.instances) == 1:\n",
"return self.instances[0]\n",
"VAR_70 = sha256(VAR_47.encode('utf8')).digest()\n",
"VAR_71 = int.from_bytes(VAR_70, byteorder='little')\n",
"VAR_72 = VAR_71 % len(self.instances)\n",
"return self.instances[VAR_72]\n"
] | [
"def get_instance(self, key: str) ->str:...\n",
"\"\"\"docstring\"\"\"\n",
"if not self.instances:\n",
"return 'master'\n",
"if len(self.instances) == 1:\n",
"return self.instances[0]\n",
"dest_hash = sha256(key.encode('utf8')).digest()\n",
"dest_int = int.from_bytes(dest_hash, byteorder='little')\n",
"remainder = dest_int % len(self.instances)\n",
"return self.instances[remainder]\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Condition",
"Return'",
"Condition",
"Return'",
"Assign'",
"Assign'",
"Assign'",
"Return'"
] |
[
"def FUNC_11(VAR_37):...\n",
"VAR_4, VAR_39 = VAR_37\n",
"VAR_25.write(VAR_4)\n",
"return VAR_39\n"
] | [
"def write_to(r):...\n",
"data, response = r\n",
"output_stream.write(data)\n",
"return response\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Expr'",
"Return'"
] |
[
"@VAR_2.route('/get_tags_json', methods=['GET'])...\n",
"return calibre_db.get_typeahead(db.Tags, request.args.get('q'), tag_filter=\n tags_filters())\n"
] | [
"@web.route('/get_tags_json', methods=['GET'])...\n",
"return calibre_db.get_typeahead(db.Tags, request.args.get('q'), tag_filter=\n tags_filters())\n"
] | [
0,
0
] | [
"Condition",
"Return'"
] |
[
"@login_required()...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_142 = VAR_5.getObject('Image', VAR_26)\n",
"if VAR_142 is None:\n",
"VAR_181 = VAR_142.loadOriginalMetadata()\n",
"VAR_182 = ['[Global Metadata]']\n",
"VAR_182.extend([('%s=%s' % (kv[0], kv[1])) for kv in VAR_181[1]])\n",
"VAR_182.append('[Series Metadata]')\n",
"VAR_182.extend([('%s=%s' % (kv[0], kv[1])) for kv in VAR_181[2]])\n",
"VAR_183 = '\\n'.join(VAR_182)\n",
"VAR_174 = HttpResponse(VAR_183)\n",
"VAR_174['Content-Type'] = 'application/force-download'\n",
"VAR_174['Content-Length'] = len(VAR_183)\n",
"VAR_174['Content-Disposition'] = 'attachment; filename=Original_Metadata.txt'\n",
"return VAR_174\n"
] | [
"@login_required()...\n",
"\"\"\"docstring\"\"\"\n",
"image = conn.getObject('Image', imageId)\n",
"if image is None:\n",
"om = image.loadOriginalMetadata()\n",
"txtLines = ['[Global Metadata]']\n",
"txtLines.extend([('%s=%s' % (kv[0], kv[1])) for kv in om[1]])\n",
"txtLines.append('[Series Metadata]')\n",
"txtLines.extend([('%s=%s' % (kv[0], kv[1])) for kv in om[2]])\n",
"rspText = '\\n'.join(txtLines)\n",
"rsp = HttpResponse(rspText)\n",
"rsp['Content-Type'] = 'application/force-download'\n",
"rsp['Content-Length'] = len(rspText)\n",
"rsp['Content-Disposition'] = 'attachment; filename=Original_Metadata.txt'\n",
"return rsp\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Docstring",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Return'"
] |
[
"@VAR_1.url_value_preprocessor...\n",
"g.beancount_file_slug = VAR_9.pop('bfile', None) if VAR_9 else None\n",
"if not VAR_1.config.get('LEDGERS'):\n",
"FUNC_2()\n",
"if g.beancount_file_slug:\n",
"if g.beancount_file_slug not in VAR_1.config['LEDGERS']:\n",
"if not any(g.beancount_file_slug == FUNC_0(VAR_5) for VAR_5 in VAR_1.config\n",
"g.ledger = VAR_1.config['LEDGERS'][g.beancount_file_slug]\n",
"abort(404)\n",
"FUNC_1(VAR_1.config['LEDGERS'].values())\n",
"g.conversion = request.args.get('conversion', 'at_cost')\n",
"g.interval = Interval.get(request.args.get('interval', 'month'))\n"
] | [
"@app.url_value_preprocessor...\n",
"g.beancount_file_slug = values.pop('bfile', None) if values else None\n",
"if not app.config.get('LEDGERS'):\n",
"_load_file()\n",
"if g.beancount_file_slug:\n",
"if g.beancount_file_slug not in app.config['LEDGERS']:\n",
"if not any(g.beancount_file_slug == ledger_slug(ledger) for ledger in app.\n",
"g.ledger = app.config['LEDGERS'][g.beancount_file_slug]\n",
"abort(404)\n",
"update_ledger_slugs(app.config['LEDGERS'].values())\n",
"g.conversion = request.args.get('conversion', 'at_cost')\n",
"g.interval = Interval.get(request.args.get('interval', 'month'))\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Assign'",
"Condition",
"Expr'",
"Condition",
"Condition",
"For",
"Assign'",
"Expr'",
"Expr'",
"Assign'",
"Assign'"
] |
[
"@CLASS_4('backend-warning')...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_20 = jinja.render('backend-warning.html', distribution=version.\n distribution(), Distribution=version.Distribution, version=\n pkg_resources.parse_version, title='Legacy backend warning')\n",
"return 'text/html', VAR_20\n"
] | [
"@add_handler('backend-warning')...\n",
"\"\"\"docstring\"\"\"\n",
"src = jinja.render('backend-warning.html', distribution=version.\n distribution(), Distribution=version.Distribution, version=\n pkg_resources.parse_version, title='Legacy backend warning')\n",
"return 'text/html', src\n"
] | [
0,
0,
0,
0
] | [
"Condition",
"Docstring",
"Assign'",
"Return'"
] |
[
"def __init__(self, VAR_1=False, VAR_2=False, VAR_3=None, **VAR_4):...\n",
"\"\"\"docstring\"\"\"\n",
"super(CLASS_0, self).__init__(**kwargs)\n",
"self.ignore_login_fail = VAR_1\n",
"self.setGroupContext = VAR_2\n",
"self.login_redirect = VAR_3\n"
] | [
"def __init__(self, ignore_login_fail=False, setGroupContext=False,...\n",
"\"\"\"docstring\"\"\"\n",
"super(login_required, self).__init__(**kwargs)\n",
"self.ignore_login_fail = ignore_login_fail\n",
"self.setGroupContext = setGroupContext\n",
"self.login_redirect = login_redirect\n"
] | [
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Expr'",
"Assign'",
"Assign'",
"Assign'"
] |
[
"def FUNC_34(self):...\n",
"if VAR_53.session['admin']:\n",
"VAR_105 = self.userdb.getUserList()\n",
"return json.dumps({'time': 0, 'userlist': []})\n",
"for VAR_51 in VAR_105:\n",
"if VAR_51['id'] == VAR_53.session['userid']:\n",
"VAR_106 = lambda VAR_51: VAR_51['last_time_online']\n",
"VAR_51['deletable'] = False\n",
"VAR_119 = self.useroptions.forUser(VAR_51['id'])\n",
"VAR_105 = sorted(VAR_105, key=sortfunc, reverse=True)\n",
"VAR_120 = VAR_119.getOptionValue('last_time_online')\n",
"return json.dumps({'time': int(time.time()), 'userlist': VAR_105})\n",
"VAR_121 = VAR_119.getOptionValue('media.may_download')\n",
"VAR_51['last_time_online'] = VAR_120\n",
"VAR_51['may_download'] = VAR_121\n"
] | [
"def api_getuserlist(self):...\n",
"if cherrypy.session['admin']:\n",
"userlist = self.userdb.getUserList()\n",
"return json.dumps({'time': 0, 'userlist': []})\n",
"for user in userlist:\n",
"if user['id'] == cherrypy.session['userid']:\n",
"sortfunc = lambda user: user['last_time_online']\n",
"user['deletable'] = False\n",
"user_options = self.useroptions.forUser(user['id'])\n",
"userlist = sorted(userlist, key=sortfunc, reverse=True)\n",
"t = user_options.getOptionValue('last_time_online')\n",
"return json.dumps({'time': int(time.time()), 'userlist': userlist})\n",
"may_download = user_options.getOptionValue('media.may_download')\n",
"user['last_time_online'] = t\n",
"user['may_download'] = may_download\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Condition",
"Assign'",
"Return'",
"For",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Return'",
"Assign'",
"Assign'",
"Assign'"
] |
[
"@CLASS_4('tabs')...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_21 = collections.defaultdict(list)\n",
"for win_id, window in objreg.window_registry.items():\n",
"if sip.isdeleted(window):\n",
"VAR_20 = jinja.render('tabs.html', title='Tabs', tab_list_by_window=tabs)\n",
"VAR_33 = objreg.get('tabbed-browser', scope='window', window=win_id)\n",
"return 'text/html', VAR_20\n",
"for tab in VAR_33.widgets():\n",
"if tab.url() not in [QUrl('qute://tabs/'), QUrl('qute://tabs')]:\n",
"VAR_45 = tab.url().toDisplayString()\n",
"VAR_21[str(win_id)].append((tab.title(), VAR_45))\n"
] | [
"@add_handler('tabs')...\n",
"\"\"\"docstring\"\"\"\n",
"tabs = collections.defaultdict(list)\n",
"for win_id, window in objreg.window_registry.items():\n",
"if sip.isdeleted(window):\n",
"html = jinja.render('tabs.html', title='Tabs', tab_list_by_window=tabs)\n",
"tabbed_browser = objreg.get('tabbed-browser', scope='window', window=win_id)\n",
"return 'text/html', html\n",
"for tab in tabbed_browser.widgets():\n",
"if tab.url() not in [QUrl('qute://tabs/'), QUrl('qute://tabs')]:\n",
"urlstr = tab.url().toDisplayString()\n",
"tabs[str(win_id)].append((tab.title(), urlstr))\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Docstring",
"Assign'",
"For",
"Condition",
"Assign'",
"Assign'",
"Return'",
"For",
"Condition",
"Assign'",
"Expr'"
] |
[
"def FUNC_14(self, VAR_36):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_36 = self._substitute_comments('', VAR_36)\n",
"VAR_36 = VAR_36.replace('\\\\', '')\n",
"VAR_36 = VAR_6('', VAR_36)\n",
"VAR_36 = VAR_36.lower()\n",
"if 'javascript:' in VAR_36:\n",
"return True\n",
"if 'expression(' in VAR_36:\n",
"return True\n",
"if '@import' in VAR_36:\n",
"return True\n",
"if '</noscript' in VAR_36:\n",
"return True\n",
"if VAR_2(VAR_36):\n",
"return True\n",
"return False\n"
] | [
"def _has_sneaky_javascript(self, style):...\n",
"\"\"\"docstring\"\"\"\n",
"style = self._substitute_comments('', style)\n",
"style = style.replace('\\\\', '')\n",
"style = _substitute_whitespace('', style)\n",
"style = style.lower()\n",
"if 'javascript:' in style:\n",
"return True\n",
"if 'expression(' in style:\n",
"return True\n",
"if '@import' in style:\n",
"return True\n",
"if '</noscript' in style:\n",
"return True\n",
"if _looks_like_tag_content(style):\n",
"return True\n",
"return False\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Return'",
"Condition",
"Return'",
"Condition",
"Return'",
"Condition",
"Return'",
"Condition",
"Return'",
"Return'"
] |
[
"def FUNC_41(self, VAR_131=False):...\n",
"\"\"\"docstring\"\"\"\n",
"if not self.settings.allow_basic_login:\n",
"return False, False, False\n",
"VAR_258 = VAR_263.request.env.http_authorization\n",
"if VAR_131:\n",
"if callable(VAR_131):\n",
"if not VAR_258 or not VAR_258[:6].lower() == 'basic ':\n",
"VAR_131 = VAR_131()\n",
"if isinstance(VAR_131, string_types):\n",
"if VAR_131:\n",
"VAR_124, VAR_207, VAR_132 = base64.b64decode(VAR_258[6:]).partition(b':')\n",
"VAR_363 = HTTP(401, 'Not Authorized', **{'WWW-Authenticate': \n 'Basic realm=\"' + basic_realm + '\"'})\n",
"VAR_449 = to_unicode(VAR_131)\n",
"if VAR_131 is True:\n",
"return True, False, False\n",
"VAR_259 = VAR_207 and self.login_bare(VAR_124, VAR_132)\n",
"VAR_449 = '' + VAR_263.request.application\n",
"if not VAR_259 and VAR_131:\n",
"return True, True, VAR_259\n"
] | [
"def basic(self, basic_auth_realm=False):...\n",
"\"\"\"docstring\"\"\"\n",
"if not self.settings.allow_basic_login:\n",
"return False, False, False\n",
"basic = current.request.env.http_authorization\n",
"if basic_auth_realm:\n",
"if callable(basic_auth_realm):\n",
"if not basic or not basic[:6].lower() == 'basic ':\n",
"basic_auth_realm = basic_auth_realm()\n",
"if isinstance(basic_auth_realm, string_types):\n",
"if basic_auth_realm:\n",
"username, sep, password = base64.b64decode(basic[6:]).partition(b':')\n",
"http_401 = HTTP(401, 'Not Authorized', **{'WWW-Authenticate': \n 'Basic realm=\"' + basic_realm + '\"'})\n",
"basic_realm = to_unicode(basic_auth_realm)\n",
"if basic_auth_realm is True:\n",
"return True, False, False\n",
"is_valid_user = sep and self.login_bare(username, password)\n",
"basic_realm = '' + current.request.application\n",
"if not is_valid_user and basic_auth_realm:\n",
"return True, True, is_valid_user\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Condition",
"Return'",
"Assign'",
"Condition",
"Condition",
"Condition",
"Assign'",
"Condition",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Return'",
"Assign'",
"Assign'",
"Condition",
"Return'"
] |
[
"def FUNC_59(self, VAR_142=100):...\n",
"\"\"\"docstring\"\"\"\n",
"if not self.user:\n",
"redirect(self.settings.login_url)\n",
"if not self.settings.bulk_register_enabled:\n",
"return HTTP(404)\n",
"VAR_7 = SQLFORM.factory(VAR_1('subject', 'string', VAR_5=self.messages.\n bulk_invite_subject, VAR_279=IS_NOT_EMPTY()), VAR_1('emails', 'text',\n VAR_279=IS_NOT_EMPTY()), VAR_1('message', 'text', VAR_5=self.messages.\n bulk_invite_body, VAR_279=IS_NOT_EMPTY()), VAR_273=self.settings.formstyle)\n",
"if VAR_7.process().accepted:\n",
"VAR_375 = re.compile('[^\\\\s\\'\"@<>,;:]+\\\\@[^\\\\s\\'\"@<>,;:]+').findall(VAR_7.\n vars.emails)\n",
"return VAR_7\n",
"VAR_376 = []\n",
"VAR_377 = []\n",
"VAR_378 = []\n",
"for email in VAR_375[:VAR_142]:\n",
"if self.table_user()(email=email):\n",
"VAR_377 += VAR_375[VAR_142:]\n",
"VAR_378.append(email)\n",
"VAR_141 = self.register_bare(email=email)\n",
"VAR_7 = DIV(H4('Emails sent'), UL(*[A(VAR_30, _href='mailto:' + x) for\n VAR_30 in VAR_376]), H4('Emails failed'), UL(*[A(VAR_30, _href=\n 'mailto:' + x) for VAR_30 in VAR_377]), H4('Emails existing'), UL(*[A(\n VAR_30, _href='mailto:' + x) for VAR_30 in VAR_378]))\n",
"if self.email_registration(VAR_7.vars.subject, VAR_7.vars.message, VAR_141):\n",
"VAR_376.append(email)\n",
"VAR_377.append(email)\n"
] | [
"def bulk_register(self, max_emails=100):...\n",
"\"\"\"docstring\"\"\"\n",
"if not self.user:\n",
"redirect(self.settings.login_url)\n",
"if not self.settings.bulk_register_enabled:\n",
"return HTTP(404)\n",
"form = SQLFORM.factory(Field('subject', 'string', default=self.messages.\n bulk_invite_subject, requires=IS_NOT_EMPTY()), Field('emails', 'text',\n requires=IS_NOT_EMPTY()), Field('message', 'text', default=self.\n messages.bulk_invite_body, requires=IS_NOT_EMPTY()), formstyle=self.\n settings.formstyle)\n",
"if form.process().accepted:\n",
"emails = re.compile('[^\\\\s\\'\"@<>,;:]+\\\\@[^\\\\s\\'\"@<>,;:]+').findall(form.\n vars.emails)\n",
"return form\n",
"emails_sent = []\n",
"emails_fail = []\n",
"emails_exist = []\n",
"for email in emails[:max_emails]:\n",
"if self.table_user()(email=email):\n",
"emails_fail += emails[max_emails:]\n",
"emails_exist.append(email)\n",
"user = self.register_bare(email=email)\n",
"form = DIV(H4('Emails sent'), UL(*[A(x, _href='mailto:' + x) for x in\n emails_sent]), H4('Emails failed'), UL(*[A(x, _href='mailto:' + x) for\n x in emails_fail]), H4('Emails existing'), UL(*[A(x, _href='mailto:' +\n x) for x in emails_exist]))\n",
"if self.email_registration(form.vars.subject, form.vars.message, user):\n",
"emails_sent.append(email)\n",
"emails_fail.append(email)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Condition",
"Expr'",
"Condition",
"Return'",
"Assign'",
"Condition",
"Assign'",
"Return'",
"Assign'",
"Assign'",
"Assign'",
"For",
"Condition",
"AugAssign'",
"Expr'",
"Assign'",
"Assign'",
"Condition",
"Expr'",
"Expr'"
] |
[
"def FUNC_20(VAR_56, VAR_46):...\n",
"return os.path.exists(VAR_56) and os.access(VAR_56, VAR_46\n ) and not os.path.isdir(VAR_56)\n"
] | [
"def _access_check(fn, mode):...\n",
"return os.path.exists(fn) and os.access(fn, mode) and not os.path.isdir(fn)\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
[
"def FUNC_11(self):...\n",
"VAR_5 = self._makeContext()\n",
"self.assertTrue(VAR_5.evaluate('x | nothing') is None)\n"
] | [
"def test_evaluate_alternative_first_missing(self):...\n",
"ec = self._makeContext()\n",
"self.assertTrue(ec.evaluate('x | nothing') is None)\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Expr'"
] |
[
"def FUNC_31():...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_3 = FUNC_5()\n",
"VAR_5 = '/'.join(request.args)\n",
"VAR_43.title = request.args[-1]\n",
"VAR_71 = read_dict(apath(VAR_5, VAR_122=request))\n",
"if '__corrupted__' in VAR_71:\n",
"VAR_26 = SPAN(VAR_71['__corrupted__'], VAR_144='error')\n",
"VAR_72 = sorted(VAR_71.keys(), VAR_143=lambda x: to_native(x).lower())\n",
"return dict(VAR_5=filename, VAR_26=form)\n",
"VAR_73 = []\n",
"VAR_73.append(H2(T('Original/Translation')))\n",
"for VAR_143 in VAR_72:\n",
"VAR_10 = md5_hash(VAR_143)\n",
"VAR_73.append(DIV(INPUT(_type='submit', _value=T('update'), VAR_144=\n 'btn btn-primary'), VAR_144='controls'))\n",
"VAR_140 = VAR_71[VAR_143]\n",
"VAR_26 = FORM(*VAR_73)\n",
"VAR_141, VAR_142, VAR_143 = VAR_143.partition('\\x01')\n",
"if VAR_26.accepts(request.vars, keepvalues=True):\n",
"if VAR_142:\n",
"VAR_147 = dict()\n",
"return dict(VAR_3=request.args[0], VAR_5=filename, VAR_26=form)\n",
"VAR_141 = SPAN(VAR_141 + ': ', VAR_144='tm_ftag')\n",
"VAR_145, VAR_141 = VAR_141, ''\n",
"for VAR_143 in VAR_72:\n",
"VAR_145 = VAR_143\n",
"VAR_144 = 'untranslated' if VAR_145 == VAR_140 else 'translated'\n",
"VAR_10 = md5_hash(VAR_143)\n",
"write_dict(apath(VAR_5, VAR_122=request), VAR_147)\n",
"if len(VAR_140) <= 40:\n",
"if VAR_26.vars[VAR_10] == chr(127):\n",
"session.flash = T('file saved on %(time)s', dict(time=time.ctime()))\n",
"VAR_187 = INPUT(_type='text', _name=name, VAR_9=s, _size=70, VAR_144=_class)\n",
"VAR_187 = TEXTAREA(_name=name, VAR_9=s, _cols=70, _rows=5, VAR_144=_class)\n",
"VAR_147[VAR_143] = VAR_26.vars[VAR_10]\n",
"redirect(URL(VAR_122=request, VAR_98=request.args))\n",
"VAR_145 = VAR_140 != VAR_145 and VAR_145 or B(VAR_145)\n",
"VAR_146 = DIV(LABEL(VAR_141, VAR_145, _style='font-weight:normal;'), CAT(\n VAR_187, '\\n', TAG.BUTTON(T('delete'), _onclick='return delkey(\"%s\")' %\n name, VAR_144='btn')), _id=name, VAR_144='span6 well well-small')\n",
"VAR_73.append(DIV(VAR_146, VAR_144='row-fluid'))\n"
] | [
"def edit_language():...\n",
"\"\"\"docstring\"\"\"\n",
"app = get_app()\n",
"filename = '/'.join(request.args)\n",
"response.title = request.args[-1]\n",
"strings = read_dict(apath(filename, r=request))\n",
"if '__corrupted__' in strings:\n",
"form = SPAN(strings['__corrupted__'], _class='error')\n",
"keys = sorted(strings.keys(), key=lambda x: to_native(x).lower())\n",
"return dict(filename=filename, form=form)\n",
"rows = []\n",
"rows.append(H2(T('Original/Translation')))\n",
"for key in keys:\n",
"name = md5_hash(key)\n",
"rows.append(DIV(INPUT(_type='submit', _value=T('update'), _class=\n 'btn btn-primary'), _class='controls'))\n",
"s = strings[key]\n",
"form = FORM(*rows)\n",
"prefix, sep, key = key.partition('\\x01')\n",
"if form.accepts(request.vars, keepvalues=True):\n",
"if sep:\n",
"strs = dict()\n",
"return dict(app=request.args[0], filename=filename, form=form)\n",
"prefix = SPAN(prefix + ': ', _class='tm_ftag')\n",
"k, prefix = prefix, ''\n",
"for key in keys:\n",
"k = key\n",
"_class = 'untranslated' if k == s else 'translated'\n",
"name = md5_hash(key)\n",
"write_dict(apath(filename, r=request), strs)\n",
"if len(s) <= 40:\n",
"if form.vars[name] == chr(127):\n",
"session.flash = T('file saved on %(time)s', dict(time=time.ctime()))\n",
"elem = INPUT(_type='text', _name=name, value=s, _size=70, _class=_class)\n",
"elem = TEXTAREA(_name=name, value=s, _cols=70, _rows=5, _class=_class)\n",
"strs[key] = form.vars[name]\n",
"redirect(URL(r=request, args=request.args))\n",
"k = s != k and k or B(k)\n",
"new_row = DIV(LABEL(prefix, k, _style='font-weight:normal;'), CAT(elem,\n '\\n', TAG.BUTTON(T('delete'), _onclick='return delkey(\"%s\")' % name,\n _class='btn')), _id=name, _class='span6 well well-small')\n",
"rows.append(DIV(new_row, _class='row-fluid'))\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Return'",
"Assign'",
"Expr'",
"For",
"Assign'",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Condition",
"Assign'",
"Return'",
"Assign'",
"Assign'",
"For",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Condition",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Assign'",
"Expr'"
] |
[
"def FUNC_17(VAR_16, VAR_17, VAR_18):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_42 = {}\n",
"VAR_43 = FUNC_13(VAR_16)\n",
"VAR_44 = FUNC_14(VAR_17)\n",
"VAR_45 = FUNC_15(VAR_18)\n",
"for VAR_71, (filename, variable_name) in VAR_43.items():\n",
"VAR_70 = np.load(file_io.FileIO(filename, mode='rb'), allow_pickle=True)\n",
"for VAR_71, py_expr_evaluated in VAR_44.items():\n",
"if variable_name:\n",
"if VAR_71 in VAR_42:\n",
"for VAR_71, VAR_41 in VAR_45.items():\n",
"if isinstance(VAR_70, np.ndarray):\n",
"if isinstance(VAR_70, np.lib.npyio.NpzFile):\n",
"logging.warn('string' % VAR_71)\n",
"VAR_42[VAR_71] = py_expr_evaluated\n",
"if VAR_71 in VAR_42:\n",
"return VAR_42\n",
"logging.warn(\n 'Input file %s contains a single ndarray. Name key \"%s\" ignored.' % (\n filename, variable_name))\n",
"if variable_name in VAR_70:\n",
"VAR_80 = VAR_70.files\n",
"VAR_42[VAR_71] = VAR_70\n",
"logging.warn(\n 'input_key %s has been specified in multiple options. Value in --input_examples will be used.'\n % VAR_71)\n",
"VAR_42[VAR_71] = VAR_41\n",
"VAR_42[VAR_71] = VAR_70\n",
"VAR_42[VAR_71] = VAR_70[variable_name]\n",
"if len(VAR_80) != 1:\n",
"VAR_42[VAR_71] = VAR_70[VAR_80[0]]\n"
] | [
"def load_inputs_from_input_arg_string(inputs_str, input_exprs_str,...\n",
"\"\"\"docstring\"\"\"\n",
"tensor_key_feed_dict = {}\n",
"inputs = preprocess_inputs_arg_string(inputs_str)\n",
"input_exprs = preprocess_input_exprs_arg_string(input_exprs_str)\n",
"input_examples = preprocess_input_examples_arg_string(input_examples_str)\n",
"for input_tensor_key, (filename, variable_name) in inputs.items():\n",
"data = np.load(file_io.FileIO(filename, mode='rb'), allow_pickle=True)\n",
"for input_tensor_key, py_expr_evaluated in input_exprs.items():\n",
"if variable_name:\n",
"if input_tensor_key in tensor_key_feed_dict:\n",
"for input_tensor_key, example in input_examples.items():\n",
"if isinstance(data, np.ndarray):\n",
"if isinstance(data, np.lib.npyio.NpzFile):\n",
"logging.warn(\n 'input_key %s has been specified with both --inputs and --input_exprs options. Value in --input_exprs will be used.'\n % input_tensor_key)\n",
"tensor_key_feed_dict[input_tensor_key] = py_expr_evaluated\n",
"if input_tensor_key in tensor_key_feed_dict:\n",
"return tensor_key_feed_dict\n",
"logging.warn(\n 'Input file %s contains a single ndarray. Name key \"%s\" ignored.' % (\n filename, variable_name))\n",
"if variable_name in data:\n",
"variable_name_list = data.files\n",
"tensor_key_feed_dict[input_tensor_key] = data\n",
"logging.warn(\n 'input_key %s has been specified in multiple options. Value in --input_examples will be used.'\n % input_tensor_key)\n",
"tensor_key_feed_dict[input_tensor_key] = example\n",
"tensor_key_feed_dict[input_tensor_key] = data\n",
"tensor_key_feed_dict[input_tensor_key] = data[variable_name]\n",
"if len(variable_name_list) != 1:\n",
"tensor_key_feed_dict[input_tensor_key] = data[variable_name_list[0]]\n"
] | [
0,
0,
0,
0,
5,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"For",
"Assign'",
"For",
"Condition",
"Condition",
"For",
"Condition",
"Condition",
"Expr'",
"Assign'",
"Condition",
"Return'",
"Expr'",
"Condition",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Assign'"
] |
[
"def FUNC_9(self, *VAR_4, **VAR_5):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_54 = VAR_4[0] if VAR_4 else ''\n",
"if not VAR_54 in self.handlers:\n",
"return \"Error: no such action. '%s'\" % VAR_54\n",
"VAR_55 = self.handlers[VAR_54]\n",
"VAR_56 = not ('noauth' in dir(VAR_55) and VAR_55.noauth)\n",
"if VAR_56 and not self.isAuthorized():\n",
"VAR_57 = {}\n",
"if 'data' in VAR_5:\n",
"VAR_57 = json.loads(VAR_5['data'])\n",
"VAR_58 = 'binary' in dir(VAR_55) and VAR_55.binary\n",
"if VAR_58:\n",
"return VAR_55(**handler_args)\n",
"return json.dumps({'data': VAR_55(**handler_args)})\n"
] | [
"def api(self, *args, **kwargs):...\n",
"\"\"\"docstring\"\"\"\n",
"action = args[0] if args else ''\n",
"if not action in self.handlers:\n",
"return \"Error: no such action. '%s'\" % action\n",
"handler = self.handlers[action]\n",
"needsAuth = not ('noauth' in dir(handler) and handler.noauth)\n",
"if needsAuth and not self.isAuthorized():\n",
"handler_args = {}\n",
"if 'data' in kwargs:\n",
"handler_args = json.loads(kwargs['data'])\n",
"is_binary = 'binary' in dir(handler) and handler.binary\n",
"if is_binary:\n",
"return handler(**handler_args)\n",
"return json.dumps({'data': handler(**handler_args)})\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Condition",
"Return'",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Condition",
"Return'",
"Return'"
] |
[
"@staticmethod...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_29 = oai.OpenAPI.parse_obj(VAR_13)\n",
"return GeneratorError(header='Failed to parse OpenAPI document', detail=str(e))\n",
"if VAR_29.components is None or VAR_29.components.schemas is None:\n",
"VAR_12 = CLASS_4()\n",
"VAR_12 = CLASS_4.build(VAR_12=openapi.components.schemas)\n",
"VAR_22 = CLASS_1.from_data(VAR_5=openapi.paths)\n",
"VAR_23 = EnumProperty.get_all_enums()\n",
"return CLASS_5(title=openapi.info.title, description=openapi.info.\n description, version=openapi.info.version, VAR_22=\n endpoint_collections_by_tag, VAR_12=schemas, VAR_23=enums)\n"
] | [
"@staticmethod...\n",
"\"\"\"docstring\"\"\"\n",
"openapi = oai.OpenAPI.parse_obj(d)\n",
"return GeneratorError(header='Failed to parse OpenAPI document', detail=str(e))\n",
"if openapi.components is None or openapi.components.schemas is None:\n",
"schemas = Schemas()\n",
"schemas = Schemas.build(schemas=openapi.components.schemas)\n",
"endpoint_collections_by_tag = EndpointCollection.from_data(data=openapi.paths)\n",
"enums = EnumProperty.get_all_enums()\n",
"return GeneratorData(title=openapi.info.title, description=openapi.info.\n description, version=openapi.info.version, endpoint_collections_by_tag=\n endpoint_collections_by_tag, schemas=schemas, enums=enums)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Docstring",
"Assign'",
"Return'",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Return'"
] |
[
"def FUNC_23(self, VAR_0):...\n",
"import pathlib\n",
"from jinja2 import Template\n",
"from openapi_python_client import GeneratorData, Project\n",
"VAR_3 = VAR_0.MagicMock(autospec=GeneratorData, title='My Test API')\n",
"VAR_52 = VAR_0.MagicMock(autospec=str)\n",
"VAR_53 = VAR_0.MagicMock(autospec=str)\n",
"VAR_54 = VAR_0.MagicMock()\n",
"VAR_55 = VAR_0.MagicMock()\n",
"VAR_3.endpoint_collections_by_tag = {VAR_52: VAR_54, VAR_53: VAR_55}\n",
"VAR_8 = Project(VAR_3=openapi)\n",
"VAR_8.package_dir = VAR_0.MagicMock()\n",
"VAR_56 = VAR_0.MagicMock(autospec=pathlib.Path)\n",
"VAR_57 = VAR_0.MagicMock()\n",
"VAR_58 = VAR_0.MagicMock(autospec=pathlib.Path)\n",
"VAR_59 = VAR_0.MagicMock(autospec=pathlib.Path)\n",
"VAR_60 = VAR_0.MagicMock(autospec=pathlib.Path)\n",
"VAR_61 = VAR_0.MagicMock(autospec=pathlib.Path)\n",
"VAR_62 = VAR_0.MagicMock(autospec=pathlib.Path)\n",
"VAR_63 = VAR_0.MagicMock(autospec=pathlib.Path)\n",
"VAR_64 = {'__init__.py': VAR_58, f'{VAR_52}.py': VAR_59, f'{VAR_53}.py': VAR_60\n }\n",
"VAR_65 = {'__init__.py': VAR_61, f'{VAR_52}.py': VAR_62, f'{VAR_53}.py': VAR_63\n }\n",
"VAR_66 = VAR_0.MagicMock(autospec=pathlib.Path)\n",
"VAR_66.__truediv__.side_effect = lambda VAR_40: VAR_64[VAR_40]\n",
"VAR_67 = VAR_0.MagicMock(autospec=pathlib.Path)\n",
"VAR_67.__truediv__.side_effect = lambda VAR_40: VAR_65[VAR_40]\n",
"VAR_68 = {'client.py': VAR_57, 'api': VAR_66, 'async_api': VAR_67,\n 'errors.py': VAR_56}\n",
"VAR_8.package_dir.__truediv__.side_effect = lambda VAR_40: VAR_68[VAR_40]\n",
"VAR_69 = VAR_0.MagicMock(autospec=Template)\n",
"VAR_70 = VAR_0.MagicMock(autospec=Template)\n",
"VAR_71 = VAR_0.MagicMock(autospec=Template)\n",
"VAR_72 = VAR_0.MagicMock(autospec=Template)\n",
"VAR_27 = {'client.pyi': VAR_69, 'errors.pyi': VAR_70, 'endpoint_module.pyi':\n VAR_71, 'async_endpoint_module.pyi': VAR_72}\n",
"VAR_0.patch.object(VAR_8.env, 'get_template', autospec=True, side_effect=lambda\n x: templates[x])\n",
"VAR_73 = {VAR_54: VAR_0.MagicMock(), VAR_55: VAR_0.MagicMock()}\n",
"VAR_71.render.side_effect = lambda collection: VAR_73[collection]\n",
"VAR_74 = {VAR_54: VAR_0.MagicMock(), VAR_55: VAR_0.MagicMock()}\n",
"VAR_72.render.side_effect = lambda collection: VAR_74[collection]\n",
"VAR_8._build_api()\n",
"VAR_8.package_dir.__truediv__.assert_has_calls([VAR_0.call(key) for key in\n VAR_68])\n",
"VAR_8.env.get_template.assert_has_calls([VAR_0.call(key) for key in VAR_27])\n",
"VAR_69.render.assert_called_once()\n",
"VAR_57.write_text.assert_called_once_with(VAR_69.render())\n",
"VAR_70.render.assert_called_once()\n",
"VAR_56.write_text.assert_called_once_with(VAR_70.render())\n",
"VAR_66.mkdir.assert_called_once()\n",
"VAR_66.__truediv__.assert_has_calls([VAR_0.call(key) for key in VAR_64])\n",
"VAR_58.write_text.assert_called_once_with(\n '\"\"\" Contains synchronous methods for accessing the API \"\"\"')\n",
"VAR_71.render.assert_has_calls([VAR_0.call(collection=collection_1), VAR_0.\n call(collection=collection_2)])\n",
"VAR_59.write_text.assert_called_once_with(VAR_73[VAR_54])\n",
"VAR_60.write_text.assert_called_once_with(VAR_73[VAR_55])\n",
"VAR_67.mkdir.assert_called_once()\n",
"VAR_67.__truediv__.assert_has_calls([VAR_0.call(key) for key in VAR_65])\n",
"VAR_61.write_text.assert_called_once_with(\n '\"\"\" Contains async methods for accessing the API \"\"\"')\n",
"VAR_72.render.assert_has_calls([VAR_0.call(collection=collection_1), VAR_0.\n call(collection=collection_2)])\n",
"VAR_62.write_text.assert_called_once_with(VAR_74[VAR_54])\n",
"VAR_63.write_text.assert_called_once_with(VAR_74[VAR_55])\n"
] | [
"def test__build_api(self, mocker):...\n",
"import pathlib\n",
"from jinja2 import Template\n",
"from openapi_python_client import GeneratorData, Project\n",
"openapi = mocker.MagicMock(autospec=GeneratorData, title='My Test API')\n",
"tag_1 = mocker.MagicMock(autospec=str)\n",
"tag_2 = mocker.MagicMock(autospec=str)\n",
"collection_1 = mocker.MagicMock()\n",
"collection_2 = mocker.MagicMock()\n",
"openapi.endpoint_collections_by_tag = {tag_1: collection_1, tag_2: collection_2\n }\n",
"project = Project(openapi=openapi)\n",
"project.package_dir = mocker.MagicMock()\n",
"api_errors = mocker.MagicMock(autospec=pathlib.Path)\n",
"client_path = mocker.MagicMock()\n",
"api_init = mocker.MagicMock(autospec=pathlib.Path)\n",
"collection_1_path = mocker.MagicMock(autospec=pathlib.Path)\n",
"collection_2_path = mocker.MagicMock(autospec=pathlib.Path)\n",
"async_api_init = mocker.MagicMock(autospec=pathlib.Path)\n",
"async_collection_1_path = mocker.MagicMock(autospec=pathlib.Path)\n",
"async_collection_2_path = mocker.MagicMock(autospec=pathlib.Path)\n",
"api_paths = {'__init__.py': api_init, f'{tag_1}.py': collection_1_path,\n f'{tag_2}.py': collection_2_path}\n",
"async_api_paths = {'__init__.py': async_api_init, f'{tag_1}.py':\n async_collection_1_path, f'{tag_2}.py': async_collection_2_path}\n",
"api_dir = mocker.MagicMock(autospec=pathlib.Path)\n",
"api_dir.__truediv__.side_effect = lambda x: api_paths[x]\n",
"async_api_dir = mocker.MagicMock(autospec=pathlib.Path)\n",
"async_api_dir.__truediv__.side_effect = lambda x: async_api_paths[x]\n",
"package_paths = {'client.py': client_path, 'api': api_dir, 'async_api':\n async_api_dir, 'errors.py': api_errors}\n",
"project.package_dir.__truediv__.side_effect = lambda x: package_paths[x]\n",
"client_template = mocker.MagicMock(autospec=Template)\n",
"errors_template = mocker.MagicMock(autospec=Template)\n",
"endpoint_template = mocker.MagicMock(autospec=Template)\n",
"async_endpoint_template = mocker.MagicMock(autospec=Template)\n",
"templates = {'client.pyi': client_template, 'errors.pyi': errors_template,\n 'endpoint_module.pyi': endpoint_template, 'async_endpoint_module.pyi':\n async_endpoint_template}\n",
"mocker.patch.object(project.env, 'get_template', autospec=True, side_effect\n =lambda x: templates[x])\n",
"endpoint_renders = {collection_1: mocker.MagicMock(), collection_2: mocker.\n MagicMock()}\n",
"endpoint_template.render.side_effect = lambda collection: endpoint_renders[\n collection]\n",
"async_endpoint_renders = {collection_1: mocker.MagicMock(), collection_2:\n mocker.MagicMock()}\n",
"async_endpoint_template.render.side_effect = (lambda collection:\n async_endpoint_renders[collection])\n",
"project._build_api()\n",
"project.package_dir.__truediv__.assert_has_calls([mocker.call(key) for key in\n package_paths])\n",
"project.env.get_template.assert_has_calls([mocker.call(key) for key in\n templates])\n",
"client_template.render.assert_called_once()\n",
"client_path.write_text.assert_called_once_with(client_template.render())\n",
"errors_template.render.assert_called_once()\n",
"api_errors.write_text.assert_called_once_with(errors_template.render())\n",
"api_dir.mkdir.assert_called_once()\n",
"api_dir.__truediv__.assert_has_calls([mocker.call(key) for key in api_paths])\n",
"api_init.write_text.assert_called_once_with(\n '\"\"\" Contains synchronous methods for accessing the API \"\"\"')\n",
"endpoint_template.render.assert_has_calls([mocker.call(collection=\n collection_1), mocker.call(collection=collection_2)])\n",
"collection_1_path.write_text.assert_called_once_with(endpoint_renders[\n collection_1])\n",
"collection_2_path.write_text.assert_called_once_with(endpoint_renders[\n collection_2])\n",
"async_api_dir.mkdir.assert_called_once()\n",
"async_api_dir.__truediv__.assert_has_calls([mocker.call(key) for key in\n async_api_paths])\n",
"async_api_init.write_text.assert_called_once_with(\n '\"\"\" Contains async methods for accessing the API \"\"\"')\n",
"async_endpoint_template.render.assert_has_calls([mocker.call(collection=\n collection_1), mocker.call(collection=collection_2)])\n",
"async_collection_1_path.write_text.assert_called_once_with(\n async_endpoint_renders[collection_1])\n",
"async_collection_2_path.write_text.assert_called_once_with(\n async_endpoint_renders[collection_2])\n"
] | [
0,
0,
0,
0,
0,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Import'",
"ImportFrom'",
"ImportFrom'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'"
] |
[
"@log_function...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_2 = FUNC_1('/groups/%s/categories/%s', VAR_30, VAR_33)\n",
"return self.client.post_json(VAR_5=destination, VAR_2=path, VAR_3={\n 'requester_user_id': requester_user_id}, VAR_39=content, VAR_15=True)\n"
] | [
"@log_function...\n",
"\"\"\"docstring\"\"\"\n",
"path = _create_v1_path('/groups/%s/categories/%s', group_id, category_id)\n",
"return self.client.post_json(destination=destination, path=path, args={\n 'requester_user_id': requester_user_id}, data=content, ignore_backoff=True)\n"
] | [
0,
0,
0,
0
] | [
"Condition",
"Docstring",
"Assign'",
"Return'"
] |
[
"def FUNC_33(self):...\n",
"VAR_39 = 's0_0_0_0_0_0_0_0_0'\n",
"VAR_22, VAR_23 = self.make_request('GET', \n '/rooms/%s/messages?access_token=x&from=%s' % (self.room_id, VAR_39))\n",
"self.assertEquals(200, VAR_23.code)\n",
"self.assertTrue('start' in VAR_23.json_body)\n",
"self.assertEquals(VAR_39, VAR_23.json_body['start'])\n",
"self.assertTrue('chunk' in VAR_23.json_body)\n",
"self.assertTrue('end' in VAR_23.json_body)\n"
] | [
"def test_stream_token_is_accepted_for_fwd_pagianation(self):...\n",
"token = 's0_0_0_0_0_0_0_0_0'\n",
"request, channel = self.make_request('GET', \n '/rooms/%s/messages?access_token=x&from=%s' % (self.room_id, token))\n",
"self.assertEquals(200, channel.code)\n",
"self.assertTrue('start' in channel.json_body)\n",
"self.assertEquals(token, channel.json_body['start'])\n",
"self.assertTrue('chunk' in channel.json_body)\n",
"self.assertTrue('end' in channel.json_body)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'"
] |
[
"def FUNC_2(self, VAR_5, VAR_6, VAR_3=None):...\n",
"if VAR_3 is None:\n",
"VAR_3 = {}\n",
"if self.rel.to in self.admin_site._registry:\n",
"VAR_3['class'] = 'vManyToManyRawIdAdminField'\n",
"if VAR_6:\n",
"VAR_6 = ','.join([force_text(VAR_39) for VAR_39 in VAR_6])\n",
"VAR_6 = ''\n",
"return super(CLASS_8, self).render(VAR_5, VAR_6, VAR_3)\n"
] | [
"def render(self, name, value, attrs=None):...\n",
"if attrs is None:\n",
"attrs = {}\n",
"if self.rel.to in self.admin_site._registry:\n",
"attrs['class'] = 'vManyToManyRawIdAdminField'\n",
"if value:\n",
"value = ','.join([force_text(v) for v in value])\n",
"value = ''\n",
"return super(ManyToManyRawIdWidget, self).render(name, value, attrs)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Return'"
] |
[
"@CLASS_4('spawn-output')...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_20 = jinja.render('pre.html', title='spawn output', content=spawn_output)\n",
"return 'text/html', VAR_20\n"
] | [
"@add_handler('spawn-output')...\n",
"\"\"\"docstring\"\"\"\n",
"html = jinja.render('pre.html', title='spawn output', content=spawn_output)\n",
"return 'text/html', html\n"
] | [
0,
0,
0,
0
] | [
"Condition",
"Docstring",
"Assign'",
"Return'"
] |
[
"def FUNC_9():...\n",
"VAR_16 = 0\n",
"for _ in f.readlines():\n",
"VAR_16 += 1\n",
"yield '<?xml version=\"1.0\" encoding=\"UTF-8\" ?>\\n'\n",
"if VAR_8 is not None and VAR_16 >= VAR_8:\n",
"if VAR_1:\n",
"yield '<?xml-stylesheet type=\"text/xsl\" href=\"/scopelist.xsl\" ?>\\n'\n",
"yield '<objectlist count=\"{:d}\">\\n'.format(VAR_16)\n",
"VAR_17 = 0\n",
"for VAR_13 in f.readlines():\n",
"VAR_13 = VAR_13.strip()\n",
"yield '</objectlist>\\n'\n",
"yield FUNC_4(VAR_9=path) + '\\n'\n",
"VAR_17 += 1\n",
"if VAR_8 is not None and VAR_17 >= VAR_8:\n"
] | [
"def generate():...\n",
"num_entries = 0\n",
"for _ in f.readlines():\n",
"num_entries += 1\n",
"yield '<?xml version=\"1.0\" encoding=\"UTF-8\" ?>\\n'\n",
"if limit is not None and num_entries >= limit:\n",
"if STYLE:\n",
"yield '<?xml-stylesheet type=\"text/xsl\" href=\"/scopelist.xsl\" ?>\\n'\n",
"yield '<objectlist count=\"{:d}\">\\n'.format(num_entries)\n",
"count = 0\n",
"for path in f.readlines():\n",
"path = path.strip()\n",
"yield '</objectlist>\\n'\n",
"yield _get_object_element(object_path=path) + '\\n'\n",
"count += 1\n",
"if limit is not None and count >= limit:\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"For",
"AugAssign'",
"Expr'",
"Condition",
"Condition",
"Expr'",
"Expr'",
"Assign'",
"For",
"Assign'",
"Expr'",
"Expr'",
"AugAssign'",
"Condition"
] |
[
"def FUNC_51(VAR_62, VAR_9, VAR_109=0):...\n",
"\"\"\"docstring\"\"\"\n",
"if VAR_12.exists(VAR_62, VAR_9):\n",
"FUNC_50(VAR_62, VAR_9, VAR_109=force)\n"
] | [
"def delete_doc_if_exists(doctype, name, force=0):...\n",
"\"\"\"docstring\"\"\"\n",
"if db.exists(doctype, name):\n",
"delete_doc(doctype, name, force=force)\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Condition",
"Expr'"
] |
[
"def __init__(self, VAR_198=None, **VAR_182):...\n",
"if not VAR_198:\n",
"self.__dict__.clear()\n",
"VAR_222 = self.__getattr__(VAR_198)\n",
"VAR_222.installed = True\n",
"VAR_222.update((VAR_346, v) for VAR_346, v in VAR_182.items() if VAR_346 not in\n VAR_222)\n"
] | [
"def __init__(self, plugin=None, **defaults):...\n",
"if not plugin:\n",
"self.__dict__.clear()\n",
"settings = self.__getattr__(plugin)\n",
"settings.installed = True\n",
"settings.update((k, v) for k, v in defaults.items() if k not in settings)\n"
] | [
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Condition",
"Expr'",
"Assign'",
"Assign'",
"Expr'"
] |
[
"@FUNC_0...\n",
"return Keyring(self)\n"
] | [
"@cache_in_self...\n",
"return Keyring(self)\n"
] | [
0,
0
] | [
"Condition",
"Return'"
] |
[
"@VAR_1.route('/ajax/xchange', methods=['POST'])...\n",
"VAR_79 = VAR_32.get_json().get('xchange')\n",
"if VAR_79:\n",
"for val in VAR_79:\n",
"return ''\n",
"VAR_39 = False\n",
"return json.dumps({'success': True})\n",
"VAR_17 = calibre_db.get_book(val)\n",
"VAR_115 = VAR_17.title\n",
"VAR_116 = calibre_db.order_authors(VAR_17)\n",
"VAR_53 = []\n",
"for VAR_38 in VAR_116.authors:\n",
"VAR_53.append(VAR_38.name.replace('|', ','))\n",
"VAR_91 = FUNC_27(VAR_17, ' '.join(VAR_53))\n",
"VAR_62, VAR_92 = FUNC_28(VAR_17, VAR_115)\n",
"if VAR_92 or VAR_91:\n",
"VAR_90 = VAR_17.id\n",
"if config.config_use_google_drive:\n",
"VAR_39 = True\n",
"gdriveutils.updateGdriveCalibreFromLocal()\n",
"if VAR_90:\n",
"helper.update_dir_stucture(VAR_90, config.config_calibre_dir, VAR_62[0])\n",
"if VAR_39:\n",
"VAR_17.last_modified = datetime.utcnow()\n",
"calibre_db.session.commit()\n",
"calibre_db.session.rollback()\n",
"if config.config_use_google_drive:\n",
"VAR_2.error('Database error: %s', e)\n",
"gdriveutils.updateGdriveCalibreFromLocal()\n",
"return json.dumps({'success': False})\n"
] | [
"@editbook.route('/ajax/xchange', methods=['POST'])...\n",
"vals = request.get_json().get('xchange')\n",
"if vals:\n",
"for val in vals:\n",
"return ''\n",
"modif_date = False\n",
"return json.dumps({'success': True})\n",
"book = calibre_db.get_book(val)\n",
"authors = book.title\n",
"entries = calibre_db.order_authors(book)\n",
"author_names = []\n",
"for authr in entries.authors:\n",
"author_names.append(authr.name.replace('|', ','))\n",
"title_change = handle_title_on_edit(book, ' '.join(author_names))\n",
"input_authors, authorchange = handle_author_on_edit(book, authors)\n",
"if authorchange or title_change:\n",
"edited_books_id = book.id\n",
"if config.config_use_google_drive:\n",
"modif_date = True\n",
"gdriveutils.updateGdriveCalibreFromLocal()\n",
"if edited_books_id:\n",
"helper.update_dir_stucture(edited_books_id, config.config_calibre_dir,\n input_authors[0])\n",
"if modif_date:\n",
"book.last_modified = datetime.utcnow()\n",
"calibre_db.session.commit()\n",
"calibre_db.session.rollback()\n",
"if config.config_use_google_drive:\n",
"log.error('Database error: %s', e)\n",
"gdriveutils.updateGdriveCalibreFromLocal()\n",
"return json.dumps({'success': False})\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Assign'",
"Condition",
"For",
"Return'",
"Assign'",
"Return'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"For",
"Expr'",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Expr'",
"Condition",
"Expr'",
"Condition",
"Assign'",
"Expr'",
"Expr'",
"Condition",
"Expr'",
"Expr'",
"Return'"
] |
[
"async def FUNC_15(VAR_20):...\n",
""
] | [
"async def get_key(key_server):...\n",
""
] | [
0,
0
] | [
"AsyncFunctionDef'",
"Condition"
] |
[
"def FUNC_14(self, VAR_30: bytes) ->None:...\n",
"self.stream.write(VAR_30)\n",
"self.length += len(VAR_30)\n",
"if self.max_size is not None and self.length >= self.max_size:\n",
"self.deferred.errback(SynapseError(502, \n 'Requested file is too large > %r bytes' % (self.max_size,), Codes.\n TOO_LARGE))\n",
"self.deferred = defer.Deferred()\n",
"self.transport.loseConnection()\n"
] | [
"def dataReceived(self, data: bytes) ->None:...\n",
"self.stream.write(data)\n",
"self.length += len(data)\n",
"if self.max_size is not None and self.length >= self.max_size:\n",
"self.deferred.errback(SynapseError(502, \n 'Requested file is too large > %r bytes' % (self.max_size,), Codes.\n TOO_LARGE))\n",
"self.deferred = defer.Deferred()\n",
"self.transport.loseConnection()\n"
] | [
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Expr'",
"AugAssign'",
"Condition",
"Expr'",
"Assign'",
"Expr'"
] |
[
"def FUNC_16(self):...\n",
"VAR_5 = self._makeContext()\n",
"self.assertEqual(VAR_5.evaluate('x | string:x'), 'x')\n"
] | [
"def test_hybrid_with_string_expression(self):...\n",
"ec = self._makeContext()\n",
"self.assertEqual(ec.evaluate('x | string:x'), 'x')\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Expr'"
] |
[
"def FUNC_38(self, VAR_2='password'):...\n",
"VAR_17 = reverse('login')\n",
"for bad_url in ('http://example.com', 'https://example.com',\n",
"VAR_30 = '%(url)s?%(next)s=%(bad_url)s' % {'url': VAR_17, 'next':\n REDIRECT_FIELD_NAME, 'bad_url': urlquote(bad_url)}\n",
"for good_url in ('/view/?param=http://example.com',\n",
"VAR_3 = self.client.post(VAR_30, {'username': 'testclient', 'password': VAR_2})\n",
"VAR_31 = '%(url)s?%(next)s=%(good_url)s' % {'url': VAR_17, 'next':\n REDIRECT_FIELD_NAME, 'good_url': urlquote(good_url)}\n",
"self.assertEqual(VAR_3.status_code, 302)\n",
"VAR_3 = self.client.post(VAR_31, {'username': 'testclient', 'password': VAR_2})\n",
"self.assertFalse(bad_url in VAR_3.url, '%s should be blocked' % bad_url)\n",
"self.assertEqual(VAR_3.status_code, 302)\n",
"self.assertTrue(good_url in VAR_3.url, '%s should be allowed' % good_url)\n"
] | [
"def test_security_check(self, password='password'):...\n",
"login_url = reverse('login')\n",
"for bad_url in ('http://example.com', 'https://example.com',\n",
"nasty_url = '%(url)s?%(next)s=%(bad_url)s' % {'url': login_url, 'next':\n REDIRECT_FIELD_NAME, 'bad_url': urlquote(bad_url)}\n",
"for good_url in ('/view/?param=http://example.com',\n",
"response = self.client.post(nasty_url, {'username': 'testclient',\n 'password': password})\n",
"safe_url = '%(url)s?%(next)s=%(good_url)s' % {'url': login_url, 'next':\n REDIRECT_FIELD_NAME, 'good_url': urlquote(good_url)}\n",
"self.assertEqual(response.status_code, 302)\n",
"response = self.client.post(safe_url, {'username': 'testclient', 'password':\n password})\n",
"self.assertFalse(bad_url in response.url, '%s should be blocked' % bad_url)\n",
"self.assertEqual(response.status_code, 302)\n",
"self.assertTrue(good_url in response.url, '%s should be allowed' % good_url)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"For",
"Assign'",
"For",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Expr'",
"Expr'",
"Expr'"
] |
[
"def FUNC_19(self, VAR_6, VAR_8, VAR_10=False):...\n",
"VAR_13, VAR_15 = [], []\n",
"VAR_25 = frappe.desk.query_report.run(self.name, VAR_6=filters, VAR_8=user,\n VAR_10=ignore_prepared_report)\n",
"for d in VAR_25.get('columns'):\n",
"if isinstance(d, dict):\n",
"VAR_15 += VAR_25.get('result')\n",
"VAR_36 = frappe._dict(d)\n",
"VAR_37, VAR_38 = 'Data', None\n",
"return VAR_13, VAR_15\n",
"if not VAR_36.fieldname:\n",
"VAR_11 = d.split(':')\n",
"VAR_36.fieldname = VAR_36.label\n",
"VAR_13.append(VAR_36)\n",
"if len(VAR_11) > 1:\n",
"if VAR_11[1]:\n",
"VAR_13.append(frappe._dict(VAR_32=parts[0], VAR_37=fieldtype, fieldname=\n parts[0], VAR_38=options))\n",
"VAR_37, VAR_38 = VAR_11[1], None\n",
"if VAR_37 and '/' in VAR_37:\n",
"VAR_37, VAR_38 = VAR_37.split('/')\n"
] | [
"def run_query_report(self, filters, user, ignore_prepared_report=False):...\n",
"columns, result = [], []\n",
"data = frappe.desk.query_report.run(self.name, filters=filters, user=user,\n ignore_prepared_report=ignore_prepared_report)\n",
"for d in data.get('columns'):\n",
"if isinstance(d, dict):\n",
"result += data.get('result')\n",
"col = frappe._dict(d)\n",
"fieldtype, options = 'Data', None\n",
"return columns, result\n",
"if not col.fieldname:\n",
"parts = d.split(':')\n",
"col.fieldname = col.label\n",
"columns.append(col)\n",
"if len(parts) > 1:\n",
"if parts[1]:\n",
"columns.append(frappe._dict(label=parts[0], fieldtype=fieldtype, fieldname=\n parts[0], options=options))\n",
"fieldtype, options = parts[1], None\n",
"if fieldtype and '/' in fieldtype:\n",
"fieldtype, options = fieldtype.split('/')\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"For",
"Condition",
"AugAssign'",
"Assign'",
"Assign'",
"Return'",
"Condition",
"Assign'",
"Assign'",
"Expr'",
"Condition",
"Condition",
"Expr'",
"Assign'",
"Condition",
"Assign'"
] |
[
"def FUNC_1(VAR_2):...\n",
"if VAR_2.is_public:\n",
"return True\n",
"if current_user.is_anonymous or VAR_2.user_id != current_user.id:\n",
"VAR_1.error('User is unauthorized to view non-public shelf: %s', VAR_2)\n",
"return True\n",
"return False\n"
] | [
"def check_shelf_view_permissions(cur_shelf):...\n",
"if cur_shelf.is_public:\n",
"return True\n",
"if current_user.is_anonymous or cur_shelf.user_id != current_user.id:\n",
"log.error('User is unauthorized to view non-public shelf: %s', cur_shelf)\n",
"return True\n",
"return False\n"
] | [
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Condition",
"Return'",
"Condition",
"Expr'",
"Return'",
"Return'"
] |
[
"def __init__(self, VAR_58, *VAR_6, **VAR_7):...\n",
"self.project = VAR_58\n",
"super().__init__(*VAR_6, **kwargs)\n",
"self.fields['group'].queryset = VAR_58.defined_groups.all()\n"
] | [
"def __init__(self, project, *args, **kwargs):...\n",
"self.project = project\n",
"super().__init__(*args, **kwargs)\n",
"self.fields['group'].queryset = project.defined_groups.all()\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Expr'",
"Assign'"
] |
[
"def FUNC_13(self, VAR_5):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_8 = getattr(self, 'create_%s_field' % VAR_5, None)\n",
"if VAR_8:\n",
"return VAR_8\n",
"import inspect\n",
"VAR_10 = [f[0] for f in inspect.getmembers(self.__class__, inspect.\n isfunction) if f[0].startswith('create_') and f[0].endswith('_field')]\n"
] | [
"def get_create_field_function(self, type):...\n",
"\"\"\"docstring\"\"\"\n",
"create_field_function = getattr(self, 'create_%s_field' % type, None)\n",
"if create_field_function:\n",
"return create_field_function\n",
"import inspect\n",
"method_list = [f[0] for f in inspect.getmembers(self.__class__, inspect.\n isfunction) if f[0].startswith('create_') and f[0].endswith('_field')]\n"
] | [
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Condition",
"Return'",
"Import'",
"Assign'"
] |
[
"def FUNC_4(VAR_14):...\n",
"VAR_23 = []\n",
"for r in VAR_14:\n",
"VAR_25 = {'value': r[0], 'description': ', '.join(unique(cstr(d) for d in r if\n d)[1:])}\n",
"return VAR_23\n",
"VAR_23.append(VAR_25)\n"
] | [
"def build_for_autosuggest(res):...\n",
"results = []\n",
"for r in res:\n",
"out = {'value': r[0], 'description': ', '.join(unique(cstr(d) for d in r if\n d)[1:])}\n",
"return results\n",
"results.append(out)\n"
] | [
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"For",
"Assign'",
"Return'",
"Expr'"
] |
[
"def FUNC_1(VAR_1):...\n",
"VAR_8 = magic.from_buffer(VAR_1, mime=True)\n",
"if VAR_8[:5] != 'image':\n",
"return VAR_1, False\n",
"return VAR_1, True\n"
] | [
"def allowed_file(enc_data):...\n",
"mimetype = magic.from_buffer(enc_data, mime=True)\n",
"if mimetype[:5] != 'image':\n",
"return enc_data, False\n",
"return enc_data, True\n"
] | [
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Condition",
"Return'",
"Return'"
] |
[
"def FUNC_8(self, VAR_3, VAR_4):...\n",
"VAR_17 = self.default_config()\n",
"VAR_17['require_auth_for_profile_requests'] = True\n",
"VAR_17['limit_profile_requests_to_users_who_share_rooms'] = True\n",
"self.hs = self.setup_test_homeserver(VAR_17=config)\n",
"return self.hs\n"
] | [
"def make_homeserver(self, reactor, clock):...\n",
"config = self.default_config()\n",
"config['require_auth_for_profile_requests'] = True\n",
"config['limit_profile_requests_to_users_who_share_rooms'] = True\n",
"self.hs = self.setup_test_homeserver(config=config)\n",
"return self.hs\n"
] | [
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Return'"
] |
[
"def FUNC_20(self, VAR_3):...\n",
"\"\"\"docstring\"\"\"\n",
"if not VAR_3 or not self.index_urls:\n",
"return None\n",
"for u in self.index_urls:\n",
"VAR_94 = remove_auth_from_url(u).rstrip('/') + '/'\n",
"if VAR_3.startswith(VAR_94):\n",
"return u\n"
] | [
"def _get_index_url(self, url):...\n",
"\"\"\"docstring\"\"\"\n",
"if not url or not self.index_urls:\n",
"return None\n",
"for u in self.index_urls:\n",
"prefix = remove_auth_from_url(u).rstrip('/') + '/'\n",
"if url.startswith(prefix):\n",
"return u\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Condition",
"Return'",
"For",
"Assign'",
"Condition",
"Return'"
] |
[
"def FUNC_0(self, VAR_3, VAR_4):...\n",
"self.push_attempts = []\n",
"VAR_5 = Mock()\n",
"def FUNC_9(VAR_6, VAR_7):...\n",
"VAR_25 = Deferred()\n",
"self.push_attempts.append((VAR_25, VAR_6, VAR_7))\n",
"return make_deferred_yieldable(VAR_25)\n"
] | [
"def make_homeserver(self, reactor, clock):...\n",
"self.push_attempts = []\n",
"m = Mock()\n",
"def post_json_get_json(url, body):...\n",
"d = Deferred()\n",
"self.push_attempts.append((d, url, body))\n",
"return make_deferred_yieldable(d)\n"
] | [
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"FunctionDef'",
"Assign'",
"Expr'",
"Return'"
] |
[
"def FUNC_16(self, VAR_1):...\n",
"VAR_5 = VAR_1.MagicMock()\n",
"VAR_26 = VAR_1.MagicMock()\n",
"VAR_23 = oai.Schema(anyOf=[{'type': 'number', 'default': '0.0'}, {'type':\n 'integer', 'default': '0'}])\n",
"VAR_33 = VAR_1.patch(f'{VAR_0}.UnionProperty')\n",
"VAR_32 = VAR_1.patch(f'{VAR_0}.FloatProperty')\n",
"VAR_34 = VAR_1.patch(f'{VAR_0}.IntProperty')\n",
"from openapi_python_client.parser.properties import property_from_data\n",
"VAR_4 = property_from_data(VAR_5=name, VAR_26=required, VAR_23=data)\n",
"VAR_32.assert_called_once_with(VAR_5=name, VAR_26=required, default='0.0')\n",
"VAR_34.assert_called_once_with(VAR_5=name, VAR_26=required, default='0')\n",
"VAR_33.assert_called_once_with(VAR_5=name, VAR_26=required, default=None,\n inner_properties=[FloatProperty.return_value, IntProperty.return_value])\n",
"assert VAR_4 == VAR_33.return_value\n"
] | [
"def test_property_from_data_union(self, mocker):...\n",
"name = mocker.MagicMock()\n",
"required = mocker.MagicMock()\n",
"data = oai.Schema(anyOf=[{'type': 'number', 'default': '0.0'}, {'type':\n 'integer', 'default': '0'}])\n",
"UnionProperty = mocker.patch(f'{MODULE_NAME}.UnionProperty')\n",
"FloatProperty = mocker.patch(f'{MODULE_NAME}.FloatProperty')\n",
"IntProperty = mocker.patch(f'{MODULE_NAME}.IntProperty')\n",
"from openapi_python_client.parser.properties import property_from_data\n",
"p = property_from_data(name=name, required=required, data=data)\n",
"FloatProperty.assert_called_once_with(name=name, required=required, default\n ='0.0')\n",
"IntProperty.assert_called_once_with(name=name, required=required, default='0')\n",
"UnionProperty.assert_called_once_with(name=name, required=required, default\n =None, inner_properties=[FloatProperty.return_value, IntProperty.\n return_value])\n",
"assert p == UnionProperty.return_value\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"ImportFrom'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Assert'"
] |
[
"async def FUNC_5(self):...\n",
"return '1a2b3c4d5e6f'\n"
] | [
"async def get_resolved_ref(self):...\n",
"return '1a2b3c4d5e6f'\n"
] | [
0,
0
] | [
"AsyncFunctionDef'",
"Return'"
] |
[
"@VAR_5.route('/scope/<gididx>')...\n",
"VAR_12 = 'GIDIDX' + VAR_6.upper()\n",
"def FUNC_9():...\n",
"yield '<?xml version=\"1.0\" encoding=\"UTF-8\" ?>\\n'\n",
"if VAR_1:\n",
"yield '<?xml-stylesheet type=\"text/xsl\" href=\"/scopelist.xsl\" ?>\\n'\n",
"yield '<objectlist>\\n'\n",
"for line in f:\n",
"VAR_10 = line.strip()\n",
"yield '</objectlist>\\n'\n",
"VAR_13 = str(FUNC_5(VAR_10))\n",
"VAR_16 = Headers([('Content-Type', 'text/xml')])\n",
"VAR_25 = FUNC_7(VAR_13)\n",
"print('Error parsing {}. {}. Skip.'.format(VAR_10, str(e)), file=sys.stderr)\n",
"return Response(stream_with_context(FUNC_9()), status='200 OK', VAR_16=headers)\n",
"VAR_26 = float(VAR_25['format']['duration'])\n",
"VAR_27 = int(ceil(VAR_26 / VAR_7))\n",
"yield '<count adjust=\"{}\"/>\\n'.format(VAR_27)\n",
"for clip in range(VAR_27):\n",
"yield FUNC_4(VAR_9=clip * stride, VAR_8=span, VAR_10=video) + '\\n'\n"
] | [
"@scope_blueprint.route('/scope/<gididx>')...\n",
"index = 'GIDIDX' + gididx.upper()\n",
"def generate():...\n",
"yield '<?xml version=\"1.0\" encoding=\"UTF-8\" ?>\\n'\n",
"if STYLE:\n",
"yield '<?xml-stylesheet type=\"text/xsl\" href=\"/scopelist.xsl\" ?>\\n'\n",
"yield '<objectlist>\\n'\n",
"for line in f:\n",
"video = line.strip()\n",
"yield '</objectlist>\\n'\n",
"video_path = str(_get_obj_absolute_path(video))\n",
"headers = Headers([('Content-Type', 'text/xml')])\n",
"video_meta = _ffprobe(video_path)\n",
"print('Error parsing {}. {}. Skip.'.format(video, str(e)), file=sys.stderr)\n",
"return Response(stream_with_context(generate()), status='200 OK', headers=\n headers)\n",
"length_sec = float(video_meta['format']['duration'])\n",
"num_clips = int(ceil(length_sec / stride))\n",
"yield '<count adjust=\"{}\"/>\\n'.format(num_clips)\n",
"for clip in range(num_clips):\n",
"yield _get_object_element(start=clip * stride, span=span, video=video) + '\\n'\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Assign'",
"FunctionDef'",
"Expr'",
"Condition",
"Expr'",
"Expr'",
"For",
"Assign'",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Return'",
"Assign'",
"Assign'",
"Expr'",
"For",
"Expr'"
] |
[
"def FUNC_23(VAR_9, VAR_5, VAR_10):...\n",
"VAR_13 = calibre_db.session.query(db.Ratings).filter(db.Ratings.id == VAR_5\n ).first()\n",
"VAR_63, VAR_68, VAR_65 = calibre_db.fill_indexpage(VAR_9, 0, db.Books, db.\n Books.ratings.any(db.Ratings.id == VAR_5), [VAR_10[0][0]])\n",
"if VAR_13 and VAR_13.rating <= 10:\n",
"return render_title_template('index.html', VAR_68=random, VAR_65=pagination,\n VAR_63=entries, id=book_id, VAR_150=_(u'Rating: %(rating)s stars',\n rating=int(name.rating / 2)), VAR_9='ratings', VAR_10=order[1])\n",
"abort(404)\n"
] | [
"def render_ratings_books(page, book_id, order):...\n",
"name = calibre_db.session.query(db.Ratings).filter(db.Ratings.id == book_id\n ).first()\n",
"entries, random, pagination = calibre_db.fill_indexpage(page, 0, db.Books,\n db.Books.ratings.any(db.Ratings.id == book_id), [order[0][0]])\n",
"if name and name.rating <= 10:\n",
"return render_title_template('index.html', random=random, pagination=\n pagination, entries=entries, id=book_id, title=_(\n u'Rating: %(rating)s stars', rating=int(name.rating / 2)), page=\n 'ratings', order=order[1])\n",
"abort(404)\n"
] | [
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Condition",
"Return'",
"Expr'"
] |
Subsets and Splits
No saved queries yet
Save your SQL queries to embed, download, and access them later. Queries will appear here once saved.