lines
sequencelengths
1
383
raw_lines
sequencelengths
1
383
label
sequencelengths
1
383
type
sequencelengths
1
383
[ "def FUNC_4(VAR_1: Any) ->str:...\n", "assert isinstance(VAR_1, type), repr(VAR_1)\n", "if VAR_1.__module__.startswith('pytorch_lightning.'):\n", "return f'pl.{VAR_1.__name__}'\n", "return f'{VAR_1.__module__}.{VAR_1.__qualname__}'\n" ]
[ "def _get_abbrev_qualified_cls_name(cls: Any) ->str:...\n", "assert isinstance(cls, type), repr(cls)\n", "if cls.__module__.startswith('pytorch_lightning.'):\n", "return f'pl.{cls.__name__}'\n", "return f'{cls.__module__}.{cls.__qualname__}'\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assert'", "Condition", "Return'", "Return'" ]
[ "@VAR_2.route('/ajax/simulatedbchange', methods=['POST'])...\n", "VAR_84, VAR_85 = FUNC_47()\n", "return Response(json.dumps({'change': VAR_84, 'valid': VAR_85}), mimetype=\n 'application/json')\n" ]
[ "@admi.route('/ajax/simulatedbchange', methods=['POST'])...\n", "db_change, db_valid = _db_simulate_change()\n", "return Response(json.dumps({'change': db_change, 'valid': db_valid}),\n mimetype='application/json')\n" ]
[ 0, 0, 0 ]
[ "Condition", "Assign'", "Return'" ]
[ "def FUNC_47(self, VAR_12):...\n", "VAR_22, VAR_23 = self.make_request('GET',\n '/_matrix/client/r0/rooms/{}/state/m.room.member/{}'.format(self.\n room_id, self.second_user_id), VAR_16=self.creator_tok)\n", "self.assertEqual(VAR_23.code, 200, VAR_23.result)\n", "VAR_57 = VAR_23.json_body\n", "self.assertEqual(VAR_57.get('reason'), VAR_12, VAR_23.result)\n" ]
[ "def _check_for_reason(self, reason):...\n", "request, channel = self.make_request('GET',\n '/_matrix/client/r0/rooms/{}/state/m.room.member/{}'.format(self.\n room_id, self.second_user_id), access_token=self.creator_tok)\n", "self.assertEqual(channel.code, 200, channel.result)\n", "event_content = channel.json_body\n", "self.assertEqual(event_content.get('reason'), reason, channel.result)\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Expr'", "Assign'", "Expr'" ]
[ "def FUNC_1(VAR_6: list[FavaLedger]) ->None:...\n", "\"\"\"docstring\"\"\"\n", "VAR_38: dict[str, FavaLedger] = {}\n", "for VAR_5 in VAR_6:\n", "VAR_36 = FUNC_0(VAR_5)\n", "VAR_1.config['LEDGERS'] = VAR_38\n", "VAR_37 = next_key(VAR_36, VAR_38)\n", "VAR_38[VAR_37] = VAR_5\n" ]
[ "def update_ledger_slugs(ledgers: list[FavaLedger]) ->None:...\n", "\"\"\"docstring\"\"\"\n", "ledgers_by_slug: dict[str, FavaLedger] = {}\n", "for ledger in ledgers:\n", "slug = ledger_slug(ledger)\n", "app.config['LEDGERS'] = ledgers_by_slug\n", "unique_key = next_key(slug, ledgers_by_slug)\n", "ledgers_by_slug[unique_key] = ledger\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "AnnAssign'", "For", "Assign'", "Assign'", "Assign'", "Assign'" ]
[ "def FUNC_118(VAR_120=None):...\n", "VAR_184 = {}\n", "for VAR_219 in ([VAR_120] if VAR_120 else FUNC_64(VAR_122=True)):\n", "VAR_219 = 'frappe' if VAR_219 == 'webnotes' else VAR_219\n", "return VAR_184\n", "VAR_226 = FUNC_55(VAR_219 + '.hooks')\n", "if VAR_1.flags.in_install_app:\n", "for VAR_46 in dir(VAR_226):\n", "print('Could not find app \"{0}\"'.format(VAR_120))\n", "if not VAR_46.startswith('_'):\n", "if not VAR_16:\n", "FUNC_67(VAR_184, VAR_46, getattr(VAR_226, VAR_46))\n", "sys.exit(1)\n" ]
[ "def load_app_hooks(app_name=None):...\n", "hooks = {}\n", "for app in ([app_name] if app_name else get_installed_apps(sort=True)):\n", "app = 'frappe' if app == 'webnotes' else app\n", "return hooks\n", "app_hooks = get_module(app + '.hooks')\n", "if local.flags.in_install_app:\n", "for key in dir(app_hooks):\n", "print('Could not find app \"{0}\"'.format(app_name))\n", "if not key.startswith('_'):\n", "if not request:\n", "append_hook(hooks, key, getattr(app_hooks, key))\n", "sys.exit(1)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "For", "Assign'", "Return'", "Assign'", "Condition", "For", "Expr'", "Condition", "Condition", "Expr'", "Expr'" ]
[ "def FUNC_51(self):...\n", "\"\"\"docstring\"\"\"\n", "self._doc_before_save = None\n", "if not self.is_new():\n", "self._doc_before_save = frappe.get_doc(self.doctype, self.name)\n", "self._doc_before_save = None\n", "frappe.clear_last_message()\n" ]
[ "def load_doc_before_save(self):...\n", "\"\"\"docstring\"\"\"\n", "self._doc_before_save = None\n", "if not self.is_new():\n", "self._doc_before_save = frappe.get_doc(self.doctype, self.name)\n", "self._doc_before_save = None\n", "frappe.clear_last_message()\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Condition", "Assign'", "Assign'", "Expr'" ]
[ "def FUNC_1(VAR_0, VAR_1: FlaskClient):...\n", "VAR_0.config['THEME_CONF']['use_custom_css'] = True\n", "VAR_6 = 'custom.css'\n", "VAR_7 = \"\"\"\n body {\n color: red\n }\n \"\"\"\n", "os.mkdir(f\"{VAR_0.config['USER_DIR']}/css/\")\n", "f.write(VAR_7)\n", "VAR_0.config['THEME_CONF']['custom_css_file'] = VAR_6\n", "VAR_8 = VAR_1.get('/static/custom.css')\n", "assert VAR_7.encode('utf-8') in VAR_8.data\n", "VAR_0.config['THEME_CONF']['use_custom_css'] = False\n" ]
[ "def test_get_custom_css(test_app, client: FlaskClient):...\n", "test_app.config['THEME_CONF']['use_custom_css'] = True\n", "css_file = 'custom.css'\n", "css_contents = \"\"\"\n body {\n color: red\n }\n \"\"\"\n", "os.mkdir(f\"{test_app.config['USER_DIR']}/css/\")\n", "f.write(css_contents)\n", "test_app.config['THEME_CONF']['custom_css_file'] = css_file\n", "resp = client.get('/static/custom.css')\n", "assert css_contents.encode('utf-8') in resp.data\n", "test_app.config['THEME_CONF']['use_custom_css'] = False\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'", "Assign'", "Assign'", "Assert'", "Assign'" ]
[ "def FUNC_33(VAR_30, VAR_35='\\n'):...\n", "print(VAR_29 + VAR_30, VAR_35=end)\n" ]
[ "def in_print(s, end='\\n'):...\n", "print(indent_str + s, end=end)\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Expr'" ]
[ "def __init__(self, VAR_26: str, VAR_27: Callable[..., Any], *, VAR_29:...\n", "if isinstance(VAR_13, enum.IntEnum):\n", "VAR_13 = int(VAR_13)\n", "self.path = VAR_26\n", "self.endpoint = VAR_27\n", "self.name = get_name(VAR_27) if VAR_28 is None else VAR_28\n", "self.path_regex, self.path_format, self.param_convertors = compile_path(VAR_26)\n", "if VAR_37 is None:\n", "VAR_37 = ['GET']\n", "self.methods: Set[str] = set([method.upper() for method in VAR_37])\n", "self.unique_id = generate_operation_id_for_path(VAR_28=self.name, VAR_26=\n self.path_format, method=list(methods)[0])\n", "self.response_model = VAR_29\n", "if self.response_model:\n", "assert VAR_13 not in STATUS_CODES_WITH_NO_BODY, f'Status code {VAR_13} must not have a response body'\n", "self.response_field = None\n", "VAR_73 = 'Response_' + self.unique_id\n", "self.secure_cloned_response_field = None\n", "self.response_field = create_response_field(VAR_28=response_name, type_=\n self.response_model)\n", "self.status_code = VAR_13\n", "self.secure_cloned_response_field: Optional[ModelField] = create_cloned_field(\n self.response_field)\n", "self.tags = VAR_30 or []\n", "if VAR_31:\n", "self.dependencies = list(VAR_31)\n", "self.dependencies = []\n", "self.summary = VAR_32\n", "self.description = VAR_33 or inspect.cleandoc(self.endpoint.__doc__ or '')\n", "self.description = self.description.split('\\x0c')[0]\n", "self.response_description = VAR_34\n", "self.responses = VAR_35 or {}\n", "VAR_59 = {}\n", "for VAR_75, VAR_72 in self.responses.items():\n", "assert isinstance(VAR_72, dict), 'An additional response must be a dict'\n", "if VAR_59:\n", "VAR_74 = VAR_72.get('model')\n", "self.response_fields: Dict[Union[int, str], ModelField] = VAR_59\n", "self.response_fields = {}\n", "if VAR_74:\n", "self.deprecated = VAR_36\n", "assert VAR_75 not in STATUS_CODES_WITH_NO_BODY, f'Status code {VAR_75} must not have a response body'\n", "self.operation_id = VAR_38\n", "VAR_73 = f'Response_{VAR_75}_{self.unique_id}'\n", "self.response_model_include = VAR_16\n", "VAR_15 = create_response_field(VAR_28=response_name, type_=model)\n", "self.response_model_exclude = VAR_17\n", "VAR_59[VAR_75] = VAR_15\n", "self.response_model_by_alias = VAR_18\n", "self.response_model_exclude_unset = VAR_19\n", "self.response_model_exclude_defaults = VAR_20\n", "self.response_model_exclude_none = VAR_21\n", "self.include_in_schema = VAR_39\n", "self.response_class = VAR_14\n", "assert callable(VAR_27), 'An endpoint must be a callable'\n", "self.dependant = get_dependant(VAR_26=self.path_format, call=self.endpoint)\n", "for depends in self.dependencies[::-1]:\n", "self.dependant.dependencies.insert(0, get_parameterless_sub_dependant(\n depends=depends, VAR_26=self.path_format))\n", "self.body_field = get_body_field(VAR_10=self.dependant, VAR_28=self.unique_id)\n", "self.dependency_overrides_provider = VAR_22\n", "self.callbacks = VAR_40\n", "self.app = request_response(self.get_route_handler())\n" ]
[ "def __init__(self, path: str, endpoint: Callable[..., Any], *,...\n", "if isinstance(status_code, enum.IntEnum):\n", "status_code = int(status_code)\n", "self.path = path\n", "self.endpoint = endpoint\n", "self.name = get_name(endpoint) if name is None else name\n", "self.path_regex, self.path_format, self.param_convertors = compile_path(path)\n", "if methods is None:\n", "methods = ['GET']\n", "self.methods: Set[str] = set([method.upper() for method in methods])\n", "self.unique_id = generate_operation_id_for_path(name=self.name, path=self.\n path_format, method=list(methods)[0])\n", "self.response_model = response_model\n", "if self.response_model:\n", "assert status_code not in STATUS_CODES_WITH_NO_BODY, f'Status code {status_code} must not have a response body'\n", "self.response_field = None\n", "response_name = 'Response_' + self.unique_id\n", "self.secure_cloned_response_field = None\n", "self.response_field = create_response_field(name=response_name, type_=self.\n response_model)\n", "self.status_code = status_code\n", "self.secure_cloned_response_field: Optional[ModelField] = create_cloned_field(\n self.response_field)\n", "self.tags = tags or []\n", "if dependencies:\n", "self.dependencies = list(dependencies)\n", "self.dependencies = []\n", "self.summary = summary\n", "self.description = description or inspect.cleandoc(self.endpoint.__doc__ or '')\n", "self.description = self.description.split('\\x0c')[0]\n", "self.response_description = response_description\n", "self.responses = responses or {}\n", "response_fields = {}\n", "for additional_status_code, response in self.responses.items():\n", "assert isinstance(response, dict), 'An additional response must be a dict'\n", "if response_fields:\n", "model = response.get('model')\n", "self.response_fields: Dict[Union[int, str], ModelField] = response_fields\n", "self.response_fields = {}\n", "if model:\n", "self.deprecated = deprecated\n", "assert additional_status_code not in STATUS_CODES_WITH_NO_BODY, f'Status code {additional_status_code} must not have a response body'\n", "self.operation_id = operation_id\n", "response_name = f'Response_{additional_status_code}_{self.unique_id}'\n", "self.response_model_include = response_model_include\n", "response_field = create_response_field(name=response_name, type_=model)\n", "self.response_model_exclude = response_model_exclude\n", "response_fields[additional_status_code] = response_field\n", "self.response_model_by_alias = response_model_by_alias\n", "self.response_model_exclude_unset = response_model_exclude_unset\n", "self.response_model_exclude_defaults = response_model_exclude_defaults\n", "self.response_model_exclude_none = response_model_exclude_none\n", "self.include_in_schema = include_in_schema\n", "self.response_class = response_class\n", "assert callable(endpoint), 'An endpoint must be a callable'\n", "self.dependant = get_dependant(path=self.path_format, call=self.endpoint)\n", "for depends in self.dependencies[::-1]:\n", "self.dependant.dependencies.insert(0, get_parameterless_sub_dependant(\n depends=depends, path=self.path_format))\n", "self.body_field = get_body_field(dependant=self.dependant, name=self.unique_id)\n", "self.dependency_overrides_provider = dependency_overrides_provider\n", "self.callbacks = callbacks\n", "self.app = request_response(self.get_route_handler())\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "AnnAssign'", "Assign'", "Assign'", "Condition", "Assert'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "AnnAssign'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "For", "Assert'", "Condition", "Assign'", "AnnAssign'", "Assign'", "Condition", "Assign'", "Assert'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assert'", "Assign'", "For", "Expr'", "Assign'", "Assign'", "Assign'", "Assign'" ]
[ "@defer.inlineCallbacks...\n", "VAR_18 = yield FUNC_7(VAR_11)\n", "for member in VAR_18:\n", "if VAR_14 is not None and member == VAR_14:\n", "if VAR_6.is_mine(member):\n", "if VAR_12 is not None:\n", "if VAR_13 is not None:\n", "VAR_12.add(member)\n", "VAR_13.add(member.domain)\n" ]
[ "@defer.inlineCallbacks...\n", "members = yield get_room_members(room_id)\n", "for member in members:\n", "if ignore_user is not None and member == ignore_user:\n", "if hs.is_mine(member):\n", "if localusers is not None:\n", "if remotedomains is not None:\n", "localusers.add(member)\n", "remotedomains.add(member.domain)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Assign'", "For", "Condition", "Condition", "Condition", "Condition", "Expr'", "Expr'" ]
[ "def FUNC_3(self):...\n", "VAR_10 = self.get_success(self.handler.check_device_registered(VAR_5=\n '@boris:foo', VAR_6='fco', initial_device_display_name='display name'))\n", "self.assertEqual(VAR_10, 'fco')\n", "VAR_11 = self.get_success(self.handler.store.get_device('@boris:foo', 'fco'))\n", "self.assertEqual(VAR_11['display_name'], 'display name')\n" ]
[ "def test_device_is_created_if_doesnt_exist(self):...\n", "res = self.get_success(self.handler.check_device_registered(user_id=\n '@boris:foo', device_id='fco', initial_device_display_name='display name'))\n", "self.assertEqual(res, 'fco')\n", "dev = self.get_success(self.handler.store.get_device('@boris:foo', 'fco'))\n", "self.assertEqual(dev['display_name'], 'display name')\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Expr'", "Assign'", "Expr'" ]
[ "@VAR_2.route('/ajax/listbooks')...\n", "VAR_79 = int(request.args.get('offset') or 0)\n", "VAR_19 = int(request.args.get('limit') or config.config_books_per_page)\n", "VAR_80 = request.args.get('search')\n", "VAR_7 = request.args.get('sort', 'id')\n", "VAR_10 = request.args.get('order', '').lower()\n", "VAR_81 = None\n", "VAR_76 = tuple()\n", "if VAR_7 == 'state':\n", "VAR_81 = json.loads(request.args.get('state', '[]'))\n", "if VAR_7 == 'tags':\n", "VAR_82 = VAR_83 = calibre_db.session.query(db.Books).filter(calibre_db.\n common_filters(False)).count()\n", "VAR_10 = [db.Tags.name.asc()] if VAR_10 == 'asc' else [db.Tags.name.desc()]\n", "if VAR_7 == 'series':\n", "if VAR_81 is not None:\n", "VAR_76 = db.books_tags_link, db.Books.id == db.books_tags_link.c.book, db.Tags\n", "VAR_10 = [db.Series.name.asc()] if VAR_10 == 'asc' else [db.Series.name.desc()]\n", "if VAR_7 == 'publishers':\n", "if VAR_80:\n", "if VAR_80:\n", "VAR_76 = (db.books_series_link, db.Books.id == db.books_series_link.c.book,\n db.Series)\n", "VAR_10 = [db.Publishers.name.asc()] if VAR_10 == 'asc' else [db.Publishers.\n name.desc()]\n", "if VAR_7 == 'authors':\n", "VAR_127 = calibre_db.search_query(VAR_80).all()\n", "VAR_127 = calibre_db.session.query(db.Books).filter(calibre_db.common_filters()\n ).all()\n", "VAR_63, VAR_83, VAR_64 = calibre_db.get_search_results(VAR_80, VAR_79, [\n VAR_10, ''], VAR_19, *VAR_76)\n", "VAR_63, VAR_64, VAR_64 = calibre_db.fill_indexpage(int(VAR_79) / int(VAR_19\n ) + 1, VAR_19, db.Books, True, VAR_10, *VAR_76)\n", "VAR_76 = (db.books_publishers_link, db.Books.id == db.books_publishers_link\n .c.book, db.Publishers)\n", "VAR_10 = [db.Authors.name.asc(), db.Series.name, db.Books.series_index\n ] if VAR_10 == 'asc' else [db.Authors.name.desc(), db.Series.name.desc(\n ), db.Books.series_index.desc()]\n", "if VAR_7 == 'languages':\n", "VAR_83 = len(VAR_127)\n", "VAR_63 = calibre_db.get_checkbox_sorted(VAR_127, VAR_81, VAR_79, VAR_19, VAR_10\n )\n", "for entry in VAR_63:\n", "VAR_76 = (db.books_authors_link, db.Books.id == db.books_authors_link.c.\n book, db.Authors, db.books_series_link, db.Books.id == db.\n books_series_link.c.book, db.Series)\n", "VAR_10 = [db.Languages.lang_code.asc()] if VAR_10 == 'asc' else [db.\n Languages.lang_code.desc()]\n", "if VAR_10 and VAR_7 in ['sort', 'title', 'authors_sort', 'series_index']:\n", "for FUNC_31 in range(0, len(entry.languages)):\n", "VAR_84 = {'totalNotFiltered': VAR_82, 'total': VAR_83, 'rows': VAR_63}\n", "VAR_76 = (db.books_languages_link, db.Books.id == db.books_languages_link.c\n .book, db.Languages)\n", "VAR_10 = [text(VAR_7 + ' ' + VAR_10)]\n", "if not VAR_81:\n", "entry.languages[FUNC_31].language_name = isoLanguages.get_language_name(\n get_locale(), entry.languages[FUNC_31].lang_code)\n", "VAR_85 = json.dumps(VAR_84, cls=db.AlchemyEncoder)\n", "VAR_10 = [db.Books.timestamp.desc()]\n", "VAR_86 = make_response(VAR_85)\n", "VAR_86.headers['Content-Type'] = 'application/json; charset=utf-8'\n", "return VAR_86\n" ]
[ "@web.route('/ajax/listbooks')...\n", "off = int(request.args.get('offset') or 0)\n", "limit = int(request.args.get('limit') or config.config_books_per_page)\n", "search = request.args.get('search')\n", "sort = request.args.get('sort', 'id')\n", "order = request.args.get('order', '').lower()\n", "state = None\n", "join = tuple()\n", "if sort == 'state':\n", "state = json.loads(request.args.get('state', '[]'))\n", "if sort == 'tags':\n", "total_count = filtered_count = calibre_db.session.query(db.Books).filter(\n calibre_db.common_filters(False)).count()\n", "order = [db.Tags.name.asc()] if order == 'asc' else [db.Tags.name.desc()]\n", "if sort == 'series':\n", "if state is not None:\n", "join = db.books_tags_link, db.Books.id == db.books_tags_link.c.book, db.Tags\n", "order = [db.Series.name.asc()] if order == 'asc' else [db.Series.name.desc()]\n", "if sort == 'publishers':\n", "if search:\n", "if search:\n", "join = (db.books_series_link, db.Books.id == db.books_series_link.c.book,\n db.Series)\n", "order = [db.Publishers.name.asc()] if order == 'asc' else [db.Publishers.\n name.desc()]\n", "if sort == 'authors':\n", "books = calibre_db.search_query(search).all()\n", "books = calibre_db.session.query(db.Books).filter(calibre_db.common_filters()\n ).all()\n", "entries, filtered_count, __ = calibre_db.get_search_results(search, off, [\n order, ''], limit, *join)\n", "entries, __, __ = calibre_db.fill_indexpage(int(off) / int(limit) + 1,\n limit, db.Books, True, order, *join)\n", "join = (db.books_publishers_link, db.Books.id == db.books_publishers_link.c\n .book, db.Publishers)\n", "order = [db.Authors.name.asc(), db.Series.name, db.Books.series_index\n ] if order == 'asc' else [db.Authors.name.desc(), db.Series.name.desc(),\n db.Books.series_index.desc()]\n", "if sort == 'languages':\n", "filtered_count = len(books)\n", "entries = calibre_db.get_checkbox_sorted(books, state, off, limit, order)\n", "for entry in entries:\n", "join = (db.books_authors_link, db.Books.id == db.books_authors_link.c.book,\n db.Authors, db.books_series_link, db.Books.id == db.books_series_link.c\n .book, db.Series)\n", "order = [db.Languages.lang_code.asc()] if order == 'asc' else [db.Languages\n .lang_code.desc()]\n", "if order and sort in ['sort', 'title', 'authors_sort', 'series_index']:\n", "for index in range(0, len(entry.languages)):\n", "table_entries = {'totalNotFiltered': total_count, 'total': filtered_count,\n 'rows': entries}\n", "join = (db.books_languages_link, db.Books.id == db.books_languages_link.c.\n book, db.Languages)\n", "order = [text(sort + ' ' + order)]\n", "if not state:\n", "entry.languages[index].language_name = isoLanguages.get_language_name(\n get_locale(), entry.languages[index].lang_code)\n", "js_list = json.dumps(table_entries, cls=db.AlchemyEncoder)\n", "order = [db.Books.timestamp.desc()]\n", "response = make_response(js_list)\n", "response.headers['Content-Type'] = 'application/json; charset=utf-8'\n", "return response\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Assign'", "Condition", "Condition", "Assign'", "Assign'", "Condition", "Condition", "Condition", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "For", "Assign'", "Assign'", "Condition", "For", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "def __init__(self, VAR_3=None):...\n", "VAR_28 = {'class': 'vCommaSeparatedIntegerField'}\n", "if VAR_3 is not None:\n", "VAR_28.update(VAR_3)\n", "super(CLASS_16, self).__init__(VAR_3=final_attrs)\n" ]
[ "def __init__(self, attrs=None):...\n", "final_attrs = {'class': 'vCommaSeparatedIntegerField'}\n", "if attrs is not None:\n", "final_attrs.update(attrs)\n", "super(AdminCommaSeparatedIntegerFieldWidget, self).__init__(attrs=final_attrs)\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Condition", "Expr'", "Expr'" ]
[ "def FUNC_32(VAR_87):...\n", "VAR_86.update(to_bytes(VAR_87, encoding='utf-8', errors='replace'))\n" ]
[ "def hash_update(value):...\n", "hash.update(to_bytes(value, encoding='utf-8', errors='replace'))\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Expr'" ]
[ "def FUNC_9(self):...\n", "if not self.get('roles') and self.is_standard == 'No':\n", "VAR_5 = frappe.get_meta(self.ref_doctype)\n", "if not VAR_5.istable:\n", "VAR_35 = [{'role': d.role} for d in VAR_5.permissions if d.permlevel == 0]\n", "self.set('roles', VAR_35)\n" ]
[ "def set_doctype_roles(self):...\n", "if not self.get('roles') and self.is_standard == 'No':\n", "meta = frappe.get_meta(self.ref_doctype)\n", "if not meta.istable:\n", "roles = [{'role': d.role} for d in meta.permissions if d.permlevel == 0]\n", "self.set('roles', roles)\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Assign'", "Condition", "Assign'", "Expr'" ]
[ "def FUNC_10(VAR_0, VAR_1):...\n", "self.query_handlers[VAR_0] = VAR_1\n" ]
[ "def register_query_handler(query_type, handler):...\n", "self.query_handlers[query_type] = handler\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Assign'" ]
[ "def FUNC_26(VAR_9, VAR_13, VAR_10):...\n", "" ]
[ "def render_language_books(page, name, order):...\n", "" ]
[ 0, 0 ]
[ "FunctionDef'", "Condition" ]
[ "def __unicode__(self):...\n", "return f\"<Text '{self.offendingText}' shows up in both lt and local-lt>\"\n" ]
[ "def __unicode__(self):...\n", "return f\"<Text '{self.offendingText}' shows up in both lt and local-lt>\"\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_0(self, VAR_2, VAR_3):...\n", "VAR_4 = self.setup_test_homeserver('server', http_client=None)\n", "self.handler = VAR_4.get_device_handler()\n", "self.registration = VAR_4.get_registration_handler()\n", "self.auth = VAR_4.get_auth()\n", "self.store = VAR_4.get_datastore()\n", "return VAR_4\n" ]
[ "def make_homeserver(self, reactor, clock):...\n", "hs = self.setup_test_homeserver('server', http_client=None)\n", "self.handler = hs.get_device_handler()\n", "self.registration = hs.get_registration_handler()\n", "self.auth = hs.get_auth()\n", "self.store = hs.get_datastore()\n", "return hs\n" ]
[ 0, 4, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "@VAR_1.route('/ajax/delete/<int:book_id>')...\n", "return Response(FUNC_13(VAR_15, '', True), mimetype='application/json')\n" ]
[ "@editbook.route('/ajax/delete/<int:book_id>')...\n", "return Response(delete_book(book_id, '', True), mimetype='application/json')\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "def FUNC_0(self):...\n", "VAR_5 = super().default_config()\n", "VAR_5['send_federation'] = False\n", "return VAR_5\n" ]
[ "def default_config(self):...\n", "conf = super().default_config()\n", "conf['send_federation'] = False\n", "return conf\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Return'" ]
[ "async def FUNC_17(VAR_46: 'TrainingDataImporter', VAR_21: Union[Path, Text]...\n", "\"\"\"docstring\"\"\"\n", "VAR_18 = Path(VAR_21) / DEFAULT_CORE_SUBDIRECTORY_NAME\n", "VAR_52 = await VAR_46.get_domain()\n", "VAR_52.persist(VAR_18 / DEFAULT_DOMAIN_PATH)\n" ]
[ "async def update_model_with_new_domain(importer: 'TrainingDataImporter',...\n", "\"\"\"docstring\"\"\"\n", "model_path = Path(unpacked_model_path) / DEFAULT_CORE_SUBDIRECTORY_NAME\n", "domain = await importer.get_domain()\n", "domain.persist(model_path / DEFAULT_DOMAIN_PATH)\n" ]
[ 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Assign'", "Expr'" ]
[ "def FUNC_11(self):...\n", "\"\"\"docstring\"\"\"\n", "VAR_6 = self._make_request(b'GET', b'/res/')\n", "self.assertEqual(VAR_6.result['code'], b'200')\n", "self.assertEqual(VAR_6.result['body'], b'/res/')\n" ]
[ "def test_known_request(self):...\n", "\"\"\"docstring\"\"\"\n", "channel = self._make_request(b'GET', b'/res/')\n", "self.assertEqual(channel.result['code'], b'200')\n", "self.assertEqual(channel.result['body'], b'/res/')\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Expr'", "Expr'" ]
[ "def __str__(self):...\n", "return '%s' % self.title\n" ]
[ "def __str__(self):...\n", "return '%s' % self.title\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "@wraps(VAR_10)...\n", "VAR_1 = self.__class__\n", "if VAR_2:\n", "VAR_42 = [arg[0] for arg in FUNC_3(VAR_1)]\n", "VAR_39 = vars(FUNC_2(VAR_1))\n", "VAR_3.update(dict(zip(VAR_42, VAR_2)))\n", "VAR_3 = dict(list(VAR_39.items()) + list(VAR_3.items()))\n", "return VAR_10(self, **kwargs)\n" ]
[ "@wraps(fn)...\n", "cls = self.__class__\n", "if args:\n", "cls_arg_names = [arg[0] for arg in get_init_arguments_and_types(cls)]\n", "env_variables = vars(parse_env_variables(cls))\n", "kwargs.update(dict(zip(cls_arg_names, args)))\n", "kwargs = dict(list(env_variables.items()) + list(kwargs.items()))\n", "return fn(self, **kwargs)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Assign'", "Condition", "Assign'", "Assign'", "Expr'", "Assign'", "Return'" ]
[ "def FUNC_7(self):...\n", "return self.spec\n" ]
[ "def get_repo_url(self):...\n", "return self.spec\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "from __future__ import unicode_literals\n", "from django import template\n", "from djblets.gravatars import get_gravatar_url, get_gravatar_url_for_email\n", "from djblets.util.decorators import basictag\n", "VAR_0 = template.Library()\n", "@VAR_0.tag...\n", "\"\"\"docstring\"\"\"\n", "VAR_5 = get_gravatar_url(VAR_1['request'], VAR_2, VAR_3)\n", "if VAR_5:\n", "return '<img src=\"%s\" width=\"%s\" height=\"%s\" alt=\"%s\" class=\"gravatar\"/>' % (\n VAR_5, VAR_3, VAR_3, VAR_2.get_full_name() or VAR_2.username)\n", "return ''\n" ]
[ "from __future__ import unicode_literals\n", "from django import template\n", "from djblets.gravatars import get_gravatar_url, get_gravatar_url_for_email\n", "from djblets.util.decorators import basictag\n", "register = template.Library()\n", "@register.tag...\n", "\"\"\"docstring\"\"\"\n", "url = get_gravatar_url(context['request'], user, size)\n", "if url:\n", "return '<img src=\"%s\" width=\"%s\" height=\"%s\" alt=\"%s\" class=\"gravatar\"/>' % (\n url, size, size, user.get_full_name() or user.username)\n", "return ''\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 2, 0 ]
[ "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Assign'", "Condition", "Docstring", "Assign'", "Condition", "Return'", "Return'" ]
[ "def FUNC_140(self):...\n", "\"\"\"docstring\"\"\"\n", "if not self.wiki_menu_items and self.settings.controller and self.settings.function:\n", "self.wiki_menu_items = self.menu(self.settings.controller, self.settings.\n function)\n", "VAR_263.response.menu += self.wiki_menu_items\n" ]
[ "def automenu(self):...\n", "\"\"\"docstring\"\"\"\n", "if not self.wiki_menu_items and self.settings.controller and self.settings.function:\n", "self.wiki_menu_items = self.menu(self.settings.controller, self.settings.\n function)\n", "current.response.menu += self.wiki_menu_items\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Assign'", "AugAssign'" ]
[ "@app.route('/plugin_assets/<string:name>/<path:filename>')...\n", "return redirect(url_for('plugin.' + VAR_23 + '.static', VAR_32=filename))\n" ]
[ "@app.route('/plugin_assets/<string:name>/<path:filename>')...\n", "return redirect(url_for('plugin.' + name + '.static', filename=filename))\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "def FUNC_1(self, VAR_3, VAR_4, VAR_11):...\n", "self.user_id = self.register_user('user', 'password')\n", "self.tok = self.login('user', 'password')\n", "self.room_id = self.helper.create_room_as(self.user_id, VAR_52=self.tok,\n is_public=False)\n", "self.other_user_id = self.register_user('user2', 'password')\n", "self.other_tok = self.login('user2', 'password')\n", "self.helper.invite(self.room_id, self.user_id, self.other_user_id, VAR_52=\n self.tok)\n", "self.helper.join(self.room_id, self.other_user_id, VAR_52=self.other_tok)\n" ]
[ "def prepare(self, reactor, clock, homeserver):...\n", "self.user_id = self.register_user('user', 'password')\n", "self.tok = self.login('user', 'password')\n", "self.room_id = self.helper.create_room_as(self.user_id, tok=self.tok,\n is_public=False)\n", "self.other_user_id = self.register_user('user2', 'password')\n", "self.other_tok = self.login('user2', 'password')\n", "self.helper.invite(self.room_id, self.user_id, self.other_user_id, tok=self.tok\n )\n", "self.helper.join(self.room_id, self.other_user_id, tok=self.other_tok)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'" ]
[ "@app.after_request...\n", "VAR_1.headers['Content-Security-Policy'\n ] = \"default-src 'self' 'unsafe-inline' 'unsafe-eval';\"\n", "if request.endpoint == 'editbook.edit_book':\n", "VAR_1.headers['Content-Security-Policy'] += 'img-src * data:'\n", "VAR_1.headers['X-Content-Type-Options'] = 'nosniff'\n", "VAR_1.headers['X-Frame-Options'] = 'SAMEORIGIN'\n", "VAR_1.headers['X-XSS-Protection'] = '1; mode=block'\n", "VAR_1.headers['Strict-Transport-Security'\n ] = 'max-age=31536000; includeSubDomains'\n", "return VAR_1\n" ]
[ "@app.after_request...\n", "resp.headers['Content-Security-Policy'\n ] = \"default-src 'self' 'unsafe-inline' 'unsafe-eval';\"\n", "if request.endpoint == 'editbook.edit_book':\n", "resp.headers['Content-Security-Policy'] += 'img-src * data:'\n", "resp.headers['X-Content-Type-Options'] = 'nosniff'\n", "resp.headers['X-Frame-Options'] = 'SAMEORIGIN'\n", "resp.headers['X-XSS-Protection'] = '1; mode=block'\n", "resp.headers['Strict-Transport-Security'\n ] = 'max-age=31536000; includeSubDomains'\n", "return resp\n" ]
[ 0, 3, 0, 3, 0, 0, 0, 0, 0 ]
[ "Condition", "Assign'", "Condition", "AugAssign'", "Assign'", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_23(self):...\n", "VAR_5 = self._makeContext()\n", "self.assertEqual(VAR_5.evaluate('path: '), None)\n" ]
[ "def test_empty_path_expression_explicit_with_trailing_whitespace(self):...\n", "ec = self._makeContext()\n", "self.assertEqual(ec.evaluate('path: '), None)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Expr'" ]
[ "def FUNC_1(VAR_1, VAR_2=False):...\n", "\"\"\"docstring\"\"\"\n", "for hook in VAR_25.get_hooks('override_whitelisted_methods', {}).get(VAR_1, []\n", "VAR_1 = hook\n", "if run_server_script_api(VAR_1):\n", "return None\n", "VAR_3 = FUNC_11(VAR_1)\n", "if VAR_25.local.conf.developer_mode:\n", "if VAR_2:\n", "VAR_25.respond_as_web_page(title='Invalid Method', html='Method not found',\n indicator_color='red', http_status_code=404)\n", "return\n", "VAR_3 = VAR_3.queue\n", "FUNC_3(VAR_3)\n", "FUNC_2(VAR_3)\n", "return VAR_25.call(VAR_3, **frappe.form_dict)\n" ]
[ "def execute_cmd(cmd, from_async=False):...\n", "\"\"\"docstring\"\"\"\n", "for hook in frappe.get_hooks('override_whitelisted_methods', {}).get(cmd, []):\n", "cmd = hook\n", "if run_server_script_api(cmd):\n", "return None\n", "method = get_attr(cmd)\n", "if frappe.local.conf.developer_mode:\n", "if from_async:\n", "frappe.respond_as_web_page(title='Invalid Method', html='Method not found',\n indicator_color='red', http_status_code=404)\n", "return\n", "method = method.queue\n", "is_whitelisted(method)\n", "is_valid_http_method(method)\n", "return frappe.call(method, **frappe.form_dict)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2, 2, 0 ]
[ "FunctionDef'", "Docstring", "For", "Assign'", "Condition", "Return'", "Assign'", "Condition", "Condition", "Expr'", "Return'", "Assign'", "Expr'", "Expr'", "Return'" ]
[ "def FUNC_19(self):...\n", "\"\"\"docstring\"\"\"\n", "if not self.assigned_to:\n", "return _('Unassigned')\n", "if self.assigned_to.get_full_name():\n", "return self.assigned_to.get_full_name()\n", "return self.assigned_to.get_username()\n" ]
[ "def _get_assigned_to(self):...\n", "\"\"\"docstring\"\"\"\n", "if not self.assigned_to:\n", "return _('Unassigned')\n", "if self.assigned_to.get_full_name():\n", "return self.assigned_to.get_full_name()\n", "return self.assigned_to.get_username()\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Return'", "Condition", "Return'", "Return'" ]
[ "def FUNC_9(VAR_18, VAR_19, VAR_20, VAR_21):...\n", "VAR_69 = settings().getBoolean(['server', 'firstRun'])\n", "VAR_70 = {}\n", "for VAR_131 in LOCALES:\n", "VAR_41 = [permission.as_dict() for permission in Permissions.all()]\n", "VAR_70[VAR_131.language] = {'language': VAR_131.language, 'display':\n VAR_131.display_name, 'english': VAR_131.english_name}\n", "VAR_0.exception('Error while collecting available locales')\n", "VAR_71 = list(sorted(full_extension_tree().keys()))\n", "VAR_72 = list(map(lambda ext: f'.{ext}', get_all_extensions()))\n", "VAR_43 = {'debug': debug, 'firstRun': VAR_69, 'version': {'number': VERSION,\n 'display': DISPLAY_VERSION, 'branch': BRANCH}, 'python_version':\n get_python_version_string(), 'templates': VAR_18, 'pluginNames': VAR_19,\n 'locales': VAR_70, 'permissions': VAR_41, 'supportedFiletypes': VAR_71,\n 'supportedExtensions': VAR_72}\n", "VAR_43.update(VAR_20)\n", "return VAR_43\n" ]
[ "def _get_render_kwargs(templates, plugin_names, plugin_vars, now):...\n", "first_run = settings().getBoolean(['server', 'firstRun'])\n", "locales = {}\n", "for loc in LOCALES:\n", "permissions = [permission.as_dict() for permission in Permissions.all()]\n", "locales[loc.language] = {'language': loc.language, 'display': loc.\n display_name, 'english': loc.english_name}\n", "_logger.exception('Error while collecting available locales')\n", "filetypes = list(sorted(full_extension_tree().keys()))\n", "extensions = list(map(lambda ext: f'.{ext}', get_all_extensions()))\n", "render_kwargs = {'debug': debug, 'firstRun': first_run, 'version': {\n 'number': VERSION, 'display': DISPLAY_VERSION, 'branch': BRANCH},\n 'python_version': get_python_version_string(), 'templates': templates,\n 'pluginNames': plugin_names, 'locales': locales, 'permissions':\n permissions, 'supportedFiletypes': filetypes, 'supportedExtensions':\n extensions}\n", "render_kwargs.update(plugin_vars)\n", "return render_kwargs\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "For", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "Assign'", "Expr'", "Return'" ]
[ "def FUNC_49():...\n", "\"\"\"docstring\"\"\"\n", "VAR_3 = FUNC_5()\n", "update_all_languages(apath(VAR_3, VAR_122=request))\n", "session.flash = T('Language files (static strings) updated')\n", "redirect(URL('design', VAR_98=app, VAR_157='languages'))\n" ]
[ "def update_languages():...\n", "\"\"\"docstring\"\"\"\n", "app = get_app()\n", "update_all_languages(apath(app, r=request))\n", "session.flash = T('Language files (static strings) updated')\n", "redirect(URL('design', args=app, anchor='languages'))\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Expr'", "Assign'", "Expr'" ]
[ "def FUNC_22(self):...\n", "assert join('/', '/') == '/'\n", "assert join(None, '/') == '/'\n", "assert join('/', None) == '/'\n" ]
[ "def test_it_doesnt_eat_single_slash(self):...\n", "assert join('/', '/') == '/'\n", "assert join(None, '/') == '/'\n", "assert join('/', None) == '/'\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Assert'", "Assert'", "Assert'" ]
[ "def FUNC_0(VAR_0):...\n", "VAR_1 = os.path.dirname(os.path.abspath(__file__))\n", "return open(os.path.join(VAR_1, VAR_0)).read()\n" ]
[ "def _read(fname):...\n", "here = os.path.dirname(os.path.abspath(__file__))\n", "return open(os.path.join(here, fname)).read()\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Return'" ]
[ "def FUNC_15(self):...\n", "VAR_26 = np.array([[1], [2]])\n", "VAR_31 = os.path.join(test.get_temp_dir(), 'input.npz')\n", "np.savez(VAR_31, VAR_53=x0)\n", "VAR_20 = 'x=' + VAR_31 + '[a];y=' + VAR_31\n", "VAR_30 = saved_model_cli.load_inputs_from_input_arg_string(VAR_20, '', '')\n", "self.assertTrue(np.all(VAR_30['x'] == VAR_26))\n", "self.assertTrue(np.all(VAR_30['y'] == VAR_26))\n" ]
[ "def testInputParserNPZ(self):...\n", "x0 = np.array([[1], [2]])\n", "input_path = os.path.join(test.get_temp_dir(), 'input.npz')\n", "np.savez(input_path, a=x0)\n", "input_str = 'x=' + input_path + '[a];y=' + input_path\n", "feed_dict = saved_model_cli.load_inputs_from_input_arg_string(input_str, '', ''\n )\n", "self.assertTrue(np.all(feed_dict['x'] == x0))\n", "self.assertTrue(np.all(feed_dict['y'] == x0))\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "Expr'", "Expr'" ]
[ "def FUNC_26(VAR_37, VAR_38):...\n", "VAR_76 = FUNC_25(VAR_38 + '/v1/tunnel-request' if VAR_38 is not None else VAR_3\n )\n", "if VAR_76 and VAR_76.code == 200:\n", "VAR_97 = json.loads(VAR_76.read().decode('utf-8'))[0]\n", "return create_tunnel(VAR_97, VAR_2, VAR_37)\n" ]
[ "def setup_tunnel(local_server_port, endpoint):...\n", "response = url_request(endpoint + '/v1/tunnel-request' if endpoint is not\n None else GRADIO_API_SERVER)\n", "if response and response.code == 200:\n", "payload = json.loads(response.read().decode('utf-8'))[0]\n", "return create_tunnel(payload, LOCALHOST_NAME, local_server_port)\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Condition", "Assign'", "Return'" ]
[ "def FUNC_5(self, VAR_24):...\n", "\"\"\"docstring\"\"\"\n", "VAR_79 = []\n", "VAR_79.append(('/deprecatedwidgets/(.*)', CLASS_0))\n", "VAR_79.extend(FUNC_1('tree.handlers'))\n", "VAR_79.extend([('/login', VAR_24['login_handler_class'])])\n", "VAR_79.extend([('/logout', VAR_24['logout_handler_class'])])\n", "VAR_79.extend(FUNC_1('files.handlers'))\n", "VAR_79.extend(FUNC_1('notebook.handlers'))\n", "VAR_79.extend(FUNC_1('nbconvert.handlers'))\n", "VAR_79.extend(FUNC_1('kernelspecs.handlers'))\n", "VAR_79.extend(FUNC_1('edit.handlers'))\n", "VAR_79.extend(FUNC_1('services.api.handlers'))\n", "VAR_79.extend(FUNC_1('services.config.handlers'))\n", "VAR_79.extend(FUNC_1('services.kernels.handlers'))\n", "VAR_79.extend(FUNC_1('services.contents.handlers'))\n", "VAR_79.extend(FUNC_1('services.sessions.handlers'))\n", "VAR_79.extend(FUNC_1('services.nbconvert.handlers'))\n", "VAR_79.extend(FUNC_1('services.kernelspecs.handlers'))\n", "VAR_79.extend(FUNC_1('services.security.handlers'))\n", "import ipywidgets\n", "app_log.warn('ipywidgets package not installed. Widgets are unavailable.')\n", "VAR_79.append(('/nbextensions/(.*)', FileFindHandler, {'path': VAR_24[\n 'nbextensions_path'], 'no_cache_paths': ['/']}))\n", "VAR_79.append(('/nbextensions/widgets/(.*)', FileFindHandler, {'path':\n ipywidgets.find_static_assets(), 'no_cache_paths': ['/']}))\n", "VAR_79.append(('/custom/(.*)', FileFindHandler, {'path': VAR_24[\n 'static_custom_path'], 'no_cache_paths': ['/']}))\n", "VAR_79.extend(FUNC_1('base.handlers'))\n", "VAR_79.append(('/?', web.RedirectHandler, {'url': VAR_24['default_url'],\n 'permanent': False}))\n", "VAR_85 = []\n", "for handler in VAR_79:\n", "VAR_98 = url_path_join(VAR_24['base_url'], handler[0])\n", "VAR_85.append(('(.*)', Template404))\n", "VAR_99 = tuple([VAR_98] + list(handler[1:]))\n", "return VAR_85\n", "VAR_85.append(VAR_99)\n" ]
[ "def init_handlers(self, settings):...\n", "\"\"\"docstring\"\"\"\n", "handlers = []\n", "handlers.append(('/deprecatedwidgets/(.*)', DeprecationHandler))\n", "handlers.extend(load_handlers('tree.handlers'))\n", "handlers.extend([('/login', settings['login_handler_class'])])\n", "handlers.extend([('/logout', settings['logout_handler_class'])])\n", "handlers.extend(load_handlers('files.handlers'))\n", "handlers.extend(load_handlers('notebook.handlers'))\n", "handlers.extend(load_handlers('nbconvert.handlers'))\n", "handlers.extend(load_handlers('kernelspecs.handlers'))\n", "handlers.extend(load_handlers('edit.handlers'))\n", "handlers.extend(load_handlers('services.api.handlers'))\n", "handlers.extend(load_handlers('services.config.handlers'))\n", "handlers.extend(load_handlers('services.kernels.handlers'))\n", "handlers.extend(load_handlers('services.contents.handlers'))\n", "handlers.extend(load_handlers('services.sessions.handlers'))\n", "handlers.extend(load_handlers('services.nbconvert.handlers'))\n", "handlers.extend(load_handlers('services.kernelspecs.handlers'))\n", "handlers.extend(load_handlers('services.security.handlers'))\n", "import ipywidgets\n", "app_log.warn('ipywidgets package not installed. Widgets are unavailable.')\n", "handlers.append(('/nbextensions/(.*)', FileFindHandler, {'path': settings[\n 'nbextensions_path'], 'no_cache_paths': ['/']}))\n", "handlers.append(('/nbextensions/widgets/(.*)', FileFindHandler, {'path':\n ipywidgets.find_static_assets(), 'no_cache_paths': ['/']}))\n", "handlers.append(('/custom/(.*)', FileFindHandler, {'path': settings[\n 'static_custom_path'], 'no_cache_paths': ['/']}))\n", "handlers.extend(load_handlers('base.handlers'))\n", "handlers.append(('/?', web.RedirectHandler, {'url': settings['default_url'],\n 'permanent': False}))\n", "new_handlers = []\n", "for handler in handlers:\n", "pattern = url_path_join(settings['base_url'], handler[0])\n", "new_handlers.append(('(.*)', Template404))\n", "new_handler = tuple([pattern] + list(handler[1:]))\n", "return new_handlers\n", "new_handlers.append(new_handler)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Import'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Assign'", "For", "Assign'", "Expr'", "Assign'", "Return'", "Expr'" ]
[ "@app.after_request...\n", "VAR_1.headers['Content-Security-Policy'] = (\"default-src 'self'\" + ''.join(\n [(' ' + host) for host in config.config_trustedhosts.strip().split(',')\n ]) +\n \" 'unsafe-inline' 'unsafe-eval'; font-src 'self' data:; img-src 'self' data:\"\n )\n", "if request.endpoint == 'editbook.edit_book' or config.config_use_google_drive:\n", "VAR_1.headers['Content-Security-Policy'] += ' *'\n", "if request.endpoint == 'web.read_book':\n", "VAR_1.headers['X-Content-Type-Options'] = 'nosniff'\n", "VAR_1.headers['Content-Security-Policy'\n ] += \" blob:;style-src-elem 'self' blob: 'unsafe-inline';\"\n", "VAR_1.headers['X-Frame-Options'] = 'SAMEORIGIN'\n", "VAR_1.headers['X-XSS-Protection'] = '1; mode=block'\n", "VAR_1.headers['Strict-Transport-Security'] = 'max-age=31536000;'\n", "return VAR_1\n" ]
[ "@app.after_request...\n", "resp.headers['Content-Security-Policy'] = (\"default-src 'self'\" + ''.join([\n (' ' + host) for host in config.config_trustedhosts.strip().split(',')]\n ) +\n \" 'unsafe-inline' 'unsafe-eval'; font-src 'self' data:; img-src 'self' data:\"\n )\n", "if request.endpoint == 'editbook.edit_book' or config.config_use_google_drive:\n", "resp.headers['Content-Security-Policy'] += ' *'\n", "if request.endpoint == 'web.read_book':\n", "resp.headers['X-Content-Type-Options'] = 'nosniff'\n", "resp.headers['Content-Security-Policy'\n ] += \" blob:;style-src-elem 'self' blob: 'unsafe-inline';\"\n", "resp.headers['X-Frame-Options'] = 'SAMEORIGIN'\n", "resp.headers['X-XSS-Protection'] = '1; mode=block'\n", "resp.headers['Strict-Transport-Security'] = 'max-age=31536000;'\n", "return resp\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Assign'", "Condition", "AugAssign'", "Condition", "Assign'", "AugAssign'", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "@VAR_0.route('/plugins')...\n", "\"\"\"docstring\"\"\"\n", "VAR_110 = VAR_8\n", "flash(str(err), 'danger')\n", "return render_template('plugins.html', VAR_110=plugins_info)\n", "return redirect(url_for('dashboard'))\n" ]
[ "@gui.route('/plugins')...\n", "\"\"\"docstring\"\"\"\n", "plugins_info = plugins\n", "flash(str(err), 'danger')\n", "return render_template('plugins.html', plugins_info=plugins_info)\n", "return redirect(url_for('dashboard'))\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Expr'", "Return'", "Return'" ]
[ "def FUNC_121(*VAR_79, **VAR_42):...\n", "if VAR_13.read_from_replica:\n", "FUNC_6()\n", "VAR_225 = VAR_129(*VAR_79, **get_newargs(fn, kwargs))\n", "if VAR_1 and hasattr(VAR_1, 'primary_db'):\n", "return VAR_225\n", "VAR_1.db.close()\n", "VAR_1.db = VAR_1.primary_db\n" ]
[ "def wrapper_fn(*args, **kwargs):...\n", "if conf.read_from_replica:\n", "connect_replica()\n", "retval = fn(*args, **get_newargs(fn, kwargs))\n", "if local and hasattr(local, 'primary_db'):\n", "return retval\n", "local.db.close()\n", "local.db = local.primary_db\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Expr'", "Assign'", "Condition", "Return'", "Expr'", "Assign'" ]
[ "def FUNC_2(self):...\n", "VAR_9 = test.test_src_dir_path(VAR_0)\n", "self.parser = saved_model_cli.create_parser()\n", "VAR_10 = self.parser.parse_args(['show', '--dir', VAR_9, '--all'])\n", "saved_model_cli.show(VAR_10)\n", "VAR_11 = out.getvalue().strip()\n", "VAR_12 = 'string'\n", "self.maxDiff = None\n", "self.assertMultiLineEqual(VAR_11, VAR_12)\n", "self.assertEqual(err.getvalue().strip(), '')\n" ]
[ "def testShowCommandAll(self):...\n", "base_path = test.test_src_dir_path(SAVED_MODEL_PATH)\n", "self.parser = saved_model_cli.create_parser()\n", "args = self.parser.parse_args(['show', '--dir', base_path, '--all'])\n", "saved_model_cli.show(args)\n", "output = out.getvalue().strip()\n", "exp_out = \"\"\"MetaGraphDef with tag-set: 'serve' contains the following SignatureDefs:\n\nsignature_def['classify_x2_to_y3']:\n The given SavedModel SignatureDef contains the following input(s):\n inputs['inputs'] tensor_info:\n dtype: DT_FLOAT\n shape: (-1, 1)\n name: x2:0\n The given SavedModel SignatureDef contains the following output(s):\n outputs['scores'] tensor_info:\n dtype: DT_FLOAT\n shape: (-1, 1)\n name: y3:0\n Method name is: tensorflow/serving/classify\n\nsignature_def['classify_x_to_y']:\n The given SavedModel SignatureDef contains the following input(s):\n inputs['inputs'] tensor_info:\n dtype: DT_STRING\n shape: unknown_rank\n name: tf_example:0\n The given SavedModel SignatureDef contains the following output(s):\n outputs['scores'] tensor_info:\n dtype: DT_FLOAT\n shape: (-1, 1)\n name: y:0\n Method name is: tensorflow/serving/classify\n\nsignature_def['regress_x2_to_y3']:\n The given SavedModel SignatureDef contains the following input(s):\n inputs['inputs'] tensor_info:\n dtype: DT_FLOAT\n shape: (-1, 1)\n name: x2:0\n The given SavedModel SignatureDef contains the following output(s):\n outputs['outputs'] tensor_info:\n dtype: DT_FLOAT\n shape: (-1, 1)\n name: y3:0\n Method name is: tensorflow/serving/regress\n\nsignature_def['regress_x_to_y']:\n The given SavedModel SignatureDef contains the following input(s):\n inputs['inputs'] tensor_info:\n dtype: DT_STRING\n shape: unknown_rank\n name: tf_example:0\n The given SavedModel SignatureDef contains the following output(s):\n outputs['outputs'] tensor_info:\n dtype: DT_FLOAT\n shape: (-1, 1)\n name: y:0\n Method name is: tensorflow/serving/regress\n\nsignature_def['regress_x_to_y2']:\n The given SavedModel SignatureDef contains the following input(s):\n inputs['inputs'] tensor_info:\n dtype: DT_STRING\n shape: unknown_rank\n name: tf_example:0\n The given SavedModel SignatureDef contains the following output(s):\n outputs['outputs'] tensor_info:\n dtype: DT_FLOAT\n shape: (-1, 1)\n name: y2:0\n Method name is: tensorflow/serving/regress\n\nsignature_def['serving_default']:\n The given SavedModel SignatureDef contains the following input(s):\n inputs['x'] tensor_info:\n dtype: DT_FLOAT\n shape: (-1, 1)\n name: x:0\n The given SavedModel SignatureDef contains the following output(s):\n outputs['y'] tensor_info:\n dtype: DT_FLOAT\n shape: (-1, 1)\n name: y:0\n Method name is: tensorflow/serving/predict\"\"\"\n", "self.maxDiff = None\n", "self.assertMultiLineEqual(output, exp_out)\n", "self.assertEqual(err.getvalue().strip(), '')\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'" ]
[ "def FUNC_18(VAR_5):...\n", "if VAR_5 != '0':\n", "if not calibre_db.session.query(db.Custom_Columns).filter(db.Custom_Columns\n", "return True\n", "return False\n" ]
[ "def check_valid_read_column(column):...\n", "if column != '0':\n", "if not calibre_db.session.query(db.Custom_Columns).filter(db.Custom_Columns\n", "return True\n", "return False\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Condition", "Return'", "Return'" ]
[ "def FUNC_27(VAR_10, VAR_11):...\n", "VAR_69 = VAR_11()\n", "if VAR_69 == ['']:\n", "VAR_69 = []\n", "if not VAR_10['add_element'] in VAR_69:\n", "VAR_69 += [VAR_10['add_element']]\n", "return ','.join(VAR_69)\n" ]
[ "def restriction_addition(element, list_func):...\n", "elementlist = list_func()\n", "if elementlist == ['']:\n", "elementlist = []\n", "if not element['add_element'] in elementlist:\n", "elementlist += [element['add_element']]\n", "return ','.join(elementlist)\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Condition", "Assign'", "Condition", "AugAssign'", "Return'" ]
[ "@defer.inlineCallbacks...\n", "VAR_4 = defer.ensureDeferred(self.handler.set_displayname(self.frank,\n synapse.types.create_requester(self.bob), 'Frank Jr.'))\n", "yield self.assertFailure(VAR_4, AuthError)\n" ]
[ "@defer.inlineCallbacks...\n", "d = defer.ensureDeferred(self.handler.set_displayname(self.frank, synapse.\n types.create_requester(self.bob), 'Frank Jr.'))\n", "yield self.assertFailure(d, AuthError)\n" ]
[ 0, 0, 0 ]
[ "Condition", "Assign'", "Expr'" ]
[ "@property...\n", "return ''.join(self.rawLines)\n" ]
[ "@property...\n", "return ''.join(self.rawLines)\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "def __repr__(self):...\n", "return '<User %r>' % self.name\n" ]
[ "def __repr__(self):...\n", "return '<User %r>' % self.name\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_35(self):...\n", "self.login()\n", "VAR_3 = self.client.post('/password_change/custom/', {'old_password':\n 'password', 'new_password1': 'password1', 'new_password2': 'password1'})\n", "self.assertEqual(VAR_3.status_code, 302)\n", "self.assertURLEqual(VAR_3.url, '/custom/')\n" ]
[ "def test_password_change_redirect_custom(self):...\n", "self.login()\n", "response = self.client.post('/password_change/custom/', {'old_password':\n 'password', 'new_password1': 'password1', 'new_password2': 'password1'})\n", "self.assertEqual(response.status_code, 302)\n", "self.assertURLEqual(response.url, '/custom/')\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Assign'", "Expr'", "Expr'" ]
[ "@log_function...\n", "\"\"\"docstring\"\"\"\n", "VAR_2 = FUNC_1('/groups/%s/roles/%s', VAR_30, VAR_34)\n", "return self.client.get_json(VAR_5=destination, VAR_2=path, VAR_3={\n 'requester_user_id': requester_user_id}, VAR_15=True)\n" ]
[ "@log_function...\n", "\"\"\"docstring\"\"\"\n", "path = _create_v1_path('/groups/%s/roles/%s', group_id, role_id)\n", "return self.client.get_json(destination=destination, path=path, args={\n 'requester_user_id': requester_user_id}, ignore_backoff=True)\n" ]
[ 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Return'" ]
[ "def FUNC_74(self, VAR_44=None):...\n", "\"\"\"docstring\"\"\"\n", "VAR_77 = self.get_signature()\n", "if file_lock.lock_exists(VAR_77):\n", "VAR_96 = True\n", "file_lock.create_lock(VAR_77)\n", "if VAR_44:\n", "for i in range(VAR_44):\n", "if VAR_96:\n", "time.sleep(1)\n", "if not file_lock.lock_exists(VAR_77):\n", "VAR_96 = False\n" ]
[ "def lock(self, timeout=None):...\n", "\"\"\"docstring\"\"\"\n", "signature = self.get_signature()\n", "if file_lock.lock_exists(signature):\n", "lock_exists = True\n", "file_lock.create_lock(signature)\n", "if timeout:\n", "for i in range(timeout):\n", "if lock_exists:\n", "time.sleep(1)\n", "if not file_lock.lock_exists(signature):\n", "lock_exists = False\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Condition", "Assign'", "Expr'", "Condition", "For", "Condition", "Expr'", "Condition", "Assign'" ]
[ "@defer.inlineCallbacks...\n", "VAR_50 = yield VAR_4.verify_json_for_server(*VAR_2)\n", "return VAR_50\n" ]
[ "@defer.inlineCallbacks...\n", "rv1 = yield kr.verify_json_for_server(*args)\n", "return rv1\n" ]
[ 0, 0, 0 ]
[ "Condition", "Assign'", "Return'" ]
[ "def FUNC_25():...\n", "return ''\n" ]
[ "def keepalive():...\n", "return ''\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_24(self, VAR_47: str) ->str:...\n", "\"\"\"docstring\"\"\"\n", "if not self.instances:\n", "return 'master'\n", "if len(self.instances) == 1:\n", "return self.instances[0]\n", "VAR_70 = sha256(VAR_47.encode('utf8')).digest()\n", "VAR_71 = int.from_bytes(VAR_70, byteorder='little')\n", "VAR_72 = VAR_71 % len(self.instances)\n", "return self.instances[VAR_72]\n" ]
[ "def get_instance(self, key: str) ->str:...\n", "\"\"\"docstring\"\"\"\n", "if not self.instances:\n", "return 'master'\n", "if len(self.instances) == 1:\n", "return self.instances[0]\n", "dest_hash = sha256(key.encode('utf8')).digest()\n", "dest_int = int.from_bytes(dest_hash, byteorder='little')\n", "remainder = dest_int % len(self.instances)\n", "return self.instances[remainder]\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Return'", "Condition", "Return'", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_11(VAR_37):...\n", "VAR_4, VAR_39 = VAR_37\n", "VAR_25.write(VAR_4)\n", "return VAR_39\n" ]
[ "def write_to(r):...\n", "data, response = r\n", "output_stream.write(data)\n", "return response\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Expr'", "Return'" ]
[ "@VAR_2.route('/get_tags_json', methods=['GET'])...\n", "return calibre_db.get_typeahead(db.Tags, request.args.get('q'), tag_filter=\n tags_filters())\n" ]
[ "@web.route('/get_tags_json', methods=['GET'])...\n", "return calibre_db.get_typeahead(db.Tags, request.args.get('q'), tag_filter=\n tags_filters())\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "@login_required()...\n", "\"\"\"docstring\"\"\"\n", "VAR_142 = VAR_5.getObject('Image', VAR_26)\n", "if VAR_142 is None:\n", "VAR_181 = VAR_142.loadOriginalMetadata()\n", "VAR_182 = ['[Global Metadata]']\n", "VAR_182.extend([('%s=%s' % (kv[0], kv[1])) for kv in VAR_181[1]])\n", "VAR_182.append('[Series Metadata]')\n", "VAR_182.extend([('%s=%s' % (kv[0], kv[1])) for kv in VAR_181[2]])\n", "VAR_183 = '\\n'.join(VAR_182)\n", "VAR_174 = HttpResponse(VAR_183)\n", "VAR_174['Content-Type'] = 'application/force-download'\n", "VAR_174['Content-Length'] = len(VAR_183)\n", "VAR_174['Content-Disposition'] = 'attachment; filename=Original_Metadata.txt'\n", "return VAR_174\n" ]
[ "@login_required()...\n", "\"\"\"docstring\"\"\"\n", "image = conn.getObject('Image', imageId)\n", "if image is None:\n", "om = image.loadOriginalMetadata()\n", "txtLines = ['[Global Metadata]']\n", "txtLines.extend([('%s=%s' % (kv[0], kv[1])) for kv in om[1]])\n", "txtLines.append('[Series Metadata]')\n", "txtLines.extend([('%s=%s' % (kv[0], kv[1])) for kv in om[2]])\n", "rspText = '\\n'.join(txtLines)\n", "rsp = HttpResponse(rspText)\n", "rsp['Content-Type'] = 'application/force-download'\n", "rsp['Content-Length'] = len(rspText)\n", "rsp['Content-Disposition'] = 'attachment; filename=Original_Metadata.txt'\n", "return rsp\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Condition", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "@VAR_1.url_value_preprocessor...\n", "g.beancount_file_slug = VAR_9.pop('bfile', None) if VAR_9 else None\n", "if not VAR_1.config.get('LEDGERS'):\n", "FUNC_2()\n", "if g.beancount_file_slug:\n", "if g.beancount_file_slug not in VAR_1.config['LEDGERS']:\n", "if not any(g.beancount_file_slug == FUNC_0(VAR_5) for VAR_5 in VAR_1.config\n", "g.ledger = VAR_1.config['LEDGERS'][g.beancount_file_slug]\n", "abort(404)\n", "FUNC_1(VAR_1.config['LEDGERS'].values())\n", "g.conversion = request.args.get('conversion', 'at_cost')\n", "g.interval = Interval.get(request.args.get('interval', 'month'))\n" ]
[ "@app.url_value_preprocessor...\n", "g.beancount_file_slug = values.pop('bfile', None) if values else None\n", "if not app.config.get('LEDGERS'):\n", "_load_file()\n", "if g.beancount_file_slug:\n", "if g.beancount_file_slug not in app.config['LEDGERS']:\n", "if not any(g.beancount_file_slug == ledger_slug(ledger) for ledger in app.\n", "g.ledger = app.config['LEDGERS'][g.beancount_file_slug]\n", "abort(404)\n", "update_ledger_slugs(app.config['LEDGERS'].values())\n", "g.conversion = request.args.get('conversion', 'at_cost')\n", "g.interval = Interval.get(request.args.get('interval', 'month'))\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Assign'", "Condition", "Expr'", "Condition", "Condition", "For", "Assign'", "Expr'", "Expr'", "Assign'", "Assign'" ]
[ "@CLASS_4('backend-warning')...\n", "\"\"\"docstring\"\"\"\n", "VAR_20 = jinja.render('backend-warning.html', distribution=version.\n distribution(), Distribution=version.Distribution, version=\n pkg_resources.parse_version, title='Legacy backend warning')\n", "return 'text/html', VAR_20\n" ]
[ "@add_handler('backend-warning')...\n", "\"\"\"docstring\"\"\"\n", "src = jinja.render('backend-warning.html', distribution=version.\n distribution(), Distribution=version.Distribution, version=\n pkg_resources.parse_version, title='Legacy backend warning')\n", "return 'text/html', src\n" ]
[ 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Return'" ]
[ "def __init__(self, VAR_1=False, VAR_2=False, VAR_3=None, **VAR_4):...\n", "\"\"\"docstring\"\"\"\n", "super(CLASS_0, self).__init__(**kwargs)\n", "self.ignore_login_fail = VAR_1\n", "self.setGroupContext = VAR_2\n", "self.login_redirect = VAR_3\n" ]
[ "def __init__(self, ignore_login_fail=False, setGroupContext=False,...\n", "\"\"\"docstring\"\"\"\n", "super(login_required, self).__init__(**kwargs)\n", "self.ignore_login_fail = ignore_login_fail\n", "self.setGroupContext = setGroupContext\n", "self.login_redirect = login_redirect\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Expr'", "Assign'", "Assign'", "Assign'" ]
[ "def FUNC_34(self):...\n", "if VAR_53.session['admin']:\n", "VAR_105 = self.userdb.getUserList()\n", "return json.dumps({'time': 0, 'userlist': []})\n", "for VAR_51 in VAR_105:\n", "if VAR_51['id'] == VAR_53.session['userid']:\n", "VAR_106 = lambda VAR_51: VAR_51['last_time_online']\n", "VAR_51['deletable'] = False\n", "VAR_119 = self.useroptions.forUser(VAR_51['id'])\n", "VAR_105 = sorted(VAR_105, key=sortfunc, reverse=True)\n", "VAR_120 = VAR_119.getOptionValue('last_time_online')\n", "return json.dumps({'time': int(time.time()), 'userlist': VAR_105})\n", "VAR_121 = VAR_119.getOptionValue('media.may_download')\n", "VAR_51['last_time_online'] = VAR_120\n", "VAR_51['may_download'] = VAR_121\n" ]
[ "def api_getuserlist(self):...\n", "if cherrypy.session['admin']:\n", "userlist = self.userdb.getUserList()\n", "return json.dumps({'time': 0, 'userlist': []})\n", "for user in userlist:\n", "if user['id'] == cherrypy.session['userid']:\n", "sortfunc = lambda user: user['last_time_online']\n", "user['deletable'] = False\n", "user_options = self.useroptions.forUser(user['id'])\n", "userlist = sorted(userlist, key=sortfunc, reverse=True)\n", "t = user_options.getOptionValue('last_time_online')\n", "return json.dumps({'time': int(time.time()), 'userlist': userlist})\n", "may_download = user_options.getOptionValue('media.may_download')\n", "user['last_time_online'] = t\n", "user['may_download'] = may_download\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Assign'", "Return'", "For", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Return'", "Assign'", "Assign'", "Assign'" ]
[ "@CLASS_4('tabs')...\n", "\"\"\"docstring\"\"\"\n", "VAR_21 = collections.defaultdict(list)\n", "for win_id, window in objreg.window_registry.items():\n", "if sip.isdeleted(window):\n", "VAR_20 = jinja.render('tabs.html', title='Tabs', tab_list_by_window=tabs)\n", "VAR_33 = objreg.get('tabbed-browser', scope='window', window=win_id)\n", "return 'text/html', VAR_20\n", "for tab in VAR_33.widgets():\n", "if tab.url() not in [QUrl('qute://tabs/'), QUrl('qute://tabs')]:\n", "VAR_45 = tab.url().toDisplayString()\n", "VAR_21[str(win_id)].append((tab.title(), VAR_45))\n" ]
[ "@add_handler('tabs')...\n", "\"\"\"docstring\"\"\"\n", "tabs = collections.defaultdict(list)\n", "for win_id, window in objreg.window_registry.items():\n", "if sip.isdeleted(window):\n", "html = jinja.render('tabs.html', title='Tabs', tab_list_by_window=tabs)\n", "tabbed_browser = objreg.get('tabbed-browser', scope='window', window=win_id)\n", "return 'text/html', html\n", "for tab in tabbed_browser.widgets():\n", "if tab.url() not in [QUrl('qute://tabs/'), QUrl('qute://tabs')]:\n", "urlstr = tab.url().toDisplayString()\n", "tabs[str(win_id)].append((tab.title(), urlstr))\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "For", "Condition", "Assign'", "Assign'", "Return'", "For", "Condition", "Assign'", "Expr'" ]
[ "def FUNC_14(self, VAR_36):...\n", "\"\"\"docstring\"\"\"\n", "VAR_36 = self._substitute_comments('', VAR_36)\n", "VAR_36 = VAR_36.replace('\\\\', '')\n", "VAR_36 = VAR_6('', VAR_36)\n", "VAR_36 = VAR_36.lower()\n", "if 'javascript:' in VAR_36:\n", "return True\n", "if 'expression(' in VAR_36:\n", "return True\n", "if '@import' in VAR_36:\n", "return True\n", "if '</noscript' in VAR_36:\n", "return True\n", "if VAR_2(VAR_36):\n", "return True\n", "return False\n" ]
[ "def _has_sneaky_javascript(self, style):...\n", "\"\"\"docstring\"\"\"\n", "style = self._substitute_comments('', style)\n", "style = style.replace('\\\\', '')\n", "style = _substitute_whitespace('', style)\n", "style = style.lower()\n", "if 'javascript:' in style:\n", "return True\n", "if 'expression(' in style:\n", "return True\n", "if '@import' in style:\n", "return True\n", "if '</noscript' in style:\n", "return True\n", "if _looks_like_tag_content(style):\n", "return True\n", "return False\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Return'", "Condition", "Return'", "Condition", "Return'", "Condition", "Return'", "Condition", "Return'", "Return'" ]
[ "def FUNC_41(self, VAR_131=False):...\n", "\"\"\"docstring\"\"\"\n", "if not self.settings.allow_basic_login:\n", "return False, False, False\n", "VAR_258 = VAR_263.request.env.http_authorization\n", "if VAR_131:\n", "if callable(VAR_131):\n", "if not VAR_258 or not VAR_258[:6].lower() == 'basic ':\n", "VAR_131 = VAR_131()\n", "if isinstance(VAR_131, string_types):\n", "if VAR_131:\n", "VAR_124, VAR_207, VAR_132 = base64.b64decode(VAR_258[6:]).partition(b':')\n", "VAR_363 = HTTP(401, 'Not Authorized', **{'WWW-Authenticate': \n 'Basic realm=\"' + basic_realm + '\"'})\n", "VAR_449 = to_unicode(VAR_131)\n", "if VAR_131 is True:\n", "return True, False, False\n", "VAR_259 = VAR_207 and self.login_bare(VAR_124, VAR_132)\n", "VAR_449 = '' + VAR_263.request.application\n", "if not VAR_259 and VAR_131:\n", "return True, True, VAR_259\n" ]
[ "def basic(self, basic_auth_realm=False):...\n", "\"\"\"docstring\"\"\"\n", "if not self.settings.allow_basic_login:\n", "return False, False, False\n", "basic = current.request.env.http_authorization\n", "if basic_auth_realm:\n", "if callable(basic_auth_realm):\n", "if not basic or not basic[:6].lower() == 'basic ':\n", "basic_auth_realm = basic_auth_realm()\n", "if isinstance(basic_auth_realm, string_types):\n", "if basic_auth_realm:\n", "username, sep, password = base64.b64decode(basic[6:]).partition(b':')\n", "http_401 = HTTP(401, 'Not Authorized', **{'WWW-Authenticate': \n 'Basic realm=\"' + basic_realm + '\"'})\n", "basic_realm = to_unicode(basic_auth_realm)\n", "if basic_auth_realm is True:\n", "return True, False, False\n", "is_valid_user = sep and self.login_bare(username, password)\n", "basic_realm = '' + current.request.application\n", "if not is_valid_user and basic_auth_realm:\n", "return True, True, is_valid_user\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Return'", "Assign'", "Condition", "Condition", "Condition", "Assign'", "Condition", "Condition", "Assign'", "Assign'", "Assign'", "Condition", "Return'", "Assign'", "Assign'", "Condition", "Return'" ]
[ "def FUNC_59(self, VAR_142=100):...\n", "\"\"\"docstring\"\"\"\n", "if not self.user:\n", "redirect(self.settings.login_url)\n", "if not self.settings.bulk_register_enabled:\n", "return HTTP(404)\n", "VAR_7 = SQLFORM.factory(VAR_1('subject', 'string', VAR_5=self.messages.\n bulk_invite_subject, VAR_279=IS_NOT_EMPTY()), VAR_1('emails', 'text',\n VAR_279=IS_NOT_EMPTY()), VAR_1('message', 'text', VAR_5=self.messages.\n bulk_invite_body, VAR_279=IS_NOT_EMPTY()), VAR_273=self.settings.formstyle)\n", "if VAR_7.process().accepted:\n", "VAR_375 = re.compile('[^\\\\s\\'\"@<>,;:]+\\\\@[^\\\\s\\'\"@<>,;:]+').findall(VAR_7.\n vars.emails)\n", "return VAR_7\n", "VAR_376 = []\n", "VAR_377 = []\n", "VAR_378 = []\n", "for email in VAR_375[:VAR_142]:\n", "if self.table_user()(email=email):\n", "VAR_377 += VAR_375[VAR_142:]\n", "VAR_378.append(email)\n", "VAR_141 = self.register_bare(email=email)\n", "VAR_7 = DIV(H4('Emails sent'), UL(*[A(VAR_30, _href='mailto:' + x) for\n VAR_30 in VAR_376]), H4('Emails failed'), UL(*[A(VAR_30, _href=\n 'mailto:' + x) for VAR_30 in VAR_377]), H4('Emails existing'), UL(*[A(\n VAR_30, _href='mailto:' + x) for VAR_30 in VAR_378]))\n", "if self.email_registration(VAR_7.vars.subject, VAR_7.vars.message, VAR_141):\n", "VAR_376.append(email)\n", "VAR_377.append(email)\n" ]
[ "def bulk_register(self, max_emails=100):...\n", "\"\"\"docstring\"\"\"\n", "if not self.user:\n", "redirect(self.settings.login_url)\n", "if not self.settings.bulk_register_enabled:\n", "return HTTP(404)\n", "form = SQLFORM.factory(Field('subject', 'string', default=self.messages.\n bulk_invite_subject, requires=IS_NOT_EMPTY()), Field('emails', 'text',\n requires=IS_NOT_EMPTY()), Field('message', 'text', default=self.\n messages.bulk_invite_body, requires=IS_NOT_EMPTY()), formstyle=self.\n settings.formstyle)\n", "if form.process().accepted:\n", "emails = re.compile('[^\\\\s\\'\"@<>,;:]+\\\\@[^\\\\s\\'\"@<>,;:]+').findall(form.\n vars.emails)\n", "return form\n", "emails_sent = []\n", "emails_fail = []\n", "emails_exist = []\n", "for email in emails[:max_emails]:\n", "if self.table_user()(email=email):\n", "emails_fail += emails[max_emails:]\n", "emails_exist.append(email)\n", "user = self.register_bare(email=email)\n", "form = DIV(H4('Emails sent'), UL(*[A(x, _href='mailto:' + x) for x in\n emails_sent]), H4('Emails failed'), UL(*[A(x, _href='mailto:' + x) for\n x in emails_fail]), H4('Emails existing'), UL(*[A(x, _href='mailto:' +\n x) for x in emails_exist]))\n", "if self.email_registration(form.vars.subject, form.vars.message, user):\n", "emails_sent.append(email)\n", "emails_fail.append(email)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Expr'", "Condition", "Return'", "Assign'", "Condition", "Assign'", "Return'", "Assign'", "Assign'", "Assign'", "For", "Condition", "AugAssign'", "Expr'", "Assign'", "Assign'", "Condition", "Expr'", "Expr'" ]
[ "def FUNC_20(VAR_56, VAR_46):...\n", "return os.path.exists(VAR_56) and os.access(VAR_56, VAR_46\n ) and not os.path.isdir(VAR_56)\n" ]
[ "def _access_check(fn, mode):...\n", "return os.path.exists(fn) and os.access(fn, mode) and not os.path.isdir(fn)\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_11(self):...\n", "VAR_5 = self._makeContext()\n", "self.assertTrue(VAR_5.evaluate('x | nothing') is None)\n" ]
[ "def test_evaluate_alternative_first_missing(self):...\n", "ec = self._makeContext()\n", "self.assertTrue(ec.evaluate('x | nothing') is None)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Expr'" ]
[ "def FUNC_31():...\n", "\"\"\"docstring\"\"\"\n", "VAR_3 = FUNC_5()\n", "VAR_5 = '/'.join(request.args)\n", "VAR_43.title = request.args[-1]\n", "VAR_71 = read_dict(apath(VAR_5, VAR_122=request))\n", "if '__corrupted__' in VAR_71:\n", "VAR_26 = SPAN(VAR_71['__corrupted__'], VAR_144='error')\n", "VAR_72 = sorted(VAR_71.keys(), VAR_143=lambda x: to_native(x).lower())\n", "return dict(VAR_5=filename, VAR_26=form)\n", "VAR_73 = []\n", "VAR_73.append(H2(T('Original/Translation')))\n", "for VAR_143 in VAR_72:\n", "VAR_10 = md5_hash(VAR_143)\n", "VAR_73.append(DIV(INPUT(_type='submit', _value=T('update'), VAR_144=\n 'btn btn-primary'), VAR_144='controls'))\n", "VAR_140 = VAR_71[VAR_143]\n", "VAR_26 = FORM(*VAR_73)\n", "VAR_141, VAR_142, VAR_143 = VAR_143.partition('\\x01')\n", "if VAR_26.accepts(request.vars, keepvalues=True):\n", "if VAR_142:\n", "VAR_147 = dict()\n", "return dict(VAR_3=request.args[0], VAR_5=filename, VAR_26=form)\n", "VAR_141 = SPAN(VAR_141 + ': ', VAR_144='tm_ftag')\n", "VAR_145, VAR_141 = VAR_141, ''\n", "for VAR_143 in VAR_72:\n", "VAR_145 = VAR_143\n", "VAR_144 = 'untranslated' if VAR_145 == VAR_140 else 'translated'\n", "VAR_10 = md5_hash(VAR_143)\n", "write_dict(apath(VAR_5, VAR_122=request), VAR_147)\n", "if len(VAR_140) <= 40:\n", "if VAR_26.vars[VAR_10] == chr(127):\n", "session.flash = T('file saved on %(time)s', dict(time=time.ctime()))\n", "VAR_187 = INPUT(_type='text', _name=name, VAR_9=s, _size=70, VAR_144=_class)\n", "VAR_187 = TEXTAREA(_name=name, VAR_9=s, _cols=70, _rows=5, VAR_144=_class)\n", "VAR_147[VAR_143] = VAR_26.vars[VAR_10]\n", "redirect(URL(VAR_122=request, VAR_98=request.args))\n", "VAR_145 = VAR_140 != VAR_145 and VAR_145 or B(VAR_145)\n", "VAR_146 = DIV(LABEL(VAR_141, VAR_145, _style='font-weight:normal;'), CAT(\n VAR_187, '\\n', TAG.BUTTON(T('delete'), _onclick='return delkey(\"%s\")' %\n name, VAR_144='btn')), _id=name, VAR_144='span6 well well-small')\n", "VAR_73.append(DIV(VAR_146, VAR_144='row-fluid'))\n" ]
[ "def edit_language():...\n", "\"\"\"docstring\"\"\"\n", "app = get_app()\n", "filename = '/'.join(request.args)\n", "response.title = request.args[-1]\n", "strings = read_dict(apath(filename, r=request))\n", "if '__corrupted__' in strings:\n", "form = SPAN(strings['__corrupted__'], _class='error')\n", "keys = sorted(strings.keys(), key=lambda x: to_native(x).lower())\n", "return dict(filename=filename, form=form)\n", "rows = []\n", "rows.append(H2(T('Original/Translation')))\n", "for key in keys:\n", "name = md5_hash(key)\n", "rows.append(DIV(INPUT(_type='submit', _value=T('update'), _class=\n 'btn btn-primary'), _class='controls'))\n", "s = strings[key]\n", "form = FORM(*rows)\n", "prefix, sep, key = key.partition('\\x01')\n", "if form.accepts(request.vars, keepvalues=True):\n", "if sep:\n", "strs = dict()\n", "return dict(app=request.args[0], filename=filename, form=form)\n", "prefix = SPAN(prefix + ': ', _class='tm_ftag')\n", "k, prefix = prefix, ''\n", "for key in keys:\n", "k = key\n", "_class = 'untranslated' if k == s else 'translated'\n", "name = md5_hash(key)\n", "write_dict(apath(filename, r=request), strs)\n", "if len(s) <= 40:\n", "if form.vars[name] == chr(127):\n", "session.flash = T('file saved on %(time)s', dict(time=time.ctime()))\n", "elem = INPUT(_type='text', _name=name, value=s, _size=70, _class=_class)\n", "elem = TEXTAREA(_name=name, value=s, _cols=70, _rows=5, _class=_class)\n", "strs[key] = form.vars[name]\n", "redirect(URL(r=request, args=request.args))\n", "k = s != k and k or B(k)\n", "new_row = DIV(LABEL(prefix, k, _style='font-weight:normal;'), CAT(elem,\n '\\n', TAG.BUTTON(T('delete'), _onclick='return delkey(\"%s\")' % name,\n _class='btn')), _id=name, _class='span6 well well-small')\n", "rows.append(DIV(new_row, _class='row-fluid'))\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Return'", "Assign'", "Expr'", "For", "Assign'", "Expr'", "Assign'", "Assign'", "Assign'", "Condition", "Condition", "Assign'", "Return'", "Assign'", "Assign'", "For", "Assign'", "Assign'", "Assign'", "Expr'", "Condition", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "Expr'" ]
[ "def FUNC_17(VAR_16, VAR_17, VAR_18):...\n", "\"\"\"docstring\"\"\"\n", "VAR_42 = {}\n", "VAR_43 = FUNC_13(VAR_16)\n", "VAR_44 = FUNC_14(VAR_17)\n", "VAR_45 = FUNC_15(VAR_18)\n", "for VAR_71, (filename, variable_name) in VAR_43.items():\n", "VAR_70 = np.load(file_io.FileIO(filename, mode='rb'), allow_pickle=True)\n", "for VAR_71, py_expr_evaluated in VAR_44.items():\n", "if variable_name:\n", "if VAR_71 in VAR_42:\n", "for VAR_71, VAR_41 in VAR_45.items():\n", "if isinstance(VAR_70, np.ndarray):\n", "if isinstance(VAR_70, np.lib.npyio.NpzFile):\n", "logging.warn('string' % VAR_71)\n", "VAR_42[VAR_71] = py_expr_evaluated\n", "if VAR_71 in VAR_42:\n", "return VAR_42\n", "logging.warn(\n 'Input file %s contains a single ndarray. Name key \"%s\" ignored.' % (\n filename, variable_name))\n", "if variable_name in VAR_70:\n", "VAR_80 = VAR_70.files\n", "VAR_42[VAR_71] = VAR_70\n", "logging.warn(\n 'input_key %s has been specified in multiple options. Value in --input_examples will be used.'\n % VAR_71)\n", "VAR_42[VAR_71] = VAR_41\n", "VAR_42[VAR_71] = VAR_70\n", "VAR_42[VAR_71] = VAR_70[variable_name]\n", "if len(VAR_80) != 1:\n", "VAR_42[VAR_71] = VAR_70[VAR_80[0]]\n" ]
[ "def load_inputs_from_input_arg_string(inputs_str, input_exprs_str,...\n", "\"\"\"docstring\"\"\"\n", "tensor_key_feed_dict = {}\n", "inputs = preprocess_inputs_arg_string(inputs_str)\n", "input_exprs = preprocess_input_exprs_arg_string(input_exprs_str)\n", "input_examples = preprocess_input_examples_arg_string(input_examples_str)\n", "for input_tensor_key, (filename, variable_name) in inputs.items():\n", "data = np.load(file_io.FileIO(filename, mode='rb'), allow_pickle=True)\n", "for input_tensor_key, py_expr_evaluated in input_exprs.items():\n", "if variable_name:\n", "if input_tensor_key in tensor_key_feed_dict:\n", "for input_tensor_key, example in input_examples.items():\n", "if isinstance(data, np.ndarray):\n", "if isinstance(data, np.lib.npyio.NpzFile):\n", "logging.warn(\n 'input_key %s has been specified with both --inputs and --input_exprs options. Value in --input_exprs will be used.'\n % input_tensor_key)\n", "tensor_key_feed_dict[input_tensor_key] = py_expr_evaluated\n", "if input_tensor_key in tensor_key_feed_dict:\n", "return tensor_key_feed_dict\n", "logging.warn(\n 'Input file %s contains a single ndarray. Name key \"%s\" ignored.' % (\n filename, variable_name))\n", "if variable_name in data:\n", "variable_name_list = data.files\n", "tensor_key_feed_dict[input_tensor_key] = data\n", "logging.warn(\n 'input_key %s has been specified in multiple options. Value in --input_examples will be used.'\n % input_tensor_key)\n", "tensor_key_feed_dict[input_tensor_key] = example\n", "tensor_key_feed_dict[input_tensor_key] = data\n", "tensor_key_feed_dict[input_tensor_key] = data[variable_name]\n", "if len(variable_name_list) != 1:\n", "tensor_key_feed_dict[input_tensor_key] = data[variable_name_list[0]]\n" ]
[ 0, 0, 0, 0, 5, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Assign'", "Assign'", "For", "Assign'", "For", "Condition", "Condition", "For", "Condition", "Condition", "Expr'", "Assign'", "Condition", "Return'", "Expr'", "Condition", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "Assign'", "Condition", "Assign'" ]
[ "def FUNC_9(self, *VAR_4, **VAR_5):...\n", "\"\"\"docstring\"\"\"\n", "VAR_54 = VAR_4[0] if VAR_4 else ''\n", "if not VAR_54 in self.handlers:\n", "return \"Error: no such action. '%s'\" % VAR_54\n", "VAR_55 = self.handlers[VAR_54]\n", "VAR_56 = not ('noauth' in dir(VAR_55) and VAR_55.noauth)\n", "if VAR_56 and not self.isAuthorized():\n", "VAR_57 = {}\n", "if 'data' in VAR_5:\n", "VAR_57 = json.loads(VAR_5['data'])\n", "VAR_58 = 'binary' in dir(VAR_55) and VAR_55.binary\n", "if VAR_58:\n", "return VAR_55(**handler_args)\n", "return json.dumps({'data': VAR_55(**handler_args)})\n" ]
[ "def api(self, *args, **kwargs):...\n", "\"\"\"docstring\"\"\"\n", "action = args[0] if args else ''\n", "if not action in self.handlers:\n", "return \"Error: no such action. '%s'\" % action\n", "handler = self.handlers[action]\n", "needsAuth = not ('noauth' in dir(handler) and handler.noauth)\n", "if needsAuth and not self.isAuthorized():\n", "handler_args = {}\n", "if 'data' in kwargs:\n", "handler_args = json.loads(kwargs['data'])\n", "is_binary = 'binary' in dir(handler) and handler.binary\n", "if is_binary:\n", "return handler(**handler_args)\n", "return json.dumps({'data': handler(**handler_args)})\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Condition", "Return'", "Assign'", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Assign'", "Condition", "Return'", "Return'" ]
[ "@staticmethod...\n", "\"\"\"docstring\"\"\"\n", "VAR_29 = oai.OpenAPI.parse_obj(VAR_13)\n", "return GeneratorError(header='Failed to parse OpenAPI document', detail=str(e))\n", "if VAR_29.components is None or VAR_29.components.schemas is None:\n", "VAR_12 = CLASS_4()\n", "VAR_12 = CLASS_4.build(VAR_12=openapi.components.schemas)\n", "VAR_22 = CLASS_1.from_data(VAR_5=openapi.paths)\n", "VAR_23 = EnumProperty.get_all_enums()\n", "return CLASS_5(title=openapi.info.title, description=openapi.info.\n description, version=openapi.info.version, VAR_22=\n endpoint_collections_by_tag, VAR_12=schemas, VAR_23=enums)\n" ]
[ "@staticmethod...\n", "\"\"\"docstring\"\"\"\n", "openapi = oai.OpenAPI.parse_obj(d)\n", "return GeneratorError(header='Failed to parse OpenAPI document', detail=str(e))\n", "if openapi.components is None or openapi.components.schemas is None:\n", "schemas = Schemas()\n", "schemas = Schemas.build(schemas=openapi.components.schemas)\n", "endpoint_collections_by_tag = EndpointCollection.from_data(data=openapi.paths)\n", "enums = EnumProperty.get_all_enums()\n", "return GeneratorData(title=openapi.info.title, description=openapi.info.\n description, version=openapi.info.version, endpoint_collections_by_tag=\n endpoint_collections_by_tag, schemas=schemas, enums=enums)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Return'", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_23(self, VAR_0):...\n", "import pathlib\n", "from jinja2 import Template\n", "from openapi_python_client import GeneratorData, Project\n", "VAR_3 = VAR_0.MagicMock(autospec=GeneratorData, title='My Test API')\n", "VAR_52 = VAR_0.MagicMock(autospec=str)\n", "VAR_53 = VAR_0.MagicMock(autospec=str)\n", "VAR_54 = VAR_0.MagicMock()\n", "VAR_55 = VAR_0.MagicMock()\n", "VAR_3.endpoint_collections_by_tag = {VAR_52: VAR_54, VAR_53: VAR_55}\n", "VAR_8 = Project(VAR_3=openapi)\n", "VAR_8.package_dir = VAR_0.MagicMock()\n", "VAR_56 = VAR_0.MagicMock(autospec=pathlib.Path)\n", "VAR_57 = VAR_0.MagicMock()\n", "VAR_58 = VAR_0.MagicMock(autospec=pathlib.Path)\n", "VAR_59 = VAR_0.MagicMock(autospec=pathlib.Path)\n", "VAR_60 = VAR_0.MagicMock(autospec=pathlib.Path)\n", "VAR_61 = VAR_0.MagicMock(autospec=pathlib.Path)\n", "VAR_62 = VAR_0.MagicMock(autospec=pathlib.Path)\n", "VAR_63 = VAR_0.MagicMock(autospec=pathlib.Path)\n", "VAR_64 = {'__init__.py': VAR_58, f'{VAR_52}.py': VAR_59, f'{VAR_53}.py': VAR_60\n }\n", "VAR_65 = {'__init__.py': VAR_61, f'{VAR_52}.py': VAR_62, f'{VAR_53}.py': VAR_63\n }\n", "VAR_66 = VAR_0.MagicMock(autospec=pathlib.Path)\n", "VAR_66.__truediv__.side_effect = lambda VAR_40: VAR_64[VAR_40]\n", "VAR_67 = VAR_0.MagicMock(autospec=pathlib.Path)\n", "VAR_67.__truediv__.side_effect = lambda VAR_40: VAR_65[VAR_40]\n", "VAR_68 = {'client.py': VAR_57, 'api': VAR_66, 'async_api': VAR_67,\n 'errors.py': VAR_56}\n", "VAR_8.package_dir.__truediv__.side_effect = lambda VAR_40: VAR_68[VAR_40]\n", "VAR_69 = VAR_0.MagicMock(autospec=Template)\n", "VAR_70 = VAR_0.MagicMock(autospec=Template)\n", "VAR_71 = VAR_0.MagicMock(autospec=Template)\n", "VAR_72 = VAR_0.MagicMock(autospec=Template)\n", "VAR_27 = {'client.pyi': VAR_69, 'errors.pyi': VAR_70, 'endpoint_module.pyi':\n VAR_71, 'async_endpoint_module.pyi': VAR_72}\n", "VAR_0.patch.object(VAR_8.env, 'get_template', autospec=True, side_effect=lambda\n x: templates[x])\n", "VAR_73 = {VAR_54: VAR_0.MagicMock(), VAR_55: VAR_0.MagicMock()}\n", "VAR_71.render.side_effect = lambda collection: VAR_73[collection]\n", "VAR_74 = {VAR_54: VAR_0.MagicMock(), VAR_55: VAR_0.MagicMock()}\n", "VAR_72.render.side_effect = lambda collection: VAR_74[collection]\n", "VAR_8._build_api()\n", "VAR_8.package_dir.__truediv__.assert_has_calls([VAR_0.call(key) for key in\n VAR_68])\n", "VAR_8.env.get_template.assert_has_calls([VAR_0.call(key) for key in VAR_27])\n", "VAR_69.render.assert_called_once()\n", "VAR_57.write_text.assert_called_once_with(VAR_69.render())\n", "VAR_70.render.assert_called_once()\n", "VAR_56.write_text.assert_called_once_with(VAR_70.render())\n", "VAR_66.mkdir.assert_called_once()\n", "VAR_66.__truediv__.assert_has_calls([VAR_0.call(key) for key in VAR_64])\n", "VAR_58.write_text.assert_called_once_with(\n '\"\"\" Contains synchronous methods for accessing the API \"\"\"')\n", "VAR_71.render.assert_has_calls([VAR_0.call(collection=collection_1), VAR_0.\n call(collection=collection_2)])\n", "VAR_59.write_text.assert_called_once_with(VAR_73[VAR_54])\n", "VAR_60.write_text.assert_called_once_with(VAR_73[VAR_55])\n", "VAR_67.mkdir.assert_called_once()\n", "VAR_67.__truediv__.assert_has_calls([VAR_0.call(key) for key in VAR_65])\n", "VAR_61.write_text.assert_called_once_with(\n '\"\"\" Contains async methods for accessing the API \"\"\"')\n", "VAR_72.render.assert_has_calls([VAR_0.call(collection=collection_1), VAR_0.\n call(collection=collection_2)])\n", "VAR_62.write_text.assert_called_once_with(VAR_74[VAR_54])\n", "VAR_63.write_text.assert_called_once_with(VAR_74[VAR_55])\n" ]
[ "def test__build_api(self, mocker):...\n", "import pathlib\n", "from jinja2 import Template\n", "from openapi_python_client import GeneratorData, Project\n", "openapi = mocker.MagicMock(autospec=GeneratorData, title='My Test API')\n", "tag_1 = mocker.MagicMock(autospec=str)\n", "tag_2 = mocker.MagicMock(autospec=str)\n", "collection_1 = mocker.MagicMock()\n", "collection_2 = mocker.MagicMock()\n", "openapi.endpoint_collections_by_tag = {tag_1: collection_1, tag_2: collection_2\n }\n", "project = Project(openapi=openapi)\n", "project.package_dir = mocker.MagicMock()\n", "api_errors = mocker.MagicMock(autospec=pathlib.Path)\n", "client_path = mocker.MagicMock()\n", "api_init = mocker.MagicMock(autospec=pathlib.Path)\n", "collection_1_path = mocker.MagicMock(autospec=pathlib.Path)\n", "collection_2_path = mocker.MagicMock(autospec=pathlib.Path)\n", "async_api_init = mocker.MagicMock(autospec=pathlib.Path)\n", "async_collection_1_path = mocker.MagicMock(autospec=pathlib.Path)\n", "async_collection_2_path = mocker.MagicMock(autospec=pathlib.Path)\n", "api_paths = {'__init__.py': api_init, f'{tag_1}.py': collection_1_path,\n f'{tag_2}.py': collection_2_path}\n", "async_api_paths = {'__init__.py': async_api_init, f'{tag_1}.py':\n async_collection_1_path, f'{tag_2}.py': async_collection_2_path}\n", "api_dir = mocker.MagicMock(autospec=pathlib.Path)\n", "api_dir.__truediv__.side_effect = lambda x: api_paths[x]\n", "async_api_dir = mocker.MagicMock(autospec=pathlib.Path)\n", "async_api_dir.__truediv__.side_effect = lambda x: async_api_paths[x]\n", "package_paths = {'client.py': client_path, 'api': api_dir, 'async_api':\n async_api_dir, 'errors.py': api_errors}\n", "project.package_dir.__truediv__.side_effect = lambda x: package_paths[x]\n", "client_template = mocker.MagicMock(autospec=Template)\n", "errors_template = mocker.MagicMock(autospec=Template)\n", "endpoint_template = mocker.MagicMock(autospec=Template)\n", "async_endpoint_template = mocker.MagicMock(autospec=Template)\n", "templates = {'client.pyi': client_template, 'errors.pyi': errors_template,\n 'endpoint_module.pyi': endpoint_template, 'async_endpoint_module.pyi':\n async_endpoint_template}\n", "mocker.patch.object(project.env, 'get_template', autospec=True, side_effect\n =lambda x: templates[x])\n", "endpoint_renders = {collection_1: mocker.MagicMock(), collection_2: mocker.\n MagicMock()}\n", "endpoint_template.render.side_effect = lambda collection: endpoint_renders[\n collection]\n", "async_endpoint_renders = {collection_1: mocker.MagicMock(), collection_2:\n mocker.MagicMock()}\n", "async_endpoint_template.render.side_effect = (lambda collection:\n async_endpoint_renders[collection])\n", "project._build_api()\n", "project.package_dir.__truediv__.assert_has_calls([mocker.call(key) for key in\n package_paths])\n", "project.env.get_template.assert_has_calls([mocker.call(key) for key in\n templates])\n", "client_template.render.assert_called_once()\n", "client_path.write_text.assert_called_once_with(client_template.render())\n", "errors_template.render.assert_called_once()\n", "api_errors.write_text.assert_called_once_with(errors_template.render())\n", "api_dir.mkdir.assert_called_once()\n", "api_dir.__truediv__.assert_has_calls([mocker.call(key) for key in api_paths])\n", "api_init.write_text.assert_called_once_with(\n '\"\"\" Contains synchronous methods for accessing the API \"\"\"')\n", "endpoint_template.render.assert_has_calls([mocker.call(collection=\n collection_1), mocker.call(collection=collection_2)])\n", "collection_1_path.write_text.assert_called_once_with(endpoint_renders[\n collection_1])\n", "collection_2_path.write_text.assert_called_once_with(endpoint_renders[\n collection_2])\n", "async_api_dir.mkdir.assert_called_once()\n", "async_api_dir.__truediv__.assert_has_calls([mocker.call(key) for key in\n async_api_paths])\n", "async_api_init.write_text.assert_called_once_with(\n '\"\"\" Contains async methods for accessing the API \"\"\"')\n", "async_endpoint_template.render.assert_has_calls([mocker.call(collection=\n collection_1), mocker.call(collection=collection_2)])\n", "async_collection_1_path.write_text.assert_called_once_with(\n async_endpoint_renders[collection_1])\n", "async_collection_2_path.write_text.assert_called_once_with(\n async_endpoint_renders[collection_2])\n" ]
[ 0, 0, 0, 0, 0, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Import'", "ImportFrom'", "ImportFrom'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'" ]
[ "@log_function...\n", "\"\"\"docstring\"\"\"\n", "VAR_2 = FUNC_1('/groups/%s/categories/%s', VAR_30, VAR_33)\n", "return self.client.post_json(VAR_5=destination, VAR_2=path, VAR_3={\n 'requester_user_id': requester_user_id}, VAR_39=content, VAR_15=True)\n" ]
[ "@log_function...\n", "\"\"\"docstring\"\"\"\n", "path = _create_v1_path('/groups/%s/categories/%s', group_id, category_id)\n", "return self.client.post_json(destination=destination, path=path, args={\n 'requester_user_id': requester_user_id}, data=content, ignore_backoff=True)\n" ]
[ 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Return'" ]
[ "def FUNC_33(self):...\n", "VAR_39 = 's0_0_0_0_0_0_0_0_0'\n", "VAR_22, VAR_23 = self.make_request('GET', \n '/rooms/%s/messages?access_token=x&from=%s' % (self.room_id, VAR_39))\n", "self.assertEquals(200, VAR_23.code)\n", "self.assertTrue('start' in VAR_23.json_body)\n", "self.assertEquals(VAR_39, VAR_23.json_body['start'])\n", "self.assertTrue('chunk' in VAR_23.json_body)\n", "self.assertTrue('end' in VAR_23.json_body)\n" ]
[ "def test_stream_token_is_accepted_for_fwd_pagianation(self):...\n", "token = 's0_0_0_0_0_0_0_0_0'\n", "request, channel = self.make_request('GET', \n '/rooms/%s/messages?access_token=x&from=%s' % (self.room_id, token))\n", "self.assertEquals(200, channel.code)\n", "self.assertTrue('start' in channel.json_body)\n", "self.assertEquals(token, channel.json_body['start'])\n", "self.assertTrue('chunk' in channel.json_body)\n", "self.assertTrue('end' in channel.json_body)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'" ]
[ "def FUNC_2(self, VAR_5, VAR_6, VAR_3=None):...\n", "if VAR_3 is None:\n", "VAR_3 = {}\n", "if self.rel.to in self.admin_site._registry:\n", "VAR_3['class'] = 'vManyToManyRawIdAdminField'\n", "if VAR_6:\n", "VAR_6 = ','.join([force_text(VAR_39) for VAR_39 in VAR_6])\n", "VAR_6 = ''\n", "return super(CLASS_8, self).render(VAR_5, VAR_6, VAR_3)\n" ]
[ "def render(self, name, value, attrs=None):...\n", "if attrs is None:\n", "attrs = {}\n", "if self.rel.to in self.admin_site._registry:\n", "attrs['class'] = 'vManyToManyRawIdAdminField'\n", "if value:\n", "value = ','.join([force_text(v) for v in value])\n", "value = ''\n", "return super(ManyToManyRawIdWidget, self).render(name, value, attrs)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Assign'", "Return'" ]
[ "@CLASS_4('spawn-output')...\n", "\"\"\"docstring\"\"\"\n", "VAR_20 = jinja.render('pre.html', title='spawn output', content=spawn_output)\n", "return 'text/html', VAR_20\n" ]
[ "@add_handler('spawn-output')...\n", "\"\"\"docstring\"\"\"\n", "html = jinja.render('pre.html', title='spawn output', content=spawn_output)\n", "return 'text/html', html\n" ]
[ 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Return'" ]
[ "def FUNC_9():...\n", "VAR_16 = 0\n", "for _ in f.readlines():\n", "VAR_16 += 1\n", "yield '<?xml version=\"1.0\" encoding=\"UTF-8\" ?>\\n'\n", "if VAR_8 is not None and VAR_16 >= VAR_8:\n", "if VAR_1:\n", "yield '<?xml-stylesheet type=\"text/xsl\" href=\"/scopelist.xsl\" ?>\\n'\n", "yield '<objectlist count=\"{:d}\">\\n'.format(VAR_16)\n", "VAR_17 = 0\n", "for VAR_13 in f.readlines():\n", "VAR_13 = VAR_13.strip()\n", "yield '</objectlist>\\n'\n", "yield FUNC_4(VAR_9=path) + '\\n'\n", "VAR_17 += 1\n", "if VAR_8 is not None and VAR_17 >= VAR_8:\n" ]
[ "def generate():...\n", "num_entries = 0\n", "for _ in f.readlines():\n", "num_entries += 1\n", "yield '<?xml version=\"1.0\" encoding=\"UTF-8\" ?>\\n'\n", "if limit is not None and num_entries >= limit:\n", "if STYLE:\n", "yield '<?xml-stylesheet type=\"text/xsl\" href=\"/scopelist.xsl\" ?>\\n'\n", "yield '<objectlist count=\"{:d}\">\\n'.format(num_entries)\n", "count = 0\n", "for path in f.readlines():\n", "path = path.strip()\n", "yield '</objectlist>\\n'\n", "yield _get_object_element(object_path=path) + '\\n'\n", "count += 1\n", "if limit is not None and count >= limit:\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "For", "AugAssign'", "Expr'", "Condition", "Condition", "Expr'", "Expr'", "Assign'", "For", "Assign'", "Expr'", "Expr'", "AugAssign'", "Condition" ]
[ "def FUNC_51(VAR_62, VAR_9, VAR_109=0):...\n", "\"\"\"docstring\"\"\"\n", "if VAR_12.exists(VAR_62, VAR_9):\n", "FUNC_50(VAR_62, VAR_9, VAR_109=force)\n" ]
[ "def delete_doc_if_exists(doctype, name, force=0):...\n", "\"\"\"docstring\"\"\"\n", "if db.exists(doctype, name):\n", "delete_doc(doctype, name, force=force)\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Expr'" ]
[ "def __init__(self, VAR_198=None, **VAR_182):...\n", "if not VAR_198:\n", "self.__dict__.clear()\n", "VAR_222 = self.__getattr__(VAR_198)\n", "VAR_222.installed = True\n", "VAR_222.update((VAR_346, v) for VAR_346, v in VAR_182.items() if VAR_346 not in\n VAR_222)\n" ]
[ "def __init__(self, plugin=None, **defaults):...\n", "if not plugin:\n", "self.__dict__.clear()\n", "settings = self.__getattr__(plugin)\n", "settings.installed = True\n", "settings.update((k, v) for k, v in defaults.items() if k not in settings)\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Expr'", "Assign'", "Assign'", "Expr'" ]
[ "@FUNC_0...\n", "return Keyring(self)\n" ]
[ "@cache_in_self...\n", "return Keyring(self)\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "@VAR_1.route('/ajax/xchange', methods=['POST'])...\n", "VAR_79 = VAR_32.get_json().get('xchange')\n", "if VAR_79:\n", "for val in VAR_79:\n", "return ''\n", "VAR_39 = False\n", "return json.dumps({'success': True})\n", "VAR_17 = calibre_db.get_book(val)\n", "VAR_115 = VAR_17.title\n", "VAR_116 = calibre_db.order_authors(VAR_17)\n", "VAR_53 = []\n", "for VAR_38 in VAR_116.authors:\n", "VAR_53.append(VAR_38.name.replace('|', ','))\n", "VAR_91 = FUNC_27(VAR_17, ' '.join(VAR_53))\n", "VAR_62, VAR_92 = FUNC_28(VAR_17, VAR_115)\n", "if VAR_92 or VAR_91:\n", "VAR_90 = VAR_17.id\n", "if config.config_use_google_drive:\n", "VAR_39 = True\n", "gdriveutils.updateGdriveCalibreFromLocal()\n", "if VAR_90:\n", "helper.update_dir_stucture(VAR_90, config.config_calibre_dir, VAR_62[0])\n", "if VAR_39:\n", "VAR_17.last_modified = datetime.utcnow()\n", "calibre_db.session.commit()\n", "calibre_db.session.rollback()\n", "if config.config_use_google_drive:\n", "VAR_2.error('Database error: %s', e)\n", "gdriveutils.updateGdriveCalibreFromLocal()\n", "return json.dumps({'success': False})\n" ]
[ "@editbook.route('/ajax/xchange', methods=['POST'])...\n", "vals = request.get_json().get('xchange')\n", "if vals:\n", "for val in vals:\n", "return ''\n", "modif_date = False\n", "return json.dumps({'success': True})\n", "book = calibre_db.get_book(val)\n", "authors = book.title\n", "entries = calibre_db.order_authors(book)\n", "author_names = []\n", "for authr in entries.authors:\n", "author_names.append(authr.name.replace('|', ','))\n", "title_change = handle_title_on_edit(book, ' '.join(author_names))\n", "input_authors, authorchange = handle_author_on_edit(book, authors)\n", "if authorchange or title_change:\n", "edited_books_id = book.id\n", "if config.config_use_google_drive:\n", "modif_date = True\n", "gdriveutils.updateGdriveCalibreFromLocal()\n", "if edited_books_id:\n", "helper.update_dir_stucture(edited_books_id, config.config_calibre_dir,\n input_authors[0])\n", "if modif_date:\n", "book.last_modified = datetime.utcnow()\n", "calibre_db.session.commit()\n", "calibre_db.session.rollback()\n", "if config.config_use_google_drive:\n", "log.error('Database error: %s', e)\n", "gdriveutils.updateGdriveCalibreFromLocal()\n", "return json.dumps({'success': False})\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Assign'", "Condition", "For", "Return'", "Assign'", "Return'", "Assign'", "Assign'", "Assign'", "Assign'", "For", "Expr'", "Assign'", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Expr'", "Condition", "Expr'", "Condition", "Assign'", "Expr'", "Expr'", "Condition", "Expr'", "Expr'", "Return'" ]
[ "async def FUNC_15(VAR_20):...\n", "" ]
[ "async def get_key(key_server):...\n", "" ]
[ 0, 0 ]
[ "AsyncFunctionDef'", "Condition" ]
[ "def FUNC_14(self, VAR_30: bytes) ->None:...\n", "self.stream.write(VAR_30)\n", "self.length += len(VAR_30)\n", "if self.max_size is not None and self.length >= self.max_size:\n", "self.deferred.errback(SynapseError(502, \n 'Requested file is too large > %r bytes' % (self.max_size,), Codes.\n TOO_LARGE))\n", "self.deferred = defer.Deferred()\n", "self.transport.loseConnection()\n" ]
[ "def dataReceived(self, data: bytes) ->None:...\n", "self.stream.write(data)\n", "self.length += len(data)\n", "if self.max_size is not None and self.length >= self.max_size:\n", "self.deferred.errback(SynapseError(502, \n 'Requested file is too large > %r bytes' % (self.max_size,), Codes.\n TOO_LARGE))\n", "self.deferred = defer.Deferred()\n", "self.transport.loseConnection()\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "AugAssign'", "Condition", "Expr'", "Assign'", "Expr'" ]
[ "def FUNC_16(self):...\n", "VAR_5 = self._makeContext()\n", "self.assertEqual(VAR_5.evaluate('x | string:x'), 'x')\n" ]
[ "def test_hybrid_with_string_expression(self):...\n", "ec = self._makeContext()\n", "self.assertEqual(ec.evaluate('x | string:x'), 'x')\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Expr'" ]
[ "def FUNC_38(self, VAR_2='password'):...\n", "VAR_17 = reverse('login')\n", "for bad_url in ('http://example.com', 'https://example.com',\n", "VAR_30 = '%(url)s?%(next)s=%(bad_url)s' % {'url': VAR_17, 'next':\n REDIRECT_FIELD_NAME, 'bad_url': urlquote(bad_url)}\n", "for good_url in ('/view/?param=http://example.com',\n", "VAR_3 = self.client.post(VAR_30, {'username': 'testclient', 'password': VAR_2})\n", "VAR_31 = '%(url)s?%(next)s=%(good_url)s' % {'url': VAR_17, 'next':\n REDIRECT_FIELD_NAME, 'good_url': urlquote(good_url)}\n", "self.assertEqual(VAR_3.status_code, 302)\n", "VAR_3 = self.client.post(VAR_31, {'username': 'testclient', 'password': VAR_2})\n", "self.assertFalse(bad_url in VAR_3.url, '%s should be blocked' % bad_url)\n", "self.assertEqual(VAR_3.status_code, 302)\n", "self.assertTrue(good_url in VAR_3.url, '%s should be allowed' % good_url)\n" ]
[ "def test_security_check(self, password='password'):...\n", "login_url = reverse('login')\n", "for bad_url in ('http://example.com', 'https://example.com',\n", "nasty_url = '%(url)s?%(next)s=%(bad_url)s' % {'url': login_url, 'next':\n REDIRECT_FIELD_NAME, 'bad_url': urlquote(bad_url)}\n", "for good_url in ('/view/?param=http://example.com',\n", "response = self.client.post(nasty_url, {'username': 'testclient',\n 'password': password})\n", "safe_url = '%(url)s?%(next)s=%(good_url)s' % {'url': login_url, 'next':\n REDIRECT_FIELD_NAME, 'good_url': urlquote(good_url)}\n", "self.assertEqual(response.status_code, 302)\n", "response = self.client.post(safe_url, {'username': 'testclient', 'password':\n password})\n", "self.assertFalse(bad_url in response.url, '%s should be blocked' % bad_url)\n", "self.assertEqual(response.status_code, 302)\n", "self.assertTrue(good_url in response.url, '%s should be allowed' % good_url)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "For", "Assign'", "For", "Assign'", "Assign'", "Expr'", "Assign'", "Expr'", "Expr'", "Expr'" ]
[ "def FUNC_19(self, VAR_6, VAR_8, VAR_10=False):...\n", "VAR_13, VAR_15 = [], []\n", "VAR_25 = frappe.desk.query_report.run(self.name, VAR_6=filters, VAR_8=user,\n VAR_10=ignore_prepared_report)\n", "for d in VAR_25.get('columns'):\n", "if isinstance(d, dict):\n", "VAR_15 += VAR_25.get('result')\n", "VAR_36 = frappe._dict(d)\n", "VAR_37, VAR_38 = 'Data', None\n", "return VAR_13, VAR_15\n", "if not VAR_36.fieldname:\n", "VAR_11 = d.split(':')\n", "VAR_36.fieldname = VAR_36.label\n", "VAR_13.append(VAR_36)\n", "if len(VAR_11) > 1:\n", "if VAR_11[1]:\n", "VAR_13.append(frappe._dict(VAR_32=parts[0], VAR_37=fieldtype, fieldname=\n parts[0], VAR_38=options))\n", "VAR_37, VAR_38 = VAR_11[1], None\n", "if VAR_37 and '/' in VAR_37:\n", "VAR_37, VAR_38 = VAR_37.split('/')\n" ]
[ "def run_query_report(self, filters, user, ignore_prepared_report=False):...\n", "columns, result = [], []\n", "data = frappe.desk.query_report.run(self.name, filters=filters, user=user,\n ignore_prepared_report=ignore_prepared_report)\n", "for d in data.get('columns'):\n", "if isinstance(d, dict):\n", "result += data.get('result')\n", "col = frappe._dict(d)\n", "fieldtype, options = 'Data', None\n", "return columns, result\n", "if not col.fieldname:\n", "parts = d.split(':')\n", "col.fieldname = col.label\n", "columns.append(col)\n", "if len(parts) > 1:\n", "if parts[1]:\n", "columns.append(frappe._dict(label=parts[0], fieldtype=fieldtype, fieldname=\n parts[0], options=options))\n", "fieldtype, options = parts[1], None\n", "if fieldtype and '/' in fieldtype:\n", "fieldtype, options = fieldtype.split('/')\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "For", "Condition", "AugAssign'", "Assign'", "Assign'", "Return'", "Condition", "Assign'", "Assign'", "Expr'", "Condition", "Condition", "Expr'", "Assign'", "Condition", "Assign'" ]
[ "def FUNC_1(VAR_2):...\n", "if VAR_2.is_public:\n", "return True\n", "if current_user.is_anonymous or VAR_2.user_id != current_user.id:\n", "VAR_1.error('User is unauthorized to view non-public shelf: %s', VAR_2)\n", "return True\n", "return False\n" ]
[ "def check_shelf_view_permissions(cur_shelf):...\n", "if cur_shelf.is_public:\n", "return True\n", "if current_user.is_anonymous or cur_shelf.user_id != current_user.id:\n", "log.error('User is unauthorized to view non-public shelf: %s', cur_shelf)\n", "return True\n", "return False\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Return'", "Condition", "Expr'", "Return'", "Return'" ]
[ "def __init__(self, VAR_58, *VAR_6, **VAR_7):...\n", "self.project = VAR_58\n", "super().__init__(*VAR_6, **kwargs)\n", "self.fields['group'].queryset = VAR_58.defined_groups.all()\n" ]
[ "def __init__(self, project, *args, **kwargs):...\n", "self.project = project\n", "super().__init__(*args, **kwargs)\n", "self.fields['group'].queryset = project.defined_groups.all()\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Expr'", "Assign'" ]
[ "def FUNC_13(self, VAR_5):...\n", "\"\"\"docstring\"\"\"\n", "VAR_8 = getattr(self, 'create_%s_field' % VAR_5, None)\n", "if VAR_8:\n", "return VAR_8\n", "import inspect\n", "VAR_10 = [f[0] for f in inspect.getmembers(self.__class__, inspect.\n isfunction) if f[0].startswith('create_') and f[0].endswith('_field')]\n" ]
[ "def get_create_field_function(self, type):...\n", "\"\"\"docstring\"\"\"\n", "create_field_function = getattr(self, 'create_%s_field' % type, None)\n", "if create_field_function:\n", "return create_field_function\n", "import inspect\n", "method_list = [f[0] for f in inspect.getmembers(self.__class__, inspect.\n isfunction) if f[0].startswith('create_') and f[0].endswith('_field')]\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Condition", "Return'", "Import'", "Assign'" ]
[ "def FUNC_4(VAR_14):...\n", "VAR_23 = []\n", "for r in VAR_14:\n", "VAR_25 = {'value': r[0], 'description': ', '.join(unique(cstr(d) for d in r if\n d)[1:])}\n", "return VAR_23\n", "VAR_23.append(VAR_25)\n" ]
[ "def build_for_autosuggest(res):...\n", "results = []\n", "for r in res:\n", "out = {'value': r[0], 'description': ', '.join(unique(cstr(d) for d in r if\n d)[1:])}\n", "return results\n", "results.append(out)\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "For", "Assign'", "Return'", "Expr'" ]
[ "def FUNC_1(VAR_1):...\n", "VAR_8 = magic.from_buffer(VAR_1, mime=True)\n", "if VAR_8[:5] != 'image':\n", "return VAR_1, False\n", "return VAR_1, True\n" ]
[ "def allowed_file(enc_data):...\n", "mimetype = magic.from_buffer(enc_data, mime=True)\n", "if mimetype[:5] != 'image':\n", "return enc_data, False\n", "return enc_data, True\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Condition", "Return'", "Return'" ]
[ "def FUNC_8(self, VAR_3, VAR_4):...\n", "VAR_17 = self.default_config()\n", "VAR_17['require_auth_for_profile_requests'] = True\n", "VAR_17['limit_profile_requests_to_users_who_share_rooms'] = True\n", "self.hs = self.setup_test_homeserver(VAR_17=config)\n", "return self.hs\n" ]
[ "def make_homeserver(self, reactor, clock):...\n", "config = self.default_config()\n", "config['require_auth_for_profile_requests'] = True\n", "config['limit_profile_requests_to_users_who_share_rooms'] = True\n", "self.hs = self.setup_test_homeserver(config=config)\n", "return self.hs\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_20(self, VAR_3):...\n", "\"\"\"docstring\"\"\"\n", "if not VAR_3 or not self.index_urls:\n", "return None\n", "for u in self.index_urls:\n", "VAR_94 = remove_auth_from_url(u).rstrip('/') + '/'\n", "if VAR_3.startswith(VAR_94):\n", "return u\n" ]
[ "def _get_index_url(self, url):...\n", "\"\"\"docstring\"\"\"\n", "if not url or not self.index_urls:\n", "return None\n", "for u in self.index_urls:\n", "prefix = remove_auth_from_url(u).rstrip('/') + '/'\n", "if url.startswith(prefix):\n", "return u\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Return'", "For", "Assign'", "Condition", "Return'" ]
[ "def FUNC_0(self, VAR_3, VAR_4):...\n", "self.push_attempts = []\n", "VAR_5 = Mock()\n", "def FUNC_9(VAR_6, VAR_7):...\n", "VAR_25 = Deferred()\n", "self.push_attempts.append((VAR_25, VAR_6, VAR_7))\n", "return make_deferred_yieldable(VAR_25)\n" ]
[ "def make_homeserver(self, reactor, clock):...\n", "self.push_attempts = []\n", "m = Mock()\n", "def post_json_get_json(url, body):...\n", "d = Deferred()\n", "self.push_attempts.append((d, url, body))\n", "return make_deferred_yieldable(d)\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "FunctionDef'", "Assign'", "Expr'", "Return'" ]
[ "def FUNC_16(self, VAR_1):...\n", "VAR_5 = VAR_1.MagicMock()\n", "VAR_26 = VAR_1.MagicMock()\n", "VAR_23 = oai.Schema(anyOf=[{'type': 'number', 'default': '0.0'}, {'type':\n 'integer', 'default': '0'}])\n", "VAR_33 = VAR_1.patch(f'{VAR_0}.UnionProperty')\n", "VAR_32 = VAR_1.patch(f'{VAR_0}.FloatProperty')\n", "VAR_34 = VAR_1.patch(f'{VAR_0}.IntProperty')\n", "from openapi_python_client.parser.properties import property_from_data\n", "VAR_4 = property_from_data(VAR_5=name, VAR_26=required, VAR_23=data)\n", "VAR_32.assert_called_once_with(VAR_5=name, VAR_26=required, default='0.0')\n", "VAR_34.assert_called_once_with(VAR_5=name, VAR_26=required, default='0')\n", "VAR_33.assert_called_once_with(VAR_5=name, VAR_26=required, default=None,\n inner_properties=[FloatProperty.return_value, IntProperty.return_value])\n", "assert VAR_4 == VAR_33.return_value\n" ]
[ "def test_property_from_data_union(self, mocker):...\n", "name = mocker.MagicMock()\n", "required = mocker.MagicMock()\n", "data = oai.Schema(anyOf=[{'type': 'number', 'default': '0.0'}, {'type':\n 'integer', 'default': '0'}])\n", "UnionProperty = mocker.patch(f'{MODULE_NAME}.UnionProperty')\n", "FloatProperty = mocker.patch(f'{MODULE_NAME}.FloatProperty')\n", "IntProperty = mocker.patch(f'{MODULE_NAME}.IntProperty')\n", "from openapi_python_client.parser.properties import property_from_data\n", "p = property_from_data(name=name, required=required, data=data)\n", "FloatProperty.assert_called_once_with(name=name, required=required, default\n ='0.0')\n", "IntProperty.assert_called_once_with(name=name, required=required, default='0')\n", "UnionProperty.assert_called_once_with(name=name, required=required, default\n =None, inner_properties=[FloatProperty.return_value, IntProperty.\n return_value])\n", "assert p == UnionProperty.return_value\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "ImportFrom'", "Assign'", "Expr'", "Expr'", "Expr'", "Assert'" ]
[ "async def FUNC_5(self):...\n", "return '1a2b3c4d5e6f'\n" ]
[ "async def get_resolved_ref(self):...\n", "return '1a2b3c4d5e6f'\n" ]
[ 0, 0 ]
[ "AsyncFunctionDef'", "Return'" ]
[ "@VAR_5.route('/scope/<gididx>')...\n", "VAR_12 = 'GIDIDX' + VAR_6.upper()\n", "def FUNC_9():...\n", "yield '<?xml version=\"1.0\" encoding=\"UTF-8\" ?>\\n'\n", "if VAR_1:\n", "yield '<?xml-stylesheet type=\"text/xsl\" href=\"/scopelist.xsl\" ?>\\n'\n", "yield '<objectlist>\\n'\n", "for line in f:\n", "VAR_10 = line.strip()\n", "yield '</objectlist>\\n'\n", "VAR_13 = str(FUNC_5(VAR_10))\n", "VAR_16 = Headers([('Content-Type', 'text/xml')])\n", "VAR_25 = FUNC_7(VAR_13)\n", "print('Error parsing {}. {}. Skip.'.format(VAR_10, str(e)), file=sys.stderr)\n", "return Response(stream_with_context(FUNC_9()), status='200 OK', VAR_16=headers)\n", "VAR_26 = float(VAR_25['format']['duration'])\n", "VAR_27 = int(ceil(VAR_26 / VAR_7))\n", "yield '<count adjust=\"{}\"/>\\n'.format(VAR_27)\n", "for clip in range(VAR_27):\n", "yield FUNC_4(VAR_9=clip * stride, VAR_8=span, VAR_10=video) + '\\n'\n" ]
[ "@scope_blueprint.route('/scope/<gididx>')...\n", "index = 'GIDIDX' + gididx.upper()\n", "def generate():...\n", "yield '<?xml version=\"1.0\" encoding=\"UTF-8\" ?>\\n'\n", "if STYLE:\n", "yield '<?xml-stylesheet type=\"text/xsl\" href=\"/scopelist.xsl\" ?>\\n'\n", "yield '<objectlist>\\n'\n", "for line in f:\n", "video = line.strip()\n", "yield '</objectlist>\\n'\n", "video_path = str(_get_obj_absolute_path(video))\n", "headers = Headers([('Content-Type', 'text/xml')])\n", "video_meta = _ffprobe(video_path)\n", "print('Error parsing {}. {}. Skip.'.format(video, str(e)), file=sys.stderr)\n", "return Response(stream_with_context(generate()), status='200 OK', headers=\n headers)\n", "length_sec = float(video_meta['format']['duration'])\n", "num_clips = int(ceil(length_sec / stride))\n", "yield '<count adjust=\"{}\"/>\\n'.format(num_clips)\n", "for clip in range(num_clips):\n", "yield _get_object_element(start=clip * stride, span=span, video=video) + '\\n'\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Assign'", "FunctionDef'", "Expr'", "Condition", "Expr'", "Expr'", "For", "Assign'", "Expr'", "Assign'", "Assign'", "Assign'", "Expr'", "Return'", "Assign'", "Assign'", "Expr'", "For", "Expr'" ]
[ "def FUNC_23(VAR_9, VAR_5, VAR_10):...\n", "VAR_13 = calibre_db.session.query(db.Ratings).filter(db.Ratings.id == VAR_5\n ).first()\n", "VAR_63, VAR_68, VAR_65 = calibre_db.fill_indexpage(VAR_9, 0, db.Books, db.\n Books.ratings.any(db.Ratings.id == VAR_5), [VAR_10[0][0]])\n", "if VAR_13 and VAR_13.rating <= 10:\n", "return render_title_template('index.html', VAR_68=random, VAR_65=pagination,\n VAR_63=entries, id=book_id, VAR_150=_(u'Rating: %(rating)s stars',\n rating=int(name.rating / 2)), VAR_9='ratings', VAR_10=order[1])\n", "abort(404)\n" ]
[ "def render_ratings_books(page, book_id, order):...\n", "name = calibre_db.session.query(db.Ratings).filter(db.Ratings.id == book_id\n ).first()\n", "entries, random, pagination = calibre_db.fill_indexpage(page, 0, db.Books,\n db.Books.ratings.any(db.Ratings.id == book_id), [order[0][0]])\n", "if name and name.rating <= 10:\n", "return render_title_template('index.html', random=random, pagination=\n pagination, entries=entries, id=book_id, title=_(\n u'Rating: %(rating)s stars', rating=int(name.rating / 2)), page=\n 'ratings', order=order[1])\n", "abort(404)\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Condition", "Return'", "Expr'" ]