lines
sequencelengths
1
383
raw_lines
sequencelengths
1
383
label
sequencelengths
1
383
type
sequencelengths
1
383
[ "def FUNC_3(self) ->bool:...\n", "\"\"\"docstring\"\"\"\n", "return False\n" ]
[ "def hasDirectory(self) ->bool:...\n", "\"\"\"docstring\"\"\"\n", "return False\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Return'" ]
[ "def FUNC_4(VAR_17):...\n", "VAR_39 = []\n", "for VAR_23, VAR_17 in pluginManager.get_hooks(VAR_17).items():\n", "return VAR_39\n", "VAR_128 = VAR_17()\n", "VAR_0.exception(f'Error fetching theming CSS to include from plugin {VAR_23}',\n extra={'plugin': name})\n", "if isinstance(VAR_128, (tuple, list)):\n", "VAR_39 += VAR_128\n" ]
[ "def _add_additional_assets(hook):...\n", "result = []\n", "for name, hook in pluginManager.get_hooks(hook).items():\n", "return result\n", "assets = hook()\n", "_logger.exception(f'Error fetching theming CSS to include from plugin {name}',\n extra={'plugin': name})\n", "if isinstance(assets, (tuple, list)):\n", "result += assets\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "For", "Return'", "Assign'", "Expr'", "Condition", "AugAssign'" ]
[ "def FUNC_11(VAR_11):...\n", "return os.path.join(VAR_3, VAR_11)\n" ]
[ "def _get_index_absolute_path(index):...\n", "return os.path.join(INDEXDIR, index)\n" ]
[ 0, 1 ]
[ "FunctionDef'", "Return'" ]
[ "@pytest.mark.linux...\n", "\"\"\"docstring\"\"\"\n", "VAR_11 = ['--temp-basedir'] + FUNC_0(VAR_4.config) + ['/home/user/föö.html']\n", "VAR_6.start(VAR_11, VAR_16={'LC_ALL': 'C'}, wait_focus=False)\n", "if not VAR_4.config.webengine:\n", "VAR_18 = VAR_6.wait_for(message=\n 'Error while loading *: Error opening /*: No such file or directory')\n", "VAR_6.wait_for(message=\n \"load status for <* tab_id=* url='*/f*.html'>: LoadStatus.error\")\n", "VAR_18.expected = True\n" ]
[ "@pytest.mark.linux...\n", "\"\"\"docstring\"\"\"\n", "args = ['--temp-basedir'] + _base_args(request.config) + ['/home/user/föö.html'\n ]\n", "quteproc_new.start(args, env={'LC_ALL': 'C'}, wait_focus=False)\n", "if not request.config.webengine:\n", "line = quteproc_new.wait_for(message=\n 'Error while loading *: Error opening /*: No such file or directory')\n", "quteproc_new.wait_for(message=\n \"load status for <* tab_id=* url='*/f*.html'>: LoadStatus.error\")\n", "line.expected = True\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Expr'", "Condition", "Assign'", "Expr'", "Assign'" ]
[ "def FUNC_46(VAR_134):...\n", "def FUNC_47(VAR_136, VAR_38):...\n", "return VAR_134(VAR_136, VAR_38)\n", "VAR_0.exception('Error while extracting sorting keys for template {}'.\n format(VAR_117))\n", "return FUNC_47\n", "return None\n" ]
[ "def create_safe_extractor(extractor):...\n", "def f(x, k):...\n", "return extractor(x, k)\n", "_logger.exception('Error while extracting sorting keys for template {}'.\n format(t))\n", "return f\n", "return None\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "FunctionDef'", "Return'", "Expr'", "Return'", "Return'" ]
[ "@staticmethod...\n", "VAR_26 = VAR_13(VAR_24.text)\n", "if not VAR_26:\n" ]
[ "@staticmethod...\n", "is_valid = function(document.text)\n", "if not is_valid:\n" ]
[ 0, 0, 0 ]
[ "Condition", "Assign'", "Condition" ]
[ "def FUNC_14(VAR_3):...\n", "\"\"\"docstring\"\"\"\n", "VAR_17 = QUrlQuery(VAR_3)\n", "VAR_24 = VAR_17.queryItemValue('option', QUrl.FullyDecoded)\n", "VAR_25 = VAR_17.queryItemValue('value', QUrl.FullyDecoded)\n", "if VAR_24 == 'content.javascript.enabled' and VAR_25 == 'false':\n", "VAR_40 = (\n 'Refusing to disable javascript via qute://settings as it needs javascript support.'\n )\n", "config.instance.set_str(VAR_24, VAR_25, save_yaml=True)\n", "message.error(str(e))\n", "message.error(VAR_40)\n", "return 'text/html', b'ok'\n", "return 'text/html', b'error: ' + str(e).encode('utf-8')\n", "return 'text/html', b'error: ' + VAR_40.encode('utf-8')\n" ]
[ "def _qute_settings_set(url):...\n", "\"\"\"docstring\"\"\"\n", "query = QUrlQuery(url)\n", "option = query.queryItemValue('option', QUrl.FullyDecoded)\n", "value = query.queryItemValue('value', QUrl.FullyDecoded)\n", "if option == 'content.javascript.enabled' and value == 'false':\n", "msg = (\n 'Refusing to disable javascript via qute://settings as it needs javascript support.'\n )\n", "config.instance.set_str(option, value, save_yaml=True)\n", "message.error(str(e))\n", "message.error(msg)\n", "return 'text/html', b'ok'\n", "return 'text/html', b'error: ' + str(e).encode('utf-8')\n", "return 'text/html', b'error: ' + msg.encode('utf-8')\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Expr'", "Expr'", "Expr'", "Return'", "Return'", "Return'" ]
[ "def FUNC_44(self):...\n", "return json.dumps(audiotranscode.getEncoders())\n" ]
[ "def api_getencoders(self):...\n", "return json.dumps(audiotranscode.getEncoders())\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "@VAR_8.route('/api/', methods=['GET'])...\n", "VAR_50 = [type(inp) for inp in VAR_8.interface.input_components]\n", "VAR_51 = [type(out) for out in VAR_8.interface.output_components]\n", "VAR_52, VAR_53 = FUNC_11(VAR_50, 'input')\n", "VAR_54, VAR_55 = FUNC_11(VAR_51, 'output')\n", "VAR_56 = [type(inp).__name__ for inp in VAR_8.interface.input_components]\n", "VAR_57 = [type(out).__name__ for out in VAR_8.interface.output_components]\n", "VAR_58 = [inp.generate_sample() for inp in VAR_8.interface.input_components]\n", "VAR_59 = {'inputs': VAR_56, 'outputs': VAR_57, 'len_inputs': len(VAR_50),\n 'len_outputs': len(VAR_51), 'inputs_lower': [name.lower() for name in\n VAR_56], 'outputs_lower': [name.lower() for name in VAR_57],\n 'input_types': VAR_53, 'output_types': VAR_55, 'input_types_doc':\n VAR_52, 'output_types_doc': VAR_54, 'sample_inputs': VAR_58}\n", "return render_template('api_docs.html', **docs)\n" ]
[ "@app.route('/api/', methods=['GET'])...\n", "inputs = [type(inp) for inp in app.interface.input_components]\n", "outputs = [type(out) for out in app.interface.output_components]\n", "input_types_doc, input_types = get_types(inputs, 'input')\n", "output_types_doc, output_types = get_types(outputs, 'output')\n", "input_names = [type(inp).__name__ for inp in app.interface.input_components]\n", "output_names = [type(out).__name__ for out in app.interface.output_components]\n", "sample_inputs = [inp.generate_sample() for inp in app.interface.\n input_components]\n", "docs = {'inputs': input_names, 'outputs': output_names, 'len_inputs': len(\n inputs), 'len_outputs': len(outputs), 'inputs_lower': [name.lower() for\n name in input_names], 'outputs_lower': [name.lower() for name in\n output_names], 'input_types': input_types, 'output_types': output_types,\n 'input_types_doc': input_types_doc, 'output_types_doc':\n output_types_doc, 'sample_inputs': sample_inputs}\n", "return render_template('api_docs.html', **docs)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "def __getstate__(self):...\n", "return self\n" ]
[ "def __getstate__(self):...\n", "return self\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_38(self):...\n", "from django.urls import reverse\n", "return str(reverse('helpdesk:list')) + '?kbitem=' + str(self.pk)\n" ]
[ "def query_url(self):...\n", "from django.urls import reverse\n", "return str(reverse('helpdesk:list')) + '?kbitem=' + str(self.pk)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "ImportFrom'", "Return'" ]
[ "@login_required()...\n", "\"\"\"docstring\"\"\"\n", "VAR_6['viewport_server'] = VAR_25 is not None and VAR_350('webindex'\n ) + VAR_25 or VAR_350('webindex')\n", "VAR_6['viewport_server'] = VAR_6['viewport_server'].rstrip('/')\n", "return webgateway_views.full_viewer(VAR_2, VAR_41, **kwargs)\n" ]
[ "@login_required()...\n", "\"\"\"docstring\"\"\"\n", "kwargs['viewport_server'] = share_id is not None and reverse('webindex'\n ) + share_id or reverse('webindex')\n", "kwargs['viewport_server'] = kwargs['viewport_server'].rstrip('/')\n", "return webgateway_views.full_viewer(request, iid, **kwargs)\n" ]
[ 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Assign'", "Return'" ]
[ "@VAR_2.route('/downloadlist')...\n", "if VAR_87.get_view_property('download', 'dir') == 'desc':\n", "VAR_10 = ub.User.name.desc()\n", "VAR_10 = ub.User.name.asc()\n", "VAR_109 = 0\n", "VAR_109 = 1\n", "if VAR_87.check_visibility(constants.SIDEBAR_DOWNLOAD) and VAR_87.role_admin():\n", "VAR_63 = ub.session.query(ub.User, func.count(ub.Downloads.book_id).label(\n 'count')).join(ub.Downloads).group_by(ub.Downloads.user_id).order_by(VAR_10\n ).all()\n", "abort(404)\n", "VAR_107 = ub.session.query(func.upper(func.substr(ub.User.name, 1, 1)).\n label('char')).filter(ub.User.role.op('&')(constants.ROLE_ANONYMOUS) !=\n constants.ROLE_ANONYMOUS).group_by(func.upper(func.substr(ub.User.name,\n 1, 1))).all()\n", "return render_title_template('list.html', VAR_63=entries, folder=\n 'web.books_list', VAR_107=charlist, VAR_150=_(u'Downloads'), VAR_9=\n 'downloadlist', VAR_8='download', VAR_10=order_no)\n" ]
[ "@web.route('/downloadlist')...\n", "if current_user.get_view_property('download', 'dir') == 'desc':\n", "order = ub.User.name.desc()\n", "order = ub.User.name.asc()\n", "order_no = 0\n", "order_no = 1\n", "if current_user.check_visibility(constants.SIDEBAR_DOWNLOAD\n", "entries = ub.session.query(ub.User, func.count(ub.Downloads.book_id).label(\n 'count')).join(ub.Downloads).group_by(ub.Downloads.user_id).order_by(order\n ).all()\n", "abort(404)\n", "charlist = ub.session.query(func.upper(func.substr(ub.User.name, 1, 1)).\n label('char')).filter(ub.User.role.op('&')(constants.ROLE_ANONYMOUS) !=\n constants.ROLE_ANONYMOUS).group_by(func.upper(func.substr(ub.User.name,\n 1, 1))).all()\n", "return render_title_template('list.html', entries=entries, folder=\n 'web.books_list', charlist=charlist, title=_(u'Downloads'), page=\n 'downloadlist', data='download', order=order_no)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Expr'", "Assign'", "Return'" ]
[ "def FUNC_28(self):...\n", "self.parser = saved_model_cli.create_parser()\n", "VAR_9 = test.test_src_dir_path(VAR_0)\n", "VAR_41 = os.path.join(test.get_temp_dir(), 'new_dir')\n", "VAR_10 = self.parser.parse_args(['run', '--dir', VAR_9, '--tag_set',\n 'serve', '--signature_def', 'regress_x_to_y', '--input_examples',\n 'inputs={\"x\":8.0,\"x2\":5.0}', '--outdir', VAR_41])\n", "saved_model_cli.run(VAR_10)\n" ]
[ "def testRunCommandInputExamplesNotListError(self):...\n", "self.parser = saved_model_cli.create_parser()\n", "base_path = test.test_src_dir_path(SAVED_MODEL_PATH)\n", "output_dir = os.path.join(test.get_temp_dir(), 'new_dir')\n", "args = self.parser.parse_args(['run', '--dir', base_path, '--tag_set',\n 'serve', '--signature_def', 'regress_x_to_y', '--input_examples',\n 'inputs={\"x\":8.0,\"x2\":5.0}', '--outdir', output_dir])\n", "saved_model_cli.run(args)\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'" ]
[ "def FUNC_2(self, VAR_1, VAR_2=None):...\n", "print('matched define search: not implemented')\n", "self.write('')\n" ]
[ "def get(self, arg, word=None):...\n", "print('matched define search: not implemented')\n", "self.write('')\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Expr'" ]
[ "def FUNC_5(self):...\n", "self.get_success(self.inject_room_member(self.room1, self.u_alice,\n Membership.JOIN))\n", "VAR_16 = self.get_success(self.inject_message(self.room1, self.u_alice, 't'))\n", "VAR_14 = self.get_success(self.store.get_event(VAR_16.event_id))\n", "self.assertObjectHasAttributes({'type': EventTypes.Message, 'user_id': self\n .u_alice.to_string(), 'content': {'body': 't', 'msgtype': 'message'}},\n VAR_14)\n", "self.assertFalse('redacted_because' in VAR_14.unsigned)\n", "VAR_10 = 'Because I said so'\n", "self.get_success(self.inject_redaction(self.room1, VAR_16.event_id, self.\n u_alice, VAR_10))\n", "VAR_14 = self.get_success(self.store.get_event(VAR_16.event_id))\n", "self.assertEqual(VAR_16.event_id, VAR_14.event_id)\n", "self.assertTrue('redacted_because' in VAR_14.unsigned)\n", "self.assertObjectHasAttributes({'type': EventTypes.Message, 'user_id': self\n .u_alice.to_string(), 'content': {}}, VAR_14)\n", "self.assertObjectHasAttributes({'type': EventTypes.Redaction, 'user_id':\n self.u_alice.to_string(), 'content': {'reason': VAR_10}}, VAR_14.\n unsigned['redacted_because'])\n" ]
[ "def test_redact(self):...\n", "self.get_success(self.inject_room_member(self.room1, self.u_alice,\n Membership.JOIN))\n", "msg_event = self.get_success(self.inject_message(self.room1, self.u_alice, 't')\n )\n", "event = self.get_success(self.store.get_event(msg_event.event_id))\n", "self.assertObjectHasAttributes({'type': EventTypes.Message, 'user_id': self\n .u_alice.to_string(), 'content': {'body': 't', 'msgtype': 'message'}},\n event)\n", "self.assertFalse('redacted_because' in event.unsigned)\n", "reason = 'Because I said so'\n", "self.get_success(self.inject_redaction(self.room1, msg_event.event_id, self\n .u_alice, reason))\n", "event = self.get_success(self.store.get_event(msg_event.event_id))\n", "self.assertEqual(msg_event.event_id, event.event_id)\n", "self.assertTrue('redacted_because' in event.unsigned)\n", "self.assertObjectHasAttributes({'type': EventTypes.Message, 'user_id': self\n .u_alice.to_string(), 'content': {}}, event)\n", "self.assertObjectHasAttributes({'type': EventTypes.Redaction, 'user_id':\n self.u_alice.to_string(), 'content': {'reason': reason}}, event.\n unsigned['redacted_because'])\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Assign'", "Assign'", "Expr'", "Expr'", "Assign'", "Expr'", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'" ]
[ "def FUNC_17(self):...\n", "return {'id': self.id}\n" ]
[ "def to_json(self):...\n", "return {'id': self.id}\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_1(VAR_1: Type['pl.Trainer'], VAR_4: Union[ArgumentParser, Namespace]...\n", "\"\"\"docstring\"\"\"\n", "VAR_2 = VAR_4.parse_args() if isinstance(VAR_4, ArgumentParser) else VAR_4\n", "VAR_16 = {arg: (VAR_36, VAR_32) for arg, VAR_36, VAR_32 in FUNC_3(VAR_1)}\n", "VAR_17 = {}\n", "for VAR_28, VAR_43 in vars(VAR_2).items():\n", "if VAR_28 in VAR_16 and VAR_43 is None:\n", "return Namespace(**modified_args)\n", "VAR_36, VAR_32 = VAR_16[VAR_28]\n", "VAR_17[VAR_28] = VAR_43\n", "if bool in VAR_36 and isinstance(VAR_32, bool):\n", "VAR_43 = True\n" ]
[ "def parse_argparser(cls: Type['pl.Trainer'], arg_parser: Union[...\n", "\"\"\"docstring\"\"\"\n", "args = arg_parser.parse_args() if isinstance(arg_parser, ArgumentParser\n ) else arg_parser\n", "types_default = {arg: (arg_types, arg_default) for arg, arg_types,\n arg_default in get_init_arguments_and_types(cls)}\n", "modified_args = {}\n", "for k, v in vars(args).items():\n", "if k in types_default and v is None:\n", "return Namespace(**modified_args)\n", "arg_types, arg_default = types_default[k]\n", "modified_args[k] = v\n", "if bool in arg_types and isinstance(arg_default, bool):\n", "v = True\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Assign'", "Assign'", "For", "Condition", "Return'", "Assign'", "Assign'", "Condition", "Assign'" ]
[ "def FUNC_4(VAR_9: Any, VAR_10: int) ->Any:...\n", "\"\"\"docstring\"\"\"\n", "if VAR_10 == 0:\n", "if isinstance(VAR_9, dict):\n", "return {key: FUNC_4(VAR_4, VAR_10 - 1) for key, VAR_4 in VAR_9.items()}\n", "if isinstance(VAR_9, list):\n", "return [FUNC_4(item, VAR_10 - 1) for item in VAR_9]\n", "if isinstance(VAR_9, str):\n", "return FUNC_1(VAR_9)\n", "return VAR_9\n" ]
[ "def validate_json(json: Any, max_depth: int) ->Any:...\n", "\"\"\"docstring\"\"\"\n", "if max_depth == 0:\n", "if isinstance(json, dict):\n", "return {key: validate_json(value, max_depth - 1) for key, value in json.items()\n }\n", "if isinstance(json, list):\n", "return [validate_json(item, max_depth - 1) for item in json]\n", "if isinstance(json, str):\n", "return validate_html_strict(json)\n", "return json\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Condition", "Return'", "Condition", "Return'", "Condition", "Return'", "Return'" ]
[ "def FUNC_117(VAR_10):...\n", "def FUNC_168(*VAR_9, **VAR_13):...\n", "VAR_27 = None\n", "VAR_27 = VAR_10(*VAR_9, **b)\n", "thread.start_new_thread(VAR_26, (VAR_27,))\n", "return FUNC_168\n", "return VAR_27\n" ]
[ "def _completion(f):...\n", "def __completion(*a, **b):...\n", "d = None\n", "d = f(*a, **b)\n", "thread.start_new_thread(callback, (d,))\n", "return __completion\n", "return d\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "FunctionDef'", "Assign'", "Assign'", "Expr'", "Return'", "Return'" ]
[ "def FUNC_29(VAR_90, VAR_51=False):...\n", "\"\"\"docstring\"\"\"\n", "if VAR_1.flags.in_test:\n", "return\n", "if not isinstance(VAR_90, (tuple, list)):\n", "VAR_90 = VAR_90,\n", "VAR_90 = set(VAR_90)\n", "VAR_182 = set(FUNC_24())\n", "if not VAR_90.intersection(VAR_182):\n", "if VAR_51:\n", "FUNC_15(FUNC_0('This action is only allowed for {}').format(FUNC_101(', '.\n join(VAR_90))), FUNC_0('Not Permitted'))\n" ]
[ "def only_for(roles, message=False):...\n", "\"\"\"docstring\"\"\"\n", "if local.flags.in_test:\n", "return\n", "if not isinstance(roles, (tuple, list)):\n", "roles = roles,\n", "roles = set(roles)\n", "myroles = set(get_roles())\n", "if not roles.intersection(myroles):\n", "if message:\n", "msgprint(_('This action is only allowed for {}').format(bold(', '.join(\n roles))), _('Not Permitted'))\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Return'", "Condition", "Assign'", "Assign'", "Assign'", "Condition", "Condition", "Expr'" ]
[ "def FUNC_10(VAR_5, VAR_2, VAR_6, **VAR_7):...\n", "\"\"\"docstring\"\"\"\n", "if VAR_6:\n", "CLASS_15.objects.create(VAR_74=instance)\n" ]
[ "def create_usersettings(sender, instance, created, **kwargs):...\n", "\"\"\"docstring\"\"\"\n", "if created:\n", "UserSettings.objects.create(user=instance)\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Expr'" ]
[ "def FUNC_28(self):...\n", "VAR_11 = self.folder.t\n", "VAR_11.write(\"<p tal:define='p a//b' />\")\n", "VAR_11()\n" ]
[ "def testBadExpression(self):...\n", "t = self.folder.t\n", "t.write(\"<p tal:define='p a//b' />\")\n", "t()\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Expr'", "Expr'" ]
[ "@defer.inlineCallbacks...\n", "self.mock_resource = MockHttpResource(prefix=PATH_PREFIX)\n", "self.mock_handler = Mock(spec=['get_displayname', 'set_displayname',\n 'get_avatar_url', 'set_avatar_url', 'check_profile_query_allowed'])\n", "self.mock_handler.get_displayname.return_value = defer.succeed(Mock())\n", "self.mock_handler.set_displayname.return_value = defer.succeed(Mock())\n", "self.mock_handler.get_avatar_url.return_value = defer.succeed(Mock())\n", "self.mock_handler.set_avatar_url.return_value = defer.succeed(Mock())\n", "self.mock_handler.check_profile_query_allowed.return_value = defer.succeed(Mock\n ())\n", "VAR_5 = yield setup_test_homeserver(self.addCleanup, 'test', http_client=\n None, resource_for_client=self.mock_resource, federation=Mock(),\n federation_client=Mock(), profile_handler=self.mock_handler)\n", "async def FUNC_20(VAR_9=None, VAR_10=False):...\n", "return synapse.types.create_requester(VAR_0)\n" ]
[ "@defer.inlineCallbacks...\n", "self.mock_resource = MockHttpResource(prefix=PATH_PREFIX)\n", "self.mock_handler = Mock(spec=['get_displayname', 'set_displayname',\n 'get_avatar_url', 'set_avatar_url', 'check_profile_query_allowed'])\n", "self.mock_handler.get_displayname.return_value = defer.succeed(Mock())\n", "self.mock_handler.set_displayname.return_value = defer.succeed(Mock())\n", "self.mock_handler.get_avatar_url.return_value = defer.succeed(Mock())\n", "self.mock_handler.set_avatar_url.return_value = defer.succeed(Mock())\n", "self.mock_handler.check_profile_query_allowed.return_value = defer.succeed(Mock\n ())\n", "hs = yield setup_test_homeserver(self.addCleanup, 'test', http_client=None,\n resource_for_client=self.mock_resource, federation=Mock(),\n federation_client=Mock(), profile_handler=self.mock_handler)\n", "async def _get_user_by_req(request=None, allow_guest=False):...\n", "return synapse.types.create_requester(myid)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 4, 0, 0 ]
[ "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "AsyncFunctionDef'", "Return'" ]
[ "def FUNC_23(self):...\n", "VAR_19 = '{\"topic\":\"Seasons\",\"subtopic\":\"Summer\"}'\n", "VAR_22, VAR_23 = self.make_request('PUT', self.path, VAR_19)\n", "self.assertEquals(200, VAR_23.code, msg=channel.result['body'])\n", "VAR_22, VAR_23 = self.make_request('GET', self.path)\n", "self.assertEquals(200, VAR_23.code, msg=channel.result['body'])\n", "self.assert_dict(json.loads(VAR_19), VAR_23.json_body)\n" ]
[ "def test_rooms_topic_with_extra_keys(self):...\n", "content = '{\"topic\":\"Seasons\",\"subtopic\":\"Summer\"}'\n", "request, channel = self.make_request('PUT', self.path, content)\n", "self.assertEquals(200, channel.code, msg=channel.result['body'])\n", "request, channel = self.make_request('GET', self.path)\n", "self.assertEquals(200, channel.code, msg=channel.result['body'])\n", "self.assert_dict(json.loads(content), channel.json_body)\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Expr'", "Assign'", "Expr'", "Expr'" ]
[ "def FUNC_2(VAR_4: Dict[str, Any]) ->Union[datetime, date]:...\n", "VAR_5: Union[datetime, date]\n", "VAR_5 = datetime.fromisoformat(VAR_0['aCamelDateTime'])\n", "VAR_5 = date.fromisoformat(VAR_0['aCamelDateTime'])\n", "return VAR_5\n", "return VAR_5\n" ]
[ "def _parse_a_camel_date_time(data: Dict[str, Any]) ->Union[datetime, date]:...\n", "a_camel_date_time: Union[datetime, date]\n", "a_camel_date_time = datetime.fromisoformat(d['aCamelDateTime'])\n", "a_camel_date_time = date.fromisoformat(d['aCamelDateTime'])\n", "return a_camel_date_time\n", "return a_camel_date_time\n" ]
[ 0, 5, 5, 5, 0, 0 ]
[ "FunctionDef'", "AnnAssign'", "Assign'", "Assign'", "Return'", "Return'" ]
[ "def FUNC_49(self, VAR_11=False):...\n", "\"\"\"docstring\"\"\"\n", "frappe.delete_doc(self.doctype, self.name, VAR_11=ignore_permissions, flags\n =self.flags)\n" ]
[ "def delete(self, ignore_permissions=False):...\n", "\"\"\"docstring\"\"\"\n", "frappe.delete_doc(self.doctype, self.name, ignore_permissions=\n ignore_permissions, flags=self.flags)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Expr'" ]
[ "def FUNC_51(VAR_21, VAR_26):...\n", "if VAR_26:\n", "if config.config_read_column:\n", "return VAR_21\n", "if VAR_26 == 'True':\n", "if VAR_26 == 'True':\n", "VAR_3.error(u'Custom Column No.%d is not existing in calibre database',\n config.config_read_column)\n", "VAR_21 = VAR_21.join(ub.ReadBook, db.Books.id == ub.ReadBook.book_id,\n isouter=True).filter(ub.ReadBook.user_id == int(VAR_87.id), ub.ReadBook\n .read_status == ub.ReadBook.STATUS_FINISHED)\n", "VAR_21 = VAR_21.join(ub.ReadBook, db.Books.id == ub.ReadBook.book_id,\n isouter=True).filter(ub.ReadBook.user_id == int(VAR_87.id), coalesce(ub\n .ReadBook.read_status, 0) != ub.ReadBook.STATUS_FINISHED)\n", "VAR_21 = VAR_21.join(db.cc_classes[config.config_read_column], isouter=True\n ).filter(db.cc_classes[config.config_read_column].value == True)\n", "VAR_21 = VAR_21.join(db.cc_classes[config.config_read_column], isouter=True\n ).filter(coalesce(db.cc_classes[config.config_read_column].value, False\n ) != True)\n", "flash(_('Custom Column No.%(column)d is not existing in calibre database',\n column=config.config_read_column), category='error')\n", "return VAR_21\n" ]
[ "def adv_search_read_status(q, read_status):...\n", "if read_status:\n", "if config.config_read_column:\n", "return q\n", "if read_status == 'True':\n", "if read_status == 'True':\n", "log.error(u'Custom Column No.%d is not existing in calibre database',\n config.config_read_column)\n", "q = q.join(ub.ReadBook, db.Books.id == ub.ReadBook.book_id, isouter=True\n ).filter(ub.ReadBook.user_id == int(current_user.id), ub.ReadBook.\n read_status == ub.ReadBook.STATUS_FINISHED)\n", "q = q.join(ub.ReadBook, db.Books.id == ub.ReadBook.book_id, isouter=True\n ).filter(ub.ReadBook.user_id == int(current_user.id), coalesce(ub.\n ReadBook.read_status, 0) != ub.ReadBook.STATUS_FINISHED)\n", "q = q.join(db.cc_classes[config.config_read_column], isouter=True).filter(\n db.cc_classes[config.config_read_column].value == True)\n", "q = q.join(db.cc_classes[config.config_read_column], isouter=True).filter(\n coalesce(db.cc_classes[config.config_read_column].value, False) != True)\n", "flash(_('Custom Column No.%(column)d is not existing in calibre database',\n column=config.config_read_column), category='error')\n", "return q\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Condition", "Return'", "Condition", "Condition", "Expr'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Return'" ]
[ "@staticmethod...\n", "return VAR_418.path.splitext(VAR_10)[-1].lower() in ('.png', '.jpg',\n '.jpeg', '.gif', '.tiff')\n" ]
[ "@staticmethod...\n", "return os.path.splitext(f)[-1].lower() in ('.png', '.jpg', '.jpeg', '.gif',\n '.tiff')\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "def FUNC_39(self):...\n", "from zope.interface.verify import verifyObject\n", "from zope.tal.interfaces import ITALExpressionEngine\n", "verifyObject(ITALExpressionEngine, self._makeOne())\n" ]
[ "def test_instance_conforms_to_ITALExpressionEngine(self):...\n", "from zope.interface.verify import verifyObject\n", "from zope.tal.interfaces import ITALExpressionEngine\n", "verifyObject(ITALExpressionEngine, self._makeOne())\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "ImportFrom'", "ImportFrom'", "Expr'" ]
[ "@FUNC_0...\n", "return RoomListHandler(self)\n" ]
[ "@cache_in_self...\n", "return RoomListHandler(self)\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "def __init__(self, VAR_25, VAR_26, VAR_11):...\n", "super().__init__()\n", "self.offendingText = VAR_25\n", "self.allTexts = VAR_26\n", "self.el = VAR_11\n" ]
[ "def __init__(self, offendingText, allTexts, el):...\n", "super().__init__()\n", "self.offendingText = offendingText\n", "self.allTexts = allTexts\n", "self.el = el\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Assign'", "Assign'", "Assign'" ]
[ "def FUNC_33(self):...\n", "\"\"\"docstring\"\"\"\n", "return json.dumps(cherry.config['media.playable'])\n" ]
[ "def api_getplayables(self):...\n", "\"\"\"docstring\"\"\"\n", "return json.dumps(cherry.config['media.playable'])\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Return'" ]
[ "def FUNC_43(self):...\n", "\"\"\"docstring\"\"\"\n" ]
[ "def init_components(self):...\n", "\"\"\"docstring\"\"\"\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Docstring" ]
[ "def FUNC_0():...\n", "return tornado.web.Application([('/autofuzzy(.*)', CLASS_2), ('/auto(.*)',\n CLASS_0), ('/fuzzy(.*)', CLASS_1), ('/define(.*)', CLASS_4), (\n '/nodeslike(.*)', CLASS_3), ('/(.*)', tornado.web.StaticFileHandler, {\n 'path': VAR_0, 'default_filename': 'rtxcomplete.html'})],\n compress_response=True)\n" ]
[ "def make_https_app():...\n", "return tornado.web.Application([('/autofuzzy(.*)', autofuzzySearch), (\n '/auto(.*)', autoSearch), ('/fuzzy(.*)', fuzzySearch), ('/define(.*)',\n defineSearch), ('/nodeslike(.*)', nodesLikeSearch), ('/(.*)', tornado.\n web.StaticFileHandler, {'path': root, 'default_filename':\n 'rtxcomplete.html'})], compress_response=True)\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_3(VAR_3, VAR_4, VAR_5):...\n", "VAR_9 = []\n", "for inp_element in VAR_5:\n", "VAR_83 = False\n", "return VAR_9\n", "for c_elements in VAR_3:\n", "if VAR_4 == 'languages':\n", "if not VAR_83:\n", "VAR_100 = c_elements.lang_code\n", "if VAR_4 == 'custom':\n", "VAR_9.append(inp_element)\n", "if inp_element == VAR_100:\n", "VAR_100 = c_elements.value\n", "VAR_100 = c_elements.name\n", "VAR_83 = True\n" ]
[ "def search_objects_add(db_book_object, db_type, input_elements):...\n", "add_elements = []\n", "for inp_element in input_elements:\n", "found = False\n", "return add_elements\n", "for c_elements in db_book_object:\n", "if db_type == 'languages':\n", "if not found:\n", "type_elements = c_elements.lang_code\n", "if db_type == 'custom':\n", "add_elements.append(inp_element)\n", "if inp_element == type_elements:\n", "type_elements = c_elements.value\n", "type_elements = c_elements.name\n", "found = True\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "For", "Assign'", "Return'", "For", "Condition", "Condition", "Assign'", "Condition", "Expr'", "Condition", "Assign'", "Assign'", "Assign'" ]
[ "def FUNC_18(self):...\n", "self.assertFormfield(models.Member, 'gender', forms.Select)\n" ]
[ "def testFieldWithChoices(self):...\n", "self.assertFormfield(models.Member, 'gender', forms.Select)\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Expr'" ]
[ "@require_POST...\n", "VAR_3 = get_object_or_404(Topic, VAR_2=topic_id)\n", "VAR_4 = FavoriteForm(user=request.user, VAR_3=topic, data=request.POST)\n", "if VAR_4.is_valid():\n", "VAR_4.save()\n", "messages.error(VAR_0, utils.render_form_errors(VAR_4))\n", "return redirect(VAR_0.POST.get('next', VAR_3.get_absolute_url()))\n" ]
[ "@require_POST...\n", "topic = get_object_or_404(Topic, pk=topic_id)\n", "form = FavoriteForm(user=request.user, topic=topic, data=request.POST)\n", "if form.is_valid():\n", "form.save()\n", "messages.error(request, utils.render_form_errors(form))\n", "return redirect(request.POST.get('next', topic.get_absolute_url()))\n" ]
[ 0, 0, 0, 0, 0, 0, 4 ]
[ "Condition", "Assign'", "Assign'", "Condition", "Expr'", "Expr'", "Return'" ]
[ "def FUNC_3(self, VAR_3):...\n", "os.umask(0)\n", "VAR_153 = 'helpdesk/attachments/kb/{category}/{kbi}'.format(VAR_99=self.\n kbitem.category, kbi=self.kbitem.id)\n", "VAR_154 = os.path.join(settings.MEDIA_ROOT, VAR_153)\n", "if settings.DEFAULT_FILE_STORAGE == 'django.core.files.storage.FileSystemStorage':\n", "if not os.path.exists(VAR_154):\n", "return os.path.join(VAR_153, VAR_3)\n", "os.makedirs(VAR_154, 511)\n" ]
[ "def attachment_path(self, filename):...\n", "os.umask(0)\n", "path = 'helpdesk/attachments/kb/{category}/{kbi}'.format(category=self.\n kbitem.category, kbi=self.kbitem.id)\n", "att_path = os.path.join(settings.MEDIA_ROOT, path)\n", "if settings.DEFAULT_FILE_STORAGE == 'django.core.files.storage.FileSystemStorage':\n", "if not os.path.exists(att_path):\n", "return os.path.join(path, filename)\n", "os.makedirs(att_path, 511)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Assign'", "Assign'", "Condition", "Condition", "Return'", "Expr'" ]
[ "def FUNC_9(self):...\n", "VAR_19 = meta_graph_pb2.TensorInfo()\n", "VAR_19.dtype = types_pb2.DT_FLOAT_REF\n", "saved_model_cli._print_tensor_info(VAR_19)\n", "self.assertTrue('DT_FLOAT_REF' in out.getvalue().strip())\n", "self.assertEqual(err.getvalue().strip(), '')\n" ]
[ "def testPrintREFTypeTensor(self):...\n", "ref_tensor_info = meta_graph_pb2.TensorInfo()\n", "ref_tensor_info.dtype = types_pb2.DT_FLOAT_REF\n", "saved_model_cli._print_tensor_info(ref_tensor_info)\n", "self.assertTrue('DT_FLOAT_REF' in out.getvalue().strip())\n", "self.assertEqual(err.getvalue().strip(), '')\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'" ]
[ "def FUNC_25(self, VAR_23):...\n", "VAR_53.session['playlist'] = VAR_23\n" ]
[ "def api_rememberplaylist(self, playlist):...\n", "cherrypy.session['playlist'] = playlist\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Assign'" ]
[ "def FUNC_80(self, VAR_2):...\n", "\"\"\"docstring\"\"\"\n", "return self.handle_not_logged_in(VAR_2)\n" ]
[ "def get(self, request):...\n", "\"\"\"docstring\"\"\"\n", "return self.handle_not_logged_in(request)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Return'" ]
[ "def FUNC_1(VAR_3: Inventory, VAR_4: (datetime.date | None)=None) ->Any:...\n", "\"\"\"docstring\"\"\"\n", "return cost_or_value_without_context(VAR_3, g.conversion, g.ledger.\n price_map, VAR_4)\n" ]
[ "def cost_or_value(inventory: Inventory, date: (datetime.date | None)=None...\n", "\"\"\"docstring\"\"\"\n", "return cost_or_value_without_context(inventory, g.conversion, g.ledger.\n price_map, date)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Return'" ]
[ "def FUNC_38(VAR_47):...\n", "return VAR_47\n" ]
[ "def fake_wrapper_session(sess):...\n", "return sess\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_20(VAR_25, VAR_17, VAR_26=False, VAR_27=None):...\n", "VAR_59 = VAR_25.split(',')\n", "VAR_60 = []\n", "if not VAR_26:\n", "VAR_61 = isoLanguages.get_language_codes(get_locale(), VAR_59, VAR_60)\n", "VAR_61 = isoLanguages.get_valid_language_codes(get_locale(), VAR_59, VAR_60)\n", "for l in VAR_60:\n", "VAR_2.error('%s is not a valid language', l)\n", "if VAR_26 and len(VAR_61) == 1:\n", "if isinstance(VAR_27, list):\n", "if VAR_61[0] != current_user.filter_language(\n", "VAR_61 = helper.uniq(VAR_61)\n", "VAR_27.append(l)\n", "flash(_(u'%(langname)s is not a valid language', langname=l), category=\n 'warning')\n", "VAR_61[0] = calibre_db.session.query(db.Languages).filter(db.Languages.\n lang_code == current_user.filter_language()).first().lang_code\n", "return FUNC_7(VAR_61, VAR_17.languages, db.Languages, calibre_db.session,\n 'languages')\n" ]
[ "def edit_book_languages(languages, book, upload=False, invalid=None):...\n", "input_languages = languages.split(',')\n", "unknown_languages = []\n", "if not upload:\n", "input_l = isoLanguages.get_language_codes(get_locale(), input_languages,\n unknown_languages)\n", "input_l = isoLanguages.get_valid_language_codes(get_locale(),\n input_languages, unknown_languages)\n", "for l in unknown_languages:\n", "log.error('%s is not a valid language', l)\n", "if upload and len(input_l) == 1:\n", "if isinstance(invalid, list):\n", "if input_l[0] != current_user.filter_language(\n", "input_l = helper.uniq(input_l)\n", "invalid.append(l)\n", "flash(_(u'%(langname)s is not a valid language', langname=l), category=\n 'warning')\n", "input_l[0] = calibre_db.session.query(db.Languages).filter(db.Languages.\n lang_code == current_user.filter_language()).first().lang_code\n", "return modify_database_object(input_l, book.languages, db.Languages,\n calibre_db.session, 'languages')\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "For", "Expr'", "Condition", "Condition", "Condition", "Assign'", "Expr'", "Expr'", "Assign'", "Return'" ]
[ "async def FUNC_6(self):...\n", "if not hasattr(self, 'record_id'):\n", "self.record_id = await self.get_resolved_ref()\n", "VAR_36 = self.spec.split('zenodo')[0] + 'zenodo.' + self.record_id\n", "return VAR_36\n" ]
[ "async def get_resolved_spec(self):...\n", "if not hasattr(self, 'record_id'):\n", "self.record_id = await self.get_resolved_ref()\n", "resolved_spec = self.spec.split('zenodo')[0] + 'zenodo.' + self.record_id\n", "return resolved_spec\n" ]
[ 0, 0, 0, 0, 0 ]
[ "AsyncFunctionDef'", "Condition", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_26(self):...\n", "VAR_9 = self._makeEngine()\n", "VAR_5 = self._makeContext()\n", "VAR_10 = VAR_5.evaluate(VAR_9.compile('string:x'))\n", "self.assertEqual(VAR_10, 'x')\n", "self.assertIsInstance(VAR_10, str)\n" ]
[ "def test_unicode(self):...\n", "eng = self._makeEngine()\n", "ec = self._makeContext()\n", "result = ec.evaluate(eng.compile('string:x'))\n", "self.assertEqual(result, 'x')\n", "self.assertIsInstance(result, str)\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'" ]
[ "def FUNC_0(VAR_2):...\n", "return Static(template('obj', VAR_2=Symbol(obj), mode='eval'))\n" ]
[ "def static(obj):...\n", "return Static(template('obj', obj=Symbol(obj), mode='eval'))\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_1(VAR_6):...\n", "if isinstance(VAR_6, (UnicodeEncodeError, UnicodeTranslateError)):\n", "VAR_28 = []\n", "return xmlcharrefreplace_errors(VAR_6)\n", "VAR_29 = []\n", "VAR_30 = False\n", "for i, c in enumerate(VAR_6.object[VAR_6.start:VAR_6.end]):\n", "if VAR_30:\n", "for cp in VAR_29:\n", "VAR_30 = False\n", "VAR_33 = i + VAR_6.start\n", "VAR_34 = VAR_4.get(cp)\n", "return ''.join(VAR_28), VAR_6.end\n", "if utils.isSurrogatePair(VAR_6.object[VAR_33:min([VAR_6.end, VAR_33 + 2])]):\n", "if VAR_34:\n", "VAR_36 = utils.surrogatePairToCodepoint(VAR_6.object[VAR_33:VAR_33 + 2])\n", "VAR_36 = ord(c)\n", "VAR_28.append('&')\n", "VAR_28.append('&#x%s;' % hex(cp)[2:])\n", "VAR_30 = True\n", "VAR_29.append(VAR_36)\n", "VAR_28.append(VAR_34)\n", "if not VAR_34.endswith(';'):\n", "VAR_28.append(';')\n" ]
[ "def htmlentityreplace_errors(exc):...\n", "if isinstance(exc, (UnicodeEncodeError, UnicodeTranslateError)):\n", "res = []\n", "return xmlcharrefreplace_errors(exc)\n", "codepoints = []\n", "skip = False\n", "for i, c in enumerate(exc.object[exc.start:exc.end]):\n", "if skip:\n", "for cp in codepoints:\n", "skip = False\n", "index = i + exc.start\n", "e = encode_entity_map.get(cp)\n", "return ''.join(res), exc.end\n", "if utils.isSurrogatePair(exc.object[index:min([exc.end, index + 2])]):\n", "if e:\n", "codepoint = utils.surrogatePairToCodepoint(exc.object[index:index + 2])\n", "codepoint = ord(c)\n", "res.append('&')\n", "res.append('&#x%s;' % hex(cp)[2:])\n", "skip = True\n", "codepoints.append(codepoint)\n", "res.append(e)\n", "if not e.endswith(';'):\n", "res.append(';')\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Assign'", "Return'", "Assign'", "Assign'", "For", "Condition", "For", "Assign'", "Assign'", "Assign'", "Return'", "Condition", "Condition", "Assign'", "Assign'", "Expr'", "Expr'", "Assign'", "Expr'", "Expr'", "Condition", "Expr'" ]
[ "@CLASS_4('bindings')...\n", "\"\"\"docstring\"\"\"\n", "VAR_26 = {}\n", "VAR_27 = config.val.bindings.default\n", "VAR_28 = set(VAR_27.keys()).union(config.val.bindings.commands)\n", "VAR_28.remove('normal')\n", "VAR_28 = ['normal'] + sorted(list(VAR_28))\n", "for VAR_41 in VAR_28:\n", "VAR_26[VAR_41] = config.key_instance.get_bindings_for(VAR_41)\n", "VAR_20 = jinja.render('bindings.html', title='Bindings', VAR_26=bindings)\n", "return 'text/html', VAR_20\n" ]
[ "@add_handler('bindings')...\n", "\"\"\"docstring\"\"\"\n", "bindings = {}\n", "defaults = config.val.bindings.default\n", "modes = set(defaults.keys()).union(config.val.bindings.commands)\n", "modes.remove('normal')\n", "modes = ['normal'] + sorted(list(modes))\n", "for mode in modes:\n", "bindings[mode] = config.key_instance.get_bindings_for(mode)\n", "src = jinja.render('bindings.html', title='Bindings', bindings=bindings)\n", "return 'text/html', src\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "For", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_20(VAR_66: DecoratedCallable) ->DecoratedCallable:...\n", "self.add_api_route(VAR_26, VAR_66, VAR_29=response_model, VAR_13=\n status_code, VAR_30=tags, VAR_31=dependencies, VAR_32=summary, VAR_33=\n description, VAR_34=response_description, VAR_35=responses, VAR_36=\n deprecated, VAR_37=methods, VAR_38=operation_id, VAR_16=\n response_model_include, VAR_17=response_model_exclude, VAR_18=\n response_model_by_alias, VAR_19=response_model_exclude_unset, VAR_20=\n response_model_exclude_defaults, VAR_21=response_model_exclude_none,\n VAR_39=include_in_schema, VAR_14=response_class, VAR_28=name, VAR_40=\n callbacks)\n", "return VAR_66\n" ]
[ "def decorator(func: DecoratedCallable) ->DecoratedCallable:...\n", "self.add_api_route(path, func, response_model=response_model, status_code=\n status_code, tags=tags, dependencies=dependencies, summary=summary,\n description=description, response_description=response_description,\n responses=responses, deprecated=deprecated, methods=methods,\n operation_id=operation_id, response_model_include=\n response_model_include, response_model_exclude=response_model_exclude,\n response_model_by_alias=response_model_by_alias,\n response_model_exclude_unset=response_model_exclude_unset,\n response_model_exclude_defaults=response_model_exclude_defaults,\n response_model_exclude_none=response_model_exclude_none,\n include_in_schema=include_in_schema, response_class=response_class,\n name=name, callbacks=callbacks)\n", "return func\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Return'" ]
[ "@VAR_1.route('/<bfile>/statement/', methods=['GET'])...\n", "\"\"\"docstring\"\"\"\n", "VAR_27 = request.args.get('entry_hash', '')\n", "VAR_28 = request.args.get('key', '')\n", "VAR_29 = g.ledger.statement_path(VAR_27, VAR_28)\n", "return send_file_inline(VAR_29)\n" ]
[ "@app.route('/<bfile>/statement/', methods=['GET'])...\n", "\"\"\"docstring\"\"\"\n", "entry_hash = request.args.get('entry_hash', '')\n", "key = request.args.get('key', '')\n", "document_path = g.ledger.statement_path(entry_hash, key)\n", "return send_file_inline(document_path)\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_61(self):...\n", "\"\"\"docstring\"\"\"\n", "return self._blitzcon.getUser().omeName\n" ]
[ "def getName(self):...\n", "\"\"\"docstring\"\"\"\n", "return self._blitzcon.getUser().omeName\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Return'" ]
[ "def FUNC_125(self, VAR_10):...\n", "\"\"\"docstring\"\"\"\n", "return VAR_418.path.islink(VAR_10) and not self.in_base(VAR_10)\n" ]
[ "def issymlink_out(self, f):...\n", "\"\"\"docstring\"\"\"\n", "return os.path.islink(f) and not self.in_base(f)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Return'" ]
[ "def FUNC_12(self):...\n", "" ]
[ "def finalize_options(self):...\n", "" ]
[ 0, 0 ]
[ "FunctionDef'", "Condition" ]
[ "\"\"\"\nDecorators for use with the webclient application.\n\"\"\"\n", "import logging\n", "import omeroweb.decorators\n", "from omero import constants\n", "from django.http import HttpResponse\n", "from django.conf import settings\n", "from django.core.urlresolvers import reverse\n", "from django.core.urlresolvers import NoReverseMatch\n", "from omeroweb.webclient.forms import GlobalSearchForm\n", "from omeroweb.utils import reverse_with_params\n", "from omeroweb.webgateway.marshal import eventContextMarshal\n", "VAR_0 = logging.getLogger(__name__)\n", "\"\"\"\n webclient specific extension of the OMERO.web login_required() decorator.\n \"\"\"\n", "def __init__(self, VAR_1=False, VAR_2=False, VAR_3=None, **VAR_4):...\n", "\"\"\"docstring\"\"\"\n", "super(CLASS_0, self).__init__(**kwargs)\n", "self.ignore_login_fail = VAR_1\n", "self.setGroupContext = VAR_2\n", "self.login_redirect = VAR_3\n", "def FUNC_0(self, VAR_5, VAR_6):...\n", "\"\"\"docstring\"\"\"\n", "super(CLASS_0, self).on_logged_in(VAR_5, VAR_6)\n", "self.prepare_session(VAR_5)\n", "if self.setGroupContext:\n", "if VAR_5.session.get('active_group'):\n", "def FUNC_1(self, VAR_5, VAR_7, VAR_8=None):...\n", "VAR_6.SERVICE_OPTS.setOmeroGroup(VAR_5.session.get('active_group'))\n", "VAR_6.SERVICE_OPTS.setOmeroGroup(VAR_6.getEventContext().groupId)\n", "\"\"\"docstring\"\"\"\n", "if self.ignore_login_fail:\n", "return HttpResponse('Connection Failed')\n", "if self.login_redirect is not None:\n", "return super(CLASS_0, self).on_not_logged_in(VAR_5, VAR_7, VAR_8)\n", "VAR_7 = reverse(self.login_redirect)\n" ]
[ "\"\"\"\nDecorators for use with the webclient application.\n\"\"\"\n", "import logging\n", "import omeroweb.decorators\n", "from omero import constants\n", "from django.http import HttpResponse\n", "from django.conf import settings\n", "from django.core.urlresolvers import reverse\n", "from django.core.urlresolvers import NoReverseMatch\n", "from omeroweb.webclient.forms import GlobalSearchForm\n", "from omeroweb.utils import reverse_with_params\n", "from omeroweb.webgateway.marshal import eventContextMarshal\n", "logger = logging.getLogger(__name__)\n", "\"\"\"\n webclient specific extension of the OMERO.web login_required() decorator.\n \"\"\"\n", "def __init__(self, ignore_login_fail=False, setGroupContext=False,...\n", "\"\"\"docstring\"\"\"\n", "super(login_required, self).__init__(**kwargs)\n", "self.ignore_login_fail = ignore_login_fail\n", "self.setGroupContext = setGroupContext\n", "self.login_redirect = login_redirect\n", "def on_logged_in(self, request, conn):...\n", "\"\"\"docstring\"\"\"\n", "super(login_required, self).on_logged_in(request, conn)\n", "self.prepare_session(request)\n", "if self.setGroupContext:\n", "if request.session.get('active_group'):\n", "def on_not_logged_in(self, request, url, error=None):...\n", "conn.SERVICE_OPTS.setOmeroGroup(request.session.get('active_group'))\n", "conn.SERVICE_OPTS.setOmeroGroup(conn.getEventContext().groupId)\n", "\"\"\"docstring\"\"\"\n", "if self.ignore_login_fail:\n", "return HttpResponse('Connection Failed')\n", "if self.login_redirect is not None:\n", "return super(login_required, self).on_not_logged_in(request, url, error)\n", "url = reverse(self.login_redirect)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Expr'", "Import'", "Import'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Assign'", "Expr'", "FunctionDef'", "Docstring", "Expr'", "Assign'", "Assign'", "Assign'", "FunctionDef'", "Docstring", "Expr'", "Expr'", "Condition", "Condition", "FunctionDef'", "Expr'", "Expr'", "Docstring", "Condition", "Return'", "Condition", "Return'", "Assign'" ]
[ "def FUNC_36(VAR_51):...\n", "for VAR_105 in VAR_51:\n", "VAR_16.write(VAR_105)\n", "yield VAR_105\n" ]
[ "def written_chunks(chunks):...\n", "for chunk in chunks:\n", "content_file.write(chunk)\n", "yield chunk\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "For", "Expr'", "Expr'" ]
[ "def FUNC_7(VAR_0):...\n", "import subprocess\n", "from openapi_python_client import GeneratorData, Project\n", "VAR_12 = VAR_0.patch('subprocess.run')\n", "VAR_3 = VAR_0.MagicMock(autospec=GeneratorData, title='My Test API')\n", "VAR_8 = Project(VAR_3=openapi)\n", "VAR_8.project_dir = VAR_0.MagicMock(autospec=pathlib.Path)\n", "VAR_8._reformat()\n", "VAR_12.assert_has_calls([VAR_0.call('isort .', cwd=project.project_dir,\n shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE), VAR_0.call\n ('black .', cwd=project.project_dir, shell=True, stdout=subprocess.PIPE,\n stderr=subprocess.PIPE)])\n" ]
[ "def test__reformat(mocker):...\n", "import subprocess\n", "from openapi_python_client import GeneratorData, Project\n", "sub_run = mocker.patch('subprocess.run')\n", "openapi = mocker.MagicMock(autospec=GeneratorData, title='My Test API')\n", "project = Project(openapi=openapi)\n", "project.project_dir = mocker.MagicMock(autospec=pathlib.Path)\n", "project._reformat()\n", "sub_run.assert_has_calls([mocker.call('isort .', cwd=project.project_dir,\n shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE), mocker.\n call('black .', cwd=project.project_dir, shell=True, stdout=subprocess.\n PIPE, stderr=subprocess.PIPE)])\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Import'", "ImportFrom'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'" ]
[ "@VAR_2.route('/ratings')...\n", "if VAR_87.check_visibility(constants.SIDEBAR_RATING):\n", "if VAR_87.get_view_property('ratings', 'dir') == 'desc':\n", "abort(404)\n", "VAR_10 = db.Ratings.rating.desc()\n", "VAR_10 = db.Ratings.rating.asc()\n", "VAR_109 = 0\n", "VAR_109 = 1\n", "VAR_63 = calibre_db.session.query(db.Ratings, func.count(\n 'books_ratings_link.book').label('count'), (db.Ratings.rating / 2).\n label('name')).join(db.books_ratings_link).join(db.Books).filter(calibre_db\n .common_filters()).group_by(text('books_ratings_link.rating')).order_by(\n VAR_10).all()\n", "return render_title_template('list.html', VAR_63=entries, folder=\n 'web.books_list', VAR_107=list(), VAR_149=_(u'Ratings list'), VAR_9=\n 'ratingslist', VAR_8='ratings', VAR_10=order_no)\n" ]
[ "@web.route('/ratings')...\n", "if current_user.check_visibility(constants.SIDEBAR_RATING):\n", "if current_user.get_view_property('ratings', 'dir') == 'desc':\n", "abort(404)\n", "order = db.Ratings.rating.desc()\n", "order = db.Ratings.rating.asc()\n", "order_no = 0\n", "order_no = 1\n", "entries = calibre_db.session.query(db.Ratings, func.count(\n 'books_ratings_link.book').label('count'), (db.Ratings.rating / 2).\n label('name')).join(db.books_ratings_link).join(db.Books).filter(calibre_db\n .common_filters()).group_by(text('books_ratings_link.rating')).order_by(\n order).all()\n", "return render_title_template('list.html', entries=entries, folder=\n 'web.books_list', charlist=list(), title=_(u'Ratings list'), page=\n 'ratingslist', data='ratings', order=order_no)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Condition", "Condition", "Expr'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "@parameterized.named_parameters(('non_tfrt', False))...\n", "self.parser = saved_model_cli.create_parser()\n", "VAR_10 = test.test_src_dir_path(VAR_0)\n", "VAR_45 = np.array([[1], [2]])\n", "VAR_46 = np.zeros((6, 3))\n", "VAR_32 = os.path.join(test.get_temp_dir(), 'testRunCommand_inputs.npz')\n", "np.savez(VAR_32, VAR_27=x, VAR_28=x_notused)\n", "VAR_47 = os.path.join(test.get_temp_dir(), 'outputs.npy')\n", "if os.path.exists(VAR_47):\n", "os.remove(VAR_47)\n", "VAR_11 = self.parser.parse_args(['run', '--dir', VAR_10, '--tag_set',\n 'serve', '--signature_def', 'regress_x2_to_y3', '--inputs', 'inputs=' +\n VAR_32 + '[x0]', '--outdir', test.get_temp_dir()] + (['--use_tfrt'] if\n VAR_5 else []))\n", "saved_model_cli.run(VAR_11)\n", "VAR_43 = np.load(VAR_47)\n", "VAR_44 = np.array([[3.5], [4.0]])\n", "self.assertAllClose(VAR_44, VAR_43)\n" ]
[ "@parameterized.named_parameters(('non_tfrt', False))...\n", "self.parser = saved_model_cli.create_parser()\n", "base_path = test.test_src_dir_path(SAVED_MODEL_PATH)\n", "x = np.array([[1], [2]])\n", "x_notused = np.zeros((6, 3))\n", "input_path = os.path.join(test.get_temp_dir(), 'testRunCommand_inputs.npz')\n", "np.savez(input_path, x0=x, x1=x_notused)\n", "output_file = os.path.join(test.get_temp_dir(), 'outputs.npy')\n", "if os.path.exists(output_file):\n", "os.remove(output_file)\n", "args = self.parser.parse_args(['run', '--dir', base_path, '--tag_set',\n 'serve', '--signature_def', 'regress_x2_to_y3', '--inputs', 'inputs=' +\n input_path + '[x0]', '--outdir', test.get_temp_dir()] + (['--use_tfrt'] if\n use_tfrt else []))\n", "saved_model_cli.run(args)\n", "y_actual = np.load(output_file)\n", "y_expected = np.array([[3.5], [4.0]])\n", "self.assertAllClose(y_expected, y_actual)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "Condition", "Expr'", "Assign'", "Expr'", "Assign'", "Assign'", "Expr'" ]
[ "async def FUNC_9(self, VAR_14):...\n", "if 'notify' not in VAR_14['actions']:\n", "return True\n", "VAR_16 = push_rule_evaluator.tweaks_for_actions(VAR_14['actions'])\n", "VAR_17 = await push_tools.get_badge_count(self.hs.get_datastore(), self.\n user_id, group_by_room=self._group_unread_count_by_room)\n", "VAR_15 = await self.store.get_event(VAR_14['event_id'], allow_none=True)\n", "if VAR_15 is None:\n", "return True\n", "VAR_21 = await self.dispatch_push(VAR_15, VAR_16, VAR_17)\n", "if VAR_21 is False:\n", "return False\n", "if isinstance(VAR_21, list) or isinstance(VAR_21, tuple):\n", "for pk in VAR_21:\n", "return True\n", "if pk != self.pushkey:\n", "VAR_0.warning(\"Ignoring rejected pushkey %s because we didn't send it\", pk)\n", "VAR_0.info('Pushkey %s was rejected: removing', pk)\n", "await self.hs.remove_pusher(self.app_id, pk, self.user_id)\n" ]
[ "async def _process_one(self, push_action):...\n", "if 'notify' not in push_action['actions']:\n", "return True\n", "tweaks = push_rule_evaluator.tweaks_for_actions(push_action['actions'])\n", "badge = await push_tools.get_badge_count(self.hs.get_datastore(), self.\n user_id, group_by_room=self._group_unread_count_by_room)\n", "event = await self.store.get_event(push_action['event_id'], allow_none=True)\n", "if event is None:\n", "return True\n", "rejected = await self.dispatch_push(event, tweaks, badge)\n", "if rejected is False:\n", "return False\n", "if isinstance(rejected, list) or isinstance(rejected, tuple):\n", "for pk in rejected:\n", "return True\n", "if pk != self.pushkey:\n", "logger.warning(\"Ignoring rejected pushkey %s because we didn't send it\", pk)\n", "logger.info('Pushkey %s was rejected: removing', pk)\n", "await self.hs.remove_pusher(self.app_id, pk, self.user_id)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "AsyncFunctionDef'", "Condition", "Return'", "Assign'", "Assign'", "Assign'", "Condition", "Return'", "Assign'", "Condition", "Return'", "Condition", "For", "Return'", "Condition", "Expr'", "Expr'", "Expr'" ]
[ "def FUNC_37(self):...\n", "if not sys.platform.startswith('win') and sys.stdin.isatty():\n", "signal.signal(signal.SIGINT, self._handle_sigint)\n", "signal.signal(signal.SIGTERM, self._signal_stop)\n", "if hasattr(signal, 'SIGUSR1'):\n", "signal.signal(signal.SIGUSR1, self._signal_info)\n", "if hasattr(signal, 'SIGINFO'):\n", "signal.signal(signal.SIGINFO, self._signal_info)\n" ]
[ "def init_signal(self):...\n", "if not sys.platform.startswith('win') and sys.stdin.isatty():\n", "signal.signal(signal.SIGINT, self._handle_sigint)\n", "signal.signal(signal.SIGTERM, self._signal_stop)\n", "if hasattr(signal, 'SIGUSR1'):\n", "signal.signal(signal.SIGUSR1, self._signal_info)\n", "if hasattr(signal, 'SIGINFO'):\n", "signal.signal(signal.SIGINFO, self._signal_info)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "For", "Expr'", "Expr'", "Condition", "Expr'", "Condition", "Expr'" ]
[ "@contextlib.contextmanager...\n", "VAR_1, VAR_2 = StringIO(), StringIO()\n", "VAR_3, VAR_4 = VAR_8.stdout, VAR_8.stderr\n", "VAR_8.stdout, VAR_8.stderr = VAR_1, VAR_2\n", "VAR_8.stdout, VAR_8.stderr = VAR_3, VAR_4\n", "yield VAR_8.stdout, VAR_8.stderr\n" ]
[ "@contextlib.contextmanager...\n", "new_out, new_err = StringIO(), StringIO()\n", "old_out, old_err = sys.stdout, sys.stderr\n", "sys.stdout, sys.stderr = new_out, new_err\n", "sys.stdout, sys.stderr = old_out, old_err\n", "yield sys.stdout, sys.stderr\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'" ]
[ "def FUNC_18(self):...\n", "self.reactor.advance(1000000000000)\n", "VAR_5 = self.helper.create_room_as(self.user_id)\n", "self.get_success(self.presence_handler.set_state(UserID.from_string(\n '@test:server'), {'presence': PresenceState.ONLINE}))\n", "self.get_success(self.presence_handler.set_state(UserID.from_string(\n '@test2:server'), {'presence': PresenceState.ONLINE}))\n", "self._add_new_user(VAR_5, '@alice:server2')\n", "self._add_new_user(VAR_5, '@bob:server3')\n", "self.reactor.pump([0])\n", "self.federation_sender.reset_mock()\n", "self.helper.join(VAR_5, '@test2:server')\n", "self.reactor.pump([0])\n", "self.federation_sender.send_presence.assert_not_called()\n", "VAR_14 = self.get_success(self.presence_handler.current_state_for_user(\n '@test2:server'))\n", "self.assertEqual(VAR_14.state, PresenceState.ONLINE)\n", "self.federation_sender.send_presence_to_destinations.assert_called_once_with(\n destinations={'server2', 'server3'}, states=[expected_state])\n" ]
[ "def test_remote_gets_presence_when_local_user_joins(self):...\n", "self.reactor.advance(1000000000000)\n", "room_id = self.helper.create_room_as(self.user_id)\n", "self.get_success(self.presence_handler.set_state(UserID.from_string(\n '@test:server'), {'presence': PresenceState.ONLINE}))\n", "self.get_success(self.presence_handler.set_state(UserID.from_string(\n '@test2:server'), {'presence': PresenceState.ONLINE}))\n", "self._add_new_user(room_id, '@alice:server2')\n", "self._add_new_user(room_id, '@bob:server3')\n", "self.reactor.pump([0])\n", "self.federation_sender.reset_mock()\n", "self.helper.join(room_id, '@test2:server')\n", "self.reactor.pump([0])\n", "self.federation_sender.send_presence.assert_not_called()\n", "expected_state = self.get_success(self.presence_handler.\n current_state_for_user('@test2:server'))\n", "self.assertEqual(expected_state.state, PresenceState.ONLINE)\n", "self.federation_sender.send_presence_to_destinations.assert_called_once_with(\n destinations={'server2', 'server3'}, states=[expected_state])\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Assign'", "Expr'", "Expr'" ]
[ "def FUNC_55(self):...\n", "\"\"\"docstring\"\"\"\n", "VAR_54 = json.dumps({'search_categories': {'room_events': {'search_term':\n 'label', 'filter': self.FILTER_NOT_LABELS}}})\n", "self._send_labelled_messages_in_room()\n", "VAR_22, VAR_23 = self.make_request('POST', '/search?access_token=%s' % self\n .tok, VAR_54)\n", "VAR_49 = VAR_23.json_body['search_categories']['room_events']['results']\n", "self.assertEqual(len(VAR_49), 4, [result['result']['content'] for result in\n VAR_49])\n", "self.assertEqual(VAR_49[0]['result']['content']['body'], 'without label',\n VAR_49[0]['result']['content']['body'])\n", "self.assertEqual(VAR_49[1]['result']['content']['body'], 'without label',\n VAR_49[1]['result']['content']['body'])\n", "self.assertEqual(VAR_49[2]['result']['content']['body'], 'with wrong label',\n VAR_49[2]['result']['content']['body'])\n", "self.assertEqual(VAR_49[3]['result']['content']['body'],\n 'with two wrong labels', VAR_49[3]['result']['content']['body'])\n" ]
[ "def test_search_filter_not_labels(self):...\n", "\"\"\"docstring\"\"\"\n", "request_data = json.dumps({'search_categories': {'room_events': {\n 'search_term': 'label', 'filter': self.FILTER_NOT_LABELS}}})\n", "self._send_labelled_messages_in_room()\n", "request, channel = self.make_request('POST', '/search?access_token=%s' %\n self.tok, request_data)\n", "results = channel.json_body['search_categories']['room_events']['results']\n", "self.assertEqual(len(results), 4, [result['result']['content'] for result in\n results])\n", "self.assertEqual(results[0]['result']['content']['body'], 'without label',\n results[0]['result']['content']['body'])\n", "self.assertEqual(results[1]['result']['content']['body'], 'without label',\n results[1]['result']['content']['body'])\n", "self.assertEqual(results[2]['result']['content']['body'],\n 'with wrong label', results[2]['result']['content']['body'])\n", "self.assertEqual(results[3]['result']['content']['body'],\n 'with two wrong labels', results[3]['result']['content']['body'])\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Expr'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'" ]
[ "def FUNC_2(self) ->str:...\n", "\"\"\"docstring\"\"\"\n", "if self.required:\n", "return self.reference.class_name\n", "return f'Optional[{self.reference.class_name}]'\n" ]
[ "def get_type_string(self) ->str:...\n", "\"\"\"docstring\"\"\"\n", "if self.required:\n", "return self.reference.class_name\n", "return f'Optional[{self.reference.class_name}]'\n" ]
[ 0, 0, 5, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Return'", "Return'" ]
[ "@ratelimit(field='email', rate='5/5m')...\n", "if VAR_7.method == 'POST' and VAR_7.is_limited():\n", "return redirect(reverse('spirit:user:auth:password-reset'))\n", "return VAR_3(VAR_7, **kwargs)\n" ]
[ "@ratelimit(field='email', rate='5/5m')...\n", "if request.method == 'POST' and request.is_limited():\n", "return redirect(reverse('spirit:user:auth:password-reset'))\n", "return _password_reset_view(request, **kwargs)\n" ]
[ 0, 0, 0, 0 ]
[ "Condition", "Condition", "Return'", "Return'" ]
[ "def FUNC_10(VAR_18=16):...\n", "\"\"\"docstring\"\"\"\n", "random.seed()\n", "assert isinstance(VAR_18, int)\n", "VAR_27 = '%0' + str(2 * VAR_18) + 'x'\n", "return VAR_27 % random.getrandbits(VAR_18 * 8)\n" ]
[ "def generate_random_key(length=16):...\n", "\"\"\"docstring\"\"\"\n", "random.seed()\n", "assert isinstance(length, int)\n", "format_string = '%0' + str(2 * length) + 'x'\n", "return format_string % random.getrandbits(length * 8)\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Expr'", "Assert'", "Assign'", "Return'" ]
[ "def FUNC_1(self, VAR_3):...\n", "\"\"\"docstring\"\"\"\n", "self.mongo_query = {}\n", "self.ops = []\n", "self.current_value = None\n", "self.generic_visit(VAR_3)\n", "if self.mongo_query == {}:\n" ]
[ "def visit_Module(self, node):...\n", "\"\"\"docstring\"\"\"\n", "self.mongo_query = {}\n", "self.ops = []\n", "self.current_value = None\n", "self.generic_visit(node)\n", "if self.mongo_query == {}:\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Assign'", "Expr'", "Condition" ]
[ "import logging\n", "import urllib\n", "from collections import defaultdict\n", "import attr\n", "from signedjson.key import decode_verify_key_bytes, encode_verify_key_base64, is_signing_algorithm_supported\n", "from signedjson.sign import SignatureVerifyException, encode_canonical_json, signature_ids, verify_signed_json\n", "from unpaddedbase64 import decode_base64\n", "from twisted.internet import defer\n", "from synapse.api.errors import Codes, HttpResponseException, RequestSendFailed, SynapseError\n", "from synapse.logging.context import PreserveLoggingContext, make_deferred_yieldable, preserve_fn, run_in_background\n", "from synapse.storage.keys import FetchKeyResult\n", "from synapse.util import unwrapFirstError\n", "from synapse.util.async_helpers import yieldable_gather_results\n", "from synapse.util.metrics import Measure\n", "from synapse.util.retryutils import NotRetryingDestination\n", "VAR_0 = logging.getLogger(__name__)\n", "\"\"\"string\"\"\"\n", "VAR_2 = attr.ib()\n", "VAR_3 = attr.ib()\n", "VAR_4 = attr.ib()\n", "VAR_5 = attr.ib()\n", "VAR_6 = attr.ib(init=False)\n", "VAR_7 = attr.ib(default=attr.Factory(defer.Deferred))\n", "def __attrs_post_init__(self):...\n", "self.key_ids = signature_ids(self.json_object, self.server_name)\n", "def __init__(self, VAR_8, VAR_9=None):...\n", "self.clock = VAR_8.get_clock()\n", "if VAR_9 is None:\n", "VAR_9 = CLASS_4(VAR_8), CLASS_6(VAR_8), CLASS_7(VAR_8)\n", "self._key_fetchers = VAR_9\n", "self.key_downloads = {}\n", "def FUNC_1(self, VAR_2, VAR_3, VAR_10, VAR_5):...\n", "\"\"\"docstring\"\"\"\n", "VAR_22 = CLASS_0(VAR_2, VAR_3, VAR_10, VAR_5)\n", "VAR_23 = VAR_22,\n", "return make_deferred_yieldable(self._verify_objects(VAR_23)[0])\n" ]
[ "import logging\n", "import urllib\n", "from collections import defaultdict\n", "import attr\n", "from signedjson.key import decode_verify_key_bytes, encode_verify_key_base64, is_signing_algorithm_supported\n", "from signedjson.sign import SignatureVerifyException, encode_canonical_json, signature_ids, verify_signed_json\n", "from unpaddedbase64 import decode_base64\n", "from twisted.internet import defer\n", "from synapse.api.errors import Codes, HttpResponseException, RequestSendFailed, SynapseError\n", "from synapse.logging.context import PreserveLoggingContext, make_deferred_yieldable, preserve_fn, run_in_background\n", "from synapse.storage.keys import FetchKeyResult\n", "from synapse.util import unwrapFirstError\n", "from synapse.util.async_helpers import yieldable_gather_results\n", "from synapse.util.metrics import Measure\n", "from synapse.util.retryutils import NotRetryingDestination\n", "logger = logging.getLogger(__name__)\n", "\"\"\"\n A request to verify a JSON object.\n\n Attributes:\n server_name(str): The name of the server to verify against.\n\n key_ids(set[str]): The set of key_ids to that could be used to verify the\n JSON object\n\n json_object(dict): The JSON object to verify.\n\n minimum_valid_until_ts (int): time at which we require the signing key to\n be valid. (0 implies we don't care)\n\n key_ready (Deferred[str, str, nacl.signing.VerifyKey]):\n A deferred (server_name, key_id, verify_key) tuple that resolves when\n a verify key has been fetched. The deferreds' callbacks are run with no\n logcontext.\n\n If we are unable to find a key which satisfies the request, the deferred\n errbacks with an M_UNAUTHORIZED SynapseError.\n \"\"\"\n", "server_name = attr.ib()\n", "json_object = attr.ib()\n", "minimum_valid_until_ts = attr.ib()\n", "request_name = attr.ib()\n", "key_ids = attr.ib(init=False)\n", "key_ready = attr.ib(default=attr.Factory(defer.Deferred))\n", "def __attrs_post_init__(self):...\n", "self.key_ids = signature_ids(self.json_object, self.server_name)\n", "def __init__(self, hs, key_fetchers=None):...\n", "self.clock = hs.get_clock()\n", "if key_fetchers is None:\n", "key_fetchers = StoreKeyFetcher(hs), PerspectivesKeyFetcher(hs\n ), ServerKeyFetcher(hs)\n", "self._key_fetchers = key_fetchers\n", "self.key_downloads = {}\n", "def verify_json_for_server(self, server_name, json_object, validity_time,...\n", "\"\"\"docstring\"\"\"\n", "req = VerifyJsonRequest(server_name, json_object, validity_time, request_name)\n", "requests = req,\n", "return make_deferred_yieldable(self._verify_objects(requests)[0])\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Import'", "Import'", "ImportFrom'", "Import'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Assign'", "Expr'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "FunctionDef'", "Assign'", "FunctionDef'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "FunctionDef'", "Docstring", "Assign'", "Assign'", "Return'" ]
[ "@VAR_2.route('/admin/dbconfig', methods=['GET', 'POST'])...\n", "if request.method == 'POST':\n", "return FUNC_48()\n", "return FUNC_51()\n" ]
[ "@admi.route('/admin/dbconfig', methods=['GET', 'POST'])...\n", "if request.method == 'POST':\n", "return _db_configuration_update_helper()\n", "return _db_configuration_result()\n" ]
[ 0, 0, 0, 0 ]
[ "Condition", "Condition", "Return'", "Return'" ]
[ "def FUNC_5(self):...\n", "VAR_5 = self._makeContext()\n", "self.assertEqual(VAR_5.evaluate('one'), 1)\n" ]
[ "def test_evaluate_simple_path_binding(self):...\n", "ec = self._makeContext()\n", "self.assertEqual(ec.evaluate('one'), 1)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Expr'" ]
[ "\"\"\"Tests for SavedModelCLI tool.\"\"\"\n", "import contextlib\n", "import os\n", "import pickle\n", "import platform\n", "import shutil\n", "import sys\n", "from absl.testing import parameterized\n", "import numpy as np\n", "from six import StringIO\n", "from tensorflow.core.example import example_pb2\n", "from tensorflow.core.framework import types_pb2\n", "from tensorflow.core.protobuf import meta_graph_pb2\n", "from tensorflow.python.debug.wrappers import local_cli_wrapper\n", "from tensorflow.python.eager import def_function\n", "from tensorflow.python.framework import constant_op\n", "from tensorflow.python.framework import dtypes\n", "from tensorflow.python.framework import tensor_spec\n", "from tensorflow.python.lib.io import file_io\n", "from tensorflow.python.ops import variables\n", "from tensorflow.python.platform import test\n", "from tensorflow.python.platform import tf_logging as logging\n", "from tensorflow.python.saved_model import save\n", "from tensorflow.python.tools import saved_model_cli\n", "from tensorflow.python.training.tracking import tracking\n", "VAR_0 = 'cc/saved_model/testdata/half_plus_two/00000123'\n", "@contextlib.contextmanager...\n", "VAR_1, VAR_2 = StringIO(), StringIO()\n", "VAR_3, VAR_4 = VAR_9.stdout, VAR_9.stderr\n", "VAR_9.stdout, VAR_9.stderr = VAR_1, VAR_2\n", "VAR_9.stdout, VAR_9.stderr = VAR_3, VAR_4\n", "def FUNC_1(self):...\n", "yield VAR_9.stdout, VAR_9.stderr\n", "super(CLASS_0, self).setUp()\n", "if platform.system() == 'Windows':\n", "self.skipTest('Skipping failing tests on Windows.')\n", "def FUNC_2(self):...\n", "VAR_10 = test.test_src_dir_path(VAR_0)\n", "self.parser = saved_model_cli.create_parser()\n", "VAR_11 = self.parser.parse_args(['show', '--dir', VAR_10, '--all'])\n", "saved_model_cli.show(VAR_11)\n", "VAR_12 = out.getvalue().strip()\n", "VAR_13 = 'string'\n", "self.maxDiff = None\n", "self.assertMultiLineEqual(VAR_12, VAR_13)\n", "self.assertEqual(err.getvalue().strip(), '')\n", "def FUNC_3(self):...\n", "\"\"\"Model with callable polymorphic functions specified.\"\"\"\n", "@def_function.function...\n", "if VAR_56:\n", "return VAR_54 + VAR_55\n", "return VAR_54 * VAR_55\n" ]
[ "\"\"\"Tests for SavedModelCLI tool.\"\"\"\n", "import contextlib\n", "import os\n", "import pickle\n", "import platform\n", "import shutil\n", "import sys\n", "from absl.testing import parameterized\n", "import numpy as np\n", "from six import StringIO\n", "from tensorflow.core.example import example_pb2\n", "from tensorflow.core.framework import types_pb2\n", "from tensorflow.core.protobuf import meta_graph_pb2\n", "from tensorflow.python.debug.wrappers import local_cli_wrapper\n", "from tensorflow.python.eager import def_function\n", "from tensorflow.python.framework import constant_op\n", "from tensorflow.python.framework import dtypes\n", "from tensorflow.python.framework import tensor_spec\n", "from tensorflow.python.lib.io import file_io\n", "from tensorflow.python.ops import variables\n", "from tensorflow.python.platform import test\n", "from tensorflow.python.platform import tf_logging as logging\n", "from tensorflow.python.saved_model import save\n", "from tensorflow.python.tools import saved_model_cli\n", "from tensorflow.python.training.tracking import tracking\n", "SAVED_MODEL_PATH = 'cc/saved_model/testdata/half_plus_two/00000123'\n", "@contextlib.contextmanager...\n", "new_out, new_err = StringIO(), StringIO()\n", "old_out, old_err = sys.stdout, sys.stderr\n", "sys.stdout, sys.stderr = new_out, new_err\n", "sys.stdout, sys.stderr = old_out, old_err\n", "def setUp(self):...\n", "yield sys.stdout, sys.stderr\n", "super(SavedModelCLITestCase, self).setUp()\n", "if platform.system() == 'Windows':\n", "self.skipTest('Skipping failing tests on Windows.')\n", "def testShowCommandAll(self):...\n", "base_path = test.test_src_dir_path(SAVED_MODEL_PATH)\n", "self.parser = saved_model_cli.create_parser()\n", "args = self.parser.parse_args(['show', '--dir', base_path, '--all'])\n", "saved_model_cli.show(args)\n", "output = out.getvalue().strip()\n", "exp_out = \"\"\"MetaGraphDef with tag-set: 'serve' contains the following SignatureDefs:\n\nsignature_def['classify_x2_to_y3']:\n The given SavedModel SignatureDef contains the following input(s):\n inputs['inputs'] tensor_info:\n dtype: DT_FLOAT\n shape: (-1, 1)\n name: x2:0\n The given SavedModel SignatureDef contains the following output(s):\n outputs['scores'] tensor_info:\n dtype: DT_FLOAT\n shape: (-1, 1)\n name: y3:0\n Method name is: tensorflow/serving/classify\n\nsignature_def['classify_x_to_y']:\n The given SavedModel SignatureDef contains the following input(s):\n inputs['inputs'] tensor_info:\n dtype: DT_STRING\n shape: unknown_rank\n name: tf_example:0\n The given SavedModel SignatureDef contains the following output(s):\n outputs['scores'] tensor_info:\n dtype: DT_FLOAT\n shape: (-1, 1)\n name: y:0\n Method name is: tensorflow/serving/classify\n\nsignature_def['regress_x2_to_y3']:\n The given SavedModel SignatureDef contains the following input(s):\n inputs['inputs'] tensor_info:\n dtype: DT_FLOAT\n shape: (-1, 1)\n name: x2:0\n The given SavedModel SignatureDef contains the following output(s):\n outputs['outputs'] tensor_info:\n dtype: DT_FLOAT\n shape: (-1, 1)\n name: y3:0\n Method name is: tensorflow/serving/regress\n\nsignature_def['regress_x_to_y']:\n The given SavedModel SignatureDef contains the following input(s):\n inputs['inputs'] tensor_info:\n dtype: DT_STRING\n shape: unknown_rank\n name: tf_example:0\n The given SavedModel SignatureDef contains the following output(s):\n outputs['outputs'] tensor_info:\n dtype: DT_FLOAT\n shape: (-1, 1)\n name: y:0\n Method name is: tensorflow/serving/regress\n\nsignature_def['regress_x_to_y2']:\n The given SavedModel SignatureDef contains the following input(s):\n inputs['inputs'] tensor_info:\n dtype: DT_STRING\n shape: unknown_rank\n name: tf_example:0\n The given SavedModel SignatureDef contains the following output(s):\n outputs['outputs'] tensor_info:\n dtype: DT_FLOAT\n shape: (-1, 1)\n name: y2:0\n Method name is: tensorflow/serving/regress\n\nsignature_def['serving_default']:\n The given SavedModel SignatureDef contains the following input(s):\n inputs['x'] tensor_info:\n dtype: DT_FLOAT\n shape: (-1, 1)\n name: x:0\n The given SavedModel SignatureDef contains the following output(s):\n outputs['y'] tensor_info:\n dtype: DT_FLOAT\n shape: (-1, 1)\n name: y:0\n Method name is: tensorflow/serving/predict\"\"\"\n", "self.maxDiff = None\n", "self.assertMultiLineEqual(output, exp_out)\n", "self.assertEqual(err.getvalue().strip(), '')\n", "def testShowAllWithFunctions(self):...\n", "\"\"\"Model with callable polymorphic functions specified.\"\"\"\n", "@def_function.function...\n", "if c:\n", "return a + b\n", "return a * b\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Expr'", "Import'", "Import'", "Import'", "Import'", "Import'", "Import'", "ImportFrom'", "Import'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "FunctionDef'", "Expr'", "Expr'", "For", "Expr'", "FunctionDef'", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'", "FunctionDef'", "Expr'", "Condition", "Condition", "Return'", "Return'" ]
[ "def __call__(self):...\n", "" ]
[ "def __call__(self):...\n", "" ]
[ 0, 0 ]
[ "FunctionDef'", "Condition" ]
[ "@VAR_0.route('/credentials')...\n", "\"\"\"docstring\"\"\"\n", "VAR_108 = FUNC_58('/internal/credentials', 'get')\n", "flash(str(err), 'danger')\n", "return render_template('credentials.html', credential_info_list=\n credentials_info)\n", "return redirect(url_for('dashboard'))\n" ]
[ "@gui.route('/credentials')...\n", "\"\"\"docstring\"\"\"\n", "credentials_info = query_internal_api('/internal/credentials', 'get')\n", "flash(str(err), 'danger')\n", "return render_template('credentials.html', credential_info_list=\n credentials_info)\n", "return redirect(url_for('dashboard'))\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Expr'", "Return'", "Return'" ]
[ "def FUNC_30():...\n", "VAR_82 = FUNC_13(VAR_1[VAR_14], FUNC_26)\n", "VAR_107 = FUNC_25(VAR_82)\n", "VAR_108 = userManager.has_been_customized()\n", "VAR_43 = FUNC_9(VAR_82, VAR_2, VAR_3, VAR_21)\n", "VAR_43.update({'enableWebcam': VAR_54, 'enableTemperatureGraph': VAR_55,\n 'enableAccessControl': True, 'accessControlActive': VAR_108,\n 'enableLoadingAnimation': VAR_52, 'enableSdSupport': VAR_53,\n 'sockJsConnectTimeout': VAR_56 * 1000, 'wizard': VAR_107, 'online':\n connectivityChecker.online, 'now': VAR_21})\n", "def FUNC_42():...\n", "VAR_129 = make_response(render_template('index.jinja2', **render_kwargs))\n", "if VAR_107:\n", "VAR_129 = util.flask.add_non_caching_response_headers(VAR_129)\n", "return VAR_129\n" ]
[ "def default_view():...\n", "filtered_templates = _filter_templates(_templates[locale],\n default_template_filter)\n", "wizard = wizard_active(filtered_templates)\n", "accesscontrol_active = userManager.has_been_customized()\n", "render_kwargs = _get_render_kwargs(filtered_templates, _plugin_names,\n _plugin_vars, now)\n", "render_kwargs.update({'enableWebcam': enable_webcam,\n 'enableTemperatureGraph': enable_temperature_graph,\n 'enableAccessControl': True, 'accessControlActive':\n accesscontrol_active, 'enableLoadingAnimation':\n enable_loading_animation, 'enableSdSupport': enable_sd_support,\n 'sockJsConnectTimeout': sockjs_connect_timeout * 1000, 'wizard': wizard,\n 'online': connectivityChecker.online, 'now': now})\n", "def make_default_ui():...\n", "r = make_response(render_template('index.jinja2', **render_kwargs))\n", "if wizard:\n", "r = util.flask.add_non_caching_response_headers(r)\n", "return r\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "FunctionDef'", "Assign'", "Condition", "Assign'", "Return'" ]
[ "def FUNC_8(self, VAR_3):...\n", "\"\"\"docstring\"\"\"\n", "self.current_value = VAR_3.n\n" ]
[ "def visit_Num(self, node):...\n", "\"\"\"docstring\"\"\"\n", "self.current_value = node.n\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'" ]
[ "@CLASS_4('javascript')...\n", "\"\"\"docstring\"\"\"\n", "VAR_15 = VAR_3.path()\n", "if VAR_15:\n", "VAR_15 = 'javascript' + os.sep.join(VAR_15.split('/'))\n", "return 'text/html', utils.read_file(VAR_15, binary=False)\n" ]
[ "@add_handler('javascript')...\n", "\"\"\"docstring\"\"\"\n", "path = url.path()\n", "if path:\n", "path = 'javascript' + os.sep.join(path.split('/'))\n", "return 'text/html', utils.read_file(path, binary=False)\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Condition", "Assign'", "Return'" ]
[ "def FUNC_49(VAR_62):...\n", "import frappe.modules\n", "return frappe.modules.load_doctype_module(VAR_62)\n" ]
[ "def get_meta_module(doctype):...\n", "import frappe.modules\n", "return frappe.modules.load_doctype_module(doctype)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Import'", "Return'" ]
[ "def FUNC_0(VAR_0: Team, VAR_1: HttpRequest) ->bool:...\n", "VAR_3 = ['127.0.0.1', 'localhost']\n", "for VAR_2 in VAR_0.app_urls:\n", "VAR_7 = FUNC_2(VAR_2)\n", "VAR_4 = FUNC_2(VAR_1.headers.get('Origin'))\n", "if VAR_7:\n", "VAR_5 = FUNC_2(VAR_1.headers.get('Referer'))\n", "VAR_3.append(VAR_7)\n", "for permitted_domain in VAR_3:\n", "if '*' in permitted_domain:\n", "return False\n", "VAR_12 = '^{}$'.format(permitted_domain.replace('.', '\\\\.').replace('*',\n '(.*)'))\n", "if permitted_domain == VAR_4 or permitted_domain == VAR_5:\n", "if VAR_4 and re.search(VAR_12, VAR_4) or VAR_5 and re.search(VAR_12, VAR_5):\n", "return True\n", "return True\n" ]
[ "def on_permitted_domain(team: Team, request: HttpRequest) ->bool:...\n", "permitted_domains = ['127.0.0.1', 'localhost']\n", "for url in team.app_urls:\n", "hostname = parse_domain(url)\n", "origin = parse_domain(request.headers.get('Origin'))\n", "if hostname:\n", "referer = parse_domain(request.headers.get('Referer'))\n", "permitted_domains.append(hostname)\n", "for permitted_domain in permitted_domains:\n", "if '*' in permitted_domain:\n", "return False\n", "pattern = '^{}$'.format(permitted_domain.replace('.', '\\\\.').replace('*',\n '(.*)'))\n", "if permitted_domain == origin or permitted_domain == referer:\n", "if origin and re.search(pattern, origin) or referer and re.search(pattern,\n", "return True\n", "return True\n" ]
[ 0, 0, 0, 4, 4, 4, 4, 4, 0, 0, 0, 4, 4, 4, 0, 4 ]
[ "FunctionDef'", "Assign'", "For", "Assign'", "Assign'", "Condition", "Assign'", "Expr'", "For", "Condition", "Return'", "Assign'", "Condition", "Condition", "Return'", "Return'" ]
[ "def FUNC_17(VAR_1):...\n", "VAR_6 = CLASS_1()\n", "VAR_6.name = 'admin'\n", "VAR_6.role = constants.ADMIN_USER_ROLES\n", "VAR_6.sidebar_view = constants.ADMIN_USER_SIDEBAR\n", "VAR_6.password = generate_password_hash(constants.DEFAULT_PASSWORD)\n", "VAR_1.add(VAR_6)\n", "VAR_1.commit()\n", "VAR_1.rollback()\n" ]
[ "def create_admin_user(session):...\n", "user = User()\n", "user.name = 'admin'\n", "user.role = constants.ADMIN_USER_ROLES\n", "user.sidebar_view = constants.ADMIN_USER_SIDEBAR\n", "user.password = generate_password_hash(constants.DEFAULT_PASSWORD)\n", "session.add(user)\n", "session.commit()\n", "session.rollback()\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'" ]
[ "def FUNC_20(self):...\n", "VAR_5 = {'not_rooms': ['!secretbase:unknown']}\n", "VAR_6 = FUNC_0(sender='@foo:bar', type='m.room.message', room_id=\n '!anothersecretbase:unknown')\n", "self.assertTrue(Filter(VAR_5).check(VAR_6))\n" ]
[ "def test_definition_not_rooms_works_with_unknowns(self):...\n", "definition = {'not_rooms': ['!secretbase:unknown']}\n", "event = MockEvent(sender='@foo:bar', type='m.room.message', room_id=\n '!anothersecretbase:unknown')\n", "self.assertTrue(Filter(definition).check(event))\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Expr'" ]
[ "def __setstate__(self, VAR_174):...\n", "self.update(VAR_174)\n" ]
[ "def __setstate__(self, d):...\n", "self.update(d)\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Expr'" ]
[ "def FUNC_7(self):...\n", "VAR_53 = VAR_1(VAR_2, 'notebook', 'static', 'base', 'js', 'namespace.js')\n", "VAR_49 = f.readlines()\n", "VAR_59 = False\n", "for VAR_68 in VAR_49:\n", "if VAR_68.strip().startswith('Jupyter.version'):\n", "if not VAR_59:\n", "VAR_68 = ' Jupyter.version = \"{0}\";\\n'.format(VAR_7)\n", "f.write(VAR_68)\n", "VAR_59 = True\n" ]
[ "def run(self):...\n", "nsfile = pjoin(repo_root, 'notebook', 'static', 'base', 'js', 'namespace.js')\n", "lines = f.readlines()\n", "found = False\n", "for line in lines:\n", "if line.strip().startswith('Jupyter.version'):\n", "if not found:\n", "line = ' Jupyter.version = \"{0}\";\\n'.format(version)\n", "f.write(line)\n", "found = True\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "For", "Condition", "Condition", "Assign'", "Expr'", "Assign'" ]
[ "def FUNC_34(self):...\n", "return str(self.id)\n" ]
[ "def get_id(self):...\n", "return str(self.id)\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_48(VAR_62, VAR_108=True):...\n", "\"\"\"docstring\"\"\"\n", "import frappe.model.meta\n", "return frappe.model.meta.get_meta(VAR_62, VAR_108=cached)\n" ]
[ "def get_meta(doctype, cached=True):...\n", "\"\"\"docstring\"\"\"\n", "import frappe.model.meta\n", "return frappe.model.meta.get_meta(doctype, cached=cached)\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Import'", "Return'" ]
[ "@log_function...\n", "\"\"\"docstring\"\"\"\n", "VAR_2 = FUNC_1('/groups/%s/invited_users', VAR_30)\n", "return self.client.get_json(VAR_5=destination, VAR_2=path, VAR_3={\n 'requester_user_id': requester_user_id}, VAR_15=True)\n" ]
[ "@log_function...\n", "\"\"\"docstring\"\"\"\n", "path = _create_v1_path('/groups/%s/invited_users', group_id)\n", "return self.client.get_json(destination=destination, path=path, args={\n 'requester_user_id': requester_user_id}, ignore_backoff=True)\n" ]
[ 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Return'" ]
[ "def FUNC_168(*VAR_9, **VAR_13):...\n", "VAR_27 = None\n", "VAR_27 = VAR_10(*VAR_9, **b)\n", "thread.start_new_thread(VAR_26, (VAR_27,))\n", "return VAR_27\n" ]
[ "def __completion(*a, **b):...\n", "d = None\n", "d = f(*a, **b)\n", "thread.start_new_thread(callback, (d,))\n", "return d\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Expr'", "Return'" ]
[ "def FUNC_91(VAR_62, VAR_9, VAR_152=None, VAR_146=None, VAR_147=None,...\n", "from frappe.utils import scrub_urls\n", "if not VAR_152:\n", "VAR_152 = VAR_9\n", "VAR_152 = VAR_152.replace(' ', '').replace('/', '-')\n", "VAR_199 = VAR_12.get_singles_dict('Print Settings')\n", "VAR_200 = VAR_1.lang\n", "if VAR_4:\n", "VAR_1.lang = VAR_4\n", "VAR_1.flags.ignore_print_permissions = True\n", "VAR_150 = not VAR_81\n", "VAR_42 = dict(VAR_146=print_format, VAR_147=style, VAR_136=html, VAR_94=doc,\n VAR_150=no_letterhead, VAR_151=password)\n", "VAR_61 = ''\n", "if int(VAR_199.send_print_as_pdf or 0):\n", "VAR_213 = '.pdf'\n", "VAR_213 = '.html'\n", "VAR_42['as_pdf'] = True\n", "VAR_61 = scrub_urls(FUNC_90(VAR_62, VAR_9, **kwargs)).encode('utf-8')\n", "VAR_61 = FUNC_90(VAR_62, VAR_9, **kwargs)\n", "VAR_179 = {'fname': VAR_152 + VAR_213, 'fcontent': VAR_61}\n", "VAR_1.flags.ignore_print_permissions = False\n", "VAR_1.lang = VAR_200\n", "return VAR_179\n" ]
[ "def attach_print(doctype, name, file_name=None, print_format=None, style=...\n", "from frappe.utils import scrub_urls\n", "if not file_name:\n", "file_name = name\n", "file_name = file_name.replace(' ', '').replace('/', '-')\n", "print_settings = db.get_singles_dict('Print Settings')\n", "_lang = local.lang\n", "if lang:\n", "local.lang = lang\n", "local.flags.ignore_print_permissions = True\n", "no_letterhead = not print_letterhead\n", "kwargs = dict(print_format=print_format, style=style, html=html, doc=doc,\n no_letterhead=no_letterhead, password=password)\n", "content = ''\n", "if int(print_settings.send_print_as_pdf or 0):\n", "ext = '.pdf'\n", "ext = '.html'\n", "kwargs['as_pdf'] = True\n", "content = scrub_urls(get_print(doctype, name, **kwargs)).encode('utf-8')\n", "content = get_print(doctype, name, **kwargs)\n", "out = {'fname': file_name + ext, 'fcontent': content}\n", "local.flags.ignore_print_permissions = False\n", "local.lang = _lang\n", "return out\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "ImportFrom'", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_33():...\n", "\"\"\"docstring\"\"\"\n", "VAR_3 = FUNC_5()\n", "VAR_79 = FUNC_3(apath('%s/ABOUT' % VAR_3, VAR_122=request))\n", "VAR_80 = FUNC_3(apath('%s/LICENSE' % VAR_3, VAR_122=request))\n", "return dict(VAR_3=app, VAR_79=MARKMIN(about), VAR_80=MARKMIN(license),\n VAR_6=report_progress(app))\n" ]
[ "def about():...\n", "\"\"\"docstring\"\"\"\n", "app = get_app()\n", "about = safe_read(apath('%s/ABOUT' % app, r=request))\n", "license = safe_read(apath('%s/LICENSE' % app, r=request))\n", "return dict(app=app, about=MARKMIN(about), license=MARKMIN(license),\n progress=report_progress(app))\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_109(VAR_170):...\n", "from frappe.utils import parse_json\n", "return FUNC_109(VAR_170)\n" ]
[ "def parse_json(val):...\n", "from frappe.utils import parse_json\n", "return parse_json(val)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "ImportFrom'", "Return'" ]
[ "import re\n", "from twisted.internet.defer import Deferred\n", "from twisted.web.resource import Resource\n", "from synapse.api.errors import Codes, RedirectException, SynapseError\n", "from synapse.config.server import parse_listener_def\n", "from synapse.http.server import DirectServeHtmlResource, JsonResource, OptionsResource\n", "from synapse.http.site import SynapseSite\n", "from synapse.logging.context import make_deferred_yieldable\n", "from synapse.util import Clock\n", "from tests import unittest\n", "from tests.server import FakeSite, ThreadedMemoryReactorClock, make_request, setup_test_homeserver\n", "def FUNC_0(self):...\n", "self.reactor = ThreadedMemoryReactorClock()\n", "self.hs_clock = Clock(self.reactor)\n", "self.homeserver = setup_test_homeserver(self.addCleanup, http_client=None,\n clock=self.hs_clock, reactor=self.reactor)\n", "def FUNC_1(self):...\n", "\"\"\"docstring\"\"\"\n", "VAR_2 = {}\n", "def FUNC_15(VAR_3, **VAR_4):...\n", "VAR_2.update(VAR_4)\n", "return 200, VAR_4\n" ]
[ "import re\n", "from twisted.internet.defer import Deferred\n", "from twisted.web.resource import Resource\n", "from synapse.api.errors import Codes, RedirectException, SynapseError\n", "from synapse.config.server import parse_listener_def\n", "from synapse.http.server import DirectServeHtmlResource, JsonResource, OptionsResource\n", "from synapse.http.site import SynapseSite\n", "from synapse.logging.context import make_deferred_yieldable\n", "from synapse.util import Clock\n", "from tests import unittest\n", "from tests.server import FakeSite, ThreadedMemoryReactorClock, make_request, setup_test_homeserver\n", "def setUp(self):...\n", "self.reactor = ThreadedMemoryReactorClock()\n", "self.hs_clock = Clock(self.reactor)\n", "self.homeserver = setup_test_homeserver(self.addCleanup, http_client=None,\n clock=self.hs_clock, reactor=self.reactor)\n", "def test_handler_for_request(self):...\n", "\"\"\"docstring\"\"\"\n", "got_kwargs = {}\n", "def _callback(request, **kwargs):...\n", "got_kwargs.update(kwargs)\n", "return 200, kwargs\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 4, 0, 0, 0, 0, 0, 0 ]
[ "Import'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "FunctionDef'", "Assign'", "Assign'", "Assign'", "FunctionDef'", "Docstring", "Assign'", "FunctionDef'", "Expr'", "Return'" ]
[ "def FUNC_21(self):...\n", "return [self.cleaned_data['lang']]\n" ]
[ "def clean_lang(self):...\n", "return [self.cleaned_data['lang']]\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "@CLASS_4('configdiff')...\n", "\"\"\"docstring\"\"\"\n", "if VAR_3.path() == '/old':\n", "VAR_32 = config.instance.dump_userconfig().encode('utf-8')\n", "return 'text/html', configdiff.get_diff()\n", "VAR_8 = b'Failed to read old config: ' + str(e.strerror).encode('utf-8')\n", "return 'text/plain', VAR_32\n", "return 'text/plain', VAR_8\n" ]
[ "@add_handler('configdiff')...\n", "\"\"\"docstring\"\"\"\n", "if url.path() == '/old':\n", "data = config.instance.dump_userconfig().encode('utf-8')\n", "return 'text/html', configdiff.get_diff()\n", "error = b'Failed to read old config: ' + str(e.strerror).encode('utf-8')\n", "return 'text/plain', data\n", "return 'text/plain', error\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Condition", "Assign'", "Return'", "Assign'", "Return'", "Return'" ]
[ "def FUNC_15(self, VAR_36: Failure) ->None:...\n", "if VAR_36.check(ResponseDone):\n", "self.deferred.callback(self.length)\n", "if VAR_36.check(PotentialDataLoss):\n", "self.deferred.callback(self.length)\n", "self.deferred.errback(VAR_36)\n" ]
[ "def connectionLost(self, reason: Failure) ->None:...\n", "if reason.check(ResponseDone):\n", "self.deferred.callback(self.length)\n", "if reason.check(PotentialDataLoss):\n", "self.deferred.callback(self.length)\n", "self.deferred.errback(reason)\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Expr'", "Condition", "Expr'", "Expr'" ]
[ "def FUNC_1(self):...\n", "self.edit_unit('Hello, world!\\n', 'Nazdar svete!\\n')\n" ]
[ "def add_change(self):...\n", "self.edit_unit('Hello, world!\\n', 'Nazdar svete!\\n')\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Expr'" ]
[ "@VAR_7.route('/meta/<path:object_path>')...\n", "VAR_36 = FUNC_10(VAR_19)\n", "VAR_37 = dict()\n", "VAR_37['_gt_label'] = VAR_36.split('/')[-2]\n", "return jsonify(VAR_37)\n" ]
[ "@scope_blueprint.route('/meta/<path:object_path>')...\n", "path = _get_obj_absolute_path(object_path)\n", "attrs = dict()\n", "attrs['_gt_label'] = path.split('/')[-2]\n", "return jsonify(attrs)\n" ]
[ 0, 0, 0, 0, 0 ]
[ "Condition", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_2(VAR_3):...\n", "\"\"\"docstring\"\"\"\n", "VAR_20 = VAR_3.strip().split()\n", "if len(VAR_20) > 0:\n", "return ' ' + ' '.join(VAR_20) + ' '\n", "return ''\n" ]
[ "def pad_string(text):...\n", "\"\"\"docstring\"\"\"\n", "words = text.strip().split()\n", "if len(words) > 0:\n", "return ' ' + ' '.join(words) + ' '\n", "return ''\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Condition", "Return'", "Return'" ]
[ "async def FUNC_24(VAR_23):...\n", "self.assertEqual(VAR_23, {'server1': {FUNC_0(VAR_20): 1500}})\n", "return {'server1': {FUNC_0(VAR_20): FetchKeyResult(get_verify_key(VAR_20), \n 1200)}}\n" ]
[ "async def get_keys(keys_to_fetch):...\n", "self.assertEqual(keys_to_fetch, {'server1': {get_key_id(key1): 1500}})\n", "return {'server1': {get_key_id(key1): FetchKeyResult(get_verify_key(key1), \n 1200)}}\n" ]
[ 0, 0, 0 ]
[ "AsyncFunctionDef'", "Expr'", "Return'" ]
[ "def __init__(self, *VAR_17, **VAR_18):...\n", "super(CLASS_0, self).__init__(*VAR_17, **kwargs)\n", "VAR_49 = self.spec.split('/')\n", "self.user, self.gist_id, *VAR_50 = VAR_49\n", "if len(VAR_49) > 2:\n", "self.unresolved_ref = VAR_49[2]\n", "self.unresolved_ref = ''\n" ]
[ "def __init__(self, *args, **kwargs):...\n", "super(RepoProvider, self).__init__(*args, **kwargs)\n", "parts = self.spec.split('/')\n", "self.user, self.gist_id, *_ = parts\n", "if len(parts) > 2:\n", "self.unresolved_ref = parts[2]\n", "self.unresolved_ref = ''\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'" ]