lines
sequencelengths 1
444
| raw_lines
sequencelengths 1
444
| label
sequencelengths 1
444
| type
sequencelengths 1
444
|
---|---|---|---|
[
"def FUNC_6(self, VAR_33):...\n",
"VAR_78 = []\n",
"for v in self.splitter.split(VAR_33):\n",
"VAR_21 = self.id_re.match(v)\n",
"return VAR_78\n",
"if VAR_21:\n",
"VAR_108 = CLASS_5(None, False).run(VAR_21.group(1))\n",
"if VAR_108:\n",
"VAR_78.append(VAR_108)\n"
] | [
"def run(self, val):...\n",
"res = []\n",
"for v in self.splitter.split(val):\n",
"link_id = self.id_re.match(v)\n",
"return res\n",
"if link_id:\n",
"l = VLink(None, False).run(link_id.group(1))\n",
"if l:\n",
"res.append(l)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"For",
"Assign'",
"Return'",
"Condition",
"Assign'",
"Condition",
"Expr'"
] |
[
"def FUNC_0(VAR_2):...\n",
"if VAR_2.user.is_authenticated():\n",
"VAR_33 = VAR_2.GET.get('next', reverse('home'))\n",
"VAR_1.error('login() - got anonymous user: %s' % VAR_2.META)\n",
"return HttpResponseRedirect(VAR_33)\n",
"return HttpResponseServerError(\n 'Internet Server error. Please contact %s for assistance.' % VAR_0)\n"
] | [
"def login(request):...\n",
"if request.user.is_authenticated():\n",
"next_url = request.GET.get('next', reverse('home'))\n",
"logger.error('login() - got anonymous user: %s' % request.META)\n",
"return HttpResponseRedirect(next_url)\n",
"return HttpResponseServerError(\n 'Internet Server error. Please contact %s for assistance.' % BDR_EMAIL)\n"
] | [
0,
0,
6,
0,
0,
0
] | [
"FunctionDef'",
"Condition",
"Assign'",
"Expr'",
"Return'",
"Return'"
] |
[
"def FUNC_2(self):...\n",
"return sqlite\n"
] | [
"def dbapiModule(self):...\n",
"return sqlite\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
[
"def FUNC_11(self, VAR_9, *VAR_13, **VAR_14):...\n",
"VAR_14['partial'] = True\n",
"return self.update(VAR_9, *VAR_13, **kwargs)\n"
] | [
"def partial_update(self, request, *args, **kwargs):...\n",
"kwargs['partial'] = True\n",
"return self.update(request, *args, **kwargs)\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Return'"
] |
[
"def __init__(self, VAR_25=settings.CLICKHOUSE_SERVER.split(':')[0], VAR_26=...\n",
"self.host = VAR_25\n",
"self.port = VAR_26\n",
"self.connect_timeout = VAR_27\n",
"self.send_receive_timeout = VAR_28\n",
"self.client_settings = VAR_30\n",
"self.metrics = VAR_31\n",
"self.pool = queue.LifoQueue(VAR_29)\n",
"for _ in range(VAR_29):\n",
"self.pool.put(None)\n"
] | [
"def __init__(self, host=settings.CLICKHOUSE_SERVER.split(':')[0], port=int(...\n",
"self.host = host\n",
"self.port = port\n",
"self.connect_timeout = connect_timeout\n",
"self.send_receive_timeout = send_receive_timeout\n",
"self.client_settings = client_settings\n",
"self.metrics = metrics\n",
"self.pool = queue.LifoQueue(max_pool_size)\n",
"for _ in range(max_pool_size):\n",
"self.pool.put(None)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"For",
"Expr'"
] |
[
"def FUNC_13(self, *VAR_13, **VAR_14):...\n",
"for VAR_9 in VAR_13:\n",
"self._set_log_item(VAR_9)\n",
"for VAR_10, VAR_9 in VAR_14.items():\n",
"self._set_log_item(VAR_9, VAR_10=name)\n"
] | [
"def set_log(self, *logs, **kwlogs):...\n",
"for item in logs:\n",
"self._set_log_item(item)\n",
"for name, item in kwlogs.items():\n",
"self._set_log_item(item, name=name)\n"
] | [
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"For",
"Expr'",
"For",
"Expr'"
] |
[
"def FUNC_8(self):...\n",
"url_helper.urllib2.urlopen(mox.IgnoreArg(), mox.IgnoreArg(), timeout=mox.\n IgnoreArg()).AndRaise(urllib2.URLError('url'))\n",
"self._mox.ReplayAll()\n",
"self.assertIsNone(url_helper.UrlOpen('url', max_tries=1))\n",
"self._mox.VerifyAll()\n"
] | [
"def testUrlOpenFailure(self):...\n",
"url_helper.urllib2.urlopen(mox.IgnoreArg(), mox.IgnoreArg(), timeout=mox.\n IgnoreArg()).AndRaise(urllib2.URLError('url'))\n",
"self._mox.ReplayAll()\n",
"self.assertIsNone(url_helper.UrlOpen('url', max_tries=1))\n",
"self._mox.VerifyAll()\n"
] | [
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Expr'",
"Expr'",
"Expr'",
"Expr'"
] |
[
"def __init__(self, VAR_3):...\n",
"\"\"\"docstring\"\"\"\n",
"super(CLASS_0, self).__init__()\n",
"os.makedirs(path.join(VAR_1, DATABASES_FOLDER_NAME), exist_ok=True)\n",
"self.filename = path.join(VAR_1, DATABASES_FOLDER_NAME, VAR_3 + '.db')\n",
"self.lock = Lock()\n",
"VAR_13 = {'type': 0, 'meta': 'str', 'path': 'str', 'mod_time': 0, 'file_id':\n 'str'}\n",
"if VAR_13:\n",
"if path.isfile(self.filename):\n",
"for i in VAR_13.keys():\n",
"self.createTable(VAR_13)\n",
"self._addColumn(i, VAR_13[i])\n"
] | [
"def __init__(self, filename):...\n",
"\"\"\"docstring\"\"\"\n",
"super(FileDB, self).__init__()\n",
"os.makedirs(path.join(SCRIPT_FOLDER, DATABASES_FOLDER_NAME), exist_ok=True)\n",
"self.filename = path.join(SCRIPT_FOLDER, DATABASES_FOLDER_NAME, filename +\n '.db')\n",
"self.lock = Lock()\n",
"initial = {'type': 0, 'meta': 'str', 'path': 'str', 'mod_time': 0,\n 'file_id': 'str'}\n",
"if initial:\n",
"if path.isfile(self.filename):\n",
"for i in initial.keys():\n",
"self.createTable(initial)\n",
"self._addColumn(i, initial[i])\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Expr'",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Condition",
"For",
"Expr'",
"Expr'"
] |
[
"def FUNC_11(VAR_16, VAR_17, VAR_18=True):...\n",
"if isinstance(VAR_16, CLASS_1):\n",
"VAR_16.flags[VAR_17] = VAR_18\n",
"if FUNC_10(VAR_16):\n",
"return VAR_16\n",
"VAR_16 = CLASS_1(VAR_16)\n",
"return [VAR_19(v, VAR_17, VAR_18=flag_value) for v in VAR_16]\n",
"VAR_16.flags[VAR_17] = VAR_18\n",
"return VAR_16\n"
] | [
"def flag(value, flag_type, flag_value=True):...\n",
"if isinstance(value, AnnotatedString):\n",
"value.flags[flag_type] = flag_value\n",
"if not_iterable(value):\n",
"return value\n",
"value = AnnotatedString(value)\n",
"return [flag(v, flag_type, flag_value=flag_value) for v in value]\n",
"value.flags[flag_type] = flag_value\n",
"return value\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Condition",
"Assign'",
"Condition",
"Return'",
"Assign'",
"Return'",
"Assign'",
"Return'"
] |
[
"def __init__(self, VAR_7, VAR_57, VAR_58=True, **VAR_16):...\n",
"self.nav = VAR_57\n",
"self.remember = VAR_58\n",
"VAR_7 = VAR_57.get_param, VAR_7\n",
"CLASS_0.__init__(self, VAR_7, **kw)\n"
] | [
"def __init__(self, param, menu_cls, remember=True, **kw):...\n",
"self.nav = menu_cls\n",
"self.remember = remember\n",
"param = menu_cls.get_param, param\n",
"Validator.__init__(self, param, **kw)\n"
] | [
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'",
"Expr'"
] |
[
"def FUNC_4(self, VAR_10, VAR_9):...\n",
"VAR_0.info('Updating table c{} in the MySQL database'.format(VAR_9))\n",
"VAR_0.info('Requesting for MySQL connection')\n",
"VAR_11 = self.get_connection()\n",
"VAR_18 = VAR_11.cursor()\n",
"for l in VAR_10:\n",
"if not isinstance(l, Listing):\n",
"return 0\n",
"VAR_0.error('TypeError: Expected a Listing instance')\n",
"VAR_0.debug('Generating SQL command')\n",
"VAR_0.error('Skipping this listing')\n",
"VAR_19 = self.gen_sql_insert(l, VAR_9)\n",
"if VAR_19 == -1:\n",
"VAR_0.error('Skipping the listing')\n",
"VAR_0.debug('Executing SQL command')\n",
"VAR_11.rollback()\n",
"VAR_18.execute(VAR_19)\n",
"VAR_0.error('Failed to add a listing to table c{:d}'.format(VAR_9))\n",
"VAR_0.debug('Committing changes to the database')\n",
"VAR_0.error('Rolled back the database changes')\n",
"VAR_11.commit()\n",
"VAR_0.info('Successfully added a listing to table c{:d}'.format(VAR_9))\n"
] | [
"def update_table(self, listings, cat_id):...\n",
"logger.info('Updating table c{} in the MySQL database'.format(cat_id))\n",
"logger.info('Requesting for MySQL connection')\n",
"db = self.get_connection()\n",
"cursor = db.cursor()\n",
"for l in listings:\n",
"if not isinstance(l, Listing):\n",
"return 0\n",
"logger.error('TypeError: Expected a Listing instance')\n",
"logger.debug('Generating SQL command')\n",
"logger.error('Skipping this listing')\n",
"sql = self.gen_sql_insert(l, cat_id)\n",
"if sql == -1:\n",
"logger.error('Skipping the listing')\n",
"logger.debug('Executing SQL command')\n",
"db.rollback()\n",
"cursor.execute(sql)\n",
"logger.error('Failed to add a listing to table c{:d}'.format(cat_id))\n",
"logger.debug('Committing changes to the database')\n",
"logger.error('Rolled back the database changes')\n",
"db.commit()\n",
"logger.info('Successfully added a listing to table c{:d}'.format(cat_id))\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
4,
0,
4,
0,
0,
0,
0,
4,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Expr'",
"Expr'",
"Assign'",
"Assign'",
"For",
"Condition",
"Return'",
"Expr'",
"Expr'",
"Expr'",
"Assign'",
"Condition",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'"
] |
[
"def FUNC_8(VAR_13, VAR_16):...\n",
"VAR_35 = subprocess.Popen([VAR_13], stdin=subprocess.PIPE, VAR_36=\n subprocess.PIPE, VAR_37=subprocess.PIPE)\n",
"VAR_36, VAR_37 = VAR_35.communicate(json.dumps(VAR_16))\n",
"VAR_35.wait()\n",
"if VAR_35.returncode != 0:\n",
"return VAR_36\n"
] | [
"def render_executable(path, config):...\n",
"p = subprocess.Popen([path], stdin=subprocess.PIPE, stdout=subprocess.PIPE,\n stderr=subprocess.PIPE)\n",
"stdout, stderr = p.communicate(json.dumps(config))\n",
"p.wait()\n",
"if p.returncode != 0:\n",
"return stdout\n"
] | [
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Expr'",
"Condition",
"Return'"
] |
[
"def FUNC_20():...\n",
"VAR_61 = Popen(['git', 'ls-tree', '-r', '--name-only', 'HEAD'], stdout=PIPE,\n VAR_37=env)\n",
"VAR_62 = VAR_61.communicate()[0]\n",
"if VAR_61.returncode != 0 or not VAR_62.strip():\n",
"return None\n",
"return [e.strip() for e in VAR_62.splitlines() if os.path.exists(e)]\n"
] | [
"def tracked_files():...\n",
"p = Popen(['git', 'ls-tree', '-r', '--name-only', 'HEAD'], stdout=PIPE, env=env\n )\n",
"out = p.communicate()[0]\n",
"if p.returncode != 0 or not out.strip():\n",
"return None\n",
"return [e.strip() for e in out.splitlines() if os.path.exists(e)]\n"
] | [
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Condition",
"Return'",
"Return'"
] |
[
"def __init__(VAR_39, VAR_40, VAR_41, VAR_42, VAR_48, VAR_43, VAR_44, VAR_45,...\n",
"VAR_39.returncode = None\n",
"VAR_39._out_file = VAR_7.path.join(self.root_dir, 'work',\n 'task_runner_out.json')\n",
"VAR_6 = [sys.executable, bot_main.THIS_FILE, 'task_runner',\n '--swarming-server', VAR_3, '--in-file', VAR_7.path.join(self.root_dir,\n 'work', 'task_runner_in.json'), '--out-file', VAR_39._out_file,\n '--cost-usd-hour', '3600.0', '--start', '100.0', '--min-free-space',\n str(int((os_utilities.get_min_free_space(bot_main.THIS_FILE) + 250.0) *\n 1024 * 1024))]\n",
"self.assertEqual(VAR_6, VAR_40)\n",
"self.assertEqual(True, VAR_41)\n",
"self.assertEqual(self.bot.base_dir, VAR_42)\n",
"self.assertEqual('24', VAR_48['SWARMING_TASK_ID'])\n",
"self.assertTrue(VAR_43)\n",
"self.assertEqual(subprocess42.STDOUT, VAR_44)\n",
"self.assertEqual(subprocess42.PIPE, VAR_45)\n",
"self.assertEqual(sys.platform != 'win32', VAR_46)\n"
] | [
"def __init__(self2, cmd, detached, cwd, env, stdout, stderr, stdin, close_fds):...\n",
"self2.returncode = None\n",
"self2._out_file = os.path.join(self.root_dir, 'work', 'task_runner_out.json')\n",
"expected = [sys.executable, bot_main.THIS_FILE, 'task_runner',\n '--swarming-server', url, '--in-file', os.path.join(self.root_dir,\n 'work', 'task_runner_in.json'), '--out-file', self2._out_file,\n '--cost-usd-hour', '3600.0', '--start', '100.0', '--min-free-space',\n str(int((os_utilities.get_min_free_space(bot_main.THIS_FILE) + 250.0) *\n 1024 * 1024))]\n",
"self.assertEqual(expected, cmd)\n",
"self.assertEqual(True, detached)\n",
"self.assertEqual(self.bot.base_dir, cwd)\n",
"self.assertEqual('24', env['SWARMING_TASK_ID'])\n",
"self.assertTrue(stdout)\n",
"self.assertEqual(subprocess42.STDOUT, stderr)\n",
"self.assertEqual(subprocess42.PIPE, stdin)\n",
"self.assertEqual(sys.platform != 'win32', close_fds)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'"
] |
[
"def FUNC_6(self):...\n",
"return True\n"
] | [
"def Done(self):...\n",
"return True\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
[
"def FUNC_40(self):...\n",
""
] | [
"def _update_sortkey(self):...\n",
""
] | [
0,
0
] | [
"FunctionDef'",
"Condition"
] |
[
"def FUNC_0(VAR_3):...\n",
"if not VAR_3:\n",
"return VAR_3\n",
"if VAR_1.match(VAR_3):\n",
"return VAR_3\n",
"return u'{}`{}`'.format(*VAR_2.match(VAR_3).groups())\n"
] | [
"def escape_col(col):...\n",
"if not col:\n",
"return col\n",
"if ESCAPE_RE.match(col):\n",
"return col\n",
"return u'{}`{}`'.format(*NEGATE_RE.match(col).groups())\n"
] | [
0,
0,
0,
4,
0,
0
] | [
"FunctionDef'",
"Condition",
"Return'",
"Condition",
"Return'",
"Return'"
] |
[
"def FUNC_1(VAR_0=None, VAR_1=None):...\n",
"\"\"\"docstring\"\"\"\n",
"if VAR_0 is None:\n",
"VAR_0 = active_view()\n",
"if not get_settings(VAR_0, 'anaconda_go_linting', True):\n",
"return\n",
"if VAR_0.file_name() in anaconda_sublime.ANACONDA['DISABLED']:\n",
"anaconda_sublime.erase_lint_marks(VAR_0)\n",
"VAR_2 = FUNC_4(VAR_0)\n",
"return\n",
"VAR_3 = {'vid': VAR_0.id(), 'code': VAR_0.substr(st3_sublime.Region(0,\n VAR_0.size())), 'settings': VAR_2, 'filepath': VAR_0.file_name(),\n 'method': 'fast_lint', 'handler': 'anaGonda', 'go_env': {'GOROOT': go.\n GOROOT, 'GOPATH': go.GOPATH, 'CGO_ENABLED': go.CGO_ENABLED}}\n",
"VAR_4 = partial(anaconda_sublime.parse_results, **dict(code='go'))\n",
"if VAR_1 is None:\n",
"Worker().execute(Callback(on_success=callback), **data)\n",
"Worker().execute(Callback(partial(VAR_1, VAR_4)), **data)\n"
] | [
"def fast_linters(view=None, hook=None):...\n",
"\"\"\"docstring\"\"\"\n",
"if view is None:\n",
"view = active_view()\n",
"if not get_settings(view, 'anaconda_go_linting', True):\n",
"return\n",
"if view.file_name() in anaconda_sublime.ANACONDA['DISABLED']:\n",
"anaconda_sublime.erase_lint_marks(view)\n",
"settings = _get_settings(view)\n",
"return\n",
"data = {'vid': view.id(), 'code': view.substr(st3_sublime.Region(0, view.\n size())), 'settings': settings, 'filepath': view.file_name(), 'method':\n 'fast_lint', 'handler': 'anaGonda', 'go_env': {'GOROOT': go.GOROOT,\n 'GOPATH': go.GOPATH, 'CGO_ENABLED': go.CGO_ENABLED}}\n",
"callback = partial(anaconda_sublime.parse_results, **dict(code='go'))\n",
"if hook is None:\n",
"Worker().execute(Callback(on_success=callback), **data)\n",
"Worker().execute(Callback(partial(hook, callback)), **data)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Condition",
"Assign'",
"Condition",
"Return'",
"Condition",
"Expr'",
"Assign'",
"Return'",
"Assign'",
"Assign'",
"Condition",
"Expr'",
"Expr'"
] |
[
"def FUNC_6(self):...\n",
"if not (VAR_101.user_is_loggedin and VAR_101.site.is_moderator(VAR_101.user\n",
"abort(403, 'forbidden')\n"
] | [
"def run(self):...\n",
"if not (c.user_is_loggedin and c.site.is_moderator(c.user) or c.user_is_admin):\n",
"abort(403, 'forbidden')\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Condition",
"Expr'"
] |
[
"def FUNC_17(self, VAR_14):...\n",
""
] | [
"def is_relation(self, col_name):...\n",
""
] | [
0,
0
] | [
"FunctionDef'",
"Condition"
] |
[
"def __str__(self):...\n",
"VAR_23 = '{type}: {from_to_url}'\n",
"if self.redirect_type in ['prefix', 'page', 'exact']:\n",
"return VAR_23.format(type=self.get_redirect_type_display(), from_to_url=\n self.get_from_to_url_display())\n",
"return ugettext('Redirect: {}'.format(self.get_redirect_type_display()))\n"
] | [
"def __str__(self):...\n",
"redirect_text = '{type}: {from_to_url}'\n",
"if self.redirect_type in ['prefix', 'page', 'exact']:\n",
"return redirect_text.format(type=self.get_redirect_type_display(),\n from_to_url=self.get_from_to_url_display())\n",
"return ugettext('Redirect: {}'.format(self.get_redirect_type_display()))\n"
] | [
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Condition",
"Return'",
"Return'"
] |
[
"def FUNC_0(self, VAR_25, VAR_26):...\n",
"self._logger.info('Got Tribler core error: %s' % VAR_25)\n",
"if VAR_25 == QNetworkReply.ConnectionRefusedError:\n",
"if self.failed_attempts == 40:\n",
"self.failed_attempts += 1\n",
"if VAR_26:\n",
"self.connect_timer = QTimer()\n",
"self.connect_timer.setSingleShot(True)\n",
"self.connect_timer.timeout.connect(self.connect)\n",
"self.connect_timer.start(500)\n"
] | [
"def on_error(self, error, reschedule_on_err):...\n",
"self._logger.info('Got Tribler core error: %s' % error)\n",
"if error == QNetworkReply.ConnectionRefusedError:\n",
"if self.failed_attempts == 40:\n",
"self.failed_attempts += 1\n",
"if reschedule_on_err:\n",
"self.connect_timer = QTimer()\n",
"self.connect_timer.setSingleShot(True)\n",
"self.connect_timer.timeout.connect(self.connect)\n",
"self.connect_timer.start(500)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Expr'",
"Condition",
"Condition",
"AugAssign'",
"Condition",
"Assign'",
"Expr'",
"Expr'",
"Expr'"
] |
[
"def FUNC_38(VAR_65, VAR_66):...\n",
"VAR_79 = ElementTree.parse(VAR_66).getroot()\n",
"if VAR_79.tag != 'plugin':\n",
"return VAR_79.find('name').text\n"
] | [
"def process_info_file(cp_elem, info_file):...\n",
"plugin_info = ElementTree.parse(info_file).getroot()\n",
"if plugin_info.tag != 'plugin':\n",
"return plugin_info.find('name').text\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Condition",
"Return'"
] |
[
"def FUNC_4(self):...\n",
"if self.redirected_to is None:\n",
"return None\n",
"VAR_14 = self.redirected_to.split('?')\n",
"if len(VAR_14) != 2:\n",
"VAR_0.warning(\"Redirected to an unexpected page: `%s'\", self.redirected_to)\n",
"VAR_20 = decrypt_number(VAR_14[-1])\n",
"VAR_0.warning(\"Unable to decrypt submission id from page: `%s'\", self.\n redirected_to)\n",
"return VAR_20\n",
"return None\n",
"return None\n"
] | [
"def get_submission_id(self):...\n",
"if self.redirected_to is None:\n",
"return None\n",
"p = self.redirected_to.split('?')\n",
"if len(p) != 2:\n",
"logger.warning(\"Redirected to an unexpected page: `%s'\", self.redirected_to)\n",
"submission_id = decrypt_number(p[-1])\n",
"logger.warning(\"Unable to decrypt submission id from page: `%s'\", self.\n redirected_to)\n",
"return submission_id\n",
"return None\n",
"return None\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Condition",
"Return'",
"Assign'",
"Condition",
"Expr'",
"Assign'",
"Expr'",
"Return'",
"Return'",
"Return'"
] |
[
"async def FUNC_2():...\n",
""
] | [
"async def is_connect_db():...\n",
""
] | [
0,
0
] | [
"AsyncFunctionDef'",
"Condition"
] |
[
"def FUNC_2(VAR_15, VAR_2, VAR_3='text', VAR_4=bconfig....\n",
"\"\"\"docstring\"\"\"\n",
"VAR_43 = reader.get_cache(VAR_2)\n",
"if not VAR_43:\n",
"reader.set_cache(VAR_2, reader.get_regular_expressions(VAR_2, rebuild=\n rebuild_cache, VAR_7=no_cache))\n",
"VAR_44 = VAR_43[0]\n",
"VAR_43 = reader.get_cache(VAR_2)\n",
"VAR_45 = VAR_43[1]\n",
"VAR_15 = normalizer.cut_references(VAR_15)\n",
"VAR_17 = normalizer.normalize_fulltext('\\n'.join(VAR_15))\n",
"if VAR_6 == 'partial':\n",
"VAR_17 = FUNC_26(VAR_17)\n",
"VAR_22 = None\n",
"if VAR_8:\n",
"VAR_22 = FUNC_6(VAR_44, VAR_45, VAR_17)\n",
"VAR_23 = {}\n",
"if VAR_11:\n",
"VAR_23 = FUNC_5(VAR_17)\n",
"VAR_20 = FUNC_3(VAR_44, VAR_17)\n",
"VAR_21 = FUNC_4(VAR_45, VAR_17, VAR_20)\n",
"if VAR_10:\n",
"VAR_20 = FUNC_22(FUNC_20(VAR_20))\n",
"VAR_20 = FUNC_22(VAR_20)\n",
"VAR_21 = FUNC_20(VAR_21)\n",
"return FUNC_7(VAR_20, VAR_21, VAR_2, VAR_22, VAR_23, VAR_3, VAR_4, VAR_5,\n VAR_10)\n"
] | [
"def get_keywords_from_text(text_lines, taxonomy_name, output_mode='text',...\n",
"\"\"\"docstring\"\"\"\n",
"cache = reader.get_cache(taxonomy_name)\n",
"if not cache:\n",
"reader.set_cache(taxonomy_name, reader.get_regular_expressions(\n taxonomy_name, rebuild=rebuild_cache, no_cache=no_cache))\n",
"_skw = cache[0]\n",
"cache = reader.get_cache(taxonomy_name)\n",
"_ckw = cache[1]\n",
"text_lines = normalizer.cut_references(text_lines)\n",
"fulltext = normalizer.normalize_fulltext('\\n'.join(text_lines))\n",
"if match_mode == 'partial':\n",
"fulltext = _get_partial_text(fulltext)\n",
"author_keywords = None\n",
"if with_author_keywords:\n",
"author_keywords = extract_author_keywords(_skw, _ckw, fulltext)\n",
"acronyms = {}\n",
"if extract_acronyms:\n",
"acronyms = extract_abbreviations(fulltext)\n",
"single_keywords = extract_single_keywords(_skw, fulltext)\n",
"composite_keywords = extract_composite_keywords(_ckw, fulltext, single_keywords\n )\n",
"if only_core_tags:\n",
"single_keywords = clean_before_output(_filter_core_keywors(single_keywords))\n",
"single_keywords = clean_before_output(single_keywords)\n",
"composite_keywords = _filter_core_keywors(composite_keywords)\n",
"return get_keywords_output(single_keywords, composite_keywords,\n taxonomy_name, author_keywords, acronyms, output_mode, output_limit,\n spires, only_core_tags)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Docstring",
"Assign'",
"Condition",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Return'"
] |
[
"def FUNC_20(self, VAR_16):...\n",
"return self.meta.get_field(VAR_16).options\n"
] | [
"def get_table_field_doctype(self, fieldname):...\n",
"return self.meta.get_field(fieldname).options\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
[
"def FUNC_0():...\n",
"VAR_13 = VAR_0.path.dirname(VAR_0.path.abspath(__file__))\n",
"return VAR_0.path.join(VAR_13, 'server/ycmd.py')\n"
] | [
"def _PathToServerScript():...\n",
"dir_of_current_script = os.path.dirname(os.path.abspath(__file__))\n",
"return os.path.join(dir_of_current_script, 'server/ycmd.py')\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Return'"
] |
[
"def FUNC_2(self):...\n",
"VAR_14 = utils.GetUnusedLocalhostPort()\n",
"self._temp_options_filename = options_file.name\n",
"json.dump(dict(self._user_options), options_file)\n",
"options_file.flush()\n",
"VAR_21 = [utils.PathToPythonInterpreter(), FUNC_0(), '--port={0}'.format(\n VAR_14), '--options_file={0}'.format(options_file.name), '--log={0}'.\n format(self._user_options['server_log_level']),\n '--idle_suicide_seconds={0}'.format(VAR_4)]\n",
"if not self._user_options['server_use_vim_stdout']:\n",
"VAR_26 = VAR_0.path.join(utils.PathToTempDir(), 'server_{port}_{std}.log')\n",
"self._server_popen = utils.SafePopen(VAR_21, stdout=PIPE, stderr=PIPE)\n",
"self._server_stdout = VAR_26.format(port=server_port, std='stdout')\n",
"BaseRequest.server_location = 'http://localhost:' + str(VAR_14)\n",
"self._server_stderr = VAR_26.format(port=server_port, std='stderr')\n",
"self._NotifyUserIfServerCrashed()\n",
"VAR_21.append('--stdout={0}'.format(self._server_stdout))\n",
"VAR_21.append('--stderr={0}'.format(self._server_stderr))\n",
"if self._user_options['server_keep_logfiles']:\n",
"VAR_21.append('--keep_logfiles')\n"
] | [
"def _SetupServer(self):...\n",
"server_port = utils.GetUnusedLocalhostPort()\n",
"self._temp_options_filename = options_file.name\n",
"json.dump(dict(self._user_options), options_file)\n",
"options_file.flush()\n",
"args = [utils.PathToPythonInterpreter(), _PathToServerScript(),\n '--port={0}'.format(server_port), '--options_file={0}'.format(\n options_file.name), '--log={0}'.format(self._user_options[\n 'server_log_level']), '--idle_suicide_seconds={0}'.format(\n SERVER_IDLE_SUICIDE_SECONDS)]\n",
"if not self._user_options['server_use_vim_stdout']:\n",
"filename_format = os.path.join(utils.PathToTempDir(), 'server_{port}_{std}.log'\n )\n",
"self._server_popen = utils.SafePopen(args, stdout=PIPE, stderr=PIPE)\n",
"self._server_stdout = filename_format.format(port=server_port, std='stdout')\n",
"BaseRequest.server_location = 'http://localhost:' + str(server_port)\n",
"self._server_stderr = filename_format.format(port=server_port, std='stderr')\n",
"self._NotifyUserIfServerCrashed()\n",
"args.append('--stdout={0}'.format(self._server_stdout))\n",
"args.append('--stderr={0}'.format(self._server_stderr))\n",
"if self._user_options['server_keep_logfiles']:\n",
"args.append('--keep_logfiles')\n"
] | [
0,
0,
7,
7,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Condition",
"Expr'"
] |
[
"def FUNC_7(self):...\n",
"VAR_6 = self.client.get('/api/apps')\n",
"VAR_8 = VAR_6.data['results'][0]['id']\n",
"self.assertTrue(self.client.login(VAR_1='autotest-2', VAR_2='password'))\n",
"VAR_6 = self.client.get('/api/apps/{}/perms'.format(VAR_8), content_type=\n 'application/json')\n",
"self.assertEqual(VAR_6.status_code, 403)\n"
] | [
"def test_list_errors(self):...\n",
"response = self.client.get('/api/apps')\n",
"app_id = response.data['results'][0]['id']\n",
"self.assertTrue(self.client.login(username='autotest-2', password='password'))\n",
"response = self.client.get('/api/apps/{}/perms'.format(app_id),\n content_type='application/json')\n",
"self.assertEqual(response.status_code, 403)\n"
] | [
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Expr'"
] |
[
"def FUNC_32(self, VAR_81):...\n",
"def FUNC_39(VAR_101):...\n",
"VAR_101.shellcmd = VAR_81\n",
"return VAR_101\n"
] | [
"def shellcmd(self, cmd):...\n",
"def decorate(ruleinfo):...\n",
"ruleinfo.shellcmd = cmd\n",
"return ruleinfo\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"FunctionDef'",
"Assign'",
"Return'"
] |
[
"def FUNC_1(self):...\n",
"VAR_4 = CLASS_2(id=0)\n",
"VAR_5 = EntityCache(entities=[])\n",
"VAR_5.add(VAR_4)\n",
"self.assert_true(VAR_5.get_by_id(VAR_4.id) is VAR_4)\n",
"self.assert_true(VAR_5.has_id(VAR_4.id))\n",
"self.assert_true(VAR_5.get_by_slug(VAR_4.slug) is VAR_4)\n",
"self.assert_true(VAR_5.has_slug(VAR_4.slug))\n",
"VAR_6 = CLASS_2(id=0)\n",
"VAR_7 = 'FROBNIC'\n",
"VAR_6.text = VAR_7\n",
"VAR_5.replace(VAR_6)\n",
"self.assert_equal(VAR_5.get_by_id(VAR_4.id).text, VAR_7)\n",
"self.assert_equal(VAR_5.get_all(), [VAR_4])\n",
"self.assert_equal(list(VAR_5.retrieve()), [VAR_4])\n",
"VAR_5.remove(VAR_4)\n",
"self.assert_is_none(VAR_5.get_by_id(VAR_4.id))\n",
"self.assert_is_none(VAR_5.get_by_slug(VAR_4.slug))\n"
] | [
"def test_basics(self):...\n",
"ent = MyEntity(id=0)\n",
"cache = EntityCache(entities=[])\n",
"cache.add(ent)\n",
"self.assert_true(cache.get_by_id(ent.id) is ent)\n",
"self.assert_true(cache.has_id(ent.id))\n",
"self.assert_true(cache.get_by_slug(ent.slug) is ent)\n",
"self.assert_true(cache.has_slug(ent.slug))\n",
"ent1 = MyEntity(id=0)\n",
"txt = 'FROBNIC'\n",
"ent1.text = txt\n",
"cache.replace(ent1)\n",
"self.assert_equal(cache.get_by_id(ent.id).text, txt)\n",
"self.assert_equal(cache.get_all(), [ent])\n",
"self.assert_equal(list(cache.retrieve()), [ent])\n",
"cache.remove(ent)\n",
"self.assert_is_none(cache.get_by_id(ent.id))\n",
"self.assert_is_none(cache.get_by_slug(ent.slug))\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'"
] |
[
"def FUNC_12(VAR_24):...\n",
"for user, VAR_20, id1, id2 in VAR_23.findall(VAR_24):\n",
"VAR_21 = id1 + id2\n",
"FUNC_9(VAR_20, VAR_21, user)\n"
] | [
"def rtfu(urls):...\n",
"for user, domain, id1, id2 in r_di.findall(urls):\n",
"id_ = id1 + id2\n",
"remove_target(domain, id_, user)\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"For",
"Assign'",
"Expr'"
] |
[
"def FUNC_0(self, VAR_0='crimemap'):...\n",
"return pymysql.connect(host='localhost', user=dbconfig.db_user, password=\n dbconfig.db_password, db=database)\n"
] | [
"def connect(self, database='crimemap'):...\n",
"return pymysql.connect(host='localhost', user=dbconfig.db_user, password=\n dbconfig.db_password, db=database)\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
[
"def FUNC_7(self):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_1 = self.bindings\n",
"VAR_7 = VAR_1['TEST_APP_COMPONENT_NAME']\n",
"VAR_15 = '{app}-{stack}-v000'.format(app=self.TEST_APP, stack=bindings[\n 'TEST_STACK'])\n",
"VAR_8 = VAR_1['TEST_AWS_REGION']\n",
"VAR_9 = [VAR_8 + 'a', VAR_8 + 'b']\n",
"VAR_12 = self.agent.make_json_payload_from_kwargs(job=[{'type':\n 'createServerGroup', 'cloudProvider': 'aws', 'application': self.\n TEST_APP, 'credentials': bindings['AWS_CREDENTIALS'], 'strategy': '',\n 'capacity': {'min': 2, 'max': 2, 'desired': 2},\n 'targetHealthyDeployPercentage': 100, 'loadBalancers': [\n load_balancer_name], 'cooldown': 8, 'healthCheckType': 'EC2',\n 'healthCheckGracePeriod': 40, 'instanceMonitoring': False,\n 'ebsOptimized': False, 'iamRole': bindings['AWS_IAM_ROLE'],\n 'terminationPolicies': ['Default'], 'availabilityZones': {region:\n avail_zones}, 'keyPair': bindings['AWS_CREDENTIALS'] + '-keypair',\n 'suspendedProcesses': [], 'subnetType': 'internal (defaultvpc)',\n 'securityGroups': [bindings['TEST_AWS_SECURITY_GROUP_ID']],\n 'virtualizationType': 'paravirtual', 'stack': bindings['TEST_STACK'],\n 'freeFormDetails': '', 'amiName': bindings['TEST_AWS_AMI'],\n 'instanceType': 'm1.small', 'useSourceCapacity': False, 'account':\n bindings['AWS_CREDENTIALS'], 'user': '[anonymous]'}], description=\n 'Create Server Group in ' + group_name, application=self.TEST_APP)\n",
"VAR_13 = aws.AwsContractBuilder(self.aws_observer)\n",
"VAR_13.new_clause_builder('Auto Server Group Added', retryable_for_secs=30\n ).collect_resources('autoscaling', 'describe-auto-scaling-groups', args\n =['--auto-scaling-group-names', group_name]).contains_path_value(\n 'AutoScalingGroups', {'MaxSize': 2})\n",
"return st.OperationContract(self.new_post_operation(title=\n 'create_server_group', data=payload, path='tasks'), VAR_6=builder.build())\n"
] | [
"def create_server_group(self):...\n",
"\"\"\"docstring\"\"\"\n",
"bindings = self.bindings\n",
"load_balancer_name = bindings['TEST_APP_COMPONENT_NAME']\n",
"group_name = '{app}-{stack}-v000'.format(app=self.TEST_APP, stack=bindings[\n 'TEST_STACK'])\n",
"region = bindings['TEST_AWS_REGION']\n",
"avail_zones = [region + 'a', region + 'b']\n",
"payload = self.agent.make_json_payload_from_kwargs(job=[{'type':\n 'createServerGroup', 'cloudProvider': 'aws', 'application': self.\n TEST_APP, 'credentials': bindings['AWS_CREDENTIALS'], 'strategy': '',\n 'capacity': {'min': 2, 'max': 2, 'desired': 2},\n 'targetHealthyDeployPercentage': 100, 'loadBalancers': [\n load_balancer_name], 'cooldown': 8, 'healthCheckType': 'EC2',\n 'healthCheckGracePeriod': 40, 'instanceMonitoring': False,\n 'ebsOptimized': False, 'iamRole': bindings['AWS_IAM_ROLE'],\n 'terminationPolicies': ['Default'], 'availabilityZones': {region:\n avail_zones}, 'keyPair': bindings['AWS_CREDENTIALS'] + '-keypair',\n 'suspendedProcesses': [], 'subnetType': 'internal (defaultvpc)',\n 'securityGroups': [bindings['TEST_AWS_SECURITY_GROUP_ID']],\n 'virtualizationType': 'paravirtual', 'stack': bindings['TEST_STACK'],\n 'freeFormDetails': '', 'amiName': bindings['TEST_AWS_AMI'],\n 'instanceType': 'm1.small', 'useSourceCapacity': False, 'account':\n bindings['AWS_CREDENTIALS'], 'user': '[anonymous]'}], description=\n 'Create Server Group in ' + group_name, application=self.TEST_APP)\n",
"builder = aws.AwsContractBuilder(self.aws_observer)\n",
"builder.new_clause_builder('Auto Server Group Added', retryable_for_secs=30\n ).collect_resources('autoscaling', 'describe-auto-scaling-groups', args\n =['--auto-scaling-group-names', group_name]).contains_path_value(\n 'AutoScalingGroups', {'MaxSize': 2})\n",
"return st.OperationContract(self.new_post_operation(title=\n 'create_server_group', data=payload, path='tasks'), contract=builder.\n build())\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Return'"
] |
[
"from os.path import join\n",
"from datetime import datetime\n",
"from flask import request, redirect, url_for, render_template, flash, g, current_app\n",
"from werkzeug.utils import secure_filename\n",
"from flask_security import login_required, current_user\n",
"from benwaonline.database import db\n",
"from benwaonline.models import Post, Tag, Comment, Preview, Image\n",
"from benwaonline.gallery import gallery\n",
"from benwaonline.gallery.forms import CommentForm, PostForm\n",
"@gallery.before_request...\n",
"g.user = current_user\n",
"@gallery.route('/gallery/')...\n",
"if VAR_0 == 'all':\n",
"VAR_9 = Post.query.all()\n",
"VAR_10 = VAR_0.split(' ')\n",
"VAR_0 = Tag.query.all()\n",
"VAR_9 = []\n",
"return render_template('gallery.html', VAR_9=posts, VAR_0=tags)\n",
"for s in VAR_10:\n",
"VAR_18 = Post.query.filter(Post.tags.any(name=s))\n",
"VAR_9.extend(VAR_18)\n"
] | [
"from os.path import join\n",
"from datetime import datetime\n",
"from flask import request, redirect, url_for, render_template, flash, g, current_app\n",
"from werkzeug.utils import secure_filename\n",
"from flask_security import login_required, current_user\n",
"from benwaonline.database import db\n",
"from benwaonline.models import Post, Tag, Comment, Preview, Image\n",
"from benwaonline.gallery import gallery\n",
"from benwaonline.gallery.forms import CommentForm, PostForm\n",
"@gallery.before_request...\n",
"g.user = current_user\n",
"@gallery.route('/gallery/')...\n",
"if tags == 'all':\n",
"posts = Post.query.all()\n",
"split = tags.split(' ')\n",
"tags = Tag.query.all()\n",
"posts = []\n",
"return render_template('gallery.html', posts=posts, tags=tags)\n",
"for s in split:\n",
"results = Post.query.filter(Post.tags.any(name=s))\n",
"posts.extend(results)\n"
] | [
0,
0,
6,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"For",
"Assign'",
"Condition",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Return'",
"For",
"Assign'",
"Expr'"
] |
[
"@VAR_0.route('/list')...\n",
"if VAR_3 is None:\n",
"VAR_3 = request.args.get('start', None, float)\n",
"if VAR_4 is None:\n",
"VAR_4 = request.args.get('end', None, float)\n",
"if VAR_5 is None:\n",
"VAR_5 = request.args.get('limit', 100, int)\n",
"if VAR_6 is None:\n",
"VAR_6 = request.args.get('format', 'plain', str)\n",
"if VAR_7 is None:\n",
"VAR_6 = request.args.get('download', 'no')\n",
"if VAR_8 is None:\n",
"VAR_8 = request.args.get('degree', None, int)\n",
"if VAR_5 > 1000:\n",
"VAR_5 = 1000\n",
"if VAR_5 < 0:\n",
"VAR_5 = 100\n",
"if VAR_3 is None and VAR_4 is None:\n",
"VAR_4 = 1000\n",
"VAR_5 = int(VAR_5)\n",
"VAR_11 = 'WHERE 1=1 '\n",
"if VAR_4 is not None:\n",
"VAR_4 = str(VAR_4)\n",
"if VAR_3 is None:\n",
"if '.' in VAR_4:\n",
"VAR_11 += ' AND zero <= ' + VAR_4\n",
"if VAR_4 is None:\n",
"VAR_4 = VAR_4 + '999'\n",
"if VAR_8 is not None and VAR_8 != '':\n",
"VAR_3 = float(VAR_3)\n",
"VAR_11 += ' AND zero >= {} AND zero <= {}'.format(VAR_3, VAR_4)\n",
"VAR_11 += ' AND degree = ' + str(VAR_8)\n",
"if VAR_4 is None:\n",
"VAR_11 += ' AND zero >= ' + str(VAR_3)\n",
"VAR_14 = (\n 'SELECT * FROM (SELECT * FROM zeros {} ORDER BY zero ASC LIMIT {}) ORDER BY zero DESC'\n .format(VAR_11, VAR_5))\n",
"VAR_14 = 'SELECT * FROM zeros {} ORDER BY zero DESC LIMIT {}'.format(VAR_11,\n VAR_5)\n",
"VAR_12 = sqlite3.connect(VAR_2 + 'first_zeros.db').cursor()\n",
"VAR_12.execute(VAR_14)\n",
"VAR_13 = flask.Response(' '.join([str(x) for x in row]) + '\\n' for row in\n VAR_12)\n",
"VAR_13.headers['content-type'] = 'text/plain'\n",
"if VAR_7 == 'yes':\n",
"VAR_13.headers['content-disposition'] = 'attachment; filename=zetazeros'\n",
"return VAR_13\n"
] | [
"@FirstZeros.route('/list')...\n",
"if start is None:\n",
"start = request.args.get('start', None, float)\n",
"if end is None:\n",
"end = request.args.get('end', None, float)\n",
"if limit is None:\n",
"limit = request.args.get('limit', 100, int)\n",
"if fmt is None:\n",
"fmt = request.args.get('format', 'plain', str)\n",
"if download is None:\n",
"fmt = request.args.get('download', 'no')\n",
"if degree is None:\n",
"degree = request.args.get('degree', None, int)\n",
"if limit > 1000:\n",
"limit = 1000\n",
"if limit < 0:\n",
"limit = 100\n",
"if start is None and end is None:\n",
"end = 1000\n",
"limit = int(limit)\n",
"where_clause = 'WHERE 1=1 '\n",
"if end is not None:\n",
"end = str(end)\n",
"if start is None:\n",
"if '.' in end:\n",
"where_clause += ' AND zero <= ' + end\n",
"if end is None:\n",
"end = end + '999'\n",
"if degree is not None and degree != '':\n",
"start = float(start)\n",
"where_clause += ' AND zero >= {} AND zero <= {}'.format(start, end)\n",
"where_clause += ' AND degree = ' + str(degree)\n",
"if end is None:\n",
"where_clause += ' AND zero >= ' + str(start)\n",
"query = (\n 'SELECT * FROM (SELECT * FROM zeros {} ORDER BY zero ASC LIMIT {}) ORDER BY zero DESC'\n .format(where_clause, limit))\n",
"query = 'SELECT * FROM zeros {} ORDER BY zero DESC LIMIT {}'.format(\n where_clause, limit)\n",
"c = sqlite3.connect(data_location + 'first_zeros.db').cursor()\n",
"c.execute(query)\n",
"response = flask.Response(' '.join([str(x) for x in row]) + '\\n' for row in c)\n",
"response.headers['content-type'] = 'text/plain'\n",
"if download == 'yes':\n",
"response.headers['content-disposition'] = 'attachment; filename=zetazeros'\n",
"return response\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
4,
0,
4,
4,
0,
4,
0,
0,
4,
4,
0,
4,
0,
0,
0,
4,
0,
0,
0,
0,
0
] | [
"Condition",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Condition",
"AugAssign'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"AugAssign'",
"AugAssign'",
"Condition",
"AugAssign'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Return'"
] |
[
"def FUNC_5(self):...\n",
"self._check_duplicate_key(VAR_0)\n"
] | [
"def test_rsa_duplicate_key(self):...\n",
"self._check_duplicate_key(RSA_PUBKEY)\n"
] | [
0,
0
] | [
"FunctionDef'",
"Expr'"
] |
[
"def FUNC_33(VAR_57, VAR_22):...\n",
"if not len(VAR_57):\n",
"VAR_56.append('/%s' % ('/'.join(VAR_22),))\n",
"if VAR_57[0] == '.':\n",
"FUNC_33(VAR_57[1:], VAR_22)\n",
"if VAR_57[0] == '..':\n",
"FUNC_33(VAR_57[1:], VAR_22[:-1])\n",
"VAR_65 = [x[VAR_1] for x in self.get_path('/'.join(VAR_22))]\n",
"VAR_74 = [x for x in VAR_65 if fnmatch.fnmatchcase(x, VAR_57[0])]\n",
"for match in VAR_74:\n",
"FUNC_33(VAR_57[1:], VAR_22 + [match])\n"
] | [
"def foo(p, cwd):...\n",
"if not len(p):\n",
"found.append('/%s' % ('/'.join(cwd),))\n",
"if p[0] == '.':\n",
"foo(p[1:], cwd)\n",
"if p[0] == '..':\n",
"foo(p[1:], cwd[:-1])\n",
"names = [x[A_NAME] for x in self.get_path('/'.join(cwd))]\n",
"matches = [x for x in names if fnmatch.fnmatchcase(x, p[0])]\n",
"for match in matches:\n",
"foo(p[1:], cwd + [match])\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Condition",
"Expr'",
"Condition",
"Expr'",
"Condition",
"Expr'",
"Assign'",
"Assign'",
"For",
"Expr'"
] |
[
"def FUNC_13(self):...\n",
"VAR_4 = CachedContent(self.course_instance)\n",
"def FUNC_17(VAR_5):...\n",
"VAR_11 = VAR_5['children']\n",
"if VAR_11:\n",
"return itertools.chain.from_iterable([FUNC_17(child) for child in VAR_11])\n",
"return VAR_5['id'],\n"
] | [
"def __get_exercises(self):...\n",
"content = CachedContent(self.course_instance)\n",
"def get_descendant_ids(node):...\n",
"children = node['children']\n",
"if children:\n",
"return itertools.chain.from_iterable([get_descendant_ids(child) for child in\n children])\n",
"return node['id'],\n"
] | [
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"FunctionDef'",
"Assign'",
"Condition",
"Return'",
"Return'"
] |
[
"def FUNC_3(self, VAR_8):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_19 = self._cliq_get_cluster_info(VAR_8)\n",
"VAR_20 = []\n",
"for vip in VAR_19.findall('response/cluster/vip'):\n",
"VAR_20.append(vip.attrib.get('ipAddress'))\n",
"if len(VAR_20) == 1:\n",
"return VAR_20[0]\n",
"VAR_21 = etree.tostring(VAR_19)\n",
"VAR_22 = _(\n 'Unexpected number of virtual ips for cluster %(cluster_name)s. Result=%(_xml)s'\n ) % {'cluster_name': VAR_8, '_xml': VAR_21}\n"
] | [
"def _cliq_get_cluster_vip(self, cluster_name):...\n",
"\"\"\"docstring\"\"\"\n",
"cluster_xml = self._cliq_get_cluster_info(cluster_name)\n",
"vips = []\n",
"for vip in cluster_xml.findall('response/cluster/vip'):\n",
"vips.append(vip.attrib.get('ipAddress'))\n",
"if len(vips) == 1:\n",
"return vips[0]\n",
"_xml = etree.tostring(cluster_xml)\n",
"msg = _(\n 'Unexpected number of virtual ips for cluster %(cluster_name)s. Result=%(_xml)s'\n ) % {'cluster_name': cluster_name, '_xml': _xml}\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Assign'",
"For",
"Expr'",
"Condition",
"Return'",
"Assign'",
"Assign'"
] |
[
"def FUNC_24(self, VAR_25):...\n",
"for VAR_24 in self.databases_to_save:\n",
"VAR_24.database_manager.save_database()\n",
"self.quit_dialog.destroy()\n",
"self.application.quit()\n"
] | [
"def on_quit_button_clicked(self, button):...\n",
"for db in self.databases_to_save:\n",
"db.database_manager.save_database()\n",
"self.quit_dialog.destroy()\n",
"self.application.quit()\n"
] | [
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"For",
"Expr'",
"Expr'",
"Expr'"
] |
[
"def FUNC_0(self):...\n",
"VAR_1, VAR_2 = 'firstuser', 'password'\n",
"VAR_3 = '[email protected]'\n",
"VAR_4 = {'username': VAR_1, 'password': VAR_2, 'email': VAR_3}\n",
"VAR_5 = '/api/auth/register'\n",
"VAR_6 = self.client.post(VAR_5, json.dumps(VAR_4), content_type=\n 'application/json')\n",
"self.assertEqual(VAR_6.status_code, 201)\n",
"self.assertTrue(VAR_6.data['is_superuser'])\n",
"VAR_1, VAR_2 = 'seconduser', 'password'\n",
"VAR_3 = '[email protected]'\n",
"VAR_4 = {'username': VAR_1, 'password': VAR_2, 'email': VAR_3}\n",
"VAR_5 = '/api/auth/register'\n",
"VAR_6 = self.client.post(VAR_5, json.dumps(VAR_4), content_type=\n 'application/json')\n",
"self.assertEqual(VAR_6.status_code, 201)\n",
"self.assertFalse(VAR_6.data['is_superuser'])\n"
] | [
"def test_first_signup(self):...\n",
"username, password = 'firstuser', 'password'\n",
"email = '[email protected]'\n",
"submit = {'username': username, 'password': password, 'email': email}\n",
"url = '/api/auth/register'\n",
"response = self.client.post(url, json.dumps(submit), content_type=\n 'application/json')\n",
"self.assertEqual(response.status_code, 201)\n",
"self.assertTrue(response.data['is_superuser'])\n",
"username, password = 'seconduser', 'password'\n",
"email = '[email protected]'\n",
"submit = {'username': username, 'password': password, 'email': email}\n",
"url = '/api/auth/register'\n",
"response = self.client.post(url, json.dumps(submit), content_type=\n 'application/json')\n",
"self.assertEqual(response.status_code, 201)\n",
"self.assertFalse(response.data['is_superuser'])\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Expr'"
] |
[
"def __init__(self, VAR_7=True, VAR_8=False):...\n",
"\"\"\"docstring\"\"\"\n",
"super().__init__(description='ParlAI parser.', allow_abbrev=False,\n conflict_handler='resolve')\n",
"self.register('type', 'bool', FUNC_0)\n",
"self.register('type', 'class', FUNC_1)\n",
"self.parlai_home = VAR_20.path.dirname(VAR_20.path.dirname(VAR_20.path.\n dirname(VAR_20.path.realpath(__file__))))\n",
"VAR_20.environ['PARLAI_HOME'] = self.parlai_home\n",
"self.add_arg = self.add_argument\n",
"self.cli_args = sys.argv\n",
"self.overridable = {}\n",
"if VAR_7:\n",
"self.add_parlai_args()\n",
"if VAR_8:\n",
"self.add_model_args()\n"
] | [
"def __init__(self, add_parlai_args=True, add_model_args=False):...\n",
"\"\"\"docstring\"\"\"\n",
"super().__init__(description='ParlAI parser.', allow_abbrev=False,\n conflict_handler='resolve')\n",
"self.register('type', 'bool', str2bool)\n",
"self.register('type', 'class', str2class)\n",
"self.parlai_home = os.path.dirname(os.path.dirname(os.path.dirname(os.path.\n realpath(__file__))))\n",
"os.environ['PARLAI_HOME'] = self.parlai_home\n",
"self.add_arg = self.add_argument\n",
"self.cli_args = sys.argv\n",
"self.overridable = {}\n",
"if add_parlai_args:\n",
"self.add_parlai_args()\n",
"if add_model_args:\n",
"self.add_model_args()\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Expr'",
"Expr'",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Expr'",
"Condition",
"Expr'"
] |
[
"def FUNC_22(self, VAR_28):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_58 = VAR_28,\n",
"VAR_36 = ''\n",
"if self.execute('SELECT name FROM history WHERE nzo_id=?', VAR_58):\n",
"return VAR_36\n",
"VAR_36 = self.c.fetchone().get('name')\n"
] | [
"def get_name(self, nzo_id):...\n",
"\"\"\"docstring\"\"\"\n",
"t = nzo_id,\n",
"name = ''\n",
"if self.execute('SELECT name FROM history WHERE nzo_id=?', t):\n",
"return name\n",
"name = self.c.fetchone().get('name')\n"
] | [
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Assign'",
"Condition",
"Return'",
"Assign'"
] |
[
"def FUNC_13(self, VAR_16):...\n",
"\"\"\"docstring\"\"\"\n",
"self._check_volume(VAR_16)\n",
"VAR_0.warn(_('Volume %s was not found while trying to delete it'), VAR_16[\n 'name'])\n",
"self._eql_execute('volume', 'select', VAR_16['name'], 'offline')\n",
"VAR_0.error(_('Failed to delete volume %s'), VAR_16['name'])\n",
"self._eql_execute('volume', 'delete', VAR_16['name'])\n"
] | [
"def delete_volume(self, volume):...\n",
"\"\"\"docstring\"\"\"\n",
"self._check_volume(volume)\n",
"LOG.warn(_('Volume %s was not found while trying to delete it'), volume['name']\n )\n",
"self._eql_execute('volume', 'select', volume['name'], 'offline')\n",
"LOG.error(_('Failed to delete volume %s'), volume['name'])\n",
"self._eql_execute('volume', 'delete', volume['name'])\n"
] | [
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'"
] |
[
"@property...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_15 = set()\n",
"for VAR_28, VAR_27 in zip(self.output, self.rule.output):\n",
"if VAR_28 in self.dynamic_output:\n",
"VAR_16 = defaultdict(list)\n",
"for VAR_28, w in self.expand_dynamic(VAR_27, VAR_13=self.wildcards, VAR_14=\n",
"for combination in VAR_15:\n",
"VAR_15.add(tuple(w.items()))\n",
"for name, value in combination:\n",
"return VAR_16\n",
"VAR_16[name].append(value)\n"
] | [
"@property...\n",
"\"\"\"docstring\"\"\"\n",
"combinations = set()\n",
"for f, f_ in zip(self.output, self.rule.output):\n",
"if f in self.dynamic_output:\n",
"wildcards = defaultdict(list)\n",
"for f, w in self.expand_dynamic(f_, restriction=self.wildcards, omit_value=\n",
"for combination in combinations:\n",
"combinations.add(tuple(w.items()))\n",
"for name, value in combination:\n",
"return wildcards\n",
"wildcards[name].append(value)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Docstring",
"Assign'",
"For",
"Condition",
"Assign'",
"For",
"For",
"Expr'",
"For",
"Return'",
"Expr'"
] |
[
"def FUNC_19(self, VAR_38):...\n",
"\"\"\"docstring\"\"\"\n",
"return self._parse_conditions(self.OR_CONDITION, VAR_38)\n"
] | [
"def _parse_or(self, data):...\n",
"\"\"\"docstring\"\"\"\n",
"return self._parse_conditions(self.OR_CONDITION, data)\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Return'"
] |
[
"def FUNC_42(self):...\n",
"self.cursor.execute('create table t1(c1 int identity(1, 1), c2 varchar(50))')\n",
"for i in range(3):\n",
"self.cursor.execute('insert into t1(c2) values (?)', 'string%s' % i)\n",
"self.cursor.execute('create view t2 as select * from t1')\n",
"self.cursor.execute('select * from t2')\n",
"VAR_29 = self.cursor.fetchall()\n",
"self.assertTrue(VAR_29 is not None)\n",
"self.assertTrue(VAR_23(VAR_29) == 3)\n"
] | [
"def test_view_select(self):...\n",
"self.cursor.execute('create table t1(c1 int identity(1, 1), c2 varchar(50))')\n",
"for i in range(3):\n",
"self.cursor.execute('insert into t1(c2) values (?)', 'string%s' % i)\n",
"self.cursor.execute('create view t2 as select * from t1')\n",
"self.cursor.execute('select * from t2')\n",
"rows = self.cursor.fetchall()\n",
"self.assertTrue(rows is not None)\n",
"self.assertTrue(len(rows) == 3)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Expr'",
"For",
"Expr'",
"Expr'",
"Expr'",
"Assign'",
"Expr'",
"Expr'"
] |
[
"def FUNC_18(self):...\n",
"self.run_test_case(self.scenario.delete_load_balancer())\n"
] | [
"def test_z_delete_load_balancer(self):...\n",
"self.run_test_case(self.scenario.delete_load_balancer())\n"
] | [
0,
0
] | [
"FunctionDef'",
"Expr'"
] |
[
"@VAR_0.route('/upload', methods=['POST'])...\n",
"VAR_8 = request.files['file']\n",
"VAR_9 = MySQL.get_connection(VAR_3)\n",
"if VAR_8:\n",
"VAR_16 = os.path.splitext(VAR_8.filename)[0] + str(int(time.time())) + '.log'\n",
"VAR_9.close()\n",
"VAR_17 = logging.FileHandler(os.path.join(VAR_1, VAR_16), 'w')\n",
"Config.setup_logging(VAR_17)\n",
"VAR_18 = os.path.join(VAR_2, VAR_8.filename)\n",
"VAR_8.save(VAR_18)\n",
"Importer.run(VAR_18, VAR_9)\n",
"logger.removeHandler(VAR_17)\n",
"VAR_17.close()\n",
"return jsonify({'name': VAR_8.filename, 'log': VAR_16})\n"
] | [
"@app.route('/upload', methods=['POST'])...\n",
"datafile = request.files['file']\n",
"c = MySQL.get_connection(DATABASE)\n",
"if datafile:\n",
"logfile = os.path.splitext(datafile.filename)[0] + str(int(time.time())\n ) + '.log'\n",
"c.close()\n",
"f = logging.FileHandler(os.path.join(LOG_DIR, logfile), 'w')\n",
"Config.setup_logging(f)\n",
"filepath = os.path.join(UPLOADS_DIR, datafile.filename)\n",
"datafile.save(filepath)\n",
"Importer.run(filepath, c)\n",
"logger.removeHandler(f)\n",
"f.close()\n",
"return jsonify({'name': datafile.filename, 'log': logfile})\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Expr'",
"Assign'",
"Expr'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Return'"
] |
[
"def __bool__(self):...\n",
"return False\n"
] | [
"def __bool__(self):...\n",
"return False\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
[
"def FUNC_0(VAR_0=None):...\n",
"VAR_1 = FUNC_1()\n",
"VAR_2 = FUNC_2(VAR_0)\n",
"return VAR_1, VAR_2\n"
] | [
"def execute(filters=None):...\n",
"columns = get_columns()\n",
"data = get_data(filters)\n",
"return columns, data\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Return'"
] |
[
"@property...\n",
"\"\"\"docstring\"\"\"\n",
"return self.config.get(self.section, 'cert_chain')\n"
] | [
"@property...\n",
"\"\"\"docstring\"\"\"\n",
"return self.config.get(self.section, 'cert_chain')\n"
] | [
0,
0,
0
] | [
"Condition",
"Docstring",
"Return'"
] |
[
"def FUNC_1(self, VAR_0):...\n",
"VAR_19 = dict(self.query_params).copy()\n",
"for VAR_26, VAR_4 in VAR_19.items():\n",
"if type(VAR_4) == list:\n",
"VAR_19.update(VAR_0)\n",
"VAR_19[VAR_26] = VAR_4[0]\n",
"for VAR_26, VAR_4 in VAR_19.copy().items():\n",
"if VAR_4 is None:\n",
"return '?{}'.format(urlencode(VAR_19))\n"
] | [
"def get_url_with_query_params(self, new_query_dict):...\n",
"query = dict(self.query_params).copy()\n",
"for key, value in query.items():\n",
"if type(value) == list:\n",
"query.update(new_query_dict)\n",
"query[key] = value[0]\n",
"for key, value in query.copy().items():\n",
"if value is None:\n",
"return '?{}'.format(urlencode(query))\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"For",
"Condition",
"Expr'",
"Assign'",
"For",
"Condition",
"Return'"
] |
[
"def __init__(self):...\n",
"self.__cache_map = defaultdict(CLASS_0)\n"
] | [
"def __init__(self):...\n",
"self.__cache_map = defaultdict(EntityCache)\n"
] | [
0,
0
] | [
"FunctionDef'",
"Assign'"
] |
[
"def FUNC_0(self):...\n",
"\"\"\"docstring\"\"\"\n",
"\"\"\"string\"\"\"\n",
"VAR_6 = request.args.get('username')\n",
"VAR_9 = request.args.get('password')\n",
"if VAR_6.isalnum():\n",
"VAR_16 = db_interac.authenticate(VAR_6, VAR_9)\n",
"VAR_4['error'] = 'username must be alphanumeric'\n",
"VAR_4 = {}\n",
"VAR_4['userId'] = None\n",
"if not VAR_16[0]:\n",
"VAR_4['firstName'] = None\n",
"VAR_4['error'] = 'user could not be authenticated'\n",
"VAR_4['userId'] = VAR_16[1]\n",
"VAR_4['lastName'] = None\n",
"return VAR_4, 401\n",
"VAR_4['firstName'] = VAR_16[2]\n",
"return VAR_4, 401\n",
"VAR_4['lastName'] = VAR_16[3]\n",
"return VAR_4, 200\n"
] | [
"def get(self):...\n",
"\"\"\"docstring\"\"\"\n",
"\"\"\"\n # Check the request comes from appropriate location.\n if not utils.validate_ip(request.remote_addr)\n return {}, 403\n \"\"\"\n",
"username = request.args.get('username')\n",
"password = request.args.get('password')\n",
"if username.isalnum():\n",
"response = db_interac.authenticate(username, password)\n",
"return_obj['error'] = 'username must be alphanumeric'\n",
"return_obj = {}\n",
"return_obj['userId'] = None\n",
"if not response[0]:\n",
"return_obj['firstName'] = None\n",
"return_obj['error'] = 'user could not be authenticated'\n",
"return_obj['userId'] = response[1]\n",
"return_obj['lastName'] = None\n",
"return return_obj, 401\n",
"return_obj['firstName'] = response[2]\n",
"return return_obj, 401\n",
"return_obj['lastName'] = response[3]\n",
"return return_obj, 200\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Expr'",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Return'",
"Assign'",
"Return'",
"Assign'",
"Return'"
] |
[
"def FUNC_1(self):...\n",
"\"\"\"docstring\"\"\"\n",
"return bool(self.wildcard_names)\n"
] | [
"def has_wildcards(self):...\n",
"\"\"\"docstring\"\"\"\n",
"return bool(self.wildcard_names)\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Return'"
] |
[
"def FUNC_2(self):...\n",
"return self.best_submission\n"
] | [
"def get_best_submission(self):...\n",
"return self.best_submission\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
[
"def FUNC_6(self):...\n",
"\"\"\"docstring\"\"\"\n",
"self.check_image_exists()\n",
"if not self.security_groups:\n",
"self.security_groups[self.tenant_id] = self.verify(25, self.\n _create_security_group, 1, 'Security group can not be created.',\n 'security group creation', self.compute_client)\n",
"VAR_5 = rand_name('ost1_test-server-smoke-')\n",
"VAR_6 = [self.security_groups[self.tenant_id].name]\n",
"VAR_7 = self.verify(200, self._create_server, 2,\n 'Creating instance using the new security group has failed.',\n 'image creation', self.compute_client, VAR_5, VAR_6)\n",
"self.verify(30, self._delete_server, 3, 'Server can not be deleted.',\n 'server deletion', VAR_7)\n"
] | [
"def test_004_create_servers(self):...\n",
"\"\"\"docstring\"\"\"\n",
"self.check_image_exists()\n",
"if not self.security_groups:\n",
"self.security_groups[self.tenant_id] = self.verify(25, self.\n _create_security_group, 1, 'Security group can not be created.',\n 'security group creation', self.compute_client)\n",
"name = rand_name('ost1_test-server-smoke-')\n",
"security_groups = [self.security_groups[self.tenant_id].name]\n",
"server = self.verify(200, self._create_server, 2,\n 'Creating instance using the new security group has failed.',\n 'image creation', self.compute_client, name, security_groups)\n",
"self.verify(30, self._delete_server, 3, 'Server can not be deleted.',\n 'server deletion', server)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Expr'",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Expr'"
] |
[
"def FUNC_6(self, VAR_18):...\n",
"if not VAR_18:\n",
"self.error()\n",
"return VAR_18\n"
] | [
"def run(self, item):...\n",
"if not item:\n",
"self.error()\n",
"return item\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"Condition",
"Expr'",
"Return'"
] |
[
"@classmethod...\n",
"super(CLASS_0, VAR_0).setUpTestData()\n",
"VAR_0.user = UserFactory(email='[email protected]')\n",
"VAR_0.user.set_password('testing')\n",
"VAR_0.user.save()\n"
] | [
"@classmethod...\n",
"super(TestNavigation, cls).setUpTestData()\n",
"cls.user = UserFactory(email='[email protected]')\n",
"cls.user.set_password('testing')\n",
"cls.user.save()\n"
] | [
0,
0,
0,
0,
0
] | [
"Condition",
"Expr'",
"Assign'",
"Expr'",
"Expr'"
] |
[
"def FUNC_11(self):...\n",
"self.assertRaises(frappe.DataError, DatabaseQuery('DocType').execute,\n fields=['name', 'issingle, version()'], limit_start=0, limit_page_length=1)\n",
"self.assertRaises(frappe.DataError, DatabaseQuery('DocType').execute,\n fields=['name',\n 'issingle, IF(issingle=1, (select name from tabUser), count(name))'],\n limit_start=0, limit_page_length=1)\n",
"self.assertRaises(frappe.DataError, DatabaseQuery('DocType').execute,\n fields=['name', 'issingle, (select count(*) from tabSessions)'],\n limit_start=0, limit_page_length=1)\n",
"self.assertRaises(frappe.DataError, DatabaseQuery('DocType').execute,\n fields=['name',\n \"issingle, SELECT LOCATE('', `tabUser`.`user`) AS user;\"], limit_start=\n 0, limit_page_length=1)\n",
"self.assertRaises(frappe.DataError, DatabaseQuery('DocType').execute,\n fields=['name',\n 'issingle, IF(issingle=1, (SELECT name from tabUser), count(*))'],\n limit_start=0, limit_page_length=1)\n",
"VAR_3 = DatabaseQuery('DocType').execute(fields=['name', 'issingle',\n 'count(name)'], limit_start=0, limit_page_length=1)\n",
"self.assertTrue('count(name)' in VAR_3[0])\n",
"VAR_3 = DatabaseQuery('DocType').execute(fields=['name', 'issingle',\n \"locate('', name) as _relevance\"], limit_start=0, limit_page_length=1)\n",
"self.assertTrue('_relevance' in VAR_3[0])\n",
"VAR_3 = DatabaseQuery('DocType').execute(fields=['name', 'issingle',\n 'date(creation) as creation'], limit_start=0, limit_page_length=1)\n",
"self.assertTrue('creation' in VAR_3[0])\n",
"VAR_3 = DatabaseQuery('DocType').execute(fields=['name', 'issingle',\n 'datediff(modified, creation) as date_diff'], limit_start=0,\n limit_page_length=1)\n",
"self.assertTrue('date_diff' in VAR_3[0])\n"
] | [
"def test_query_fields_sanitizer(self):...\n",
"self.assertRaises(frappe.DataError, DatabaseQuery('DocType').execute,\n fields=['name', 'issingle, version()'], limit_start=0, limit_page_length=1)\n",
"self.assertRaises(frappe.DataError, DatabaseQuery('DocType').execute,\n fields=['name',\n 'issingle, IF(issingle=1, (select name from tabUser), count(name))'],\n limit_start=0, limit_page_length=1)\n",
"self.assertRaises(frappe.DataError, DatabaseQuery('DocType').execute,\n fields=['name', 'issingle, (select count(*) from tabSessions)'],\n limit_start=0, limit_page_length=1)\n",
"self.assertRaises(frappe.DataError, DatabaseQuery('DocType').execute,\n fields=['name',\n \"issingle, SELECT LOCATE('', `tabUser`.`user`) AS user;\"], limit_start=\n 0, limit_page_length=1)\n",
"self.assertRaises(frappe.DataError, DatabaseQuery('DocType').execute,\n fields=['name',\n 'issingle, IF(issingle=1, (SELECT name from tabUser), count(*))'],\n limit_start=0, limit_page_length=1)\n",
"data = DatabaseQuery('DocType').execute(fields=['name', 'issingle',\n 'count(name)'], limit_start=0, limit_page_length=1)\n",
"self.assertTrue('count(name)' in data[0])\n",
"data = DatabaseQuery('DocType').execute(fields=['name', 'issingle',\n \"locate('', name) as _relevance\"], limit_start=0, limit_page_length=1)\n",
"self.assertTrue('_relevance' in data[0])\n",
"data = DatabaseQuery('DocType').execute(fields=['name', 'issingle',\n 'date(creation) as creation'], limit_start=0, limit_page_length=1)\n",
"self.assertTrue('creation' in data[0])\n",
"data = DatabaseQuery('DocType').execute(fields=['name', 'issingle',\n 'datediff(modified, creation) as date_diff'], limit_start=0,\n limit_page_length=1)\n",
"self.assertTrue('date_diff' in data[0])\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Assign'",
"Expr'",
"Assign'",
"Expr'",
"Assign'",
"Expr'",
"Assign'",
"Expr'"
] |
[
"@inlineCallbacks...\n",
""
] | [
"@inlineCallbacks...\n",
""
] | [
0,
0
] | [
"Condition",
"Condition"
] |
[
"@VAR_0.route('/login', methods=['POST'])...\n",
"VAR_19 = CLASS_0.query.get(request.form['email'])\n",
"if VAR_19:\n",
"if request.form['password'] == VAR_19.pwd:\n",
"return render_template('unsuccessful-login.html')\n",
"VAR_19.authenticated = True\n",
"VAR_2.session.add(VAR_19)\n",
"VAR_2.session.commit()\n",
"login_user(VAR_19, remember=True)\n",
"return redirect(VAR_4)\n"
] | [
"@app.route('/login', methods=['POST'])...\n",
"user = User.query.get(request.form['email'])\n",
"if user:\n",
"if request.form['password'] == user.pwd:\n",
"return render_template('unsuccessful-login.html')\n",
"user.authenticated = True\n",
"db.session.add(user)\n",
"db.session.commit()\n",
"login_user(user, remember=True)\n",
"return redirect(url_prefix)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Assign'",
"Condition",
"For",
"Return'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Return'"
] |
[
"@decorators.require_cronjob...\n",
"VAR_25 = utils.utcnow() - on_error.ERROR_TIME_TO_LIVE\n",
"VAR_18 = models.Error.query(models.Error.created_ts < VAR_25,\n default_options=ndb.QueryOptions(keys_only=True))\n",
"VAR_26 = len(ndb.delete_multi(VAR_18))\n",
"self.response.headers['Content-Type'] = 'text/plain; charset=utf-8'\n",
"self.response.write(str(VAR_26))\n"
] | [
"@decorators.require_cronjob...\n",
"old_cutoff = utils.utcnow() - on_error.ERROR_TIME_TO_LIVE\n",
"items = models.Error.query(models.Error.created_ts < old_cutoff,\n default_options=ndb.QueryOptions(keys_only=True))\n",
"out = len(ndb.delete_multi(items))\n",
"self.response.headers['Content-Type'] = 'text/plain; charset=utf-8'\n",
"self.response.write(str(out))\n"
] | [
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Expr'"
] |
[
"def FUNC_17(self, VAR_18):...\n",
"\"\"\"docstring\"\"\"\n",
"self._eql_execute('volume', 'select', VAR_18['volume_name'], 'snapshot',\n 'delete', VAR_18['name'])\n",
"VAR_0.error(_('Failed to delete snapshot %(snap)s of volume %(vol)s'), {\n 'snap': VAR_18['name'], 'vol': VAR_18['volume_name']})\n"
] | [
"def delete_snapshot(self, snapshot):...\n",
"\"\"\"docstring\"\"\"\n",
"self._eql_execute('volume', 'select', snapshot['volume_name'], 'snapshot',\n 'delete', snapshot['name'])\n",
"LOG.error(_('Failed to delete snapshot %(snap)s of volume %(vol)s'), {\n 'snap': snapshot['name'], 'vol': snapshot['volume_name']})\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Expr'",
"Expr'"
] |
[
"def FUNC_11(self):...\n",
"VAR_3 = 5\n",
"for i in range(VAR_3):\n",
"VAR_11 = urllib.urlencode({url_helper.swarm_constants.COUNT_KEY: i})\n",
"self._mox.ReplayAll()\n",
"url_helper.urllib2.urlopen(mox.IgnoreArg(), VAR_11, timeout=mox.IgnoreArg()\n ).AndRaise(urllib2.URLError('url'))\n",
"self.assertEqual(url_helper.UrlOpen('url', max_tries=attempts), None)\n",
"if i != VAR_3 - 1:\n",
"self._mox.VerifyAll()\n",
"time.sleep(mox.IgnoreArg())\n"
] | [
"def testEnsureCountKeyIncludedInOpen(self):...\n",
"attempts = 5\n",
"for i in range(attempts):\n",
"encoded_data = urllib.urlencode({url_helper.swarm_constants.COUNT_KEY: i})\n",
"self._mox.ReplayAll()\n",
"url_helper.urllib2.urlopen(mox.IgnoreArg(), encoded_data, timeout=mox.\n IgnoreArg()).AndRaise(urllib2.URLError('url'))\n",
"self.assertEqual(url_helper.UrlOpen('url', max_tries=attempts), None)\n",
"if i != attempts - 1:\n",
"self._mox.VerifyAll()\n",
"time.sleep(mox.IgnoreArg())\n"
] | [
0,
0,
0,
0,
0,
5,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"For",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Condition",
"Expr'",
"Expr'"
] |
[
"def FUNC_16(self, VAR_9, VAR_7):...\n",
"VAR_18 = threading.Thread(target=self.__execute_interactive, kwargs={\n 'command': command, 'job_id': job_id})\n",
"VAR_18.daemon = True\n",
"VAR_18.start()\n"
] | [
"def _execute_interactive(self, command, job_id):...\n",
"thread = threading.Thread(target=self.__execute_interactive, kwargs={\n 'command': command, 'job_id': job_id})\n",
"thread.daemon = True\n",
"thread.start()\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Expr'"
] |
[
"from django import template\n",
"from django.utils import timezone\n",
"from lib.errors import TagUsageError\n",
"from ..cache import CachedNews\n",
"from ..models import News\n",
"VAR_0 = template.Library()\n",
"@VAR_0.inclusion_tag('news/user_news.html', takes_context=True)...\n",
"if not 'instance' in VAR_1:\n",
"if not 'now' in VAR_1:\n",
"VAR_1['now'] = timezone.now()\n",
"if not 'course_news' in VAR_1:\n",
"VAR_1['course_news'] = CachedNews(VAR_1['instance'])\n",
"VAR_7 = VAR_1['course_news']\n",
"if VAR_1['is_course_staff']:\n",
"VAR_9, VAR_7 = VAR_7.for_staff()\n",
"VAR_10 = VAR_1['request'].user\n",
"VAR_8 = 0\n",
"VAR_9, VAR_7 = VAR_7.for_user(not VAR_10.is_authenticated() or VAR_10.\n userprofile.is_external)\n",
"for VAR_11 in VAR_7:\n",
"VAR_8 += 1\n",
"return {'is_course_staff': VAR_1['is_course_staff'], 'now': VAR_1['now'],\n 'alerts': VAR_9, 'news': VAR_7, 'more': VAR_3}\n",
"VAR_11['collapsed'] = VAR_8 > VAR_2\n",
"if VAR_3 > 0 and VAR_8 == VAR_3:\n",
"VAR_11['begin_more'] = True\n"
] | [
"from django import template\n",
"from django.utils import timezone\n",
"from lib.errors import TagUsageError\n",
"from ..cache import CachedNews\n",
"from ..models import News\n",
"register = template.Library()\n",
"@register.inclusion_tag('news/user_news.html', takes_context=True)...\n",
"if not 'instance' in context:\n",
"if not 'now' in context:\n",
"context['now'] = timezone.now()\n",
"if not 'course_news' in context:\n",
"context['course_news'] = CachedNews(context['instance'])\n",
"news = context['course_news']\n",
"if context['is_course_staff']:\n",
"alerts, news = news.for_staff()\n",
"user = context['request'].user\n",
"i = 0\n",
"alerts, news = news.for_user(not user.is_authenticated() or user.\n userprofile.is_external)\n",
"for item in news:\n",
"i += 1\n",
"return {'is_course_staff': context['is_course_staff'], 'now': context['now'\n ], 'alerts': alerts, 'news': news, 'more': more}\n",
"item['collapsed'] = i > num\n",
"if more > 0 and i == more:\n",
"item['begin_more'] = True\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
4,
0,
0,
0,
0,
0,
0
] | [
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"Assign'",
"Condition",
"Condition",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"For",
"AugAssign'",
"Return'",
"Assign'",
"Condition",
"Assign'"
] |
[
"def FUNC_4(self):...\n",
"assert self.initialised, 'Initialise the map from a file first!'\n",
"for VAR_23 in range(self.mapHeight):\n",
"for VAR_22 in range(self.mapWidth):\n",
"if self.stateMap[VAR_22][VAR_23].isTraversable():\n",
"self.stateMap[VAR_22][VAR_23].setTurningNeighbours(self.nearestNodes(VAR_22,\n VAR_23))\n"
] | [
"def genConnections(self):...\n",
"assert self.initialised, 'Initialise the map from a file first!'\n",
"for row in range(self.mapHeight):\n",
"for col in range(self.mapWidth):\n",
"if self.stateMap[col][row].isTraversable():\n",
"self.stateMap[col][row].setTurningNeighbours(self.nearestNodes(col, row))\n"
] | [
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assert'",
"For",
"For",
"Condition",
"Expr'"
] |
[
"def FUNC_3(VAR_5, VAR_6):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_15 = {}\n",
"VAR_15['new_status'] = 'active' if VAR_6 == 'unfreeze' else 'inactive'\n",
"VAR_16 = 'inactive' if VAR_6 == 'unfreeze' else 'active'\n",
"VAR_15['fields'] = pickle.dumps({'scope': VAR_5.key(), 'status': VAR_16})\n",
"for pattern in VAR_0:\n",
"responses.startTask(FUNC_6(pattern), VAR_24=new_context)\n",
"VAR_15['fields'] = pickle.dumps({'program': VAR_5.key(), 'status': VAR_16})\n",
"for pattern in VAR_1:\n",
"responses.startTask(FUNC_6(pattern), VAR_24=new_context)\n",
"VAR_15 = {}\n",
"VAR_16 = 'inactive' if VAR_6 == 'unfreeze' else ['active', 'new']\n",
"VAR_15['fields'] = pickle.dumps({'scope': VAR_5.key(), 'status': VAR_16})\n",
"responses.startTask(FUNC_6(VAR_2[0]), VAR_24=new_context)\n",
"return\n"
] | [
"def _processProgramFreezing(program_entity, mode):...\n",
"\"\"\"docstring\"\"\"\n",
"new_context = {}\n",
"new_context['new_status'] = 'active' if mode == 'unfreeze' else 'inactive'\n",
"old_status = 'inactive' if mode == 'unfreeze' else 'active'\n",
"new_context['fields'] = pickle.dumps({'scope': program_entity.key(),\n 'status': old_status})\n",
"for pattern in ROLE_PER_SCOPE_MODELS_URL_PATTERNS:\n",
"responses.startTask(_constructRequestURL(pattern), context=new_context)\n",
"new_context['fields'] = pickle.dumps({'program': program_entity.key(),\n 'status': old_status})\n",
"for pattern in ROLE_PER_PROGRAM_MODELS_URL_PATTERNS:\n",
"responses.startTask(_constructRequestURL(pattern), context=new_context)\n",
"new_context = {}\n",
"old_status = 'inactive' if mode == 'unfreeze' else ['active', 'new']\n",
"new_context['fields'] = pickle.dumps({'scope': program_entity.key(),\n 'status': old_status})\n",
"responses.startTask(_constructRequestURL(ORG_MODEL_URL_PATTERNS[0]),\n context=new_context)\n",
"return\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"For",
"Expr'",
"Assign'",
"For",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Return'"
] |
[
"def FUNC_10(self, VAR_18):...\n",
"VAR_30 = VAR_18.uuid\n",
"VAR_41 = VAR_18.file_name\n",
"VAR_42 = VAR_18.owner\n",
"VAR_43 = VAR_18.upload_time\n",
"VAR_44 = list()\n",
"VAR_45 = list()\n",
"for VAR_65 in VAR_18.sub_items:\n",
"if VAR_65.is_dir:\n",
"VAR_46 = \"'{\" + ', '.join(i for i in VAR_44) + \"}'\"\n",
"VAR_44.append('\"%s\"' % VAR_65.uuid)\n",
"VAR_76 = '{%s, %s, %s, %s, %s}' % ('\"%s\"' % VAR_65.uuid, '\"%s\"' % VAR_65.\n file_name, '\"%s\"' % VAR_65.owner, '\"%f\"' % VAR_65.upload_time, '\"%s\"' %\n VAR_65.f_uuid)\n",
"VAR_47 = \"'{\" + ', '.join(i for i in VAR_45) + \"}'\"\n",
"VAR_45.append(VAR_76)\n",
"return VAR_30, VAR_41, VAR_42, VAR_43, VAR_46, VAR_47\n"
] | [
"def _sqlify_fsnode(self, item):...\n",
"n_uuid = item.uuid\n",
"n_file_name = item.file_name\n",
"n_owner = item.owner\n",
"n_upload_time = item.upload_time\n",
"n_sub_folders = list()\n",
"n_sub_files = list()\n",
"for i_sub in item.sub_items:\n",
"if i_sub.is_dir:\n",
"n_sub_folders_str = \"'{\" + ', '.join(i for i in n_sub_folders) + \"}'\"\n",
"n_sub_folders.append('\"%s\"' % i_sub.uuid)\n",
"s_attr = '{%s, %s, %s, %s, %s}' % ('\"%s\"' % i_sub.uuid, '\"%s\"' % i_sub.\n file_name, '\"%s\"' % i_sub.owner, '\"%f\"' % i_sub.upload_time, '\"%s\"' %\n i_sub.f_uuid)\n",
"n_sub_files_str = \"'{\" + ', '.join(i for i in n_sub_files) + \"}'\"\n",
"n_sub_files.append(s_attr)\n",
"return n_uuid, n_file_name, n_owner, n_upload_time, n_sub_folders_str, n_sub_files_str\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"For",
"Condition",
"Assign'",
"Expr'",
"Assign'",
"Assign'",
"Expr'",
"Return'"
] |
[
"def FUNC_21(self):...\n",
"@eqlx.with_timeout...\n",
"return 'no timeout'\n"
] | [
"def test_with_timeout(self):...\n",
"@eqlx.with_timeout...\n",
"return 'no timeout'\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Condition",
"Return'"
] |
[
"def FUNC_0(self, VAR_5):...\n",
"VAR_14 = 'string'.format(VAR_5)\n",
"self.cur.execute(VAR_14)\n",
"return self.cur.fetchall()\n"
] | [
"def get_roster(self, server_id):...\n",
"sql = (\n \"\"\"SELECT username, role\n FROM roles\n WHERE roles.server_id = {0};\n \"\"\"\n .format(server_id))\n",
"self.cur.execute(sql)\n",
"return self.cur.fetchall()\n"
] | [
0,
4,
4,
0
] | [
"FunctionDef'",
"Assign'",
"Expr'",
"Return'"
] |
[
"@property...\n",
"return hasattr(self, 'extractedtext')\n"
] | [
"@property...\n",
"return hasattr(self, 'extractedtext')\n"
] | [
0,
0
] | [
"Condition",
"Return'"
] |
[
"def FUNC_22(self, VAR_21, VAR_40):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_57 = self.getfile(VAR_21)\n",
"if VAR_57 == False:\n",
"VAR_57[VAR_6] = FUNC_30.S_IFMT(VAR_57[VAR_6]) | VAR_40\n"
] | [
"def chmod(self, path, perm):...\n",
"\"\"\"docstring\"\"\"\n",
"p = self.getfile(path)\n",
"if p == False:\n",
"p[A_MODE] = stat.S_IFMT(p[A_MODE]) | perm\n"
] | [
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Condition",
"Assign'"
] |
[
"def FUNC_1(VAR_1=None):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_4 = 'df -i'\n",
"if VAR_1 is not None:\n",
"VAR_4 = VAR_4 + ' -' + VAR_1\n",
"VAR_2 = {}\n",
"VAR_3 = __salt__['cmd.run'](VAR_4).splitlines()\n",
"for line in VAR_3:\n",
"if line.startswith('Filesystem'):\n",
"return VAR_2\n",
"VAR_5 = line.split()\n",
"if not VAR_5:\n",
"if __grains__['kernel'] == 'OpenBSD':\n",
"VAR_0.warn('Problem parsing inode usage information')\n",
"VAR_2[VAR_5[8]] = {'inodes': int(VAR_5[5]) + int(VAR_5[6]), 'used': VAR_5[5\n ], 'free': VAR_5[6], 'use': VAR_5[7], 'filesystem': VAR_5[0]}\n",
"VAR_2[VAR_5[5]] = {'inodes': VAR_5[1], 'used': VAR_5[2], 'free': VAR_5[3],\n 'use': VAR_5[4], 'filesystem': VAR_5[0]}\n",
"VAR_2 = {}\n"
] | [
"def inodeusage(args=None):...\n",
"\"\"\"docstring\"\"\"\n",
"cmd = 'df -i'\n",
"if args is not None:\n",
"cmd = cmd + ' -' + args\n",
"ret = {}\n",
"out = __salt__['cmd.run'](cmd).splitlines()\n",
"for line in out:\n",
"if line.startswith('Filesystem'):\n",
"return ret\n",
"comps = line.split()\n",
"if not comps:\n",
"if __grains__['kernel'] == 'OpenBSD':\n",
"log.warn('Problem parsing inode usage information')\n",
"ret[comps[8]] = {'inodes': int(comps[5]) + int(comps[6]), 'used': comps[5],\n 'free': comps[6], 'use': comps[7], 'filesystem': comps[0]}\n",
"ret[comps[5]] = {'inodes': comps[1], 'used': comps[2], 'free': comps[3],\n 'use': comps[4], 'filesystem': comps[0]}\n",
"ret = {}\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"For",
"Condition",
"Return'",
"Assign'",
"Condition",
"Condition",
"Expr'",
"Assign'",
"Assign'",
"Assign'"
] |
[
"def FUNC_13(self, VAR_22):...\n",
"\"\"\"docstring\"\"\"\n",
"logging.warning(\"\"\"Authorization error.\n%s\nPeer: %s\nIP: %s\"\"\", VAR_22, api.\n get_peer_identity().to_bytes(), self.request.remote_addr)\n",
"self.abort(403, detail=str(error))\n"
] | [
"def authorization_error(self, error):...\n",
"\"\"\"docstring\"\"\"\n",
"logging.warning(\"\"\"Authorization error.\n%s\nPeer: %s\nIP: %s\"\"\", error, api.\n get_peer_identity().to_bytes(), self.request.remote_addr)\n",
"self.abort(403, detail=str(error))\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Expr'",
"Expr'"
] |
[
"def FUNC_0(VAR_0=None):...\n",
"VAR_1 = FUNC_1()\n",
"VAR_2 = FUNC_2(VAR_0)\n",
"return VAR_1, VAR_2\n"
] | [
"def execute(filters=None):...\n",
"columns = get_columns()\n",
"data = get_data(filters)\n",
"return columns, data\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Return'"
] |
[
"def FUNC_0(VAR_1):...\n",
"VAR_6 = FUNC_3(VAR_1)\n",
"VAR_7 = VAR_6.cursor(MySQLdb.cursors.DictCursor)\n",
"VAR_8 = (\n \"SELECT * FROM questions WHERE status='not-started' OR status='in-progress' ORDER BY ts DESC;\"\n )\n",
"VAR_7.execute(VAR_8)\n",
"VAR_9 = []\n",
"while True:\n",
"VAR_10 = VAR_7.fetchone()\n",
"if VAR_10 == None:\n",
"VAR_9.append(\n \"<input type='submit' name=questionSubmit value='Answer Selected Question'>\"\n )\n",
"VAR_9.append('string'.format(VAR_2=row['id'], question=row['question'],\n status=row['status'], ts=row['ts']))\n",
"return '\\n'.join(VAR_9)\n",
"if VAR_10['status'] == 'in-progress':\n",
"VAR_9.append('<p>In-Progress Answer: {curr_answer}'.format(curr_answer=row[\n 'answer']))\n",
"VAR_9.append('</div>')\n"
] | [
"def makeQuestionSelect(database):...\n",
"conn = dbConnect(database)\n",
"curs = conn.cursor(MySQLdb.cursors.DictCursor)\n",
"statement = (\n \"SELECT * FROM questions WHERE status='not-started' OR status='in-progress' ORDER BY ts DESC;\"\n )\n",
"curs.execute(statement)\n",
"lines = []\n",
"while True:\n",
"row = curs.fetchone()\n",
"if row == None:\n",
"lines.append(\n \"<input type='submit' name=questionSubmit value='Answer Selected Question'>\"\n )\n",
"lines.append(\n \"\"\"<div style='border:2px solid black;'><input type='radio' name='q_selection' value={id}> Question: {question}\n<p>Status: {status}\n<p>Time submitted: {ts}\"\"\"\n .format(id=row['id'], question=row['question'], status=row['status'],\n ts=row['ts']))\n",
"return '\\n'.join(lines)\n",
"if row['status'] == 'in-progress':\n",
"lines.append('<p>In-Progress Answer: {curr_answer}'.format(curr_answer=row[\n 'answer']))\n",
"lines.append('</div>')\n"
] | [
0,
0,
0,
0,
4,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Expr'",
"Expr'",
"Return'",
"Condition",
"Expr'",
"Expr'"
] |
[
"def FUNC_7(self):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_4 = self.parse_body()\n",
"VAR_9 = VAR_4.get('version', None)\n",
"VAR_24 = VAR_4.get('dimensions', {})\n",
"VAR_25 = VAR_4.get('state', {})\n",
"VAR_14 = None\n",
"if VAR_24.get('id'):\n",
"VAR_52 = VAR_24['id']\n",
"if bool(VAR_24.get('quarantined')) or bool(VAR_25.get('quarantined')):\n",
"if isinstance(VAR_52, list) and len(VAR_52) == 1 and isinstance(VAR_52[0],\n",
"return VAR_4, VAR_14, VAR_9, VAR_25, VAR_24, 'Bot self-quarantined'\n",
"VAR_26 = None\n",
"VAR_14 = VAR_24['id'][0]\n",
"for _ in [0]:\n",
"VAR_26 = FUNC_1(self.EXPECTED_KEYS, VAR_4, 'keys')\n",
"if VAR_26:\n",
"if VAR_26:\n",
"VAR_50 = \"\"\"Quarantined Bot\nhttps://%s/restricted/bot/%s\n%s\"\"\" % (app_identity\n .get_default_version_hostname(), VAR_14, VAR_26)\n",
"VAR_27 = bot_management.get_settings_key(VAR_14).get()\n",
"VAR_26 = FUNC_4(self.REQUIRED_STATE_KEYS, VAR_25, 'state')\n",
"ereporter2.log_request(self.request, VAR_5='bot', VAR_8=line)\n",
"if bool(VAR_27 and VAR_27.quarantined):\n",
"if VAR_26:\n",
"return VAR_4, VAR_14, VAR_9, VAR_25, VAR_24, VAR_26\n",
"return VAR_4, VAR_14, VAR_9, VAR_25, VAR_24, 'Quarantined by admin'\n",
"return VAR_4, VAR_14, VAR_9, VAR_25, VAR_24, None\n",
"if not VAR_14:\n",
"VAR_26 = 'Missing bot id'\n",
"if not all(isinstance(key, unicode) and isinstance(values, list) and all(\n",
"VAR_26 = \"\"\"Invalid dimensions type:\n%s\"\"\" % json.dumps(VAR_24, sort_keys=\n True, indent=2, separators=(',', ': '))\n",
"VAR_53 = task_to_run.dimensions_powerset_count(VAR_24)\n",
"if VAR_53 > task_to_run.MAX_DIMENSIONS:\n",
"VAR_26 = 'Dimensions product %d is too high' % VAR_53\n",
"if not isinstance(VAR_25.get('lease_expiration_ts'), (None.__class__, int)):\n",
"VAR_26 = 'lease_expiration_ts (%r) must be int or None' % VAR_25[\n 'lease_expiration_ts']\n"
] | [
"def _process(self):...\n",
"\"\"\"docstring\"\"\"\n",
"request = self.parse_body()\n",
"version = request.get('version', None)\n",
"dimensions = request.get('dimensions', {})\n",
"state = request.get('state', {})\n",
"bot_id = None\n",
"if dimensions.get('id'):\n",
"dimension_id = dimensions['id']\n",
"if bool(dimensions.get('quarantined')) or bool(state.get('quarantined')):\n",
"if isinstance(dimension_id, list) and len(dimension_id) == 1 and isinstance(\n",
"return request, bot_id, version, state, dimensions, 'Bot self-quarantined'\n",
"quarantined_msg = None\n",
"bot_id = dimensions['id'][0]\n",
"for _ in [0]:\n",
"quarantined_msg = has_unexpected_keys(self.EXPECTED_KEYS, request, 'keys')\n",
"if quarantined_msg:\n",
"if quarantined_msg:\n",
"line = \"\"\"Quarantined Bot\nhttps://%s/restricted/bot/%s\n%s\"\"\" % (app_identity\n .get_default_version_hostname(), bot_id, quarantined_msg)\n",
"bot_settings = bot_management.get_settings_key(bot_id).get()\n",
"quarantined_msg = has_missing_keys(self.REQUIRED_STATE_KEYS, state, 'state')\n",
"ereporter2.log_request(self.request, source='bot', message=line)\n",
"if bool(bot_settings and bot_settings.quarantined):\n",
"if quarantined_msg:\n",
"return request, bot_id, version, state, dimensions, quarantined_msg\n",
"return request, bot_id, version, state, dimensions, 'Quarantined by admin'\n",
"return request, bot_id, version, state, dimensions, None\n",
"if not bot_id:\n",
"quarantined_msg = 'Missing bot id'\n",
"if not all(isinstance(key, unicode) and isinstance(values, list) and all(\n",
"quarantined_msg = \"\"\"Invalid dimensions type:\n%s\"\"\" % json.dumps(dimensions,\n sort_keys=True, indent=2, separators=(',', ': '))\n",
"dimensions_count = task_to_run.dimensions_powerset_count(dimensions)\n",
"if dimensions_count > task_to_run.MAX_DIMENSIONS:\n",
"quarantined_msg = 'Dimensions product %d is too high' % dimensions_count\n",
"if not isinstance(state.get('lease_expiration_ts'), (None.__class__, int)):\n",
"quarantined_msg = 'lease_expiration_ts (%r) must be int or None' % state[\n 'lease_expiration_ts']\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Condition",
"Return'",
"Assign'",
"Assign'",
"For",
"Assign'",
"Condition",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Condition",
"Condition",
"Return'",
"Return'",
"Return'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Assign'"
] |
[
"def FUNC_12(self, VAR_21):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_24 = self.getfile(VAR_21)\n",
"return False\n",
"return VAR_24[VAR_2] == VAR_11\n"
] | [
"def islink(self, path):...\n",
"\"\"\"docstring\"\"\"\n",
"f = self.getfile(path)\n",
"return False\n",
"return f[A_TYPE] == T_LINK\n"
] | [
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Return'",
"Return'"
] |
[
"def FUNC_0(self, VAR_1):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_11 = json.dumps(VAR_1)\n",
"VAR_11 = json.dumps(fix_unicode_dict(VAR_1))\n",
"if len(self.events_requests) == 0:\n",
"return\n",
"[VAR_9.write(VAR_11 + '\\n') for VAR_9 in self.events_requests]\n"
] | [
"def write_data(self, message):...\n",
"\"\"\"docstring\"\"\"\n",
"message_str = json.dumps(message)\n",
"message_str = json.dumps(fix_unicode_dict(message))\n",
"if len(self.events_requests) == 0:\n",
"return\n",
"[request.write(message_str + '\\n') for request in self.events_requests]\n"
] | [
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Assign'",
"Condition",
"Return'",
"Expr'"
] |
[
"def __unicode__(self):...\n",
"return unicode(self.nickname())\n"
] | [
"def __unicode__(self):...\n",
"return unicode(self.nickname())\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
[
"def FUNC_29(self):...\n",
"if VAR_54.flags.in_import:\n",
"return\n",
"for VAR_25 in self.meta.get_select_fields():\n",
"if VAR_25.fieldname == 'naming_series' or not (self.get(VAR_25.fieldname) and\n",
"VAR_56 = (VAR_25.options or '').split('\\n')\n",
"if not filter(None, VAR_56):\n",
"self.set(VAR_25.fieldname, cstr(self.get(VAR_25.fieldname)).strip())\n",
"VAR_9 = self.get(VAR_25.fieldname)\n",
"if VAR_9 not in VAR_56 and not (VAR_54.flags.in_test and VAR_9.startswith(\n",
"VAR_67 = _('Row #{0}:').format(self.idx) if self.get('parentfield') else ''\n",
"VAR_37 = _(self.meta.get_label(VAR_25.fieldname))\n",
"VAR_68 = '\", \"'.join(_(each) for each in VAR_56)\n",
"VAR_54.throw(_('{0} {1} cannot be \"{2}\". It should be one of \"{3}\"').format\n (VAR_67, VAR_37, VAR_9, VAR_68))\n"
] | [
"def _validate_selects(self):...\n",
"if frappe.flags.in_import:\n",
"return\n",
"for df in self.meta.get_select_fields():\n",
"if df.fieldname == 'naming_series' or not (self.get(df.fieldname) and df.\n",
"options = (df.options or '').split('\\n')\n",
"if not filter(None, options):\n",
"self.set(df.fieldname, cstr(self.get(df.fieldname)).strip())\n",
"value = self.get(df.fieldname)\n",
"if value not in options and not (frappe.flags.in_test and value.startswith(\n",
"prefix = _('Row #{0}:').format(self.idx) if self.get('parentfield') else ''\n",
"label = _(self.meta.get_label(df.fieldname))\n",
"comma_options = '\", \"'.join(_(each) for each in options)\n",
"frappe.throw(_('{0} {1} cannot be \"{2}\". It should be one of \"{3}\"').format\n (prefix, label, value, comma_options))\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Condition",
"Return'",
"For",
"Condition",
"Assign'",
"Condition",
"Expr'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Expr'"
] |
[
"def FUNC_9(self, VAR_13, VAR_14='result'):...\n",
"self.cursor.execute('SELECT COUNT(*) FROM log_' + VAR_13 + ' WHERE ' +\n VAR_14 + '_id=?', [getattr(self, VAR_14)['id']])\n",
"return self.cursor.fetchone()['COUNT(*)']\n"
] | [
"def get_count(self, item, item_from='result'):...\n",
"self.cursor.execute('SELECT COUNT(*) FROM log_' + item + ' WHERE ' +\n item_from + '_id=?', [getattr(self, item_from)['id']])\n",
"return self.cursor.fetchone()['COUNT(*)']\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Expr'",
"Return'"
] |
[
"def FUNC_5(*VAR_12):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_22 = np.asarray(VAR_12)\n",
"return VAR_22 / np.linalg.norm(VAR_22)\n"
] | [
"def ket(*args):...\n",
"\"\"\"docstring\"\"\"\n",
"state = np.asarray(args)\n",
"return state / np.linalg.norm(state)\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Return'"
] |
[
"@utils.synchronized('3par', external=True)...\n",
"self.common.client_login()\n",
"self.common.delete_snapshot(VAR_9)\n",
"self.common.client_logout()\n"
] | [
"@utils.synchronized('3par', external=True)...\n",
"self.common.client_login()\n",
"self.common.delete_snapshot(snapshot)\n",
"self.common.client_logout()\n"
] | [
0,
0,
0,
0
] | [
"Condition",
"Expr'",
"Expr'",
"Expr'"
] |
[
"def FUNC_11(VAR_1, VAR_2):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_10 = False\n",
"VAR_11 = db.connection.cursor(db.pymysql.cursors.DictCursor)\n",
"VAR_11.execute('SELECT type FROM achievement_definitions WHERE id = %s', VAR_1)\n",
"VAR_9 = VAR_11.fetchone()\n",
"if VAR_9['type'] != 'STANDARD':\n",
"VAR_11.execute('string', (VAR_1, VAR_2))\n",
"VAR_14 = VAR_11.fetchone()\n",
"VAR_15 = 'UNLOCKED'\n",
"VAR_10 = not VAR_14 or VAR_14['state'] != 'UNLOCKED'\n",
"VAR_11.execute('string', {'player_id': VAR_2, 'achievement_id': VAR_1,\n 'state': VAR_15})\n",
"return dict(VAR_10=newly_unlocked)\n"
] | [
"def unlock_achievement(achievement_id, player_id):...\n",
"\"\"\"docstring\"\"\"\n",
"newly_unlocked = False\n",
"cursor = db.connection.cursor(db.pymysql.cursors.DictCursor)\n",
"cursor.execute('SELECT type FROM achievement_definitions WHERE id = %s',\n achievement_id)\n",
"achievement = cursor.fetchone()\n",
"if achievement['type'] != 'STANDARD':\n",
"cursor.execute(\n \"\"\"SELECT\n state\n FROM player_achievements\n WHERE achievement_id = %s AND player_id = %s\"\"\"\n , (achievement_id, player_id))\n",
"player_achievement = cursor.fetchone()\n",
"new_state = 'UNLOCKED'\n",
"newly_unlocked = not player_achievement or player_achievement['state'\n ] != 'UNLOCKED'\n",
"cursor.execute(\n \"\"\"INSERT INTO player_achievements (player_id, achievement_id, state)\n VALUES\n (%(player_id)s, %(achievement_id)s, %(state)s)\n ON DUPLICATE KEY UPDATE\n state = VALUES(state)\"\"\"\n , {'player_id': player_id, 'achievement_id': achievement_id, 'state':\n new_state})\n",
"return dict(newly_unlocked=newly_unlocked)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Condition",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Return'"
] |
[
"def FUNC_8(self, VAR_43):...\n",
"VAR_82 = self.divex.match(VAR_43)\n",
"if VAR_82 is None:\n",
"VAR_83 = VAR_82.group(1)\n",
"VAR_84 = VAR_82.group(2)\n",
"VAR_85 = 'mod'\n",
"VAR_86 = int(VAR_82.group(3))\n",
"VAR_41 = VAR_82.group(4)\n",
"VAR_87 = self._parse_poly(VAR_83, VAR_41)\n",
"VAR_88 = self._parse_poly(VAR_84, VAR_41)\n",
"VAR_89 = polylib.IntField(VAR_86)\n",
"VAR_90 = polylib.FieldPoly(VAR_89, VAR_87)\n",
"VAR_91 = polylib.FieldPoly(VAR_89, VAR_88)\n",
"return VAR_90, VAR_85, VAR_91\n"
] | [
"def _parse_instruction(self, s):...\n",
"match = self.divex.match(s)\n",
"if match is None:\n",
"poly1 = match.group(1)\n",
"poly2 = match.group(2)\n",
"instruction = 'mod'\n",
"p = int(match.group(3))\n",
"var = match.group(4)\n",
"cs1 = self._parse_poly(poly1, var)\n",
"cs2 = self._parse_poly(poly2, var)\n",
"field = polylib.IntField(p)\n",
"p1 = polylib.FieldPoly(field, cs1)\n",
"p2 = polylib.FieldPoly(field, cs2)\n",
"return p1, instruction, p2\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Return'"
] |
[
"def __str__(self):...\n",
"return self.name\n"
] | [
"def __str__(self):...\n",
"return self.name\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
[
"def FUNC_17(self, VAR_50, **VAR_18):...\n",
"VAR_57 = getattr(self._record, VAR_50.lower())\n",
"VAR_57.assert_called_once_with(**kwargs)\n"
] | [
"def assert_called_once(self, verb, **kwargs):...\n",
"mock_call = getattr(self._record, verb.lower())\n",
"mock_call.assert_called_once_with(**kwargs)\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Expr'"
] |
[
"def FUNC_39(VAR_62):...\n",
"return json.loads(VAR_62) if VAR_21 else VAR_62\n"
] | [
"def get_str(_str):...\n",
"return json.loads(_str) if as_json else _str\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
[
"def FUNC_1(VAR_1, VAR_2=False, VAR_3=False):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_7 = {}\n",
"if VAR_2 and VAR_3:\n",
"VAR_0.error(e)\n",
"return VAR_7\n",
"for VAR_11 in FUNC_3(VAR_1):\n",
"if not VAR_2:\n",
"VAR_7[VAR_11.username] = VAR_11.signed\n",
"if VAR_3:\n",
"for VAR_11 in FUNC_4(VAR_1):\n",
"for VAR_11 in FUNC_5(VAR_1):\n",
"VAR_7[VAR_11.username] = VAR_11.signed\n",
"VAR_7[VAR_11.username] = bool(VAR_11.signed)\n"
] | [
"def get_signed_packets(member, intro=False, onfloor=False):...\n",
"\"\"\"docstring\"\"\"\n",
"signed_packets = {}\n",
"if intro and onfloor:\n",
"LOGGER.error(e)\n",
"return signed_packets\n",
"for signature in query_signed_intromember(member):\n",
"if not intro:\n",
"signed_packets[signature.username] = signature.signed\n",
"if onfloor:\n",
"for signature in query_signed_upperclassman(member):\n",
"for signature in query_signed_alumni(member):\n",
"signed_packets[signature.username] = signature.signed\n",
"signed_packets[signature.username] = bool(signature.signed)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Condition",
"Expr'",
"Return'",
"For",
"Condition",
"Assign'",
"Condition",
"For",
"For",
"Assign'",
"Assign'"
] |
[
"def FUNC_22(self, VAR_21=None):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_22 = {}\n",
"self.clean_fields(VAR_21)\n",
"VAR_22 = e.update_error_dict(VAR_22)\n",
"self.clean()\n",
"VAR_22 = e.update_error_dict(VAR_22)\n",
"if VAR_22:\n"
] | [
"def full_clean(self, exclude=None):...\n",
"\"\"\"docstring\"\"\"\n",
"errors = {}\n",
"self.clean_fields(exclude)\n",
"errors = e.update_error_dict(errors)\n",
"self.clean()\n",
"errors = e.update_error_dict(errors)\n",
"if errors:\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Expr'",
"Assign'",
"Expr'",
"Assign'",
"Condition"
] |
[
"def FUNC_61():...\n",
"self.cursor.execute('create table t1 (word varchar (100))')\n",
"VAR_39 = set(['a'])\n",
"self.cursor.executemany('insert into t1 (word) values (?)', [VAR_39])\n"
] | [
"def f():...\n",
"self.cursor.execute('create table t1 (word varchar (100))')\n",
"words = set(['a'])\n",
"self.cursor.executemany('insert into t1 (word) values (?)', [words])\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"Expr'",
"Assign'",
"Expr'"
] |
[
"@staticmethod...\n",
"if VAR_5 == 'north':\n",
"return 'east'\n",
"if VAR_5 == 'east':\n",
"return 'north'\n",
"if VAR_5 == 'south':\n",
"return 'west'\n",
"if VAR_5 == 'west':\n",
"return 'south'\n"
] | [
"@staticmethod...\n",
"if direction == 'north':\n",
"return 'east'\n",
"if direction == 'east':\n",
"return 'north'\n",
"if direction == 'south':\n",
"return 'west'\n",
"if direction == 'west':\n",
"return 'south'\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Condition",
"Return'",
"Condition",
"Return'",
"Condition",
"Return'",
"Condition",
"Return'"
] |
[
"def FUNC_2(self, VAR_6, VAR_7):...\n",
"\"\"\"docstring\"\"\"\n",
"self._logger.error('Error with job %d: %s', VAR_6, VAR_7)\n",
"self.put('worker/%s' % VAR_6, data={'log': message, 'returncode': 1, 'host':\n socket.gethostbyaddr(socket.getfqdn())})\n",
"self._logger.exception('Error trying to PUT back abort message')\n"
] | [
"def _abort(self, job_id, message):...\n",
"\"\"\"docstring\"\"\"\n",
"self._logger.error('Error with job %d: %s', job_id, message)\n",
"self.put('worker/%s' % job_id, data={'log': message, 'returncode': 1,\n 'host': socket.gethostbyaddr(socket.getfqdn())})\n",
"self._logger.exception('Error trying to PUT back abort message')\n"
] | [
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Expr'",
"Expr'",
"Expr'"
] |
[
"@view_config(route_name='new-note', permission='edit', renderer=...\n",
"VAR_2 = db.note.Note(None, user_id=request.user.user_id, content=request.\n params.get('note', ''))\n",
"if VAR_0.method == 'POST':\n",
"FUNC_5(VAR_2, VAR_0)\n",
"return dict(title='Ny anteckning', VAR_2=note)\n",
"return HTTPFound(location=request.route_url('notes'))\n"
] | [
"@view_config(route_name='new-note', permission='edit', renderer=...\n",
"note = db.note.Note(None, user_id=request.user.user_id, content=request.\n params.get('note', ''))\n",
"if request.method == 'POST':\n",
"_save_or_create_note(note, request)\n",
"return dict(title='Ny anteckning', note=note)\n",
"return HTTPFound(location=request.route_url('notes'))\n"
] | [
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Assign'",
"Condition",
"Expr'",
"Return'",
"Return'"
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.