lines
sequencelengths
1
444
raw_lines
sequencelengths
1
444
label
sequencelengths
1
444
type
sequencelengths
1
444
[ "def FUNC_2(VAR_1, VAR_2=1, VAR_3=False, VAR_4=[]):...\n", "\"\"\"docstring\"\"\"\n", "VAR_17 = VAR_1.data_group.type\n", "VAR_18, VAR_19 = get_extracted_models(VAR_17)\n", "VAR_20 = hasattr(VAR_1, 'extractedtext')\n", "def FUNC_3(VAR_21, VAR_22, VAR_23=BaseInlineFormSet, VAR_24=forms.ModelForm,...\n", "VAR_32 = VAR_22.detail_fields()\n", "if VAR_4:\n", "VAR_32 = [in_field for in_field in VAR_32 if not in_field in VAR_4]\n", "return forms.inlineformset_factory(VAR_21=parent_model, VAR_22=model,\n VAR_27=formset_fields, VAR_23=formset, VAR_24=form, VAR_2=extra, VAR_3=\n can_delete)\n" ]
[ "def create_detail_formset(document, extra=1, can_delete=False, exclude=[]):...\n", "\"\"\"docstring\"\"\"\n", "group_type = document.data_group.type\n", "parent, child = get_extracted_models(group_type)\n", "extracted = hasattr(document, 'extractedtext')\n", "def make_formset(parent_model, model, formset=BaseInlineFormSet, form=forms...\n", "formset_fields = model.detail_fields()\n", "if exclude:\n", "formset_fields = [in_field for in_field in formset_fields if not in_field in\n exclude]\n", "return forms.inlineformset_factory(parent_model=parent_model, model=model,\n fields=formset_fields, formset=formset, form=form, extra=extra,\n can_delete=can_delete)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Assign'", "For", "Assign'", "Condition", "Assign'", "Return'" ]
[ "def __call__(self, VAR_44):...\n", "self.p = VAR_44\n", "self.log = VAR_44.log\n", "self.inter_sleep = VAR_44.inter_sleep\n", "self.running = VAR_44.running\n", "self.p.sig_sock.setsockopt(zmq.SUBSCRIBE, b'WipeManager')\n", "self.p.wz.set_sig_handler(b'WipeManager', b'passthrough', self.send_passthrough\n )\n", "if self.c.tcount > 0:\n", "self.pc = ProcessContext(self.p.name, self.p.ctx, self.c.router_addr, VAR_10)\n", "if self.c.ecount > 0:\n", "self.spawnqueue = Queue()\n", "self.spawn_evaluators()\n", "while self.running.is_set():\n", "self.log.exception(e)\n", "self.terminate()\n", "self.load_bumplimit_set()\n", "if self.c.tcount == 0:\n", "self.join_threads()\n", "self.load_targets()\n", "self.inter_sleep(5)\n", "self.pc.check_waiting()\n", "if self.c.tcount > 0:\n", "self.load_users()\n", "VAR_56 = self.read_newproxies()\n", "self.save_users()\n", "self.spawn_wipethreads()\n", "if not VAR_56:\n", "self.save_targets()\n", "self.inter_sleep(5)\n", "self.add_spawns(VAR_56)\n" ]
[ "def __call__(self, parent):...\n", "self.p = parent\n", "self.log = parent.log\n", "self.inter_sleep = parent.inter_sleep\n", "self.running = parent.running\n", "self.p.sig_sock.setsockopt(zmq.SUBSCRIBE, b'WipeManager')\n", "self.p.wz.set_sig_handler(b'WipeManager', b'passthrough', self.send_passthrough\n )\n", "if self.c.tcount > 0:\n", "self.pc = ProcessContext(self.p.name, self.p.ctx, self.c.router_addr,\n noproxy_rp)\n", "if self.c.ecount > 0:\n", "self.spawnqueue = Queue()\n", "self.spawn_evaluators()\n", "while self.running.is_set():\n", "self.log.exception(e)\n", "self.terminate()\n", "self.load_bumplimit_set()\n", "if self.c.tcount == 0:\n", "self.join_threads()\n", "self.load_targets()\n", "self.inter_sleep(5)\n", "self.pc.check_waiting()\n", "if self.c.tcount > 0:\n", "self.load_users()\n", "new = self.read_newproxies()\n", "self.save_users()\n", "self.spawn_wipethreads()\n", "if not new:\n", "self.save_targets()\n", "self.inter_sleep(5)\n", "self.add_spawns(new)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'", "Condition", "Assign'", "Condition", "Assign'", "Expr'", "Condition", "Expr'", "Expr'", "Expr'", "Condition", "Expr'", "Expr'", "Expr'", "Expr'", "Condition", "Expr'", "Assign'", "Expr'", "Expr'", "Condition", "Expr'", "Expr'", "Expr'" ]
[ "def FUNC_23(self, VAR_16):...\n", "\"\"\"docstring\"\"\"\n", "VAR_35 = self.common._cli_run('showvlun -a -showcols Port', None)\n", "VAR_36 = {}\n", "for VAR_15 in VAR_16:\n", "VAR_36[VAR_15] = 0\n", "VAR_37 = None\n", "if VAR_35:\n", "VAR_35 = VAR_35[1:]\n", "return VAR_37\n", "for line in VAR_35:\n", "VAR_15 = line.strip()\n", "VAR_41 = sys.maxint\n", "if VAR_15 in VAR_36:\n", "for VAR_15, count in VAR_36.iteritems():\n", "VAR_36[VAR_15] = VAR_36[VAR_15] + 1\n", "if count < VAR_41:\n", "VAR_37 = VAR_15\n", "VAR_41 = count\n" ]
[ "def _get_least_used_nsp(self, nspss):...\n", "\"\"\"docstring\"\"\"\n", "result = self.common._cli_run('showvlun -a -showcols Port', None)\n", "nsp_counts = {}\n", "for nsp in nspss:\n", "nsp_counts[nsp] = 0\n", "current_least_used_nsp = None\n", "if result:\n", "result = result[1:]\n", "return current_least_used_nsp\n", "for line in result:\n", "nsp = line.strip()\n", "current_smallest_count = sys.maxint\n", "if nsp in nsp_counts:\n", "for nsp, count in nsp_counts.iteritems():\n", "nsp_counts[nsp] = nsp_counts[nsp] + 1\n", "if count < current_smallest_count:\n", "current_least_used_nsp = nsp\n", "current_smallest_count = count\n" ]
[ 0, 0, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "For", "Assign'", "Assign'", "Condition", "Assign'", "Return'", "For", "Assign'", "Assign'", "Condition", "For", "Assign'", "Condition", "Assign'", "Assign'" ]
[ "def FUNC_1(VAR_2, VAR_4):...\n", "VAR_8 = VAR_4.connect()\n", "VAR_9 = VAR_8.cursor()\n", "VAR_12 = 'string'.format(VAR_2)\n", "VAR_9.execute(VAR_12)\n", "VAR_13 = VAR_9.fetchall()\n", "return VAR_13\n" ]
[ "def getCalendarList(username, sqlInstance):...\n", "conn = sqlInstance.connect()\n", "cursor = conn.cursor()\n", "getCalendarDetails = (\n \"SELECT DISTINCT Calendars.calendarId, Calendars.name, Calendars.day FROM Users, Calendars, TimeSlots WHERE Calendars.calendarId = TimeSlots.calendarId AND (Calendars.userId = Users.userId OR TimeSlots.userId = Users.userId) AND Users.userEmail = '{0}'\"\n .format(username))\n", "cursor.execute(getCalendarDetails)\n", "result = cursor.fetchall()\n", "return result\n" ]
[ 0, 0, 0, 4, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "Return'" ]
[ "def FUNC_13():...\n", "\"\"\"docstring\"\"\"\n", "log.info('Stopping worker(s)')\n", "for pid in VAR_20:\n", "if pid is not None:\n", "os.kill(pid, VAR_7.SIGTERM)\n" ]
[ "def kill_children():...\n", "\"\"\"docstring\"\"\"\n", "log.info('Stopping worker(s)')\n", "for pid in child_pids:\n", "if pid is not None:\n", "os.kill(pid, signal.SIGTERM)\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Expr'", "For", "Condition", "Expr'" ]
[ "def FUNC_19(self):...\n", "VAR_7, VAR_12 = self.make_xsrf_handling_app()\n", "self.mock_get_current_identity(model.Identity(model.IDENTITY_USER,\n '[email protected]'))\n", "VAR_17 = VAR_7.get('/request').body\n", "self.mock_get_current_identity(model.Identity(model.IDENTITY_USER,\n '[email protected]'))\n", "VAR_15 = VAR_7.post('/request', expect_errors=True)\n", "self.assertEqual(403, VAR_15.status_int)\n", "self.assertFalse(VAR_12)\n" ]
[ "def test_xsrf_token_identity_matters(self):...\n", "app, calls = self.make_xsrf_handling_app()\n", "self.mock_get_current_identity(model.Identity(model.IDENTITY_USER,\n '[email protected]'))\n", "token = app.get('/request').body\n", "self.mock_get_current_identity(model.Identity(model.IDENTITY_USER,\n '[email protected]'))\n", "response = app.post('/request', expect_errors=True)\n", "self.assertEqual(403, response.status_int)\n", "self.assertFalse(calls)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Expr'", "Assign'", "Expr'", "Assign'", "Expr'", "Expr'" ]
[ "def FUNC_11(self):...\n", "return u'Nullable({})'.format(self.inner_type.for_schema())\n" ]
[ "def for_schema(self):...\n", "return u'Nullable({})'.format(self.inner_type.for_schema())\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_0(self, VAR_4={}):...\n", "self.connection.request(self.verb, self.path, VAR_4=headers)\n", "return self.connection.getresponse()\n" ]
[ "def run(self, headers={}):...\n", "self.connection.request(self.verb, self.path, headers=headers)\n", "return self.connection.getresponse()\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Return'" ]
[ "import sys\n", "import tkinter as tk\n", "import Controller as dbc\n", "from PIL import Image, ImageTk\n", "from tkinter import font\n", "def __init__(self):...\n", "tk.Tk.__init__(self)\n", "self.title('Raspi-Telxon')\n", "self.titleFont = font.Font(family='Helvetica', size=24)\n", "self.itemFont = font.Font(family='Helvetica', size=18)\n", "VAR_7 = tk.Frame(self)\n", "VAR_7.pack(side='top', fill='both', expand=True)\n", "VAR_7.grid_rowconfigure(0, weight=1)\n", "VAR_7.grid_columnconfigure(0, weight=1)\n", "self.frames = {}\n", "self.result = ''\n", "self.container = VAR_7\n", "for VAR_1 in (CLASS_1, VAR_9):\n", "VAR_2 = VAR_1(VAR_7, self)\n", "self.show_frame(CLASS_1)\n", "self.frames[VAR_1] = VAR_2\n", "def FUNC_0(self, VAR_1):...\n", "VAR_2.grid(row=0, column=0, sticky='nsew')\n", "VAR_8 = VAR_9(self.container, self)\n", "self.frames[VAR_9] = VAR_8\n", "VAR_8.grid(row=0, column=0, sticky='nsew')\n", "self.show_frame(VAR_8)\n", "def FUNC_1(self, VAR_2):...\n", "print('remove_frame: ' + str(VAR_2))\n", "self.frames.pop(VAR_2, None)\n", "def FUNC_2(self, VAR_3):...\n", "VAR_2 = self.frames[VAR_3]\n", "VAR_2.tkraise()\n", "def FUNC_3(self):...\n", "VAR_10 = VAR_11(self.container, self)\n", "self.frames[VAR_11] = VAR_10\n", "VAR_10.grid(row=0, column=0, sticky='nsew')\n", "self.show_frame(VAR_11)\n", "def FUNC_4(self, VAR_4):...\n", "self.result = VAR_4\n", "def FUNC_5(self):...\n", "return self.result\n" ]
[ "import sys\n", "import tkinter as tk\n", "import Controller as dbc\n", "from PIL import Image, ImageTk\n", "from tkinter import font\n", "def __init__(self):...\n", "tk.Tk.__init__(self)\n", "self.title('Raspi-Telxon')\n", "self.titleFont = font.Font(family='Helvetica', size=24)\n", "self.itemFont = font.Font(family='Helvetica', size=18)\n", "container = tk.Frame(self)\n", "container.pack(side='top', fill='both', expand=True)\n", "container.grid_rowconfigure(0, weight=1)\n", "container.grid_columnconfigure(0, weight=1)\n", "self.frames = {}\n", "self.result = ''\n", "self.container = container\n", "for F in (StartPage, SearchPage):\n", "frame = F(container, self)\n", "self.show_frame(StartPage)\n", "self.frames[F] = frame\n", "def create_frame(self, F):...\n", "frame.grid(row=0, column=0, sticky='nsew')\n", "new_frame = SearchPage(self.container, self)\n", "self.frames[SearchPage] = new_frame\n", "new_frame.grid(row=0, column=0, sticky='nsew')\n", "self.show_frame(new_frame)\n", "def remove_frame(self, frame):...\n", "print('remove_frame: ' + str(frame))\n", "self.frames.pop(frame, None)\n", "def show_frame(self, cont):...\n", "frame = self.frames[cont]\n", "frame.tkraise()\n", "def custom_frame(self):...\n", "result_frame = ResultsPage(self.container, self)\n", "self.frames[ResultsPage] = result_frame\n", "result_frame.grid(row=0, column=0, sticky='nsew')\n", "self.show_frame(ResultsPage)\n", "def set_result(self, result):...\n", "self.result = result\n", "def get_result(self):...\n", "return self.result\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 4, 4, 4, 4, 0, 4, 4, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Import'", "Import'", "Import'", "ImportFrom'", "ImportFrom'", "FunctionDef'", "Expr'", "Expr'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'", "Assign'", "Assign'", "Assign'", "For", "Assign'", "Expr'", "Assign'", "FunctionDef'", "Expr'", "Assign'", "Assign'", "Expr'", "Expr'", "FunctionDef'", "Expr'", "Expr'", "FunctionDef'", "Assign'", "Expr'", "FunctionDef'", "Assign'", "Assign'", "Expr'", "Expr'", "FunctionDef'", "Assign'", "FunctionDef'", "Return'" ]
[ "from __future__ import unicode_literals\n", "from django.db import models, migrations\n", "VAR_0 = [('exercise', '0022_auto_20160906_1401'), ('notification',\n '0001_initial')]\n", "VAR_1 = [migrations.AddField(model_name='notification', name='submission',\n field=models.ForeignKey(to='exercise.Submission', blank=True, null=True\n ), preserve_default=True), migrations.AlterField(model_name=\n 'notification', name='notification', field=models.TextField(blank=True),\n preserve_default=True), migrations.AlterField(model_name='notification',\n name='sender', field=models.ForeignKey(related_name=\n 'sent_notifications', to='userprofile.UserProfile', blank=True, null=\n True), preserve_default=True), migrations.AlterField(model_name=\n 'notification', name='subject', field=models.CharField(blank=True,\n max_length=255), preserve_default=True)]\n" ]
[ "from __future__ import unicode_literals\n", "from django.db import models, migrations\n", "dependencies = [('exercise', '0022_auto_20160906_1401'), ('notification',\n '0001_initial')]\n", "operations = [migrations.AddField(model_name='notification', name=\n 'submission', field=models.ForeignKey(to='exercise.Submission', blank=\n True, null=True), preserve_default=True), migrations.AlterField(\n model_name='notification', name='notification', field=models.TextField(\n blank=True), preserve_default=True), migrations.AlterField(model_name=\n 'notification', name='sender', field=models.ForeignKey(related_name=\n 'sent_notifications', to='userprofile.UserProfile', blank=True, null=\n True), preserve_default=True), migrations.AlterField(model_name=\n 'notification', name='subject', field=models.CharField(blank=True,\n max_length=255), preserve_default=True)]\n" ]
[ 0, 0, 0, 4 ]
[ "ImportFrom'", "ImportFrom'", "Assign'", "Assign'" ]
[ "def __init__(self, *VAR_2):...\n", "self._path = VAR_2\n" ]
[ "def __init__(self, *path):...\n", "self._path = path\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Assign'" ]
[ "def FUNC_6(self, VAR_19):...\n", "return VAR_19 if VAR_19 and self.measure.match(VAR_19) else ''\n" ]
[ "def run(self, value):...\n", "return value if value and self.measure.match(value) else ''\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_4(self):...\n", "self.assertTrue(self.client.login(VAR_1='autotest-1', VAR_2='password'))\n" ]
[ "def setUp(self):...\n", "self.assertTrue(self.client.login(username='autotest-1', password='password'))\n" ]
[ 0, 5 ]
[ "FunctionDef'", "Expr'" ]
[ "def FUNC_12():...\n", "VAR_18 = requests.get(FUNC_2('healthy'))\n", "VAR_18.raise_for_status()\n", "return VAR_18.json()\n" ]
[ "def _ServerIsHealthy():...\n", "response = requests.get(_BuildUri('healthy'))\n", "response.raise_for_status()\n", "return response.json()\n" ]
[ 0, 7, 0, 0 ]
[ "FunctionDef'", "Assign'", "Expr'", "Return'" ]
[ "from itertools import chain\n", "from datetime import datetime\n", "from model_utils.managers import InheritanceManager\n", "from django.db import models\n", "from django.core.exceptions import ValidationError\n", "from django import forms\n", "from django.urls import reverse\n", "from .common_info import CommonInfo\n", "VAR_3 = VAR_0.OneToOneField('DataDocument', on_delete=models.CASCADE,\n primary_key=True)\n", "VAR_4 = VAR_0.CharField(max_length=500, null=True, blank=True)\n", "VAR_5 = VAR_0.CharField(max_length=25, null=True, blank=True)\n", "VAR_6 = VAR_0.CharField(max_length=50, null=True, blank=True)\n", "VAR_7 = VAR_0.ForeignKey('Script', on_delete=models.CASCADE,\n limit_choices_to={'script_type': 'EX'})\n", "VAR_8 = VAR_0.BooleanField(default=False, verbose_name='QA approved')\n", "VAR_9 = VAR_0.BooleanField(default=False, verbose_name='QA edited')\n", "VAR_10 = VAR_0.DateTimeField(null=True, blank=True, verbose_name=\n 'QA approval date')\n", "VAR_11 = VAR_0.ForeignKey('auth.User', on_delete=models.SET_NULL,\n verbose_name='QA approved by', null=True, blank=True)\n", "VAR_12 = VAR_0.ForeignKey('QAGroup', verbose_name='QA group', on_delete=\n models.SET_NULL, null=True, blank=True)\n", "VAR_13 = InheritanceManager()\n", "def __str__(self):...\n", "return str(self.data_document)\n" ]
[ "from itertools import chain\n", "from datetime import datetime\n", "from model_utils.managers import InheritanceManager\n", "from django.db import models\n", "from django.core.exceptions import ValidationError\n", "from django import forms\n", "from django.urls import reverse\n", "from .common_info import CommonInfo\n", "data_document = models.OneToOneField('DataDocument', on_delete=models.\n CASCADE, primary_key=True)\n", "prod_name = models.CharField(max_length=500, null=True, blank=True)\n", "doc_date = models.CharField(max_length=25, null=True, blank=True)\n", "rev_num = models.CharField(max_length=50, null=True, blank=True)\n", "extraction_script = models.ForeignKey('Script', on_delete=models.CASCADE,\n limit_choices_to={'script_type': 'EX'})\n", "qa_checked = models.BooleanField(default=False, verbose_name='QA approved')\n", "qa_edited = models.BooleanField(default=False, verbose_name='QA edited')\n", "qa_approved_date = models.DateTimeField(null=True, blank=True, verbose_name\n ='QA approval date')\n", "qa_approved_by = models.ForeignKey('auth.User', on_delete=models.SET_NULL,\n verbose_name='QA approved by', null=True, blank=True)\n", "qa_group = models.ForeignKey('QAGroup', verbose_name='QA group', on_delete=\n models.SET_NULL, null=True, blank=True)\n", "objects = InheritanceManager()\n", "def __str__(self):...\n", "return str(self.data_document)\n" ]
[ 6, 6, 0, 0, 6, 6, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "FunctionDef'", "Return'" ]
[ "from flask import request\n", "from flask_jwt_extended import jwt_required\n", "from flask_restplus import Namespace, Resource, fields\n", "from utilities import database_utilities\n", "VAR_0 = Namespace('admins', description='Information relating to system admins'\n )\n", "def FUNC_0(self):...\n", "\"\"\"docstring\"\"\"\n", "return database_utilities.execute_query('select * from admins')\n" ]
[ "from flask import request\n", "from flask_jwt_extended import jwt_required\n", "from flask_restplus import Namespace, Resource, fields\n", "from utilities import database_utilities\n", "api = Namespace('admins', description='Information relating to system admins')\n", "def get(self):...\n", "\"\"\"docstring\"\"\"\n", "return database_utilities.execute_query('select * from admins')\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Assign'", "FunctionDef'", "Docstring", "Return'" ]
[ "def FUNC_7(self, VAR_9):...\n", "\"\"\"docstring\"\"\"\n", "VAR_30 = False\n", "for rec in self:\n", "VAR_27 = rec._next_states()\n", "VAR_31 = self.env['crapo.transition'].search([('from_state', '=', VAR_30.id\n ), ('to_state', '=', VAR_9)], limit=1)\n", "if rec.state.id == VAR_9:\n", "return VAR_31\n", "VAR_30 = rec.state\n", "if not VAR_27:\n", "if VAR_9 not in VAR_27.ids:\n", "if VAR_30 is not False and VAR_30 != rec.state:\n", "VAR_30 = rec.state\n" ]
[ "def _get_transition(self, target_state_id):...\n", "\"\"\"docstring\"\"\"\n", "current_state = False\n", "for rec in self:\n", "next_states = rec._next_states()\n", "transition = self.env['crapo.transition'].search([('from_state', '=',\n current_state.id), ('to_state', '=', target_state_id)], limit=1)\n", "if rec.state.id == target_state_id:\n", "return transition\n", "current_state = rec.state\n", "if not next_states:\n", "if target_state_id not in next_states.ids:\n", "if current_state is not False and current_state != rec.state:\n", "current_state = rec.state\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "For", "Assign'", "Assign'", "Condition", "Return'", "Assign'", "Condition", "Condition", "Condition", "Assign'" ]
[ "def FUNC_16(VAR_0, VAR_5, VAR_14):...\n", "\"\"\"docstring\"\"\"\n", "VAR_23 = pd.read_sql_table(VAR_0, db.engine)\n", "print('EQUAL FREQUENCY DISCRETIZATION FAILED:\\n' + str(e))\n", "VAR_25 = len(VAR_23[VAR_5])\n", "VAR_26 = VAR_25 // VAR_14\n", "VAR_27 = list(VAR_23[VAR_5].sort_values())\n", "VAR_28 = 0\n", "VAR_29 = []\n", "while VAR_28 < VAR_25:\n", "if VAR_29[-1] != VAR_27[-1] and len(VAR_29) == VAR_14 + 1:\n", "VAR_29.append(VAR_27[VAR_28])\n", "VAR_29[-1] = VAR_27[-1]\n", "if VAR_29[-1] != VAR_27[-1] and len(VAR_29) != VAR_14 + 1:\n", "VAR_28 += VAR_26\n", "VAR_29[0] = VAR_29[0] - VAR_29[0] * 0.001\n", "VAR_29.append(VAR_27[-1])\n", "VAR_29[-1] = VAR_29[-1] + VAR_29[-1] * 0.001\n", "VAR_24 = VAR_5 + '_' + str(VAR_14) + '_eq_freq_intervals'\n", "FUNC_15(VAR_0, VAR_5, VAR_29, VAR_23, VAR_24)\n" ]
[ "def discretize_eq_freq(table_name, attr, intervals):...\n", "\"\"\"docstring\"\"\"\n", "df = pd.read_sql_table(table_name, db.engine)\n", "print('EQUAL FREQUENCY DISCRETIZATION FAILED:\\n' + str(e))\n", "attr_length = len(df[attr])\n", "elements_per_interval = attr_length // intervals\n", "sorted_data = list(df[attr].sort_values())\n", "selector = 0\n", "edge_list = []\n", "while selector < attr_length:\n", "if edge_list[-1] != sorted_data[-1] and len(edge_list) == intervals + 1:\n", "edge_list.append(sorted_data[selector])\n", "edge_list[-1] = sorted_data[-1]\n", "if edge_list[-1] != sorted_data[-1] and len(edge_list) != intervals + 1:\n", "selector += elements_per_interval\n", "edge_list[0] = edge_list[0] - edge_list[0] * 0.001\n", "edge_list.append(sorted_data[-1])\n", "edge_list[-1] = edge_list[-1] + edge_list[-1] * 0.001\n", "column_name = attr + '_' + str(intervals) + '_eq_freq_intervals'\n", "discretize_width(table_name, attr, edge_list, df, column_name)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Expr'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Condition", "Expr'", "Assign'", "Condition", "AugAssign'", "Assign'", "Expr'", "Assign'", "Assign'", "Expr'" ]
[ "def FUNC_3(VAR_2, VAR_3):...\n", "VAR_2.browser = webdriver.PhantomJS()\n", "VAR_2.browser.set_window_size(1280, 1024)\n", "VAR_2.browser.implicitly_wait(DEFAULT_IMPLICIT_WAIT_TIMEOUT_IN_S)\n", "VAR_2.browser.set_page_load_timeout(60)\n", "VAR_2.browser.get(HOMEPAGE_URL)\n" ]
[ "def before_feature(context, feature):...\n", "context.browser = webdriver.PhantomJS()\n", "context.browser.set_window_size(1280, 1024)\n", "context.browser.implicitly_wait(DEFAULT_IMPLICIT_WAIT_TIMEOUT_IN_S)\n", "context.browser.set_page_load_timeout(60)\n", "context.browser.get(HOMEPAGE_URL)\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'" ]
[ "@wraps(VAR_1)...\n", "VAR_17 = make_response(VAR_1(*VAR_6, **kwargs))\n", "VAR_17.headers['Last-Modified'] = datetime.now()\n", "VAR_17.headers['Cache-Control'\n ] = 'no-store, no-cache, must-revalidate, post-check=0, pre-check=0, max-age=0'\n", "VAR_17.headers['Pragma'] = 'no-cache'\n", "VAR_17.headers['Expires'] = '-1'\n", "return VAR_17\n" ]
[ "@wraps(view)...\n", "response = make_response(view(*args, **kwargs))\n", "response.headers['Last-Modified'] = datetime.now()\n", "response.headers['Cache-Control'\n ] = 'no-store, no-cache, must-revalidate, post-check=0, pre-check=0, max-age=0'\n", "response.headers['Pragma'] = 'no-cache'\n", "response.headers['Expires'] = '-1'\n", "return response\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_31():...\n", "yield self._bin_path\n", "if self._is_jdk:\n", "yield os.path.join(self.home, 'jre', 'bin')\n" ]
[ "def bin_paths():...\n", "yield self._bin_path\n", "if self._is_jdk:\n", "yield os.path.join(self.home, 'jre', 'bin')\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Condition", "Expr'" ]
[ "def FUNC_3(VAR_0, VAR_4, VAR_5):...\n", "VAR_7 = psycopg2.connect(dbname=pg_connection['database'], user=\n pg_connection['user'], password=pg_connection['password'], host=\n pg_connection['host'])\n", "VAR_8 = VAR_7.cursor()\n", "VAR_8.execute(\n \"\"\"\n UPDATE quests\n SET completed = '%s'\n WHERE id = %s;\n \"\"\",\n (VAR_5, VAR_4))\n", "VAR_7.commit()\n", "VAR_8.close()\n", "VAR_7.close()\n" ]
[ "def complete_quest(pg_connection, quest_id, completion):...\n", "conn = psycopg2.connect(dbname=pg_connection['database'], user=\n pg_connection['user'], password=pg_connection['password'], host=\n pg_connection['host'])\n", "cur = conn.cursor()\n", "cur.execute(\n \"\"\"\n UPDATE quests\n SET completed = '%s'\n WHERE id = %s;\n \"\"\",\n (completion, quest_id))\n", "conn.commit()\n", "cur.close()\n", "conn.close()\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'" ]
[ "def FUNC_2(self, VAR_8, VAR_9, VAR_10, VAR_5, VAR_11):...\n", "VAR_14 = 'string'.format(VAR_8, VAR_9, VAR_10, VAR_5, VAR_11)\n", "self.cur.execute(VAR_14)\n", "self.conn.commit()\n" ]
[ "def create_event(self, title, start_time, time_zone, server_id, description):...\n", "sql = (\n \"\"\"INSERT INTO events (title, start_time, time_zone, server_id, description)\n VALUES ('{0}', '{1}', '{2}', '{3}', '{4}')\n \"\"\"\n .format(title, start_time, time_zone, server_id, description))\n", "self.cur.execute(sql)\n", "self.conn.commit()\n" ]
[ 0, 4, 0, 0 ]
[ "FunctionDef'", "Assign'", "Expr'", "Expr'" ]
[ "def FUNC_15(self):...\n", "\"\"\"docstring\"\"\"\n", "VAR_7 = BokChoyTestSuite('', num_processes=2, verbosity=3)\n", "BokChoyTestSuite.verbosity_processes_string(VAR_7)\n" ]
[ "def test_invalid_verbosity_and_processes(self):...\n", "\"\"\"docstring\"\"\"\n", "suite = BokChoyTestSuite('', num_processes=2, verbosity=3)\n", "BokChoyTestSuite.verbosity_processes_string(suite)\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Expr'" ]
[ "def FUNC_4(VAR_0, VAR_6):...\n", "VAR_7 = psycopg2.connect(dbname=pg_connection['database'], user=\n pg_connection['user'], password=pg_connection['password'], host=\n pg_connection['host'])\n", "VAR_8 = VAR_7.cursor()\n", "VAR_8.execute(VAR_6)\n", "VAR_9 = list(VAR_8.fetchall())\n", "VAR_8.close()\n", "VAR_7.close()\n", "return VAR_9\n" ]
[ "def retrieve_quest_data(pg_connection, query):...\n", "conn = psycopg2.connect(dbname=pg_connection['database'], user=\n pg_connection['user'], password=pg_connection['password'], host=\n pg_connection['host'])\n", "cur = conn.cursor()\n", "cur.execute(query)\n", "results = list(cur.fetchall())\n", "cur.close()\n", "conn.close()\n", "return results\n" ]
[ 0, 0, 0, 4, 4, 0, 0, 4 ]
[ "FunctionDef'", "Assign'", "Assign'", "Expr'", "Assign'", "Expr'", "Expr'", "Return'" ]
[ "def FUNC_14(self):...\n", "return self._get_ancestors().values()\n" ]
[ "def get_ancestors(self):...\n", "return self._get_ancestors().values()\n" ]
[ 0, 1 ]
[ "FunctionDef'", "Return'" ]
[ "@data(('9981', 'B0046A36-3F1C-11E9-9A87-005056AA6F11', 0, 0, 0), ('9982',...\n", "VAR_13 = {'sender': 'Alex Roy <[email protected]>', 'subject':\n 'DO NOT MODIFY MESSAGE BELOW - JUST HIT `SEND`', 'date':\n 'Tue, 7 May 2019 17:34:17 +0000', 'content':\n f'job_number={VAR_6}string{VAR_7}\\r\\n'}\n", "VAR_14 = 'string'\n", "VAR_15 = 'string'\n", "if VAR_9 or VAR_10:\n", "conn.cursor().execute(VAR_14.format(VAR_6, VAR_9))\n", "if VAR_8:\n", "if VAR_9:\n", "VAR_30 = pd.read_sql(f'SELECT * FROM df_dilfo WHERE job_number={VAR_6}', conn)\n", "conn.cursor().execute(VAR_15.format(VAR_6, 1))\n", "conn.cursor().execute(VAR_15.format(VAR_6, 0))\n", "VAR_31 = pd.read_sql(f'SELECT * FROM df_matched WHERE job_number={VAR_6}', conn\n )\n", "process_as_form(VAR_13)\n", "VAR_32 = pd.read_sql(f'SELECT * FROM df_dilfo WHERE job_number={VAR_6}', conn)\n", "VAR_33 = pd.read_sql(f'SELECT * FROM df_matched WHERE job_number={VAR_6}', conn\n )\n", "self.assertEqual(len(VAR_32), 1)\n", "self.assertEqual(bool(VAR_32.iloc[0].closed), bool(VAR_9 or VAR_7))\n", "self.assertEqual(any(VAR_33.ground_truth), bool(VAR_9 or VAR_7))\n", "self.assertEqual(len(VAR_31) + bool(VAR_7 and not VAR_9), len(VAR_33))\n", "self.assertEqual(list(VAR_31.columns), list(VAR_33.columns))\n", "self.assertEqual(list(VAR_30.columns), list(VAR_32.columns))\n" ]
[ "@data(('9981', 'B0046A36-3F1C-11E9-9A87-005056AA6F11', 0, 0, 0), ('9982',...\n", "email_obj = {'sender': 'Alex Roy <[email protected]>', 'subject':\n 'DO NOT MODIFY MESSAGE BELOW - JUST HIT `SEND`', 'date':\n 'Tue, 7 May 2019 17:34:17 +0000', 'content':\n f'job_number={job_number}&title=TEST_ENTRY&city=Ottawa&address=2562+Del+Zotto+Ave.%2C+Ottawa%2C+Ontario&contractor=GCN&engineer=Goodkey&owner=Douglas+Stalker&quality=2&cc_email=&link_to_cert={dcn_key}\\r\\n'\n }\n", "fake_dilfo_insert = \"\"\"\n INSERT INTO df_dilfo (job_number, receiver_email, closed)\n VALUES ({}, '[email protected]', {})\n \"\"\"\n", "fake_match_insert = \"\"\"\n INSERT INTO df_matched (job_number, verifier, ground_truth)\n VALUES ({}, '[email protected]', {})\n \"\"\"\n", "if was_prev_closed or was_prev_tracked:\n", "conn.cursor().execute(fake_dilfo_insert.format(job_number, was_prev_closed))\n", "if was_prev_matched:\n", "if was_prev_closed:\n", "df_dilfo_pre = pd.read_sql(\n f'SELECT * FROM df_dilfo WHERE job_number={job_number}', conn)\n", "conn.cursor().execute(fake_match_insert.format(job_number, 1))\n", "conn.cursor().execute(fake_match_insert.format(job_number, 0))\n", "df_matched_pre = pd.read_sql(\n f'SELECT * FROM df_matched WHERE job_number={job_number}', conn)\n", "process_as_form(email_obj)\n", "df_dilfo_post = pd.read_sql(\n f'SELECT * FROM df_dilfo WHERE job_number={job_number}', conn)\n", "df_matched_post = pd.read_sql(\n f'SELECT * FROM df_matched WHERE job_number={job_number}', conn)\n", "self.assertEqual(len(df_dilfo_post), 1)\n", "self.assertEqual(bool(df_dilfo_post.iloc[0].closed), bool(was_prev_closed or\n dcn_key))\n", "self.assertEqual(any(df_matched_post.ground_truth), bool(was_prev_closed or\n dcn_key))\n", "self.assertEqual(len(df_matched_pre) + bool(dcn_key and not was_prev_closed\n ), len(df_matched_post))\n", "self.assertEqual(list(df_matched_pre.columns), list(df_matched_post.columns))\n", "self.assertEqual(list(df_dilfo_pre.columns), list(df_dilfo_post.columns))\n" ]
[ 0, 0, 4, 0, 0, 4, 0, 0, 4, 4, 4, 4, 0, 4, 4, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Assign'", "Assign'", "Assign'", "Condition", "Expr'", "Condition", "Condition", "Assign'", "Expr'", "Expr'", "Assign'", "Expr'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'" ]
[ "def FUNC_2(self):...\n", "VAR_14 = utils.GetUnusedLocalhostPort()\n", "self._temp_options_filename = options_file.name\n", "json.dump(dict(self._user_options), options_file)\n", "options_file.flush()\n", "VAR_21 = [utils.PathToPythonInterpreter(), FUNC_0(), '--port={0}'.format(\n VAR_14), '--options_file={0}'.format(options_file.name), '--log={0}'.\n format(self._user_options['server_log_level']),\n '--idle_suicide_seconds={0}'.format(VAR_4)]\n", "if not self._user_options['server_use_vim_stdout']:\n", "VAR_26 = VAR_0.path.join(utils.PathToTempDir(), 'server_{port}_{std}.log')\n", "self._server_popen = utils.SafePopen(VAR_21, stdout=PIPE, stderr=PIPE)\n", "self._server_stdout = VAR_26.format(port=server_port, std='stdout')\n", "BaseRequest.server_location = 'http://localhost:' + str(VAR_14)\n", "self._server_stderr = VAR_26.format(port=server_port, std='stderr')\n", "self._NotifyUserIfServerCrashed()\n", "VAR_21.append('--stdout={0}'.format(self._server_stdout))\n", "VAR_21.append('--stderr={0}'.format(self._server_stderr))\n", "if self._user_options['server_keep_logfiles']:\n", "VAR_21.append('--keep_logfiles')\n" ]
[ "def _SetupServer(self):...\n", "server_port = utils.GetUnusedLocalhostPort()\n", "self._temp_options_filename = options_file.name\n", "json.dump(dict(self._user_options), options_file)\n", "options_file.flush()\n", "args = [utils.PathToPythonInterpreter(), _PathToServerScript(),\n '--port={0}'.format(server_port), '--options_file={0}'.format(\n options_file.name), '--log={0}'.format(self._user_options[\n 'server_log_level']), '--idle_suicide_seconds={0}'.format(\n SERVER_IDLE_SUICIDE_SECONDS)]\n", "if not self._user_options['server_use_vim_stdout']:\n", "filename_format = os.path.join(utils.PathToTempDir(), 'server_{port}_{std}.log'\n )\n", "self._server_popen = utils.SafePopen(args, stdout=PIPE, stderr=PIPE)\n", "self._server_stdout = filename_format.format(port=server_port, std='stdout')\n", "BaseRequest.server_location = 'http://localhost:' + str(server_port)\n", "self._server_stderr = filename_format.format(port=server_port, std='stderr')\n", "self._NotifyUserIfServerCrashed()\n", "args.append('--stdout={0}'.format(self._server_stdout))\n", "args.append('--stderr={0}'.format(self._server_stderr))\n", "if self._user_options['server_keep_logfiles']:\n", "args.append('--keep_logfiles')\n" ]
[ 0, 0, 7, 7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Expr'", "Expr'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'", "Condition", "Expr'" ]
[ "def FUNC_2(VAR_1):...\n", "\"\"\"docstring\"\"\"\n", "VAR_4 = FUNC_1(VAR_1)\n", "json.dump(VAR_4, sys.stdout, indent=2)\n" ]
[ "def dump_yml_filedir(root_dir):...\n", "\"\"\"docstring\"\"\"\n", "result = load_yml_filedir(root_dir)\n", "json.dump(result, sys.stdout, indent=2)\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Expr'" ]
[ "def FUNC_6(self, VAR_13, VAR_14='tests'):...\n", "\"\"\"docstring\"\"\"\n", "self._logger.csv(VAR_13 + '_mappose', ['X', 'Y', 'yaw'], [self.map_pos.x,\n self.map_pos.y, self.map_angle], VAR_14=folder)\n" ]
[ "def csvLogMap(self, test_name, folder='tests'):...\n", "\"\"\"docstring\"\"\"\n", "self._logger.csv(test_name + '_mappose', ['X', 'Y', 'yaw'], [self.map_pos.x,\n self.map_pos.y, self.map_angle], folder=folder)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Expr'" ]
[ "def FUNC_18(self):...\n", "return FUNC_18.dumps(self.properties())\n" ]
[ "def json(self):...\n", "return json.dumps(self.properties())\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_0(VAR_2):...\n", "\"\"\"docstring\"\"\"\n", "if VAR_2.request.method == 'GET':\n", "VAR_6 = VAR_2.request.session.get_csrf_token()\n", "if VAR_2.request.cookies.get('XSRF-TOKEN') != VAR_6:\n", "VAR_2.response.set_cookie(VAR_0, VAR_6)\n" ]
[ "def on_GET_request_setup_csrf_cookie(ev):...\n", "\"\"\"docstring\"\"\"\n", "if ev.request.method == 'GET':\n", "token = ev.request.session.get_csrf_token()\n", "if ev.request.cookies.get('XSRF-TOKEN') != token:\n", "ev.response.set_cookie(COOKIE_NAME, token)\n" ]
[ 0, 0, 0, 0, 0, 5 ]
[ "FunctionDef'", "Docstring", "Condition", "Assign'", "Condition", "Expr'" ]
[ "import psycopg2\n", "from pyMetricServer.config import *\n", "import time\n", "VAR_0 = psycopg2.connect(host=DATABASE_HOST, port=DATABASE_PORT, user=\n DATABASE_USER, password=DATABASE_PASS, VAR_0=DATABASE_NAME)\n", "VAR_1 = VAR_0.cursor()\n", "VAR_1.execute('string')\n", "VAR_1.execute('string')\n", "VAR_0.commit()\n", "VAR_1.close()\n", "def FUNC_0(VAR_2=None, VAR_3=None, VAR_4=None, VAR_5=None, VAR_6=None,...\n", "VAR_8 = []\n", "VAR_1 = VAR_0.cursor()\n", "VAR_9 = []\n", "VAR_10 = 'SELECT Id, Time, Origin, Key, Value FROM log_metric '\n", "if VAR_2 != None or VAR_3 != None or VAR_4 != None or VAR_5 != None:\n", "VAR_10 += 'WHERE '\n", "if VAR_2 != None:\n", "VAR_10 += 'Time >= %s AND '\n", "if VAR_3 != None:\n", "VAR_9.append(VAR_2)\n", "VAR_10 += 'Time <= %s AND '\n", "if VAR_4 != None:\n", "VAR_9.append(VAR_3)\n", "VAR_10 += 'Origin = %s AND '\n", "if VAR_5 != None:\n", "VAR_9.append(VAR_4)\n", "VAR_10 += 'Key = %s AND '\n", "VAR_10 = VAR_10.strip('AND ')\n", "VAR_9.append(VAR_5)\n", "VAR_10 += ' '\n", "if VAR_7 != None and VAR_7[0] != None:\n", "if VAR_7[1]:\n", "if VAR_6 != None:\n", "VAR_10 += 'ORDER BY %s DESC ' % VAR_7[0]\n", "VAR_10 += 'ORDER BY %s ASC ' % VAR_7[0]\n", "VAR_10 += 'LIMIT %s '\n", "VAR_1.execute(VAR_10, tuple(VAR_9))\n", "VAR_9.append(VAR_6)\n", "for row in VAR_1:\n", "VAR_8.append({'Id': str(row[0]), 'Time': str(row[1]), 'Origin': str(row[2]),\n 'Key': str(row[3]), 'Value': str(row[4])})\n", "return VAR_8\n" ]
[ "import psycopg2\n", "from pyMetricServer.config import *\n", "import time\n", "database = psycopg2.connect(host=DATABASE_HOST, port=DATABASE_PORT, user=\n DATABASE_USER, password=DATABASE_PASS, database=DATABASE_NAME)\n", "cursor = database.cursor()\n", "cursor.execute(\n 'CREATE TABLE IF NOT EXISTS log_messages (Id BIGSERIAL, Time INTEGER, Origin TEXT, Message TEXT, Type INTEGER);'\n )\n", "cursor.execute(\n 'CREATE TABLE IF NOT EXISTS log_metric (Id BIGSERIAL, Time INTEGER, Origin TEXT, Key TEXT, Value DOUBLE PRECISION)'\n )\n", "database.commit()\n", "cursor.close()\n", "def getMetric(timefrom=None, timeto=None, origin=None, key=None, count=None,...\n", "results = []\n", "cursor = database.cursor()\n", "params = []\n", "query = 'SELECT Id, Time, Origin, Key, Value FROM log_metric '\n", "if timefrom != None or timeto != None or origin != None or key != None:\n", "query += 'WHERE '\n", "if timefrom != None:\n", "query += 'Time >= %s AND '\n", "if timeto != None:\n", "params.append(timefrom)\n", "query += 'Time <= %s AND '\n", "if origin != None:\n", "params.append(timeto)\n", "query += 'Origin = %s AND '\n", "if key != None:\n", "params.append(origin)\n", "query += 'Key = %s AND '\n", "query = query.strip('AND ')\n", "params.append(key)\n", "query += ' '\n", "if order != None and order[0] != None:\n", "if order[1]:\n", "if count != None:\n", "query += 'ORDER BY %s DESC ' % order[0]\n", "query += 'ORDER BY %s ASC ' % order[0]\n", "query += 'LIMIT %s '\n", "cursor.execute(query, tuple(params))\n", "params.append(count)\n", "for row in cursor:\n", "results.append({'Id': str(row[0]), 'Time': str(row[1]), 'Origin': str(row[2\n ]), 'Key': str(row[3]), 'Value': str(row[4])})\n", "return results\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 4, 0, 4, 4, 0, 0, 0, 0, 0, 0 ]
[ "Import'", "ImportFrom'", "Import'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "AugAssign'", "Condition", "AugAssign'", "Condition", "Expr'", "AugAssign'", "Condition", "Expr'", "AugAssign'", "Condition", "Expr'", "AugAssign'", "Assign'", "Expr'", "AugAssign'", "Condition", "Condition", "Condition", "AugAssign'", "AugAssign'", "AugAssign'", "Expr'", "Expr'", "For", "Expr'", "Return'" ]
[ "def FUNC_7(self, VAR_5):...\n", "\"\"\"docstring\"\"\"\n", "VAR_13 = (KeyError, ValueError, TypeError, AttributeError, SQLAlchemyError,\n DokomoError)\n", "if isinstance(VAR_5, tornado.web.HTTPError):\n", "VAR_33 = exc.HttpError(VAR_5.log_message)\n", "if isinstance(VAR_5, VAR_13):\n", "VAR_33.status = VAR_5.status_code\n", "VAR_5 = exc.BadRequest(VAR_5)\n", "return super().handle_error(VAR_5)\n", "VAR_5 = VAR_33\n" ]
[ "def handle_error(self, err):...\n", "\"\"\"docstring\"\"\"\n", "understood = (KeyError, ValueError, TypeError, AttributeError,\n SQLAlchemyError, DokomoError)\n", "if isinstance(err, tornado.web.HTTPError):\n", "restless_error = exc.HttpError(err.log_message)\n", "if isinstance(err, understood):\n", "restless_error.status = err.status_code\n", "err = exc.BadRequest(err)\n", "return super().handle_error(err)\n", "err = restless_error\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Assign'", "Return'", "Assign'" ]
[ "def FUNC_2(VAR_0, VAR_1):...\n", "VAR_5 = frappe.session.user\n", "VAR_6 = get_leave_allocation_records(VAR_0.to_date)\n", "VAR_7 = get_leave_allocation_records(VAR_0.from_date)\n", "VAR_8 = frappe.get_all('Employee', VAR_0={'status': 'Active', 'company':\n filters.company}, fields=['name', 'employee_name', 'department', 'user_id']\n )\n", "VAR_4 = []\n", "for employee in VAR_8:\n", "VAR_12 = FUNC_3(employee.department)\n", "return VAR_4\n", "if len(VAR_12) and VAR_5 in VAR_12 or VAR_5 in ['Administrator', employee.\n", "VAR_13 = [employee.name, employee.employee_name, employee.department]\n", "for leave_type in VAR_1:\n", "VAR_14 = get_approved_leaves_for_period(employee.name, leave_type, VAR_0.\n from_date, VAR_0.to_date)\n", "VAR_4.append(VAR_13)\n", "VAR_15 = get_leave_balance_on(employee.name, leave_type, VAR_0.from_date,\n VAR_7.get(employee.name, frappe._dict()))\n", "VAR_16 = get_leave_balance_on(employee.name, leave_type, VAR_0.to_date,\n VAR_6.get(employee.name, frappe._dict()))\n", "VAR_13 += [VAR_15, VAR_14, VAR_16]\n" ]
[ "def get_data(filters, leave_types):...\n", "user = frappe.session.user\n", "allocation_records_based_on_to_date = get_leave_allocation_records(filters.\n to_date)\n", "allocation_records_based_on_from_date = get_leave_allocation_records(filters\n .from_date)\n", "active_employees = frappe.get_all('Employee', filters={'status': 'Active',\n 'company': filters.company}, fields=['name', 'employee_name',\n 'department', 'user_id'])\n", "data = []\n", "for employee in active_employees:\n", "leave_approvers = get_approvers(employee.department)\n", "return data\n", "if len(leave_approvers) and user in leave_approvers or user in ['Administrator'\n", "row = [employee.name, employee.employee_name, employee.department]\n", "for leave_type in leave_types:\n", "leaves_taken = get_approved_leaves_for_period(employee.name, leave_type,\n filters.from_date, filters.to_date)\n", "data.append(row)\n", "opening = get_leave_balance_on(employee.name, leave_type, filters.from_date,\n allocation_records_based_on_from_date.get(employee.name, frappe._dict()))\n", "closing = get_leave_balance_on(employee.name, leave_type, filters.to_date,\n allocation_records_based_on_to_date.get(employee.name, frappe._dict()))\n", "row += [opening, leaves_taken, closing]\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 4, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "For", "Assign'", "Return'", "Condition", "Assign'", "For", "Assign'", "Expr'", "Assign'", "Assign'", "AugAssign'" ]
[ "@property...\n", "\"\"\"docstring\"\"\"\n", "return set(VAR_28 for VAR_28 in self.input if not VAR_28.exists and not \n VAR_28 in self.subworkflow_input)\n" ]
[ "@property...\n", "\"\"\"docstring\"\"\"\n", "return set(f for f in self.input if not f.exists and not f in self.\n subworkflow_input)\n" ]
[ 0, 0, 0 ]
[ "Condition", "Docstring", "Return'" ]
[ "def FUNC_0(VAR_0):...\n", "\"\"\"docstring\"\"\"\n", "VAR_1 = True if VAR_0.find(':') != -1 else False\n", "if VAR_0[-4:] == '.rpm':\n", "VAR_0 = VAR_0[:-4]\n", "VAR_2 = VAR_0.rfind('.')\n", "VAR_3 = VAR_0[VAR_2 + 1:]\n", "VAR_4 = VAR_0[:VAR_2].rfind('-')\n", "VAR_5 = VAR_0[VAR_4 + 1:VAR_2]\n", "if VAR_1:\n", "VAR_10 = VAR_0[:VAR_4].rfind(':')\n", "VAR_10 = VAR_0[:VAR_4].rfind('-')\n", "VAR_6 = VAR_0[VAR_10 + 1:VAR_4]\n", "if VAR_1:\n", "VAR_11 = VAR_0[:VAR_10].rfind('-')\n", "VAR_11 = VAR_10\n", "VAR_12 = VAR_0[VAR_11 + 1:VAR_10]\n", "VAR_12 = '0'\n", "VAR_7 = VAR_0[:VAR_11]\n", "return VAR_7, VAR_6, VAR_5, VAR_12, VAR_3\n" ]
[ "def split_filename(filename):...\n", "\"\"\"docstring\"\"\"\n", "is_epoch = True if filename.find(':') != -1 else False\n", "if filename[-4:] == '.rpm':\n", "filename = filename[:-4]\n", "arch_index = filename.rfind('.')\n", "arch = filename[arch_index + 1:]\n", "rel_index = filename[:arch_index].rfind('-')\n", "rel = filename[rel_index + 1:arch_index]\n", "if is_epoch:\n", "ver_index = filename[:rel_index].rfind(':')\n", "ver_index = filename[:rel_index].rfind('-')\n", "ver = filename[ver_index + 1:rel_index]\n", "if is_epoch:\n", "epoch_index = filename[:ver_index].rfind('-')\n", "epoch_index = ver_index\n", "epoch = filename[epoch_index + 1:ver_index]\n", "epoch = '0'\n", "name = filename[:epoch_index]\n", "return name, ver, rel, epoch, arch\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "\"\"\"string\"\"\"\n", "from __future__ import unicode_literals\n", "from django.conf import settings\n", "from django.conf.urls import include\n", "from django.conf.urls import patterns\n", "from django.conf.urls import url\n", "from api import routers\n", "from api import views\n", "VAR_0 = routers.ApiRouter()\n", "VAR_1 = patterns('', url('^', include(VAR_0.urls)), url(\n '^apps/(?P<id>{})/config/?'.format(settings.APP_URL_REGEX), views.\n AppConfigViewSet.as_view({'get': 'retrieve', 'post': 'create'})), url(\n '^apps/(?P<id>{})/builds/(?P<uuid>[-_\\\\w]+)/?'.format(settings.\n APP_URL_REGEX), views.AppBuildViewSet.as_view({'get': 'retrieve'})),\n url('^apps/(?P<id>{})/builds/?'.format(settings.APP_URL_REGEX), views.\n AppBuildViewSet.as_view({'get': 'list', 'post': 'create'})), url(\n '^apps/(?P<id>{})/releases/v(?P<version>[0-9]+)/?'.format(settings.\n APP_URL_REGEX), views.AppReleaseViewSet.as_view({'get': 'retrieve'})),\n url('^apps/(?P<id>{})/releases/rollback/?'.format(settings.\n APP_URL_REGEX), views.AppReleaseViewSet.as_view({'post': 'rollback'})),\n url('^apps/(?P<id>{})/releases/?'.format(settings.APP_URL_REGEX), views\n .AppReleaseViewSet.as_view({'get': 'list'})), url(\n '^apps/(?P<id>{})/containers/(?P<type>[-_\\\\w]+)/(?P<num>[-_\\\\w]+)/?'.\n format(settings.APP_URL_REGEX), views.AppContainerViewSet.as_view({\n 'get': 'retrieve'})), url(\n '^apps/(?P<id>{})/containers/(?P<type>[-_\\\\w.]+)/?'.format(settings.\n APP_URL_REGEX), views.AppContainerViewSet.as_view({'get': 'list'})),\n url('^apps/(?P<id>{})/containers/?'.format(settings.APP_URL_REGEX),\n views.AppContainerViewSet.as_view({'get': 'list'})), url(\n '^apps/(?P<id>{})/domains/(?P<domain>[-\\\\._\\\\w]+)/?'.format(settings.\n APP_URL_REGEX), views.DomainViewSet.as_view({'delete': 'destroy'})),\n url('^apps/(?P<id>{})/domains/?'.format(settings.APP_URL_REGEX), views.\n DomainViewSet.as_view({'post': 'create', 'get': 'list'})), url(\n '^apps/(?P<id>{})/scale/?'.format(settings.APP_URL_REGEX), views.\n AppViewSet.as_view({'post': 'scale'})), url('^apps/(?P<id>{})/logs/?'.\n format(settings.APP_URL_REGEX), views.AppViewSet.as_view({'get': 'logs'\n })), url('^apps/(?P<id>{})/run/?'.format(settings.APP_URL_REGEX), views\n .AppViewSet.as_view({'post': 'run'})), url(\n '^apps/(?P<id>{})/perms/(?P<username>[-_\\\\w]+)/?'.format(settings.\n APP_URL_REGEX), views.AppPermsViewSet.as_view({'delete': 'destroy'})),\n url('^apps/(?P<id>{})/perms/?'.format(settings.APP_URL_REGEX), views.\n AppPermsViewSet.as_view({'get': 'list', 'post': 'create'})), url(\n '^apps/(?P<id>{})/?'.format(settings.APP_URL_REGEX), views.AppViewSet.\n as_view({'get': 'retrieve', 'delete': 'destroy'})), url('^apps/?',\n views.AppViewSet.as_view({'get': 'list', 'post': 'create'})), url(\n '^keys/(?P<id>.+)/?', views.KeyViewSet.as_view({'get': 'retrieve',\n 'delete': 'destroy'})), url('^keys/?', views.KeyViewSet.as_view({'get':\n 'list', 'post': 'create'})), url('^hooks/push/?', views.PushHookViewSet\n .as_view({'post': 'create'})), url('^hooks/build/?', views.\n BuildHookViewSet.as_view({'post': 'create'})), url('^hooks/config/?',\n views.ConfigHookViewSet.as_view({'post': 'create'})), url(\n '^auth/register/?', views.UserRegistrationView.as_view({'post':\n 'create'})), url('^auth/cancel/?', views.UserCancellationView.as_view({\n 'delete': 'destroy'})), url('^auth/', include('rest_framework.urls',\n namespace='rest_framework')), url('^generate-api-key/',\n 'rest_framework.authtoken.views.obtain_auth_token'), url(\n '^admin/perms/(?P<username>[-_\\\\w]+)/?', views.AdminPermsViewSet.\n as_view({'delete': 'destroy'})), url('^admin/perms/?', views.\n AdminPermsViewSet.as_view({'get': 'list', 'post': 'create'})))\n" ]
[ "\"\"\"\nRESTful URL patterns and routing for the Deis API app.\n\n\nApplications\n============\n\n.. http:get:: /api/apps/(string:id)/\n\n Retrieve a :class:`~api.models.App` by its `id`.\n\n.. http:delete:: /api/apps/(string:id)/\n\n Destroy a :class:`~api.models.App` by its `id`.\n\n.. http:get:: /api/apps/\n\n List all :class:`~api.models.App`\\\\s.\n\n.. http:post:: /api/apps/\n\n Create a new :class:`~api.models.App`.\n\n\nApplication Release Components\n------------------------------\n\n.. http:get:: /api/apps/(string:id)/config/\n\n List all :class:`~api.models.Config`\\\\s.\n\n.. http:post:: /api/apps/(string:id)/config/\n\n Create a new :class:`~api.models.Config`.\n\n.. http:get:: /api/apps/(string:id)/builds/(string:uuid)/\n\n Retrieve a :class:`~api.models.Build` by its `uuid`.\n\n.. http:get:: /api/apps/(string:id)/builds/\n\n List all :class:`~api.models.Build`\\\\s.\n\n.. http:post:: /api/apps/(string:id)/builds/\n\n Create a new :class:`~api.models.Build`.\n\n.. http:get:: /api/apps/(string:id)/releases/(int:version)/\n\n Retrieve a :class:`~api.models.Release` by its `version`.\n\n.. http:get:: /api/apps/(string:id)/releases/\n\n List all :class:`~api.models.Release`\\\\s.\n\n.. http:post:: /api/apps/(string:id)/releases/rollback/\n\n Rollback to a previous :class:`~api.models.Release`.\n\n\nApplication Infrastructure\n--------------------------\n\n.. http:get:: /api/apps/(string:id)/containers/(string:type)/(int:num)/\n\n List all :class:`~api.models.Container`\\\\s.\n\n.. http:get:: /api/apps/(string:id)/containers/(string:type)/\n\n List all :class:`~api.models.Container`\\\\s.\n\n.. http:get:: /api/apps/(string:id)/containers/\n\n List all :class:`~api.models.Container`\\\\s.\n\n\nApplication Domains\n-------------------\n\n\n.. http:delete:: /api/apps/(string:id)/domains/(string:hostname)\n\n Destroy a :class:`~api.models.Domain` by its `hostname`\n\n.. http:get:: /api/apps/(string:id)/domains/\n\n List all :class:`~api.models.Domain`\\\\s.\n\n.. http:post:: /api/apps/(string:id)/domains/\n\n Create a new :class:`~api.models.Domain`\\\\s.\n\n\nApplication Actions\n-------------------\n\n.. http:post:: /api/apps/(string:id)/scale/\n\n See also\n :meth:`AppViewSet.scale() <api.views.AppViewSet.scale>`\n\n.. http:get:: /api/apps/(string:id)/logs/\n\n See also\n :meth:`AppViewSet.logs() <api.views.AppViewSet.logs>`\n\n.. http:post:: /api/apps/(string:id)/run/\n\n See also\n :meth:`AppViewSet.run() <api.views.AppViewSet.run>`\n\n\nApplication Sharing\n===================\n\n.. http:delete:: /api/apps/(string:id)/perms/(string:username)/\n\n Destroy an app permission by its `username`.\n\n.. http:get:: /api/apps/(string:id)/perms/\n\n List all permissions granted to this app.\n\n.. http:post:: /api/apps/(string:id)/perms/\n\n Create a new app permission.\n\n\nKeys\n====\n\n.. http:get:: /api/keys/(string:id)/\n\n Retrieve a :class:`~api.models.Key` by its `id`.\n\n.. http:delete:: /api/keys/(string:id)/\n\n Destroy a :class:`~api.models.Key` by its `id`.\n\n.. http:get:: /api/keys/\n\n List all :class:`~api.models.Key`\\\\s.\n\n.. http:post:: /api/keys/\n\n Create a new :class:`~api.models.Key`.\n\n\nAPI Hooks\n=========\n\n.. http:post:: /api/hooks/push/\n\n Create a new :class:`~api.models.Push`.\n\n.. http:post:: /api/hooks/build/\n\n Create a new :class:`~api.models.Build`.\n\n.. http:post:: /api/hooks/config/\n\n Retrieve latest application :class:`~api.models.Config`.\n\n\nAuth\n====\n\n.. http:post:: /api/auth/register/\n\n Create a new User.\n\n.. http:delete:: /api/auth/register/\n\n Destroy the logged-in User.\n\n.. http:post:: /api/auth/login\n\n Authenticate for the REST framework.\n\n.. http:post:: /api/auth/logout\n\n Clear authentication for the REST framework.\n\n.. http:get:: /api/generate-api-key/\n\n Generate an API key.\n\n\nAdmin Sharing\n=============\n\n.. http:delete:: /api/admin/perms/(string:username)/\n\n Destroy an admin permission by its `username`.\n\n.. http:get:: /api/admin/perms/\n\n List all admin permissions granted.\n\n.. http:post:: /api/admin/perms/\n\n Create a new admin permission.\n\n\"\"\"\n", "from __future__ import unicode_literals\n", "from django.conf import settings\n", "from django.conf.urls import include\n", "from django.conf.urls import patterns\n", "from django.conf.urls import url\n", "from api import routers\n", "from api import views\n", "router = routers.ApiRouter()\n", "urlpatterns = patterns('', url('^', include(router.urls)), url(\n '^apps/(?P<id>{})/config/?'.format(settings.APP_URL_REGEX), views.\n AppConfigViewSet.as_view({'get': 'retrieve', 'post': 'create'})), url(\n '^apps/(?P<id>{})/builds/(?P<uuid>[-_\\\\w]+)/?'.format(settings.\n APP_URL_REGEX), views.AppBuildViewSet.as_view({'get': 'retrieve'})),\n url('^apps/(?P<id>{})/builds/?'.format(settings.APP_URL_REGEX), views.\n AppBuildViewSet.as_view({'get': 'list', 'post': 'create'})), url(\n '^apps/(?P<id>{})/releases/v(?P<version>[0-9]+)/?'.format(settings.\n APP_URL_REGEX), views.AppReleaseViewSet.as_view({'get': 'retrieve'})),\n url('^apps/(?P<id>{})/releases/rollback/?'.format(settings.\n APP_URL_REGEX), views.AppReleaseViewSet.as_view({'post': 'rollback'})),\n url('^apps/(?P<id>{})/releases/?'.format(settings.APP_URL_REGEX), views\n .AppReleaseViewSet.as_view({'get': 'list'})), url(\n '^apps/(?P<id>{})/containers/(?P<type>[-_\\\\w]+)/(?P<num>[-_\\\\w]+)/?'.\n format(settings.APP_URL_REGEX), views.AppContainerViewSet.as_view({\n 'get': 'retrieve'})), url(\n '^apps/(?P<id>{})/containers/(?P<type>[-_\\\\w.]+)/?'.format(settings.\n APP_URL_REGEX), views.AppContainerViewSet.as_view({'get': 'list'})),\n url('^apps/(?P<id>{})/containers/?'.format(settings.APP_URL_REGEX),\n views.AppContainerViewSet.as_view({'get': 'list'})), url(\n '^apps/(?P<id>{})/domains/(?P<domain>[-\\\\._\\\\w]+)/?'.format(settings.\n APP_URL_REGEX), views.DomainViewSet.as_view({'delete': 'destroy'})),\n url('^apps/(?P<id>{})/domains/?'.format(settings.APP_URL_REGEX), views.\n DomainViewSet.as_view({'post': 'create', 'get': 'list'})), url(\n '^apps/(?P<id>{})/scale/?'.format(settings.APP_URL_REGEX), views.\n AppViewSet.as_view({'post': 'scale'})), url('^apps/(?P<id>{})/logs/?'.\n format(settings.APP_URL_REGEX), views.AppViewSet.as_view({'get': 'logs'\n })), url('^apps/(?P<id>{})/run/?'.format(settings.APP_URL_REGEX), views\n .AppViewSet.as_view({'post': 'run'})), url(\n '^apps/(?P<id>{})/perms/(?P<username>[-_\\\\w]+)/?'.format(settings.\n APP_URL_REGEX), views.AppPermsViewSet.as_view({'delete': 'destroy'})),\n url('^apps/(?P<id>{})/perms/?'.format(settings.APP_URL_REGEX), views.\n AppPermsViewSet.as_view({'get': 'list', 'post': 'create'})), url(\n '^apps/(?P<id>{})/?'.format(settings.APP_URL_REGEX), views.AppViewSet.\n as_view({'get': 'retrieve', 'delete': 'destroy'})), url('^apps/?',\n views.AppViewSet.as_view({'get': 'list', 'post': 'create'})), url(\n '^keys/(?P<id>.+)/?', views.KeyViewSet.as_view({'get': 'retrieve',\n 'delete': 'destroy'})), url('^keys/?', views.KeyViewSet.as_view({'get':\n 'list', 'post': 'create'})), url('^hooks/push/?', views.PushHookViewSet\n .as_view({'post': 'create'})), url('^hooks/build/?', views.\n BuildHookViewSet.as_view({'post': 'create'})), url('^hooks/config/?',\n views.ConfigHookViewSet.as_view({'post': 'create'})), url(\n '^auth/register/?', views.UserRegistrationView.as_view({'post':\n 'create'})), url('^auth/cancel/?', views.UserCancellationView.as_view({\n 'delete': 'destroy'})), url('^auth/', include('rest_framework.urls',\n namespace='rest_framework')), url('^generate-api-key/',\n 'rest_framework.authtoken.views.obtain_auth_token'), url(\n '^admin/perms/(?P<username>[-_\\\\w]+)/?', views.AdminPermsViewSet.\n as_view({'delete': 'destroy'})), url('^admin/perms/?', views.\n AdminPermsViewSet.as_view({'get': 'list', 'post': 'create'})))\n" ]
[ 5, 0, 0, 0, 0, 0, 0, 0, 0, 5 ]
[ "Expr'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Assign'", "Assign'" ]
[ "@user_passes_test(user_is_superuser)...\n", "VAR_8 = Project.objects.all()\n", "for VAR_17 in VAR_8:\n", "create_testcases_celery.delay(VAR_17.id)\n", "return redirect('testcases:testcases')\n" ]
[ "@user_passes_test(user_is_superuser)...\n", "projects = Project.objects.all()\n", "for project in projects:\n", "create_testcases_celery.delay(project.id)\n", "return redirect('testcases:testcases')\n" ]
[ 0, 0, 0, 0, 0 ]
[ "Condition", "Assign'", "For", "Expr'", "Return'" ]
[ "def FUNC_11(self):...\n", "VAR_20 = []\n", "VAR_18 = threading.Event()\n", "self.mock(VAR_18, 'wait', self.fail)\n", "self.mock(bot_main, 'run_manifest', self.fail)\n", "self.mock(bot_main, 'update_bot', lambda *VAR_24: VAR_20.append(VAR_24))\n", "self.expected_requests([(\n 'https://localhost:1/auth/api/v1/accounts/self/xsrf_token', {'data': {},\n 'headers': {'X-XSRF-Token-Request': '1'}}, {'xsrf_token': 'token'}), (\n 'https://localhost:1/swarming/api/v1/bot/poll', {'data': self.\n attributes, 'headers': {'X-XSRF-Token': 'token'}}, {'cmd': 'update',\n 'version': '123'})])\n", "self.assertTrue(bot_main.poll_server(self.bot, VAR_18))\n", "self.assertEqual([(self.bot, '123')], VAR_20)\n" ]
[ "def test_poll_server_update(self):...\n", "update = []\n", "bit = threading.Event()\n", "self.mock(bit, 'wait', self.fail)\n", "self.mock(bot_main, 'run_manifest', self.fail)\n", "self.mock(bot_main, 'update_bot', lambda *args: update.append(args))\n", "self.expected_requests([(\n 'https://localhost:1/auth/api/v1/accounts/self/xsrf_token', {'data': {},\n 'headers': {'X-XSRF-Token-Request': '1'}}, {'xsrf_token': 'token'}), (\n 'https://localhost:1/swarming/api/v1/bot/poll', {'data': self.\n attributes, 'headers': {'X-XSRF-Token': 'token'}}, {'cmd': 'update',\n 'version': '123'})])\n", "self.assertTrue(bot_main.poll_server(self.bot, bit))\n", "self.assertEqual([(self.bot, '123')], update)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'" ]
[ "def __new__(VAR_10, VAR_11, VAR_12, VAR_13):...\n", "for VAR_47 in webapp2.WSGIApplication.allowed_methods:\n", "VAR_48 = VAR_13.get(VAR_47.lower())\n", "return type.__new__(VAR_10, VAR_11, VAR_12, VAR_13)\n", "if VAR_48 and not api.is_decorated(VAR_48):\n" ]
[ "def __new__(mcs, name, bases, attributes):...\n", "for method in webapp2.WSGIApplication.allowed_methods:\n", "func = attributes.get(method.lower())\n", "return type.__new__(mcs, name, bases, attributes)\n", "if func and not api.is_decorated(func):\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "For", "Assign'", "Return'", "Condition" ]
[ "def FUNC_8(self):...\n", "self._WriteHeader('text/html', VAR_22=404)\n", "self.wfile.write('<html><body>%s not found</body></html>' % self.path)\n" ]
[ "def _DoNotFound(self):...\n", "self._WriteHeader('text/html', status_code=404)\n", "self.wfile.write('<html><body>%s not found</body></html>' % self.path)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Expr'" ]
[ "def FUNC_2(self, VAR_5):...\n", "if self.closed:\n", "if self._adapter is not None:\n", "assert self._session is None\n" ]
[ "def start_debugging(self, launchcfg):...\n", "if self.closed:\n", "if self._adapter is not None:\n", "assert self._session is None\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Condition", "Assert'" ]
[ "def __init__(self, VAR_0=None):...\n", "super(CLASS_0, self).__init__()\n", "if VAR_0:\n", "self._init_graph(VAR_0)\n" ]
[ "def __init__(self, workflow_job=None):...\n", "super(WorkflowDAG, self).__init__()\n", "if workflow_job:\n", "self._init_graph(workflow_job)\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Condition", "Expr'" ]
[ "def __missing__(self, VAR_30):...\n", "VAR_66 = self[VAR_30] = VAR_31(*VAR_30)\n", "return VAR_66\n" ]
[ "def __missing__(self, key):...\n", "ret = self[key] = f(*key)\n", "return ret\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Return'" ]
[ "def FUNC_4(self):...\n", "if hasattr(self, 'extractedcpcat'):\n", "return self.extractedcpcat\n", "return self\n" ]
[ "def pull_out_cp(self):...\n", "if hasattr(self, 'extractedcpcat'):\n", "return self.extractedcpcat\n", "return self\n" ]
[ 0, 6, 6, 6 ]
[ "FunctionDef'", "Condition", "Return'", "Return'" ]
[ "def FUNC_12(VAR_7):...\n", "@wraps(VAR_7)...\n", "if not app.config.get('USE_AUTH'):\n", "return VAR_7(*VAR_9, **kwargs)\n", "FUNC_0()\n", "return VAR_7(*VAR_9, **kwargs)\n", "logging.warning('require_logout(): calling log_out()')\n", "VAR_16 = VAR_1.log_out()\n", "if VAR_16.headers.get('Location') == url_for('goodbye'):\n", "return VAR_7(*VAR_9, **kwargs)\n", "return VAR_16\n" ]
[ "def require_logout_for_goodbye(f):...\n", "@wraps(f)...\n", "if not app.config.get('USE_AUTH'):\n", "return f(*args, **kwargs)\n", "get_logged_in_user()\n", "return f(*args, **kwargs)\n", "logging.warning('require_logout(): calling log_out()')\n", "resp = user_mod.log_out()\n", "if resp.headers.get('Location') == url_for('goodbye'):\n", "return f(*args, **kwargs)\n", "return resp\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Condition", "Return'", "Expr'", "Return'", "Expr'", "Assign'", "For", "Return'", "Return'" ]
[ "def FUNC_5(VAR_2, VAR_6):...\n", "for _privilege in VAR_0[VAR_2]:\n", "if fnmatch.fnmatch(VAR_6, _privilege):\n", "return False\n", "return True\n" ]
[ "def user_type_has_privilege(user_type, privilege):...\n", "for _privilege in PRIVILEGES[user_type]:\n", "if fnmatch.fnmatch(privilege, _privilege):\n", "return False\n", "return True\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "For", "Condition", "Return'", "Return'" ]
[ "@staticmethod...\n", "" ]
[ "@staticmethod...\n", "" ]
[ 0, 0 ]
[ "Condition", "Condition" ]
[ "def FUNC_2(self, VAR_8):...\n", "\"\"\"docstring\"\"\"\n", "VAR_5 = {}\n", "VAR_5['clusterName'] = VAR_8\n", "VAR_5['searchDepth'] = '1'\n", "VAR_5['verbose'] = '0'\n", "VAR_18 = self._cliq_run_xml('getClusterInfo', VAR_5)\n", "return VAR_18\n" ]
[ "def _cliq_get_cluster_info(self, cluster_name):...\n", "\"\"\"docstring\"\"\"\n", "cliq_args = {}\n", "cliq_args['clusterName'] = cluster_name\n", "cliq_args['searchDepth'] = '1'\n", "cliq_args['verbose'] = '0'\n", "result_xml = self._cliq_run_xml('getClusterInfo', cliq_args)\n", "return result_xml\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "@property...\n", "return list(self)\n" ]
[ "@property...\n", "return list(self)\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "def FUNC_0(VAR_6):...\n", "VAR_10 = VAR_6.find('http:')\n", "VAR_11 = VAR_6.find('https:')\n", "if VAR_11 == -1:\n", "if VAR_10 == -1:\n", "if VAR_10 == -1:\n", "return None\n", "return VAR_6[VAR_10:]\n", "return VAR_6[VAR_11:]\n", "return VAR_6[min(VAR_10, VAR_11):]\n" ]
[ "def ParseURL(s):...\n", "http_idx = s.find('http:')\n", "https_idx = s.find('https:')\n", "if https_idx == -1:\n", "if http_idx == -1:\n", "if http_idx == -1:\n", "return None\n", "return s[http_idx:]\n", "return s[https_idx:]\n", "return s[min(http_idx, https_idx):]\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Condition", "Condition", "Condition", "Return'", "Return'", "Return'", "Return'" ]
[ "def FUNC_24(self, VAR_9):...\n", "\"\"\"docstring\"\"\"\n", "def FUNC_25(VAR_10):...\n", "self.events_requests.remove(VAR_9)\n", "self.events_requests.append(VAR_9)\n", "VAR_9.notifyFinish().addCallbacks(FUNC_25, FUNC_25)\n", "VAR_9.write(json.dumps({'type': 'events_start', 'event': {'tribler_started':\n self.session.lm.initComplete, 'version': version_id}}) + '\\n')\n", "return server.NOT_DONE_YET\n" ]
[ "def render_GET(self, request):...\n", "\"\"\"docstring\"\"\"\n", "def on_request_finished(_):...\n", "self.events_requests.remove(request)\n", "self.events_requests.append(request)\n", "request.notifyFinish().addCallbacks(on_request_finished, on_request_finished)\n", "request.write(json.dumps({'type': 'events_start', 'event': {\n 'tribler_started': self.session.lm.initComplete, 'version': version_id}\n }) + '\\n')\n", "return server.NOT_DONE_YET\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "FunctionDef'", "Expr'", "Expr'", "Expr'", "Expr'", "Return'" ]
[ "def __init__(self, *VAR_2, **VAR_3):...\n", "super(CLASS_0, self).__init__(*VAR_2, **kwargs)\n", "self.cluster_vip = None\n" ]
[ "def __init__(self, *args, **kwargs):...\n", "super(HpSanISCSIDriver, self).__init__(*args, **kwargs)\n", "self.cluster_vip = None\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Assign'" ]
[ "def FUNC_7(self):...\n", "return self.data\n" ]
[ "def to_json(self):...\n", "return self.data\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_4(self):...\n", "\"\"\"docstring\"\"\"\n", "VAR_11 = VAR_0[:]\n", "VAR_11.extend([(Allow, user.login, ('view_company', 'edit_company',\n 'view.file', 'list_customers', 'add_customer', 'list_projects',\n 'add_project', 'list_estimations', 'list_invoices',\n 'edit_commercial_handling', 'list_expenses', 'add.expense',\n 'list_sale_products', 'add_sale_product', 'list_treasury_files',\n 'list_activities', 'list_workshops')) for user in self.employees])\n", "return VAR_11\n" ]
[ "def get_company_acl(self):...\n", "\"\"\"docstring\"\"\"\n", "acl = DEFAULT_PERM[:]\n", "acl.extend([(Allow, user.login, ('view_company', 'edit_company',\n 'view.file', 'list_customers', 'add_customer', 'list_projects',\n 'add_project', 'list_estimations', 'list_invoices',\n 'edit_commercial_handling', 'list_expenses', 'add.expense',\n 'list_sale_products', 'add_sale_product', 'list_treasury_files',\n 'list_activities', 'list_workshops')) for user in self.employees])\n", "return acl\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Expr'", "Return'" ]
[ "def FUNC_9(self, VAR_9, VAR_10):...\n", "if self._IsServerAlive():\n", "return FUNC_9(VAR_9, VAR_10)\n" ]
[ "def SendCommandRequest(self, arguments, completer):...\n", "if self._IsServerAlive():\n", "return SendCommandRequest(arguments, completer)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Return'" ]
[ "def FUNC_1(self):...\n", "return 'load statement for task %s (%s)' % (self.task_id, self.url)\n" ]
[ "def describe(self):...\n", "return 'load statement for task %s (%s)' % (self.task_id, self.url)\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_7(self, VAR_12, VAR_13):...\n", "self._credentials[VAR_12] = VAR_13\n" ]
[ "def add_user(self, username, password):...\n", "self._credentials[username] = password\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Assign'" ]
[ "def FUNC_6(VAR_12, VAR_3, VAR_4=-1):...\n", "VAR_7 = VAR_12[0]\n", "VAR_24 = VAR_12[1]\n", "if not VAR_24:\n", "return VAR_24\n", "if VAR_7 == -2 ** 255 and VAR_24 == -1:\n", "return -2 ** 255\n", "return copysign(abs(VAR_7 / VAR_24), VAR_7 / VAR_24)\n" ]
[ "def signedDiv(params, symbols, symId=-1):...\n", "x = params[0]\n", "y = params[1]\n", "if not y:\n", "return y\n", "if x == -2 ** 255 and y == -1:\n", "return -2 ** 255\n", "return copysign(abs(x / y), x / y)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Condition", "Return'", "Condition", "Return'", "Return'" ]
[ "def FUNC_1(self):...\n", "VAR_6 = self.client.get('/api/apps')\n", "self.assertEqual(VAR_6.status_code, 200)\n", "self.assertEqual(len(VAR_6.data['results']), 2)\n", "VAR_8 = VAR_6.data['results'][0]['id']\n", "VAR_5 = '/api/apps/{}/perms'.format(VAR_8)\n", "VAR_7 = {'username': 'autotest-2'}\n", "VAR_6 = self.client.post(VAR_5, json.dumps(VAR_7), content_type=\n 'application/json')\n", "self.assertEqual(VAR_6.status_code, 201)\n", "VAR_6 = self.client.get('/api/apps/{}/perms'.format(VAR_8), content_type=\n 'application/json')\n", "self.assertEqual(VAR_6.data, {'users': ['autotest-2']})\n" ]
[ "def test_list(self):...\n", "response = self.client.get('/api/apps')\n", "self.assertEqual(response.status_code, 200)\n", "self.assertEqual(len(response.data['results']), 2)\n", "app_id = response.data['results'][0]['id']\n", "url = '/api/apps/{}/perms'.format(app_id)\n", "body = {'username': 'autotest-2'}\n", "response = self.client.post(url, json.dumps(body), content_type=\n 'application/json')\n", "self.assertEqual(response.status_code, 201)\n", "response = self.client.get('/api/apps/{}/perms'.format(app_id),\n content_type='application/json')\n", "self.assertEqual(response.data, {'users': ['autotest-2']})\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 5, 0 ]
[ "FunctionDef'", "Assign'", "Expr'", "Expr'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "Expr'" ]
[ "from flaskext.mysql import MySQL\n", "def FUNC_0(VAR_0, VAR_1, VAR_2):...\n", "VAR_3 = VAR_2.connect()\n", "VAR_4 = VAR_3.cursor()\n", "VAR_5 = \"SELECT * FROM Users WHERE userEmail = '{0}'\".format(VAR_0)\n", "VAR_4.execute(VAR_5)\n", "VAR_6 = VAR_4.fetchone()\n", "if VAR_6 is not None:\n", "return 'Fail'\n", "VAR_7 = \"INSERT INTO Users (userEmail, password) VALUES('{0}', '{1}');\".format(\n VAR_0, VAR_1)\n", "VAR_4.execute(VAR_7)\n", "VAR_3.commit()\n", "return 'Success'\n" ]
[ "from flaskext.mysql import MySQL\n", "def registerUser(username, password, sqlHandle):...\n", "conn = sqlHandle.connect()\n", "cursor = conn.cursor()\n", "userCheckQuery = \"SELECT * FROM Users WHERE userEmail = '{0}'\".format(username)\n", "cursor.execute(userCheckQuery)\n", "result = cursor.fetchone()\n", "if result is not None:\n", "return 'Fail'\n", "queryString = (\"INSERT INTO Users (userEmail, password) VALUES('{0}', '{1}');\"\n .format(username, password))\n", "cursor.execute(queryString)\n", "conn.commit()\n", "return 'Success'\n" ]
[ 0, 0, 0, 0, 4, 0, 0, 0, 0, 4, 0, 0, 0 ]
[ "ImportFrom'", "FunctionDef'", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "Condition", "Return'", "Assign'", "Expr'", "Expr'", "Return'" ]
[ "@wraps(VAR_55)...\n", "def FUNC_27(VAR_6):...\n", "if callable(VAR_6):\n", "return VAR_6()\n", "if request.method == 'POST':\n", "return VAR_55(*VAR_69, **kwargs)\n", "VAR_10 = json.loads(request.data)\n", "VAR_54 = {'error': {'type': 'schema', 'message': str(e)}, 'schema': VAR_24}\n", "schemas.validate(VAR_10, VAR_24)\n", "return json.dumps(VAR_54, sort_keys=True, indent=4, default=default_encode\n ), 400, {'Content-Type': 'application/json'}\n", "VAR_70['validated_body'] = VAR_10\n", "if VAR_70.get('timer'):\n", "VAR_70['timer'].mark('validate_schema')\n" ]
[ "@wraps(func)...\n", "def default_encode(value):...\n", "if callable(value):\n", "return value()\n", "if request.method == 'POST':\n", "return func(*args, **kwargs)\n", "body = json.loads(request.data)\n", "result = {'error': {'type': 'schema', 'message': str(e)}, 'schema': schema}\n", "schemas.validate(body, schema)\n", "return json.dumps(result, sort_keys=True, indent=4, default=default_encode\n ), 400, {'Content-Type': 'application/json'}\n", "kwargs['validated_body'] = body\n", "if kwargs.get('timer'):\n", "kwargs['timer'].mark('validate_schema')\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "FunctionDef'", "Condition", "Return'", "Condition", "Return'", "Assign'", "Assign'", "Expr'", "Return'", "Assign'", "Condition", "Expr'" ]
[ "@memoized_property...\n", "return self.get_options().native_source_extensions\n" ]
[ "@memoized_property...\n", "return self.get_options().native_source_extensions\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "@jwt_required...\n", "if VAR_5:\n", "VAR_7 = Table.filter_by(email=None, VAR_5=user_id)\n", "VAR_8 = {'results': Table.query(), 'status': 'success'}\n", "if len(VAR_7) < 1:\n", "return jsonify(VAR_8), 200\n", "VAR_8 = {'results': 'User not found', 'status': 'fail'}\n", "VAR_8 = {'results': VAR_7, 'status': 'success'}\n", "return make_response(jsonify(VAR_8)), 404\n", "return jsonify(VAR_8), 200\n" ]
[ "@jwt_required...\n", "if user_id:\n", "user = Table.filter_by(email=None, user_id=user_id)\n", "response_object = {'results': Table.query(), 'status': 'success'}\n", "if len(user) < 1:\n", "return jsonify(response_object), 200\n", "response_object = {'results': 'User not found', 'status': 'fail'}\n", "response_object = {'results': user, 'status': 'success'}\n", "return make_response(jsonify(response_object)), 404\n", "return jsonify(response_object), 200\n" ]
[ 0, 0, 4, 4, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Condition", "Assign'", "Assign'", "Condition", "Return'", "Assign'", "Assign'", "Return'", "Return'" ]
[ "@property...\n", "\"\"\"docstring\"\"\"\n", "if self._parsed_content_packages:\n", "return self._parsed_content_packages\n", "VAR_6 = self.config.get(self.section, 'content_packages')\n", "VAR_5 = []\n", "for VAR_7 in VAR_6.split(','):\n", "VAR_7 = VAR_7.strip()\n", "self._parsed_content_packages = VAR_5\n", "VAR_8, VAR_9, VAR_10 = VAR_7.partition(':')\n", "return VAR_5\n", "VAR_5.append((VAR_8, VAR_10))\n" ]
[ "@property...\n", "\"\"\"docstring\"\"\"\n", "if self._parsed_content_packages:\n", "return self._parsed_content_packages\n", "value = self.config.get(self.section, 'content_packages')\n", "res = []\n", "for this in value.split(','):\n", "this = this.strip()\n", "self._parsed_content_packages = res\n", "name, _sep, path = this.partition(':')\n", "return res\n", "res.append((name, path))\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Condition", "Return'", "Assign'", "Assign'", "For", "Assign'", "Assign'", "Assign'", "Return'", "Expr'" ]
[ "def FUNC_7(*VAR_15, **VAR_16):...\n", "\"\"\"docstring\"\"\"\n", "return VAR_0(VAR_1, *VAR_15, **kwargs)\n" ]
[ "def new_object(*args, **kwargs):...\n", "\"\"\"docstring\"\"\"\n", "return cls(name, *args, **kwargs)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Return'" ]
[ "def FUNC_0(self, VAR_0, VAR_1=False):...\n", "if not VAR_0.get('database'):\n", "VAR_0['database'] = '%s.db' % self._model.sqlDatabaseName()\n" ]
[ "def augmentDatabaseArgs(self, args, pool=False):...\n", "if not args.get('database'):\n", "args['database'] = '%s.db' % self._model.sqlDatabaseName()\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Assign'" ]
[ "def FUNC_25(self, VAR_21):...\n", "\"\"\"docstring\"\"\"\n", "VAR_57 = self.getfile(VAR_21, VAR_23=False)\n", "if VAR_57 == False:\n", "if not VAR_57[VAR_6] & FUNC_30.S_IFLNK:\n", "return VAR_57[VAR_9]\n" ]
[ "def readlink(self, path):...\n", "\"\"\"docstring\"\"\"\n", "p = self.getfile(path, follow_symlinks=False)\n", "if p == False:\n", "if not p[A_MODE] & stat.S_IFLNK:\n", "return p[A_TARGET]\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Condition", "Condition", "Return'" ]
[ "def FUNC_9(self):...\n", "\"\"\"docstring\"\"\"\n", "return self.meta.iptc_keys\n" ]
[ "def get_meta_used(self):...\n", "\"\"\"docstring\"\"\"\n", "return self.meta.iptc_keys\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Return'" ]
[ "def FUNC_1(self, VAR_0):...\n", "\"\"\"docstring\"\"\"\n", "VAR_14 = subprocess.Popen(VAR_0, VAR_1=subprocess.PIPE, VAR_2=subprocess.\n PIPE, shell=True)\n", "VAR_4, VAR_15 = VAR_14.communicate()\n", "VAR_16 = VAR_14.returncode\n", "logging.debug('cmd: %s, stdout: %s, stderr: %s, ret: %s', VAR_0, VAR_4,\n VAR_15, VAR_16)\n", "if VAR_16 == 0:\n", "return VAR_4\n" ]
[ "def _exec_cmd(self, cmd):...\n", "\"\"\"docstring\"\"\"\n", "proc = subprocess.Popen(cmd, stdout=subprocess.PIPE, stderr=subprocess.PIPE,\n shell=True)\n", "out, err = proc.communicate()\n", "ret = proc.returncode\n", "logging.debug('cmd: %s, stdout: %s, stderr: %s, ret: %s', cmd, out, err, ret)\n", "if ret == 0:\n", "return out\n" ]
[ 0, 0, 2, 0, 0, 2, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Assign'", "Expr'", "Condition", "Return'" ]
[ "def FUNC_38(self):...\n", "\"\"\"docstring\"\"\"\n", "VAR_2 = {'SEVERITY': {'HIGH': 3, 'MEDIUM': 1, 'LOW': 14}, 'CONFIDENCE': {\n 'HIGH': 17, 'LOW': 1}}\n", "self.check_example('subprocess_shell.py', VAR_2)\n" ]
[ "def test_subprocess_shell(self):...\n", "\"\"\"docstring\"\"\"\n", "expect = {'SEVERITY': {'HIGH': 3, 'MEDIUM': 1, 'LOW': 14}, 'CONFIDENCE': {\n 'HIGH': 17, 'LOW': 1}}\n", "self.check_example('subprocess_shell.py', expect)\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Expr'" ]
[ "def FUNC_2(self):...\n", "return self.config['runtime.gallery_dir']\n" ]
[ "def get_base_path(self):...\n", "return self.config['runtime.gallery_dir']\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "@validate(VUser(), VAR_0=VEditMeetup('id'))...\n", "return BoringPage(pagename='Edit Meetup', VAR_27=EditMeetup(meetup, title=\n meetup.title, description=meetup.description, location=meetup.location,\n latitude=meetup.latitude, longitude=meetup.longitude, timestamp=int(\n meetup.timestamp * 1000), tzoffset=meetup.tzoffset)).render()\n" ]
[ "@validate(VUser(), meetup=VEditMeetup('id'))...\n", "return BoringPage(pagename='Edit Meetup', content=EditMeetup(meetup, title=\n meetup.title, description=meetup.description, location=meetup.location,\n latitude=meetup.latitude, longitude=meetup.longitude, timestamp=int(\n meetup.timestamp * 1000), tzoffset=meetup.tzoffset)).render()\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "@utils.synchronized('3par', external=True)...\n", "\"\"\"docstring\"\"\"\n", "self.common.client_login()\n", "VAR_18 = self.common.create_volume_from_snapshot(VAR_6, VAR_8)\n", "self.common.client_logout()\n", "return {'metadata': VAR_18}\n" ]
[ "@utils.synchronized('3par', external=True)...\n", "\"\"\"docstring\"\"\"\n", "self.common.client_login()\n", "metadata = self.common.create_volume_from_snapshot(volume, snapshot)\n", "self.common.client_logout()\n", "return {'metadata': metadata}\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Expr'", "Assign'", "Expr'", "Return'" ]
[ "import os.path\n", "import re\n", "import motor.motor_tornado\n", "from argon2 import PasswordHasher\n", "from pymongo import MongoClient\n", "import random\n", "import tornado.httpserver\n", "import tornado.ioloop\n", "import tornado.options\n", "import tornado.web\n", "import pymongo\n", "from tornado.options import define, options\n", "define('port', default=8100, help='run on the given port', type=int)\n", "\"\"\"string\"\"\"\n", "def FUNC_0(self):...\n", "return self.get_secure_cookie('user')\n" ]
[ "import os.path\n", "import re\n", "import motor.motor_tornado\n", "from argon2 import PasswordHasher\n", "from pymongo import MongoClient\n", "import random\n", "import tornado.httpserver\n", "import tornado.ioloop\n", "import tornado.options\n", "import tornado.web\n", "import pymongo\n", "from tornado.options import define, options\n", "define('port', default=8100, help='run on the given port', type=int)\n", "\"\"\" BaseHandler():\nClass that'll be used later when @tornado.web.authenticated is needed for POST requests.\n\"\"\"\n", "def get_current_user(self):...\n", "return self.get_secure_cookie('user')\n" ]
[ 0, 0, 0, 6, 0, 0, 0, 0, 0, 0, 0, 0, 0, 6, 0, 0 ]
[ "Import'", "Import'", "Import'", "ImportFrom'", "ImportFrom'", "Import'", "Import'", "Import'", "Import'", "Import'", "Import'", "ImportFrom'", "Expr'", "Expr'", "FunctionDef'", "Return'" ]
[ "def FUNC_13(VAR_10):...\n", "\"\"\"docstring\"\"\"\n", "VAR_41 = str(VAR_10) + SECRET_KEY\n", "VAR_35 = hashlib.sha256()\n", "VAR_35.update(VAR_41.encode('utf-8'))\n", "VAR_35 = VAR_35.digest()\n", "VAR_42 = base64.b64encode(VAR_35).decode('utf-8')\n", "return VAR_42\n" ]
[ "def csrf_hash(csrf):...\n", "\"\"\"docstring\"\"\"\n", "enc = str(csrf) + SECRET_KEY\n", "m = hashlib.sha256()\n", "m.update(enc.encode('utf-8'))\n", "m = m.digest()\n", "encrypted_csrf = base64.b64encode(m).decode('utf-8')\n", "return encrypted_csrf\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_40(self):...\n", "self.compilation_ko(\"\"\"\nif (true) {\n\n}\n\"\"\")\n" ]
[ "def test_misplaced_parenthesis(self):...\n", "self.compilation_ko(\"\"\"\nif (true) {\n\n}\n\"\"\")\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Expr'" ]
[ "@staticmethod...\n", "return mongo.db.analysis.find_one({'info.id': int(VAR_1)}, sort=[('_id',\n pymongo.DESCENDING)])\n" ]
[ "@staticmethod...\n", "return mongo.db.analysis.find_one({'info.id': int(task_id)}, sort=[('_id',\n pymongo.DESCENDING)])\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "def FUNC_5(self, VAR_8, VAR_6):...\n", "self.command(VAR_8, VAR_6, 'Error halting DUT', False)\n", "db.log_event('Information', 'Debugger', 'Halt DUT')\n" ]
[ "def halt_dut(self, halt_command, expected_output):...\n", "self.command(halt_command, expected_output, 'Error halting DUT', False)\n", "db.log_event('Information', 'Debugger', 'Halt DUT')\n" ]
[ 0, 0, 2 ]
[ "FunctionDef'", "Expr'", "Expr'" ]
[ "@VAR_1.route('/unique_ingredients/<cuisine_id>')...\n", "VAR_0.info('GET find_unique_ingredients_from_cuisine query')\n", "if VAR_9 in VAR_5:\n", "VAR_29, VAR_30 = VAR_5[VAR_9]\n", "VAR_33 = int(VAR_9)\n", "VAR_0.error(\n 'Error translating cuisine_id to int in find_unique_ingredients_from_cuisine, passed value: %s'\n % VAR_9)\n", "VAR_18 = VAR_3.find_unique_ingredients_of_cuisine(VAR_33, 500)\n", "if datetime.now() < VAR_29 + VAR_6:\n", "return None\n", "if VAR_18 == -1:\n", "return VAR_30\n", "return None\n", "if len(simplejson.loads(VAR_18)) == 0:\n", "VAR_18 = VAR_3.find_unique_ingredients_of_cuisine(VAR_33, 250)\n", "VAR_5[VAR_9] = datetime.now(), VAR_18\n", "if VAR_18 == -1:\n", "return VAR_18\n", "return None\n", "VAR_5[VAR_9] = datetime.now(), VAR_18\n", "return VAR_18\n" ]
[ "@app.route('/unique_ingredients/<cuisine_id>')...\n", "logger.info('GET find_unique_ingredients_from_cuisine query')\n", "if cuisine_id in unique_ingredients_cache:\n", "insert_time, data = unique_ingredients_cache[cuisine_id]\n", "cuisine_id_int = int(cuisine_id)\n", "logger.error(\n 'Error translating cuisine_id to int in find_unique_ingredients_from_cuisine, passed value: %s'\n % cuisine_id)\n", "query_res = database.find_unique_ingredients_of_cuisine(cuisine_id_int, 500)\n", "if datetime.now() < insert_time + cache_persistence_time:\n", "return None\n", "if query_res == -1:\n", "return data\n", "return None\n", "if len(simplejson.loads(query_res)) == 0:\n", "query_res = database.find_unique_ingredients_of_cuisine(cuisine_id_int, 250)\n", "unique_ingredients_cache[cuisine_id] = datetime.now(), query_res\n", "if query_res == -1:\n", "return query_res\n", "return None\n", "unique_ingredients_cache[cuisine_id] = datetime.now(), query_res\n", "return query_res\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Expr'", "Condition", "Assign'", "Assign'", "Expr'", "Assign'", "Condition", "Return'", "Condition", "Return'", "Return'", "Condition", "Assign'", "Assign'", "Condition", "Return'", "Return'", "Assign'", "Return'" ]
[ "def FUNC_37(self):...\n", "\"\"\"docstring\"\"\"\n", "VAR_2 = {'SEVERITY': {'LOW': 1, 'MEDIUM': 10, 'HIGH': 7}, 'CONFIDENCE': {\n 'LOW': 0, 'MEDIUM': 11, 'HIGH': 7}}\n", "self.check_example('ssl-insecure-version.py', VAR_2)\n" ]
[ "def test_ssl_insecure_version(self):...\n", "\"\"\"docstring\"\"\"\n", "expect = {'SEVERITY': {'LOW': 1, 'MEDIUM': 10, 'HIGH': 7}, 'CONFIDENCE': {\n 'LOW': 0, 'MEDIUM': 11, 'HIGH': 7}}\n", "self.check_example('ssl-insecure-version.py', expect)\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Expr'" ]
[ "def FUNC_0(self, VAR_3):...\n", "VAR_4 = super().value_from_object(VAR_3)\n", "return self.to_python(VAR_4)\n" ]
[ "def value_from_object(self, obj):...\n", "value = super().value_from_object(obj)\n", "return self.to_python(value)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Return'" ]
[ "def FUNC_1(VAR_14, VAR_2, VAR_3='text', VAR_4=bconfig....\n", "\"\"\"docstring\"\"\"\n", "VAR_0.info('Analyzing keywords for local file %s.' % VAR_14)\n", "VAR_15 = extractor.text_lines_from_local_file(VAR_14)\n", "return FUNC_2(VAR_15, VAR_2, VAR_3=output_mode, VAR_4=output_limit, VAR_5=\n spires, VAR_6=match_mode, VAR_7=no_cache, VAR_8=with_author_keywords,\n VAR_9=rebuild_cache, VAR_10=only_core_tags, VAR_11=extract_acronyms)\n" ]
[ "def get_keywords_from_local_file(local_file, taxonomy_name, output_mode=...\n", "\"\"\"docstring\"\"\"\n", "log.info('Analyzing keywords for local file %s.' % local_file)\n", "text_lines = extractor.text_lines_from_local_file(local_file)\n", "return get_keywords_from_text(text_lines, taxonomy_name, output_mode=\n output_mode, output_limit=output_limit, spires=spires, match_mode=\n match_mode, no_cache=no_cache, with_author_keywords=\n with_author_keywords, rebuild_cache=rebuild_cache, only_core_tags=\n only_core_tags, extract_acronyms=extract_acronyms)\n" ]
[ 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Expr'", "Assign'", "Return'" ]
[ "def FUNC_10(self):...\n", "\"\"\"docstring\"\"\"\n", "VAR_8 = time.time()\n", "tournament.check_version((2, 4))\n", "VAR_9 = time.time()\n", "VAR_10 = round(VAR_9 - VAR_8, 1)\n", "self.assertEqual(VAR_10, 3.0)\n" ]
[ "def test_wait_time(self):...\n", "\"\"\"docstring\"\"\"\n", "start = time.time()\n", "tournament.check_version((2, 4))\n", "end = time.time()\n", "count = round(end - start, 1)\n", "self.assertEqual(count, 3.0)\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Expr'", "Assign'", "Assign'", "Expr'" ]
[ "def FUNC_1(VAR_4):...\n", "\"\"\"docstring\"\"\"\n", "return VAR_4.__dict__\n" ]
[ "def serialize_to_json(object):...\n", "\"\"\"docstring\"\"\"\n", "return object.__dict__\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Return'" ]
[ "@web.authenticated...\n", "self.get(VAR_1, VAR_2=False)\n" ]
[ "@web.authenticated...\n", "self.get(path, include_body=False)\n" ]
[ 0, 5 ]
[ "Condition", "Expr'" ]
[ "def FUNC_16(VAR_16):...\n", "\"\"\"docstring\"\"\"\n", "VAR_30 = VAR_19(VAR_16, 'dynamic')\n", "VAR_31 = [VAR_30] if FUNC_10(VAR_30) else VAR_30\n", "for VAR_3 in VAR_31:\n", "VAR_52 = list(VAR_5.finditer(VAR_3))\n", "return VAR_30\n", "for VAR_29 in VAR_52:\n", "if VAR_29.group('constraint'):\n" ]
[ "def dynamic(value):...\n", "\"\"\"docstring\"\"\"\n", "annotated = flag(value, 'dynamic')\n", "tocheck = [annotated] if not_iterable(annotated) else annotated\n", "for file in tocheck:\n", "matches = list(_wildcard_regex.finditer(file))\n", "return annotated\n", "for match in matches:\n", "if match.group('constraint'):\n" ]
[ 0, 0, 7, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "For", "Assign'", "Return'", "For", "Condition" ]
[ "def FUNC_4(VAR_0):...\n", "\"\"\"docstring\"\"\"\n", "VAR_28 = CLASS_1(VAR_0)\n", "VAR_29, VAR_30 = VAR_28.get()\n", "VAR_31 = VAR_0.GET.get('tags')\n", "VAR_32 = VAR_0.GET.get('a')\n", "if VAR_32:\n", "VAR_75 = CLASS_2(VAR_30=obj, VAR_41=q_tag)\n", "VAR_33 = VAR_30.tag.all().order_by('pk')\n", "getattr(VAR_75, VAR_32)()\n", "VAR_34 = TestPlanTag.objects.filter(tag__in=all_tags).values('tag').annotate(\n num_plans=Count('tag')).order_by('tag')\n", "VAR_35 = TestCaseTag.objects.filter(tag__in=all_tags).values('tag').annotate(\n num_cases=Count('tag')).order_by('tag')\n", "VAR_36 = TestRunTag.objects.filter(tag__in=all_tags).values('tag').annotate(\n num_runs=Count('tag')).order_by('tag')\n", "VAR_37 = CLASS_3('num_plans', VAR_34)\n", "VAR_38 = CLASS_3('num_cases', VAR_35)\n", "VAR_39 = CLASS_3('num_runs', VAR_36)\n", "for VAR_44 in VAR_33:\n", "VAR_44.num_plans = VAR_37.calculate_tag_count(VAR_44)\n", "VAR_40 = {'tags': VAR_33, 'object': VAR_30}\n", "VAR_44.num_cases = VAR_38.calculate_tag_count(VAR_44)\n", "return render(VAR_0, VAR_29, VAR_40)\n", "VAR_44.num_runs = VAR_39.calculate_tag_count(VAR_44)\n" ]
[ "def tags(request):...\n", "\"\"\"docstring\"\"\"\n", "tag_objects = _TagObjects(request)\n", "template_name, obj = tag_objects.get()\n", "q_tag = request.GET.get('tags')\n", "q_action = request.GET.get('a')\n", "if q_action:\n", "tag_actions = _TagActions(obj=obj, tag_name=q_tag)\n", "all_tags = obj.tag.all().order_by('pk')\n", "getattr(tag_actions, q_action)()\n", "test_plan_tags = TestPlanTag.objects.filter(tag__in=all_tags).values('tag'\n ).annotate(num_plans=Count('tag')).order_by('tag')\n", "test_case_tags = TestCaseTag.objects.filter(tag__in=all_tags).values('tag'\n ).annotate(num_cases=Count('tag')).order_by('tag')\n", "test_run_tags = TestRunTag.objects.filter(tag__in=all_tags).values('tag'\n ).annotate(num_runs=Count('tag')).order_by('tag')\n", "plan_counter = _TagCounter('num_plans', test_plan_tags)\n", "case_counter = _TagCounter('num_cases', test_case_tags)\n", "run_counter = _TagCounter('num_runs', test_run_tags)\n", "for tag in all_tags:\n", "tag.num_plans = plan_counter.calculate_tag_count(tag)\n", "context_data = {'tags': all_tags, 'object': obj}\n", "tag.num_cases = case_counter.calculate_tag_count(tag)\n", "return render(request, template_name, context_data)\n", "tag.num_runs = run_counter.calculate_tag_count(tag)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "For", "Assign'", "Assign'", "Assign'", "Return'", "Assign'" ]
[ "def FUNC_5(self):...\n", "\"\"\"docstring\"\"\"\n", "self.read = True\n", "db.session.commit()\n" ]
[ "def mark_as_read(self):...\n", "\"\"\"docstring\"\"\"\n", "self.read = True\n", "db.session.commit()\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Expr'" ]
[ "def FUNC_10(self):...\n", "if self._hosts_cache is None:\n", "self._hosts_cache = self._get_hosts()\n", "return self._hosts_cache\n" ]
[ "def get_hosts(self):...\n", "if self._hosts_cache is None:\n", "self._hosts_cache = self._get_hosts()\n", "return self._hosts_cache\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Assign'", "Return'" ]
[ "def FUNC_3(self):...\n", "VAR_15 = []\n", "VAR_14 = datetime.now(timezone(timedelta(hours=+9), 'JST'))\n", "VAR_14 = VAR_14.replace(hour=0, minute=0, second=0, microsecond=0)\n", "for day in range(30)[::-1]:\n", "VAR_25 = VAR_14 - timedelta(days=day)\n", "return VAR_15\n", "VAR_26 = VAR_25.strftime('%Y-%m-%d')\n", "VAR_15.append(VAR_26)\n" ]
[ "def thirty_day_labels(self):...\n", "labels = []\n", "today = datetime.now(timezone(timedelta(hours=+9), 'JST'))\n", "today = today.replace(hour=0, minute=0, second=0, microsecond=0)\n", "for day in range(30)[::-1]:\n", "date = today - timedelta(days=day)\n", "return labels\n", "label = date.strftime('%Y-%m-%d')\n", "labels.append(label)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "For", "Assign'", "Return'", "Assign'", "Expr'" ]
[ "def FUNC_30(self):...\n", "VAR_44 = Tag.objects.get(name=self.tag_name)\n", "self.obj.remove_tag(VAR_44)\n" ]
[ "def remove(self):...\n", "tag = Tag.objects.get(name=self.tag_name)\n", "self.obj.remove_tag(tag)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Expr'" ]
[ "def FUNC_8(self, VAR_14):...\n", "" ]
[ "def is_file(self, col_name):...\n", "" ]
[ 0, 0 ]
[ "FunctionDef'", "Condition" ]
[ "def FUNC_1(self, VAR_5, VAR_6):...\n", "\"\"\"docstring\"\"\"\n", "if VAR_5 not in ('strings', 'files'):\n", "VAR_0.error(\"Bad parameter '%s' for submit_type\", VAR_5)\n", "VAR_13 = Folders.create_temp()\n", "return False\n", "VAR_14 = {'data': [], 'errors': []}\n", "if VAR_5 == 'strings':\n", "for VAR_4 in VAR_6:\n", "if VAR_5 == 'files':\n", "self._handle_string(VAR_14, VAR_13, VAR_4)\n", "for entry in VAR_6:\n", "return Database().add_submit(VAR_13, VAR_5, VAR_14)\n", "VAR_22 = Storage.get_filename_from_path(entry['name'])\n", "VAR_18 = Files.create(VAR_13, VAR_22, entry['data'])\n", "VAR_14['data'].append({'type': 'file', 'data': VAR_18})\n" ]
[ "def pre(self, submit_type, data):...\n", "\"\"\"docstring\"\"\"\n", "if submit_type not in ('strings', 'files'):\n", "log.error(\"Bad parameter '%s' for submit_type\", submit_type)\n", "path_tmp = Folders.create_temp()\n", "return False\n", "submit_data = {'data': [], 'errors': []}\n", "if submit_type == 'strings':\n", "for line in data:\n", "if submit_type == 'files':\n", "self._handle_string(submit_data, path_tmp, line)\n", "for entry in data:\n", "return Database().add_submit(path_tmp, submit_type, submit_data)\n", "filename = Storage.get_filename_from_path(entry['name'])\n", "filepath = Files.create(path_tmp, filename, entry['data'])\n", "submit_data['data'].append({'type': 'file', 'data': filepath})\n" ]
[ 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Expr'", "Assign'", "Return'", "Assign'", "Condition", "For", "Condition", "Expr'", "For", "Return'", "Assign'", "Assign'", "Expr'" ]
[ "def FUNC_32(self, VAR_32, VAR_29):...\n", "\"\"\"docstring\"\"\"\n", "VAR_24 = self.getfile(VAR_32)\n", "if VAR_24 == False:\n", "return\n", "if VAR_24[VAR_2] != VAR_13:\n", "return\n", "VAR_24[VAR_5] = VAR_29\n" ]
[ "def update_size(self, filename, size):...\n", "\"\"\"docstring\"\"\"\n", "f = self.getfile(filename)\n", "if f == False:\n", "return\n", "if f[A_TYPE] != T_FILE:\n", "return\n", "f[A_SIZE] = size\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Condition", "Return'", "Condition", "Return'", "Assign'" ]
[ "def FUNC_28(self):...\n", "self.compilation_ko('require [];')\n" ]
[ "def test_empty_string_list(self):...\n", "self.compilation_ko('require [];')\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Expr'" ]
[ "@validate(VUser(), VAR_0=VEditMeetup('id'))...\n", "return BoringPage(pagename='Edit Meetup', VAR_27=EditMeetup(meetup, title=\n meetup.title, description=meetup.description, location=meetup.location,\n latitude=meetup.latitude, longitude=meetup.longitude, timestamp=int(\n meetup.timestamp * 1000), tzoffset=meetup.tzoffset)).render()\n" ]
[ "@validate(VUser(), meetup=VEditMeetup('id'))...\n", "return BoringPage(pagename='Edit Meetup', content=EditMeetup(meetup, title=\n meetup.title, description=meetup.description, location=meetup.location,\n latitude=meetup.latitude, longitude=meetup.longitude, timestamp=int(\n meetup.timestamp * 1000), tzoffset=meetup.tzoffset)).render()\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "def FUNC_6(self, VAR_23):...\n", "VAR_82 = CLASS_46.run(self, VAR_23)\n", "if not VAR_82 or not hasattr(VAR_82, 'email') or not VAR_82.email:\n", "return self.error(errors.NO_EMAIL_FOR_USER)\n", "return VAR_82\n" ]
[ "def run(self, name):...\n", "user = VExistingUname.run(self, name)\n", "if not user or not hasattr(user, 'email') or not user.email:\n", "return self.error(errors.NO_EMAIL_FOR_USER)\n", "return user\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Condition", "Return'", "Return'" ]
[ "def FUNC_0(self):...\n", "self.assertTrue(self.client.login(username='autotest', password='password'))\n" ]
[ "def setUp(self):...\n", "self.assertTrue(self.client.login(username='autotest', password='password'))\n" ]
[ 0, 5 ]
[ "FunctionDef'", "Expr'" ]