lines
sequencelengths 1
444
| raw_lines
sequencelengths 1
444
| label
sequencelengths 1
444
| type
sequencelengths 1
444
|
---|---|---|---|
[
"def FUNC_2(VAR_1, VAR_2=1, VAR_3=False, VAR_4=[]):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_17 = VAR_1.data_group.type\n",
"VAR_18, VAR_19 = get_extracted_models(VAR_17)\n",
"VAR_20 = hasattr(VAR_1, 'extractedtext')\n",
"def FUNC_3(VAR_21, VAR_22, VAR_23=BaseInlineFormSet, VAR_24=forms.ModelForm,...\n",
"VAR_32 = VAR_22.detail_fields()\n",
"if VAR_4:\n",
"VAR_32 = [in_field for in_field in VAR_32 if not in_field in VAR_4]\n",
"return forms.inlineformset_factory(VAR_21=parent_model, VAR_22=model,\n VAR_27=formset_fields, VAR_23=formset, VAR_24=form, VAR_2=extra, VAR_3=\n can_delete)\n"
] | [
"def create_detail_formset(document, extra=1, can_delete=False, exclude=[]):...\n",
"\"\"\"docstring\"\"\"\n",
"group_type = document.data_group.type\n",
"parent, child = get_extracted_models(group_type)\n",
"extracted = hasattr(document, 'extractedtext')\n",
"def make_formset(parent_model, model, formset=BaseInlineFormSet, form=forms...\n",
"formset_fields = model.detail_fields()\n",
"if exclude:\n",
"formset_fields = [in_field for in_field in formset_fields if not in_field in\n exclude]\n",
"return forms.inlineformset_factory(parent_model=parent_model, model=model,\n fields=formset_fields, formset=formset, form=form, extra=extra,\n can_delete=can_delete)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Assign'",
"Assign'",
"For",
"Assign'",
"Condition",
"Assign'",
"Return'"
] |
[
"def __call__(self, VAR_44):...\n",
"self.p = VAR_44\n",
"self.log = VAR_44.log\n",
"self.inter_sleep = VAR_44.inter_sleep\n",
"self.running = VAR_44.running\n",
"self.p.sig_sock.setsockopt(zmq.SUBSCRIBE, b'WipeManager')\n",
"self.p.wz.set_sig_handler(b'WipeManager', b'passthrough', self.send_passthrough\n )\n",
"if self.c.tcount > 0:\n",
"self.pc = ProcessContext(self.p.name, self.p.ctx, self.c.router_addr, VAR_10)\n",
"if self.c.ecount > 0:\n",
"self.spawnqueue = Queue()\n",
"self.spawn_evaluators()\n",
"while self.running.is_set():\n",
"self.log.exception(e)\n",
"self.terminate()\n",
"self.load_bumplimit_set()\n",
"if self.c.tcount == 0:\n",
"self.join_threads()\n",
"self.load_targets()\n",
"self.inter_sleep(5)\n",
"self.pc.check_waiting()\n",
"if self.c.tcount > 0:\n",
"self.load_users()\n",
"VAR_56 = self.read_newproxies()\n",
"self.save_users()\n",
"self.spawn_wipethreads()\n",
"if not VAR_56:\n",
"self.save_targets()\n",
"self.inter_sleep(5)\n",
"self.add_spawns(VAR_56)\n"
] | [
"def __call__(self, parent):...\n",
"self.p = parent\n",
"self.log = parent.log\n",
"self.inter_sleep = parent.inter_sleep\n",
"self.running = parent.running\n",
"self.p.sig_sock.setsockopt(zmq.SUBSCRIBE, b'WipeManager')\n",
"self.p.wz.set_sig_handler(b'WipeManager', b'passthrough', self.send_passthrough\n )\n",
"if self.c.tcount > 0:\n",
"self.pc = ProcessContext(self.p.name, self.p.ctx, self.c.router_addr,\n noproxy_rp)\n",
"if self.c.ecount > 0:\n",
"self.spawnqueue = Queue()\n",
"self.spawn_evaluators()\n",
"while self.running.is_set():\n",
"self.log.exception(e)\n",
"self.terminate()\n",
"self.load_bumplimit_set()\n",
"if self.c.tcount == 0:\n",
"self.join_threads()\n",
"self.load_targets()\n",
"self.inter_sleep(5)\n",
"self.pc.check_waiting()\n",
"if self.c.tcount > 0:\n",
"self.load_users()\n",
"new = self.read_newproxies()\n",
"self.save_users()\n",
"self.spawn_wipethreads()\n",
"if not new:\n",
"self.save_targets()\n",
"self.inter_sleep(5)\n",
"self.add_spawns(new)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Expr'",
"Condition",
"Expr'",
"Expr'",
"Expr'",
"Condition",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Condition",
"Expr'",
"Assign'",
"Expr'",
"Expr'",
"Condition",
"Expr'",
"Expr'",
"Expr'"
] |
[
"def FUNC_23(self, VAR_16):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_35 = self.common._cli_run('showvlun -a -showcols Port', None)\n",
"VAR_36 = {}\n",
"for VAR_15 in VAR_16:\n",
"VAR_36[VAR_15] = 0\n",
"VAR_37 = None\n",
"if VAR_35:\n",
"VAR_35 = VAR_35[1:]\n",
"return VAR_37\n",
"for line in VAR_35:\n",
"VAR_15 = line.strip()\n",
"VAR_41 = sys.maxint\n",
"if VAR_15 in VAR_36:\n",
"for VAR_15, count in VAR_36.iteritems():\n",
"VAR_36[VAR_15] = VAR_36[VAR_15] + 1\n",
"if count < VAR_41:\n",
"VAR_37 = VAR_15\n",
"VAR_41 = count\n"
] | [
"def _get_least_used_nsp(self, nspss):...\n",
"\"\"\"docstring\"\"\"\n",
"result = self.common._cli_run('showvlun -a -showcols Port', None)\n",
"nsp_counts = {}\n",
"for nsp in nspss:\n",
"nsp_counts[nsp] = 0\n",
"current_least_used_nsp = None\n",
"if result:\n",
"result = result[1:]\n",
"return current_least_used_nsp\n",
"for line in result:\n",
"nsp = line.strip()\n",
"current_smallest_count = sys.maxint\n",
"if nsp in nsp_counts:\n",
"for nsp, count in nsp_counts.iteritems():\n",
"nsp_counts[nsp] = nsp_counts[nsp] + 1\n",
"if count < current_smallest_count:\n",
"current_least_used_nsp = nsp\n",
"current_smallest_count = count\n"
] | [
0,
0,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Assign'",
"For",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Return'",
"For",
"Assign'",
"Assign'",
"Condition",
"For",
"Assign'",
"Condition",
"Assign'",
"Assign'"
] |
[
"def FUNC_1(VAR_2, VAR_4):...\n",
"VAR_8 = VAR_4.connect()\n",
"VAR_9 = VAR_8.cursor()\n",
"VAR_12 = 'string'.format(VAR_2)\n",
"VAR_9.execute(VAR_12)\n",
"VAR_13 = VAR_9.fetchall()\n",
"return VAR_13\n"
] | [
"def getCalendarList(username, sqlInstance):...\n",
"conn = sqlInstance.connect()\n",
"cursor = conn.cursor()\n",
"getCalendarDetails = (\n \"SELECT DISTINCT Calendars.calendarId, Calendars.name, Calendars.day FROM Users, Calendars, TimeSlots WHERE Calendars.calendarId = TimeSlots.calendarId AND (Calendars.userId = Users.userId OR TimeSlots.userId = Users.userId) AND Users.userEmail = '{0}'\"\n .format(username))\n",
"cursor.execute(getCalendarDetails)\n",
"result = cursor.fetchall()\n",
"return result\n"
] | [
0,
0,
0,
4,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Return'"
] |
[
"def FUNC_13():...\n",
"\"\"\"docstring\"\"\"\n",
"log.info('Stopping worker(s)')\n",
"for pid in VAR_20:\n",
"if pid is not None:\n",
"os.kill(pid, VAR_7.SIGTERM)\n"
] | [
"def kill_children():...\n",
"\"\"\"docstring\"\"\"\n",
"log.info('Stopping worker(s)')\n",
"for pid in child_pids:\n",
"if pid is not None:\n",
"os.kill(pid, signal.SIGTERM)\n"
] | [
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Expr'",
"For",
"Condition",
"Expr'"
] |
[
"def FUNC_19(self):...\n",
"VAR_7, VAR_12 = self.make_xsrf_handling_app()\n",
"self.mock_get_current_identity(model.Identity(model.IDENTITY_USER,\n '[email protected]'))\n",
"VAR_17 = VAR_7.get('/request').body\n",
"self.mock_get_current_identity(model.Identity(model.IDENTITY_USER,\n '[email protected]'))\n",
"VAR_15 = VAR_7.post('/request', expect_errors=True)\n",
"self.assertEqual(403, VAR_15.status_int)\n",
"self.assertFalse(VAR_12)\n"
] | [
"def test_xsrf_token_identity_matters(self):...\n",
"app, calls = self.make_xsrf_handling_app()\n",
"self.mock_get_current_identity(model.Identity(model.IDENTITY_USER,\n '[email protected]'))\n",
"token = app.get('/request').body\n",
"self.mock_get_current_identity(model.Identity(model.IDENTITY_USER,\n '[email protected]'))\n",
"response = app.post('/request', expect_errors=True)\n",
"self.assertEqual(403, response.status_int)\n",
"self.assertFalse(calls)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Expr'",
"Assign'",
"Expr'",
"Assign'",
"Expr'",
"Expr'"
] |
[
"def FUNC_11(self):...\n",
"return u'Nullable({})'.format(self.inner_type.for_schema())\n"
] | [
"def for_schema(self):...\n",
"return u'Nullable({})'.format(self.inner_type.for_schema())\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
[
"def FUNC_0(self, VAR_4={}):...\n",
"self.connection.request(self.verb, self.path, VAR_4=headers)\n",
"return self.connection.getresponse()\n"
] | [
"def run(self, headers={}):...\n",
"self.connection.request(self.verb, self.path, headers=headers)\n",
"return self.connection.getresponse()\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Expr'",
"Return'"
] |
[
"import sys\n",
"import tkinter as tk\n",
"import Controller as dbc\n",
"from PIL import Image, ImageTk\n",
"from tkinter import font\n",
"def __init__(self):...\n",
"tk.Tk.__init__(self)\n",
"self.title('Raspi-Telxon')\n",
"self.titleFont = font.Font(family='Helvetica', size=24)\n",
"self.itemFont = font.Font(family='Helvetica', size=18)\n",
"VAR_7 = tk.Frame(self)\n",
"VAR_7.pack(side='top', fill='both', expand=True)\n",
"VAR_7.grid_rowconfigure(0, weight=1)\n",
"VAR_7.grid_columnconfigure(0, weight=1)\n",
"self.frames = {}\n",
"self.result = ''\n",
"self.container = VAR_7\n",
"for VAR_1 in (CLASS_1, VAR_9):\n",
"VAR_2 = VAR_1(VAR_7, self)\n",
"self.show_frame(CLASS_1)\n",
"self.frames[VAR_1] = VAR_2\n",
"def FUNC_0(self, VAR_1):...\n",
"VAR_2.grid(row=0, column=0, sticky='nsew')\n",
"VAR_8 = VAR_9(self.container, self)\n",
"self.frames[VAR_9] = VAR_8\n",
"VAR_8.grid(row=0, column=0, sticky='nsew')\n",
"self.show_frame(VAR_8)\n",
"def FUNC_1(self, VAR_2):...\n",
"print('remove_frame: ' + str(VAR_2))\n",
"self.frames.pop(VAR_2, None)\n",
"def FUNC_2(self, VAR_3):...\n",
"VAR_2 = self.frames[VAR_3]\n",
"VAR_2.tkraise()\n",
"def FUNC_3(self):...\n",
"VAR_10 = VAR_11(self.container, self)\n",
"self.frames[VAR_11] = VAR_10\n",
"VAR_10.grid(row=0, column=0, sticky='nsew')\n",
"self.show_frame(VAR_11)\n",
"def FUNC_4(self, VAR_4):...\n",
"self.result = VAR_4\n",
"def FUNC_5(self):...\n",
"return self.result\n"
] | [
"import sys\n",
"import tkinter as tk\n",
"import Controller as dbc\n",
"from PIL import Image, ImageTk\n",
"from tkinter import font\n",
"def __init__(self):...\n",
"tk.Tk.__init__(self)\n",
"self.title('Raspi-Telxon')\n",
"self.titleFont = font.Font(family='Helvetica', size=24)\n",
"self.itemFont = font.Font(family='Helvetica', size=18)\n",
"container = tk.Frame(self)\n",
"container.pack(side='top', fill='both', expand=True)\n",
"container.grid_rowconfigure(0, weight=1)\n",
"container.grid_columnconfigure(0, weight=1)\n",
"self.frames = {}\n",
"self.result = ''\n",
"self.container = container\n",
"for F in (StartPage, SearchPage):\n",
"frame = F(container, self)\n",
"self.show_frame(StartPage)\n",
"self.frames[F] = frame\n",
"def create_frame(self, F):...\n",
"frame.grid(row=0, column=0, sticky='nsew')\n",
"new_frame = SearchPage(self.container, self)\n",
"self.frames[SearchPage] = new_frame\n",
"new_frame.grid(row=0, column=0, sticky='nsew')\n",
"self.show_frame(new_frame)\n",
"def remove_frame(self, frame):...\n",
"print('remove_frame: ' + str(frame))\n",
"self.frames.pop(frame, None)\n",
"def show_frame(self, cont):...\n",
"frame = self.frames[cont]\n",
"frame.tkraise()\n",
"def custom_frame(self):...\n",
"result_frame = ResultsPage(self.container, self)\n",
"self.frames[ResultsPage] = result_frame\n",
"result_frame.grid(row=0, column=0, sticky='nsew')\n",
"self.show_frame(ResultsPage)\n",
"def set_result(self, result):...\n",
"self.result = result\n",
"def get_result(self):...\n",
"return self.result\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
4,
4,
4,
4,
0,
4,
4,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Import'",
"Import'",
"Import'",
"ImportFrom'",
"ImportFrom'",
"FunctionDef'",
"Expr'",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"For",
"Assign'",
"Expr'",
"Assign'",
"FunctionDef'",
"Expr'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"FunctionDef'",
"Expr'",
"Expr'",
"FunctionDef'",
"Assign'",
"Expr'",
"FunctionDef'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"FunctionDef'",
"Assign'",
"FunctionDef'",
"Return'"
] |
[
"from __future__ import unicode_literals\n",
"from django.db import models, migrations\n",
"VAR_0 = [('exercise', '0022_auto_20160906_1401'), ('notification',\n '0001_initial')]\n",
"VAR_1 = [migrations.AddField(model_name='notification', name='submission',\n field=models.ForeignKey(to='exercise.Submission', blank=True, null=True\n ), preserve_default=True), migrations.AlterField(model_name=\n 'notification', name='notification', field=models.TextField(blank=True),\n preserve_default=True), migrations.AlterField(model_name='notification',\n name='sender', field=models.ForeignKey(related_name=\n 'sent_notifications', to='userprofile.UserProfile', blank=True, null=\n True), preserve_default=True), migrations.AlterField(model_name=\n 'notification', name='subject', field=models.CharField(blank=True,\n max_length=255), preserve_default=True)]\n"
] | [
"from __future__ import unicode_literals\n",
"from django.db import models, migrations\n",
"dependencies = [('exercise', '0022_auto_20160906_1401'), ('notification',\n '0001_initial')]\n",
"operations = [migrations.AddField(model_name='notification', name=\n 'submission', field=models.ForeignKey(to='exercise.Submission', blank=\n True, null=True), preserve_default=True), migrations.AlterField(\n model_name='notification', name='notification', field=models.TextField(\n blank=True), preserve_default=True), migrations.AlterField(model_name=\n 'notification', name='sender', field=models.ForeignKey(related_name=\n 'sent_notifications', to='userprofile.UserProfile', blank=True, null=\n True), preserve_default=True), migrations.AlterField(model_name=\n 'notification', name='subject', field=models.CharField(blank=True,\n max_length=255), preserve_default=True)]\n"
] | [
0,
0,
0,
4
] | [
"ImportFrom'",
"ImportFrom'",
"Assign'",
"Assign'"
] |
[
"def __init__(self, *VAR_2):...\n",
"self._path = VAR_2\n"
] | [
"def __init__(self, *path):...\n",
"self._path = path\n"
] | [
0,
0
] | [
"FunctionDef'",
"Assign'"
] |
[
"def FUNC_6(self, VAR_19):...\n",
"return VAR_19 if VAR_19 and self.measure.match(VAR_19) else ''\n"
] | [
"def run(self, value):...\n",
"return value if value and self.measure.match(value) else ''\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
[
"def FUNC_4(self):...\n",
"self.assertTrue(self.client.login(VAR_1='autotest-1', VAR_2='password'))\n"
] | [
"def setUp(self):...\n",
"self.assertTrue(self.client.login(username='autotest-1', password='password'))\n"
] | [
0,
5
] | [
"FunctionDef'",
"Expr'"
] |
[
"def FUNC_12():...\n",
"VAR_18 = requests.get(FUNC_2('healthy'))\n",
"VAR_18.raise_for_status()\n",
"return VAR_18.json()\n"
] | [
"def _ServerIsHealthy():...\n",
"response = requests.get(_BuildUri('healthy'))\n",
"response.raise_for_status()\n",
"return response.json()\n"
] | [
0,
7,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Expr'",
"Return'"
] |
[
"from itertools import chain\n",
"from datetime import datetime\n",
"from model_utils.managers import InheritanceManager\n",
"from django.db import models\n",
"from django.core.exceptions import ValidationError\n",
"from django import forms\n",
"from django.urls import reverse\n",
"from .common_info import CommonInfo\n",
"VAR_3 = VAR_0.OneToOneField('DataDocument', on_delete=models.CASCADE,\n primary_key=True)\n",
"VAR_4 = VAR_0.CharField(max_length=500, null=True, blank=True)\n",
"VAR_5 = VAR_0.CharField(max_length=25, null=True, blank=True)\n",
"VAR_6 = VAR_0.CharField(max_length=50, null=True, blank=True)\n",
"VAR_7 = VAR_0.ForeignKey('Script', on_delete=models.CASCADE,\n limit_choices_to={'script_type': 'EX'})\n",
"VAR_8 = VAR_0.BooleanField(default=False, verbose_name='QA approved')\n",
"VAR_9 = VAR_0.BooleanField(default=False, verbose_name='QA edited')\n",
"VAR_10 = VAR_0.DateTimeField(null=True, blank=True, verbose_name=\n 'QA approval date')\n",
"VAR_11 = VAR_0.ForeignKey('auth.User', on_delete=models.SET_NULL,\n verbose_name='QA approved by', null=True, blank=True)\n",
"VAR_12 = VAR_0.ForeignKey('QAGroup', verbose_name='QA group', on_delete=\n models.SET_NULL, null=True, blank=True)\n",
"VAR_13 = InheritanceManager()\n",
"def __str__(self):...\n",
"return str(self.data_document)\n"
] | [
"from itertools import chain\n",
"from datetime import datetime\n",
"from model_utils.managers import InheritanceManager\n",
"from django.db import models\n",
"from django.core.exceptions import ValidationError\n",
"from django import forms\n",
"from django.urls import reverse\n",
"from .common_info import CommonInfo\n",
"data_document = models.OneToOneField('DataDocument', on_delete=models.\n CASCADE, primary_key=True)\n",
"prod_name = models.CharField(max_length=500, null=True, blank=True)\n",
"doc_date = models.CharField(max_length=25, null=True, blank=True)\n",
"rev_num = models.CharField(max_length=50, null=True, blank=True)\n",
"extraction_script = models.ForeignKey('Script', on_delete=models.CASCADE,\n limit_choices_to={'script_type': 'EX'})\n",
"qa_checked = models.BooleanField(default=False, verbose_name='QA approved')\n",
"qa_edited = models.BooleanField(default=False, verbose_name='QA edited')\n",
"qa_approved_date = models.DateTimeField(null=True, blank=True, verbose_name\n ='QA approval date')\n",
"qa_approved_by = models.ForeignKey('auth.User', on_delete=models.SET_NULL,\n verbose_name='QA approved by', null=True, blank=True)\n",
"qa_group = models.ForeignKey('QAGroup', verbose_name='QA group', on_delete=\n models.SET_NULL, null=True, blank=True)\n",
"objects = InheritanceManager()\n",
"def __str__(self):...\n",
"return str(self.data_document)\n"
] | [
6,
6,
0,
0,
6,
6,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"FunctionDef'",
"Return'"
] |
[
"from flask import request\n",
"from flask_jwt_extended import jwt_required\n",
"from flask_restplus import Namespace, Resource, fields\n",
"from utilities import database_utilities\n",
"VAR_0 = Namespace('admins', description='Information relating to system admins'\n )\n",
"def FUNC_0(self):...\n",
"\"\"\"docstring\"\"\"\n",
"return database_utilities.execute_query('select * from admins')\n"
] | [
"from flask import request\n",
"from flask_jwt_extended import jwt_required\n",
"from flask_restplus import Namespace, Resource, fields\n",
"from utilities import database_utilities\n",
"api = Namespace('admins', description='Information relating to system admins')\n",
"def get(self):...\n",
"\"\"\"docstring\"\"\"\n",
"return database_utilities.execute_query('select * from admins')\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0
] | [
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"Assign'",
"FunctionDef'",
"Docstring",
"Return'"
] |
[
"def FUNC_7(self, VAR_9):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_30 = False\n",
"for rec in self:\n",
"VAR_27 = rec._next_states()\n",
"VAR_31 = self.env['crapo.transition'].search([('from_state', '=', VAR_30.id\n ), ('to_state', '=', VAR_9)], limit=1)\n",
"if rec.state.id == VAR_9:\n",
"return VAR_31\n",
"VAR_30 = rec.state\n",
"if not VAR_27:\n",
"if VAR_9 not in VAR_27.ids:\n",
"if VAR_30 is not False and VAR_30 != rec.state:\n",
"VAR_30 = rec.state\n"
] | [
"def _get_transition(self, target_state_id):...\n",
"\"\"\"docstring\"\"\"\n",
"current_state = False\n",
"for rec in self:\n",
"next_states = rec._next_states()\n",
"transition = self.env['crapo.transition'].search([('from_state', '=',\n current_state.id), ('to_state', '=', target_state_id)], limit=1)\n",
"if rec.state.id == target_state_id:\n",
"return transition\n",
"current_state = rec.state\n",
"if not next_states:\n",
"if target_state_id not in next_states.ids:\n",
"if current_state is not False and current_state != rec.state:\n",
"current_state = rec.state\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"For",
"Assign'",
"Assign'",
"Condition",
"Return'",
"Assign'",
"Condition",
"Condition",
"Condition",
"Assign'"
] |
[
"def FUNC_16(VAR_0, VAR_5, VAR_14):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_23 = pd.read_sql_table(VAR_0, db.engine)\n",
"print('EQUAL FREQUENCY DISCRETIZATION FAILED:\\n' + str(e))\n",
"VAR_25 = len(VAR_23[VAR_5])\n",
"VAR_26 = VAR_25 // VAR_14\n",
"VAR_27 = list(VAR_23[VAR_5].sort_values())\n",
"VAR_28 = 0\n",
"VAR_29 = []\n",
"while VAR_28 < VAR_25:\n",
"if VAR_29[-1] != VAR_27[-1] and len(VAR_29) == VAR_14 + 1:\n",
"VAR_29.append(VAR_27[VAR_28])\n",
"VAR_29[-1] = VAR_27[-1]\n",
"if VAR_29[-1] != VAR_27[-1] and len(VAR_29) != VAR_14 + 1:\n",
"VAR_28 += VAR_26\n",
"VAR_29[0] = VAR_29[0] - VAR_29[0] * 0.001\n",
"VAR_29.append(VAR_27[-1])\n",
"VAR_29[-1] = VAR_29[-1] + VAR_29[-1] * 0.001\n",
"VAR_24 = VAR_5 + '_' + str(VAR_14) + '_eq_freq_intervals'\n",
"FUNC_15(VAR_0, VAR_5, VAR_29, VAR_23, VAR_24)\n"
] | [
"def discretize_eq_freq(table_name, attr, intervals):...\n",
"\"\"\"docstring\"\"\"\n",
"df = pd.read_sql_table(table_name, db.engine)\n",
"print('EQUAL FREQUENCY DISCRETIZATION FAILED:\\n' + str(e))\n",
"attr_length = len(df[attr])\n",
"elements_per_interval = attr_length // intervals\n",
"sorted_data = list(df[attr].sort_values())\n",
"selector = 0\n",
"edge_list = []\n",
"while selector < attr_length:\n",
"if edge_list[-1] != sorted_data[-1] and len(edge_list) == intervals + 1:\n",
"edge_list.append(sorted_data[selector])\n",
"edge_list[-1] = sorted_data[-1]\n",
"if edge_list[-1] != sorted_data[-1] and len(edge_list) != intervals + 1:\n",
"selector += elements_per_interval\n",
"edge_list[0] = edge_list[0] - edge_list[0] * 0.001\n",
"edge_list.append(sorted_data[-1])\n",
"edge_list[-1] = edge_list[-1] + edge_list[-1] * 0.001\n",
"column_name = attr + '_' + str(intervals) + '_eq_freq_intervals'\n",
"discretize_width(table_name, attr, edge_list, df, column_name)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Condition",
"Expr'",
"Assign'",
"Condition",
"AugAssign'",
"Assign'",
"Expr'",
"Assign'",
"Assign'",
"Expr'"
] |
[
"def FUNC_3(VAR_2, VAR_3):...\n",
"VAR_2.browser = webdriver.PhantomJS()\n",
"VAR_2.browser.set_window_size(1280, 1024)\n",
"VAR_2.browser.implicitly_wait(DEFAULT_IMPLICIT_WAIT_TIMEOUT_IN_S)\n",
"VAR_2.browser.set_page_load_timeout(60)\n",
"VAR_2.browser.get(HOMEPAGE_URL)\n"
] | [
"def before_feature(context, feature):...\n",
"context.browser = webdriver.PhantomJS()\n",
"context.browser.set_window_size(1280, 1024)\n",
"context.browser.implicitly_wait(DEFAULT_IMPLICIT_WAIT_TIMEOUT_IN_S)\n",
"context.browser.set_page_load_timeout(60)\n",
"context.browser.get(HOMEPAGE_URL)\n"
] | [
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Expr'"
] |
[
"@wraps(VAR_1)...\n",
"VAR_17 = make_response(VAR_1(*VAR_6, **kwargs))\n",
"VAR_17.headers['Last-Modified'] = datetime.now()\n",
"VAR_17.headers['Cache-Control'\n ] = 'no-store, no-cache, must-revalidate, post-check=0, pre-check=0, max-age=0'\n",
"VAR_17.headers['Pragma'] = 'no-cache'\n",
"VAR_17.headers['Expires'] = '-1'\n",
"return VAR_17\n"
] | [
"@wraps(view)...\n",
"response = make_response(view(*args, **kwargs))\n",
"response.headers['Last-Modified'] = datetime.now()\n",
"response.headers['Cache-Control'\n ] = 'no-store, no-cache, must-revalidate, post-check=0, pre-check=0, max-age=0'\n",
"response.headers['Pragma'] = 'no-cache'\n",
"response.headers['Expires'] = '-1'\n",
"return response\n"
] | [
0,
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Return'"
] |
[
"def FUNC_31():...\n",
"yield self._bin_path\n",
"if self._is_jdk:\n",
"yield os.path.join(self.home, 'jre', 'bin')\n"
] | [
"def bin_paths():...\n",
"yield self._bin_path\n",
"if self._is_jdk:\n",
"yield os.path.join(self.home, 'jre', 'bin')\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"Expr'",
"Condition",
"Expr'"
] |
[
"def FUNC_3(VAR_0, VAR_4, VAR_5):...\n",
"VAR_7 = psycopg2.connect(dbname=pg_connection['database'], user=\n pg_connection['user'], password=pg_connection['password'], host=\n pg_connection['host'])\n",
"VAR_8 = VAR_7.cursor()\n",
"VAR_8.execute(\n \"\"\"\n UPDATE quests\n SET completed = '%s'\n WHERE id = %s;\n \"\"\",\n (VAR_5, VAR_4))\n",
"VAR_7.commit()\n",
"VAR_8.close()\n",
"VAR_7.close()\n"
] | [
"def complete_quest(pg_connection, quest_id, completion):...\n",
"conn = psycopg2.connect(dbname=pg_connection['database'], user=\n pg_connection['user'], password=pg_connection['password'], host=\n pg_connection['host'])\n",
"cur = conn.cursor()\n",
"cur.execute(\n \"\"\"\n UPDATE quests\n SET completed = '%s'\n WHERE id = %s;\n \"\"\",\n (completion, quest_id))\n",
"conn.commit()\n",
"cur.close()\n",
"conn.close()\n"
] | [
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Expr'"
] |
[
"def FUNC_2(self, VAR_8, VAR_9, VAR_10, VAR_5, VAR_11):...\n",
"VAR_14 = 'string'.format(VAR_8, VAR_9, VAR_10, VAR_5, VAR_11)\n",
"self.cur.execute(VAR_14)\n",
"self.conn.commit()\n"
] | [
"def create_event(self, title, start_time, time_zone, server_id, description):...\n",
"sql = (\n \"\"\"INSERT INTO events (title, start_time, time_zone, server_id, description)\n VALUES ('{0}', '{1}', '{2}', '{3}', '{4}')\n \"\"\"\n .format(title, start_time, time_zone, server_id, description))\n",
"self.cur.execute(sql)\n",
"self.conn.commit()\n"
] | [
0,
4,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Expr'",
"Expr'"
] |
[
"def FUNC_15(self):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_7 = BokChoyTestSuite('', num_processes=2, verbosity=3)\n",
"BokChoyTestSuite.verbosity_processes_string(VAR_7)\n"
] | [
"def test_invalid_verbosity_and_processes(self):...\n",
"\"\"\"docstring\"\"\"\n",
"suite = BokChoyTestSuite('', num_processes=2, verbosity=3)\n",
"BokChoyTestSuite.verbosity_processes_string(suite)\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Expr'"
] |
[
"def FUNC_4(VAR_0, VAR_6):...\n",
"VAR_7 = psycopg2.connect(dbname=pg_connection['database'], user=\n pg_connection['user'], password=pg_connection['password'], host=\n pg_connection['host'])\n",
"VAR_8 = VAR_7.cursor()\n",
"VAR_8.execute(VAR_6)\n",
"VAR_9 = list(VAR_8.fetchall())\n",
"VAR_8.close()\n",
"VAR_7.close()\n",
"return VAR_9\n"
] | [
"def retrieve_quest_data(pg_connection, query):...\n",
"conn = psycopg2.connect(dbname=pg_connection['database'], user=\n pg_connection['user'], password=pg_connection['password'], host=\n pg_connection['host'])\n",
"cur = conn.cursor()\n",
"cur.execute(query)\n",
"results = list(cur.fetchall())\n",
"cur.close()\n",
"conn.close()\n",
"return results\n"
] | [
0,
0,
0,
4,
4,
0,
0,
4
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Expr'",
"Expr'",
"Return'"
] |
[
"def FUNC_14(self):...\n",
"return self._get_ancestors().values()\n"
] | [
"def get_ancestors(self):...\n",
"return self._get_ancestors().values()\n"
] | [
0,
1
] | [
"FunctionDef'",
"Return'"
] |
[
"@data(('9981', 'B0046A36-3F1C-11E9-9A87-005056AA6F11', 0, 0, 0), ('9982',...\n",
"VAR_13 = {'sender': 'Alex Roy <[email protected]>', 'subject':\n 'DO NOT MODIFY MESSAGE BELOW - JUST HIT `SEND`', 'date':\n 'Tue, 7 May 2019 17:34:17 +0000', 'content':\n f'job_number={VAR_6}string{VAR_7}\\r\\n'}\n",
"VAR_14 = 'string'\n",
"VAR_15 = 'string'\n",
"if VAR_9 or VAR_10:\n",
"conn.cursor().execute(VAR_14.format(VAR_6, VAR_9))\n",
"if VAR_8:\n",
"if VAR_9:\n",
"VAR_30 = pd.read_sql(f'SELECT * FROM df_dilfo WHERE job_number={VAR_6}', conn)\n",
"conn.cursor().execute(VAR_15.format(VAR_6, 1))\n",
"conn.cursor().execute(VAR_15.format(VAR_6, 0))\n",
"VAR_31 = pd.read_sql(f'SELECT * FROM df_matched WHERE job_number={VAR_6}', conn\n )\n",
"process_as_form(VAR_13)\n",
"VAR_32 = pd.read_sql(f'SELECT * FROM df_dilfo WHERE job_number={VAR_6}', conn)\n",
"VAR_33 = pd.read_sql(f'SELECT * FROM df_matched WHERE job_number={VAR_6}', conn\n )\n",
"self.assertEqual(len(VAR_32), 1)\n",
"self.assertEqual(bool(VAR_32.iloc[0].closed), bool(VAR_9 or VAR_7))\n",
"self.assertEqual(any(VAR_33.ground_truth), bool(VAR_9 or VAR_7))\n",
"self.assertEqual(len(VAR_31) + bool(VAR_7 and not VAR_9), len(VAR_33))\n",
"self.assertEqual(list(VAR_31.columns), list(VAR_33.columns))\n",
"self.assertEqual(list(VAR_30.columns), list(VAR_32.columns))\n"
] | [
"@data(('9981', 'B0046A36-3F1C-11E9-9A87-005056AA6F11', 0, 0, 0), ('9982',...\n",
"email_obj = {'sender': 'Alex Roy <[email protected]>', 'subject':\n 'DO NOT MODIFY MESSAGE BELOW - JUST HIT `SEND`', 'date':\n 'Tue, 7 May 2019 17:34:17 +0000', 'content':\n f'job_number={job_number}&title=TEST_ENTRY&city=Ottawa&address=2562+Del+Zotto+Ave.%2C+Ottawa%2C+Ontario&contractor=GCN&engineer=Goodkey&owner=Douglas+Stalker&quality=2&cc_email=&link_to_cert={dcn_key}\\r\\n'\n }\n",
"fake_dilfo_insert = \"\"\"\n INSERT INTO df_dilfo (job_number, receiver_email, closed)\n VALUES ({}, '[email protected]', {})\n \"\"\"\n",
"fake_match_insert = \"\"\"\n INSERT INTO df_matched (job_number, verifier, ground_truth)\n VALUES ({}, '[email protected]', {})\n \"\"\"\n",
"if was_prev_closed or was_prev_tracked:\n",
"conn.cursor().execute(fake_dilfo_insert.format(job_number, was_prev_closed))\n",
"if was_prev_matched:\n",
"if was_prev_closed:\n",
"df_dilfo_pre = pd.read_sql(\n f'SELECT * FROM df_dilfo WHERE job_number={job_number}', conn)\n",
"conn.cursor().execute(fake_match_insert.format(job_number, 1))\n",
"conn.cursor().execute(fake_match_insert.format(job_number, 0))\n",
"df_matched_pre = pd.read_sql(\n f'SELECT * FROM df_matched WHERE job_number={job_number}', conn)\n",
"process_as_form(email_obj)\n",
"df_dilfo_post = pd.read_sql(\n f'SELECT * FROM df_dilfo WHERE job_number={job_number}', conn)\n",
"df_matched_post = pd.read_sql(\n f'SELECT * FROM df_matched WHERE job_number={job_number}', conn)\n",
"self.assertEqual(len(df_dilfo_post), 1)\n",
"self.assertEqual(bool(df_dilfo_post.iloc[0].closed), bool(was_prev_closed or\n dcn_key))\n",
"self.assertEqual(any(df_matched_post.ground_truth), bool(was_prev_closed or\n dcn_key))\n",
"self.assertEqual(len(df_matched_pre) + bool(dcn_key and not was_prev_closed\n ), len(df_matched_post))\n",
"self.assertEqual(list(df_matched_pre.columns), list(df_matched_post.columns))\n",
"self.assertEqual(list(df_dilfo_pre.columns), list(df_dilfo_post.columns))\n"
] | [
0,
0,
4,
0,
0,
4,
0,
0,
4,
4,
4,
4,
0,
4,
4,
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Expr'",
"Condition",
"Condition",
"Assign'",
"Expr'",
"Expr'",
"Assign'",
"Expr'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'"
] |
[
"def FUNC_2(self):...\n",
"VAR_14 = utils.GetUnusedLocalhostPort()\n",
"self._temp_options_filename = options_file.name\n",
"json.dump(dict(self._user_options), options_file)\n",
"options_file.flush()\n",
"VAR_21 = [utils.PathToPythonInterpreter(), FUNC_0(), '--port={0}'.format(\n VAR_14), '--options_file={0}'.format(options_file.name), '--log={0}'.\n format(self._user_options['server_log_level']),\n '--idle_suicide_seconds={0}'.format(VAR_4)]\n",
"if not self._user_options['server_use_vim_stdout']:\n",
"VAR_26 = VAR_0.path.join(utils.PathToTempDir(), 'server_{port}_{std}.log')\n",
"self._server_popen = utils.SafePopen(VAR_21, stdout=PIPE, stderr=PIPE)\n",
"self._server_stdout = VAR_26.format(port=server_port, std='stdout')\n",
"BaseRequest.server_location = 'http://localhost:' + str(VAR_14)\n",
"self._server_stderr = VAR_26.format(port=server_port, std='stderr')\n",
"self._NotifyUserIfServerCrashed()\n",
"VAR_21.append('--stdout={0}'.format(self._server_stdout))\n",
"VAR_21.append('--stderr={0}'.format(self._server_stderr))\n",
"if self._user_options['server_keep_logfiles']:\n",
"VAR_21.append('--keep_logfiles')\n"
] | [
"def _SetupServer(self):...\n",
"server_port = utils.GetUnusedLocalhostPort()\n",
"self._temp_options_filename = options_file.name\n",
"json.dump(dict(self._user_options), options_file)\n",
"options_file.flush()\n",
"args = [utils.PathToPythonInterpreter(), _PathToServerScript(),\n '--port={0}'.format(server_port), '--options_file={0}'.format(\n options_file.name), '--log={0}'.format(self._user_options[\n 'server_log_level']), '--idle_suicide_seconds={0}'.format(\n SERVER_IDLE_SUICIDE_SECONDS)]\n",
"if not self._user_options['server_use_vim_stdout']:\n",
"filename_format = os.path.join(utils.PathToTempDir(), 'server_{port}_{std}.log'\n )\n",
"self._server_popen = utils.SafePopen(args, stdout=PIPE, stderr=PIPE)\n",
"self._server_stdout = filename_format.format(port=server_port, std='stdout')\n",
"BaseRequest.server_location = 'http://localhost:' + str(server_port)\n",
"self._server_stderr = filename_format.format(port=server_port, std='stderr')\n",
"self._NotifyUserIfServerCrashed()\n",
"args.append('--stdout={0}'.format(self._server_stdout))\n",
"args.append('--stderr={0}'.format(self._server_stderr))\n",
"if self._user_options['server_keep_logfiles']:\n",
"args.append('--keep_logfiles')\n"
] | [
0,
0,
7,
7,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Condition",
"Expr'"
] |
[
"def FUNC_2(VAR_1):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_4 = FUNC_1(VAR_1)\n",
"json.dump(VAR_4, sys.stdout, indent=2)\n"
] | [
"def dump_yml_filedir(root_dir):...\n",
"\"\"\"docstring\"\"\"\n",
"result = load_yml_filedir(root_dir)\n",
"json.dump(result, sys.stdout, indent=2)\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Expr'"
] |
[
"def FUNC_6(self, VAR_13, VAR_14='tests'):...\n",
"\"\"\"docstring\"\"\"\n",
"self._logger.csv(VAR_13 + '_mappose', ['X', 'Y', 'yaw'], [self.map_pos.x,\n self.map_pos.y, self.map_angle], VAR_14=folder)\n"
] | [
"def csvLogMap(self, test_name, folder='tests'):...\n",
"\"\"\"docstring\"\"\"\n",
"self._logger.csv(test_name + '_mappose', ['X', 'Y', 'yaw'], [self.map_pos.x,\n self.map_pos.y, self.map_angle], folder=folder)\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Expr'"
] |
[
"def FUNC_18(self):...\n",
"return FUNC_18.dumps(self.properties())\n"
] | [
"def json(self):...\n",
"return json.dumps(self.properties())\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
[
"def FUNC_0(VAR_2):...\n",
"\"\"\"docstring\"\"\"\n",
"if VAR_2.request.method == 'GET':\n",
"VAR_6 = VAR_2.request.session.get_csrf_token()\n",
"if VAR_2.request.cookies.get('XSRF-TOKEN') != VAR_6:\n",
"VAR_2.response.set_cookie(VAR_0, VAR_6)\n"
] | [
"def on_GET_request_setup_csrf_cookie(ev):...\n",
"\"\"\"docstring\"\"\"\n",
"if ev.request.method == 'GET':\n",
"token = ev.request.session.get_csrf_token()\n",
"if ev.request.cookies.get('XSRF-TOKEN') != token:\n",
"ev.response.set_cookie(COOKIE_NAME, token)\n"
] | [
0,
0,
0,
0,
0,
5
] | [
"FunctionDef'",
"Docstring",
"Condition",
"Assign'",
"Condition",
"Expr'"
] |
[
"import psycopg2\n",
"from pyMetricServer.config import *\n",
"import time\n",
"VAR_0 = psycopg2.connect(host=DATABASE_HOST, port=DATABASE_PORT, user=\n DATABASE_USER, password=DATABASE_PASS, VAR_0=DATABASE_NAME)\n",
"VAR_1 = VAR_0.cursor()\n",
"VAR_1.execute('string')\n",
"VAR_1.execute('string')\n",
"VAR_0.commit()\n",
"VAR_1.close()\n",
"def FUNC_0(VAR_2=None, VAR_3=None, VAR_4=None, VAR_5=None, VAR_6=None,...\n",
"VAR_8 = []\n",
"VAR_1 = VAR_0.cursor()\n",
"VAR_9 = []\n",
"VAR_10 = 'SELECT Id, Time, Origin, Key, Value FROM log_metric '\n",
"if VAR_2 != None or VAR_3 != None or VAR_4 != None or VAR_5 != None:\n",
"VAR_10 += 'WHERE '\n",
"if VAR_2 != None:\n",
"VAR_10 += 'Time >= %s AND '\n",
"if VAR_3 != None:\n",
"VAR_9.append(VAR_2)\n",
"VAR_10 += 'Time <= %s AND '\n",
"if VAR_4 != None:\n",
"VAR_9.append(VAR_3)\n",
"VAR_10 += 'Origin = %s AND '\n",
"if VAR_5 != None:\n",
"VAR_9.append(VAR_4)\n",
"VAR_10 += 'Key = %s AND '\n",
"VAR_10 = VAR_10.strip('AND ')\n",
"VAR_9.append(VAR_5)\n",
"VAR_10 += ' '\n",
"if VAR_7 != None and VAR_7[0] != None:\n",
"if VAR_7[1]:\n",
"if VAR_6 != None:\n",
"VAR_10 += 'ORDER BY %s DESC ' % VAR_7[0]\n",
"VAR_10 += 'ORDER BY %s ASC ' % VAR_7[0]\n",
"VAR_10 += 'LIMIT %s '\n",
"VAR_1.execute(VAR_10, tuple(VAR_9))\n",
"VAR_9.append(VAR_6)\n",
"for row in VAR_1:\n",
"VAR_8.append({'Id': str(row[0]), 'Time': str(row[1]), 'Origin': str(row[2]),\n 'Key': str(row[3]), 'Value': str(row[4])})\n",
"return VAR_8\n"
] | [
"import psycopg2\n",
"from pyMetricServer.config import *\n",
"import time\n",
"database = psycopg2.connect(host=DATABASE_HOST, port=DATABASE_PORT, user=\n DATABASE_USER, password=DATABASE_PASS, database=DATABASE_NAME)\n",
"cursor = database.cursor()\n",
"cursor.execute(\n 'CREATE TABLE IF NOT EXISTS log_messages (Id BIGSERIAL, Time INTEGER, Origin TEXT, Message TEXT, Type INTEGER);'\n )\n",
"cursor.execute(\n 'CREATE TABLE IF NOT EXISTS log_metric (Id BIGSERIAL, Time INTEGER, Origin TEXT, Key TEXT, Value DOUBLE PRECISION)'\n )\n",
"database.commit()\n",
"cursor.close()\n",
"def getMetric(timefrom=None, timeto=None, origin=None, key=None, count=None,...\n",
"results = []\n",
"cursor = database.cursor()\n",
"params = []\n",
"query = 'SELECT Id, Time, Origin, Key, Value FROM log_metric '\n",
"if timefrom != None or timeto != None or origin != None or key != None:\n",
"query += 'WHERE '\n",
"if timefrom != None:\n",
"query += 'Time >= %s AND '\n",
"if timeto != None:\n",
"params.append(timefrom)\n",
"query += 'Time <= %s AND '\n",
"if origin != None:\n",
"params.append(timeto)\n",
"query += 'Origin = %s AND '\n",
"if key != None:\n",
"params.append(origin)\n",
"query += 'Key = %s AND '\n",
"query = query.strip('AND ')\n",
"params.append(key)\n",
"query += ' '\n",
"if order != None and order[0] != None:\n",
"if order[1]:\n",
"if count != None:\n",
"query += 'ORDER BY %s DESC ' % order[0]\n",
"query += 'ORDER BY %s ASC ' % order[0]\n",
"query += 'LIMIT %s '\n",
"cursor.execute(query, tuple(params))\n",
"params.append(count)\n",
"for row in cursor:\n",
"results.append({'Id': str(row[0]), 'Time': str(row[1]), 'Origin': str(row[2\n ]), 'Key': str(row[3]), 'Value': str(row[4])})\n",
"return results\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
4,
0,
4,
4,
0,
0,
0,
0,
0,
0
] | [
"Import'",
"ImportFrom'",
"Import'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"AugAssign'",
"Condition",
"AugAssign'",
"Condition",
"Expr'",
"AugAssign'",
"Condition",
"Expr'",
"AugAssign'",
"Condition",
"Expr'",
"AugAssign'",
"Assign'",
"Expr'",
"AugAssign'",
"Condition",
"Condition",
"Condition",
"AugAssign'",
"AugAssign'",
"AugAssign'",
"Expr'",
"Expr'",
"For",
"Expr'",
"Return'"
] |
[
"def FUNC_7(self, VAR_5):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_13 = (KeyError, ValueError, TypeError, AttributeError, SQLAlchemyError,\n DokomoError)\n",
"if isinstance(VAR_5, tornado.web.HTTPError):\n",
"VAR_33 = exc.HttpError(VAR_5.log_message)\n",
"if isinstance(VAR_5, VAR_13):\n",
"VAR_33.status = VAR_5.status_code\n",
"VAR_5 = exc.BadRequest(VAR_5)\n",
"return super().handle_error(VAR_5)\n",
"VAR_5 = VAR_33\n"
] | [
"def handle_error(self, err):...\n",
"\"\"\"docstring\"\"\"\n",
"understood = (KeyError, ValueError, TypeError, AttributeError,\n SQLAlchemyError, DokomoError)\n",
"if isinstance(err, tornado.web.HTTPError):\n",
"restless_error = exc.HttpError(err.log_message)\n",
"if isinstance(err, understood):\n",
"restless_error.status = err.status_code\n",
"err = exc.BadRequest(err)\n",
"return super().handle_error(err)\n",
"err = restless_error\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Return'",
"Assign'"
] |
[
"def FUNC_2(VAR_0, VAR_1):...\n",
"VAR_5 = frappe.session.user\n",
"VAR_6 = get_leave_allocation_records(VAR_0.to_date)\n",
"VAR_7 = get_leave_allocation_records(VAR_0.from_date)\n",
"VAR_8 = frappe.get_all('Employee', VAR_0={'status': 'Active', 'company':\n filters.company}, fields=['name', 'employee_name', 'department', 'user_id']\n )\n",
"VAR_4 = []\n",
"for employee in VAR_8:\n",
"VAR_12 = FUNC_3(employee.department)\n",
"return VAR_4\n",
"if len(VAR_12) and VAR_5 in VAR_12 or VAR_5 in ['Administrator', employee.\n",
"VAR_13 = [employee.name, employee.employee_name, employee.department]\n",
"for leave_type in VAR_1:\n",
"VAR_14 = get_approved_leaves_for_period(employee.name, leave_type, VAR_0.\n from_date, VAR_0.to_date)\n",
"VAR_4.append(VAR_13)\n",
"VAR_15 = get_leave_balance_on(employee.name, leave_type, VAR_0.from_date,\n VAR_7.get(employee.name, frappe._dict()))\n",
"VAR_16 = get_leave_balance_on(employee.name, leave_type, VAR_0.to_date,\n VAR_6.get(employee.name, frappe._dict()))\n",
"VAR_13 += [VAR_15, VAR_14, VAR_16]\n"
] | [
"def get_data(filters, leave_types):...\n",
"user = frappe.session.user\n",
"allocation_records_based_on_to_date = get_leave_allocation_records(filters.\n to_date)\n",
"allocation_records_based_on_from_date = get_leave_allocation_records(filters\n .from_date)\n",
"active_employees = frappe.get_all('Employee', filters={'status': 'Active',\n 'company': filters.company}, fields=['name', 'employee_name',\n 'department', 'user_id'])\n",
"data = []\n",
"for employee in active_employees:\n",
"leave_approvers = get_approvers(employee.department)\n",
"return data\n",
"if len(leave_approvers) and user in leave_approvers or user in ['Administrator'\n",
"row = [employee.name, employee.employee_name, employee.department]\n",
"for leave_type in leave_types:\n",
"leaves_taken = get_approved_leaves_for_period(employee.name, leave_type,\n filters.from_date, filters.to_date)\n",
"data.append(row)\n",
"opening = get_leave_balance_on(employee.name, leave_type, filters.from_date,\n allocation_records_based_on_from_date.get(employee.name, frappe._dict()))\n",
"closing = get_leave_balance_on(employee.name, leave_type, filters.to_date,\n allocation_records_based_on_to_date.get(employee.name, frappe._dict()))\n",
"row += [opening, leaves_taken, closing]\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
4,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"For",
"Assign'",
"Return'",
"Condition",
"Assign'",
"For",
"Assign'",
"Expr'",
"Assign'",
"Assign'",
"AugAssign'"
] |
[
"@property...\n",
"\"\"\"docstring\"\"\"\n",
"return set(VAR_28 for VAR_28 in self.input if not VAR_28.exists and not \n VAR_28 in self.subworkflow_input)\n"
] | [
"@property...\n",
"\"\"\"docstring\"\"\"\n",
"return set(f for f in self.input if not f.exists and not f in self.\n subworkflow_input)\n"
] | [
0,
0,
0
] | [
"Condition",
"Docstring",
"Return'"
] |
[
"def FUNC_0(VAR_0):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_1 = True if VAR_0.find(':') != -1 else False\n",
"if VAR_0[-4:] == '.rpm':\n",
"VAR_0 = VAR_0[:-4]\n",
"VAR_2 = VAR_0.rfind('.')\n",
"VAR_3 = VAR_0[VAR_2 + 1:]\n",
"VAR_4 = VAR_0[:VAR_2].rfind('-')\n",
"VAR_5 = VAR_0[VAR_4 + 1:VAR_2]\n",
"if VAR_1:\n",
"VAR_10 = VAR_0[:VAR_4].rfind(':')\n",
"VAR_10 = VAR_0[:VAR_4].rfind('-')\n",
"VAR_6 = VAR_0[VAR_10 + 1:VAR_4]\n",
"if VAR_1:\n",
"VAR_11 = VAR_0[:VAR_10].rfind('-')\n",
"VAR_11 = VAR_10\n",
"VAR_12 = VAR_0[VAR_11 + 1:VAR_10]\n",
"VAR_12 = '0'\n",
"VAR_7 = VAR_0[:VAR_11]\n",
"return VAR_7, VAR_6, VAR_5, VAR_12, VAR_3\n"
] | [
"def split_filename(filename):...\n",
"\"\"\"docstring\"\"\"\n",
"is_epoch = True if filename.find(':') != -1 else False\n",
"if filename[-4:] == '.rpm':\n",
"filename = filename[:-4]\n",
"arch_index = filename.rfind('.')\n",
"arch = filename[arch_index + 1:]\n",
"rel_index = filename[:arch_index].rfind('-')\n",
"rel = filename[rel_index + 1:arch_index]\n",
"if is_epoch:\n",
"ver_index = filename[:rel_index].rfind(':')\n",
"ver_index = filename[:rel_index].rfind('-')\n",
"ver = filename[ver_index + 1:rel_index]\n",
"if is_epoch:\n",
"epoch_index = filename[:ver_index].rfind('-')\n",
"epoch_index = ver_index\n",
"epoch = filename[epoch_index + 1:ver_index]\n",
"epoch = '0'\n",
"name = filename[:epoch_index]\n",
"return name, ver, rel, epoch, arch\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Return'"
] |
[
"\"\"\"string\"\"\"\n",
"from __future__ import unicode_literals\n",
"from django.conf import settings\n",
"from django.conf.urls import include\n",
"from django.conf.urls import patterns\n",
"from django.conf.urls import url\n",
"from api import routers\n",
"from api import views\n",
"VAR_0 = routers.ApiRouter()\n",
"VAR_1 = patterns('', url('^', include(VAR_0.urls)), url(\n '^apps/(?P<id>{})/config/?'.format(settings.APP_URL_REGEX), views.\n AppConfigViewSet.as_view({'get': 'retrieve', 'post': 'create'})), url(\n '^apps/(?P<id>{})/builds/(?P<uuid>[-_\\\\w]+)/?'.format(settings.\n APP_URL_REGEX), views.AppBuildViewSet.as_view({'get': 'retrieve'})),\n url('^apps/(?P<id>{})/builds/?'.format(settings.APP_URL_REGEX), views.\n AppBuildViewSet.as_view({'get': 'list', 'post': 'create'})), url(\n '^apps/(?P<id>{})/releases/v(?P<version>[0-9]+)/?'.format(settings.\n APP_URL_REGEX), views.AppReleaseViewSet.as_view({'get': 'retrieve'})),\n url('^apps/(?P<id>{})/releases/rollback/?'.format(settings.\n APP_URL_REGEX), views.AppReleaseViewSet.as_view({'post': 'rollback'})),\n url('^apps/(?P<id>{})/releases/?'.format(settings.APP_URL_REGEX), views\n .AppReleaseViewSet.as_view({'get': 'list'})), url(\n '^apps/(?P<id>{})/containers/(?P<type>[-_\\\\w]+)/(?P<num>[-_\\\\w]+)/?'.\n format(settings.APP_URL_REGEX), views.AppContainerViewSet.as_view({\n 'get': 'retrieve'})), url(\n '^apps/(?P<id>{})/containers/(?P<type>[-_\\\\w.]+)/?'.format(settings.\n APP_URL_REGEX), views.AppContainerViewSet.as_view({'get': 'list'})),\n url('^apps/(?P<id>{})/containers/?'.format(settings.APP_URL_REGEX),\n views.AppContainerViewSet.as_view({'get': 'list'})), url(\n '^apps/(?P<id>{})/domains/(?P<domain>[-\\\\._\\\\w]+)/?'.format(settings.\n APP_URL_REGEX), views.DomainViewSet.as_view({'delete': 'destroy'})),\n url('^apps/(?P<id>{})/domains/?'.format(settings.APP_URL_REGEX), views.\n DomainViewSet.as_view({'post': 'create', 'get': 'list'})), url(\n '^apps/(?P<id>{})/scale/?'.format(settings.APP_URL_REGEX), views.\n AppViewSet.as_view({'post': 'scale'})), url('^apps/(?P<id>{})/logs/?'.\n format(settings.APP_URL_REGEX), views.AppViewSet.as_view({'get': 'logs'\n })), url('^apps/(?P<id>{})/run/?'.format(settings.APP_URL_REGEX), views\n .AppViewSet.as_view({'post': 'run'})), url(\n '^apps/(?P<id>{})/perms/(?P<username>[-_\\\\w]+)/?'.format(settings.\n APP_URL_REGEX), views.AppPermsViewSet.as_view({'delete': 'destroy'})),\n url('^apps/(?P<id>{})/perms/?'.format(settings.APP_URL_REGEX), views.\n AppPermsViewSet.as_view({'get': 'list', 'post': 'create'})), url(\n '^apps/(?P<id>{})/?'.format(settings.APP_URL_REGEX), views.AppViewSet.\n as_view({'get': 'retrieve', 'delete': 'destroy'})), url('^apps/?',\n views.AppViewSet.as_view({'get': 'list', 'post': 'create'})), url(\n '^keys/(?P<id>.+)/?', views.KeyViewSet.as_view({'get': 'retrieve',\n 'delete': 'destroy'})), url('^keys/?', views.KeyViewSet.as_view({'get':\n 'list', 'post': 'create'})), url('^hooks/push/?', views.PushHookViewSet\n .as_view({'post': 'create'})), url('^hooks/build/?', views.\n BuildHookViewSet.as_view({'post': 'create'})), url('^hooks/config/?',\n views.ConfigHookViewSet.as_view({'post': 'create'})), url(\n '^auth/register/?', views.UserRegistrationView.as_view({'post':\n 'create'})), url('^auth/cancel/?', views.UserCancellationView.as_view({\n 'delete': 'destroy'})), url('^auth/', include('rest_framework.urls',\n namespace='rest_framework')), url('^generate-api-key/',\n 'rest_framework.authtoken.views.obtain_auth_token'), url(\n '^admin/perms/(?P<username>[-_\\\\w]+)/?', views.AdminPermsViewSet.\n as_view({'delete': 'destroy'})), url('^admin/perms/?', views.\n AdminPermsViewSet.as_view({'get': 'list', 'post': 'create'})))\n"
] | [
"\"\"\"\nRESTful URL patterns and routing for the Deis API app.\n\n\nApplications\n============\n\n.. http:get:: /api/apps/(string:id)/\n\n Retrieve a :class:`~api.models.App` by its `id`.\n\n.. http:delete:: /api/apps/(string:id)/\n\n Destroy a :class:`~api.models.App` by its `id`.\n\n.. http:get:: /api/apps/\n\n List all :class:`~api.models.App`\\\\s.\n\n.. http:post:: /api/apps/\n\n Create a new :class:`~api.models.App`.\n\n\nApplication Release Components\n------------------------------\n\n.. http:get:: /api/apps/(string:id)/config/\n\n List all :class:`~api.models.Config`\\\\s.\n\n.. http:post:: /api/apps/(string:id)/config/\n\n Create a new :class:`~api.models.Config`.\n\n.. http:get:: /api/apps/(string:id)/builds/(string:uuid)/\n\n Retrieve a :class:`~api.models.Build` by its `uuid`.\n\n.. http:get:: /api/apps/(string:id)/builds/\n\n List all :class:`~api.models.Build`\\\\s.\n\n.. http:post:: /api/apps/(string:id)/builds/\n\n Create a new :class:`~api.models.Build`.\n\n.. http:get:: /api/apps/(string:id)/releases/(int:version)/\n\n Retrieve a :class:`~api.models.Release` by its `version`.\n\n.. http:get:: /api/apps/(string:id)/releases/\n\n List all :class:`~api.models.Release`\\\\s.\n\n.. http:post:: /api/apps/(string:id)/releases/rollback/\n\n Rollback to a previous :class:`~api.models.Release`.\n\n\nApplication Infrastructure\n--------------------------\n\n.. http:get:: /api/apps/(string:id)/containers/(string:type)/(int:num)/\n\n List all :class:`~api.models.Container`\\\\s.\n\n.. http:get:: /api/apps/(string:id)/containers/(string:type)/\n\n List all :class:`~api.models.Container`\\\\s.\n\n.. http:get:: /api/apps/(string:id)/containers/\n\n List all :class:`~api.models.Container`\\\\s.\n\n\nApplication Domains\n-------------------\n\n\n.. http:delete:: /api/apps/(string:id)/domains/(string:hostname)\n\n Destroy a :class:`~api.models.Domain` by its `hostname`\n\n.. http:get:: /api/apps/(string:id)/domains/\n\n List all :class:`~api.models.Domain`\\\\s.\n\n.. http:post:: /api/apps/(string:id)/domains/\n\n Create a new :class:`~api.models.Domain`\\\\s.\n\n\nApplication Actions\n-------------------\n\n.. http:post:: /api/apps/(string:id)/scale/\n\n See also\n :meth:`AppViewSet.scale() <api.views.AppViewSet.scale>`\n\n.. http:get:: /api/apps/(string:id)/logs/\n\n See also\n :meth:`AppViewSet.logs() <api.views.AppViewSet.logs>`\n\n.. http:post:: /api/apps/(string:id)/run/\n\n See also\n :meth:`AppViewSet.run() <api.views.AppViewSet.run>`\n\n\nApplication Sharing\n===================\n\n.. http:delete:: /api/apps/(string:id)/perms/(string:username)/\n\n Destroy an app permission by its `username`.\n\n.. http:get:: /api/apps/(string:id)/perms/\n\n List all permissions granted to this app.\n\n.. http:post:: /api/apps/(string:id)/perms/\n\n Create a new app permission.\n\n\nKeys\n====\n\n.. http:get:: /api/keys/(string:id)/\n\n Retrieve a :class:`~api.models.Key` by its `id`.\n\n.. http:delete:: /api/keys/(string:id)/\n\n Destroy a :class:`~api.models.Key` by its `id`.\n\n.. http:get:: /api/keys/\n\n List all :class:`~api.models.Key`\\\\s.\n\n.. http:post:: /api/keys/\n\n Create a new :class:`~api.models.Key`.\n\n\nAPI Hooks\n=========\n\n.. http:post:: /api/hooks/push/\n\n Create a new :class:`~api.models.Push`.\n\n.. http:post:: /api/hooks/build/\n\n Create a new :class:`~api.models.Build`.\n\n.. http:post:: /api/hooks/config/\n\n Retrieve latest application :class:`~api.models.Config`.\n\n\nAuth\n====\n\n.. http:post:: /api/auth/register/\n\n Create a new User.\n\n.. http:delete:: /api/auth/register/\n\n Destroy the logged-in User.\n\n.. http:post:: /api/auth/login\n\n Authenticate for the REST framework.\n\n.. http:post:: /api/auth/logout\n\n Clear authentication for the REST framework.\n\n.. http:get:: /api/generate-api-key/\n\n Generate an API key.\n\n\nAdmin Sharing\n=============\n\n.. http:delete:: /api/admin/perms/(string:username)/\n\n Destroy an admin permission by its `username`.\n\n.. http:get:: /api/admin/perms/\n\n List all admin permissions granted.\n\n.. http:post:: /api/admin/perms/\n\n Create a new admin permission.\n\n\"\"\"\n",
"from __future__ import unicode_literals\n",
"from django.conf import settings\n",
"from django.conf.urls import include\n",
"from django.conf.urls import patterns\n",
"from django.conf.urls import url\n",
"from api import routers\n",
"from api import views\n",
"router = routers.ApiRouter()\n",
"urlpatterns = patterns('', url('^', include(router.urls)), url(\n '^apps/(?P<id>{})/config/?'.format(settings.APP_URL_REGEX), views.\n AppConfigViewSet.as_view({'get': 'retrieve', 'post': 'create'})), url(\n '^apps/(?P<id>{})/builds/(?P<uuid>[-_\\\\w]+)/?'.format(settings.\n APP_URL_REGEX), views.AppBuildViewSet.as_view({'get': 'retrieve'})),\n url('^apps/(?P<id>{})/builds/?'.format(settings.APP_URL_REGEX), views.\n AppBuildViewSet.as_view({'get': 'list', 'post': 'create'})), url(\n '^apps/(?P<id>{})/releases/v(?P<version>[0-9]+)/?'.format(settings.\n APP_URL_REGEX), views.AppReleaseViewSet.as_view({'get': 'retrieve'})),\n url('^apps/(?P<id>{})/releases/rollback/?'.format(settings.\n APP_URL_REGEX), views.AppReleaseViewSet.as_view({'post': 'rollback'})),\n url('^apps/(?P<id>{})/releases/?'.format(settings.APP_URL_REGEX), views\n .AppReleaseViewSet.as_view({'get': 'list'})), url(\n '^apps/(?P<id>{})/containers/(?P<type>[-_\\\\w]+)/(?P<num>[-_\\\\w]+)/?'.\n format(settings.APP_URL_REGEX), views.AppContainerViewSet.as_view({\n 'get': 'retrieve'})), url(\n '^apps/(?P<id>{})/containers/(?P<type>[-_\\\\w.]+)/?'.format(settings.\n APP_URL_REGEX), views.AppContainerViewSet.as_view({'get': 'list'})),\n url('^apps/(?P<id>{})/containers/?'.format(settings.APP_URL_REGEX),\n views.AppContainerViewSet.as_view({'get': 'list'})), url(\n '^apps/(?P<id>{})/domains/(?P<domain>[-\\\\._\\\\w]+)/?'.format(settings.\n APP_URL_REGEX), views.DomainViewSet.as_view({'delete': 'destroy'})),\n url('^apps/(?P<id>{})/domains/?'.format(settings.APP_URL_REGEX), views.\n DomainViewSet.as_view({'post': 'create', 'get': 'list'})), url(\n '^apps/(?P<id>{})/scale/?'.format(settings.APP_URL_REGEX), views.\n AppViewSet.as_view({'post': 'scale'})), url('^apps/(?P<id>{})/logs/?'.\n format(settings.APP_URL_REGEX), views.AppViewSet.as_view({'get': 'logs'\n })), url('^apps/(?P<id>{})/run/?'.format(settings.APP_URL_REGEX), views\n .AppViewSet.as_view({'post': 'run'})), url(\n '^apps/(?P<id>{})/perms/(?P<username>[-_\\\\w]+)/?'.format(settings.\n APP_URL_REGEX), views.AppPermsViewSet.as_view({'delete': 'destroy'})),\n url('^apps/(?P<id>{})/perms/?'.format(settings.APP_URL_REGEX), views.\n AppPermsViewSet.as_view({'get': 'list', 'post': 'create'})), url(\n '^apps/(?P<id>{})/?'.format(settings.APP_URL_REGEX), views.AppViewSet.\n as_view({'get': 'retrieve', 'delete': 'destroy'})), url('^apps/?',\n views.AppViewSet.as_view({'get': 'list', 'post': 'create'})), url(\n '^keys/(?P<id>.+)/?', views.KeyViewSet.as_view({'get': 'retrieve',\n 'delete': 'destroy'})), url('^keys/?', views.KeyViewSet.as_view({'get':\n 'list', 'post': 'create'})), url('^hooks/push/?', views.PushHookViewSet\n .as_view({'post': 'create'})), url('^hooks/build/?', views.\n BuildHookViewSet.as_view({'post': 'create'})), url('^hooks/config/?',\n views.ConfigHookViewSet.as_view({'post': 'create'})), url(\n '^auth/register/?', views.UserRegistrationView.as_view({'post':\n 'create'})), url('^auth/cancel/?', views.UserCancellationView.as_view({\n 'delete': 'destroy'})), url('^auth/', include('rest_framework.urls',\n namespace='rest_framework')), url('^generate-api-key/',\n 'rest_framework.authtoken.views.obtain_auth_token'), url(\n '^admin/perms/(?P<username>[-_\\\\w]+)/?', views.AdminPermsViewSet.\n as_view({'delete': 'destroy'})), url('^admin/perms/?', views.\n AdminPermsViewSet.as_view({'get': 'list', 'post': 'create'})))\n"
] | [
5,
0,
0,
0,
0,
0,
0,
0,
0,
5
] | [
"Expr'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"Assign'",
"Assign'"
] |
[
"@user_passes_test(user_is_superuser)...\n",
"VAR_8 = Project.objects.all()\n",
"for VAR_17 in VAR_8:\n",
"create_testcases_celery.delay(VAR_17.id)\n",
"return redirect('testcases:testcases')\n"
] | [
"@user_passes_test(user_is_superuser)...\n",
"projects = Project.objects.all()\n",
"for project in projects:\n",
"create_testcases_celery.delay(project.id)\n",
"return redirect('testcases:testcases')\n"
] | [
0,
0,
0,
0,
0
] | [
"Condition",
"Assign'",
"For",
"Expr'",
"Return'"
] |
[
"def FUNC_11(self):...\n",
"VAR_20 = []\n",
"VAR_18 = threading.Event()\n",
"self.mock(VAR_18, 'wait', self.fail)\n",
"self.mock(bot_main, 'run_manifest', self.fail)\n",
"self.mock(bot_main, 'update_bot', lambda *VAR_24: VAR_20.append(VAR_24))\n",
"self.expected_requests([(\n 'https://localhost:1/auth/api/v1/accounts/self/xsrf_token', {'data': {},\n 'headers': {'X-XSRF-Token-Request': '1'}}, {'xsrf_token': 'token'}), (\n 'https://localhost:1/swarming/api/v1/bot/poll', {'data': self.\n attributes, 'headers': {'X-XSRF-Token': 'token'}}, {'cmd': 'update',\n 'version': '123'})])\n",
"self.assertTrue(bot_main.poll_server(self.bot, VAR_18))\n",
"self.assertEqual([(self.bot, '123')], VAR_20)\n"
] | [
"def test_poll_server_update(self):...\n",
"update = []\n",
"bit = threading.Event()\n",
"self.mock(bit, 'wait', self.fail)\n",
"self.mock(bot_main, 'run_manifest', self.fail)\n",
"self.mock(bot_main, 'update_bot', lambda *args: update.append(args))\n",
"self.expected_requests([(\n 'https://localhost:1/auth/api/v1/accounts/self/xsrf_token', {'data': {},\n 'headers': {'X-XSRF-Token-Request': '1'}}, {'xsrf_token': 'token'}), (\n 'https://localhost:1/swarming/api/v1/bot/poll', {'data': self.\n attributes, 'headers': {'X-XSRF-Token': 'token'}}, {'cmd': 'update',\n 'version': '123'})])\n",
"self.assertTrue(bot_main.poll_server(self.bot, bit))\n",
"self.assertEqual([(self.bot, '123')], update)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'"
] |
[
"def __new__(VAR_10, VAR_11, VAR_12, VAR_13):...\n",
"for VAR_47 in webapp2.WSGIApplication.allowed_methods:\n",
"VAR_48 = VAR_13.get(VAR_47.lower())\n",
"return type.__new__(VAR_10, VAR_11, VAR_12, VAR_13)\n",
"if VAR_48 and not api.is_decorated(VAR_48):\n"
] | [
"def __new__(mcs, name, bases, attributes):...\n",
"for method in webapp2.WSGIApplication.allowed_methods:\n",
"func = attributes.get(method.lower())\n",
"return type.__new__(mcs, name, bases, attributes)\n",
"if func and not api.is_decorated(func):\n"
] | [
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"For",
"Assign'",
"Return'",
"Condition"
] |
[
"def FUNC_8(self):...\n",
"self._WriteHeader('text/html', VAR_22=404)\n",
"self.wfile.write('<html><body>%s not found</body></html>' % self.path)\n"
] | [
"def _DoNotFound(self):...\n",
"self._WriteHeader('text/html', status_code=404)\n",
"self.wfile.write('<html><body>%s not found</body></html>' % self.path)\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Expr'",
"Expr'"
] |
[
"def FUNC_2(self, VAR_5):...\n",
"if self.closed:\n",
"if self._adapter is not None:\n",
"assert self._session is None\n"
] | [
"def start_debugging(self, launchcfg):...\n",
"if self.closed:\n",
"if self._adapter is not None:\n",
"assert self._session is None\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"Condition",
"Condition",
"Assert'"
] |
[
"def __init__(self, VAR_0=None):...\n",
"super(CLASS_0, self).__init__()\n",
"if VAR_0:\n",
"self._init_graph(VAR_0)\n"
] | [
"def __init__(self, workflow_job=None):...\n",
"super(WorkflowDAG, self).__init__()\n",
"if workflow_job:\n",
"self._init_graph(workflow_job)\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"Expr'",
"Condition",
"Expr'"
] |
[
"def __missing__(self, VAR_30):...\n",
"VAR_66 = self[VAR_30] = VAR_31(*VAR_30)\n",
"return VAR_66\n"
] | [
"def __missing__(self, key):...\n",
"ret = self[key] = f(*key)\n",
"return ret\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Return'"
] |
[
"def FUNC_4(self):...\n",
"if hasattr(self, 'extractedcpcat'):\n",
"return self.extractedcpcat\n",
"return self\n"
] | [
"def pull_out_cp(self):...\n",
"if hasattr(self, 'extractedcpcat'):\n",
"return self.extractedcpcat\n",
"return self\n"
] | [
0,
6,
6,
6
] | [
"FunctionDef'",
"Condition",
"Return'",
"Return'"
] |
[
"def FUNC_12(VAR_7):...\n",
"@wraps(VAR_7)...\n",
"if not app.config.get('USE_AUTH'):\n",
"return VAR_7(*VAR_9, **kwargs)\n",
"FUNC_0()\n",
"return VAR_7(*VAR_9, **kwargs)\n",
"logging.warning('require_logout(): calling log_out()')\n",
"VAR_16 = VAR_1.log_out()\n",
"if VAR_16.headers.get('Location') == url_for('goodbye'):\n",
"return VAR_7(*VAR_9, **kwargs)\n",
"return VAR_16\n"
] | [
"def require_logout_for_goodbye(f):...\n",
"@wraps(f)...\n",
"if not app.config.get('USE_AUTH'):\n",
"return f(*args, **kwargs)\n",
"get_logged_in_user()\n",
"return f(*args, **kwargs)\n",
"logging.warning('require_logout(): calling log_out()')\n",
"resp = user_mod.log_out()\n",
"if resp.headers.get('Location') == url_for('goodbye'):\n",
"return f(*args, **kwargs)\n",
"return resp\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Condition",
"Condition",
"Return'",
"Expr'",
"Return'",
"Expr'",
"Assign'",
"For",
"Return'",
"Return'"
] |
[
"def FUNC_5(VAR_2, VAR_6):...\n",
"for _privilege in VAR_0[VAR_2]:\n",
"if fnmatch.fnmatch(VAR_6, _privilege):\n",
"return False\n",
"return True\n"
] | [
"def user_type_has_privilege(user_type, privilege):...\n",
"for _privilege in PRIVILEGES[user_type]:\n",
"if fnmatch.fnmatch(privilege, _privilege):\n",
"return False\n",
"return True\n"
] | [
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"For",
"Condition",
"Return'",
"Return'"
] |
[
"@staticmethod...\n",
""
] | [
"@staticmethod...\n",
""
] | [
0,
0
] | [
"Condition",
"Condition"
] |
[
"def FUNC_2(self, VAR_8):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_5 = {}\n",
"VAR_5['clusterName'] = VAR_8\n",
"VAR_5['searchDepth'] = '1'\n",
"VAR_5['verbose'] = '0'\n",
"VAR_18 = self._cliq_run_xml('getClusterInfo', VAR_5)\n",
"return VAR_18\n"
] | [
"def _cliq_get_cluster_info(self, cluster_name):...\n",
"\"\"\"docstring\"\"\"\n",
"cliq_args = {}\n",
"cliq_args['clusterName'] = cluster_name\n",
"cliq_args['searchDepth'] = '1'\n",
"cliq_args['verbose'] = '0'\n",
"result_xml = self._cliq_run_xml('getClusterInfo', cliq_args)\n",
"return result_xml\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Return'"
] |
[
"@property...\n",
"return list(self)\n"
] | [
"@property...\n",
"return list(self)\n"
] | [
0,
0
] | [
"Condition",
"Return'"
] |
[
"def FUNC_0(VAR_6):...\n",
"VAR_10 = VAR_6.find('http:')\n",
"VAR_11 = VAR_6.find('https:')\n",
"if VAR_11 == -1:\n",
"if VAR_10 == -1:\n",
"if VAR_10 == -1:\n",
"return None\n",
"return VAR_6[VAR_10:]\n",
"return VAR_6[VAR_11:]\n",
"return VAR_6[min(VAR_10, VAR_11):]\n"
] | [
"def ParseURL(s):...\n",
"http_idx = s.find('http:')\n",
"https_idx = s.find('https:')\n",
"if https_idx == -1:\n",
"if http_idx == -1:\n",
"if http_idx == -1:\n",
"return None\n",
"return s[http_idx:]\n",
"return s[https_idx:]\n",
"return s[min(http_idx, https_idx):]\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Condition",
"Condition",
"Condition",
"Return'",
"Return'",
"Return'",
"Return'"
] |
[
"def FUNC_24(self, VAR_9):...\n",
"\"\"\"docstring\"\"\"\n",
"def FUNC_25(VAR_10):...\n",
"self.events_requests.remove(VAR_9)\n",
"self.events_requests.append(VAR_9)\n",
"VAR_9.notifyFinish().addCallbacks(FUNC_25, FUNC_25)\n",
"VAR_9.write(json.dumps({'type': 'events_start', 'event': {'tribler_started':\n self.session.lm.initComplete, 'version': version_id}}) + '\\n')\n",
"return server.NOT_DONE_YET\n"
] | [
"def render_GET(self, request):...\n",
"\"\"\"docstring\"\"\"\n",
"def on_request_finished(_):...\n",
"self.events_requests.remove(request)\n",
"self.events_requests.append(request)\n",
"request.notifyFinish().addCallbacks(on_request_finished, on_request_finished)\n",
"request.write(json.dumps({'type': 'events_start', 'event': {\n 'tribler_started': self.session.lm.initComplete, 'version': version_id}\n }) + '\\n')\n",
"return server.NOT_DONE_YET\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"FunctionDef'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Return'"
] |
[
"def __init__(self, *VAR_2, **VAR_3):...\n",
"super(CLASS_0, self).__init__(*VAR_2, **kwargs)\n",
"self.cluster_vip = None\n"
] | [
"def __init__(self, *args, **kwargs):...\n",
"super(HpSanISCSIDriver, self).__init__(*args, **kwargs)\n",
"self.cluster_vip = None\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Expr'",
"Assign'"
] |
[
"def FUNC_7(self):...\n",
"return self.data\n"
] | [
"def to_json(self):...\n",
"return self.data\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
[
"def FUNC_4(self):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_11 = VAR_0[:]\n",
"VAR_11.extend([(Allow, user.login, ('view_company', 'edit_company',\n 'view.file', 'list_customers', 'add_customer', 'list_projects',\n 'add_project', 'list_estimations', 'list_invoices',\n 'edit_commercial_handling', 'list_expenses', 'add.expense',\n 'list_sale_products', 'add_sale_product', 'list_treasury_files',\n 'list_activities', 'list_workshops')) for user in self.employees])\n",
"return VAR_11\n"
] | [
"def get_company_acl(self):...\n",
"\"\"\"docstring\"\"\"\n",
"acl = DEFAULT_PERM[:]\n",
"acl.extend([(Allow, user.login, ('view_company', 'edit_company',\n 'view.file', 'list_customers', 'add_customer', 'list_projects',\n 'add_project', 'list_estimations', 'list_invoices',\n 'edit_commercial_handling', 'list_expenses', 'add.expense',\n 'list_sale_products', 'add_sale_product', 'list_treasury_files',\n 'list_activities', 'list_workshops')) for user in self.employees])\n",
"return acl\n"
] | [
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Expr'",
"Return'"
] |
[
"def FUNC_9(self, VAR_9, VAR_10):...\n",
"if self._IsServerAlive():\n",
"return FUNC_9(VAR_9, VAR_10)\n"
] | [
"def SendCommandRequest(self, arguments, completer):...\n",
"if self._IsServerAlive():\n",
"return SendCommandRequest(arguments, completer)\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Condition",
"Return'"
] |
[
"def FUNC_1(self):...\n",
"return 'load statement for task %s (%s)' % (self.task_id, self.url)\n"
] | [
"def describe(self):...\n",
"return 'load statement for task %s (%s)' % (self.task_id, self.url)\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
[
"def FUNC_7(self, VAR_12, VAR_13):...\n",
"self._credentials[VAR_12] = VAR_13\n"
] | [
"def add_user(self, username, password):...\n",
"self._credentials[username] = password\n"
] | [
0,
0
] | [
"FunctionDef'",
"Assign'"
] |
[
"def FUNC_6(VAR_12, VAR_3, VAR_4=-1):...\n",
"VAR_7 = VAR_12[0]\n",
"VAR_24 = VAR_12[1]\n",
"if not VAR_24:\n",
"return VAR_24\n",
"if VAR_7 == -2 ** 255 and VAR_24 == -1:\n",
"return -2 ** 255\n",
"return copysign(abs(VAR_7 / VAR_24), VAR_7 / VAR_24)\n"
] | [
"def signedDiv(params, symbols, symId=-1):...\n",
"x = params[0]\n",
"y = params[1]\n",
"if not y:\n",
"return y\n",
"if x == -2 ** 255 and y == -1:\n",
"return -2 ** 255\n",
"return copysign(abs(x / y), x / y)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Condition",
"Return'",
"Condition",
"Return'",
"Return'"
] |
[
"def FUNC_1(self):...\n",
"VAR_6 = self.client.get('/api/apps')\n",
"self.assertEqual(VAR_6.status_code, 200)\n",
"self.assertEqual(len(VAR_6.data['results']), 2)\n",
"VAR_8 = VAR_6.data['results'][0]['id']\n",
"VAR_5 = '/api/apps/{}/perms'.format(VAR_8)\n",
"VAR_7 = {'username': 'autotest-2'}\n",
"VAR_6 = self.client.post(VAR_5, json.dumps(VAR_7), content_type=\n 'application/json')\n",
"self.assertEqual(VAR_6.status_code, 201)\n",
"VAR_6 = self.client.get('/api/apps/{}/perms'.format(VAR_8), content_type=\n 'application/json')\n",
"self.assertEqual(VAR_6.data, {'users': ['autotest-2']})\n"
] | [
"def test_list(self):...\n",
"response = self.client.get('/api/apps')\n",
"self.assertEqual(response.status_code, 200)\n",
"self.assertEqual(len(response.data['results']), 2)\n",
"app_id = response.data['results'][0]['id']\n",
"url = '/api/apps/{}/perms'.format(app_id)\n",
"body = {'username': 'autotest-2'}\n",
"response = self.client.post(url, json.dumps(body), content_type=\n 'application/json')\n",
"self.assertEqual(response.status_code, 201)\n",
"response = self.client.get('/api/apps/{}/perms'.format(app_id),\n content_type='application/json')\n",
"self.assertEqual(response.data, {'users': ['autotest-2']})\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
5,
0
] | [
"FunctionDef'",
"Assign'",
"Expr'",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Expr'"
] |
[
"from flaskext.mysql import MySQL\n",
"def FUNC_0(VAR_0, VAR_1, VAR_2):...\n",
"VAR_3 = VAR_2.connect()\n",
"VAR_4 = VAR_3.cursor()\n",
"VAR_5 = \"SELECT * FROM Users WHERE userEmail = '{0}'\".format(VAR_0)\n",
"VAR_4.execute(VAR_5)\n",
"VAR_6 = VAR_4.fetchone()\n",
"if VAR_6 is not None:\n",
"return 'Fail'\n",
"VAR_7 = \"INSERT INTO Users (userEmail, password) VALUES('{0}', '{1}');\".format(\n VAR_0, VAR_1)\n",
"VAR_4.execute(VAR_7)\n",
"VAR_3.commit()\n",
"return 'Success'\n"
] | [
"from flaskext.mysql import MySQL\n",
"def registerUser(username, password, sqlHandle):...\n",
"conn = sqlHandle.connect()\n",
"cursor = conn.cursor()\n",
"userCheckQuery = \"SELECT * FROM Users WHERE userEmail = '{0}'\".format(username)\n",
"cursor.execute(userCheckQuery)\n",
"result = cursor.fetchone()\n",
"if result is not None:\n",
"return 'Fail'\n",
"queryString = (\"INSERT INTO Users (userEmail, password) VALUES('{0}', '{1}');\"\n .format(username, password))\n",
"cursor.execute(queryString)\n",
"conn.commit()\n",
"return 'Success'\n"
] | [
0,
0,
0,
0,
4,
0,
0,
0,
0,
4,
0,
0,
0
] | [
"ImportFrom'",
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Condition",
"Return'",
"Assign'",
"Expr'",
"Expr'",
"Return'"
] |
[
"@wraps(VAR_55)...\n",
"def FUNC_27(VAR_6):...\n",
"if callable(VAR_6):\n",
"return VAR_6()\n",
"if request.method == 'POST':\n",
"return VAR_55(*VAR_69, **kwargs)\n",
"VAR_10 = json.loads(request.data)\n",
"VAR_54 = {'error': {'type': 'schema', 'message': str(e)}, 'schema': VAR_24}\n",
"schemas.validate(VAR_10, VAR_24)\n",
"return json.dumps(VAR_54, sort_keys=True, indent=4, default=default_encode\n ), 400, {'Content-Type': 'application/json'}\n",
"VAR_70['validated_body'] = VAR_10\n",
"if VAR_70.get('timer'):\n",
"VAR_70['timer'].mark('validate_schema')\n"
] | [
"@wraps(func)...\n",
"def default_encode(value):...\n",
"if callable(value):\n",
"return value()\n",
"if request.method == 'POST':\n",
"return func(*args, **kwargs)\n",
"body = json.loads(request.data)\n",
"result = {'error': {'type': 'schema', 'message': str(e)}, 'schema': schema}\n",
"schemas.validate(body, schema)\n",
"return json.dumps(result, sort_keys=True, indent=4, default=default_encode\n ), 400, {'Content-Type': 'application/json'}\n",
"kwargs['validated_body'] = body\n",
"if kwargs.get('timer'):\n",
"kwargs['timer'].mark('validate_schema')\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Condition",
"FunctionDef'",
"Condition",
"Return'",
"Condition",
"Return'",
"Assign'",
"Assign'",
"Expr'",
"Return'",
"Assign'",
"Condition",
"Expr'"
] |
[
"@memoized_property...\n",
"return self.get_options().native_source_extensions\n"
] | [
"@memoized_property...\n",
"return self.get_options().native_source_extensions\n"
] | [
0,
0
] | [
"Condition",
"Return'"
] |
[
"@jwt_required...\n",
"if VAR_5:\n",
"VAR_7 = Table.filter_by(email=None, VAR_5=user_id)\n",
"VAR_8 = {'results': Table.query(), 'status': 'success'}\n",
"if len(VAR_7) < 1:\n",
"return jsonify(VAR_8), 200\n",
"VAR_8 = {'results': 'User not found', 'status': 'fail'}\n",
"VAR_8 = {'results': VAR_7, 'status': 'success'}\n",
"return make_response(jsonify(VAR_8)), 404\n",
"return jsonify(VAR_8), 200\n"
] | [
"@jwt_required...\n",
"if user_id:\n",
"user = Table.filter_by(email=None, user_id=user_id)\n",
"response_object = {'results': Table.query(), 'status': 'success'}\n",
"if len(user) < 1:\n",
"return jsonify(response_object), 200\n",
"response_object = {'results': 'User not found', 'status': 'fail'}\n",
"response_object = {'results': user, 'status': 'success'}\n",
"return make_response(jsonify(response_object)), 404\n",
"return jsonify(response_object), 200\n"
] | [
0,
0,
4,
4,
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Condition",
"Assign'",
"Assign'",
"Condition",
"Return'",
"Assign'",
"Assign'",
"Return'",
"Return'"
] |
[
"@property...\n",
"\"\"\"docstring\"\"\"\n",
"if self._parsed_content_packages:\n",
"return self._parsed_content_packages\n",
"VAR_6 = self.config.get(self.section, 'content_packages')\n",
"VAR_5 = []\n",
"for VAR_7 in VAR_6.split(','):\n",
"VAR_7 = VAR_7.strip()\n",
"self._parsed_content_packages = VAR_5\n",
"VAR_8, VAR_9, VAR_10 = VAR_7.partition(':')\n",
"return VAR_5\n",
"VAR_5.append((VAR_8, VAR_10))\n"
] | [
"@property...\n",
"\"\"\"docstring\"\"\"\n",
"if self._parsed_content_packages:\n",
"return self._parsed_content_packages\n",
"value = self.config.get(self.section, 'content_packages')\n",
"res = []\n",
"for this in value.split(','):\n",
"this = this.strip()\n",
"self._parsed_content_packages = res\n",
"name, _sep, path = this.partition(':')\n",
"return res\n",
"res.append((name, path))\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Docstring",
"Condition",
"Return'",
"Assign'",
"Assign'",
"For",
"Assign'",
"Assign'",
"Assign'",
"Return'",
"Expr'"
] |
[
"def FUNC_7(*VAR_15, **VAR_16):...\n",
"\"\"\"docstring\"\"\"\n",
"return VAR_0(VAR_1, *VAR_15, **kwargs)\n"
] | [
"def new_object(*args, **kwargs):...\n",
"\"\"\"docstring\"\"\"\n",
"return cls(name, *args, **kwargs)\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Return'"
] |
[
"def FUNC_0(self, VAR_0, VAR_1=False):...\n",
"if not VAR_0.get('database'):\n",
"VAR_0['database'] = '%s.db' % self._model.sqlDatabaseName()\n"
] | [
"def augmentDatabaseArgs(self, args, pool=False):...\n",
"if not args.get('database'):\n",
"args['database'] = '%s.db' % self._model.sqlDatabaseName()\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Condition",
"Assign'"
] |
[
"def FUNC_25(self, VAR_21):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_57 = self.getfile(VAR_21, VAR_23=False)\n",
"if VAR_57 == False:\n",
"if not VAR_57[VAR_6] & FUNC_30.S_IFLNK:\n",
"return VAR_57[VAR_9]\n"
] | [
"def readlink(self, path):...\n",
"\"\"\"docstring\"\"\"\n",
"p = self.getfile(path, follow_symlinks=False)\n",
"if p == False:\n",
"if not p[A_MODE] & stat.S_IFLNK:\n",
"return p[A_TARGET]\n"
] | [
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Condition",
"Condition",
"Return'"
] |
[
"def FUNC_9(self):...\n",
"\"\"\"docstring\"\"\"\n",
"return self.meta.iptc_keys\n"
] | [
"def get_meta_used(self):...\n",
"\"\"\"docstring\"\"\"\n",
"return self.meta.iptc_keys\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Return'"
] |
[
"def FUNC_1(self, VAR_0):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_14 = subprocess.Popen(VAR_0, VAR_1=subprocess.PIPE, VAR_2=subprocess.\n PIPE, shell=True)\n",
"VAR_4, VAR_15 = VAR_14.communicate()\n",
"VAR_16 = VAR_14.returncode\n",
"logging.debug('cmd: %s, stdout: %s, stderr: %s, ret: %s', VAR_0, VAR_4,\n VAR_15, VAR_16)\n",
"if VAR_16 == 0:\n",
"return VAR_4\n"
] | [
"def _exec_cmd(self, cmd):...\n",
"\"\"\"docstring\"\"\"\n",
"proc = subprocess.Popen(cmd, stdout=subprocess.PIPE, stderr=subprocess.PIPE,\n shell=True)\n",
"out, err = proc.communicate()\n",
"ret = proc.returncode\n",
"logging.debug('cmd: %s, stdout: %s, stderr: %s, ret: %s', cmd, out, err, ret)\n",
"if ret == 0:\n",
"return out\n"
] | [
0,
0,
2,
0,
0,
2,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Condition",
"Return'"
] |
[
"def FUNC_38(self):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_2 = {'SEVERITY': {'HIGH': 3, 'MEDIUM': 1, 'LOW': 14}, 'CONFIDENCE': {\n 'HIGH': 17, 'LOW': 1}}\n",
"self.check_example('subprocess_shell.py', VAR_2)\n"
] | [
"def test_subprocess_shell(self):...\n",
"\"\"\"docstring\"\"\"\n",
"expect = {'SEVERITY': {'HIGH': 3, 'MEDIUM': 1, 'LOW': 14}, 'CONFIDENCE': {\n 'HIGH': 17, 'LOW': 1}}\n",
"self.check_example('subprocess_shell.py', expect)\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Expr'"
] |
[
"def FUNC_2(self):...\n",
"return self.config['runtime.gallery_dir']\n"
] | [
"def get_base_path(self):...\n",
"return self.config['runtime.gallery_dir']\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
[
"@validate(VUser(), VAR_0=VEditMeetup('id'))...\n",
"return BoringPage(pagename='Edit Meetup', VAR_27=EditMeetup(meetup, title=\n meetup.title, description=meetup.description, location=meetup.location,\n latitude=meetup.latitude, longitude=meetup.longitude, timestamp=int(\n meetup.timestamp * 1000), tzoffset=meetup.tzoffset)).render()\n"
] | [
"@validate(VUser(), meetup=VEditMeetup('id'))...\n",
"return BoringPage(pagename='Edit Meetup', content=EditMeetup(meetup, title=\n meetup.title, description=meetup.description, location=meetup.location,\n latitude=meetup.latitude, longitude=meetup.longitude, timestamp=int(\n meetup.timestamp * 1000), tzoffset=meetup.tzoffset)).render()\n"
] | [
0,
0
] | [
"Condition",
"Return'"
] |
[
"@utils.synchronized('3par', external=True)...\n",
"\"\"\"docstring\"\"\"\n",
"self.common.client_login()\n",
"VAR_18 = self.common.create_volume_from_snapshot(VAR_6, VAR_8)\n",
"self.common.client_logout()\n",
"return {'metadata': VAR_18}\n"
] | [
"@utils.synchronized('3par', external=True)...\n",
"\"\"\"docstring\"\"\"\n",
"self.common.client_login()\n",
"metadata = self.common.create_volume_from_snapshot(volume, snapshot)\n",
"self.common.client_logout()\n",
"return {'metadata': metadata}\n"
] | [
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Docstring",
"Expr'",
"Assign'",
"Expr'",
"Return'"
] |
[
"import os.path\n",
"import re\n",
"import motor.motor_tornado\n",
"from argon2 import PasswordHasher\n",
"from pymongo import MongoClient\n",
"import random\n",
"import tornado.httpserver\n",
"import tornado.ioloop\n",
"import tornado.options\n",
"import tornado.web\n",
"import pymongo\n",
"from tornado.options import define, options\n",
"define('port', default=8100, help='run on the given port', type=int)\n",
"\"\"\"string\"\"\"\n",
"def FUNC_0(self):...\n",
"return self.get_secure_cookie('user')\n"
] | [
"import os.path\n",
"import re\n",
"import motor.motor_tornado\n",
"from argon2 import PasswordHasher\n",
"from pymongo import MongoClient\n",
"import random\n",
"import tornado.httpserver\n",
"import tornado.ioloop\n",
"import tornado.options\n",
"import tornado.web\n",
"import pymongo\n",
"from tornado.options import define, options\n",
"define('port', default=8100, help='run on the given port', type=int)\n",
"\"\"\" BaseHandler():\nClass that'll be used later when @tornado.web.authenticated is needed for POST requests.\n\"\"\"\n",
"def get_current_user(self):...\n",
"return self.get_secure_cookie('user')\n"
] | [
0,
0,
0,
6,
0,
0,
0,
0,
0,
0,
0,
0,
0,
6,
0,
0
] | [
"Import'",
"Import'",
"Import'",
"ImportFrom'",
"ImportFrom'",
"Import'",
"Import'",
"Import'",
"Import'",
"Import'",
"Import'",
"ImportFrom'",
"Expr'",
"Expr'",
"FunctionDef'",
"Return'"
] |
[
"def FUNC_13(VAR_10):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_41 = str(VAR_10) + SECRET_KEY\n",
"VAR_35 = hashlib.sha256()\n",
"VAR_35.update(VAR_41.encode('utf-8'))\n",
"VAR_35 = VAR_35.digest()\n",
"VAR_42 = base64.b64encode(VAR_35).decode('utf-8')\n",
"return VAR_42\n"
] | [
"def csrf_hash(csrf):...\n",
"\"\"\"docstring\"\"\"\n",
"enc = str(csrf) + SECRET_KEY\n",
"m = hashlib.sha256()\n",
"m.update(enc.encode('utf-8'))\n",
"m = m.digest()\n",
"encrypted_csrf = base64.b64encode(m).decode('utf-8')\n",
"return encrypted_csrf\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Assign'",
"Return'"
] |
[
"def FUNC_40(self):...\n",
"self.compilation_ko(\"\"\"\nif (true) {\n\n}\n\"\"\")\n"
] | [
"def test_misplaced_parenthesis(self):...\n",
"self.compilation_ko(\"\"\"\nif (true) {\n\n}\n\"\"\")\n"
] | [
0,
0
] | [
"FunctionDef'",
"Expr'"
] |
[
"@staticmethod...\n",
"return mongo.db.analysis.find_one({'info.id': int(VAR_1)}, sort=[('_id',\n pymongo.DESCENDING)])\n"
] | [
"@staticmethod...\n",
"return mongo.db.analysis.find_one({'info.id': int(task_id)}, sort=[('_id',\n pymongo.DESCENDING)])\n"
] | [
0,
0
] | [
"Condition",
"Return'"
] |
[
"def FUNC_5(self, VAR_8, VAR_6):...\n",
"self.command(VAR_8, VAR_6, 'Error halting DUT', False)\n",
"db.log_event('Information', 'Debugger', 'Halt DUT')\n"
] | [
"def halt_dut(self, halt_command, expected_output):...\n",
"self.command(halt_command, expected_output, 'Error halting DUT', False)\n",
"db.log_event('Information', 'Debugger', 'Halt DUT')\n"
] | [
0,
0,
2
] | [
"FunctionDef'",
"Expr'",
"Expr'"
] |
[
"@VAR_1.route('/unique_ingredients/<cuisine_id>')...\n",
"VAR_0.info('GET find_unique_ingredients_from_cuisine query')\n",
"if VAR_9 in VAR_5:\n",
"VAR_29, VAR_30 = VAR_5[VAR_9]\n",
"VAR_33 = int(VAR_9)\n",
"VAR_0.error(\n 'Error translating cuisine_id to int in find_unique_ingredients_from_cuisine, passed value: %s'\n % VAR_9)\n",
"VAR_18 = VAR_3.find_unique_ingredients_of_cuisine(VAR_33, 500)\n",
"if datetime.now() < VAR_29 + VAR_6:\n",
"return None\n",
"if VAR_18 == -1:\n",
"return VAR_30\n",
"return None\n",
"if len(simplejson.loads(VAR_18)) == 0:\n",
"VAR_18 = VAR_3.find_unique_ingredients_of_cuisine(VAR_33, 250)\n",
"VAR_5[VAR_9] = datetime.now(), VAR_18\n",
"if VAR_18 == -1:\n",
"return VAR_18\n",
"return None\n",
"VAR_5[VAR_9] = datetime.now(), VAR_18\n",
"return VAR_18\n"
] | [
"@app.route('/unique_ingredients/<cuisine_id>')...\n",
"logger.info('GET find_unique_ingredients_from_cuisine query')\n",
"if cuisine_id in unique_ingredients_cache:\n",
"insert_time, data = unique_ingredients_cache[cuisine_id]\n",
"cuisine_id_int = int(cuisine_id)\n",
"logger.error(\n 'Error translating cuisine_id to int in find_unique_ingredients_from_cuisine, passed value: %s'\n % cuisine_id)\n",
"query_res = database.find_unique_ingredients_of_cuisine(cuisine_id_int, 500)\n",
"if datetime.now() < insert_time + cache_persistence_time:\n",
"return None\n",
"if query_res == -1:\n",
"return data\n",
"return None\n",
"if len(simplejson.loads(query_res)) == 0:\n",
"query_res = database.find_unique_ingredients_of_cuisine(cuisine_id_int, 250)\n",
"unique_ingredients_cache[cuisine_id] = datetime.now(), query_res\n",
"if query_res == -1:\n",
"return query_res\n",
"return None\n",
"unique_ingredients_cache[cuisine_id] = datetime.now(), query_res\n",
"return query_res\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Expr'",
"Condition",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Condition",
"Return'",
"Condition",
"Return'",
"Return'",
"Condition",
"Assign'",
"Assign'",
"Condition",
"Return'",
"Return'",
"Assign'",
"Return'"
] |
[
"def FUNC_37(self):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_2 = {'SEVERITY': {'LOW': 1, 'MEDIUM': 10, 'HIGH': 7}, 'CONFIDENCE': {\n 'LOW': 0, 'MEDIUM': 11, 'HIGH': 7}}\n",
"self.check_example('ssl-insecure-version.py', VAR_2)\n"
] | [
"def test_ssl_insecure_version(self):...\n",
"\"\"\"docstring\"\"\"\n",
"expect = {'SEVERITY': {'LOW': 1, 'MEDIUM': 10, 'HIGH': 7}, 'CONFIDENCE': {\n 'LOW': 0, 'MEDIUM': 11, 'HIGH': 7}}\n",
"self.check_example('ssl-insecure-version.py', expect)\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Expr'"
] |
[
"def FUNC_0(self, VAR_3):...\n",
"VAR_4 = super().value_from_object(VAR_3)\n",
"return self.to_python(VAR_4)\n"
] | [
"def value_from_object(self, obj):...\n",
"value = super().value_from_object(obj)\n",
"return self.to_python(value)\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Return'"
] |
[
"def FUNC_1(VAR_14, VAR_2, VAR_3='text', VAR_4=bconfig....\n",
"\"\"\"docstring\"\"\"\n",
"VAR_0.info('Analyzing keywords for local file %s.' % VAR_14)\n",
"VAR_15 = extractor.text_lines_from_local_file(VAR_14)\n",
"return FUNC_2(VAR_15, VAR_2, VAR_3=output_mode, VAR_4=output_limit, VAR_5=\n spires, VAR_6=match_mode, VAR_7=no_cache, VAR_8=with_author_keywords,\n VAR_9=rebuild_cache, VAR_10=only_core_tags, VAR_11=extract_acronyms)\n"
] | [
"def get_keywords_from_local_file(local_file, taxonomy_name, output_mode=...\n",
"\"\"\"docstring\"\"\"\n",
"log.info('Analyzing keywords for local file %s.' % local_file)\n",
"text_lines = extractor.text_lines_from_local_file(local_file)\n",
"return get_keywords_from_text(text_lines, taxonomy_name, output_mode=\n output_mode, output_limit=output_limit, spires=spires, match_mode=\n match_mode, no_cache=no_cache, with_author_keywords=\n with_author_keywords, rebuild_cache=rebuild_cache, only_core_tags=\n only_core_tags, extract_acronyms=extract_acronyms)\n"
] | [
0,
0,
0,
0,
0
] | [
"Condition",
"Docstring",
"Expr'",
"Assign'",
"Return'"
] |
[
"def FUNC_10(self):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_8 = time.time()\n",
"tournament.check_version((2, 4))\n",
"VAR_9 = time.time()\n",
"VAR_10 = round(VAR_9 - VAR_8, 1)\n",
"self.assertEqual(VAR_10, 3.0)\n"
] | [
"def test_wait_time(self):...\n",
"\"\"\"docstring\"\"\"\n",
"start = time.time()\n",
"tournament.check_version((2, 4))\n",
"end = time.time()\n",
"count = round(end - start, 1)\n",
"self.assertEqual(count, 3.0)\n"
] | [
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Expr'",
"Assign'",
"Assign'",
"Expr'"
] |
[
"def FUNC_1(VAR_4):...\n",
"\"\"\"docstring\"\"\"\n",
"return VAR_4.__dict__\n"
] | [
"def serialize_to_json(object):...\n",
"\"\"\"docstring\"\"\"\n",
"return object.__dict__\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Return'"
] |
[
"@web.authenticated...\n",
"self.get(VAR_1, VAR_2=False)\n"
] | [
"@web.authenticated...\n",
"self.get(path, include_body=False)\n"
] | [
0,
5
] | [
"Condition",
"Expr'"
] |
[
"def FUNC_16(VAR_16):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_30 = VAR_19(VAR_16, 'dynamic')\n",
"VAR_31 = [VAR_30] if FUNC_10(VAR_30) else VAR_30\n",
"for VAR_3 in VAR_31:\n",
"VAR_52 = list(VAR_5.finditer(VAR_3))\n",
"return VAR_30\n",
"for VAR_29 in VAR_52:\n",
"if VAR_29.group('constraint'):\n"
] | [
"def dynamic(value):...\n",
"\"\"\"docstring\"\"\"\n",
"annotated = flag(value, 'dynamic')\n",
"tocheck = [annotated] if not_iterable(annotated) else annotated\n",
"for file in tocheck:\n",
"matches = list(_wildcard_regex.finditer(file))\n",
"return annotated\n",
"for match in matches:\n",
"if match.group('constraint'):\n"
] | [
0,
0,
7,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Assign'",
"For",
"Assign'",
"Return'",
"For",
"Condition"
] |
[
"def FUNC_4(VAR_0):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_28 = CLASS_1(VAR_0)\n",
"VAR_29, VAR_30 = VAR_28.get()\n",
"VAR_31 = VAR_0.GET.get('tags')\n",
"VAR_32 = VAR_0.GET.get('a')\n",
"if VAR_32:\n",
"VAR_75 = CLASS_2(VAR_30=obj, VAR_41=q_tag)\n",
"VAR_33 = VAR_30.tag.all().order_by('pk')\n",
"getattr(VAR_75, VAR_32)()\n",
"VAR_34 = TestPlanTag.objects.filter(tag__in=all_tags).values('tag').annotate(\n num_plans=Count('tag')).order_by('tag')\n",
"VAR_35 = TestCaseTag.objects.filter(tag__in=all_tags).values('tag').annotate(\n num_cases=Count('tag')).order_by('tag')\n",
"VAR_36 = TestRunTag.objects.filter(tag__in=all_tags).values('tag').annotate(\n num_runs=Count('tag')).order_by('tag')\n",
"VAR_37 = CLASS_3('num_plans', VAR_34)\n",
"VAR_38 = CLASS_3('num_cases', VAR_35)\n",
"VAR_39 = CLASS_3('num_runs', VAR_36)\n",
"for VAR_44 in VAR_33:\n",
"VAR_44.num_plans = VAR_37.calculate_tag_count(VAR_44)\n",
"VAR_40 = {'tags': VAR_33, 'object': VAR_30}\n",
"VAR_44.num_cases = VAR_38.calculate_tag_count(VAR_44)\n",
"return render(VAR_0, VAR_29, VAR_40)\n",
"VAR_44.num_runs = VAR_39.calculate_tag_count(VAR_44)\n"
] | [
"def tags(request):...\n",
"\"\"\"docstring\"\"\"\n",
"tag_objects = _TagObjects(request)\n",
"template_name, obj = tag_objects.get()\n",
"q_tag = request.GET.get('tags')\n",
"q_action = request.GET.get('a')\n",
"if q_action:\n",
"tag_actions = _TagActions(obj=obj, tag_name=q_tag)\n",
"all_tags = obj.tag.all().order_by('pk')\n",
"getattr(tag_actions, q_action)()\n",
"test_plan_tags = TestPlanTag.objects.filter(tag__in=all_tags).values('tag'\n ).annotate(num_plans=Count('tag')).order_by('tag')\n",
"test_case_tags = TestCaseTag.objects.filter(tag__in=all_tags).values('tag'\n ).annotate(num_cases=Count('tag')).order_by('tag')\n",
"test_run_tags = TestRunTag.objects.filter(tag__in=all_tags).values('tag'\n ).annotate(num_runs=Count('tag')).order_by('tag')\n",
"plan_counter = _TagCounter('num_plans', test_plan_tags)\n",
"case_counter = _TagCounter('num_cases', test_case_tags)\n",
"run_counter = _TagCounter('num_runs', test_run_tags)\n",
"for tag in all_tags:\n",
"tag.num_plans = plan_counter.calculate_tag_count(tag)\n",
"context_data = {'tags': all_tags, 'object': obj}\n",
"tag.num_cases = case_counter.calculate_tag_count(tag)\n",
"return render(request, template_name, context_data)\n",
"tag.num_runs = run_counter.calculate_tag_count(tag)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"For",
"Assign'",
"Assign'",
"Assign'",
"Return'",
"Assign'"
] |
[
"def FUNC_5(self):...\n",
"\"\"\"docstring\"\"\"\n",
"self.read = True\n",
"db.session.commit()\n"
] | [
"def mark_as_read(self):...\n",
"\"\"\"docstring\"\"\"\n",
"self.read = True\n",
"db.session.commit()\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Expr'"
] |
[
"def FUNC_10(self):...\n",
"if self._hosts_cache is None:\n",
"self._hosts_cache = self._get_hosts()\n",
"return self._hosts_cache\n"
] | [
"def get_hosts(self):...\n",
"if self._hosts_cache is None:\n",
"self._hosts_cache = self._get_hosts()\n",
"return self._hosts_cache\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"Condition",
"Assign'",
"Return'"
] |
[
"def FUNC_3(self):...\n",
"VAR_15 = []\n",
"VAR_14 = datetime.now(timezone(timedelta(hours=+9), 'JST'))\n",
"VAR_14 = VAR_14.replace(hour=0, minute=0, second=0, microsecond=0)\n",
"for day in range(30)[::-1]:\n",
"VAR_25 = VAR_14 - timedelta(days=day)\n",
"return VAR_15\n",
"VAR_26 = VAR_25.strftime('%Y-%m-%d')\n",
"VAR_15.append(VAR_26)\n"
] | [
"def thirty_day_labels(self):...\n",
"labels = []\n",
"today = datetime.now(timezone(timedelta(hours=+9), 'JST'))\n",
"today = today.replace(hour=0, minute=0, second=0, microsecond=0)\n",
"for day in range(30)[::-1]:\n",
"date = today - timedelta(days=day)\n",
"return labels\n",
"label = date.strftime('%Y-%m-%d')\n",
"labels.append(label)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'",
"For",
"Assign'",
"Return'",
"Assign'",
"Expr'"
] |
[
"def FUNC_30(self):...\n",
"VAR_44 = Tag.objects.get(name=self.tag_name)\n",
"self.obj.remove_tag(VAR_44)\n"
] | [
"def remove(self):...\n",
"tag = Tag.objects.get(name=self.tag_name)\n",
"self.obj.remove_tag(tag)\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Expr'"
] |
[
"def FUNC_8(self, VAR_14):...\n",
""
] | [
"def is_file(self, col_name):...\n",
""
] | [
0,
0
] | [
"FunctionDef'",
"Condition"
] |
[
"def FUNC_1(self, VAR_5, VAR_6):...\n",
"\"\"\"docstring\"\"\"\n",
"if VAR_5 not in ('strings', 'files'):\n",
"VAR_0.error(\"Bad parameter '%s' for submit_type\", VAR_5)\n",
"VAR_13 = Folders.create_temp()\n",
"return False\n",
"VAR_14 = {'data': [], 'errors': []}\n",
"if VAR_5 == 'strings':\n",
"for VAR_4 in VAR_6:\n",
"if VAR_5 == 'files':\n",
"self._handle_string(VAR_14, VAR_13, VAR_4)\n",
"for entry in VAR_6:\n",
"return Database().add_submit(VAR_13, VAR_5, VAR_14)\n",
"VAR_22 = Storage.get_filename_from_path(entry['name'])\n",
"VAR_18 = Files.create(VAR_13, VAR_22, entry['data'])\n",
"VAR_14['data'].append({'type': 'file', 'data': VAR_18})\n"
] | [
"def pre(self, submit_type, data):...\n",
"\"\"\"docstring\"\"\"\n",
"if submit_type not in ('strings', 'files'):\n",
"log.error(\"Bad parameter '%s' for submit_type\", submit_type)\n",
"path_tmp = Folders.create_temp()\n",
"return False\n",
"submit_data = {'data': [], 'errors': []}\n",
"if submit_type == 'strings':\n",
"for line in data:\n",
"if submit_type == 'files':\n",
"self._handle_string(submit_data, path_tmp, line)\n",
"for entry in data:\n",
"return Database().add_submit(path_tmp, submit_type, submit_data)\n",
"filename = Storage.get_filename_from_path(entry['name'])\n",
"filepath = Files.create(path_tmp, filename, entry['data'])\n",
"submit_data['data'].append({'type': 'file', 'data': filepath})\n"
] | [
0,
0,
0,
0,
0,
0,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Condition",
"Expr'",
"Assign'",
"Return'",
"Assign'",
"Condition",
"For",
"Condition",
"Expr'",
"For",
"Return'",
"Assign'",
"Assign'",
"Expr'"
] |
[
"def FUNC_32(self, VAR_32, VAR_29):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_24 = self.getfile(VAR_32)\n",
"if VAR_24 == False:\n",
"return\n",
"if VAR_24[VAR_2] != VAR_13:\n",
"return\n",
"VAR_24[VAR_5] = VAR_29\n"
] | [
"def update_size(self, filename, size):...\n",
"\"\"\"docstring\"\"\"\n",
"f = self.getfile(filename)\n",
"if f == False:\n",
"return\n",
"if f[A_TYPE] != T_FILE:\n",
"return\n",
"f[A_SIZE] = size\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Condition",
"Return'",
"Condition",
"Return'",
"Assign'"
] |
[
"def FUNC_28(self):...\n",
"self.compilation_ko('require [];')\n"
] | [
"def test_empty_string_list(self):...\n",
"self.compilation_ko('require [];')\n"
] | [
0,
0
] | [
"FunctionDef'",
"Expr'"
] |
[
"@validate(VUser(), VAR_0=VEditMeetup('id'))...\n",
"return BoringPage(pagename='Edit Meetup', VAR_27=EditMeetup(meetup, title=\n meetup.title, description=meetup.description, location=meetup.location,\n latitude=meetup.latitude, longitude=meetup.longitude, timestamp=int(\n meetup.timestamp * 1000), tzoffset=meetup.tzoffset)).render()\n"
] | [
"@validate(VUser(), meetup=VEditMeetup('id'))...\n",
"return BoringPage(pagename='Edit Meetup', content=EditMeetup(meetup, title=\n meetup.title, description=meetup.description, location=meetup.location,\n latitude=meetup.latitude, longitude=meetup.longitude, timestamp=int(\n meetup.timestamp * 1000), tzoffset=meetup.tzoffset)).render()\n"
] | [
0,
0
] | [
"Condition",
"Return'"
] |
[
"def FUNC_6(self, VAR_23):...\n",
"VAR_82 = CLASS_46.run(self, VAR_23)\n",
"if not VAR_82 or not hasattr(VAR_82, 'email') or not VAR_82.email:\n",
"return self.error(errors.NO_EMAIL_FOR_USER)\n",
"return VAR_82\n"
] | [
"def run(self, name):...\n",
"user = VExistingUname.run(self, name)\n",
"if not user or not hasattr(user, 'email') or not user.email:\n",
"return self.error(errors.NO_EMAIL_FOR_USER)\n",
"return user\n"
] | [
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Condition",
"Return'",
"Return'"
] |
[
"def FUNC_0(self):...\n",
"self.assertTrue(self.client.login(username='autotest', password='password'))\n"
] | [
"def setUp(self):...\n",
"self.assertTrue(self.client.login(username='autotest', password='password'))\n"
] | [
0,
5
] | [
"FunctionDef'",
"Expr'"
] |