lines
sequencelengths
1
444
raw_lines
sequencelengths
1
444
label
sequencelengths
1
444
type
sequencelengths
1
444
[ "def FUNC_12():...\n", "VAR_20 = []\n", "for i in range(int(VAR_2)):\n", "VAR_26 = subprocess.Popen([sys.executable, __file__], stdout=subprocess.\n PIPE, stderr=subprocess.PIPE)\n", "VAR_20.append(VAR_26.pid)\n", "log.info('Started worker with pid %s' % VAR_26.pid)\n" ]
[ "def run_children():...\n", "child_pids = []\n", "for i in range(int(no_subprocess)):\n", "proc = subprocess.Popen([sys.executable, __file__], stdout=subprocess.PIPE,\n stderr=subprocess.PIPE)\n", "child_pids.append(proc.pid)\n", "log.info('Started worker with pid %s' % proc.pid)\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "For", "Assign'", "Expr'", "Expr'" ]
[ "@VAR_0.route('/tutorial')...\n", "return ''\n" ]
[ "@app.route('/tutorial')...\n", "return ''\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "def FUNC_26(self, VAR_37):...\n", "while self.running.is_set():\n", "self.log.exception('Exception \"%s\" raised on create_spawn', e)\n", "VAR_58 = VAR_37.pop()\n", "return\n", "self.proxylist.add(VAR_58)\n", "for spawn in FUNC_7(VAR_58[0], VAR_58[1], self.pc, self.get_userqueue):\n", "self.log.info('Created spawn %s', spawn.name)\n", "self.spawnqueue.put(spawn, False)\n" ]
[ "def add_spawns(self, proxypairs):...\n", "while self.running.is_set():\n", "self.log.exception('Exception \"%s\" raised on create_spawn', e)\n", "proxypair = proxypairs.pop()\n", "return\n", "self.proxylist.add(proxypair)\n", "for spawn in create_spawn(proxypair[0], proxypair[1], self.pc, self.\n", "self.log.info('Created spawn %s', spawn.name)\n", "self.spawnqueue.put(spawn, False)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Expr'", "Assign'", "Return'", "Expr'", "For", "Expr'", "Expr'" ]
[ "def __init__(self, VAR_7, *VAR_15, **VAR_16):...\n", "CLASS_0.__init__(self, VAR_7, *VAR_15, **kw)\n" ]
[ "def __init__(self, param, *a, **kw):...\n", "Validator.__init__(self, param, *a, **kw)\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Expr'" ]
[ "def FUNC_24(self, VAR_21):...\n", "\"\"\"docstring\"\"\"\n", "return VAR_21 in self.sos_info['presets']\n" ]
[ "def _preset_exists(self, preset):...\n", "\"\"\"docstring\"\"\"\n", "return preset in self.sos_info['presets']\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Return'" ]
[ "def FUNC_7(self):...\n", "VAR_17 = shellutil.run_get_output('/sbin/pidof dhclient')\n", "return VAR_17[1] if VAR_17[0] == 0 else None\n" ]
[ "def get_dhcp_pid(self):...\n", "ret = shellutil.run_get_output('/sbin/pidof dhclient')\n", "return ret[1] if ret[0] == 0 else None\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Return'" ]
[ "def FUNC_39(VAR_62):...\n", "return json.loads(VAR_62) if VAR_21 else VAR_62\n" ]
[ "def get_str(_str):...\n", "return json.loads(_str) if as_json else _str\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_3(VAR_6, VAR_3, VAR_4, VAR_5=None):...\n", "\"\"\"docstring\"\"\"\n", "if VAR_5 is None:\n", "VAR_5 = ray.DriverID.nil()\n", "VAR_35 = ray.gcs_utils.construct_error_message(VAR_5, VAR_3, VAR_4, time.time()\n )\n", "VAR_6.execute_command('RAY.TABLE_APPEND', ray.gcs_utils.TablePrefix.\n ERROR_INFO, ray.gcs_utils.TablePubsub.ERROR_INFO, VAR_5.binary(), VAR_35)\n" ]
[ "def push_error_to_driver_through_redis(redis_client, error_type, message,...\n", "\"\"\"docstring\"\"\"\n", "if driver_id is None:\n", "driver_id = ray.DriverID.nil()\n", "error_data = ray.gcs_utils.construct_error_message(driver_id, error_type,\n message, time.time())\n", "redis_client.execute_command('RAY.TABLE_APPEND', ray.gcs_utils.TablePrefix.\n ERROR_INFO, ray.gcs_utils.TablePubsub.ERROR_INFO, driver_id.binary(),\n error_data)\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Assign'", "Assign'", "Expr'" ]
[ "def FUNC_1():...\n", "\"\"\"docstring\"\"\"\n", "VAR_3 = 'testing inband sql injection on parameter '\n", "VAR_3 += \"'%s'\" % kb.injParameter\n", "logger.info(VAR_3)\n", "VAR_4 = ''\n", "VAR_0 = agent.prefixQuery(' UNION ALL SELECT NULL')\n", "for VAR_1 in (queries[kb.dbms].comment, ''):\n", "VAR_4 = FUNC_0(VAR_0, VAR_1)\n", "if kb.unionCount:\n", "if VAR_4:\n", "VAR_3 = 'the target url could be affected by an '\n", "VAR_8 = 'the target url is not affected by an '\n", "setUnion(VAR_1, VAR_4.count('NULL'))\n", "VAR_3 += 'inband sql injection vulnerability'\n", "VAR_8 += 'inband sql injection vulnerability'\n", "logger.info(VAR_3)\n", "logger.warn(VAR_8)\n", "return VAR_4\n" ]
[ "def unionTest():...\n", "\"\"\"docstring\"\"\"\n", "logMsg = 'testing inband sql injection on parameter '\n", "logMsg += \"'%s'\" % kb.injParameter\n", "logger.info(logMsg)\n", "value = ''\n", "query = agent.prefixQuery(' UNION ALL SELECT NULL')\n", "for comment in (queries[kb.dbms].comment, ''):\n", "value = __effectiveUnionTest(query, comment)\n", "if kb.unionCount:\n", "if value:\n", "logMsg = 'the target url could be affected by an '\n", "warnMsg = 'the target url is not affected by an '\n", "setUnion(comment, value.count('NULL'))\n", "logMsg += 'inband sql injection vulnerability'\n", "warnMsg += 'inband sql injection vulnerability'\n", "logger.info(logMsg)\n", "logger.warn(warnMsg)\n", "return value\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "AugAssign'", "Expr'", "Assign'", "Assign'", "For", "Assign'", "Condition", "Condition", "Assign'", "Assign'", "Expr'", "AugAssign'", "AugAssign'", "Expr'", "Expr'", "Return'" ]
[ "def FUNC_2(self, VAR_1):...\n", "VAR_2 = self.connect()\n", "VAR_3 = \"INSERT INTO crimes (description) VALUES ('{}');\".format(VAR_1)\n", "VAR_2.close()\n", "cursor.execute(VAR_3)\n", "VAR_2.commit()\n" ]
[ "def add_input(self, data):...\n", "connection = self.connect()\n", "query = \"INSERT INTO crimes (description) VALUES ('{}');\".format(data)\n", "connection.close()\n", "cursor.execute(query)\n", "connection.commit()\n" ]
[ 0, 0, 4, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'" ]
[ "\"\"\"API over the nova service.\n\"\"\"\n", "from collections import OrderedDict\n", "from django.http import HttpResponse\n", "from django.template.defaultfilters import slugify\n", "from django.utils import http as utils_http\n", "from django.utils.translation import ugettext_lazy as _\n", "from django.views import generic\n", "from horizon import exceptions as hz_exceptions\n", "from novaclient import exceptions\n", "from openstack_dashboard import api\n", "from openstack_dashboard.api.rest import json_encoder\n", "from openstack_dashboard.api.rest import urls\n", "from openstack_dashboard.api.rest import utils as rest_utils\n", "from openstack_dashboard.usage import quotas\n", "\"\"\"API for nova snapshots.\n \"\"\"\n", "VAR_0 = 'nova/snapshots/$'\n", "@rest_utils.ajax(data_required=True)...\n", "VAR_9 = VAR_1.DATA['instance_id']\n", "VAR_6 = VAR_1.DATA['name']\n", "VAR_10 = api.nova.snapshot_create(VAR_1, VAR_9=instance_id, VAR_6=name)\n", "return VAR_10\n" ]
[ "\"\"\"API over the nova service.\n\"\"\"\n", "from collections import OrderedDict\n", "from django.http import HttpResponse\n", "from django.template.defaultfilters import slugify\n", "from django.utils import http as utils_http\n", "from django.utils.translation import ugettext_lazy as _\n", "from django.views import generic\n", "from horizon import exceptions as hz_exceptions\n", "from novaclient import exceptions\n", "from openstack_dashboard import api\n", "from openstack_dashboard.api.rest import json_encoder\n", "from openstack_dashboard.api.rest import urls\n", "from openstack_dashboard.api.rest import utils as rest_utils\n", "from openstack_dashboard.usage import quotas\n", "\"\"\"API for nova snapshots.\n \"\"\"\n", "url_regex = 'nova/snapshots/$'\n", "@rest_utils.ajax(data_required=True)...\n", "instance_id = request.DATA['instance_id']\n", "name = request.DATA['name']\n", "result = api.nova.snapshot_create(request, instance_id=instance_id, name=name)\n", "return result\n" ]
[ 5, 0, 5, 5, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Expr'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Expr'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_4(VAR_6):...\n", "VAR_8 = VAR_6.submission.exercise if VAR_6.submission else None\n", "return {'id': VAR_6.id, 'submission_id': VAR_6.submission.id if VAR_6.\n submission else 0, 'name': '{} {}, {}'.format(VAR_6.course_instance.\n course.code, str(VAR_8.parent) if VAR_8 and VAR_8.parent else VAR_6.\n course_instance.instance_name, str(VAR_8) if VAR_8 else VAR_6.subject),\n 'link': VAR_6.get_display_url()}\n" ]
[ "def notification_entry(n):...\n", "exercise = n.submission.exercise if n.submission else None\n", "return {'id': n.id, 'submission_id': n.submission.id if n.submission else 0,\n 'name': '{} {}, {}'.format(n.course_instance.course.code, str(exercise.\n parent) if exercise and exercise.parent else n.course_instance.\n instance_name, str(exercise) if exercise else n.subject), 'link': n.\n get_display_url()}\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Return'" ]
[ "def FUNC_7(self, VAR_11):...\n", "\"\"\"docstring\"\"\"\n" ]
[ "def create_snapshot(self, snapshot):...\n", "\"\"\"docstring\"\"\"\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Docstring" ]
[ "def FUNC_24(self, VAR_2, VAR_3, VAR_4, *VAR_6):...\n", "VAR_47 = self.browse(VAR_2, VAR_3, VAR_4)\n", "for t in VAR_47:\n", "self.write(VAR_2, VAR_3, [t.id], {'state': 'open'})\n", "return True\n" ]
[ "def do_open(self, cr, uid, ids, *args):...\n", "tasks = self.browse(cr, uid, ids)\n", "for t in tasks:\n", "self.write(cr, uid, [t.id], {'state': 'open'})\n", "return True\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "For", "Expr'", "Return'" ]
[ "def FUNC_17(self):...\n", "\"\"\"docstring\"\"\"\n", "db.session.delete(self)\n", "db.session.commit()\n" ]
[ "def delete(self):...\n", "\"\"\"docstring\"\"\"\n", "db.session.delete(self)\n", "db.session.commit()\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Expr'", "Expr'" ]
[ "def __init__(self, VAR_7, VAR_20=True, *VAR_15, **VAR_16):...\n", "CLASS_5.__init__(self, VAR_7, *VAR_15, VAR_20=redirect, **kw)\n" ]
[ "def __init__(self, param, redirect=True, *a, **kw):...\n", "VLink.__init__(self, param, *a, redirect=redirect, **kw)\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Expr'" ]
[ "def FUNC_6(self, VAR_14):...\n", "self.icinga_lines.append(VAR_14)\n" ]
[ "def write_line(self, line):...\n", "self.icinga_lines.append(line)\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Expr'" ]
[ "def FUNC_9():...\n", "\"\"\"docstring\"\"\"\n", "VAR_35 = FUNC_8('tasks')\n", "VAR_36 = re.findall('string', VAR_35.text, re.X)\n", "VAR_37 = {}\n", "for VAR_46 in VAR_36:\n", "VAR_57, VAR_58, VAR_59 = VAR_46\n", "return VAR_37\n", "VAR_57 = int(VAR_57)\n", "VAR_37[VAR_58] = {'title': VAR_59, 'id': VAR_57}\n" ]
[ "def get_tasks():...\n", "\"\"\"docstring\"\"\"\n", "r = admin_req('tasks')\n", "groups = re.findall(\n \"\"\"\n <tr>\\\\s*\n <td><a\\\\s+href=\"./task/(\\\\d+)\">(.*)</a></td>\\\\s*\n <td>(.*)</td>\\\\s*\n \"\"\"\n , r.text, re.X)\n", "tasks = {}\n", "for g in groups:\n", "id, name, title = g\n", "return tasks\n", "id = int(id)\n", "tasks[name] = {'title': title, 'id': id}\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Assign'", "For", "Assign'", "Return'", "Assign'", "Assign'" ]
[ "def __init__(self, VAR_1, VAR_11):...\n", "self.name = VAR_1\n", "self.shots = VAR_11\n", "self._out = None\n", "self._queue = []\n", "self._observe = None\n" ]
[ "def __init__(self, name, shots):...\n", "self.name = name\n", "self.shots = shots\n", "self._out = None\n", "self._queue = []\n", "self._observe = None\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'" ]
[ "def FUNC_5(self):...\n", "" ]
[ "def test_run_interactive_shell_command_kwargs_delegation(self):...\n", "" ]
[ 0, 0 ]
[ "FunctionDef'", "Condition" ]
[ "def FUNC_23(self):...\n", "VAR_22 = model.Identity.from_bytes('user:[email protected]')\n", "@classmethod...\n", "return [lambda _request: VAR_22]\n" ]
[ "def test_delegation_token(self):...\n", "peer_ident = model.Identity.from_bytes('user:[email protected]')\n", "@classmethod...\n", "return [lambda _request: peer_ident]\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Condition", "Return'" ]
[ "@property...\n", "return self._result_getter()\n" ]
[ "@property...\n", "return self._result_getter()\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "def FUNC_16(self, VAR_9=True):...\n", "\"\"\"docstring\"\"\"\n", "self._kill_process_type(VAR_13.PROCESS_TYPE_REDIS_SERVER, VAR_9=check_alive)\n" ]
[ "def kill_redis(self, check_alive=True):...\n", "\"\"\"docstring\"\"\"\n", "self._kill_process_type(ray_constants.PROCESS_TYPE_REDIS_SERVER,\n check_alive=check_alive)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Expr'" ]
[ "def FUNC_12(self, VAR_10=None):...\n", "\"\"\"docstring\"\"\"\n", "VAR_30 = vars(self.parse_known_args(VAR_15=True)[0])\n", "VAR_13 = VAR_30.get('image_mode', None)\n", "if VAR_13 is not None and VAR_13 != 'none':\n", "self.add_image_args(VAR_13)\n", "VAR_12 = VAR_30.get('task', None)\n", "if VAR_12 is not None:\n", "self.add_task_args(VAR_12)\n", "VAR_31 = VAR_30.get('evaltask', None)\n", "if VAR_31 is not None:\n", "self.add_task_args(VAR_31)\n", "VAR_11 = VAR_30.get('model', None)\n", "if VAR_11 is not None:\n", "self.add_model_subargs(VAR_11)\n", "self.set_defaults(**self._defaults)\n" ]
[ "def add_extra_args(self, args=None):...\n", "\"\"\"docstring\"\"\"\n", "parsed = vars(self.parse_known_args(nohelp=True)[0])\n", "image_mode = parsed.get('image_mode', None)\n", "if image_mode is not None and image_mode != 'none':\n", "self.add_image_args(image_mode)\n", "task = parsed.get('task', None)\n", "if task is not None:\n", "self.add_task_args(task)\n", "evaltask = parsed.get('evaltask', None)\n", "if evaltask is not None:\n", "self.add_task_args(evaltask)\n", "model = parsed.get('model', None)\n", "if model is not None:\n", "self.add_model_subargs(model)\n", "self.set_defaults(**self._defaults)\n" ]
[ 0, 0, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Condition", "Expr'", "Assign'", "Condition", "Expr'", "Assign'", "Condition", "Expr'", "Assign'", "Condition", "Expr'", "Expr'" ]
[ "def FUNC_2(VAR_7):...\n", "return urlparse.urljoin(CLASS_0.server_location, VAR_7)\n" ]
[ "def _BuildUri(handler):...\n", "return urlparse.urljoin(BaseRequest.server_location, handler)\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def __init__(self, VAR_0):...\n", "self.MAX_P_CHECKS = 5\n", "self.MAX_CRAWLS = 1\n", "self.MAX_PATH_LENGTH = 50\n", "self.TARGET = 'Philosophy'\n", "self.DOMAIN = 'https://en.wikipedia.org'\n", "self.start_wiki = 'Special:Random' if not VAR_0 else VAR_0\n", "self.path_lengths = []\n", "self.wiki_to_target_length = {}\n", "self.completed_path = 0\n", "self.invalid_path = 0\n" ]
[ "def __init__(self, wiki):...\n", "self.MAX_P_CHECKS = 5\n", "self.MAX_CRAWLS = 1\n", "self.MAX_PATH_LENGTH = 50\n", "self.TARGET = 'Philosophy'\n", "self.DOMAIN = 'https://en.wikipedia.org'\n", "self.start_wiki = 'Special:Random' if not wiki else wiki\n", "self.path_lengths = []\n", "self.wiki_to_target_length = {}\n", "self.completed_path = 0\n", "self.invalid_path = 0\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'" ]
[ "def __next__(self):...\n", "if self._resource_iters:\n", "if self._field_iters:\n", "if self._field_iters[-1]:\n", "if self.current_resource:\n", "VAR_7 = self._field_iters[-1][0]\n", "self._field_iters.pop()\n", "if hasattr(self, 'on_exit'):\n", "VAR_13, VAR_15 = VAR_6(self._resource_iters[-1])\n", "self._path.pop()\n", "if VAR_13 is not None:\n", "self._resource_iters.append(VAR_7.item_iter_from_object(self.current_resource))\n", "self.on_exit()\n", "self._resource_iters.pop()\n", "VAR_8, VAR_8, VAR_7 = self._path[-1]\n", "self._field_iters.append(list(VAR_15._meta.composite_fields))\n", "self._path.append((CLASS_0, CLASS_0, VAR_7.name))\n", "self._resource_stack.pop()\n", "self._path[-1] = VAR_13, CLASS_0, VAR_7\n", "self._resource_stack[-1] = VAR_15\n", "self._resource_stack.append(None)\n", "return VAR_6(self)\n", "if hasattr(self, 'on_enter'):\n", "self._field_iters[-1].pop(0)\n", "self.on_enter()\n", "return VAR_15\n" ]
[ "def __next__(self):...\n", "if self._resource_iters:\n", "if self._field_iters:\n", "if self._field_iters[-1]:\n", "if self.current_resource:\n", "field = self._field_iters[-1][0]\n", "self._field_iters.pop()\n", "if hasattr(self, 'on_exit'):\n", "key, next_resource = next(self._resource_iters[-1])\n", "self._path.pop()\n", "if key is not None:\n", "self._resource_iters.append(field.item_iter_from_object(self.current_resource))\n", "self.on_exit()\n", "self._resource_iters.pop()\n", "_, _, field = self._path[-1]\n", "self._field_iters.append(list(next_resource._meta.composite_fields))\n", "self._path.append((NotSupplied, NotSupplied, field.name))\n", "self._resource_stack.pop()\n", "self._path[-1] = key, NotSupplied, field\n", "self._resource_stack[-1] = next_resource\n", "self._resource_stack.append(None)\n", "return next(self)\n", "if hasattr(self, 'on_enter'):\n", "self._field_iters[-1].pop(0)\n", "self.on_enter()\n", "return next_resource\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Condition", "Condition", "Condition", "Assign'", "Expr'", "Condition", "Assign'", "Expr'", "Condition", "Expr'", "Expr'", "Expr'", "Assign'", "Expr'", "Expr'", "Expr'", "Assign'", "Assign'", "Expr'", "Return'", "Condition", "Expr'", "Expr'", "Return'" ]
[ "def FUNC_5(self):...\n", "if hasattr(self, 'extractedhhdoc'):\n", "return self.extractedhhdoc\n", "return self\n" ]
[ "def pull_out_hh(self):...\n", "if hasattr(self, 'extractedhhdoc'):\n", "return self.extractedhhdoc\n", "return self\n" ]
[ 0, 6, 6, 0 ]
[ "FunctionDef'", "Condition", "Return'", "Return'" ]
[ "import re, requests, argparse, logging, os, coloredlogs, datetime, getpass, tempfile, itertools, json, concurrent.futures, random\n", "from utils import *\n", "from UploadForm import UploadForm\n", "from threading import Lock\n", "VAR_0 = '0.5.0'\n", "logging.basicConfig(datefmt='[%m/%d/%Y-%H:%M:%S]')\n", "VAR_1 = logging.getLogger('fuxploider')\n", "coloredlogs.install(VAR_1=logger, fmt=\n '%(asctime)s %(levelname)s - %(message)s', level=logging.INFO)\n", "logging.getLogger('requests').setLevel(logging.ERROR)\n", "VAR_2 = 'payloads'\n", "VAR_31 = json.loads(fd.read())\n", "VAR_3 = [x['templateName'] for x in VAR_31]\n", "VAR_4 = 'string'\n", "VAR_4 += \"\"\"\n\tDefault templates are the following (name - description) : \"\"\"\n", "for VAR_46 in VAR_31:\n", "VAR_4 += \"\\n\\t * '\" + VAR_46['templateName'] + \"' - \" + VAR_46['description']\n", "VAR_5 = argparse.ArgumentParser(epilog=templatesSection, description=\n __doc__, formatter_class=argparse.RawTextHelpFormatter)\n", "VAR_5.add_argument('-d', '--data', metavar='postData', dest='data', help=\n 'Additionnal data to be transmitted via POST method. Example : -d \"key1=value1&key2=value2\"'\n , type=valid_postData)\n", "VAR_5.add_argument('--proxy', metavar='proxyUrl', dest='proxy', help=\n 'Proxy information. Example : --proxy \"user:[email protected]:8080\"',\n type=valid_proxyString)\n", "VAR_5.add_argument('--proxy-creds', metavar='credentials', nargs='?', const\n =True, dest='proxyCreds', help=\n \"Prompt for proxy credentials at runtime. Format : 'user:pass'\", type=\n valid_proxyCreds)\n", "VAR_5.add_argument('-f', '--filesize', metavar='integer', nargs=1, default=\n ['10'], dest='size', help=\n 'File size to use for files to be created and uploaded (in kB).')\n", "VAR_5.add_argument('--cookies', metavar='omnomnom', nargs=1, dest='cookies',\n help=\n 'Cookies to use with HTTP requests. Example : PHPSESSID=aef45aef45afeaef45aef45&JSESSID=AQSEJHQSQSG'\n , type=valid_postData)\n", "VAR_5.add_argument('--uploads-path', default=[None], metavar='path', nargs=\n 1, dest='uploadsPath', help=\n \"Path on the remote server where uploads are put. Example : '/tmp/uploads/'\"\n )\n", "VAR_5.add_argument('-t', '--template', metavar='templateName', nargs=1,\n dest='template', help=\n 'Malicious payload to use for code execution detection. Default is to use every known templates. For a complete list of templates, see the TEMPLATE section.'\n )\n", "VAR_5.add_argument('-r', '--regex-override', metavar='regex', nargs=1, dest\n ='regexOverride', help=\n 'Specify a regular expression to detect code execution. Overrides the default code execution detection regex defined in the template in use.'\n , type=valid_regex)\n", "VAR_6 = VAR_5.add_argument_group('Required named arguments')\n", "VAR_6.add_argument('-u', '--url', metavar='target', dest='url', required=\n True, help=\n 'Web page URL containing the file upload form to be tested. Example : http://test.com/index.html?action=upload'\n , type=valid_url)\n", "VAR_6.add_argument('--not-regex', metavar='regex', help=\n 'Regex matching an upload failure', type=valid_regex, dest='notRegex')\n", "VAR_6.add_argument('--true-regex', metavar='regex', help=\n 'Regex matching an upload success', type=valid_regex, dest='trueRegex')\n", "VAR_7 = VAR_5.add_mutually_exclusive_group()\n", "VAR_7.add_argument('-l', '--legit-extensions', metavar='listOfExtensions',\n dest='legitExtensions', nargs=1, help=\n \"Legit extensions expected, for a normal use of the form, comma separated. Example : 'jpg,png,bmp'\"\n )\n", "VAR_7.add_argument('-n', metavar='n', nargs=1, default=['100'], dest='n',\n help='Number of common extensions to use. Example : -n 100', type=\n valid_nArg)\n", "VAR_8 = VAR_5.add_mutually_exclusive_group()\n", "VAR_8.add_argument('-v', action='store_true', required=False, dest=\n 'verbose', help='Verbose mode')\n", "VAR_8.add_argument('-vv', action='store_true', required=False, dest=\n 'veryVerbose', help='Very verbose mode')\n", "VAR_8.add_argument('-vvv', action='store_true', required=False, dest=\n 'veryVeryVerbose', help='Much verbose, very log, wow.')\n", "VAR_5.add_argument('-s', '--skip-recon', action='store_true', required=\n False, dest='skipRecon', help=\n 'Skip recon phase, where fuxploider tries to determine what extensions are expected and filtered by the server. Needs -l switch.'\n )\n", "VAR_5.add_argument('-y', action='store_true', required=False, dest=\n 'detectAllEntryPoints', help=\n 'Force detection of every entry points. Will not stop at first code exec found.'\n )\n", "VAR_5.add_argument('-T', '--threads', metavar='Threads', nargs=1, dest=\n 'nbThreads', help='Number of parallel tasks (threads).', type=int,\n default=[4])\n", "VAR_9 = VAR_5.add_mutually_exclusive_group()\n", "VAR_9.add_argument('-U', '--user-agent', metavar='useragent', nargs=1, dest\n ='userAgent', help='User-agent to use while requesting the target.',\n type=str, default=[requests.utils.default_user_agent()])\n", "VAR_9.add_argument('--random-user-agent', action='store_true', required=\n False, dest='randomUserAgent', help=\n 'Use a random user-agent while requesting the target.')\n", "VAR_10 = VAR_5.add_argument_group('Manual Form Detection arguments')\n", "VAR_10.add_argument('-m', '--manual-form-detection', action='store_true',\n dest='manualFormDetection', help=\n 'Disable automatic form detection. Useful when automatic detection fails due to: (1) Form loaded using Javascript (2) Multiple file upload forms in URL.'\n )\n", "VAR_10.add_argument('--input-name', metavar='image', dest='inputName', help\n ='Name of input for file. Example: <input type=\"file\" name=\"image\">')\n", "VAR_10.add_argument('--form-action', default='', metavar='upload.php', dest\n ='formAction', help=\n 'Path of form action. Example: <form method=\"POST\" action=\"upload.php\">')\n", "VAR_11 = VAR_5.parse_args()\n", "VAR_11.uploadsPath = VAR_11.uploadsPath[0]\n", "VAR_11.nbThreads = VAR_11.nbThreads[0]\n", "VAR_11.userAgent = VAR_11.userAgent[0]\n", "if VAR_11.randomUserAgent:\n", "VAR_45 = 0\n", "if VAR_11.template:\n", "for l in fd:\n", "VAR_11.template = VAR_11.template[0]\n", "if VAR_11.regexOverride:\n", "VAR_45 += 1\n", "fd.seek(0)\n", "if VAR_11.template not in VAR_3:\n", "for VAR_46 in VAR_31:\n", "VAR_11.verbosity = 0\n", "VAR_45 = random.randint(0, VAR_45)\n", "logging.warning('Unknown template : %s', VAR_11.template)\n", "VAR_31 = [[x for x in VAR_31 if x['templateName'] == VAR_11.template][0]]\n", "VAR_46['codeExecRegex'] = VAR_11.regexOverride[0]\n", "if VAR_11.verbose:\n", "for i in range(0, VAR_45):\n", "VAR_22 = input('Use default templates instead ? [Y/n]')\n", "VAR_11.verbosity = 1\n", "if VAR_11.veryVerbose:\n", "VAR_11.userAgent = fd.readline()[:-1]\n", "if not VAR_22.lower().startswith('y'):\n", "VAR_11.verbosity = 2\n", "if VAR_11.veryVeryVerbose:\n", "exit()\n", "VAR_11.verbosity = 3\n", "VAR_1.verbosity = VAR_11.verbosity\n", "if VAR_11.verbosity > 0:\n", "coloredlogs.install(VAR_1=logger, fmt=\n '%(asctime)s %(levelname)s - %(message)s', level=logging.DEBUG)\n", "if VAR_11.proxyCreds and VAR_11.proxy == None:\n", "VAR_5.error('--proxy-creds must be used with --proxy.')\n", "if VAR_11.skipRecon and VAR_11.legitExtensions == None:\n", "VAR_5.error(\n '-s switch needs -l switch. Cannot skip recon phase without any known entry point.'\n )\n", "VAR_11.n = int(VAR_11.n[0])\n", "VAR_11.size = int(VAR_11.size[0])\n", "VAR_11.size = 1024 * VAR_11.size\n", "if not VAR_11.notRegex and not VAR_11.trueRegex:\n", "VAR_5.error(\n 'At least one detection method must be provided, either with --not-regex or with --true-regex.'\n )\n", "if VAR_11.legitExtensions:\n", "VAR_11.legitExtensions = VAR_11.legitExtensions[0].split(',')\n", "if VAR_11.cookies:\n", "VAR_11.cookies = postDataFromStringToJSON(VAR_11.cookies[0])\n", "if VAR_11.manualFormDetection and VAR_11.inputName is None:\n", "VAR_5.error('--manual-form-detection requires --input-name')\n", "print('string' + VAR_0 + 'string')\n", "if VAR_11.proxyCreds == True:\n", "VAR_11.proxyCreds = {}\n", "VAR_12 = datetime.datetime.now()\n", "VAR_11.proxyCreds['username'] = input('Proxy username : ')\n", "print('[*] starting at ' + str(VAR_12.hour) + ':' + str(VAR_12.minute) +\n ':' + str(VAR_12.second))\n", "VAR_11.proxyCreds['password'] = getpass.getpass('Proxy password : ')\n", "VAR_13 = 'mimeTypes.basic'\n", "VAR_14 = loadExtensions('file', VAR_13)\n", "VAR_15 = []\n", "if VAR_11.legitExtensions:\n", "VAR_11.legitExtensions = [x.lower() for x in VAR_11.legitExtensions]\n", "VAR_11.legitExtensions = VAR_15\n", "VAR_32 = [VAR_20[0] for VAR_20 in VAR_14]\n", "VAR_16 = postDataFromStringToJSON(VAR_11.data)\n", "for VAR_21 in VAR_11.legitExtensions:\n", "VAR_17 = requests.Session()\n", "if VAR_21 in VAR_32:\n", "if VAR_11.cookies:\n", "VAR_15.append(VAR_21)\n", "logging.warning(\n \"Extension %s can't be found as a valid/known extension with associated mime type.\"\n , VAR_21)\n", "for VAR_47 in VAR_11.cookies.keys():\n", "VAR_17.headers = {'User-Agent': VAR_11.userAgent}\n", "VAR_17.cookies[VAR_47] = VAR_11.cookies[VAR_47]\n", "VAR_17.trust_env = False\n", "if VAR_11.proxy:\n", "if VAR_11.proxy['username'] and VAR_11.proxy['password'] and VAR_11.proxyCreds:\n", "if VAR_11.manualFormDetection:\n", "logging.warning('string')\n", "if VAR_11.proxyCreds:\n", "if VAR_11.formAction == '':\n", "VAR_37 = UploadForm(VAR_11.notRegex, VAR_11.trueRegex, VAR_17, VAR_11.size,\n VAR_16, VAR_11.uploadsPath)\n", "VAR_48 = VAR_11.proxyCreds['username']\n", "VAR_48 = VAR_11.proxy['username']\n", "VAR_1.warning('string')\n", "VAR_37 = UploadForm(VAR_11.notRegex, VAR_11.trueRegex, VAR_17, VAR_11.size,\n VAR_16, VAR_11.uploadsPath, VAR_11.url, VAR_11.formAction, VAR_11.inputName\n )\n", "VAR_37.setup(VAR_11.url)\n", "VAR_49 = VAR_11.proxyCreds['password']\n", "VAR_49 = VAR_11.proxy['password']\n", "VAR_37.threads = VAR_11.nbThreads\n", "VAR_33 = VAR_11.proxy['protocol']\n", "VAR_18 = VAR_37.uploadUrl\n", "VAR_34 = VAR_11.proxy['hostname']\n", "VAR_19 = {'name': VAR_37.inputName}\n", "VAR_35 = VAR_11.proxy['port']\n", "VAR_20 = datetime.datetime.now()\n", "VAR_36 = ''\n", "if not VAR_11.skipRecon:\n", "if VAR_33 != None:\n", "if len(VAR_11.legitExtensions) > 0:\n", "VAR_1.info(\n '### Skipping detection of valid extensions, using provided extensions instead (%s)'\n , VAR_11.legitExtensions)\n", "VAR_36 += VAR_33 + '://'\n", "VAR_36 += 'http://'\n", "VAR_51 = VAR_37.detectValidExtensions(VAR_14, VAR_11.n, VAR_11.legitExtensions)\n", "VAR_51 = VAR_37.detectValidExtensions(VAR_14, VAR_11.n)\n", "VAR_37.validExtensions = VAR_11.legitExtensions\n", "if VAR_48 != None and VAR_49 != None:\n", "VAR_1.info('### Tried %s extensions, %s are valid.', VAR_51, len(VAR_37.\n validExtensions))\n", "if VAR_37.validExtensions == []:\n", "VAR_36 += VAR_48 + ':' + VAR_49 + '@'\n", "VAR_36 += VAR_34\n", "VAR_1.error('No valid extension found.')\n", "VAR_21 = datetime.datetime.now()\n", "if VAR_35 != None:\n", "exit()\n", "print('Extensions detection : ' + str(VAR_21 - VAR_20))\n", "VAR_36 += ':' + VAR_35\n", "if VAR_33 == 'https':\n", "VAR_22 = input('Start uploading payloads ? [Y/n] : ')\n", "VAR_50 = {'https': VAR_36}\n", "VAR_50 = {'http': VAR_36, 'https': VAR_36}\n", "VAR_37.shouldLog = True\n", "VAR_17.proxies.update(VAR_50)\n", "if VAR_22.lower().startswith('y') or VAR_22 == '':\n", "exit('Exiting.')\n", "VAR_23 = []\n", "VAR_37.stopThreads = True\n", "VAR_38 = json.loads(rawTechniques.read())\n", "VAR_1.info(\n '### Starting code execution detection (messing with file extensions and mime types...)'\n )\n", "VAR_24 = datetime.datetime.now()\n", "VAR_25 = 0\n", "VAR_26 = []\n", "VAR_27 = {}\n", "for VAR_40 in VAR_31:\n", "VAR_39 = open(VAR_2 + '/' + VAR_40['filename'], 'rb')\n", "VAR_28 = False\n", "VAR_27[VAR_40['templateName']] = VAR_39.read()\n", "VAR_29 = 0\n", "VAR_39.close()\n", "VAR_44 = []\n", "VAR_41 = VAR_40['nastyExt']\n", "for VAR_20 in VAR_26:\n", "VAR_28 = True\n", "VAR_30 = datetime.datetime.now()\n", "VAR_42 = getMime(VAR_14, VAR_41)\n", "VAR_52 = VAR_20['suffix']\n", "for future in concurrent.futures.as_completed(VAR_44):\n", "executor.shutdown(wait=False)\n", "print()\n", "VAR_43 = VAR_40['extVariants']\n", "VAR_53 = VAR_20['mime']\n", "VAR_57 = future.result()\n", "executor._threads.clear()\n", "logging.info('%s entry point(s) found using %s HTTP requests.', VAR_25,\n VAR_37.httpRequests)\n", "for VAR_46 in VAR_38:\n", "VAR_54 = VAR_27[VAR_20['templateName']]\n", "VAR_29 += 1\n", "concurrent.futures.thread._threads_queues.clear()\n", "print('Found the following entry points : ')\n", "for nastyVariant in ([VAR_41] + VAR_43):\n", "VAR_55 = [VAR_46['codeExecRegex'] for VAR_46 in VAR_31 if VAR_46[\n 'templateName'] == VAR_20['templateName']][0]\n", "if not VAR_28:\n", "VAR_1.setLevel(logging.CRITICAL)\n", "print(VAR_23)\n", "for legitExt in VAR_37.validExtensions:\n", "VAR_56 = executor.submit(VAR_37.submitTestCase, VAR_52, VAR_53, VAR_54, VAR_55)\n", "if VAR_57['codeExec']:\n", "VAR_1.verbosity = -1\n", "VAR_58 = getMime(VAR_14, legitExt)\n", "VAR_56.a = VAR_20\n", "VAR_59 = future.a\n", "VAR_53 = VAR_58 if VAR_46['mime'] == 'legit' else VAR_42\n", "VAR_44.append(VAR_56)\n", "logging.info(\"\\x1b[1m\\x1b[42mCode execution obtained ('%s','%s','%s')\\x1b[m\",\n VAR_59['suffix'], VAR_59['mime'], VAR_59['templateName'])\n", "VAR_52 = VAR_46['suffix'].replace('$legitExt$', legitExt).replace('$nastyExt$',\n nastyVariant)\n", "VAR_25 += 1\n", "VAR_26.append({'suffix': VAR_52, 'mime': VAR_53, 'templateName': VAR_40[\n 'templateName']})\n", "VAR_23.append(VAR_59)\n", "if not VAR_11.detectAllEntryPoints:\n" ]
[ "import re, requests, argparse, logging, os, coloredlogs, datetime, getpass, tempfile, itertools, json, concurrent.futures, random\n", "from utils import *\n", "from UploadForm import UploadForm\n", "from threading import Lock\n", "version = '0.5.0'\n", "logging.basicConfig(datefmt='[%m/%d/%Y-%H:%M:%S]')\n", "logger = logging.getLogger('fuxploider')\n", "coloredlogs.install(logger=logger, fmt=\n '%(asctime)s %(levelname)s - %(message)s', level=logging.INFO)\n", "logging.getLogger('requests').setLevel(logging.ERROR)\n", "templatesFolder = 'payloads'\n", "templates = json.loads(fd.read())\n", "templatesNames = [x['templateName'] for x in templates]\n", "templatesSection = \"\"\"[TEMPLATES]\nTemplates are malicious payloads meant to be uploaded on the scanned remote server. Code execution detection is done based on the expected output of the payload.\"\"\"\n", "templatesSection += \"\"\"\n\tDefault templates are the following (name - description) : \"\"\"\n", "for t in templates:\n", "templatesSection += \"\\n\\t * '\" + t['templateName'] + \"' - \" + t['description']\n", "parser = argparse.ArgumentParser(epilog=templatesSection, description=\n __doc__, formatter_class=argparse.RawTextHelpFormatter)\n", "parser.add_argument('-d', '--data', metavar='postData', dest='data', help=\n 'Additionnal data to be transmitted via POST method. Example : -d \"key1=value1&key2=value2\"'\n , type=valid_postData)\n", "parser.add_argument('--proxy', metavar='proxyUrl', dest='proxy', help=\n 'Proxy information. Example : --proxy \"user:[email protected]:8080\"',\n type=valid_proxyString)\n", "parser.add_argument('--proxy-creds', metavar='credentials', nargs='?',\n const=True, dest='proxyCreds', help=\n \"Prompt for proxy credentials at runtime. Format : 'user:pass'\", type=\n valid_proxyCreds)\n", "parser.add_argument('-f', '--filesize', metavar='integer', nargs=1, default\n =['10'], dest='size', help=\n 'File size to use for files to be created and uploaded (in kB).')\n", "parser.add_argument('--cookies', metavar='omnomnom', nargs=1, dest=\n 'cookies', help=\n 'Cookies to use with HTTP requests. Example : PHPSESSID=aef45aef45afeaef45aef45&JSESSID=AQSEJHQSQSG'\n , type=valid_postData)\n", "parser.add_argument('--uploads-path', default=[None], metavar='path', nargs\n =1, dest='uploadsPath', help=\n \"Path on the remote server where uploads are put. Example : '/tmp/uploads/'\"\n )\n", "parser.add_argument('-t', '--template', metavar='templateName', nargs=1,\n dest='template', help=\n 'Malicious payload to use for code execution detection. Default is to use every known templates. For a complete list of templates, see the TEMPLATE section.'\n )\n", "parser.add_argument('-r', '--regex-override', metavar='regex', nargs=1,\n dest='regexOverride', help=\n 'Specify a regular expression to detect code execution. Overrides the default code execution detection regex defined in the template in use.'\n , type=valid_regex)\n", "requiredNamedArgs = parser.add_argument_group('Required named arguments')\n", "requiredNamedArgs.add_argument('-u', '--url', metavar='target', dest='url',\n required=True, help=\n 'Web page URL containing the file upload form to be tested. Example : http://test.com/index.html?action=upload'\n , type=valid_url)\n", "requiredNamedArgs.add_argument('--not-regex', metavar='regex', help=\n 'Regex matching an upload failure', type=valid_regex, dest='notRegex')\n", "requiredNamedArgs.add_argument('--true-regex', metavar='regex', help=\n 'Regex matching an upload success', type=valid_regex, dest='trueRegex')\n", "exclusiveArgs = parser.add_mutually_exclusive_group()\n", "exclusiveArgs.add_argument('-l', '--legit-extensions', metavar=\n 'listOfExtensions', dest='legitExtensions', nargs=1, help=\n \"Legit extensions expected, for a normal use of the form, comma separated. Example : 'jpg,png,bmp'\"\n )\n", "exclusiveArgs.add_argument('-n', metavar='n', nargs=1, default=['100'],\n dest='n', help='Number of common extensions to use. Example : -n 100',\n type=valid_nArg)\n", "exclusiveVerbosityArgs = parser.add_mutually_exclusive_group()\n", "exclusiveVerbosityArgs.add_argument('-v', action='store_true', required=\n False, dest='verbose', help='Verbose mode')\n", "exclusiveVerbosityArgs.add_argument('-vv', action='store_true', required=\n False, dest='veryVerbose', help='Very verbose mode')\n", "exclusiveVerbosityArgs.add_argument('-vvv', action='store_true', required=\n False, dest='veryVeryVerbose', help='Much verbose, very log, wow.')\n", "parser.add_argument('-s', '--skip-recon', action='store_true', required=\n False, dest='skipRecon', help=\n 'Skip recon phase, where fuxploider tries to determine what extensions are expected and filtered by the server. Needs -l switch.'\n )\n", "parser.add_argument('-y', action='store_true', required=False, dest=\n 'detectAllEntryPoints', help=\n 'Force detection of every entry points. Will not stop at first code exec found.'\n )\n", "parser.add_argument('-T', '--threads', metavar='Threads', nargs=1, dest=\n 'nbThreads', help='Number of parallel tasks (threads).', type=int,\n default=[4])\n", "exclusiveUserAgentsArgs = parser.add_mutually_exclusive_group()\n", "exclusiveUserAgentsArgs.add_argument('-U', '--user-agent', metavar=\n 'useragent', nargs=1, dest='userAgent', help=\n 'User-agent to use while requesting the target.', type=str, default=[\n requests.utils.default_user_agent()])\n", "exclusiveUserAgentsArgs.add_argument('--random-user-agent', action=\n 'store_true', required=False, dest='randomUserAgent', help=\n 'Use a random user-agent while requesting the target.')\n", "manualFormArgs = parser.add_argument_group('Manual Form Detection arguments')\n", "manualFormArgs.add_argument('-m', '--manual-form-detection', action=\n 'store_true', dest='manualFormDetection', help=\n 'Disable automatic form detection. Useful when automatic detection fails due to: (1) Form loaded using Javascript (2) Multiple file upload forms in URL.'\n )\n", "manualFormArgs.add_argument('--input-name', metavar='image', dest=\n 'inputName', help=\n 'Name of input for file. Example: <input type=\"file\" name=\"image\">')\n", "manualFormArgs.add_argument('--form-action', default='', metavar=\n 'upload.php', dest='formAction', help=\n 'Path of form action. Example: <form method=\"POST\" action=\"upload.php\">')\n", "args = parser.parse_args()\n", "args.uploadsPath = args.uploadsPath[0]\n", "args.nbThreads = args.nbThreads[0]\n", "args.userAgent = args.userAgent[0]\n", "if args.randomUserAgent:\n", "nb = 0\n", "if args.template:\n", "for l in fd:\n", "args.template = args.template[0]\n", "if args.regexOverride:\n", "nb += 1\n", "fd.seek(0)\n", "if args.template not in templatesNames:\n", "for t in templates:\n", "args.verbosity = 0\n", "nb = random.randint(0, nb)\n", "logging.warning('Unknown template : %s', args.template)\n", "templates = [[x for x in templates if x['templateName'] == args.template][0]]\n", "t['codeExecRegex'] = args.regexOverride[0]\n", "if args.verbose:\n", "for i in range(0, nb):\n", "cont = input('Use default templates instead ? [Y/n]')\n", "args.verbosity = 1\n", "if args.veryVerbose:\n", "args.userAgent = fd.readline()[:-1]\n", "if not cont.lower().startswith('y'):\n", "args.verbosity = 2\n", "if args.veryVeryVerbose:\n", "exit()\n", "args.verbosity = 3\n", "logger.verbosity = args.verbosity\n", "if args.verbosity > 0:\n", "coloredlogs.install(logger=logger, fmt=\n '%(asctime)s %(levelname)s - %(message)s', level=logging.DEBUG)\n", "if args.proxyCreds and args.proxy == None:\n", "parser.error('--proxy-creds must be used with --proxy.')\n", "if args.skipRecon and args.legitExtensions == None:\n", "parser.error(\n '-s switch needs -l switch. Cannot skip recon phase without any known entry point.'\n )\n", "args.n = int(args.n[0])\n", "args.size = int(args.size[0])\n", "args.size = 1024 * args.size\n", "if not args.notRegex and not args.trueRegex:\n", "parser.error(\n 'At least one detection method must be provided, either with --not-regex or with --true-regex.'\n )\n", "if args.legitExtensions:\n", "args.legitExtensions = args.legitExtensions[0].split(',')\n", "if args.cookies:\n", "args.cookies = postDataFromStringToJSON(args.cookies[0])\n", "if args.manualFormDetection and args.inputName is None:\n", "parser.error('--manual-form-detection requires --input-name')\n", "print(\n \"\"\"\u001b[1;32m\n \n ___ _ _ _ \n| _|_ _ _ _ ___| |___|_|_| |___ ___ \n| _| | |_'_| . | | . | | . | -_| _|\n|_| |___|_,_| _|_|___|_|___|___|_| \n |_| \n\n\u001b[1m\u001b[42m{version \"\"\"\n + version +\n \"\"\"}\u001b[m\n\n\u001b[m[!] legal disclaimer : Usage of fuxploider for attacking targets without prior mutual consent is illegal. It is the end user's responsibility to obey all applicable local, state and federal laws. Developers assume no liability and are not responsible for any misuse or damage caused by this program\n\t\"\"\"\n )\n", "if args.proxyCreds == True:\n", "args.proxyCreds = {}\n", "now = datetime.datetime.now()\n", "args.proxyCreds['username'] = input('Proxy username : ')\n", "print('[*] starting at ' + str(now.hour) + ':' + str(now.minute) + ':' +\n str(now.second))\n", "args.proxyCreds['password'] = getpass.getpass('Proxy password : ')\n", "mimeFile = 'mimeTypes.basic'\n", "extensions = loadExtensions('file', mimeFile)\n", "tmpLegitExt = []\n", "if args.legitExtensions:\n", "args.legitExtensions = [x.lower() for x in args.legitExtensions]\n", "args.legitExtensions = tmpLegitExt\n", "foundExt = [a[0] for a in extensions]\n", "postData = postDataFromStringToJSON(args.data)\n", "for b in args.legitExtensions:\n", "s = requests.Session()\n", "if b in foundExt:\n", "if args.cookies:\n", "tmpLegitExt.append(b)\n", "logging.warning(\n \"Extension %s can't be found as a valid/known extension with associated mime type.\"\n , b)\n", "for key in args.cookies.keys():\n", "s.headers = {'User-Agent': args.userAgent}\n", "s.cookies[key] = args.cookies[key]\n", "s.trust_env = False\n", "if args.proxy:\n", "if args.proxy['username'] and args.proxy['password'] and args.proxyCreds:\n", "if args.manualFormDetection:\n", "logging.warning(\n 'Proxy username and password provided by the --proxy-creds switch replaces credentials provided using the --proxy switch'\n )\n", "if args.proxyCreds:\n", "if args.formAction == '':\n", "up = UploadForm(args.notRegex, args.trueRegex, s, args.size, postData, args\n .uploadsPath)\n", "proxyUser = args.proxyCreds['username']\n", "proxyUser = args.proxy['username']\n", "logger.warning(\n 'Using Manual Form Detection and no action specified with --form-action. Defaulting to empty string - meaning form action will be set to --url parameter.'\n )\n", "up = UploadForm(args.notRegex, args.trueRegex, s, args.size, postData, args\n .uploadsPath, args.url, args.formAction, args.inputName)\n", "up.setup(args.url)\n", "proxyPass = args.proxyCreds['password']\n", "proxyPass = args.proxy['password']\n", "up.threads = args.nbThreads\n", "proxyProtocol = args.proxy['protocol']\n", "uploadURL = up.uploadUrl\n", "proxyHostname = args.proxy['hostname']\n", "fileInput = {'name': up.inputName}\n", "proxyPort = args.proxy['port']\n", "a = datetime.datetime.now()\n", "proxy = ''\n", "if not args.skipRecon:\n", "if proxyProtocol != None:\n", "if len(args.legitExtensions) > 0:\n", "logger.info(\n '### Skipping detection of valid extensions, using provided extensions instead (%s)'\n , args.legitExtensions)\n", "proxy += proxyProtocol + '://'\n", "proxy += 'http://'\n", "n = up.detectValidExtensions(extensions, args.n, args.legitExtensions)\n", "n = up.detectValidExtensions(extensions, args.n)\n", "up.validExtensions = args.legitExtensions\n", "if proxyUser != None and proxyPass != None:\n", "logger.info('### Tried %s extensions, %s are valid.', n, len(up.\n validExtensions))\n", "if up.validExtensions == []:\n", "proxy += proxyUser + ':' + proxyPass + '@'\n", "proxy += proxyHostname\n", "logger.error('No valid extension found.')\n", "b = datetime.datetime.now()\n", "if proxyPort != None:\n", "exit()\n", "print('Extensions detection : ' + str(b - a))\n", "proxy += ':' + proxyPort\n", "if proxyProtocol == 'https':\n", "cont = input('Start uploading payloads ? [Y/n] : ')\n", "proxies = {'https': proxy}\n", "proxies = {'http': proxy, 'https': proxy}\n", "up.shouldLog = True\n", "s.proxies.update(proxies)\n", "if cont.lower().startswith('y') or cont == '':\n", "exit('Exiting.')\n", "entryPoints = []\n", "up.stopThreads = True\n", "techniques = json.loads(rawTechniques.read())\n", "logger.info(\n '### Starting code execution detection (messing with file extensions and mime types...)'\n )\n", "c = datetime.datetime.now()\n", "nbOfEntryPointsFound = 0\n", "attempts = []\n", "templatesData = {}\n", "for template in templates:\n", "templatefd = open(templatesFolder + '/' + template['filename'], 'rb')\n", "stopThreads = False\n", "templatesData[template['templateName']] = templatefd.read()\n", "attemptsTested = 0\n", "templatefd.close()\n", "futures = []\n", "nastyExt = template['nastyExt']\n", "for a in attempts:\n", "stopThreads = True\n", "d = datetime.datetime.now()\n", "nastyMime = getMime(extensions, nastyExt)\n", "suffix = a['suffix']\n", "for future in concurrent.futures.as_completed(futures):\n", "executor.shutdown(wait=False)\n", "print()\n", "nastyExtVariants = template['extVariants']\n", "mime = a['mime']\n", "res = future.result()\n", "executor._threads.clear()\n", "logging.info('%s entry point(s) found using %s HTTP requests.',\n nbOfEntryPointsFound, up.httpRequests)\n", "for t in techniques:\n", "payload = templatesData[a['templateName']]\n", "attemptsTested += 1\n", "concurrent.futures.thread._threads_queues.clear()\n", "print('Found the following entry points : ')\n", "for nastyVariant in ([nastyExt] + nastyExtVariants):\n", "codeExecRegex = [t['codeExecRegex'] for t in templates if t['templateName'] ==\n a['templateName']][0]\n", "if not stopThreads:\n", "logger.setLevel(logging.CRITICAL)\n", "print(entryPoints)\n", "for legitExt in up.validExtensions:\n", "f = executor.submit(up.submitTestCase, suffix, mime, payload, codeExecRegex)\n", "if res['codeExec']:\n", "logger.verbosity = -1\n", "legitMime = getMime(extensions, legitExt)\n", "f.a = a\n", "foundEntryPoint = future.a\n", "mime = legitMime if t['mime'] == 'legit' else nastyMime\n", "futures.append(f)\n", "logging.info(\"\\x1b[1m\\x1b[42mCode execution obtained ('%s','%s','%s')\\x1b[m\",\n foundEntryPoint['suffix'], foundEntryPoint['mime'], foundEntryPoint[\n 'templateName'])\n", "suffix = t['suffix'].replace('$legitExt$', legitExt).replace('$nastyExt$',\n nastyVariant)\n", "nbOfEntryPointsFound += 1\n", "attempts.append({'suffix': suffix, 'mime': mime, 'templateName': template[\n 'templateName']})\n", "entryPoints.append(foundEntryPoint)\n", "if not args.detectAllEntryPoints:\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 7, 0, 0, 0, 7, 0, 0, 0, 0, 7, 0, 0, 0, 0, 7, 0, 0, 0, 0, 7, 0, 0, 0, 0, 7, 7, 0, 0, 7, 0, 0, 7, 0, 0, 7, 0, 7, 0, 0 ]
[ "Import'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Assign'", "Expr'", "Assign'", "Expr'", "Expr'", "Assign'", "Assign'", "Assign'", "Assign'", "AugAssign'", "For", "AugAssign'", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Assign'", "Expr'", "Expr'", "Expr'", "Assign'", "Expr'", "Expr'", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Assign'", "Expr'", "Expr'", "Assign'", "Expr'", "Expr'", "Expr'", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Condition", "For", "Assign'", "Condition", "AugAssign'", "Expr'", "Condition", "For", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "Condition", "For", "Assign'", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Condition", "Expr'", "Assign'", "Assign'", "Condition", "Expr'", "Condition", "Expr'", "Condition", "Expr'", "Assign'", "Assign'", "Assign'", "Condition", "Expr'", "Condition", "Assign'", "Condition", "Assign'", "Condition", "Expr'", "Expr'", "Condition", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "For", "Assign'", "Condition", "Condition", "Expr'", "Expr'", "For", "Assign'", "Assign'", "Assign'", "Condition", "Condition", "Condition", "Expr'", "Condition", "For", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "Expr'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Condition", "Condition", "Expr'", "AugAssign'", "AugAssign'", "Assign'", "Assign'", "Assign'", "Condition", "Expr'", "Condition", "AugAssign'", "AugAssign'", "Expr'", "Assign'", "Condition", "Expr'", "Expr'", "AugAssign'", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Condition", "Expr'", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "Assign'", "Assign'", "For", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "For", "Assign'", "Assign'", "Assign'", "Assign'", "For", "Expr'", "Expr'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'", "For", "Assign'", "AugAssign'", "Expr'", "Expr'", "For", "Assign'", "Condition", "Expr'", "Expr'", "For", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'", "Assign'", "AugAssign'", "Expr'", "Expr'", "Condition" ]
[ "def FUNC_32(VAR_19):...\n", "return dict((expr, vimsupport.VimExpressionToPythonType(expr)) for expr in\n VAR_19)\n" ]
[ "def BuildExtraConfData(extra_conf_vim_data):...\n", "return dict((expr, vimsupport.VimExpressionToPythonType(expr)) for expr in\n extra_conf_vim_data)\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_1(VAR_0, VAR_1=False):...\n", "\"\"\"docstring\"\"\"\n", "VAR_34 = VAR_0.split('\\n')\n", "if VAR_1:\n", "VAR_34 = VAR_34[0:1] + VAR_34[3:]\n", "return '\\n'.join(VAR_34)\n" ]
[ "def format_error_message(exception_message, task_exception=False):...\n", "\"\"\"docstring\"\"\"\n", "lines = exception_message.split('\\n')\n", "if task_exception:\n", "lines = lines[0:1] + lines[3:]\n", "return '\\n'.join(lines)\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Condition", "Assign'", "Return'" ]
[ "def __init__(self, VAR_46=50, VAR_47=100):...\n", "\"\"\"docstring\"\"\"\n", "self.regex = re.compile(\n '((?P<value>.+)(?P=value){{{min_repeat},{max_repeat}}})$'.format(VAR_46\n =min_repeat - 1, VAR_47=max_repeat - 1))\n" ]
[ "def __init__(self, min_repeat=50, max_repeat=100):...\n", "\"\"\"docstring\"\"\"\n", "self.regex = re.compile(\n '((?P<value>.+)(?P=value){{{min_repeat},{max_repeat}}})$'.format(\n min_repeat=min_repeat - 1, max_repeat=max_repeat - 1))\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'" ]
[ "@VAR_0.route('/users/<id>')...\n", "VAR_2 = connectToMySQL('users_db')\n", "VAR_4 = VAR_2.query_db('SELECT * FROM users WHERE id = {};'.format(VAR_1))\n", "return render_template('user.html', VAR_4=user[0])\n" ]
[ "@app.route('/users/<id>')...\n", "mysql = connectToMySQL('users_db')\n", "user = mysql.query_db('SELECT * FROM users WHERE id = {};'.format(id))\n", "return render_template('user.html', user=user[0])\n" ]
[ 0, 0, 4, 0 ]
[ "Condition", "Assign'", "Assign'", "Return'" ]
[ "def __init__(self, VAR_7):...\n", "self._user_options = VAR_7\n", "self._user_notified_about_crash = False\n", "self._diag_interface = DiagnosticInterface(VAR_7)\n", "self._omnicomp = OmniCompleter(VAR_7)\n", "self._latest_completion_request = None\n", "self._latest_file_parse_request = None\n", "self._server_stdout = None\n", "self._server_stderr = None\n", "self._server_popen = None\n", "self._filetypes_with_keywords_loaded = set()\n", "self._temp_options_filename = None\n", "self._ycmd_keepalive = YcmdKeepalive()\n", "self._SetupServer()\n", "self._ycmd_keepalive.Start()\n" ]
[ "def __init__(self, user_options):...\n", "self._user_options = user_options\n", "self._user_notified_about_crash = False\n", "self._diag_interface = DiagnosticInterface(user_options)\n", "self._omnicomp = OmniCompleter(user_options)\n", "self._latest_completion_request = None\n", "self._latest_file_parse_request = None\n", "self._server_stdout = None\n", "self._server_stderr = None\n", "self._server_popen = None\n", "self._filetypes_with_keywords_loaded = set()\n", "self._temp_options_filename = None\n", "self._ycmd_keepalive = YcmdKeepalive()\n", "self._SetupServer()\n", "self._ycmd_keepalive.Start()\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 7, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'" ]
[ "import os\n", "import sys\n", "import zlib\n", "import time\n", "import logging as logger\n", "import urlparse\n", "from datetime import datetime, timedelta\n", "import cPickle as pickle\n", "logger.info('cPickle module not available')\n", "sys.setrecursionlimit(10000)\n", "import pickle\n", "\"\"\"string\"\"\"\n", "def __init__(self, VAR_0='cache', VAR_1=timedelta(days=30), VAR_2=True):...\n", "\"\"\"docstring\"\"\"\n", "self.cache_dir = VAR_0\n", "self.expires = VAR_1\n", "self.compress = VAR_2\n", "def __getitem__(self, VAR_3):...\n", "\"\"\"docstring\"\"\"\n", "VAR_6 = self.url_to_path(VAR_3)\n", "if os.path.exists(VAR_6):\n", "VAR_8 = fp.read()\n", "def __setitem__(self, VAR_3, VAR_4):...\n", "if self.compress:\n", "\"\"\"docstring\"\"\"\n", "logger.info('Loading...')\n", "VAR_4 = pickle.loads(VAR_8)\n", "VAR_6 = self.url_to_path(VAR_3)\n", "VAR_8 = zlib.decompress(VAR_8)\n", "return VAR_4\n", "VAR_7 = os.path.dirname(VAR_6)\n", "if not os.path.exists(VAR_7):\n", "os.makedirs(VAR_7)\n", "VAR_8 = pickle.dumps(VAR_4)\n", "if self.compress:\n", "logger.info('Saving...')\n", "fp.write(VAR_8)\n", "VAR_8 = zlib.compress(VAR_8)\n", "def __delitem__(self, VAR_3):...\n", "\"\"\"docstring\"\"\"\n", "VAR_6 = self.url_to_path(VAR_3)\n", "os.remove(VAR_6)\n", "def FUNC_0(self, VAR_3):...\n", "os.removedirs(os.path.dirname(VAR_6))\n", "\"\"\"docstring\"\"\"\n", "VAR_9 = urlparse.urlsplit(VAR_3)\n", "VAR_6 = VAR_9.path\n", "if not VAR_6:\n", "VAR_6 = '/index.html'\n", "if VAR_6.endswith('/'):\n", "VAR_10 = VAR_9.netloc + VAR_6 + VAR_9.query\n", "VAR_6 += 'index.html'\n", "VAR_10 = '/'.join(segment[:255] for segment in VAR_10.split('/'))\n", "return os.path.join(self.cache_dir, VAR_10)\n" ]
[ "import os\n", "import sys\n", "import zlib\n", "import time\n", "import logging as logger\n", "import urlparse\n", "from datetime import datetime, timedelta\n", "import cPickle as pickle\n", "logger.info('cPickle module not available')\n", "sys.setrecursionlimit(10000)\n", "import pickle\n", "\"\"\"\n Dictionary interface that stores cached\n values in the file system rather than in memory.\n The file path is formed from an md5 hash of the key.\n \"\"\"\n", "def __init__(self, cache_dir='cache', expires=timedelta(days=30), compress=True...\n", "\"\"\"docstring\"\"\"\n", "self.cache_dir = cache_dir\n", "self.expires = expires\n", "self.compress = compress\n", "def __getitem__(self, url):...\n", "\"\"\"docstring\"\"\"\n", "path = self.url_to_path(url)\n", "if os.path.exists(path):\n", "data = fp.read()\n", "def __setitem__(self, url, result):...\n", "if self.compress:\n", "\"\"\"docstring\"\"\"\n", "logger.info('Loading...')\n", "result = pickle.loads(data)\n", "path = self.url_to_path(url)\n", "data = zlib.decompress(data)\n", "return result\n", "folder = os.path.dirname(path)\n", "if not os.path.exists(folder):\n", "os.makedirs(folder)\n", "data = pickle.dumps(result)\n", "if self.compress:\n", "logger.info('Saving...')\n", "fp.write(data)\n", "data = zlib.compress(data)\n", "def __delitem__(self, url):...\n", "\"\"\"docstring\"\"\"\n", "path = self.url_to_path(url)\n", "os.remove(path)\n", "def url_to_path(self, url):...\n", "os.removedirs(os.path.dirname(path))\n", "\"\"\"docstring\"\"\"\n", "components = urlparse.urlsplit(url)\n", "path = components.path\n", "if not path:\n", "path = '/index.html'\n", "if path.endswith('/'):\n", "filename = components.netloc + path + components.query\n", "path += 'index.html'\n", "filename = '/'.join(segment[:255] for segment in filename.split('/'))\n", "return os.path.join(self.cache_dir, filename)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Import'", "Import'", "Import'", "Import'", "Import'", "Import'", "ImportFrom'", "Import'", "Expr'", "Expr'", "Import'", "Expr'", "FunctionDef'", "Docstring", "Assign'", "Assign'", "Assign'", "FunctionDef'", "Docstring", "Assign'", "Condition", "Assign'", "FunctionDef'", "Condition", "Docstring", "Expr'", "Assign'", "Assign'", "Assign'", "Return'", "Assign'", "Condition", "Expr'", "Assign'", "Condition", "Expr'", "Expr'", "Assign'", "FunctionDef'", "Docstring", "Assign'", "Expr'", "FunctionDef'", "Expr'", "Docstring", "Assign'", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "AugAssign'", "Assign'", "Return'" ]
[ "def FUNC_22(VAR_21):...\n", "\"\"\"docstring\"\"\"\n", "VAR_45 = FUNC_21(VAR_21)\n", "if not isinstance(VAR_45, dict):\n", "return VAR_45\n" ]
[ "def load_configfile(configpath):...\n", "\"\"\"docstring\"\"\"\n", "config = _load_configfile(configpath)\n", "if not isinstance(config, dict):\n", "return config\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Condition", "Return'" ]
[ "def FUNC_23(self, VAR_16):...\n", "\"\"\"docstring\"\"\"\n", "VAR_35 = self.common._cli_run('showvlun -a -showcols Port', None)\n", "VAR_36 = {}\n", "for VAR_15 in VAR_16:\n", "VAR_36[VAR_15] = 0\n", "VAR_37 = None\n", "if VAR_35:\n", "VAR_35 = VAR_35[1:]\n", "return VAR_37\n", "for line in VAR_35:\n", "VAR_15 = line.strip()\n", "VAR_41 = sys.maxint\n", "if VAR_15 in VAR_36:\n", "for VAR_15, count in VAR_36.iteritems():\n", "VAR_36[VAR_15] = VAR_36[VAR_15] + 1\n", "if count < VAR_41:\n", "VAR_37 = VAR_15\n", "VAR_41 = count\n" ]
[ "def _get_least_used_nsp(self, nspss):...\n", "\"\"\"docstring\"\"\"\n", "result = self.common._cli_run('showvlun -a -showcols Port', None)\n", "nsp_counts = {}\n", "for nsp in nspss:\n", "nsp_counts[nsp] = 0\n", "current_least_used_nsp = None\n", "if result:\n", "result = result[1:]\n", "return current_least_used_nsp\n", "for line in result:\n", "nsp = line.strip()\n", "current_smallest_count = sys.maxint\n", "if nsp in nsp_counts:\n", "for nsp, count in nsp_counts.iteritems():\n", "nsp_counts[nsp] = nsp_counts[nsp] + 1\n", "if count < current_smallest_count:\n", "current_least_used_nsp = nsp\n", "current_smallest_count = count\n" ]
[ 0, 0, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "For", "Assign'", "Assign'", "Condition", "Assign'", "Return'", "For", "Assign'", "Assign'", "Condition", "For", "Assign'", "Condition", "Assign'", "Assign'" ]
[ "def FUNC_4(VAR_12, VAR_16, VAR_17, VAR_18, VAR_19):...\n", "\"\"\"docstring\"\"\"\n", "VAR_16 = VAR_16.copy()\n", "if VAR_17 is not None:\n", "VAR_16['exit_code'] = VAR_17\n", "if VAR_18:\n", "VAR_16['output'] = base64.b64encode(VAR_18)\n", "VAR_35 = VAR_12.url_read_json('/swarming/api/v1/bot/task_update/%s' %\n VAR_16['task_id'], VAR_31=params)\n", "VAR_16['output_chunk_start'] = VAR_19\n", "logging.debug('post_update() = %s', VAR_35)\n", "if VAR_35.get('error'):\n" ]
[ "def post_update(swarming_server, params, exit_code, stdout, output_chunk_start...\n", "\"\"\"docstring\"\"\"\n", "params = params.copy()\n", "if exit_code is not None:\n", "params['exit_code'] = exit_code\n", "if stdout:\n", "params['output'] = base64.b64encode(stdout)\n", "resp = swarming_server.url_read_json('/swarming/api/v1/bot/task_update/%s' %\n params['task_id'], data=params)\n", "params['output_chunk_start'] = output_chunk_start\n", "logging.debug('post_update() = %s', resp)\n", "if resp.get('error'):\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 5, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Expr'", "Condition" ]
[ "def FUNC_4(self):...\n", "\"\"\"docstring\"\"\"\n", "self.cursor.execute('create table t1(n int)')\n", "self.cursor.execute('insert into t1 values (?)', 1)\n", "self.cursor.execute('insert into t1 values (?)', 2)\n", "self.cursor.execute('insert into t1 values (?)', 3)\n", "for i in range(3):\n", "self.cursor.execute('select n from t1 where n < ?', 10)\n", "self.cursor.execute('select n from t1 where n < 3')\n" ]
[ "def test_multiple_bindings(self):...\n", "\"\"\"docstring\"\"\"\n", "self.cursor.execute('create table t1(n int)')\n", "self.cursor.execute('insert into t1 values (?)', 1)\n", "self.cursor.execute('insert into t1 values (?)', 2)\n", "self.cursor.execute('insert into t1 values (?)', 3)\n", "for i in range(3):\n", "self.cursor.execute('select n from t1 where n < ?', 10)\n", "self.cursor.execute('select n from t1 where n < 3')\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Expr'", "Expr'", "Expr'", "Expr'", "For", "Expr'", "Expr'" ]
[ "def FUNC_6(self):...\n", "self.mock(bot_main, 'get_remote', lambda : self.server)\n", "VAR_9 = []\n", "def FUNC_21(VAR_10):...\n", "VAR_9.append(1)\n", "return False\n" ]
[ "def test_setup_bot(self):...\n", "self.mock(bot_main, 'get_remote', lambda : self.server)\n", "setup_bots = []\n", "def setup_bot(_bot):...\n", "setup_bots.append(1)\n", "return False\n" ]
[ 0, 5, 0, 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Assign'", "FunctionDef'", "Expr'", "Return'" ]
[ "@staticmethod...\n", "\"\"\"docstring\"\"\"\n", "log.info(\n 'Evaluating last active users with date of last time when they used bot...'\n )\n", "VAR_10 = f'string{VAR_6}'\n", "VAR_19 = db.execute_query(VAR_10)\n", "log.error(\n 'Cannot get the last active users because of some problems with the database'\n )\n", "VAR_13 = VAR_19.fetchall()\n", "return VAR_13\n" ]
[ "@staticmethod...\n", "\"\"\"docstring\"\"\"\n", "log.info(\n 'Evaluating last active users with date of last time when they used bot...'\n )\n", "query = (\n f'SELECT p.chat_id, u.first_name, u.nickname, u.last_name, u.language FROM photo_queries_table2 p INNER JOIN users u ON p.chat_id = u.chat_id GROUP BY u.chat_id, u.first_name, u.nickname, u.last_name, u.language ORDER BY MAX(time)DESC LIMIT {limit}'\n )\n", "cursor = db.execute_query(query)\n", "log.error(\n 'Cannot get the last active users because of some problems with the database'\n )\n", "last_active_users = cursor.fetchall()\n", "return last_active_users\n" ]
[ 0, 0, 0, 4, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Expr'", "Assign'", "Assign'", "Expr'", "Assign'", "Return'" ]
[ "def FUNC_20(self):...\n", "VAR_3 = self.current_email()\n", "if not self.passes_email_suffix(VAR_3):\n", "VAR_40 = 'User {!r} does not have email suffix {!r}'.format(VAR_3, self.\n allowed_email_suffix)\n", "if not self.passes_email_whitelist(VAR_3):\n", "VAR_40 = 'User not in whitelist: {!r}'.format(VAR_3, self.\n allowed_email_whitelist)\n", "return True\n" ]
[ "def check_authorization(self):...\n", "email = self.current_email()\n", "if not self.passes_email_suffix(email):\n", "msg = 'User {!r} does not have email suffix {!r}'.format(email, self.\n allowed_email_suffix)\n", "if not self.passes_email_whitelist(email):\n", "msg = 'User not in whitelist: {!r}'.format(email, self.allowed_email_whitelist)\n", "return True\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Return'" ]
[ "def FUNC_13(self):...\n", "\"\"\"docstring\"\"\"\n", "VAR_2 = {'SEVERITY': {'MEDIUM': 3}, 'CONFIDENCE': {'MEDIUM': 3}}\n", "self.check_example('hardcoded-tmp.py', VAR_2)\n" ]
[ "def test_hardcoded_tmp(self):...\n", "\"\"\"docstring\"\"\"\n", "expect = {'SEVERITY': {'MEDIUM': 3}, 'CONFIDENCE': {'MEDIUM': 3}}\n", "self.check_example('hardcoded-tmp.py', expect)\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Expr'" ]
[ "def FUNC_3(self, VAR_5):...\n", "self.common = self._init_common()\n", "self._check_flags()\n", "self.common.do_setup(VAR_5)\n" ]
[ "def do_setup(self, context):...\n", "self.common = self._init_common()\n", "self._check_flags()\n", "self.common.do_setup(context)\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Expr'", "Expr'" ]
[ "def FUNC_5(VAR_13, VAR_11, VAR_10):...\n", "\"\"\"docstring\"\"\"\n", "VAR_37 = VAR_10.setdefault('alias_cache', [])\n", "if VAR_13 == VAR_11:\n", "return VAR_13\n", "if VAR_11 in VAR_37:\n", "return VAR_11\n", "VAR_37.append(VAR_11)\n", "return u'({} AS {})'.format(VAR_13, VAR_11)\n" ]
[ "def alias_expr(expr, alias, body):...\n", "\"\"\"docstring\"\"\"\n", "alias_cache = body.setdefault('alias_cache', [])\n", "if expr == alias:\n", "return expr\n", "if alias in alias_cache:\n", "return alias\n", "alias_cache.append(alias)\n", "return u'({} AS {})'.format(expr, alias)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Condition", "Return'", "Condition", "Return'", "Expr'", "Return'" ]
[ "@VAR_0.route('/signup/<remote_app>/', methods=['GET', 'POST'])...\n", "\"\"\"docstring\"\"\"\n", "if VAR_1 not in signup_handlers:\n", "return abort(404)\n", "VAR_3 = signup_handlers[VAR_1]['view']()\n", "return abort(404) if VAR_3 is None else VAR_3\n" ]
[ "@blueprint.route('/signup/<remote_app>/', methods=['GET', 'POST'])...\n", "\"\"\"docstring\"\"\"\n", "if remote_app not in signup_handlers:\n", "return abort(404)\n", "res = signup_handlers[remote_app]['view']()\n", "return abort(404) if res is None else res\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Condition", "Return'", "Assign'", "Return'" ]
[ "@VAR_2.route('/templates/<path:path>')...\n", "return send_from_directory('templates', VAR_6)\n" ]
[ "@endpoints.route('/templates/<path:path>')...\n", "return send_from_directory('templates', path)\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "@VAR_1.route('/facnonfac', methods=['GET', 'POST'])...\n", "VAR_9 = FUNC_14(1)\n", "VAR_10 = [[VAR_3[0], VAR_3[1], VAR_3[2]] for VAR_3 in VAR_9]\n", "return render_template('demo2.html', VAR_8=query1, rows=results)\n" ]
[ "@app.route('/facnonfac', methods=['GET', 'POST'])...\n", "z = demo(1)\n", "results = [[x[0], x[1], x[2]] for x in z]\n", "return render_template('demo2.html', query=query1, rows=results)\n" ]
[ 0, 0, 0, 0 ]
[ "Condition", "Assign'", "Assign'", "Return'" ]
[ "@property...\n", "if self._resource_stack:\n", "return self._resource_stack[-1]\n" ]
[ "@property...\n", "if self._resource_stack:\n", "return self._resource_stack[-1]\n" ]
[ 0, 0, 0 ]
[ "Condition", "Condition", "Return'" ]
[ "def FUNC_14(self, VAR_32, VAR_33, VAR_30):...\n", "\"\"\"docstring\"\"\"\n", "if VAR_33 & os.O_WRONLY == os.O_WRONLY or VAR_33 & os.O_RDWR == os.O_RDWR:\n", "VAR_68 = VAR_30 & ~111\n", "if VAR_33 & os.O_RDONLY == os.O_RDONLY:\n", "VAR_69 = '%s/%s_sftp_%s' % (self.cfg.get('honeypot', 'download_path'), time\n .strftime('%Y%m%d-%H%M%S'), re.sub('[^A-Za-z0-9]', '_', VAR_32))\n", "return None\n", "return None\n", "self.mkfile(VAR_32, 0, 0, 0, FUNC_30.S_IFREG | VAR_30)\n", "VAR_34 = os.open(VAR_69, VAR_33, VAR_68)\n", "self.update_realfile(self.getfile(VAR_32), VAR_69)\n", "self.tempfiles[VAR_34] = VAR_69\n", "self.filenames[VAR_34] = VAR_32\n", "return VAR_34\n" ]
[ "def open(self, filename, openFlags, mode):...\n", "\"\"\"docstring\"\"\"\n", "if openFlags & os.O_WRONLY == os.O_WRONLY or openFlags & os.O_RDWR == os.O_RDWR:\n", "hostmode = mode & ~111\n", "if openFlags & os.O_RDONLY == os.O_RDONLY:\n", "hostfile = '%s/%s_sftp_%s' % (self.cfg.get('honeypot', 'download_path'),\n time.strftime('%Y%m%d-%H%M%S'), re.sub('[^A-Za-z0-9]', '_', filename))\n", "return None\n", "return None\n", "self.mkfile(filename, 0, 0, 0, stat.S_IFREG | mode)\n", "fd = os.open(hostfile, openFlags, hostmode)\n", "self.update_realfile(self.getfile(filename), hostfile)\n", "self.tempfiles[fd] = hostfile\n", "self.filenames[fd] = filename\n", "return fd\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Assign'", "Condition", "Assign'", "Return'", "Return'", "Expr'", "Assign'", "Expr'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_5(VAR_1, *VAR_2, **VAR_3):...\n", "def FUNC_6():...\n", "if not request.method in VAR_0:\n", "return True\n", "if not request.environ.get('AUTH_TYPE') == 'cookie':\n", "return True\n", "if config.get('skip_authentication'):\n", "return True\n", "if request.environ.get('HTTP_REFERER'):\n", "VAR_7 = urlparse(request.environ.get('HTTP_REFERER'))\n", "if request.method == 'GET' and FUNC_4():\n", "VAR_8 = VAR_7.hostname\n", "return True\n", "return False\n", "if VAR_7.port:\n", "VAR_8 += ':' + str(VAR_7.port)\n", "if VAR_8.endswith(request.environ['adhocracy.domain']):\n", "if request.method != 'GET':\n", "return True\n" ]
[ "def _decorate(f, *a, **kw):...\n", "def check():...\n", "if not request.method in methods:\n", "return True\n", "if not request.environ.get('AUTH_TYPE') == 'cookie':\n", "return True\n", "if config.get('skip_authentication'):\n", "return True\n", "if request.environ.get('HTTP_REFERER'):\n", "ref_url = urlparse(request.environ.get('HTTP_REFERER'))\n", "if request.method == 'GET' and has_token():\n", "ref_host = ref_url.hostname\n", "return True\n", "return False\n", "if ref_url.port:\n", "ref_host += ':' + str(ref_url.port)\n", "if ref_host.endswith(request.environ['adhocracy.domain']):\n", "if request.method != 'GET':\n", "return True\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 5, 0, 0, 0, 0, 0, 0, 5, 0 ]
[ "FunctionDef'", "FunctionDef'", "Condition", "Return'", "Condition", "Return'", "Condition", "Return'", "Condition", "Assign'", "Condition", "Assign'", "Return'", "Return'", "Condition", "AugAssign'", "Condition", "Condition", "Return'" ]
[ "@property...\n", "return self._subworkflows.values()\n" ]
[ "@property...\n", "return self._subworkflows.values()\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "@classmethod...\n", "return [FUNC_27, FUNC_28]\n" ]
[ "@classmethod...\n", "return [not_applicable, applicable]\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "def FUNC_5(VAR_0, VAR_7):...\n", "\"\"\"docstring\"\"\"\n", "VAR_6.append(CLASS_1(VAR_0, VAR_7))\n" ]
[ "def add_menu(name, href):...\n", "\"\"\"docstring\"\"\"\n", "_MENU_ITEMS.append(_MenuItem(name, href))\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Expr'" ]
[ "def FUNC_8(self):...\n", "self.mock(time, 'time', lambda : 126.0)\n", "def FUNC_22(VAR_14, VAR_15):...\n", "VAR_37 = VAR_14.state['sleep_streak']\n", "self.assertEqual(VAR_14.remote, self.server)\n", "if VAR_37 == 5:\n", "return False\n" ]
[ "def test_run_bot(self):...\n", "self.mock(time, 'time', lambda : 126.0)\n", "def poll_server(botobj, _):...\n", "sleep_streak = botobj.state['sleep_streak']\n", "self.assertEqual(botobj.remote, self.server)\n", "if sleep_streak == 5:\n", "return False\n" ]
[ 0, 0, 0, 0, 5, 0, 0 ]
[ "FunctionDef'", "Expr'", "FunctionDef'", "Assign'", "Expr'", "Condition", "Return'" ]
[ "def FUNC_24(VAR_40):...\n", "for VAR_18 in VAR_40.sub_items:\n", "if VAR_18.parent:\n", "return\n", "VAR_18.parent = VAR_40\n", "FUNC_24(VAR_18)\n" ]
[ "def iterate_fsnode(node):...\n", "for item in node.sub_items:\n", "if item.parent:\n", "return\n", "item.parent = node\n", "iterate_fsnode(item)\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "For", "Condition", "Return'", "Assign'", "Expr'" ]
[ "def FUNC_16(self):...\n", "VAR_25 = c.get('/')\n", "self.assertEqual(VAR_25.status_code, 302)\n", "self.assertTrue(VAR_25.location.startswith(self.app.config[\n 'TOKEN_SERVICE_URL']))\n" ]
[ "def test_no_cookie(self):...\n", "resp = c.get('/')\n", "self.assertEqual(resp.status_code, 302)\n", "self.assertTrue(resp.location.startswith(self.app.config['TOKEN_SERVICE_URL']))\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Expr'", "Expr'" ]
[ "def FUNC_1(VAR_6, VAR_2, VAR_3):...\n", "\"\"\"docstring\"\"\"\n", "VAR_9 = [f\"{'VULNERABLE_STRING'!r}\"]\n", "VAR_6 = VAR_4.get(VAR_0, VAR_1=sqli)\n", "VAR_10 = 1\n", "VAR_11 = VAR_2.index('NULL')\n", "while not VAR_6.ok:\n", "VAR_7 = VAR_2.index('NULL', VAR_11)\n", "print(f'Column {VAR_10} contains inserted text')\n", "VAR_2[VAR_7] = VAR_9[0]\n", "return VAR_7\n", "VAR_3['category'] = f\"Lifestyle{' '.join(VAR_2)}\"\n", "VAR_6 = VAR_4.get(VAR_0, VAR_1=sqli)\n", "if not VAR_6.ok:\n", "VAR_2[VAR_7] = 'NULL'\n", "VAR_11 = VAR_7 + 1\n", "VAR_10 += 1\n" ]
[ "def sqli_union_lab_2(lab2, null, sqli):...\n", "\"\"\"docstring\"\"\"\n", "secret_value = [f\"{'VULNERABLE_STRING'!r}\"]\n", "lab2 = api_session.get(url, params=sqli)\n", "column = 1\n", "step = null.index('NULL')\n", "while not lab2.ok:\n", "index = null.index('NULL', step)\n", "print(f'Column {column} contains inserted text')\n", "null[index] = secret_value[0]\n", "return index\n", "sqli['category'] = f\"Lifestyle{' '.join(null)}\"\n", "lab2 = api_session.get(url, params=sqli)\n", "if not lab2.ok:\n", "null[index] = 'NULL'\n", "step = index + 1\n", "column += 1\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Expr'", "Assign'", "Return'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "AugAssign'" ]
[ "@VAR_0.route('/messages')...\n", "if request.method == 'GET':\n", "return db_getMessage(VAR_1)\n" ]
[ "@messages_module.route('/messages')...\n", "if request.method == 'GET':\n", "return db_getMessage(dialog_id)\n" ]
[ 0, 0, 0 ]
[ "Condition", "Condition", "Return'" ]
[ "from osv import osv\n", "from tools.translate import _\n", "VAR_0 = 'pos.close.statement'\n", "VAR_1 = 'Close Statements'\n", "def FUNC_0(self, VAR_2, VAR_3, VAR_4, VAR_5):...\n", "\"\"\"docstring\"\"\"\n", "VAR_6 = self.pool.get('res.users').browse(VAR_2, VAR_3, VAR_3).company_id.id\n", "VAR_7 = []\n", "VAR_8 = self.pool.get('ir.model.data')\n", "VAR_9 = self.pool.get('account.bank.statement')\n", "VAR_10 = self.pool.get('account.journal')\n", "VAR_2.execute(\n 'select DISTINCT journal_id from pos_journal_users where user_id=%d order by journal_id'\n % VAR_3)\n", "VAR_11 = map(lambda x1: x1[0], VAR_2.fetchall())\n", "VAR_2.execute('string' % ','.join(map(lambda x: \"'\" + str(x) + \"'\", VAR_11)))\n", "VAR_12 = map(lambda x1: x1[0], VAR_2.fetchall())\n", "for journal in VAR_10.browse(VAR_2, VAR_3, VAR_12):\n", "VAR_4 = VAR_9.search(VAR_2, VAR_3, [('state', '!=', 'confirm'), ('user_id',\n '=', VAR_3), ('journal_id', '=', journal.id)])\n", "VAR_13 = self.pool.get('ir.model.data')\n", "if not VAR_4:\n", "VAR_14 = VAR_13._get_id(VAR_2, VAR_3, 'account', 'view_bank_statement_tree')\n", "VAR_7.append(VAR_4[0])\n", "VAR_15 = VAR_13._get_id(VAR_2, VAR_3, 'account', 'view_bank_statement_form2')\n", "if not journal.check_dtls:\n", "if VAR_14:\n", "VAR_9.button_confirm_cash(VAR_2, VAR_3, VAR_4, VAR_5)\n", "VAR_14 = VAR_13.browse(VAR_2, VAR_3, VAR_14, VAR_5=context).res_id\n", "if VAR_15:\n", "VAR_15 = VAR_13.browse(VAR_2, VAR_3, VAR_15, VAR_5=context).res_id\n", "return {'domain': \"[('id','in',\" + str(VAR_7) + ')]', 'name':\n 'Close Statements', 'view_type': 'form', 'view_mode': 'tree,form',\n 'res_model': 'account.bank.statement', 'views': [(VAR_14, 'tree'), (\n VAR_15, 'form')], 'type': 'ir.actions.act_window'}\n" ]
[ "from osv import osv\n", "from tools.translate import _\n", "_name = 'pos.close.statement'\n", "_description = 'Close Statements'\n", "def close_statement(self, cr, uid, ids, context):...\n", "\"\"\"docstring\"\"\"\n", "company_id = self.pool.get('res.users').browse(cr, uid, uid).company_id.id\n", "list_statement = []\n", "mod_obj = self.pool.get('ir.model.data')\n", "statement_obj = self.pool.get('account.bank.statement')\n", "journal_obj = self.pool.get('account.journal')\n", "cr.execute(\n 'select DISTINCT journal_id from pos_journal_users where user_id=%d order by journal_id'\n % uid)\n", "j_ids = map(lambda x1: x1[0], cr.fetchall())\n", "cr.execute(\n \"\"\" select id from account_journal\n where auto_cash='True' and type='cash'\n and id in (%s)\"\"\"\n % ','.join(map(lambda x: \"'\" + str(x) + \"'\", j_ids)))\n", "journal_ids = map(lambda x1: x1[0], cr.fetchall())\n", "for journal in journal_obj.browse(cr, uid, journal_ids):\n", "ids = statement_obj.search(cr, uid, [('state', '!=', 'confirm'), ('user_id',\n '=', uid), ('journal_id', '=', journal.id)])\n", "data_obj = self.pool.get('ir.model.data')\n", "if not ids:\n", "id2 = data_obj._get_id(cr, uid, 'account', 'view_bank_statement_tree')\n", "list_statement.append(ids[0])\n", "id3 = data_obj._get_id(cr, uid, 'account', 'view_bank_statement_form2')\n", "if not journal.check_dtls:\n", "if id2:\n", "statement_obj.button_confirm_cash(cr, uid, ids, context)\n", "id2 = data_obj.browse(cr, uid, id2, context=context).res_id\n", "if id3:\n", "id3 = data_obj.browse(cr, uid, id3, context=context).res_id\n", "return {'domain': \"[('id','in',\" + str(list_statement) + ')]', 'name':\n 'Close Statements', 'view_type': 'form', 'view_mode': 'tree,form',\n 'res_model': 'account.bank.statement', 'views': [(id2, 'tree'), (id3,\n 'form')], 'type': 'ir.actions.act_window'}\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 4, 4, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "ImportFrom'", "ImportFrom'", "Assign'", "Assign'", "FunctionDef'", "Docstring", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "Expr'", "Assign'", "For", "Assign'", "Assign'", "Condition", "Assign'", "Expr'", "Assign'", "Condition", "Condition", "Expr'", "Assign'", "Condition", "Assign'", "Return'" ]
[ "def FUNC_6(self):...\n", "\"\"\"docstring\"\"\"\n", "if not self.issue_gcs_flushes:\n", "return\n", "if self.gcs_flush_policy is None:\n", "VAR_45 = self.redis.get('gcs_flushing_policy')\n", "if not self.gcs_flush_policy.should_flush(self.redis_shard):\n", "if VAR_45 is None:\n", "return\n", "VAR_29 = self.gcs_flush_policy.num_entries_to_flush()\n", "return\n", "self.gcs_flush_policy = pickle.loads(VAR_45)\n", "VAR_30 = self.redis_shard.execute_command('HEAD.FLUSH {}'.format(VAR_29))\n", "VAR_0.info('Monitor: num_flushed {}'.format(VAR_30))\n", "ray.experimental.flush_redis_unsafe(self.redis)\n", "self.gcs_flush_policy.record_flush()\n" ]
[ "def _maybe_flush_gcs(self):...\n", "\"\"\"docstring\"\"\"\n", "if not self.issue_gcs_flushes:\n", "return\n", "if self.gcs_flush_policy is None:\n", "serialized = self.redis.get('gcs_flushing_policy')\n", "if not self.gcs_flush_policy.should_flush(self.redis_shard):\n", "if serialized is None:\n", "return\n", "max_entries_to_flush = self.gcs_flush_policy.num_entries_to_flush()\n", "return\n", "self.gcs_flush_policy = pickle.loads(serialized)\n", "num_flushed = self.redis_shard.execute_command('HEAD.FLUSH {}'.format(\n max_entries_to_flush))\n", "logger.info('Monitor: num_flushed {}'.format(num_flushed))\n", "ray.experimental.flush_redis_unsafe(self.redis)\n", "self.gcs_flush_policy.record_flush()\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Return'", "Condition", "Assign'", "Condition", "Condition", "Return'", "Assign'", "Return'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'" ]
[ "def FUNC_30(self):...\n", "\"\"\"docstring\"\"\"\n", "if 'saml_data' in VAR_15:\n", "return VAR_15['saml_data']['nameid']\n" ]
[ "def _current_user_nameid(self):...\n", "\"\"\"docstring\"\"\"\n", "if 'saml_data' in session:\n", "return session['saml_data']['nameid']\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Return'" ]
[ "@VAR_4...\n", "self.print_prompt()\n" ]
[ "@deferred...\n", "self.print_prompt()\n" ]
[ 0, 0 ]
[ "Condition", "Expr'" ]
[ "@VAR_2.route('/delete_comment/<pic_name>', methods=['POST', 'GET'])...\n", "if request.method == 'POST':\n", "VAR_4 = (\"DELETE FROM commentsTable\\t\\t\\t\\tWHERE comment_id = '%s'\" %\n request.form['delete_comment'])\n", "VAR_1.execute(VAR_4)\n", "VAR_0.commit()\n", "return redirect(url_for('view_comment', VAR_3=pic_name))\n" ]
[ "@app.route('/delete_comment/<pic_name>', methods=['POST', 'GET'])...\n", "if request.method == 'POST':\n", "query = (\"DELETE FROM commentsTable\\t\\t\\t\\tWHERE comment_id = '%s'\" %\n request.form['delete_comment'])\n", "cursor.execute(query)\n", "db.commit()\n", "return redirect(url_for('view_comment', pic_name=pic_name))\n" ]
[ 4, 0, 4, 0, 0, 0 ]
[ "Condition", "Condition", "Assign'", "Expr'", "Expr'", "Return'" ]
[ "def FUNC_2(self, VAR_3, VAR_4):...\n", "return self.has_object_permission(VAR_3, VAR_4, VAR_4.instance)\n" ]
[ "def has_permission(self, request, view):...\n", "return self.has_object_permission(request, view, view.instance)\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_11(self):...\n", "VAR_9 = 'foo'\n", "VAR_7 = BokChoyTestSuite('', VAR_9=test_dir)\n", "self.assertEqual(VAR_7.cmd, self._expected_command(VAR_1=test_dir))\n" ]
[ "def test_test_dir(self):...\n", "test_dir = 'foo'\n", "suite = BokChoyTestSuite('', test_dir=test_dir)\n", "self.assertEqual(suite.cmd, self._expected_command(name=test_dir))\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Expr'" ]
[ "def FUNC_39(VAR_62):...\n", "return json.loads(VAR_62) if VAR_21 else VAR_62\n" ]
[ "def get_str(_str):...\n", "return json.loads(_str) if as_json else _str\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_3(VAR_13, VAR_14):...\n", "VAR_28.info('writing %s', VAR_13)\n", "if os.path.exists(VAR_13):\n", "VAR_41 = os.stat(VAR_13)\n", "VAR_42, VAR_43, VAR_44 = 420, -1, -1\n", "VAR_42, VAR_43, VAR_44 = VAR_41.st_mode, VAR_41.st_uid, VAR_41.st_gid\n", "VAR_32 = os.path.dirname(VAR_13)\n", "os.path.exists(VAR_32) or os.makedirs(VAR_32)\n", "newfile.write(VAR_14)\n", "os.chmod(newfile.name, VAR_42)\n", "os.chown(newfile.name, VAR_43, VAR_44)\n", "os.rename(newfile.name, VAR_13)\n" ]
[ "def write_file(path, contents):...\n", "logger.info('writing %s', path)\n", "if os.path.exists(path):\n", "stat = os.stat(path)\n", "mode, uid, gid = 420, -1, -1\n", "mode, uid, gid = stat.st_mode, stat.st_uid, stat.st_gid\n", "d = os.path.dirname(path)\n", "os.path.exists(d) or os.makedirs(d)\n", "newfile.write(contents)\n", "os.chmod(newfile.name, mode)\n", "os.chown(newfile.name, uid, gid)\n", "os.rename(newfile.name, path)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'" ]
[ "def FUNC_0(VAR_0):...\n", "VAR_12 = ['select', 'delete', 'drop', 'update', 'case', 'and', 'or', 'like']\n", "def FUNC_6():...\n", "VAR_13.throw(_('Invalid Search Field'), VAR_13.DataError)\n", "if len(VAR_0) >= 3:\n", "if '=' in VAR_0:\n", "FUNC_6()\n", "if ' --' in VAR_0:\n", "FUNC_6()\n", "if any(' {0} '.format(keyword) in VAR_0.split() for keyword in VAR_12):\n", "FUNC_6()\n", "if any(keyword in VAR_0.split() for keyword in VAR_12):\n", "FUNC_6()\n" ]
[ "def sanitize_searchfield(searchfield):...\n", "blacklisted_keywords = ['select', 'delete', 'drop', 'update', 'case', 'and',\n 'or', 'like']\n", "def _raise_exception():...\n", "frappe.throw(_('Invalid Search Field'), frappe.DataError)\n", "if len(searchfield) >= 3:\n", "if '=' in searchfield:\n", "_raise_exception()\n", "if ' --' in searchfield:\n", "_raise_exception()\n", "if any(' {0} '.format(keyword) in searchfield.split() for keyword in\n", "_raise_exception()\n", "if any(keyword in searchfield.split() for keyword in blacklisted_keywords):\n", "_raise_exception()\n" ]
[ 0, 0, 0, 4, 0, 0, 4, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "FunctionDef'", "Expr'", "Condition", "Condition", "Expr'", "Condition", "Expr'", "For", "Expr'", "For", "Expr'" ]
[ "def FUNC_14(self):...\n", "self._test_health(FUNC_0, cluster.ClusterHealth.GREEN)\n" ]
[ "def test_green_health(self):...\n", "self._test_health(_validate_conn_up, cluster.ClusterHealth.GREEN)\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Expr'" ]
[ "def FUNC_0():...\n", "VAR_0 = CLASS_0()\n", "VAR_0.run()\n" ]
[ "def main():...\n", "sh = Shell()\n", "sh.run()\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Expr'" ]
[ "def FUNC_6(self, *VAR_4, **VAR_5):...\n", "VAR_11 = self.REQUEST\n", "self._zmi_page_request()\n", "if not VAR_11.get('lang'):\n", "VAR_11.set('lang', self.getLanguage(VAR_11))\n", "if not VAR_11.get('manage_lang') in self.getLocale().get_manage_langs():\n", "VAR_11.set('manage_lang', self.get_manage_lang())\n" ]
[ "def f_standard_html_request(self, *args, **kwargs):...\n", "request = self.REQUEST\n", "self._zmi_page_request()\n", "if not request.get('lang'):\n", "request.set('lang', self.getLanguage(request))\n", "if not request.get('manage_lang') in self.getLocale().get_manage_langs():\n", "request.set('manage_lang', self.get_manage_lang())\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Expr'", "Condition", "Expr'", "Condition", "Expr'" ]
[ "def FUNC_7(VAR_9, VAR_10):...\n", "\"\"\"docstring\"\"\"\n", "VAR_3 = HttpRequest()\n", "VAR_3.method = 'GET'\n", "VAR_3.session = VAR_10\n", "VAR_26, VAR_27, VAR_28 = resolve(VAR_9)\n", "VAR_29 = VAR_26(VAR_3, *VAR_27, **kwargs)\n", "return VAR_29.content\n" ]
[ "def _local_server_get(url, session):...\n", "\"\"\"docstring\"\"\"\n", "request = HttpRequest()\n", "request.method = 'GET'\n", "request.session = session\n", "view, args, kwargs = resolve(url)\n", "response = view(request, *args, **kwargs)\n", "return response.content\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_14(self, VAR_17):...\n", "VAR_21 = fs.split_path(VAR_17.path)\n", "VAR_22 = len(VAR_21)\n", "VAR_23 = min(self._lensplitpath, VAR_22)\n", "VAR_25 = len(set(VAR_17.parameters.keys()) - self._targetparam)\n", "return self._splitpath[:VAR_23] == VAR_21[:VAR_23] and VAR_25 < 2\n" ]
[ "def does_intersect_path(self, pathctx):...\n", "testpath = fs.split_path(pathctx.path)\n", "lentestpath = len(testpath)\n", "lenpath = min(self._lensplitpath, lentestpath)\n", "extra_count = len(set(pathctx.parameters.keys()) - self._targetparam)\n", "return self._splitpath[:lenpath] == testpath[:lenpath] and extra_count < 2\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_6(self):...\n", "return True\n" ]
[ "def Done(self):...\n", "return True\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_34(self):...\n", "return self.replace(self.dynamic_fill, '{*}')\n" ]
[ "def format_dynamic(self):...\n", "return self.replace(self.dynamic_fill, '{*}')\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_42(self):...\n", "VAR_54 = 0\n", "for VAR_40, VAR_41 in sorted(self._names.items(), VAR_44=lambda item: item[\n", "VAR_57, VAR_42 = VAR_41\n", "for item in self[VAR_54:]:\n", "if VAR_42 is None:\n", "yield None, item\n", "VAR_42 = VAR_57 + 1\n", "if VAR_57 > VAR_54:\n", "for item in self[VAR_54:VAR_57]:\n", "yield VAR_40, getattr(self, VAR_40)\n", "yield None, item\n", "VAR_54 = VAR_42\n" ]
[ "def allitems(self):...\n", "next = 0\n", "for name, index in sorted(self._names.items(), key=lambda item: item[1][0]):\n", "start, end = index\n", "for item in self[next:]:\n", "if end is None:\n", "yield None, item\n", "end = start + 1\n", "if start > next:\n", "for item in self[next:start]:\n", "yield name, getattr(self, name)\n", "yield None, item\n", "next = end\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "For", "Assign'", "For", "Condition", "Expr'", "Assign'", "Condition", "For", "Expr'", "Expr'", "Assign'" ]
[ "def FUNC_36(VAR_16, VAR_22, VAR_23, VAR_29):...\n", "VAR_53 = 'string'.format(VAR_22, VAR_23, VAR_29)\n", "VAR_54 = VAR_16.exec(VAR_53)\n", "VAR_18 = [VAR_41[0] for VAR_41 in VAR_54]\n", "return VAR_18, VAR_23\n" ]
[ "def get_n_tournaments_after_date(db, scene, date, limit):...\n", "sql = (\n \"select url, date from matches where scene='{}' and date>='{}' group by url, date order by date desc limit {};\"\n .format(scene, date, limit))\n", "res = db.exec(sql)\n", "urls = [r[0] for r in res]\n", "return urls, date\n" ]
[ 0, 4, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_35(VAR_29):...\n", "VAR_40 = VAR_29.group('name')\n", "VAR_16 = VAR_11[VAR_40]\n", "if VAR_15:\n", "if VAR_13 and VAR_16 == VAR_14:\n", "return '{{{}}}'.format(VAR_40)\n", "if VAR_12:\n", "return str(VAR_16)\n", "return VAR_14\n" ]
[ "def format_match(match):...\n", "name = match.group('name')\n", "value = wildcards[name]\n", "if keep_dynamic:\n", "if fail_dynamic and value == dynamic_fill:\n", "return '{{{}}}'.format(name)\n", "if fill_missing:\n", "return str(value)\n", "return dynamic_fill\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Condition", "Condition", "Return'", "Condition", "Return'", "Return'" ]
[ "def FUNC_4(self):...\n", "for clause in self._ruleorder:\n", "for rulename in clause:\n", "if not self.is_rule(rulename):\n" ]
[ "def check(self):...\n", "for clause in self._ruleorder:\n", "for rulename in clause:\n", "if not self.is_rule(rulename):\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "For", "For", "Condition" ]
[ "@property...\n", "\"\"\"docstring\"\"\"\n", "return self.config.get(self.section, 'dashboard_link')\n" ]
[ "@property...\n", "\"\"\"docstring\"\"\"\n", "return self.config.get(self.section, 'dashboard_link')\n" ]
[ 0, 0, 0 ]
[ "Condition", "Docstring", "Return'" ]
[ "@classmethod...\n", "return '-S-Xfatal-warnings', '-C-Werror'\n" ]
[ "@classmethod...\n", "return '-S-Xfatal-warnings', '-C-Werror'\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "def FUNC_17(self, VAR_15):...\n", "\"\"\"docstring\"\"\"\n", "return LooseVersion(self.sos_info['version']) >= VAR_15\n" ]
[ "def check_sos_version(self, ver):...\n", "\"\"\"docstring\"\"\"\n", "return LooseVersion(self.sos_info['version']) >= ver\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Return'" ]
[ "def __iter__(self):...\n", "" ]
[ "def __iter__(self):...\n", "" ]
[ 0, 0 ]
[ "FunctionDef'", "Condition" ]
[ "def FUNC_3(self):...\n", "self._mox.ReplayAll()\n", "self.assertEqual(url_helper.UrlOpen('url', wait_duration=-1), None)\n", "self._mox.VerifyAll()\n" ]
[ "def testUrlOpenInvalidWaitDuration(self):...\n", "self._mox.ReplayAll()\n", "self.assertEqual(url_helper.UrlOpen('url', wait_duration=-1), None)\n", "self._mox.VerifyAll()\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Expr'", "Expr'" ]
[ "@utils.synchronized('3par', external=True)...\n", "self.common.client_login()\n", "self.common.delete_volume(VAR_7)\n", "self.common.client_logout()\n" ]
[ "@utils.synchronized('3par', external=True)...\n", "self.common.client_login()\n", "self.common.delete_volume(volume)\n", "self.common.client_logout()\n" ]
[ 0, 0, 0, 0 ]
[ "Condition", "Expr'", "Expr'", "Expr'" ]
[ "def FUNC_21(VAR_55):...\n", "@wraps(VAR_55)...\n", "def FUNC_27(VAR_6):...\n", "if callable(VAR_6):\n", "return VAR_6()\n", "if request.method == 'POST':\n", "return VAR_55(*VAR_69, **kwargs)\n", "VAR_10 = json.loads(request.data)\n", "VAR_54 = {'error': {'type': 'schema', 'message': str(e)}, 'schema': VAR_24}\n", "schemas.validate(VAR_10, VAR_24)\n", "return json.dumps(VAR_54, sort_keys=True, indent=4, default=default_encode\n ), 400, {'Content-Type': 'application/json'}\n", "VAR_70['validated_body'] = VAR_10\n", "if VAR_70.get('timer'):\n", "VAR_70['timer'].mark('validate_schema')\n" ]
[ "def decorator(func):...\n", "@wraps(func)...\n", "def default_encode(value):...\n", "if callable(value):\n", "return value()\n", "if request.method == 'POST':\n", "return func(*args, **kwargs)\n", "body = json.loads(request.data)\n", "result = {'error': {'type': 'schema', 'message': str(e)}, 'schema': schema}\n", "schemas.validate(body, schema)\n", "return json.dumps(result, sort_keys=True, indent=4, default=default_encode\n ), 400, {'Content-Type': 'application/json'}\n", "kwargs['validated_body'] = body\n", "if kwargs.get('timer'):\n", "kwargs['timer'].mark('validate_schema')\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "FunctionDef'", "Condition", "Return'", "Condition", "Return'", "Assign'", "Assign'", "Expr'", "Return'", "Assign'", "Condition", "Expr'" ]
[ "def FUNC_36():...\n", "yield os.path.join(self.home, 'lib')\n", "if self.jdk:\n", "yield os.path.join(self.home, 'jre', 'lib')\n" ]
[ "def lib_paths():...\n", "yield os.path.join(self.home, 'lib')\n", "if self.jdk:\n", "yield os.path.join(self.home, 'jre', 'lib')\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Condition", "Expr'" ]
[ "def FUNC_2(self):...\n", "self.events.append('on_exit: %s' % self.path)\n" ]
[ "def on_exit(self):...\n", "self.events.append('on_exit: %s' % self.path)\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Expr'" ]
[ "import unittest\n", "import pandas as pd\n", "import numpy as np\n", "import datetime\n", "from ddt import ddt, data, unpack\n", "from scraper import scrape\n", "from wrangler import clean_job_number, clean_pub_date, clean_city, clean_company_name, get_acronyms, get_street_number, get_street_name, clean_title, wrangle\n", "from matcher import match\n", "from communicator import communicate\n", "from inbox_scanner import process_as_form, process_as_reply\n", "from ml import build_train_set, train_model, validate_model\n", "from db_tools import create_connection\n", "from test.test_setup import create_test_db\n", "import os\n", "VAR_0 = 0.7\n", "@data((' ', ''), ('\\n #2404\\n', '2404'), ('no. 2404', '2404'), ('# 2404',...\n", "VAR_12 = clean_job_number(VAR_1)\n", "self.assertEqual(VAR_2, VAR_12)\n", "@data((' ', ''), ('\\n 2019-02-20\\n', '2019-02-20'), ('2019-02-20',...\n", "VAR_12 = clean_pub_date(VAR_1)\n", "self.assertEqual(VAR_2, VAR_12)\n", "@data((' ', ''), (...\n", "VAR_12 = clean_city(VAR_1)\n", "self.assertEqual(VAR_2, VAR_12)\n", "@data((' ', ''), ('Frecon', 'frecon'), ('Frecon', 'frecon'), (...\n", "VAR_12 = clean_company_name(VAR_1)\n", "self.assertEqual(VAR_2, VAR_12)\n", "@data((' ', []), ('Ron Eastern Construction Ltd. (RECL)', ['RECL']), (...\n", "VAR_12 = get_acronyms(VAR_1)\n", "self.assertEqual(VAR_2, VAR_12)\n", "VAR_3 = ('123 Fake St.', '123', 'fake'), ('12 Carrière Rd', '12', 'carriere'\n ), ('8-1230 marenger street', '1230', 'marenger'), (\n 'apt. 8, 1230 marenger street', '1230', 'marenger'), (\n '8-1230 marenger street, apt. 8, ', '1230', 'marenger'), (\n '1230 apt. 8, marenger street', '1230', 'marenger'), (\n '1010 talbot st. unit #1', '1010', 'talbot'), ('6250 st albans court',\n '6250', 'albans'), ('6250 saint albans court', '6250', 'albans'), (\n '6250 st. albans', '6250', 'albans'), ('6250 st-albans CRT', '6250',\n 'albans'), ('string', '451', 'smyth'), ('145 Jean-Jacques Lussier',\n '145', 'jean-jacques'), ('Edwardsburgh/Cardinal', '', '')\n", "@data(*VAR_3)...\n", "VAR_12 = get_street_number(VAR_1)\n", "self.assertEqual(VAR_4, VAR_12)\n", "@data(*VAR_3)...\n", "VAR_12 = get_street_name(VAR_1)\n", "self.assertEqual(VAR_5, VAR_12)\n", "@data((' ', ''), ('test', 'test'), ('testé', 'teste'))...\n", "VAR_12 = clean_title(VAR_1)\n", "self.assertEqual(VAR_2, VAR_12)\n", "def FUNC_8():...\n", "for filename in ['cert_db.sqlite3', 'rf_model.pkl', 'rf_features.pkl']:\n", "create_test_db()\n", "os.rename(filename, 'temp_' + filename)\n", "for filename in ['cert_db.sqlite3', 'rf_model.pkl', 'rf_features.pkl']:\n", "def FUNC_9():...\n", "os.rename('test_' + filename, filename)\n", "for filename in ['cert_db.sqlite3', 'rf_model.pkl', 'rf_features.pkl']:\n", "@data(('9981', 'B0046A36-3F1C-11E9-9A87-005056AA6F11', 0, 0, 0), ('9982',...\n", "os.rename('temp_' + filename, filename)\n", "os.remove(filename)\n", "VAR_13 = {'sender': 'Alex Roy <[email protected]>', 'subject':\n 'DO NOT MODIFY MESSAGE BELOW - JUST HIT `SEND`', 'date':\n 'Tue, 7 May 2019 17:34:17 +0000', 'content':\n f'job_number={VAR_6}string{VAR_7}\\r\\n'}\n", "VAR_14 = 'string'\n", "VAR_15 = 'string'\n", "if VAR_9 or VAR_10:\n", "conn.cursor().execute(VAR_14.format(VAR_6, VAR_9))\n", "if VAR_8:\n", "if VAR_9:\n", "VAR_30 = pd.read_sql(f'SELECT * FROM df_dilfo WHERE job_number={VAR_6}', conn)\n", "conn.cursor().execute(VAR_15.format(VAR_6, 1))\n", "conn.cursor().execute(VAR_15.format(VAR_6, 0))\n", "VAR_31 = pd.read_sql(f'SELECT * FROM df_matched WHERE job_number={VAR_6}', conn\n )\n", "process_as_form(VAR_13)\n", "VAR_32 = pd.read_sql(f'SELECT * FROM df_dilfo WHERE job_number={VAR_6}', conn)\n", "VAR_33 = pd.read_sql(f'SELECT * FROM df_matched WHERE job_number={VAR_6}', conn\n )\n", "self.assertEqual(len(VAR_32), 1)\n", "self.assertEqual(bool(VAR_32.iloc[0].closed), bool(VAR_9 or VAR_7))\n", "self.assertEqual(any(VAR_33.ground_truth), bool(VAR_9 or VAR_7))\n", "self.assertEqual(len(VAR_31) + bool(VAR_7 and not VAR_9), len(VAR_33))\n", "self.assertEqual(list(VAR_31.columns), list(VAR_33.columns))\n", "self.assertEqual(list(VAR_30.columns), list(VAR_32.columns))\n", "@data(('9991', 'B0046A36-3F1C-11E9-9A87-005056AA6F01', 0, 0, 0), ('9992',...\n", "VAR_13 = {'sender': 'Alex Roy <[email protected]>', 'subject':\n f'Re: [EXTERNAL] Upcoming Holdback Release: #{VAR_6}', 'date':\n 'Thu, 30 May 2019 00:41:05 +0000', 'content':\n f'{VAR_11}string{VAR_6}string{VAR_6}string{VAR_7}string'}\n", "VAR_14 = \"\"\"\n INSERT INTO df_dilfo (job_number, closed)\n VALUES ({}, {})\n \"\"\"\n", "VAR_15 = \"\"\"\n INSERT INTO df_matched (job_number, ground_truth)\n VALUES ({}, {})\n \"\"\"\n", "conn.cursor().execute(VAR_14.format(VAR_6, VAR_9))\n", "if VAR_8:\n", "if VAR_9:\n", "VAR_30 = pd.read_sql(f'SELECT * FROM df_dilfo WHERE job_number={VAR_6}', conn)\n", "conn.cursor().execute(VAR_15.format(VAR_6, 1))\n", "conn.cursor().execute(VAR_15.format(VAR_6, 0))\n", "VAR_31 = pd.read_sql(f'SELECT * FROM df_matched WHERE job_number={VAR_6}', conn\n )\n", "process_as_reply(VAR_13)\n", "VAR_32 = pd.read_sql(f'SELECT * FROM df_dilfo WHERE job_number={VAR_6}', conn)\n", "VAR_33 = pd.read_sql(f'SELECT * FROM df_matched WHERE job_number={VAR_6}', conn\n )\n", "self.assertEqual(len(VAR_30), len(VAR_32))\n", "self.assertEqual(VAR_32.iloc[0].closed, VAR_9 or VAR_11)\n", "self.assertEqual(any(VAR_33.ground_truth), VAR_9 or VAR_11)\n", "self.assertEqual(len(VAR_31) + (not VAR_9), len(VAR_33))\n", "self.assertEqual(list(VAR_31.columns), list(VAR_33.columns))\n", "self.assertEqual(list(VAR_30.columns), list(VAR_32.columns))\n", "def FUNC_12(self):...\n", "for filename in ['cert_db.sqlite3', 'rf_model.pkl', 'rf_features.pkl']:\n", "create_test_db()\n", "os.rename(filename, 'temp_' + filename)\n", "for filename in ['cert_db.sqlite3', 'rf_model.pkl', 'rf_features.pkl']:\n", "def FUNC_13(self):...\n", "os.rename('test_' + filename, filename)\n", "for filename in ['cert_db.sqlite3', 'rf_model.pkl', 'rf_features.pkl']:\n", "def FUNC_14(self):...\n", "os.rename('temp_' + filename, filename)\n", "os.remove(filename)\n", "VAR_16 = 3\n", "VAR_17 = scrape(limit=test_limit, test=True, since='week_ago')\n", "self.assertEqual(len(VAR_17), VAR_16)\n", "VAR_18 = 'SELECT * FROM df_dilfo WHERE closed=0 LIMIT 1'\n", "VAR_34 = pd.read_sql(VAR_18, conn).iloc[0]\n", "communicate(VAR_17, VAR_34, test=True)\n", "def FUNC_15(self):...\n", "build_train_set()\n", "train_model(VAR_0=prob_thresh)\n", "VAR_19 = 'string'\n", "VAR_35 = pd.read_sql(VAR_19, conn)\n", "VAR_20 = scrape(ref=test_df_dilfo)\n", "VAR_21 = match(df_dilfo=test_df_dilfo, df_web=test_web_df, test=True, VAR_0\n =prob_thresh, version='new')\n", "VAR_22 = int(len(VAR_21) ** 0.5)\n", "VAR_23 = VAR_21[VAR_21.pred_match == 1].title.nunique()\n", "self.assertTrue(VAR_23 == VAR_22, msg=\n f'qty_found_matches({qty_found_matches}) not equal qty_actual_matches({qty_actual_matches})'\n )\n", "VAR_24 = len(VAR_21[VAR_21.pred_match == 1]) - VAR_23\n", "self.assertTrue(VAR_24 <= round(VAR_22 * 0.25, 1), msg=\n f'found too many false positives ({false_positives}) out of total test projects ({qty_actual_matches})'\n )\n", "VAR_25 = pd.DataFrame({'job_number': '2387', 'city': 'Ottawa', 'address':\n '2562 Del Zotto Ave., Ottawa, Ontario', 'title':\n 'DWS Building Expansion', 'owner': 'Douglas Stalker', 'contractor':\n 'GNC', 'engineer': 'Goodkey', 'receiver_email': '[email protected]',\n 'cc_email': '', 'quality': '2', 'closed': '0'}, index=range(1))\n", "VAR_26 = pd.DataFrame({'pub_date': '2019-03-06', 'city': 'Ottawa-Carleton',\n 'address': '2562 Del Zotto Avenue, Gloucester, Ontario', 'title':\n 'Construct a 1 storey storage addition to a 2 storey office/industrial building'\n , 'owner': 'Doug Stalker, DWS Roofing', 'contractor':\n 'GNC Constructors Inc.', 'engineer': None, 'dcn_key':\n 'B0046A36-3F1C-11E9-9A87-005056AA6F02'}, index=range(1))\n", "VAR_27, VAR_28 = match(df_dilfo=sample_dilfo, df_web=sample_web, test=True,\n version='new').iloc[0][['pred_match', 'pred_prob']]\n", "self.assertTrue(VAR_27, msg=\n f'Project #{sample_dilfo.job_number} did not match successfully. Match probability returned was {prob}.'\n )\n", "VAR_21 = match(df_dilfo=sample_dilfo, since='2019-03-05', until=\n '2019-03-07', test=True, version='new')\n", "VAR_29 = VAR_21[VAR_21.contractor == 'gnc'].iloc[0].pred_prob\n", "self.assertTrue(round(VAR_28, 2) == round(VAR_29, 2))\n", "validate_model(VAR_0=prob_thresh, test=True)\n", "if __name__ == '__main__':\n", "for filename in ['cert_db.sqlite3', 'rf_model.pkl', 'rf_features.pkl']:\n", "unittest.main(verbosity=2)\n", "os.rename('temp_' + filename, filename)\n" ]
[ "import unittest\n", "import pandas as pd\n", "import numpy as np\n", "import datetime\n", "from ddt import ddt, data, unpack\n", "from scraper import scrape\n", "from wrangler import clean_job_number, clean_pub_date, clean_city, clean_company_name, get_acronyms, get_street_number, get_street_name, clean_title, wrangle\n", "from matcher import match\n", "from communicator import communicate\n", "from inbox_scanner import process_as_form, process_as_reply\n", "from ml import build_train_set, train_model, validate_model\n", "from db_tools import create_connection\n", "from test.test_setup import create_test_db\n", "import os\n", "prob_thresh = 0.7\n", "@data((' ', ''), ('\\n #2404\\n', '2404'), ('no. 2404', '2404'), ('# 2404',...\n", "output_string = clean_job_number(input_string)\n", "self.assertEqual(desired_string, output_string)\n", "@data((' ', ''), ('\\n 2019-02-20\\n', '2019-02-20'), ('2019-02-20',...\n", "output_string = clean_pub_date(input_string)\n", "self.assertEqual(desired_string, output_string)\n", "@data((' ', ''), (...\n", "output_string = clean_city(input_string)\n", "self.assertEqual(desired_string, output_string)\n", "@data((' ', ''), ('Frecon', 'frecon'), ('Frecon', 'frecon'), (...\n", "output_string = clean_company_name(input_string)\n", "self.assertEqual(desired_string, output_string)\n", "@data((' ', []), ('Ron Eastern Construction Ltd. (RECL)', ['RECL']), (...\n", "output_string = get_acronyms(input_string)\n", "self.assertEqual(desired_string, output_string)\n", "address_test_data = ('123 Fake St.', '123', 'fake'), ('12 Carrière Rd',\n '12', 'carriere'), ('8-1230 marenger street', '1230', 'marenger'), (\n 'apt. 8, 1230 marenger street', '1230', 'marenger'), (\n '8-1230 marenger street, apt. 8, ', '1230', 'marenger'), (\n '1230 apt. 8, marenger street', '1230', 'marenger'), (\n '1010 talbot st. unit #1', '1010', 'talbot'), ('6250 st albans court',\n '6250', 'albans'), ('6250 saint albans court', '6250', 'albans'), (\n '6250 st. albans', '6250', 'albans'), ('6250 st-albans CRT', '6250',\n 'albans'), (\n 'University of Ottawa, Faculty of Medicine and Faculty of Health Sciences, Roger Guindon Hall, 451 Smyth Road, Ottawa, Ontario K1H 8L1'\n , '451', 'smyth'), ('145 Jean-Jacques Lussier', '145', 'jean-jacques'), (\n 'Edwardsburgh/Cardinal', '', '')\n", "@data(*address_test_data)...\n", "output_string = get_street_number(input_string)\n", "self.assertEqual(desired_string1, output_string)\n", "@data(*address_test_data)...\n", "output_string = get_street_name(input_string)\n", "self.assertEqual(desired_string2, output_string)\n", "@data((' ', ''), ('test', 'test'), ('testé', 'teste'))...\n", "output_string = clean_title(input_string)\n", "self.assertEqual(desired_string, output_string)\n", "def setUpClass():...\n", "for filename in ['cert_db.sqlite3', 'rf_model.pkl', 'rf_features.pkl']:\n", "create_test_db()\n", "os.rename(filename, 'temp_' + filename)\n", "for filename in ['cert_db.sqlite3', 'rf_model.pkl', 'rf_features.pkl']:\n", "def tearDownClass():...\n", "os.rename('test_' + filename, filename)\n", "for filename in ['cert_db.sqlite3', 'rf_model.pkl', 'rf_features.pkl']:\n", "@data(('9981', 'B0046A36-3F1C-11E9-9A87-005056AA6F11', 0, 0, 0), ('9982',...\n", "os.rename('temp_' + filename, filename)\n", "os.remove(filename)\n", "email_obj = {'sender': 'Alex Roy <[email protected]>', 'subject':\n 'DO NOT MODIFY MESSAGE BELOW - JUST HIT `SEND`', 'date':\n 'Tue, 7 May 2019 17:34:17 +0000', 'content':\n f'job_number={job_number}&title=TEST_ENTRY&city=Ottawa&address=2562+Del+Zotto+Ave.%2C+Ottawa%2C+Ontario&contractor=GCN&engineer=Goodkey&owner=Douglas+Stalker&quality=2&cc_email=&link_to_cert={dcn_key}\\r\\n'\n }\n", "fake_dilfo_insert = \"\"\"\n INSERT INTO df_dilfo (job_number, receiver_email, closed)\n VALUES ({}, '[email protected]', {})\n \"\"\"\n", "fake_match_insert = \"\"\"\n INSERT INTO df_matched (job_number, verifier, ground_truth)\n VALUES ({}, '[email protected]', {})\n \"\"\"\n", "if was_prev_closed or was_prev_tracked:\n", "conn.cursor().execute(fake_dilfo_insert.format(job_number, was_prev_closed))\n", "if was_prev_matched:\n", "if was_prev_closed:\n", "df_dilfo_pre = pd.read_sql(\n f'SELECT * FROM df_dilfo WHERE job_number={job_number}', conn)\n", "conn.cursor().execute(fake_match_insert.format(job_number, 1))\n", "conn.cursor().execute(fake_match_insert.format(job_number, 0))\n", "df_matched_pre = pd.read_sql(\n f'SELECT * FROM df_matched WHERE job_number={job_number}', conn)\n", "process_as_form(email_obj)\n", "df_dilfo_post = pd.read_sql(\n f'SELECT * FROM df_dilfo WHERE job_number={job_number}', conn)\n", "df_matched_post = pd.read_sql(\n f'SELECT * FROM df_matched WHERE job_number={job_number}', conn)\n", "self.assertEqual(len(df_dilfo_post), 1)\n", "self.assertEqual(bool(df_dilfo_post.iloc[0].closed), bool(was_prev_closed or\n dcn_key))\n", "self.assertEqual(any(df_matched_post.ground_truth), bool(was_prev_closed or\n dcn_key))\n", "self.assertEqual(len(df_matched_pre) + bool(dcn_key and not was_prev_closed\n ), len(df_matched_post))\n", "self.assertEqual(list(df_matched_pre.columns), list(df_matched_post.columns))\n", "self.assertEqual(list(df_dilfo_pre.columns), list(df_dilfo_post.columns))\n", "@data(('9991', 'B0046A36-3F1C-11E9-9A87-005056AA6F01', 0, 0, 0), ('9992',...\n", "email_obj = {'sender': 'Alex Roy <[email protected]>', 'subject':\n f'Re: [EXTERNAL] Upcoming Holdback Release: #{job_number}', 'date':\n 'Thu, 30 May 2019 00:41:05 +0000', 'content':\n f\"{ground_truth}\\r\\n\\r\\nAlex Roy\\r\\nDilfo Mechanical\\r\\n(613) 899-9324\\r\\n\\r\\n________________________________\\r\\nFrom: Dilfo HBR Bot <[email protected]>\\r\\nSent: Wednesday, May 29, 2019 8:40 PM\\r\\nTo: Alex Roy\\r\\nSubject: [EXTERNAL] #{job_number} - Upcoming Holdback Release\\r\\n\\r\\nHi Alex,\\r\\n\\r\\nYou're receiving this e-mail notification because you added the project #{job_number} - DWS Building Expansion to the watchlist of upcoming holdback releases. \\r\\n\\r\\nBefore going any further, please follow the link below to make sure the algorithm correctly matched the project in question:\\r\\nhttps://link.spamstopshere.net/u/f544cec5/3CEdd3OC6RGV00Hm8I9C_g?u=https%3A%2F%2Fcanada.constructconnect.com%2Fdcn%2Fcertificates-and-notices%2F%2F{dcn_key}\\r\\n\\r\\nIf it's the right project, then the certificate was just published this past Wednesday on March 6, 2019. This means a valid holdback release invoice could be submitted as of:\\r\\nA) April 20, 2019 if the contract was signed before October 1, 2019 or;\\r\\nB) May 5, 2019 if the contract was signed since then.\\r\\n\\r\\nPlease be aware this is a fully automated message. The info presented above could be erroneous.\\r\\nYou can help improve the matching algorithms by replying to this e-mail with a simple `1` or `0` to confirm whether or not the linked certificate represents the project in question.\\r\\n\\r\\nThanks,\\r\\nDilfo HBR Bot\\r\\n\"\n }\n", "fake_dilfo_insert = \"\"\"\n INSERT INTO df_dilfo (job_number, closed)\n VALUES ({}, {})\n \"\"\"\n", "fake_match_insert = \"\"\"\n INSERT INTO df_matched (job_number, ground_truth)\n VALUES ({}, {})\n \"\"\"\n", "conn.cursor().execute(fake_dilfo_insert.format(job_number, was_prev_closed))\n", "if was_prev_matched:\n", "if was_prev_closed:\n", "df_dilfo_pre = pd.read_sql(\n f'SELECT * FROM df_dilfo WHERE job_number={job_number}', conn)\n", "conn.cursor().execute(fake_match_insert.format(job_number, 1))\n", "conn.cursor().execute(fake_match_insert.format(job_number, 0))\n", "df_matched_pre = pd.read_sql(\n f'SELECT * FROM df_matched WHERE job_number={job_number}', conn)\n", "process_as_reply(email_obj)\n", "df_dilfo_post = pd.read_sql(\n f'SELECT * FROM df_dilfo WHERE job_number={job_number}', conn)\n", "df_matched_post = pd.read_sql(\n f'SELECT * FROM df_matched WHERE job_number={job_number}', conn)\n", "self.assertEqual(len(df_dilfo_pre), len(df_dilfo_post))\n", "self.assertEqual(df_dilfo_post.iloc[0].closed, was_prev_closed or ground_truth)\n", "self.assertEqual(any(df_matched_post.ground_truth), was_prev_closed or\n ground_truth)\n", "self.assertEqual(len(df_matched_pre) + (not was_prev_closed), len(\n df_matched_post))\n", "self.assertEqual(list(df_matched_pre.columns), list(df_matched_post.columns))\n", "self.assertEqual(list(df_dilfo_pre.columns), list(df_dilfo_post.columns))\n", "def setUp(self):...\n", "for filename in ['cert_db.sqlite3', 'rf_model.pkl', 'rf_features.pkl']:\n", "create_test_db()\n", "os.rename(filename, 'temp_' + filename)\n", "for filename in ['cert_db.sqlite3', 'rf_model.pkl', 'rf_features.pkl']:\n", "def tearDown(self):...\n", "os.rename('test_' + filename, filename)\n", "for filename in ['cert_db.sqlite3', 'rf_model.pkl', 'rf_features.pkl']:\n", "def test_scarpe_to_communicate(self):...\n", "os.rename('temp_' + filename, filename)\n", "os.remove(filename)\n", "test_limit = 3\n", "web_df = scrape(limit=test_limit, test=True, since='week_ago')\n", "self.assertEqual(len(web_df), test_limit)\n", "match_first_query = 'SELECT * FROM df_dilfo WHERE closed=0 LIMIT 1'\n", "dilfo_row = pd.read_sql(match_first_query, conn).iloc[0]\n", "communicate(web_df, dilfo_row, test=True)\n", "def test_truth_table(self):...\n", "build_train_set()\n", "train_model(prob_thresh=prob_thresh)\n", "match_query = \"\"\"\n SELECT \n df_dilfo.job_number,\n df_dilfo.city,\n df_dilfo.address,\n df_dilfo.title,\n df_dilfo.owner,\n df_dilfo.contractor,\n df_dilfo.engineer,\n df_dilfo.receiver_email,\n df_dilfo.cc_email,\n df_dilfo.quality,\n df_matched.dcn_key,\n df_matched.ground_truth\n FROM \n df_dilfo \n LEFT JOIN \n df_matched\n ON \n df_dilfo.job_number=df_matched.job_number\n WHERE \n df_dilfo.closed=1\n AND\n df_matched.ground_truth=1\n AND \n df_matched.validate=0\n \"\"\"\n", "test_df_dilfo = pd.read_sql(match_query, conn)\n", "test_web_df = scrape(ref=test_df_dilfo)\n", "results = match(df_dilfo=test_df_dilfo, df_web=test_web_df, test=True,\n prob_thresh=prob_thresh, version='new')\n", "qty_actual_matches = int(len(results) ** 0.5)\n", "qty_found_matches = results[results.pred_match == 1].title.nunique()\n", "self.assertTrue(qty_found_matches == qty_actual_matches, msg=\n f'qty_found_matches({qty_found_matches}) not equal qty_actual_matches({qty_actual_matches})'\n )\n", "false_positives = len(results[results.pred_match == 1]) - qty_found_matches\n", "self.assertTrue(false_positives <= round(qty_actual_matches * 0.25, 1), msg\n =\n f'found too many false positives ({false_positives}) out of total test projects ({qty_actual_matches})'\n )\n", "sample_dilfo = pd.DataFrame({'job_number': '2387', 'city': 'Ottawa',\n 'address': '2562 Del Zotto Ave., Ottawa, Ontario', 'title':\n 'DWS Building Expansion', 'owner': 'Douglas Stalker', 'contractor':\n 'GNC', 'engineer': 'Goodkey', 'receiver_email': '[email protected]',\n 'cc_email': '', 'quality': '2', 'closed': '0'}, index=range(1))\n", "sample_web = pd.DataFrame({'pub_date': '2019-03-06', 'city':\n 'Ottawa-Carleton', 'address':\n '2562 Del Zotto Avenue, Gloucester, Ontario', 'title':\n 'Construct a 1 storey storage addition to a 2 storey office/industrial building'\n , 'owner': 'Doug Stalker, DWS Roofing', 'contractor':\n 'GNC Constructors Inc.', 'engineer': None, 'dcn_key':\n 'B0046A36-3F1C-11E9-9A87-005056AA6F02'}, index=range(1))\n", "is_match, prob = match(df_dilfo=sample_dilfo, df_web=sample_web, test=True,\n version='new').iloc[0][['pred_match', 'pred_prob']]\n", "self.assertTrue(is_match, msg=\n f'Project #{sample_dilfo.job_number} did not match successfully. Match probability returned was {prob}.'\n )\n", "results = match(df_dilfo=sample_dilfo, since='2019-03-05', until=\n '2019-03-07', test=True, version='new')\n", "prob_from_db_cert = results[results.contractor == 'gnc'].iloc[0].pred_prob\n", "self.assertTrue(round(prob, 2) == round(prob_from_db_cert, 2))\n", "validate_model(prob_thresh=prob_thresh, test=True)\n", "if __name__ == '__main__':\n", "for filename in ['cert_db.sqlite3', 'rf_model.pkl', 'rf_features.pkl']:\n", "unittest.main(verbosity=2)\n", "os.rename('temp_' + filename, filename)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 4, 0, 0, 4, 0, 0, 4, 4, 4, 4, 0, 4, 4, 0, 0, 0, 0, 0, 0, 0, 0, 4, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Import'", "Import'", "Import'", "Import'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Import'", "Assign'", "Condition", "Assign'", "Expr'", "Condition", "Assign'", "Expr'", "Condition", "Assign'", "Expr'", "Condition", "Assign'", "Expr'", "Condition", "Assign'", "Expr'", "Assign'", "Condition", "Assign'", "Expr'", "Condition", "Assign'", "Expr'", "Condition", "Assign'", "Expr'", "FunctionDef'", "For", "Expr'", "Expr'", "For", "FunctionDef'", "Expr'", "For", "Condition", "Expr'", "Expr'", "Assign'", "Assign'", "Assign'", "Condition", "Expr'", "Condition", "Condition", "Assign'", "Expr'", "Expr'", "Assign'", "Expr'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Condition", "Assign'", "Assign'", "Assign'", "Expr'", "Condition", "Condition", "Assign'", "Expr'", "Expr'", "Assign'", "Expr'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "FunctionDef'", "For", "Expr'", "Expr'", "For", "FunctionDef'", "Expr'", "For", "FunctionDef'", "Expr'", "Expr'", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "Expr'", "FunctionDef'", "Expr'", "Expr'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "Expr'", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "Expr'", "Expr'", "Condition", "For", "Expr'", "Expr'" ]
[ "@defer.inlineCallbacks...\n", "self.obj_path = yield db.transact(lambda : [db.ref(db.get_root()['oms_root'])]\n )()\n" ]
[ "@defer.inlineCallbacks...\n", "self.obj_path = yield db.transact(lambda : [db.ref(db.get_root()['oms_root'])]\n )()\n" ]
[ 0, 0 ]
[ "Condition", "Assign'" ]
[ "@staticmethod...\n", "return {'files': ['file_opened', 'file_read'], 'registry': ['regkey_opened',\n 'regkey_written', 'regkey_read'], 'mutexes': ['mutex'], 'directories':\n ['directory_created', 'directory_removed', 'directory_enumerated'],\n 'processes': ['command_line', 'dll_loaded']}\n" ]
[ "@staticmethod...\n", "return {'files': ['file_opened', 'file_read'], 'registry': ['regkey_opened',\n 'regkey_written', 'regkey_read'], 'mutexes': ['mutex'], 'directories':\n ['directory_created', 'directory_removed', 'directory_enumerated'],\n 'processes': ['command_line', 'dll_loaded']}\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "def FUNC_7(VAR_17, *VAR_18, **VAR_19):...\n", "\"\"\"docstring\"\"\"\n", "VAR_20 = VAR_17.POST\n", "VAR_21 = VAR_20.get('new_status')\n", "if not VAR_21:\n", "if not VAR_8 or not callable(VAR_8):\n", "if not 'fields' in VAR_20:\n", "return error_handler.logErrorAndReturnOK(\n 'No valid status can be set by the manageModelStatus.')\n", "error_handler.logErrorAndReturnOK(\n 'No fields to filter on found for manageModelStatus.')\n", "VAR_22 = pickle.loads(str(VAR_20['fields']))\n", "VAR_23 = VAR_7.getForFields(VAR_22, limit=BATCH_SIZE)\n", "for VAR_9 in VAR_23:\n", "if VAR_21:\n", "db.put(VAR_23)\n", "VAR_25 = VAR_21\n", "VAR_25 = VAR_8(VAR_9)\n", "if len(VAR_23) == VAR_4:\n", "VAR_9.status = VAR_25\n", "VAR_24 = VAR_20.copy()\n", "return responses.terminateTask()\n", "return responses.startTask(VAR_17.path, VAR_24=context)\n" ]
[ "def manageModelsStatus(request, *args, **kwargs):...\n", "\"\"\"docstring\"\"\"\n", "post_dict = request.POST\n", "new_status = post_dict.get('new_status')\n", "if not new_status:\n", "if not status_retriever or not callable(status_retriever):\n", "if not 'fields' in post_dict:\n", "return error_handler.logErrorAndReturnOK(\n 'No valid status can be set by the manageModelStatus.')\n", "error_handler.logErrorAndReturnOK(\n 'No fields to filter on found for manageModelStatus.')\n", "fields = pickle.loads(str(post_dict['fields']))\n", "entities = entity_logic.getForFields(fields, limit=BATCH_SIZE)\n", "for entity in entities:\n", "if new_status:\n", "db.put(entities)\n", "status = new_status\n", "status = status_retriever(entity)\n", "if len(entities) == BATCH_SIZE:\n", "entity.status = status\n", "context = post_dict.copy()\n", "return responses.terminateTask()\n", "return responses.startTask(request.path, context=context)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 5 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Condition", "Condition", "Condition", "Return'", "Expr'", "Assign'", "Assign'", "For", "Condition", "Expr'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Return'", "Return'" ]
[ "def FUNC_14(self, VAR_19, VAR_20):...\n", "\"\"\"docstring\"\"\"\n", "if not self.reached_corner[self.corner_counter]:\n", "self.reached_corner[self.corner_counter] = self.navloc.goToPosition(VAR_20[\n self.corner_counter][0] * VAR_19, VAR_20[self.corner_counter][1] * VAR_19)\n", "self.logArrival('corner ' + str(self.corner_counter), VAR_20[self.\n corner_counter][0] * VAR_19, VAR_20[self.corner_counter][1] * VAR_19)\n", "if self.corner_counter == len(self.reached_corner) - 1:\n", "self.reached_corner = [False] * len(self.reached_corner)\n", "self.corner_counter = (self.corner_counter + 1) % len(self.reached_corner)\n" ]
[ "def testSquare(self, length, corners):...\n", "\"\"\"docstring\"\"\"\n", "if not self.reached_corner[self.corner_counter]:\n", "self.reached_corner[self.corner_counter] = self.navloc.goToPosition(corners\n [self.corner_counter][0] * length, corners[self.corner_counter][1] * length\n )\n", "self.logArrival('corner ' + str(self.corner_counter), corners[self.\n corner_counter][0] * length, corners[self.corner_counter][1] * length)\n", "if self.corner_counter == len(self.reached_corner) - 1:\n", "self.reached_corner = [False] * len(self.reached_corner)\n", "self.corner_counter = (self.corner_counter + 1) % len(self.reached_corner)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Assign'", "Expr'", "Condition", "Assign'", "Assign'" ]
[ "def __init__(self, VAR_18, VAR_35=10000, VAR_36=errors.BAD_COMMENT, VAR_37=...\n", "CLASS_0.__init__(self, VAR_18, **kw)\n", "self.length = VAR_35\n", "self.len_error = VAR_37\n", "self.emp_error = VAR_36\n" ]
[ "def __init__(self, item, length=10000, empty_error=errors.BAD_COMMENT,...\n", "Validator.__init__(self, item, **kw)\n", "self.length = length\n", "self.len_error = length_error\n", "self.emp_error = empty_error\n" ]
[ 0, 0, 0, 0, 0 ]
[ "Condition", "Expr'", "Assign'", "Assign'", "Assign'" ]
[ "def __init__(self, VAR_64):...\n", "self.func = VAR_64\n", "self.shellcmd = None\n", "self.norun = False\n", "self.input = None\n", "self.output = None\n", "self.params = None\n", "self.message = None\n", "self.benchmark = None\n", "self.threads = None\n", "self.resources = None\n", "self.priority = None\n", "self.version = None\n", "self.log = None\n", "self.docstring = None\n" ]
[ "def __init__(self, func):...\n", "self.func = func\n", "self.shellcmd = None\n", "self.norun = False\n", "self.input = None\n", "self.output = None\n", "self.params = None\n", "self.message = None\n", "self.benchmark = None\n", "self.threads = None\n", "self.resources = None\n", "self.priority = None\n", "self.version = None\n", "self.log = None\n", "self.docstring = None\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Docstring" ]
[ "def FUNC_3(self):...\n", "self.expected_requests([(\n 'http://localhost/auth/api/v1/accounts/self/xsrf_token', {'data': {},\n 'headers': {'X-XSRF-Token-Request': '1'}}, {'expiration_sec': 100,\n 'xsrf_token': 'token'}), ('http://localhost/a', {'data': {'foo': 'bar'},\n 'headers': {'X-XSRF-Token': 'token'}}, 'bar', None), (\n 'http://localhost/auth/api/v1/accounts/self/xsrf_token', {'data': {},\n 'headers': {'X-XSRF-Token-Request': '1'}}, {'expiration_sec': 100,\n 'xsrf_token': 'token2'}), ('http://localhost/a', {'data': {'foo': 'bar'\n }, 'headers': {'X-XSRF-Token': 'token2'}}, 'foo', None)])\n", "VAR_1 = xsrf_client._utcnow()\n", "VAR_0 = xsrf_client.XsrfRemote('http://localhost/')\n", "VAR_0.url_read('/a', data={'foo': 'bar'})\n", "self.mock(xsrf_client, '_utcnow', lambda : VAR_1 + datetime.timedelta(\n seconds=91))\n", "VAR_0.url_read('/a', data={'foo': 'bar'})\n" ]
[ "def testXsrfRemoteRefresh(self):...\n", "self.expected_requests([(\n 'http://localhost/auth/api/v1/accounts/self/xsrf_token', {'data': {},\n 'headers': {'X-XSRF-Token-Request': '1'}}, {'expiration_sec': 100,\n 'xsrf_token': 'token'}), ('http://localhost/a', {'data': {'foo': 'bar'},\n 'headers': {'X-XSRF-Token': 'token'}}, 'bar', None), (\n 'http://localhost/auth/api/v1/accounts/self/xsrf_token', {'data': {},\n 'headers': {'X-XSRF-Token-Request': '1'}}, {'expiration_sec': 100,\n 'xsrf_token': 'token2'}), ('http://localhost/a', {'data': {'foo': 'bar'\n }, 'headers': {'X-XSRF-Token': 'token2'}}, 'foo', None)])\n", "now = xsrf_client._utcnow()\n", "remote = xsrf_client.XsrfRemote('http://localhost/')\n", "remote.url_read('/a', data={'foo': 'bar'})\n", "self.mock(xsrf_client, '_utcnow', lambda : now + datetime.timedelta(seconds=91)\n )\n", "remote.url_read('/a', data={'foo': 'bar'})\n" ]
[ 0, 5, 5, 0, 5, 5, 0 ]
[ "FunctionDef'", "Expr'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'" ]
[ "def FUNC_8(self):...\n", "\"\"\"docstring\"\"\"\n", "VAR_14 = self.session.get('https://www.kijiji.ca/m-my-ads.html')\n", "VAR_27 = FUNC_0(VAR_14.text, 'userId')\n", "VAR_28 = (\n 'https://www.kijiji.ca/j-get-my-ads.json?_=1&currentOffset=0&isPromoting=false&show=ACTIVE&user={}'\n .format(VAR_27))\n", "VAR_17 = self.session.get(VAR_28)\n", "VAR_29 = json.loads(VAR_17.text)\n", "VAR_30 = [entry['id'] for entry in VAR_29['myAdEntries']]\n", "VAR_31 = [entry['title'] for entry in VAR_29['myAdEntries']]\n", "return zip(VAR_31, VAR_30)\n" ]
[ "def get_all_ads(self):...\n", "\"\"\"docstring\"\"\"\n", "resp = self.session.get('https://www.kijiji.ca/m-my-ads.html')\n", "user_id = get_token(resp.text, 'userId')\n", "my_ads_url = (\n 'https://www.kijiji.ca/j-get-my-ads.json?_=1&currentOffset=0&isPromoting=false&show=ACTIVE&user={}'\n .format(user_id))\n", "my_ads_page = self.session.get(my_ads_url)\n", "my_ads_tree = json.loads(my_ads_page.text)\n", "ad_ids = [entry['id'] for entry in my_ads_tree['myAdEntries']]\n", "ad_names = [entry['title'] for entry in my_ads_tree['myAdEntries']]\n", "return zip(ad_names, ad_ids)\n" ]
[ 0, 0, 0, 5, 5, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_0(self, VAR_2, VAR_3, VAR_4, VAR_5):...\n", "\"\"\"docstring\"\"\"\n", "VAR_6 = []\n", "VAR_7 = self.pool.get('ir.model.data')\n", "VAR_8 = self.pool.get('res.users').browse(VAR_2, VAR_3, VAR_3).company_id.id\n", "VAR_9 = self.pool.get('account.bank.statement')\n", "VAR_10 = self.pool.get('ir.sequence')\n", "VAR_11 = self.pool.get('account.journal')\n", "VAR_2.execute(\n 'select DISTINCT journal_id from pos_journal_users where user_id=%d order by journal_id'\n % VAR_3)\n", "VAR_12 = map(lambda x1: x1[0], VAR_2.fetchall())\n", "VAR_2.execute('string' % ','.join(map(lambda x: \"'\" + str(x) + \"'\", VAR_12)))\n", "VAR_13 = map(lambda x1: x1[0], VAR_2.fetchall())\n", "for journal in VAR_11.browse(VAR_2, VAR_3, VAR_13):\n", "VAR_4 = VAR_9.search(VAR_2, VAR_3, [('state', '!=', 'confirm'), ('user_id',\n '=', VAR_3), ('journal_id', '=', journal.id)])\n", "VAR_14 = self.pool.get('ir.model.data')\n", "if len(VAR_4):\n", "VAR_15 = VAR_14._get_id(VAR_2, VAR_3, 'account', 'view_bank_statement_tree')\n", "VAR_17 = ''\n", "VAR_16 = VAR_14._get_id(VAR_2, VAR_3, 'account', 'view_bank_statement_form2')\n", "if journal.sequence_id:\n", "if VAR_15:\n", "VAR_17 = VAR_10.get_id(VAR_2, VAR_3, journal.sequence_id.id)\n", "VAR_17 = VAR_10.get(VAR_2, VAR_3, 'account.bank.statement')\n", "VAR_15 = VAR_14.browse(VAR_2, VAR_3, VAR_15, VAR_5=context).res_id\n", "if VAR_16:\n", "VAR_18 = VAR_9.create(VAR_2, VAR_3, {'journal_id': journal.id, 'company_id':\n VAR_8, 'user_id': VAR_3, 'state': 'open', 'name': VAR_17,\n 'starting_details_ids': VAR_9._get_cash_close_box_lines(VAR_2, VAR_3, [])})\n", "VAR_16 = VAR_14.browse(VAR_2, VAR_3, VAR_16, VAR_5=context).res_id\n", "return {'domain': \"[('state','=','open')]\", 'name': 'Open Statement',\n 'view_type': 'form', 'view_mode': 'tree,form', 'res_model':\n 'account.bank.statement', 'views': [(VAR_15, 'tree'), (VAR_16, 'form')],\n 'type': 'ir.actions.act_window'}\n", "VAR_9.button_open(VAR_2, VAR_3, [VAR_18], VAR_5)\n" ]
[ "def open_statement(self, cr, uid, ids, context):...\n", "\"\"\"docstring\"\"\"\n", "list_statement = []\n", "mod_obj = self.pool.get('ir.model.data')\n", "company_id = self.pool.get('res.users').browse(cr, uid, uid).company_id.id\n", "statement_obj = self.pool.get('account.bank.statement')\n", "sequence_obj = self.pool.get('ir.sequence')\n", "journal_obj = self.pool.get('account.journal')\n", "cr.execute(\n 'select DISTINCT journal_id from pos_journal_users where user_id=%d order by journal_id'\n % uid)\n", "j_ids = map(lambda x1: x1[0], cr.fetchall())\n", "cr.execute(\n \"\"\" select id from account_journal\n where auto_cash='True' and type='cash'\n and id in (%s)\"\"\"\n % ','.join(map(lambda x: \"'\" + str(x) + \"'\", j_ids)))\n", "journal_ids = map(lambda x1: x1[0], cr.fetchall())\n", "for journal in journal_obj.browse(cr, uid, journal_ids):\n", "ids = statement_obj.search(cr, uid, [('state', '!=', 'confirm'), ('user_id',\n '=', uid), ('journal_id', '=', journal.id)])\n", "data_obj = self.pool.get('ir.model.data')\n", "if len(ids):\n", "id2 = data_obj._get_id(cr, uid, 'account', 'view_bank_statement_tree')\n", "number = ''\n", "id3 = data_obj._get_id(cr, uid, 'account', 'view_bank_statement_form2')\n", "if journal.sequence_id:\n", "if id2:\n", "number = sequence_obj.get_id(cr, uid, journal.sequence_id.id)\n", "number = sequence_obj.get(cr, uid, 'account.bank.statement')\n", "id2 = data_obj.browse(cr, uid, id2, context=context).res_id\n", "if id3:\n", "statement_id = statement_obj.create(cr, uid, {'journal_id': journal.id,\n 'company_id': company_id, 'user_id': uid, 'state': 'open', 'name':\n number, 'starting_details_ids': statement_obj._get_cash_close_box_lines\n (cr, uid, [])})\n", "id3 = data_obj.browse(cr, uid, id3, context=context).res_id\n", "return {'domain': \"[('state','=','open')]\", 'name': 'Open Statement',\n 'view_type': 'form', 'view_mode': 'tree,form', 'res_model':\n 'account.bank.statement', 'views': [(id2, 'tree'), (id3, 'form')],\n 'type': 'ir.actions.act_window'}\n", "statement_obj.button_open(cr, uid, [statement_id], context)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 4, 4, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "Expr'", "Assign'", "For", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Condition", "Condition", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Return'", "Expr'" ]
[ "@staticmethod...\n", "def FUNC_13(VAR_9, VAR_7, VAR_11, VAR_10):...\n", "if VAR_11 == 'POST':\n", "return CLASS_0.session.post(FUNC_2(VAR_7), VAR_9=ToUtf8Json(data), headers=\n _HEADERS, VAR_10=timeout)\n", "if VAR_11 == 'GET':\n", "return CLASS_0.session.get(FUNC_2(VAR_7), headers=_HEADERS, VAR_10=timeout)\n", "@retries(5, delay=0.5, backoff=1.5)...\n", "if VAR_11 == 'POST':\n", "return requests.post(FUNC_2(VAR_7), VAR_9=ToUtf8Json(data), headers=_HEADERS)\n", "if VAR_11 == 'GET':\n", "return requests.get(FUNC_2(VAR_7), headers=_HEADERS)\n", "if not FUNC_3():\n", "return VAR_1.submit(FUNC_14, VAR_9, VAR_7, VAR_11)\n", "return FUNC_13(VAR_9, VAR_7, VAR_11, VAR_10)\n" ]
[ "@staticmethod...\n", "def SendRequest(data, handler, method, timeout):...\n", "if method == 'POST':\n", "return BaseRequest.session.post(_BuildUri(handler), data=ToUtf8Json(data),\n headers=_HEADERS, timeout=timeout)\n", "if method == 'GET':\n", "return BaseRequest.session.get(_BuildUri(handler), headers=_HEADERS,\n timeout=timeout)\n", "@retries(5, delay=0.5, backoff=1.5)...\n", "if method == 'POST':\n", "return requests.post(_BuildUri(handler), data=ToUtf8Json(data), headers=\n _HEADERS)\n", "if method == 'GET':\n", "return requests.get(_BuildUri(handler), headers=_HEADERS)\n", "if not _CheckServerIsHealthyWithCache():\n", "return _EXECUTOR.submit(DelayedSendRequest, data, handler, method)\n", "return SendRequest(data, handler, method, timeout)\n" ]
[ 0, 0, 0, 7, 0, 7, 0, 0, 7, 0, 0, 0, 0, 0 ]
[ "Condition", "FunctionDef'", "Condition", "Return'", "Condition", "Return'", "Condition", "Condition", "Return'", "Condition", "Return'", "Condition", "Return'", "Return'" ]