lines
sequencelengths
1
383
raw_lines
sequencelengths
1
383
label
sequencelengths
1
383
type
sequencelengths
1
383
[ "async def FUNC_18(VAR_3):...\n", "VAR_3.write(b'response')\n", "VAR_3.finish()\n" ]
[ "async def callback(request):...\n", "request.write(b'response')\n", "request.finish()\n" ]
[ 0, 0, 0 ]
[ "AsyncFunctionDef'", "Expr'", "Expr'" ]
[ "import re\n", "import json\n", "import base64\n", "import warnings\n", "from functools import wraps\n", "import omero\n", "import omero.clients\n", "from past.builtins import unicode\n", "from django.http import HttpResponse, HttpResponseBadRequest, HttpResponseServerError, JsonResponse, HttpResponseForbidden\n", "from django.http import HttpResponseRedirect, HttpResponseNotAllowed, Http404, StreamingHttpResponse, HttpResponseNotFound\n", "from django.views.decorators.http import require_POST\n", "from django.views.decorators.debug import sensitive_post_parameters\n", "from django.utils.decorators import method_decorator\n", "from django.core.urlresolvers import reverse, NoReverseMatch\n", "from django.conf import settings\n", "from wsgiref.util import FileWrapper\n", "from omero.rtypes import rlong, unwrap\n", "from omero.constants.namespaces import NSBULKANNOTATIONS\n", "from .util import points_string_to_XY_list, xy_list_to_bbox\n", "from .plategrid import PlateGrid\n", "from omeroweb.version import omeroweb_buildyear as build_year\n", "from .marshal import imageMarshal, shapeMarshal, rgb_int2rgba\n", "from django.contrib.staticfiles.templatetags.staticfiles import static\n", "from django.views.generic import View\n", "from django.shortcuts import render\n", "from omeroweb.webadmin.forms import LoginForm\n", "from omeroweb.decorators import get_client_ip, is_public_user\n", "from omeroweb.webadmin.webadmin_utils import upgradeCheck\n", "from hashlib import md5\n", "from md5 import md5\n", "import long\n", "VAR_178 = int\n", "from io import BytesIO\n", "import tempfile\n", "from omero import ApiUsageException\n", "from omero.util.decorators import timeit, TimeIt\n", "from omeroweb.httprsp import HttpJavascriptResponse, HttpJavascriptResponseServerError\n", "from omeroweb.connector import Server\n", "import glob\n", "from omeroweb.webgateway.webgateway_cache import webgateway_cache, CacheBase, webgateway_tempfile\n", "import logging\n", "import os\n", "import traceback\n", "import time\n", "import zipfile\n", "import shutil\n", "from omeroweb.decorators import login_required, ConnCleaningHttpResponse\n", "from omeroweb.connector import Connector\n", "from omeroweb.webgateway.util import zip_archived_files, LUTS_IN_PNG\n", "from omeroweb.webgateway.util import get_longs, getIntOrDefault\n", "VAR_0 = CacheBase()\n", "VAR_1 = logging.getLogger(__name__)\n", "from PIL import Image\n", "import numpy\n", "VAR_1.error('No numpy installed')\n", "def FUNC_0(VAR_2):...\n", "import Image\n", "VAR_1.error('No Pillow installed')\n", "from PIL import ImageDraw\n", "VAR_48 = True\n", "VAR_48 = False\n", "\"\"\"docstring\"\"\"\n", "import ImageDraw\n", "return HttpResponse('Welcome to webgateway')\n" ]
[ "import re\n", "import json\n", "import base64\n", "import warnings\n", "from functools import wraps\n", "import omero\n", "import omero.clients\n", "from past.builtins import unicode\n", "from django.http import HttpResponse, HttpResponseBadRequest, HttpResponseServerError, JsonResponse, HttpResponseForbidden\n", "from django.http import HttpResponseRedirect, HttpResponseNotAllowed, Http404, StreamingHttpResponse, HttpResponseNotFound\n", "from django.views.decorators.http import require_POST\n", "from django.views.decorators.debug import sensitive_post_parameters\n", "from django.utils.decorators import method_decorator\n", "from django.core.urlresolvers import reverse, NoReverseMatch\n", "from django.conf import settings\n", "from wsgiref.util import FileWrapper\n", "from omero.rtypes import rlong, unwrap\n", "from omero.constants.namespaces import NSBULKANNOTATIONS\n", "from .util import points_string_to_XY_list, xy_list_to_bbox\n", "from .plategrid import PlateGrid\n", "from omeroweb.version import omeroweb_buildyear as build_year\n", "from .marshal import imageMarshal, shapeMarshal, rgb_int2rgba\n", "from django.contrib.staticfiles.templatetags.staticfiles import static\n", "from django.views.generic import View\n", "from django.shortcuts import render\n", "from omeroweb.webadmin.forms import LoginForm\n", "from omeroweb.decorators import get_client_ip, is_public_user\n", "from omeroweb.webadmin.webadmin_utils import upgradeCheck\n", "from hashlib import md5\n", "from md5 import md5\n", "import long\n", "long = int\n", "from io import BytesIO\n", "import tempfile\n", "from omero import ApiUsageException\n", "from omero.util.decorators import timeit, TimeIt\n", "from omeroweb.httprsp import HttpJavascriptResponse, HttpJavascriptResponseServerError\n", "from omeroweb.connector import Server\n", "import glob\n", "from omeroweb.webgateway.webgateway_cache import webgateway_cache, CacheBase, webgateway_tempfile\n", "import logging\n", "import os\n", "import traceback\n", "import time\n", "import zipfile\n", "import shutil\n", "from omeroweb.decorators import login_required, ConnCleaningHttpResponse\n", "from omeroweb.connector import Connector\n", "from omeroweb.webgateway.util import zip_archived_files, LUTS_IN_PNG\n", "from omeroweb.webgateway.util import get_longs, getIntOrDefault\n", "cache = CacheBase()\n", "logger = logging.getLogger(__name__)\n", "from PIL import Image\n", "import numpy\n", "logger.error('No numpy installed')\n", "def index(request):...\n", "import Image\n", "logger.error('No Pillow installed')\n", "from PIL import ImageDraw\n", "numpyInstalled = True\n", "numpyInstalled = False\n", "\"\"\"docstring\"\"\"\n", "import ImageDraw\n", "return HttpResponse('Welcome to webgateway')\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Import'", "Import'", "Import'", "Import'", "ImportFrom'", "Import'", "Import'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Import'", "Assign'", "ImportFrom'", "Import'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Import'", "ImportFrom'", "Import'", "Import'", "Import'", "Import'", "Import'", "Import'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Assign'", "Assign'", "ImportFrom'", "Import'", "Expr'", "FunctionDef'", "Import'", "Expr'", "ImportFrom'", "Assign'", "Assign'", "Docstring", "Import'", "Return'" ]
[ "def __getstate__(self):...\n", "return self\n" ]
[ "def __getstate__(self):...\n", "return self\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "async def FUNC_24(VAR_23):...\n", "self.assertEqual(VAR_23, {'server1': {FUNC_0(VAR_20): 1500}})\n", "return {'server1': {FUNC_0(VAR_20): FetchKeyResult(get_verify_key(VAR_20), \n 1200)}}\n" ]
[ "async def get_keys(keys_to_fetch):...\n", "self.assertEqual(keys_to_fetch, {'server1': {get_key_id(key1): 1500}})\n", "return {'server1': {get_key_id(key1): FetchKeyResult(get_verify_key(key1), \n 1200)}}\n" ]
[ 0, 0, 0 ]
[ "AsyncFunctionDef'", "Expr'", "Return'" ]
[ "from __future__ import absolute_import, division, unicode_literals\n", "from six import text_type\n", "import re\n", "from ..constants import voidElements, booleanAttributes, spaceCharacters\n", "from ..constants import rcdataElements, entities, xmlEntities\n", "from .. import utils\n", "from xml.sax.saxutils import escape\n", "VAR_0 = ''.join(VAR_0)\n", "VAR_1 = re.compile('[' + VAR_0 + '\"\\'=<>`]')\n", "from codecs import register_error, xmlcharrefreplace_errors\n", "VAR_3 = 'strict'\n", "VAR_3 = 'htmlentityreplace'\n", "VAR_7 = False\n", "VAR_4 = {}\n", "VAR_8 = '\"'\n", "VAR_5 = len('\\U0010ffff') == 1\n", "VAR_9 = True\n", "for VAR_38, VAR_32 in list(entities.items()):\n", "VAR_10 = True\n", "if VAR_5 and len(VAR_32) > 1 or not VAR_5 and len(VAR_32) > 2:\n", "def FUNC_1(VAR_6):...\n", "VAR_11 = True\n", "if VAR_32 != '&':\n", "if isinstance(VAR_6, (UnicodeEncodeError, UnicodeTranslateError)):\n", "VAR_12 = False\n", "if len(VAR_32) == 2:\n", "VAR_28 = []\n", "return xmlcharrefreplace_errors(VAR_6)\n", "VAR_13 = True\n", "VAR_32 = utils.surrogatePairToCodepoint(VAR_32)\n", "VAR_32 = ord(VAR_32)\n", "VAR_29 = []\n", "VAR_14 = False\n", "if VAR_32 not in VAR_4 or VAR_38.islower():\n", "VAR_30 = False\n", "VAR_15 = False\n", "VAR_4[VAR_32] = VAR_38\n", "for i, c in enumerate(VAR_6.object[VAR_6.start:VAR_6.end]):\n", "VAR_16 = True\n", "if VAR_30:\n", "for cp in VAR_29:\n", "VAR_17 = False\n", "VAR_30 = False\n", "VAR_33 = i + VAR_6.start\n", "VAR_34 = VAR_4.get(cp)\n", "return ''.join(VAR_28), VAR_6.end\n", "VAR_18 = True\n", "if utils.isSurrogatePair(VAR_6.object[VAR_33:min([VAR_6.end, VAR_33 + 2])]):\n", "if VAR_34:\n", "VAR_19 = False\n", "VAR_36 = utils.surrogatePairToCodepoint(VAR_6.object[VAR_33:VAR_33 + 2])\n", "VAR_36 = ord(c)\n", "VAR_28.append('&')\n", "VAR_28.append('&#x%s;' % hex(cp)[2:])\n", "VAR_20 = False\n", "VAR_30 = True\n", "VAR_29.append(VAR_36)\n", "VAR_28.append(VAR_34)\n", "VAR_21 = ('quote_attr_values', 'quote_char', 'use_best_quote_char',\n 'omit_optional_tags', 'minimize_boolean_attributes',\n 'use_trailing_solidus', 'space_before_trailing_solidus',\n 'escape_lt_in_attrs', 'escape_rcdata', 'resolve_entities',\n 'alphabetical_attributes', 'inject_meta_charset', 'strip_whitespace',\n 'sanitize')\n", "if not VAR_34.endswith(';'):\n", "def __init__(self, **VAR_22):...\n", "VAR_28.append(';')\n", "\"\"\"docstring\"\"\"\n", "if 'quote_char' in VAR_22:\n", "self.use_best_quote_char = False\n", "for attr in self.options:\n", "setattr(self, attr, VAR_22.get(attr, getattr(self, attr)))\n", "self.errors = []\n", "self.strict = False\n", "def FUNC_2(self, VAR_23):...\n", "assert isinstance(VAR_23, text_type)\n", "if self.encoding:\n", "return VAR_23.encode(self.encoding, VAR_3)\n", "return VAR_23\n" ]
[ "from __future__ import absolute_import, division, unicode_literals\n", "from six import text_type\n", "import re\n", "from ..constants import voidElements, booleanAttributes, spaceCharacters\n", "from ..constants import rcdataElements, entities, xmlEntities\n", "from .. import utils\n", "from xml.sax.saxutils import escape\n", "spaceCharacters = ''.join(spaceCharacters)\n", "quoteAttributeSpec = re.compile('[' + spaceCharacters + '\"\\'=<>`]')\n", "from codecs import register_error, xmlcharrefreplace_errors\n", "unicode_encode_errors = 'strict'\n", "unicode_encode_errors = 'htmlentityreplace'\n", "quote_attr_values = False\n", "encode_entity_map = {}\n", "quote_char = '\"'\n", "is_ucs4 = len('\\U0010ffff') == 1\n", "use_best_quote_char = True\n", "for k, v in list(entities.items()):\n", "omit_optional_tags = True\n", "if is_ucs4 and len(v) > 1 or not is_ucs4 and len(v) > 2:\n", "def htmlentityreplace_errors(exc):...\n", "minimize_boolean_attributes = True\n", "if v != '&':\n", "if isinstance(exc, (UnicodeEncodeError, UnicodeTranslateError)):\n", "use_trailing_solidus = False\n", "if len(v) == 2:\n", "res = []\n", "return xmlcharrefreplace_errors(exc)\n", "space_before_trailing_solidus = True\n", "v = utils.surrogatePairToCodepoint(v)\n", "v = ord(v)\n", "codepoints = []\n", "escape_lt_in_attrs = False\n", "if v not in encode_entity_map or k.islower():\n", "skip = False\n", "escape_rcdata = False\n", "encode_entity_map[v] = k\n", "for i, c in enumerate(exc.object[exc.start:exc.end]):\n", "resolve_entities = True\n", "if skip:\n", "for cp in codepoints:\n", "alphabetical_attributes = False\n", "skip = False\n", "index = i + exc.start\n", "e = encode_entity_map.get(cp)\n", "return ''.join(res), exc.end\n", "inject_meta_charset = True\n", "if utils.isSurrogatePair(exc.object[index:min([exc.end, index + 2])]):\n", "if e:\n", "strip_whitespace = False\n", "codepoint = utils.surrogatePairToCodepoint(exc.object[index:index + 2])\n", "codepoint = ord(c)\n", "res.append('&')\n", "res.append('&#x%s;' % hex(cp)[2:])\n", "sanitize = False\n", "skip = True\n", "codepoints.append(codepoint)\n", "res.append(e)\n", "options = ('quote_attr_values', 'quote_char', 'use_best_quote_char',\n 'omit_optional_tags', 'minimize_boolean_attributes',\n 'use_trailing_solidus', 'space_before_trailing_solidus',\n 'escape_lt_in_attrs', 'escape_rcdata', 'resolve_entities',\n 'alphabetical_attributes', 'inject_meta_charset', 'strip_whitespace',\n 'sanitize')\n", "if not e.endswith(';'):\n", "def __init__(self, **kwargs):...\n", "res.append(';')\n", "\"\"\"docstring\"\"\"\n", "if 'quote_char' in kwargs:\n", "self.use_best_quote_char = False\n", "for attr in self.options:\n", "setattr(self, attr, kwargs.get(attr, getattr(self, attr)))\n", "self.errors = []\n", "self.strict = False\n", "def encode(self, string):...\n", "assert isinstance(string, text_type)\n", "if self.encoding:\n", "return string.encode(self.encoding, unicode_encode_errors)\n", "return string\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 2, 0, 0, 0, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "ImportFrom'", "ImportFrom'", "Import'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Assign'", "Assign'", "ImportFrom'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "For", "Assign'", "Condition", "FunctionDef'", "Assign'", "Condition", "Condition", "Assign'", "Condition", "Assign'", "Return'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "For", "Assign'", "Condition", "For", "Assign'", "Assign'", "Assign'", "Assign'", "Return'", "Assign'", "Condition", "Condition", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'", "Assign'", "Assign'", "Expr'", "Expr'", "Assign'", "Condition", "FunctionDef'", "Expr'", "Docstring", "Condition", "Assign'", "For", "Expr'", "Assign'", "Assign'", "FunctionDef'", "Assert'", "Condition", "Return'", "Return'" ]
[ "import errno\n", "import logging\n", "import os\n", "import shutil\n", "from typing import IO, Dict, List, Optional, Tuple\n", "import twisted.internet.error\n", "import twisted.web.http\n", "from twisted.web.http import Request\n", "from twisted.web.resource import Resource\n", "from synapse.api.errors import FederationDeniedError, HttpResponseException, NotFoundError, RequestSendFailed, SynapseError\n", "from synapse.config._base import ConfigError\n", "from synapse.logging.context import defer_to_thread\n", "from synapse.metrics.background_process_metrics import run_as_background_process\n", "from synapse.util.async_helpers import Linearizer\n", "from synapse.util.retryutils import NotRetryingDestination\n", "from synapse.util.stringutils import random_string\n", "from ._base import FileInfo, Responder, get_filename_from_headers, respond_404, respond_with_responder\n", "from .config_resource import MediaConfigResource\n", "from .download_resource import DownloadResource\n", "from .filepath import MediaFilePaths\n", "from .media_storage import MediaStorage\n", "from .preview_url_resource import PreviewUrlResource\n", "from .storage_provider import StorageProviderWrapper\n", "from .thumbnail_resource import ThumbnailResource\n", "from .thumbnailer import Thumbnailer, ThumbnailError\n", "from .upload_resource import UploadResource\n", "VAR_0 = logging.getLogger(__name__)\n", "VAR_1 = 60 * 1000\n", "def __init__(self, VAR_2):...\n", "self.hs = VAR_2\n", "self.auth = VAR_2.get_auth()\n", "self.client = VAR_2.get_http_client()\n", "self.clock = VAR_2.get_clock()\n", "self.server_name = VAR_2.hostname\n", "self.store = VAR_2.get_datastore()\n", "self.max_upload_size = VAR_2.config.max_upload_size\n", "self.max_image_pixels = VAR_2.config.max_image_pixels\n", "self.primary_base_path = VAR_2.config.media_store_path\n", "self.filepaths = MediaFilePaths(self.primary_base_path)\n", "self.dynamic_thumbnails = VAR_2.config.dynamic_thumbnails\n", "self.thumbnail_requirements = VAR_2.config.thumbnail_requirements\n", "self.remote_media_linearizer = Linearizer(VAR_11='media_remote')\n", "self.recently_accessed_remotes = set()\n", "self.recently_accessed_locals = set()\n", "self.federation_domain_whitelist = VAR_2.config.federation_domain_whitelist\n", "VAR_23 = []\n", "for clz, provider_config, wrapper_config in VAR_2.config.media_storage_providers:\n", "VAR_42 = clz(VAR_2, provider_config)\n", "self.media_storage = MediaStorage(self.hs, self.primary_base_path, self.\n filepaths, VAR_23)\n", "VAR_43 = StorageProviderWrapper(VAR_42, store_local=wrapper_config.\n store_local, store_remote=wrapper_config.store_remote,\n store_synchronous=wrapper_config.store_synchronous)\n", "self.clock.looping_call(self._start_update_recently_accessed, VAR_1)\n", "VAR_23.append(VAR_43)\n", "def FUNC_0(self):...\n", "return run_as_background_process('update_recently_accessed_media', self.\n _update_recently_accessed)\n" ]
[ "import errno\n", "import logging\n", "import os\n", "import shutil\n", "from typing import IO, Dict, List, Optional, Tuple\n", "import twisted.internet.error\n", "import twisted.web.http\n", "from twisted.web.http import Request\n", "from twisted.web.resource import Resource\n", "from synapse.api.errors import FederationDeniedError, HttpResponseException, NotFoundError, RequestSendFailed, SynapseError\n", "from synapse.config._base import ConfigError\n", "from synapse.logging.context import defer_to_thread\n", "from synapse.metrics.background_process_metrics import run_as_background_process\n", "from synapse.util.async_helpers import Linearizer\n", "from synapse.util.retryutils import NotRetryingDestination\n", "from synapse.util.stringutils import random_string\n", "from ._base import FileInfo, Responder, get_filename_from_headers, respond_404, respond_with_responder\n", "from .config_resource import MediaConfigResource\n", "from .download_resource import DownloadResource\n", "from .filepath import MediaFilePaths\n", "from .media_storage import MediaStorage\n", "from .preview_url_resource import PreviewUrlResource\n", "from .storage_provider import StorageProviderWrapper\n", "from .thumbnail_resource import ThumbnailResource\n", "from .thumbnailer import Thumbnailer, ThumbnailError\n", "from .upload_resource import UploadResource\n", "logger = logging.getLogger(__name__)\n", "UPDATE_RECENTLY_ACCESSED_TS = 60 * 1000\n", "def __init__(self, hs):...\n", "self.hs = hs\n", "self.auth = hs.get_auth()\n", "self.client = hs.get_http_client()\n", "self.clock = hs.get_clock()\n", "self.server_name = hs.hostname\n", "self.store = hs.get_datastore()\n", "self.max_upload_size = hs.config.max_upload_size\n", "self.max_image_pixels = hs.config.max_image_pixels\n", "self.primary_base_path = hs.config.media_store_path\n", "self.filepaths = MediaFilePaths(self.primary_base_path)\n", "self.dynamic_thumbnails = hs.config.dynamic_thumbnails\n", "self.thumbnail_requirements = hs.config.thumbnail_requirements\n", "self.remote_media_linearizer = Linearizer(name='media_remote')\n", "self.recently_accessed_remotes = set()\n", "self.recently_accessed_locals = set()\n", "self.federation_domain_whitelist = hs.config.federation_domain_whitelist\n", "storage_providers = []\n", "for clz, provider_config, wrapper_config in hs.config.media_storage_providers:\n", "backend = clz(hs, provider_config)\n", "self.media_storage = MediaStorage(self.hs, self.primary_base_path, self.\n filepaths, storage_providers)\n", "provider = StorageProviderWrapper(backend, store_local=wrapper_config.\n store_local, store_remote=wrapper_config.store_remote,\n store_synchronous=wrapper_config.store_synchronous)\n", "self.clock.looping_call(self._start_update_recently_accessed,\n UPDATE_RECENTLY_ACCESSED_TS)\n", "storage_providers.append(provider)\n", "def _start_update_recently_accessed(self):...\n", "return run_as_background_process('update_recently_accessed_media', self.\n _update_recently_accessed)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 4, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Import'", "Import'", "Import'", "Import'", "ImportFrom'", "Import'", "Import'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Assign'", "Assign'", "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "For", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'", "FunctionDef'", "Return'" ]
[ "def FUNC_142(self, VAR_82):...\n", "return (VAR_82 or '').replace('://HOSTNAME', '://%s' % self.host)\n" ]
[ "def fix_hostname(self, body):...\n", "return (body or '').replace('://HOSTNAME', '://%s' % self.host)\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def __init__(self):...\n", "VAR_56 = def_function.function(self.multiply, input_signature=[tensor_spec.\n TensorSpec(shape=(), dtype=dtypes.float32), tensor_spec.TensorSpec(\n shape=(), dtype=dtypes.float32)])\n", "self.pure_concrete_function = VAR_56.get_concrete_function()\n", "super(CLASS_2, self).__init__()\n" ]
[ "def __init__(self):...\n", "function = def_function.function(self.multiply, input_signature=[\n tensor_spec.TensorSpec(shape=(), dtype=dtypes.float32), tensor_spec.\n TensorSpec(shape=(), dtype=dtypes.float32)])\n", "self.pure_concrete_function = function.get_concrete_function()\n", "super(DummyModel, self).__init__()\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Expr'" ]
[ "def FUNC_10(self):...\n", "VAR_20 = 'input1=/path/file.txt[ab3];input2=file2'\n", "VAR_21 = 'input3=np.zeros([2,2]);input4=[4,5]'\n", "VAR_22 = saved_model_cli.preprocess_inputs_arg_string(VAR_20)\n", "VAR_23 = saved_model_cli.preprocess_input_exprs_arg_string(VAR_21)\n", "self.assertTrue(VAR_22['input1'] == ('/path/file.txt', 'ab3'))\n", "self.assertTrue(VAR_22['input2'] == ('file2', None))\n", "print(VAR_23['input3'])\n", "self.assertAllClose(VAR_23['input3'], np.zeros([2, 2]))\n", "self.assertAllClose(VAR_23['input4'], [4, 5])\n", "self.assertTrue(len(VAR_22) == 2)\n", "self.assertTrue(len(VAR_23) == 2)\n" ]
[ "def testInputPreProcessFormats(self):...\n", "input_str = 'input1=/path/file.txt[ab3];input2=file2'\n", "input_expr_str = 'input3=np.zeros([2,2]);input4=[4,5]'\n", "input_dict = saved_model_cli.preprocess_inputs_arg_string(input_str)\n", "input_expr_dict = saved_model_cli.preprocess_input_exprs_arg_string(\n input_expr_str)\n", "self.assertTrue(input_dict['input1'] == ('/path/file.txt', 'ab3'))\n", "self.assertTrue(input_dict['input2'] == ('file2', None))\n", "print(input_expr_dict['input3'])\n", "self.assertAllClose(input_expr_dict['input3'], np.zeros([2, 2]))\n", "self.assertAllClose(input_expr_dict['input4'], [4, 5])\n", "self.assertTrue(len(input_dict) == 2)\n", "self.assertTrue(len(input_expr_dict) == 2)\n" ]
[ 0, 0, 0, 0, 5, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'" ]
[ "def FUNC_5(self):...\n", "from openapi_python_client.parser.properties import EnumProperty\n", "VAR_23 = ['abc', '123', 'a23', 'abc']\n", "VAR_27.values_from_list(VAR_23)\n" ]
[ "def test_values_from_list_duplicate(self):...\n", "from openapi_python_client.parser.properties import EnumProperty\n", "data = ['abc', '123', 'a23', 'abc']\n", "EnumProperty.values_from_list(data)\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "ImportFrom'", "Assign'", "Expr'" ]
[ "def FUNC_21(self):...\n", "VAR_32 = '/rooms/%s/send/m.room.message/mid1' % urlparse.quote(self.room_id)\n", "VAR_22, VAR_23 = self.make_request('PUT', VAR_32, b'{}')\n", "self.assertEquals(400, VAR_23.code, msg=channel.result['body'])\n", "VAR_22, VAR_23 = self.make_request('PUT', VAR_32, b'{\"_name\":\"bo\"}')\n", "self.assertEquals(400, VAR_23.code, msg=channel.result['body'])\n", "VAR_22, VAR_23 = self.make_request('PUT', VAR_32, b'{\"nao')\n", "self.assertEquals(400, VAR_23.code, msg=channel.result['body'])\n", "VAR_22, VAR_23 = self.make_request('PUT', VAR_32,\n b'[{\"_name\":\"bo\"},{\"_name\":\"jill\"}]')\n", "self.assertEquals(400, VAR_23.code, msg=channel.result['body'])\n", "VAR_22, VAR_23 = self.make_request('PUT', VAR_32, b'text only')\n", "self.assertEquals(400, VAR_23.code, msg=channel.result['body'])\n", "VAR_22, VAR_23 = self.make_request('PUT', VAR_32, b'')\n", "self.assertEquals(400, VAR_23.code, msg=channel.result['body'])\n" ]
[ "def test_invalid_puts(self):...\n", "path = '/rooms/%s/send/m.room.message/mid1' % urlparse.quote(self.room_id)\n", "request, channel = self.make_request('PUT', path, b'{}')\n", "self.assertEquals(400, channel.code, msg=channel.result['body'])\n", "request, channel = self.make_request('PUT', path, b'{\"_name\":\"bo\"}')\n", "self.assertEquals(400, channel.code, msg=channel.result['body'])\n", "request, channel = self.make_request('PUT', path, b'{\"nao')\n", "self.assertEquals(400, channel.code, msg=channel.result['body'])\n", "request, channel = self.make_request('PUT', path,\n b'[{\"_name\":\"bo\"},{\"_name\":\"jill\"}]')\n", "self.assertEquals(400, channel.code, msg=channel.result['body'])\n", "request, channel = self.make_request('PUT', path, b'text only')\n", "self.assertEquals(400, channel.code, msg=channel.result['body'])\n", "request, channel = self.make_request('PUT', path, b'')\n", "self.assertEquals(400, channel.code, msg=channel.result['body'])\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Expr'", "Assign'", "Expr'", "Assign'", "Expr'", "Assign'", "Expr'", "Assign'", "Expr'", "Assign'", "Expr'" ]
[ "def __init__(self, *VAR_13, **VAR_14):...\n", "VAR_80 = VAR_14.pop('retries', 0)\n", "VAR_81 = VAR_14.pop('cache', None)\n", "VAR_82 = VAR_14.pop('insecure_hosts', [])\n", "VAR_27 = VAR_14.pop('index_urls', None)\n", "super(CLASS_4, self).__init__(*VAR_13, **kwargs)\n", "self.headers['User-Agent'] = FUNC_1()\n", "self.auth = CLASS_0(VAR_27=index_urls)\n", "VAR_80 = urllib3.Retry(total=retries, status_forcelist=[500, 503, 520, 527],\n backoff_factor=0.25)\n", "if VAR_81:\n", "VAR_100 = CacheControlAdapter(VAR_81=SafeFileCache(cache, use_dir_lock=True\n ), max_retries=retries)\n", "VAR_100 = HTTPAdapter(max_retries=retries)\n", "VAR_83 = CLASS_3(max_retries=retries)\n", "self.mount('https://', VAR_100)\n", "self.mount('http://', VAR_83)\n", "self.mount('file://', CLASS_1())\n", "for VAR_60 in VAR_82:\n", "self.mount('https://{}/'.format(VAR_60), VAR_83)\n" ]
[ "def __init__(self, *args, **kwargs):...\n", "retries = kwargs.pop('retries', 0)\n", "cache = kwargs.pop('cache', None)\n", "insecure_hosts = kwargs.pop('insecure_hosts', [])\n", "index_urls = kwargs.pop('index_urls', None)\n", "super(PipSession, self).__init__(*args, **kwargs)\n", "self.headers['User-Agent'] = user_agent()\n", "self.auth = MultiDomainBasicAuth(index_urls=index_urls)\n", "retries = urllib3.Retry(total=retries, status_forcelist=[500, 503, 520, 527\n ], backoff_factor=0.25)\n", "if cache:\n", "secure_adapter = CacheControlAdapter(cache=SafeFileCache(cache,\n use_dir_lock=True), max_retries=retries)\n", "secure_adapter = HTTPAdapter(max_retries=retries)\n", "insecure_adapter = InsecureHTTPAdapter(max_retries=retries)\n", "self.mount('https://', secure_adapter)\n", "self.mount('http://', insecure_adapter)\n", "self.mount('file://', LocalFSAdapter())\n", "for host in insecure_hosts:\n", "self.mount('https://{}/'.format(host), insecure_adapter)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'", "For", "Expr'" ]
[ "def FUNC_7(VAR_1):...\n", "VAR_20.throw(_('Invalid Search Field {0}').format(VAR_1), VAR_20.DataError)\n" ]
[ "def _raise_exception(searchfield):...\n", "frappe.throw(_('Invalid Search Field {0}').format(searchfield), frappe.\n DataError)\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Expr'" ]
[ "def FUNC_71(self, VAR_43, VAR_26):...\n", "if not self.get('__onload'):\n", "self.set('__onload', frappe._dict())\n", "self.get('__onload')[VAR_43] = VAR_26\n" ]
[ "def set_onload(self, key, value):...\n", "if not self.get('__onload'):\n", "self.set('__onload', frappe._dict())\n", "self.get('__onload')[key] = value\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Expr'", "Assign'" ]
[ "def FUNC_3(self):...\n", "VAR_15, VAR_16 = self.make_request('PUT', '/rooms/%s/typing/%s' % (self.\n room_id, self.user_id), b'{\"typing\": false}')\n", "self.assertEquals(200, VAR_16.code)\n" ]
[ "def test_set_not_typing(self):...\n", "request, channel = self.make_request('PUT', '/rooms/%s/typing/%s' % (self.\n room_id, self.user_id), b'{\"typing\": false}')\n", "self.assertEquals(200, channel.code)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Expr'" ]
[ "@VAR_1.route('/redv')...\n", "VAR_12 = request.args.get('url')\n", "VAR_3 = urllib2.build_opener()\n", "VAR_4 = victim_headers()\n", "VAR_3.addheaders = VAR_4\n", "VAR_5 = victim_inject_code(VAR_3.open(VAR_12).read(), 'vscript')\n", "return VAR_5\n" ]
[ "@app.route('/redv')...\n", "url = request.args.get('url')\n", "opener = urllib2.build_opener()\n", "headers = victim_headers()\n", "opener.addheaders = headers\n", "html = victim_inject_code(opener.open(url).read(), 'vscript')\n", "return html\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_1(self):...\n", "return create_resource_tree({'/_matrix/key/v2': KeyApiV2Resource(self.hs)},\n root_resource=NoResource())\n" ]
[ "def create_test_resource(self):...\n", "return create_resource_tree({'/_matrix/key/v2': KeyApiV2Resource(self.hs)},\n root_resource=NoResource())\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "@VAR_2.route('/ajax/addrestriction/<int:res_type>', methods=['POST'])...\n", "return FUNC_31(VAR_8, 0)\n" ]
[ "@admi.route('/ajax/addrestriction/<int:res_type>', methods=['POST'])...\n", "return add_restriction(res_type, 0)\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "import sys\n", "from datetime import datetime\n", "from flask import Blueprint, flash, redirect, request, url_for\n", "from flask_babel import gettext as _\n", "from flask_login import current_user, login_required\n", "from sqlalchemy.exc import InvalidRequestError, OperationalError\n", "from sqlalchemy.sql.expression import func, true\n", "from . import calibre_db, config, db, logger, ub\n", "from .render_template import render_title_template\n", "from .usermanagement import login_required_if_no_ano\n", "VAR_0 = Blueprint('shelf', __name__)\n", "VAR_1 = logger.create()\n", "def FUNC_0(VAR_2):...\n", "if not VAR_2.is_public and not VAR_2.user_id == int(current_user.id):\n", "VAR_1.error('User %s not allowed to edit shelf %s', current_user, VAR_2)\n", "if VAR_2.is_public and not current_user.role_edit_shelfs():\n", "return False\n", "VAR_1.info('User %s not allowed to edit public shelves', current_user)\n", "return True\n", "return False\n" ]
[ "import sys\n", "from datetime import datetime\n", "from flask import Blueprint, flash, redirect, request, url_for\n", "from flask_babel import gettext as _\n", "from flask_login import current_user, login_required\n", "from sqlalchemy.exc import InvalidRequestError, OperationalError\n", "from sqlalchemy.sql.expression import func, true\n", "from . import calibre_db, config, db, logger, ub\n", "from .render_template import render_title_template\n", "from .usermanagement import login_required_if_no_ano\n", "shelf = Blueprint('shelf', __name__)\n", "log = logger.create()\n", "def check_shelf_edit_permissions(cur_shelf):...\n", "if not cur_shelf.is_public and not cur_shelf.user_id == int(current_user.id):\n", "log.error('User %s not allowed to edit shelf %s', current_user, cur_shelf)\n", "if cur_shelf.is_public and not current_user.role_edit_shelfs():\n", "return False\n", "log.info('User %s not allowed to edit public shelves', current_user)\n", "return True\n", "return False\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Import'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Assign'", "Assign'", "FunctionDef'", "Condition", "Expr'", "Condition", "Return'", "Expr'", "Return'", "Return'" ]
[ "def FUNC_26(self, VAR_2, VAR_3):...\n", "VAR_3.setattr('flask.request.host_url', 'http://example.com')\n", "assert _validate_redirect_url('http://fail.com') is False\n", "VAR_3.undo()\n" ]
[ "def test_it_fails_with_invalid_netloc(self, app, monkeypatch):...\n", "monkeypatch.setattr('flask.request.host_url', 'http://example.com')\n", "assert _validate_redirect_url('http://fail.com') is False\n", "monkeypatch.undo()\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Assert'", "Expr'" ]
[ "def FUNC_70(VAR_5, VAR_44, VAR_45, VAR_46, VAR_47, VAR_48, VAR_11):...\n", "\"\"\"docstring\"\"\"\n", "VAR_73 = omero.sys.ParametersI()\n", "VAR_74 = VAR_5.getQueryService()\n", "VAR_44 = set(VAR_44)\n", "VAR_45 = set(VAR_45)\n", "VAR_46 = set(VAR_46)\n", "VAR_219 = set([])\n", "VAR_48 = set(VAR_48)\n", "VAR_47 = set(VAR_47)\n", "if VAR_44:\n", "VAR_73.map = {}\n", "if VAR_47:\n", "VAR_73.map['pids'] = rlist([rlong(x) for x in list(VAR_44)])\n", "VAR_73.map = {}\n", "if VAR_45:\n", "VAR_75 = 'string'\n", "VAR_73.map['sids'] = rlist([rlong(x) for x in VAR_47])\n", "VAR_73.map = {}\n", "if VAR_48:\n", "for e in VAR_74.projection(VAR_75, VAR_73, VAR_5.SERVICE_OPTS):\n", "VAR_75 = 'string'\n", "VAR_73.map['dids'] = rlist([rlong(x) for x in VAR_45])\n", "VAR_73.map = {}\n", "if VAR_219:\n", "VAR_45.add(e[0].val)\n", "for e in VAR_74.projection(VAR_75, VAR_73, VAR_5.SERVICE_OPTS):\n", "VAR_75 = 'string'\n", "VAR_73.map['plids'] = rlist([rlong(x) for x in VAR_48])\n", "VAR_73.map = {}\n", "VAR_220 = set([])\n", "VAR_48.add(e[0].val)\n", "for e in VAR_74.projection(VAR_75, VAR_73, VAR_5.SERVICE_OPTS):\n", "VAR_75 = 'string'\n", "VAR_73.map['fsids'] = rlist([rlong(x) for x in VAR_219])\n", "VAR_221 = False\n", "VAR_46.add(e[0].val)\n", "for e in VAR_74.projection(VAR_75, VAR_73, VAR_5.SERVICE_OPTS):\n", "VAR_75 = 'string'\n", "if VAR_46:\n", "if e[1] is not None:\n", "VAR_46.add(e[0].val)\n", "for e in VAR_74.projection(VAR_75, VAR_73, VAR_5.SERVICE_OPTS):\n", "VAR_73.map = {'iids': rlist([rlong(x) for x in VAR_46])}\n", "VAR_222 = set([])\n", "VAR_219.add(e[1].val)\n", "VAR_46.add(e[0].val)\n", "VAR_330 = ''\n", "if VAR_45:\n", "if VAR_45:\n", "VAR_73.map = {'dids': rlist([rlong(x) for x in VAR_45])}\n", "VAR_223 = {'remove': {'project': list(VAR_44), 'dataset': list(VAR_45),\n 'screen': list(VAR_47), 'plate': list(VAR_48), 'image': list(VAR_46)},\n 'childless': {'project': list(VAR_222), 'dataset': list(VAR_220),\n 'orphaned': VAR_221}}\n", "VAR_73.map['dids'] = rlist([rlong(x) for x in VAR_45])\n", "VAR_75 = 'string' % VAR_330\n", "VAR_331 = ''\n", "return VAR_223\n", "VAR_330 = 'string'\n", "for e in VAR_74.projection(VAR_75, VAR_73, VAR_5.SERVICE_OPTS):\n", "if VAR_44:\n", "if e:\n", "VAR_73.map['pids'] = rlist([rlong(x) for x in VAR_44])\n", "VAR_75 = 'string' % VAR_331\n", "VAR_220.add(e[0].val)\n", "VAR_221 = True\n", "VAR_331 = 'and pdlink.parent.id not in (:pids)'\n", "for e in VAR_74.projection(VAR_75, VAR_73, VAR_5.SERVICE_OPTS):\n", "VAR_222.add(e[0].val)\n" ]
[ "def getAllObjects(conn, project_ids, dataset_ids, image_ids, screen_ids,...\n", "\"\"\"docstring\"\"\"\n", "params = omero.sys.ParametersI()\n", "qs = conn.getQueryService()\n", "project_ids = set(project_ids)\n", "dataset_ids = set(dataset_ids)\n", "image_ids = set(image_ids)\n", "fileset_ids = set([])\n", "plate_ids = set(plate_ids)\n", "screen_ids = set(screen_ids)\n", "if project_ids:\n", "params.map = {}\n", "if screen_ids:\n", "params.map['pids'] = rlist([rlong(x) for x in list(project_ids)])\n", "params.map = {}\n", "if dataset_ids:\n", "q = \"\"\"\n select pdlink.child.id\n from ProjectDatasetLink pdlink\n where pdlink.parent.id in (:pids)\n \"\"\"\n", "params.map['sids'] = rlist([rlong(x) for x in screen_ids])\n", "params.map = {}\n", "if plate_ids:\n", "for e in qs.projection(q, params, conn.SERVICE_OPTS):\n", "q = \"\"\"\n select splink.child.id\n from ScreenPlateLink splink\n where splink.parent.id in (:sids)\n \"\"\"\n", "params.map['dids'] = rlist([rlong(x) for x in dataset_ids])\n", "params.map = {}\n", "if fileset_ids:\n", "dataset_ids.add(e[0].val)\n", "for e in qs.projection(q, params, conn.SERVICE_OPTS):\n", "q = \"\"\"\n select dilink.child.id,\n dilink.child.fileset.id\n from DatasetImageLink dilink\n where dilink.parent.id in (:dids)\n \"\"\"\n", "params.map['plids'] = rlist([rlong(x) for x in plate_ids])\n", "params.map = {}\n", "extra_dataset_ids = set([])\n", "plate_ids.add(e[0].val)\n", "for e in qs.projection(q, params, conn.SERVICE_OPTS):\n", "q = \"\"\"\n select ws.image.id\n from WellSample ws\n join ws.plateAcquisition pa\n where pa.plate.id in (:plids)\n \"\"\"\n", "params.map['fsids'] = rlist([rlong(x) for x in fileset_ids])\n", "extra_orphaned = False\n", "image_ids.add(e[0].val)\n", "for e in qs.projection(q, params, conn.SERVICE_OPTS):\n", "q = \"\"\"\n select image.id\n from Image image\n left outer join image.datasetLinks dilink\n where image.fileset.id in (select fs.id\n from Image im\n join im.fileset fs\n where fs.id in (:fsids)\n group by fs.id\n having count(im.id)>1)\n \"\"\"\n", "if image_ids:\n", "if e[1] is not None:\n", "image_ids.add(e[0].val)\n", "for e in qs.projection(q, params, conn.SERVICE_OPTS):\n", "params.map = {'iids': rlist([rlong(x) for x in image_ids])}\n", "extra_project_ids = set([])\n", "fileset_ids.add(e[1].val)\n", "image_ids.add(e[0].val)\n", "exclude_datasets = ''\n", "if dataset_ids:\n", "if dataset_ids:\n", "params.map = {'dids': rlist([rlong(x) for x in dataset_ids])}\n", "result = {'remove': {'project': list(project_ids), 'dataset': list(\n dataset_ids), 'screen': list(screen_ids), 'plate': list(plate_ids),\n 'image': list(image_ids)}, 'childless': {'project': list(\n extra_project_ids), 'dataset': list(extra_dataset_ids), 'orphaned':\n extra_orphaned}}\n", "params.map['dids'] = rlist([rlong(x) for x in dataset_ids])\n", "q = (\n \"\"\"\n select distinct dilink.parent.id\n from Image image\n left outer join image.datasetLinks dilink\n where image.id in (:iids)\n %s\n and (select count(dilink2.child.id)\n from DatasetImageLink dilink2\n where dilink2.parent.id = dilink.parent.id\n and dilink2.child.id not in (:iids)) = 0\n \"\"\"\n % exclude_datasets)\n", "exclude_projects = ''\n", "return result\n", "exclude_datasets = \"\"\"\n and (\n dilink.parent.id not in (:dids)\n or dilink.parent.id = null\n )\n \"\"\"\n", "for e in qs.projection(q, params, conn.SERVICE_OPTS):\n", "if project_ids:\n", "if e:\n", "params.map['pids'] = rlist([rlong(x) for x in project_ids])\n", "q = (\n \"\"\"\n select distinct pdlink.parent.id\n from ProjectDatasetLink pdlink\n where pdlink.child.id in (:dids)\n %s\n and (select count(pdlink2.child.id)\n from ProjectDatasetLink pdlink2\n where pdlink2.parent.id = pdlink.parent.id\n and pdlink2.child.id not in (:dids)) = 0\n \"\"\"\n % exclude_projects)\n", "extra_dataset_ids.add(e[0].val)\n", "extra_orphaned = True\n", "exclude_projects = 'and pdlink.parent.id not in (:pids)'\n", "for e in qs.projection(q, params, conn.SERVICE_OPTS):\n", "extra_project_ids.add(e[0].val)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Condition", "For", "Assign'", "Assign'", "Assign'", "Condition", "Expr'", "For", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "For", "Assign'", "Assign'", "Assign'", "Expr'", "For", "Assign'", "Condition", "Condition", "Expr'", "For", "Assign'", "Assign'", "Expr'", "Expr'", "Assign'", "Condition", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Return'", "Assign'", "For", "Condition", "Condition", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "For", "Expr'" ]
[ "def FUNC_63(self, VAR_141):...\n", "VAR_275 = str(int(time.time())) + '-' + web2py_uuid()\n", "VAR_276 = self.url(self.settings.function, VAR_11=('reset_password',),\n VAR_96={'key': reset_password_key}, VAR_97=True)\n", "VAR_27 = dict(VAR_141)\n", "VAR_27.update(dict(VAR_199=reset_password_key, VAR_276=link))\n", "if self.settings.mailer and self.settings.mailer.send(VAR_35=user.email,\n", "VAR_141.update_record(VAR_275=reset_password_key)\n", "return False\n", "return True\n" ]
[ "def email_reset_password(self, user):...\n", "reset_password_key = str(int(time.time())) + '-' + web2py_uuid()\n", "link = self.url(self.settings.function, args=('reset_password',), vars={\n 'key': reset_password_key}, scheme=True)\n", "d = dict(user)\n", "d.update(dict(key=reset_password_key, link=link))\n", "if self.settings.mailer and self.settings.mailer.send(to=user.email,\n", "user.update_record(reset_password_key=reset_password_key)\n", "return False\n", "return True\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Expr'", "Condition", "Expr'", "Return'", "Return'" ]
[ "def FUNC_35(self):...\n", "return self.file.file.size\n" ]
[ "def get_size(self):...\n", "return self.file.file.size\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_65():...\n", "\"\"\"docstring\"\"\"\n", "if not hasattr(VAR_1, 'doc_events_hooks'):\n", "VAR_184 = FUNC_66('doc_events', {})\n", "return VAR_1.doc_events_hooks\n", "VAR_179 = {}\n", "for VAR_46, VAR_105 in iteritems(VAR_184):\n", "if isinstance(VAR_46, tuple):\n", "VAR_1.doc_events_hooks = VAR_179\n", "for VAR_62 in VAR_46:\n", "FUNC_67(VAR_179, VAR_46, VAR_105)\n", "FUNC_67(VAR_179, VAR_62, VAR_105)\n" ]
[ "def get_doc_hooks():...\n", "\"\"\"docstring\"\"\"\n", "if not hasattr(local, 'doc_events_hooks'):\n", "hooks = get_hooks('doc_events', {})\n", "return local.doc_events_hooks\n", "out = {}\n", "for key, value in iteritems(hooks):\n", "if isinstance(key, tuple):\n", "local.doc_events_hooks = out\n", "for doctype in key:\n", "append_hook(out, key, value)\n", "append_hook(out, doctype, value)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Assign'", "Return'", "Assign'", "For", "Condition", "Assign'", "For", "Expr'", "Expr'" ]
[ "def FUNC_20(self, VAR_44):...\n", "VAR_69 = re.sub('^ *| *\\\\| *$', '', VAR_44.group(1))\n", "VAR_69 = re.split(' *\\\\| *', VAR_69)\n", "VAR_81 = re.sub(' *|\\\\| *$', '', VAR_44.group(2))\n", "VAR_81 = re.split(' *\\\\| *', VAR_81)\n", "for VAR_101, VAR_100 in enumerate(VAR_81):\n", "if re.search('^ *-+: *$', VAR_100):\n", "VAR_79 = {'type': 'table', 'header': VAR_69, 'align': VAR_81}\n", "VAR_81[VAR_101] = 'right'\n", "if re.search('^ *:-+: *$', VAR_100):\n", "return VAR_79\n", "VAR_81[VAR_101] = 'center'\n", "if re.search('^ *:-+ *$', VAR_100):\n", "VAR_81[VAR_101] = 'left'\n", "VAR_81[VAR_101] = None\n" ]
[ "def _process_table(self, m):...\n", "header = re.sub('^ *| *\\\\| *$', '', m.group(1))\n", "header = re.split(' *\\\\| *', header)\n", "align = re.sub(' *|\\\\| *$', '', m.group(2))\n", "align = re.split(' *\\\\| *', align)\n", "for i, v in enumerate(align):\n", "if re.search('^ *-+: *$', v):\n", "item = {'type': 'table', 'header': header, 'align': align}\n", "align[i] = 'right'\n", "if re.search('^ *:-+: *$', v):\n", "return item\n", "align[i] = 'center'\n", "if re.search('^ *:-+ *$', v):\n", "align[i] = 'left'\n", "align[i] = None\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "For", "Condition", "Assign'", "Assign'", "Condition", "Return'", "Assign'", "Condition", "Assign'", "Assign'" ]
[ "def FUNC_41(self):...\n", "return self.renderer.paragraph(self.tok_text())\n" ]
[ "def output_text(self):...\n", "return self.renderer.paragraph(self.tok_text())\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_51(self, VAR_5, VAR_35):...\n", "\"\"\"docstring\"\"\"\n", "VAR_2 = FUNC_1('/get_groups_publicised')\n", "VAR_19 = {'user_ids': VAR_35}\n", "return self.client.post_json(VAR_5=destination, VAR_2=path, VAR_39=content,\n VAR_15=True)\n" ]
[ "def bulk_get_publicised_groups(self, destination, user_ids):...\n", "\"\"\"docstring\"\"\"\n", "path = _create_v1_path('/get_groups_publicised')\n", "content = {'user_ids': user_ids}\n", "return self.client.post_json(destination=destination, path=path, data=\n content, ignore_backoff=True)\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Return'" ]
[ "async def FUNC_13(self, VAR_3: Optional[str], VAR_4: str, VAR_18: str,...\n", "\"\"\"docstring\"\"\"\n", "VAR_36 = self._get_thumbnail_requirements(VAR_5)\n", "if not VAR_36:\n", "return None\n", "VAR_34 = await self.media_storage.ensure_media_is_in_local_cache(FileInfo(\n VAR_3, VAR_18, VAR_17=url_cache))\n", "VAR_12 = Thumbnailer(VAR_34)\n", "VAR_0.warning(\n 'Unable to generate thumbnails for remote media %s from %s of type %s: %s',\n VAR_4, VAR_3, VAR_5, e)\n", "VAR_32 = VAR_12.width\n", "return None\n", "VAR_33 = VAR_12.height\n", "if VAR_32 * VAR_33 >= self.max_image_pixels:\n", "VAR_0.info('Image too large to thumbnail %r x %r > %r', VAR_32, VAR_33,\n self.max_image_pixels)\n", "if VAR_12.transpose_method is not None:\n", "return None\n", "VAR_32, VAR_33 = await defer_to_thread(self.hs.get_reactor(), VAR_12.transpose)\n", "VAR_37 = {}\n", "for r_width, r_height, r_method, VAR_53 in VAR_36:\n", "if r_method == 'crop':\n", "for (VAR_13, VAR_14, VAR_16), VAR_15 in VAR_37.items():\n", "VAR_37.setdefault((r_width, r_height, VAR_53), r_method)\n", "if r_method == 'scale':\n", "if VAR_15 == 'crop':\n", "return {'width': VAR_32, 'height': VAR_33}\n", "VAR_13, VAR_14 = VAR_12.aspect(r_width, r_height)\n", "VAR_35 = await defer_to_thread(self.hs.get_reactor(), VAR_12.crop, VAR_13,\n VAR_14, VAR_16)\n", "if VAR_15 == 'scale':\n", "VAR_13 = min(VAR_32, VAR_13)\n", "if not VAR_35:\n", "VAR_35 = await defer_to_thread(self.hs.get_reactor(), VAR_12.scale, VAR_13,\n VAR_14, VAR_16)\n", "VAR_0.error('Unrecognized method: %r', VAR_15)\n", "VAR_14 = min(VAR_33, VAR_14)\n", "VAR_26 = FileInfo(VAR_3=server_name, VAR_18=file_id, thumbnail=True,\n thumbnail_width=t_width, thumbnail_height=t_height, thumbnail_method=\n t_method, thumbnail_type=t_type, VAR_17=url_cache)\n", "VAR_37[VAR_13, VAR_14, VAR_53] = r_method\n", "await self.media_storage.write_to_file(VAR_35, f)\n", "VAR_35.close()\n", "VAR_46 = os.path.getsize(VAR_27)\n", "await finish()\n", "if VAR_3:\n", "await self.store.store_local_thumbnail(VAR_4, VAR_13, VAR_14, VAR_16,\n VAR_15, VAR_46)\n", "await self.store.store_remote_media_thumbnail(VAR_3, VAR_4, VAR_18, VAR_13,\n VAR_14, VAR_16, VAR_15, VAR_46)\n", "VAR_54 = await self.store.get_remote_media_thumbnail(VAR_3, VAR_4, VAR_13,\n VAR_14, VAR_16)\n", "if not VAR_54:\n" ]
[ "async def _generate_thumbnails(self, server_name: Optional[str], media_id:...\n", "\"\"\"docstring\"\"\"\n", "requirements = self._get_thumbnail_requirements(media_type)\n", "if not requirements:\n", "return None\n", "input_path = await self.media_storage.ensure_media_is_in_local_cache(FileInfo\n (server_name, file_id, url_cache=url_cache))\n", "thumbnailer = Thumbnailer(input_path)\n", "logger.warning(\n 'Unable to generate thumbnails for remote media %s from %s of type %s: %s',\n media_id, server_name, media_type, e)\n", "m_width = thumbnailer.width\n", "return None\n", "m_height = thumbnailer.height\n", "if m_width * m_height >= self.max_image_pixels:\n", "logger.info('Image too large to thumbnail %r x %r > %r', m_width, m_height,\n self.max_image_pixels)\n", "if thumbnailer.transpose_method is not None:\n", "return None\n", "m_width, m_height = await defer_to_thread(self.hs.get_reactor(),\n thumbnailer.transpose)\n", "thumbnails = {}\n", "for r_width, r_height, r_method, r_type in requirements:\n", "if r_method == 'crop':\n", "for (t_width, t_height, t_type), t_method in thumbnails.items():\n", "thumbnails.setdefault((r_width, r_height, r_type), r_method)\n", "if r_method == 'scale':\n", "if t_method == 'crop':\n", "return {'width': m_width, 'height': m_height}\n", "t_width, t_height = thumbnailer.aspect(r_width, r_height)\n", "t_byte_source = await defer_to_thread(self.hs.get_reactor(), thumbnailer.\n crop, t_width, t_height, t_type)\n", "if t_method == 'scale':\n", "t_width = min(m_width, t_width)\n", "if not t_byte_source:\n", "t_byte_source = await defer_to_thread(self.hs.get_reactor(), thumbnailer.\n scale, t_width, t_height, t_type)\n", "logger.error('Unrecognized method: %r', t_method)\n", "t_height = min(m_height, t_height)\n", "file_info = FileInfo(server_name=server_name, file_id=file_id, thumbnail=\n True, thumbnail_width=t_width, thumbnail_height=t_height,\n thumbnail_method=t_method, thumbnail_type=t_type, url_cache=url_cache)\n", "thumbnails[t_width, t_height, r_type] = r_method\n", "await self.media_storage.write_to_file(t_byte_source, f)\n", "t_byte_source.close()\n", "t_len = os.path.getsize(fname)\n", "await finish()\n", "if server_name:\n", "await self.store.store_local_thumbnail(media_id, t_width, t_height, t_type,\n t_method, t_len)\n", "await self.store.store_remote_media_thumbnail(server_name, media_id,\n file_id, t_width, t_height, t_type, t_method, t_len)\n", "thumbnail_exists = await self.store.get_remote_media_thumbnail(server_name,\n media_id, t_width, t_height, t_type)\n", "if not thumbnail_exists:\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Condition", "Return'", "Assign'", "Assign'", "Expr'", "Assign'", "Return'", "Assign'", "Condition", "Expr'", "Condition", "Return'", "Assign'", "Assign'", "For", "Condition", "For", "Expr'", "Condition", "Condition", "Return'", "Assign'", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Expr'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'", "Assign'", "Expr'", "Condition", "Expr'", "Expr'", "Assign'", "Condition" ]
[ "def FUNC_13(self, VAR_26: str, *, VAR_29: Optional[Type[Any]]=None, VAR_13:...\n", "return self.api_route(VAR_26=path, VAR_29=response_model, VAR_13=\n status_code, VAR_30=tags, VAR_31=dependencies, VAR_32=summary, VAR_33=\n description, VAR_34=response_description, VAR_35=responses, VAR_36=\n deprecated, VAR_37=['PUT'], VAR_38=operation_id, VAR_16=\n response_model_include, VAR_17=response_model_exclude, VAR_18=\n response_model_by_alias, VAR_19=response_model_exclude_unset, VAR_20=\n response_model_exclude_defaults, VAR_21=response_model_exclude_none,\n VAR_39=include_in_schema, VAR_14=response_class, VAR_28=name, VAR_40=\n callbacks)\n" ]
[ "def put(self, path: str, *, response_model: Optional[Type[Any]]=None,...\n", "return self.api_route(path=path, response_model=response_model, status_code\n =status_code, tags=tags, dependencies=dependencies, summary=summary,\n description=description, response_description=response_description,\n responses=responses, deprecated=deprecated, methods=['PUT'],\n operation_id=operation_id, response_model_include=\n response_model_include, response_model_exclude=response_model_exclude,\n response_model_by_alias=response_model_by_alias,\n response_model_exclude_unset=response_model_exclude_unset,\n response_model_exclude_defaults=response_model_exclude_defaults,\n response_model_exclude_none=response_model_exclude_none,\n include_in_schema=include_in_schema, response_class=response_class,\n name=name, callbacks=callbacks)\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "def FUNC_84(VAR_14):...\n", "for VAR_13 in VAR_43:\n", "VAR_21 = getattr(self.rules, VAR_13)\n", "return False\n", "VAR_44 = VAR_21.match(VAR_14)\n", "if not VAR_44:\n", "self.line_match = VAR_44\n", "VAR_87 = getattr(self, 'output_%s' % VAR_13)(VAR_44)\n", "if VAR_87 is not None:\n", "return VAR_44, VAR_87\n" ]
[ "def manipulate(text):...\n", "for key in rules:\n", "pattern = getattr(self.rules, key)\n", "return False\n", "m = pattern.match(text)\n", "if not m:\n", "self.line_match = m\n", "out = getattr(self, 'output_%s' % key)(m)\n", "if out is not None:\n", "return m, out\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "For", "Assign'", "Return'", "Assign'", "Condition", "Assign'", "Assign'", "Condition", "Return'" ]
[ "def FUNC_3(self):...\n", "\"\"\"docstring\"\"\"\n", "def FUNC_16(*VAR_8):...\n", "def FUNC_15(VAR_3, **VAR_4):...\n", "VAR_15 = Deferred()\n", "VAR_15.addCallback(FUNC_16)\n", "self.reactor.callLater(1, VAR_15.callback, True)\n", "return make_deferred_yieldable(VAR_15)\n" ]
[ "def test_callback_indirect_exception(self):...\n", "\"\"\"docstring\"\"\"\n", "def _throw(*args):...\n", "def _callback(request, **kwargs):...\n", "d = Deferred()\n", "d.addCallback(_throw)\n", "self.reactor.callLater(1, d.callback, True)\n", "return make_deferred_yieldable(d)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "FunctionDef'", "FunctionDef'", "Assign'", "Expr'", "Expr'", "Return'" ]
[ "def FUNC_0(self):...\n", "VAR_0 = self.getHeader(b'host')\n", "if VAR_0:\n", "if VAR_0[0] == '[':\n", "return self.getHost().host.encode('ascii')\n", "return VAR_0.split(']', 1)[0] + ']'\n", "return VAR_0.split(':', 1)[0].encode('ascii')\n" ]
[ "def new_getRequestHostname(self):...\n", "host = self.getHeader(b'host')\n", "if host:\n", "if host[0] == '[':\n", "return self.getHost().host.encode('ascii')\n", "return host.split(']', 1)[0] + ']'\n", "return host.split(':', 1)[0].encode('ascii')\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Condition", "Condition", "Return'", "Return'", "Return'" ]
[ "def FUNC_28(self, VAR_4, VAR_39, VAR_40):...\n", "\"\"\"docstring\"\"\"\n", "if not os.path.isabs(VAR_40):\n", "self.notebook_dir = os.path.abspath(VAR_40)\n", "if not os.path.isdir(VAR_40):\n", "return\n", "self.config.FileContentsManager.root_dir = VAR_40\n", "self.config.MappingKernelManager.root_dir = VAR_40\n" ]
[ "def _notebook_dir_changed(self, name, old, new):...\n", "\"\"\"docstring\"\"\"\n", "if not os.path.isabs(new):\n", "self.notebook_dir = os.path.abspath(new)\n", "if not os.path.isdir(new):\n", "return\n", "self.config.FileContentsManager.root_dir = new\n", "self.config.MappingKernelManager.root_dir = new\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Assign'", "Condition", "Return'", "Assign'", "Assign'" ]
[ "def FUNC_30(VAR_16, VAR_35, VAR_36=True):...\n", "VAR_63 = VAR_35.split('&')\n", "VAR_63 = list(map(lambda it: it.strip().replace(',', '|'), VAR_63))\n", "VAR_63 = helper.uniq(VAR_63)\n", "if VAR_63 == ['']:\n", "VAR_63 = [_(u'Unknown')]\n", "VAR_64 = FUNC_7(VAR_63, VAR_16.authors, db.Authors, calibre_db.session,\n 'author')\n", "VAR_65 = list()\n", "for inp in VAR_63:\n", "VAR_90 = calibre_db.session.query(db.Authors).filter(db.Authors.name == inp\n ).first()\n", "VAR_66 = ' & '.join(VAR_65)\n", "if not VAR_90:\n", "if VAR_16.author_sort != VAR_66 and VAR_36:\n", "VAR_90 = helper.get_sorted_author(inp)\n", "VAR_90 = VAR_90.sort\n", "VAR_16.author_sort = VAR_66\n", "return VAR_63, VAR_64\n", "VAR_65.append(helper.get_sorted_author(VAR_90))\n", "VAR_64 = True\n" ]
[ "def handle_author_on_edit(book, author_name, update_stored=True):...\n", "input_authors = author_name.split('&')\n", "input_authors = list(map(lambda it: it.strip().replace(',', '|'),\n input_authors))\n", "input_authors = helper.uniq(input_authors)\n", "if input_authors == ['']:\n", "input_authors = [_(u'Unknown')]\n", "change = modify_database_object(input_authors, book.authors, db.Authors,\n calibre_db.session, 'author')\n", "sort_authors_list = list()\n", "for inp in input_authors:\n", "stored_author = calibre_db.session.query(db.Authors).filter(db.Authors.name ==\n inp).first()\n", "sort_authors = ' & '.join(sort_authors_list)\n", "if not stored_author:\n", "if book.author_sort != sort_authors and update_stored:\n", "stored_author = helper.get_sorted_author(inp)\n", "stored_author = stored_author.sort\n", "book.author_sort = sort_authors\n", "return input_authors, change\n", "sort_authors_list.append(helper.get_sorted_author(stored_author))\n", "change = True\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "For", "Assign'", "Assign'", "Condition", "Condition", "Assign'", "Assign'", "Assign'", "Return'", "Expr'", "Assign'" ]
[ "def FUNC_55(self):...\n", "\"\"\"docstring\"\"\"\n", "if frappe.flags.in_patch:\n", "return\n", "frappe.publish_realtime('doc_update', {'modified': self.modified, 'doctype':\n self.doctype, 'name': self.name}, VAR_2=self.doctype, docname=self.name,\n after_commit=True)\n", "if not self.meta.get('read_only') and not self.meta.get('issingle'\n", "VAR_90 = {'doctype': self.doctype, 'name': self.name, 'user': frappe.\n session.user}\n", "frappe.publish_realtime('list_update', VAR_90, after_commit=True)\n" ]
[ "def notify_update(self):...\n", "\"\"\"docstring\"\"\"\n", "if frappe.flags.in_patch:\n", "return\n", "frappe.publish_realtime('doc_update', {'modified': self.modified, 'doctype':\n self.doctype, 'name': self.name}, doctype=self.doctype, docname=self.\n name, after_commit=True)\n", "if not self.meta.get('read_only') and not self.meta.get('issingle'\n", "data = {'doctype': self.doctype, 'name': self.name, 'user': frappe.session.user\n }\n", "frappe.publish_realtime('list_update', data, after_commit=True)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Return'", "Expr'", "Condition", "Assign'", "Expr'" ]
[ "def FUNC_88(VAR_212):...\n", "VAR_408 = VAR_5.getAnnotationLinks('Image', parent_ids=imageIds)\n", "VAR_409 = {}\n", "VAR_410 = {}\n", "for VAR_432 in VAR_212:\n", "VAR_409[VAR_432] = []\n", "for VAR_344 in VAR_408:\n", "VAR_383 = VAR_344.getChild()\n", "VAR_411 = []\n", "if VAR_383._obj.__class__ == omero.model.TagAnnotationI:\n", "for VAR_432 in VAR_212:\n", "VAR_410[VAR_383.id] = VAR_383\n", "VAR_411.append({'id': VAR_432, 'tags': VAR_409[VAR_432]})\n", "VAR_164 = []\n", "VAR_409[VAR_344.getParent().id].append(VAR_383)\n", "for tId, VAR_433 in VAR_410.items():\n", "VAR_164.append(VAR_433)\n", "return VAR_411, VAR_164\n" ]
[ "def loadImageTags(imageIds):...\n", "tagLinks = conn.getAnnotationLinks('Image', parent_ids=imageIds)\n", "linkMap = {}\n", "tagMap = {}\n", "for iId in imageIds:\n", "linkMap[iId] = []\n", "for link in tagLinks:\n", "c = link.getChild()\n", "imageTags = []\n", "if c._obj.__class__ == omero.model.TagAnnotationI:\n", "for iId in imageIds:\n", "tagMap[c.id] = c\n", "imageTags.append({'id': iId, 'tags': linkMap[iId]})\n", "tags = []\n", "linkMap[link.getParent().id].append(c)\n", "for tId, t in tagMap.items():\n", "tags.append(t)\n", "return imageTags, tags\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "For", "Assign'", "For", "Assign'", "Assign'", "Condition", "For", "Assign'", "Expr'", "Assign'", "Expr'", "For", "Expr'", "Return'" ]
[ "def FUNC_9(self):...\n", "return 'figshare-{}'.format(self.record_id)\n" ]
[ "def get_build_slug(self):...\n", "return 'figshare-{}'.format(self.record_id)\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_10(VAR_22=False):...\n", "VAR_14 = g.locale.language if g.locale else 'en'\n", "if not VAR_22 and VAR_1.get(VAR_14\n", "return VAR_1[VAR_14], VAR_2, VAR_3\n", "VAR_69 = settings().getBoolean(['server', 'firstRun'])\n", "VAR_18 = defaultdict(lambda : {'order': [], 'entries': {}})\n", "VAR_73 = {'navbar': {'div': lambda VAR_136: 'navbar_plugin_' + VAR_136,\n 'template': lambda VAR_136: VAR_136 + '_navbar.jinja2', 'to_entry': lambda\n VAR_10: VAR_10}, 'sidebar': {'div': lambda VAR_136: 'sidebar_plugin_' +\n VAR_136, 'template': lambda VAR_136: VAR_136 + '_sidebar.jinja2',\n 'to_entry': lambda VAR_10: (VAR_10['name'], VAR_10)}, 'tab': {'div': lambda\n VAR_136: 'tab_plugin_' + VAR_136, 'template': lambda VAR_136: VAR_136 +\n '_tab.jinja2', 'to_entry': lambda VAR_10: (VAR_10['name'], VAR_10)},\n 'settings': {'div': lambda VAR_136: 'settings_plugin_' + VAR_136,\n 'template': lambda VAR_136: VAR_136 + '_settings.jinja2', 'to_entry': \n lambda VAR_10: (VAR_10['name'], VAR_10)}, 'usersettings': {'div': lambda\n VAR_136: 'usersettings_plugin_' + VAR_136, 'template': lambda VAR_136: \n VAR_136 + '_usersettings.jinja2', 'to_entry': lambda VAR_10: (VAR_10[\n 'name'], VAR_10)}, 'wizard': {'div': lambda VAR_136: 'wizard_plugin_' +\n VAR_136, 'template': lambda VAR_136: VAR_136 + '_wizard.jinja2',\n 'to_entry': lambda VAR_10: (VAR_10['name'], VAR_10)}, 'about': {'div': \n lambda VAR_136: 'about_plugin_' + VAR_136, 'template': lambda VAR_136: \n VAR_136 + '_about.jinja2', 'to_entry': lambda VAR_10: (VAR_10['name'],\n VAR_10)}, 'generic': {'template': lambda VAR_136: VAR_136 + '.jinja2',\n 'to_entry': lambda VAR_10: VAR_10}}\n", "def FUNC_31(VAR_37, VAR_38):...\n", "if VAR_37[1].get('_key', None) == 'plugin_corewizard_acl':\n", "return f'0:{to_unicode(VAR_37[0])}'\n", "if VAR_37[1].get('mandatory', False):\n", "return f'1:{to_unicode(VAR_37[0])}'\n", "return f'2:{to_unicode(VAR_37[0])}'\n" ]
[ "def fetch_template_data(refresh=False):...\n", "locale = g.locale.language if g.locale else 'en'\n", "if not refresh and _templates.get(locale\n", "return _templates[locale], _plugin_names, _plugin_vars\n", "first_run = settings().getBoolean(['server', 'firstRun'])\n", "templates = defaultdict(lambda : {'order': [], 'entries': {}})\n", "template_rules = {'navbar': {'div': lambda x: 'navbar_plugin_' + x,\n 'template': lambda x: x + '_navbar.jinja2', 'to_entry': lambda data:\n data}, 'sidebar': {'div': lambda x: 'sidebar_plugin_' + x, 'template': \n lambda x: x + '_sidebar.jinja2', 'to_entry': lambda data: (data['name'],\n data)}, 'tab': {'div': lambda x: 'tab_plugin_' + x, 'template': lambda\n x: x + '_tab.jinja2', 'to_entry': lambda data: (data['name'], data)},\n 'settings': {'div': lambda x: 'settings_plugin_' + x, 'template': lambda\n x: x + '_settings.jinja2', 'to_entry': lambda data: (data['name'], data\n )}, 'usersettings': {'div': lambda x: 'usersettings_plugin_' + x,\n 'template': lambda x: x + '_usersettings.jinja2', 'to_entry': lambda\n data: (data['name'], data)}, 'wizard': {'div': lambda x: \n 'wizard_plugin_' + x, 'template': lambda x: x + '_wizard.jinja2',\n 'to_entry': lambda data: (data['name'], data)}, 'about': {'div': lambda\n x: 'about_plugin_' + x, 'template': lambda x: x + '_about.jinja2',\n 'to_entry': lambda data: (data['name'], data)}, 'generic': {'template':\n lambda x: x + '.jinja2', 'to_entry': lambda data: data}}\n", "def wizard_key_extractor(d, k):...\n", "if d[1].get('_key', None) == 'plugin_corewizard_acl':\n", "return f'0:{to_unicode(d[0])}'\n", "if d[1].get('mandatory', False):\n", "return f'1:{to_unicode(d[0])}'\n", "return f'2:{to_unicode(d[0])}'\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Condition", "Return'", "Assign'", "Assign'", "Assign'", "FunctionDef'", "Condition", "Return'", "Condition", "Return'", "Return'" ]
[ "def __call__(self, *VAR_4, VAR_5: Any=None, VAR_6: Optional[str]=None, **VAR_7...\n", "\"\"\"docstring\"\"\"\n", "return self.create_instance(*VAR_4, VAR_5=default, VAR_6=key, **kwargs)\n" ]
[ "def __call__(self, *args, default: Any=None, key: Optional[str]=None, **kwargs...\n", "\"\"\"docstring\"\"\"\n", "return self.create_instance(*args, default=default, key=key, **kwargs)\n" ]
[ 0, 0, 0 ]
[ "Condition", "Docstring", "Return'" ]
[ "def FUNC_2():...\n", "VAR_10 = VAR_0.post('/items/', data={'name': 'Foo', 'price': 50.5})\n", "assert VAR_10.status_code == 422, VAR_10.text\n", "assert VAR_10.json() == {'detail': [{'ctx': {'colno': 1, 'doc':\n 'name=Foo&price=50.5', 'lineno': 1, 'msg': 'Expecting value', 'pos': 0},\n 'loc': ['body', 0], 'msg': 'Expecting value: line 1 column 1 (char 0)',\n 'type': 'value_error.jsondecode'}]}\n", "VAR_10 = VAR_0.post('/items/', json={'test': 'test2'})\n", "assert VAR_10.status_code == 400, VAR_10.text\n", "assert VAR_10.json() == {'detail': 'There was an error parsing the body'}\n" ]
[ "def test_post_broken_body():...\n", "response = client.post('/items/', data={'name': 'Foo', 'price': 50.5})\n", "assert response.status_code == 422, response.text\n", "assert response.json() == {'detail': [{'ctx': {'colno': 1, 'doc':\n 'name=Foo&price=50.5', 'lineno': 1, 'msg': 'Expecting value', 'pos': 0},\n 'loc': ['body', 0], 'msg': 'Expecting value: line 1 column 1 (char 0)',\n 'type': 'value_error.jsondecode'}]}\n", "response = client.post('/items/', json={'test': 'test2'})\n", "assert response.status_code == 400, response.text\n", "assert response.json() == {'detail': 'There was an error parsing the body'}\n" ]
[ 0, 3, 0, 3, 0, 0, 3 ]
[ "FunctionDef'", "Assign'", "Assert'", "Assert'", "Assign'", "Assert'", "Assert'" ]
[ "def FUNC_0(self):...\n", "self.reactor = ThreadedMemoryReactorClock()\n", "self.hs_clock = Clock(self.reactor)\n", "self.homeserver = setup_test_homeserver(self.addCleanup, http_client=None,\n clock=self.hs_clock, reactor=self.reactor)\n" ]
[ "def setUp(self):...\n", "self.reactor = ThreadedMemoryReactorClock()\n", "self.hs_clock = Clock(self.reactor)\n", "self.homeserver = setup_test_homeserver(self.addCleanup, http_client=None,\n clock=self.hs_clock, reactor=self.reactor)\n" ]
[ 0, 0, 0, 4 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'" ]
[ "@VAR_4.route('/get-quiz')...\n", "if 'quiz_count' not in VAR_17:\n", "VAR_17['quiz_count'] = 0\n", "if 'solved_quiz' not in VAR_17:\n", "VAR_17['total_munhak'] = len(VAR_3)\n", "VAR_17['solved_quiz'] = []\n", "VAR_17['result'] = None\n", "VAR_18 = VAR_17['quiz_count'] + 1\n", "VAR_19 = VAR_17['solved_quiz']\n", "if 'current_munhak' not in VAR_17 or VAR_17['current_munhak'] is None:\n", "VAR_15 = copy.deepcopy(VAR_3)\n", "VAR_16 = {'quiz_no': VAR_18, 'type': '객관식', 'category': VAR_17[\n 'current_munhak']['category'], 'hint': VAR_17['current_munhak']['hint'],\n 'options': [f\"{munhak_row['writer']}, 『{munhak_row['title']}』\" for\n munhak_row in VAR_17['options']], 'total_munhak': len(VAR_3)}\n", "VAR_24 = [munhak_row for munhak_row in VAR_15 if munhak_row.munhak_seq not in\n VAR_19]\n", "print(VAR_16)\n", "if len(VAR_24) == 0:\n", "return render_template('quiz/quiz.html', VAR_16=data)\n", "VAR_17['result'] = True\n", "VAR_25 = random.choice(VAR_24)\n", "return redirect(url_for('result'))\n", "for _ in [munhak_row for munhak_row in VAR_15 if munhak_row.title == VAR_25\n", "VAR_15.remove(_)\n", "random.shuffle(VAR_15)\n", "VAR_26 = VAR_15[0:3] + [VAR_25]\n", "random.shuffle(VAR_26)\n", "VAR_21 = VAR_26.index(VAR_25)\n", "print(VAR_21)\n", "VAR_17['correct'] = VAR_21\n", "VAR_27 = random.choice(VAR_25.keywords)\n", "VAR_27 = VAR_27.replace('\\\\', '')\n", "VAR_17['current_munhak'] = {'munhak_seq': VAR_25.munhak_seq, 'source':\n VAR_25.source, 'category': VAR_25.category, 'hint': VAR_27, 'title':\n VAR_25.title, 'writer': VAR_25.writer}\n", "VAR_17['options'] = [munhak_row._asdict() for munhak_row in VAR_26]\n", "VAR_16 = {'quiz_no': VAR_18, 'type': '객관식', 'category': VAR_25.category,\n 'hint': VAR_27, 'options': [\n f'{munhak_row.writer}, 『{munhak_row.title}』' for munhak_row in VAR_26],\n 'total_munhak': len(VAR_3)}\n", "print(VAR_16)\n", "return render_template('quiz/quiz.html', VAR_16=data)\n" ]
[ "@app.route('/get-quiz')...\n", "if 'quiz_count' not in session:\n", "session['quiz_count'] = 0\n", "if 'solved_quiz' not in session:\n", "session['total_munhak'] = len(munhak_rows_data)\n", "session['solved_quiz'] = []\n", "session['result'] = None\n", "quiz_no = session['quiz_count'] + 1\n", "solved_quiz = session['solved_quiz']\n", "if 'current_munhak' not in session or session['current_munhak'] is None:\n", "munhak_rows = copy.deepcopy(munhak_rows_data)\n", "data = {'quiz_no': quiz_no, 'type': '객관식', 'category': session[\n 'current_munhak']['category'], 'hint': session['current_munhak']['hint'\n ], 'options': [f\"{munhak_row['writer']}, 『{munhak_row['title']}』\" for\n munhak_row in session['options']], 'total_munhak': len(munhak_rows_data)}\n", "not_solved_munhak_rows = [munhak_row for munhak_row in munhak_rows if \n munhak_row.munhak_seq not in solved_quiz]\n", "print(data)\n", "if len(not_solved_munhak_rows) == 0:\n", "return render_template('quiz/quiz.html', data=data)\n", "session['result'] = True\n", "correct_munhak_row = random.choice(not_solved_munhak_rows)\n", "return redirect(url_for('result'))\n", "for _ in [munhak_row for munhak_row in munhak_rows if munhak_row.title ==\n", "munhak_rows.remove(_)\n", "random.shuffle(munhak_rows)\n", "option_munhak_rows = munhak_rows[0:3] + [correct_munhak_row]\n", "random.shuffle(option_munhak_rows)\n", "correct = option_munhak_rows.index(correct_munhak_row)\n", "print(correct)\n", "session['correct'] = correct\n", "hint = random.choice(correct_munhak_row.keywords)\n", "hint = hint.replace('\\\\', '')\n", "session['current_munhak'] = {'munhak_seq': correct_munhak_row.munhak_seq,\n 'source': correct_munhak_row.source, 'category': correct_munhak_row.\n category, 'hint': hint, 'title': correct_munhak_row.title, 'writer':\n correct_munhak_row.writer}\n", "session['options'] = [munhak_row._asdict() for munhak_row in option_munhak_rows\n ]\n", "data = {'quiz_no': quiz_no, 'type': '객관식', 'category': correct_munhak_row.\n category, 'hint': hint, 'options': [\n f'{munhak_row.writer}, 『{munhak_row.title}』' for munhak_row in\n option_munhak_rows], 'total_munhak': len(munhak_rows_data)}\n", "print(data)\n", "return render_template('quiz/quiz.html', data=data)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Condition", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Expr'", "Condition", "Return'", "Assign'", "Assign'", "Return'", "For", "Expr'", "Expr'", "Assign'", "Expr'", "Assign'", "Expr'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Return'" ]
[ "async def FUNC_1(*, VAR_4: Optional[ModelField]=None, VAR_5: Any, VAR_6:...\n", "if VAR_4:\n", "VAR_51 = []\n", "return jsonable_encoder(VAR_5)\n", "VAR_5 = FUNC_0(VAR_5, VAR_1=exclude_unset, VAR_2=exclude_defaults, VAR_3=\n exclude_none)\n", "if VAR_9:\n", "VAR_67, VAR_68 = VAR_4.validate(VAR_5, {}, loc=('response',))\n", "VAR_67, VAR_68 = await run_in_threadpool(VAR_4.validate, VAR_5, {}, loc=(\n 'response',))\n", "if isinstance(VAR_68, ErrorWrapper):\n", "VAR_51.append(VAR_68)\n", "if isinstance(VAR_68, list):\n", "if VAR_51:\n", "VAR_51.extend(VAR_68)\n", "return jsonable_encoder(VAR_67, VAR_6=include, VAR_7=exclude, VAR_8=\n by_alias, VAR_1=exclude_unset, VAR_2=exclude_defaults, VAR_3=exclude_none)\n" ]
[ "async def serialize_response(*, field: Optional[ModelField]=None,...\n", "if field:\n", "errors = []\n", "return jsonable_encoder(response_content)\n", "response_content = _prepare_response_content(response_content,\n exclude_unset=exclude_unset, exclude_defaults=exclude_defaults,\n exclude_none=exclude_none)\n", "if is_coroutine:\n", "value, errors_ = field.validate(response_content, {}, loc=('response',))\n", "value, errors_ = await run_in_threadpool(field.validate, response_content,\n {}, loc=('response',))\n", "if isinstance(errors_, ErrorWrapper):\n", "errors.append(errors_)\n", "if isinstance(errors_, list):\n", "if errors:\n", "errors.extend(errors_)\n", "return jsonable_encoder(value, include=include, exclude=exclude, by_alias=\n by_alias, exclude_unset=exclude_unset, exclude_defaults=\n exclude_defaults, exclude_none=exclude_none)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Condition", "Assign'", "Return'", "Assign'", "Condition", "Assign'", "Assign'", "Condition", "Expr'", "Condition", "Condition", "Expr'", "Return'" ]
[ "def FUNC_1(VAR_2):...\n", "\"\"\"docstring\"\"\"\n", "VAR_19 = enumerate(VAR_2, start=1)\n", "return [('%d) %s' % item) for item in VAR_19]\n" ]
[ "def enumerate_string_list(strings):...\n", "\"\"\"docstring\"\"\"\n", "numbered_strings = enumerate(strings, start=1)\n", "return [('%d) %s' % item) for item in numbered_strings]\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Return'" ]
[ "def FUNC_10(self, VAR_11):...\n", "if not VAR_53.session['admin']:\n", "VAR_62 = self.useroptions.forUser(self.getUserId())\n", "for f in VAR_11:\n", "if not VAR_62.getOptionValue('media.may_download'):\n", "if '/../' in f:\n", "VAR_59 = cherry.config['media.maximum_download_size']\n", "return 'not_permitted'\n", "return 'invalid_file'\n", "if self.model.file_size_within_limit(VAR_11, VAR_59):\n", "return str(e)\n", "return 'ok'\n", "return 'too_big'\n" ]
[ "def download_check_files(self, filelist):...\n", "if not cherrypy.session['admin']:\n", "uo = self.useroptions.forUser(self.getUserId())\n", "for f in filelist:\n", "if not uo.getOptionValue('media.may_download'):\n", "if '/../' in f:\n", "size_limit = cherry.config['media.maximum_download_size']\n", "return 'not_permitted'\n", "return 'invalid_file'\n", "if self.model.file_size_within_limit(filelist, size_limit):\n", "return str(e)\n", "return 'ok'\n", "return 'too_big'\n" ]
[ 0, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Assign'", "For", "Condition", "Condition", "Assign'", "Return'", "Return'", "Condition", "Return'", "Return'", "Return'" ]
[ "def FUNC_20(self):...\n", "return Language.objects.exclude(Q(translation__component=self.component) |\n Q(VAR_9=self.component))\n" ]
[ "def get_lang_objects(self):...\n", "return Language.objects.exclude(Q(translation__component=self.component) |\n Q(component=self.component))\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def __init__(self, VAR_63, VAR_157=None, VAR_158='markmin', VAR_159=False,...\n", "VAR_222 = self.settings = VAR_63.settings.wiki\n", "\"\"\"string\"\"\"\n", "VAR_328 = set(['markmin', 'html'])\n", "VAR_329 = False\n", "if VAR_158 == 'multiple':\n", "VAR_158 = {}\n", "if isinstance(VAR_158, dict):\n", "[VAR_328.add(VAR_199) for VAR_199 in VAR_158]\n", "VAR_222.render = VAR_158\n", "VAR_329 = True\n", "VAR_330 = VAR_222.manage_permissions = VAR_159\n", "VAR_222.force_prefix = VAR_160\n", "VAR_222.restrict_search = VAR_161\n", "VAR_222.extra = VAR_163 or {}\n", "VAR_222.menu_groups = VAR_164\n", "VAR_222.templates = VAR_165\n", "VAR_222.controller = VAR_104\n", "VAR_222.function = VAR_105\n", "VAR_222.groups = list(VAR_63.user_groups.values()\n ) if VAR_167 is None else VAR_167\n", "VAR_101 = VAR_63.db\n", "self.env = VAR_157 or {}\n", "self.env['component'] = CLASS_8.component\n", "self.auth = VAR_63\n", "self.wiki_menu_items = None\n", "if self.auth.user:\n", "self.settings.force_prefix = VAR_160 % self.auth.user\n", "self.settings.force_prefix = VAR_160\n", "self.host = VAR_263.request.env.http_host\n", "VAR_331 = [('wiki_page', {'args': [VAR_1('slug', VAR_279=[IS_SLUG(),\n IS_NOT_IN_DB(db, 'wiki_page.slug')], writable=False), VAR_1('title',\n length=255, unique=True), VAR_1('body', 'text', notnull=True), VAR_1(\n 'tags', 'list:string'), VAR_1('can_read', 'list:string', writable=perms,\n readable=perms, VAR_5=[Wiki.everybody]), VAR_1('can_edit',\n 'list:string', writable=perms, readable=perms, VAR_5=[Wiki.everybody]),\n VAR_1('changelog'), VAR_1('html', 'text', compute=self.get_renderer(),\n readable=False, writable=False), VAR_1('render', VAR_5='markmin',\n readable=show_engine, writable=show_engine, VAR_279=IS_EMPTY_OR(\n IS_IN_SET(engines))), VAR_63.signature], 'vars': {'format': '%(title)s',\n 'migrate': VAR_126}}), ('wiki_tag', {'args': [VAR_1('name'), VAR_1(\n 'wiki_page', 'reference wiki_page'), VAR_63.signature], 'vars': {\n 'format': '%(title)s', 'migrate': VAR_126}}), ('wiki_media', {'args': [\n VAR_1('wiki_page', 'reference wiki_page'), VAR_1('title', VAR_91=True),\n VAR_1('filename', 'upload', VAR_91=True), VAR_63.signature], 'vars': {\n 'format': '%(title)s', 'migrate': VAR_126}})]\n", "for VAR_199, VAR_179 in VAR_331:\n", "VAR_11 = []\n", "if self.settings.templates is None and not self.settings.manage_permissions:\n", "if VAR_199 not in VAR_101.tables():\n", "self.settings.templates = VAR_101.wiki_page.tags.contains('template'\n ) & VAR_101.wiki_page.can_read.contains('everybody')\n", "def FUNC_169(VAR_212, VAR_215, VAR_101=VAR_101):...\n", "VAR_360 = VAR_63.settings.extra_fields\n", "for VAR_440 in (VAR_212.tags or []):\n", "if VAR_360:\n", "VAR_440 = VAR_440.strip().lower()\n", "def FUNC_170(VAR_332, VAR_212, VAR_101=VAR_101):...\n", "if VAR_199 in VAR_360:\n", "VAR_11 += VAR_179['args']\n", "if VAR_440:\n", "VAR_212 = VAR_332.select(VAR_175=(0, 1)).first()\n", "if VAR_360[VAR_199]:\n", "VAR_101.define_table(VAR_199, *VAR_11, **value['vars'])\n", "VAR_101.wiki_tag.insert(VAR_148=tag, wiki_page=id)\n", "VAR_101(VAR_101.wiki_tag.wiki_page == VAR_212.id).delete()\n", "for VAR_177 in VAR_360[VAR_199]:\n", "for VAR_440 in (VAR_212.tags or []):\n", "VAR_11.append(VAR_177)\n", "VAR_440 = VAR_440.strip().lower()\n", "VAR_101.wiki_page._after_insert.append(FUNC_169)\n", "if VAR_440:\n", "VAR_101.wiki_page._after_update.append(FUNC_170)\n", "VAR_101.wiki_tag.insert(VAR_148=tag, wiki_page=page.id)\n", "if VAR_63.user and check_credentials(VAR_263.request, gae_login=False\n", "VAR_405 = VAR_101.auth_group(VAR_146='wiki_editor')\n", "VAR_222.lock_keys = True\n", "VAR_406 = VAR_405.id if VAR_405 else VAR_101.auth_group.insert(VAR_146=\n 'wiki_editor')\n", "VAR_63.add_membership(VAR_406)\n" ]
[ "def __init__(self, auth, env=None, render='markmin', manage_permissions=...\n", "settings = self.settings = auth.settings.wiki\n", "\"\"\"\n Args:\n render:\n\n - \"markmin\"\n - \"html\"\n - `<function>` : Sets a custom render function\n - `dict(html=<function>, markmin=...)`: dict(...) allows\n multiple custom render functions\n - \"multiple\" : Is the same as `{}`. It enables per-record\n formats using builtins\n\n \"\"\"\n", "engines = set(['markmin', 'html'])\n", "show_engine = False\n", "if render == 'multiple':\n", "render = {}\n", "if isinstance(render, dict):\n", "[engines.add(key) for key in render]\n", "settings.render = render\n", "show_engine = True\n", "perms = settings.manage_permissions = manage_permissions\n", "settings.force_prefix = force_prefix\n", "settings.restrict_search = restrict_search\n", "settings.extra = extra or {}\n", "settings.menu_groups = menu_groups\n", "settings.templates = templates\n", "settings.controller = controller\n", "settings.function = function\n", "settings.groups = list(auth.user_groups.values()) if groups is None else groups\n", "db = auth.db\n", "self.env = env or {}\n", "self.env['component'] = Wiki.component\n", "self.auth = auth\n", "self.wiki_menu_items = None\n", "if self.auth.user:\n", "self.settings.force_prefix = force_prefix % self.auth.user\n", "self.settings.force_prefix = force_prefix\n", "self.host = current.request.env.http_host\n", "table_definitions = [('wiki_page', {'args': [Field('slug', requires=[\n IS_SLUG(), IS_NOT_IN_DB(db, 'wiki_page.slug')], writable=False), Field(\n 'title', length=255, unique=True), Field('body', 'text', notnull=True),\n Field('tags', 'list:string'), Field('can_read', 'list:string', writable\n =perms, readable=perms, default=[Wiki.everybody]), Field('can_edit',\n 'list:string', writable=perms, readable=perms, default=[Wiki.everybody]\n ), Field('changelog'), Field('html', 'text', compute=self.get_renderer(\n ), readable=False, writable=False), Field('render', default='markmin',\n readable=show_engine, writable=show_engine, requires=IS_EMPTY_OR(\n IS_IN_SET(engines))), auth.signature], 'vars': {'format': '%(title)s',\n 'migrate': migrate}}), ('wiki_tag', {'args': [Field('name'), Field(\n 'wiki_page', 'reference wiki_page'), auth.signature], 'vars': {'format':\n '%(title)s', 'migrate': migrate}}), ('wiki_media', {'args': [Field(\n 'wiki_page', 'reference wiki_page'), Field('title', required=True),\n Field('filename', 'upload', required=True), auth.signature], 'vars': {\n 'format': '%(title)s', 'migrate': migrate}})]\n", "for key, value in table_definitions:\n", "args = []\n", "if self.settings.templates is None and not self.settings.manage_permissions:\n", "if key not in db.tables():\n", "self.settings.templates = db.wiki_page.tags.contains('template'\n ) & db.wiki_page.can_read.contains('everybody')\n", "def update_tags_insert(page, id, db=db):...\n", "extra_fields = auth.settings.extra_fields\n", "for tag in (page.tags or []):\n", "if extra_fields:\n", "tag = tag.strip().lower()\n", "def update_tags_update(dbset, page, db=db):...\n", "if key in extra_fields:\n", "args += value['args']\n", "if tag:\n", "page = dbset.select(limitby=(0, 1)).first()\n", "if extra_fields[key]:\n", "db.define_table(key, *args, **value['vars'])\n", "db.wiki_tag.insert(name=tag, wiki_page=id)\n", "db(db.wiki_tag.wiki_page == page.id).delete()\n", "for field in extra_fields[key]:\n", "for tag in (page.tags or []):\n", "args.append(field)\n", "tag = tag.strip().lower()\n", "db.wiki_page._after_insert.append(update_tags_insert)\n", "if tag:\n", "db.wiki_page._after_update.append(update_tags_update)\n", "db.wiki_tag.insert(name=tag, wiki_page=page.id)\n", "if auth.user and check_credentials(current.request, gae_login=False\n", "group = db.auth_group(role='wiki_editor')\n", "settings.lock_keys = True\n", "gid = group.id if group else db.auth_group.insert(role='wiki_editor')\n", "auth.add_membership(gid)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Assign'", "Expr'", "Assign'", "Assign'", "Condition", "Assign'", "Condition", "Expr'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "For", "Assign'", "Condition", "Condition", "Assign'", "FunctionDef'", "Assign'", "For", "Condition", "Assign'", "FunctionDef'", "Condition", "AugAssign'", "Condition", "Assign'", "Condition", "Expr'", "Expr'", "Expr'", "For", "For", "Expr'", "Assign'", "Expr'", "Condition", "Expr'", "Expr'", "Condition", "Assign'", "Assign'", "Assign'", "Expr'" ]
[ "@staticmethod...\n", "return CLASS_3.archive(VAR_7, VAR_154=archive_table, VAR_122=current_record)\n" ]
[ "@staticmethod...\n", "return Auth.archive(form, archive_table=archive_table, current_record=\n current_record)\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "def FUNC_4(VAR_9):...\n", "VAR_13 = FUNC_6(VAR_9)\n", "if os.path.isfile(VAR_13 + ATTR_SUFFIX):\n", "return '<object id={} src={} meta={} />'.format(quoteattr(url_for(\n '.get_object_id', VAR_9=object_path)), quoteattr(FUNC_5(VAR_9)),\n quoteattr(url_for('.get_object_meta', VAR_9=object_path)))\n", "return '<object id={} src={} />'.format(quoteattr(url_for('.get_object_id',\n VAR_9=object_path)), quoteattr(FUNC_5(VAR_9)))\n" ]
[ "def _get_object_element(object_path):...\n", "path = _get_obj_absolute_path(object_path)\n", "if os.path.isfile(path + ATTR_SUFFIX):\n", "return '<object id={} src={} meta={} />'.format(quoteattr(url_for(\n '.get_object_id', object_path=object_path)), quoteattr(\n _get_object_src_uri(object_path)), quoteattr(url_for('.get_object_meta',\n object_path=object_path)))\n", "return '<object id={} src={} />'.format(quoteattr(url_for('.get_object_id',\n object_path=object_path)), quoteattr(_get_object_src_uri(object_path)))\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Condition", "Return'", "Return'" ]
[ "def FUNC_1(self):...\n", "assert controller_name(CLASS_8) == 'user'\n" ]
[ "def test_it_strips_controller(self):...\n", "assert controller_name(UserController) == 'user'\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Assert'" ]
[ "async def FUNC_5(self) ->List[Server]:...\n", "\"\"\"docstring\"\"\"\n", "if self._parsed_uri.scheme != b'matrix':\n", "return [Server(VAR_1=self._parsed_uri.host, VAR_19=self._parsed_uri.port)]\n", "VAR_12 = urllib.parse.urlparse(self._parsed_uri.toBytes())\n", "VAR_1 = VAR_12.hostname\n", "VAR_19 = VAR_12.port\n", "if VAR_19 or FUNC_0(VAR_1):\n", "return [Server(VAR_1, VAR_19 or 8448)]\n", "VAR_18 = await self._srv_resolver.resolve_service(b'_matrix._tcp.' + VAR_1)\n", "if VAR_18:\n", "return VAR_18\n", "return [Server(VAR_1, 8448)]\n" ]
[ "async def _resolve_server(self) ->List[Server]:...\n", "\"\"\"docstring\"\"\"\n", "if self._parsed_uri.scheme != b'matrix':\n", "return [Server(host=self._parsed_uri.host, port=self._parsed_uri.port)]\n", "parsed_uri = urllib.parse.urlparse(self._parsed_uri.toBytes())\n", "host = parsed_uri.hostname\n", "port = parsed_uri.port\n", "if port or _is_ip_literal(host):\n", "return [Server(host, port or 8448)]\n", "server_list = await self._srv_resolver.resolve_service(b'_matrix._tcp.' + host)\n", "if server_list:\n", "return server_list\n", "return [Server(host, 8448)]\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "AsyncFunctionDef'", "Docstring", "Condition", "Return'", "Assign'", "Assign'", "Assign'", "Condition", "Return'", "Assign'", "Condition", "Return'", "Return'" ]
[ "@FUNC_0...\n", "return ReceiptsHandler(self)\n" ]
[ "@cache_in_self...\n", "return ReceiptsHandler(self)\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "def FUNC_37(self):...\n", "if not test.is_built_with_xla():\n", "self.skipTest('Skipping test because XLA is not compiled in.')\n", "self.parser = saved_model_cli.create_parser()\n", "VAR_10 = test.test_src_dir_path(VAR_0)\n", "VAR_42 = os.path.join(test.get_temp_dir(), 'aot_compile_cpu_dir')\n", "VAR_11 = self.parser.parse_args(['aot_compile_cpu', '--dir', VAR_10,\n '--tag_set', 'serve', '--output_prefix', VAR_42, '--cpp_class',\n 'Compiled', '--signature_def_key', 'MISSING'])\n", "saved_model_cli.aot_compile_cpu(VAR_11)\n" ]
[ "def testAOTCompileCPUWrongSignatureDefKey(self):...\n", "if not test.is_built_with_xla():\n", "self.skipTest('Skipping test because XLA is not compiled in.')\n", "self.parser = saved_model_cli.create_parser()\n", "base_path = test.test_src_dir_path(SAVED_MODEL_PATH)\n", "output_dir = os.path.join(test.get_temp_dir(), 'aot_compile_cpu_dir')\n", "args = self.parser.parse_args(['aot_compile_cpu', '--dir', base_path,\n '--tag_set', 'serve', '--output_prefix', output_dir, '--cpp_class',\n 'Compiled', '--signature_def_key', 'MISSING'])\n", "saved_model_cli.aot_compile_cpu(args)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Expr'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'" ]
[ "@def_function.function(input_signature=[tensor_spec.TensorSpec(shape=(2048,...\n", "return {'res': VAR_45 + self.var}\n" ]
[ "@def_function.function(input_signature=[tensor_spec.TensorSpec(shape=(2048,...\n", "return {'res': x + self.var}\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "from django.http.response import HttpResponseRedirect\n", "from shuup.utils.excs import Problem\n", "from shuup.xtheme.editing import set_edit_mode\n", "def FUNC_0(VAR_0, VAR_1):...\n", "\"\"\"docstring\"\"\"\n", "VAR_2 = VAR_0.POST.get('path') or VAR_0.META.get('HTTP_REFERER') or '/'\n", "if VAR_1 == 'edit_on' or VAR_1 == 'edit_off':\n", "set_edit_mode(VAR_0, VAR_1.endswith('_on'))\n", "def FUNC_1(VAR_0):...\n", "return HttpResponseRedirect(VAR_2)\n", "\"\"\"docstring\"\"\"\n", "VAR_1 = VAR_0.POST.get('command')\n", "if VAR_1:\n", "VAR_3 = FUNC_0(VAR_0, VAR_1)\n", "if VAR_3:\n", "return VAR_3\n" ]
[ "from django.http.response import HttpResponseRedirect\n", "from shuup.utils.excs import Problem\n", "from shuup.xtheme.editing import set_edit_mode\n", "def handle_command(request, command):...\n", "\"\"\"docstring\"\"\"\n", "path = request.POST.get('path') or request.META.get('HTTP_REFERER') or '/'\n", "if command == 'edit_on' or command == 'edit_off':\n", "set_edit_mode(request, command.endswith('_on'))\n", "def command_dispatch(request):...\n", "return HttpResponseRedirect(path)\n", "\"\"\"docstring\"\"\"\n", "command = request.POST.get('command')\n", "if command:\n", "response = handle_command(request, command)\n", "if response:\n", "return response\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "ImportFrom'", "ImportFrom'", "ImportFrom'", "FunctionDef'", "Docstring", "Assign'", "Condition", "Expr'", "FunctionDef'", "Return'", "Docstring", "Assign'", "Condition", "Assign'", "Condition", "Return'" ]
[ "def FUNC_12(self):...\n", "self.force = bool(self.force)\n" ]
[ "def finalize_options(self):...\n", "self.force = bool(self.force)\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Assign'" ]
[ "def FUNC_38(self):...\n", "VAR_33 = models.Individual.objects.create(name='Subject #1')\n", "models.Individual.objects.create(name='Child', parent=subject1)\n", "VAR_29 = models.Individual._meta.get_field('parent').rel\n", "VAR_27 = widgets.ForeignKeyRawIdWidget(VAR_29, widget_admin_site)\n", "self.assertHTMLEqual(conditional_escape(VAR_27.render('individual_widget',\n VAR_33.pk, attrs={})), 'string' % {'subj1pk': VAR_33.pk})\n" ]
[ "def test_fk_to_self_model_not_in_admin(self):...\n", "subject1 = models.Individual.objects.create(name='Subject #1')\n", "models.Individual.objects.create(name='Child', parent=subject1)\n", "rel = models.Individual._meta.get_field('parent').rel\n", "w = widgets.ForeignKeyRawIdWidget(rel, widget_admin_site)\n", "self.assertHTMLEqual(conditional_escape(w.render('individual_widget',\n subject1.pk, attrs={})), \n '<input type=\"text\" name=\"individual_widget\" value=\"%(subj1pk)s\" />&nbsp;<strong>Individual object</strong>'\n % {'subj1pk': subject1.pk})\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Expr'", "Assign'", "Assign'", "Expr'" ]
[ "def FUNC_5(VAR_11):...\n", "from ..h import find, textContent\n", "if VAR_11.get('data-lt') == '':\n", "return []\n", "if VAR_11.get('data-lt'):\n", "VAR_33 = VAR_11.get('data-lt')\n", "if VAR_11.tag in ('dfn', 'a'):\n", "if VAR_33 in ['|', '||', '|||']:\n", "VAR_24 = [textContent(VAR_11).strip()]\n", "if VAR_11.tag in ('h2', 'h3', 'h4', 'h5', 'h6'):\n", "VAR_24 = [VAR_33]\n", "VAR_24 = [x.strip() for x in VAR_33.split('|')]\n", "if VAR_11.get('data-local-lt'):\n", "VAR_24 = [textContent(find('.content', VAR_11)).strip()]\n", "VAR_31 = [x.strip() for x in VAR_11.get('data-local-lt').split('|')]\n", "VAR_24 = [re.sub('\\\\s+', ' ', x) for x in VAR_24 if x != '']\n", "for VAR_10 in VAR_31:\n", "return VAR_24\n", "if VAR_10 in VAR_24:\n", "VAR_24 += VAR_31\n" ]
[ "def linkTextsFromElement(el):...\n", "from ..h import find, textContent\n", "if el.get('data-lt') == '':\n", "return []\n", "if el.get('data-lt'):\n", "rawText = el.get('data-lt')\n", "if el.tag in ('dfn', 'a'):\n", "if rawText in ['|', '||', '|||']:\n", "texts = [textContent(el).strip()]\n", "if el.tag in ('h2', 'h3', 'h4', 'h5', 'h6'):\n", "texts = [rawText]\n", "texts = [x.strip() for x in rawText.split('|')]\n", "if el.get('data-local-lt'):\n", "texts = [textContent(find('.content', el)).strip()]\n", "localTexts = [x.strip() for x in el.get('data-local-lt').split('|')]\n", "texts = [re.sub('\\\\s+', ' ', x) for x in texts if x != '']\n", "for text in localTexts:\n", "return texts\n", "if text in texts:\n", "texts += localTexts\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "ImportFrom'", "Condition", "Return'", "Condition", "Assign'", "Condition", "Condition", "Assign'", "Condition", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "For", "Return'", "Condition", "AugAssign'" ]
[ "def FUNC_16(VAR_0, VAR_1: FlaskClient):...\n", "\"\"\"docstring\"\"\"\n", "VAR_8 = VAR_1.post('/folders/create', data={'parent_dir': '', 'new_dir':\n 'testing'}, follow_redirects=True)\n", "assert VAR_8.status_code == 200\n", "assert request.args.get('path') == 'testing'\n", "assert 'testing' in get_dirs()\n", "assert b'Folder successfully created' in VAR_8.data\n" ]
[ "def test_create_dir(test_app, client: FlaskClient):...\n", "\"\"\"docstring\"\"\"\n", "resp = client.post('/folders/create', data={'parent_dir': '', 'new_dir':\n 'testing'}, follow_redirects=True)\n", "assert resp.status_code == 200\n", "assert request.args.get('path') == 'testing'\n", "assert 'testing' in get_dirs()\n", "assert b'Folder successfully created' in resp.data\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assert'", "Assert'", "Assert'", "Assert'" ]
[ "@VAR_2.route('/ajax/deletedomain', methods=['POST'])...\n", "" ]
[ "@admi.route('/ajax/deletedomain', methods=['POST'])...\n", "" ]
[ 0, 0 ]
[ "Condition", "Condition" ]
[ "def FUNC_37(self, VAR_15):...\n", "VAR_76 = VAR_53.session['userid'] == VAR_15\n", "if VAR_53.session['admin'] and not VAR_76:\n", "VAR_109 = self.userdb.deleteUser(VAR_15)\n", "return \"You didn't think that would work, did you?\"\n", "return 'success' if VAR_109 else 'failed'\n" ]
[ "def api_userdelete(self, userid):...\n", "is_self = cherrypy.session['userid'] == userid\n", "if cherrypy.session['admin'] and not is_self:\n", "deleted = self.userdb.deleteUser(userid)\n", "return \"You didn't think that would work, did you?\"\n", "return 'success' if deleted else 'failed'\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Condition", "Assign'", "Return'", "Return'" ]
[ "def FUNC_84(VAR_72, *VAR_73):...\n", "def FUNC_86(self, VAR_25, *VAR_0, **VAR_1):...\n", "FUNC_83(self, VAR_72(self, *VAR_0, **kwargs))\n", "for VAR_6 in VAR_73:\n", "FUNC_83(self, VAR_6(self, VAR_25, *VAR_0, **kwargs))\n", "return self._return_value\n" ]
[ "def compose(fn, *hooks):...\n", "def runner(self, method, *args, **kwargs):...\n", "add_to_return_value(self, fn(self, *args, **kwargs))\n", "for f in hooks:\n", "add_to_return_value(self, f(self, method, *args, **kwargs))\n", "return self._return_value\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "FunctionDef'", "Expr'", "For", "Expr'", "Return'" ]
[ "def FUNC_15(self, VAR_23):...\n", "VAR_66 = type(VAR_23)\n", "if isinstance(VAR_23, VAR_72):\n", "VAR_56 = fromstring(VAR_23)\n", "VAR_56 = copy.deepcopy(VAR_23)\n", "self(VAR_56)\n", "return _transform_result(VAR_66, VAR_56)\n" ]
[ "def clean_html(self, html):...\n", "result_type = type(html)\n", "if isinstance(html, basestring):\n", "doc = fromstring(html)\n", "doc = copy.deepcopy(html)\n", "self(doc)\n", "return _transform_result(result_type, doc)\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Condition", "Assign'", "Assign'", "Expr'", "Return'" ]
[ "def FUNC_55(self, VAR_14):...\n", "\"\"\"docstring\"\"\"\n", "return '<em>%s</em>' % VAR_14\n" ]
[ "def emphasis(self, text):...\n", "\"\"\"docstring\"\"\"\n", "return '<em>%s</em>' % text\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Return'" ]
[ "def FUNC_147():...\n", "self.bar = [(VAR_250[0]['name'], False, VAR_250[0]['href'], [])]\n", "for VAR_305 in VAR_250:\n", "self.bar[0][3].append((VAR_305['name'], False, VAR_305['href']))\n" ]
[ "def menu():...\n", "self.bar = [(items[0]['name'], False, items[0]['href'], [])]\n", "for item in items:\n", "self.bar[0][3].append((item['name'], False, item['href']))\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "For", "Expr'" ]
[ "@classmethod...\n", "VAR_15.check_file(VAR_4, VAR_16)\n", "return file_stream.read()\n" ]
[ "@classmethod...\n", "cls.check_file(file_path, config_name)\n", "return file_stream.read()\n" ]
[ 0, 0, 0 ]
[ "Condition", "Expr'", "Return'" ]
[ "def FUNC_15():...\n", "random.seed(VAR_24)\n", "return list(map(lambda x: x + VAR_5 * VAR_41, sorted(random.sample(list(\n range(VAR_5)), VAR_40))))\n" ]
[ "def generate_mix_indices():...\n", "random.seed(seed)\n", "return list(map(lambda x: x + ITEMS_PER_ITERATION * iteration_count, sorted\n (random.sample(list(range(ITEMS_PER_ITERATION)), mix_per_iteration))))\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Return'" ]
[ "@VAR_0.route('/api/jobs/<int:job_id>', methods=['GET'])...\n", "\"\"\"docstring\"\"\"\n", "VAR_94 = FUNC_58(f'/internal/jobs/{VAR_9}', 'get')\n", "return jsonify({'success': False, 'message': str(err)}), 400\n", "return jsonify(VAR_94)\n" ]
[ "@gui.route('/api/jobs/<int:job_id>', methods=['GET'])...\n", "\"\"\"docstring\"\"\"\n", "job_info = query_internal_api(f'/internal/jobs/{job_id}', 'get')\n", "return jsonify({'success': False, 'message': str(err)}), 400\n", "return jsonify(job_info)\n" ]
[ 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Return'", "Return'" ]
[ "@VAR_0.route('/settings', methods=('GET', 'POST'))...\n", "return redirect('settings/general')\n" ]
[ "@blueprint.route('/settings', methods=('GET', 'POST'))...\n", "return redirect('settings/general')\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "def FUNC_34(self, VAR_60, VAR_61, VAR_62, VAR_63=False):...\n", "VAR_89 = self._scheme, VAR_60, VAR_61, None, None, None\n", "VAR_3 = urllib_parse.urlunparse(VAR_89)\n", "VAR_102 = {'Content-Type': 'text/xml'}\n", "VAR_1.critical('HTTP error %s while getting %s', exc.response.status_code,\n VAR_3)\n", "VAR_86 = self._session.post(VAR_3, VAR_23=request_body, VAR_102=headers,\n VAR_34=True)\n", "VAR_86.raise_for_status()\n", "self.verbose = VAR_63\n", "return self.parse_response(VAR_86.raw)\n" ]
[ "def request(self, host, handler, request_body, verbose=False):...\n", "parts = self._scheme, host, handler, None, None, None\n", "url = urllib_parse.urlunparse(parts)\n", "headers = {'Content-Type': 'text/xml'}\n", "logger.critical('HTTP error %s while getting %s', exc.response.status_code, url\n )\n", "response = self._session.post(url, data=request_body, headers=headers,\n stream=True)\n", "response.raise_for_status()\n", "self.verbose = verbose\n", "return self.parse_response(response.raw)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "Expr'", "Assign'", "Return'" ]
[ "def FUNC_41(*VAR_79, **VAR_42):...\n", "if VAR_79 and len(VAR_79) > 1 and isinstance(VAR_79[1], text_type):\n", "VAR_46 = FUNC_42(VAR_79[0], VAR_79[1])\n", "VAR_94 = FUNC_45(*VAR_79, **kwargs)\n", "VAR_94 = VAR_1.document_cache.get(VAR_46)\n", "return VAR_94\n", "if VAR_94:\n", "return VAR_94\n", "VAR_94 = FUNC_10().hget('document_cache', VAR_46)\n", "if VAR_94:\n", "VAR_94 = FUNC_45(VAR_94)\n", "VAR_1.document_cache[VAR_46] = VAR_94\n", "return VAR_94\n" ]
[ "def get_cached_doc(*args, **kwargs):...\n", "if args and len(args) > 1 and isinstance(args[1], text_type):\n", "key = get_document_cache_key(args[0], args[1])\n", "doc = get_doc(*args, **kwargs)\n", "doc = local.document_cache.get(key)\n", "return doc\n", "if doc:\n", "return doc\n", "doc = cache().hget('document_cache', key)\n", "if doc:\n", "doc = get_doc(doc)\n", "local.document_cache[key] = doc\n", "return doc\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Assign'", "Assign'", "Assign'", "Return'", "Condition", "Return'", "Assign'", "Condition", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_17(self):...\n", "VAR_5 = self._makeContext()\n", "self.assertEqual(VAR_5.evaluate('x | string:$one'), '1')\n" ]
[ "def test_hybrid_with_string_expression_and_expansion(self):...\n", "ec = self._makeContext()\n", "self.assertEqual(ec.evaluate('x | string:$one'), '1')\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Expr'" ]
[ "@log_function...\n", "\"\"\"docstring\"\"\"\n", "VAR_2 = FUNC_1('/groups/%s/categories/%s', VAR_30, VAR_33)\n", "return self.client.delete_json(VAR_5=destination, VAR_2=path, VAR_3={\n 'requester_user_id': requester_user_id}, VAR_15=True)\n" ]
[ "@log_function...\n", "\"\"\"docstring\"\"\"\n", "path = _create_v1_path('/groups/%s/categories/%s', group_id, category_id)\n", "return self.client.delete_json(destination=destination, path=path, args={\n 'requester_user_id': requester_user_id}, ignore_backoff=True)\n" ]
[ 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Return'" ]
[ "import copy\n", "import json\n", "import re\n", "from datetime import date, datetime, timedelta\n", "from typing import Dict, List\n", "from crispy_forms.bootstrap import InlineCheckboxes, InlineRadios, Tab, TabHolder\n", "from crispy_forms.helper import FormHelper\n", "from crispy_forms.layout import Div, Field, Fieldset, Layout\n", "from django import forms\n", "from django.conf import settings\n", "from django.core.exceptions import NON_FIELD_ERRORS, PermissionDenied, ValidationError\n", "from django.core.validators import FileExtensionValidator\n", "from django.db.models import Q\n", "from django.forms import model_to_dict\n", "from django.forms.models import ModelChoiceIterator\n", "from django.forms.utils import from_current_timezone\n", "from django.template.loader import render_to_string\n", "from django.urls import reverse\n", "from django.utils import timezone\n", "from django.utils.http import urlencode\n", "from django.utils.safestring import mark_safe\n", "from django.utils.translation import gettext\n", "from django.utils.translation import gettext_lazy as _\n", "from translation_finder import DiscoveryResult, discover\n", "from weblate.auth.models import Group, User\n", "from weblate.checks.flags import Flags\n", "from weblate.checks.models import CHECKS\n", "from weblate.checks.utils import highlight_string\n", "from weblate.formats.models import EXPORTERS, FILE_FORMATS\n", "from weblate.glossary.forms import GlossaryAddMixin\n", "from weblate.lang.data import BASIC_LANGUAGES\n", "from weblate.lang.models import Language\n", "from weblate.machinery import MACHINE_TRANSLATION_SERVICES\n", "from weblate.trans.defines import COMPONENT_NAME_LENGTH, REPO_LENGTH\n", "from weblate.trans.filter import FILTERS, get_filter_choice\n", "from weblate.trans.models import Announcement, Change, Component, Label, Project, ProjectToken, Unit\n", "from weblate.trans.specialchars import RTL_CHARS_DATA, get_special_chars\n", "from weblate.trans.util import check_upload_method_permissions, is_repo_link\n", "from weblate.trans.validators import validate_check_flags\n", "from weblate.utils.antispam import is_spam\n", "from weblate.utils.errors import report_error\n", "from weblate.utils.forms import ColorWidget, ContextDiv, EmailField, SearchField, SortedSelect, SortedSelectMultiple, UsernameField\n", "from weblate.utils.hash import checksum_to_hash, hash_to_checksum\n", "from weblate.utils.search import parse_query\n", "from weblate.utils.state import STATE_APPROVED, STATE_CHOICES, STATE_EMPTY, STATE_FUZZY, STATE_READONLY, STATE_TRANSLATED\n", "from weblate.utils.validators import validate_file_extension\n", "from weblate.vcs.models import VCS_REGISTRY\n", "VAR_0 = \"\"\"\n<button class=\"btn btn-default {0}\" title=\"{1}\" {2}>{3}</button>\n\"\"\"\n", "VAR_1 = 'string'\n", "VAR_2 = \"\"\"\n<div class=\"btn-group btn-group-xs\" {0}>{1}</div>\n\"\"\"\n", "VAR_3 = \"\"\"\n<div class=\"btn-toolbar pull-right flip editor-toolbar\">{0}</div>\n\"\"\"\n", "def __init__(self, **VAR_7):...\n", "VAR_7['attrs'] = {'dir': 'auto', 'class':\n 'markdown-editor highlight-editor', 'data-mode': 'markdown'}\n", "super().__init__(**kwargs)\n", "def __init__(self, VAR_12=True, **VAR_7):...\n", "VAR_21 = {'type': 'date'}\n", "if VAR_12:\n", "VAR_21['data-provide'] = 'datepicker'\n", "super().__init__(VAR_21=attrs, VAR_36='%Y-%m-%d', **kwargs)\n", "VAR_21['data-date-format'] = 'yyyy-mm-dd'\n", "def __init__(self, VAR_12=True, **VAR_7):...\n", "if 'widget' not in VAR_7:\n", "VAR_7['widget'] = CLASS_1(VAR_12=datepicker)\n", "super().__init__(**kwargs)\n", "def FUNC_3(self, VAR_13):...\n", "\"\"\"docstring\"\"\"\n", "VAR_13 = super().to_python(VAR_13)\n", "if isinstance(VAR_13, date):\n", "return from_current_timezone(datetime(VAR_13.year, VAR_13.month, VAR_13.day,\n 0, 0, 0))\n", "return VAR_13\n" ]
[ "import copy\n", "import json\n", "import re\n", "from datetime import date, datetime, timedelta\n", "from typing import Dict, List\n", "from crispy_forms.bootstrap import InlineCheckboxes, InlineRadios, Tab, TabHolder\n", "from crispy_forms.helper import FormHelper\n", "from crispy_forms.layout import Div, Field, Fieldset, Layout\n", "from django import forms\n", "from django.conf import settings\n", "from django.core.exceptions import NON_FIELD_ERRORS, PermissionDenied, ValidationError\n", "from django.core.validators import FileExtensionValidator\n", "from django.db.models import Q\n", "from django.forms import model_to_dict\n", "from django.forms.models import ModelChoiceIterator\n", "from django.forms.utils import from_current_timezone\n", "from django.template.loader import render_to_string\n", "from django.urls import reverse\n", "from django.utils import timezone\n", "from django.utils.http import urlencode\n", "from django.utils.safestring import mark_safe\n", "from django.utils.translation import gettext\n", "from django.utils.translation import gettext_lazy as _\n", "from translation_finder import DiscoveryResult, discover\n", "from weblate.auth.models import Group, User\n", "from weblate.checks.flags import Flags\n", "from weblate.checks.models import CHECKS\n", "from weblate.checks.utils import highlight_string\n", "from weblate.formats.models import EXPORTERS, FILE_FORMATS\n", "from weblate.glossary.forms import GlossaryAddMixin\n", "from weblate.lang.data import BASIC_LANGUAGES\n", "from weblate.lang.models import Language\n", "from weblate.machinery import MACHINE_TRANSLATION_SERVICES\n", "from weblate.trans.defines import COMPONENT_NAME_LENGTH, REPO_LENGTH\n", "from weblate.trans.filter import FILTERS, get_filter_choice\n", "from weblate.trans.models import Announcement, Change, Component, Label, Project, ProjectToken, Unit\n", "from weblate.trans.specialchars import RTL_CHARS_DATA, get_special_chars\n", "from weblate.trans.util import check_upload_method_permissions, is_repo_link\n", "from weblate.trans.validators import validate_check_flags\n", "from weblate.utils.antispam import is_spam\n", "from weblate.utils.errors import report_error\n", "from weblate.utils.forms import ColorWidget, ContextDiv, EmailField, SearchField, SortedSelect, SortedSelectMultiple, UsernameField\n", "from weblate.utils.hash import checksum_to_hash, hash_to_checksum\n", "from weblate.utils.search import parse_query\n", "from weblate.utils.state import STATE_APPROVED, STATE_CHOICES, STATE_EMPTY, STATE_FUZZY, STATE_READONLY, STATE_TRANSLATED\n", "from weblate.utils.validators import validate_file_extension\n", "from weblate.vcs.models import VCS_REGISTRY\n", "BUTTON_TEMPLATE = \"\"\"\n<button class=\"btn btn-default {0}\" title=\"{1}\" {2}>{3}</button>\n\"\"\"\n", "RADIO_TEMPLATE = \"\"\"\n<label class=\"btn btn-default {0}\" title=\"{1}\">\n<input type=\"radio\" name=\"{2}\" value=\"{3}\" {4}/>\n{5}\n</label>\n\"\"\"\n", "GROUP_TEMPLATE = \"\"\"\n<div class=\"btn-group btn-group-xs\" {0}>{1}</div>\n\"\"\"\n", "TOOLBAR_TEMPLATE = \"\"\"\n<div class=\"btn-toolbar pull-right flip editor-toolbar\">{0}</div>\n\"\"\"\n", "def __init__(self, **kwargs):...\n", "kwargs['attrs'] = {'dir': 'auto', 'class':\n 'markdown-editor highlight-editor', 'data-mode': 'markdown'}\n", "super().__init__(**kwargs)\n", "def __init__(self, datepicker=True, **kwargs):...\n", "attrs = {'type': 'date'}\n", "if datepicker:\n", "attrs['data-provide'] = 'datepicker'\n", "super().__init__(attrs=attrs, format='%Y-%m-%d', **kwargs)\n", "attrs['data-date-format'] = 'yyyy-mm-dd'\n", "def __init__(self, datepicker=True, **kwargs):...\n", "if 'widget' not in kwargs:\n", "kwargs['widget'] = WeblateDateInput(datepicker=datepicker)\n", "super().__init__(**kwargs)\n", "def to_python(self, value):...\n", "\"\"\"docstring\"\"\"\n", "value = super().to_python(value)\n", "if isinstance(value, date):\n", "return from_current_timezone(datetime(value.year, value.month, value.day, 0,\n 0, 0))\n", "return value\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Import'", "Import'", "Import'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Assign'", "Assign'", "Assign'", "Assign'", "FunctionDef'", "Assign'", "Expr'", "FunctionDef'", "Assign'", "Condition", "Assign'", "Expr'", "Assign'", "FunctionDef'", "Condition", "Assign'", "Expr'", "FunctionDef'", "Docstring", "Assign'", "Condition", "Return'", "Return'" ]
[ "@pytest.fixture...\n", "VAR_3 = tempfile.mkdtemp()\n", "VAR_4 = os.path.join(VAR_3, '%s.txt' % VAR_0.node.name)\n", "f.write(VAR_0.node.name)\n", "return VAR_4\n" ]
[ "@pytest.fixture...\n", "path = tempfile.mkdtemp()\n", "file_name = os.path.join(path, '%s.txt' % request.node.name)\n", "f.write(request.node.name)\n", "return file_name\n" ]
[ 1, 1, 1, 0, 1 ]
[ "Condition", "Assign'", "Assign'", "Expr'", "Return'" ]
[ "def __call__(self):...\n", "" ]
[ "def __call__(self):...\n", "" ]
[ 0, 0 ]
[ "FunctionDef'", "Condition" ]
[ "def FUNC_5(self):...\n", "from ...plugins.CookieAuthHelper import CookieAuthHelper\n", "return CookieAuthHelper\n" ]
[ "def _getTargetClass(self):...\n", "from ...plugins.CookieAuthHelper import CookieAuthHelper\n", "return CookieAuthHelper\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "ImportFrom'", "Return'" ]
[ "def FUNC_3(self, *VAR_4, **VAR_5):...\n", "self.getBaseUrl(VAR_3=True)\n", "VAR_48 = 0 == self.userdb.getUserCount()\n", "VAR_49 = self.mainpage\n", "if 'devel' in VAR_5:\n", "VAR_49 = readRes('res/devel.html')\n", "if 'login' in VAR_5:\n", "self.loginpage = readRes('res/login.html')\n", "VAR_6 = VAR_5.get('username', '')\n", "if VAR_48:\n", "self.firstrunpage = readRes('res/firstrun.html')\n", "VAR_7 = VAR_5.get('password', '')\n", "return self.firstrunpage\n", "if self.isAuthorized():\n", "VAR_89 = VAR_5.get('login', '')\n", "return VAR_49\n", "return self.loginpage\n", "if VAR_89 == 'login':\n", "self.session_auth(VAR_6, VAR_7)\n", "if VAR_89 == 'create admin user':\n", "if VAR_53.session['username']:\n", "if VAR_48:\n", "VAR_6 = VAR_53.session['username']\n", "if VAR_6.strip() and VAR_7.strip():\n", "return \"No, you can't.\"\n", "log.i(_('user {name} just logged in.').format(VAR_81=username))\n", "self.userdb.addUser(VAR_6, VAR_7, True)\n", "self.session_auth(VAR_6, VAR_7)\n", "return VAR_49\n" ]
[ "def index(self, *args, **kwargs):...\n", "self.getBaseUrl(redirect_unencrypted=True)\n", "firstrun = 0 == self.userdb.getUserCount()\n", "show_page = self.mainpage\n", "if 'devel' in kwargs:\n", "show_page = readRes('res/devel.html')\n", "if 'login' in kwargs:\n", "self.loginpage = readRes('res/login.html')\n", "username = kwargs.get('username', '')\n", "if firstrun:\n", "self.firstrunpage = readRes('res/firstrun.html')\n", "password = kwargs.get('password', '')\n", "return self.firstrunpage\n", "if self.isAuthorized():\n", "login_action = kwargs.get('login', '')\n", "return show_page\n", "return self.loginpage\n", "if login_action == 'login':\n", "self.session_auth(username, password)\n", "if login_action == 'create admin user':\n", "if cherrypy.session['username']:\n", "if firstrun:\n", "username = cherrypy.session['username']\n", "if username.strip() and password.strip():\n", "return \"No, you can't.\"\n", "log.i(_('user {name} just logged in.').format(name=username))\n", "self.userdb.addUser(username, password, True)\n", "self.session_auth(username, password)\n", "return show_page\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Assign'", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Return'", "Condition", "Assign'", "Return'", "Return'", "Condition", "Expr'", "Condition", "Condition", "Condition", "Assign'", "Condition", "Return'", "Expr'", "Expr'", "Expr'", "Return'" ]
[ "def FUNC_5(self):...\n", "if not getattr(self, 'latest', None):\n", "self.latest = frappe.get_doc(self.doctype, self.name)\n", "return self.latest\n" ]
[ "def get_latest(self):...\n", "if not getattr(self, 'latest', None):\n", "self.latest = frappe.get_doc(self.doctype, self.name)\n", "return self.latest\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Assign'", "Return'" ]
[ "def FUNC_5(VAR_11):...\n", "\"\"\"docstring\"\"\"\n", "VAR_29 = VAR_11.get_command_obj('build_py')\n", "VAR_11.package_data = FUNC_2()\n", "VAR_29.finalize_options()\n" ]
[ "def update_package_data(distribution):...\n", "\"\"\"docstring\"\"\"\n", "build_py = distribution.get_command_obj('build_py')\n", "distribution.package_data = find_package_data()\n", "build_py.finalize_options()\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Expr'" ]
[ "def FUNC_11(VAR_16):...\n", "return {VAR_22.domain for VAR_22 in self.room_members}\n" ]
[ "def get_joined_hosts_for_room(room_id):...\n", "return {member.domain for member in self.room_members}\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_0(VAR_2):...\n", "return Static(template('obj', VAR_2=Symbol(obj), mode='eval'))\n" ]
[ "def static(obj):...\n", "return Static(template('obj', obj=Symbol(obj), mode='eval'))\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_0(**VAR_1):...\n", "VAR_0.clear()\n" ]
[ "def clear_view_cache(**kwargs):...\n", "_VIEW_CACHE.clear()\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Expr'" ]
[ "@app.route('/login', methods=['GET', 'POST'])...\n", "VAR_9 = forms.UserForm()\n", "if VAR_9.validate_on_submit():\n", "VAR_33 = get_db()\n", "return render_template('users/login.html', VAR_9=form, title='Login')\n", "VAR_34 = VAR_33.search((Query().username == VAR_9.username.data) & (Query()\n .type == 'user'))\n", "if VAR_34 and check_password_hash(VAR_34[0]['hashed_password'], VAR_9.\n", "VAR_34 = User.from_db(VAR_34[0])\n", "flash('Invalid credentials', 'error')\n", "login_user(VAR_34, remember=True)\n", "return redirect('/login')\n", "flash('Login successful!', 'success')\n", "VAR_39 = request.args.get('next')\n", "return redirect(VAR_39 or '/')\n" ]
[ "@app.route('/login', methods=['GET', 'POST'])...\n", "form = forms.UserForm()\n", "if form.validate_on_submit():\n", "db = get_db()\n", "return render_template('users/login.html', form=form, title='Login')\n", "user = db.search((Query().username == form.username.data) & (Query().type ==\n 'user'))\n", "if user and check_password_hash(user[0]['hashed_password'], form.password.data\n", "user = User.from_db(user[0])\n", "flash('Invalid credentials', 'error')\n", "login_user(user, remember=True)\n", "return redirect('/login')\n", "flash('Login successful!', 'success')\n", "next_url = request.args.get('next')\n", "return redirect(next_url or '/')\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 4 ]
[ "Condition", "Assign'", "Condition", "Assign'", "Return'", "Assign'", "Condition", "Assign'", "Expr'", "Expr'", "Return'", "Expr'", "Assign'", "Return'" ]
[ "def FUNC_111(VAR_129):...\n", "from frappe.desk.search import validate_and_sanitize_search_inputs as func\n", "return func(VAR_129)\n" ]
[ "def validate_and_sanitize_search_inputs(fn):...\n", "from frappe.desk.search import validate_and_sanitize_search_inputs as func\n", "return func(fn)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "ImportFrom'", "Return'" ]
[ "import os\n", "import datetime\n", "from xml.sax.saxutils import quoteattr\n", "from flask import Blueprint, url_for, Response, stream_with_context, send_file, jsonify\n", "from werkzeug.datastructures import Headers\n", "from opendiamond.dataretriever.util import ATTR_SUFFIX\n", "VAR_0 = 'collection'\n", "VAR_1 = False\n", "VAR_2 = True\n", "VAR_3 = VAR_4 = None\n", "def FUNC_0(VAR_5):...\n", "VAR_3 = VAR_5.indexdir\n", "VAR_4 = VAR_5.dataroot\n", "VAR_6 = Blueprint('diamond_store', __name__)\n", "@VAR_6.route('/<gididx>')...\n", "VAR_11 = 'GIDIDX' + VAR_7.upper()\n", "VAR_11 = FUNC_7(VAR_11)\n", "def FUNC_9():...\n", "VAR_16 = 0\n", "for _ in f.readlines():\n", "VAR_16 += 1\n", "yield '<?xml version=\"1.0\" encoding=\"UTF-8\" ?>\\n'\n", "if VAR_8 is not None and VAR_16 >= VAR_8:\n", "if VAR_1:\n", "yield '<?xml-stylesheet type=\"text/xsl\" href=\"/scopelist.xsl\" ?>\\n'\n", "yield '<objectlist count=\"{:d}\">\\n'.format(VAR_16)\n", "VAR_17 = 0\n", "for VAR_13 in f.readlines():\n", "VAR_13 = VAR_13.strip()\n", "yield '</objectlist>\\n'\n", "yield FUNC_4(VAR_9=path) + '\\n'\n", "VAR_12 = Headers([('Content-Type', 'text/xml')])\n", "VAR_17 += 1\n", "return Response(stream_with_context(FUNC_9()), status='200 OK', VAR_12=headers)\n", "if VAR_8 is not None and VAR_17 >= VAR_8:\n" ]
[ "import os\n", "import datetime\n", "from xml.sax.saxutils import quoteattr\n", "from flask import Blueprint, url_for, Response, stream_with_context, send_file, jsonify\n", "from werkzeug.datastructures import Headers\n", "from opendiamond.dataretriever.util import ATTR_SUFFIX\n", "BASEURL = 'collection'\n", "STYLE = False\n", "LOCAL_OBJ_URI = True\n", "INDEXDIR = DATAROOT = None\n", "def init(config):...\n", "INDEXDIR = config.indexdir\n", "DATAROOT = config.dataroot\n", "scope_blueprint = Blueprint('diamond_store', __name__)\n", "@scope_blueprint.route('/<gididx>')...\n", "index = 'GIDIDX' + gididx.upper()\n", "index = _get_index_absolute_path(index)\n", "def generate():...\n", "num_entries = 0\n", "for _ in f.readlines():\n", "num_entries += 1\n", "yield '<?xml version=\"1.0\" encoding=\"UTF-8\" ?>\\n'\n", "if limit is not None and num_entries >= limit:\n", "if STYLE:\n", "yield '<?xml-stylesheet type=\"text/xsl\" href=\"/scopelist.xsl\" ?>\\n'\n", "yield '<objectlist count=\"{:d}\">\\n'.format(num_entries)\n", "count = 0\n", "for path in f.readlines():\n", "path = path.strip()\n", "yield '</objectlist>\\n'\n", "yield _get_object_element(object_path=path) + '\\n'\n", "headers = Headers([('Content-Type', 'text/xml')])\n", "count += 1\n", "return Response(stream_with_context(generate()), status='200 OK', headers=\n headers)\n", "if limit is not None and count >= limit:\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Import'", "Import'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Assign'", "Assign'", "Assign'", "Assign'", "FunctionDef'", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "FunctionDef'", "Assign'", "For", "AugAssign'", "Expr'", "Condition", "Condition", "Expr'", "Expr'", "Assign'", "For", "Assign'", "Expr'", "Expr'", "Assign'", "AugAssign'", "Return'", "Condition" ]
[ "def FUNC_30(self):...\n", "self.login()\n", "VAR_3 = self.client.post('/password_change/', {'old_password': 'password',\n 'new_password1': 'password1', 'new_password2': 'donuts'})\n", "self.assertFormError(VAR_3, SetPasswordForm.error_messages['password_mismatch']\n )\n" ]
[ "def test_password_change_fails_with_mismatched_passwords(self):...\n", "self.login()\n", "response = self.client.post('/password_change/', {'old_password':\n 'password', 'new_password1': 'password1', 'new_password2': 'donuts'})\n", "self.assertFormError(response, SetPasswordForm.error_messages[\n 'password_mismatch'])\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Assign'", "Expr'" ]
[ "from mock import Mock\n", "from twisted.internet import defer\n", "import synapse.types\n", "from synapse.api.errors import AuthError, SynapseError\n", "from synapse.types import UserID\n", "from tests import unittest\n", "from tests.test_utils import make_awaitable\n", "from tests.utils import setup_test_homeserver\n", "\"\"\" Tests profile management. \"\"\"\n", "@defer.inlineCallbacks...\n", "self.mock_federation = Mock()\n", "self.mock_registry = Mock()\n", "self.query_handlers = {}\n", "def FUNC_10(VAR_0, VAR_1):...\n", "self.query_handlers[VAR_0] = VAR_1\n", "self.mock_registry.register_query_handler = FUNC_10\n", "VAR_2 = yield setup_test_homeserver(self.addCleanup, http_client=None,\n resource_for_federation=Mock(), federation_client=self.mock_federation,\n federation_server=Mock(), federation_registry=self.mock_registry)\n", "self.store = VAR_2.get_datastore()\n", "self.frank = UserID.from_string('@1234ABCD:test')\n", "self.bob = UserID.from_string('@4567:test')\n", "self.alice = UserID.from_string('@alice:remote')\n", "yield defer.ensureDeferred(self.store.create_profile(self.frank.localpart))\n", "self.handler = VAR_2.get_profile_handler()\n", "self.hs = VAR_2\n", "@defer.inlineCallbacks...\n", "yield defer.ensureDeferred(self.store.set_profile_displayname(self.frank.\n localpart, 'Frank'))\n", "VAR_3 = yield defer.ensureDeferred(self.handler.get_displayname(self.frank))\n", "self.assertEquals('Frank', VAR_3)\n", "@defer.inlineCallbacks...\n", "yield defer.ensureDeferred(self.handler.set_displayname(self.frank, synapse\n .types.create_requester(self.frank), 'Frank Jr.'))\n", "self.assertEquals((yield defer.ensureDeferred(self.store.\n get_profile_displayname(self.frank.localpart))), 'Frank Jr.')\n", "yield defer.ensureDeferred(self.handler.set_displayname(self.frank, synapse\n .types.create_requester(self.frank), 'Frank'))\n", "self.assertEquals((yield defer.ensureDeferred(self.store.\n get_profile_displayname(self.frank.localpart))), 'Frank')\n", "@defer.inlineCallbacks...\n", "self.hs.config.enable_set_displayname = False\n", "yield defer.ensureDeferred(self.store.set_profile_displayname(self.frank.\n localpart, 'Frank'))\n", "self.assertEquals((yield defer.ensureDeferred(self.store.\n get_profile_displayname(self.frank.localpart))), 'Frank')\n", "VAR_4 = defer.ensureDeferred(self.handler.set_displayname(self.frank,\n synapse.types.create_requester(self.frank), 'Frank Jr.'))\n", "yield self.assertFailure(VAR_4, SynapseError)\n", "@defer.inlineCallbacks...\n", "VAR_4 = defer.ensureDeferred(self.handler.set_displayname(self.frank,\n synapse.types.create_requester(self.bob), 'Frank Jr.'))\n", "yield self.assertFailure(VAR_4, AuthError)\n", "@defer.inlineCallbacks...\n", "self.mock_federation.make_query.return_value = make_awaitable({\n 'displayname': 'Alice'})\n", "VAR_3 = yield defer.ensureDeferred(self.handler.get_displayname(self.alice))\n", "self.assertEquals(VAR_3, 'Alice')\n", "self.mock_federation.make_query.assert_called_with(destination='remote',\n VAR_0='profile', args={'user_id': '@alice:remote', 'field':\n 'displayname'}, ignore_backoff=True)\n", "@defer.inlineCallbacks...\n", "yield defer.ensureDeferred(self.store.create_profile('caroline'))\n", "yield defer.ensureDeferred(self.store.set_profile_displayname('caroline',\n 'Caroline'))\n", "VAR_5 = yield defer.ensureDeferred(self.query_handlers['profile']({\n 'user_id': '@caroline:test', 'field': 'displayname'}))\n", "self.assertEquals({'displayname': 'Caroline'}, VAR_5)\n", "@defer.inlineCallbacks...\n", "yield defer.ensureDeferred(self.store.set_profile_avatar_url(self.frank.\n localpart, 'http://my.server/me.png'))\n", "VAR_6 = yield defer.ensureDeferred(self.handler.get_avatar_url(self.frank))\n", "self.assertEquals('http://my.server/me.png', VAR_6)\n", "@defer.inlineCallbacks...\n", "yield defer.ensureDeferred(self.handler.set_avatar_url(self.frank, synapse.\n types.create_requester(self.frank), 'http://my.server/pic.gif'))\n", "self.assertEquals((yield defer.ensureDeferred(self.store.\n get_profile_avatar_url(self.frank.localpart))), 'http://my.server/pic.gif')\n", "yield defer.ensureDeferred(self.handler.set_avatar_url(self.frank, synapse.\n types.create_requester(self.frank), 'http://my.server/me.png'))\n", "self.assertEquals((yield defer.ensureDeferred(self.store.\n get_profile_avatar_url(self.frank.localpart))), 'http://my.server/me.png')\n", "@defer.inlineCallbacks...\n", "self.hs.config.enable_set_avatar_url = False\n", "yield defer.ensureDeferred(self.store.set_profile_avatar_url(self.frank.\n localpart, 'http://my.server/me.png'))\n", "self.assertEquals((yield defer.ensureDeferred(self.store.\n get_profile_avatar_url(self.frank.localpart))), 'http://my.server/me.png')\n", "VAR_4 = defer.ensureDeferred(self.handler.set_avatar_url(self.frank,\n synapse.types.create_requester(self.frank), 'http://my.server/pic.gif'))\n", "yield self.assertFailure(VAR_4, SynapseError)\n" ]
[ "from mock import Mock\n", "from twisted.internet import defer\n", "import synapse.types\n", "from synapse.api.errors import AuthError, SynapseError\n", "from synapse.types import UserID\n", "from tests import unittest\n", "from tests.test_utils import make_awaitable\n", "from tests.utils import setup_test_homeserver\n", "\"\"\" Tests profile management. \"\"\"\n", "@defer.inlineCallbacks...\n", "self.mock_federation = Mock()\n", "self.mock_registry = Mock()\n", "self.query_handlers = {}\n", "def register_query_handler(query_type, handler):...\n", "self.query_handlers[query_type] = handler\n", "self.mock_registry.register_query_handler = register_query_handler\n", "hs = yield setup_test_homeserver(self.addCleanup, http_client=None,\n resource_for_federation=Mock(), federation_client=self.mock_federation,\n federation_server=Mock(), federation_registry=self.mock_registry)\n", "self.store = hs.get_datastore()\n", "self.frank = UserID.from_string('@1234ABCD:test')\n", "self.bob = UserID.from_string('@4567:test')\n", "self.alice = UserID.from_string('@alice:remote')\n", "yield defer.ensureDeferred(self.store.create_profile(self.frank.localpart))\n", "self.handler = hs.get_profile_handler()\n", "self.hs = hs\n", "@defer.inlineCallbacks...\n", "yield defer.ensureDeferred(self.store.set_profile_displayname(self.frank.\n localpart, 'Frank'))\n", "displayname = yield defer.ensureDeferred(self.handler.get_displayname(self.\n frank))\n", "self.assertEquals('Frank', displayname)\n", "@defer.inlineCallbacks...\n", "yield defer.ensureDeferred(self.handler.set_displayname(self.frank, synapse\n .types.create_requester(self.frank), 'Frank Jr.'))\n", "self.assertEquals((yield defer.ensureDeferred(self.store.\n get_profile_displayname(self.frank.localpart))), 'Frank Jr.')\n", "yield defer.ensureDeferred(self.handler.set_displayname(self.frank, synapse\n .types.create_requester(self.frank), 'Frank'))\n", "self.assertEquals((yield defer.ensureDeferred(self.store.\n get_profile_displayname(self.frank.localpart))), 'Frank')\n", "@defer.inlineCallbacks...\n", "self.hs.config.enable_set_displayname = False\n", "yield defer.ensureDeferred(self.store.set_profile_displayname(self.frank.\n localpart, 'Frank'))\n", "self.assertEquals((yield defer.ensureDeferred(self.store.\n get_profile_displayname(self.frank.localpart))), 'Frank')\n", "d = defer.ensureDeferred(self.handler.set_displayname(self.frank, synapse.\n types.create_requester(self.frank), 'Frank Jr.'))\n", "yield self.assertFailure(d, SynapseError)\n", "@defer.inlineCallbacks...\n", "d = defer.ensureDeferred(self.handler.set_displayname(self.frank, synapse.\n types.create_requester(self.bob), 'Frank Jr.'))\n", "yield self.assertFailure(d, AuthError)\n", "@defer.inlineCallbacks...\n", "self.mock_federation.make_query.return_value = make_awaitable({\n 'displayname': 'Alice'})\n", "displayname = yield defer.ensureDeferred(self.handler.get_displayname(self.\n alice))\n", "self.assertEquals(displayname, 'Alice')\n", "self.mock_federation.make_query.assert_called_with(destination='remote',\n query_type='profile', args={'user_id': '@alice:remote', 'field':\n 'displayname'}, ignore_backoff=True)\n", "@defer.inlineCallbacks...\n", "yield defer.ensureDeferred(self.store.create_profile('caroline'))\n", "yield defer.ensureDeferred(self.store.set_profile_displayname('caroline',\n 'Caroline'))\n", "response = yield defer.ensureDeferred(self.query_handlers['profile']({\n 'user_id': '@caroline:test', 'field': 'displayname'}))\n", "self.assertEquals({'displayname': 'Caroline'}, response)\n", "@defer.inlineCallbacks...\n", "yield defer.ensureDeferred(self.store.set_profile_avatar_url(self.frank.\n localpart, 'http://my.server/me.png'))\n", "avatar_url = yield defer.ensureDeferred(self.handler.get_avatar_url(self.frank)\n )\n", "self.assertEquals('http://my.server/me.png', avatar_url)\n", "@defer.inlineCallbacks...\n", "yield defer.ensureDeferred(self.handler.set_avatar_url(self.frank, synapse.\n types.create_requester(self.frank), 'http://my.server/pic.gif'))\n", "self.assertEquals((yield defer.ensureDeferred(self.store.\n get_profile_avatar_url(self.frank.localpart))), 'http://my.server/pic.gif')\n", "yield defer.ensureDeferred(self.handler.set_avatar_url(self.frank, synapse.\n types.create_requester(self.frank), 'http://my.server/me.png'))\n", "self.assertEquals((yield defer.ensureDeferred(self.store.\n get_profile_avatar_url(self.frank.localpart))), 'http://my.server/me.png')\n", "@defer.inlineCallbacks...\n", "self.hs.config.enable_set_avatar_url = False\n", "yield defer.ensureDeferred(self.store.set_profile_avatar_url(self.frank.\n localpart, 'http://my.server/me.png'))\n", "self.assertEquals((yield defer.ensureDeferred(self.store.\n get_profile_avatar_url(self.frank.localpart))), 'http://my.server/me.png')\n", "d = defer.ensureDeferred(self.handler.set_avatar_url(self.frank, synapse.\n types.create_requester(self.frank), 'http://my.server/pic.gif'))\n", "yield self.assertFailure(d, SynapseError)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 4, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "ImportFrom'", "ImportFrom'", "Import'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Expr'", "Condition", "Assign'", "Assign'", "Assign'", "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "Condition", "Expr'", "Assign'", "Expr'", "Condition", "Expr'", "Expr'", "Expr'", "Expr'", "Condition", "Assign'", "Expr'", "Expr'", "Assign'", "Expr'", "Condition", "Assign'", "Expr'", "Condition", "Assign'", "Assign'", "Expr'", "Expr'", "Condition", "Expr'", "Expr'", "Assign'", "Expr'", "Condition", "Expr'", "Assign'", "Expr'", "Condition", "Expr'", "Expr'", "Expr'", "Expr'", "Condition", "Assign'", "Expr'", "Expr'", "Assign'", "Expr'" ]
[ "@VAR_7.route('/base/<baseidx>/keywords/<params>')...\n", "print('Enter Scope')\n", "sys.stdout.flush()\n", "VAR_23 = []\n", "VAR_24, VAR_25 = FUNC_3(VAR_13)\n", "if VAR_9 != '0':\n", "VAR_38 = FUNC_11('GIDIDX' + VAR_9.upper())\n", "VAR_26 = bool(VAR_17 and VAR_23)\n", "VAR_23 = list(f.readlines())\n", "if VAR_23:\n", "if VAR_16 > 0:\n", "VAR_39 = len(VAR_23)\n", "VAR_39 = len(VAR_17)\n", "VAR_23 = VAR_23[VAR_15:VAR_15 + VAR_16]\n", "if VAR_15 > 0:\n", "random.seed(VAR_24)\n", "VAR_23 = VAR_17.copy()\n", "VAR_39 = len(VAR_23)\n", "VAR_23 = VAR_23[VAR_15:]\n", "VAR_27 = 0\n", "if VAR_26:\n", "random.Random(VAR_24).shuffle(VAR_17)\n", "def FUNC_13():...\n", "VAR_27 = int(VAR_25 * VAR_39)\n", "yield '<?xml version=\"1.0\" encoding=\"UTF-8\" ?>\\n'\n", "VAR_39 = VAR_39 + VAR_27\n", "if VAR_1:\n", "yield '<?xml-stylesheet type=\"text/xsl\" href=\"/scopelist.xsl\" ?>\\n'\n", "yield '<objectlist count=\"{:d}\">\\n'.format(VAR_39)\n", "VAR_40 = 0\n", "VAR_41 = 0\n", "if VAR_26:\n", "VAR_40 = int(VAR_25 * VAR_5)\n", "VAR_42 = []\n", "VAR_44 = cycle(VAR_17)\n", "def FUNC_15():...\n", "random.seed(VAR_24)\n", "return list(map(lambda x: x + VAR_5 * VAR_41, sorted(random.sample(list(\n range(VAR_5)), VAR_40))))\n" ]
[ "@scope_blueprint.route('/base/<baseidx>/keywords/<params>')...\n", "print('Enter Scope')\n", "sys.stdout.flush()\n", "base_list = []\n", "seed, percentage = decode_params(params)\n", "if baseidx != '0':\n", "base_index = _get_index_absolute_path('GIDIDX' + baseidx.upper())\n", "make_cocktail = bool(mixer_list and base_list)\n", "base_list = list(f.readlines())\n", "if base_list:\n", "if limit > 0:\n", "total_entries = len(base_list)\n", "total_entries = len(mixer_list)\n", "base_list = base_list[start:start + limit]\n", "if start > 0:\n", "random.seed(seed)\n", "base_list = mixer_list.copy()\n", "total_entries = len(base_list)\n", "base_list = base_list[start:]\n", "total_sample = 0\n", "if make_cocktail:\n", "random.Random(seed).shuffle(mixer_list)\n", "def generate():...\n", "total_sample = int(percentage * total_entries)\n", "yield '<?xml version=\"1.0\" encoding=\"UTF-8\" ?>\\n'\n", "total_entries = total_entries + total_sample\n", "if STYLE:\n", "yield '<?xml-stylesheet type=\"text/xsl\" href=\"/scopelist.xsl\" ?>\\n'\n", "yield '<objectlist count=\"{:d}\">\\n'.format(total_entries)\n", "mix_per_iteration = 0\n", "iteration_count = 0\n", "if make_cocktail:\n", "mix_per_iteration = int(percentage * ITEMS_PER_ITERATION)\n", "mix_indices = []\n", "pool = cycle(mixer_list)\n", "def generate_mix_indices():...\n", "random.seed(seed)\n", "return list(map(lambda x: x + ITEMS_PER_ITERATION * iteration_count, sorted\n (random.sample(list(range(ITEMS_PER_ITERATION)), mix_per_iteration))))\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Expr'", "Expr'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Condition", "Condition", "Assign'", "Assign'", "Assign'", "Condition", "Expr'", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Expr'", "FunctionDef'", "Assign'", "Expr'", "Assign'", "Condition", "Expr'", "Expr'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "FunctionDef'", "Expr'", "Return'" ]
[ "__author__ = 'Gina Häußge <[email protected]>'\n", "__license__ = (\n 'GNU Affero General Public License http://www.gnu.org/licenses/agpl.html')\n", "__copyright__ = (\n 'Copyright (C) 2015 The OctoPrint Project - Released under terms of the AGPLv3 License'\n )\n", "import base64\n", "import datetime\n", "import logging\n", "import os\n", "import re\n", "from collections import defaultdict\n", "from flask import Response, abort, g, make_response, redirect, render_template, request, send_from_directory, url_for\n", "import octoprint.plugin\n", "from octoprint.access.permissions import OctoPrintPermission, Permissions\n", "from octoprint.filemanager import full_extension_tree, get_all_extensions\n", "from octoprint.server import BRANCH, DISPLAY_VERSION, LOCALES, NOT_MODIFIED, VERSION, app, debug, gettext, groupManager, pluginManager, preemptiveCache, userManager\n", "from octoprint.server.util import has_permissions, require_login_with\n", "from octoprint.settings import settings\n", "from octoprint.util import sv, to_bytes, to_unicode\n", "from octoprint.util.version import get_python_version_string\n", "from . import util\n", "VAR_0 = logging.getLogger(__name__)\n", "VAR_1 = {}\n", "VAR_2 = None\n", "VAR_3 = None\n", "VAR_4 = re.compile('[a-z_]+')\n", "VAR_5 = re.compile('[a-zA-Z_-]+')\n", "def FUNC_0(VAR_6=None, VAR_7=None):...\n", "if VAR_6 is None:\n", "VAR_6 = request.url_root\n", "VAR_35 = not settings().getBoolean(['devel', 'cache', 'preemptive']\n ) or VAR_6 in settings().get(['server', 'preemptiveCache', 'exceptions']\n ) or not (VAR_6.startswith('http://') or VAR_6.startswith('https://'))\n", "VAR_36 = request.headers.get('X-Preemptive-Recording', 'no') == 'yes'\n", "if callable(VAR_7):\n", "return VAR_36 or VAR_35 or VAR_7()\n", "return VAR_36 or VAR_35\n" ]
[ "__author__ = 'Gina Häußge <[email protected]>'\n", "__license__ = (\n 'GNU Affero General Public License http://www.gnu.org/licenses/agpl.html')\n", "__copyright__ = (\n 'Copyright (C) 2015 The OctoPrint Project - Released under terms of the AGPLv3 License'\n )\n", "import base64\n", "import datetime\n", "import logging\n", "import os\n", "import re\n", "from collections import defaultdict\n", "from flask import Response, abort, g, make_response, redirect, render_template, request, send_from_directory, url_for\n", "import octoprint.plugin\n", "from octoprint.access.permissions import OctoPrintPermission, Permissions\n", "from octoprint.filemanager import full_extension_tree, get_all_extensions\n", "from octoprint.server import BRANCH, DISPLAY_VERSION, LOCALES, NOT_MODIFIED, VERSION, app, debug, gettext, groupManager, pluginManager, preemptiveCache, userManager\n", "from octoprint.server.util import has_permissions, require_login_with\n", "from octoprint.settings import settings\n", "from octoprint.util import sv, to_bytes, to_unicode\n", "from octoprint.util.version import get_python_version_string\n", "from . import util\n", "_logger = logging.getLogger(__name__)\n", "_templates = {}\n", "_plugin_names = None\n", "_plugin_vars = None\n", "_valid_id_re = re.compile('[a-z_]+')\n", "_valid_div_re = re.compile('[a-zA-Z_-]+')\n", "def _preemptive_unless(base_url=None, additional_unless=None):...\n", "if base_url is None:\n", "base_url = request.url_root\n", "disabled_for_root = not settings().getBoolean(['devel', 'cache', 'preemptive']\n ) or base_url in settings().get(['server', 'preemptiveCache', 'exceptions']\n ) or not (base_url.startswith('http://') or base_url.startswith('https://')\n )\n", "recording_disabled = request.headers.get('X-Preemptive-Recording', 'no'\n ) == 'yes'\n", "if callable(additional_unless):\n", "return recording_disabled or disabled_for_root or additional_unless()\n", "return recording_disabled or disabled_for_root\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Assign'", "Assign'", "Assign'", "Import'", "Import'", "Import'", "Import'", "Import'", "ImportFrom'", "ImportFrom'", "Import'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "FunctionDef'", "Condition", "Assign'", "Assign'", "Assign'", "Condition", "Return'", "Return'" ]
[ "@log_function...\n", "\"\"\"docstring\"\"\"\n", "VAR_2 = FUNC_1('/groups/%s/users', VAR_30)\n", "return self.client.get_json(VAR_5=destination, VAR_2=path, VAR_3={\n 'requester_user_id': requester_user_id}, VAR_15=True)\n" ]
[ "@log_function...\n", "\"\"\"docstring\"\"\"\n", "path = _create_v1_path('/groups/%s/users', group_id)\n", "return self.client.get_json(destination=destination, path=path, args={\n 'requester_user_id': requester_user_id}, ignore_backoff=True)\n" ]
[ 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Return'" ]
[ "def FUNC_9(self, VAR_25):...\n", "if callable(getattr(VAR_25, 'as_view', None)):\n", "VAR_25 = VAR_25.as_view()\n", "@six.wraps(VAR_25)...\n", "VAR_38 = self._get_unauth_reason(VAR_20)\n", "if VAR_38:\n", "return self._get_unauth_response(VAR_20, VAR_38)\n", "return VAR_25(VAR_20, *VAR_36, **kwargs)\n" ]
[ "def wrap_with_permissions(self, view_func):...\n", "if callable(getattr(view_func, 'as_view', None)):\n", "view_func = view_func.as_view()\n", "@six.wraps(view_func)...\n", "unauth_reason = self._get_unauth_reason(request)\n", "if unauth_reason:\n", "return self._get_unauth_response(request, unauth_reason)\n", "return view_func(request, *args, **kwargs)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Assign'", "Condition", "Assign'", "Condition", "Return'", "Return'" ]
[ "def FUNC_40():...\n", "VAR_90 = util.flask.check_lastmodified(VAR_103)\n", "VAR_89 = util.flask.check_etag(VAR_104)\n", "return VAR_90 and VAR_89\n" ]
[ "def check_etag_and_lastmodified():...\n", "lastmodified_ok = util.flask.check_lastmodified(current_lastmodified)\n", "etag_ok = util.flask.check_etag(current_etag)\n", "return lastmodified_ok and etag_ok\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_6():...\n", "VAR_26 = mysql.connector.connect(user=DB_USER, password=DB_PASSWORD, host=\n DB_HOST, database=DB_DBNAME, port=DB_PORT)\n", "VAR_27 = VAR_26.cursor()\n", "VAR_27.execute(VAR_20, VAR_22)\n", "yield '<?xml version=\"1.0\" encoding=\"UTF-8\" ?>\\n'\n", "if VAR_1:\n", "yield '<?xml-stylesheet type=\"text/xsl\" href=\"/scopelist.xsl\" ?>\\n'\n", "yield '<objectlist>\\n'\n", "for VAR_17, VAR_18, VAR_19 in VAR_27:\n", "yield '<count adjust=\"1\"/>\\n'\n", "yield '</objectlist>\\n'\n", "yield FUNC_5(VAR_13, VAR_17, VAR_18, VAR_19) + '\\n'\n" ]
[ "def generate():...\n", "cnx = mysql.connector.connect(user=DB_USER, password=DB_PASSWORD, host=\n DB_HOST, database=DB_DBNAME, port=DB_PORT)\n", "cursor = cnx.cursor()\n", "cursor.execute(query, substitutes)\n", "yield '<?xml version=\"1.0\" encoding=\"UTF-8\" ?>\\n'\n", "if STYLE:\n", "yield '<?xml-stylesheet type=\"text/xsl\" href=\"/scopelist.xsl\" ?>\\n'\n", "yield '<objectlist>\\n'\n", "for seq_no, rel_path, download_link in cursor:\n", "yield '<count adjust=\"1\"/>\\n'\n", "yield '</objectlist>\\n'\n", "yield _get_object_element(dataset, seq_no, rel_path, download_link) + '\\n'\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Expr'", "Expr'", "Condition", "Expr'", "Expr'", "For", "Expr'", "Expr'", "Expr'" ]
[ "def FUNC_8():...\n", "\"\"\"docstring\"\"\"\n", "VAR_34 = VAR_1(VAR_12, 'components', 'bootstrap', 'less', 'print.less')\n", "VAR_49 = f.readlines()\n", "for ix, VAR_68 in enumerate(VAR_49):\n", "if 'Black prints faster' in VAR_68:\n", "VAR_35 = VAR_49.pop(ix)\n", "print('Removed line', ix, 'from bootstrap print.less:')\n", "print('-', VAR_35)\n", "print()\n", "f.writelines(VAR_49)\n" ]
[ "def patch_out_bootstrap_bw_print():...\n", "\"\"\"docstring\"\"\"\n", "print_less = pjoin(static, 'components', 'bootstrap', 'less', 'print.less')\n", "lines = f.readlines()\n", "for ix, line in enumerate(lines):\n", "if 'Black prints faster' in line:\n", "rmed = lines.pop(ix)\n", "print('Removed line', ix, 'from bootstrap print.less:')\n", "print('-', rmed)\n", "print()\n", "f.writelines(lines)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "For", "Condition", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'" ]
[ "def FUNC_3(self):...\n", "\"\"\"docstring\"\"\"\n", "self.load_from_db()\n" ]
[ "def reload(self):...\n", "\"\"\"docstring\"\"\"\n", "self.load_from_db()\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Expr'" ]
[ "def FUNC_5(self):...\n", "if not getattr(self, 'latest', None):\n", "self.latest = frappe.get_doc(self.doctype, self.name)\n", "return self.latest\n" ]
[ "def get_latest(self):...\n", "if not getattr(self, 'latest', None):\n", "self.latest = frappe.get_doc(self.doctype, self.name)\n", "return self.latest\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Assign'", "Return'" ]
[ "def FUNC_32(VAR_62, VAR_10=None, VAR_92=False):...\n", "if VAR_92:\n", "return False\n", "if not VAR_10:\n", "VAR_10 = VAR_1.session.user\n", "import frappe.permissions\n", "VAR_183 = frappe.permissions.get_role_permissions(VAR_62, VAR_10=user)\n", "if VAR_183.get('select') and not VAR_183.get('read'):\n", "return True\n", "return False\n" ]
[ "def only_has_select_perm(doctype, user=None, ignore_permissions=False):...\n", "if ignore_permissions:\n", "return False\n", "if not user:\n", "user = local.session.user\n", "import frappe.permissions\n", "permissions = frappe.permissions.get_role_permissions(doctype, user=user)\n", "if permissions.get('select') and not permissions.get('read'):\n", "return True\n", "return False\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Return'", "Condition", "Assign'", "Import'", "Assign'", "Condition", "Return'", "Return'" ]
[ "def __init__(self, *VAR_2):...\n", "super().__init__(*VAR_2)\n", "self.user_consent_version = None\n", "self.user_consent_template_dir = None\n", "self.user_consent_server_notice_content = None\n", "self.user_consent_server_notice_to_guests = False\n", "self.block_events_without_consent_error = None\n", "self.user_consent_at_registration = False\n", "self.user_consent_policy_name = 'Privacy Policy'\n" ]
[ "def __init__(self, *args):...\n", "super().__init__(*args)\n", "self.user_consent_version = None\n", "self.user_consent_template_dir = None\n", "self.user_consent_server_notice_content = None\n", "self.user_consent_server_notice_to_guests = False\n", "self.block_events_without_consent_error = None\n", "self.user_consent_at_registration = False\n", "self.user_consent_policy_name = 'Privacy Policy'\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'" ]
[ "def FUNC_12(self, VAR_12):...\n", "if not self.isAuthorized():\n", "VAR_11 = [VAR_69 for VAR_69 in json.loads(unquote(VAR_12))]\n", "VAR_61 = self.download_check_files(VAR_11)\n", "if VAR_61 == 'ok':\n", "FUNC_0()\n", "return VAR_61\n", "VAR_82 = 'application/x-zip-compressed'\n", "VAR_53.response.headers['Content-Type'] = VAR_82\n", "VAR_83 = 'attachment; filename=\"music.zip\"'\n", "VAR_53.response.headers['Content-Disposition'] = VAR_83\n", "VAR_84 = cherry.config['media.basedir']\n", "VAR_98 = [os.path.join(VAR_84, f) for f in VAR_11]\n", "return zipstream.ZipStream(VAR_98)\n" ]
[ "def download(self, value):...\n", "if not self.isAuthorized():\n", "filelist = [filepath for filepath in json.loads(unquote(value))]\n", "dlstatus = self.download_check_files(filelist)\n", "if dlstatus == 'ok':\n", "_save_and_release_session()\n", "return dlstatus\n", "zipmime = 'application/x-zip-compressed'\n", "cherrypy.response.headers['Content-Type'] = zipmime\n", "zipname = 'attachment; filename=\"music.zip\"'\n", "cherrypy.response.headers['Content-Disposition'] = zipname\n", "basedir = cherry.config['media.basedir']\n", "fullpath_filelist = [os.path.join(basedir, f) for f in filelist]\n", "return zipstream.ZipStream(fullpath_filelist)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Assign'", "Assign'", "Condition", "Expr'", "Return'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Return'" ]