lines
sequencelengths
1
383
raw_lines
sequencelengths
1
383
label
sequencelengths
1
383
type
sequencelengths
1
383
[ "def FUNC_2(self):...\n", "\"\"\"docstring\"\"\"\n", "return format_html(\"\"\"<ul{0}>\n{1}\n</ul>\"\"\", flatatt(self.attrs),\n format_html_join('\\n', '<li>{0}</li>', ((force_text(w),) for w in self)))\n" ]
[ "def render(self):...\n", "\"\"\"docstring\"\"\"\n", "return format_html(\"\"\"<ul{0}>\n{1}\n</ul>\"\"\", flatatt(self.attrs),\n format_html_join('\\n', '<li>{0}</li>', ((force_text(w),) for w in self)))\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Return'" ]
[ "def FUNC_25(self):...\n", "return self._has_role(constants.ROLE_UPLOAD)\n" ]
[ "def role_upload(self):...\n", "return self._has_role(constants.ROLE_UPLOAD)\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_1(self):...\n", "self.assertFormfield(models.Event, 'start_date', widgets.AdminDateWidget)\n" ]
[ "def testDateField(self):...\n", "self.assertFormfield(models.Event, 'start_date', widgets.AdminDateWidget)\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Expr'" ]
[ "import logging\n", "import urllib.parse\n", "from io import BytesIO\n", "from typing import TYPE_CHECKING, Any, BinaryIO, Dict, Iterable, List, Mapping, Optional, Sequence, Tuple, Union\n", "import treq\n", "from canonicaljson import encode_canonical_json\n", "from netaddr import IPAddress, IPSet\n", "from prometheus_client import Counter\n", "from zope.interface import implementer, provider\n", "from OpenSSL import SSL\n", "from OpenSSL.SSL import VERIFY_NONE\n", "from twisted.internet import defer, error as twisted_error, protocol, ssl\n", "from twisted.internet.interfaces import IAddress, IHostResolution, IReactorPluggableNameResolver, IResolutionReceiver\n", "from twisted.internet.task import Cooperator\n", "from twisted.python.failure import Failure\n", "from twisted.web._newclient import ResponseDone\n", "from twisted.web.client import Agent, HTTPConnectionPool, ResponseNeverReceived, readBody\n", "from twisted.web.http import PotentialDataLoss\n", "from twisted.web.http_headers import Headers\n", "from twisted.web.iweb import IAgent, IBodyProducer, IResponse\n", "from synapse.api.errors import Codes, HttpResponseException, SynapseError\n", "from synapse.http import QuieterFileBodyProducer, RequestTimedOutError, redact_uri\n", "from synapse.http.proxyagent import ProxyAgent\n", "from synapse.logging.context import make_deferred_yieldable\n", "from synapse.logging.opentracing import set_tag, start_active_span, tags\n", "from synapse.util import json_decoder\n", "from synapse.util.async_helpers import timeout_deferred\n", "if TYPE_CHECKING:\n", "from synapse.app.homeserver import HomeServer\n", "VAR_0 = logging.getLogger(__name__)\n", "VAR_1 = Counter('synapse_http_client_requests', '', ['method'])\n", "VAR_2 = Counter('synapse_http_client_responses', '', ['method', 'code'])\n", "VAR_3 = Union[Mapping[str, 'RawHeaderValue'], Mapping[bytes, 'RawHeaderValue']]\n", "VAR_4 = Sequence[Union[str, bytes]]\n", "VAR_5 = Union[str, bytes, Iterable[Union[str, bytes]]]\n", "VAR_6 = Union[Mapping[str, VAR_5], Mapping[bytes, VAR_5]]\n", "def FUNC_0(VAR_7: IPAddress, VAR_8: Optional[IPSet], VAR_9: IPSet) ->bool:...\n", "\"\"\"docstring\"\"\"\n", "if VAR_7 in VAR_9:\n", "if VAR_8 is None or VAR_7 not in VAR_8:\n", "return False\n", "return True\n" ]
[ "import logging\n", "import urllib.parse\n", "from io import BytesIO\n", "from typing import TYPE_CHECKING, Any, BinaryIO, Dict, Iterable, List, Mapping, Optional, Sequence, Tuple, Union\n", "import treq\n", "from canonicaljson import encode_canonical_json\n", "from netaddr import IPAddress, IPSet\n", "from prometheus_client import Counter\n", "from zope.interface import implementer, provider\n", "from OpenSSL import SSL\n", "from OpenSSL.SSL import VERIFY_NONE\n", "from twisted.internet import defer, error as twisted_error, protocol, ssl\n", "from twisted.internet.interfaces import IAddress, IHostResolution, IReactorPluggableNameResolver, IResolutionReceiver\n", "from twisted.internet.task import Cooperator\n", "from twisted.python.failure import Failure\n", "from twisted.web._newclient import ResponseDone\n", "from twisted.web.client import Agent, HTTPConnectionPool, ResponseNeverReceived, readBody\n", "from twisted.web.http import PotentialDataLoss\n", "from twisted.web.http_headers import Headers\n", "from twisted.web.iweb import IAgent, IBodyProducer, IResponse\n", "from synapse.api.errors import Codes, HttpResponseException, SynapseError\n", "from synapse.http import QuieterFileBodyProducer, RequestTimedOutError, redact_uri\n", "from synapse.http.proxyagent import ProxyAgent\n", "from synapse.logging.context import make_deferred_yieldable\n", "from synapse.logging.opentracing import set_tag, start_active_span, tags\n", "from synapse.util import json_decoder\n", "from synapse.util.async_helpers import timeout_deferred\n", "if TYPE_CHECKING:\n", "from synapse.app.homeserver import HomeServer\n", "logger = logging.getLogger(__name__)\n", "outgoing_requests_counter = Counter('synapse_http_client_requests', '', [\n 'method'])\n", "incoming_responses_counter = Counter('synapse_http_client_responses', '', [\n 'method', 'code'])\n", "RawHeaders = Union[Mapping[str, 'RawHeaderValue'], Mapping[bytes,\n 'RawHeaderValue']]\n", "RawHeaderValue = Sequence[Union[str, bytes]]\n", "QueryParamValue = Union[str, bytes, Iterable[Union[str, bytes]]]\n", "QueryParams = Union[Mapping[str, QueryParamValue], Mapping[bytes,\n QueryParamValue]]\n", "def check_against_blacklist(ip_address: IPAddress, ip_whitelist: Optional[...\n", "\"\"\"docstring\"\"\"\n", "if ip_address in ip_blacklist:\n", "if ip_whitelist is None or ip_address not in ip_whitelist:\n", "return False\n", "return True\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Import'", "Import'", "ImportFrom'", "ImportFrom'", "Import'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Condition", "ImportFrom'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "FunctionDef'", "Docstring", "Condition", "Condition", "Return'", "Return'" ]
[ "def FUNC_38(self):...\n", "from zope.interface.verify import verifyClass\n", "from zope.tal.interfaces import ITALExpressionEngine\n", "verifyClass(ITALExpressionEngine, self._getTargetClass())\n" ]
[ "def test_class_conforms_to_ITALExpressionEngine(self):...\n", "from zope.interface.verify import verifyClass\n", "from zope.tal.interfaces import ITALExpressionEngine\n", "verifyClass(ITALExpressionEngine, self._getTargetClass())\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "ImportFrom'", "ImportFrom'", "Expr'" ]
[ "@login_required()...\n", "" ]
[ "@login_required()...\n", "" ]
[ 0, 0 ]
[ "Condition", "Condition" ]
[ "\"\"\"\nForm Widget classes specific to the Django admin site.\n\"\"\"\n", "from __future__ import unicode_literals\n", "import copy\n", "from django import forms\n", "from django.contrib.admin.templatetags.admin_static import static\n", "from django.core.urlresolvers import reverse\n", "from django.forms.widgets import RadioFieldRenderer\n", "from django.forms.util import flatatt\n", "from django.utils.html import escape, format_html, format_html_join, smart_urlquote\n", "from django.utils.text import Truncator\n", "from django.utils.translation import ugettext as _\n", "from django.utils.safestring import mark_safe\n", "from django.utils.encoding import force_text\n", "from django.utils import six\n", "\"\"\"string\"\"\"\n", "@property...\n", "VAR_26 = ['core.js', 'SelectBox.js', 'SelectFilter2.js']\n", "return forms.Media(VAR_26=[static('admin/js/%s' % path) for path in js])\n" ]
[ "\"\"\"\nForm Widget classes specific to the Django admin site.\n\"\"\"\n", "from __future__ import unicode_literals\n", "import copy\n", "from django import forms\n", "from django.contrib.admin.templatetags.admin_static import static\n", "from django.core.urlresolvers import reverse\n", "from django.forms.widgets import RadioFieldRenderer\n", "from django.forms.util import flatatt\n", "from django.utils.html import escape, format_html, format_html_join, smart_urlquote\n", "from django.utils.text import Truncator\n", "from django.utils.translation import ugettext as _\n", "from django.utils.safestring import mark_safe\n", "from django.utils.encoding import force_text\n", "from django.utils import six\n", "\"\"\"\n A SelectMultiple with a JavaScript filter interface.\n\n Note that the resulting JavaScript assumes that the jsi18n\n catalog has been loaded in the page\n \"\"\"\n", "@property...\n", "js = ['core.js', 'SelectBox.js', 'SelectFilter2.js']\n", "return forms.Media(js=[static('admin/js/%s' % path) for path in js])\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Expr'", "ImportFrom'", "Import'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Expr'", "Condition", "Assign'", "Return'" ]
[ "async def FUNC_6(*VAR_9, **VAR_10):...\n", "return None\n" ]
[ "async def _insert_client_ip(*args, **kwargs):...\n", "return None\n" ]
[ 0, 0 ]
[ "AsyncFunctionDef'", "Return'" ]
[ "def FUNC_88(VAR_212):...\n", "VAR_408 = VAR_5.getAnnotationLinks('Image', parent_ids=imageIds)\n", "VAR_409 = {}\n", "VAR_410 = {}\n", "for VAR_432 in VAR_212:\n", "VAR_409[VAR_432] = []\n", "for VAR_344 in VAR_408:\n", "VAR_383 = VAR_344.getChild()\n", "VAR_411 = []\n", "if VAR_383._obj.__class__ == omero.model.TagAnnotationI:\n", "for VAR_432 in VAR_212:\n", "VAR_410[VAR_383.id] = VAR_383\n", "VAR_411.append({'id': VAR_432, 'tags': VAR_409[VAR_432]})\n", "VAR_164 = []\n", "VAR_409[VAR_344.getParent().id].append(VAR_383)\n", "for tId, VAR_433 in VAR_410.items():\n", "VAR_164.append(VAR_433)\n", "return VAR_411, VAR_164\n" ]
[ "def loadImageTags(imageIds):...\n", "tagLinks = conn.getAnnotationLinks('Image', parent_ids=imageIds)\n", "linkMap = {}\n", "tagMap = {}\n", "for iId in imageIds:\n", "linkMap[iId] = []\n", "for link in tagLinks:\n", "c = link.getChild()\n", "imageTags = []\n", "if c._obj.__class__ == omero.model.TagAnnotationI:\n", "for iId in imageIds:\n", "tagMap[c.id] = c\n", "imageTags.append({'id': iId, 'tags': linkMap[iId]})\n", "tags = []\n", "linkMap[link.getParent().id].append(c)\n", "for tId, t in tagMap.items():\n", "tags.append(t)\n", "return imageTags, tags\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "For", "Assign'", "For", "Assign'", "Assign'", "Condition", "For", "Assign'", "Expr'", "Assign'", "Expr'", "For", "Expr'", "Return'" ]
[ "def FUNC_14(VAR_3):...\n", "\"\"\"docstring\"\"\"\n", "if not VAR_16:\n", "if VAR_13.get('logging') or False:\n", "VAR_21.append(FUNC_1(VAR_3))\n", "print(repr(VAR_3))\n" ]
[ "def log(msg):...\n", "\"\"\"docstring\"\"\"\n", "if not request:\n", "if conf.get('logging') or False:\n", "debug_log.append(as_unicode(msg))\n", "print(repr(msg))\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Condition", "Expr'", "Expr'" ]
[ "def FUNC_5(self):...\n", "conditional.processConditionals(self)\n", "self.fillContainers = locateFillContainers(self)\n", "lint.exampleIDs(self)\n", "boilerplate.addBikeshedVersion(self)\n", "boilerplate.addCanonicalURL(self)\n", "boilerplate.addFavicon(self)\n", "boilerplate.addSpecVersion(self)\n", "boilerplate.addStatusSection(self)\n", "boilerplate.addLogo(self)\n", "boilerplate.addCopyright(self)\n", "boilerplate.addSpecMetadataSection(self)\n", "boilerplate.addAbstract(self)\n", "boilerplate.addExpiryNotice(self)\n", "boilerplate.addObsoletionNotice(self)\n", "boilerplate.addAtRisk(self)\n", "addNoteHeaders(self)\n", "boilerplate.removeUnwantedBoilerplate(self)\n", "wpt.processWptElements(self)\n", "shorthands.run(self)\n", "inlineTags.processTags(self)\n", "canonicalizeShortcuts(self)\n", "addImplicitAlgorithms(self)\n", "fixManualDefTables(self)\n", "headings.processHeadings(self)\n", "checkVarHygiene(self)\n", "processIssuesAndExamples(self)\n", "idl.markupIDL(self)\n", "inlineRemoteIssues(self)\n", "addImageSize(self)\n", "processBiblioLinks(self)\n", "processDfns(self)\n", "idl.processIDL(self)\n", "dfns.annotateDfns(self)\n", "formatArgumentdefTables(self)\n", "formatElementdefTables(self)\n", "processAutolinks(self)\n", "biblio.dedupBiblioReferences(self)\n", "verifyUsageOfAllLocalBiblios(self)\n", "caniuse.addCanIUsePanels(self)\n", "boilerplate.addIndexSection(self)\n", "boilerplate.addExplicitIndexes(self)\n", "boilerplate.addStyles(self)\n", "boilerplate.addReferencesSection(self)\n", "boilerplate.addPropertyIndex(self)\n", "boilerplate.addIDLSection(self)\n", "boilerplate.addIssuesSection(self)\n", "boilerplate.addCustomBoilerplate(self)\n", "headings.processHeadings(self, 'all')\n", "boilerplate.removeUnwantedBoilerplate(self)\n", "boilerplate.addTOCSection(self)\n", "addSelfLinks(self)\n", "processAutolinks(self)\n", "boilerplate.addAnnotations(self)\n", "boilerplate.removeUnwantedBoilerplate(self)\n", "mdnspeclinks.addMdnPanels(self)\n", "highlight.addSyntaxHighlighting(self)\n", "boilerplate.addBikeshedBoilerplate(self)\n", "fingerprinting.addTrackingVector(self)\n", "fixIntraDocumentReferences(self)\n", "fixInterDocumentReferences(self)\n", "removeMultipleLinks(self)\n", "forceCrossorigin(self)\n", "lint.brokenLinks(self)\n", "lint.accidental2119(self)\n", "lint.missingExposed(self)\n", "lint.requiredIDs(self)\n", "lint.unusedInternalDfns(self)\n", "cleanupHTML(self)\n", "if self.md.prepTR:\n", "for VAR_19 in findAll(\"[rel ~= 'icon']\", self):\n", "return self\n", "removeNode(VAR_19)\n", "for VAR_19 in findAll('link', self):\n", "if VAR_19.get('href').startswith('https://www.w3.org/StyleSheets/TR'):\n", "for VAR_19 in findAll('a', self):\n", "appendChild(find('head', self), VAR_19)\n", "VAR_35 = VAR_19.get('href', '')\n", "extensions.BSPrepTR(self)\n", "if VAR_35.startswith('http://www.w3.org') or VAR_35.startswith(\n", "VAR_19.set('href', 'https' + VAR_35[4:])\n", "VAR_17 = VAR_19.text or ''\n", "if VAR_17.startswith('http://www.w3.org') or VAR_17.startswith(\n", "VAR_19.text = 'https' + VAR_17[4:]\n" ]
[ "def processDocument(self):...\n", "conditional.processConditionals(self)\n", "self.fillContainers = locateFillContainers(self)\n", "lint.exampleIDs(self)\n", "boilerplate.addBikeshedVersion(self)\n", "boilerplate.addCanonicalURL(self)\n", "boilerplate.addFavicon(self)\n", "boilerplate.addSpecVersion(self)\n", "boilerplate.addStatusSection(self)\n", "boilerplate.addLogo(self)\n", "boilerplate.addCopyright(self)\n", "boilerplate.addSpecMetadataSection(self)\n", "boilerplate.addAbstract(self)\n", "boilerplate.addExpiryNotice(self)\n", "boilerplate.addObsoletionNotice(self)\n", "boilerplate.addAtRisk(self)\n", "addNoteHeaders(self)\n", "boilerplate.removeUnwantedBoilerplate(self)\n", "wpt.processWptElements(self)\n", "shorthands.run(self)\n", "inlineTags.processTags(self)\n", "canonicalizeShortcuts(self)\n", "addImplicitAlgorithms(self)\n", "fixManualDefTables(self)\n", "headings.processHeadings(self)\n", "checkVarHygiene(self)\n", "processIssuesAndExamples(self)\n", "idl.markupIDL(self)\n", "inlineRemoteIssues(self)\n", "addImageSize(self)\n", "processBiblioLinks(self)\n", "processDfns(self)\n", "idl.processIDL(self)\n", "dfns.annotateDfns(self)\n", "formatArgumentdefTables(self)\n", "formatElementdefTables(self)\n", "processAutolinks(self)\n", "biblio.dedupBiblioReferences(self)\n", "verifyUsageOfAllLocalBiblios(self)\n", "caniuse.addCanIUsePanels(self)\n", "boilerplate.addIndexSection(self)\n", "boilerplate.addExplicitIndexes(self)\n", "boilerplate.addStyles(self)\n", "boilerplate.addReferencesSection(self)\n", "boilerplate.addPropertyIndex(self)\n", "boilerplate.addIDLSection(self)\n", "boilerplate.addIssuesSection(self)\n", "boilerplate.addCustomBoilerplate(self)\n", "headings.processHeadings(self, 'all')\n", "boilerplate.removeUnwantedBoilerplate(self)\n", "boilerplate.addTOCSection(self)\n", "addSelfLinks(self)\n", "processAutolinks(self)\n", "boilerplate.addAnnotations(self)\n", "boilerplate.removeUnwantedBoilerplate(self)\n", "mdnspeclinks.addMdnPanels(self)\n", "highlight.addSyntaxHighlighting(self)\n", "boilerplate.addBikeshedBoilerplate(self)\n", "fingerprinting.addTrackingVector(self)\n", "fixIntraDocumentReferences(self)\n", "fixInterDocumentReferences(self)\n", "removeMultipleLinks(self)\n", "forceCrossorigin(self)\n", "lint.brokenLinks(self)\n", "lint.accidental2119(self)\n", "lint.missingExposed(self)\n", "lint.requiredIDs(self)\n", "lint.unusedInternalDfns(self)\n", "cleanupHTML(self)\n", "if self.md.prepTR:\n", "for el in findAll(\"[rel ~= 'icon']\", self):\n", "return self\n", "removeNode(el)\n", "for el in findAll('link', self):\n", "if el.get('href').startswith('https://www.w3.org/StyleSheets/TR'):\n", "for el in findAll('a', self):\n", "appendChild(find('head', self), el)\n", "href = el.get('href', '')\n", "extensions.BSPrepTR(self)\n", "if href.startswith('http://www.w3.org') or href.startswith(\n", "el.set('href', 'https' + href[4:])\n", "text = el.text or ''\n", "if text.startswith('http://www.w3.org') or text.startswith(\n", "el.text = 'https' + text[4:]\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Condition", "For", "Return'", "Expr'", "For", "Condition", "For", "Expr'", "Assign'", "Expr'", "Condition", "Expr'", "Assign'", "Condition", "Assign'" ]
[ "@wraps(VAR_4)...\n", "if VAR_87.role_viewer():\n", "return VAR_4(*VAR_49, **kwargs)\n", "abort(403)\n" ]
[ "@wraps(f)...\n", "if current_user.role_viewer():\n", "return f(*args, **kwargs)\n", "abort(403)\n" ]
[ 0, 0, 0, 0 ]
[ "Condition", "Condition", "Return'", "Expr'" ]
[ "def FUNC_0(self):...\n", "from openapi_python_client.parser.properties import StringProperty\n", "VAR_4 = VAR_35(VAR_5='test', VAR_26=True, default=None)\n", "assert VAR_4.get_type_string() == 'str'\n", "VAR_4.required = False\n", "assert VAR_4.get_type_string() == 'Optional[str]'\n" ]
[ "def test_get_type_string(self):...\n", "from openapi_python_client.parser.properties import StringProperty\n", "p = StringProperty(name='test', required=True, default=None)\n", "assert p.get_type_string() == 'str'\n", "p.required = False\n", "assert p.get_type_string() == 'Optional[str]'\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "ImportFrom'", "Assign'", "Assert'", "Assign'", "Assert'" ]
[ "def FUNC_26(self, VAR_15, **VAR_14):...\n", "\"\"\"docstring\"\"\"\n", "if VAR_15.status_code == 401:\n", "VAR_1.warning('401 Error, Credentials not correct for %s', VAR_15.request.url)\n" ]
[ "def warn_on_401(self, resp, **kwargs):...\n", "\"\"\"docstring\"\"\"\n", "if resp.status_code == 401:\n", "logger.warning('401 Error, Credentials not correct for %s', resp.request.url)\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Expr'" ]
[ "@override_config({'send_federation': True})...\n", "self.room_members = [VAR_0, VAR_1, VAR_2]\n", "from synapse.handlers.typing import RoomMember\n", "VAR_22 = RoomMember(VAR_3, VAR_0.to_string())\n", "self.handler._member_typing_until[VAR_22] = 1002000\n", "self.handler._room_typing[VAR_3] = {VAR_0.to_string()}\n", "self.assertEquals(self.event_source.get_current_key(), 0)\n", "self.get_success(self.handler.stopped_typing(target_user=U_APPLE, requester\n =create_requester(U_APPLE), VAR_16=ROOM_ID))\n", "self.on_new_event.assert_has_calls([call('typing_key', 1, rooms=[ROOM_ID])])\n", "VAR_19 = self.hs.get_http_client().put_json\n", "VAR_19.assert_called_once_with('farm', path=\n '/_matrix/federation/v1/send/1000000', data=_expect_edu_transaction(\n 'm.typing', content={'room_id': ROOM_ID, 'user_id': U_APPLE.to_string(),\n 'typing': False}), json_data_callback=ANY, long_retries=True,\n backoff_on_404=True, try_trailing_slash_on_400=True)\n", "self.assertEquals(self.event_source.get_current_key(), 1)\n", "VAR_18 = self.get_success(self.event_source.get_new_events(room_ids=[\n ROOM_ID], from_key=0))\n", "self.assertEquals(VAR_18[0], [{'type': 'm.typing', 'room_id': VAR_3,\n 'content': {'user_ids': []}}])\n" ]
[ "@override_config({'send_federation': True})...\n", "self.room_members = [U_APPLE, U_BANANA, U_ONION]\n", "from synapse.handlers.typing import RoomMember\n", "member = RoomMember(ROOM_ID, U_APPLE.to_string())\n", "self.handler._member_typing_until[member] = 1002000\n", "self.handler._room_typing[ROOM_ID] = {U_APPLE.to_string()}\n", "self.assertEquals(self.event_source.get_current_key(), 0)\n", "self.get_success(self.handler.stopped_typing(target_user=U_APPLE, requester\n =create_requester(U_APPLE), room_id=ROOM_ID))\n", "self.on_new_event.assert_has_calls([call('typing_key', 1, rooms=[ROOM_ID])])\n", "put_json = self.hs.get_http_client().put_json\n", "put_json.assert_called_once_with('farm', path=\n '/_matrix/federation/v1/send/1000000', data=_expect_edu_transaction(\n 'm.typing', content={'room_id': ROOM_ID, 'user_id': U_APPLE.to_string(),\n 'typing': False}), json_data_callback=ANY, long_retries=True,\n backoff_on_404=True, try_trailing_slash_on_400=True)\n", "self.assertEquals(self.event_source.get_current_key(), 1)\n", "events = self.get_success(self.event_source.get_new_events(room_ids=[\n ROOM_ID], from_key=0))\n", "self.assertEquals(events[0], [{'type': 'm.typing', 'room_id': ROOM_ID,\n 'content': {'user_ids': []}}])\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 4, 0, 0, 0, 0 ]
[ "Condition", "Assign'", "ImportFrom'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'", "Assign'", "Expr'", "Expr'", "Assign'", "Expr'" ]
[ "def FUNC_30(VAR_27):...\n", "print(VAR_26 + VAR_27)\n" ]
[ "def in_print(s):...\n", "print(indent_str + s)\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Expr'" ]
[ "def __init__(self, VAR_8, *VAR_6, **VAR_7):...\n", "super().__init__(*VAR_6, **kwargs)\n", "self.fields['lang'].choices += Language.objects.have_translation().as_choices()\n", "self.fields['project'].choices += [(VAR_58.slug, VAR_58.name) for VAR_58 in\n VAR_8.user.allowed_projects]\n" ]
[ "def __init__(self, request, *args, **kwargs):...\n", "super().__init__(*args, **kwargs)\n", "self.fields['lang'].choices += Language.objects.have_translation().as_choices()\n", "self.fields['project'].choices += [(project.slug, project.name) for project in\n request.user.allowed_projects]\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "AugAssign'", "AugAssign'" ]
[ "import re\n", "import stringcase\n", "def FUNC_0(VAR_0: str) ->str:...\n", "return re.sub('[^\\\\w _-]+', '', VAR_0)\n" ]
[ "import re\n", "import stringcase\n", "def _sanitize(value: str) ->str:...\n", "return re.sub('[^\\\\w _-]+', '', value)\n" ]
[ 0, 0, 0, 5 ]
[ "Import'", "Import'", "FunctionDef'", "Return'" ]
[ "@pytest.mark.posix...\n", "assert filescheme.parent_dir(VAR_9) == VAR_11\n" ]
[ "@pytest.mark.posix...\n", "assert filescheme.parent_dir(directory) == parent\n" ]
[ 0, 0 ]
[ "Condition", "Assert'" ]
[ "def FUNC_1(VAR_2):...\n", "@wraps(VAR_2)...\n", "if current_user.role_edit() or current_user.role_admin():\n", "return VAR_2(*VAR_48, **kwargs)\n", "abort(403)\n", "return FUNC_46\n" ]
[ "def edit_required(f):...\n", "@wraps(f)...\n", "if current_user.role_edit() or current_user.role_admin():\n", "return f(*args, **kwargs)\n", "abort(403)\n", "return inner\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Condition", "Return'", "Expr'", "Return'" ]
[ "def FUNC_22(self):...\n", "VAR_5 = self._makeContext()\n", "self.assertEqual(VAR_5.evaluate('path:'), None)\n" ]
[ "def test_empty_path_expression_explicit(self):...\n", "ec = self._makeContext()\n", "self.assertEqual(ec.evaluate('path:'), None)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Expr'" ]
[ "def FUNC_156():...\n", "\"\"\"docstring\"\"\"\n", "VAR_355 = {'prefix': VAR_113, 'user': VAR_116 if self.user_id else None}\n", "for i in VAR_250:\n", "if i['name'] == VAR_28('Log In'):\n", "self.bar = VAR_355\n", "VAR_346 = 'login'\n", "if i['name'] == VAR_28('Sign Up'):\n", "VAR_355[VAR_346] = i['href']\n", "VAR_346 = 'register'\n", "if i['name'] == VAR_28('Lost password?'):\n", "VAR_346 = 'request_reset_password'\n", "if i['name'] == VAR_28('Forgot username?'):\n", "VAR_346 = 'retrieve_username'\n", "if i['name'] == VAR_28('Log Out'):\n", "VAR_346 = 'logout'\n", "if i['name'] == VAR_28('Profile'):\n", "VAR_346 = 'profile'\n", "if i['name'] == VAR_28('Password'):\n", "VAR_346 = 'change_password'\n" ]
[ "def bare():...\n", "\"\"\"docstring\"\"\"\n", "bare = {'prefix': prefix, 'user': user_identifier if self.user_id else None}\n", "for i in items:\n", "if i['name'] == T('Log In'):\n", "self.bar = bare\n", "k = 'login'\n", "if i['name'] == T('Sign Up'):\n", "bare[k] = i['href']\n", "k = 'register'\n", "if i['name'] == T('Lost password?'):\n", "k = 'request_reset_password'\n", "if i['name'] == T('Forgot username?'):\n", "k = 'retrieve_username'\n", "if i['name'] == T('Log Out'):\n", "k = 'logout'\n", "if i['name'] == T('Profile'):\n", "k = 'profile'\n", "if i['name'] == T('Password'):\n", "k = 'change_password'\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "For", "Condition", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Condition", "Assign'" ]
[ "@log_function...\n", "VAR_142 = {'signed': VAR_44}\n", "VAR_45 = {'type': VAR_188.Member, 'content': {'membership': Membership.\n INVITE, 'third_party_invite': VAR_142}, 'room_id': VAR_11, 'sender':\n VAR_42, 'state_key': VAR_43}\n", "if await self.auth.check_host_in_room(VAR_11, self.hs.hostname):\n", "VAR_28 = await self.store.get_room_version_id(VAR_11)\n", "VAR_181 = {VAR_204.split(':', 1)[-1] for VAR_204 in (VAR_42, VAR_11)}\n", "VAR_93 = self.event_builder_factory.new(VAR_28, VAR_45)\n", "await self.federation_client.forward_third_party_invite(VAR_181, VAR_11, VAR_45\n )\n", "EventValidator().validate_builder(VAR_93)\n", "VAR_1, VAR_32 = await self.event_creation_handler.create_new_client_event(\n VAR_93=builder)\n", "VAR_1, VAR_32 = await self.add_display_name_to_third_party_invite(VAR_28,\n VAR_45, VAR_1, VAR_32)\n", "EventValidator().validate_new(VAR_1, self.config)\n", "VAR_1.internal_metadata.send_on_behalf_of = self.hs.hostname\n", "await self.auth.check_from_context(VAR_28, VAR_1, VAR_32)\n", "VAR_0.warning('Denying new third party invite %r because %s', VAR_1, VAR_170)\n", "await self._check_signature(VAR_1, VAR_32)\n", "VAR_143 = self.hs.get_room_member_handler()\n", "await VAR_143.send_membership_event(None, VAR_1, VAR_32)\n" ]
[ "@log_function...\n", "third_party_invite = {'signed': signed}\n", "event_dict = {'type': EventTypes.Member, 'content': {'membership':\n Membership.INVITE, 'third_party_invite': third_party_invite}, 'room_id':\n room_id, 'sender': sender_user_id, 'state_key': target_user_id}\n", "if await self.auth.check_host_in_room(room_id, self.hs.hostname):\n", "room_version = await self.store.get_room_version_id(room_id)\n", "destinations = {x.split(':', 1)[-1] for x in (sender_user_id, room_id)}\n", "builder = self.event_builder_factory.new(room_version, event_dict)\n", "await self.federation_client.forward_third_party_invite(destinations,\n room_id, event_dict)\n", "EventValidator().validate_builder(builder)\n", "event, context = await self.event_creation_handler.create_new_client_event(\n builder=builder)\n", "event, context = await self.add_display_name_to_third_party_invite(room_version\n , event_dict, event, context)\n", "EventValidator().validate_new(event, self.config)\n", "event.internal_metadata.send_on_behalf_of = self.hs.hostname\n", "await self.auth.check_from_context(room_version, event, context)\n", "logger.warning('Denying new third party invite %r because %s', event, e)\n", "await self._check_signature(event, context)\n", "member_handler = self.hs.get_room_member_handler()\n", "await member_handler.send_membership_event(None, event, context)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'", "Assign'", "Assign'", "Expr'", "Assign'", "Expr'", "Expr'", "Expr'", "Assign'", "Expr'" ]
[ "@log_function...\n", "\"\"\"docstring\"\"\"\n", "VAR_2 = FUNC_1('/user/keys/query')\n", "VAR_19 = await self.client.post_json(VAR_5=destination, VAR_2=path, VAR_39=\n query_content, VAR_8=timeout)\n", "return VAR_19\n" ]
[ "@log_function...\n", "\"\"\"docstring\"\"\"\n", "path = _create_v1_path('/user/keys/query')\n", "content = await self.client.post_json(destination=destination, path=path,\n data=query_content, timeout=timeout)\n", "return content\n" ]
[ 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_8(self, VAR_4):...\n", "self.get_success(self.store.create_room_alias_association(self.room_alias,\n self.room_id, ['test'], VAR_4))\n" ]
[ "def _create_alias(self, user):...\n", "self.get_success(self.store.create_room_alias_association(self.room_alias,\n self.room_id, ['test'], user))\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Expr'" ]
[ "@ensure_csrf_cookie...\n", "\"\"\"docstring\"\"\"\n", "VAR_10 = SlashSeparatedCourseKey.from_deprecated_string(VAR_10)\n", "VAR_65 = get_course_by_id(VAR_10)\n", "VAR_136 = has_access(VAR_9.user, 'instructor', VAR_65)\n", "VAR_137 = has_forum_access(VAR_9.user, VAR_10, FORUM_ROLE_ADMINISTRATOR)\n", "VAR_141 = VAR_9.GET.get('unique_student_identifier')\n", "VAR_64 = VAR_9.GET.get('rolename')\n", "VAR_57 = VAR_9.GET.get('action')\n", "if not (VAR_137 or VAR_136):\n", "return HttpResponseBadRequest(\n 'Operation requires staff & forum admin or instructor access')\n", "if VAR_64 == FORUM_ROLE_ADMINISTRATOR and not VAR_136:\n", "return HttpResponseBadRequest('Operation requires instructor access.')\n", "if VAR_64 not in [FORUM_ROLE_ADMINISTRATOR, FORUM_ROLE_MODERATOR,\n", "return HttpResponseBadRequest(strip_tags(\"Unrecognized rolename '{}'.\".\n format(VAR_64)))\n", "VAR_19 = get_student_from_identifier(VAR_141)\n", "update_forum_role(VAR_10, VAR_19, VAR_64, VAR_57)\n", "return HttpResponseBadRequest('Role does not exist.')\n", "VAR_63 = {'course_id': VAR_10.to_deprecated_string(), 'action': VAR_57}\n", "return JsonResponse(VAR_63)\n" ]
[ "@ensure_csrf_cookie...\n", "\"\"\"docstring\"\"\"\n", "course_id = SlashSeparatedCourseKey.from_deprecated_string(course_id)\n", "course = get_course_by_id(course_id)\n", "has_instructor_access = has_access(request.user, 'instructor', course)\n", "has_forum_admin = has_forum_access(request.user, course_id,\n FORUM_ROLE_ADMINISTRATOR)\n", "unique_student_identifier = request.GET.get('unique_student_identifier')\n", "rolename = request.GET.get('rolename')\n", "action = request.GET.get('action')\n", "if not (has_forum_admin or has_instructor_access):\n", "return HttpResponseBadRequest(\n 'Operation requires staff & forum admin or instructor access')\n", "if rolename == FORUM_ROLE_ADMINISTRATOR and not has_instructor_access:\n", "return HttpResponseBadRequest('Operation requires instructor access.')\n", "if rolename not in [FORUM_ROLE_ADMINISTRATOR, FORUM_ROLE_MODERATOR,\n", "return HttpResponseBadRequest(strip_tags(\"Unrecognized rolename '{}'.\".\n format(rolename)))\n", "user = get_student_from_identifier(unique_student_identifier)\n", "update_forum_role(course_id, user, rolename, action)\n", "return HttpResponseBadRequest('Role does not exist.')\n", "response_payload = {'course_id': course_id.to_deprecated_string(), 'action':\n action}\n", "return JsonResponse(response_payload)\n" ]
[ 0, 0, 0, 0, 0, 0, 3, 3, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Return'", "Condition", "Return'", "Condition", "Return'", "Assign'", "Expr'", "Return'", "Assign'", "Return'" ]
[ "@def_function.function...\n", "return VAR_49 + 2 * VAR_55\n" ]
[ "@def_function.function...\n", "return y + 2 * c\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "def FUNC_70(VAR_12, VAR_29):...\n", "VAR_95 = FUNC_69(VAR_12, VAR_29)\n", "return FUNC_68(VAR_12, VAR_95)\n" ]
[ "def extract_user_identifier(user, filtr):...\n", "dynamic_field = extract_dynamic_field_from_filter(user, filtr)\n", "return extract_user_data_from_field(user, dynamic_field)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Return'" ]
[ "def FUNC_0(VAR_2: str, VAR_3: pathlib.Path) ->pathlib.Path:...\n", "VAR_12 = (VAR_3 / VAR_2).absolute()\n", "if not str(VAR_12).startswith(str(VAR_3) + '/'):\n", "return VAR_12\n" ]
[ "def sanitized_join(path: str, root: pathlib.Path) ->pathlib.Path:...\n", "result = (root / path).absolute()\n", "if not str(result).startswith(str(root) + '/'):\n", "return result\n" ]
[ 0, 1, 1, 1 ]
[ "FunctionDef'", "Assign'", "Condition", "Return'" ]
[ "def FUNC_0(VAR_5):...\n", "if VAR_3(VAR_5):\n", "return None\n", "return VAR_4(VAR_5)\n" ]
[ "def _is_javascript_scheme(s):...\n", "if _is_image_dataurl(s):\n", "return None\n", "return _is_possibly_malicious_scheme(s)\n" ]
[ 0, 2, 2, 2 ]
[ "FunctionDef'", "Condition", "Return'", "Return'" ]
[ "def FUNC_18(self, VAR_17, VAR_18):...\n", "if not VAR_53.session['admin']:\n", "VAR_66 = albumArtFilePath(VAR_17)\n", "VAR_64 = albumartfetcher.AlbumArtFetcher()\n", "VAR_20, VAR_67 = VAR_64.retrieveData(VAR_18)\n", "self.albumartcache_save(VAR_66, VAR_20)\n" ]
[ "def api_albumart_set(self, directory, imageurl):...\n", "if not cherrypy.session['admin']:\n", "b64imgpath = albumArtFilePath(directory)\n", "fetcher = albumartfetcher.AlbumArtFetcher()\n", "data, header = fetcher.retrieveData(imageurl)\n", "self.albumartcache_save(b64imgpath, data)\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Assign'", "Assign'", "Assign'", "Expr'" ]
[ "def FUNC_4(VAR_6: EscapableEntity) ->str:...\n", "\"\"\"docstring\"\"\"\n", "if isinstance(VAR_6, HTML):\n", "VAR_6 = VAR_6.__html__()\n", "if not isinstance(VAR_6, str):\n", "return u'%s' % VAR_6\n", "VAR_6 = ensure_str(VAR_6)\n", "while True:\n", "VAR_9 = VAR_6.find('<')\n", "if VAR_9 == -1:\n", "VAR_10 = VAR_6.find('>', VAR_9)\n", "return VAR_6.replace('&nbsp;', ' ')\n", "if VAR_10 == -1:\n", "VAR_6 = VAR_6[0:VAR_9] + VAR_6[VAR_10 + 1:]\n" ]
[ "def strip_tags(ht: EscapableEntity) ->str:...\n", "\"\"\"docstring\"\"\"\n", "if isinstance(ht, HTML):\n", "ht = ht.__html__()\n", "if not isinstance(ht, str):\n", "return u'%s' % ht\n", "ht = ensure_str(ht)\n", "while True:\n", "x = ht.find('<')\n", "if x == -1:\n", "y = ht.find('>', x)\n", "return ht.replace('&nbsp;', ' ')\n", "if y == -1:\n", "ht = ht[0:x] + ht[y + 1:]\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Assign'", "Condition", "Return'", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Return'", "Condition", "Assign'" ]
[ "def FUNC_3(self, VAR_5, *VAR_6, **VAR_4):...\n", "VAR_16 = VAR_5.POST.get('command')\n", "if VAR_16:\n", "VAR_24 = getattr(self, 'dispatch_%s' % VAR_16, None)\n", "if VAR_5.POST.get('save') and self.form and self.form.is_valid():\n", "if not callable(VAR_24):\n", "self.form.save()\n", "return self.get(VAR_5, *VAR_6, **kwargs)\n", "VAR_25 = dict(VAR_5.POST.items())\n", "self.save_layout()\n", "VAR_26 = VAR_24(**dispatch_kwargs)\n", "self.build_form()\n", "if VAR_26:\n", "if VAR_5.POST.get('publish') == '1':\n", "return VAR_26\n", "self.request.method = 'GET'\n", "return self.dispatch_publish()\n", "self.build_form()\n", "return super(CLASS_0, self).get(VAR_5, *VAR_6, **kwargs)\n" ]
[ "def post(self, request, *args, **kwargs):...\n", "command = request.POST.get('command')\n", "if command:\n", "dispatcher = getattr(self, 'dispatch_%s' % command, None)\n", "if request.POST.get('save') and self.form and self.form.is_valid():\n", "if not callable(dispatcher):\n", "self.form.save()\n", "return self.get(request, *args, **kwargs)\n", "dispatch_kwargs = dict(request.POST.items())\n", "self.save_layout()\n", "rv = dispatcher(**dispatch_kwargs)\n", "self.build_form()\n", "if rv:\n", "if request.POST.get('publish') == '1':\n", "return rv\n", "self.request.method = 'GET'\n", "return self.dispatch_publish()\n", "self.build_form()\n", "return super(EditorView, self).get(request, *args, **kwargs)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Condition", "Assign'", "For", "Condition", "Expr'", "Return'", "Assign'", "Expr'", "Assign'", "Expr'", "Condition", "Condition", "Return'", "Assign'", "Return'", "Expr'", "Return'" ]
[ "@action('copyrole', lazy_gettext('Copy Role'), lazy_gettext(...\n", "self.update_redirect()\n", "for VAR_21 in VAR_33:\n", "VAR_49 = VAR_21.__class__()\n", "return redirect(self.get_redirect())\n", "VAR_49.name = VAR_21.name\n", "VAR_49.permissions = VAR_21.permissions\n", "VAR_49.name = VAR_49.name + ' copy'\n", "self.datamodel.add(VAR_49)\n" ]
[ "@action('copyrole', lazy_gettext('Copy Role'), lazy_gettext(...\n", "self.update_redirect()\n", "for item in items:\n", "new_role = item.__class__()\n", "return redirect(self.get_redirect())\n", "new_role.name = item.name\n", "new_role.permissions = item.permissions\n", "new_role.name = new_role.name + ' copy'\n", "self.datamodel.add(new_role)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Expr'", "For", "Assign'", "Return'", "Assign'", "Assign'", "Assign'", "Expr'" ]
[ "def FUNC_1(self, **VAR_4):...\n", "return VAR_0\n" ]
[ "def generate_config_section(self, **kwargs):...\n", "return DEFAULT_CONFIG\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "async def FUNC_76(*VAR_20, **VAR_21):...\n", "return None\n" ]
[ "async def _insert_client_ip(*args, **kwargs):...\n", "return None\n" ]
[ 0, 0 ]
[ "AsyncFunctionDef'", "Return'" ]
[ "def FUNC_0(self) ->Dict[str, Any]:...\n", "VAR_1 = self.an_enum_value.value\n", "if isinstance(self.a_camel_date_time, datetime):\n", "VAR_5 = self.a_camel_date_time.isoformat()\n", "VAR_5 = self.a_camel_date_time.isoformat()\n", "VAR_2 = self.a_date.isoformat()\n", "if self.nested_list_of_enums is None:\n", "VAR_6 = None\n", "VAR_6 = []\n", "VAR_3 = self.some_dict\n", "for nested_list_of_enums_item_data in self.nested_list_of_enums:\n", "return {'an_enum_value': VAR_1, 'aCamelDateTime': VAR_5, 'a_date': VAR_2,\n 'nested_list_of_enums': VAR_6, 'some_dict': VAR_3}\n", "VAR_7 = []\n", "for nested_list_of_enums_item_item_data in nested_list_of_enums_item_data:\n", "VAR_8 = nested_list_of_enums_item_item_data.value\n", "VAR_6.append(VAR_7)\n", "VAR_7.append(VAR_8)\n" ]
[ "def to_dict(self) ->Dict[str, Any]:...\n", "an_enum_value = self.an_enum_value.value\n", "if isinstance(self.a_camel_date_time, datetime):\n", "a_camel_date_time = self.a_camel_date_time.isoformat()\n", "a_camel_date_time = self.a_camel_date_time.isoformat()\n", "a_date = self.a_date.isoformat()\n", "if self.nested_list_of_enums is None:\n", "nested_list_of_enums = None\n", "nested_list_of_enums = []\n", "some_dict = self.some_dict\n", "for nested_list_of_enums_item_data in self.nested_list_of_enums:\n", "return {'an_enum_value': an_enum_value, 'aCamelDateTime': a_camel_date_time,\n 'a_date': a_date, 'nested_list_of_enums': nested_list_of_enums,\n 'some_dict': some_dict}\n", "nested_list_of_enums_item = []\n", "for nested_list_of_enums_item_item_data in nested_list_of_enums_item_data:\n", "nested_list_of_enums_item_item = nested_list_of_enums_item_item_data.value\n", "nested_list_of_enums.append(nested_list_of_enums_item)\n", "nested_list_of_enums_item.append(nested_list_of_enums_item_item)\n" ]
[ 0, 0, 5, 0, 0, 0, 0, 0, 0, 5, 0, 5, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "For", "Return'", "Assign'", "For", "Assign'", "Expr'", "Expr'" ]
[ "def FUNC_62(VAR_95):...\n", "VAR_32 = re.compile('^plugin_' + VAR_96 + '(/.*|\\\\..*)?$')\n", "return [VAR_70 for VAR_70 in VAR_95 if VAR_70 and VAR_32.match(VAR_70)]\n" ]
[ "def filter_plugins(items):...\n", "regex = re.compile('^plugin_' + plugin + '(/.*|\\\\..*)?$')\n", "return [item for item in items if item and regex.match(item)]\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Return'" ]
[ "def FUNC_9():...\n", "yield '<?xml version=\"1.0\" encoding=\"UTF-8\" ?>\\n'\n", "if VAR_1:\n", "yield '<?xml-stylesheet type=\"text/xsl\" href=\"/scopelist.xsl\" ?>\\n'\n", "yield '<objectlist>\\n'\n", "for line in f:\n", "VAR_10 = line.strip()\n", "yield '</objectlist>\\n'\n", "VAR_13 = str(FUNC_5(VAR_10))\n", "VAR_25 = FUNC_7(VAR_13)\n", "print('Error parsing {}. {}. Skip.'.format(VAR_10, str(e)), file=sys.stderr)\n", "VAR_26 = float(VAR_25['format']['duration'])\n", "VAR_27 = int(ceil(VAR_26 / VAR_7))\n", "yield '<count adjust=\"{}\"/>\\n'.format(VAR_27)\n", "for clip in range(VAR_27):\n", "yield FUNC_4(VAR_9=clip * stride, VAR_8=span, VAR_10=video) + '\\n'\n" ]
[ "def generate():...\n", "yield '<?xml version=\"1.0\" encoding=\"UTF-8\" ?>\\n'\n", "if STYLE:\n", "yield '<?xml-stylesheet type=\"text/xsl\" href=\"/scopelist.xsl\" ?>\\n'\n", "yield '<objectlist>\\n'\n", "for line in f:\n", "video = line.strip()\n", "yield '</objectlist>\\n'\n", "video_path = str(_get_obj_absolute_path(video))\n", "video_meta = _ffprobe(video_path)\n", "print('Error parsing {}. {}. Skip.'.format(video, str(e)), file=sys.stderr)\n", "length_sec = float(video_meta['format']['duration'])\n", "num_clips = int(ceil(length_sec / stride))\n", "yield '<count adjust=\"{}\"/>\\n'.format(num_clips)\n", "for clip in range(num_clips):\n", "yield _get_object_element(start=clip * stride, span=span, video=video) + '\\n'\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Condition", "Expr'", "Expr'", "For", "Assign'", "Expr'", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "Expr'", "For", "Expr'" ]
[ "@login_required()...\n", "\"\"\"docstring\"\"\"\n", "VAR_157 = []\n", "VAR_158 = VAR_8.getRoiService()\n", "VAR_62 = VAR_158.findByImage(VAR_178(VAR_36), None, VAR_8.SERVICE_OPTS)\n", "for VAR_53 in VAR_62.rois:\n", "VAR_241 = {}\n", "VAR_157.sort(VAR_79=lambda x: x['id'])\n", "VAR_241['id'] = VAR_53.getId().getValue()\n", "return VAR_157\n", "VAR_63 = []\n", "for VAR_3 in VAR_53.copyShapes():\n", "if VAR_3 is None:\n", "VAR_63.sort(VAR_79=lambda x: '%03d%03d' % (x.get('theZ', -1), x.get('theT',\n -1)))\n", "VAR_63.append(shapeMarshal(VAR_3))\n", "VAR_241['shapes'] = VAR_63\n", "VAR_157.append(VAR_241)\n" ]
[ "@login_required()...\n", "\"\"\"docstring\"\"\"\n", "rois = []\n", "roiService = conn.getRoiService()\n", "result = roiService.findByImage(long(imageId), None, conn.SERVICE_OPTS)\n", "for r in result.rois:\n", "roi = {}\n", "rois.sort(key=lambda x: x['id'])\n", "roi['id'] = r.getId().getValue()\n", "return rois\n", "shapes = []\n", "for s in r.copyShapes():\n", "if s is None:\n", "shapes.sort(key=lambda x: '%03d%03d' % (x.get('theZ', -1), x.get('theT', -1)))\n", "shapes.append(shapeMarshal(s))\n", "roi['shapes'] = shapes\n", "rois.append(roi)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Assign'", "Assign'", "For", "Assign'", "Expr'", "Assign'", "Return'", "Assign'", "For", "Condition", "Expr'", "Expr'", "Assign'", "Expr'" ]
[ "def __init__(self, **VAR_22):...\n", "\"\"\"docstring\"\"\"\n", "if 'quote_char' in VAR_22:\n", "self.use_best_quote_char = False\n", "for attr in self.options:\n", "setattr(self, attr, VAR_22.get(attr, getattr(self, attr)))\n", "self.errors = []\n", "self.strict = False\n" ]
[ "def __init__(self, **kwargs):...\n", "\"\"\"docstring\"\"\"\n", "if 'quote_char' in kwargs:\n", "self.use_best_quote_char = False\n", "for attr in self.options:\n", "setattr(self, attr, kwargs.get(attr, getattr(self, attr)))\n", "self.errors = []\n", "self.strict = False\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Assign'", "For", "Expr'", "Assign'", "Assign'" ]
[ "def FUNC_14(VAR_3):...\n", "\"\"\"docstring\"\"\"\n", "if not VAR_16:\n", "if VAR_13.get('logging') or False:\n", "VAR_21.append(FUNC_1(VAR_3))\n", "print(repr(VAR_3))\n" ]
[ "def log(msg):...\n", "\"\"\"docstring\"\"\"\n", "if not request:\n", "if conf.get('logging') or False:\n", "debug_log.append(as_unicode(msg))\n", "print(repr(msg))\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Condition", "Expr'", "Expr'" ]
[ "def FUNC_20(self):...\n", "VAR_5, VAR_15 = self._test_confirm_start()\n", "VAR_3 = self.client.post(VAR_15, {'new_password1': 'anewpassword',\n 'new_password2': 'x'})\n", "self.assertFormError(VAR_3, SetPasswordForm.error_messages['password_mismatch']\n )\n" ]
[ "def test_confirm_different_passwords(self):...\n", "url, path = self._test_confirm_start()\n", "response = self.client.post(path, {'new_password1': 'anewpassword',\n 'new_password2': 'x'})\n", "self.assertFormError(response, SetPasswordForm.error_messages[\n 'password_mismatch'])\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Expr'" ]
[ "def FUNC_9(self):...\n", "self.register_user('id_admin', 'pass', admin=True)\n", "VAR_8 = self.login('id_admin', 'pass')\n", "self.register_user('id_nonadmin', 'pass', admin=False)\n", "VAR_22 = self.login('id_nonadmin', 'pass')\n", "VAR_23 = self.helper.upload_media(self.upload_resource, self.image_data,\n tok=admin_user_tok)\n", "VAR_24 = VAR_23['content_uri'][6:]\n", "VAR_25, VAR_26 = VAR_24.split('/')\n", "VAR_12, VAR_13 = make_request(self.reactor, FakeSite(self.download_resource\n ), 'GET', VAR_24, shorthand=False, VAR_7=non_admin_user_tok)\n", "self.assertEqual(200, int(VAR_13.code), msg=channel.result['body'])\n", "VAR_0 = '/_synapse/admin/v1/media/quarantine/%s/%s' % (urllib.parse.quote(\n VAR_25), urllib.parse.quote(VAR_26))\n", "VAR_12, VAR_13 = self.make_request('POST', VAR_0, VAR_7=admin_user_tok)\n", "self.pump(1.0)\n", "self.assertEqual(200, int(VAR_13.code), msg=channel.result['body'])\n", "self._ensure_quarantined(VAR_8, VAR_24)\n" ]
[ "def test_quarantine_media_by_id(self):...\n", "self.register_user('id_admin', 'pass', admin=True)\n", "admin_user_tok = self.login('id_admin', 'pass')\n", "self.register_user('id_nonadmin', 'pass', admin=False)\n", "non_admin_user_tok = self.login('id_nonadmin', 'pass')\n", "response = self.helper.upload_media(self.upload_resource, self.image_data,\n tok=admin_user_tok)\n", "server_name_and_media_id = response['content_uri'][6:]\n", "server_name, media_id = server_name_and_media_id.split('/')\n", "request, channel = make_request(self.reactor, FakeSite(self.\n download_resource), 'GET', server_name_and_media_id, shorthand=False,\n access_token=non_admin_user_tok)\n", "self.assertEqual(200, int(channel.code), msg=channel.result['body'])\n", "url = '/_synapse/admin/v1/media/quarantine/%s/%s' % (urllib.parse.quote(\n server_name), urllib.parse.quote(media_id))\n", "request, channel = self.make_request('POST', url, access_token=admin_user_tok)\n", "self.pump(1.0)\n", "self.assertEqual(200, int(channel.code), msg=channel.result['body'])\n", "self._ensure_quarantined(admin_user_tok, server_name_and_media_id)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Assign'", "Expr'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'" ]
[ "def FUNC_2(self, VAR_2, VAR_3, VAR_4):...\n", "self.store = VAR_4.get_datastore()\n", "self.admin_user = self.register_user('admin', 'pass', admin=True)\n", "self.admin_user_tok = self.login('admin', 'pass')\n", "self.other_user = self.register_user('user', 'pass')\n", "self.other_user_token = self.login('user', 'pass')\n" ]
[ "def prepare(self, reactor, clock, hs):...\n", "self.store = hs.get_datastore()\n", "self.admin_user = self.register_user('admin', 'pass', admin=True)\n", "self.admin_user_tok = self.login('admin', 'pass')\n", "self.other_user = self.register_user('user', 'pass')\n", "self.other_user_token = self.login('user', 'pass')\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'" ]
[ "@VAR_2.route('/shutdown')...\n", "VAR_32 = VAR_119(request.args.get('parameter').strip())\n", "VAR_33 = {}\n", "if VAR_32 in (0, 1):\n", "calibre_db.dispose()\n", "if VAR_32 == 2:\n", "ub.dispose()\n", "VAR_0.warning('reconnecting to calibre database')\n", "VAR_33['text'] = _(u'Unknown command')\n", "if VAR_32 == 0:\n", "calibre_db.reconnect_db(config, ub.app_DB_path)\n", "return json.dumps(VAR_33), 400\n", "VAR_33['text'] = _(u'Server restarted, please reload page')\n", "VAR_33['text'] = _(u'Performing shutdown of server, please close window')\n", "VAR_33['text'] = _(u'Reconnect successful')\n", "web_server.stop(VAR_32 == 0)\n", "return json.dumps(VAR_33)\n", "return json.dumps(VAR_33)\n" ]
[ "@admi.route('/shutdown')...\n", "task = int(request.args.get('parameter').strip())\n", "showtext = {}\n", "if task in (0, 1):\n", "calibre_db.dispose()\n", "if task == 2:\n", "ub.dispose()\n", "log.warning('reconnecting to calibre database')\n", "showtext['text'] = _(u'Unknown command')\n", "if task == 0:\n", "calibre_db.reconnect_db(config, ub.app_DB_path)\n", "return json.dumps(showtext), 400\n", "showtext['text'] = _(u'Server restarted, please reload page')\n", "showtext['text'] = _(u'Performing shutdown of server, please close window')\n", "showtext['text'] = _(u'Reconnect successful')\n", "web_server.stop(task == 0)\n", "return json.dumps(showtext)\n", "return json.dumps(showtext)\n" ]
[ 0, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Assign'", "Assign'", "Condition", "Expr'", "Condition", "Expr'", "Expr'", "Assign'", "Condition", "Expr'", "Return'", "Assign'", "Assign'", "Assign'", "Expr'", "Return'", "Return'" ]
[ "def FUNC_20(self):...\n", "VAR_27 = np.array([[1], [2]])\n", "VAR_32 = os.path.join(test.get_temp_dir(), 'input.npz')\n", "np.savez(VAR_32, VAR_54=x0)\n", "VAR_28 = np.ones([2, 10])\n", "VAR_21 = 'x0=' + VAR_32 + '[a]'\n", "VAR_22 = 'x0=np.ones([2,10])'\n", "VAR_31 = saved_model_cli.load_inputs_from_input_arg_string(VAR_21, VAR_22, '')\n", "self.assertTrue(np.all(VAR_31['x0'] == VAR_28))\n" ]
[ "def testInputParserBothDuplicate(self):...\n", "x0 = np.array([[1], [2]])\n", "input_path = os.path.join(test.get_temp_dir(), 'input.npz')\n", "np.savez(input_path, a=x0)\n", "x1 = np.ones([2, 10])\n", "input_str = 'x0=' + input_path + '[a]'\n", "input_expr_str = 'x0=np.ones([2,10])'\n", "feed_dict = saved_model_cli.load_inputs_from_input_arg_string(input_str,\n input_expr_str, '')\n", "self.assertTrue(np.all(feed_dict['x0'] == x1))\n" ]
[ 0, 5, 5, 5, 5, 5, 5, 5, 5 ]
[ "FunctionDef'", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'" ]
[ "@staticmethod...\n", "FUNC_18()\n" ]
[ "@staticmethod...\n", "_callback()\n" ]
[ 0, 0 ]
[ "Condition", "Expr'" ]
[ "@staticmethod...\n", "\"\"\"docstring\"\"\"\n", "VAR_18 = CLASS_4()\n", "for VAR_11, VAR_5 in VAR_12.items():\n", "if isinstance(VAR_5, oai.Reference):\n", "return VAR_18\n", "VAR_18.errors.append(ParseError(VAR_5=data, detail=\n 'Reference schemas are not supported.'))\n", "if VAR_5.enum is not None:\n", "EnumProperty(VAR_11=name, title=data.title or name, VAR_26=True, default=\n data.default, values=EnumProperty.values_from_list(data.enum))\n", "VAR_28 = CLASS_3.from_data(VAR_5=data, VAR_11=name)\n", "if isinstance(VAR_28, ParseError):\n", "VAR_18.errors.append(VAR_28)\n", "VAR_18.models[VAR_28.reference.class_name] = VAR_28\n" ]
[ "@staticmethod...\n", "\"\"\"docstring\"\"\"\n", "result = Schemas()\n", "for name, data in schemas.items():\n", "if isinstance(data, oai.Reference):\n", "return result\n", "result.errors.append(ParseError(data=data, detail=\n 'Reference schemas are not supported.'))\n", "if data.enum is not None:\n", "EnumProperty(name=name, title=data.title or name, required=True, default=\n data.default, values=EnumProperty.values_from_list(data.enum))\n", "s = Model.from_data(data=data, name=name)\n", "if isinstance(s, ParseError):\n", "result.errors.append(s)\n", "result.models[s.reference.class_name] = s\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "For", "Condition", "Return'", "Expr'", "Condition", "Expr'", "Assign'", "Condition", "Expr'", "Assign'" ]
[ "def FUNC_12(self):...\n", "p('Exported terms:')\n", "for VAR_19 in findAll('[data-export]', self):\n", "for term in config.linkTextsFromElement(VAR_19):\n", "p('Unexported terms:')\n", "p(' ' + term)\n", "for VAR_19 in findAll('[data-noexport]', self):\n", "for term in config.linkTextsFromElement(VAR_19):\n", "p(' ' + term)\n" ]
[ "def printTargets(self):...\n", "p('Exported terms:')\n", "for el in findAll('[data-export]', self):\n", "for term in config.linkTextsFromElement(el):\n", "p('Unexported terms:')\n", "p(' ' + term)\n", "for el in findAll('[data-noexport]', self):\n", "for term in config.linkTextsFromElement(el):\n", "p(' ' + term)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "For", "For", "Expr'", "Expr'", "For", "For", "Expr'" ]
[ "def FUNC_17(self):...\n", "VAR_27 = np.ones([2, 10])\n", "VAR_38 = np.array([[1], [2], [3]])\n", "VAR_39 = np.mgrid[0:5, 0:5]\n", "VAR_40 = [[3], [4]]\n", "VAR_21 = (\n 'x1=np.ones([2,10]);x2=np.array([[1],[2],[3]]);x3=np.mgrid[0:5,0:5];x4=[[3],[4]]'\n )\n", "VAR_30 = saved_model_cli.load_inputs_from_input_arg_string('', VAR_21, '')\n", "self.assertTrue(np.all(VAR_30['x1'] == VAR_27))\n", "self.assertTrue(np.all(VAR_30['x2'] == VAR_38))\n", "self.assertTrue(np.all(VAR_30['x3'] == VAR_39))\n", "self.assertTrue(np.all(VAR_30['x4'] == VAR_40))\n" ]
[ "def testInputParserPythonExpression(self):...\n", "x1 = np.ones([2, 10])\n", "x2 = np.array([[1], [2], [3]])\n", "x3 = np.mgrid[0:5, 0:5]\n", "x4 = [[3], [4]]\n", "input_expr_str = (\n 'x1=np.ones([2,10]);x2=np.array([[1],[2],[3]]);x3=np.mgrid[0:5,0:5];x4=[[3],[4]]'\n )\n", "feed_dict = saved_model_cli.load_inputs_from_input_arg_string('',\n input_expr_str, '')\n", "self.assertTrue(np.all(feed_dict['x1'] == x1))\n", "self.assertTrue(np.all(feed_dict['x2'] == x2))\n", "self.assertTrue(np.all(feed_dict['x3'] == x3))\n", "self.assertTrue(np.all(feed_dict['x4'] == x4))\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'" ]
[ "def FUNC_98(self, VAR_10):...\n", "\"\"\"docstring\"\"\"\n", "self.json_procedures[VAR_10.__name__] = VAR_10\n", "return VAR_10\n" ]
[ "def json(self, f):...\n", "\"\"\"docstring\"\"\"\n", "self.json_procedures[f.__name__] = f\n", "return f\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Return'" ]
[ "def FUNC_2(self, VAR_1, VAR_2=None):...\n", "" ]
[ "def get(self, arg, word=None):...\n", "" ]
[ 0, 0 ]
[ "FunctionDef'", "Condition" ]
[ "def FUNC_13(VAR_13, VAR_14, VAR_15, VAR_16=None):...\n", "" ]
[ "def create_and_login_user(org, name, email, picture=None):...\n", "" ]
[ 0, 0 ]
[ "FunctionDef'", "Condition" ]
[ "def FUNC_3(self, *, VAR_4: str) ->Set[str]:...\n", "\"\"\"docstring\"\"\"\n", "VAR_8 = super().get_imports(VAR_4=prefix)\n", "VAR_8.add('from typing import Dict')\n", "if self.default is not None:\n", "VAR_8.add('from dataclasses import field')\n", "return VAR_8\n", "VAR_8.add('from typing import cast')\n" ]
[ "def get_imports(self, *, prefix: str) ->Set[str]:...\n", "\"\"\"docstring\"\"\"\n", "imports = super().get_imports(prefix=prefix)\n", "imports.add('from typing import Dict')\n", "if self.default is not None:\n", "imports.add('from dataclasses import field')\n", "return imports\n", "imports.add('from typing import cast')\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Expr'", "Condition", "Expr'", "Return'", "Expr'" ]
[ "def FUNC_118(*VAR_11, **VAR_351):...\n", "" ]
[ "def f(*args, **kwargs):...\n", "" ]
[ 0, 0 ]
[ "FunctionDef'", "Condition" ]
[ "def FUNC_12(VAR_23, VAR_24, VAR_27, VAR_28=None, VAR_29=1):...\n", "if 'mandatory' in VAR_27:\n", "for mandatory in VAR_27['mandatory']:\n", "if VAR_28 is None:\n", "if mandatory not in VAR_28:\n", "VAR_28 = {}\n", "VAR_10 = dict(VAR_28)\n", "return None\n", "if 'suffix' not in VAR_10 and VAR_29 > 1:\n", "VAR_10['suffix'] = '_%d' % VAR_29\n", "if 'div' in VAR_10:\n", "VAR_10['_div'] = VAR_10['div']\n", "if 'div' in VAR_27:\n", "if VAR_10.get('template'):\n", "VAR_10['_div'] = VAR_27['div'](VAR_23)\n", "VAR_10['template'] = VAR_24.template_folder_key + '/' + VAR_10['template']\n", "VAR_10['template'] = VAR_24.template_folder_key + '/' + VAR_27['template'](\n VAR_23)\n", "if 'suffix' in VAR_10:\n", "if VAR_10.get('template_header'):\n", "VAR_10['_div'] = VAR_10['_div'] + VAR_10['suffix']\n", "if not VAR_5.match(VAR_10['_div']):\n", "VAR_10['template_header'] = VAR_24.template_folder_key + '/' + VAR_10[\n 'template_header']\n", "if 'name' not in VAR_10:\n", "VAR_0.warning(\n 'Template config {} contains invalid div identifier {}, skipping it'.\n format(VAR_23, VAR_10['_div']))\n", "VAR_10['name'] = VAR_24._plugin_name\n", "if 'custom_bindings' not in VAR_10 or VAR_10['custom_bindings']:\n", "return None\n", "VAR_123 = 'allowBindings: true'\n", "VAR_10['_key'] = 'plugin_' + VAR_23\n", "if 'data_bind' in VAR_10:\n", "if 'suffix' in VAR_10:\n", "VAR_123 = VAR_123 + ', ' + VAR_10['data_bind']\n", "VAR_123 = VAR_123.replace('\"', '\\\\\"')\n", "VAR_10['_key'] += VAR_10['suffix']\n", "VAR_10['_plugin'] = VAR_23\n", "VAR_10['data_bind'] = VAR_123\n", "return VAR_10\n" ]
[ "def _process_template_config(name, implementation, rule, config=None, counter=1...\n", "if 'mandatory' in rule:\n", "for mandatory in rule['mandatory']:\n", "if config is None:\n", "if mandatory not in config:\n", "config = {}\n", "data = dict(config)\n", "return None\n", "if 'suffix' not in data and counter > 1:\n", "data['suffix'] = '_%d' % counter\n", "if 'div' in data:\n", "data['_div'] = data['div']\n", "if 'div' in rule:\n", "if data.get('template'):\n", "data['_div'] = rule['div'](name)\n", "data['template'] = implementation.template_folder_key + '/' + data['template']\n", "data['template'] = implementation.template_folder_key + '/' + rule['template'](\n name)\n", "if 'suffix' in data:\n", "if data.get('template_header'):\n", "data['_div'] = data['_div'] + data['suffix']\n", "if not _valid_div_re.match(data['_div']):\n", "data['template_header'] = implementation.template_folder_key + '/' + data[\n 'template_header']\n", "if 'name' not in data:\n", "_logger.warning(\n 'Template config {} contains invalid div identifier {}, skipping it'.\n format(name, data['_div']))\n", "data['name'] = implementation._plugin_name\n", "if 'custom_bindings' not in data or data['custom_bindings']:\n", "return None\n", "data_bind = 'allowBindings: true'\n", "data['_key'] = 'plugin_' + name\n", "if 'data_bind' in data:\n", "if 'suffix' in data:\n", "data_bind = data_bind + ', ' + data['data_bind']\n", "data_bind = data_bind.replace('\"', '\\\\\"')\n", "data['_key'] += data['suffix']\n", "data['_plugin'] = name\n", "data['data_bind'] = data_bind\n", "return data\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "For", "Condition", "Condition", "Assign'", "Assign'", "Return'", "Condition", "Assign'", "Condition", "Assign'", "Condition", "Condition", "Assign'", "Assign'", "Assign'", "Condition", "Condition", "Assign'", "Condition", "Assign'", "Condition", "Expr'", "Assign'", "Condition", "Return'", "Assign'", "Assign'", "Condition", "Condition", "Assign'", "Assign'", "AugAssign'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_27(self):...\n", "for test_str in ('Iñtërnâtiônàlizætiøn', \"1234'\", -1234):\n", "VAR_21 = self.client.post('%s/admin_widgets/event/add/' % self.admin_root,\n {'main_band': test_str})\n", "self.assertContains(VAR_21,\n 'Select a valid choice. That choice is not one of the available choices.')\n" ]
[ "def test_invalid_target_id(self):...\n", "for test_str in ('Iñtërnâtiônàlizætiøn', \"1234'\", -1234):\n", "response = self.client.post('%s/admin_widgets/event/add/' % self.admin_root,\n {'main_band': test_str})\n", "self.assertContains(response,\n 'Select a valid choice. That choice is not one of the available choices.')\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "For", "Assign'", "Expr'" ]
[ "def FUNC_9(self):...\n", "return '-'.join(p.replace('-', '_-') for p in self.namespace.split('/'))\n" ]
[ "def get_build_slug(self):...\n", "return '-'.join(p.replace('-', '_-') for p in self.namespace.split('/'))\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "from django.contrib.auth.decorators import login_required\n", "from django.shortcuts import render, redirect, get_object_or_404\n", "from ...core.utils.views import is_post, post_data\n", "from ..models import Comment\n", "from .forms import FlagForm\n", "@login_required...\n", "VAR_2 = get_object_or_404(Comment, pk=comment_id)\n", "VAR_3 = FlagForm(user=request.user, VAR_2=comment, data=post_data(request))\n", "if is_post(VAR_0) and VAR_3.is_valid():\n", "VAR_3.save()\n", "return render(VAR_0=request, template_name=\n 'spirit/comment/flag/create.html', context={'form': form, 'comment':\n comment})\n", "return redirect(VAR_0.POST.get('next', VAR_2.get_absolute_url()))\n" ]
[ "from django.contrib.auth.decorators import login_required\n", "from django.shortcuts import render, redirect, get_object_or_404\n", "from ...core.utils.views import is_post, post_data\n", "from ..models import Comment\n", "from .forms import FlagForm\n", "@login_required...\n", "comment = get_object_or_404(Comment, pk=comment_id)\n", "form = FlagForm(user=request.user, comment=comment, data=post_data(request))\n", "if is_post(request) and form.is_valid():\n", "form.save()\n", "return render(request=request, template_name=\n 'spirit/comment/flag/create.html', context={'form': form, 'comment':\n comment})\n", "return redirect(request.POST.get('next', comment.get_absolute_url()))\n" ]
[ 0, 0, 4, 0, 0, 0, 0, 0, 0, 0, 0, 4 ]
[ "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Condition", "Assign'", "Assign'", "Condition", "Expr'", "Return'", "Return'" ]
[ "def FUNC_30(VAR_16):...\n", "VAR_35 = PerspectivesKeyFetcher(self.hs)\n", "VAR_23 = {VAR_34: {'key1': 0}}\n", "self.expect_outgoing_key_query(VAR_34, 'key1', VAR_16)\n", "return self.get_success(VAR_35.get_keys(VAR_23))\n" ]
[ "def get_key_from_perspectives(response):...\n", "fetcher = PerspectivesKeyFetcher(self.hs)\n", "keys_to_fetch = {SERVER_NAME: {'key1': 0}}\n", "self.expect_outgoing_key_query(SERVER_NAME, 'key1', response)\n", "return self.get_success(fetcher.get_keys(keys_to_fetch))\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Expr'", "Return'" ]
[ "def FUNC_14(self, VAR_6):...\n", "if not self.query:\n", "frappe.throw(_('Must specify a Query to run'), title=_('Report Document Error')\n )\n", "if not self.query.lower().startswith('select'):\n", "frappe.throw(_('Query must be a SELECT'), title=_('Report Document Error'))\n", "VAR_15 = [list(t) for t in frappe.db.sql(self.query, VAR_6, debug=True)]\n", "VAR_13 = self.get_columns() or [cstr(c[0]) for c in frappe.db.get_description()\n ]\n", "return [VAR_13, VAR_15]\n" ]
[ "def execute_query_report(self, filters):...\n", "if not self.query:\n", "frappe.throw(_('Must specify a Query to run'), title=_('Report Document Error')\n )\n", "if not self.query.lower().startswith('select'):\n", "frappe.throw(_('Query must be a SELECT'), title=_('Report Document Error'))\n", "result = [list(t) for t in frappe.db.sql(self.query, filters, debug=True)]\n", "columns = self.get_columns() or [cstr(c[0]) for c in frappe.db.\n get_description()]\n", "return [columns, result]\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Expr'", "Condition", "Expr'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_15(VAR_3, **VAR_4):...\n", "VAR_2.update(VAR_4)\n", "return 200, VAR_4\n" ]
[ "def _callback(request, **kwargs):...\n", "got_kwargs.update(kwargs)\n", "return 200, kwargs\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Return'" ]
[ "def FUNC_7(self):...\n", "return logging.INFO\n" ]
[ "def _log_level_default(self):...\n", "return logging.INFO\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_53(self):...\n", "frappe.clear_document_cache(self.doctype, self.name)\n" ]
[ "def clear_cache(self):...\n", "frappe.clear_document_cache(self.doctype, self.name)\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Expr'" ]
[ "def FUNC_43(self, VAR_21, VAR_22, VAR_20):...\n", "if 'view' in VAR_92:\n", "if not VAR_92['view'].get(VAR_21):\n", "return None\n", "VAR_92['view'][VAR_21] = dict()\n", "VAR_92['view'][VAR_21][VAR_22] = VAR_20\n" ]
[ "def set_view_property(self, page, prop, value):...\n", "if 'view' in flask_session:\n", "if not flask_session['view'].get(page):\n", "return None\n", "flask_session['view'][page] = dict()\n", "flask_session['view'][page][prop] = value\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Condition", "Return'", "Assign'", "Assign'" ]
[ "def FUNC_3(self):...\n", "self.transitiveDependencies.clear()\n", "self.assembleDocument()\n", "self.processDocument()\n" ]
[ "def preprocess(self):...\n", "self.transitiveDependencies.clear()\n", "self.assembleDocument()\n", "self.processDocument()\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Expr'", "Expr'" ]
[ "def FUNC_20(self, VAR_6, VAR_7, VAR_8):...\n", "VAR_12 = json.loads(self.json)\n", "VAR_13 = self.get_standard_report_columns(VAR_12)\n", "VAR_15 = []\n", "VAR_26, VAR_27, VAR_14 = self.get_standard_report_order_by(VAR_12)\n", "VAR_28 = frappe.get_list(self.ref_doctype, fields=[(get_group_by_field(\n group_by_args, c[1]) if c[0] == '_aggregate_column' and group_by_args else\n Report._format([c[1], c[0]])) for c in columns], VAR_6=self.\n get_standard_report_filters(params, filters), VAR_26=order_by, VAR_27=\n group_by, as_list=True, VAR_7=limit, VAR_8=user)\n", "VAR_13 = self.build_standard_report_columns(VAR_13, VAR_14)\n", "VAR_15 = VAR_15 + [list(d) for d in VAR_28]\n", "if VAR_12.get('add_totals_row'):\n", "VAR_15 = append_totals_row(VAR_15)\n", "return VAR_13, VAR_15\n" ]
[ "def run_standard_report(self, filters, limit, user):...\n", "params = json.loads(self.json)\n", "columns = self.get_standard_report_columns(params)\n", "result = []\n", "order_by, group_by, group_by_args = self.get_standard_report_order_by(params)\n", "_result = frappe.get_list(self.ref_doctype, fields=[(get_group_by_field(\n group_by_args, c[1]) if c[0] == '_aggregate_column' and group_by_args else\n Report._format([c[1], c[0]])) for c in columns], filters=self.\n get_standard_report_filters(params, filters), order_by=order_by,\n group_by=group_by, as_list=True, limit=limit, user=user)\n", "columns = self.build_standard_report_columns(columns, group_by_args)\n", "result = result + [list(d) for d in _result]\n", "if params.get('add_totals_row'):\n", "result = append_totals_row(result)\n", "return columns, result\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Return'" ]
[ "@log_function...\n", "\"\"\"docstring\"\"\"\n", "VAR_2 = FUNC_1('/groups/%s/users/%s/accept_invite', VAR_30, VAR_16)\n", "return self.client.post_json(VAR_5=destination, VAR_2=path, VAR_39=content,\n VAR_15=True)\n" ]
[ "@log_function...\n", "\"\"\"docstring\"\"\"\n", "path = _create_v1_path('/groups/%s/users/%s/accept_invite', group_id, user_id)\n", "return self.client.post_json(destination=destination, path=path, data=\n content, ignore_backoff=True)\n" ]
[ 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Return'" ]
[ "def FUNC_31(self):...\n", "VAR_5 = self._makeContext()\n", "VAR_5.evaluate('context/__class__')\n", "VAR_5.evaluate('nocall: random/_itertools/repeat')\n", "VAR_5.evaluate('random/_itertools/repeat/foobar')\n" ]
[ "def test_underscore_traversal(self):...\n", "ec = self._makeContext()\n", "ec.evaluate('context/__class__')\n", "ec.evaluate('nocall: random/_itertools/repeat')\n", "ec.evaluate('random/_itertools/repeat/foobar')\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Expr'", "Expr'", "Expr'" ]
[ "def FUNC_104():...\n", "from frappe.core.doctype.domain_settings.domain_settings import get_active_domains\n", "return FUNC_104()\n" ]
[ "def get_active_domains():...\n", "from frappe.core.doctype.domain_settings.domain_settings import get_active_domains\n", "return get_active_domains()\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "ImportFrom'", "Return'" ]
[ "def FUNC_33(self):...\n", "self.parser = saved_model_cli.create_parser()\n", "VAR_9 = test.test_src_dir_path(VAR_0)\n", "VAR_44 = np.array([[1], [2]])\n", "VAR_45 = np.zeros((6, 3))\n", "VAR_31 = os.path.join(test.get_temp_dir(), 'testRunCommandNewOutdir_inputs.npz'\n )\n", "VAR_41 = os.path.join(test.get_temp_dir(), 'new_dir')\n", "if os.path.isdir(VAR_41):\n", "shutil.rmtree(VAR_41)\n", "np.savez(VAR_31, VAR_26=x, VAR_27=x_notused)\n", "VAR_10 = self.parser.parse_args(['run', '--dir', VAR_9, '--tag_set',\n 'serve', '--signature_def', 'serving_default', '--inputs', 'x=' +\n VAR_31 + '[x0]', '--outdir', VAR_41, '--tf_debug'])\n", "def FUNC_38(VAR_47):...\n", "return VAR_47\n" ]
[ "def testRunCommandWithDebuggerEnabled(self):...\n", "self.parser = saved_model_cli.create_parser()\n", "base_path = test.test_src_dir_path(SAVED_MODEL_PATH)\n", "x = np.array([[1], [2]])\n", "x_notused = np.zeros((6, 3))\n", "input_path = os.path.join(test.get_temp_dir(),\n 'testRunCommandNewOutdir_inputs.npz')\n", "output_dir = os.path.join(test.get_temp_dir(), 'new_dir')\n", "if os.path.isdir(output_dir):\n", "shutil.rmtree(output_dir)\n", "np.savez(input_path, x0=x, x1=x_notused)\n", "args = self.parser.parse_args(['run', '--dir', base_path, '--tag_set',\n 'serve', '--signature_def', 'serving_default', '--inputs', 'x=' +\n input_path + '[x0]', '--outdir', output_dir, '--tf_debug'])\n", "def fake_wrapper_session(sess):...\n", "return sess\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Expr'", "Expr'", "Assign'", "FunctionDef'", "Return'" ]
[ "def FUNC_105(VAR_62, VAR_9, VAR_166=None, VAR_167=False, VAR_168=True):...\n", "\"\"\"docstring\"\"\"\n", "VAR_202 = FUNC_48(VAR_62)\n", "if VAR_202.track_changes:\n", "VAR_216 = VAR_12.get_all('Version', VAR_106={'ref_doctype': doctype,\n 'docname': name, 'order_by': 'creation' if head else None, 'limit':\n limit}, VAR_34=1)\n", "if VAR_168:\n", "from frappe.chat.util import squashify, dictify, safe_json_loads\n", "VAR_217 = []\n", "for VAR_9 in VAR_216:\n", "VAR_9 = squashify(VAR_9)\n", "return VAR_217\n", "VAR_94 = FUNC_45('Version', VAR_9)\n", "VAR_224 = VAR_94.data\n", "VAR_224 = safe_json_loads(VAR_224)\n", "VAR_224 = dictify(dict(version=data, VAR_10=doc.owner, creation=doc.creation))\n", "VAR_217.append(VAR_224)\n" ]
[ "def get_version(doctype, name, limit=None, head=False, raise_err=True):...\n", "\"\"\"docstring\"\"\"\n", "meta = get_meta(doctype)\n", "if meta.track_changes:\n", "names = db.get_all('Version', filters={'ref_doctype': doctype, 'docname':\n name, 'order_by': 'creation' if head else None, 'limit': limit}, as_list=1)\n", "if raise_err:\n", "from frappe.chat.util import squashify, dictify, safe_json_loads\n", "versions = []\n", "for name in names:\n", "name = squashify(name)\n", "return versions\n", "doc = get_doc('Version', name)\n", "data = doc.data\n", "data = safe_json_loads(data)\n", "data = dictify(dict(version=data, user=doc.owner, creation=doc.creation))\n", "versions.append(data)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Condition", "Assign'", "Condition", "ImportFrom'", "Assign'", "For", "Assign'", "Return'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'" ]
[ "@ensure_csrf_cookie...\n", "\"\"\"docstring\"\"\"\n", "VAR_10 = SlashSeparatedCourseKey.from_deprecated_string(VAR_10)\n", "VAR_131 = 'bulk_course_email'\n", "VAR_133 = instructor_task.api.get_instructor_task_history(VAR_10, VAR_131=\n task_type)\n", "VAR_63 = {'emails': map(extract_email_features, VAR_133)}\n", "return JsonResponse(VAR_63)\n" ]
[ "@ensure_csrf_cookie...\n", "\"\"\"docstring\"\"\"\n", "course_id = SlashSeparatedCourseKey.from_deprecated_string(course_id)\n", "task_type = 'bulk_course_email'\n", "emails = instructor_task.api.get_instructor_task_history(course_id,\n task_type=task_type)\n", "response_payload = {'emails': map(extract_email_features, emails)}\n", "return JsonResponse(response_payload)\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_6(self):...\n", "return self.kw_project\n" ]
[ "def get_kwargs(self):...\n", "return self.kw_project\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "@wraps(VAR_28)...\n", "VAR_1.debug('jsonp')\n", "VAR_19 = VAR_9.get('server_id', None)\n", "VAR_315 = 500\n", "if VAR_19 is None and VAR_2.session.get('connector'):\n", "if isinstance(ex, omero.SecurityViolation):\n", "VAR_19 = VAR_2.session['connector'].server_id\n", "VAR_9['server_id'] = VAR_19\n", "VAR_315 = 403\n", "if isinstance(ex, omero.ApiUsageException):\n", "VAR_54 = VAR_28(VAR_2, *VAR_116, **kwargs)\n", "VAR_316 = traceback.format_exc()\n", "VAR_315 = 400\n", "if VAR_9.get('_raw', False):\n", "VAR_1.debug(VAR_316)\n", "return VAR_54\n", "if isinstance(VAR_54, HttpResponse):\n", "if VAR_9.get('_raw', False) or VAR_9.get('_internal', False):\n", "return VAR_54\n", "VAR_203 = VAR_2.GET.get('callback', None)\n", "return JsonResponse({'message': str(ex), 'stacktrace': VAR_316}, VAR_315=status\n )\n", "if VAR_203 is not None and not VAR_9.get('_internal', False):\n", "VAR_54 = json.dumps(VAR_54)\n", "if VAR_9.get('_internal', False):\n", "VAR_54 = '%s(%s)' % (VAR_203, VAR_54)\n", "return VAR_54\n", "VAR_281 = type(VAR_54) is dict\n", "return HttpJavascriptResponse(VAR_54)\n", "return JsonResponse(VAR_54, VAR_281=safe)\n" ]
[ "@wraps(f)...\n", "logger.debug('jsonp')\n", "server_id = kwargs.get('server_id', None)\n", "status = 500\n", "if server_id is None and request.session.get('connector'):\n", "if isinstance(ex, omero.SecurityViolation):\n", "server_id = request.session['connector'].server_id\n", "kwargs['server_id'] = server_id\n", "status = 403\n", "if isinstance(ex, omero.ApiUsageException):\n", "rv = f(request, *args, **kwargs)\n", "trace = traceback.format_exc()\n", "status = 400\n", "if kwargs.get('_raw', False):\n", "logger.debug(trace)\n", "return rv\n", "if isinstance(rv, HttpResponse):\n", "if kwargs.get('_raw', False) or kwargs.get('_internal', False):\n", "return rv\n", "c = request.GET.get('callback', None)\n", "return JsonResponse({'message': str(ex), 'stacktrace': trace}, status=status)\n", "if c is not None and not kwargs.get('_internal', False):\n", "rv = json.dumps(rv)\n", "if kwargs.get('_internal', False):\n", "rv = '%s(%s)' % (c, rv)\n", "return rv\n", "safe = type(rv) is dict\n", "return HttpJavascriptResponse(rv)\n", "return JsonResponse(rv, safe=safe)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Expr'", "Assign'", "Assign'", "Condition", "Condition", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Condition", "Expr'", "Return'", "Condition", "Condition", "Return'", "Assign'", "Return'", "Condition", "Assign'", "Condition", "Assign'", "Return'", "Assign'", "Return'", "Return'" ]
[ "def FUNC_0(self, VAR_1, VAR_2):...\n", "VAR_3 = self.setup_test_homeserver(resource_for_federation=Mock(),\n http_client=None)\n", "return VAR_3\n" ]
[ "def make_homeserver(self, reactor, clock):...\n", "hs = self.setup_test_homeserver(resource_for_federation=Mock(), http_client\n =None)\n", "return hs\n" ]
[ 0, 4, 0 ]
[ "FunctionDef'", "Assign'", "Return'" ]
[ "@FUNC_0...\n", "if self.config.worker_app:\n", "return GroupsLocalWorkerHandler(self)\n", "return GroupsLocalHandler(self)\n" ]
[ "@cache_in_self...\n", "if self.config.worker_app:\n", "return GroupsLocalWorkerHandler(self)\n", "return GroupsLocalHandler(self)\n" ]
[ 0, 0, 0, 0 ]
[ "Condition", "Condition", "Return'", "Return'" ]
[ "def __init__(self, *VAR_17, **VAR_18):...\n", "super().__init__(*VAR_17, **kwargs)\n", "self.url, VAR_10 = self.spec.split('/', 1)\n", "self.repo = urllib.parse.unquote(self.url)\n", "self.unresolved_ref = urllib.parse.unquote(VAR_10)\n", "if not self.unresolved_ref:\n" ]
[ "def __init__(self, *args, **kwargs):...\n", "super().__init__(*args, **kwargs)\n", "self.url, unresolved_ref = self.spec.split('/', 1)\n", "self.repo = urllib.parse.unquote(self.url)\n", "self.unresolved_ref = urllib.parse.unquote(unresolved_ref)\n", "if not self.unresolved_ref:\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Assign'", "Assign'", "Assign'", "Condition" ]
[ "def FUNC_52(self):...\n", "\"\"\"docstring\"\"\"\n", "VAR_69 = self.get_doc_before_save()\n", "if self._action == 'save':\n", "self.run_method('on_update')\n", "if self._action == 'submit':\n", "self.clear_cache()\n", "self.run_method('on_update')\n", "if self._action == 'cancel':\n", "self.notify_update()\n", "self.run_method('on_submit')\n", "self.run_method('on_cancel')\n", "if self._action == 'update_after_submit':\n", "update_global_search(self)\n", "self.check_no_back_links_exist()\n", "self.run_method('on_update_after_submit')\n", "self.save_version()\n", "self.run_method('on_change')\n", "if (self.doctype, self.name) in frappe.flags.currently_saving:\n", "frappe.flags.currently_saving.remove((self.doctype, self.name))\n", "self.latest = None\n" ]
[ "def run_post_save_methods(self):...\n", "\"\"\"docstring\"\"\"\n", "doc_before_save = self.get_doc_before_save()\n", "if self._action == 'save':\n", "self.run_method('on_update')\n", "if self._action == 'submit':\n", "self.clear_cache()\n", "self.run_method('on_update')\n", "if self._action == 'cancel':\n", "self.notify_update()\n", "self.run_method('on_submit')\n", "self.run_method('on_cancel')\n", "if self._action == 'update_after_submit':\n", "update_global_search(self)\n", "self.check_no_back_links_exist()\n", "self.run_method('on_update_after_submit')\n", "self.save_version()\n", "self.run_method('on_change')\n", "if (self.doctype, self.name) in frappe.flags.currently_saving:\n", "frappe.flags.currently_saving.remove((self.doctype, self.name))\n", "self.latest = None\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Condition", "Expr'", "Condition", "Expr'", "Expr'", "Condition", "Expr'", "Expr'", "Expr'", "Condition", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Condition", "Expr'", "Assign'" ]
[ "@FUNC_0...\n", "return DeactivateAccountHandler(self)\n" ]
[ "@cache_in_self...\n", "return DeactivateAccountHandler(self)\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "def FUNC_12(VAR_14: Union[Text, Path], VAR_4: Any) ->None:...\n", "\"\"\"docstring\"\"\"\n", "import jsonpickle.ext.numpy as jsonpickle_numpy\n", "import jsonpickle\n", "jsonpickle_numpy.register_handlers()\n", "rasa.shared.utils.io.write_text_file(jsonpickle.dumps(VAR_4), VAR_14)\n" ]
[ "def json_pickle(file_name: Union[Text, Path], obj: Any) ->None:...\n", "\"\"\"docstring\"\"\"\n", "import jsonpickle.ext.numpy as jsonpickle_numpy\n", "import jsonpickle\n", "jsonpickle_numpy.register_handlers()\n", "rasa.shared.utils.io.write_text_file(jsonpickle.dumps(obj), file_name)\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Import'", "Import'", "Expr'", "Expr'" ]
[ "def FUNC_76(self):...\n", "VAR_90 = self.token['align']\n", "VAR_91 = len(VAR_90)\n", "VAR_92 = self.renderer.placeholder()\n", "VAR_69 = self.renderer.placeholder()\n", "for VAR_101, value in enumerate(self.token['header']):\n", "VAR_81 = VAR_90[VAR_101] if VAR_101 < VAR_91 else None\n", "VAR_69 += self.renderer.table_row(VAR_92)\n", "VAR_71 = {'header': True, 'align': VAR_81}\n", "VAR_67 = self.renderer.placeholder()\n", "VAR_92 += self.renderer.table_cell(self.inline(value), **flags)\n", "for VAR_101, row in enumerate(self.token['cells']):\n", "VAR_92 = self.renderer.placeholder()\n", "return self.renderer.table(VAR_69, VAR_67)\n", "for j, value in enumerate(row):\n", "VAR_81 = VAR_90[j] if j < VAR_91 else None\n", "VAR_67 += self.renderer.table_row(VAR_92)\n", "VAR_71 = {'header': False, 'align': VAR_81}\n", "VAR_92 += self.renderer.table_cell(self.inline(value), **flags)\n" ]
[ "def output_table(self):...\n", "aligns = self.token['align']\n", "aligns_length = len(aligns)\n", "cell = self.renderer.placeholder()\n", "header = self.renderer.placeholder()\n", "for i, value in enumerate(self.token['header']):\n", "align = aligns[i] if i < aligns_length else None\n", "header += self.renderer.table_row(cell)\n", "flags = {'header': True, 'align': align}\n", "body = self.renderer.placeholder()\n", "cell += self.renderer.table_cell(self.inline(value), **flags)\n", "for i, row in enumerate(self.token['cells']):\n", "cell = self.renderer.placeholder()\n", "return self.renderer.table(header, body)\n", "for j, value in enumerate(row):\n", "align = aligns[j] if j < aligns_length else None\n", "body += self.renderer.table_row(cell)\n", "flags = {'header': False, 'align': align}\n", "cell += self.renderer.table_cell(self.inline(value), **flags)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "For", "Assign'", "AugAssign'", "Assign'", "Assign'", "AugAssign'", "For", "Assign'", "Return'", "For", "Assign'", "AugAssign'", "Assign'", "AugAssign'" ]
[ "def FUNC_2(self, VAR_1):...\n", "from openapi_python_client.parser.properties import FileProperty\n", "VAR_5 = VAR_1.MagicMock()\n", "VAR_1.patch('openapi_python_client.utils.snake_case')\n", "VAR_9 = 'blah'\n", "VAR_4 = VAR_38(VAR_5=name, VAR_26=True, default=None)\n", "assert VAR_4.get_imports(VAR_9=prefix) == {f'from {VAR_9}.types import File',\n 'from dataclasses import astuple'}\n", "VAR_4.required = False\n", "assert VAR_4.get_imports(VAR_9=prefix) == {'from typing import Optional',\n f'from {VAR_9}.types import File', 'from dataclasses import astuple'}\n" ]
[ "def test_get_imports(self, mocker):...\n", "from openapi_python_client.parser.properties import FileProperty\n", "name = mocker.MagicMock()\n", "mocker.patch('openapi_python_client.utils.snake_case')\n", "prefix = 'blah'\n", "p = FileProperty(name=name, required=True, default=None)\n", "assert p.get_imports(prefix=prefix) == {f'from {prefix}.types import File',\n 'from dataclasses import astuple'}\n", "p.required = False\n", "assert p.get_imports(prefix=prefix) == {'from typing import Optional',\n f'from {prefix}.types import File', 'from dataclasses import astuple'}\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "ImportFrom'", "Assign'", "Expr'", "Assign'", "Assign'", "Assert'", "Assign'", "Assert'" ]
[ "@require_POST...\n", "VAR_3 = get_object_or_404(Topic.objects.for_access(VAR_0.user), VAR_2=topic_id)\n", "VAR_4 = NotificationCreationForm(user=request.user, VAR_3=topic, data=\n request.POST)\n", "if VAR_4.is_valid():\n", "VAR_4.save()\n", "messages.error(VAR_0, utils.render_form_errors(VAR_4))\n", "return redirect(VAR_0.POST.get('next', VAR_3.get_absolute_url()))\n" ]
[ "@require_POST...\n", "topic = get_object_or_404(Topic.objects.for_access(request.user), pk=topic_id)\n", "form = NotificationCreationForm(user=request.user, topic=topic, data=\n request.POST)\n", "if form.is_valid():\n", "form.save()\n", "messages.error(request, utils.render_form_errors(form))\n", "return redirect(request.POST.get('next', topic.get_absolute_url()))\n" ]
[ 0, 0, 0, 0, 0, 0, 4 ]
[ "Condition", "Assign'", "Assign'", "Condition", "Expr'", "Expr'", "Return'" ]
[ "def FUNC_4(self):...\n", "VAR_4 = self._makeEngine()\n", "for p in ('x', 'x/y', 'x/y/z'):\n", "VAR_4.compile(p)\n", "VAR_4.compile('path:a|b|c/d/e')\n", "VAR_4.compile('string:Fred')\n", "VAR_4.compile('string:A$B')\n", "VAR_4.compile('string:a ${x/y} b ${y/z} c')\n", "VAR_4.compile('python: 2 + 2')\n", "VAR_4.compile(\"\"\"python: 2 \n+\n 2\n\"\"\")\n" ]
[ "def test_compile(self):...\n", "e = self._makeEngine()\n", "for p in ('x', 'x/y', 'x/y/z'):\n", "e.compile(p)\n", "e.compile('path:a|b|c/d/e')\n", "e.compile('string:Fred')\n", "e.compile('string:A$B')\n", "e.compile('string:a ${x/y} b ${y/z} c')\n", "e.compile('python: 2 + 2')\n", "e.compile(\"\"\"python: 2 \n+\n 2\n\"\"\")\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "For", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'" ]
[ "def FUNC_17(VAR_5):...\n", "VAR_19 = filescheme.dirbrowser_html(VAR_5).decode('utf-8')\n", "VAR_20 = bs4.BeautifulSoup(VAR_19, 'html.parser')\n", "print(VAR_20.prettify())\n", "VAR_21 = VAR_20('div', id='dirbrowserContainer')[0]\n", "VAR_42 = VAR_21('ul', class_='parent')\n", "if not VAR_42:\n", "VAR_11 = None\n", "VAR_11 = VAR_42[0].li.a.string\n", "VAR_15 = []\n", "VAR_16 = []\n", "for li in VAR_21('ul', class_='folders')[0]('li'):\n", "VAR_8 = self.Item(VAR_17=li.a['href'], VAR_18=str(li.a.string))\n", "for li in VAR_21('ul', class_='files')[0]('li'):\n", "VAR_15.append(VAR_8)\n", "VAR_8 = self.Item(VAR_17=li.a['href'], VAR_18=str(li.a.string))\n", "return self.Parsed(VAR_11=parent, VAR_15=folders, VAR_16=files)\n", "VAR_16.append(VAR_8)\n" ]
[ "def parse(path):...\n", "html = filescheme.dirbrowser_html(path).decode('utf-8')\n", "soup = bs4.BeautifulSoup(html, 'html.parser')\n", "print(soup.prettify())\n", "container = soup('div', id='dirbrowserContainer')[0]\n", "parent_elem = container('ul', class_='parent')\n", "if not parent_elem:\n", "parent = None\n", "parent = parent_elem[0].li.a.string\n", "folders = []\n", "files = []\n", "for li in container('ul', class_='folders')[0]('li'):\n", "item = self.Item(link=li.a['href'], text=str(li.a.string))\n", "for li in container('ul', class_='files')[0]('li'):\n", "folders.append(item)\n", "item = self.Item(link=li.a['href'], text=str(li.a.string))\n", "return self.Parsed(parent=parent, folders=folders, files=files)\n", "files.append(item)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "For", "Assign'", "For", "Expr'", "Assign'", "Return'", "Expr'" ]
[ "def FUNC_5(VAR_22: Text, VAR_23: Text, VAR_24: Optional[Fingerprint]=None...\n", "\"\"\"docstring\"\"\"\n", "import tarfile\n", "if VAR_24:\n", "FUNC_11(VAR_22, VAR_24)\n", "VAR_43 = os.path.dirname(VAR_23)\n", "if not os.path.exists(VAR_43):\n", "os.makedirs(VAR_43)\n", "for elem in os.scandir(VAR_22):\n", "tar.add(elem.path, arcname=elem.name)\n", "shutil.rmtree(VAR_22)\n", "return VAR_23\n" ]
[ "def create_package_rasa(training_directory: Text, output_filename: Text,...\n", "\"\"\"docstring\"\"\"\n", "import tarfile\n", "if fingerprint:\n", "persist_fingerprint(training_directory, fingerprint)\n", "output_directory = os.path.dirname(output_filename)\n", "if not os.path.exists(output_directory):\n", "os.makedirs(output_directory)\n", "for elem in os.scandir(training_directory):\n", "tar.add(elem.path, arcname=elem.name)\n", "shutil.rmtree(training_directory)\n", "return output_filename\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Import'", "Condition", "Expr'", "Assign'", "Condition", "Expr'", "For", "Expr'", "Expr'", "Return'" ]
[ "@expose('/login/', methods=['GET', 'POST'])...\n", "if g.user is not None and g.user.is_authenticated:\n", "return redirect(self.appbuilder.get_url_for_index)\n", "VAR_10 = LoginForm_db()\n", "if VAR_10.validate_on_submit():\n", "VAR_50 = self.appbuilder.sm.auth_user_db(VAR_10.username.data, VAR_10.\n password.data)\n", "return self.render_template(self.login_template, VAR_37=self.title, VAR_10=\n form, appbuilder=self.appbuilder)\n", "if not VAR_50:\n", "flash(as_unicode(self.invalid_login_message), 'warning')\n", "login_user(VAR_50, remember=False)\n", "return redirect(self.appbuilder.get_url_for_login)\n", "return redirect(self.appbuilder.get_url_for_index)\n" ]
[ "@expose('/login/', methods=['GET', 'POST'])...\n", "if g.user is not None and g.user.is_authenticated:\n", "return redirect(self.appbuilder.get_url_for_index)\n", "form = LoginForm_db()\n", "if form.validate_on_submit():\n", "user = self.appbuilder.sm.auth_user_db(form.username.data, form.password.data)\n", "return self.render_template(self.login_template, title=self.title, form=\n form, appbuilder=self.appbuilder)\n", "if not user:\n", "flash(as_unicode(self.invalid_login_message), 'warning')\n", "login_user(user, remember=False)\n", "return redirect(self.appbuilder.get_url_for_login)\n", "return redirect(self.appbuilder.get_url_for_index)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Condition", "Return'", "Assign'", "Condition", "Assign'", "Return'", "Condition", "Expr'", "Expr'", "Return'", "Return'" ]
[ "@VAR_1.after_request...\n", "\"\"\"docstring\"\"\"\n", "if VAR_1.config.get('INCOGNITO') and VAR_13.content_type.startswith('text/html'\n", "VAR_40 = (request.endpoint == 'report' and request.view_args is not None and\n request.view_args['report_name'] == 'editor')\n", "return VAR_13\n", "if not VAR_40:\n", "VAR_44 = VAR_13.get_data(as_text=True)\n", "VAR_13.set_data(replace_numbers(VAR_44))\n" ]
[ "@app.after_request...\n", "\"\"\"docstring\"\"\"\n", "if app.config.get('INCOGNITO') and response.content_type.startswith('text/html'\n", "is_editor = (request.endpoint == 'report' and request.view_args is not None and\n request.view_args['report_name'] == 'editor')\n", "return response\n", "if not is_editor:\n", "original_text = response.get_data(as_text=True)\n", "response.set_data(replace_numbers(original_text))\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Condition", "Assign'", "Return'", "Condition", "Assign'", "Expr'" ]
[ "def FUNC_0(VAR_4, VAR_5=None):...\n", "VAR_11 = VAR_263.request.args\n", "if VAR_4 < 0 and len(VAR_11) >= -VAR_4:\n", "return VAR_11[VAR_4]\n", "if VAR_4 >= 0 and len(VAR_11) > VAR_4:\n", "return VAR_11[VAR_4]\n", "return VAR_5\n" ]
[ "def getarg(position, default=None):...\n", "args = current.request.args\n", "if position < 0 and len(args) >= -position:\n", "return args[position]\n", "if position >= 0 and len(args) > position:\n", "return args[position]\n", "return default\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Condition", "Return'", "Condition", "Return'", "Return'" ]
[ "def FUNC_7(self):...\n", "VAR_9 = test.test_src_dir_path(VAR_0)\n", "self.parser = saved_model_cli.create_parser()\n", "VAR_10 = self.parser.parse_args(['show', '--dir', VAR_9, '--tag_set',\n 'badtagset'])\n", "saved_model_cli.show(VAR_10)\n" ]
[ "def testShowCommandErrorNoTagSet(self):...\n", "base_path = test.test_src_dir_path(SAVED_MODEL_PATH)\n", "self.parser = saved_model_cli.create_parser()\n", "args = self.parser.parse_args(['show', '--dir', base_path, '--tag_set',\n 'badtagset'])\n", "saved_model_cli.show(args)\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Expr'" ]
[ "def FUNC_68():...\n", "\"\"\"docstring\"\"\"\n", "VAR_192 = FUNC_10()\n", "if VAR_13.db_name:\n", "VAR_1.app_modules = VAR_192.get_value('app_modules')\n", "if not (VAR_1.app_modules and VAR_1.module_app):\n", "VAR_1.module_app = VAR_192.get_value('module_app')\n", "VAR_1.module_app, VAR_1.app_modules = {}, {}\n", "for VAR_219 in FUNC_63(True):\n", "if VAR_219 == 'webnotes':\n", "if VAR_13.db_name:\n", "VAR_219 = 'frappe'\n", "VAR_1.app_modules.setdefault(VAR_219, [])\n", "VAR_192.set_value('app_modules', VAR_1.app_modules)\n", "for VAR_91 in FUNC_62(VAR_219):\n", "VAR_192.set_value('module_app', VAR_1.module_app)\n", "VAR_91 = FUNC_56(VAR_91)\n", "VAR_1.module_app[VAR_91] = VAR_219\n", "VAR_1.app_modules[VAR_219].append(VAR_91)\n" ]
[ "def setup_module_map():...\n", "\"\"\"docstring\"\"\"\n", "_cache = cache()\n", "if conf.db_name:\n", "local.app_modules = _cache.get_value('app_modules')\n", "if not (local.app_modules and local.module_app):\n", "local.module_app = _cache.get_value('module_app')\n", "local.module_app, local.app_modules = {}, {}\n", "for app in get_all_apps(True):\n", "if app == 'webnotes':\n", "if conf.db_name:\n", "app = 'frappe'\n", "local.app_modules.setdefault(app, [])\n", "_cache.set_value('app_modules', local.app_modules)\n", "for module in get_module_list(app):\n", "_cache.set_value('module_app', local.module_app)\n", "module = scrub(module)\n", "local.module_app[module] = app\n", "local.app_modules[app].append(module)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Assign'", "For", "Condition", "Condition", "Assign'", "Expr'", "Expr'", "For", "Expr'", "Assign'", "Assign'", "Expr'" ]
[ "async def FUNC_28(self, VAR_5: str, VAR_11: str, VAR_3: List[EventBase],...\n", "\"\"\"docstring\"\"\"\n", "VAR_104 = {}\n", "for VAR_170 in itertools.chain(VAR_3, VAR_2):\n", "VAR_170.internal_metadata.outlier = True\n", "VAR_60 = {VAR_170.event_id: VAR_170 for VAR_170 in itertools.chain(VAR_3,\n VAR_2, [VAR_1])}\n", "VAR_169 = await self.state_handler.compute_event_context(VAR_170)\n", "VAR_105 = None\n", "VAR_104[VAR_170.event_id] = VAR_169\n", "for VAR_170 in VAR_3:\n", "if (VAR_170.type, VAR_170.state_key) == (VAR_188.Create, ''):\n", "if VAR_105 is None:\n", "VAR_105 = VAR_170\n", "VAR_106 = VAR_105.content.get('room_version', RoomVersions.V1.identifier)\n", "if VAR_28.identifier != VAR_106:\n", "VAR_107 = set()\n", "for VAR_170 in itertools.chain(VAR_3, VAR_2, [VAR_1]):\n", "for VAR_160 in VAR_170.auth_event_ids():\n", "for VAR_160 in VAR_107:\n", "if VAR_160 not in VAR_60:\n", "VAR_171 = await self.federation_client.get_pdu([VAR_5], VAR_160, VAR_28=\n room_version, outlier=True, timeout=10000)\n", "for VAR_170 in itertools.chain(VAR_3, VAR_2, [VAR_1]):\n", "VAR_107.add(VAR_160)\n", "if VAR_171 and VAR_171.event_id == VAR_160:\n", "VAR_172 = {(VAR_60[VAR_160].type, VAR_60[VAR_160].state_key): VAR_60[\n VAR_160] for VAR_160 in VAR_170.auth_event_ids() if VAR_160 in VAR_60}\n", "await self.persist_events_and_notify(VAR_11, [(VAR_170, VAR_104[VAR_170.\n event_id]) for VAR_170 in itertools.chain(VAR_3, VAR_2)])\n", "VAR_60[VAR_160] = VAR_171\n", "VAR_0.info('Failed to find auth event %r', VAR_160)\n", "if VAR_105:\n", "VAR_108 = await self.state_handler.compute_event_context(VAR_1, old_state=state\n )\n", "VAR_172[VAR_188.Create, ''] = VAR_105\n", "event_auth.check(VAR_28, VAR_170, VAR_3=auth_for_e)\n", "VAR_0.warning('Rejecting %s because %s', VAR_170.event_id, err.msg)\n", "return await self.persist_events_and_notify(VAR_11, [(VAR_1, VAR_108)])\n", "if VAR_170 == VAR_1:\n", "VAR_104[VAR_170.event_id].rejected = RejectedReason.AUTH_ERROR\n" ]
[ "async def _persist_auth_tree(self, origin: str, room_id: str, auth_events:...\n", "\"\"\"docstring\"\"\"\n", "events_to_context = {}\n", "for e in itertools.chain(auth_events, state):\n", "e.internal_metadata.outlier = True\n", "event_map = {e.event_id: e for e in itertools.chain(auth_events, state, [\n event])}\n", "ctx = await self.state_handler.compute_event_context(e)\n", "create_event = None\n", "events_to_context[e.event_id] = ctx\n", "for e in auth_events:\n", "if (e.type, e.state_key) == (EventTypes.Create, ''):\n", "if create_event is None:\n", "create_event = e\n", "room_version_id = create_event.content.get('room_version', RoomVersions.V1.\n identifier)\n", "if room_version.identifier != room_version_id:\n", "missing_auth_events = set()\n", "for e in itertools.chain(auth_events, state, [event]):\n", "for e_id in e.auth_event_ids():\n", "for e_id in missing_auth_events:\n", "if e_id not in event_map:\n", "m_ev = await self.federation_client.get_pdu([origin], e_id, room_version=\n room_version, outlier=True, timeout=10000)\n", "for e in itertools.chain(auth_events, state, [event]):\n", "missing_auth_events.add(e_id)\n", "if m_ev and m_ev.event_id == e_id:\n", "auth_for_e = {(event_map[e_id].type, event_map[e_id].state_key): event_map[\n e_id] for e_id in e.auth_event_ids() if e_id in event_map}\n", "await self.persist_events_and_notify(room_id, [(e, events_to_context[e.\n event_id]) for e in itertools.chain(auth_events, state)])\n", "event_map[e_id] = m_ev\n", "logger.info('Failed to find auth event %r', e_id)\n", "if create_event:\n", "new_event_context = await self.state_handler.compute_event_context(event,\n old_state=state)\n", "auth_for_e[EventTypes.Create, ''] = create_event\n", "event_auth.check(room_version, e, auth_events=auth_for_e)\n", "logger.warning('Rejecting %s because %s', e.event_id, err.msg)\n", "return await self.persist_events_and_notify(room_id, [(event,\n new_event_context)])\n", "if e == event:\n", "events_to_context[e.event_id].rejected = RejectedReason.AUTH_ERROR\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "For", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "For", "Condition", "Condition", "Assign'", "Assign'", "Condition", "Assign'", "For", "For", "For", "Condition", "Assign'", "For", "Expr'", "Condition", "Assign'", "Expr'", "Assign'", "Expr'", "Condition", "Assign'", "Assign'", "Expr'", "Expr'", "Return'", "Condition", "Assign'" ]
[ "def FUNC_4(self) ->Sequence[GeneratorError]:...\n", "\"\"\"docstring\"\"\"\n", "print(f'Generating {self.project_name}')\n", "self.project_dir.mkdir()\n", "return [GeneratorError(detail=\n 'Directory already exists. Delete it or use the update command.')]\n", "self._create_package()\n", "self._build_metadata()\n", "self._build_models()\n", "self._build_api()\n", "self._reformat()\n", "return self._get_errors()\n" ]
[ "def build(self) ->Sequence[GeneratorError]:...\n", "\"\"\"docstring\"\"\"\n", "print(f'Generating {self.project_name}')\n", "self.project_dir.mkdir()\n", "return [GeneratorError(detail=\n 'Directory already exists. Delete it or use the update command.')]\n", "self._create_package()\n", "self._build_metadata()\n", "self._build_models()\n", "self._build_api()\n", "self._reformat()\n", "return self._get_errors()\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Expr'", "Expr'", "Return'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Return'" ]
[ "def FUNC_29(self):...\n", "VAR_5 = self._makeContext()\n", "self.assertIs(VAR_5.evaluate('nocall:open'), open)\n" ]
[ "def test_open_in_path_expr(self):...\n", "ec = self._makeContext()\n", "self.assertIs(ec.evaluate('nocall:open'), open)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Expr'" ]
[ "def FUNC_7(self, VAR_13):...\n", "if VAR_13 is None:\n", "if not isinstance(self.inputSource, FileInputSource):\n", "return VAR_13\n", "VAR_13 = '-'\n", "if self.inputSource.sourceName.endswith('.bs'):\n", "VAR_13 = self.inputSource.sourceName[0:-3] + '.html'\n", "if self.inputSource.sourceName.endswith('.src.html'):\n", "VAR_13 = self.inputSource.sourceName[0:-9] + '.html'\n", "VAR_13 = '-'\n" ]
[ "def fixMissingOutputFilename(self, outputFilename):...\n", "if outputFilename is None:\n", "if not isinstance(self.inputSource, FileInputSource):\n", "return outputFilename\n", "outputFilename = '-'\n", "if self.inputSource.sourceName.endswith('.bs'):\n", "outputFilename = self.inputSource.sourceName[0:-3] + '.html'\n", "if self.inputSource.sourceName.endswith('.src.html'):\n", "outputFilename = self.inputSource.sourceName[0:-9] + '.html'\n", "outputFilename = '-'\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Condition", "Return'", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Assign'" ]
[ "@CLASS_4('gpl')...\n", "\"\"\"docstring\"\"\"\n", "return 'text/html', utils.read_file('html/license.html')\n" ]
[ "@add_handler('gpl')...\n", "\"\"\"docstring\"\"\"\n", "return 'text/html', utils.read_file('html/license.html')\n" ]
[ 0, 0, 0 ]
[ "Condition", "Docstring", "Return'" ]
[ "def FUNC_3(self, VAR_23):...\n", "assert isinstance(VAR_23, text_type)\n", "if self.encoding:\n", "return VAR_23.encode(self.encoding, 'strict')\n", "return VAR_23\n" ]
[ "def encodeStrict(self, string):...\n", "assert isinstance(string, text_type)\n", "if self.encoding:\n", "return string.encode(self.encoding, 'strict')\n", "return string\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assert'", "Condition", "Return'", "Return'" ]
[ "@VAR_0.route('/api/templates', methods=['GET'])...\n", "\"\"\"docstring\"\"\"\n", "VAR_73 = {'application': request.args.get('application'), 'backend':\n request.args.get('backend'), 'recent': request.args.get('recent'),\n 'length': request.args.get('length'), 'offset': request.args.get('offset')}\n", "VAR_88 = FUNC_58('/internal/templates', 'get', VAR_73=params)\n", "return jsonify({'success': False, 'message': str(err)}), 400\n", "return jsonify(VAR_88)\n" ]
[ "@gui.route('/api/templates', methods=['GET'])...\n", "\"\"\"docstring\"\"\"\n", "params = {'application': request.args.get('application'), 'backend':\n request.args.get('backend'), 'recent': request.args.get('recent'),\n 'length': request.args.get('length'), 'offset': request.args.get('offset')}\n", "templates_info = query_internal_api('/internal/templates', 'get', params=params\n )\n", "return jsonify({'success': False, 'message': str(err)}), 400\n", "return jsonify(templates_info)\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Assign'", "Return'", "Return'" ]