|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
from antlr4 import CommonTokenStream, DFA, PredictionContextCache, Lexer, LexerATNSimulator, ParserRuleContext, TerminalNode |
|
from antlr4.InputStream import InputStream |
|
from antlr4.Parser import Parser |
|
from antlr4.RuleContext import RuleContext |
|
from antlr4.Token import Token |
|
from antlr4.atn.ATNDeserializer import ATNDeserializer |
|
from antlr4.error.ErrorListener import ErrorListener |
|
from antlr4.error.Errors import LexerNoViableAltException |
|
from antlr4.tree.Tree import ParseTree |
|
from antlr4.tree.Trees import Trees |
|
from io import StringIO |
|
|
|
|
|
def serializedATN(): |
|
with StringIO() as buf: |
|
buf.write("\3\u0430\ud6d1\u8206\uad2d\u4417\uaef1\u8d80\uaadd\2\n") |
|
buf.write("\64\b\1\4\2\t\2\4\3\t\3\4\4\t\4\4\5\t\5\4\6\t\6\4\7\t") |
|
buf.write("\7\4\b\t\b\4\t\t\t\3\2\3\2\3\2\3\3\3\3\3\4\3\4\3\5\3\5") |
|
buf.write("\3\6\3\6\7\6\37\n\6\f\6\16\6\"\13\6\3\6\3\6\3\7\3\7\5") |
|
buf.write("\7(\n\7\3\b\3\b\3\t\3\t\7\t.\n\t\f\t\16\t\61\13\t\3\t") |
|
buf.write("\3\t\3/\2\n\3\5\5\6\7\7\t\b\13\t\r\2\17\2\21\n\3\2\4\7") |
|
buf.write("\2\62;aa\u00b9\u00b9\u0302\u0371\u2041\u2042\17\2C\\c") |
|
buf.write("|\u00c2\u00d8\u00da\u00f8\u00fa\u0301\u0372\u037f\u0381") |
|
buf.write("\u2001\u200e\u200f\u2072\u2191\u2c02\u2ff1\u3003\ud801") |
|
buf.write("\uf902\ufdd1\ufdf2\uffff\64\2\3\3\2\2\2\2\5\3\2\2\2\2") |
|
buf.write("\7\3\2\2\2\2\t\3\2\2\2\2\13\3\2\2\2\2\21\3\2\2\2\3\23") |
|
buf.write("\3\2\2\2\5\26\3\2\2\2\7\30\3\2\2\2\t\32\3\2\2\2\13\34") |
|
buf.write("\3\2\2\2\r\'\3\2\2\2\17)\3\2\2\2\21+\3\2\2\2\23\24\7\61") |
|
buf.write("\2\2\24\25\7\61\2\2\25\4\3\2\2\2\26\27\7\61\2\2\27\6\3") |
|
buf.write("\2\2\2\30\31\7,\2\2\31\b\3\2\2\2\32\33\7#\2\2\33\n\3\2") |
|
buf.write("\2\2\34 \5\17\b\2\35\37\5\r\7\2\36\35\3\2\2\2\37\"\3\2") |
|
buf.write("\2\2 \36\3\2\2\2 !\3\2\2\2!#\3\2\2\2\" \3\2\2\2#$\b\6") |
|
buf.write("\2\2$\f\3\2\2\2%(\5\17\b\2&(\t\2\2\2\'%\3\2\2\2\'&\3\2") |
|
buf.write("\2\2(\16\3\2\2\2)*\t\3\2\2*\20\3\2\2\2+/\7)\2\2,.\13\2") |
|
buf.write("\2\2-,\3\2\2\2.\61\3\2\2\2/\60\3\2\2\2/-\3\2\2\2\60\62") |
|
buf.write("\3\2\2\2\61/\3\2\2\2\62\63\7)\2\2\63\22\3\2\2\2\6\2 \'") |
|
buf.write("/\3\3\6\2") |
|
return buf.getvalue() |
|
|
|
|
|
class XPathLexer(Lexer): |
|
|
|
atn = ATNDeserializer().deserialize(serializedATN()) |
|
|
|
decisionsToDFA = [ DFA(ds, i) for i, ds in enumerate(atn.decisionToState) ] |
|
|
|
|
|
TOKEN_REF = 1 |
|
RULE_REF = 2 |
|
ANYWHERE = 3 |
|
ROOT = 4 |
|
WILDCARD = 5 |
|
BANG = 6 |
|
ID = 7 |
|
STRING = 8 |
|
|
|
modeNames = [ "DEFAULT_MODE" ] |
|
|
|
literalNames = [ "<INVALID>", |
|
"'//'", "'/'", "'*'", "'!'" ] |
|
|
|
symbolicNames = [ "<INVALID>", |
|
"TOKEN_REF", "RULE_REF", "ANYWHERE", "ROOT", "WILDCARD", "BANG", |
|
"ID", "STRING" ] |
|
|
|
ruleNames = [ "ANYWHERE", "ROOT", "WILDCARD", "BANG", "ID", "NameChar", |
|
"NameStartChar", "STRING" ] |
|
|
|
grammarFileName = "XPathLexer.g4" |
|
|
|
def __init__(self, input=None): |
|
super().__init__(input) |
|
self.checkVersion("4.9.1") |
|
self._interp = LexerATNSimulator(self, self.atn, self.decisionsToDFA, PredictionContextCache()) |
|
self._actions = None |
|
self._predicates = None |
|
|
|
|
|
def action(self, localctx:RuleContext, ruleIndex:int, actionIndex:int): |
|
if self._actions is None: |
|
actions = dict() |
|
actions[4] = self.ID_action |
|
self._actions = actions |
|
_action = self._actions.get(ruleIndex, None) |
|
if _action is not None: |
|
_action(localctx, actionIndex) |
|
else: |
|
raise Exception("No registered action for: %d" % ruleIndex) |
|
|
|
def ID_action(self, localctx:RuleContext , actionIndex:int): |
|
if actionIndex == 0: |
|
char = self.text[0] |
|
if char.isupper(): |
|
self.type = XPathLexer.TOKEN_REF |
|
else: |
|
self.type = XPathLexer.RULE_REF |
|
|
|
class XPath(object): |
|
|
|
WILDCARD = "*" |
|
NOT = "!" |
|
|
|
def __init__(self, parser:Parser, path:str): |
|
self.parser = parser |
|
self.path = path |
|
self.elements = self.split(path) |
|
|
|
def split(self, path:str): |
|
input = InputStream(path) |
|
lexer = XPathLexer(input) |
|
def recover(self, e): |
|
raise e |
|
lexer.recover = recover |
|
lexer.removeErrorListeners() |
|
lexer.addErrorListener(ErrorListener()) |
|
tokenStream = CommonTokenStream(lexer) |
|
try: |
|
tokenStream.fill() |
|
except LexerNoViableAltException as e: |
|
pos = lexer.column |
|
msg = "Invalid tokens or characters at index %d in path '%s'" % (pos, path) |
|
raise Exception(msg, e) |
|
|
|
tokens = iter(tokenStream.tokens) |
|
elements = list() |
|
for el in tokens: |
|
invert = False |
|
anywhere = False |
|
|
|
if el.type in [XPathLexer.ROOT, XPathLexer.ANYWHERE]: |
|
anywhere = el.type == XPathLexer.ANYWHERE |
|
next_el = next(tokens, None) |
|
if not next_el: |
|
raise Exception('Missing element after %s' % el.getText()) |
|
else: |
|
el = next_el |
|
|
|
if el.type == XPathLexer.BANG: |
|
invert = True |
|
next_el = next(tokens, None) |
|
if not next_el: |
|
raise Exception('Missing element after %s' % el.getText()) |
|
else: |
|
el = next_el |
|
|
|
if el.type in [XPathLexer.TOKEN_REF, XPathLexer.RULE_REF, XPathLexer.WILDCARD, XPathLexer.STRING]: |
|
element = self.getXPathElement(el, anywhere) |
|
element.invert = invert |
|
elements.append(element) |
|
elif el.type==Token.EOF: |
|
break |
|
else: |
|
raise Exception("Unknown path element %s" % lexer.symbolicNames[el.type]) |
|
return elements |
|
|
|
|
|
|
|
|
|
|
|
|
|
def getXPathElement(self, wordToken:Token, anywhere:bool): |
|
if wordToken.type==Token.EOF: |
|
raise Exception("Missing path element at end of path") |
|
|
|
word = wordToken.text |
|
if wordToken.type==XPathLexer.WILDCARD : |
|
return XPathWildcardAnywhereElement() if anywhere else XPathWildcardElement() |
|
|
|
elif wordToken.type in [XPathLexer.TOKEN_REF, XPathLexer.STRING]: |
|
tsource = self.parser.getTokenStream().tokenSource |
|
|
|
ttype = Token.INVALID_TYPE |
|
if wordToken.type == XPathLexer.TOKEN_REF: |
|
if word in tsource.ruleNames: |
|
ttype = tsource.ruleNames.index(word) + 1 |
|
else: |
|
if word in tsource.literalNames: |
|
ttype = tsource.literalNames.index(word) |
|
|
|
if ttype == Token.INVALID_TYPE: |
|
raise Exception("%s at index %d isn't a valid token name" % (word, wordToken.tokenIndex)) |
|
return XPathTokenAnywhereElement(word, ttype) if anywhere else XPathTokenElement(word, ttype) |
|
|
|
else: |
|
ruleIndex = self.parser.ruleNames.index(word) if word in self.parser.ruleNames else -1 |
|
|
|
if ruleIndex == -1: |
|
raise Exception("%s at index %d isn't a valid rule name" % (word, wordToken.tokenIndex)) |
|
return XPathRuleAnywhereElement(word, ruleIndex) if anywhere else XPathRuleElement(word, ruleIndex) |
|
|
|
|
|
@staticmethod |
|
def findAll(tree:ParseTree, xpath:str, parser:Parser): |
|
p = XPath(parser, xpath) |
|
return p.evaluate(tree) |
|
|
|
|
|
|
|
|
|
|
|
|
|
def evaluate(self, t:ParseTree): |
|
dummyRoot = ParserRuleContext() |
|
dummyRoot.children = [t] |
|
|
|
work = [dummyRoot] |
|
for element in self.elements: |
|
work_next = list() |
|
for node in work: |
|
if not isinstance(node, TerminalNode) and node.children: |
|
|
|
|
|
|
|
matching = element.evaluate(node) |
|
|
|
|
|
|
|
matching = filter(lambda m: m not in work_next, matching) |
|
|
|
work_next.extend(matching) |
|
work = work_next |
|
|
|
return work |
|
|
|
|
|
class XPathElement(object): |
|
|
|
def __init__(self, nodeName:str): |
|
self.nodeName = nodeName |
|
self.invert = False |
|
|
|
def __str__(self): |
|
return type(self).__name__ + "[" + ("!" if self.invert else "") + self.nodeName + "]" |
|
|
|
|
|
|
|
|
|
|
|
|
|
class XPathRuleAnywhereElement(XPathElement): |
|
|
|
def __init__(self, ruleName:str, ruleIndex:int): |
|
super().__init__(ruleName) |
|
self.ruleIndex = ruleIndex |
|
|
|
def evaluate(self, t:ParseTree): |
|
|
|
return filter(lambda c: isinstance(c, ParserRuleContext) and (self.invert ^ (c.getRuleIndex() == self.ruleIndex)), Trees.descendants(t)) |
|
|
|
class XPathRuleElement(XPathElement): |
|
|
|
def __init__(self, ruleName:str, ruleIndex:int): |
|
super().__init__(ruleName) |
|
self.ruleIndex = ruleIndex |
|
|
|
def evaluate(self, t:ParseTree): |
|
|
|
return filter(lambda c: isinstance(c, ParserRuleContext) and (self.invert ^ (c.getRuleIndex() == self.ruleIndex)), Trees.getChildren(t)) |
|
|
|
class XPathTokenAnywhereElement(XPathElement): |
|
|
|
def __init__(self, ruleName:str, tokenType:int): |
|
super().__init__(ruleName) |
|
self.tokenType = tokenType |
|
|
|
def evaluate(self, t:ParseTree): |
|
|
|
return filter(lambda c: isinstance(c, TerminalNode) and (self.invert ^ (c.symbol.type == self.tokenType)), Trees.descendants(t)) |
|
|
|
class XPathTokenElement(XPathElement): |
|
|
|
def __init__(self, ruleName:str, tokenType:int): |
|
super().__init__(ruleName) |
|
self.tokenType = tokenType |
|
|
|
def evaluate(self, t:ParseTree): |
|
|
|
return filter(lambda c: isinstance(c, TerminalNode) and (self.invert ^ (c.symbol.type == self.tokenType)), Trees.getChildren(t)) |
|
|
|
|
|
class XPathWildcardAnywhereElement(XPathElement): |
|
|
|
def __init__(self): |
|
super().__init__(XPath.WILDCARD) |
|
|
|
def evaluate(self, t:ParseTree): |
|
if self.invert: |
|
return list() |
|
else: |
|
return Trees.descendants(t) |
|
|
|
|
|
class XPathWildcardElement(XPathElement): |
|
|
|
def __init__(self): |
|
super().__init__(XPath.WILDCARD) |
|
|
|
|
|
def evaluate(self, t:ParseTree): |
|
if self.invert: |
|
return list() |
|
else: |
|
return Trees.getChildren(t) |
|
|