| Index: third_party/google-endpoints/pyparsing.py
|
| diff --git a/third_party/google-endpoints/pyparsing.py b/third_party/google-endpoints/pyparsing.py
|
| new file mode 100644
|
| index 0000000000000000000000000000000000000000..a21224359e8899ba74d53f647a838aaf3ddf5dde
|
| --- /dev/null
|
| +++ b/third_party/google-endpoints/pyparsing.py
|
| @@ -0,0 +1,5696 @@
|
| +# module pyparsing.py
|
| +#
|
| +# Copyright (c) 2003-2016 Paul T. McGuire
|
| +#
|
| +# Permission is hereby granted, free of charge, to any person obtaining
|
| +# a copy of this software and associated documentation files (the
|
| +# "Software"), to deal in the Software without restriction, including
|
| +# without limitation the rights to use, copy, modify, merge, publish,
|
| +# distribute, sublicense, and/or sell copies of the Software, and to
|
| +# permit persons to whom the Software is furnished to do so, subject to
|
| +# the following conditions:
|
| +#
|
| +# The above copyright notice and this permission notice shall be
|
| +# included in all copies or substantial portions of the Software.
|
| +#
|
| +# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
|
| +# EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
|
| +# MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.
|
| +# IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY
|
| +# CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT,
|
| +# TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE
|
| +# SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
|
| +#
|
| +
|
| +__doc__ = \
|
| +"""
|
| +pyparsing module - Classes and methods to define and execute parsing grammars
|
| +
|
| +The pyparsing module is an alternative approach to creating and executing simple grammars,
|
| +vs. the traditional lex/yacc approach, or the use of regular expressions. With pyparsing, you
|
| +don't need to learn a new syntax for defining grammars or matching expressions - the parsing module
|
| +provides a library of classes that you use to construct the grammar directly in Python.
|
| +
|
| +Here is a program to parse "Hello, World!" (or any greeting of the form
|
| +C{"<salutation>, <addressee>!"}), built up using L{Word}, L{Literal}, and L{And} elements
|
| +(L{'+'<ParserElement.__add__>} operator gives L{And} expressions, strings are auto-converted to
|
| +L{Literal} expressions)::
|
| +
|
| + from pyparsing import Word, alphas
|
| +
|
| + # define grammar of a greeting
|
| + greet = Word(alphas) + "," + Word(alphas) + "!"
|
| +
|
| + hello = "Hello, World!"
|
| + print (hello, "->", greet.parseString(hello))
|
| +
|
| +The program outputs the following::
|
| +
|
| + Hello, World! -> ['Hello', ',', 'World', '!']
|
| +
|
| +The Python representation of the grammar is quite readable, owing to the self-explanatory
|
| +class names, and the use of '+', '|' and '^' operators.
|
| +
|
| +The L{ParseResults} object returned from L{ParserElement.parseString<ParserElement.parseString>} can be accessed as a nested list, a dictionary, or an
|
| +object with named attributes.
|
| +
|
| +The pyparsing module handles some of the problems that are typically vexing when writing text parsers:
|
| + - extra or missing whitespace (the above program will also handle "Hello,World!", "Hello , World !", etc.)
|
| + - quoted strings
|
| + - embedded comments
|
| +"""
|
| +
|
| +__version__ = "2.1.10"
|
| +__versionTime__ = "07 Oct 2016 01:31 UTC"
|
| +__author__ = "Paul McGuire <ptmcg@users.sourceforge.net>"
|
| +
|
| +import string
|
| +from weakref import ref as wkref
|
| +import copy
|
| +import sys
|
| +import warnings
|
| +import re
|
| +import sre_constants
|
| +import collections
|
| +import pprint
|
| +import traceback
|
| +import types
|
| +from datetime import datetime
|
| +
|
| +try:
|
| + from _thread import RLock
|
| +except ImportError:
|
| + from threading import RLock
|
| +
|
| +try:
|
| + from collections import OrderedDict as _OrderedDict
|
| +except ImportError:
|
| + try:
|
| + from ordereddict import OrderedDict as _OrderedDict
|
| + except ImportError:
|
| + _OrderedDict = None
|
| +
|
| +#~ sys.stderr.write( "testing pyparsing module, version %s, %s\n" % (__version__,__versionTime__ ) )
|
| +
|
| +__all__ = [
|
| +'And', 'CaselessKeyword', 'CaselessLiteral', 'CharsNotIn', 'Combine', 'Dict', 'Each', 'Empty',
|
| +'FollowedBy', 'Forward', 'GoToColumn', 'Group', 'Keyword', 'LineEnd', 'LineStart', 'Literal',
|
| +'MatchFirst', 'NoMatch', 'NotAny', 'OneOrMore', 'OnlyOnce', 'Optional', 'Or',
|
| +'ParseBaseException', 'ParseElementEnhance', 'ParseException', 'ParseExpression', 'ParseFatalException',
|
| +'ParseResults', 'ParseSyntaxException', 'ParserElement', 'QuotedString', 'RecursiveGrammarException',
|
| +'Regex', 'SkipTo', 'StringEnd', 'StringStart', 'Suppress', 'Token', 'TokenConverter',
|
| +'White', 'Word', 'WordEnd', 'WordStart', 'ZeroOrMore',
|
| +'alphanums', 'alphas', 'alphas8bit', 'anyCloseTag', 'anyOpenTag', 'cStyleComment', 'col',
|
| +'commaSeparatedList', 'commonHTMLEntity', 'countedArray', 'cppStyleComment', 'dblQuotedString',
|
| +'dblSlashComment', 'delimitedList', 'dictOf', 'downcaseTokens', 'empty', 'hexnums',
|
| +'htmlComment', 'javaStyleComment', 'line', 'lineEnd', 'lineStart', 'lineno',
|
| +'makeHTMLTags', 'makeXMLTags', 'matchOnlyAtCol', 'matchPreviousExpr', 'matchPreviousLiteral',
|
| +'nestedExpr', 'nullDebugAction', 'nums', 'oneOf', 'opAssoc', 'operatorPrecedence', 'printables',
|
| +'punc8bit', 'pythonStyleComment', 'quotedString', 'removeQuotes', 'replaceHTMLEntity',
|
| +'replaceWith', 'restOfLine', 'sglQuotedString', 'srange', 'stringEnd',
|
| +'stringStart', 'traceParseAction', 'unicodeString', 'upcaseTokens', 'withAttribute',
|
| +'indentedBlock', 'originalTextFor', 'ungroup', 'infixNotation','locatedExpr', 'withClass',
|
| +'CloseMatch', 'tokenMap', 'pyparsing_common',
|
| +]
|
| +
|
| +system_version = tuple(sys.version_info)[:3]
|
| +PY_3 = system_version[0] == 3
|
| +if PY_3:
|
| + _MAX_INT = sys.maxsize
|
| + basestring = str
|
| + unichr = chr
|
| + _ustr = str
|
| +
|
| + # build list of single arg builtins, that can be used as parse actions
|
| + singleArgBuiltins = [sum, len, sorted, reversed, list, tuple, set, any, all, min, max]
|
| +
|
| +else:
|
| + _MAX_INT = sys.maxint
|
| + range = xrange
|
| +
|
| + def _ustr(obj):
|
| + """Drop-in replacement for str(obj) that tries to be Unicode friendly. It first tries
|
| + str(obj). If that fails with a UnicodeEncodeError, then it tries unicode(obj). It
|
| + then < returns the unicode object | encodes it with the default encoding | ... >.
|
| + """
|
| + if isinstance(obj,unicode):
|
| + return obj
|
| +
|
| + try:
|
| + # If this works, then _ustr(obj) has the same behaviour as str(obj), so
|
| + # it won't break any existing code.
|
| + return str(obj)
|
| +
|
| + except UnicodeEncodeError:
|
| + # Else encode it
|
| + ret = unicode(obj).encode(sys.getdefaultencoding(), 'xmlcharrefreplace')
|
| + xmlcharref = Regex('&#\d+;')
|
| + xmlcharref.setParseAction(lambda t: '\\u' + hex(int(t[0][2:-1]))[2:])
|
| + return xmlcharref.transformString(ret)
|
| +
|
| + # build list of single arg builtins, tolerant of Python version, that can be used as parse actions
|
| + singleArgBuiltins = []
|
| + import __builtin__
|
| + for fname in "sum len sorted reversed list tuple set any all min max".split():
|
| + try:
|
| + singleArgBuiltins.append(getattr(__builtin__,fname))
|
| + except AttributeError:
|
| + continue
|
| +
|
| +_generatorType = type((y for y in range(1)))
|
| +
|
| +def _xml_escape(data):
|
| + """Escape &, <, >, ", ', etc. in a string of data."""
|
| +
|
| + # ampersand must be replaced first
|
| + from_symbols = '&><"\''
|
| + to_symbols = ('&'+s+';' for s in "amp gt lt quot apos".split())
|
| + for from_,to_ in zip(from_symbols, to_symbols):
|
| + data = data.replace(from_, to_)
|
| + return data
|
| +
|
| +class _Constants(object):
|
| + pass
|
| +
|
| +alphas = string.ascii_uppercase + string.ascii_lowercase
|
| +nums = "0123456789"
|
| +hexnums = nums + "ABCDEFabcdef"
|
| +alphanums = alphas + nums
|
| +_bslash = chr(92)
|
| +printables = "".join(c for c in string.printable if c not in string.whitespace)
|
| +
|
| +class ParseBaseException(Exception):
|
| + """base exception class for all parsing runtime exceptions"""
|
| + # Performance tuning: we construct a *lot* of these, so keep this
|
| + # constructor as small and fast as possible
|
| + def __init__( self, pstr, loc=0, msg=None, elem=None ):
|
| + self.loc = loc
|
| + if msg is None:
|
| + self.msg = pstr
|
| + self.pstr = ""
|
| + else:
|
| + self.msg = msg
|
| + self.pstr = pstr
|
| + self.parserElement = elem
|
| + self.args = (pstr, loc, msg)
|
| +
|
| + @classmethod
|
| + def _from_exception(cls, pe):
|
| + """
|
| + internal factory method to simplify creating one type of ParseException
|
| + from another - avoids having __init__ signature conflicts among subclasses
|
| + """
|
| + return cls(pe.pstr, pe.loc, pe.msg, pe.parserElement)
|
| +
|
| + def __getattr__( self, aname ):
|
| + """supported attributes by name are:
|
| + - lineno - returns the line number of the exception text
|
| + - col - returns the column number of the exception text
|
| + - line - returns the line containing the exception text
|
| + """
|
| + if( aname == "lineno" ):
|
| + return lineno( self.loc, self.pstr )
|
| + elif( aname in ("col", "column") ):
|
| + return col( self.loc, self.pstr )
|
| + elif( aname == "line" ):
|
| + return line( self.loc, self.pstr )
|
| + else:
|
| + raise AttributeError(aname)
|
| +
|
| + def __str__( self ):
|
| + return "%s (at char %d), (line:%d, col:%d)" % \
|
| + ( self.msg, self.loc, self.lineno, self.column )
|
| + def __repr__( self ):
|
| + return _ustr(self)
|
| + def markInputline( self, markerString = ">!<" ):
|
| + """Extracts the exception line from the input string, and marks
|
| + the location of the exception with a special symbol.
|
| + """
|
| + line_str = self.line
|
| + line_column = self.column - 1
|
| + if markerString:
|
| + line_str = "".join((line_str[:line_column],
|
| + markerString, line_str[line_column:]))
|
| + return line_str.strip()
|
| + def __dir__(self):
|
| + return "lineno col line".split() + dir(type(self))
|
| +
|
| +class ParseException(ParseBaseException):
|
| + """
|
| + Exception thrown when parse expressions don't match class;
|
| + supported attributes by name are:
|
| + - lineno - returns the line number of the exception text
|
| + - col - returns the column number of the exception text
|
| + - line - returns the line containing the exception text
|
| +
|
| + Example::
|
| + try:
|
| + Word(nums).setName("integer").parseString("ABC")
|
| + except ParseException as pe:
|
| + print(pe)
|
| + print("column: {}".format(pe.col))
|
| +
|
| + prints::
|
| + Expected integer (at char 0), (line:1, col:1)
|
| + column: 1
|
| + """
|
| + pass
|
| +
|
| +class ParseFatalException(ParseBaseException):
|
| + """user-throwable exception thrown when inconsistent parse content
|
| + is found; stops all parsing immediately"""
|
| + pass
|
| +
|
| +class ParseSyntaxException(ParseFatalException):
|
| + """just like L{ParseFatalException}, but thrown internally when an
|
| + L{ErrorStop<And._ErrorStop>} ('-' operator) indicates that parsing is to stop
|
| + immediately because an unbacktrackable syntax error has been found"""
|
| + pass
|
| +
|
| +#~ class ReparseException(ParseBaseException):
|
| + #~ """Experimental class - parse actions can raise this exception to cause
|
| + #~ pyparsing to reparse the input string:
|
| + #~ - with a modified input string, and/or
|
| + #~ - with a modified start location
|
| + #~ Set the values of the ReparseException in the constructor, and raise the
|
| + #~ exception in a parse action to cause pyparsing to use the new string/location.
|
| + #~ Setting the values as None causes no change to be made.
|
| + #~ """
|
| + #~ def __init_( self, newstring, restartLoc ):
|
| + #~ self.newParseText = newstring
|
| + #~ self.reparseLoc = restartLoc
|
| +
|
| +class RecursiveGrammarException(Exception):
|
| + """exception thrown by L{ParserElement.validate} if the grammar could be improperly recursive"""
|
| + def __init__( self, parseElementList ):
|
| + self.parseElementTrace = parseElementList
|
| +
|
| + def __str__( self ):
|
| + return "RecursiveGrammarException: %s" % self.parseElementTrace
|
| +
|
| +class _ParseResultsWithOffset(object):
|
| + def __init__(self,p1,p2):
|
| + self.tup = (p1,p2)
|
| + def __getitem__(self,i):
|
| + return self.tup[i]
|
| + def __repr__(self):
|
| + return repr(self.tup[0])
|
| + def setOffset(self,i):
|
| + self.tup = (self.tup[0],i)
|
| +
|
| +class ParseResults(object):
|
| + """
|
| + Structured parse results, to provide multiple means of access to the parsed data:
|
| + - as a list (C{len(results)})
|
| + - by list index (C{results[0], results[1]}, etc.)
|
| + - by attribute (C{results.<resultsName>} - see L{ParserElement.setResultsName})
|
| +
|
| + Example::
|
| + integer = Word(nums)
|
| + date_str = (integer.setResultsName("year") + '/'
|
| + + integer.setResultsName("month") + '/'
|
| + + integer.setResultsName("day"))
|
| + # equivalent form:
|
| + # date_str = integer("year") + '/' + integer("month") + '/' + integer("day")
|
| +
|
| + # parseString returns a ParseResults object
|
| + result = date_str.parseString("1999/12/31")
|
| +
|
| + def test(s, fn=repr):
|
| + print("%s -> %s" % (s, fn(eval(s))))
|
| + test("list(result)")
|
| + test("result[0]")
|
| + test("result['month']")
|
| + test("result.day")
|
| + test("'month' in result")
|
| + test("'minutes' in result")
|
| + test("result.dump()", str)
|
| + prints::
|
| + list(result) -> ['1999', '/', '12', '/', '31']
|
| + result[0] -> '1999'
|
| + result['month'] -> '12'
|
| + result.day -> '31'
|
| + 'month' in result -> True
|
| + 'minutes' in result -> False
|
| + result.dump() -> ['1999', '/', '12', '/', '31']
|
| + - day: 31
|
| + - month: 12
|
| + - year: 1999
|
| + """
|
| + def __new__(cls, toklist=None, name=None, asList=True, modal=True ):
|
| + if isinstance(toklist, cls):
|
| + return toklist
|
| + retobj = object.__new__(cls)
|
| + retobj.__doinit = True
|
| + return retobj
|
| +
|
| + # Performance tuning: we construct a *lot* of these, so keep this
|
| + # constructor as small and fast as possible
|
| + def __init__( self, toklist=None, name=None, asList=True, modal=True, isinstance=isinstance ):
|
| + if self.__doinit:
|
| + self.__doinit = False
|
| + self.__name = None
|
| + self.__parent = None
|
| + self.__accumNames = {}
|
| + self.__asList = asList
|
| + self.__modal = modal
|
| + if toklist is None:
|
| + toklist = []
|
| + if isinstance(toklist, list):
|
| + self.__toklist = toklist[:]
|
| + elif isinstance(toklist, _generatorType):
|
| + self.__toklist = list(toklist)
|
| + else:
|
| + self.__toklist = [toklist]
|
| + self.__tokdict = dict()
|
| +
|
| + if name is not None and name:
|
| + if not modal:
|
| + self.__accumNames[name] = 0
|
| + if isinstance(name,int):
|
| + name = _ustr(name) # will always return a str, but use _ustr for consistency
|
| + self.__name = name
|
| + if not (isinstance(toklist, (type(None), basestring, list)) and toklist in (None,'',[])):
|
| + if isinstance(toklist,basestring):
|
| + toklist = [ toklist ]
|
| + if asList:
|
| + if isinstance(toklist,ParseResults):
|
| + self[name] = _ParseResultsWithOffset(toklist.copy(),0)
|
| + else:
|
| + self[name] = _ParseResultsWithOffset(ParseResults(toklist[0]),0)
|
| + self[name].__name = name
|
| + else:
|
| + try:
|
| + self[name] = toklist[0]
|
| + except (KeyError,TypeError,IndexError):
|
| + self[name] = toklist
|
| +
|
| + def __getitem__( self, i ):
|
| + if isinstance( i, (int,slice) ):
|
| + return self.__toklist[i]
|
| + else:
|
| + if i not in self.__accumNames:
|
| + return self.__tokdict[i][-1][0]
|
| + else:
|
| + return ParseResults([ v[0] for v in self.__tokdict[i] ])
|
| +
|
| + def __setitem__( self, k, v, isinstance=isinstance ):
|
| + if isinstance(v,_ParseResultsWithOffset):
|
| + self.__tokdict[k] = self.__tokdict.get(k,list()) + [v]
|
| + sub = v[0]
|
| + elif isinstance(k,(int,slice)):
|
| + self.__toklist[k] = v
|
| + sub = v
|
| + else:
|
| + self.__tokdict[k] = self.__tokdict.get(k,list()) + [_ParseResultsWithOffset(v,0)]
|
| + sub = v
|
| + if isinstance(sub,ParseResults):
|
| + sub.__parent = wkref(self)
|
| +
|
| + def __delitem__( self, i ):
|
| + if isinstance(i,(int,slice)):
|
| + mylen = len( self.__toklist )
|
| + del self.__toklist[i]
|
| +
|
| + # convert int to slice
|
| + if isinstance(i, int):
|
| + if i < 0:
|
| + i += mylen
|
| + i = slice(i, i+1)
|
| + # get removed indices
|
| + removed = list(range(*i.indices(mylen)))
|
| + removed.reverse()
|
| + # fixup indices in token dictionary
|
| + for name,occurrences in self.__tokdict.items():
|
| + for j in removed:
|
| + for k, (value, position) in enumerate(occurrences):
|
| + occurrences[k] = _ParseResultsWithOffset(value, position - (position > j))
|
| + else:
|
| + del self.__tokdict[i]
|
| +
|
| + def __contains__( self, k ):
|
| + return k in self.__tokdict
|
| +
|
| + def __len__( self ): return len( self.__toklist )
|
| + def __bool__(self): return ( not not self.__toklist )
|
| + __nonzero__ = __bool__
|
| + def __iter__( self ): return iter( self.__toklist )
|
| + def __reversed__( self ): return iter( self.__toklist[::-1] )
|
| + def _iterkeys( self ):
|
| + if hasattr(self.__tokdict, "iterkeys"):
|
| + return self.__tokdict.iterkeys()
|
| + else:
|
| + return iter(self.__tokdict)
|
| +
|
| + def _itervalues( self ):
|
| + return (self[k] for k in self._iterkeys())
|
| +
|
| + def _iteritems( self ):
|
| + return ((k, self[k]) for k in self._iterkeys())
|
| +
|
| + if PY_3:
|
| + keys = _iterkeys
|
| + """Returns an iterator of all named result keys (Python 3.x only)."""
|
| +
|
| + values = _itervalues
|
| + """Returns an iterator of all named result values (Python 3.x only)."""
|
| +
|
| + items = _iteritems
|
| + """Returns an iterator of all named result key-value tuples (Python 3.x only)."""
|
| +
|
| + else:
|
| + iterkeys = _iterkeys
|
| + """Returns an iterator of all named result keys (Python 2.x only)."""
|
| +
|
| + itervalues = _itervalues
|
| + """Returns an iterator of all named result values (Python 2.x only)."""
|
| +
|
| + iteritems = _iteritems
|
| + """Returns an iterator of all named result key-value tuples (Python 2.x only)."""
|
| +
|
| + def keys( self ):
|
| + """Returns all named result keys (as a list in Python 2.x, as an iterator in Python 3.x)."""
|
| + return list(self.iterkeys())
|
| +
|
| + def values( self ):
|
| + """Returns all named result values (as a list in Python 2.x, as an iterator in Python 3.x)."""
|
| + return list(self.itervalues())
|
| +
|
| + def items( self ):
|
| + """Returns all named result key-values (as a list of tuples in Python 2.x, as an iterator in Python 3.x)."""
|
| + return list(self.iteritems())
|
| +
|
| + def haskeys( self ):
|
| + """Since keys() returns an iterator, this method is helpful in bypassing
|
| + code that looks for the existence of any defined results names."""
|
| + return bool(self.__tokdict)
|
| +
|
| + def pop( self, *args, **kwargs):
|
| + """
|
| + Removes and returns item at specified index (default=C{last}).
|
| + Supports both C{list} and C{dict} semantics for C{pop()}. If passed no
|
| + argument or an integer argument, it will use C{list} semantics
|
| + and pop tokens from the list of parsed tokens. If passed a
|
| + non-integer argument (most likely a string), it will use C{dict}
|
| + semantics and pop the corresponding value from any defined
|
| + results names. A second default return value argument is
|
| + supported, just as in C{dict.pop()}.
|
| +
|
| + Example::
|
| + def remove_first(tokens):
|
| + tokens.pop(0)
|
| + print(OneOrMore(Word(nums)).parseString("0 123 321")) # -> ['0', '123', '321']
|
| + print(OneOrMore(Word(nums)).addParseAction(remove_first).parseString("0 123 321")) # -> ['123', '321']
|
| +
|
| + label = Word(alphas)
|
| + patt = label("LABEL") + OneOrMore(Word(nums))
|
| + print(patt.parseString("AAB 123 321").dump())
|
| +
|
| + # Use pop() in a parse action to remove named result (note that corresponding value is not
|
| + # removed from list form of results)
|
| + def remove_LABEL(tokens):
|
| + tokens.pop("LABEL")
|
| + return tokens
|
| + patt.addParseAction(remove_LABEL)
|
| + print(patt.parseString("AAB 123 321").dump())
|
| + prints::
|
| + ['AAB', '123', '321']
|
| + - LABEL: AAB
|
| +
|
| + ['AAB', '123', '321']
|
| + """
|
| + if not args:
|
| + args = [-1]
|
| + for k,v in kwargs.items():
|
| + if k == 'default':
|
| + args = (args[0], v)
|
| + else:
|
| + raise TypeError("pop() got an unexpected keyword argument '%s'" % k)
|
| + if (isinstance(args[0], int) or
|
| + len(args) == 1 or
|
| + args[0] in self):
|
| + index = args[0]
|
| + ret = self[index]
|
| + del self[index]
|
| + return ret
|
| + else:
|
| + defaultvalue = args[1]
|
| + return defaultvalue
|
| +
|
| + def get(self, key, defaultValue=None):
|
| + """
|
| + Returns named result matching the given key, or if there is no
|
| + such name, then returns the given C{defaultValue} or C{None} if no
|
| + C{defaultValue} is specified.
|
| +
|
| + Similar to C{dict.get()}.
|
| +
|
| + Example::
|
| + integer = Word(nums)
|
| + date_str = integer("year") + '/' + integer("month") + '/' + integer("day")
|
| +
|
| + result = date_str.parseString("1999/12/31")
|
| + print(result.get("year")) # -> '1999'
|
| + print(result.get("hour", "not specified")) # -> 'not specified'
|
| + print(result.get("hour")) # -> None
|
| + """
|
| + if key in self:
|
| + return self[key]
|
| + else:
|
| + return defaultValue
|
| +
|
| + def insert( self, index, insStr ):
|
| + """
|
| + Inserts new element at location index in the list of parsed tokens.
|
| +
|
| + Similar to C{list.insert()}.
|
| +
|
| + Example::
|
| + print(OneOrMore(Word(nums)).parseString("0 123 321")) # -> ['0', '123', '321']
|
| +
|
| + # use a parse action to insert the parse location in the front of the parsed results
|
| + def insert_locn(locn, tokens):
|
| + tokens.insert(0, locn)
|
| + print(OneOrMore(Word(nums)).addParseAction(insert_locn).parseString("0 123 321")) # -> [0, '0', '123', '321']
|
| + """
|
| + self.__toklist.insert(index, insStr)
|
| + # fixup indices in token dictionary
|
| + for name,occurrences in self.__tokdict.items():
|
| + for k, (value, position) in enumerate(occurrences):
|
| + occurrences[k] = _ParseResultsWithOffset(value, position + (position > index))
|
| +
|
| + def append( self, item ):
|
| + """
|
| + Add single element to end of ParseResults list of elements.
|
| +
|
| + Example::
|
| + print(OneOrMore(Word(nums)).parseString("0 123 321")) # -> ['0', '123', '321']
|
| +
|
| + # use a parse action to compute the sum of the parsed integers, and add it to the end
|
| + def append_sum(tokens):
|
| + tokens.append(sum(map(int, tokens)))
|
| + print(OneOrMore(Word(nums)).addParseAction(append_sum).parseString("0 123 321")) # -> ['0', '123', '321', 444]
|
| + """
|
| + self.__toklist.append(item)
|
| +
|
| + def extend( self, itemseq ):
|
| + """
|
| + Add sequence of elements to end of ParseResults list of elements.
|
| +
|
| + Example::
|
| + patt = OneOrMore(Word(alphas))
|
| +
|
| + # use a parse action to append the reverse of the matched strings, to make a palindrome
|
| + def make_palindrome(tokens):
|
| + tokens.extend(reversed([t[::-1] for t in tokens]))
|
| + return ''.join(tokens)
|
| + print(patt.addParseAction(make_palindrome).parseString("lskdj sdlkjf lksd")) # -> 'lskdjsdlkjflksddsklfjkldsjdksl'
|
| + """
|
| + if isinstance(itemseq, ParseResults):
|
| + self += itemseq
|
| + else:
|
| + self.__toklist.extend(itemseq)
|
| +
|
| + def clear( self ):
|
| + """
|
| + Clear all elements and results names.
|
| + """
|
| + del self.__toklist[:]
|
| + self.__tokdict.clear()
|
| +
|
| + def __getattr__( self, name ):
|
| + try:
|
| + return self[name]
|
| + except KeyError:
|
| + return ""
|
| +
|
| + if name in self.__tokdict:
|
| + if name not in self.__accumNames:
|
| + return self.__tokdict[name][-1][0]
|
| + else:
|
| + return ParseResults([ v[0] for v in self.__tokdict[name] ])
|
| + else:
|
| + return ""
|
| +
|
| + def __add__( self, other ):
|
| + ret = self.copy()
|
| + ret += other
|
| + return ret
|
| +
|
| + def __iadd__( self, other ):
|
| + if other.__tokdict:
|
| + offset = len(self.__toklist)
|
| + addoffset = lambda a: offset if a<0 else a+offset
|
| + otheritems = other.__tokdict.items()
|
| + otherdictitems = [(k, _ParseResultsWithOffset(v[0],addoffset(v[1])) )
|
| + for (k,vlist) in otheritems for v in vlist]
|
| + for k,v in otherdictitems:
|
| + self[k] = v
|
| + if isinstance(v[0],ParseResults):
|
| + v[0].__parent = wkref(self)
|
| +
|
| + self.__toklist += other.__toklist
|
| + self.__accumNames.update( other.__accumNames )
|
| + return self
|
| +
|
| + def __radd__(self, other):
|
| + if isinstance(other,int) and other == 0:
|
| + # useful for merging many ParseResults using sum() builtin
|
| + return self.copy()
|
| + else:
|
| + # this may raise a TypeError - so be it
|
| + return other + self
|
| +
|
| + def __repr__( self ):
|
| + return "(%s, %s)" % ( repr( self.__toklist ), repr( self.__tokdict ) )
|
| +
|
| + def __str__( self ):
|
| + return '[' + ', '.join(_ustr(i) if isinstance(i, ParseResults) else repr(i) for i in self.__toklist) + ']'
|
| +
|
| + def _asStringList( self, sep='' ):
|
| + out = []
|
| + for item in self.__toklist:
|
| + if out and sep:
|
| + out.append(sep)
|
| + if isinstance( item, ParseResults ):
|
| + out += item._asStringList()
|
| + else:
|
| + out.append( _ustr(item) )
|
| + return out
|
| +
|
| + def asList( self ):
|
| + """
|
| + Returns the parse results as a nested list of matching tokens, all converted to strings.
|
| +
|
| + Example::
|
| + patt = OneOrMore(Word(alphas))
|
| + result = patt.parseString("sldkj lsdkj sldkj")
|
| + # even though the result prints in string-like form, it is actually a pyparsing ParseResults
|
| + print(type(result), result) # -> <class 'pyparsing.ParseResults'> ['sldkj', 'lsdkj', 'sldkj']
|
| +
|
| + # Use asList() to create an actual list
|
| + result_list = result.asList()
|
| + print(type(result_list), result_list) # -> <class 'list'> ['sldkj', 'lsdkj', 'sldkj']
|
| + """
|
| + return [res.asList() if isinstance(res,ParseResults) else res for res in self.__toklist]
|
| +
|
| + def asDict( self ):
|
| + """
|
| + Returns the named parse results as a nested dictionary.
|
| +
|
| + Example::
|
| + integer = Word(nums)
|
| + date_str = integer("year") + '/' + integer("month") + '/' + integer("day")
|
| +
|
| + result = date_str.parseString('12/31/1999')
|
| + print(type(result), repr(result)) # -> <class 'pyparsing.ParseResults'> (['12', '/', '31', '/', '1999'], {'day': [('1999', 4)], 'year': [('12', 0)], 'month': [('31', 2)]})
|
| +
|
| + result_dict = result.asDict()
|
| + print(type(result_dict), repr(result_dict)) # -> <class 'dict'> {'day': '1999', 'year': '12', 'month': '31'}
|
| +
|
| + # even though a ParseResults supports dict-like access, sometime you just need to have a dict
|
| + import json
|
| + print(json.dumps(result)) # -> Exception: TypeError: ... is not JSON serializable
|
| + print(json.dumps(result.asDict())) # -> {"month": "31", "day": "1999", "year": "12"}
|
| + """
|
| + if PY_3:
|
| + item_fn = self.items
|
| + else:
|
| + item_fn = self.iteritems
|
| +
|
| + def toItem(obj):
|
| + if isinstance(obj, ParseResults):
|
| + if obj.haskeys():
|
| + return obj.asDict()
|
| + else:
|
| + return [toItem(v) for v in obj]
|
| + else:
|
| + return obj
|
| +
|
| + return dict((k,toItem(v)) for k,v in item_fn())
|
| +
|
| + def copy( self ):
|
| + """
|
| + Returns a new copy of a C{ParseResults} object.
|
| + """
|
| + ret = ParseResults( self.__toklist )
|
| + ret.__tokdict = self.__tokdict.copy()
|
| + ret.__parent = self.__parent
|
| + ret.__accumNames.update( self.__accumNames )
|
| + ret.__name = self.__name
|
| + return ret
|
| +
|
| + def asXML( self, doctag=None, namedItemsOnly=False, indent="", formatted=True ):
|
| + """
|
| + (Deprecated) Returns the parse results as XML. Tags are created for tokens and lists that have defined results names.
|
| + """
|
| + nl = "\n"
|
| + out = []
|
| + namedItems = dict((v[1],k) for (k,vlist) in self.__tokdict.items()
|
| + for v in vlist)
|
| + nextLevelIndent = indent + " "
|
| +
|
| + # collapse out indents if formatting is not desired
|
| + if not formatted:
|
| + indent = ""
|
| + nextLevelIndent = ""
|
| + nl = ""
|
| +
|
| + selfTag = None
|
| + if doctag is not None:
|
| + selfTag = doctag
|
| + else:
|
| + if self.__name:
|
| + selfTag = self.__name
|
| +
|
| + if not selfTag:
|
| + if namedItemsOnly:
|
| + return ""
|
| + else:
|
| + selfTag = "ITEM"
|
| +
|
| + out += [ nl, indent, "<", selfTag, ">" ]
|
| +
|
| + for i,res in enumerate(self.__toklist):
|
| + if isinstance(res,ParseResults):
|
| + if i in namedItems:
|
| + out += [ res.asXML(namedItems[i],
|
| + namedItemsOnly and doctag is None,
|
| + nextLevelIndent,
|
| + formatted)]
|
| + else:
|
| + out += [ res.asXML(None,
|
| + namedItemsOnly and doctag is None,
|
| + nextLevelIndent,
|
| + formatted)]
|
| + else:
|
| + # individual token, see if there is a name for it
|
| + resTag = None
|
| + if i in namedItems:
|
| + resTag = namedItems[i]
|
| + if not resTag:
|
| + if namedItemsOnly:
|
| + continue
|
| + else:
|
| + resTag = "ITEM"
|
| + xmlBodyText = _xml_escape(_ustr(res))
|
| + out += [ nl, nextLevelIndent, "<", resTag, ">",
|
| + xmlBodyText,
|
| + "</", resTag, ">" ]
|
| +
|
| + out += [ nl, indent, "</", selfTag, ">" ]
|
| + return "".join(out)
|
| +
|
| + def __lookup(self,sub):
|
| + for k,vlist in self.__tokdict.items():
|
| + for v,loc in vlist:
|
| + if sub is v:
|
| + return k
|
| + return None
|
| +
|
| + def getName(self):
|
| + """
|
| + Returns the results name for this token expression. Useful when several
|
| + different expressions might match at a particular location.
|
| +
|
| + Example::
|
| + integer = Word(nums)
|
| + ssn_expr = Regex(r"\d\d\d-\d\d-\d\d\d\d")
|
| + house_number_expr = Suppress('#') + Word(nums, alphanums)
|
| + user_data = (Group(house_number_expr)("house_number")
|
| + | Group(ssn_expr)("ssn")
|
| + | Group(integer)("age"))
|
| + user_info = OneOrMore(user_data)
|
| +
|
| + result = user_info.parseString("22 111-22-3333 #221B")
|
| + for item in result:
|
| + print(item.getName(), ':', item[0])
|
| + prints::
|
| + age : 22
|
| + ssn : 111-22-3333
|
| + house_number : 221B
|
| + """
|
| + if self.__name:
|
| + return self.__name
|
| + elif self.__parent:
|
| + par = self.__parent()
|
| + if par:
|
| + return par.__lookup(self)
|
| + else:
|
| + return None
|
| + elif (len(self) == 1 and
|
| + len(self.__tokdict) == 1 and
|
| + next(iter(self.__tokdict.values()))[0][1] in (0,-1)):
|
| + return next(iter(self.__tokdict.keys()))
|
| + else:
|
| + return None
|
| +
|
| + def dump(self, indent='', depth=0, full=True):
|
| + """
|
| + Diagnostic method for listing out the contents of a C{ParseResults}.
|
| + Accepts an optional C{indent} argument so that this string can be embedded
|
| + in a nested display of other data.
|
| +
|
| + Example::
|
| + integer = Word(nums)
|
| + date_str = integer("year") + '/' + integer("month") + '/' + integer("day")
|
| +
|
| + result = date_str.parseString('12/31/1999')
|
| + print(result.dump())
|
| + prints::
|
| + ['12', '/', '31', '/', '1999']
|
| + - day: 1999
|
| + - month: 31
|
| + - year: 12
|
| + """
|
| + out = []
|
| + NL = '\n'
|
| + out.append( indent+_ustr(self.asList()) )
|
| + if full:
|
| + if self.haskeys():
|
| + items = sorted((str(k), v) for k,v in self.items())
|
| + for k,v in items:
|
| + if out:
|
| + out.append(NL)
|
| + out.append( "%s%s- %s: " % (indent,(' '*depth), k) )
|
| + if isinstance(v,ParseResults):
|
| + if v:
|
| + out.append( v.dump(indent,depth+1) )
|
| + else:
|
| + out.append(_ustr(v))
|
| + else:
|
| + out.append(repr(v))
|
| + elif any(isinstance(vv,ParseResults) for vv in self):
|
| + v = self
|
| + for i,vv in enumerate(v):
|
| + if isinstance(vv,ParseResults):
|
| + out.append("\n%s%s[%d]:\n%s%s%s" % (indent,(' '*(depth)),i,indent,(' '*(depth+1)),vv.dump(indent,depth+1) ))
|
| + else:
|
| + out.append("\n%s%s[%d]:\n%s%s%s" % (indent,(' '*(depth)),i,indent,(' '*(depth+1)),_ustr(vv)))
|
| +
|
| + return "".join(out)
|
| +
|
| + def pprint(self, *args, **kwargs):
|
| + """
|
| + Pretty-printer for parsed results as a list, using the C{pprint} module.
|
| + Accepts additional positional or keyword args as defined for the
|
| + C{pprint.pprint} method. (U{http://docs.python.org/3/library/pprint.html#pprint.pprint})
|
| +
|
| + Example::
|
| + ident = Word(alphas, alphanums)
|
| + num = Word(nums)
|
| + func = Forward()
|
| + term = ident | num | Group('(' + func + ')')
|
| + func <<= ident + Group(Optional(delimitedList(term)))
|
| + result = func.parseString("fna a,b,(fnb c,d,200),100")
|
| + result.pprint(width=40)
|
| + prints::
|
| + ['fna',
|
| + ['a',
|
| + 'b',
|
| + ['(', 'fnb', ['c', 'd', '200'], ')'],
|
| + '100']]
|
| + """
|
| + pprint.pprint(self.asList(), *args, **kwargs)
|
| +
|
| + # add support for pickle protocol
|
| + def __getstate__(self):
|
| + return ( self.__toklist,
|
| + ( self.__tokdict.copy(),
|
| + self.__parent is not None and self.__parent() or None,
|
| + self.__accumNames,
|
| + self.__name ) )
|
| +
|
| + def __setstate__(self,state):
|
| + self.__toklist = state[0]
|
| + (self.__tokdict,
|
| + par,
|
| + inAccumNames,
|
| + self.__name) = state[1]
|
| + self.__accumNames = {}
|
| + self.__accumNames.update(inAccumNames)
|
| + if par is not None:
|
| + self.__parent = wkref(par)
|
| + else:
|
| + self.__parent = None
|
| +
|
| + def __getnewargs__(self):
|
| + return self.__toklist, self.__name, self.__asList, self.__modal
|
| +
|
| + def __dir__(self):
|
| + return (dir(type(self)) + list(self.keys()))
|
| +
|
| +collections.MutableMapping.register(ParseResults)
|
| +
|
| +def col (loc,strg):
|
| + """Returns current column within a string, counting newlines as line separators.
|
| + The first column is number 1.
|
| +
|
| + Note: the default parsing behavior is to expand tabs in the input string
|
| + before starting the parsing process. See L{I{ParserElement.parseString}<ParserElement.parseString>} for more information
|
| + on parsing strings containing C{<TAB>}s, and suggested methods to maintain a
|
| + consistent view of the parsed string, the parse location, and line and column
|
| + positions within the parsed string.
|
| + """
|
| + s = strg
|
| + return 1 if 0<loc<len(s) and s[loc-1] == '\n' else loc - s.rfind("\n", 0, loc)
|
| +
|
| +def lineno(loc,strg):
|
| + """Returns current line number within a string, counting newlines as line separators.
|
| + The first line is number 1.
|
| +
|
| + Note: the default parsing behavior is to expand tabs in the input string
|
| + before starting the parsing process. See L{I{ParserElement.parseString}<ParserElement.parseString>} for more information
|
| + on parsing strings containing C{<TAB>}s, and suggested methods to maintain a
|
| + consistent view of the parsed string, the parse location, and line and column
|
| + positions within the parsed string.
|
| + """
|
| + return strg.count("\n",0,loc) + 1
|
| +
|
| +def line( loc, strg ):
|
| + """Returns the line of text containing loc within a string, counting newlines as line separators.
|
| + """
|
| + lastCR = strg.rfind("\n", 0, loc)
|
| + nextCR = strg.find("\n", loc)
|
| + if nextCR >= 0:
|
| + return strg[lastCR+1:nextCR]
|
| + else:
|
| + return strg[lastCR+1:]
|
| +
|
| +def _defaultStartDebugAction( instring, loc, expr ):
|
| + print (("Match " + _ustr(expr) + " at loc " + _ustr(loc) + "(%d,%d)" % ( lineno(loc,instring), col(loc,instring) )))
|
| +
|
| +def _defaultSuccessDebugAction( instring, startloc, endloc, expr, toks ):
|
| + print ("Matched " + _ustr(expr) + " -> " + str(toks.asList()))
|
| +
|
| +def _defaultExceptionDebugAction( instring, loc, expr, exc ):
|
| + print ("Exception raised:" + _ustr(exc))
|
| +
|
| +def nullDebugAction(*args):
|
| + """'Do-nothing' debug action, to suppress debugging output during parsing."""
|
| + pass
|
| +
|
| +# Only works on Python 3.x - nonlocal is toxic to Python 2 installs
|
| +#~ 'decorator to trim function calls to match the arity of the target'
|
| +#~ def _trim_arity(func, maxargs=3):
|
| + #~ if func in singleArgBuiltins:
|
| + #~ return lambda s,l,t: func(t)
|
| + #~ limit = 0
|
| + #~ foundArity = False
|
| + #~ def wrapper(*args):
|
| + #~ nonlocal limit,foundArity
|
| + #~ while 1:
|
| + #~ try:
|
| + #~ ret = func(*args[limit:])
|
| + #~ foundArity = True
|
| + #~ return ret
|
| + #~ except TypeError:
|
| + #~ if limit == maxargs or foundArity:
|
| + #~ raise
|
| + #~ limit += 1
|
| + #~ continue
|
| + #~ return wrapper
|
| +
|
| +# this version is Python 2.x-3.x cross-compatible
|
| +'decorator to trim function calls to match the arity of the target'
|
| +def _trim_arity(func, maxargs=2):
|
| + if func in singleArgBuiltins:
|
| + return lambda s,l,t: func(t)
|
| + limit = [0]
|
| + foundArity = [False]
|
| +
|
| + # traceback return data structure changed in Py3.5 - normalize back to plain tuples
|
| + if system_version[:2] >= (3,5):
|
| + def extract_stack(limit=0):
|
| + # special handling for Python 3.5.0 - extra deep call stack by 1
|
| + offset = -3 if system_version == (3,5,0) else -2
|
| + frame_summary = traceback.extract_stack(limit=-offset+limit-1)[offset]
|
| + return [(frame_summary.filename, frame_summary.lineno)]
|
| + def extract_tb(tb, limit=0):
|
| + frames = traceback.extract_tb(tb, limit=limit)
|
| + frame_summary = frames[-1]
|
| + return [(frame_summary.filename, frame_summary.lineno)]
|
| + else:
|
| + extract_stack = traceback.extract_stack
|
| + extract_tb = traceback.extract_tb
|
| +
|
| + # synthesize what would be returned by traceback.extract_stack at the call to
|
| + # user's parse action 'func', so that we don't incur call penalty at parse time
|
| +
|
| + LINE_DIFF = 6
|
| + # IF ANY CODE CHANGES, EVEN JUST COMMENTS OR BLANK LINES, BETWEEN THE NEXT LINE AND
|
| + # THE CALL TO FUNC INSIDE WRAPPER, LINE_DIFF MUST BE MODIFIED!!!!
|
| + this_line = extract_stack(limit=2)[-1]
|
| + pa_call_line_synth = (this_line[0], this_line[1]+LINE_DIFF)
|
| +
|
| + def wrapper(*args):
|
| + while 1:
|
| + try:
|
| + ret = func(*args[limit[0]:])
|
| + foundArity[0] = True
|
| + return ret
|
| + except TypeError:
|
| + # re-raise TypeErrors if they did not come from our arity testing
|
| + if foundArity[0]:
|
| + raise
|
| + else:
|
| + try:
|
| + tb = sys.exc_info()[-1]
|
| + if not extract_tb(tb, limit=2)[-1][:2] == pa_call_line_synth:
|
| + raise
|
| + finally:
|
| + del tb
|
| +
|
| + if limit[0] <= maxargs:
|
| + limit[0] += 1
|
| + continue
|
| + raise
|
| +
|
| + # copy func name to wrapper for sensible debug output
|
| + func_name = "<parse action>"
|
| + try:
|
| + func_name = getattr(func, '__name__',
|
| + getattr(func, '__class__').__name__)
|
| + except Exception:
|
| + func_name = str(func)
|
| + wrapper.__name__ = func_name
|
| +
|
| + return wrapper
|
| +
|
| +class ParserElement(object):
|
| + """Abstract base level parser element class."""
|
| + DEFAULT_WHITE_CHARS = " \n\t\r"
|
| + verbose_stacktrace = False
|
| +
|
| + @staticmethod
|
| + def setDefaultWhitespaceChars( chars ):
|
| + r"""
|
| + Overrides the default whitespace chars
|
| +
|
| + Example::
|
| + # default whitespace chars are space, <TAB> and newline
|
| + OneOrMore(Word(alphas)).parseString("abc def\nghi jkl") # -> ['abc', 'def', 'ghi', 'jkl']
|
| +
|
| + # change to just treat newline as significant
|
| + ParserElement.setDefaultWhitespaceChars(" \t")
|
| + OneOrMore(Word(alphas)).parseString("abc def\nghi jkl") # -> ['abc', 'def']
|
| + """
|
| + ParserElement.DEFAULT_WHITE_CHARS = chars
|
| +
|
| + @staticmethod
|
| + def inlineLiteralsUsing(cls):
|
| + """
|
| + Set class to be used for inclusion of string literals into a parser.
|
| +
|
| + Example::
|
| + # default literal class used is Literal
|
| + integer = Word(nums)
|
| + date_str = integer("year") + '/' + integer("month") + '/' + integer("day")
|
| +
|
| + date_str.parseString("1999/12/31") # -> ['1999', '/', '12', '/', '31']
|
| +
|
| +
|
| + # change to Suppress
|
| + ParserElement.inlineLiteralsUsing(Suppress)
|
| + date_str = integer("year") + '/' + integer("month") + '/' + integer("day")
|
| +
|
| + date_str.parseString("1999/12/31") # -> ['1999', '12', '31']
|
| + """
|
| + ParserElement._literalStringClass = cls
|
| +
|
| + def __init__( self, savelist=False ):
|
| + self.parseAction = list()
|
| + self.failAction = None
|
| + #~ self.name = "<unknown>" # don't define self.name, let subclasses try/except upcall
|
| + self.strRepr = None
|
| + self.resultsName = None
|
| + self.saveAsList = savelist
|
| + self.skipWhitespace = True
|
| + self.whiteChars = ParserElement.DEFAULT_WHITE_CHARS
|
| + self.copyDefaultWhiteChars = True
|
| + self.mayReturnEmpty = False # used when checking for left-recursion
|
| + self.keepTabs = False
|
| + self.ignoreExprs = list()
|
| + self.debug = False
|
| + self.streamlined = False
|
| + self.mayIndexError = True # used to optimize exception handling for subclasses that don't advance parse index
|
| + self.errmsg = ""
|
| + self.modalResults = True # used to mark results names as modal (report only last) or cumulative (list all)
|
| + self.debugActions = ( None, None, None ) #custom debug actions
|
| + self.re = None
|
| + self.callPreparse = True # used to avoid redundant calls to preParse
|
| + self.callDuringTry = False
|
| +
|
| + def copy( self ):
|
| + """
|
| + Make a copy of this C{ParserElement}. Useful for defining different parse actions
|
| + for the same parsing pattern, using copies of the original parse element.
|
| +
|
| + Example::
|
| + integer = Word(nums).setParseAction(lambda toks: int(toks[0]))
|
| + integerK = integer.copy().addParseAction(lambda toks: toks[0]*1024) + Suppress("K")
|
| + integerM = integer.copy().addParseAction(lambda toks: toks[0]*1024*1024) + Suppress("M")
|
| +
|
| + print(OneOrMore(integerK | integerM | integer).parseString("5K 100 640K 256M"))
|
| + prints::
|
| + [5120, 100, 655360, 268435456]
|
| + Equivalent form of C{expr.copy()} is just C{expr()}::
|
| + integerM = integer().addParseAction(lambda toks: toks[0]*1024*1024) + Suppress("M")
|
| + """
|
| + cpy = copy.copy( self )
|
| + cpy.parseAction = self.parseAction[:]
|
| + cpy.ignoreExprs = self.ignoreExprs[:]
|
| + if self.copyDefaultWhiteChars:
|
| + cpy.whiteChars = ParserElement.DEFAULT_WHITE_CHARS
|
| + return cpy
|
| +
|
| + def setName( self, name ):
|
| + """
|
| + Define name for this expression, makes debugging and exception messages clearer.
|
| +
|
| + Example::
|
| + Word(nums).parseString("ABC") # -> Exception: Expected W:(0123...) (at char 0), (line:1, col:1)
|
| + Word(nums).setName("integer").parseString("ABC") # -> Exception: Expected integer (at char 0), (line:1, col:1)
|
| + """
|
| + self.name = name
|
| + self.errmsg = "Expected " + self.name
|
| + if hasattr(self,"exception"):
|
| + self.exception.msg = self.errmsg
|
| + return self
|
| +
|
| + def setResultsName( self, name, listAllMatches=False ):
|
| + """
|
| + Define name for referencing matching tokens as a nested attribute
|
| + of the returned parse results.
|
| + NOTE: this returns a *copy* of the original C{ParserElement} object;
|
| + this is so that the client can define a basic element, such as an
|
| + integer, and reference it in multiple places with different names.
|
| +
|
| + You can also set results names using the abbreviated syntax,
|
| + C{expr("name")} in place of C{expr.setResultsName("name")} -
|
| + see L{I{__call__}<__call__>}.
|
| +
|
| + Example::
|
| + date_str = (integer.setResultsName("year") + '/'
|
| + + integer.setResultsName("month") + '/'
|
| + + integer.setResultsName("day"))
|
| +
|
| + # equivalent form:
|
| + date_str = integer("year") + '/' + integer("month") + '/' + integer("day")
|
| + """
|
| + newself = self.copy()
|
| + if name.endswith("*"):
|
| + name = name[:-1]
|
| + listAllMatches=True
|
| + newself.resultsName = name
|
| + newself.modalResults = not listAllMatches
|
| + return newself
|
| +
|
| + def setBreak(self,breakFlag = True):
|
| + """Method to invoke the Python pdb debugger when this element is
|
| + about to be parsed. Set C{breakFlag} to True to enable, False to
|
| + disable.
|
| + """
|
| + if breakFlag:
|
| + _parseMethod = self._parse
|
| + def breaker(instring, loc, doActions=True, callPreParse=True):
|
| + import pdb
|
| + pdb.set_trace()
|
| + return _parseMethod( instring, loc, doActions, callPreParse )
|
| + breaker._originalParseMethod = _parseMethod
|
| + self._parse = breaker
|
| + else:
|
| + if hasattr(self._parse,"_originalParseMethod"):
|
| + self._parse = self._parse._originalParseMethod
|
| + return self
|
| +
|
| + def setParseAction( self, *fns, **kwargs ):
|
| + """
|
| + Define action to perform when successfully matching parse element definition.
|
| + Parse action fn is a callable method with 0-3 arguments, called as C{fn(s,loc,toks)},
|
| + C{fn(loc,toks)}, C{fn(toks)}, or just C{fn()}, where:
|
| + - s = the original string being parsed (see note below)
|
| + - loc = the location of the matching substring
|
| + - toks = a list of the matched tokens, packaged as a C{L{ParseResults}} object
|
| + If the functions in fns modify the tokens, they can return them as the return
|
| + value from fn, and the modified list of tokens will replace the original.
|
| + Otherwise, fn does not need to return any value.
|
| +
|
| + Optional keyword arguments:
|
| + - callDuringTry = (default=C{False}) indicate if parse action should be run during lookaheads and alternate testing
|
| +
|
| + Note: the default parsing behavior is to expand tabs in the input string
|
| + before starting the parsing process. See L{I{parseString}<parseString>} for more information
|
| + on parsing strings containing C{<TAB>}s, and suggested methods to maintain a
|
| + consistent view of the parsed string, the parse location, and line and column
|
| + positions within the parsed string.
|
| +
|
| + Example::
|
| + integer = Word(nums)
|
| + date_str = integer + '/' + integer + '/' + integer
|
| +
|
| + date_str.parseString("1999/12/31") # -> ['1999', '/', '12', '/', '31']
|
| +
|
| + # use parse action to convert to ints at parse time
|
| + integer = Word(nums).setParseAction(lambda toks: int(toks[0]))
|
| + date_str = integer + '/' + integer + '/' + integer
|
| +
|
| + # note that integer fields are now ints, not strings
|
| + date_str.parseString("1999/12/31") # -> [1999, '/', 12, '/', 31]
|
| + """
|
| + self.parseAction = list(map(_trim_arity, list(fns)))
|
| + self.callDuringTry = kwargs.get("callDuringTry", False)
|
| + return self
|
| +
|
| + def addParseAction( self, *fns, **kwargs ):
|
| + """
|
| + Add parse action to expression's list of parse actions. See L{I{setParseAction}<setParseAction>}.
|
| +
|
| + See examples in L{I{copy}<copy>}.
|
| + """
|
| + self.parseAction += list(map(_trim_arity, list(fns)))
|
| + self.callDuringTry = self.callDuringTry or kwargs.get("callDuringTry", False)
|
| + return self
|
| +
|
| + def addCondition(self, *fns, **kwargs):
|
| + """Add a boolean predicate function to expression's list of parse actions. See
|
| + L{I{setParseAction}<setParseAction>} for function call signatures. Unlike C{setParseAction},
|
| + functions passed to C{addCondition} need to return boolean success/fail of the condition.
|
| +
|
| + Optional keyword arguments:
|
| + - message = define a custom message to be used in the raised exception
|
| + - fatal = if True, will raise ParseFatalException to stop parsing immediately; otherwise will raise ParseException
|
| +
|
| + Example::
|
| + integer = Word(nums).setParseAction(lambda toks: int(toks[0]))
|
| + year_int = integer.copy()
|
| + year_int.addCondition(lambda toks: toks[0] >= 2000, message="Only support years 2000 and later")
|
| + date_str = year_int + '/' + integer + '/' + integer
|
| +
|
| + result = date_str.parseString("1999/12/31") # -> Exception: Only support years 2000 and later (at char 0), (line:1, col:1)
|
| + """
|
| + msg = kwargs.get("message", "failed user-defined condition")
|
| + exc_type = ParseFatalException if kwargs.get("fatal", False) else ParseException
|
| + for fn in fns:
|
| + def pa(s,l,t):
|
| + if not bool(_trim_arity(fn)(s,l,t)):
|
| + raise exc_type(s,l,msg)
|
| + self.parseAction.append(pa)
|
| + self.callDuringTry = self.callDuringTry or kwargs.get("callDuringTry", False)
|
| + return self
|
| +
|
| + def setFailAction( self, fn ):
|
| + """Define action to perform if parsing fails at this expression.
|
| + Fail acton fn is a callable function that takes the arguments
|
| + C{fn(s,loc,expr,err)} where:
|
| + - s = string being parsed
|
| + - loc = location where expression match was attempted and failed
|
| + - expr = the parse expression that failed
|
| + - err = the exception thrown
|
| + The function returns no value. It may throw C{L{ParseFatalException}}
|
| + if it is desired to stop parsing immediately."""
|
| + self.failAction = fn
|
| + return self
|
| +
|
| + def _skipIgnorables( self, instring, loc ):
|
| + exprsFound = True
|
| + while exprsFound:
|
| + exprsFound = False
|
| + for e in self.ignoreExprs:
|
| + try:
|
| + while 1:
|
| + loc,dummy = e._parse( instring, loc )
|
| + exprsFound = True
|
| + except ParseException:
|
| + pass
|
| + return loc
|
| +
|
| + def preParse( self, instring, loc ):
|
| + if self.ignoreExprs:
|
| + loc = self._skipIgnorables( instring, loc )
|
| +
|
| + if self.skipWhitespace:
|
| + wt = self.whiteChars
|
| + instrlen = len(instring)
|
| + while loc < instrlen and instring[loc] in wt:
|
| + loc += 1
|
| +
|
| + return loc
|
| +
|
| + def parseImpl( self, instring, loc, doActions=True ):
|
| + return loc, []
|
| +
|
| + def postParse( self, instring, loc, tokenlist ):
|
| + return tokenlist
|
| +
|
| + #~ @profile
|
| + def _parseNoCache( self, instring, loc, doActions=True, callPreParse=True ):
|
| + debugging = ( self.debug ) #and doActions )
|
| +
|
| + if debugging or self.failAction:
|
| + #~ print ("Match",self,"at loc",loc,"(%d,%d)" % ( lineno(loc,instring), col(loc,instring) ))
|
| + if (self.debugActions[0] ):
|
| + self.debugActions[0]( instring, loc, self )
|
| + if callPreParse and self.callPreparse:
|
| + preloc = self.preParse( instring, loc )
|
| + else:
|
| + preloc = loc
|
| + tokensStart = preloc
|
| + try:
|
| + try:
|
| + loc,tokens = self.parseImpl( instring, preloc, doActions )
|
| + except IndexError:
|
| + raise ParseException( instring, len(instring), self.errmsg, self )
|
| + except ParseBaseException as err:
|
| + #~ print ("Exception raised:", err)
|
| + if self.debugActions[2]:
|
| + self.debugActions[2]( instring, tokensStart, self, err )
|
| + if self.failAction:
|
| + self.failAction( instring, tokensStart, self, err )
|
| + raise
|
| + else:
|
| + if callPreParse and self.callPreparse:
|
| + preloc = self.preParse( instring, loc )
|
| + else:
|
| + preloc = loc
|
| + tokensStart = preloc
|
| + if self.mayIndexError or loc >= len(instring):
|
| + try:
|
| + loc,tokens = self.parseImpl( instring, preloc, doActions )
|
| + except IndexError:
|
| + raise ParseException( instring, len(instring), self.errmsg, self )
|
| + else:
|
| + loc,tokens = self.parseImpl( instring, preloc, doActions )
|
| +
|
| + tokens = self.postParse( instring, loc, tokens )
|
| +
|
| + retTokens = ParseResults( tokens, self.resultsName, asList=self.saveAsList, modal=self.modalResults )
|
| + if self.parseAction and (doActions or self.callDuringTry):
|
| + if debugging:
|
| + try:
|
| + for fn in self.parseAction:
|
| + tokens = fn( instring, tokensStart, retTokens )
|
| + if tokens is not None:
|
| + retTokens = ParseResults( tokens,
|
| + self.resultsName,
|
| + asList=self.saveAsList and isinstance(tokens,(ParseResults,list)),
|
| + modal=self.modalResults )
|
| + except ParseBaseException as err:
|
| + #~ print "Exception raised in user parse action:", err
|
| + if (self.debugActions[2] ):
|
| + self.debugActions[2]( instring, tokensStart, self, err )
|
| + raise
|
| + else:
|
| + for fn in self.parseAction:
|
| + tokens = fn( instring, tokensStart, retTokens )
|
| + if tokens is not None:
|
| + retTokens = ParseResults( tokens,
|
| + self.resultsName,
|
| + asList=self.saveAsList and isinstance(tokens,(ParseResults,list)),
|
| + modal=self.modalResults )
|
| +
|
| + if debugging:
|
| + #~ print ("Matched",self,"->",retTokens.asList())
|
| + if (self.debugActions[1] ):
|
| + self.debugActions[1]( instring, tokensStart, loc, self, retTokens )
|
| +
|
| + return loc, retTokens
|
| +
|
| + def tryParse( self, instring, loc ):
|
| + try:
|
| + return self._parse( instring, loc, doActions=False )[0]
|
| + except ParseFatalException:
|
| + raise ParseException( instring, loc, self.errmsg, self)
|
| +
|
| + def canParseNext(self, instring, loc):
|
| + try:
|
| + self.tryParse(instring, loc)
|
| + except (ParseException, IndexError):
|
| + return False
|
| + else:
|
| + return True
|
| +
|
| + class _UnboundedCache(object):
|
| + def __init__(self):
|
| + cache = {}
|
| + self.not_in_cache = not_in_cache = object()
|
| +
|
| + def get(self, key):
|
| + return cache.get(key, not_in_cache)
|
| +
|
| + def set(self, key, value):
|
| + cache[key] = value
|
| +
|
| + def clear(self):
|
| + cache.clear()
|
| +
|
| + self.get = types.MethodType(get, self)
|
| + self.set = types.MethodType(set, self)
|
| + self.clear = types.MethodType(clear, self)
|
| +
|
| + if _OrderedDict is not None:
|
| + class _FifoCache(object):
|
| + def __init__(self, size):
|
| + self.not_in_cache = not_in_cache = object()
|
| +
|
| + cache = _OrderedDict()
|
| +
|
| + def get(self, key):
|
| + return cache.get(key, not_in_cache)
|
| +
|
| + def set(self, key, value):
|
| + cache[key] = value
|
| + if len(cache) > size:
|
| + cache.popitem(False)
|
| +
|
| + def clear(self):
|
| + cache.clear()
|
| +
|
| + self.get = types.MethodType(get, self)
|
| + self.set = types.MethodType(set, self)
|
| + self.clear = types.MethodType(clear, self)
|
| +
|
| + else:
|
| + class _FifoCache(object):
|
| + def __init__(self, size):
|
| + self.not_in_cache = not_in_cache = object()
|
| +
|
| + cache = {}
|
| + key_fifo = collections.deque([], size)
|
| +
|
| + def get(self, key):
|
| + return cache.get(key, not_in_cache)
|
| +
|
| + def set(self, key, value):
|
| + cache[key] = value
|
| + if len(cache) > size:
|
| + cache.pop(key_fifo.popleft(), None)
|
| + key_fifo.append(key)
|
| +
|
| + def clear(self):
|
| + cache.clear()
|
| + key_fifo.clear()
|
| +
|
| + self.get = types.MethodType(get, self)
|
| + self.set = types.MethodType(set, self)
|
| + self.clear = types.MethodType(clear, self)
|
| +
|
| + # argument cache for optimizing repeated calls when backtracking through recursive expressions
|
| + packrat_cache = {} # this is set later by enabledPackrat(); this is here so that resetCache() doesn't fail
|
| + packrat_cache_lock = RLock()
|
| + packrat_cache_stats = [0, 0]
|
| +
|
| + # this method gets repeatedly called during backtracking with the same arguments -
|
| + # we can cache these arguments and save ourselves the trouble of re-parsing the contained expression
|
| + def _parseCache( self, instring, loc, doActions=True, callPreParse=True ):
|
| + HIT, MISS = 0, 1
|
| + lookup = (self, instring, loc, callPreParse, doActions)
|
| + with ParserElement.packrat_cache_lock:
|
| + cache = ParserElement.packrat_cache
|
| + value = cache.get(lookup)
|
| + if value is cache.not_in_cache:
|
| + ParserElement.packrat_cache_stats[MISS] += 1
|
| + try:
|
| + value = self._parseNoCache(instring, loc, doActions, callPreParse)
|
| + except ParseBaseException as pe:
|
| + # cache a copy of the exception, without the traceback
|
| + cache.set(lookup, pe.__class__(*pe.args))
|
| + raise
|
| + else:
|
| + cache.set(lookup, (value[0], value[1].copy()))
|
| + return value
|
| + else:
|
| + ParserElement.packrat_cache_stats[HIT] += 1
|
| + if isinstance(value, Exception):
|
| + raise value
|
| + return (value[0], value[1].copy())
|
| +
|
| + _parse = _parseNoCache
|
| +
|
| + @staticmethod
|
| + def resetCache():
|
| + ParserElement.packrat_cache.clear()
|
| + ParserElement.packrat_cache_stats[:] = [0] * len(ParserElement.packrat_cache_stats)
|
| +
|
| + _packratEnabled = False
|
| + @staticmethod
|
| + def enablePackrat(cache_size_limit=128):
|
| + """Enables "packrat" parsing, which adds memoizing to the parsing logic.
|
| + Repeated parse attempts at the same string location (which happens
|
| + often in many complex grammars) can immediately return a cached value,
|
| + instead of re-executing parsing/validating code. Memoizing is done of
|
| + both valid results and parsing exceptions.
|
| +
|
| + Parameters:
|
| + - cache_size_limit - (default=C{128}) - if an integer value is provided
|
| + will limit the size of the packrat cache; if None is passed, then
|
| + the cache size will be unbounded; if 0 is passed, the cache will
|
| + be effectively disabled.
|
| +
|
| + This speedup may break existing programs that use parse actions that
|
| + have side-effects. For this reason, packrat parsing is disabled when
|
| + you first import pyparsing. To activate the packrat feature, your
|
| + program must call the class method C{ParserElement.enablePackrat()}. If
|
| + your program uses C{psyco} to "compile as you go", you must call
|
| + C{enablePackrat} before calling C{psyco.full()}. If you do not do this,
|
| + Python will crash. For best results, call C{enablePackrat()} immediately
|
| + after importing pyparsing.
|
| +
|
| + Example::
|
| + import pyparsing
|
| + pyparsing.ParserElement.enablePackrat()
|
| + """
|
| + if not ParserElement._packratEnabled:
|
| + ParserElement._packratEnabled = True
|
| + if cache_size_limit is None:
|
| + ParserElement.packrat_cache = ParserElement._UnboundedCache()
|
| + else:
|
| + ParserElement.packrat_cache = ParserElement._FifoCache(cache_size_limit)
|
| + ParserElement._parse = ParserElement._parseCache
|
| +
|
| + def parseString( self, instring, parseAll=False ):
|
| + """
|
| + Execute the parse expression with the given string.
|
| + This is the main interface to the client code, once the complete
|
| + expression has been built.
|
| +
|
| + If you want the grammar to require that the entire input string be
|
| + successfully parsed, then set C{parseAll} to True (equivalent to ending
|
| + the grammar with C{L{StringEnd()}}).
|
| +
|
| + Note: C{parseString} implicitly calls C{expandtabs()} on the input string,
|
| + in order to report proper column numbers in parse actions.
|
| + If the input string contains tabs and
|
| + the grammar uses parse actions that use the C{loc} argument to index into the
|
| + string being parsed, you can ensure you have a consistent view of the input
|
| + string by:
|
| + - calling C{parseWithTabs} on your grammar before calling C{parseString}
|
| + (see L{I{parseWithTabs}<parseWithTabs>})
|
| + - define your parse action using the full C{(s,loc,toks)} signature, and
|
| + reference the input string using the parse action's C{s} argument
|
| + - explictly expand the tabs in your input string before calling
|
| + C{parseString}
|
| +
|
| + Example::
|
| + Word('a').parseString('aaaaabaaa') # -> ['aaaaa']
|
| + Word('a').parseString('aaaaabaaa', parseAll=True) # -> Exception: Expected end of text
|
| + """
|
| + ParserElement.resetCache()
|
| + if not self.streamlined:
|
| + self.streamline()
|
| + #~ self.saveAsList = True
|
| + for e in self.ignoreExprs:
|
| + e.streamline()
|
| + if not self.keepTabs:
|
| + instring = instring.expandtabs()
|
| + try:
|
| + loc, tokens = self._parse( instring, 0 )
|
| + if parseAll:
|
| + loc = self.preParse( instring, loc )
|
| + se = Empty() + StringEnd()
|
| + se._parse( instring, loc )
|
| + except ParseBaseException as exc:
|
| + if ParserElement.verbose_stacktrace:
|
| + raise
|
| + else:
|
| + # catch and re-raise exception from here, clears out pyparsing internal stack trace
|
| + raise exc
|
| + else:
|
| + return tokens
|
| +
|
| + def scanString( self, instring, maxMatches=_MAX_INT, overlap=False ):
|
| + """
|
| + Scan the input string for expression matches. Each match will return the
|
| + matching tokens, start location, and end location. May be called with optional
|
| + C{maxMatches} argument, to clip scanning after 'n' matches are found. If
|
| + C{overlap} is specified, then overlapping matches will be reported.
|
| +
|
| + Note that the start and end locations are reported relative to the string
|
| + being parsed. See L{I{parseString}<parseString>} for more information on parsing
|
| + strings with embedded tabs.
|
| +
|
| + Example::
|
| + source = "sldjf123lsdjjkf345sldkjf879lkjsfd987"
|
| + print(source)
|
| + for tokens,start,end in Word(alphas).scanString(source):
|
| + print(' '*start + '^'*(end-start))
|
| + print(' '*start + tokens[0])
|
| +
|
| + prints::
|
| +
|
| + sldjf123lsdjjkf345sldkjf879lkjsfd987
|
| + ^^^^^
|
| + sldjf
|
| + ^^^^^^^
|
| + lsdjjkf
|
| + ^^^^^^
|
| + sldkjf
|
| + ^^^^^^
|
| + lkjsfd
|
| + """
|
| + if not self.streamlined:
|
| + self.streamline()
|
| + for e in self.ignoreExprs:
|
| + e.streamline()
|
| +
|
| + if not self.keepTabs:
|
| + instring = _ustr(instring).expandtabs()
|
| + instrlen = len(instring)
|
| + loc = 0
|
| + preparseFn = self.preParse
|
| + parseFn = self._parse
|
| + ParserElement.resetCache()
|
| + matches = 0
|
| + try:
|
| + while loc <= instrlen and matches < maxMatches:
|
| + try:
|
| + preloc = preparseFn( instring, loc )
|
| + nextLoc,tokens = parseFn( instring, preloc, callPreParse=False )
|
| + except ParseException:
|
| + loc = preloc+1
|
| + else:
|
| + if nextLoc > loc:
|
| + matches += 1
|
| + yield tokens, preloc, nextLoc
|
| + if overlap:
|
| + nextloc = preparseFn( instring, loc )
|
| + if nextloc > loc:
|
| + loc = nextLoc
|
| + else:
|
| + loc += 1
|
| + else:
|
| + loc = nextLoc
|
| + else:
|
| + loc = preloc+1
|
| + except ParseBaseException as exc:
|
| + if ParserElement.verbose_stacktrace:
|
| + raise
|
| + else:
|
| + # catch and re-raise exception from here, clears out pyparsing internal stack trace
|
| + raise exc
|
| +
|
| + def transformString( self, instring ):
|
| + """
|
| + Extension to C{L{scanString}}, to modify matching text with modified tokens that may
|
| + be returned from a parse action. To use C{transformString}, define a grammar and
|
| + attach a parse action to it that modifies the returned token list.
|
| + Invoking C{transformString()} on a target string will then scan for matches,
|
| + and replace the matched text patterns according to the logic in the parse
|
| + action. C{transformString()} returns the resulting transformed string.
|
| +
|
| + Example::
|
| + wd = Word(alphas)
|
| + wd.setParseAction(lambda toks: toks[0].title())
|
| +
|
| + print(wd.transformString("now is the winter of our discontent made glorious summer by this sun of york."))
|
| + Prints::
|
| + Now Is The Winter Of Our Discontent Made Glorious Summer By This Sun Of York.
|
| + """
|
| + out = []
|
| + lastE = 0
|
| + # force preservation of <TAB>s, to minimize unwanted transformation of string, and to
|
| + # keep string locs straight between transformString and scanString
|
| + self.keepTabs = True
|
| + try:
|
| + for t,s,e in self.scanString( instring ):
|
| + out.append( instring[lastE:s] )
|
| + if t:
|
| + if isinstance(t,ParseResults):
|
| + out += t.asList()
|
| + elif isinstance(t,list):
|
| + out += t
|
| + else:
|
| + out.append(t)
|
| + lastE = e
|
| + out.append(instring[lastE:])
|
| + out = [o for o in out if o]
|
| + return "".join(map(_ustr,_flatten(out)))
|
| + except ParseBaseException as exc:
|
| + if ParserElement.verbose_stacktrace:
|
| + raise
|
| + else:
|
| + # catch and re-raise exception from here, clears out pyparsing internal stack trace
|
| + raise exc
|
| +
|
| + def searchString( self, instring, maxMatches=_MAX_INT ):
|
| + """
|
| + Another extension to C{L{scanString}}, simplifying the access to the tokens found
|
| + to match the given parse expression. May be called with optional
|
| + C{maxMatches} argument, to clip searching after 'n' matches are found.
|
| +
|
| + Example::
|
| + # a capitalized word starts with an uppercase letter, followed by zero or more lowercase letters
|
| + cap_word = Word(alphas.upper(), alphas.lower())
|
| +
|
| + print(cap_word.searchString("More than Iron, more than Lead, more than Gold I need Electricity"))
|
| + prints::
|
| + ['More', 'Iron', 'Lead', 'Gold', 'I']
|
| + """
|
| + try:
|
| + return ParseResults([ t for t,s,e in self.scanString( instring, maxMatches ) ])
|
| + except ParseBaseException as exc:
|
| + if ParserElement.verbose_stacktrace:
|
| + raise
|
| + else:
|
| + # catch and re-raise exception from here, clears out pyparsing internal stack trace
|
| + raise exc
|
| +
|
| + def split(self, instring, maxsplit=_MAX_INT, includeSeparators=False):
|
| + """
|
| + Generator method to split a string using the given expression as a separator.
|
| + May be called with optional C{maxsplit} argument, to limit the number of splits;
|
| + and the optional C{includeSeparators} argument (default=C{False}), if the separating
|
| + matching text should be included in the split results.
|
| +
|
| + Example::
|
| + punc = oneOf(list(".,;:/-!?"))
|
| + print(list(punc.split("This, this?, this sentence, is badly punctuated!")))
|
| + prints::
|
| + ['This', ' this', '', ' this sentence', ' is badly punctuated', '']
|
| + """
|
| + splits = 0
|
| + last = 0
|
| + for t,s,e in self.scanString(instring, maxMatches=maxsplit):
|
| + yield instring[last:s]
|
| + if includeSeparators:
|
| + yield t[0]
|
| + last = e
|
| + yield instring[last:]
|
| +
|
| + def __add__(self, other ):
|
| + """
|
| + Implementation of + operator - returns C{L{And}}. Adding strings to a ParserElement
|
| + converts them to L{Literal}s by default.
|
| +
|
| + Example::
|
| + greet = Word(alphas) + "," + Word(alphas) + "!"
|
| + hello = "Hello, World!"
|
| + print (hello, "->", greet.parseString(hello))
|
| + Prints::
|
| + Hello, World! -> ['Hello', ',', 'World', '!']
|
| + """
|
| + if isinstance( other, basestring ):
|
| + other = ParserElement._literalStringClass( other )
|
| + if not isinstance( other, ParserElement ):
|
| + warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
|
| + SyntaxWarning, stacklevel=2)
|
| + return None
|
| + return And( [ self, other ] )
|
| +
|
| + def __radd__(self, other ):
|
| + """
|
| + Implementation of + operator when left operand is not a C{L{ParserElement}}
|
| + """
|
| + if isinstance( other, basestring ):
|
| + other = ParserElement._literalStringClass( other )
|
| + if not isinstance( other, ParserElement ):
|
| + warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
|
| + SyntaxWarning, stacklevel=2)
|
| + return None
|
| + return other + self
|
| +
|
| + def __sub__(self, other):
|
| + """
|
| + Implementation of - operator, returns C{L{And}} with error stop
|
| + """
|
| + if isinstance( other, basestring ):
|
| + other = ParserElement._literalStringClass( other )
|
| + if not isinstance( other, ParserElement ):
|
| + warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
|
| + SyntaxWarning, stacklevel=2)
|
| + return None
|
| + return And( [ self, And._ErrorStop(), other ] )
|
| +
|
| + def __rsub__(self, other ):
|
| + """
|
| + Implementation of - operator when left operand is not a C{L{ParserElement}}
|
| + """
|
| + if isinstance( other, basestring ):
|
| + other = ParserElement._literalStringClass( other )
|
| + if not isinstance( other, ParserElement ):
|
| + warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
|
| + SyntaxWarning, stacklevel=2)
|
| + return None
|
| + return other - self
|
| +
|
| + def __mul__(self,other):
|
| + """
|
| + Implementation of * operator, allows use of C{expr * 3} in place of
|
| + C{expr + expr + expr}. Expressions may also me multiplied by a 2-integer
|
| + tuple, similar to C{{min,max}} multipliers in regular expressions. Tuples
|
| + may also include C{None} as in:
|
| + - C{expr*(n,None)} or C{expr*(n,)} is equivalent
|
| + to C{expr*n + L{ZeroOrMore}(expr)}
|
| + (read as "at least n instances of C{expr}")
|
| + - C{expr*(None,n)} is equivalent to C{expr*(0,n)}
|
| + (read as "0 to n instances of C{expr}")
|
| + - C{expr*(None,None)} is equivalent to C{L{ZeroOrMore}(expr)}
|
| + - C{expr*(1,None)} is equivalent to C{L{OneOrMore}(expr)}
|
| +
|
| + Note that C{expr*(None,n)} does not raise an exception if
|
| + more than n exprs exist in the input stream; that is,
|
| + C{expr*(None,n)} does not enforce a maximum number of expr
|
| + occurrences. If this behavior is desired, then write
|
| + C{expr*(None,n) + ~expr}
|
| + """
|
| + if isinstance(other,int):
|
| + minElements, optElements = other,0
|
| + elif isinstance(other,tuple):
|
| + other = (other + (None, None))[:2]
|
| + if other[0] is None:
|
| + other = (0, other[1])
|
| + if isinstance(other[0],int) and other[1] is None:
|
| + if other[0] == 0:
|
| + return ZeroOrMore(self)
|
| + if other[0] == 1:
|
| + return OneOrMore(self)
|
| + else:
|
| + return self*other[0] + ZeroOrMore(self)
|
| + elif isinstance(other[0],int) and isinstance(other[1],int):
|
| + minElements, optElements = other
|
| + optElements -= minElements
|
| + else:
|
| + raise TypeError("cannot multiply 'ParserElement' and ('%s','%s') objects", type(other[0]),type(other[1]))
|
| + else:
|
| + raise TypeError("cannot multiply 'ParserElement' and '%s' objects", type(other))
|
| +
|
| + if minElements < 0:
|
| + raise ValueError("cannot multiply ParserElement by negative value")
|
| + if optElements < 0:
|
| + raise ValueError("second tuple value must be greater or equal to first tuple value")
|
| + if minElements == optElements == 0:
|
| + raise ValueError("cannot multiply ParserElement by 0 or (0,0)")
|
| +
|
| + if (optElements):
|
| + def makeOptionalList(n):
|
| + if n>1:
|
| + return Optional(self + makeOptionalList(n-1))
|
| + else:
|
| + return Optional(self)
|
| + if minElements:
|
| + if minElements == 1:
|
| + ret = self + makeOptionalList(optElements)
|
| + else:
|
| + ret = And([self]*minElements) + makeOptionalList(optElements)
|
| + else:
|
| + ret = makeOptionalList(optElements)
|
| + else:
|
| + if minElements == 1:
|
| + ret = self
|
| + else:
|
| + ret = And([self]*minElements)
|
| + return ret
|
| +
|
| + def __rmul__(self, other):
|
| + return self.__mul__(other)
|
| +
|
| + def __or__(self, other ):
|
| + """
|
| + Implementation of | operator - returns C{L{MatchFirst}}
|
| + """
|
| + if isinstance( other, basestring ):
|
| + other = ParserElement._literalStringClass( other )
|
| + if not isinstance( other, ParserElement ):
|
| + warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
|
| + SyntaxWarning, stacklevel=2)
|
| + return None
|
| + return MatchFirst( [ self, other ] )
|
| +
|
| + def __ror__(self, other ):
|
| + """
|
| + Implementation of | operator when left operand is not a C{L{ParserElement}}
|
| + """
|
| + if isinstance( other, basestring ):
|
| + other = ParserElement._literalStringClass( other )
|
| + if not isinstance( other, ParserElement ):
|
| + warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
|
| + SyntaxWarning, stacklevel=2)
|
| + return None
|
| + return other | self
|
| +
|
| + def __xor__(self, other ):
|
| + """
|
| + Implementation of ^ operator - returns C{L{Or}}
|
| + """
|
| + if isinstance( other, basestring ):
|
| + other = ParserElement._literalStringClass( other )
|
| + if not isinstance( other, ParserElement ):
|
| + warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
|
| + SyntaxWarning, stacklevel=2)
|
| + return None
|
| + return Or( [ self, other ] )
|
| +
|
| + def __rxor__(self, other ):
|
| + """
|
| + Implementation of ^ operator when left operand is not a C{L{ParserElement}}
|
| + """
|
| + if isinstance( other, basestring ):
|
| + other = ParserElement._literalStringClass( other )
|
| + if not isinstance( other, ParserElement ):
|
| + warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
|
| + SyntaxWarning, stacklevel=2)
|
| + return None
|
| + return other ^ self
|
| +
|
| + def __and__(self, other ):
|
| + """
|
| + Implementation of & operator - returns C{L{Each}}
|
| + """
|
| + if isinstance( other, basestring ):
|
| + other = ParserElement._literalStringClass( other )
|
| + if not isinstance( other, ParserElement ):
|
| + warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
|
| + SyntaxWarning, stacklevel=2)
|
| + return None
|
| + return Each( [ self, other ] )
|
| +
|
| + def __rand__(self, other ):
|
| + """
|
| + Implementation of & operator when left operand is not a C{L{ParserElement}}
|
| + """
|
| + if isinstance( other, basestring ):
|
| + other = ParserElement._literalStringClass( other )
|
| + if not isinstance( other, ParserElement ):
|
| + warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
|
| + SyntaxWarning, stacklevel=2)
|
| + return None
|
| + return other & self
|
| +
|
| + def __invert__( self ):
|
| + """
|
| + Implementation of ~ operator - returns C{L{NotAny}}
|
| + """
|
| + return NotAny( self )
|
| +
|
| + def __call__(self, name=None):
|
| + """
|
| + Shortcut for C{L{setResultsName}}, with C{listAllMatches=False}.
|
| +
|
| + If C{name} is given with a trailing C{'*'} character, then C{listAllMatches} will be
|
| + passed as C{True}.
|
| +
|
| + If C{name} is omitted, same as calling C{L{copy}}.
|
| +
|
| + Example::
|
| + # these are equivalent
|
| + userdata = Word(alphas).setResultsName("name") + Word(nums+"-").setResultsName("socsecno")
|
| + userdata = Word(alphas)("name") + Word(nums+"-")("socsecno")
|
| + """
|
| + if name is not None:
|
| + return self.setResultsName(name)
|
| + else:
|
| + return self.copy()
|
| +
|
| + def suppress( self ):
|
| + """
|
| + Suppresses the output of this C{ParserElement}; useful to keep punctuation from
|
| + cluttering up returned output.
|
| + """
|
| + return Suppress( self )
|
| +
|
| + def leaveWhitespace( self ):
|
| + """
|
| + Disables the skipping of whitespace before matching the characters in the
|
| + C{ParserElement}'s defined pattern. This is normally only used internally by
|
| + the pyparsing module, but may be needed in some whitespace-sensitive grammars.
|
| + """
|
| + self.skipWhitespace = False
|
| + return self
|
| +
|
| + def setWhitespaceChars( self, chars ):
|
| + """
|
| + Overrides the default whitespace chars
|
| + """
|
| + self.skipWhitespace = True
|
| + self.whiteChars = chars
|
| + self.copyDefaultWhiteChars = False
|
| + return self
|
| +
|
| + def parseWithTabs( self ):
|
| + """
|
| + Overrides default behavior to expand C{<TAB>}s to spaces before parsing the input string.
|
| + Must be called before C{parseString} when the input grammar contains elements that
|
| + match C{<TAB>} characters.
|
| + """
|
| + self.keepTabs = True
|
| + return self
|
| +
|
| + def ignore( self, other ):
|
| + """
|
| + Define expression to be ignored (e.g., comments) while doing pattern
|
| + matching; may be called repeatedly, to define multiple comment or other
|
| + ignorable patterns.
|
| +
|
| + Example::
|
| + patt = OneOrMore(Word(alphas))
|
| + patt.parseString('ablaj /* comment */ lskjd') # -> ['ablaj']
|
| +
|
| + patt.ignore(cStyleComment)
|
| + patt.parseString('ablaj /* comment */ lskjd') # -> ['ablaj', 'lskjd']
|
| + """
|
| + if isinstance(other, basestring):
|
| + other = Suppress(other)
|
| +
|
| + if isinstance( other, Suppress ):
|
| + if other not in self.ignoreExprs:
|
| + self.ignoreExprs.append(other)
|
| + else:
|
| + self.ignoreExprs.append( Suppress( other.copy() ) )
|
| + return self
|
| +
|
| + def setDebugActions( self, startAction, successAction, exceptionAction ):
|
| + """
|
| + Enable display of debugging messages while doing pattern matching.
|
| + """
|
| + self.debugActions = (startAction or _defaultStartDebugAction,
|
| + successAction or _defaultSuccessDebugAction,
|
| + exceptionAction or _defaultExceptionDebugAction)
|
| + self.debug = True
|
| + return self
|
| +
|
| + def setDebug( self, flag=True ):
|
| + """
|
| + Enable display of debugging messages while doing pattern matching.
|
| + Set C{flag} to True to enable, False to disable.
|
| +
|
| + Example::
|
| + wd = Word(alphas).setName("alphaword")
|
| + integer = Word(nums).setName("numword")
|
| + term = wd | integer
|
| +
|
| + # turn on debugging for wd
|
| + wd.setDebug()
|
| +
|
| + OneOrMore(term).parseString("abc 123 xyz 890")
|
| +
|
| + prints::
|
| + Match alphaword at loc 0(1,1)
|
| + Matched alphaword -> ['abc']
|
| + Match alphaword at loc 3(1,4)
|
| + Exception raised:Expected alphaword (at char 4), (line:1, col:5)
|
| + Match alphaword at loc 7(1,8)
|
| + Matched alphaword -> ['xyz']
|
| + Match alphaword at loc 11(1,12)
|
| + Exception raised:Expected alphaword (at char 12), (line:1, col:13)
|
| + Match alphaword at loc 15(1,16)
|
| + Exception raised:Expected alphaword (at char 15), (line:1, col:16)
|
| +
|
| + The output shown is that produced by the default debug actions - custom debug actions can be
|
| + specified using L{setDebugActions}. Prior to attempting
|
| + to match the C{wd} expression, the debugging message C{"Match <exprname> at loc <n>(<line>,<col>)"}
|
| + is shown. Then if the parse succeeds, a C{"Matched"} message is shown, or an C{"Exception raised"}
|
| + message is shown. Also note the use of L{setName} to assign a human-readable name to the expression,
|
| + which makes debugging and exception messages easier to understand - for instance, the default
|
| + name created for the C{Word} expression without calling C{setName} is C{"W:(ABCD...)"}.
|
| + """
|
| + if flag:
|
| + self.setDebugActions( _defaultStartDebugAction, _defaultSuccessDebugAction, _defaultExceptionDebugAction )
|
| + else:
|
| + self.debug = False
|
| + return self
|
| +
|
| + def __str__( self ):
|
| + return self.name
|
| +
|
| + def __repr__( self ):
|
| + return _ustr(self)
|
| +
|
| + def streamline( self ):
|
| + self.streamlined = True
|
| + self.strRepr = None
|
| + return self
|
| +
|
| + def checkRecursion( self, parseElementList ):
|
| + pass
|
| +
|
| + def validate( self, validateTrace=[] ):
|
| + """
|
| + Check defined expressions for valid structure, check for infinite recursive definitions.
|
| + """
|
| + self.checkRecursion( [] )
|
| +
|
| + def parseFile( self, file_or_filename, parseAll=False ):
|
| + """
|
| + Execute the parse expression on the given file or filename.
|
| + If a filename is specified (instead of a file object),
|
| + the entire file is opened, read, and closed before parsing.
|
| + """
|
| + try:
|
| + file_contents = file_or_filename.read()
|
| + except AttributeError:
|
| + with open(file_or_filename, "r") as f:
|
| + file_contents = f.read()
|
| + try:
|
| + return self.parseString(file_contents, parseAll)
|
| + except ParseBaseException as exc:
|
| + if ParserElement.verbose_stacktrace:
|
| + raise
|
| + else:
|
| + # catch and re-raise exception from here, clears out pyparsing internal stack trace
|
| + raise exc
|
| +
|
| + def __eq__(self,other):
|
| + if isinstance(other, ParserElement):
|
| + return self is other or vars(self) == vars(other)
|
| + elif isinstance(other, basestring):
|
| + return self.matches(other)
|
| + else:
|
| + return super(ParserElement,self)==other
|
| +
|
| + def __ne__(self,other):
|
| + return not (self == other)
|
| +
|
| + def __hash__(self):
|
| + return hash(id(self))
|
| +
|
| + def __req__(self,other):
|
| + return self == other
|
| +
|
| + def __rne__(self,other):
|
| + return not (self == other)
|
| +
|
| + def matches(self, testString, parseAll=True):
|
| + """
|
| + Method for quick testing of a parser against a test string. Good for simple
|
| + inline microtests of sub expressions while building up larger parser.
|
| +
|
| + Parameters:
|
| + - testString - to test against this expression for a match
|
| + - parseAll - (default=C{True}) - flag to pass to C{L{parseString}} when running tests
|
| +
|
| + Example::
|
| + expr = Word(nums)
|
| + assert expr.matches("100")
|
| + """
|
| + try:
|
| + self.parseString(_ustr(testString), parseAll=parseAll)
|
| + return True
|
| + except ParseBaseException:
|
| + return False
|
| +
|
| + def runTests(self, tests, parseAll=True, comment='#', fullDump=True, printResults=True, failureTests=False):
|
| + """
|
| + Execute the parse expression on a series of test strings, showing each
|
| + test, the parsed results or where the parse failed. Quick and easy way to
|
| + run a parse expression against a list of sample strings.
|
| +
|
| + Parameters:
|
| + - tests - a list of separate test strings, or a multiline string of test strings
|
| + - parseAll - (default=C{True}) - flag to pass to C{L{parseString}} when running tests
|
| + - comment - (default=C{'#'}) - expression for indicating embedded comments in the test
|
| + string; pass None to disable comment filtering
|
| + - fullDump - (default=C{True}) - dump results as list followed by results names in nested outline;
|
| + if False, only dump nested list
|
| + - printResults - (default=C{True}) prints test output to stdout
|
| + - failureTests - (default=C{False}) indicates if these tests are expected to fail parsing
|
| +
|
| + Returns: a (success, results) tuple, where success indicates that all tests succeeded
|
| + (or failed if C{failureTests} is True), and the results contain a list of lines of each
|
| + test's output
|
| +
|
| + Example::
|
| + number_expr = pyparsing_common.number.copy()
|
| +
|
| + result = number_expr.runTests('''
|
| + # unsigned integer
|
| + 100
|
| + # negative integer
|
| + -100
|
| + # float with scientific notation
|
| + 6.02e23
|
| + # integer with scientific notation
|
| + 1e-12
|
| + ''')
|
| + print("Success" if result[0] else "Failed!")
|
| +
|
| + result = number_expr.runTests('''
|
| + # stray character
|
| + 100Z
|
| + # missing leading digit before '.'
|
| + -.100
|
| + # too many '.'
|
| + 3.14.159
|
| + ''', failureTests=True)
|
| + print("Success" if result[0] else "Failed!")
|
| + prints::
|
| + # unsigned integer
|
| + 100
|
| + [100]
|
| +
|
| + # negative integer
|
| + -100
|
| + [-100]
|
| +
|
| + # float with scientific notation
|
| + 6.02e23
|
| + [6.02e+23]
|
| +
|
| + # integer with scientific notation
|
| + 1e-12
|
| + [1e-12]
|
| +
|
| + Success
|
| +
|
| + # stray character
|
| + 100Z
|
| + ^
|
| + FAIL: Expected end of text (at char 3), (line:1, col:4)
|
| +
|
| + # missing leading digit before '.'
|
| + -.100
|
| + ^
|
| + FAIL: Expected {real number with scientific notation | real number | signed integer} (at char 0), (line:1, col:1)
|
| +
|
| + # too many '.'
|
| + 3.14.159
|
| + ^
|
| + FAIL: Expected end of text (at char 4), (line:1, col:5)
|
| +
|
| + Success
|
| +
|
| + Each test string must be on a single line. If you want to test a string that spans multiple
|
| + lines, create a test like this::
|
| +
|
| + expr.runTest(r"this is a test\\n of strings that spans \\n 3 lines")
|
| +
|
| + (Note that this is a raw string literal, you must include the leading 'r'.)
|
| + """
|
| + if isinstance(tests, basestring):
|
| + tests = list(map(str.strip, tests.rstrip().splitlines()))
|
| + if isinstance(comment, basestring):
|
| + comment = Literal(comment)
|
| + allResults = []
|
| + comments = []
|
| + success = True
|
| + for t in tests:
|
| + if comment is not None and comment.matches(t, False) or comments and not t:
|
| + comments.append(t)
|
| + continue
|
| + if not t:
|
| + continue
|
| + out = ['\n'.join(comments), t]
|
| + comments = []
|
| + try:
|
| + t = t.replace(r'\n','\n')
|
| + result = self.parseString(t, parseAll=parseAll)
|
| + out.append(result.dump(full=fullDump))
|
| + success = success and not failureTests
|
| + except ParseBaseException as pe:
|
| + fatal = "(FATAL)" if isinstance(pe, ParseFatalException) else ""
|
| + if '\n' in t:
|
| + out.append(line(pe.loc, t))
|
| + out.append(' '*(col(pe.loc,t)-1) + '^' + fatal)
|
| + else:
|
| + out.append(' '*pe.loc + '^' + fatal)
|
| + out.append("FAIL: " + str(pe))
|
| + success = success and failureTests
|
| + result = pe
|
| + except Exception as exc:
|
| + out.append("FAIL-EXCEPTION: " + str(exc))
|
| + success = success and failureTests
|
| + result = exc
|
| +
|
| + if printResults:
|
| + if fullDump:
|
| + out.append('')
|
| + print('\n'.join(out))
|
| +
|
| + allResults.append((t, result))
|
| +
|
| + return success, allResults
|
| +
|
| +
|
| +class Token(ParserElement):
|
| + """
|
| + Abstract C{ParserElement} subclass, for defining atomic matching patterns.
|
| + """
|
| + def __init__( self ):
|
| + super(Token,self).__init__( savelist=False )
|
| +
|
| +
|
| +class Empty(Token):
|
| + """
|
| + An empty token, will always match.
|
| + """
|
| + def __init__( self ):
|
| + super(Empty,self).__init__()
|
| + self.name = "Empty"
|
| + self.mayReturnEmpty = True
|
| + self.mayIndexError = False
|
| +
|
| +
|
| +class NoMatch(Token):
|
| + """
|
| + A token that will never match.
|
| + """
|
| + def __init__( self ):
|
| + super(NoMatch,self).__init__()
|
| + self.name = "NoMatch"
|
| + self.mayReturnEmpty = True
|
| + self.mayIndexError = False
|
| + self.errmsg = "Unmatchable token"
|
| +
|
| + def parseImpl( self, instring, loc, doActions=True ):
|
| + raise ParseException(instring, loc, self.errmsg, self)
|
| +
|
| +
|
| +class Literal(Token):
|
| + """
|
| + Token to exactly match a specified string.
|
| +
|
| + Example::
|
| + Literal('blah').parseString('blah') # -> ['blah']
|
| + Literal('blah').parseString('blahfooblah') # -> ['blah']
|
| + Literal('blah').parseString('bla') # -> Exception: Expected "blah"
|
| +
|
| + For case-insensitive matching, use L{CaselessLiteral}.
|
| +
|
| + For keyword matching (force word break before and after the matched string),
|
| + use L{Keyword} or L{CaselessKeyword}.
|
| + """
|
| + def __init__( self, matchString ):
|
| + super(Literal,self).__init__()
|
| + self.match = matchString
|
| + self.matchLen = len(matchString)
|
| + try:
|
| + self.firstMatchChar = matchString[0]
|
| + except IndexError:
|
| + warnings.warn("null string passed to Literal; use Empty() instead",
|
| + SyntaxWarning, stacklevel=2)
|
| + self.__class__ = Empty
|
| + self.name = '"%s"' % _ustr(self.match)
|
| + self.errmsg = "Expected " + self.name
|
| + self.mayReturnEmpty = False
|
| + self.mayIndexError = False
|
| +
|
| + # Performance tuning: this routine gets called a *lot*
|
| + # if this is a single character match string and the first character matches,
|
| + # short-circuit as quickly as possible, and avoid calling startswith
|
| + #~ @profile
|
| + def parseImpl( self, instring, loc, doActions=True ):
|
| + if (instring[loc] == self.firstMatchChar and
|
| + (self.matchLen==1 or instring.startswith(self.match,loc)) ):
|
| + return loc+self.matchLen, self.match
|
| + raise ParseException(instring, loc, self.errmsg, self)
|
| +_L = Literal
|
| +ParserElement._literalStringClass = Literal
|
| +
|
| +class Keyword(Token):
|
| + """
|
| + Token to exactly match a specified string as a keyword, that is, it must be
|
| + immediately followed by a non-keyword character. Compare with C{L{Literal}}:
|
| + - C{Literal("if")} will match the leading C{'if'} in C{'ifAndOnlyIf'}.
|
| + - C{Keyword("if")} will not; it will only match the leading C{'if'} in C{'if x=1'}, or C{'if(y==2)'}
|
| + Accepts two optional constructor arguments in addition to the keyword string:
|
| + - C{identChars} is a string of characters that would be valid identifier characters,
|
| + defaulting to all alphanumerics + "_" and "$"
|
| + - C{caseless} allows case-insensitive matching, default is C{False}.
|
| +
|
| + Example::
|
| + Keyword("start").parseString("start") # -> ['start']
|
| + Keyword("start").parseString("starting") # -> Exception
|
| +
|
| + For case-insensitive matching, use L{CaselessKeyword}.
|
| + """
|
| + DEFAULT_KEYWORD_CHARS = alphanums+"_$"
|
| +
|
| + def __init__( self, matchString, identChars=None, caseless=False ):
|
| + super(Keyword,self).__init__()
|
| + if identChars is None:
|
| + identChars = Keyword.DEFAULT_KEYWORD_CHARS
|
| + self.match = matchString
|
| + self.matchLen = len(matchString)
|
| + try:
|
| + self.firstMatchChar = matchString[0]
|
| + except IndexError:
|
| + warnings.warn("null string passed to Keyword; use Empty() instead",
|
| + SyntaxWarning, stacklevel=2)
|
| + self.name = '"%s"' % self.match
|
| + self.errmsg = "Expected " + self.name
|
| + self.mayReturnEmpty = False
|
| + self.mayIndexError = False
|
| + self.caseless = caseless
|
| + if caseless:
|
| + self.caselessmatch = matchString.upper()
|
| + identChars = identChars.upper()
|
| + self.identChars = set(identChars)
|
| +
|
| + def parseImpl( self, instring, loc, doActions=True ):
|
| + if self.caseless:
|
| + if ( (instring[ loc:loc+self.matchLen ].upper() == self.caselessmatch) and
|
| + (loc >= len(instring)-self.matchLen or instring[loc+self.matchLen].upper() not in self.identChars) and
|
| + (loc == 0 or instring[loc-1].upper() not in self.identChars) ):
|
| + return loc+self.matchLen, self.match
|
| + else:
|
| + if (instring[loc] == self.firstMatchChar and
|
| + (self.matchLen==1 or instring.startswith(self.match,loc)) and
|
| + (loc >= len(instring)-self.matchLen or instring[loc+self.matchLen] not in self.identChars) and
|
| + (loc == 0 or instring[loc-1] not in self.identChars) ):
|
| + return loc+self.matchLen, self.match
|
| + raise ParseException(instring, loc, self.errmsg, self)
|
| +
|
| + def copy(self):
|
| + c = super(Keyword,self).copy()
|
| + c.identChars = Keyword.DEFAULT_KEYWORD_CHARS
|
| + return c
|
| +
|
| + @staticmethod
|
| + def setDefaultKeywordChars( chars ):
|
| + """Overrides the default Keyword chars
|
| + """
|
| + Keyword.DEFAULT_KEYWORD_CHARS = chars
|
| +
|
| +class CaselessLiteral(Literal):
|
| + """
|
| + Token to match a specified string, ignoring case of letters.
|
| + Note: the matched results will always be in the case of the given
|
| + match string, NOT the case of the input text.
|
| +
|
| + Example::
|
| + OneOrMore(CaselessLiteral("CMD")).parseString("cmd CMD Cmd10") # -> ['CMD', 'CMD', 'CMD']
|
| +
|
| + (Contrast with example for L{CaselessKeyword}.)
|
| + """
|
| + def __init__( self, matchString ):
|
| + super(CaselessLiteral,self).__init__( matchString.upper() )
|
| + # Preserve the defining literal.
|
| + self.returnString = matchString
|
| + self.name = "'%s'" % self.returnString
|
| + self.errmsg = "Expected " + self.name
|
| +
|
| + def parseImpl( self, instring, loc, doActions=True ):
|
| + if instring[ loc:loc+self.matchLen ].upper() == self.match:
|
| + return loc+self.matchLen, self.returnString
|
| + raise ParseException(instring, loc, self.errmsg, self)
|
| +
|
| +class CaselessKeyword(Keyword):
|
| + """
|
| + Caseless version of L{Keyword}.
|
| +
|
| + Example::
|
| + OneOrMore(CaselessKeyword("CMD")).parseString("cmd CMD Cmd10") # -> ['CMD', 'CMD']
|
| +
|
| + (Contrast with example for L{CaselessLiteral}.)
|
| + """
|
| + def __init__( self, matchString, identChars=None ):
|
| + super(CaselessKeyword,self).__init__( matchString, identChars, caseless=True )
|
| +
|
| + def parseImpl( self, instring, loc, doActions=True ):
|
| + if ( (instring[ loc:loc+self.matchLen ].upper() == self.caselessmatch) and
|
| + (loc >= len(instring)-self.matchLen or instring[loc+self.matchLen].upper() not in self.identChars) ):
|
| + return loc+self.matchLen, self.match
|
| + raise ParseException(instring, loc, self.errmsg, self)
|
| +
|
| +class CloseMatch(Token):
|
| + """
|
| + A variation on L{Literal} which matches "close" matches, that is,
|
| + strings with at most 'n' mismatching characters. C{CloseMatch} takes parameters:
|
| + - C{match_string} - string to be matched
|
| + - C{maxMismatches} - (C{default=1}) maximum number of mismatches allowed to count as a match
|
| +
|
| + The results from a successful parse will contain the matched text from the input string and the following named results:
|
| + - C{mismatches} - a list of the positions within the match_string where mismatches were found
|
| + - C{original} - the original match_string used to compare against the input string
|
| +
|
| + If C{mismatches} is an empty list, then the match was an exact match.
|
| +
|
| + Example::
|
| + patt = CloseMatch("ATCATCGAATGGA")
|
| + patt.parseString("ATCATCGAAXGGA") # -> (['ATCATCGAAXGGA'], {'mismatches': [[9]], 'original': ['ATCATCGAATGGA']})
|
| + patt.parseString("ATCAXCGAAXGGA") # -> Exception: Expected 'ATCATCGAATGGA' (with up to 1 mismatches) (at char 0), (line:1, col:1)
|
| +
|
| + # exact match
|
| + patt.parseString("ATCATCGAATGGA") # -> (['ATCATCGAATGGA'], {'mismatches': [[]], 'original': ['ATCATCGAATGGA']})
|
| +
|
| + # close match allowing up to 2 mismatches
|
| + patt = CloseMatch("ATCATCGAATGGA", maxMismatches=2)
|
| + patt.parseString("ATCAXCGAAXGGA") # -> (['ATCAXCGAAXGGA'], {'mismatches': [[4, 9]], 'original': ['ATCATCGAATGGA']})
|
| + """
|
| + def __init__(self, match_string, maxMismatches=1):
|
| + super(CloseMatch,self).__init__()
|
| + self.name = match_string
|
| + self.match_string = match_string
|
| + self.maxMismatches = maxMismatches
|
| + self.errmsg = "Expected %r (with up to %d mismatches)" % (self.match_string, self.maxMismatches)
|
| + self.mayIndexError = False
|
| + self.mayReturnEmpty = False
|
| +
|
| + def parseImpl( self, instring, loc, doActions=True ):
|
| + start = loc
|
| + instrlen = len(instring)
|
| + maxloc = start + len(self.match_string)
|
| +
|
| + if maxloc <= instrlen:
|
| + match_string = self.match_string
|
| + match_stringloc = 0
|
| + mismatches = []
|
| + maxMismatches = self.maxMismatches
|
| +
|
| + for match_stringloc,s_m in enumerate(zip(instring[loc:maxloc], self.match_string)):
|
| + src,mat = s_m
|
| + if src != mat:
|
| + mismatches.append(match_stringloc)
|
| + if len(mismatches) > maxMismatches:
|
| + break
|
| + else:
|
| + loc = match_stringloc + 1
|
| + results = ParseResults([instring[start:loc]])
|
| + results['original'] = self.match_string
|
| + results['mismatches'] = mismatches
|
| + return loc, results
|
| +
|
| + raise ParseException(instring, loc, self.errmsg, self)
|
| +
|
| +
|
| +class Word(Token):
|
| + """
|
| + Token for matching words composed of allowed character sets.
|
| + Defined with string containing all allowed initial characters,
|
| + an optional string containing allowed body characters (if omitted,
|
| + defaults to the initial character set), and an optional minimum,
|
| + maximum, and/or exact length. The default value for C{min} is 1 (a
|
| + minimum value < 1 is not valid); the default values for C{max} and C{exact}
|
| + are 0, meaning no maximum or exact length restriction. An optional
|
| + C{excludeChars} parameter can list characters that might be found in
|
| + the input C{bodyChars} string; useful to define a word of all printables
|
| + except for one or two characters, for instance.
|
| +
|
| + L{srange} is useful for defining custom character set strings for defining
|
| + C{Word} expressions, using range notation from regular expression character sets.
|
| +
|
| + A common mistake is to use C{Word} to match a specific literal string, as in
|
| + C{Word("Address")}. Remember that C{Word} uses the string argument to define
|
| + I{sets} of matchable characters. This expression would match "Add", "AAA",
|
| + "dAred", or any other word made up of the characters 'A', 'd', 'r', 'e', and 's'.
|
| + To match an exact literal string, use L{Literal} or L{Keyword}.
|
| +
|
| + pyparsing includes helper strings for building Words:
|
| + - L{alphas}
|
| + - L{nums}
|
| + - L{alphanums}
|
| + - L{hexnums}
|
| + - L{alphas8bit} (alphabetic characters in ASCII range 128-255 - accented, tilded, umlauted, etc.)
|
| + - L{punc8bit} (non-alphabetic characters in ASCII range 128-255 - currency, symbols, superscripts, diacriticals, etc.)
|
| + - L{printables} (any non-whitespace character)
|
| +
|
| + Example::
|
| + # a word composed of digits
|
| + integer = Word(nums) # equivalent to Word("0123456789") or Word(srange("0-9"))
|
| +
|
| + # a word with a leading capital, and zero or more lowercase
|
| + capital_word = Word(alphas.upper(), alphas.lower())
|
| +
|
| + # hostnames are alphanumeric, with leading alpha, and '-'
|
| + hostname = Word(alphas, alphanums+'-')
|
| +
|
| + # roman numeral (not a strict parser, accepts invalid mix of characters)
|
| + roman = Word("IVXLCDM")
|
| +
|
| + # any string of non-whitespace characters, except for ','
|
| + csv_value = Word(printables, excludeChars=",")
|
| + """
|
| + def __init__( self, initChars, bodyChars=None, min=1, max=0, exact=0, asKeyword=False, excludeChars=None ):
|
| + super(Word,self).__init__()
|
| + if excludeChars:
|
| + initChars = ''.join(c for c in initChars if c not in excludeChars)
|
| + if bodyChars:
|
| + bodyChars = ''.join(c for c in bodyChars if c not in excludeChars)
|
| + self.initCharsOrig = initChars
|
| + self.initChars = set(initChars)
|
| + if bodyChars :
|
| + self.bodyCharsOrig = bodyChars
|
| + self.bodyChars = set(bodyChars)
|
| + else:
|
| + self.bodyCharsOrig = initChars
|
| + self.bodyChars = set(initChars)
|
| +
|
| + self.maxSpecified = max > 0
|
| +
|
| + if min < 1:
|
| + raise ValueError("cannot specify a minimum length < 1; use Optional(Word()) if zero-length word is permitted")
|
| +
|
| + self.minLen = min
|
| +
|
| + if max > 0:
|
| + self.maxLen = max
|
| + else:
|
| + self.maxLen = _MAX_INT
|
| +
|
| + if exact > 0:
|
| + self.maxLen = exact
|
| + self.minLen = exact
|
| +
|
| + self.name = _ustr(self)
|
| + self.errmsg = "Expected " + self.name
|
| + self.mayIndexError = False
|
| + self.asKeyword = asKeyword
|
| +
|
| + if ' ' not in self.initCharsOrig+self.bodyCharsOrig and (min==1 and max==0 and exact==0):
|
| + if self.bodyCharsOrig == self.initCharsOrig:
|
| + self.reString = "[%s]+" % _escapeRegexRangeChars(self.initCharsOrig)
|
| + elif len(self.initCharsOrig) == 1:
|
| + self.reString = "%s[%s]*" % \
|
| + (re.escape(self.initCharsOrig),
|
| + _escapeRegexRangeChars(self.bodyCharsOrig),)
|
| + else:
|
| + self.reString = "[%s][%s]*" % \
|
| + (_escapeRegexRangeChars(self.initCharsOrig),
|
| + _escapeRegexRangeChars(self.bodyCharsOrig),)
|
| + if self.asKeyword:
|
| + self.reString = r"\b"+self.reString+r"\b"
|
| + try:
|
| + self.re = re.compile( self.reString )
|
| + except Exception:
|
| + self.re = None
|
| +
|
| + def parseImpl( self, instring, loc, doActions=True ):
|
| + if self.re:
|
| + result = self.re.match(instring,loc)
|
| + if not result:
|
| + raise ParseException(instring, loc, self.errmsg, self)
|
| +
|
| + loc = result.end()
|
| + return loc, result.group()
|
| +
|
| + if not(instring[ loc ] in self.initChars):
|
| + raise ParseException(instring, loc, self.errmsg, self)
|
| +
|
| + start = loc
|
| + loc += 1
|
| + instrlen = len(instring)
|
| + bodychars = self.bodyChars
|
| + maxloc = start + self.maxLen
|
| + maxloc = min( maxloc, instrlen )
|
| + while loc < maxloc and instring[loc] in bodychars:
|
| + loc += 1
|
| +
|
| + throwException = False
|
| + if loc - start < self.minLen:
|
| + throwException = True
|
| + if self.maxSpecified and loc < instrlen and instring[loc] in bodychars:
|
| + throwException = True
|
| + if self.asKeyword:
|
| + if (start>0 and instring[start-1] in bodychars) or (loc<instrlen and instring[loc] in bodychars):
|
| + throwException = True
|
| +
|
| + if throwException:
|
| + raise ParseException(instring, loc, self.errmsg, self)
|
| +
|
| + return loc, instring[start:loc]
|
| +
|
| + def __str__( self ):
|
| + try:
|
| + return super(Word,self).__str__()
|
| + except Exception:
|
| + pass
|
| +
|
| +
|
| + if self.strRepr is None:
|
| +
|
| + def charsAsStr(s):
|
| + if len(s)>4:
|
| + return s[:4]+"..."
|
| + else:
|
| + return s
|
| +
|
| + if ( self.initCharsOrig != self.bodyCharsOrig ):
|
| + self.strRepr = "W:(%s,%s)" % ( charsAsStr(self.initCharsOrig), charsAsStr(self.bodyCharsOrig) )
|
| + else:
|
| + self.strRepr = "W:(%s)" % charsAsStr(self.initCharsOrig)
|
| +
|
| + return self.strRepr
|
| +
|
| +
|
| +class Regex(Token):
|
| + """
|
| + Token for matching strings that match a given regular expression.
|
| + Defined with string specifying the regular expression in a form recognized by the inbuilt Python re module.
|
| + If the given regex contains named groups (defined using C{(?P<name>...)}), these will be preserved as
|
| + named parse results.
|
| +
|
| + Example::
|
| + realnum = Regex(r"[+-]?\d+\.\d*")
|
| + date = Regex(r'(?P<year>\d{4})-(?P<month>\d\d?)-(?P<day>\d\d?)')
|
| + # ref: http://stackoverflow.com/questions/267399/how-do-you-match-only-valid-roman-numerals-with-a-regular-expression
|
| + roman = Regex(r"M{0,4}(CM|CD|D?C{0,3})(XC|XL|L?X{0,3})(IX|IV|V?I{0,3})")
|
| + """
|
| + compiledREtype = type(re.compile("[A-Z]"))
|
| + def __init__( self, pattern, flags=0):
|
| + """The parameters C{pattern} and C{flags} are passed to the C{re.compile()} function as-is. See the Python C{re} module for an explanation of the acceptable patterns and flags."""
|
| + super(Regex,self).__init__()
|
| +
|
| + if isinstance(pattern, basestring):
|
| + if not pattern:
|
| + warnings.warn("null string passed to Regex; use Empty() instead",
|
| + SyntaxWarning, stacklevel=2)
|
| +
|
| + self.pattern = pattern
|
| + self.flags = flags
|
| +
|
| + try:
|
| + self.re = re.compile(self.pattern, self.flags)
|
| + self.reString = self.pattern
|
| + except sre_constants.error:
|
| + warnings.warn("invalid pattern (%s) passed to Regex" % pattern,
|
| + SyntaxWarning, stacklevel=2)
|
| + raise
|
| +
|
| + elif isinstance(pattern, Regex.compiledREtype):
|
| + self.re = pattern
|
| + self.pattern = \
|
| + self.reString = str(pattern)
|
| + self.flags = flags
|
| +
|
| + else:
|
| + raise ValueError("Regex may only be constructed with a string or a compiled RE object")
|
| +
|
| + self.name = _ustr(self)
|
| + self.errmsg = "Expected " + self.name
|
| + self.mayIndexError = False
|
| + self.mayReturnEmpty = True
|
| +
|
| + def parseImpl( self, instring, loc, doActions=True ):
|
| + result = self.re.match(instring,loc)
|
| + if not result:
|
| + raise ParseException(instring, loc, self.errmsg, self)
|
| +
|
| + loc = result.end()
|
| + d = result.groupdict()
|
| + ret = ParseResults(result.group())
|
| + if d:
|
| + for k in d:
|
| + ret[k] = d[k]
|
| + return loc,ret
|
| +
|
| + def __str__( self ):
|
| + try:
|
| + return super(Regex,self).__str__()
|
| + except Exception:
|
| + pass
|
| +
|
| + if self.strRepr is None:
|
| + self.strRepr = "Re:(%s)" % repr(self.pattern)
|
| +
|
| + return self.strRepr
|
| +
|
| +
|
| +class QuotedString(Token):
|
| + r"""
|
| + Token for matching strings that are delimited by quoting characters.
|
| +
|
| + Defined with the following parameters:
|
| + - quoteChar - string of one or more characters defining the quote delimiting string
|
| + - escChar - character to escape quotes, typically backslash (default=C{None})
|
| + - escQuote - special quote sequence to escape an embedded quote string (such as SQL's "" to escape an embedded ") (default=C{None})
|
| + - multiline - boolean indicating whether quotes can span multiple lines (default=C{False})
|
| + - unquoteResults - boolean indicating whether the matched text should be unquoted (default=C{True})
|
| + - endQuoteChar - string of one or more characters defining the end of the quote delimited string (default=C{None} => same as quoteChar)
|
| + - convertWhitespaceEscapes - convert escaped whitespace (C{'\t'}, C{'\n'}, etc.) to actual whitespace (default=C{True})
|
| +
|
| + Example::
|
| + qs = QuotedString('"')
|
| + print(qs.searchString('lsjdf "This is the quote" sldjf'))
|
| + complex_qs = QuotedString('{{', endQuoteChar='}}')
|
| + print(complex_qs.searchString('lsjdf {{This is the "quote"}} sldjf'))
|
| + sql_qs = QuotedString('"', escQuote='""')
|
| + print(sql_qs.searchString('lsjdf "This is the quote with ""embedded"" quotes" sldjf'))
|
| + prints::
|
| + [['This is the quote']]
|
| + [['This is the "quote"']]
|
| + [['This is the quote with "embedded" quotes']]
|
| + """
|
| + def __init__( self, quoteChar, escChar=None, escQuote=None, multiline=False, unquoteResults=True, endQuoteChar=None, convertWhitespaceEscapes=True):
|
| + super(QuotedString,self).__init__()
|
| +
|
| + # remove white space from quote chars - wont work anyway
|
| + quoteChar = quoteChar.strip()
|
| + if not quoteChar:
|
| + warnings.warn("quoteChar cannot be the empty string",SyntaxWarning,stacklevel=2)
|
| + raise SyntaxError()
|
| +
|
| + if endQuoteChar is None:
|
| + endQuoteChar = quoteChar
|
| + else:
|
| + endQuoteChar = endQuoteChar.strip()
|
| + if not endQuoteChar:
|
| + warnings.warn("endQuoteChar cannot be the empty string",SyntaxWarning,stacklevel=2)
|
| + raise SyntaxError()
|
| +
|
| + self.quoteChar = quoteChar
|
| + self.quoteCharLen = len(quoteChar)
|
| + self.firstQuoteChar = quoteChar[0]
|
| + self.endQuoteChar = endQuoteChar
|
| + self.endQuoteCharLen = len(endQuoteChar)
|
| + self.escChar = escChar
|
| + self.escQuote = escQuote
|
| + self.unquoteResults = unquoteResults
|
| + self.convertWhitespaceEscapes = convertWhitespaceEscapes
|
| +
|
| + if multiline:
|
| + self.flags = re.MULTILINE | re.DOTALL
|
| + self.pattern = r'%s(?:[^%s%s]' % \
|
| + ( re.escape(self.quoteChar),
|
| + _escapeRegexRangeChars(self.endQuoteChar[0]),
|
| + (escChar is not None and _escapeRegexRangeChars(escChar) or '') )
|
| + else:
|
| + self.flags = 0
|
| + self.pattern = r'%s(?:[^%s\n\r%s]' % \
|
| + ( re.escape(self.quoteChar),
|
| + _escapeRegexRangeChars(self.endQuoteChar[0]),
|
| + (escChar is not None and _escapeRegexRangeChars(escChar) or '') )
|
| + if len(self.endQuoteChar) > 1:
|
| + self.pattern += (
|
| + '|(?:' + ')|(?:'.join("%s[^%s]" % (re.escape(self.endQuoteChar[:i]),
|
| + _escapeRegexRangeChars(self.endQuoteChar[i]))
|
| + for i in range(len(self.endQuoteChar)-1,0,-1)) + ')'
|
| + )
|
| + if escQuote:
|
| + self.pattern += (r'|(?:%s)' % re.escape(escQuote))
|
| + if escChar:
|
| + self.pattern += (r'|(?:%s.)' % re.escape(escChar))
|
| + self.escCharReplacePattern = re.escape(self.escChar)+"(.)"
|
| + self.pattern += (r')*%s' % re.escape(self.endQuoteChar))
|
| +
|
| + try:
|
| + self.re = re.compile(self.pattern, self.flags)
|
| + self.reString = self.pattern
|
| + except sre_constants.error:
|
| + warnings.warn("invalid pattern (%s) passed to Regex" % self.pattern,
|
| + SyntaxWarning, stacklevel=2)
|
| + raise
|
| +
|
| + self.name = _ustr(self)
|
| + self.errmsg = "Expected " + self.name
|
| + self.mayIndexError = False
|
| + self.mayReturnEmpty = True
|
| +
|
| + def parseImpl( self, instring, loc, doActions=True ):
|
| + result = instring[loc] == self.firstQuoteChar and self.re.match(instring,loc) or None
|
| + if not result:
|
| + raise ParseException(instring, loc, self.errmsg, self)
|
| +
|
| + loc = result.end()
|
| + ret = result.group()
|
| +
|
| + if self.unquoteResults:
|
| +
|
| + # strip off quotes
|
| + ret = ret[self.quoteCharLen:-self.endQuoteCharLen]
|
| +
|
| + if isinstance(ret,basestring):
|
| + # replace escaped whitespace
|
| + if '\\' in ret and self.convertWhitespaceEscapes:
|
| + ws_map = {
|
| + r'\t' : '\t',
|
| + r'\n' : '\n',
|
| + r'\f' : '\f',
|
| + r'\r' : '\r',
|
| + }
|
| + for wslit,wschar in ws_map.items():
|
| + ret = ret.replace(wslit, wschar)
|
| +
|
| + # replace escaped characters
|
| + if self.escChar:
|
| + ret = re.sub(self.escCharReplacePattern,"\g<1>",ret)
|
| +
|
| + # replace escaped quotes
|
| + if self.escQuote:
|
| + ret = ret.replace(self.escQuote, self.endQuoteChar)
|
| +
|
| + return loc, ret
|
| +
|
| + def __str__( self ):
|
| + try:
|
| + return super(QuotedString,self).__str__()
|
| + except Exception:
|
| + pass
|
| +
|
| + if self.strRepr is None:
|
| + self.strRepr = "quoted string, starting with %s ending with %s" % (self.quoteChar, self.endQuoteChar)
|
| +
|
| + return self.strRepr
|
| +
|
| +
|
| +class CharsNotIn(Token):
|
| + """
|
| + Token for matching words composed of characters I{not} in a given set (will
|
| + include whitespace in matched characters if not listed in the provided exclusion set - see example).
|
| + Defined with string containing all disallowed characters, and an optional
|
| + minimum, maximum, and/or exact length. The default value for C{min} is 1 (a
|
| + minimum value < 1 is not valid); the default values for C{max} and C{exact}
|
| + are 0, meaning no maximum or exact length restriction.
|
| +
|
| + Example::
|
| + # define a comma-separated-value as anything that is not a ','
|
| + csv_value = CharsNotIn(',')
|
| + print(delimitedList(csv_value).parseString("dkls,lsdkjf,s12 34,@!#,213"))
|
| + prints::
|
| + ['dkls', 'lsdkjf', 's12 34', '@!#', '213']
|
| + """
|
| + def __init__( self, notChars, min=1, max=0, exact=0 ):
|
| + super(CharsNotIn,self).__init__()
|
| + self.skipWhitespace = False
|
| + self.notChars = notChars
|
| +
|
| + if min < 1:
|
| + raise ValueError("cannot specify a minimum length < 1; use Optional(CharsNotIn()) if zero-length char group is permitted")
|
| +
|
| + self.minLen = min
|
| +
|
| + if max > 0:
|
| + self.maxLen = max
|
| + else:
|
| + self.maxLen = _MAX_INT
|
| +
|
| + if exact > 0:
|
| + self.maxLen = exact
|
| + self.minLen = exact
|
| +
|
| + self.name = _ustr(self)
|
| + self.errmsg = "Expected " + self.name
|
| + self.mayReturnEmpty = ( self.minLen == 0 )
|
| + self.mayIndexError = False
|
| +
|
| + def parseImpl( self, instring, loc, doActions=True ):
|
| + if instring[loc] in self.notChars:
|
| + raise ParseException(instring, loc, self.errmsg, self)
|
| +
|
| + start = loc
|
| + loc += 1
|
| + notchars = self.notChars
|
| + maxlen = min( start+self.maxLen, len(instring) )
|
| + while loc < maxlen and \
|
| + (instring[loc] not in notchars):
|
| + loc += 1
|
| +
|
| + if loc - start < self.minLen:
|
| + raise ParseException(instring, loc, self.errmsg, self)
|
| +
|
| + return loc, instring[start:loc]
|
| +
|
| + def __str__( self ):
|
| + try:
|
| + return super(CharsNotIn, self).__str__()
|
| + except Exception:
|
| + pass
|
| +
|
| + if self.strRepr is None:
|
| + if len(self.notChars) > 4:
|
| + self.strRepr = "!W:(%s...)" % self.notChars[:4]
|
| + else:
|
| + self.strRepr = "!W:(%s)" % self.notChars
|
| +
|
| + return self.strRepr
|
| +
|
| +class White(Token):
|
| + """
|
| + Special matching class for matching whitespace. Normally, whitespace is ignored
|
| + by pyparsing grammars. This class is included when some whitespace structures
|
| + are significant. Define with a string containing the whitespace characters to be
|
| + matched; default is C{" \\t\\r\\n"}. Also takes optional C{min}, C{max}, and C{exact} arguments,
|
| + as defined for the C{L{Word}} class.
|
| + """
|
| + whiteStrs = {
|
| + " " : "<SPC>",
|
| + "\t": "<TAB>",
|
| + "\n": "<LF>",
|
| + "\r": "<CR>",
|
| + "\f": "<FF>",
|
| + }
|
| + def __init__(self, ws=" \t\r\n", min=1, max=0, exact=0):
|
| + super(White,self).__init__()
|
| + self.matchWhite = ws
|
| + self.setWhitespaceChars( "".join(c for c in self.whiteChars if c not in self.matchWhite) )
|
| + #~ self.leaveWhitespace()
|
| + self.name = ("".join(White.whiteStrs[c] for c in self.matchWhite))
|
| + self.mayReturnEmpty = True
|
| + self.errmsg = "Expected " + self.name
|
| +
|
| + self.minLen = min
|
| +
|
| + if max > 0:
|
| + self.maxLen = max
|
| + else:
|
| + self.maxLen = _MAX_INT
|
| +
|
| + if exact > 0:
|
| + self.maxLen = exact
|
| + self.minLen = exact
|
| +
|
| + def parseImpl( self, instring, loc, doActions=True ):
|
| + if not(instring[ loc ] in self.matchWhite):
|
| + raise ParseException(instring, loc, self.errmsg, self)
|
| + start = loc
|
| + loc += 1
|
| + maxloc = start + self.maxLen
|
| + maxloc = min( maxloc, len(instring) )
|
| + while loc < maxloc and instring[loc] in self.matchWhite:
|
| + loc += 1
|
| +
|
| + if loc - start < self.minLen:
|
| + raise ParseException(instring, loc, self.errmsg, self)
|
| +
|
| + return loc, instring[start:loc]
|
| +
|
| +
|
| +class _PositionToken(Token):
|
| + def __init__( self ):
|
| + super(_PositionToken,self).__init__()
|
| + self.name=self.__class__.__name__
|
| + self.mayReturnEmpty = True
|
| + self.mayIndexError = False
|
| +
|
| +class GoToColumn(_PositionToken):
|
| + """
|
| + Token to advance to a specific column of input text; useful for tabular report scraping.
|
| + """
|
| + def __init__( self, colno ):
|
| + super(GoToColumn,self).__init__()
|
| + self.col = colno
|
| +
|
| + def preParse( self, instring, loc ):
|
| + if col(loc,instring) != self.col:
|
| + instrlen = len(instring)
|
| + if self.ignoreExprs:
|
| + loc = self._skipIgnorables( instring, loc )
|
| + while loc < instrlen and instring[loc].isspace() and col( loc, instring ) != self.col :
|
| + loc += 1
|
| + return loc
|
| +
|
| + def parseImpl( self, instring, loc, doActions=True ):
|
| + thiscol = col( loc, instring )
|
| + if thiscol > self.col:
|
| + raise ParseException( instring, loc, "Text not in expected column", self )
|
| + newloc = loc + self.col - thiscol
|
| + ret = instring[ loc: newloc ]
|
| + return newloc, ret
|
| +
|
| +
|
| +class LineStart(_PositionToken):
|
| + """
|
| + Matches if current position is at the beginning of a line within the parse string
|
| +
|
| + Example::
|
| +
|
| + test = '''\
|
| + AAA this line
|
| + AAA and this line
|
| + AAA but not this one
|
| + B AAA and definitely not this one
|
| + '''
|
| +
|
| + for t in (LineStart() + 'AAA' + restOfLine).searchString(test):
|
| + print(t)
|
| +
|
| + Prints::
|
| + ['AAA', ' this line']
|
| + ['AAA', ' and this line']
|
| +
|
| + """
|
| + def __init__( self ):
|
| + super(LineStart,self).__init__()
|
| + self.errmsg = "Expected start of line"
|
| +
|
| + def parseImpl( self, instring, loc, doActions=True ):
|
| + if col(loc, instring) == 1:
|
| + return loc, []
|
| + raise ParseException(instring, loc, self.errmsg, self)
|
| +
|
| +class LineEnd(_PositionToken):
|
| + """
|
| + Matches if current position is at the end of a line within the parse string
|
| + """
|
| + def __init__( self ):
|
| + super(LineEnd,self).__init__()
|
| + self.setWhitespaceChars( ParserElement.DEFAULT_WHITE_CHARS.replace("\n","") )
|
| + self.errmsg = "Expected end of line"
|
| +
|
| + def parseImpl( self, instring, loc, doActions=True ):
|
| + if loc<len(instring):
|
| + if instring[loc] == "\n":
|
| + return loc+1, "\n"
|
| + else:
|
| + raise ParseException(instring, loc, self.errmsg, self)
|
| + elif loc == len(instring):
|
| + return loc+1, []
|
| + else:
|
| + raise ParseException(instring, loc, self.errmsg, self)
|
| +
|
| +class StringStart(_PositionToken):
|
| + """
|
| + Matches if current position is at the beginning of the parse string
|
| + """
|
| + def __init__( self ):
|
| + super(StringStart,self).__init__()
|
| + self.errmsg = "Expected start of text"
|
| +
|
| + def parseImpl( self, instring, loc, doActions=True ):
|
| + if loc != 0:
|
| + # see if entire string up to here is just whitespace and ignoreables
|
| + if loc != self.preParse( instring, 0 ):
|
| + raise ParseException(instring, loc, self.errmsg, self)
|
| + return loc, []
|
| +
|
| +class StringEnd(_PositionToken):
|
| + """
|
| + Matches if current position is at the end of the parse string
|
| + """
|
| + def __init__( self ):
|
| + super(StringEnd,self).__init__()
|
| + self.errmsg = "Expected end of text"
|
| +
|
| + def parseImpl( self, instring, loc, doActions=True ):
|
| + if loc < len(instring):
|
| + raise ParseException(instring, loc, self.errmsg, self)
|
| + elif loc == len(instring):
|
| + return loc+1, []
|
| + elif loc > len(instring):
|
| + return loc, []
|
| + else:
|
| + raise ParseException(instring, loc, self.errmsg, self)
|
| +
|
| +class WordStart(_PositionToken):
|
| + """
|
| + Matches if the current position is at the beginning of a Word, and
|
| + is not preceded by any character in a given set of C{wordChars}
|
| + (default=C{printables}). To emulate the C{\b} behavior of regular expressions,
|
| + use C{WordStart(alphanums)}. C{WordStart} will also match at the beginning of
|
| + the string being parsed, or at the beginning of a line.
|
| + """
|
| + def __init__(self, wordChars = printables):
|
| + super(WordStart,self).__init__()
|
| + self.wordChars = set(wordChars)
|
| + self.errmsg = "Not at the start of a word"
|
| +
|
| + def parseImpl(self, instring, loc, doActions=True ):
|
| + if loc != 0:
|
| + if (instring[loc-1] in self.wordChars or
|
| + instring[loc] not in self.wordChars):
|
| + raise ParseException(instring, loc, self.errmsg, self)
|
| + return loc, []
|
| +
|
| +class WordEnd(_PositionToken):
|
| + """
|
| + Matches if the current position is at the end of a Word, and
|
| + is not followed by any character in a given set of C{wordChars}
|
| + (default=C{printables}). To emulate the C{\b} behavior of regular expressions,
|
| + use C{WordEnd(alphanums)}. C{WordEnd} will also match at the end of
|
| + the string being parsed, or at the end of a line.
|
| + """
|
| + def __init__(self, wordChars = printables):
|
| + super(WordEnd,self).__init__()
|
| + self.wordChars = set(wordChars)
|
| + self.skipWhitespace = False
|
| + self.errmsg = "Not at the end of a word"
|
| +
|
| + def parseImpl(self, instring, loc, doActions=True ):
|
| + instrlen = len(instring)
|
| + if instrlen>0 and loc<instrlen:
|
| + if (instring[loc] in self.wordChars or
|
| + instring[loc-1] not in self.wordChars):
|
| + raise ParseException(instring, loc, self.errmsg, self)
|
| + return loc, []
|
| +
|
| +
|
| +class ParseExpression(ParserElement):
|
| + """
|
| + Abstract subclass of ParserElement, for combining and post-processing parsed tokens.
|
| + """
|
| + def __init__( self, exprs, savelist = False ):
|
| + super(ParseExpression,self).__init__(savelist)
|
| + if isinstance( exprs, _generatorType ):
|
| + exprs = list(exprs)
|
| +
|
| + if isinstance( exprs, basestring ):
|
| + self.exprs = [ ParserElement._literalStringClass( exprs ) ]
|
| + elif isinstance( exprs, collections.Iterable ):
|
| + exprs = list(exprs)
|
| + # if sequence of strings provided, wrap with Literal
|
| + if all(isinstance(expr, basestring) for expr in exprs):
|
| + exprs = map(ParserElement._literalStringClass, exprs)
|
| + self.exprs = list(exprs)
|
| + else:
|
| + try:
|
| + self.exprs = list( exprs )
|
| + except TypeError:
|
| + self.exprs = [ exprs ]
|
| + self.callPreparse = False
|
| +
|
| + def __getitem__( self, i ):
|
| + return self.exprs[i]
|
| +
|
| + def append( self, other ):
|
| + self.exprs.append( other )
|
| + self.strRepr = None
|
| + return self
|
| +
|
| + def leaveWhitespace( self ):
|
| + """Extends C{leaveWhitespace} defined in base class, and also invokes C{leaveWhitespace} on
|
| + all contained expressions."""
|
| + self.skipWhitespace = False
|
| + self.exprs = [ e.copy() for e in self.exprs ]
|
| + for e in self.exprs:
|
| + e.leaveWhitespace()
|
| + return self
|
| +
|
| + def ignore( self, other ):
|
| + if isinstance( other, Suppress ):
|
| + if other not in self.ignoreExprs:
|
| + super( ParseExpression, self).ignore( other )
|
| + for e in self.exprs:
|
| + e.ignore( self.ignoreExprs[-1] )
|
| + else:
|
| + super( ParseExpression, self).ignore( other )
|
| + for e in self.exprs:
|
| + e.ignore( self.ignoreExprs[-1] )
|
| + return self
|
| +
|
| + def __str__( self ):
|
| + try:
|
| + return super(ParseExpression,self).__str__()
|
| + except Exception:
|
| + pass
|
| +
|
| + if self.strRepr is None:
|
| + self.strRepr = "%s:(%s)" % ( self.__class__.__name__, _ustr(self.exprs) )
|
| + return self.strRepr
|
| +
|
| + def streamline( self ):
|
| + super(ParseExpression,self).streamline()
|
| +
|
| + for e in self.exprs:
|
| + e.streamline()
|
| +
|
| + # collapse nested And's of the form And( And( And( a,b), c), d) to And( a,b,c,d )
|
| + # but only if there are no parse actions or resultsNames on the nested And's
|
| + # (likewise for Or's and MatchFirst's)
|
| + if ( len(self.exprs) == 2 ):
|
| + other = self.exprs[0]
|
| + if ( isinstance( other, self.__class__ ) and
|
| + not(other.parseAction) and
|
| + other.resultsName is None and
|
| + not other.debug ):
|
| + self.exprs = other.exprs[:] + [ self.exprs[1] ]
|
| + self.strRepr = None
|
| + self.mayReturnEmpty |= other.mayReturnEmpty
|
| + self.mayIndexError |= other.mayIndexError
|
| +
|
| + other = self.exprs[-1]
|
| + if ( isinstance( other, self.__class__ ) and
|
| + not(other.parseAction) and
|
| + other.resultsName is None and
|
| + not other.debug ):
|
| + self.exprs = self.exprs[:-1] + other.exprs[:]
|
| + self.strRepr = None
|
| + self.mayReturnEmpty |= other.mayReturnEmpty
|
| + self.mayIndexError |= other.mayIndexError
|
| +
|
| + self.errmsg = "Expected " + _ustr(self)
|
| +
|
| + return self
|
| +
|
| + def setResultsName( self, name, listAllMatches=False ):
|
| + ret = super(ParseExpression,self).setResultsName(name,listAllMatches)
|
| + return ret
|
| +
|
| + def validate( self, validateTrace=[] ):
|
| + tmp = validateTrace[:]+[self]
|
| + for e in self.exprs:
|
| + e.validate(tmp)
|
| + self.checkRecursion( [] )
|
| +
|
| + def copy(self):
|
| + ret = super(ParseExpression,self).copy()
|
| + ret.exprs = [e.copy() for e in self.exprs]
|
| + return ret
|
| +
|
| +class And(ParseExpression):
|
| + """
|
| + Requires all given C{ParseExpression}s to be found in the given order.
|
| + Expressions may be separated by whitespace.
|
| + May be constructed using the C{'+'} operator.
|
| + May also be constructed using the C{'-'} operator, which will suppress backtracking.
|
| +
|
| + Example::
|
| + integer = Word(nums)
|
| + name_expr = OneOrMore(Word(alphas))
|
| +
|
| + expr = And([integer("id"),name_expr("name"),integer("age")])
|
| + # more easily written as:
|
| + expr = integer("id") + name_expr("name") + integer("age")
|
| + """
|
| +
|
| + class _ErrorStop(Empty):
|
| + def __init__(self, *args, **kwargs):
|
| + super(And._ErrorStop,self).__init__(*args, **kwargs)
|
| + self.name = '-'
|
| + self.leaveWhitespace()
|
| +
|
| + def __init__( self, exprs, savelist = True ):
|
| + super(And,self).__init__(exprs, savelist)
|
| + self.mayReturnEmpty = all(e.mayReturnEmpty for e in self.exprs)
|
| + self.setWhitespaceChars( self.exprs[0].whiteChars )
|
| + self.skipWhitespace = self.exprs[0].skipWhitespace
|
| + self.callPreparse = True
|
| +
|
| + def parseImpl( self, instring, loc, doActions=True ):
|
| + # pass False as last arg to _parse for first element, since we already
|
| + # pre-parsed the string as part of our And pre-parsing
|
| + loc, resultlist = self.exprs[0]._parse( instring, loc, doActions, callPreParse=False )
|
| + errorStop = False
|
| + for e in self.exprs[1:]:
|
| + if isinstance(e, And._ErrorStop):
|
| + errorStop = True
|
| + continue
|
| + if errorStop:
|
| + try:
|
| + loc, exprtokens = e._parse( instring, loc, doActions )
|
| + except ParseSyntaxException:
|
| + raise
|
| + except ParseBaseException as pe:
|
| + pe.__traceback__ = None
|
| + raise ParseSyntaxException._from_exception(pe)
|
| + except IndexError:
|
| + raise ParseSyntaxException(instring, len(instring), self.errmsg, self)
|
| + else:
|
| + loc, exprtokens = e._parse( instring, loc, doActions )
|
| + if exprtokens or exprtokens.haskeys():
|
| + resultlist += exprtokens
|
| + return loc, resultlist
|
| +
|
| + def __iadd__(self, other ):
|
| + if isinstance( other, basestring ):
|
| + other = ParserElement._literalStringClass( other )
|
| + return self.append( other ) #And( [ self, other ] )
|
| +
|
| + def checkRecursion( self, parseElementList ):
|
| + subRecCheckList = parseElementList[:] + [ self ]
|
| + for e in self.exprs:
|
| + e.checkRecursion( subRecCheckList )
|
| + if not e.mayReturnEmpty:
|
| + break
|
| +
|
| + def __str__( self ):
|
| + if hasattr(self,"name"):
|
| + return self.name
|
| +
|
| + if self.strRepr is None:
|
| + self.strRepr = "{" + " ".join(_ustr(e) for e in self.exprs) + "}"
|
| +
|
| + return self.strRepr
|
| +
|
| +
|
| +class Or(ParseExpression):
|
| + """
|
| + Requires that at least one C{ParseExpression} is found.
|
| + If two expressions match, the expression that matches the longest string will be used.
|
| + May be constructed using the C{'^'} operator.
|
| +
|
| + Example::
|
| + # construct Or using '^' operator
|
| +
|
| + number = Word(nums) ^ Combine(Word(nums) + '.' + Word(nums))
|
| + print(number.searchString("123 3.1416 789"))
|
| + prints::
|
| + [['123'], ['3.1416'], ['789']]
|
| + """
|
| + def __init__( self, exprs, savelist = False ):
|
| + super(Or,self).__init__(exprs, savelist)
|
| + if self.exprs:
|
| + self.mayReturnEmpty = any(e.mayReturnEmpty for e in self.exprs)
|
| + else:
|
| + self.mayReturnEmpty = True
|
| +
|
| + def parseImpl( self, instring, loc, doActions=True ):
|
| + maxExcLoc = -1
|
| + maxException = None
|
| + matches = []
|
| + for e in self.exprs:
|
| + try:
|
| + loc2 = e.tryParse( instring, loc )
|
| + except ParseException as err:
|
| + err.__traceback__ = None
|
| + if err.loc > maxExcLoc:
|
| + maxException = err
|
| + maxExcLoc = err.loc
|
| + except IndexError:
|
| + if len(instring) > maxExcLoc:
|
| + maxException = ParseException(instring,len(instring),e.errmsg,self)
|
| + maxExcLoc = len(instring)
|
| + else:
|
| + # save match among all matches, to retry longest to shortest
|
| + matches.append((loc2, e))
|
| +
|
| + if matches:
|
| + matches.sort(key=lambda x: -x[0])
|
| + for _,e in matches:
|
| + try:
|
| + return e._parse( instring, loc, doActions )
|
| + except ParseException as err:
|
| + err.__traceback__ = None
|
| + if err.loc > maxExcLoc:
|
| + maxException = err
|
| + maxExcLoc = err.loc
|
| +
|
| + if maxException is not None:
|
| + maxException.msg = self.errmsg
|
| + raise maxException
|
| + else:
|
| + raise ParseException(instring, loc, "no defined alternatives to match", self)
|
| +
|
| +
|
| + def __ixor__(self, other ):
|
| + if isinstance( other, basestring ):
|
| + other = ParserElement._literalStringClass( other )
|
| + return self.append( other ) #Or( [ self, other ] )
|
| +
|
| + def __str__( self ):
|
| + if hasattr(self,"name"):
|
| + return self.name
|
| +
|
| + if self.strRepr is None:
|
| + self.strRepr = "{" + " ^ ".join(_ustr(e) for e in self.exprs) + "}"
|
| +
|
| + return self.strRepr
|
| +
|
| + def checkRecursion( self, parseElementList ):
|
| + subRecCheckList = parseElementList[:] + [ self ]
|
| + for e in self.exprs:
|
| + e.checkRecursion( subRecCheckList )
|
| +
|
| +
|
| +class MatchFirst(ParseExpression):
|
| + """
|
| + Requires that at least one C{ParseExpression} is found.
|
| + If two expressions match, the first one listed is the one that will match.
|
| + May be constructed using the C{'|'} operator.
|
| +
|
| + Example::
|
| + # construct MatchFirst using '|' operator
|
| +
|
| + # watch the order of expressions to match
|
| + number = Word(nums) | Combine(Word(nums) + '.' + Word(nums))
|
| + print(number.searchString("123 3.1416 789")) # Fail! -> [['123'], ['3'], ['1416'], ['789']]
|
| +
|
| + # put more selective expression first
|
| + number = Combine(Word(nums) + '.' + Word(nums)) | Word(nums)
|
| + print(number.searchString("123 3.1416 789")) # Better -> [['123'], ['3.1416'], ['789']]
|
| + """
|
| + def __init__( self, exprs, savelist = False ):
|
| + super(MatchFirst,self).__init__(exprs, savelist)
|
| + if self.exprs:
|
| + self.mayReturnEmpty = any(e.mayReturnEmpty for e in self.exprs)
|
| + else:
|
| + self.mayReturnEmpty = True
|
| +
|
| + def parseImpl( self, instring, loc, doActions=True ):
|
| + maxExcLoc = -1
|
| + maxException = None
|
| + for e in self.exprs:
|
| + try:
|
| + ret = e._parse( instring, loc, doActions )
|
| + return ret
|
| + except ParseException as err:
|
| + if err.loc > maxExcLoc:
|
| + maxException = err
|
| + maxExcLoc = err.loc
|
| + except IndexError:
|
| + if len(instring) > maxExcLoc:
|
| + maxException = ParseException(instring,len(instring),e.errmsg,self)
|
| + maxExcLoc = len(instring)
|
| +
|
| + # only got here if no expression matched, raise exception for match that made it the furthest
|
| + else:
|
| + if maxException is not None:
|
| + maxException.msg = self.errmsg
|
| + raise maxException
|
| + else:
|
| + raise ParseException(instring, loc, "no defined alternatives to match", self)
|
| +
|
| + def __ior__(self, other ):
|
| + if isinstance( other, basestring ):
|
| + other = ParserElement._literalStringClass( other )
|
| + return self.append( other ) #MatchFirst( [ self, other ] )
|
| +
|
| + def __str__( self ):
|
| + if hasattr(self,"name"):
|
| + return self.name
|
| +
|
| + if self.strRepr is None:
|
| + self.strRepr = "{" + " | ".join(_ustr(e) for e in self.exprs) + "}"
|
| +
|
| + return self.strRepr
|
| +
|
| + def checkRecursion( self, parseElementList ):
|
| + subRecCheckList = parseElementList[:] + [ self ]
|
| + for e in self.exprs:
|
| + e.checkRecursion( subRecCheckList )
|
| +
|
| +
|
| +class Each(ParseExpression):
|
| + """
|
| + Requires all given C{ParseExpression}s to be found, but in any order.
|
| + Expressions may be separated by whitespace.
|
| + May be constructed using the C{'&'} operator.
|
| +
|
| + Example::
|
| + color = oneOf("RED ORANGE YELLOW GREEN BLUE PURPLE BLACK WHITE BROWN")
|
| + shape_type = oneOf("SQUARE CIRCLE TRIANGLE STAR HEXAGON OCTAGON")
|
| + integer = Word(nums)
|
| + shape_attr = "shape:" + shape_type("shape")
|
| + posn_attr = "posn:" + Group(integer("x") + ',' + integer("y"))("posn")
|
| + color_attr = "color:" + color("color")
|
| + size_attr = "size:" + integer("size")
|
| +
|
| + # use Each (using operator '&') to accept attributes in any order
|
| + # (shape and posn are required, color and size are optional)
|
| + shape_spec = shape_attr & posn_attr & Optional(color_attr) & Optional(size_attr)
|
| +
|
| + shape_spec.runTests('''
|
| + shape: SQUARE color: BLACK posn: 100, 120
|
| + shape: CIRCLE size: 50 color: BLUE posn: 50,80
|
| + color:GREEN size:20 shape:TRIANGLE posn:20,40
|
| + '''
|
| + )
|
| + prints::
|
| + shape: SQUARE color: BLACK posn: 100, 120
|
| + ['shape:', 'SQUARE', 'color:', 'BLACK', 'posn:', ['100', ',', '120']]
|
| + - color: BLACK
|
| + - posn: ['100', ',', '120']
|
| + - x: 100
|
| + - y: 120
|
| + - shape: SQUARE
|
| +
|
| +
|
| + shape: CIRCLE size: 50 color: BLUE posn: 50,80
|
| + ['shape:', 'CIRCLE', 'size:', '50', 'color:', 'BLUE', 'posn:', ['50', ',', '80']]
|
| + - color: BLUE
|
| + - posn: ['50', ',', '80']
|
| + - x: 50
|
| + - y: 80
|
| + - shape: CIRCLE
|
| + - size: 50
|
| +
|
| +
|
| + color: GREEN size: 20 shape: TRIANGLE posn: 20,40
|
| + ['color:', 'GREEN', 'size:', '20', 'shape:', 'TRIANGLE', 'posn:', ['20', ',', '40']]
|
| + - color: GREEN
|
| + - posn: ['20', ',', '40']
|
| + - x: 20
|
| + - y: 40
|
| + - shape: TRIANGLE
|
| + - size: 20
|
| + """
|
| + def __init__( self, exprs, savelist = True ):
|
| + super(Each,self).__init__(exprs, savelist)
|
| + self.mayReturnEmpty = all(e.mayReturnEmpty for e in self.exprs)
|
| + self.skipWhitespace = True
|
| + self.initExprGroups = True
|
| +
|
| + def parseImpl( self, instring, loc, doActions=True ):
|
| + if self.initExprGroups:
|
| + self.opt1map = dict((id(e.expr),e) for e in self.exprs if isinstance(e,Optional))
|
| + opt1 = [ e.expr for e in self.exprs if isinstance(e,Optional) ]
|
| + opt2 = [ e for e in self.exprs if e.mayReturnEmpty and not isinstance(e,Optional)]
|
| + self.optionals = opt1 + opt2
|
| + self.multioptionals = [ e.expr for e in self.exprs if isinstance(e,ZeroOrMore) ]
|
| + self.multirequired = [ e.expr for e in self.exprs if isinstance(e,OneOrMore) ]
|
| + self.required = [ e for e in self.exprs if not isinstance(e,(Optional,ZeroOrMore,OneOrMore)) ]
|
| + self.required += self.multirequired
|
| + self.initExprGroups = False
|
| + tmpLoc = loc
|
| + tmpReqd = self.required[:]
|
| + tmpOpt = self.optionals[:]
|
| + matchOrder = []
|
| +
|
| + keepMatching = True
|
| + while keepMatching:
|
| + tmpExprs = tmpReqd + tmpOpt + self.multioptionals + self.multirequired
|
| + failed = []
|
| + for e in tmpExprs:
|
| + try:
|
| + tmpLoc = e.tryParse( instring, tmpLoc )
|
| + except ParseException:
|
| + failed.append(e)
|
| + else:
|
| + matchOrder.append(self.opt1map.get(id(e),e))
|
| + if e in tmpReqd:
|
| + tmpReqd.remove(e)
|
| + elif e in tmpOpt:
|
| + tmpOpt.remove(e)
|
| + if len(failed) == len(tmpExprs):
|
| + keepMatching = False
|
| +
|
| + if tmpReqd:
|
| + missing = ", ".join(_ustr(e) for e in tmpReqd)
|
| + raise ParseException(instring,loc,"Missing one or more required elements (%s)" % missing )
|
| +
|
| + # add any unmatched Optionals, in case they have default values defined
|
| + matchOrder += [e for e in self.exprs if isinstance(e,Optional) and e.expr in tmpOpt]
|
| +
|
| + resultlist = []
|
| + for e in matchOrder:
|
| + loc,results = e._parse(instring,loc,doActions)
|
| + resultlist.append(results)
|
| +
|
| + finalResults = sum(resultlist, ParseResults([]))
|
| + return loc, finalResults
|
| +
|
| + def __str__( self ):
|
| + if hasattr(self,"name"):
|
| + return self.name
|
| +
|
| + if self.strRepr is None:
|
| + self.strRepr = "{" + " & ".join(_ustr(e) for e in self.exprs) + "}"
|
| +
|
| + return self.strRepr
|
| +
|
| + def checkRecursion( self, parseElementList ):
|
| + subRecCheckList = parseElementList[:] + [ self ]
|
| + for e in self.exprs:
|
| + e.checkRecursion( subRecCheckList )
|
| +
|
| +
|
| +class ParseElementEnhance(ParserElement):
|
| + """
|
| + Abstract subclass of C{ParserElement}, for combining and post-processing parsed tokens.
|
| + """
|
| + def __init__( self, expr, savelist=False ):
|
| + super(ParseElementEnhance,self).__init__(savelist)
|
| + if isinstance( expr, basestring ):
|
| + if issubclass(ParserElement._literalStringClass, Token):
|
| + expr = ParserElement._literalStringClass(expr)
|
| + else:
|
| + expr = ParserElement._literalStringClass(Literal(expr))
|
| + self.expr = expr
|
| + self.strRepr = None
|
| + if expr is not None:
|
| + self.mayIndexError = expr.mayIndexError
|
| + self.mayReturnEmpty = expr.mayReturnEmpty
|
| + self.setWhitespaceChars( expr.whiteChars )
|
| + self.skipWhitespace = expr.skipWhitespace
|
| + self.saveAsList = expr.saveAsList
|
| + self.callPreparse = expr.callPreparse
|
| + self.ignoreExprs.extend(expr.ignoreExprs)
|
| +
|
| + def parseImpl( self, instring, loc, doActions=True ):
|
| + if self.expr is not None:
|
| + return self.expr._parse( instring, loc, doActions, callPreParse=False )
|
| + else:
|
| + raise ParseException("",loc,self.errmsg,self)
|
| +
|
| + def leaveWhitespace( self ):
|
| + self.skipWhitespace = False
|
| + self.expr = self.expr.copy()
|
| + if self.expr is not None:
|
| + self.expr.leaveWhitespace()
|
| + return self
|
| +
|
| + def ignore( self, other ):
|
| + if isinstance( other, Suppress ):
|
| + if other not in self.ignoreExprs:
|
| + super( ParseElementEnhance, self).ignore( other )
|
| + if self.expr is not None:
|
| + self.expr.ignore( self.ignoreExprs[-1] )
|
| + else:
|
| + super( ParseElementEnhance, self).ignore( other )
|
| + if self.expr is not None:
|
| + self.expr.ignore( self.ignoreExprs[-1] )
|
| + return self
|
| +
|
| + def streamline( self ):
|
| + super(ParseElementEnhance,self).streamline()
|
| + if self.expr is not None:
|
| + self.expr.streamline()
|
| + return self
|
| +
|
| + def checkRecursion( self, parseElementList ):
|
| + if self in parseElementList:
|
| + raise RecursiveGrammarException( parseElementList+[self] )
|
| + subRecCheckList = parseElementList[:] + [ self ]
|
| + if self.expr is not None:
|
| + self.expr.checkRecursion( subRecCheckList )
|
| +
|
| + def validate( self, validateTrace=[] ):
|
| + tmp = validateTrace[:]+[self]
|
| + if self.expr is not None:
|
| + self.expr.validate(tmp)
|
| + self.checkRecursion( [] )
|
| +
|
| + def __str__( self ):
|
| + try:
|
| + return super(ParseElementEnhance,self).__str__()
|
| + except Exception:
|
| + pass
|
| +
|
| + if self.strRepr is None and self.expr is not None:
|
| + self.strRepr = "%s:(%s)" % ( self.__class__.__name__, _ustr(self.expr) )
|
| + return self.strRepr
|
| +
|
| +
|
| +class FollowedBy(ParseElementEnhance):
|
| + """
|
| + Lookahead matching of the given parse expression. C{FollowedBy}
|
| + does I{not} advance the parsing position within the input string, it only
|
| + verifies that the specified parse expression matches at the current
|
| + position. C{FollowedBy} always returns a null token list.
|
| +
|
| + Example::
|
| + # use FollowedBy to match a label only if it is followed by a ':'
|
| + data_word = Word(alphas)
|
| + label = data_word + FollowedBy(':')
|
| + attr_expr = Group(label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join))
|
| +
|
| + OneOrMore(attr_expr).parseString("shape: SQUARE color: BLACK posn: upper left").pprint()
|
| + prints::
|
| + [['shape', 'SQUARE'], ['color', 'BLACK'], ['posn', 'upper left']]
|
| + """
|
| + def __init__( self, expr ):
|
| + super(FollowedBy,self).__init__(expr)
|
| + self.mayReturnEmpty = True
|
| +
|
| + def parseImpl( self, instring, loc, doActions=True ):
|
| + self.expr.tryParse( instring, loc )
|
| + return loc, []
|
| +
|
| +
|
| +class NotAny(ParseElementEnhance):
|
| + """
|
| + Lookahead to disallow matching with the given parse expression. C{NotAny}
|
| + does I{not} advance the parsing position within the input string, it only
|
| + verifies that the specified parse expression does I{not} match at the current
|
| + position. Also, C{NotAny} does I{not} skip over leading whitespace. C{NotAny}
|
| + always returns a null token list. May be constructed using the '~' operator.
|
| +
|
| + Example::
|
| +
|
| + """
|
| + def __init__( self, expr ):
|
| + super(NotAny,self).__init__(expr)
|
| + #~ self.leaveWhitespace()
|
| + self.skipWhitespace = False # do NOT use self.leaveWhitespace(), don't want to propagate to exprs
|
| + self.mayReturnEmpty = True
|
| + self.errmsg = "Found unwanted token, "+_ustr(self.expr)
|
| +
|
| + def parseImpl( self, instring, loc, doActions=True ):
|
| + if self.expr.canParseNext(instring, loc):
|
| + raise ParseException(instring, loc, self.errmsg, self)
|
| + return loc, []
|
| +
|
| + def __str__( self ):
|
| + if hasattr(self,"name"):
|
| + return self.name
|
| +
|
| + if self.strRepr is None:
|
| + self.strRepr = "~{" + _ustr(self.expr) + "}"
|
| +
|
| + return self.strRepr
|
| +
|
| +class _MultipleMatch(ParseElementEnhance):
|
| + def __init__( self, expr, stopOn=None):
|
| + super(_MultipleMatch, self).__init__(expr)
|
| + self.saveAsList = True
|
| + ender = stopOn
|
| + if isinstance(ender, basestring):
|
| + ender = ParserElement._literalStringClass(ender)
|
| + self.not_ender = ~ender if ender is not None else None
|
| +
|
| + def parseImpl( self, instring, loc, doActions=True ):
|
| + self_expr_parse = self.expr._parse
|
| + self_skip_ignorables = self._skipIgnorables
|
| + check_ender = self.not_ender is not None
|
| + if check_ender:
|
| + try_not_ender = self.not_ender.tryParse
|
| +
|
| + # must be at least one (but first see if we are the stopOn sentinel;
|
| + # if so, fail)
|
| + if check_ender:
|
| + try_not_ender(instring, loc)
|
| + loc, tokens = self_expr_parse( instring, loc, doActions, callPreParse=False )
|
| + try:
|
| + hasIgnoreExprs = (not not self.ignoreExprs)
|
| + while 1:
|
| + if check_ender:
|
| + try_not_ender(instring, loc)
|
| + if hasIgnoreExprs:
|
| + preloc = self_skip_ignorables( instring, loc )
|
| + else:
|
| + preloc = loc
|
| + loc, tmptokens = self_expr_parse( instring, preloc, doActions )
|
| + if tmptokens or tmptokens.haskeys():
|
| + tokens += tmptokens
|
| + except (ParseException,IndexError):
|
| + pass
|
| +
|
| + return loc, tokens
|
| +
|
| +class OneOrMore(_MultipleMatch):
|
| + """
|
| + Repetition of one or more of the given expression.
|
| +
|
| + Parameters:
|
| + - expr - expression that must match one or more times
|
| + - stopOn - (default=C{None}) - expression for a terminating sentinel
|
| + (only required if the sentinel would ordinarily match the repetition
|
| + expression)
|
| +
|
| + Example::
|
| + data_word = Word(alphas)
|
| + label = data_word + FollowedBy(':')
|
| + attr_expr = Group(label + Suppress(':') + OneOrMore(data_word).setParseAction(' '.join))
|
| +
|
| + text = "shape: SQUARE posn: upper left color: BLACK"
|
| + OneOrMore(attr_expr).parseString(text).pprint() # Fail! read 'color' as data instead of next label -> [['shape', 'SQUARE color']]
|
| +
|
| + # use stopOn attribute for OneOrMore to avoid reading label string as part of the data
|
| + attr_expr = Group(label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join))
|
| + OneOrMore(attr_expr).parseString(text).pprint() # Better -> [['shape', 'SQUARE'], ['posn', 'upper left'], ['color', 'BLACK']]
|
| +
|
| + # could also be written as
|
| + (attr_expr * (1,)).parseString(text).pprint()
|
| + """
|
| +
|
| + def __str__( self ):
|
| + if hasattr(self,"name"):
|
| + return self.name
|
| +
|
| + if self.strRepr is None:
|
| + self.strRepr = "{" + _ustr(self.expr) + "}..."
|
| +
|
| + return self.strRepr
|
| +
|
| +class ZeroOrMore(_MultipleMatch):
|
| + """
|
| + Optional repetition of zero or more of the given expression.
|
| +
|
| + Parameters:
|
| + - expr - expression that must match zero or more times
|
| + - stopOn - (default=C{None}) - expression for a terminating sentinel
|
| + (only required if the sentinel would ordinarily match the repetition
|
| + expression)
|
| +
|
| + Example: similar to L{OneOrMore}
|
| + """
|
| + def __init__( self, expr, stopOn=None):
|
| + super(ZeroOrMore,self).__init__(expr, stopOn=stopOn)
|
| + self.mayReturnEmpty = True
|
| +
|
| + def parseImpl( self, instring, loc, doActions=True ):
|
| + try:
|
| + return super(ZeroOrMore, self).parseImpl(instring, loc, doActions)
|
| + except (ParseException,IndexError):
|
| + return loc, []
|
| +
|
| + def __str__( self ):
|
| + if hasattr(self,"name"):
|
| + return self.name
|
| +
|
| + if self.strRepr is None:
|
| + self.strRepr = "[" + _ustr(self.expr) + "]..."
|
| +
|
| + return self.strRepr
|
| +
|
| +class _NullToken(object):
|
| + def __bool__(self):
|
| + return False
|
| + __nonzero__ = __bool__
|
| + def __str__(self):
|
| + return ""
|
| +
|
| +_optionalNotMatched = _NullToken()
|
| +class Optional(ParseElementEnhance):
|
| + """
|
| + Optional matching of the given expression.
|
| +
|
| + Parameters:
|
| + - expr - expression that must match zero or more times
|
| + - default (optional) - value to be returned if the optional expression is not found.
|
| +
|
| + Example::
|
| + # US postal code can be a 5-digit zip, plus optional 4-digit qualifier
|
| + zip = Combine(Word(nums, exact=5) + Optional('-' + Word(nums, exact=4)))
|
| + zip.runTests('''
|
| + # traditional ZIP code
|
| + 12345
|
| +
|
| + # ZIP+4 form
|
| + 12101-0001
|
| +
|
| + # invalid ZIP
|
| + 98765-
|
| + ''')
|
| + prints::
|
| + # traditional ZIP code
|
| + 12345
|
| + ['12345']
|
| +
|
| + # ZIP+4 form
|
| + 12101-0001
|
| + ['12101-0001']
|
| +
|
| + # invalid ZIP
|
| + 98765-
|
| + ^
|
| + FAIL: Expected end of text (at char 5), (line:1, col:6)
|
| + """
|
| + def __init__( self, expr, default=_optionalNotMatched ):
|
| + super(Optional,self).__init__( expr, savelist=False )
|
| + self.saveAsList = self.expr.saveAsList
|
| + self.defaultValue = default
|
| + self.mayReturnEmpty = True
|
| +
|
| + def parseImpl( self, instring, loc, doActions=True ):
|
| + try:
|
| + loc, tokens = self.expr._parse( instring, loc, doActions, callPreParse=False )
|
| + except (ParseException,IndexError):
|
| + if self.defaultValue is not _optionalNotMatched:
|
| + if self.expr.resultsName:
|
| + tokens = ParseResults([ self.defaultValue ])
|
| + tokens[self.expr.resultsName] = self.defaultValue
|
| + else:
|
| + tokens = [ self.defaultValue ]
|
| + else:
|
| + tokens = []
|
| + return loc, tokens
|
| +
|
| + def __str__( self ):
|
| + if hasattr(self,"name"):
|
| + return self.name
|
| +
|
| + if self.strRepr is None:
|
| + self.strRepr = "[" + _ustr(self.expr) + "]"
|
| +
|
| + return self.strRepr
|
| +
|
| +class SkipTo(ParseElementEnhance):
|
| + """
|
| + Token for skipping over all undefined text until the matched expression is found.
|
| +
|
| + Parameters:
|
| + - expr - target expression marking the end of the data to be skipped
|
| + - include - (default=C{False}) if True, the target expression is also parsed
|
| + (the skipped text and target expression are returned as a 2-element list).
|
| + - ignore - (default=C{None}) used to define grammars (typically quoted strings and
|
| + comments) that might contain false matches to the target expression
|
| + - failOn - (default=C{None}) define expressions that are not allowed to be
|
| + included in the skipped test; if found before the target expression is found,
|
| + the SkipTo is not a match
|
| +
|
| + Example::
|
| + report = '''
|
| + Outstanding Issues Report - 1 Jan 2000
|
| +
|
| + # | Severity | Description | Days Open
|
| + -----+----------+-------------------------------------------+-----------
|
| + 101 | Critical | Intermittent system crash | 6
|
| + 94 | Cosmetic | Spelling error on Login ('log|n') | 14
|
| + 79 | Minor | System slow when running too many reports | 47
|
| + '''
|
| + integer = Word(nums)
|
| + SEP = Suppress('|')
|
| + # use SkipTo to simply match everything up until the next SEP
|
| + # - ignore quoted strings, so that a '|' character inside a quoted string does not match
|
| + # - parse action will call token.strip() for each matched token, i.e., the description body
|
| + string_data = SkipTo(SEP, ignore=quotedString)
|
| + string_data.setParseAction(tokenMap(str.strip))
|
| + ticket_expr = (integer("issue_num") + SEP
|
| + + string_data("sev") + SEP
|
| + + string_data("desc") + SEP
|
| + + integer("days_open"))
|
| +
|
| + for tkt in ticket_expr.searchString(report):
|
| + print tkt.dump()
|
| + prints::
|
| + ['101', 'Critical', 'Intermittent system crash', '6']
|
| + - days_open: 6
|
| + - desc: Intermittent system crash
|
| + - issue_num: 101
|
| + - sev: Critical
|
| + ['94', 'Cosmetic', "Spelling error on Login ('log|n')", '14']
|
| + - days_open: 14
|
| + - desc: Spelling error on Login ('log|n')
|
| + - issue_num: 94
|
| + - sev: Cosmetic
|
| + ['79', 'Minor', 'System slow when running too many reports', '47']
|
| + - days_open: 47
|
| + - desc: System slow when running too many reports
|
| + - issue_num: 79
|
| + - sev: Minor
|
| + """
|
| + def __init__( self, other, include=False, ignore=None, failOn=None ):
|
| + super( SkipTo, self ).__init__( other )
|
| + self.ignoreExpr = ignore
|
| + self.mayReturnEmpty = True
|
| + self.mayIndexError = False
|
| + self.includeMatch = include
|
| + self.asList = False
|
| + if isinstance(failOn, basestring):
|
| + self.failOn = ParserElement._literalStringClass(failOn)
|
| + else:
|
| + self.failOn = failOn
|
| + self.errmsg = "No match found for "+_ustr(self.expr)
|
| +
|
| + def parseImpl( self, instring, loc, doActions=True ):
|
| + startloc = loc
|
| + instrlen = len(instring)
|
| + expr = self.expr
|
| + expr_parse = self.expr._parse
|
| + self_failOn_canParseNext = self.failOn.canParseNext if self.failOn is not None else None
|
| + self_ignoreExpr_tryParse = self.ignoreExpr.tryParse if self.ignoreExpr is not None else None
|
| +
|
| + tmploc = loc
|
| + while tmploc <= instrlen:
|
| + if self_failOn_canParseNext is not None:
|
| + # break if failOn expression matches
|
| + if self_failOn_canParseNext(instring, tmploc):
|
| + break
|
| +
|
| + if self_ignoreExpr_tryParse is not None:
|
| + # advance past ignore expressions
|
| + while 1:
|
| + try:
|
| + tmploc = self_ignoreExpr_tryParse(instring, tmploc)
|
| + except ParseBaseException:
|
| + break
|
| +
|
| + try:
|
| + expr_parse(instring, tmploc, doActions=False, callPreParse=False)
|
| + except (ParseException, IndexError):
|
| + # no match, advance loc in string
|
| + tmploc += 1
|
| + else:
|
| + # matched skipto expr, done
|
| + break
|
| +
|
| + else:
|
| + # ran off the end of the input string without matching skipto expr, fail
|
| + raise ParseException(instring, loc, self.errmsg, self)
|
| +
|
| + # build up return values
|
| + loc = tmploc
|
| + skiptext = instring[startloc:loc]
|
| + skipresult = ParseResults(skiptext)
|
| +
|
| + if self.includeMatch:
|
| + loc, mat = expr_parse(instring,loc,doActions,callPreParse=False)
|
| + skipresult += mat
|
| +
|
| + return loc, skipresult
|
| +
|
| +class Forward(ParseElementEnhance):
|
| + """
|
| + Forward declaration of an expression to be defined later -
|
| + used for recursive grammars, such as algebraic infix notation.
|
| + When the expression is known, it is assigned to the C{Forward} variable using the '<<' operator.
|
| +
|
| + Note: take care when assigning to C{Forward} not to overlook precedence of operators.
|
| + Specifically, '|' has a lower precedence than '<<', so that::
|
| + fwdExpr << a | b | c
|
| + will actually be evaluated as::
|
| + (fwdExpr << a) | b | c
|
| + thereby leaving b and c out as parseable alternatives. It is recommended that you
|
| + explicitly group the values inserted into the C{Forward}::
|
| + fwdExpr << (a | b | c)
|
| + Converting to use the '<<=' operator instead will avoid this problem.
|
| +
|
| + See L{ParseResults.pprint} for an example of a recursive parser created using
|
| + C{Forward}.
|
| + """
|
| + def __init__( self, other=None ):
|
| + super(Forward,self).__init__( other, savelist=False )
|
| +
|
| + def __lshift__( self, other ):
|
| + if isinstance( other, basestring ):
|
| + other = ParserElement._literalStringClass(other)
|
| + self.expr = other
|
| + self.strRepr = None
|
| + self.mayIndexError = self.expr.mayIndexError
|
| + self.mayReturnEmpty = self.expr.mayReturnEmpty
|
| + self.setWhitespaceChars( self.expr.whiteChars )
|
| + self.skipWhitespace = self.expr.skipWhitespace
|
| + self.saveAsList = self.expr.saveAsList
|
| + self.ignoreExprs.extend(self.expr.ignoreExprs)
|
| + return self
|
| +
|
| + def __ilshift__(self, other):
|
| + return self << other
|
| +
|
| + def leaveWhitespace( self ):
|
| + self.skipWhitespace = False
|
| + return self
|
| +
|
| + def streamline( self ):
|
| + if not self.streamlined:
|
| + self.streamlined = True
|
| + if self.expr is not None:
|
| + self.expr.streamline()
|
| + return self
|
| +
|
| + def validate( self, validateTrace=[] ):
|
| + if self not in validateTrace:
|
| + tmp = validateTrace[:]+[self]
|
| + if self.expr is not None:
|
| + self.expr.validate(tmp)
|
| + self.checkRecursion([])
|
| +
|
| + def __str__( self ):
|
| + if hasattr(self,"name"):
|
| + return self.name
|
| + return self.__class__.__name__ + ": ..."
|
| +
|
| + # stubbed out for now - creates awful memory and perf issues
|
| + self._revertClass = self.__class__
|
| + self.__class__ = _ForwardNoRecurse
|
| + try:
|
| + if self.expr is not None:
|
| + retString = _ustr(self.expr)
|
| + else:
|
| + retString = "None"
|
| + finally:
|
| + self.__class__ = self._revertClass
|
| + return self.__class__.__name__ + ": " + retString
|
| +
|
| + def copy(self):
|
| + if self.expr is not None:
|
| + return super(Forward,self).copy()
|
| + else:
|
| + ret = Forward()
|
| + ret <<= self
|
| + return ret
|
| +
|
| +class _ForwardNoRecurse(Forward):
|
| + def __str__( self ):
|
| + return "..."
|
| +
|
| +class TokenConverter(ParseElementEnhance):
|
| + """
|
| + Abstract subclass of C{ParseExpression}, for converting parsed results.
|
| + """
|
| + def __init__( self, expr, savelist=False ):
|
| + super(TokenConverter,self).__init__( expr )#, savelist )
|
| + self.saveAsList = False
|
| +
|
| +class Combine(TokenConverter):
|
| + """
|
| + Converter to concatenate all matching tokens to a single string.
|
| + By default, the matching patterns must also be contiguous in the input string;
|
| + this can be disabled by specifying C{'adjacent=False'} in the constructor.
|
| +
|
| + Example::
|
| + real = Word(nums) + '.' + Word(nums)
|
| + print(real.parseString('3.1416')) # -> ['3', '.', '1416']
|
| + # will also erroneously match the following
|
| + print(real.parseString('3. 1416')) # -> ['3', '.', '1416']
|
| +
|
| + real = Combine(Word(nums) + '.' + Word(nums))
|
| + print(real.parseString('3.1416')) # -> ['3.1416']
|
| + # no match when there are internal spaces
|
| + print(real.parseString('3. 1416')) # -> Exception: Expected W:(0123...)
|
| + """
|
| + def __init__( self, expr, joinString="", adjacent=True ):
|
| + super(Combine,self).__init__( expr )
|
| + # suppress whitespace-stripping in contained parse expressions, but re-enable it on the Combine itself
|
| + if adjacent:
|
| + self.leaveWhitespace()
|
| + self.adjacent = adjacent
|
| + self.skipWhitespace = True
|
| + self.joinString = joinString
|
| + self.callPreparse = True
|
| +
|
| + def ignore( self, other ):
|
| + if self.adjacent:
|
| + ParserElement.ignore(self, other)
|
| + else:
|
| + super( Combine, self).ignore( other )
|
| + return self
|
| +
|
| + def postParse( self, instring, loc, tokenlist ):
|
| + retToks = tokenlist.copy()
|
| + del retToks[:]
|
| + retToks += ParseResults([ "".join(tokenlist._asStringList(self.joinString)) ], modal=self.modalResults)
|
| +
|
| + if self.resultsName and retToks.haskeys():
|
| + return [ retToks ]
|
| + else:
|
| + return retToks
|
| +
|
| +class Group(TokenConverter):
|
| + """
|
| + Converter to return the matched tokens as a list - useful for returning tokens of C{L{ZeroOrMore}} and C{L{OneOrMore}} expressions.
|
| +
|
| + Example::
|
| + ident = Word(alphas)
|
| + num = Word(nums)
|
| + term = ident | num
|
| + func = ident + Optional(delimitedList(term))
|
| + print(func.parseString("fn a,b,100")) # -> ['fn', 'a', 'b', '100']
|
| +
|
| + func = ident + Group(Optional(delimitedList(term)))
|
| + print(func.parseString("fn a,b,100")) # -> ['fn', ['a', 'b', '100']]
|
| + """
|
| + def __init__( self, expr ):
|
| + super(Group,self).__init__( expr )
|
| + self.saveAsList = True
|
| +
|
| + def postParse( self, instring, loc, tokenlist ):
|
| + return [ tokenlist ]
|
| +
|
| +class Dict(TokenConverter):
|
| + """
|
| + Converter to return a repetitive expression as a list, but also as a dictionary.
|
| + Each element can also be referenced using the first token in the expression as its key.
|
| + Useful for tabular report scraping when the first column can be used as a item key.
|
| +
|
| + Example::
|
| + data_word = Word(alphas)
|
| + label = data_word + FollowedBy(':')
|
| + attr_expr = Group(label + Suppress(':') + OneOrMore(data_word).setParseAction(' '.join))
|
| +
|
| + text = "shape: SQUARE posn: upper left color: light blue texture: burlap"
|
| + attr_expr = (label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join))
|
| +
|
| + # print attributes as plain groups
|
| + print(OneOrMore(attr_expr).parseString(text).dump())
|
| +
|
| + # instead of OneOrMore(expr), parse using Dict(OneOrMore(Group(expr))) - Dict will auto-assign names
|
| + result = Dict(OneOrMore(Group(attr_expr))).parseString(text)
|
| + print(result.dump())
|
| +
|
| + # access named fields as dict entries, or output as dict
|
| + print(result['shape'])
|
| + print(result.asDict())
|
| + prints::
|
| + ['shape', 'SQUARE', 'posn', 'upper left', 'color', 'light blue', 'texture', 'burlap']
|
| +
|
| + [['shape', 'SQUARE'], ['posn', 'upper left'], ['color', 'light blue'], ['texture', 'burlap']]
|
| + - color: light blue
|
| + - posn: upper left
|
| + - shape: SQUARE
|
| + - texture: burlap
|
| + SQUARE
|
| + {'color': 'light blue', 'posn': 'upper left', 'texture': 'burlap', 'shape': 'SQUARE'}
|
| + See more examples at L{ParseResults} of accessing fields by results name.
|
| + """
|
| + def __init__( self, expr ):
|
| + super(Dict,self).__init__( expr )
|
| + self.saveAsList = True
|
| +
|
| + def postParse( self, instring, loc, tokenlist ):
|
| + for i,tok in enumerate(tokenlist):
|
| + if len(tok) == 0:
|
| + continue
|
| + ikey = tok[0]
|
| + if isinstance(ikey,int):
|
| + ikey = _ustr(tok[0]).strip()
|
| + if len(tok)==1:
|
| + tokenlist[ikey] = _ParseResultsWithOffset("",i)
|
| + elif len(tok)==2 and not isinstance(tok[1],ParseResults):
|
| + tokenlist[ikey] = _ParseResultsWithOffset(tok[1],i)
|
| + else:
|
| + dictvalue = tok.copy() #ParseResults(i)
|
| + del dictvalue[0]
|
| + if len(dictvalue)!= 1 or (isinstance(dictvalue,ParseResults) and dictvalue.haskeys()):
|
| + tokenlist[ikey] = _ParseResultsWithOffset(dictvalue,i)
|
| + else:
|
| + tokenlist[ikey] = _ParseResultsWithOffset(dictvalue[0],i)
|
| +
|
| + if self.resultsName:
|
| + return [ tokenlist ]
|
| + else:
|
| + return tokenlist
|
| +
|
| +
|
| +class Suppress(TokenConverter):
|
| + """
|
| + Converter for ignoring the results of a parsed expression.
|
| +
|
| + Example::
|
| + source = "a, b, c,d"
|
| + wd = Word(alphas)
|
| + wd_list1 = wd + ZeroOrMore(',' + wd)
|
| + print(wd_list1.parseString(source))
|
| +
|
| + # often, delimiters that are useful during parsing are just in the
|
| + # way afterward - use Suppress to keep them out of the parsed output
|
| + wd_list2 = wd + ZeroOrMore(Suppress(',') + wd)
|
| + print(wd_list2.parseString(source))
|
| + prints::
|
| + ['a', ',', 'b', ',', 'c', ',', 'd']
|
| + ['a', 'b', 'c', 'd']
|
| + (See also L{delimitedList}.)
|
| + """
|
| + def postParse( self, instring, loc, tokenlist ):
|
| + return []
|
| +
|
| + def suppress( self ):
|
| + return self
|
| +
|
| +
|
| +class OnlyOnce(object):
|
| + """
|
| + Wrapper for parse actions, to ensure they are only called once.
|
| + """
|
| + def __init__(self, methodCall):
|
| + self.callable = _trim_arity(methodCall)
|
| + self.called = False
|
| + def __call__(self,s,l,t):
|
| + if not self.called:
|
| + results = self.callable(s,l,t)
|
| + self.called = True
|
| + return results
|
| + raise ParseException(s,l,"")
|
| + def reset(self):
|
| + self.called = False
|
| +
|
| +def traceParseAction(f):
|
| + """
|
| + Decorator for debugging parse actions.
|
| +
|
| + When the parse action is called, this decorator will print C{">> entering I{method-name}(line:I{current_source_line}, I{parse_location}, I{matched_tokens})".}
|
| + When the parse action completes, the decorator will print C{"<<"} followed by the returned value, or any exception that the parse action raised.
|
| +
|
| + Example::
|
| + wd = Word(alphas)
|
| +
|
| + @traceParseAction
|
| + def remove_duplicate_chars(tokens):
|
| + return ''.join(sorted(set(''.join(tokens)))
|
| +
|
| + wds = OneOrMore(wd).setParseAction(remove_duplicate_chars)
|
| + print(wds.parseString("slkdjs sld sldd sdlf sdljf"))
|
| + prints::
|
| + >>entering remove_duplicate_chars(line: 'slkdjs sld sldd sdlf sdljf', 0, (['slkdjs', 'sld', 'sldd', 'sdlf', 'sdljf'], {}))
|
| + <<leaving remove_duplicate_chars (ret: 'dfjkls')
|
| + ['dfjkls']
|
| + """
|
| + f = _trim_arity(f)
|
| + def z(*paArgs):
|
| + thisFunc = f.__name__
|
| + s,l,t = paArgs[-3:]
|
| + if len(paArgs)>3:
|
| + thisFunc = paArgs[0].__class__.__name__ + '.' + thisFunc
|
| + sys.stderr.write( ">>entering %s(line: '%s', %d, %r)\n" % (thisFunc,line(l,s),l,t) )
|
| + try:
|
| + ret = f(*paArgs)
|
| + except Exception as exc:
|
| + sys.stderr.write( "<<leaving %s (exception: %s)\n" % (thisFunc,exc) )
|
| + raise
|
| + sys.stderr.write( "<<leaving %s (ret: %r)\n" % (thisFunc,ret) )
|
| + return ret
|
| + try:
|
| + z.__name__ = f.__name__
|
| + except AttributeError:
|
| + pass
|
| + return z
|
| +
|
| +#
|
| +# global helpers
|
| +#
|
| +def delimitedList( expr, delim=",", combine=False ):
|
| + """
|
| + Helper to define a delimited list of expressions - the delimiter defaults to ','.
|
| + By default, the list elements and delimiters can have intervening whitespace, and
|
| + comments, but this can be overridden by passing C{combine=True} in the constructor.
|
| + If C{combine} is set to C{True}, the matching tokens are returned as a single token
|
| + string, with the delimiters included; otherwise, the matching tokens are returned
|
| + as a list of tokens, with the delimiters suppressed.
|
| +
|
| + Example::
|
| + delimitedList(Word(alphas)).parseString("aa,bb,cc") # -> ['aa', 'bb', 'cc']
|
| + delimitedList(Word(hexnums), delim=':', combine=True).parseString("AA:BB:CC:DD:EE") # -> ['AA:BB:CC:DD:EE']
|
| + """
|
| + dlName = _ustr(expr)+" ["+_ustr(delim)+" "+_ustr(expr)+"]..."
|
| + if combine:
|
| + return Combine( expr + ZeroOrMore( delim + expr ) ).setName(dlName)
|
| + else:
|
| + return ( expr + ZeroOrMore( Suppress( delim ) + expr ) ).setName(dlName)
|
| +
|
| +def countedArray( expr, intExpr=None ):
|
| + """
|
| + Helper to define a counted list of expressions.
|
| + This helper defines a pattern of the form::
|
| + integer expr expr expr...
|
| + where the leading integer tells how many expr expressions follow.
|
| + The matched tokens returns the array of expr tokens as a list - the leading count token is suppressed.
|
| +
|
| + If C{intExpr} is specified, it should be a pyparsing expression that produces an integer value.
|
| +
|
| + Example::
|
| + countedArray(Word(alphas)).parseString('2 ab cd ef') # -> ['ab', 'cd']
|
| +
|
| + # in this parser, the leading integer value is given in binary,
|
| + # '10' indicating that 2 values are in the array
|
| + binaryConstant = Word('01').setParseAction(lambda t: int(t[0], 2))
|
| + countedArray(Word(alphas), intExpr=binaryConstant).parseString('10 ab cd ef') # -> ['ab', 'cd']
|
| + """
|
| + arrayExpr = Forward()
|
| + def countFieldParseAction(s,l,t):
|
| + n = t[0]
|
| + arrayExpr << (n and Group(And([expr]*n)) or Group(empty))
|
| + return []
|
| + if intExpr is None:
|
| + intExpr = Word(nums).setParseAction(lambda t:int(t[0]))
|
| + else:
|
| + intExpr = intExpr.copy()
|
| + intExpr.setName("arrayLen")
|
| + intExpr.addParseAction(countFieldParseAction, callDuringTry=True)
|
| + return ( intExpr + arrayExpr ).setName('(len) ' + _ustr(expr) + '...')
|
| +
|
| +def _flatten(L):
|
| + ret = []
|
| + for i in L:
|
| + if isinstance(i,list):
|
| + ret.extend(_flatten(i))
|
| + else:
|
| + ret.append(i)
|
| + return ret
|
| +
|
| +def matchPreviousLiteral(expr):
|
| + """
|
| + Helper to define an expression that is indirectly defined from
|
| + the tokens matched in a previous expression, that is, it looks
|
| + for a 'repeat' of a previous expression. For example::
|
| + first = Word(nums)
|
| + second = matchPreviousLiteral(first)
|
| + matchExpr = first + ":" + second
|
| + will match C{"1:1"}, but not C{"1:2"}. Because this matches a
|
| + previous literal, will also match the leading C{"1:1"} in C{"1:10"}.
|
| + If this is not desired, use C{matchPreviousExpr}.
|
| + Do I{not} use with packrat parsing enabled.
|
| + """
|
| + rep = Forward()
|
| + def copyTokenToRepeater(s,l,t):
|
| + if t:
|
| + if len(t) == 1:
|
| + rep << t[0]
|
| + else:
|
| + # flatten t tokens
|
| + tflat = _flatten(t.asList())
|
| + rep << And(Literal(tt) for tt in tflat)
|
| + else:
|
| + rep << Empty()
|
| + expr.addParseAction(copyTokenToRepeater, callDuringTry=True)
|
| + rep.setName('(prev) ' + _ustr(expr))
|
| + return rep
|
| +
|
| +def matchPreviousExpr(expr):
|
| + """
|
| + Helper to define an expression that is indirectly defined from
|
| + the tokens matched in a previous expression, that is, it looks
|
| + for a 'repeat' of a previous expression. For example::
|
| + first = Word(nums)
|
| + second = matchPreviousExpr(first)
|
| + matchExpr = first + ":" + second
|
| + will match C{"1:1"}, but not C{"1:2"}. Because this matches by
|
| + expressions, will I{not} match the leading C{"1:1"} in C{"1:10"};
|
| + the expressions are evaluated first, and then compared, so
|
| + C{"1"} is compared with C{"10"}.
|
| + Do I{not} use with packrat parsing enabled.
|
| + """
|
| + rep = Forward()
|
| + e2 = expr.copy()
|
| + rep <<= e2
|
| + def copyTokenToRepeater(s,l,t):
|
| + matchTokens = _flatten(t.asList())
|
| + def mustMatchTheseTokens(s,l,t):
|
| + theseTokens = _flatten(t.asList())
|
| + if theseTokens != matchTokens:
|
| + raise ParseException("",0,"")
|
| + rep.setParseAction( mustMatchTheseTokens, callDuringTry=True )
|
| + expr.addParseAction(copyTokenToRepeater, callDuringTry=True)
|
| + rep.setName('(prev) ' + _ustr(expr))
|
| + return rep
|
| +
|
| +def _escapeRegexRangeChars(s):
|
| + #~ escape these chars: ^-]
|
| + for c in r"\^-]":
|
| + s = s.replace(c,_bslash+c)
|
| + s = s.replace("\n",r"\n")
|
| + s = s.replace("\t",r"\t")
|
| + return _ustr(s)
|
| +
|
| +def oneOf( strs, caseless=False, useRegex=True ):
|
| + """
|
| + Helper to quickly define a set of alternative Literals, and makes sure to do
|
| + longest-first testing when there is a conflict, regardless of the input order,
|
| + but returns a C{L{MatchFirst}} for best performance.
|
| +
|
| + Parameters:
|
| + - strs - a string of space-delimited literals, or a collection of string literals
|
| + - caseless - (default=C{False}) - treat all literals as caseless
|
| + - useRegex - (default=C{True}) - as an optimization, will generate a Regex
|
| + object; otherwise, will generate a C{MatchFirst} object (if C{caseless=True}, or
|
| + if creating a C{Regex} raises an exception)
|
| +
|
| + Example::
|
| + comp_oper = oneOf("< = > <= >= !=")
|
| + var = Word(alphas)
|
| + number = Word(nums)
|
| + term = var | number
|
| + comparison_expr = term + comp_oper + term
|
| + print(comparison_expr.searchString("B = 12 AA=23 B<=AA AA>12"))
|
| + prints::
|
| + [['B', '=', '12'], ['AA', '=', '23'], ['B', '<=', 'AA'], ['AA', '>', '12']]
|
| + """
|
| + if caseless:
|
| + isequal = ( lambda a,b: a.upper() == b.upper() )
|
| + masks = ( lambda a,b: b.upper().startswith(a.upper()) )
|
| + parseElementClass = CaselessLiteral
|
| + else:
|
| + isequal = ( lambda a,b: a == b )
|
| + masks = ( lambda a,b: b.startswith(a) )
|
| + parseElementClass = Literal
|
| +
|
| + symbols = []
|
| + if isinstance(strs,basestring):
|
| + symbols = strs.split()
|
| + elif isinstance(strs, collections.Iterable):
|
| + symbols = list(strs)
|
| + else:
|
| + warnings.warn("Invalid argument to oneOf, expected string or iterable",
|
| + SyntaxWarning, stacklevel=2)
|
| + if not symbols:
|
| + return NoMatch()
|
| +
|
| + i = 0
|
| + while i < len(symbols)-1:
|
| + cur = symbols[i]
|
| + for j,other in enumerate(symbols[i+1:]):
|
| + if ( isequal(other, cur) ):
|
| + del symbols[i+j+1]
|
| + break
|
| + elif ( masks(cur, other) ):
|
| + del symbols[i+j+1]
|
| + symbols.insert(i,other)
|
| + cur = other
|
| + break
|
| + else:
|
| + i += 1
|
| +
|
| + if not caseless and useRegex:
|
| + #~ print (strs,"->", "|".join( [ _escapeRegexChars(sym) for sym in symbols] ))
|
| + try:
|
| + if len(symbols)==len("".join(symbols)):
|
| + return Regex( "[%s]" % "".join(_escapeRegexRangeChars(sym) for sym in symbols) ).setName(' | '.join(symbols))
|
| + else:
|
| + return Regex( "|".join(re.escape(sym) for sym in symbols) ).setName(' | '.join(symbols))
|
| + except Exception:
|
| + warnings.warn("Exception creating Regex for oneOf, building MatchFirst",
|
| + SyntaxWarning, stacklevel=2)
|
| +
|
| +
|
| + # last resort, just use MatchFirst
|
| + return MatchFirst(parseElementClass(sym) for sym in symbols).setName(' | '.join(symbols))
|
| +
|
| +def dictOf( key, value ):
|
| + """
|
| + Helper to easily and clearly define a dictionary by specifying the respective patterns
|
| + for the key and value. Takes care of defining the C{L{Dict}}, C{L{ZeroOrMore}}, and C{L{Group}} tokens
|
| + in the proper order. The key pattern can include delimiting markers or punctuation,
|
| + as long as they are suppressed, thereby leaving the significant key text. The value
|
| + pattern can include named results, so that the C{Dict} results can include named token
|
| + fields.
|
| +
|
| + Example::
|
| + text = "shape: SQUARE posn: upper left color: light blue texture: burlap"
|
| + attr_expr = (label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join))
|
| + print(OneOrMore(attr_expr).parseString(text).dump())
|
| +
|
| + attr_label = label
|
| + attr_value = Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join)
|
| +
|
| + # similar to Dict, but simpler call format
|
| + result = dictOf(attr_label, attr_value).parseString(text)
|
| + print(result.dump())
|
| + print(result['shape'])
|
| + print(result.shape) # object attribute access works too
|
| + print(result.asDict())
|
| + prints::
|
| + [['shape', 'SQUARE'], ['posn', 'upper left'], ['color', 'light blue'], ['texture', 'burlap']]
|
| + - color: light blue
|
| + - posn: upper left
|
| + - shape: SQUARE
|
| + - texture: burlap
|
| + SQUARE
|
| + SQUARE
|
| + {'color': 'light blue', 'shape': 'SQUARE', 'posn': 'upper left', 'texture': 'burlap'}
|
| + """
|
| + return Dict( ZeroOrMore( Group ( key + value ) ) )
|
| +
|
| +def originalTextFor(expr, asString=True):
|
| + """
|
| + Helper to return the original, untokenized text for a given expression. Useful to
|
| + restore the parsed fields of an HTML start tag into the raw tag text itself, or to
|
| + revert separate tokens with intervening whitespace back to the original matching
|
| + input text. By default, returns astring containing the original parsed text.
|
| +
|
| + If the optional C{asString} argument is passed as C{False}, then the return value is a
|
| + C{L{ParseResults}} containing any results names that were originally matched, and a
|
| + single token containing the original matched text from the input string. So if
|
| + the expression passed to C{L{originalTextFor}} contains expressions with defined
|
| + results names, you must set C{asString} to C{False} if you want to preserve those
|
| + results name values.
|
| +
|
| + Example::
|
| + src = "this is test <b> bold <i>text</i> </b> normal text "
|
| + for tag in ("b","i"):
|
| + opener,closer = makeHTMLTags(tag)
|
| + patt = originalTextFor(opener + SkipTo(closer) + closer)
|
| + print(patt.searchString(src)[0])
|
| + prints::
|
| + ['<b> bold <i>text</i> </b>']
|
| + ['<i>text</i>']
|
| + """
|
| + locMarker = Empty().setParseAction(lambda s,loc,t: loc)
|
| + endlocMarker = locMarker.copy()
|
| + endlocMarker.callPreparse = False
|
| + matchExpr = locMarker("_original_start") + expr + endlocMarker("_original_end")
|
| + if asString:
|
| + extractText = lambda s,l,t: s[t._original_start:t._original_end]
|
| + else:
|
| + def extractText(s,l,t):
|
| + t[:] = [s[t.pop('_original_start'):t.pop('_original_end')]]
|
| + matchExpr.setParseAction(extractText)
|
| + matchExpr.ignoreExprs = expr.ignoreExprs
|
| + return matchExpr
|
| +
|
| +def ungroup(expr):
|
| + """
|
| + Helper to undo pyparsing's default grouping of And expressions, even
|
| + if all but one are non-empty.
|
| + """
|
| + return TokenConverter(expr).setParseAction(lambda t:t[0])
|
| +
|
| +def locatedExpr(expr):
|
| + """
|
| + Helper to decorate a returned token with its starting and ending locations in the input string.
|
| + This helper adds the following results names:
|
| + - locn_start = location where matched expression begins
|
| + - locn_end = location where matched expression ends
|
| + - value = the actual parsed results
|
| +
|
| + Be careful if the input text contains C{<TAB>} characters, you may want to call
|
| + C{L{ParserElement.parseWithTabs}}
|
| +
|
| + Example::
|
| + wd = Word(alphas)
|
| + for match in locatedExpr(wd).searchString("ljsdf123lksdjjf123lkkjj1222"):
|
| + print(match)
|
| + prints::
|
| + [[0, 'ljsdf', 5]]
|
| + [[8, 'lksdjjf', 15]]
|
| + [[18, 'lkkjj', 23]]
|
| + """
|
| + locator = Empty().setParseAction(lambda s,l,t: l)
|
| + return Group(locator("locn_start") + expr("value") + locator.copy().leaveWhitespace()("locn_end"))
|
| +
|
| +
|
| +# convenience constants for positional expressions
|
| +empty = Empty().setName("empty")
|
| +lineStart = LineStart().setName("lineStart")
|
| +lineEnd = LineEnd().setName("lineEnd")
|
| +stringStart = StringStart().setName("stringStart")
|
| +stringEnd = StringEnd().setName("stringEnd")
|
| +
|
| +_escapedPunc = Word( _bslash, r"\[]-*.$+^?()~ ", exact=2 ).setParseAction(lambda s,l,t:t[0][1])
|
| +_escapedHexChar = Regex(r"\\0?[xX][0-9a-fA-F]+").setParseAction(lambda s,l,t:unichr(int(t[0].lstrip(r'\0x'),16)))
|
| +_escapedOctChar = Regex(r"\\0[0-7]+").setParseAction(lambda s,l,t:unichr(int(t[0][1:],8)))
|
| +_singleChar = _escapedPunc | _escapedHexChar | _escapedOctChar | Word(printables, excludeChars=r'\]', exact=1) | Regex(r"\w", re.UNICODE)
|
| +_charRange = Group(_singleChar + Suppress("-") + _singleChar)
|
| +_reBracketExpr = Literal("[") + Optional("^").setResultsName("negate") + Group( OneOrMore( _charRange | _singleChar ) ).setResultsName("body") + "]"
|
| +
|
| +def srange(s):
|
| + r"""
|
| + Helper to easily define string ranges for use in Word construction. Borrows
|
| + syntax from regexp '[]' string range definitions::
|
| + srange("[0-9]") -> "0123456789"
|
| + srange("[a-z]") -> "abcdefghijklmnopqrstuvwxyz"
|
| + srange("[a-z$_]") -> "abcdefghijklmnopqrstuvwxyz$_"
|
| + The input string must be enclosed in []'s, and the returned string is the expanded
|
| + character set joined into a single string.
|
| + The values enclosed in the []'s may be:
|
| + - a single character
|
| + - an escaped character with a leading backslash (such as C{\-} or C{\]})
|
| + - an escaped hex character with a leading C{'\x'} (C{\x21}, which is a C{'!'} character)
|
| + (C{\0x##} is also supported for backwards compatibility)
|
| + - an escaped octal character with a leading C{'\0'} (C{\041}, which is a C{'!'} character)
|
| + - a range of any of the above, separated by a dash (C{'a-z'}, etc.)
|
| + - any combination of the above (C{'aeiouy'}, C{'a-zA-Z0-9_$'}, etc.)
|
| + """
|
| + _expanded = lambda p: p if not isinstance(p,ParseResults) else ''.join(unichr(c) for c in range(ord(p[0]),ord(p[1])+1))
|
| + try:
|
| + return "".join(_expanded(part) for part in _reBracketExpr.parseString(s).body)
|
| + except Exception:
|
| + return ""
|
| +
|
| +def matchOnlyAtCol(n):
|
| + """
|
| + Helper method for defining parse actions that require matching at a specific
|
| + column in the input text.
|
| + """
|
| + def verifyCol(strg,locn,toks):
|
| + if col(locn,strg) != n:
|
| + raise ParseException(strg,locn,"matched token not at column %d" % n)
|
| + return verifyCol
|
| +
|
| +def replaceWith(replStr):
|
| + """
|
| + Helper method for common parse actions that simply return a literal value. Especially
|
| + useful when used with C{L{transformString<ParserElement.transformString>}()}.
|
| +
|
| + Example::
|
| + num = Word(nums).setParseAction(lambda toks: int(toks[0]))
|
| + na = oneOf("N/A NA").setParseAction(replaceWith(math.nan))
|
| + term = na | num
|
| +
|
| + OneOrMore(term).parseString("324 234 N/A 234") # -> [324, 234, nan, 234]
|
| + """
|
| + return lambda s,l,t: [replStr]
|
| +
|
| +def removeQuotes(s,l,t):
|
| + """
|
| + Helper parse action for removing quotation marks from parsed quoted strings.
|
| +
|
| + Example::
|
| + # by default, quotation marks are included in parsed results
|
| + quotedString.parseString("'Now is the Winter of our Discontent'") # -> ["'Now is the Winter of our Discontent'"]
|
| +
|
| + # use removeQuotes to strip quotation marks from parsed results
|
| + quotedString.setParseAction(removeQuotes)
|
| + quotedString.parseString("'Now is the Winter of our Discontent'") # -> ["Now is the Winter of our Discontent"]
|
| + """
|
| + return t[0][1:-1]
|
| +
|
| +def tokenMap(func, *args):
|
| + """
|
| + Helper to define a parse action by mapping a function to all elements of a ParseResults list.If any additional
|
| + args are passed, they are forwarded to the given function as additional arguments after
|
| + the token, as in C{hex_integer = Word(hexnums).setParseAction(tokenMap(int, 16))}, which will convert the
|
| + parsed data to an integer using base 16.
|
| +
|
| + Example (compare the last to example in L{ParserElement.transformString}::
|
| + hex_ints = OneOrMore(Word(hexnums)).setParseAction(tokenMap(int, 16))
|
| + hex_ints.runTests('''
|
| + 00 11 22 aa FF 0a 0d 1a
|
| + ''')
|
| +
|
| + upperword = Word(alphas).setParseAction(tokenMap(str.upper))
|
| + OneOrMore(upperword).runTests('''
|
| + my kingdom for a horse
|
| + ''')
|
| +
|
| + wd = Word(alphas).setParseAction(tokenMap(str.title))
|
| + OneOrMore(wd).setParseAction(' '.join).runTests('''
|
| + now is the winter of our discontent made glorious summer by this sun of york
|
| + ''')
|
| + prints::
|
| + 00 11 22 aa FF 0a 0d 1a
|
| + [0, 17, 34, 170, 255, 10, 13, 26]
|
| +
|
| + my kingdom for a horse
|
| + ['MY', 'KINGDOM', 'FOR', 'A', 'HORSE']
|
| +
|
| + now is the winter of our discontent made glorious summer by this sun of york
|
| + ['Now Is The Winter Of Our Discontent Made Glorious Summer By This Sun Of York']
|
| + """
|
| + def pa(s,l,t):
|
| + return [func(tokn, *args) for tokn in t]
|
| +
|
| + try:
|
| + func_name = getattr(func, '__name__',
|
| + getattr(func, '__class__').__name__)
|
| + except Exception:
|
| + func_name = str(func)
|
| + pa.__name__ = func_name
|
| +
|
| + return pa
|
| +
|
| +upcaseTokens = tokenMap(lambda t: _ustr(t).upper())
|
| +"""(Deprecated) Helper parse action to convert tokens to upper case. Deprecated in favor of L{pyparsing_common.upcaseTokens}"""
|
| +
|
| +downcaseTokens = tokenMap(lambda t: _ustr(t).lower())
|
| +"""(Deprecated) Helper parse action to convert tokens to lower case. Deprecated in favor of L{pyparsing_common.downcaseTokens}"""
|
| +
|
| +def _makeTags(tagStr, xml):
|
| + """Internal helper to construct opening and closing tag expressions, given a tag name"""
|
| + if isinstance(tagStr,basestring):
|
| + resname = tagStr
|
| + tagStr = Keyword(tagStr, caseless=not xml)
|
| + else:
|
| + resname = tagStr.name
|
| +
|
| + tagAttrName = Word(alphas,alphanums+"_-:")
|
| + if (xml):
|
| + tagAttrValue = dblQuotedString.copy().setParseAction( removeQuotes )
|
| + openTag = Suppress("<") + tagStr("tag") + \
|
| + Dict(ZeroOrMore(Group( tagAttrName + Suppress("=") + tagAttrValue ))) + \
|
| + Optional("/",default=[False]).setResultsName("empty").setParseAction(lambda s,l,t:t[0]=='/') + Suppress(">")
|
| + else:
|
| + printablesLessRAbrack = "".join(c for c in printables if c not in ">")
|
| + tagAttrValue = quotedString.copy().setParseAction( removeQuotes ) | Word(printablesLessRAbrack)
|
| + openTag = Suppress("<") + tagStr("tag") + \
|
| + Dict(ZeroOrMore(Group( tagAttrName.setParseAction(downcaseTokens) + \
|
| + Optional( Suppress("=") + tagAttrValue ) ))) + \
|
| + Optional("/",default=[False]).setResultsName("empty").setParseAction(lambda s,l,t:t[0]=='/') + Suppress(">")
|
| + closeTag = Combine(_L("</") + tagStr + ">")
|
| +
|
| + openTag = openTag.setResultsName("start"+"".join(resname.replace(":"," ").title().split())).setName("<%s>" % resname)
|
| + closeTag = closeTag.setResultsName("end"+"".join(resname.replace(":"," ").title().split())).setName("</%s>" % resname)
|
| + openTag.tag = resname
|
| + closeTag.tag = resname
|
| + return openTag, closeTag
|
| +
|
| +def makeHTMLTags(tagStr):
|
| + """
|
| + Helper to construct opening and closing tag expressions for HTML, given a tag name. Matches
|
| + tags in either upper or lower case, attributes with namespaces and with quoted or unquoted values.
|
| +
|
| + Example::
|
| + text = '<td>More info at the <a href="http://pyparsing.wikispaces.com">pyparsing</a> wiki page</td>'
|
| + # makeHTMLTags returns pyparsing expressions for the opening and closing tags as a 2-tuple
|
| + a,a_end = makeHTMLTags("A")
|
| + link_expr = a + SkipTo(a_end)("link_text") + a_end
|
| +
|
| + for link in link_expr.searchString(text):
|
| + # attributes in the <A> tag (like "href" shown here) are also accessible as named results
|
| + print(link.link_text, '->', link.href)
|
| + prints::
|
| + pyparsing -> http://pyparsing.wikispaces.com
|
| + """
|
| + return _makeTags( tagStr, False )
|
| +
|
| +def makeXMLTags(tagStr):
|
| + """
|
| + Helper to construct opening and closing tag expressions for XML, given a tag name. Matches
|
| + tags only in the given upper/lower case.
|
| +
|
| + Example: similar to L{makeHTMLTags}
|
| + """
|
| + return _makeTags( tagStr, True )
|
| +
|
| +def withAttribute(*args,**attrDict):
|
| + """
|
| + Helper to create a validating parse action to be used with start tags created
|
| + with C{L{makeXMLTags}} or C{L{makeHTMLTags}}. Use C{withAttribute} to qualify a starting tag
|
| + with a required attribute value, to avoid false matches on common tags such as
|
| + C{<TD>} or C{<DIV>}.
|
| +
|
| + Call C{withAttribute} with a series of attribute names and values. Specify the list
|
| + of filter attributes names and values as:
|
| + - keyword arguments, as in C{(align="right")}, or
|
| + - as an explicit dict with C{**} operator, when an attribute name is also a Python
|
| + reserved word, as in C{**{"class":"Customer", "align":"right"}}
|
| + - a list of name-value tuples, as in ( ("ns1:class", "Customer"), ("ns2:align","right") )
|
| + For attribute names with a namespace prefix, you must use the second form. Attribute
|
| + names are matched insensitive to upper/lower case.
|
| +
|
| + If just testing for C{class} (with or without a namespace), use C{L{withClass}}.
|
| +
|
| + To verify that the attribute exists, but without specifying a value, pass
|
| + C{withAttribute.ANY_VALUE} as the value.
|
| +
|
| + Example::
|
| + html = '''
|
| + <div>
|
| + Some text
|
| + <div type="grid">1 4 0 1 0</div>
|
| + <div type="graph">1,3 2,3 1,1</div>
|
| + <div>this has no type</div>
|
| + </div>
|
| +
|
| + '''
|
| + div,div_end = makeHTMLTags("div")
|
| +
|
| + # only match div tag having a type attribute with value "grid"
|
| + div_grid = div().setParseAction(withAttribute(type="grid"))
|
| + grid_expr = div_grid + SkipTo(div | div_end)("body")
|
| + for grid_header in grid_expr.searchString(html):
|
| + print(grid_header.body)
|
| +
|
| + # construct a match with any div tag having a type attribute, regardless of the value
|
| + div_any_type = div().setParseAction(withAttribute(type=withAttribute.ANY_VALUE))
|
| + div_expr = div_any_type + SkipTo(div | div_end)("body")
|
| + for div_header in div_expr.searchString(html):
|
| + print(div_header.body)
|
| + prints::
|
| + 1 4 0 1 0
|
| +
|
| + 1 4 0 1 0
|
| + 1,3 2,3 1,1
|
| + """
|
| + if args:
|
| + attrs = args[:]
|
| + else:
|
| + attrs = attrDict.items()
|
| + attrs = [(k,v) for k,v in attrs]
|
| + def pa(s,l,tokens):
|
| + for attrName,attrValue in attrs:
|
| + if attrName not in tokens:
|
| + raise ParseException(s,l,"no matching attribute " + attrName)
|
| + if attrValue != withAttribute.ANY_VALUE and tokens[attrName] != attrValue:
|
| + raise ParseException(s,l,"attribute '%s' has value '%s', must be '%s'" %
|
| + (attrName, tokens[attrName], attrValue))
|
| + return pa
|
| +withAttribute.ANY_VALUE = object()
|
| +
|
| +def withClass(classname, namespace=''):
|
| + """
|
| + Simplified version of C{L{withAttribute}} when matching on a div class - made
|
| + difficult because C{class} is a reserved word in Python.
|
| +
|
| + Example::
|
| + html = '''
|
| + <div>
|
| + Some text
|
| + <div class="grid">1 4 0 1 0</div>
|
| + <div class="graph">1,3 2,3 1,1</div>
|
| + <div>this <div> has no class</div>
|
| + </div>
|
| +
|
| + '''
|
| + div,div_end = makeHTMLTags("div")
|
| + div_grid = div().setParseAction(withClass("grid"))
|
| +
|
| + grid_expr = div_grid + SkipTo(div | div_end)("body")
|
| + for grid_header in grid_expr.searchString(html):
|
| + print(grid_header.body)
|
| +
|
| + div_any_type = div().setParseAction(withClass(withAttribute.ANY_VALUE))
|
| + div_expr = div_any_type + SkipTo(div | div_end)("body")
|
| + for div_header in div_expr.searchString(html):
|
| + print(div_header.body)
|
| + prints::
|
| + 1 4 0 1 0
|
| +
|
| + 1 4 0 1 0
|
| + 1,3 2,3 1,1
|
| + """
|
| + classattr = "%s:class" % namespace if namespace else "class"
|
| + return withAttribute(**{classattr : classname})
|
| +
|
| +opAssoc = _Constants()
|
| +opAssoc.LEFT = object()
|
| +opAssoc.RIGHT = object()
|
| +
|
| +def infixNotation( baseExpr, opList, lpar=Suppress('('), rpar=Suppress(')') ):
|
| + """
|
| + Helper method for constructing grammars of expressions made up of
|
| + operators working in a precedence hierarchy. Operators may be unary or
|
| + binary, left- or right-associative. Parse actions can also be attached
|
| + to operator expressions. The generated parser will also recognize the use
|
| + of parentheses to override operator precedences (see example below).
|
| +
|
| + Note: if you define a deep operator list, you may see performance issues
|
| + when using infixNotation. See L{ParserElement.enablePackrat} for a
|
| + mechanism to potentially improve your parser performance.
|
| +
|
| + Parameters:
|
| + - baseExpr - expression representing the most basic element for the nested
|
| + - opList - list of tuples, one for each operator precedence level in the
|
| + expression grammar; each tuple is of the form
|
| + (opExpr, numTerms, rightLeftAssoc, parseAction), where:
|
| + - opExpr is the pyparsing expression for the operator;
|
| + may also be a string, which will be converted to a Literal;
|
| + if numTerms is 3, opExpr is a tuple of two expressions, for the
|
| + two operators separating the 3 terms
|
| + - numTerms is the number of terms for this operator (must
|
| + be 1, 2, or 3)
|
| + - rightLeftAssoc is the indicator whether the operator is
|
| + right or left associative, using the pyparsing-defined
|
| + constants C{opAssoc.RIGHT} and C{opAssoc.LEFT}.
|
| + - parseAction is the parse action to be associated with
|
| + expressions matching this operator expression (the
|
| + parse action tuple member may be omitted)
|
| + - lpar - expression for matching left-parentheses (default=C{Suppress('(')})
|
| + - rpar - expression for matching right-parentheses (default=C{Suppress(')')})
|
| +
|
| + Example::
|
| + # simple example of four-function arithmetic with ints and variable names
|
| + integer = pyparsing_common.signed_integer
|
| + varname = pyparsing_common.identifier
|
| +
|
| + arith_expr = infixNotation(integer | varname,
|
| + [
|
| + ('-', 1, opAssoc.RIGHT),
|
| + (oneOf('* /'), 2, opAssoc.LEFT),
|
| + (oneOf('+ -'), 2, opAssoc.LEFT),
|
| + ])
|
| +
|
| + arith_expr.runTests('''
|
| + 5+3*6
|
| + (5+3)*6
|
| + -2--11
|
| + ''', fullDump=False)
|
| + prints::
|
| + 5+3*6
|
| + [[5, '+', [3, '*', 6]]]
|
| +
|
| + (5+3)*6
|
| + [[[5, '+', 3], '*', 6]]
|
| +
|
| + -2--11
|
| + [[['-', 2], '-', ['-', 11]]]
|
| + """
|
| + ret = Forward()
|
| + lastExpr = baseExpr | ( lpar + ret + rpar )
|
| + for i,operDef in enumerate(opList):
|
| + opExpr,arity,rightLeftAssoc,pa = (operDef + (None,))[:4]
|
| + termName = "%s term" % opExpr if arity < 3 else "%s%s term" % opExpr
|
| + if arity == 3:
|
| + if opExpr is None or len(opExpr) != 2:
|
| + raise ValueError("if numterms=3, opExpr must be a tuple or list of two expressions")
|
| + opExpr1, opExpr2 = opExpr
|
| + thisExpr = Forward().setName(termName)
|
| + if rightLeftAssoc == opAssoc.LEFT:
|
| + if arity == 1:
|
| + matchExpr = FollowedBy(lastExpr + opExpr) + Group( lastExpr + OneOrMore( opExpr ) )
|
| + elif arity == 2:
|
| + if opExpr is not None:
|
| + matchExpr = FollowedBy(lastExpr + opExpr + lastExpr) + Group( lastExpr + OneOrMore( opExpr + lastExpr ) )
|
| + else:
|
| + matchExpr = FollowedBy(lastExpr+lastExpr) + Group( lastExpr + OneOrMore(lastExpr) )
|
| + elif arity == 3:
|
| + matchExpr = FollowedBy(lastExpr + opExpr1 + lastExpr + opExpr2 + lastExpr) + \
|
| + Group( lastExpr + opExpr1 + lastExpr + opExpr2 + lastExpr )
|
| + else:
|
| + raise ValueError("operator must be unary (1), binary (2), or ternary (3)")
|
| + elif rightLeftAssoc == opAssoc.RIGHT:
|
| + if arity == 1:
|
| + # try to avoid LR with this extra test
|
| + if not isinstance(opExpr, Optional):
|
| + opExpr = Optional(opExpr)
|
| + matchExpr = FollowedBy(opExpr.expr + thisExpr) + Group( opExpr + thisExpr )
|
| + elif arity == 2:
|
| + if opExpr is not None:
|
| + matchExpr = FollowedBy(lastExpr + opExpr + thisExpr) + Group( lastExpr + OneOrMore( opExpr + thisExpr ) )
|
| + else:
|
| + matchExpr = FollowedBy(lastExpr + thisExpr) + Group( lastExpr + OneOrMore( thisExpr ) )
|
| + elif arity == 3:
|
| + matchExpr = FollowedBy(lastExpr + opExpr1 + thisExpr + opExpr2 + thisExpr) + \
|
| + Group( lastExpr + opExpr1 + thisExpr + opExpr2 + thisExpr )
|
| + else:
|
| + raise ValueError("operator must be unary (1), binary (2), or ternary (3)")
|
| + else:
|
| + raise ValueError("operator must indicate right or left associativity")
|
| + if pa:
|
| + matchExpr.setParseAction( pa )
|
| + thisExpr <<= ( matchExpr.setName(termName) | lastExpr )
|
| + lastExpr = thisExpr
|
| + ret <<= lastExpr
|
| + return ret
|
| +
|
| +operatorPrecedence = infixNotation
|
| +"""(Deprecated) Former name of C{L{infixNotation}}, will be dropped in a future release."""
|
| +
|
| +dblQuotedString = Combine(Regex(r'"(?:[^"\n\r\\]|(?:"")|(?:\\(?:[^x]|x[0-9a-fA-F]+)))*')+'"').setName("string enclosed in double quotes")
|
| +sglQuotedString = Combine(Regex(r"'(?:[^'\n\r\\]|(?:'')|(?:\\(?:[^x]|x[0-9a-fA-F]+)))*")+"'").setName("string enclosed in single quotes")
|
| +quotedString = Combine(Regex(r'"(?:[^"\n\r\\]|(?:"")|(?:\\(?:[^x]|x[0-9a-fA-F]+)))*')+'"'|
|
| + Regex(r"'(?:[^'\n\r\\]|(?:'')|(?:\\(?:[^x]|x[0-9a-fA-F]+)))*")+"'").setName("quotedString using single or double quotes")
|
| +unicodeString = Combine(_L('u') + quotedString.copy()).setName("unicode string literal")
|
| +
|
| +def nestedExpr(opener="(", closer=")", content=None, ignoreExpr=quotedString.copy()):
|
| + """
|
| + Helper method for defining nested lists enclosed in opening and closing
|
| + delimiters ("(" and ")" are the default).
|
| +
|
| + Parameters:
|
| + - opener - opening character for a nested list (default=C{"("}); can also be a pyparsing expression
|
| + - closer - closing character for a nested list (default=C{")"}); can also be a pyparsing expression
|
| + - content - expression for items within the nested lists (default=C{None})
|
| + - ignoreExpr - expression for ignoring opening and closing delimiters (default=C{quotedString})
|
| +
|
| + If an expression is not provided for the content argument, the nested
|
| + expression will capture all whitespace-delimited content between delimiters
|
| + as a list of separate values.
|
| +
|
| + Use the C{ignoreExpr} argument to define expressions that may contain
|
| + opening or closing characters that should not be treated as opening
|
| + or closing characters for nesting, such as quotedString or a comment
|
| + expression. Specify multiple expressions using an C{L{Or}} or C{L{MatchFirst}}.
|
| + The default is L{quotedString}, but if no expressions are to be ignored,
|
| + then pass C{None} for this argument.
|
| +
|
| + Example::
|
| + data_type = oneOf("void int short long char float double")
|
| + decl_data_type = Combine(data_type + Optional(Word('*')))
|
| + ident = Word(alphas+'_', alphanums+'_')
|
| + number = pyparsing_common.number
|
| + arg = Group(decl_data_type + ident)
|
| + LPAR,RPAR = map(Suppress, "()")
|
| +
|
| + code_body = nestedExpr('{', '}', ignoreExpr=(quotedString | cStyleComment))
|
| +
|
| + c_function = (decl_data_type("type")
|
| + + ident("name")
|
| + + LPAR + Optional(delimitedList(arg), [])("args") + RPAR
|
| + + code_body("body"))
|
| + c_function.ignore(cStyleComment)
|
| +
|
| + source_code = '''
|
| + int is_odd(int x) {
|
| + return (x%2);
|
| + }
|
| +
|
| + int dec_to_hex(char hchar) {
|
| + if (hchar >= '0' && hchar <= '9') {
|
| + return (ord(hchar)-ord('0'));
|
| + } else {
|
| + return (10+ord(hchar)-ord('A'));
|
| + }
|
| + }
|
| + '''
|
| + for func in c_function.searchString(source_code):
|
| + print("%(name)s (%(type)s) args: %(args)s" % func)
|
| +
|
| + prints::
|
| + is_odd (int) args: [['int', 'x']]
|
| + dec_to_hex (int) args: [['char', 'hchar']]
|
| + """
|
| + if opener == closer:
|
| + raise ValueError("opening and closing strings cannot be the same")
|
| + if content is None:
|
| + if isinstance(opener,basestring) and isinstance(closer,basestring):
|
| + if len(opener) == 1 and len(closer)==1:
|
| + if ignoreExpr is not None:
|
| + content = (Combine(OneOrMore(~ignoreExpr +
|
| + CharsNotIn(opener+closer+ParserElement.DEFAULT_WHITE_CHARS,exact=1))
|
| + ).setParseAction(lambda t:t[0].strip()))
|
| + else:
|
| + content = (empty.copy()+CharsNotIn(opener+closer+ParserElement.DEFAULT_WHITE_CHARS
|
| + ).setParseAction(lambda t:t[0].strip()))
|
| + else:
|
| + if ignoreExpr is not None:
|
| + content = (Combine(OneOrMore(~ignoreExpr +
|
| + ~Literal(opener) + ~Literal(closer) +
|
| + CharsNotIn(ParserElement.DEFAULT_WHITE_CHARS,exact=1))
|
| + ).setParseAction(lambda t:t[0].strip()))
|
| + else:
|
| + content = (Combine(OneOrMore(~Literal(opener) + ~Literal(closer) +
|
| + CharsNotIn(ParserElement.DEFAULT_WHITE_CHARS,exact=1))
|
| + ).setParseAction(lambda t:t[0].strip()))
|
| + else:
|
| + raise ValueError("opening and closing arguments must be strings if no content expression is given")
|
| + ret = Forward()
|
| + if ignoreExpr is not None:
|
| + ret <<= Group( Suppress(opener) + ZeroOrMore( ignoreExpr | ret | content ) + Suppress(closer) )
|
| + else:
|
| + ret <<= Group( Suppress(opener) + ZeroOrMore( ret | content ) + Suppress(closer) )
|
| + ret.setName('nested %s%s expression' % (opener,closer))
|
| + return ret
|
| +
|
| +def indentedBlock(blockStatementExpr, indentStack, indent=True):
|
| + """
|
| + Helper method for defining space-delimited indentation blocks, such as
|
| + those used to define block statements in Python source code.
|
| +
|
| + Parameters:
|
| + - blockStatementExpr - expression defining syntax of statement that
|
| + is repeated within the indented block
|
| + - indentStack - list created by caller to manage indentation stack
|
| + (multiple statementWithIndentedBlock expressions within a single grammar
|
| + should share a common indentStack)
|
| + - indent - boolean indicating whether block must be indented beyond the
|
| + the current level; set to False for block of left-most statements
|
| + (default=C{True})
|
| +
|
| + A valid block must contain at least one C{blockStatement}.
|
| +
|
| + Example::
|
| + data = '''
|
| + def A(z):
|
| + A1
|
| + B = 100
|
| + G = A2
|
| + A2
|
| + A3
|
| + B
|
| + def BB(a,b,c):
|
| + BB1
|
| + def BBA():
|
| + bba1
|
| + bba2
|
| + bba3
|
| + C
|
| + D
|
| + def spam(x,y):
|
| + def eggs(z):
|
| + pass
|
| + '''
|
| +
|
| +
|
| + indentStack = [1]
|
| + stmt = Forward()
|
| +
|
| + identifier = Word(alphas, alphanums)
|
| + funcDecl = ("def" + identifier + Group( "(" + Optional( delimitedList(identifier) ) + ")" ) + ":")
|
| + func_body = indentedBlock(stmt, indentStack)
|
| + funcDef = Group( funcDecl + func_body )
|
| +
|
| + rvalue = Forward()
|
| + funcCall = Group(identifier + "(" + Optional(delimitedList(rvalue)) + ")")
|
| + rvalue << (funcCall | identifier | Word(nums))
|
| + assignment = Group(identifier + "=" + rvalue)
|
| + stmt << ( funcDef | assignment | identifier )
|
| +
|
| + module_body = OneOrMore(stmt)
|
| +
|
| + parseTree = module_body.parseString(data)
|
| + parseTree.pprint()
|
| + prints::
|
| + [['def',
|
| + 'A',
|
| + ['(', 'z', ')'],
|
| + ':',
|
| + [['A1'], [['B', '=', '100']], [['G', '=', 'A2']], ['A2'], ['A3']]],
|
| + 'B',
|
| + ['def',
|
| + 'BB',
|
| + ['(', 'a', 'b', 'c', ')'],
|
| + ':',
|
| + [['BB1'], [['def', 'BBA', ['(', ')'], ':', [['bba1'], ['bba2'], ['bba3']]]]]],
|
| + 'C',
|
| + 'D',
|
| + ['def',
|
| + 'spam',
|
| + ['(', 'x', 'y', ')'],
|
| + ':',
|
| + [[['def', 'eggs', ['(', 'z', ')'], ':', [['pass']]]]]]]
|
| + """
|
| + def checkPeerIndent(s,l,t):
|
| + if l >= len(s): return
|
| + curCol = col(l,s)
|
| + if curCol != indentStack[-1]:
|
| + if curCol > indentStack[-1]:
|
| + raise ParseFatalException(s,l,"illegal nesting")
|
| + raise ParseException(s,l,"not a peer entry")
|
| +
|
| + def checkSubIndent(s,l,t):
|
| + curCol = col(l,s)
|
| + if curCol > indentStack[-1]:
|
| + indentStack.append( curCol )
|
| + else:
|
| + raise ParseException(s,l,"not a subentry")
|
| +
|
| + def checkUnindent(s,l,t):
|
| + if l >= len(s): return
|
| + curCol = col(l,s)
|
| + if not(indentStack and curCol < indentStack[-1] and curCol <= indentStack[-2]):
|
| + raise ParseException(s,l,"not an unindent")
|
| + indentStack.pop()
|
| +
|
| + NL = OneOrMore(LineEnd().setWhitespaceChars("\t ").suppress())
|
| + INDENT = (Empty() + Empty().setParseAction(checkSubIndent)).setName('INDENT')
|
| + PEER = Empty().setParseAction(checkPeerIndent).setName('')
|
| + UNDENT = Empty().setParseAction(checkUnindent).setName('UNINDENT')
|
| + if indent:
|
| + smExpr = Group( Optional(NL) +
|
| + #~ FollowedBy(blockStatementExpr) +
|
| + INDENT + (OneOrMore( PEER + Group(blockStatementExpr) + Optional(NL) )) + UNDENT)
|
| + else:
|
| + smExpr = Group( Optional(NL) +
|
| + (OneOrMore( PEER + Group(blockStatementExpr) + Optional(NL) )) )
|
| + blockStatementExpr.ignore(_bslash + LineEnd())
|
| + return smExpr.setName('indented block')
|
| +
|
| +alphas8bit = srange(r"[\0xc0-\0xd6\0xd8-\0xf6\0xf8-\0xff]")
|
| +punc8bit = srange(r"[\0xa1-\0xbf\0xd7\0xf7]")
|
| +
|
| +anyOpenTag,anyCloseTag = makeHTMLTags(Word(alphas,alphanums+"_:").setName('any tag'))
|
| +_htmlEntityMap = dict(zip("gt lt amp nbsp quot apos".split(),'><& "\''))
|
| +commonHTMLEntity = Regex('&(?P<entity>' + '|'.join(_htmlEntityMap.keys()) +");").setName("common HTML entity")
|
| +def replaceHTMLEntity(t):
|
| + """Helper parser action to replace common HTML entities with their special characters"""
|
| + return _htmlEntityMap.get(t.entity)
|
| +
|
| +# it's easy to get these comment structures wrong - they're very common, so may as well make them available
|
| +cStyleComment = Combine(Regex(r"/\*(?:[^*]|\*(?!/))*") + '*/').setName("C style comment")
|
| +"Comment of the form C{/* ... */}"
|
| +
|
| +htmlComment = Regex(r"<!--[\s\S]*?-->").setName("HTML comment")
|
| +"Comment of the form C{<!-- ... -->}"
|
| +
|
| +restOfLine = Regex(r".*").leaveWhitespace().setName("rest of line")
|
| +dblSlashComment = Regex(r"//(?:\\\n|[^\n])*").setName("// comment")
|
| +"Comment of the form C{// ... (to end of line)}"
|
| +
|
| +cppStyleComment = Combine(Regex(r"/\*(?:[^*]|\*(?!/))*") + '*/'| dblSlashComment).setName("C++ style comment")
|
| +"Comment of either form C{L{cStyleComment}} or C{L{dblSlashComment}}"
|
| +
|
| +javaStyleComment = cppStyleComment
|
| +"Same as C{L{cppStyleComment}}"
|
| +
|
| +pythonStyleComment = Regex(r"#.*").setName("Python style comment")
|
| +"Comment of the form C{# ... (to end of line)}"
|
| +
|
| +_commasepitem = Combine(OneOrMore(Word(printables, excludeChars=',') +
|
| + Optional( Word(" \t") +
|
| + ~Literal(",") + ~LineEnd() ) ) ).streamline().setName("commaItem")
|
| +commaSeparatedList = delimitedList( Optional( quotedString.copy() | _commasepitem, default="") ).setName("commaSeparatedList")
|
| +"""(Deprecated) Predefined expression of 1 or more printable words or quoted strings, separated by commas.
|
| + This expression is deprecated in favor of L{pyparsing_common.comma_separated_list}."""
|
| +
|
| +# some other useful expressions - using lower-case class name since we are really using this as a namespace
|
| +class pyparsing_common:
|
| + """
|
| + Here are some common low-level expressions that may be useful in jump-starting parser development:
|
| + - numeric forms (L{integers<integer>}, L{reals<real>}, L{scientific notation<sci_real>})
|
| + - common L{programming identifiers<identifier>}
|
| + - network addresses (L{MAC<mac_address>}, L{IPv4<ipv4_address>}, L{IPv6<ipv6_address>})
|
| + - ISO8601 L{dates<iso8601_date>} and L{datetime<iso8601_datetime>}
|
| + - L{UUID<uuid>}
|
| + - L{comma-separated list<comma_separated_list>}
|
| + Parse actions:
|
| + - C{L{convertToInteger}}
|
| + - C{L{convertToFloat}}
|
| + - C{L{convertToDate}}
|
| + - C{L{convertToDatetime}}
|
| + - C{L{stripHTMLTags}}
|
| + - C{L{upcaseTokens}}
|
| + - C{L{downcaseTokens}}
|
| +
|
| + Example::
|
| + pyparsing_common.number.runTests('''
|
| + # any int or real number, returned as the appropriate type
|
| + 100
|
| + -100
|
| + +100
|
| + 3.14159
|
| + 6.02e23
|
| + 1e-12
|
| + ''')
|
| +
|
| + pyparsing_common.fnumber.runTests('''
|
| + # any int or real number, returned as float
|
| + 100
|
| + -100
|
| + +100
|
| + 3.14159
|
| + 6.02e23
|
| + 1e-12
|
| + ''')
|
| +
|
| + pyparsing_common.hex_integer.runTests('''
|
| + # hex numbers
|
| + 100
|
| + FF
|
| + ''')
|
| +
|
| + pyparsing_common.fraction.runTests('''
|
| + # fractions
|
| + 1/2
|
| + -3/4
|
| + ''')
|
| +
|
| + pyparsing_common.mixed_integer.runTests('''
|
| + # mixed fractions
|
| + 1
|
| + 1/2
|
| + -3/4
|
| + 1-3/4
|
| + ''')
|
| +
|
| + import uuid
|
| + pyparsing_common.uuid.setParseAction(tokenMap(uuid.UUID))
|
| + pyparsing_common.uuid.runTests('''
|
| + # uuid
|
| + 12345678-1234-5678-1234-567812345678
|
| + ''')
|
| + prints::
|
| + # any int or real number, returned as the appropriate type
|
| + 100
|
| + [100]
|
| +
|
| + -100
|
| + [-100]
|
| +
|
| + +100
|
| + [100]
|
| +
|
| + 3.14159
|
| + [3.14159]
|
| +
|
| + 6.02e23
|
| + [6.02e+23]
|
| +
|
| + 1e-12
|
| + [1e-12]
|
| +
|
| + # any int or real number, returned as float
|
| + 100
|
| + [100.0]
|
| +
|
| + -100
|
| + [-100.0]
|
| +
|
| + +100
|
| + [100.0]
|
| +
|
| + 3.14159
|
| + [3.14159]
|
| +
|
| + 6.02e23
|
| + [6.02e+23]
|
| +
|
| + 1e-12
|
| + [1e-12]
|
| +
|
| + # hex numbers
|
| + 100
|
| + [256]
|
| +
|
| + FF
|
| + [255]
|
| +
|
| + # fractions
|
| + 1/2
|
| + [0.5]
|
| +
|
| + -3/4
|
| + [-0.75]
|
| +
|
| + # mixed fractions
|
| + 1
|
| + [1]
|
| +
|
| + 1/2
|
| + [0.5]
|
| +
|
| + -3/4
|
| + [-0.75]
|
| +
|
| + 1-3/4
|
| + [1.75]
|
| +
|
| + # uuid
|
| + 12345678-1234-5678-1234-567812345678
|
| + [UUID('12345678-1234-5678-1234-567812345678')]
|
| + """
|
| +
|
| + convertToInteger = tokenMap(int)
|
| + """
|
| + Parse action for converting parsed integers to Python int
|
| + """
|
| +
|
| + convertToFloat = tokenMap(float)
|
| + """
|
| + Parse action for converting parsed numbers to Python float
|
| + """
|
| +
|
| + integer = Word(nums).setName("integer").setParseAction(convertToInteger)
|
| + """expression that parses an unsigned integer, returns an int"""
|
| +
|
| + hex_integer = Word(hexnums).setName("hex integer").setParseAction(tokenMap(int,16))
|
| + """expression that parses a hexadecimal integer, returns an int"""
|
| +
|
| + signed_integer = Regex(r'[+-]?\d+').setName("signed integer").setParseAction(convertToInteger)
|
| + """expression that parses an integer with optional leading sign, returns an int"""
|
| +
|
| + fraction = (signed_integer().setParseAction(convertToFloat) + '/' + signed_integer().setParseAction(convertToFloat)).setName("fraction")
|
| + """fractional expression of an integer divided by an integer, returns a float"""
|
| + fraction.addParseAction(lambda t: t[0]/t[-1])
|
| +
|
| + mixed_integer = (fraction | signed_integer + Optional(Optional('-').suppress() + fraction)).setName("fraction or mixed integer-fraction")
|
| + """mixed integer of the form 'integer - fraction', with optional leading integer, returns float"""
|
| + mixed_integer.addParseAction(sum)
|
| +
|
| + real = Regex(r'[+-]?\d+\.\d*').setName("real number").setParseAction(convertToFloat)
|
| + """expression that parses a floating point number and returns a float"""
|
| +
|
| + sci_real = Regex(r'[+-]?\d+([eE][+-]?\d+|\.\d*([eE][+-]?\d+)?)').setName("real number with scientific notation").setParseAction(convertToFloat)
|
| + """expression that parses a floating point number with optional scientific notation and returns a float"""
|
| +
|
| + # streamlining this expression makes the docs nicer-looking
|
| + number = (sci_real | real | signed_integer).streamline()
|
| + """any numeric expression, returns the corresponding Python type"""
|
| +
|
| + fnumber = Regex(r'[+-]?\d+\.?\d*([eE][+-]?\d+)?').setName("fnumber").setParseAction(convertToFloat)
|
| + """any int or real number, returned as float"""
|
| +
|
| + identifier = Word(alphas+'_', alphanums+'_').setName("identifier")
|
| + """typical code identifier (leading alpha or '_', followed by 0 or more alphas, nums, or '_')"""
|
| +
|
| + ipv4_address = Regex(r'(25[0-5]|2[0-4][0-9]|1?[0-9]{1,2})(\.(25[0-5]|2[0-4][0-9]|1?[0-9]{1,2})){3}').setName("IPv4 address")
|
| + "IPv4 address (C{0.0.0.0 - 255.255.255.255})"
|
| +
|
| + _ipv6_part = Regex(r'[0-9a-fA-F]{1,4}').setName("hex_integer")
|
| + _full_ipv6_address = (_ipv6_part + (':' + _ipv6_part)*7).setName("full IPv6 address")
|
| + _short_ipv6_address = (Optional(_ipv6_part + (':' + _ipv6_part)*(0,6)) + "::" + Optional(_ipv6_part + (':' + _ipv6_part)*(0,6))).setName("short IPv6 address")
|
| + _short_ipv6_address.addCondition(lambda t: sum(1 for tt in t if pyparsing_common._ipv6_part.matches(tt)) < 8)
|
| + _mixed_ipv6_address = ("::ffff:" + ipv4_address).setName("mixed IPv6 address")
|
| + ipv6_address = Combine((_full_ipv6_address | _mixed_ipv6_address | _short_ipv6_address).setName("IPv6 address")).setName("IPv6 address")
|
| + "IPv6 address (long, short, or mixed form)"
|
| +
|
| + mac_address = Regex(r'[0-9a-fA-F]{2}([:.-])[0-9a-fA-F]{2}(?:\1[0-9a-fA-F]{2}){4}').setName("MAC address")
|
| + "MAC address xx:xx:xx:xx:xx (may also have '-' or '.' delimiters)"
|
| +
|
| + @staticmethod
|
| + def convertToDate(fmt="%Y-%m-%d"):
|
| + """
|
| + Helper to create a parse action for converting parsed date string to Python datetime.date
|
| +
|
| + Params -
|
| + - fmt - format to be passed to datetime.strptime (default=C{"%Y-%m-%d"})
|
| +
|
| + Example::
|
| + date_expr = pyparsing_common.iso8601_date.copy()
|
| + date_expr.setParseAction(pyparsing_common.convertToDate())
|
| + print(date_expr.parseString("1999-12-31"))
|
| + prints::
|
| + [datetime.date(1999, 12, 31)]
|
| + """
|
| + def cvt_fn(s,l,t):
|
| + try:
|
| + return datetime.strptime(t[0], fmt).date()
|
| + except ValueError as ve:
|
| + raise ParseException(s, l, str(ve))
|
| + return cvt_fn
|
| +
|
| + @staticmethod
|
| + def convertToDatetime(fmt="%Y-%m-%dT%H:%M:%S.%f"):
|
| + """
|
| + Helper to create a parse action for converting parsed datetime string to Python datetime.datetime
|
| +
|
| + Params -
|
| + - fmt - format to be passed to datetime.strptime (default=C{"%Y-%m-%dT%H:%M:%S.%f"})
|
| +
|
| + Example::
|
| + dt_expr = pyparsing_common.iso8601_datetime.copy()
|
| + dt_expr.setParseAction(pyparsing_common.convertToDatetime())
|
| + print(dt_expr.parseString("1999-12-31T23:59:59.999"))
|
| + prints::
|
| + [datetime.datetime(1999, 12, 31, 23, 59, 59, 999000)]
|
| + """
|
| + def cvt_fn(s,l,t):
|
| + try:
|
| + return datetime.strptime(t[0], fmt)
|
| + except ValueError as ve:
|
| + raise ParseException(s, l, str(ve))
|
| + return cvt_fn
|
| +
|
| + iso8601_date = Regex(r'(?P<year>\d{4})(?:-(?P<month>\d\d)(?:-(?P<day>\d\d))?)?').setName("ISO8601 date")
|
| + "ISO8601 date (C{yyyy-mm-dd})"
|
| +
|
| + iso8601_datetime = Regex(r'(?P<year>\d{4})-(?P<month>\d\d)-(?P<day>\d\d)[T ](?P<hour>\d\d):(?P<minute>\d\d)(:(?P<second>\d\d(\.\d*)?)?)?(?P<tz>Z|[+-]\d\d:?\d\d)?').setName("ISO8601 datetime")
|
| + "ISO8601 datetime (C{yyyy-mm-ddThh:mm:ss.s(Z|+-00:00)}) - trailing seconds, milliseconds, and timezone optional; accepts separating C{'T'} or C{' '}"
|
| +
|
| + uuid = Regex(r'[0-9a-fA-F]{8}(-[0-9a-fA-F]{4}){3}-[0-9a-fA-F]{12}').setName("UUID")
|
| + "UUID (C{xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxxxxx})"
|
| +
|
| + _html_stripper = anyOpenTag.suppress() | anyCloseTag.suppress()
|
| + @staticmethod
|
| + def stripHTMLTags(s, l, tokens):
|
| + """
|
| + Parse action to remove HTML tags from web page HTML source
|
| +
|
| + Example::
|
| + # strip HTML links from normal text
|
| + text = '<td>More info at the <a href="http://pyparsing.wikispaces.com">pyparsing</a> wiki page</td>'
|
| + td,td_end = makeHTMLTags("TD")
|
| + table_text = td + SkipTo(td_end).setParseAction(pyparsing_common.stripHTMLTags)("body") + td_end
|
| +
|
| + print(table_text.parseString(text).body) # -> 'More info at the pyparsing wiki page'
|
| + """
|
| + return pyparsing_common._html_stripper.transformString(tokens[0])
|
| +
|
| + _commasepitem = Combine(OneOrMore(~Literal(",") + ~LineEnd() + Word(printables, excludeChars=',')
|
| + + Optional( White(" \t") ) ) ).streamline().setName("commaItem")
|
| + comma_separated_list = delimitedList( Optional( quotedString.copy() | _commasepitem, default="") ).setName("comma separated list")
|
| + """Predefined expression of 1 or more printable words or quoted strings, separated by commas."""
|
| +
|
| + upcaseTokens = staticmethod(tokenMap(lambda t: _ustr(t).upper()))
|
| + """Parse action to convert tokens to upper case."""
|
| +
|
| + downcaseTokens = staticmethod(tokenMap(lambda t: _ustr(t).lower()))
|
| + """Parse action to convert tokens to lower case."""
|
| +
|
| +
|
| +if __name__ == "__main__":
|
| +
|
| + selectToken = CaselessLiteral("select")
|
| + fromToken = CaselessLiteral("from")
|
| +
|
| + ident = Word(alphas, alphanums + "_$")
|
| +
|
| + columnName = delimitedList(ident, ".", combine=True).setParseAction(upcaseTokens)
|
| + columnNameList = Group(delimitedList(columnName)).setName("columns")
|
| + columnSpec = ('*' | columnNameList)
|
| +
|
| + tableName = delimitedList(ident, ".", combine=True).setParseAction(upcaseTokens)
|
| + tableNameList = Group(delimitedList(tableName)).setName("tables")
|
| +
|
| + simpleSQL = selectToken("command") + columnSpec("columns") + fromToken + tableNameList("tables")
|
| +
|
| + # demo runTests method, including embedded comments in test string
|
| + simpleSQL.runTests("""
|
| + # '*' as column list and dotted table name
|
| + select * from SYS.XYZZY
|
| +
|
| + # caseless match on "SELECT", and casts back to "select"
|
| + SELECT * from XYZZY, ABC
|
| +
|
| + # list of column names, and mixed case SELECT keyword
|
| + Select AA,BB,CC from Sys.dual
|
| +
|
| + # multiple tables
|
| + Select A, B, C from Sys.dual, Table2
|
| +
|
| + # invalid SELECT keyword - should fail
|
| + Xelect A, B, C from Sys.dual
|
| +
|
| + # incomplete command - should fail
|
| + Select
|
| +
|
| + # invalid column name - should fail
|
| + Select ^^^ frox Sys.dual
|
| +
|
| + """)
|
| +
|
| + pyparsing_common.number.runTests("""
|
| + 100
|
| + -100
|
| + +100
|
| + 3.14159
|
| + 6.02e23
|
| + 1e-12
|
| + """)
|
| +
|
| + # any int or real number, returned as float
|
| + pyparsing_common.fnumber.runTests("""
|
| + 100
|
| + -100
|
| + +100
|
| + 3.14159
|
| + 6.02e23
|
| + 1e-12
|
| + """)
|
| +
|
| + pyparsing_common.hex_integer.runTests("""
|
| + 100
|
| + FF
|
| + """)
|
| +
|
| + import uuid
|
| + pyparsing_common.uuid.setParseAction(tokenMap(uuid.UUID))
|
| + pyparsing_common.uuid.runTests("""
|
| + 12345678-1234-5678-1234-567812345678
|
| + """)
|
|
|