| Index: third_party/google-endpoints/future/backports/email/header.py
|
| diff --git a/third_party/google-endpoints/future/backports/email/header.py b/third_party/google-endpoints/future/backports/email/header.py
|
| new file mode 100644
|
| index 0000000000000000000000000000000000000000..63bf038c0223f8bfb4381f981bc37cabfab6cbd7
|
| --- /dev/null
|
| +++ b/third_party/google-endpoints/future/backports/email/header.py
|
| @@ -0,0 +1,581 @@
|
| +# Copyright (C) 2002-2007 Python Software Foundation
|
| +# Author: Ben Gertzfield, Barry Warsaw
|
| +# Contact: email-sig@python.org
|
| +
|
| +"""Header encoding and decoding functionality."""
|
| +from __future__ import unicode_literals
|
| +from __future__ import division
|
| +from __future__ import absolute_import
|
| +from future.builtins import bytes, range, str, super, zip
|
| +
|
| +__all__ = [
|
| + 'Header',
|
| + 'decode_header',
|
| + 'make_header',
|
| + ]
|
| +
|
| +import re
|
| +import binascii
|
| +
|
| +from future.backports import email
|
| +from future.backports.email import base64mime
|
| +from future.backports.email.errors import HeaderParseError
|
| +import future.backports.email.charset as _charset
|
| +
|
| +# Helpers
|
| +from future.backports.email.quoprimime import _max_append, header_decode
|
| +
|
| +Charset = _charset.Charset
|
| +
|
| +NL = '\n'
|
| +SPACE = ' '
|
| +BSPACE = b' '
|
| +SPACE8 = ' ' * 8
|
| +EMPTYSTRING = ''
|
| +MAXLINELEN = 78
|
| +FWS = ' \t'
|
| +
|
| +USASCII = Charset('us-ascii')
|
| +UTF8 = Charset('utf-8')
|
| +
|
| +# Match encoded-word strings in the form =?charset?q?Hello_World?=
|
| +ecre = re.compile(r'''
|
| + =\? # literal =?
|
| + (?P<charset>[^?]*?) # non-greedy up to the next ? is the charset
|
| + \? # literal ?
|
| + (?P<encoding>[qb]) # either a "q" or a "b", case insensitive
|
| + \? # literal ?
|
| + (?P<encoded>.*?) # non-greedy up to the next ?= is the encoded string
|
| + \?= # literal ?=
|
| + ''', re.VERBOSE | re.IGNORECASE | re.MULTILINE)
|
| +
|
| +# Field name regexp, including trailing colon, but not separating whitespace,
|
| +# according to RFC 2822. Character range is from tilde to exclamation mark.
|
| +# For use with .match()
|
| +fcre = re.compile(r'[\041-\176]+:$')
|
| +
|
| +# Find a header embedded in a putative header value. Used to check for
|
| +# header injection attack.
|
| +_embeded_header = re.compile(r'\n[^ \t]+:')
|
| +
|
| +
|
| +def decode_header(header):
|
| + """Decode a message header value without converting charset.
|
| +
|
| + Returns a list of (string, charset) pairs containing each of the decoded
|
| + parts of the header. Charset is None for non-encoded parts of the header,
|
| + otherwise a lower-case string containing the name of the character set
|
| + specified in the encoded string.
|
| +
|
| + header may be a string that may or may not contain RFC2047 encoded words,
|
| + or it may be a Header object.
|
| +
|
| + An email.errors.HeaderParseError may be raised when certain decoding error
|
| + occurs (e.g. a base64 decoding exception).
|
| + """
|
| + # If it is a Header object, we can just return the encoded chunks.
|
| + if hasattr(header, '_chunks'):
|
| + return [(_charset._encode(string, str(charset)), str(charset))
|
| + for string, charset in header._chunks]
|
| + # If no encoding, just return the header with no charset.
|
| + if not ecre.search(header):
|
| + return [(header, None)]
|
| + # First step is to parse all the encoded parts into triplets of the form
|
| + # (encoded_string, encoding, charset). For unencoded strings, the last
|
| + # two parts will be None.
|
| + words = []
|
| + for line in header.splitlines():
|
| + parts = ecre.split(line)
|
| + first = True
|
| + while parts:
|
| + unencoded = parts.pop(0)
|
| + if first:
|
| + unencoded = unencoded.lstrip()
|
| + first = False
|
| + if unencoded:
|
| + words.append((unencoded, None, None))
|
| + if parts:
|
| + charset = parts.pop(0).lower()
|
| + encoding = parts.pop(0).lower()
|
| + encoded = parts.pop(0)
|
| + words.append((encoded, encoding, charset))
|
| + # Now loop over words and remove words that consist of whitespace
|
| + # between two encoded strings.
|
| + import sys
|
| + droplist = []
|
| + for n, w in enumerate(words):
|
| + if n>1 and w[1] and words[n-2][1] and words[n-1][0].isspace():
|
| + droplist.append(n-1)
|
| + for d in reversed(droplist):
|
| + del words[d]
|
| +
|
| + # The next step is to decode each encoded word by applying the reverse
|
| + # base64 or quopri transformation. decoded_words is now a list of the
|
| + # form (decoded_word, charset).
|
| + decoded_words = []
|
| + for encoded_string, encoding, charset in words:
|
| + if encoding is None:
|
| + # This is an unencoded word.
|
| + decoded_words.append((encoded_string, charset))
|
| + elif encoding == 'q':
|
| + word = header_decode(encoded_string)
|
| + decoded_words.append((word, charset))
|
| + elif encoding == 'b':
|
| + paderr = len(encoded_string) % 4 # Postel's law: add missing padding
|
| + if paderr:
|
| + encoded_string += '==='[:4 - paderr]
|
| + try:
|
| + word = base64mime.decode(encoded_string)
|
| + except binascii.Error:
|
| + raise HeaderParseError('Base64 decoding error')
|
| + else:
|
| + decoded_words.append((word, charset))
|
| + else:
|
| + raise AssertionError('Unexpected encoding: ' + encoding)
|
| + # Now convert all words to bytes and collapse consecutive runs of
|
| + # similarly encoded words.
|
| + collapsed = []
|
| + last_word = last_charset = None
|
| + for word, charset in decoded_words:
|
| + if isinstance(word, str):
|
| + word = bytes(word, 'raw-unicode-escape')
|
| + if last_word is None:
|
| + last_word = word
|
| + last_charset = charset
|
| + elif charset != last_charset:
|
| + collapsed.append((last_word, last_charset))
|
| + last_word = word
|
| + last_charset = charset
|
| + elif last_charset is None:
|
| + last_word += BSPACE + word
|
| + else:
|
| + last_word += word
|
| + collapsed.append((last_word, last_charset))
|
| + return collapsed
|
| +
|
| +
|
| +def make_header(decoded_seq, maxlinelen=None, header_name=None,
|
| + continuation_ws=' '):
|
| + """Create a Header from a sequence of pairs as returned by decode_header()
|
| +
|
| + decode_header() takes a header value string and returns a sequence of
|
| + pairs of the format (decoded_string, charset) where charset is the string
|
| + name of the character set.
|
| +
|
| + This function takes one of those sequence of pairs and returns a Header
|
| + instance. Optional maxlinelen, header_name, and continuation_ws are as in
|
| + the Header constructor.
|
| + """
|
| + h = Header(maxlinelen=maxlinelen, header_name=header_name,
|
| + continuation_ws=continuation_ws)
|
| + for s, charset in decoded_seq:
|
| + # None means us-ascii but we can simply pass it on to h.append()
|
| + if charset is not None and not isinstance(charset, Charset):
|
| + charset = Charset(charset)
|
| + h.append(s, charset)
|
| + return h
|
| +
|
| +
|
| +class Header(object):
|
| + def __init__(self, s=None, charset=None,
|
| + maxlinelen=None, header_name=None,
|
| + continuation_ws=' ', errors='strict'):
|
| + """Create a MIME-compliant header that can contain many character sets.
|
| +
|
| + Optional s is the initial header value. If None, the initial header
|
| + value is not set. You can later append to the header with .append()
|
| + method calls. s may be a byte string or a Unicode string, but see the
|
| + .append() documentation for semantics.
|
| +
|
| + Optional charset serves two purposes: it has the same meaning as the
|
| + charset argument to the .append() method. It also sets the default
|
| + character set for all subsequent .append() calls that omit the charset
|
| + argument. If charset is not provided in the constructor, the us-ascii
|
| + charset is used both as s's initial charset and as the default for
|
| + subsequent .append() calls.
|
| +
|
| + The maximum line length can be specified explicitly via maxlinelen. For
|
| + splitting the first line to a shorter value (to account for the field
|
| + header which isn't included in s, e.g. `Subject') pass in the name of
|
| + the field in header_name. The default maxlinelen is 78 as recommended
|
| + by RFC 2822.
|
| +
|
| + continuation_ws must be RFC 2822 compliant folding whitespace (usually
|
| + either a space or a hard tab) which will be prepended to continuation
|
| + lines.
|
| +
|
| + errors is passed through to the .append() call.
|
| + """
|
| + if charset is None:
|
| + charset = USASCII
|
| + elif not isinstance(charset, Charset):
|
| + charset = Charset(charset)
|
| + self._charset = charset
|
| + self._continuation_ws = continuation_ws
|
| + self._chunks = []
|
| + if s is not None:
|
| + self.append(s, charset, errors)
|
| + if maxlinelen is None:
|
| + maxlinelen = MAXLINELEN
|
| + self._maxlinelen = maxlinelen
|
| + if header_name is None:
|
| + self._headerlen = 0
|
| + else:
|
| + # Take the separating colon and space into account.
|
| + self._headerlen = len(header_name) + 2
|
| +
|
| + def __str__(self):
|
| + """Return the string value of the header."""
|
| + self._normalize()
|
| + uchunks = []
|
| + lastcs = None
|
| + lastspace = None
|
| + for string, charset in self._chunks:
|
| + # We must preserve spaces between encoded and non-encoded word
|
| + # boundaries, which means for us we need to add a space when we go
|
| + # from a charset to None/us-ascii, or from None/us-ascii to a
|
| + # charset. Only do this for the second and subsequent chunks.
|
| + # Don't add a space if the None/us-ascii string already has
|
| + # a space (trailing or leading depending on transition)
|
| + nextcs = charset
|
| + if nextcs == _charset.UNKNOWN8BIT:
|
| + original_bytes = string.encode('ascii', 'surrogateescape')
|
| + string = original_bytes.decode('ascii', 'replace')
|
| + if uchunks:
|
| + hasspace = string and self._nonctext(string[0])
|
| + if lastcs not in (None, 'us-ascii'):
|
| + if nextcs in (None, 'us-ascii') and not hasspace:
|
| + uchunks.append(SPACE)
|
| + nextcs = None
|
| + elif nextcs not in (None, 'us-ascii') and not lastspace:
|
| + uchunks.append(SPACE)
|
| + lastspace = string and self._nonctext(string[-1])
|
| + lastcs = nextcs
|
| + uchunks.append(string)
|
| + return EMPTYSTRING.join(uchunks)
|
| +
|
| + # Rich comparison operators for equality only. BAW: does it make sense to
|
| + # have or explicitly disable <, <=, >, >= operators?
|
| + def __eq__(self, other):
|
| + # other may be a Header or a string. Both are fine so coerce
|
| + # ourselves to a unicode (of the unencoded header value), swap the
|
| + # args and do another comparison.
|
| + return other == str(self)
|
| +
|
| + def __ne__(self, other):
|
| + return not self == other
|
| +
|
| + def append(self, s, charset=None, errors='strict'):
|
| + """Append a string to the MIME header.
|
| +
|
| + Optional charset, if given, should be a Charset instance or the name
|
| + of a character set (which will be converted to a Charset instance). A
|
| + value of None (the default) means that the charset given in the
|
| + constructor is used.
|
| +
|
| + s may be a byte string or a Unicode string. If it is a byte string
|
| + (i.e. isinstance(s, str) is false), then charset is the encoding of
|
| + that byte string, and a UnicodeError will be raised if the string
|
| + cannot be decoded with that charset. If s is a Unicode string, then
|
| + charset is a hint specifying the character set of the characters in
|
| + the string. In either case, when producing an RFC 2822 compliant
|
| + header using RFC 2047 rules, the string will be encoded using the
|
| + output codec of the charset. If the string cannot be encoded to the
|
| + output codec, a UnicodeError will be raised.
|
| +
|
| + Optional `errors' is passed as the errors argument to the decode
|
| + call if s is a byte string.
|
| + """
|
| + if charset is None:
|
| + charset = self._charset
|
| + elif not isinstance(charset, Charset):
|
| + charset = Charset(charset)
|
| + if not isinstance(s, str):
|
| + input_charset = charset.input_codec or 'us-ascii'
|
| + if input_charset == _charset.UNKNOWN8BIT:
|
| + s = s.decode('us-ascii', 'surrogateescape')
|
| + else:
|
| + s = s.decode(input_charset, errors)
|
| + # Ensure that the bytes we're storing can be decoded to the output
|
| + # character set, otherwise an early error is raised.
|
| + output_charset = charset.output_codec or 'us-ascii'
|
| + if output_charset != _charset.UNKNOWN8BIT:
|
| + try:
|
| + s.encode(output_charset, errors)
|
| + except UnicodeEncodeError:
|
| + if output_charset!='us-ascii':
|
| + raise
|
| + charset = UTF8
|
| + self._chunks.append((s, charset))
|
| +
|
| + def _nonctext(self, s):
|
| + """True if string s is not a ctext character of RFC822.
|
| + """
|
| + return s.isspace() or s in ('(', ')', '\\')
|
| +
|
| + def encode(self, splitchars=';, \t', maxlinelen=None, linesep='\n'):
|
| + r"""Encode a message header into an RFC-compliant format.
|
| +
|
| + There are many issues involved in converting a given string for use in
|
| + an email header. Only certain character sets are readable in most
|
| + email clients, and as header strings can only contain a subset of
|
| + 7-bit ASCII, care must be taken to properly convert and encode (with
|
| + Base64 or quoted-printable) header strings. In addition, there is a
|
| + 75-character length limit on any given encoded header field, so
|
| + line-wrapping must be performed, even with double-byte character sets.
|
| +
|
| + Optional maxlinelen specifies the maximum length of each generated
|
| + line, exclusive of the linesep string. Individual lines may be longer
|
| + than maxlinelen if a folding point cannot be found. The first line
|
| + will be shorter by the length of the header name plus ": " if a header
|
| + name was specified at Header construction time. The default value for
|
| + maxlinelen is determined at header construction time.
|
| +
|
| + Optional splitchars is a string containing characters which should be
|
| + given extra weight by the splitting algorithm during normal header
|
| + wrapping. This is in very rough support of RFC 2822's `higher level
|
| + syntactic breaks': split points preceded by a splitchar are preferred
|
| + during line splitting, with the characters preferred in the order in
|
| + which they appear in the string. Space and tab may be included in the
|
| + string to indicate whether preference should be given to one over the
|
| + other as a split point when other split chars do not appear in the line
|
| + being split. Splitchars does not affect RFC 2047 encoded lines.
|
| +
|
| + Optional linesep is a string to be used to separate the lines of
|
| + the value. The default value is the most useful for typical
|
| + Python applications, but it can be set to \r\n to produce RFC-compliant
|
| + line separators when needed.
|
| + """
|
| + self._normalize()
|
| + if maxlinelen is None:
|
| + maxlinelen = self._maxlinelen
|
| + # A maxlinelen of 0 means don't wrap. For all practical purposes,
|
| + # choosing a huge number here accomplishes that and makes the
|
| + # _ValueFormatter algorithm much simpler.
|
| + if maxlinelen == 0:
|
| + maxlinelen = 1000000
|
| + formatter = _ValueFormatter(self._headerlen, maxlinelen,
|
| + self._continuation_ws, splitchars)
|
| + lastcs = None
|
| + hasspace = lastspace = None
|
| + for string, charset in self._chunks:
|
| + if hasspace is not None:
|
| + hasspace = string and self._nonctext(string[0])
|
| + import sys
|
| + if lastcs not in (None, 'us-ascii'):
|
| + if not hasspace or charset not in (None, 'us-ascii'):
|
| + formatter.add_transition()
|
| + elif charset not in (None, 'us-ascii') and not lastspace:
|
| + formatter.add_transition()
|
| + lastspace = string and self._nonctext(string[-1])
|
| + lastcs = charset
|
| + hasspace = False
|
| + lines = string.splitlines()
|
| + if lines:
|
| + formatter.feed('', lines[0], charset)
|
| + else:
|
| + formatter.feed('', '', charset)
|
| + for line in lines[1:]:
|
| + formatter.newline()
|
| + if charset.header_encoding is not None:
|
| + formatter.feed(self._continuation_ws, ' ' + line.lstrip(),
|
| + charset)
|
| + else:
|
| + sline = line.lstrip()
|
| + fws = line[:len(line)-len(sline)]
|
| + formatter.feed(fws, sline, charset)
|
| + if len(lines) > 1:
|
| + formatter.newline()
|
| + if self._chunks:
|
| + formatter.add_transition()
|
| + value = formatter._str(linesep)
|
| + if _embeded_header.search(value):
|
| + raise HeaderParseError("header value appears to contain "
|
| + "an embedded header: {!r}".format(value))
|
| + return value
|
| +
|
| + def _normalize(self):
|
| + # Step 1: Normalize the chunks so that all runs of identical charsets
|
| + # get collapsed into a single unicode string.
|
| + chunks = []
|
| + last_charset = None
|
| + last_chunk = []
|
| + for string, charset in self._chunks:
|
| + if charset == last_charset:
|
| + last_chunk.append(string)
|
| + else:
|
| + if last_charset is not None:
|
| + chunks.append((SPACE.join(last_chunk), last_charset))
|
| + last_chunk = [string]
|
| + last_charset = charset
|
| + if last_chunk:
|
| + chunks.append((SPACE.join(last_chunk), last_charset))
|
| + self._chunks = chunks
|
| +
|
| +
|
| +class _ValueFormatter(object):
|
| + def __init__(self, headerlen, maxlen, continuation_ws, splitchars):
|
| + self._maxlen = maxlen
|
| + self._continuation_ws = continuation_ws
|
| + self._continuation_ws_len = len(continuation_ws)
|
| + self._splitchars = splitchars
|
| + self._lines = []
|
| + self._current_line = _Accumulator(headerlen)
|
| +
|
| + def _str(self, linesep):
|
| + self.newline()
|
| + return linesep.join(self._lines)
|
| +
|
| + def __str__(self):
|
| + return self._str(NL)
|
| +
|
| + def newline(self):
|
| + end_of_line = self._current_line.pop()
|
| + if end_of_line != (' ', ''):
|
| + self._current_line.push(*end_of_line)
|
| + if len(self._current_line) > 0:
|
| + if self._current_line.is_onlyws():
|
| + self._lines[-1] += str(self._current_line)
|
| + else:
|
| + self._lines.append(str(self._current_line))
|
| + self._current_line.reset()
|
| +
|
| + def add_transition(self):
|
| + self._current_line.push(' ', '')
|
| +
|
| + def feed(self, fws, string, charset):
|
| + # If the charset has no header encoding (i.e. it is an ASCII encoding)
|
| + # then we must split the header at the "highest level syntactic break"
|
| + # possible. Note that we don't have a lot of smarts about field
|
| + # syntax; we just try to break on semi-colons, then commas, then
|
| + # whitespace. Eventually, this should be pluggable.
|
| + if charset.header_encoding is None:
|
| + self._ascii_split(fws, string, self._splitchars)
|
| + return
|
| + # Otherwise, we're doing either a Base64 or a quoted-printable
|
| + # encoding which means we don't need to split the line on syntactic
|
| + # breaks. We can basically just find enough characters to fit on the
|
| + # current line, minus the RFC 2047 chrome. What makes this trickier
|
| + # though is that we have to split at octet boundaries, not character
|
| + # boundaries but it's only safe to split at character boundaries so at
|
| + # best we can only get close.
|
| + encoded_lines = charset.header_encode_lines(string, self._maxlengths())
|
| + # The first element extends the current line, but if it's None then
|
| + # nothing more fit on the current line so start a new line.
|
| + try:
|
| + first_line = encoded_lines.pop(0)
|
| + except IndexError:
|
| + # There are no encoded lines, so we're done.
|
| + return
|
| + if first_line is not None:
|
| + self._append_chunk(fws, first_line)
|
| + try:
|
| + last_line = encoded_lines.pop()
|
| + except IndexError:
|
| + # There was only one line.
|
| + return
|
| + self.newline()
|
| + self._current_line.push(self._continuation_ws, last_line)
|
| + # Everything else are full lines in themselves.
|
| + for line in encoded_lines:
|
| + self._lines.append(self._continuation_ws + line)
|
| +
|
| + def _maxlengths(self):
|
| + # The first line's length.
|
| + yield self._maxlen - len(self._current_line)
|
| + while True:
|
| + yield self._maxlen - self._continuation_ws_len
|
| +
|
| + def _ascii_split(self, fws, string, splitchars):
|
| + # The RFC 2822 header folding algorithm is simple in principle but
|
| + # complex in practice. Lines may be folded any place where "folding
|
| + # white space" appears by inserting a linesep character in front of the
|
| + # FWS. The complication is that not all spaces or tabs qualify as FWS,
|
| + # and we are also supposed to prefer to break at "higher level
|
| + # syntactic breaks". We can't do either of these without intimate
|
| + # knowledge of the structure of structured headers, which we don't have
|
| + # here. So the best we can do here is prefer to break at the specified
|
| + # splitchars, and hope that we don't choose any spaces or tabs that
|
| + # aren't legal FWS. (This is at least better than the old algorithm,
|
| + # where we would sometimes *introduce* FWS after a splitchar, or the
|
| + # algorithm before that, where we would turn all white space runs into
|
| + # single spaces or tabs.)
|
| + parts = re.split("(["+FWS+"]+)", fws+string)
|
| + if parts[0]:
|
| + parts[:0] = ['']
|
| + else:
|
| + parts.pop(0)
|
| + for fws, part in zip(*[iter(parts)]*2):
|
| + self._append_chunk(fws, part)
|
| +
|
| + def _append_chunk(self, fws, string):
|
| + self._current_line.push(fws, string)
|
| + if len(self._current_line) > self._maxlen:
|
| + # Find the best split point, working backward from the end.
|
| + # There might be none, on a long first line.
|
| + for ch in self._splitchars:
|
| + for i in range(self._current_line.part_count()-1, 0, -1):
|
| + if ch.isspace():
|
| + fws = self._current_line[i][0]
|
| + if fws and fws[0]==ch:
|
| + break
|
| + prevpart = self._current_line[i-1][1]
|
| + if prevpart and prevpart[-1]==ch:
|
| + break
|
| + else:
|
| + continue
|
| + break
|
| + else:
|
| + fws, part = self._current_line.pop()
|
| + if self._current_line._initial_size > 0:
|
| + # There will be a header, so leave it on a line by itself.
|
| + self.newline()
|
| + if not fws:
|
| + # We don't use continuation_ws here because the whitespace
|
| + # after a header should always be a space.
|
| + fws = ' '
|
| + self._current_line.push(fws, part)
|
| + return
|
| + remainder = self._current_line.pop_from(i)
|
| + self._lines.append(str(self._current_line))
|
| + self._current_line.reset(remainder)
|
| +
|
| +
|
| +class _Accumulator(list):
|
| +
|
| + def __init__(self, initial_size=0):
|
| + self._initial_size = initial_size
|
| + super().__init__()
|
| +
|
| + def push(self, fws, string):
|
| + self.append((fws, string))
|
| +
|
| + def pop_from(self, i=0):
|
| + popped = self[i:]
|
| + self[i:] = []
|
| + return popped
|
| +
|
| + def pop(self):
|
| + if self.part_count()==0:
|
| + return ('', '')
|
| + return super().pop()
|
| +
|
| + def __len__(self):
|
| + return sum((len(fws)+len(part) for fws, part in self),
|
| + self._initial_size)
|
| +
|
| + def __str__(self):
|
| + return EMPTYSTRING.join((EMPTYSTRING.join((fws, part))
|
| + for fws, part in self))
|
| +
|
| + def reset(self, startval=None):
|
| + if startval is None:
|
| + startval = []
|
| + self[:] = startval
|
| + self._initial_size = 0
|
| +
|
| + def is_onlyws(self):
|
| + return self._initial_size==0 and (not self or str(self).isspace())
|
| +
|
| + def part_count(self):
|
| + return super().__len__()
|
|
|