Chromium Code Reviews| Index: pkg/yaml/lib/src/parser.dart |
| diff --git a/pkg/yaml/lib/src/parser.dart b/pkg/yaml/lib/src/parser.dart |
| index 94f551f21a8745a13ed98faea3d76176c06ce093..cf369ea415b37ffe8a41e0da83a0e680e2d77b54 100644 |
| --- a/pkg/yaml/lib/src/parser.dart |
| +++ b/pkg/yaml/lib/src/parser.dart |
| @@ -1,1900 +1,834 @@ |
| -// Copyright (c) 2012, the Dart project authors. Please see the AUTHORS file |
| +// Copyright (c) 2014, the Dart project authors. Please see the AUTHORS file |
| // for details. All rights reserved. Use of this source code is governed by a |
| // BSD-style license that can be found in the LICENSE file. |
| library yaml.parser; |
| -import 'dart:collection'; |
| - |
| import 'package:source_span/source_span.dart'; |
| import 'package:string_scanner/string_scanner.dart'; |
| -import 'equality.dart'; |
| -import 'model.dart'; |
| +import 'event.dart'; |
| +import 'scanner.dart'; |
| +import 'style.dart'; |
| +import 'token.dart'; |
| import 'utils.dart'; |
| +import 'yaml_document.dart'; |
| +import 'yaml_exception.dart'; |
| -/// Translates a string of characters into a YAML serialization tree. |
| -/// |
| -/// This parser is designed to closely follow the spec. All productions in the |
| -/// spec are numbered, and the corresponding methods in the parser have the same |
| -/// numbers. This is certainly not the most efficient way of parsing YAML, but |
| -/// it is the easiest to write and read in the context of the spec. |
| +/// A parser that reads [Token]s emitted by a [Scanner] and emits [Event]s. |
| /// |
| -/// Methods corresponding to productions are also named as in the spec, |
| -/// translating the name of the method (although not the annotation characters) |
| -/// into camel-case for dart style.. For example, the spec has a production |
| -/// named `nb-ns-plain-in-line`, and the method implementing it is named |
| -/// `nb_ns_plainInLine`. The exception to that rule is methods that just |
| -/// recognize character classes; these are named `is*`. |
| +/// This is based on the libyaml parser, available at |
| +/// https://github.com/yaml/libyaml/blob/master/src/parser.c. The license for |
| +/// that is available in ../../libyaml-license.txt. |
| class Parser { |
| - static const TAB = 0x9; |
| - static const LF = 0xA; |
| - static const CR = 0xD; |
| - static const SP = 0x20; |
| - static const TILDE = 0x7E; |
| - static const NEL = 0x85; |
| - static const PLUS = 0x2B; |
| - static const HYPHEN = 0x2D; |
| - static const QUESTION_MARK = 0x3F; |
| - static const COLON = 0x3A; |
| - static const COMMA = 0x2C; |
| - static const LEFT_BRACKET = 0x5B; |
| - static const RIGHT_BRACKET = 0x5D; |
| - static const LEFT_BRACE = 0x7B; |
| - static const RIGHT_BRACE = 0x7D; |
| - static const HASH = 0x23; |
| - static const AMPERSAND = 0x26; |
| - static const ASTERISK = 0x2A; |
| - static const EXCLAMATION = 0x21; |
| - static const VERTICAL_BAR = 0x7C; |
| - static const GREATER_THAN = 0x3E; |
| - static const SINGLE_QUOTE = 0x27; |
| - static const DOUBLE_QUOTE = 0x22; |
| - static const PERCENT = 0x25; |
| - static const AT = 0x40; |
| - static const GRAVE_ACCENT = 0x60; |
| - |
| - static const NULL = 0x0; |
| - static const BELL = 0x7; |
| - static const BACKSPACE = 0x8; |
| - static const VERTICAL_TAB = 0xB; |
| - static const FORM_FEED = 0xC; |
| - static const ESCAPE = 0x1B; |
| - static const SLASH = 0x2F; |
| - static const BACKSLASH = 0x5C; |
| - static const UNDERSCORE = 0x5F; |
| - static const NBSP = 0xA0; |
| - static const LINE_SEPARATOR = 0x2028; |
| - static const PARAGRAPH_SEPARATOR = 0x2029; |
| - |
| - static const NUMBER_0 = 0x30; |
| - static const NUMBER_9 = 0x39; |
| - |
| - static const LETTER_A = 0x61; |
| - static const LETTER_B = 0x62; |
| - static const LETTER_E = 0x65; |
| - static const LETTER_F = 0x66; |
| - static const LETTER_N = 0x6E; |
| - static const LETTER_R = 0x72; |
| - static const LETTER_T = 0x74; |
| - static const LETTER_U = 0x75; |
| - static const LETTER_V = 0x76; |
| - static const LETTER_X = 0x78; |
| - |
| - static const LETTER_CAP_A = 0x41; |
| - static const LETTER_CAP_F = 0x46; |
| - static const LETTER_CAP_L = 0x4C; |
| - static const LETTER_CAP_N = 0x4E; |
| - static const LETTER_CAP_P = 0x50; |
| - static const LETTER_CAP_U = 0x55; |
| - static const LETTER_CAP_X = 0x58; |
| - |
| - static const C_SEQUENCE_ENTRY = 4; |
| - static const C_MAPPING_KEY = 5; |
| - static const C_MAPPING_VALUE = 6; |
| - static const C_COLLECT_ENTRY = 7; |
| - static const C_SEQUENCE_START = 8; |
| - static const C_SEQUENCE_END = 9; |
| - static const C_MAPPING_START = 10; |
| - static const C_MAPPING_END = 11; |
| - static const C_COMMENT = 12; |
| - static const C_ANCHOR = 13; |
| - static const C_ALIAS = 14; |
| - static const C_TAG = 15; |
| - static const C_LITERAL = 16; |
| - static const C_FOLDED = 17; |
| - static const C_SINGLE_QUOTE = 18; |
| - static const C_DOUBLE_QUOTE = 19; |
| - static const C_DIRECTIVE = 20; |
| - static const C_RESERVED = 21; |
| - |
| - static const BLOCK_OUT = 0; |
| - static const BLOCK_IN = 1; |
| - static const FLOW_OUT = 2; |
| - static const FLOW_IN = 3; |
| - static const BLOCK_KEY = 4; |
| - static const FLOW_KEY = 5; |
| - |
| - static const CHOMPING_STRIP = 0; |
| - static const CHOMPING_KEEP = 1; |
| - static const CHOMPING_CLIP = 2; |
| - |
| - /// The scanner that's used to scan through the document. |
| - final SpanScanner _scanner; |
| - |
| - /// Whether we're parsing a bare document (that is, one that doesn't begin |
| - /// with `---`). Bare documents don't allow `%` immediately following |
| - /// newlines. |
| - bool _inBareDocument = false; |
| - |
| - /// The state of the scanner when it was the farthest in the document it's |
| - /// been. |
| - LineScannerState _farthestState; |
| - |
| - /// The name of the context of the farthest position that has been parsed |
| - /// successfully before backtracking. Used for error reporting. |
| - String _farthestContext = "document"; |
| - |
| - /// A stack of the names of parse contexts. Used for error reporting. |
| - final _contextStack = <String>["document"]; |
| - |
| - /// Annotations attached to ranges of the source string that add extra |
| - /// information to any errors that occur in the annotated range. |
| - final _errorAnnotations = new _RangeMap<String>(); |
| - |
| - /// The buffer containing the string currently being captured. |
| - StringBuffer _capturedString; |
| - |
| - /// The beginning of the current section of the captured string. |
| - int _captureStart; |
| - |
| - /// Whether the current string capture is being overridden. |
| - bool _capturingAs = false; |
| - |
| - Parser(String yaml, sourceUrl) |
| - : _scanner = new SpanScanner(yaml, sourceUrl: sourceUrl) { |
| - _farthestState = _scanner.state; |
| - } |
| + /// The underlying [Scanner] that generates [Token]s. |
| + final Scanner _scanner; |
| - /// Returns the character at the current position, then moves that position |
| - /// forward one character. |
| - int next() => _scanner.readChar(); |
| + /// The stack of parse states for nested contexts. |
| + final _states = new List<_State>(); |
| - /// Returns the code unit at the current position, or the character [i] |
| - /// characters after the current position. |
| - int peek([int i = 0]) => _scanner.peekChar(i); |
| + /// The current parse state. |
| + var _state = _State.STREAM_START; |
| - /// The truthiness operator. Returns `false` if [obj] is `null` or `false`, |
| - /// `true` otherwise. |
| - bool truth(obj) => obj != null && obj != false; |
| - |
| - /// Consumes the current character if it matches [matcher]. Returns the result |
| - /// of [matcher]. |
| - bool consume(bool matcher(int)) { |
| - if (matcher(peek())) { |
| - next(); |
| - return true; |
| - } |
| - return false; |
| - } |
| + /// The custom tag directives, by tag handle. |
| + final _tagDirectives = new Map<String, TagDirective>(); |
| - /// Consumes the current character if it equals [char]. |
| - bool consumeChar(int char) => consume((c) => c == char); |
| + /// Whether the parser has finished parsing. |
| + bool get isDone => _state == _State.END; |
| - /// Calls [consumer] until it returns a falsey value. Returns a list of all |
| - /// truthy return values of [consumer], or null if it didn't consume anything. |
| + /// Creates a parser that parses [source]. |
| /// |
| - /// Conceptually, repeats a production one or more times. |
| - List oneOrMore(consumer()) { |
| - var first = consumer(); |
| - if (!truth(first)) return null; |
| - var out = [first]; |
| - while (true) { |
| - var el = consumer(); |
| - if (!truth(el)) return out; |
| - out.add(el); |
| - } |
| - return null; // Unreachable. |
| - } |
| + /// [sourceUrl] can be a String or a [Uri]. |
| + Parser(String source, {sourceUrl}) |
| + : _scanner = new Scanner(source, sourceUrl: sourceUrl); |
| - /// Calls [consumer] until it returns a falsey value. Returns a list of all |
| - /// truthy return values of [consumer], or the empty list if it didn't consume |
| - /// anything. |
| - /// |
| - /// Conceptually, repeats a production any number of times. |
| - List zeroOrMore(consumer()) { |
| - var out = []; |
| - var oldPos = _scanner.position; |
| - while (true) { |
| - var el = consumer(); |
| - if (!truth(el) || oldPos == _scanner.position) return out; |
| - oldPos = _scanner.position; |
| - out.add(el); |
| + /// Consumes and returns the next event. |
| + Event parse() { |
| + try { |
| + if (isDone) throw new StateError("No more events."); |
| + var event = _stateMachine(); |
| + return event; |
| + } on StringScannerException catch (error) { |
| + throw new YamlException(error.message, error.span); |
| } |
| - return null; // Unreachable. |
| } |
| - /// Just calls [consumer] and returns its result. Used to make it explicit |
| - /// that a production is intended to be optional. |
| - zeroOrOne(consumer()) => consumer(); |
| - |
| - /// Calls each function in [consumers] until one returns a truthy value, then |
| - /// returns that. |
| - or(List<Function> consumers) { |
| - for (var c in consumers) { |
| - var res = c(); |
| - if (truth(res)) return res; |
| + /// Dispatches parsing based on the current state. |
| + Event _stateMachine() { |
| + switch (_state) { |
| + case _State.STREAM_START: |
| + return _parseStreamStart(); |
| + case _State.DOCUMENT_START: |
| + return _parseDocumentStart(); |
| + case _State.DOCUMENT_CONTENT: |
| + return _parseDocumentContent(); |
| + case _State.DOCUMENT_END: |
| + return _parseDocumentEnd(); |
| + case _State.BLOCK_NODE: |
| + return _parseNode(block: true); |
| + case _State.BLOCK_NODE_OR_INDENTLESS_SEQUENCE: |
| + return _parseNode(block: true, indentlessSequence: true); |
| + case _State.FLOW_NODE: |
| + return _parseNode(); |
| + case _State.BLOCK_SEQUENCE_FIRST_ENTRY: |
|
Bob Nystrom
2014/10/31 20:03:27
How about just putting _scanner.scan() here and re
nweiz
2014/11/04 22:19:36
Done.
|
| + return _parseBlockSequenceEntry(first: true); |
| + case _State.BLOCK_SEQUENCE_ENTRY: |
| + return _parseBlockSequenceEntry(); |
| + case _State.INDENTLESS_SEQUENCE_ENTRY: |
| + return _parseIndentlessSequenceEntry(); |
| + case _State.BLOCK_MAPPING_FIRST_KEY: |
| + return _parseBlockMappingKey(first: true); |
| + case _State.BLOCK_MAPPING_KEY: |
| + return _parseBlockMappingKey(); |
| + case _State.BLOCK_MAPPING_VALUE: |
| + return _parseBlockMappingValue(); |
| + case _State.FLOW_SEQUENCE_FIRST_ENTRY: |
| + return _parseFlowSequenceEntry(first: true); |
| + case _State.FLOW_SEQUENCE_ENTRY: |
| + return _parseFlowSequenceEntry(); |
| + case _State.FLOW_SEQUENCE_ENTRY_MAPPING_KEY: |
| + return _parseFlowSequenceEntryMappingKey(); |
| + case _State.FLOW_SEQUENCE_ENTRY_MAPPING_VALUE: |
| + return _parseFlowSequenceEntryMappingValue(); |
| + case _State.FLOW_SEQUENCE_ENTRY_MAPPING_END: |
| + return _parseFlowSequenceEntryMappingEnd(); |
| + case _State.FLOW_MAPPING_FIRST_KEY: |
| + return _parseFlowMappingKey(first: true); |
| + case _State.FLOW_MAPPING_KEY: |
| + return _parseFlowMappingKey(); |
| + case _State.FLOW_MAPPING_VALUE: |
| + return _parseFlowMappingValue(); |
| + case _State.FLOW_MAPPING_EMPTY_VALUE: |
| + return _parseFlowMappingValue(empty: true); |
| + default: |
| + throw "Unreachable"; |
| } |
| - return null; |
| } |
| - /// Calls [consumer] and returns its result, but rolls back the parser state |
| - /// if [consumer] returns a falsey value. |
| - transaction(consumer()) { |
| - var oldState = _scanner.state; |
| - var oldCaptureStart = _captureStart; |
| - String capturedSoFar = _capturedString == null ? null : |
| - _capturedString.toString(); |
| - var res = consumer(); |
| - _refreshFarthestState(); |
| - if (truth(res)) return res; |
| - |
| - _scanner.state = oldState; |
| - _captureStart = oldCaptureStart; |
| - _capturedString = capturedSoFar == null ? null : |
| - new StringBuffer(capturedSoFar); |
| - return res; |
| + /// Parses the production: |
| + /// |
| + /// stream ::= |
| + /// STREAM-START implicit_document? explicit_document* STREAM-END |
| + /// ************ |
| + Event _parseStreamStart() { |
| + var token = _scanner.scan(); |
| + assert(token.type == TokenType.STREAM_START); |
| + |
| + _state = _State.DOCUMENT_START; |
| + return new Event(EventType.STREAM_START, token.span); |
| } |
| - /// Consumes [n] characters matching [matcher], or none if there isn't a |
| - /// complete match. The first argument to [matcher] is the character code, the |
| - /// second is the index (from 0 to [n] - 1). |
| + /// Parses the productions: |
| /// |
| - /// Returns whether or not the characters were consumed. |
| - bool nAtOnce(int n, bool matcher(int c, int i)) => transaction(() { |
| - for (int i = 0; i < n; i++) { |
| - if (!consume((c) => matcher(c, i))) return false; |
| + /// implicit_document ::= block_node DOCUMENT-END* |
| + /// * |
| + /// explicit_document ::= |
| + /// DIRECTIVE* DOCUMENT-START block_node? DOCUMENT-END* |
| + /// ************************* |
| + Event _parseDocumentStart() { |
| + var token = _scanner.peek(); |
| + |
| + // libyaml requires any document beyond the first in the stream to have an |
| + // explicit document start indicator, but the spec allows it to be omitted |
| + // as long as there was an end indicator. |
| + |
| + // Parse extra document end indicators. |
| + while (token.type == TokenType.DOCUMENT_END) { |
| + _scanner.scan(); |
| + token = _scanner.peek(); |
| } |
| - return true; |
| - }); |
| - /// Consumes the exact characters in [str], or nothing. |
| - /// |
| - /// Returns whether or not the string was consumed. |
| - bool rawString(String str) => |
| - nAtOnce(str.length, (c, i) => str.codeUnitAt(i) == c); |
| - |
| - /// Consumes and returns a string of characters matching [matcher], or null if |
| - /// there are no such characters. |
| - String stringOf(bool matcher(int)) => |
| - captureString(() => oneOrMore(() => consume(matcher))); |
| - |
| - /// Calls [consumer] and returns the string that was consumed while doing so, |
| - /// or null if [consumer] returned a falsey value. Automatically wraps |
| - /// [consumer] in `transaction`. |
| - String captureString(consumer()) { |
| - // captureString calls may not be nested |
| - assert(_capturedString == null); |
| - |
| - _captureStart = _scanner.position; |
| - _capturedString = new StringBuffer(); |
| - var res = transaction(consumer); |
| - if (!truth(res)) { |
| - _captureStart = null; |
| - _capturedString = null; |
| - return null; |
| + if (token.type != TokenType.VERSION_DIRECTIVE && |
| + token.type != TokenType.TAG_DIRECTIVE && |
| + token.type != TokenType.DOCUMENT_START && |
| + token.type != TokenType.STREAM_END) { |
| + // Parse an implicit document. |
| + _processDirectives(); |
| + _states.add(_State.DOCUMENT_END); |
| + _state = _State.BLOCK_NODE; |
| + return new DocumentStartEvent(token.span.start.pointSpan()); |
| } |
| - flushCapture(); |
| - var result = _capturedString.toString(); |
| - _captureStart = null; |
| - _capturedString = null; |
| - return result; |
| - } |
| - |
| - captureAs(String replacement, consumer()) => |
| - captureAndTransform(consumer, (_) => replacement); |
| - |
| - captureAndTransform(consumer(), String transformation(String captured)) { |
| - if (_capturedString == null) return consumer(); |
| - if (_capturingAs) return consumer(); |
| - |
| - flushCapture(); |
| - _capturingAs = true; |
| - var res = consumer(); |
| - _capturingAs = false; |
| - if (!truth(res)) return res; |
| - |
| - _capturedString.write(transformation( |
| - _scanner.string.substring(_captureStart, _scanner.position))); |
| - _captureStart = _scanner.position; |
| - return res; |
| - } |
| - |
| - void flushCapture() { |
| - _capturedString.write(_scanner.string.substring( |
| - _captureStart, _scanner.position)); |
| - _captureStart = _scanner.position; |
| - } |
| - |
| - /// Adds a tag and an anchor to [node], if they're defined. |
| - Node addProps(Node node, Pair<Tag, String> props) { |
| - if (props == null || node == null) return node; |
| - if (truth(props.first)) node.tag = props.first; |
| - if (truth(props.last)) node.anchor = props.last; |
| - return node; |
| - } |
| - |
| - /// Creates a MappingNode from [pairs]. |
| - MappingNode map(List<Pair<Node, Node>> pairs, SourceSpan span) { |
| - var content = new Map<Node, Node>(); |
| - pairs.forEach((pair) => content[pair.first] = pair.last); |
| - return new MappingNode("?", content, span); |
| - } |
| - |
| - /// Runs [fn] in a context named [name]. Used for error reporting. |
| - context(String name, fn()) { |
| - try { |
| - _contextStack.add(name); |
| - return fn(); |
| - } finally { |
| - var popped = _contextStack.removeLast(); |
| - assert(popped == name); |
| + if (token.type == TokenType.STREAM_END) { |
| + _state = _State.END; |
| + _scanner.scan(); |
| + return new Event(EventType.STREAM_END, token.span); |
| } |
| - } |
| - /// Adds [message] as extra information to any errors that occur between the |
| - /// current position and the position of the cursor after running [fn]. The |
| - /// cursor is reset after [fn] is run. |
| - annotateError(String message, fn()) { |
| - var start = _scanner.position; |
| - var end; |
| - transaction(() { |
| - fn(); |
| - end = _scanner.position; |
| - return false; |
| - }); |
| - _errorAnnotations[new _Range(start, end)] = message; |
| - } |
| - |
| - /// Throws an error with additional context information. |
| - void error(String message) => |
| - _scanner.error("$message (in $_farthestContext)."); |
| + // Parse an explicit document. |
| + var start = token.span; |
| + var pair = _processDirectives(); |
| + var versionDirective = pair.first; |
| + var tagDirectives = pair.last; |
| + token = _scanner.peek(); |
| + if (token.type != TokenType.DOCUMENT_START) { |
| + throw new YamlException("Expected document start.", token.span); |
| + } |
| - /// If [result] is falsey, throws an error saying that [expected] was |
| - /// expected. |
| - expect(result, String expected) { |
| - if (truth(result)) return result; |
| - error("Expected $expected"); |
| + _states.add(_State.DOCUMENT_END); |
| + _state = _State.DOCUMENT_CONTENT; |
| + _scanner.scan(); |
| + return new DocumentStartEvent(start.expand(token.span), |
| + versionDirective: versionDirective, |
| + tagDirectives: tagDirectives, |
| + implicit: false); |
| } |
| - /// Throws an error saying that the parse failed. |
| + /// Parses the productions: |
| /// |
| - /// Uses [_farthestState] and [_farthestContext] to provide additional |
| - /// information. |
| - parseFailed() { |
| - var message = "Invalid YAML in $_farthestContext"; |
| - _refreshFarthestState(); |
| - _scanner.state = _farthestState; |
| - |
| - var extraError = _errorAnnotations[_scanner.position]; |
| - if (extraError != null) message = "$message ($extraError)"; |
| - _scanner.error("$message."); |
| - } |
| - |
| - /// Update [_farthestState] if the scanner is farther than it's been before. |
| - void _refreshFarthestState() { |
| - if (_scanner.position <= _farthestState.position) return; |
| - _farthestState = _scanner.state; |
| - } |
| - |
| - /// Returns the number of spaces after the current position. |
| - int countIndentation() { |
| - var i = 0; |
| - while (peek(i) == SP) i++; |
| - return i; |
| - } |
| - |
| - /// Returns the indentation for a block scalar. |
| - int blockScalarAdditionalIndentation(_BlockHeader header, int indent) { |
| - if (!header.autoDetectIndent) return header.additionalIndent; |
| - |
| - var maxSpaces = 0; |
| - var spaces = 0; |
| - transaction(() { |
| - do { |
| - spaces = captureString(() => zeroOrMore(() => consumeChar(SP))).length; |
| - if (spaces > maxSpaces) maxSpaces = spaces; |
| - } while (b_break()); |
| - return false; |
| - }); |
| - |
| - // If the next non-empty line isn't indented further than the start of the |
| - // block scalar, that means the scalar is going to be empty. Returning any |
| - // value > 0 will cause the parser not to consume any text. |
| - if (spaces <= indent) return 1; |
| - |
| - // It's an error for a leading empty line to be indented more than the first |
| - // non-empty line. |
| - if (maxSpaces > spaces) { |
| - _scanner.error("Leading empty lines may not be indented more than the " |
| - "first non-empty line."); |
| + /// explicit_document ::= |
| + /// DIRECTIVE* DOCUMENT-START block_node? DOCUMENT-END* |
| + /// *********** |
| + Event _parseDocumentContent() { |
| + var token = _scanner.peek(); |
| + |
| + switch (token.type) { |
| + case TokenType.VERSION_DIRECTIVE: |
| + case TokenType.TAG_DIRECTIVE: |
| + case TokenType.DOCUMENT_START: |
| + case TokenType.DOCUMENT_END: |
| + case TokenType.STREAM_END: |
| + _state = _states.removeLast(); |
| + return _processEmptyScalar(token.span.start); |
| + default: |
| + return _parseNode(block: true); |
| } |
| - |
| - return spaces - indent; |
| } |
| - /// Returns whether the current position is at the beginning of a line. |
| - bool get atStartOfLine => _scanner.column == 0; |
| - |
| - /// Given an indicator character, returns the type of that indicator (or null |
| - /// if the indicator isn't found. |
| - int indicatorType(int char) { |
| - switch (char) { |
| - case HYPHEN: return C_SEQUENCE_ENTRY; |
| - case QUESTION_MARK: return C_MAPPING_KEY; |
| - case COLON: return C_MAPPING_VALUE; |
| - case COMMA: return C_COLLECT_ENTRY; |
| - case LEFT_BRACKET: return C_SEQUENCE_START; |
| - case RIGHT_BRACKET: return C_SEQUENCE_END; |
| - case LEFT_BRACE: return C_MAPPING_START; |
| - case RIGHT_BRACE: return C_MAPPING_END; |
| - case HASH: return C_COMMENT; |
| - case AMPERSAND: return C_ANCHOR; |
| - case ASTERISK: return C_ALIAS; |
| - case EXCLAMATION: return C_TAG; |
| - case VERTICAL_BAR: return C_LITERAL; |
| - case GREATER_THAN: return C_FOLDED; |
| - case SINGLE_QUOTE: return C_SINGLE_QUOTE; |
| - case DOUBLE_QUOTE: return C_DOUBLE_QUOTE; |
| - case PERCENT: return C_DIRECTIVE; |
| - case AT: |
| - case GRAVE_ACCENT: |
| - return C_RESERVED; |
| - default: return null; |
| + /// Parses the productions: |
| + /// |
| + /// implicit_document ::= block_node DOCUMENT-END* |
| + /// ************* |
| + /// explicit_document ::= |
| + /// DIRECTIVE* DOCUMENT-START block_node? DOCUMENT-END* |
| + /// ************* |
| + Event _parseDocumentEnd() { |
| + var token = _scanner.peek(); |
| + var span = token.span.start.pointSpan(); |
| + var implicit = true; |
| + |
| + if (token.type == TokenType.DOCUMENT_END) { |
| + span = token.span; |
| + implicit = false; |
| + _scanner.scan(); |
| } |
| - } |
| - // 1 |
| - bool isPrintable(int char) { |
| - if (char == null) return false; |
| - return char == TAB || |
| - char == LF || |
| - char == CR || |
| - (char >= SP && char <= TILDE) || |
| - char == NEL || |
| - (char >= 0xA0 && char <= 0xD7FF) || |
| - (char >= 0xE000 && char <= 0xFFFD) || |
| - (char >= 0x10000 && char <= 0x10FFFF); |
| + _tagDirectives.clear(); |
| + _state = _State.DOCUMENT_START; |
| + return new DocumentEndEvent(span, implicit: implicit); |
|
Bob Nystrom
2014/10/31 20:03:27
How about just:
Event _parseDocumentEnd() {
_ta
nweiz
2014/11/04 22:19:36
Done.
|
| } |
| - // 2 |
| - bool isJson(int char) => char != null && |
| - (char == TAB || (char >= SP && char <= 0x10FFFF)); |
| - |
| - // 22 |
| - bool c_indicator(int type) => consume((c) => indicatorType(c) == type); |
| - |
| - // 23 |
| - bool isFlowIndicator(int char) { |
| - var indicator = indicatorType(char); |
| - return indicator == C_COLLECT_ENTRY || |
| - indicator == C_SEQUENCE_START || |
| - indicator == C_SEQUENCE_END || |
| - indicator == C_MAPPING_START || |
| - indicator == C_MAPPING_END; |
| - } |
| - |
| - // 26 |
| - bool isBreak(int char) => char == LF || char == CR; |
| - |
| - // 27 |
| - bool isNonBreak(int char) => isPrintable(char) && !isBreak(char); |
| - |
| - // 28 |
| - bool b_break() { |
| - if (consumeChar(CR)) { |
| - zeroOrOne(() => consumeChar(LF)); |
| - return true; |
| + /// Parses the productions: |
| + /// |
| + /// block_node_or_indentless_sequence ::= |
| + /// ALIAS |
| + /// ***** |
| + /// | properties (block_content | indentless_block_sequence)? |
| + /// ********** * |
| + /// | block_content | indentless_block_sequence |
| + /// * |
| + /// block_node ::= ALIAS |
| + /// ***** |
| + /// | properties block_content? |
| + /// ********** * |
| + /// | block_content |
| + /// * |
| + /// flow_node ::= ALIAS |
| + /// ***** |
| + /// | properties flow_content? |
| + /// ********** * |
| + /// | flow_content |
| + /// * |
| + /// properties ::= TAG ANCHOR? | ANCHOR TAG? |
| + /// ************************* |
| + /// block_content ::= block_collection | flow_collection | SCALAR |
| + /// ****** |
| + /// flow_content ::= flow_collection | SCALAR |
| + /// ****** |
| + Event _parseNode({bool block: false, bool indentlessSequence: false}) { |
| + var token = _scanner.peek(); |
| + |
| + if (token is AliasToken) { |
| + _scanner.scan(); |
| + _state = _states.removeLast(); |
| + return new AliasEvent(token.span, token.name); |
| } |
| - return consumeChar(LF); |
| - } |
| - |
| - // 29 |
| - bool b_asLineFeed() => captureAs("\n", () => b_break()); |
| - |
| - // 30 |
| - bool b_nonContent() => captureAs("", () => b_break()); |
| - |
| - // 33 |
| - bool isSpace(int char) => char == SP || char == TAB; |
| - |
| - // 34 |
| - bool isNonSpace(int char) => isNonBreak(char) && !isSpace(char); |
| - |
| - // 35 |
| - bool isDecDigit(int char) => char != null && char >= NUMBER_0 && |
| - char <= NUMBER_9; |
| - |
| - // 36 |
| - bool isHexDigit(int char) { |
| - if (char == null) return false; |
| - return isDecDigit(char) || |
| - (char >= LETTER_A && char <= LETTER_F) || |
| - (char >= LETTER_CAP_A && char <= LETTER_CAP_F); |
| - } |
| - |
| - // 41 |
| - bool c_escape() => captureAs("", () => consumeChar(BACKSLASH)); |
| - |
| - // 42 |
| - bool ns_escNull() => captureAs("\x00", () => consumeChar(NUMBER_0)); |
| - |
| - // 43 |
| - bool ns_escBell() => captureAs("\x07", () => consumeChar(LETTER_A)); |
| - |
| - // 44 |
| - bool ns_escBackspace() => captureAs("\b", () => consumeChar(LETTER_B)); |
| - |
| - // 45 |
| - bool ns_escHorizontalTab() => captureAs("\t", () { |
| - return consume((c) => c == LETTER_T || c == TAB); |
| - }); |
| - |
| - // 46 |
| - bool ns_escLineFeed() => captureAs("\n", () => consumeChar(LETTER_N)); |
| - |
| - // 47 |
| - bool ns_escVerticalTab() => captureAs("\v", () => consumeChar(LETTER_V)); |
| - |
| - // 48 |
| - bool ns_escFormFeed() => captureAs("\f", () => consumeChar(LETTER_F)); |
| - |
| - // 49 |
| - bool ns_escCarriageReturn() => captureAs("\r", () => consumeChar(LETTER_R)); |
| - |
| - // 50 |
| - bool ns_escEscape() => captureAs("\x1B", () => consumeChar(LETTER_E)); |
| - |
| - // 51 |
| - bool ns_escSpace() => consumeChar(SP); |
| - |
| - // 52 |
| - bool ns_escDoubleQuote() => consumeChar(DOUBLE_QUOTE); |
| - |
| - // 53 |
| - bool ns_escSlash() => consumeChar(SLASH); |
| - |
| - // 54 |
| - bool ns_escBackslash() => consumeChar(BACKSLASH); |
| - |
| - // 55 |
| - bool ns_escNextLine() => captureAs("\x85", () => consumeChar(LETTER_CAP_N)); |
| - // 56 |
| - bool ns_escNonBreakingSpace() => |
| - captureAs("\xA0", () => consumeChar(UNDERSCORE)); |
| - |
| - // 57 |
| - bool ns_escLineSeparator() => |
| - captureAs("\u2028", () => consumeChar(LETTER_CAP_L)); |
| - |
| - // 58 |
| - bool ns_escParagraphSeparator() => |
| - captureAs("\u2029", () => consumeChar(LETTER_CAP_P)); |
| - |
| - // 59 |
| - bool ns_esc8Bit() => ns_escNBit(LETTER_X, 2); |
| - |
| - // 60 |
| - bool ns_esc16Bit() => ns_escNBit(LETTER_U, 4); |
| - |
| - // 61 |
| - bool ns_esc32Bit() => ns_escNBit(LETTER_CAP_U, 8); |
| - |
| - // Helper method for 59 - 61 |
| - bool ns_escNBit(int char, int digits) { |
| - if (!captureAs('', () => consumeChar(char))) return false; |
| - var captured = captureAndTransform( |
| - () => nAtOnce(digits, (c, _) => isHexDigit(c)), |
| - (hex) => new String.fromCharCodes([int.parse("0x$hex")])); |
| - return expect(captured, "$digits hexidecimal digits"); |
| - } |
| - |
| - // 62 |
| - bool c_ns_escChar() => context('escape sequence', () => transaction(() { |
| - if (!truth(c_escape())) return false; |
| - return truth(or([ |
| - ns_escNull, ns_escBell, ns_escBackspace, ns_escHorizontalTab, |
| - ns_escLineFeed, ns_escVerticalTab, ns_escFormFeed, ns_escCarriageReturn, |
| - ns_escEscape, ns_escSpace, ns_escDoubleQuote, ns_escSlash, |
| - ns_escBackslash, ns_escNextLine, ns_escNonBreakingSpace, |
| - ns_escLineSeparator, ns_escParagraphSeparator, ns_esc8Bit, ns_esc16Bit, |
| - ns_esc32Bit |
| - ])); |
| - })); |
| - |
| - // 63 |
| - bool s_indent(int indent) { |
| - var result = nAtOnce(indent, (c, i) => c == SP); |
| - if (peek() == TAB) { |
| - annotateError("tab characters are not allowed as indentation in YAML", |
| - () => zeroOrMore(() => consume(isSpace))); |
| + var anchor; |
| + var tagToken; |
| + var span = token.span.start.pointSpan(); |
| + if (token is AnchorToken) { |
|
Bob Nystrom
2014/10/31 20:03:27
Lot of copy/paste here. How about local functions
nweiz
2014/11/04 22:19:36
Done.
|
| + anchor = token.name; |
| + span = token.span; |
| + _scanner.scan(); |
| + token = _scanner.peek(); |
| + |
| + if (token is TagToken) { |
| + tagToken = token; |
| + span = span.expand(token.span); |
| + _scanner.scan(); |
| + token = _scanner.peek(); |
| + } |
| + } else if (token is TagToken) { |
| + tagToken = token; |
| + span = span.expand(token.span); |
| + _scanner.scan(); |
| + token = _scanner.peek(); |
| + |
| + if (token is AnchorToken) { |
| + anchor = token.name; |
| + span = token.span; |
| + _scanner.scan(); |
| + token = _scanner.peek(); |
| + } |
| } |
| - return result; |
| - } |
| - // 64 |
| - bool s_indentLessThan(int indent) { |
| - for (int i = 0; i < indent - 1; i++) { |
| - if (!consumeChar(SP)) { |
| - if (peek() == TAB) { |
| - annotateError("tab characters are not allowed as indentation in YAML", |
| - () { |
| - for (; i < indent - 1; i++) { |
| - if (!consume(isSpace)) break; |
| - } |
| - }); |
| + var tag; |
| + if (tagToken != null) { |
| + if (tagToken.handle == null) { |
| + tag = tagToken.suffix; |
| + } else { |
| + var tagDirective = _tagDirectives[tagToken.handle]; |
| + if (tagDirective == null) { |
| + throw new YamlException("Undefined tag handle.", tagToken.span); |
| } |
| - break; |
| + |
| + tag = tagDirective.prefix + tagToken.suffix; |
| } |
| } |
| - return true; |
| - } |
| - // 65 |
| - bool s_indentLessThanOrEqualTo(int indent) => s_indentLessThan(indent + 1); |
| - |
| - // 66 |
| - bool s_separateInLine() => transaction(() { |
| - return captureAs('', () => |
| - truth(oneOrMore(() => consume(isSpace))) || atStartOfLine); |
| - }); |
| - |
| - // 67 |
| - bool s_linePrefix(int indent, int ctx) => captureAs("", () { |
| - switch (ctx) { |
| - case BLOCK_OUT: |
| - case BLOCK_IN: |
| - return s_blockLinePrefix(indent); |
| - case FLOW_OUT: |
| - case FLOW_IN: |
| - return s_flowLinePrefix(indent); |
| + if (indentlessSequence && token.type == TokenType.BLOCK_ENTRY) { |
| + _state = _State.INDENTLESS_SEQUENCE_ENTRY; |
| + return new SequenceStartEvent( |
| + span.expand(token.span), CollectionStyle.BLOCK, |
| + anchor: anchor, tag: tag); |
| } |
| - }); |
| - |
| - // 68 |
| - bool s_blockLinePrefix(int indent) => s_indent(indent); |
| - |
| - // 69 |
| - bool s_flowLinePrefix(int indent) => captureAs('', () { |
| - if (!truth(s_indent(indent))) return false; |
| - zeroOrOne(s_separateInLine); |
| - return true; |
| - }); |
| - |
| - // 70 |
| - bool l_empty(int indent, int ctx) => transaction(() { |
| - var start = or([ |
| - () => s_linePrefix(indent, ctx), |
| - () => s_indentLessThan(indent) |
| - ]); |
| - if (!truth(start)) return false; |
| - return b_asLineFeed(); |
| - }); |
| - |
| - // 71 |
| - bool b_asSpace() => captureAs(" ", () => consume(isBreak)); |
| - |
| - // 72 |
| - bool b_l_trimmed(int indent, int ctx) => transaction(() { |
| - if (!truth(b_nonContent())) return false; |
| - return truth(oneOrMore(() => captureAs("\n", () => l_empty(indent, ctx)))); |
| - }); |
| - |
| - // 73 |
| - bool b_l_folded(int indent, int ctx) => |
| - or([() => b_l_trimmed(indent, ctx), b_asSpace]); |
| - |
| - // 74 |
| - bool s_flowFolded(int indent) => transaction(() { |
| - zeroOrOne(s_separateInLine); |
| - if (!truth(b_l_folded(indent, FLOW_IN))) return false; |
| - return s_flowLinePrefix(indent); |
| - }); |
| - |
| - // 75 |
| - bool c_nb_commentText() { |
| - if (!truth(c_indicator(C_COMMENT))) return false; |
| - zeroOrMore(() => consume(isNonBreak)); |
| - return true; |
| - } |
| - // 76 |
| - bool b_comment() => _scanner.isDone || b_nonContent(); |
| + if (token is ScalarToken) { |
| + // All non-plain scalars have the "!" tag by default. |
| + if (tag == null && token.style != ScalarStyle.PLAIN) tag = "!"; |
| - // 77 |
| - bool s_b_comment() { |
| - if (truth(s_separateInLine())) { |
| - zeroOrOne(c_nb_commentText); |
| + _state = _states.removeLast(); |
| + _scanner.scan(); |
| + return new ScalarEvent( |
| + span.expand(token.span), token.value, token.style, |
| + anchor: anchor, tag: tag); |
| } |
| - return b_comment(); |
| - } |
| - // 78 |
| - bool l_comment() => transaction(() { |
| - if (!truth(s_separateInLine())) return false; |
| - zeroOrOne(c_nb_commentText); |
| - return b_comment(); |
| - }); |
| - |
| - // 79 |
| - bool s_l_comments() { |
| - if (!truth(s_b_comment()) && !atStartOfLine) return false; |
| - zeroOrMore(l_comment); |
| - return true; |
| - } |
| - |
| - // 80 |
| - bool s_separate(int indent, int ctx) { |
| - switch (ctx) { |
| - case BLOCK_OUT: |
| - case BLOCK_IN: |
| - case FLOW_OUT: |
| - case FLOW_IN: |
| - return s_separateLines(indent); |
| - case BLOCK_KEY: |
| - case FLOW_KEY: |
| - return s_separateInLine(); |
| - default: throw 'Invalid context "$ctx".'; |
| + if (token.type == TokenType.FLOW_SEQUENCE_START) { |
| + _state = _State.FLOW_SEQUENCE_FIRST_ENTRY; |
| + return new SequenceStartEvent( |
| + span.expand(token.span), CollectionStyle.FLOW, |
| + anchor: anchor, tag: tag); |
| } |
| - } |
| - // 81 |
| - bool s_separateLines(int indent) { |
| - return transaction(() => s_l_comments() && s_flowLinePrefix(indent)) || |
| - s_separateInLine(); |
| - } |
| - |
| - // 82 |
| - bool l_directive() => false; // TODO(nweiz): implement |
| - |
| - // 96 |
| - Pair<Tag, String> c_ns_properties(int indent, int ctx) { |
| - var tag, anchor; |
| - tag = c_ns_tagProperty(); |
| - if (truth(tag)) { |
| - anchor = transaction(() { |
| - if (!truth(s_separate(indent, ctx))) return null; |
| - return c_ns_anchorProperty(); |
| - }); |
| - return new Pair<Tag, String>(tag, anchor); |
| + if (token.type == TokenType.FLOW_MAPPING_START) { |
| + _state = _State.FLOW_MAPPING_FIRST_KEY; |
| + return new MappingStartEvent( |
| + span.expand(token.span), CollectionStyle.FLOW, |
| + anchor: anchor, tag: tag); |
| } |
| - anchor = c_ns_anchorProperty(); |
| - if (truth(anchor)) { |
| - tag = transaction(() { |
| - if (!truth(s_separate(indent, ctx))) return null; |
| - return c_ns_tagProperty(); |
| - }); |
| - return new Pair<Tag, String>(tag, anchor); |
| + if (block && token.type == TokenType.BLOCK_SEQUENCE_START) { |
| + _state = _State.BLOCK_SEQUENCE_FIRST_ENTRY; |
| + return new SequenceStartEvent( |
| + span.expand(token.span), CollectionStyle.BLOCK, |
| + anchor: anchor, tag: tag); |
| } |
| - return null; |
| - } |
| - |
| - // 97 |
| - Tag c_ns_tagProperty() => null; // TODO(nweiz): implement |
| - |
| - // 101 |
| - String c_ns_anchorProperty() => null; // TODO(nweiz): implement |
| - // 102 |
| - bool isAnchorChar(int char) => isNonSpace(char) && !isFlowIndicator(char); |
| + if (block && token.type == TokenType.BLOCK_MAPPING_START) { |
| + _state = _State.BLOCK_MAPPING_FIRST_KEY; |
| + return new MappingStartEvent( |
| + span.expand(token.span), CollectionStyle.BLOCK, |
| + anchor: anchor, tag: tag); |
| + } |
| - // 103 |
| - String ns_anchorName() => |
| - captureString(() => oneOrMore(() => consume(isAnchorChar))); |
| + if (anchor != null || tag != null) { |
| + _state = _states.removeLast(); |
| + return new ScalarEvent( |
| + span, '', ScalarStyle.PLAIN, |
| + anchor: anchor, tag: tag); |
| + } |
| - // 104 |
| - Node c_ns_aliasNode() { |
| - var start = _scanner.state; |
| - if (!truth(c_indicator(C_ALIAS))) return null; |
| - var name = expect(ns_anchorName(), 'anchor name'); |
| - return new AliasNode(name, _scanner.spanFrom(start)); |
| + throw new YamlException("Expected node content.", span); |
| } |
| - // 105 |
| - ScalarNode e_scalar() => new ScalarNode("?", _scanner.emptySpan, content: ""); |
| - |
| - // 106 |
| - ScalarNode e_node() => e_scalar(); |
| - |
| - // 107 |
| - bool nb_doubleChar() => or([ |
| - c_ns_escChar, |
| - () => consume((c) => isJson(c) && c != BACKSLASH && c != DOUBLE_QUOTE) |
| - ]); |
| - |
| - // 108 |
| - bool ns_doubleChar() => !isSpace(peek()) && truth(nb_doubleChar()); |
| - |
| - // 109 |
| - Node c_doubleQuoted(int indent, int ctx) => context('string', () { |
| - return transaction(() { |
| - var start = _scanner.state; |
| - if (!truth(c_indicator(C_DOUBLE_QUOTE))) return null; |
| - var contents = nb_doubleText(indent, ctx); |
| - if (!truth(c_indicator(C_DOUBLE_QUOTE))) return null; |
| - return new ScalarNode("!", _scanner.spanFrom(start), content: contents); |
| - }); |
| - }); |
| - |
| - // 110 |
| - String nb_doubleText(int indent, int ctx) => captureString(() { |
| - switch (ctx) { |
| - case FLOW_OUT: |
| - case FLOW_IN: |
| - nb_doubleMultiLine(indent); |
| - break; |
| - case BLOCK_KEY: |
| - case FLOW_KEY: |
| - nb_doubleOneLine(); |
| - break; |
| + /// Parses the productions: |
| + /// |
| + /// block_sequence ::= |
| + /// BLOCK-SEQUENCE-START (BLOCK-ENTRY block_node?)* BLOCK-END |
| + /// ******************** *********** * ********* |
| + Event _parseBlockSequenceEntry({bool first: false}) { |
| + if (first) _scanner.scan(); |
| + var token = _scanner.peek(); |
| + |
| + if (token.type == TokenType.BLOCK_ENTRY) { |
| + _scanner.scan(); |
| + token = _scanner.peek(); |
|
Bob Nystrom
2014/10/31 20:03:27
This two line pattern is really common. I assume .
nweiz
2014/11/04 22:19:36
Done.
|
| + |
| + if (token.type == TokenType.BLOCK_ENTRY || |
| + token.type == TokenType.BLOCK_END) { |
| + _state = _State.BLOCK_SEQUENCE_ENTRY; |
| + return _processEmptyScalar(token.span.end); |
| + } else { |
| + _states.add(_State.BLOCK_SEQUENCE_ENTRY); |
| + return _parseNode(block: true); |
| + } |
| } |
| - return true; |
| - }); |
| - // 111 |
| - void nb_doubleOneLine() { |
| - zeroOrMore(nb_doubleChar); |
| - } |
| + if (token.type == TokenType.BLOCK_END) { |
| + _scanner.scan(); |
| + _state = _states.removeLast(); |
| + return new Event(EventType.SEQUENCE_END, token.span); |
| + } |
| - // 112 |
| - bool s_doubleEscaped(int indent) => transaction(() { |
| - zeroOrMore(() => consume(isSpace)); |
| - if (!captureAs("", () => consumeChar(BACKSLASH))) return false; |
| - if (!truth(b_nonContent())) return false; |
| - zeroOrMore(() => captureAs("\n", () => l_empty(indent, FLOW_IN))); |
| - return s_flowLinePrefix(indent); |
| - }); |
| - |
| - // 113 |
| - bool s_doubleBreak(int indent) => or([ |
| - () => s_doubleEscaped(indent), |
| - () => s_flowFolded(indent) |
| - ]); |
| - |
| - // 114 |
| - void nb_ns_doubleInLine() { |
| - zeroOrMore(() => transaction(() { |
| - zeroOrMore(() => consume(isSpace)); |
| - return ns_doubleChar(); |
| - })); |
| + throw new YamlException("While parsing a block collection, expected '-'.", |
| + token.span.start.pointSpan()); |
| } |
| - // 115 |
| - bool s_doubleNextLine(int indent) { |
| - if (!truth(s_doubleBreak(indent))) return false; |
| - zeroOrOne(() { |
| - if (!truth(ns_doubleChar())) return; |
| - nb_ns_doubleInLine(); |
| - or([ |
| - () => s_doubleNextLine(indent), |
| - () => zeroOrMore(() => consume(isSpace)) |
| - ]); |
| - }); |
| - return true; |
| - } |
| + /// Parses the productions: |
| + /// |
| + /// indentless_sequence ::= (BLOCK-ENTRY block_node?)+ |
| + /// *********** * |
| + Event _parseIndentlessSequenceEntry() { |
| + var token = _scanner.peek(); |
| + |
| + if (token.type != TokenType.BLOCK_ENTRY) { |
| + _state = _states.removeLast(); |
| + return new Event(EventType.SEQUENCE_END, token.span.start.pointSpan()); |
| + } |
| - // 116 |
| - void nb_doubleMultiLine(int indent) { |
| - nb_ns_doubleInLine(); |
| - or([ |
| - () => s_doubleNextLine(indent), |
| - () => zeroOrMore(() => consume(isSpace)) |
| - ]); |
| + var start = token.span.start; |
| + _scanner.scan(); |
| + token = _scanner.peek(); |
| + |
| + if (token.type == TokenType.BLOCK_ENTRY || |
| + token.type == TokenType.KEY || |
| + token.type == TokenType.VALUE || |
| + token.type == TokenType.BLOCK_END) { |
| + _state = _State.INDENTLESS_SEQUENCE_ENTRY; |
| + return _processEmptyScalar(start); |
| + } else { |
| + _states.add(_State.INDENTLESS_SEQUENCE_ENTRY); |
| + return _parseNode(block: true); |
| + } |
| } |
| - // 117 |
| - bool c_quotedQuote() => captureAs("'", () => rawString("''")); |
| - |
| - // 118 |
| - bool nb_singleChar() => or([ |
| - c_quotedQuote, |
| - () => consume((c) => isJson(c) && c != SINGLE_QUOTE) |
| - ]); |
| - |
| - // 119 |
| - bool ns_singleChar() => !isSpace(peek()) && truth(nb_singleChar()); |
| - |
| - // 120 |
| - Node c_singleQuoted(int indent, int ctx) => context('string', () { |
| - return transaction(() { |
| - var start = _scanner.state; |
| - if (!truth(c_indicator(C_SINGLE_QUOTE))) return null; |
| - var contents = nb_singleText(indent, ctx); |
| - if (!truth(c_indicator(C_SINGLE_QUOTE))) return null; |
| - return new ScalarNode("!", _scanner.spanFrom(start), content: contents); |
| - }); |
| - }); |
| - |
| - // 121 |
| - String nb_singleText(int indent, int ctx) => captureString(() { |
| - switch (ctx) { |
| - case FLOW_OUT: |
| - case FLOW_IN: |
| - nb_singleMultiLine(indent); |
| - break; |
| - case BLOCK_KEY: |
| - case FLOW_KEY: |
| - nb_singleOneLine(indent); |
| - break; |
| + /// Parses the productions: |
| + /// |
| + /// block_mapping ::= BLOCK-MAPPING_START |
| + /// ******************* |
| + /// ((KEY block_node_or_indentless_sequence?)? |
| + /// *** * |
| + /// (VALUE block_node_or_indentless_sequence?)?)* |
| + /// |
| + /// BLOCK-END |
| + /// ********* |
| + Event _parseBlockMappingKey({bool first: false}) { |
| + if (first) _scanner.scan(); |
| + |
| + var token = _scanner.peek(); |
| + if (token.type == TokenType.KEY) { |
| + var start = token.span.start; |
| + _scanner.scan(); |
| + token = _scanner.peek(); |
| + |
| + if (token.type == TokenType.KEY || |
| + token.type == TokenType.VALUE || |
| + token.type == TokenType.BLOCK_END) { |
| + _state = _State.BLOCK_MAPPING_VALUE; |
| + return _processEmptyScalar(start); |
| + } else { |
| + _states.add(_State.BLOCK_MAPPING_VALUE); |
| + return _parseNode(block: true, indentlessSequence: true); |
| + } |
| } |
| - return true; |
| - }); |
| - // 122 |
| - void nb_singleOneLine(int indent) { |
| - zeroOrMore(nb_singleChar); |
| - } |
| + // libyaml doesn't allow empty keys without an explicit key indicator, but |
| + // the spec does. See example 8.18: |
| + // http://yaml.org/spec/1.2/spec.html#id2798896. |
| + if (token.type == TokenType.VALUE) { |
| + _state = _State.BLOCK_MAPPING_VALUE; |
| + return _processEmptyScalar(token.span.start); |
| + } |
| - // 123 |
| - void nb_ns_singleInLine() { |
| - zeroOrMore(() => transaction(() { |
| - zeroOrMore(() => consume(isSpace)); |
| - return ns_singleChar(); |
| - })); |
| - } |
| + if (token.type == TokenType.BLOCK_END) { |
| + _scanner.scan(); |
| + _state = _states.removeLast(); |
| + return new Event(EventType.MAPPING_END, token.span); |
| + } |
| - // 124 |
| - bool s_singleNextLine(int indent) { |
| - if (!truth(s_flowFolded(indent))) return false; |
| - zeroOrOne(() { |
| - if (!truth(ns_singleChar())) return; |
| - nb_ns_singleInLine(); |
| - or([ |
| - () => s_singleNextLine(indent), |
| - () => zeroOrMore(() => consume(isSpace)) |
| - ]); |
| - }); |
| - return true; |
| + throw new YamlException("Expected a key while parsing a block mapping.", |
| + token.span.start.pointSpan()); |
| } |
| - // 125 |
| - void nb_singleMultiLine(int indent) { |
| - nb_ns_singleInLine(); |
| - or([ |
| - () => s_singleNextLine(indent), |
| - () => zeroOrMore(() => consume(isSpace)) |
| - ]); |
| - } |
| + /// Parses the productions: |
| + /// |
| + /// block_mapping ::= BLOCK-MAPPING_START |
| + /// |
| + /// ((KEY block_node_or_indentless_sequence?)? |
| + /// |
| + /// (VALUE block_node_or_indentless_sequence?)?)* |
| + /// ***** * |
| + /// BLOCK-END |
| + /// |
| + Event _parseBlockMappingValue() { |
| + var token = _scanner.peek(); |
| - // 126 |
| - bool ns_plainFirst(int ctx) { |
| - var char = peek(); |
| - var indicator = indicatorType(char); |
| - if (indicator == C_RESERVED) { |
| - error("Reserved indicators can't start a plain scalar"); |
| + if (token.type != TokenType.VALUE) { |
| + _state = _State.BLOCK_MAPPING_KEY; |
| + return _processEmptyScalar(token.span.start); |
| } |
| - var match = (isNonSpace(char) && indicator == null) || |
| - ((indicator == C_MAPPING_KEY || |
| - indicator == C_MAPPING_VALUE || |
| - indicator == C_SEQUENCE_ENTRY) && |
| - isPlainSafe(ctx, peek(1))); |
| - |
| - if (match) next(); |
| - return match; |
| - } |
| - // 127 |
| - bool isPlainSafe(int ctx, int char) { |
| - switch (ctx) { |
| - case FLOW_OUT: |
| - case BLOCK_KEY: |
| - // 128 |
| - return isNonSpace(char); |
| - case FLOW_IN: |
| - case FLOW_KEY: |
| - // 129 |
| - return isNonSpace(char) && !isFlowIndicator(char); |
| - default: throw 'Invalid context "$ctx".'; |
| + var start = token.span.start; |
| + _scanner.scan(); |
| + token = _scanner.peek(); |
| + if (token.type == TokenType.KEY || |
| + token.type == TokenType.VALUE || |
| + token.type == TokenType.BLOCK_END) { |
| + _state = _State.BLOCK_MAPPING_KEY; |
| + return _processEmptyScalar(start); |
| + } else { |
| + _states.add(_State.BLOCK_MAPPING_KEY); |
| + return _parseNode(block: true, indentlessSequence: true); |
| } |
| } |
| - // 130 |
| - bool ns_plainChar(int ctx) { |
| - var char = peek(); |
| - var indicator = indicatorType(char); |
| - var safeChar = isPlainSafe(ctx, char) && indicator != C_MAPPING_VALUE && |
| - indicator != C_COMMENT; |
| - var nonCommentHash = isNonSpace(peek(-1)) && indicator == C_COMMENT; |
| - var nonMappingColon = indicator == C_MAPPING_VALUE && |
| - isPlainSafe(ctx, peek(1)); |
| - var match = safeChar || nonCommentHash || nonMappingColon; |
| - |
| - if (match) next(); |
| - return match; |
| - } |
| + /// Parses the productions: |
| + /// |
| + /// flow_sequence ::= FLOW-SEQUENCE-START |
| + /// ******************* |
| + /// (flow_sequence_entry FLOW-ENTRY)* |
| + /// * ********** |
| + /// flow_sequence_entry? |
| + /// * |
| + /// FLOW-SEQUENCE-END |
| + /// ***************** |
| + /// flow_sequence_entry ::= |
| + /// flow_node | KEY flow_node? (VALUE flow_node?)? |
| + /// * |
| + Event _parseFlowSequenceEntry({bool first: false}) { |
| + if (first) _scanner.scan(); |
| + var token = _scanner.peek(); |
| + |
|
Bob Nystrom
2014/10/31 20:03:27
Extra blank line.
nweiz
2014/11/04 22:19:36
Done.
|
| + |
| + if (token.type != TokenType.FLOW_SEQUENCE_END) { |
| + if (!first) { |
| + if (token.type != TokenType.FLOW_ENTRY) { |
| + throw new YamlException( |
| + "While parsing a flow sequence, expected ',' or ']'.", |
| + token.span.start.pointSpan()); |
| + } |
| - // 131 |
| - String ns_plain(int indent, int ctx) => context('plain scalar', () { |
| - return captureString(() { |
| - switch (ctx) { |
| - case FLOW_OUT: |
| - case FLOW_IN: |
| - return ns_plainMultiLine(indent, ctx); |
| - case BLOCK_KEY: |
| - case FLOW_KEY: |
| - return ns_plainOneLine(ctx); |
| - default: throw 'Invalid context "$ctx".'; |
| + _scanner.scan(); |
| + token = _scanner.peek(); |
| } |
| - }); |
| - }); |
| - |
| - // 132 |
| - void nb_ns_plainInLine(int ctx) { |
| - zeroOrMore(() => transaction(() { |
| - zeroOrMore(() => consume(isSpace)); |
| - return ns_plainChar(ctx); |
| - })); |
| - } |
| - // 133 |
| - bool ns_plainOneLine(int ctx) { |
| - if (truth(c_forbidden())) return false; |
| - if (!truth(ns_plainFirst(ctx))) return false; |
| - nb_ns_plainInLine(ctx); |
| - return true; |
| - } |
| + if (token.type == TokenType.KEY) { |
| + _state = _State.FLOW_SEQUENCE_ENTRY_MAPPING_KEY; |
| + _scanner.scan(); |
| + return new MappingStartEvent( |
| + token.span, CollectionStyle.FLOW); |
| + } else if (token.type != TokenType.FLOW_SEQUENCE_END) { |
| + _states.add(_State.FLOW_SEQUENCE_ENTRY); |
| + return _parseNode(); |
| + } |
| + } |
| - // 134 |
| - bool s_ns_plainNextLine(int indent, int ctx) => transaction(() { |
| - if (!truth(s_flowFolded(indent))) return false; |
| - if (truth(c_forbidden())) return false; |
| - if (!truth(ns_plainChar(ctx))) return false; |
| - nb_ns_plainInLine(ctx); |
| - return true; |
| - }); |
| - |
| - // 135 |
| - bool ns_plainMultiLine(int indent, int ctx) { |
| - if (!truth(ns_plainOneLine(ctx))) return false; |
| - zeroOrMore(() => s_ns_plainNextLine(indent, ctx)); |
| - return true; |
| - } |
| + _scanner.scan(); |
| + _state = _states.removeLast(); |
| + return new Event(EventType.SEQUENCE_END, token.span); |
| + } |
| - // 136 |
| - int inFlow(int ctx) { |
| - switch (ctx) { |
| - case FLOW_OUT: |
| - case FLOW_IN: |
| - return FLOW_IN; |
| - case BLOCK_KEY: |
| - case FLOW_KEY: |
| - return FLOW_KEY; |
| + /// Parses the productions: |
| + /// |
| + /// flow_sequence_entry ::= |
| + /// flow_node | KEY flow_node? (VALUE flow_node?)? |
| + /// *** * |
| + Event _parseFlowSequenceEntryMappingKey() { |
| + var token = _scanner.peek(); |
| + |
| + if (token.type == TokenType.VALUE || |
| + token.type == TokenType.FLOW_ENTRY || |
| + token.type == TokenType.FLOW_SEQUENCE_END) { |
| + // libyaml consumes the token here, but that seems like a bug, since it |
| + // always causes [_parseFlowSequenceEntryMappingValue] to emit an empty |
| + // scalar. |
| + |
| + var start = token.span.start; |
| + _state = _State.FLOW_SEQUENCE_ENTRY_MAPPING_VALUE; |
| + return _processEmptyScalar(start); |
| + } else { |
| + _states.add(_State.FLOW_SEQUENCE_ENTRY_MAPPING_VALUE); |
| + return _parseNode(); |
| } |
| - throw "unreachable"; |
| } |
| - // 137 |
| - SequenceNode c_flowSequence(int indent, int ctx) => transaction(() { |
| - var start = _scanner.state; |
| - if (!truth(c_indicator(C_SEQUENCE_START))) return null; |
| - zeroOrOne(() => s_separate(indent, ctx)); |
| - var content = zeroOrOne(() => ns_s_flowSeqEntries(indent, inFlow(ctx))); |
| - if (!truth(c_indicator(C_SEQUENCE_END))) return null; |
| - return new SequenceNode("?", new List<Node>.from(content), |
| - _scanner.spanFrom(start)); |
| - }); |
| - |
| - // 138 |
| - Iterable<Node> ns_s_flowSeqEntries(int indent, int ctx) { |
| - var first = ns_flowSeqEntry(indent, ctx); |
| - if (!truth(first)) return new Queue<Node>(); |
| - zeroOrOne(() => s_separate(indent, ctx)); |
| - |
| - var rest; |
| - if (truth(c_indicator(C_COLLECT_ENTRY))) { |
| - zeroOrOne(() => s_separate(indent, ctx)); |
| - rest = zeroOrOne(() => ns_s_flowSeqEntries(indent, ctx)); |
| + /// Parses the productions: |
| + /// |
| + /// flow_sequence_entry ::= |
| + /// flow_node | KEY flow_node? (VALUE flow_node?)? |
| + /// ***** * |
| + Event _parseFlowSequenceEntryMappingValue() { |
| + var token = _scanner.peek(); |
| + |
| + if (token.type == TokenType.VALUE) { |
| + _scanner.scan(); |
| + token = _scanner.peek(); |
| + if (token.type != TokenType.FLOW_ENTRY && |
| + token.type != TokenType.FLOW_SEQUENCE_END) { |
| + _states.add(_State.FLOW_SEQUENCE_ENTRY_MAPPING_END); |
| + return _parseNode(); |
| + } |
| } |
| - if (rest == null) rest = new Queue<Node>(); |
| - rest.addFirst(first); |
| - |
| - return rest; |
| + _state = _State.FLOW_SEQUENCE_ENTRY_MAPPING_END; |
| + return _processEmptyScalar(token.span.start); |
| } |
| - // 139 |
| - Node ns_flowSeqEntry(int indent, int ctx) => or([ |
| - () => ns_flowPair(indent, ctx), |
| - () => ns_flowNode(indent, ctx) |
| - ]); |
| - |
| - // 140 |
| - Node c_flowMapping(int indent, int ctx) { |
| - var start = _scanner.state; |
| - if (!truth(c_indicator(C_MAPPING_START))) return null; |
| - zeroOrOne(() => s_separate(indent, ctx)); |
| - var content = zeroOrOne(() => ns_s_flowMapEntries(indent, inFlow(ctx))); |
| - if (!truth(c_indicator(C_MAPPING_END))) return null; |
| - return new MappingNode("?", content, _scanner.spanFrom(start)); |
| + /// Parses the productions: |
| + /// |
| + /// flow_sequence_entry ::= |
| + /// flow_node | KEY flow_node? (VALUE flow_node?)? |
| + /// * |
| + Event _parseFlowSequenceEntryMappingEnd() { |
| + _state = _State.FLOW_SEQUENCE_ENTRY; |
| + return new Event(EventType.MAPPING_END, |
| + _scanner.peek().span.start.pointSpan()); |
| } |
| - // 141 |
| - Map ns_s_flowMapEntries(int indent, int ctx) { |
| - var first = ns_flowMapEntry(indent, ctx); |
| - if (!truth(first)) return deepEqualsMap(); |
| - zeroOrOne(() => s_separate(indent, ctx)); |
| + /// Parses the productions: |
| + /// |
| + /// flow_mapping ::= FLOW-MAPPING-START |
| + /// ****************** |
| + /// (flow_mapping_entry FLOW-ENTRY)* |
| + /// * ********** |
| + /// flow_mapping_entry? |
| + /// ****************** |
| + /// FLOW-MAPPING-END |
| + /// **************** |
| + /// flow_mapping_entry ::= |
| + /// flow_node | KEY flow_node? (VALUE flow_node?)? |
| + /// * *** * |
| + Event _parseFlowMappingKey({bool first: false}) { |
| + if (first) _scanner.scan(); |
| + var token = _scanner.peek(); |
| + |
| + if (token.type != TokenType.FLOW_MAPPING_END) { |
| + if (!first) { |
| + if (token.type != TokenType.FLOW_ENTRY) { |
| + throw new YamlException( |
| + "While parsing a flow mapping, expected ',' or '}'.", |
| + token.span.start.pointSpan()); |
| + } |
| + |
| + _scanner.scan(); |
| + token = _scanner.peek(); |
| + } |
| - var rest; |
| - if (truth(c_indicator(C_COLLECT_ENTRY))) { |
| - zeroOrOne(() => s_separate(indent, ctx)); |
| - rest = ns_s_flowMapEntries(indent, ctx); |
| + if (token.type == TokenType.KEY) { |
| + _scanner.scan(); |
| + token = _scanner.peek(); |
| + if (token.type != TokenType.VALUE && |
| + token.type != TokenType.FLOW_ENTRY && |
| + token.type != TokenType.FLOW_MAPPING_END) { |
| + _states.add(_State.FLOW_MAPPING_VALUE); |
| + return _parseNode(); |
| + } else { |
| + _state = _State.FLOW_MAPPING_VALUE; |
| + return _processEmptyScalar(token.span.start); |
| + } |
| + } else if (token.type != TokenType.FLOW_MAPPING_END) { |
| + _states.add(_State.FLOW_MAPPING_EMPTY_VALUE); |
| + return _parseNode(); |
| + } |
| } |
| - if (rest == null) rest = deepEqualsMap(); |
| + _scanner.scan(); |
| + _state = _states.removeLast(); |
| + return new Event(EventType.MAPPING_END, token.span); |
| + } |
| - // TODO(nweiz): Duplicate keys should be an error. This includes keys with |
| - // different representations but the same value (e.g. 10 vs 0xa). To make |
| - // this user-friendly we'll probably also want to associate nodes with a |
| - // source range. |
| - if (!rest.containsKey(first.first)) rest[first.first] = first.last; |
| + /// Parses the productions: |
| + /// |
| + /// flow_mapping_entry ::= |
| + /// flow_node | KEY flow_node? (VALUE flow_node?)? |
| + /// * ***** * |
| + Event _parseFlowMappingValue({bool empty: false}) { |
| + var token = _scanner.peek(); |
| + |
| + if (empty) { |
| + _state = _State.FLOW_MAPPING_KEY; |
| + return _processEmptyScalar(token.span.start); |
| + } |
| - return rest; |
| - } |
| + if (token.type == TokenType.VALUE) { |
| + _scanner.scan(); |
| + token = _scanner.peek(); |
| + if (token.type != TokenType.FLOW_ENTRY && |
| + token.type != TokenType.FLOW_MAPPING_END) { |
| + _states.add(_State.FLOW_MAPPING_KEY); |
| + return _parseNode(); |
| + } |
| + } |
| - // 142 |
| - Pair<Node, Node> ns_flowMapEntry(int indent, int ctx) => or([ |
| - () => transaction(() { |
| - if (!truth(c_indicator(C_MAPPING_KEY))) return false; |
| - if (!truth(s_separate(indent, ctx))) return false; |
| - return ns_flowMapExplicitEntry(indent, ctx); |
| - }), |
| - () => ns_flowMapImplicitEntry(indent, ctx) |
| - ]); |
| - |
| - // 143 |
| - Pair<Node, Node> ns_flowMapExplicitEntry(int indent, int ctx) => or([ |
| - () => ns_flowMapImplicitEntry(indent, ctx), |
| - () => new Pair<Node, Node>(e_node(), e_node()) |
| - ]); |
| - |
| - // 144 |
| - Pair<Node, Node> ns_flowMapImplicitEntry(int indent, int ctx) => or([ |
| - () => ns_flowMapYamlKeyEntry(indent, ctx), |
| - () => c_ns_flowMapEmptyKeyEntry(indent, ctx), |
| - () => c_ns_flowMapJsonKeyEntry(indent, ctx) |
| - ]); |
| - |
| - // 145 |
| - Pair<Node, Node> ns_flowMapYamlKeyEntry(int indent, int ctx) { |
| - var key = ns_flowYamlNode(indent, ctx); |
| - if (!truth(key)) return null; |
| - var value = or([ |
| - () => transaction(() { |
| - zeroOrOne(() => s_separate(indent, ctx)); |
| - return c_ns_flowMapSeparateValue(indent, ctx); |
| - }), |
| - e_node |
| - ]); |
| - return new Pair<Node, Node>(key, value); |
| + _state = _State.FLOW_MAPPING_KEY; |
| + return _processEmptyScalar(token.span.start); |
| } |
| - // 146 |
| - Pair<Node, Node> c_ns_flowMapEmptyKeyEntry(int indent, int ctx) { |
| - var value = c_ns_flowMapSeparateValue(indent, ctx); |
| - if (!truth(value)) return null; |
| - return new Pair<Node, Node>(e_node(), value); |
| - } |
| + /// Generate an empty scalar event. |
| + Event _processEmptyScalar(SourceLocation location) => |
| + new ScalarEvent(location.pointSpan(), '', ScalarStyle.PLAIN); |
| - // 147 |
| - Node c_ns_flowMapSeparateValue(int indent, int ctx) => transaction(() { |
| - if (!truth(c_indicator(C_MAPPING_VALUE))) return null; |
| - if (isPlainSafe(ctx, peek())) return null; |
| - |
| - return or([ |
| - () => transaction(() { |
| - if (!s_separate(indent, ctx)) return null; |
| - return ns_flowNode(indent, ctx); |
| - }), |
| - e_node |
| - ]); |
| - }); |
| - |
| - // 148 |
| - Pair<Node, Node> c_ns_flowMapJsonKeyEntry(int indent, int ctx) { |
| - var key = c_flowJsonNode(indent, ctx); |
| - if (!truth(key)) return null; |
| - var value = or([ |
| - () => transaction(() { |
| - zeroOrOne(() => s_separate(indent, ctx)); |
| - return c_ns_flowMapAdjacentValue(indent, ctx); |
| - }), |
| - e_node |
| - ]); |
| - return new Pair<Node, Node>(key, value); |
| - } |
| + /// Parses directives. |
| + Pair<VersionDirective, List<TagDirective>> _processDirectives() { |
| + var token = _scanner.peek(); |
| - // 149 |
| - Node c_ns_flowMapAdjacentValue(int indent, int ctx) { |
| - if (!truth(c_indicator(C_MAPPING_VALUE))) return null; |
| - return or([ |
| - () => transaction(() { |
| - zeroOrOne(() => s_separate(indent, ctx)); |
| - return ns_flowNode(indent, ctx); |
| - }), |
| - e_node |
| - ]); |
| - } |
| + var versionDirective; |
| + var tagDirectives = []; |
| + var reservedDirectives = []; |
| + while (token.type == TokenType.VERSION_DIRECTIVE || |
| + token.type == TokenType.TAG_DIRECTIVE) { |
| + if (token is VersionDirectiveToken) { |
| + if (versionDirective != null) { |
| + throw new YamlException("Duplicate %YAML directive.", token.span); |
| + } |
| - // 150 |
| - Node ns_flowPair(int indent, int ctx) { |
| - var start = _scanner.state; |
| - var pair = or([ |
| - () => transaction(() { |
| - if (!truth(c_indicator(C_MAPPING_KEY))) return null; |
| - if (!truth(s_separate(indent, ctx))) return null; |
| - return ns_flowMapExplicitEntry(indent, ctx); |
| - }), |
| - () => ns_flowPairEntry(indent, ctx) |
| - ]); |
| - if (!truth(pair)) return null; |
| - |
| - return map([pair], _scanner.spanFrom(start)); |
| - } |
| + if (token.major != 1 || token.minor == 0) { |
| + throw new YamlException( |
| + "Incompatible YAML document. This parser only supports YAML 1.1 " |
| + "and 1.2.", |
| + token.span); |
| + } else if (token.minor > 2) { |
| + // TODO(nweiz): Print to stderr when issue 6943 is fixed and dart:io |
| + // is available. |
| + warn("Warning: this parser only supports YAML 1.1 and 1.2.", |
| + token.span); |
| + } |
| - // 151 |
| - Pair<Node, Node> ns_flowPairEntry(int indent, int ctx) => or([ |
| - () => ns_flowPairYamlKeyEntry(indent, ctx), |
| - () => c_ns_flowMapEmptyKeyEntry(indent, ctx), |
| - () => c_ns_flowPairJsonKeyEntry(indent, ctx) |
| - ]); |
| - |
| - // 152 |
| - Pair<Node, Node> ns_flowPairYamlKeyEntry(int indent, int ctx) => |
| - transaction(() { |
| - var key = ns_s_implicitYamlKey(FLOW_KEY); |
| - if (!truth(key)) return null; |
| - var value = c_ns_flowMapSeparateValue(indent, ctx); |
| - if (!truth(value)) return null; |
| - return new Pair<Node, Node>(key, value); |
| - }); |
| - |
| - // 153 |
| - Pair<Node, Node> c_ns_flowPairJsonKeyEntry(int indent, int ctx) => |
| - transaction(() { |
| - var key = c_s_implicitJsonKey(FLOW_KEY); |
| - if (!truth(key)) return null; |
| - var value = c_ns_flowMapAdjacentValue(indent, ctx); |
| - if (!truth(value)) return null; |
| - return new Pair<Node, Node>(key, value); |
| - }); |
| - |
| - // 154 |
| - Node ns_s_implicitYamlKey(int ctx) => transaction(() { |
| - // TODO(nweiz): this is supposed to be limited to 1024 characters. |
| - |
| - // The indentation parameter is "null" since it's unused in this path |
| - var node = ns_flowYamlNode(null, ctx); |
| - if (!truth(node)) return null; |
| - zeroOrOne(s_separateInLine); |
| - return node; |
| - }); |
| - |
| - // 155 |
| - Node c_s_implicitJsonKey(int ctx) => transaction(() { |
| - // TODO(nweiz): this is supposed to be limited to 1024 characters. |
| - |
| - // The indentation parameter is "null" since it's unused in this path |
| - var node = c_flowJsonNode(null, ctx); |
| - if (!truth(node)) return null; |
| - zeroOrOne(s_separateInLine); |
| - return node; |
| - }); |
| - |
| - // 156 |
| - Node ns_flowYamlContent(int indent, int ctx) { |
| - var start = _scanner.state; |
| - var str = ns_plain(indent, ctx); |
| - if (!truth(str)) return null; |
| - return new ScalarNode("?", _scanner.spanFrom(start), content: str); |
| - } |
| + versionDirective = new VersionDirective(token.major, token.minor); |
| + } else if (token is TagDirectiveToken) { |
| + var tagDirective = new TagDirective(token.handle, token.prefix); |
| + _appendTagDirective(tagDirective, token.span); |
| + tagDirectives.add(tagDirective); |
| + } |
| - // 157 |
| - Node c_flowJsonContent(int indent, int ctx) => or([ |
| - () => c_flowSequence(indent, ctx), |
| - () => c_flowMapping(indent, ctx), |
| - () => c_singleQuoted(indent, ctx), |
| - () => c_doubleQuoted(indent, ctx) |
| - ]); |
| - |
| - // 158 |
| - Node ns_flowContent(int indent, int ctx) => or([ |
| - () => ns_flowYamlContent(indent, ctx), |
| - () => c_flowJsonContent(indent, ctx) |
| - ]); |
| - |
| - // 159 |
| - Node ns_flowYamlNode(int indent, int ctx) => or([ |
| - c_ns_aliasNode, |
| - () => ns_flowYamlContent(indent, ctx), |
| - () { |
| - var props = c_ns_properties(indent, ctx); |
| - if (!truth(props)) return null; |
| - var node = or([ |
| - () => transaction(() { |
| - if (!truth(s_separate(indent, ctx))) return null; |
| - return ns_flowYamlContent(indent, ctx); |
| - }), |
| - e_scalar |
| - ]); |
| - return addProps(node, props); |
| + _scanner.scan(); |
| + token = _scanner.peek(); |
| } |
| - ]); |
| - |
| - // 160 |
| - Node c_flowJsonNode(int indent, int ctx) => transaction(() { |
| - var props; |
| - zeroOrOne(() => transaction(() { |
| - props = c_ns_properties(indent, ctx); |
| - if (!truth(props)) return null; |
| - return s_separate(indent, ctx); |
| - })); |
| - |
| - return addProps(c_flowJsonContent(indent, ctx), props); |
| - }); |
| - |
| - // 161 |
| - Node ns_flowNode(int indent, int ctx) => or([ |
| - c_ns_aliasNode, |
| - () => ns_flowContent(indent, ctx), |
| - () => transaction(() { |
| - var props = c_ns_properties(indent, ctx); |
| - if (!truth(props)) return null; |
| - var node = or([ |
| - () => transaction(() => s_separate(indent, ctx) ? |
| - ns_flowContent(indent, ctx) : null), |
| - e_scalar]); |
| - return addProps(node, props); |
| - }) |
| - ]); |
| - |
| - // 162 |
| - _BlockHeader c_b_blockHeader() => transaction(() { |
| - var indentation = c_indentationIndicator(); |
| - var chomping = c_chompingIndicator(); |
| - if (!truth(indentation)) indentation = c_indentationIndicator(); |
| - if (!truth(s_b_comment())) return null; |
| - |
| - return new _BlockHeader(indentation, chomping); |
| - }); |
| - |
| - // 163 |
| - int c_indentationIndicator() { |
| - if (!isDecDigit(peek())) return null; |
| - return next() - NUMBER_0; |
| - } |
| - |
| - // 164 |
| - int c_chompingIndicator() { |
| - switch (peek()) { |
| - case HYPHEN: |
| - next(); |
| - return CHOMPING_STRIP; |
| - case PLUS: |
| - next(); |
| - return CHOMPING_KEEP; |
| - default: |
| - return CHOMPING_CLIP; |
| + |
| + _appendTagDirective( |
| + new TagDirective("!", "!"), |
| + token.span.start.pointSpan(), |
| + allowDuplicates: true); |
| + _appendTagDirective( |
| + new TagDirective("!!", "tag:yaml.org,2002:"), |
| + token.span.start.pointSpan(), |
| + allowDuplicates: true); |
| + |
| + return new Pair(versionDirective, tagDirectives); |
| + } |
| + |
| + /// Adds a tag directive to the directives stack. |
| + void _appendTagDirective(TagDirective newDirective, FileSpan span, |
| + {bool allowDuplicates: false}) { |
| + if (_tagDirectives.containsKey(newDirective.handle)) { |
| + if (allowDuplicates) return; |
| + throw new YamlException("Duplicate %TAG directive.", span); |
| } |
| - } |
| - // 165 |
| - bool b_chompedLast(int chomping) { |
| - if (_scanner.isDone) return true; |
| - switch (chomping) { |
| - case CHOMPING_STRIP: |
| - return b_nonContent(); |
| - case CHOMPING_CLIP: |
| - case CHOMPING_KEEP: |
| - return b_asLineFeed(); |
| - } |
| - throw "unreachable"; |
| + _tagDirectives[newDirective.handle] = newDirective; |
| } |
| +} |
| - // 166 |
| - void l_chompedEmpty(int indent, int chomping) { |
| - switch (chomping) { |
| - case CHOMPING_STRIP: |
| - case CHOMPING_CLIP: |
| - l_stripEmpty(indent); |
| - break; |
| - case CHOMPING_KEEP: |
| - l_keepEmpty(indent); |
| - break; |
| - } |
| - } |
| +/// The possible states for the parser. |
| +class _State { |
| + /// Expect [TokenType.STREAM_START]. |
| + static const STREAM_START = const _State("STREAM_START"); |
| - // 167 |
| - void l_stripEmpty(int indent) { |
| - captureAs('', () { |
| - zeroOrMore(() => transaction(() { |
| - if (!truth(s_indentLessThanOrEqualTo(indent))) return false; |
| - return b_nonContent(); |
| - })); |
| - zeroOrOne(() => l_trailComments(indent)); |
| - return true; |
| - }); |
| - } |
| + /// Expect the beginning of an implicit document. |
| + static const IMPLICIT_DOCUMENT_START = |
| + const _State("IMPLICIT_DOCUMENT_START"); |
| - // 168 |
| - void l_keepEmpty(int indent) { |
| - zeroOrMore(() => captureAs('\n', () => l_empty(indent, BLOCK_IN))); |
| - zeroOrOne(() => captureAs('', () => l_trailComments(indent))); |
| - } |
| + /// Expect [TokenType.DOCUMENT_START]. |
| + static const DOCUMENT_START = const _State("DOCUMENT_START"); |
| - // 169 |
| - bool l_trailComments(int indent) => transaction(() { |
| - if (!truth(s_indentLessThanOrEqualTo(indent))) return false; |
| - if (!truth(c_nb_commentText())) return false; |
| - if (!truth(b_comment())) return false; |
| - zeroOrMore(l_comment); |
| - return true; |
| - }); |
| - |
| - // 170 |
| - Node c_l_literal(int indent) => transaction(() { |
| - var start = _scanner.state; |
| - if (!truth(c_indicator(C_LITERAL))) return null; |
| - var header = c_b_blockHeader(); |
| - if (!truth(header)) return null; |
| - |
| - var additionalIndent = blockScalarAdditionalIndentation(header, indent); |
| - var content = l_literalContent(indent + additionalIndent, header.chomping); |
| - if (!truth(content)) return null; |
| - |
| - return new ScalarNode("!", _scanner.spanFrom(start), content: content); |
| - }); |
| - |
| - // 171 |
| - bool l_nb_literalText(int indent) => transaction(() { |
| - zeroOrMore(() => captureAs("\n", () => l_empty(indent, BLOCK_IN))); |
| - if (!truth(captureAs("", () => s_indent(indent)))) return false; |
| - return truth(oneOrMore(() => consume(isNonBreak))); |
| - }); |
| - |
| - // 172 |
| - bool b_nb_literalNext(int indent) => transaction(() { |
| - if (!truth(b_asLineFeed())) return false; |
| - return l_nb_literalText(indent); |
| - }); |
| - |
| - // 173 |
| - String l_literalContent(int indent, int chomping) => captureString(() { |
| - transaction(() { |
| - if (!truth(l_nb_literalText(indent))) return false; |
| - zeroOrMore(() => b_nb_literalNext(indent)); |
| - return b_chompedLast(chomping); |
| - }); |
| - l_chompedEmpty(indent, chomping); |
| - return true; |
| - }); |
| - |
| - // 174 |
| - Node c_l_folded(int indent) => transaction(() { |
| - var start = _scanner.state; |
| - if (!truth(c_indicator(C_FOLDED))) return null; |
| - var header = c_b_blockHeader(); |
| - if (!truth(header)) return null; |
| - |
| - var additionalIndent = blockScalarAdditionalIndentation(header, indent); |
| - var content = l_foldedContent(indent + additionalIndent, header.chomping); |
| - if (!truth(content)) return null; |
| - |
| - return new ScalarNode("!", _scanner.spanFrom(start), content: content); |
| - }); |
| - |
| - // 175 |
| - bool s_nb_foldedText(int indent) => transaction(() { |
| - if (!truth(captureAs('', () => s_indent(indent)))) return false; |
| - if (!truth(consume(isNonSpace))) return false; |
| - zeroOrMore(() => consume(isNonBreak)); |
| - return true; |
| - }); |
| - |
| - // 176 |
| - bool l_nb_foldedLines(int indent) { |
| - if (!truth(s_nb_foldedText(indent))) return false; |
| - zeroOrMore(() => transaction(() { |
| - if (!truth(b_l_folded(indent, BLOCK_IN))) return false; |
| - return s_nb_foldedText(indent); |
| - })); |
| - return true; |
| - } |
| + /// Expect the content of a document. |
| + static const DOCUMENT_CONTENT = const _State("DOCUMENT_CONTENT"); |
| - // 177 |
| - bool s_nb_spacedText(int indent) => transaction(() { |
| - if (!truth(captureAs('', () => s_indent(indent)))) return false; |
| - if (!truth(consume(isSpace))) return false; |
| - zeroOrMore(() => consume(isNonBreak)); |
| - return true; |
| - }); |
| - |
| - // 178 |
| - bool b_l_spaced(int indent) { |
| - if (!truth(b_asLineFeed())) return false; |
| - zeroOrMore(() => captureAs("\n", () => l_empty(indent, BLOCK_IN))); |
| - return true; |
| - } |
| + /// Expect [TokenType.DOCUMENT_END]. |
| + static const DOCUMENT_END = const _State("DOCUMENT_END"); |
| - // 179 |
| - bool l_nb_spacedLines(int indent) { |
| - if (!truth(s_nb_spacedText(indent))) return false; |
| - zeroOrMore(() => transaction(() { |
| - if (!truth(b_l_spaced(indent))) return false; |
| - return s_nb_spacedText(indent); |
| - })); |
| - return true; |
| - } |
| + /// Expect a block node. |
| + static const BLOCK_NODE = const _State("BLOCK_NODE"); |
| - // 180 |
| - bool l_nb_sameLines(int indent) => transaction(() { |
| - zeroOrMore(() => captureAs('\n', () => l_empty(indent, BLOCK_IN))); |
| - return or([ |
| - () => l_nb_foldedLines(indent), |
| - () => l_nb_spacedLines(indent) |
| - ]); |
| - }); |
| - |
| - // 181 |
| - bool l_nb_diffLines(int indent) { |
| - if (!truth(l_nb_sameLines(indent))) return false; |
| - zeroOrMore(() => transaction(() { |
| - if (!truth(b_asLineFeed())) return false; |
| - return l_nb_sameLines(indent); |
| - })); |
| - return true; |
| - } |
| + /// Expect a block node or indentless sequence. |
| + static const BLOCK_NODE_OR_INDENTLESS_SEQUENCE = |
| + const _State("BLOCK_NODE_OR_INDENTLESS_SEQUENCE"); |
| - // 182 |
| - String l_foldedContent(int indent, int chomping) => captureString(() { |
| - transaction(() { |
| - if (!truth(l_nb_diffLines(indent))) return false; |
| - return b_chompedLast(chomping); |
| - }); |
| - l_chompedEmpty(indent, chomping); |
| - return true; |
| - }); |
| - |
| - // 183 |
| - SequenceNode l_blockSequence(int indent) => context('sequence', () { |
| - var additionalIndent = countIndentation() - indent; |
| - if (additionalIndent <= 0) return null; |
| - |
| - var start = _scanner.state; |
| - var content = oneOrMore(() => transaction(() { |
| - if (!truth(s_indent(indent + additionalIndent))) return null; |
| - return c_l_blockSeqEntry(indent + additionalIndent); |
| - })); |
| - if (!truth(content)) return null; |
| - |
| - return new SequenceNode("?", content, _scanner.spanFrom(start)); |
| - }); |
| - |
| - // 184 |
| - Node c_l_blockSeqEntry(int indent) => transaction(() { |
| - if (!truth(c_indicator(C_SEQUENCE_ENTRY))) return null; |
| - if (isNonSpace(peek())) return null; |
| - |
| - return s_l_blockIndented(indent, BLOCK_IN); |
| - }); |
| - |
| - // 185 |
| - Node s_l_blockIndented(int indent, int ctx) { |
| - var additionalIndent = countIndentation(); |
| - return or([ |
| - () => transaction(() { |
| - if (!truth(s_indent(additionalIndent))) return null; |
| - return or([ |
| - () => ns_l_compactSequence(indent + 1 + additionalIndent), |
| - () => ns_l_compactMapping(indent + 1 + additionalIndent)]); |
| - }), |
| - () => s_l_blockNode(indent, ctx), |
| - () => s_l_comments() ? e_node() : null]); |
| - } |
| + /// Expect a flow node. |
| + static const FLOW_NODE = const _State("FLOW_NODE"); |
| - // 186 |
| - Node ns_l_compactSequence(int indent) => context('sequence', () { |
| - var start = _scanner.state; |
| - var first = c_l_blockSeqEntry(indent); |
| - if (!truth(first)) return null; |
| - |
| - var content = zeroOrMore(() => transaction(() { |
| - if (!truth(s_indent(indent))) return null; |
| - return c_l_blockSeqEntry(indent); |
| - })); |
| - content.insert(0, first); |
| - |
| - return new SequenceNode("?", content, _scanner.spanFrom(start)); |
| - }); |
| - |
| - // 187 |
| - Node l_blockMapping(int indent) => context('mapping', () { |
| - var additionalIndent = countIndentation() - indent; |
| - if (additionalIndent <= 0) return null; |
| - |
| - var start = _scanner.state; |
| - var pairs = oneOrMore(() => transaction(() { |
| - if (!truth(s_indent(indent + additionalIndent))) return null; |
| - return ns_l_blockMapEntry(indent + additionalIndent); |
| - })); |
| - if (!truth(pairs)) return null; |
| - |
| - return map(pairs, _scanner.spanFrom(start)); |
| - }); |
| - |
| - // 188 |
| - Pair<Node, Node> ns_l_blockMapEntry(int indent) => or([ |
| - () => c_l_blockMapExplicitEntry(indent), |
| - () => ns_l_blockMapImplicitEntry(indent) |
| - ]); |
| - |
| - // 189 |
| - Pair<Node, Node> c_l_blockMapExplicitEntry(int indent) { |
| - var key = c_l_blockMapExplicitKey(indent); |
| - if (!truth(key)) return null; |
| - |
| - var value = or([ |
| - () => l_blockMapExplicitValue(indent), |
| - e_node |
| - ]); |
| - |
| - return new Pair<Node, Node>(key, value); |
| - } |
| + /// Expect the first entry of a block sequence. |
| + static const BLOCK_SEQUENCE_FIRST_ENTRY = |
| + const _State("BLOCK_SEQUENCE_FIRST_ENTRY"); |
| - // 190 |
| - Node c_l_blockMapExplicitKey(int indent) => transaction(() { |
| - if (!truth(c_indicator(C_MAPPING_KEY))) return null; |
| - return s_l_blockIndented(indent, BLOCK_OUT); |
| - }); |
| - |
| - // 191 |
| - Node l_blockMapExplicitValue(int indent) => transaction(() { |
| - if (!truth(s_indent(indent))) return null; |
| - if (!truth(c_indicator(C_MAPPING_VALUE))) return null; |
| - return s_l_blockIndented(indent, BLOCK_OUT); |
| - }); |
| - |
| - // 192 |
| - Pair<Node, Node> ns_l_blockMapImplicitEntry(int indent) => transaction(() { |
| - var key = or([ns_s_blockMapImplicitKey, e_node]); |
| - var value = c_l_blockMapImplicitValue(indent); |
| - return truth(value) ? new Pair<Node, Node>(key, value) : null; |
| - }); |
| - |
| - // 193 |
| - Node ns_s_blockMapImplicitKey() => context('mapping key', () => or([ |
| - () => c_s_implicitJsonKey(BLOCK_KEY), |
| - () => ns_s_implicitYamlKey(BLOCK_KEY) |
| - ])); |
| - |
| - // 194 |
| - Node c_l_blockMapImplicitValue(int indent) => context('mapping value', () => |
| - transaction(() { |
| - if (!truth(c_indicator(C_MAPPING_VALUE))) return null; |
| - return or([ |
| - () => s_l_blockNode(indent, BLOCK_OUT), |
| - () => s_l_comments() ? e_node() : null |
| - ]); |
| - })); |
| - |
| - // 195 |
| - Node ns_l_compactMapping(int indent) => context('mapping', () { |
| - var start = _scanner.state; |
| - var first = ns_l_blockMapEntry(indent); |
| - if (!truth(first)) return null; |
| - |
| - var pairs = zeroOrMore(() => transaction(() { |
| - if (!truth(s_indent(indent))) return null; |
| - return ns_l_blockMapEntry(indent); |
| - })); |
| - pairs.insert(0, first); |
| - |
| - return map(pairs, _scanner.spanFrom(start)); |
| - }); |
| - |
| - // 196 |
| - Node s_l_blockNode(int indent, int ctx) => or([ |
| - () => s_l_blockInBlock(indent, ctx), |
| - () => s_l_flowInBlock(indent) |
| - ]); |
| - |
| - // 197 |
| - Node s_l_flowInBlock(int indent) => transaction(() { |
| - if (!truth(s_separate(indent + 1, FLOW_OUT))) return null; |
| - var node = ns_flowNode(indent + 1, FLOW_OUT); |
| - if (!truth(node)) return null; |
| - if (!truth(s_l_comments())) return null; |
| - return node; |
| - }); |
| - |
| - // 198 |
| - Node s_l_blockInBlock(int indent, int ctx) => or([ |
| - () => s_l_blockScalar(indent, ctx), |
| - () => s_l_blockCollection(indent, ctx) |
| - ]); |
| - |
| - // 199 |
| - Node s_l_blockScalar(int indent, int ctx) => transaction(() { |
| - if (!truth(s_separate(indent + 1, ctx))) return null; |
| - var props = transaction(() { |
| - var innerProps = c_ns_properties(indent + 1, ctx); |
| - if (!truth(innerProps)) return null; |
| - if (!truth(s_separate(indent + 1, ctx))) return null; |
| - return innerProps; |
| - }); |
| - |
| - var node = or([() => c_l_literal(indent), () => c_l_folded(indent)]); |
| - if (!truth(node)) return null; |
| - return addProps(node, props); |
| - }); |
| - |
| - // 200 |
| - Node s_l_blockCollection(int indent, int ctx) => transaction(() { |
| - var props = transaction(() { |
| - if (!truth(s_separate(indent + 1, ctx))) return null; |
| - return c_ns_properties(indent + 1, ctx); |
| - }); |
| - |
| - if (!truth(s_l_comments())) return null; |
| - return or([ |
| - () => l_blockSequence(seqSpaces(indent, ctx)), |
| - () => l_blockMapping(indent)]); |
| - }); |
| - |
| - // 201 |
| - int seqSpaces(int indent, int ctx) => ctx == BLOCK_OUT ? indent - 1 : indent; |
| - |
| - // 202 |
| - void l_documentPrefix() { |
| - zeroOrMore(l_comment); |
| - } |
| + /// Expect an entry of a block sequence. |
| + static const BLOCK_SEQUENCE_ENTRY = const _State("BLOCK_SEQUENCE_ENTRY"); |
| - // 203 |
| - bool c_directivesEnd() => rawString("---"); |
| - |
| - // 204 |
| - bool c_documentEnd() => rawString("..."); |
| - |
| - // 205 |
| - bool l_documentSuffix() => transaction(() { |
| - if (!truth(c_documentEnd())) return false; |
| - return s_l_comments(); |
| - }); |
| - |
| - // 206 |
| - bool c_forbidden() { |
| - if (!_inBareDocument || !atStartOfLine) return false; |
| - var forbidden = false; |
| - transaction(() { |
| - if (!truth(or([c_directivesEnd, c_documentEnd]))) return; |
| - var char = peek(); |
| - forbidden = isBreak(char) || isSpace(char) || _scanner.isDone; |
| - return; |
| - }); |
| - return forbidden; |
| - } |
| + /// Expect an entry of an indentless sequence. |
| + static const INDENTLESS_SEQUENCE_ENTRY = |
| + const _State("INDENTLESS_SEQUENCE_ENTRY"); |
| - // 207 |
| - Node l_bareDocument() { |
| - try { |
| - _inBareDocument = true; |
| - return s_l_blockNode(-1, BLOCK_IN); |
| - } finally { |
| - _inBareDocument = false; |
| - } |
| - } |
| + /// Expect the first key of a block mapping. |
| + static const BLOCK_MAPPING_FIRST_KEY = |
| + const _State("BLOCK_MAPPING_FIRST_KEY"); |
| - // 208 |
| - Node l_explicitDocument() { |
| - if (!truth(c_directivesEnd())) return null; |
| - var doc = l_bareDocument(); |
| - if (truth(doc)) return doc; |
| + /// Expect a block mapping key. |
| + static const BLOCK_MAPPING_KEY = const _State("BLOCK_MAPPING_KEY"); |
| - doc = e_node(); |
| - s_l_comments(); |
| - return doc; |
| - } |
| + /// Expect a block mapping value. |
| + static const BLOCK_MAPPING_VALUE = const _State("BLOCK_MAPPING_VALUE"); |
| - // 209 |
| - Node l_directiveDocument() { |
| - if (!truth(oneOrMore(l_directive))) return null; |
| - var doc = l_explicitDocument(); |
| - if (doc != null) return doc; |
| - parseFailed(); |
| - return null; // Unreachable. |
| - } |
| + /// Expect the first entry of a flow sequence. |
| + static const FLOW_SEQUENCE_FIRST_ENTRY = |
| + const _State("FLOW_SEQUENCE_FIRST_ENTRY"); |
| - // 210 |
| - Node l_anyDocument() => |
| - or([l_directiveDocument, l_explicitDocument, l_bareDocument]); |
| - |
| - // 211 |
| - Pair<List<Node>, SourceSpan> l_yamlStream() { |
| - var start = _scanner.state; |
| - var docs = []; |
| - zeroOrMore(l_documentPrefix); |
| - var first = zeroOrOne(l_anyDocument); |
| - if (!truth(first)) first = e_node(); |
| - docs.add(first); |
| - |
| - zeroOrMore(() { |
| - var doc; |
| - if (truth(oneOrMore(l_documentSuffix))) { |
| - zeroOrMore(l_documentPrefix); |
| - doc = zeroOrOne(l_anyDocument); |
| - } else { |
| - zeroOrMore(l_documentPrefix); |
| - doc = zeroOrOne(l_explicitDocument); |
| - } |
| - if (truth(doc)) docs.add(doc); |
| - return doc; |
| - }); |
| + /// Expect an entry of a flow sequence. |
| + static const FLOW_SEQUENCE_ENTRY = const _State("FLOW_SEQUENCE_ENTRY"); |
| - if (!_scanner.isDone) parseFailed(); |
| - return new Pair(docs, _scanner.spanFrom(start)); |
| - } |
| -} |
| + /// Expect a key of an ordered mapping. |
| + static const FLOW_SEQUENCE_ENTRY_MAPPING_KEY = |
| + const _State("FLOW_SEQUENCE_ENTRY_MAPPING_KEY"); |
| -/// The information in the header for a block scalar. |
| -class _BlockHeader { |
| - final int additionalIndent; |
| - final int chomping; |
| + /// Expect a value of an ordered mapping. |
| + static const FLOW_SEQUENCE_ENTRY_MAPPING_VALUE = |
| + const _State("FLOW_SEQUENCE_ENTRY_MAPPING_VALUE"); |
| - _BlockHeader(this.additionalIndent, this.chomping); |
| + /// Expect the and of an ordered mapping entry. |
| + static const FLOW_SEQUENCE_ENTRY_MAPPING_END = |
| + const _State("FLOW_SEQUENCE_ENTRY_MAPPING_END"); |
| - bool get autoDetectIndent => additionalIndent == null; |
| -} |
| + /// Expect the first key of a flow mapping. |
| + static const FLOW_MAPPING_FIRST_KEY = const _State("FLOW_MAPPING_FIRST_KEY"); |
| -/// A range of characters in the YAML document, from [start] to [end] |
| -/// (inclusive). |
| -class _Range { |
| - /// The first character in the range. |
| - final int start; |
| + /// Expect a key of a flow mapping. |
| + static const FLOW_MAPPING_KEY = const _State("FLOW_MAPPING_KEY"); |
| - /// The last character in the range. |
| - final int end; |
| + /// Expect a value of a flow mapping. |
| + static const FLOW_MAPPING_VALUE = const _State("FLOW_MAPPING_VALUE"); |
| - _Range(this.start, this.end); |
| + /// Expect an empty value of a flow mapping. |
| + static const FLOW_MAPPING_EMPTY_VALUE = |
| + const _State("FLOW_MAPPING_EMPTY_VALUE"); |
| - /// Returns whether or not [pos] lies within this range. |
| - bool contains(int pos) => pos >= start && pos <= end; |
| -} |
| + /// Expect nothing. |
| + static const END = const _State("END"); |
| -/// A map that associates [E] values with [_Range]s. It's efficient to create |
| -/// new associations, but finding the value associated with a position is more |
| -/// expensive. |
| -class _RangeMap<E> { |
| - /// The ranges and their associated elements. |
| - final List<Pair<_Range, E>> _contents = <Pair<_Range, E>>[]; |
| - |
| - _RangeMap(); |
| - |
| - /// Returns the value associated with the range in which [pos] lies, or null |
| - /// if there is no such range. If there's more than one such range, the most |
| - /// recently set one is used. |
| - E operator[](int pos) { |
| - // Iterate backwards through contents so the more recent range takes |
| - // precedence. |
| - for (var pair in _contents.reversed) { |
| - if (pair.first.contains(pos)) return pair.last; |
| - } |
| - return null; |
| - } |
| + final String name; |
| + |
| + const _State(this.name); |
| - /// Associates [value] with [range]. |
| - operator[]=(_Range range, E value) => |
| - _contents.add(new Pair<_Range, E>(range, value)); |
| + String toString() => name; |
| } |