Index: pkg/analyzer/lib/src/generated/incremental_scanner.dart |
diff --git a/pkg/analyzer/lib/src/generated/incremental_scanner.dart b/pkg/analyzer/lib/src/generated/incremental_scanner.dart |
deleted file mode 100644 |
index a0bd23863d99d0a317bd78d8b9965b0af6856b67..0000000000000000000000000000000000000000 |
--- a/pkg/analyzer/lib/src/generated/incremental_scanner.dart |
+++ /dev/null |
@@ -1,251 +0,0 @@ |
-// Copyright (c) 2014, the Dart project authors. Please see the AUTHORS file |
-// for details. All rights reserved. Use of this source code is governed by a |
-// BSD-style license that can be found in the LICENSE file. |
- |
-@deprecated |
-library analyzer.src.generated.incremental_scanner; |
- |
-import "dart:math" as math; |
- |
-import 'package:analyzer/dart/ast/token.dart'; |
-import 'package:analyzer/src/dart/scanner/reader.dart'; |
-import 'package:analyzer/src/dart/scanner/scanner.dart'; |
-import 'package:analyzer/src/generated/engine.dart'; |
-import 'package:analyzer/src/generated/error.dart'; |
-import 'package:analyzer/src/generated/source.dart'; |
-import 'package:analyzer/src/generated/utilities_collection.dart' show TokenMap; |
- |
-/** |
- * An `IncrementalScanner` is a scanner that scans a subset of a string and |
- * inserts the resulting tokens into the middle of an existing token stream. |
- */ |
-@deprecated |
-class IncrementalScanner { |
- /** |
- * The source being scanned. |
- */ |
- final Source source; |
- |
- /** |
- * The reader used to access the characters in the source. |
- */ |
- final CharacterReader reader; |
- |
- /** |
- * The error listener that will be informed of any errors that are found |
- * during the scan. |
- * |
- * TODO(brianwilkerson) Replace this with a list of errors so that we can |
- * update the errors. |
- */ |
- final AnalysisErrorListener errorListener; |
- |
- final AnalysisOptions _options; |
- |
- /** |
- * A map from tokens that were copied to the copies of the tokens. |
- */ |
- TokenMap _tokenMap = new TokenMap(); |
- |
- /** |
- * The token immediately to the left of the range of tokens that were |
- * modified. |
- */ |
- Token leftToken; |
- |
- /** |
- * The token immediately to the right of the range of tokens that were |
- * modified. |
- */ |
- Token rightToken; |
- |
- /** |
- * A flag indicating whether there were any non-comment tokens changed (other |
- * than having their position updated) as a result of the modification. |
- */ |
- bool hasNonWhitespaceChange = false; |
- |
- /** |
- * Initialize a newly created scanner to scan characters within the given |
- * [source]. The content of the source can be read using the given [reader]. |
- * Any errors that are found will be reported to the given [errorListener]. |
- * [_options] will determine how scanning is to be performed. |
- */ |
- IncrementalScanner( |
- this.source, this.reader, this.errorListener, this._options); |
- |
- /** |
- * Return a map from tokens that were copied to the copies of the tokens. |
- * |
- * @return a map from tokens that were copied to the copies of the tokens |
- */ |
- TokenMap get tokenMap => _tokenMap; |
- |
- /** |
- * Given the [stream] of tokens scanned from the original source, the modified |
- * source (the result of replacing one contiguous range of characters with |
- * another string of characters), and a specification of the modification that |
- * was made, update the token stream to reflect the modified source. Return |
- * the first token in the updated token stream. |
- * |
- * The [stream] is expected to be the first non-EOF token in the token stream. |
- * |
- * The modification is specified by the [index] of the first character in both |
- * the original and modified source that was affected by the modification, the |
- * number of characters removed from the original source (the [removedLength]) |
- * and the number of characters added to the modified source (the |
- * [insertedLength]). |
- */ |
- Token rescan(Token stream, int index, int removedLength, int insertedLength) { |
- Token leftEof = stream.previous; |
- // |
- // Compute the delta between the character index of characters after the |
- // modified region in the original source and the index of the corresponding |
- // character in the modified source. |
- // |
- int delta = insertedLength - removedLength; |
- // |
- // Skip past the tokens whose end is less than the replacement start. (If |
- // the replacement start is equal to the end of an existing token, then it |
- // means that the existing token might have been modified, so we need to |
- // rescan it.) |
- // |
- while (stream.type != TokenType.EOF && stream.end < index) { |
- _tokenMap.put(stream, stream); |
- stream = stream.next; |
- } |
- Token oldFirst = stream; |
- Token oldLeftToken = stream.previous; |
- leftToken = oldLeftToken; |
- // |
- // Skip past tokens until we find a token whose offset is greater than the |
- // end of the removed region. (If the end of the removed region is equal to |
- // the beginning of an existing token, then it means that the existing token |
- // might have been modified, so we need to rescan it.) |
- // |
- int removedEnd = index + (removedLength == 0 ? 0 : removedLength - 1); |
- while (stream.type != TokenType.EOF && stream.offset <= removedEnd) { |
- stream = stream.next; |
- } |
- // |
- // Figure out which region of characters actually needs to be re-scanned. |
- // |
- Token oldLast; |
- Token oldRightToken; |
- if (stream.type != TokenType.EOF && removedEnd + 1 == stream.offset) { |
- oldLast = stream; |
- stream = stream.next; |
- oldRightToken = stream; |
- } else { |
- oldLast = stream.previous; |
- oldRightToken = stream; |
- } |
- // |
- // Compute the range of characters that are known to need to be rescanned. |
- // If the index is within an existing token, then we need to start at the |
- // beginning of the token. |
- // |
- int scanStart = math.max(oldLeftToken.end, 0); |
- int scanEnd = oldRightToken.offset + delta; |
- // |
- // Rescan the characters that need to be rescanned. |
- // |
- Token replacementStart = _scanRange(scanStart, scanEnd); |
- oldLeftToken.setNext(replacementStart); |
- Token replacementEnd = _findEof(replacementStart).previous; |
- replacementEnd.setNext(stream); |
- // |
- // Apply the delta to the tokens after the last new token. |
- // |
- _updateOffsets(stream, delta); |
- rightToken = stream; |
- // |
- // If the index is immediately after an existing token and the inserted |
- // characters did not change that original token, then adjust the leftToken |
- // to be the next token. For example, in "a; c;" --> "a;b c;", the leftToken |
- // was ";", but this code advances it to "b" since "b" is the first new |
- // token. |
- // |
- Token newFirst = leftToken.next; |
- while (!identical(newFirst, rightToken) && |
- !identical(oldFirst, oldRightToken) && |
- newFirst.type != TokenType.EOF && |
- _equalTokens(oldFirst, newFirst)) { |
- _tokenMap.put(oldFirst, newFirst); |
- oldLeftToken = oldFirst; |
- oldFirst = oldFirst.next; |
- leftToken = newFirst; |
- newFirst = newFirst.next; |
- } |
- Token newLast = rightToken.previous; |
- while (!identical(newLast, leftToken) && |
- !identical(oldLast, oldLeftToken) && |
- newLast.type != TokenType.EOF && |
- _equalTokens(oldLast, newLast)) { |
- _tokenMap.put(oldLast, newLast); |
- oldRightToken = oldLast; |
- oldLast = oldLast.previous; |
- rightToken = newLast; |
- newLast = newLast.previous; |
- } |
- hasNonWhitespaceChange = !identical(leftToken.next, rightToken) || |
- !identical(oldLeftToken.next, oldRightToken); |
- // |
- // TODO(brianwilkerson) Begin tokens are not getting associated with the |
- // corresponding end tokens (because the end tokens have not been copied |
- // when we're copying the begin tokens). This could have implications for |
- // parsing. |
- // TODO(brianwilkerson) Update the lineInfo. |
- // |
- return leftEof.next; |
- } |
- |
- /** |
- * Return `true` if the [oldToken] and the [newToken] are equal to each other. |
- * For the purposes of the incremental scanner, two tokens are equal if they |
- * have the same type and lexeme. |
- */ |
- bool _equalTokens(Token oldToken, Token newToken) => |
- oldToken.type == newToken.type && |
- oldToken.length == newToken.length && |
- oldToken.lexeme == newToken.lexeme; |
- |
- /** |
- * Given a [token], return the EOF token that follows the token. |
- */ |
- Token _findEof(Token token) { |
- while (token.type != TokenType.EOF) { |
- token = token.next; |
- } |
- return token; |
- } |
- |
- /** |
- * Scan the token between the [start] (inclusive) and [end] (exclusive) |
- * offsets. |
- */ |
- Token _scanRange(int start, int end) { |
- Scanner scanner = new Scanner( |
- source, new CharacterRangeReader(reader, start, end), errorListener); |
- return scanner.tokenize(); |
- } |
- |
- /** |
- * Update the offsets of every token from the given [token] to the end of the |
- * stream by adding the given [delta]. |
- */ |
- void _updateOffsets(Token token, int delta) { |
- while (token.type != TokenType.EOF) { |
- _tokenMap.put(token, token); |
- token.offset += delta; |
- Token comment = token.precedingComments; |
- while (comment != null) { |
- comment.offset += delta; |
- comment = comment.next; |
- } |
- token = token.next; |
- } |
- _tokenMap.put(token, token); |
- token.offset += delta; |
- } |
-} |