Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(18)

Unified Diff: pkg/analyzer/test/generated/incremental_scanner_test.dart

Issue 2330803002: Remove IncrementalScanner and IncrementalParser. (Closed)
Patch Set: Created 4 years, 3 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
« no previous file with comments | « pkg/analyzer/lib/src/generated/parser.dart ('k') | pkg/analyzer/test/generated/parser_test.dart » ('j') | no next file with comments »
Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
Index: pkg/analyzer/test/generated/incremental_scanner_test.dart
diff --git a/pkg/analyzer/test/generated/incremental_scanner_test.dart b/pkg/analyzer/test/generated/incremental_scanner_test.dart
deleted file mode 100644
index 4464f5cdcab406e91ce10771f00a2d0d2bd8d836..0000000000000000000000000000000000000000
--- a/pkg/analyzer/test/generated/incremental_scanner_test.dart
+++ /dev/null
@@ -1,548 +0,0 @@
-// Copyright (c) 2014, the Dart project authors. Please see the AUTHORS file
-// for details. All rights reserved. Use of this source code is governed by a
-// BSD-style license that can be found in the LICENSE file.
-
-@deprecated
-library analyzer.test.generated.incremental_scanner_test;
-
-import 'package:analyzer/dart/ast/token.dart';
-import 'package:analyzer/src/dart/scanner/reader.dart';
-import 'package:analyzer/src/dart/scanner/scanner.dart';
-import 'package:analyzer/src/generated/engine.dart';
-import 'package:analyzer/src/generated/incremental_scanner.dart';
-import 'package:analyzer/src/generated/source.dart';
-import 'package:test_reflective_loader/test_reflective_loader.dart';
-import 'package:unittest/unittest.dart';
-
-import '../utils.dart';
-import 'test_support.dart';
-
-main() {
- initializeTestEnvironment();
- defineReflectiveTests(IncrementalScannerTest);
-}
-
-@reflectiveTest
-class IncrementalScannerTest extends EngineTestCase {
- /**
- * The first token from the token stream resulting from parsing the original
- * source, or `null` if [scan] has not been invoked.
- */
- Token _originalTokens;
-
- /**
- * The scanner used to perform incremental scanning, or `null` if [scan] has
- * not been invoked.
- */
- IncrementalScanner _incrementalScanner;
-
- /**
- * The first token from the token stream resulting from performing an
- * incremental scan, or `null` if [scan] has not been invoked.
- */
- Token _incrementalTokens;
-
- void fail_insert_beginning() {
- // This is currently reporting the changed range as being from 0 to 5, but
- // that would force us to re-parse both classes, which is clearly
- // sub-optimal.
- //
- // "class B {}"
- // "class A {} class B {}"
- _scan("", "", "class A {} ", "class B {}");
- _assertTokens(-1, 4, ["class", "A", "{", "}", "class", "B", "{", "}"]);
- expect(_incrementalScanner.hasNonWhitespaceChange, isTrue);
- }
-
- void fail_insert_comment_afterIdentifier() {
- // "a + b"
- // "a /* TODO */ + b"
- _scan("a", "", " /* TODO */", " + b");
- _assertTokens(0, 1, ["a", "+", "b"]);
- _assertComments(1, ["/* TODO */"]);
- expect(_incrementalScanner.hasNonWhitespaceChange, isFalse);
- }
-
- void fail_insert_comment_beforeIdentifier() {
- // "a + b"
- // "a + /* TODO */ b"
- _scan("a + ", "", "/* TODO */ ", "b");
- _assertTokens(1, 2, ["a", "+", "b"]);
- _assertComments(2, ["/* TODO */"]);
- expect(_incrementalScanner.hasNonWhitespaceChange, isFalse);
- }
-
- void fail_insert_inComment() {
- // "a /* TO */ b"
- // "a /* TODO */ b"
- _scan("a /* TO", "", "DO", " */ b");
- _assertTokens(0, 1, ["a", "b"]);
- _assertComments(1, ["/* TODO */"]);
- expect(_incrementalScanner.hasNonWhitespaceChange, isFalse);
- }
-
- void test_delete_identifier_beginning() {
- // "abs + b;"
- // "s + b;"
- _scan("", "ab", "", "s + b;");
- _assertTokens(-1, 1, ["s", "+", "b", ";"]);
- expect(_incrementalScanner.hasNonWhitespaceChange, isTrue);
- }
-
- void test_delete_identifier_end() {
- // "abs + b;"
- // "a + b;"
- _scan("a", "bs", "", " + b;");
- _assertTokens(-1, 1, ["a", "+", "b", ";"]);
- expect(_incrementalScanner.hasNonWhitespaceChange, isTrue);
- }
-
- void test_delete_identifier_middle() {
- // "abs + b;"
- // "as + b;"
- _scan("a", "b", "", "s + b;");
- _assertTokens(-1, 1, ["as", "+", "b", ";"]);
- expect(_incrementalScanner.hasNonWhitespaceChange, isTrue);
- }
-
- void test_delete_mergeTokens() {
- // "a + b + c;"
- // "ac;"
- _scan("a", " + b + ", "", "c;");
- _assertTokens(-1, 1, ["ac", ";"]);
- expect(_incrementalScanner.hasNonWhitespaceChange, isTrue);
- }
-
- void test_delete_whitespace() {
- // "a + b + c;"
- // "a+ b + c;"
- _scan("a", " ", "", "+ b + c;");
- _assertTokens(1, 2, ["a", "+", "b", "+", "c", ";"]);
- expect(_incrementalScanner.hasNonWhitespaceChange, isFalse);
- }
-
- void test_insert_convertOneFunctionToTwo_noOverlap() {
- // "f() {}"
- // "f() => 0; g() {}"
- _scan("f()", "", " => 0; g()", " {}");
- _assertTokens(
- 2, 9, ["f", "(", ")", "=>", "0", ";", "g", "(", ")", "{", "}"]);
- expect(_incrementalScanner.hasNonWhitespaceChange, isTrue);
- }
-
- void test_insert_convertOneFunctionToTwo_overlap() {
- // "f() {}"
- // "f() {} g() {}"
- _scan("f() {", "", "} g() {", "}");
- _assertTokens(4, 10, ["f", "(", ")", "{", "}", "g", "(", ")", "{", "}"]);
- expect(_incrementalScanner.hasNonWhitespaceChange, isTrue);
- }
-
- void test_insert_end() {
- // "class A {}"
- // "class A {} class B {}"
- _scan("class A {}", "", " class B {}", "");
- _assertTokens(3, 8, ["class", "A", "{", "}", "class", "B", "{", "}"]);
- expect(_incrementalScanner.hasNonWhitespaceChange, isTrue);
- }
-
- void test_insert_identifierAndPeriod() {
- // "a + b;"
- // "a + x.b;"
- _scan("a + ", "", "x.", "b;");
- _assertTokens(1, 4, ["a", "+", "x", ".", "b", ";"]);
- }
-
- void test_insert_inIdentifier_left_firstToken() {
- // "a + b;"
- // "xa + b;"
- _scan("", "", "x", "a + b;");
- _assertTokens(-1, 1, ["xa", "+", "b", ";"]);
- expect(_incrementalScanner.hasNonWhitespaceChange, isTrue);
- }
-
- void test_insert_inIdentifier_left_lastToken() {
- // "a + b"
- // "a + xb"
- _scan("a + ", "", "x", "b");
- _assertTokens(1, 3, ["a", "+", "xb"]);
- expect(_incrementalScanner.hasNonWhitespaceChange, isTrue);
- }
-
- void test_insert_inIdentifier_left_middleToken() {
- // "a + b;"
- // "a + xb;"
- _scan("a + ", "", "x", "b;");
- _assertTokens(1, 3, ["a", "+", "xb", ";"]);
- expect(_incrementalScanner.hasNonWhitespaceChange, isTrue);
- }
-
- void test_insert_inIdentifier_middle() {
- // "cat;"
- // "cart;"
- _scan("ca", "", "r", "t;");
- _assertTokens(-1, 1, ["cart", ";"]);
- expect(_incrementalScanner.hasNonWhitespaceChange, isTrue);
- }
-
- void test_insert_inIdentifier_right_firstToken() {
- // "a + b;"
- // "abs + b;"
- _scan("a", "", "bs", " + b;");
- _assertTokens(-1, 1, ["abs", "+", "b", ";"]);
- expect(_incrementalScanner.hasNonWhitespaceChange, isTrue);
- }
-
- void test_insert_inIdentifier_right_lastToken() {
- // "a + b"
- // "a + bc"
- _scan("a + b", "", "c", "");
- _assertTokens(1, 3, ["a", "+", "bc"]);
- expect(_incrementalScanner.hasNonWhitespaceChange, isTrue);
- }
-
- void test_insert_inIdentifier_right_middleToken() {
- // "a + b;"
- // "a + by;"
- _scan("a + b", "", "y", ";");
- _assertTokens(1, 3, ["a", "+", "by", ";"]);
- expect(_incrementalScanner.hasNonWhitespaceChange, isTrue);
- }
-
- void test_insert_newIdentifier_noSpaceBefore() {
- // "a; c;"
- // "a;b c;"
- _scan("a;", "", "b", " c;");
- _assertTokens(1, 3, ["a", ";", "b", "c", ";"]);
- expect(_incrementalScanner.hasNonWhitespaceChange, isTrue);
- }
-
- void test_insert_newIdentifier_spaceBefore() {
- // "a; c;"
- // "a; b c;"
- _scan("a; ", "", "b ", "c;");
- _assertTokens(1, 3, ["a", ";", "b", "c", ";"]);
- expect(_incrementalScanner.hasNonWhitespaceChange, isTrue);
- }
-
- void test_insert_period_afterIdentifier() {
- // "a + b;"
- // "a + b.;"
- _scan("a + b", "", ".", ";");
- _assertTokens(2, 4, ["a", "+", "b", ".", ";"]);
- }
-
- void test_insert_period_betweenIdentifiers_left() {
- // "a b;"
- // "a. b;"
- _scan("a", "", ".", " b;");
- _assertTokens(0, 2, ["a", ".", "b", ";"]);
- expect(_incrementalScanner.hasNonWhitespaceChange, isTrue);
- }
-
- void test_insert_period_betweenIdentifiers_middle() {
- // "a b;"
- // "a . b;"
- _scan("a ", "", ".", " b;");
- _assertTokens(0, 2, ["a", ".", "b", ";"]);
- expect(_incrementalScanner.hasNonWhitespaceChange, isTrue);
- }
-
- void test_insert_period_betweenIdentifiers_right() {
- // "a b;"
- // "a .b;"
- _scan("a ", "", ".", "b;");
- _assertTokens(0, 2, ["a", ".", "b", ";"]);
- }
-
- void test_insert_period_insideExistingIdentifier() {
- // "ab;"
- // "a.b;"
- _scan("a", "", ".", "b;");
- _assertTokens(-1, 3, ["a", ".", "b", ";"]);
- expect(_incrementalScanner.hasNonWhitespaceChange, isTrue);
- }
-
- void test_insert_periodAndIdentifier() {
- // "a + b;"
- // "a + b.x;"
- _scan("a + b", "", ".x", ";");
- _assertTokens(2, 5, ["a", "+", "b", ".", "x", ";"]);
- }
-
- void test_insert_splitIdentifier() {
- // "cob;"
- // "cow.b;"
- _scan("co", "", "w.", "b;");
- _assertTokens(-1, 3, ["cow", ".", "b", ";"]);
- expect(_incrementalScanner.hasNonWhitespaceChange, isTrue);
- }
-
- void test_insert_tokens_within_whitespace() {
- // "a ;"
- // "a +b ;"
- _scan("a ", "", "+b", " ;");
- _assertTokens(0, 3, ["a", "+", "b", ";"]);
- expect(_incrementalScanner.hasNonWhitespaceChange, isTrue);
- }
-
- void test_insert_whitespace_beginning_beforeToken() {
- // "a + b;"
- // " a + b;"
- _scan("", "", " ", "a + b;");
- _assertTokens(0, 1, ["a", "+", "b", ";"]);
- expect(_incrementalScanner.hasNonWhitespaceChange, isFalse);
- }
-
- void test_insert_whitespace_betweenTokens() {
- // "a + b;"
- // "a + b;"
- _scan("a ", "", " ", "+ b;");
- _assertTokens(1, 2, ["a", "+", "b", ";"]);
- expect(_incrementalScanner.hasNonWhitespaceChange, isFalse);
- }
-
- void test_insert_whitespace_end_afterToken() {
- // "a + b;"
- // "a + b; "
- _scan("a + b;", "", " ", "");
- _assertTokens(3, 4, ["a", "+", "b", ";"]);
- expect(_incrementalScanner.hasNonWhitespaceChange, isFalse);
- }
-
- void test_insert_whitespace_end_afterWhitespace() {
- // "a + b; "
- // "a + b; "
- _scan("a + b; ", "", " ", "");
- _assertTokens(3, 4, ["a", "+", "b", ";"]);
- expect(_incrementalScanner.hasNonWhitespaceChange, isFalse);
- }
-
- void test_insert_whitespace_withMultipleComments() {
- // "//comment1", "//comment2", "a + b;"
- // "//comment1", "//comment2", "a + b;"
- _scan(
- r'''
-//comment1
-//comment2
-a''',
- "",
- " ",
- " + b;");
- _assertTokens(1, 2, ["a", "+", "b", ";"]);
- expect(_incrementalScanner.hasNonWhitespaceChange, isFalse);
- }
-
- void test_replace_identifier_beginning() {
- // "bell + b;"
- // "fell + b;"
- _scan("", "b", "f", "ell + b;");
- _assertTokens(-1, 1, ["fell", "+", "b", ";"]);
- expect(_incrementalScanner.hasNonWhitespaceChange, isTrue);
- }
-
- void test_replace_identifier_end() {
- // "bell + b;"
- // "belt + b;"
- _scan("bel", "l", "t", " + b;");
- _assertTokens(-1, 1, ["belt", "+", "b", ";"]);
- expect(_incrementalScanner.hasNonWhitespaceChange, isTrue);
- }
-
- void test_replace_identifier_middle() {
- // "first + b;"
- // "frost + b;"
- _scan("f", "ir", "ro", "st + b;");
- _assertTokens(-1, 1, ["frost", "+", "b", ";"]);
- expect(_incrementalScanner.hasNonWhitespaceChange, isTrue);
- }
-
- void test_replace_multiple_partialFirstAndLast() {
- // "aa + bb;"
- // "ab * ab;"
- _scan("a", "a + b", "b * a", "b;");
- _assertTokens(-1, 3, ["ab", "*", "ab", ";"]);
- expect(_incrementalScanner.hasNonWhitespaceChange, isTrue);
- }
-
- void test_replace_operator_oneForMany() {
- // "a + b;"
- // "a * c - b;"
- _scan("a ", "+", "* c -", " b;");
- _assertTokens(0, 4, ["a", "*", "c", "-", "b", ";"]);
- expect(_incrementalScanner.hasNonWhitespaceChange, isTrue);
- }
-
- void test_replace_operator_oneForOne() {
- // "a + b;"
- // "a * b;"
- _scan("a ", "+", "*", " b;");
- _assertTokens(0, 2, ["a", "*", "b", ";"]);
- expect(_incrementalScanner.hasNonWhitespaceChange, isTrue);
- }
-
- /**
- * Assert that the comments associated with the token at the given [index]
- * have lexemes that match the given list of lexemes, both in number and in
- * content.
- */
- void _assertComments(int index, List<String> lexemes) {
- Token token = _incrementalTokens;
- for (int i = 0; i < index; i++) {
- token = token.next;
- }
- Token comment = token.precedingComments;
- if (lexemes.isEmpty) {
- expect(comment, isNull,
- reason: "No comments expected but comments found");
- }
- int count = 0;
- for (String lexeme in lexemes) {
- if (comment == null) {
- fail("Expected ${lexemes.length} comments but found $count");
- }
- expect(comment.lexeme, lexeme);
- count++;
- comment = comment.next;
- }
- if (comment != null) {
- while (comment != null) {
- count++;
- comment = comment.next;
- }
- fail("Expected ${lexemes.length} comments but found $count");
- }
- }
-
- /**
- * Assert that the [expected] token is equal to the [actual] token.
- */
- void _assertEqualTokens(Token actual, Token expected) {
- expect(actual.type, same(expected.type), reason: "Wrong type for token");
- expect(actual.lexeme, expected.lexeme, reason: "Wrong lexeme for token");
- expect(actual.offset, expected.offset,
- reason:
- "Wrong offset for token ('${actual.lexeme}' != '${expected.lexeme}')");
- expect(actual.length, expected.length,
- reason:
- "Wrong length for token ('${actual.lexeme}' != '${expected.lexeme}')");
- }
-
- /**
- * Assert that the result of the incremental scan matches the given list of
- * [lexemes] and that the left and right tokens correspond to the tokens at
- * the [leftIndex] and [rightIndex].
- */
- void _assertTokens(int leftIndex, int rightIndex, List<String> lexemes) {
- int count = lexemes.length;
- expect(leftIndex >= -1 && leftIndex < count, isTrue,
- reason: "Invalid left index");
- expect(rightIndex >= 0 && rightIndex <= count, isTrue,
- reason: "Invalid right index");
- Token leftToken = null;
- Token rightToken = null;
- Token token = _incrementalTokens;
- if (leftIndex < 0) {
- leftToken = token.previous;
- }
- for (int i = 0; i < count; i++) {
- expect(token.lexeme, lexemes[i]);
- if (i == leftIndex) {
- leftToken = token;
- }
- if (i == rightIndex) {
- rightToken = token;
- }
- token = token.next;
- }
- if (rightIndex >= count) {
- rightToken = token;
- }
- expect(token.type, same(TokenType.EOF), reason: "Too many tokens");
- if (leftIndex >= 0) {
- expect(leftToken, isNotNull);
- }
- expect(_incrementalScanner.leftToken, same(leftToken),
- reason: "Invalid left token");
- if (rightIndex >= 0) {
- expect(rightToken, isNotNull);
- }
- expect(_incrementalScanner.rightToken, same(rightToken),
- reason: "Invalid right token");
- }
-
- /**
- * Given a description of the original and modified contents, perform an
- * incremental scan of the two pieces of text. Verify that the incremental
- * scan produced the same tokens as those that would be produced by a full
- * scan of the new contents.
- *
- * The original content is the concatenation of the [prefix], [removed] and
- * [suffix] fragments. The modeified content is the concatenation of the
- * [prefix], [added] and [suffix] fragments.
- */
- void _scan(String prefix, String removed, String added, String suffix) {
- //
- // Compute the information needed to perform the test.
- //
- String originalContents = "$prefix$removed$suffix";
- String modifiedContents = "$prefix$added$suffix";
- int replaceStart = prefix.length;
- Source source = new TestSource();
- //
- // Scan the original contents.
- //
- GatheringErrorListener originalListener = new GatheringErrorListener();
- Scanner originalScanner = new Scanner(
- source, new CharSequenceReader(originalContents), originalListener);
- _originalTokens = originalScanner.tokenize();
- expect(_originalTokens, isNotNull);
- //
- // Scan the modified contents.
- //
- GatheringErrorListener modifiedListener = new GatheringErrorListener();
- Scanner modifiedScanner = new Scanner(
- source, new CharSequenceReader(modifiedContents), modifiedListener);
- Token modifiedTokens = modifiedScanner.tokenize();
- expect(modifiedTokens, isNotNull);
- //
- // Incrementally scan the modified contents.
- //
- GatheringErrorListener incrementalListener = new GatheringErrorListener();
- AnalysisOptionsImpl options = new AnalysisOptionsImpl();
- _incrementalScanner = new IncrementalScanner(source,
- new CharSequenceReader(modifiedContents), incrementalListener, options);
- _incrementalTokens = _incrementalScanner.rescan(
- _originalTokens, replaceStart, removed.length, added.length);
- //
- // Validate that the results of the incremental scan are the same as the
- // full scan of the modified source.
- //
- Token incrementalToken = _incrementalTokens;
- expect(incrementalToken, isNotNull);
- while (incrementalToken.type != TokenType.EOF &&
- modifiedTokens.type != TokenType.EOF) {
- _assertEqualTokens(incrementalToken, modifiedTokens);
- Token incrementalComment = incrementalToken.precedingComments;
- Token modifiedComment = modifiedTokens.precedingComments;
- while (incrementalComment != null && modifiedComment != null) {
- _assertEqualTokens(incrementalComment, modifiedComment);
- incrementalComment = incrementalComment.next;
- modifiedComment = modifiedComment.next;
- }
- expect(incrementalComment, isNull,
- reason:
- "Too many comment tokens preceeding '${incrementalToken.lexeme}'");
- expect(modifiedComment, isNull,
- reason:
- "Not enough comment tokens preceeding '${incrementalToken.lexeme}'");
- incrementalToken = incrementalToken.next;
- modifiedTokens = modifiedTokens.next;
- }
- expect(incrementalToken.type, same(TokenType.EOF),
- reason: "Too many tokens");
- expect(modifiedTokens.type, same(TokenType.EOF),
- reason: "Not enough tokens");
- // TODO(brianwilkerson) Verify that the errors are correct?
- }
-}
« no previous file with comments | « pkg/analyzer/lib/src/generated/parser.dart ('k') | pkg/analyzer/test/generated/parser_test.dart » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698