Index: pkg/analyzer_experimental/lib/src/generated/parser.dart |
diff --git a/pkg/analyzer_experimental/lib/src/generated/parser.dart b/pkg/analyzer_experimental/lib/src/generated/parser.dart |
index c24b42e210e3a12d7189a53cea54dbc2fd8a0d85..b03b6f2632d097032f7c67f746091c85cf3f7838 100644 |
--- a/pkg/analyzer_experimental/lib/src/generated/parser.dart |
+++ b/pkg/analyzer_experimental/lib/src/generated/parser.dart |
@@ -590,21 +590,21 @@ class Parser { |
*/ |
List<List<int>> getCodeBlockRanges(String comment) { |
List<List<int>> ranges = new List<List<int>>(); |
- int length2 = comment.length; |
+ int length = comment.length; |
int index = 0; |
if (comment.startsWith("/**") || comment.startsWith("///")) { |
index = 3; |
} |
- while (index < length2) { |
+ while (index < length) { |
int currentChar = comment.codeUnitAt(index); |
if (currentChar == 0xD || currentChar == 0xA) { |
index = index + 1; |
- while (index < length2 && Character.isWhitespace(comment.codeUnitAt(index))) { |
+ while (index < length && Character.isWhitespace(comment.codeUnitAt(index))) { |
index = index + 1; |
} |
if (JavaString.startsWithBefore(comment, "* ", index)) { |
int end = index + 6; |
- while (end < length2 && comment.codeUnitAt(end) != 0xD && comment.codeUnitAt(end) != 0xA) { |
+ while (end < length && comment.codeUnitAt(end) != 0xD && comment.codeUnitAt(end) != 0xA) { |
end = end + 1; |
} |
ranges.add(<int> [index, end]); |
@@ -613,7 +613,7 @@ class Parser { |
} else if (JavaString.startsWithBefore(comment, "[:", index)) { |
int end = comment.indexOf(":]", index + 2); |
if (end < 0) { |
- end = length2; |
+ end = length; |
} |
ranges.add(<int> [index, end]); |
index = end + 1; |
@@ -732,8 +732,8 @@ class Parser { |
if (token == null) { |
return false; |
} |
- TokenType type2 = token.type; |
- return identical(type2, TokenType.EQ) || identical(type2, TokenType.COMMA) || identical(type2, TokenType.SEMICOLON) || matches3(token, Keyword.IN); |
+ TokenType type = token.type; |
+ return identical(type, TokenType.EQ) || identical(type, TokenType.COMMA) || identical(type, TokenType.SEMICOLON) || matches3(token, Keyword.IN); |
} |
/** |
@@ -748,9 +748,9 @@ class Parser { |
* @return {@code true} if the bracketed text is followed by a link address |
*/ |
bool isLinkText(String comment, int rightIndex) { |
- int length2 = comment.length; |
+ int length = comment.length; |
int index = rightIndex + 1; |
- if (index >= length2) { |
+ if (index >= length) { |
return false; |
} |
int nextChar = comment.codeUnitAt(index); |
@@ -759,7 +759,7 @@ class Parser { |
} |
while (Character.isWhitespace(nextChar)) { |
index = index + 1; |
- if (index >= length2) { |
+ if (index >= length) { |
return false; |
} |
nextChar = comment.codeUnitAt(index); |
@@ -793,8 +793,8 @@ class Parser { |
token = token.next.next; |
} |
if (identical(token.type, TokenType.KEYWORD)) { |
- Keyword keyword2 = ((token as KeywordToken)).keyword; |
- return identical(keyword2, Keyword.CASE) || identical(keyword2, Keyword.DEFAULT); |
+ Keyword keyword = ((token as KeywordToken)).keyword; |
+ return identical(keyword, Keyword.CASE) || identical(keyword, Keyword.DEFAULT); |
} |
return false; |
} |
@@ -812,10 +812,10 @@ class Parser { |
int firstOffset = 2147483647; |
for (Token token in tokens) { |
if (token != null) { |
- int offset2 = token.offset; |
- if (offset2 < firstOffset) { |
+ int offset = token.offset; |
+ if (offset < firstOffset) { |
first = token; |
- firstOffset = offset2; |
+ firstOffset = offset; |
} |
} |
} |
@@ -865,28 +865,28 @@ class Parser { |
if (currentType != type2) { |
if (identical(type2, TokenType.GT)) { |
if (identical(currentType, TokenType.GT_GT)) { |
- int offset2 = _currentToken.offset; |
- Token first = new Token(TokenType.GT, offset2); |
- Token second = new Token(TokenType.GT, offset2 + 1); |
+ int offset = _currentToken.offset; |
+ Token first = new Token(TokenType.GT, offset); |
+ Token second = new Token(TokenType.GT, offset + 1); |
second.setNext(_currentToken.next); |
first.setNext(second); |
_currentToken.previous.setNext(first); |
_currentToken = first; |
return true; |
} else if (identical(currentType, TokenType.GT_EQ)) { |
- int offset3 = _currentToken.offset; |
- Token first = new Token(TokenType.GT, offset3); |
- Token second = new Token(TokenType.EQ, offset3 + 1); |
+ int offset = _currentToken.offset; |
+ Token first = new Token(TokenType.GT, offset); |
+ Token second = new Token(TokenType.EQ, offset + 1); |
second.setNext(_currentToken.next); |
first.setNext(second); |
_currentToken.previous.setNext(first); |
_currentToken = first; |
return true; |
} else if (identical(currentType, TokenType.GT_GT_EQ)) { |
- int offset4 = _currentToken.offset; |
- Token first = new Token(TokenType.GT, offset4); |
- Token second = new Token(TokenType.GT, offset4 + 1); |
- Token third = new Token(TokenType.EQ, offset4 + 2); |
+ int offset = _currentToken.offset; |
+ Token first = new Token(TokenType.GT, offset); |
+ Token second = new Token(TokenType.GT, offset + 1); |
+ Token third = new Token(TokenType.EQ, offset + 2); |
third.setNext(_currentToken.next); |
second.setNext(third); |
first.setNext(second); |
@@ -1723,10 +1723,10 @@ class Parser { |
List<CommentReference> references = new List<CommentReference>(); |
for (Token token in tokens) { |
String comment = token.lexeme; |
- int length2 = comment.length; |
+ int length = comment.length; |
List<List<int>> codeBlockRanges = getCodeBlockRanges(comment); |
int leftIndex = comment.indexOf('['); |
- while (leftIndex >= 0 && leftIndex + 1 < length2) { |
+ while (leftIndex >= 0 && leftIndex + 1 < length) { |
List<int> range = findRange(codeBlockRanges, leftIndex); |
if (range == null) { |
int rightIndex = comment.indexOf(']', leftIndex); |
@@ -2558,11 +2558,11 @@ class Parser { |
if (variableList == null) { |
reportError7(ParserErrorCode.MISSING_VARIABLE_IN_FOR_EACH, []); |
} else { |
- NodeList<VariableDeclaration> variables2 = variableList.variables; |
- if (variables2.length > 1) { |
- reportError7(ParserErrorCode.MULTIPLE_VARIABLES_IN_FOR_EACH, [variables2.length.toString()]); |
+ NodeList<VariableDeclaration> variables = variableList.variables; |
+ if (variables.length > 1) { |
+ reportError7(ParserErrorCode.MULTIPLE_VARIABLES_IN_FOR_EACH, [variables.length.toString()]); |
} |
- VariableDeclaration variable = variables2[0]; |
+ VariableDeclaration variable = variables[0]; |
if (variable.initializer != null) { |
reportError7(ParserErrorCode.INITIALIZED_VARIABLE_IN_FOR_EACH, []); |
} |
@@ -3313,34 +3313,34 @@ class Parser { |
} |
return parseBlock(); |
} else if (matches5(TokenType.KEYWORD) && !((_currentToken as KeywordToken)).keyword.isPseudoKeyword()) { |
- Keyword keyword2 = ((_currentToken as KeywordToken)).keyword; |
- if (identical(keyword2, Keyword.ASSERT)) { |
+ Keyword keyword = ((_currentToken as KeywordToken)).keyword; |
+ if (identical(keyword, Keyword.ASSERT)) { |
return parseAssertStatement(); |
- } else if (identical(keyword2, Keyword.BREAK)) { |
+ } else if (identical(keyword, Keyword.BREAK)) { |
return parseBreakStatement(); |
- } else if (identical(keyword2, Keyword.CONTINUE)) { |
+ } else if (identical(keyword, Keyword.CONTINUE)) { |
return parseContinueStatement(); |
- } else if (identical(keyword2, Keyword.DO)) { |
+ } else if (identical(keyword, Keyword.DO)) { |
return parseDoStatement(); |
- } else if (identical(keyword2, Keyword.FOR)) { |
+ } else if (identical(keyword, Keyword.FOR)) { |
return parseForStatement(); |
- } else if (identical(keyword2, Keyword.IF)) { |
+ } else if (identical(keyword, Keyword.IF)) { |
return parseIfStatement(); |
- } else if (identical(keyword2, Keyword.RETHROW)) { |
+ } else if (identical(keyword, Keyword.RETHROW)) { |
return new ExpressionStatement.full(parseRethrowExpression(), expect2(TokenType.SEMICOLON)); |
- } else if (identical(keyword2, Keyword.RETURN)) { |
+ } else if (identical(keyword, Keyword.RETURN)) { |
return parseReturnStatement(); |
- } else if (identical(keyword2, Keyword.SWITCH)) { |
+ } else if (identical(keyword, Keyword.SWITCH)) { |
return parseSwitchStatement(); |
- } else if (identical(keyword2, Keyword.THROW)) { |
+ } else if (identical(keyword, Keyword.THROW)) { |
return new ExpressionStatement.full(parseThrowExpression(), expect2(TokenType.SEMICOLON)); |
- } else if (identical(keyword2, Keyword.TRY)) { |
+ } else if (identical(keyword, Keyword.TRY)) { |
return parseTryStatement(); |
- } else if (identical(keyword2, Keyword.WHILE)) { |
+ } else if (identical(keyword, Keyword.WHILE)) { |
return parseWhileStatement(); |
- } else if (identical(keyword2, Keyword.VAR) || identical(keyword2, Keyword.FINAL)) { |
+ } else if (identical(keyword, Keyword.VAR) || identical(keyword, Keyword.FINAL)) { |
return parseVariableDeclarationStatement(commentAndMetadata); |
- } else if (identical(keyword2, Keyword.VOID)) { |
+ } else if (identical(keyword, Keyword.VOID)) { |
TypeName returnType = parseReturnType(); |
if (matchesIdentifier() && matchesAny(peek(), [TokenType.OPEN_PAREN, TokenType.OPEN_CURLY_BRACKET, TokenType.FUNCTION])) { |
return parseFunctionDeclarationStatement2(commentAndMetadata, returnType); |
@@ -3356,7 +3356,7 @@ class Parser { |
reportError7(ParserErrorCode.MISSING_STATEMENT, []); |
return new EmptyStatement.full(createSyntheticToken2(TokenType.SEMICOLON)); |
} |
- } else if (identical(keyword2, Keyword.CONST)) { |
+ } else if (identical(keyword, Keyword.CONST)) { |
if (matchesAny(peek(), [TokenType.LT, TokenType.OPEN_CURLY_BRACKET, TokenType.OPEN_SQUARE_BRACKET, TokenType.INDEX])) { |
return new ExpressionStatement.full(parseExpression2(), expect2(TokenType.SEMICOLON)); |
} else if (matches4(peek(), TokenType.IDENTIFIER)) { |
@@ -3368,7 +3368,7 @@ class Parser { |
} |
} |
return parseVariableDeclarationStatement(commentAndMetadata); |
- } else if (identical(keyword2, Keyword.NEW) || identical(keyword2, Keyword.TRUE) || identical(keyword2, Keyword.FALSE) || identical(keyword2, Keyword.NULL) || identical(keyword2, Keyword.SUPER) || identical(keyword2, Keyword.THIS)) { |
+ } else if (identical(keyword, Keyword.NEW) || identical(keyword, Keyword.TRUE) || identical(keyword, Keyword.FALSE) || identical(keyword, Keyword.NULL) || identical(keyword, Keyword.SUPER) || identical(keyword, Keyword.THIS)) { |
return new ExpressionStatement.full(parseExpression2(), expect2(TokenType.SEMICOLON)); |
} else { |
reportError7(ParserErrorCode.MISSING_STATEMENT, []); |
@@ -3421,9 +3421,9 @@ class Parser { |
FormalParameterList parameters = parseFormalParameterList(); |
return new FunctionTypedFormalParameter.full(commentAndMetadata.comment, commentAndMetadata.metadata, holder.type, identifier, parameters); |
} |
- TypeName type2 = holder.type; |
- if (type2 != null && matches3(type2.name.beginToken, Keyword.VOID)) { |
- reportError8(ParserErrorCode.VOID_PARAMETER, type2.name.beginToken, []); |
+ TypeName type = holder.type; |
+ if (type != null && matches3(type.name.beginToken, Keyword.VOID)) { |
+ reportError8(ParserErrorCode.VOID_PARAMETER, type.name.beginToken, []); |
} |
if (thisKeyword != null) { |
return new FieldFormalParameter.full(commentAndMetadata.comment, commentAndMetadata.metadata, holder.keyword, holder.type, thisKeyword, period, identifier); |
@@ -3459,9 +3459,9 @@ class Parser { |
} |
SimpleIdentifier name = new SimpleIdentifier.full(andAdvance); |
if (matches5(TokenType.EQ)) { |
- Token previous2 = _currentToken.previous; |
- if ((matches4(previous2, TokenType.EQ_EQ) || matches4(previous2, TokenType.BANG_EQ)) && _currentToken.offset == previous2.offset + 2) { |
- reportError7(ParserErrorCode.INVALID_OPERATOR, ["${previous2.lexeme}${_currentToken.lexeme}"]); |
+ Token previous = _currentToken.previous; |
+ if ((matches4(previous, TokenType.EQ_EQ) || matches4(previous, TokenType.BANG_EQ)) && _currentToken.offset == previous.offset + 2) { |
+ reportError7(ParserErrorCode.INVALID_OPERATOR, ["${previous.lexeme}${_currentToken.lexeme}"]); |
advance(); |
} |
} |
@@ -4263,9 +4263,9 @@ class Parser { |
return new PrefixExpression.full(operator, parseUnaryExpression()); |
} |
if (identical(operator.type, TokenType.MINUS_MINUS)) { |
- int offset2 = operator.offset; |
- Token firstOperator = new Token(TokenType.MINUS, offset2); |
- Token secondOperator = new Token(TokenType.MINUS, offset2 + 1); |
+ int offset = operator.offset; |
+ Token firstOperator = new Token(TokenType.MINUS, offset); |
+ Token secondOperator = new Token(TokenType.MINUS, offset + 1); |
secondOperator.setNext(_currentToken); |
firstOperator.setNext(secondOperator); |
operator.previous.setNext(firstOperator); |
@@ -4480,15 +4480,15 @@ class Parser { |
*/ |
Token skipFinalConstVarOrType(Token startToken) { |
if (matches3(startToken, Keyword.FINAL) || matches3(startToken, Keyword.CONST)) { |
- Token next2 = startToken.next; |
- if (matchesIdentifier2(next2.next) || matches4(next2.next, TokenType.LT) || matches3(next2.next, Keyword.THIS)) { |
- return skipTypeName(next2); |
+ Token next = startToken.next; |
+ if (matchesIdentifier2(next.next) || matches4(next.next, TokenType.LT) || matches3(next.next, Keyword.THIS)) { |
+ return skipTypeName(next); |
} |
} else if (matches3(startToken, Keyword.VAR)) { |
return startToken.next; |
} else if (matchesIdentifier2(startToken)) { |
- Token next3 = startToken.next; |
- if (matchesIdentifier2(next3) || matches4(next3, TokenType.LT) || matches3(next3, Keyword.THIS) || (matches4(next3, TokenType.PERIOD) && matchesIdentifier2(next3.next) && (matchesIdentifier2(next3.next.next) || matches4(next3.next.next, TokenType.LT) || matches3(next3.next.next, Keyword.THIS)))) { |
+ Token next = startToken.next; |
+ if (matchesIdentifier2(next) || matches4(next, TokenType.LT) || matches3(next, Keyword.THIS) || (matches4(next, TokenType.PERIOD) && matchesIdentifier2(next.next) && (matchesIdentifier2(next.next.next) || matches4(next.next.next, TokenType.LT) || matches3(next.next.next, Keyword.THIS)))) { |
return skipReturnType(startToken); |
} |
} |
@@ -4530,20 +4530,20 @@ class Parser { |
if (!matches4(startToken, TokenType.OPEN_PAREN)) { |
return null; |
} |
- Token next2 = startToken.next; |
- if (matches4(next2, TokenType.CLOSE_PAREN)) { |
- return next2.next; |
+ Token next = startToken.next; |
+ if (matches4(next, TokenType.CLOSE_PAREN)) { |
+ return next.next; |
} |
- if (matchesAny(next2, [TokenType.AT, TokenType.OPEN_SQUARE_BRACKET, TokenType.OPEN_CURLY_BRACKET]) || matches3(next2, Keyword.VOID) || (matchesIdentifier2(next2) && (matchesAny(next2.next, [TokenType.COMMA, TokenType.CLOSE_PAREN])))) { |
+ if (matchesAny(next, [TokenType.AT, TokenType.OPEN_SQUARE_BRACKET, TokenType.OPEN_CURLY_BRACKET]) || matches3(next, Keyword.VOID) || (matchesIdentifier2(next) && (matchesAny(next.next, [TokenType.COMMA, TokenType.CLOSE_PAREN])))) { |
return skipPastMatchingToken(startToken); |
} |
- if (matchesIdentifier2(next2) && matches4(next2.next, TokenType.OPEN_PAREN)) { |
- Token afterParameters = skipFormalParameterList(next2.next); |
+ if (matchesIdentifier2(next) && matches4(next.next, TokenType.OPEN_PAREN)) { |
+ Token afterParameters = skipFormalParameterList(next.next); |
if (afterParameters != null && (matchesAny(afterParameters, [TokenType.COMMA, TokenType.CLOSE_PAREN]))) { |
return skipPastMatchingToken(startToken); |
} |
} |
- Token afterType = skipFinalConstVarOrType(next2); |
+ Token afterType = skipFinalConstVarOrType(next); |
if (afterType == null) { |
return null; |
} |
@@ -4648,20 +4648,20 @@ class Parser { |
*/ |
Token skipStringInterpolation(Token startToken) { |
Token token = startToken; |
- TokenType type2 = token.type; |
- while (identical(type2, TokenType.STRING_INTERPOLATION_EXPRESSION) || identical(type2, TokenType.STRING_INTERPOLATION_IDENTIFIER)) { |
- if (identical(type2, TokenType.STRING_INTERPOLATION_EXPRESSION)) { |
+ TokenType type = token.type; |
+ while (identical(type, TokenType.STRING_INTERPOLATION_EXPRESSION) || identical(type, TokenType.STRING_INTERPOLATION_IDENTIFIER)) { |
+ if (identical(type, TokenType.STRING_INTERPOLATION_EXPRESSION)) { |
token = token.next; |
- type2 = token.type; |
+ type = token.type; |
int bracketNestingLevel = 1; |
while (bracketNestingLevel > 0) { |
- if (identical(type2, TokenType.EOF)) { |
+ if (identical(type, TokenType.EOF)) { |
return null; |
- } else if (identical(type2, TokenType.OPEN_CURLY_BRACKET)) { |
+ } else if (identical(type, TokenType.OPEN_CURLY_BRACKET)) { |
bracketNestingLevel++; |
- } else if (identical(type2, TokenType.CLOSE_CURLY_BRACKET)) { |
+ } else if (identical(type, TokenType.CLOSE_CURLY_BRACKET)) { |
bracketNestingLevel--; |
- } else if (identical(type2, TokenType.STRING)) { |
+ } else if (identical(type, TokenType.STRING)) { |
token = skipStringLiteral(token); |
if (token == null) { |
return null; |
@@ -4669,10 +4669,10 @@ class Parser { |
} else { |
token = token.next; |
} |
- type2 = token.type; |
+ type = token.type; |
} |
token = token.next; |
- type2 = token.type; |
+ type = token.type; |
} else { |
token = token.next; |
if (token.type != TokenType.IDENTIFIER) { |
@@ -4680,10 +4680,10 @@ class Parser { |
} |
token = token.next; |
} |
- type2 = token.type; |
- if (identical(type2, TokenType.STRING)) { |
+ type = token.type; |
+ if (identical(type, TokenType.STRING)) { |
token = token.next; |
- type2 = token.type; |
+ type = token.type; |
} |
} |
return token; |
@@ -4707,8 +4707,8 @@ class Parser { |
Token token = startToken; |
while (token != null && matches4(token, TokenType.STRING)) { |
token = token.next; |
- TokenType type2 = token.type; |
- if (identical(type2, TokenType.STRING_INTERPOLATION_EXPRESSION) || identical(type2, TokenType.STRING_INTERPOLATION_IDENTIFIER)) { |
+ TokenType type = token.type; |
+ if (identical(type, TokenType.STRING_INTERPOLATION_EXPRESSION) || identical(type, TokenType.STRING_INTERPOLATION_IDENTIFIER)) { |
token = skipStringInterpolation(token); |
} |
} |
@@ -4800,36 +4800,36 @@ class Parser { |
return null; |
} |
int depth = 1; |
- Token next2 = startToken.next; |
+ Token next = startToken.next; |
while (depth > 0) { |
- if (matches4(next2, TokenType.EOF)) { |
+ if (matches4(next, TokenType.EOF)) { |
return null; |
- } else if (matches4(next2, TokenType.LT)) { |
+ } else if (matches4(next, TokenType.LT)) { |
depth++; |
- } else if (matches4(next2, TokenType.GT)) { |
+ } else if (matches4(next, TokenType.GT)) { |
depth--; |
- } else if (matches4(next2, TokenType.GT_EQ)) { |
+ } else if (matches4(next, TokenType.GT_EQ)) { |
if (depth == 1) { |
- Token fakeEquals = new Token(TokenType.EQ, next2.offset + 2); |
- fakeEquals.setNextWithoutSettingPrevious(next2.next); |
+ Token fakeEquals = new Token(TokenType.EQ, next.offset + 2); |
+ fakeEquals.setNextWithoutSettingPrevious(next.next); |
return fakeEquals; |
} |
depth--; |
- } else if (matches4(next2, TokenType.GT_GT)) { |
+ } else if (matches4(next, TokenType.GT_GT)) { |
depth -= 2; |
- } else if (matches4(next2, TokenType.GT_GT_EQ)) { |
+ } else if (matches4(next, TokenType.GT_GT_EQ)) { |
if (depth < 2) { |
return null; |
} else if (depth == 2) { |
- Token fakeEquals = new Token(TokenType.EQ, next2.offset + 2); |
- fakeEquals.setNextWithoutSettingPrevious(next2.next); |
+ Token fakeEquals = new Token(TokenType.EQ, next.offset + 2); |
+ fakeEquals.setNextWithoutSettingPrevious(next.next); |
return fakeEquals; |
} |
depth -= 2; |
} |
- next2 = next2.next; |
+ next = next.next; |
} |
- return next2; |
+ return next; |
} |
/** |
@@ -4846,10 +4846,10 @@ class Parser { |
builder.appendChar(currentChar); |
return index + 1; |
} |
- int length2 = lexeme.length; |
+ int length = lexeme.length; |
int currentIndex = index + 1; |
- if (currentIndex >= length2) { |
- return length2; |
+ if (currentIndex >= length) { |
+ return length; |
} |
currentChar = lexeme.codeUnitAt(currentIndex); |
if (currentChar == 0x6E) { |
@@ -4865,9 +4865,9 @@ class Parser { |
} else if (currentChar == 0x76) { |
builder.appendChar(0xB); |
} else if (currentChar == 0x78) { |
- if (currentIndex + 2 >= length2) { |
+ if (currentIndex + 2 >= length) { |
reportError7(ParserErrorCode.INVALID_HEX_ESCAPE, []); |
- return length2; |
+ return length; |
} |
int firstDigit = lexeme.codeUnitAt(currentIndex + 1); |
int secondDigit = lexeme.codeUnitAt(currentIndex + 2); |
@@ -4879,16 +4879,16 @@ class Parser { |
return currentIndex + 3; |
} else if (currentChar == 0x75) { |
currentIndex++; |
- if (currentIndex >= length2) { |
+ if (currentIndex >= length) { |
reportError7(ParserErrorCode.INVALID_UNICODE_ESCAPE, []); |
- return length2; |
+ return length; |
} |
currentChar = lexeme.codeUnitAt(currentIndex); |
if (currentChar == 0x7B) { |
currentIndex++; |
- if (currentIndex >= length2) { |
+ if (currentIndex >= length) { |
reportError7(ParserErrorCode.INVALID_UNICODE_ESCAPE, []); |
- return length2; |
+ return length; |
} |
currentChar = lexeme.codeUnitAt(currentIndex); |
int digitCount = 0; |
@@ -4897,7 +4897,7 @@ class Parser { |
if (!isHexDigit(currentChar)) { |
reportError7(ParserErrorCode.INVALID_UNICODE_ESCAPE, []); |
currentIndex++; |
- while (currentIndex < length2 && lexeme.codeUnitAt(currentIndex) != 0x7D) { |
+ while (currentIndex < length && lexeme.codeUnitAt(currentIndex) != 0x7D) { |
currentIndex++; |
} |
return currentIndex + 1; |
@@ -4905,9 +4905,9 @@ class Parser { |
digitCount++; |
value = (value << 4) + Character.digit(currentChar, 16); |
currentIndex++; |
- if (currentIndex >= length2) { |
+ if (currentIndex >= length) { |
reportError7(ParserErrorCode.INVALID_UNICODE_ESCAPE, []); |
- return length2; |
+ return length; |
} |
currentChar = lexeme.codeUnitAt(currentIndex); |
} |
@@ -4917,9 +4917,9 @@ class Parser { |
appendScalarValue(builder, lexeme.substring(index, currentIndex + 1), value, index, currentIndex); |
return currentIndex + 1; |
} else { |
- if (currentIndex + 3 >= length2) { |
+ if (currentIndex + 3 >= length) { |
reportError7(ParserErrorCode.INVALID_UNICODE_ESCAPE, []); |
- return length2; |
+ return length; |
} |
int firstDigit = currentChar; |
int secondDigit = lexeme.codeUnitAt(currentIndex + 1); |
@@ -4991,16 +4991,16 @@ class Parser { |
if (modifiers.varKeyword != null) { |
reportError8(ParserErrorCode.CONSTRUCTOR_WITH_RETURN_TYPE, modifiers.varKeyword, []); |
} |
- Token externalKeyword2 = modifiers.externalKeyword; |
- Token constKeyword2 = modifiers.constKeyword; |
- Token factoryKeyword2 = modifiers.factoryKeyword; |
- if (externalKeyword2 != null && constKeyword2 != null && constKeyword2.offset < externalKeyword2.offset) { |
- reportError8(ParserErrorCode.EXTERNAL_AFTER_CONST, externalKeyword2, []); |
+ Token externalKeyword = modifiers.externalKeyword; |
+ Token constKeyword = modifiers.constKeyword; |
+ Token factoryKeyword = modifiers.factoryKeyword; |
+ if (externalKeyword != null && constKeyword != null && constKeyword.offset < externalKeyword.offset) { |
+ reportError8(ParserErrorCode.EXTERNAL_AFTER_CONST, externalKeyword, []); |
} |
- if (externalKeyword2 != null && factoryKeyword2 != null && factoryKeyword2.offset < externalKeyword2.offset) { |
- reportError8(ParserErrorCode.EXTERNAL_AFTER_FACTORY, externalKeyword2, []); |
+ if (externalKeyword != null && factoryKeyword != null && factoryKeyword.offset < externalKeyword.offset) { |
+ reportError8(ParserErrorCode.EXTERNAL_AFTER_FACTORY, externalKeyword, []); |
} |
- return constKeyword2; |
+ return constKeyword; |
} |
/** |
@@ -5019,31 +5019,31 @@ class Parser { |
if (modifiers.factoryKeyword != null) { |
reportError8(ParserErrorCode.NON_CONSTRUCTOR_FACTORY, modifiers.factoryKeyword, []); |
} |
- Token staticKeyword2 = modifiers.staticKeyword; |
- Token constKeyword2 = modifiers.constKeyword; |
- Token finalKeyword2 = modifiers.finalKeyword; |
- Token varKeyword2 = modifiers.varKeyword; |
- if (constKeyword2 != null) { |
- if (finalKeyword2 != null) { |
- reportError8(ParserErrorCode.CONST_AND_FINAL, finalKeyword2, []); |
+ Token staticKeyword = modifiers.staticKeyword; |
+ Token constKeyword = modifiers.constKeyword; |
+ Token finalKeyword = modifiers.finalKeyword; |
+ Token varKeyword = modifiers.varKeyword; |
+ if (constKeyword != null) { |
+ if (finalKeyword != null) { |
+ reportError8(ParserErrorCode.CONST_AND_FINAL, finalKeyword, []); |
} |
- if (varKeyword2 != null) { |
- reportError8(ParserErrorCode.CONST_AND_VAR, varKeyword2, []); |
+ if (varKeyword != null) { |
+ reportError8(ParserErrorCode.CONST_AND_VAR, varKeyword, []); |
} |
- if (staticKeyword2 != null && constKeyword2.offset < staticKeyword2.offset) { |
- reportError8(ParserErrorCode.STATIC_AFTER_CONST, staticKeyword2, []); |
+ if (staticKeyword != null && constKeyword.offset < staticKeyword.offset) { |
+ reportError8(ParserErrorCode.STATIC_AFTER_CONST, staticKeyword, []); |
} |
- } else if (finalKeyword2 != null) { |
- if (varKeyword2 != null) { |
- reportError8(ParserErrorCode.FINAL_AND_VAR, varKeyword2, []); |
+ } else if (finalKeyword != null) { |
+ if (varKeyword != null) { |
+ reportError8(ParserErrorCode.FINAL_AND_VAR, varKeyword, []); |
} |
- if (staticKeyword2 != null && finalKeyword2.offset < staticKeyword2.offset) { |
- reportError8(ParserErrorCode.STATIC_AFTER_FINAL, staticKeyword2, []); |
+ if (staticKeyword != null && finalKeyword.offset < staticKeyword.offset) { |
+ reportError8(ParserErrorCode.STATIC_AFTER_FINAL, staticKeyword, []); |
} |
- } else if (varKeyword2 != null && staticKeyword2 != null && varKeyword2.offset < staticKeyword2.offset) { |
- reportError8(ParserErrorCode.STATIC_AFTER_VAR, staticKeyword2, []); |
+ } else if (varKeyword != null && staticKeyword != null && varKeyword.offset < staticKeyword.offset) { |
+ reportError8(ParserErrorCode.STATIC_AFTER_VAR, staticKeyword, []); |
} |
- return lexicallyFirst([constKeyword2, finalKeyword2, varKeyword2]); |
+ return lexicallyFirst([constKeyword, finalKeyword, varKeyword]); |
} |
/** |
@@ -5066,10 +5066,10 @@ class Parser { |
if (modifiers.varKeyword != null) { |
reportError8(ParserErrorCode.VAR_RETURN_TYPE, modifiers.varKeyword, []); |
} |
- Token externalKeyword2 = modifiers.externalKeyword; |
- Token staticKeyword2 = modifiers.staticKeyword; |
- if (externalKeyword2 != null && staticKeyword2 != null && staticKeyword2.offset < externalKeyword2.offset) { |
- reportError8(ParserErrorCode.EXTERNAL_AFTER_STATIC, externalKeyword2, []); |
+ Token externalKeyword = modifiers.externalKeyword; |
+ Token staticKeyword = modifiers.staticKeyword; |
+ if (externalKeyword != null && staticKeyword != null && staticKeyword.offset < externalKeyword.offset) { |
+ reportError8(ParserErrorCode.EXTERNAL_AFTER_STATIC, externalKeyword, []); |
} |
} |
@@ -5145,22 +5145,22 @@ class Parser { |
if (modifiers.externalKeyword != null) { |
reportError8(ParserErrorCode.EXTERNAL_FIELD, modifiers.externalKeyword, []); |
} |
- Token constKeyword2 = modifiers.constKeyword; |
- Token finalKeyword2 = modifiers.finalKeyword; |
- Token varKeyword2 = modifiers.varKeyword; |
- if (constKeyword2 != null) { |
- if (finalKeyword2 != null) { |
- reportError8(ParserErrorCode.CONST_AND_FINAL, finalKeyword2, []); |
+ Token constKeyword = modifiers.constKeyword; |
+ Token finalKeyword = modifiers.finalKeyword; |
+ Token varKeyword = modifiers.varKeyword; |
+ if (constKeyword != null) { |
+ if (finalKeyword != null) { |
+ reportError8(ParserErrorCode.CONST_AND_FINAL, finalKeyword, []); |
} |
- if (varKeyword2 != null) { |
- reportError8(ParserErrorCode.CONST_AND_VAR, varKeyword2, []); |
+ if (varKeyword != null) { |
+ reportError8(ParserErrorCode.CONST_AND_VAR, varKeyword, []); |
} |
- } else if (finalKeyword2 != null) { |
- if (varKeyword2 != null) { |
- reportError8(ParserErrorCode.FINAL_AND_VAR, varKeyword2, []); |
+ } else if (finalKeyword != null) { |
+ if (varKeyword != null) { |
+ reportError8(ParserErrorCode.FINAL_AND_VAR, varKeyword, []); |
} |
} |
- return lexicallyFirst([constKeyword2, finalKeyword2, varKeyword2]); |
+ return lexicallyFirst([constKeyword, finalKeyword, varKeyword]); |
} |
/** |
@@ -5548,12 +5548,12 @@ class ToFormattedSourceVisitor implements ASTVisitor<Object> { |
} |
Object visitCommentReference(CommentReference node) => null; |
Object visitCompilationUnit(CompilationUnit node) { |
- ScriptTag scriptTag2 = node.scriptTag; |
- NodeList<Directive> directives2 = node.directives; |
- visit(scriptTag2); |
- String prefix = scriptTag2 == null ? "" : " "; |
- visitList7(prefix, directives2, "\n"); |
- prefix = scriptTag2 == null && directives2.isEmpty ? "" : "\n\n"; |
+ ScriptTag scriptTag = node.scriptTag; |
+ NodeList<Directive> directives = node.directives; |
+ visit(scriptTag); |
+ String prefix = scriptTag == null ? "" : " "; |
+ visitList7(prefix, directives, "\n"); |
+ prefix = scriptTag == null && directives.isEmpty ? "" : "\n\n"; |
visitList7(prefix, node.declarations, "\n"); |
return null; |
} |
@@ -5685,10 +5685,10 @@ class ToFormattedSourceVisitor implements ASTVisitor<Object> { |
Object visitFormalParameterList(FormalParameterList node) { |
String groupEnd = null; |
_writer.print('('); |
- NodeList<FormalParameter> parameters2 = node.parameters; |
- int size2 = parameters2.length; |
- for (int i = 0; i < size2; i++) { |
- FormalParameter parameter = parameters2[i]; |
+ NodeList<FormalParameter> parameters = node.parameters; |
+ int size = parameters.length; |
+ for (int i = 0; i < size; i++) { |
+ FormalParameter parameter = parameters[i]; |
if (i > 0) { |
_writer.print(", "); |
} |
@@ -5710,10 +5710,10 @@ class ToFormattedSourceVisitor implements ASTVisitor<Object> { |
return null; |
} |
Object visitForStatement(ForStatement node) { |
- Expression initialization2 = node.initialization; |
+ Expression initialization = node.initialization; |
_writer.print("for ("); |
- if (initialization2 != null) { |
- visit(initialization2); |
+ if (initialization != null) { |
+ visit(initialization); |
} else { |
visit(node.variables); |
} |
@@ -5980,12 +5980,12 @@ class ToFormattedSourceVisitor implements ASTVisitor<Object> { |
return null; |
} |
Object visitReturnStatement(ReturnStatement node) { |
- Expression expression2 = node.expression; |
- if (expression2 == null) { |
+ Expression expression = node.expression; |
+ if (expression == null) { |
_writer.print("return;"); |
} else { |
_writer.print("return "); |
- expression2.accept(this); |
+ expression.accept(this); |
_writer.print(";"); |
} |
return null; |
@@ -6214,8 +6214,8 @@ class ToFormattedSourceVisitor implements ASTVisitor<Object> { |
*/ |
void visitList5(NodeList<ASTNode> nodes, String separator) { |
if (nodes != null) { |
- int size2 = nodes.length; |
- for (int i = 0; i < size2; i++) { |
+ int size = nodes.length; |
+ for (int i = 0; i < size; i++) { |
if ("\n" == separator) { |
_writer.print("\n"); |
indent(); |
@@ -6235,9 +6235,9 @@ class ToFormattedSourceVisitor implements ASTVisitor<Object> { |
*/ |
void visitList6(NodeList<ASTNode> nodes, String separator, String suffix) { |
if (nodes != null) { |
- int size2 = nodes.length; |
- if (size2 > 0) { |
- for (int i = 0; i < size2; i++) { |
+ int size = nodes.length; |
+ if (size > 0) { |
+ for (int i = 0; i < size; i++) { |
if (i > 0) { |
_writer.print(separator); |
} |
@@ -6256,10 +6256,10 @@ class ToFormattedSourceVisitor implements ASTVisitor<Object> { |
*/ |
void visitList7(String prefix, NodeList<ASTNode> nodes, String separator) { |
if (nodes != null) { |
- int size2 = nodes.length; |
- if (size2 > 0) { |
+ int size = nodes.length; |
+ if (size > 0) { |
_writer.print(prefix); |
- for (int i = 0; i < size2; i++) { |
+ for (int i = 0; i < size; i++) { |
if (i > 0) { |
_writer.print(separator); |
} |