Index: third_party/closure_linter/closure_linter/common/tokens_test.py |
diff --git a/third_party/closure_linter/closure_linter/common/tokens_test.py b/third_party/closure_linter/closure_linter/common/tokens_test.py |
deleted file mode 100644 |
index 01ec89d01bc357df897671ca03b67eac72e6dcb3..0000000000000000000000000000000000000000 |
--- a/third_party/closure_linter/closure_linter/common/tokens_test.py |
+++ /dev/null |
@@ -1,113 +0,0 @@ |
-#!/usr/bin/env python |
-# Copyright 2011 The Closure Linter Authors. All Rights Reserved. |
-# |
-# Licensed under the Apache License, Version 2.0 (the "License"); |
-# you may not use this file except in compliance with the License. |
-# You may obtain a copy of the License at |
-# |
-# http://www.apache.org/licenses/LICENSE-2.0 |
-# |
-# Unless required by applicable law or agreed to in writing, software |
-# distributed under the License is distributed on an "AS-IS" BASIS, |
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
-# See the License for the specific language governing permissions and |
-# limitations under the License. |
- |
- |
-__author__ = 'nnaze@google.com (Nathan Naze)' |
- |
-import unittest as googletest |
-from closure_linter.common import tokens |
- |
- |
-def _CreateDummyToken(): |
- return tokens.Token('foo', None, 1, 1) |
- |
- |
-def _CreateDummyTokens(count): |
- dummy_tokens = [] |
- for _ in xrange(count): |
- dummy_tokens.append(_CreateDummyToken()) |
- return dummy_tokens |
- |
- |
-def _SetTokensAsNeighbors(neighbor_tokens): |
- for i in xrange(len(neighbor_tokens)): |
- prev_index = i - 1 |
- next_index = i + 1 |
- |
- if prev_index >= 0: |
- neighbor_tokens[i].previous = neighbor_tokens[prev_index] |
- |
- if next_index < len(neighbor_tokens): |
- neighbor_tokens[i].next = neighbor_tokens[next_index] |
- |
- |
-class TokensTest(googletest.TestCase): |
- |
- def testIsFirstInLine(self): |
- |
- # First token in file (has no previous). |
- self.assertTrue(_CreateDummyToken().IsFirstInLine()) |
- |
- a, b = _CreateDummyTokens(2) |
- _SetTokensAsNeighbors([a, b]) |
- |
- # Tokens on same line |
- a.line_number = 30 |
- b.line_number = 30 |
- |
- self.assertFalse(b.IsFirstInLine()) |
- |
- # Tokens on different lines |
- b.line_number = 31 |
- self.assertTrue(b.IsFirstInLine()) |
- |
- def testIsLastInLine(self): |
- # Last token in file (has no next). |
- self.assertTrue(_CreateDummyToken().IsLastInLine()) |
- |
- a, b = _CreateDummyTokens(2) |
- _SetTokensAsNeighbors([a, b]) |
- |
- # Tokens on same line |
- a.line_number = 30 |
- b.line_number = 30 |
- self.assertFalse(a.IsLastInLine()) |
- |
- b.line_number = 31 |
- self.assertTrue(a.IsLastInLine()) |
- |
- def testIsType(self): |
- a = tokens.Token('foo', 'fakeType1', 1, 1) |
- self.assertTrue(a.IsType('fakeType1')) |
- self.assertFalse(a.IsType('fakeType2')) |
- |
- def testIsAnyType(self): |
- a = tokens.Token('foo', 'fakeType1', 1, 1) |
- self.assertTrue(a.IsAnyType(['fakeType1', 'fakeType2'])) |
- self.assertFalse(a.IsAnyType(['fakeType3', 'fakeType4'])) |
- |
- def testRepr(self): |
- a = tokens.Token('foo', 'fakeType1', 1, 1) |
- self.assertEquals('<Token: fakeType1, "foo", None, 1, None>', str(a)) |
- |
- def testIter(self): |
- dummy_tokens = _CreateDummyTokens(5) |
- _SetTokensAsNeighbors(dummy_tokens) |
- a, b, c, d, e = dummy_tokens |
- |
- i = iter(a) |
- self.assertListEqual([a, b, c, d, e], list(i)) |
- |
- def testReverseIter(self): |
- dummy_tokens = _CreateDummyTokens(5) |
- _SetTokensAsNeighbors(dummy_tokens) |
- a, b, c, d, e = dummy_tokens |
- |
- ri = reversed(e) |
- self.assertListEqual([e, d, c, b, a], list(ri)) |
- |
- |
-if __name__ == '__main__': |
- googletest.main() |