Home | History | Annotate | Download | only in common
      1 #!/usr/bin/env python
      2 # Copyright 2011 The Closure Linter Authors. All Rights Reserved.
      3 #
      4 # Licensed under the Apache License, Version 2.0 (the "License");
      5 # you may not use this file except in compliance with the License.
      6 # You may obtain a copy of the License at
      7 #
      8 #      http://www.apache.org/licenses/LICENSE-2.0
      9 #
     10 # Unless required by applicable law or agreed to in writing, software
     11 # distributed under the License is distributed on an "AS-IS" BASIS,
     12 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
     13 # See the License for the specific language governing permissions and
     14 # limitations under the License.
     15 
     16 
     17 
     18 
     19 import unittest as googletest
     20 from closure_linter.common import tokens
     21 
     22 
     23 def _CreateDummyToken():
     24   return tokens.Token('foo', None, 1, 1)
     25 
     26 
     27 def _CreateDummyTokens(count):
     28   dummy_tokens = []
     29   for _ in xrange(count):
     30     dummy_tokens.append(_CreateDummyToken())
     31   return dummy_tokens
     32 
     33 
     34 def _SetTokensAsNeighbors(neighbor_tokens):
     35   for i in xrange(len(neighbor_tokens)):
     36     prev_index = i - 1
     37     next_index = i + 1
     38 
     39     if prev_index >= 0:
     40       neighbor_tokens[i].previous = neighbor_tokens[prev_index]
     41 
     42     if next_index < len(neighbor_tokens):
     43       neighbor_tokens[i].next = neighbor_tokens[next_index]
     44 
     45 
     46 class TokensTest(googletest.TestCase):
     47 
     48   def testIsFirstInLine(self):
     49 
     50     # First token in file (has no previous).
     51     self.assertTrue(_CreateDummyToken().IsFirstInLine())
     52 
     53     a, b = _CreateDummyTokens(2)
     54     _SetTokensAsNeighbors([a, b])
     55 
     56     # Tokens on same line
     57     a.line_number = 30
     58     b.line_number = 30
     59 
     60     self.assertFalse(b.IsFirstInLine())
     61 
     62     # Tokens on different lines
     63     b.line_number = 31
     64     self.assertTrue(b.IsFirstInLine())
     65 
     66   def testIsLastInLine(self):
     67     # Last token in file (has no next).
     68     self.assertTrue(_CreateDummyToken().IsLastInLine())
     69 
     70     a, b = _CreateDummyTokens(2)
     71     _SetTokensAsNeighbors([a, b])
     72 
     73     # Tokens on same line
     74     a.line_number = 30
     75     b.line_number = 30
     76     self.assertFalse(a.IsLastInLine())
     77 
     78     b.line_number = 31
     79     self.assertTrue(a.IsLastInLine())
     80 
     81   def testIsType(self):
     82     a = tokens.Token('foo', 'fakeType1', 1, 1)
     83     self.assertTrue(a.IsType('fakeType1'))
     84     self.assertFalse(a.IsType('fakeType2'))
     85 
     86   def testIsAnyType(self):
     87     a = tokens.Token('foo', 'fakeType1', 1, 1)
     88     self.assertTrue(a.IsAnyType(['fakeType1', 'fakeType2']))
     89     self.assertFalse(a.IsAnyType(['fakeType3', 'fakeType4']))
     90 
     91   def testRepr(self):
     92     a = tokens.Token('foo', 'fakeType1', 1, 1)
     93     self.assertEquals('<Token: fakeType1, "foo", None, 1, None>', str(a))
     94 
     95   def testIter(self):
     96     dummy_tokens = _CreateDummyTokens(5)
     97     _SetTokensAsNeighbors(dummy_tokens)
     98     a, b, c, d, e = dummy_tokens
     99 
    100     i = iter(a)
    101     self.assertListEqual([a, b, c, d, e], list(i))
    102 
    103   def testReverseIter(self):
    104     dummy_tokens = _CreateDummyTokens(5)
    105     _SetTokensAsNeighbors(dummy_tokens)
    106     a, b, c, d, e = dummy_tokens
    107 
    108     ri = reversed(e)
    109     self.assertListEqual([e, d, c, b, a], list(ri))
    110 
    111 
    112 if __name__ == '__main__':
    113   googletest.main()
    114