1 // Copyright (c) 2013 The Chromium Authors. All rights reserved. 2 // Use of this source code is governed by a BSD-style license that can be 3 // found in the LICENSE file. 4 5 #include "testing/gtest/include/gtest/gtest.h" 6 #include "tools/gn/input_file.h" 7 #include "tools/gn/token.h" 8 #include "tools/gn/tokenizer.h" 9 10 namespace { 11 12 struct TokenExpectation { 13 Token::Type type; 14 const char* value; 15 }; 16 17 template<size_t len> 18 bool CheckTokenizer(const char* input, const TokenExpectation (&expect)[len]) { 19 InputFile input_file(SourceFile("/test")); 20 input_file.SetContents(input); 21 22 Err err; 23 std::vector<Token> results = Tokenizer::Tokenize(&input_file, &err); 24 25 if (results.size() != len) 26 return false; 27 for (size_t i = 0; i < len; i++) { 28 if (expect[i].type != results[i].type()) 29 return false; 30 if (expect[i].value != results[i].value()) 31 return false; 32 } 33 return true; 34 } 35 36 } // namespace 37 38 TEST(Tokenizer, Empty) { 39 InputFile empty_string_input(SourceFile("/test")); 40 empty_string_input.SetContents(""); 41 42 Err err; 43 std::vector<Token> results = Tokenizer::Tokenize(&empty_string_input, &err); 44 EXPECT_TRUE(results.empty()); 45 46 InputFile whitespace_input(SourceFile("/test")); 47 whitespace_input.SetContents(" \n\r"); 48 49 results = Tokenizer::Tokenize(&whitespace_input, &err); 50 EXPECT_TRUE(results.empty()); 51 } 52 53 TEST(Tokenizer, Identifier) { 54 TokenExpectation one_ident[] = { 55 { Token::IDENTIFIER, "foo" } 56 }; 57 EXPECT_TRUE(CheckTokenizer(" foo ", one_ident)); 58 } 59 60 TEST(Tokenizer, Integer) { 61 TokenExpectation integers[] = { 62 { Token::INTEGER, "123" }, 63 { Token::INTEGER, "-123" } 64 }; 65 EXPECT_TRUE(CheckTokenizer(" 123 -123 ", integers)); 66 } 67 68 TEST(Tokenizer, String) { 69 TokenExpectation strings[] = { 70 { Token::STRING, "\"foo\"" }, 71 { Token::STRING, "\"bar\\\"baz\"" }, 72 { Token::STRING, "\"asdf\\\\\"" } 73 }; 74 EXPECT_TRUE(CheckTokenizer(" \"foo\" \"bar\\\"baz\" \"asdf\\\\\" ", 75 strings)); 76 } 77 78 TEST(Tokenizer, Operator) { 79 TokenExpectation operators[] = { 80 { Token::OPERATOR, "-" }, 81 { Token::OPERATOR, "+" }, 82 { Token::OPERATOR, "=" }, 83 { Token::OPERATOR, "+=" }, 84 { Token::OPERATOR, "-=" }, 85 { Token::OPERATOR, "!=" }, 86 { Token::OPERATOR, "==" }, 87 { Token::OPERATOR, "<" }, 88 { Token::OPERATOR, ">" }, 89 { Token::OPERATOR, "<=" }, 90 { Token::OPERATOR, ">=" }, 91 }; 92 EXPECT_TRUE(CheckTokenizer("- + = += -= != == < > <= >=", 93 operators)); 94 } 95 96 TEST(Tokenizer, Scoper) { 97 TokenExpectation scopers[] = { 98 { Token::SCOPER, "{" }, 99 { Token::SCOPER, "[" }, 100 { Token::SCOPER, "]" }, 101 { Token::SCOPER, "}" }, 102 { Token::SCOPER, "(" }, 103 { Token::SCOPER, ")" }, 104 }; 105 EXPECT_TRUE(CheckTokenizer("{[ ]} ()", scopers)); 106 } 107 108 TEST(Tokenizer, FunctionCall) { 109 TokenExpectation fn[] = { 110 { Token::IDENTIFIER, "fun" }, 111 { Token::SCOPER, "(" }, 112 { Token::STRING, "\"foo\"" }, 113 { Token::SCOPER, ")" }, 114 { Token::SCOPER, "{" }, 115 { Token::IDENTIFIER, "foo" }, 116 { Token::OPERATOR, "=" }, 117 { Token::INTEGER, "12" }, 118 { Token::SCOPER, "}" }, 119 }; 120 EXPECT_TRUE(CheckTokenizer("fun(\"foo\") {\nfoo = 12}", fn)); 121 } 122 123 TEST(Tokenizer, StringUnescaping) { 124 InputFile input(SourceFile("/test")); 125 input.SetContents("\"asd\\\"f\" \"\""); 126 Err err; 127 std::vector<Token> results = Tokenizer::Tokenize(&input, &err); 128 129 ASSERT_EQ(2u, results.size()); 130 EXPECT_EQ("asd\"f", results[0].StringValue()); 131 EXPECT_EQ("", results[1].StringValue()); 132 } 133 134 TEST(Tokenizer, Locations) { 135 InputFile input(SourceFile("/test")); 136 input.SetContents("1 2 \"three\"\n 4"); 137 Err err; 138 std::vector<Token> results = Tokenizer::Tokenize(&input, &err); 139 140 ASSERT_EQ(4u, results.size()); 141 ASSERT_TRUE(results[0].location() == Location(&input, 1, 1)); 142 ASSERT_TRUE(results[1].location() == Location(&input, 1, 3)); 143 ASSERT_TRUE(results[2].location() == Location(&input, 1, 5)); 144 ASSERT_TRUE(results[3].location() == Location(&input, 2, 3)); 145 } 146 147 TEST(Tokenizer, ByteOffsetOfNthLine) { 148 EXPECT_EQ(0u, Tokenizer::ByteOffsetOfNthLine("foo", 1)); 149 150 // Windows and Posix have different line endings, so check the byte at the 151 // location rather than the offset. 152 char input1[] = "aaa\nxaa\n\nya"; 153 EXPECT_EQ('x', input1[Tokenizer::ByteOffsetOfNthLine(input1, 2)]); 154 EXPECT_EQ('y', input1[Tokenizer::ByteOffsetOfNthLine(input1, 4)]); 155 156 char input2[3]; 157 input2[0] = 'a'; 158 input2[1] = '\n'; // Manually set to avoid Windows double-byte endings. 159 input2[2] = 0; 160 EXPECT_EQ(0u, Tokenizer::ByteOffsetOfNthLine(input2, 1)); 161 EXPECT_EQ(2u, Tokenizer::ByteOffsetOfNthLine(input2, 2)); 162 } 163