/external/google-breakpad/src/testing/scripts/generator/cpp/ |
tokenize.py | 74 def __init__(self, token_type, name, start, end): 75 self.token_type = token_type 147 token_type = UNKNOWN 151 token_type = NAME 159 token_type = CONSTANT 162 token_type = CONSTANT 173 token_type = SYNTAX 183 token_type = SYNTAX 186 token_type = CONSTAN [all...] |
ast.py | 549 if parts[-1].token_type == tokenize.NAME: 579 if (type_name and type_name[-1].token_type == tokenize.NAME and 580 p.token_type == tokenize.NAME): 713 if token.token_type == _INTERNAL_TOKEN: 738 if token.token_type == tokenize.NAME: 749 if next.token_type == tokenize.SYNTAX and next.name == '(': 805 elif token.token_type == tokenize.SYNTAX: 811 if (token.token_type == tokenize.NAME and 815 elif token.token_type == tokenize.PREPROCESSOR: 854 while (last_token.token_type != expected_token_type o [all...] |
/external/googletest/googlemock/scripts/generator/cpp/ |
tokenize.py | 74 def __init__(self, token_type, name, start, end): 75 self.token_type = token_type 147 token_type = UNKNOWN 151 token_type = NAME 159 token_type = CONSTANT 162 token_type = CONSTANT 173 token_type = SYNTAX 183 token_type = SYNTAX 186 token_type = CONSTAN [all...] |
ast.py | 551 if parts[-1].token_type == tokenize.NAME: 581 if (type_name and type_name[-1].token_type == tokenize.NAME and 582 p.token_type == tokenize.NAME): 714 if token.token_type == _INTERNAL_TOKEN: 739 if token.token_type == tokenize.NAME: 750 if next.token_type == tokenize.SYNTAX and next.name == '(': 806 elif token.token_type == tokenize.SYNTAX: 812 if (token.token_type == tokenize.NAME and 816 elif token.token_type == tokenize.PREPROCESSOR: 855 while (last_token.token_type != expected_token_type o [all...] |
/external/v8/testing/gmock/scripts/generator/cpp/ |
tokenize.py | 74 def __init__(self, token_type, name, start, end): 75 self.token_type = token_type 147 token_type = UNKNOWN 151 token_type = NAME 159 token_type = CONSTANT 162 token_type = CONSTANT 173 token_type = SYNTAX 183 token_type = SYNTAX 186 token_type = CONSTAN [all...] |
ast.py | 551 if parts[-1].token_type == tokenize.NAME: 581 if (type_name and type_name[-1].token_type == tokenize.NAME and 582 p.token_type == tokenize.NAME): 714 if token.token_type == _INTERNAL_TOKEN: 739 if token.token_type == tokenize.NAME: 750 if next.token_type == tokenize.SYNTAX and next.name == '(': 806 elif token.token_type == tokenize.SYNTAX: 812 if (token.token_type == tokenize.NAME and 816 elif token.token_type == tokenize.PREPROCESSOR: 855 while (last_token.token_type != expected_token_type o [all...] |
/external/chromium-trace/catapult/common/py_utils/py_utils/refactor/ |
offset_token.py | 26 def __init__(self, token_type, string, offset): 27 self._type = token_type 70 token_type, string, (srow, scol), _, _ = next_token 72 offset_tokens.append(OffsetToken(token_type, string, (0, 0))) 76 offset_tokens.append(OffsetToken(token_type, string, (0, scol - ecol))) 79 token_type, string, (srow - erow, scol)))
|
snippet.py | 97 def __init__(self, token_type, tokens): 102 self._type = token_type 107 def Create(cls, token_type, string, offset=(0, 0)): 108 return cls(token_type, 109 [offset_token.OffsetToken(token_type, string, offset)])
|
/frameworks/compile/mclinker/include/mcld/Script/ |
ScriptScanner.h | 22 mcld::ScriptParser::token_type mcld::ScriptScanner::lex( \ 43 virtual ScriptParser::token_type lex(ScriptParser::semantic_type* yylval,
|
/frameworks/compile/mclinker/lib/Script/ |
ScriptScanner.ll | 20 typedef mcld::ScriptParser::token_type token_type; 184 <LDSCRIPT,EXPRESSION>"," { return static_cast<token_type>(*yytext); } 185 <LDSCRIPT,EXPRESSION>"=" { return static_cast<token_type>(*yytext); } 186 <LDSCRIPT,EXPRESSION>"?" { return static_cast<token_type>(*yytext); } 187 <LDSCRIPT,EXPRESSION>":" { return static_cast<token_type>(*yytext); } 188 <LDSCRIPT,EXPRESSION>"|" { return static_cast<token_type>(*yytext); } 189 <LDSCRIPT,EXPRESSION>"^" { return static_cast<token_type>(*yytext); } 190 <LDSCRIPT,EXPRESSION>"&" { return static_cast<token_type>(*yytext); } 191 <LDSCRIPT,EXPRESSION>"<" { return static_cast<token_type>(*yytext); [all...] |
/external/protobuf/gtest/scripts/ |
pump.py | 141 def __init__(self, start=None, end=None, value=None, token_type=None): 151 self.token_type = token_type 155 self.start, self.value, self.token_type) 161 self.token_type) 172 for (regex, token_type) in token_table: 179 best_match_token_type = token_type 198 (start_column, length, token_type) = m 203 return MakeToken(lines, found_start, found_end, token_type) 227 def MakeToken(lines, start, end, token_type) [all...] |
/external/google-breakpad/src/testing/gtest/scripts/ |
pump.py | 140 def __init__(self, start=None, end=None, value=None, token_type=None): 150 self.token_type = token_type 154 self.start, self.value, self.token_type) 160 self.token_type) 171 for (regex, token_type) in token_table: 178 best_match_token_type = token_type 197 (start_column, length, token_type) = m 202 return MakeToken(lines, found_start, found_end, token_type) 237 def MakeToken(lines, start, end, token_type) [all...] |
/external/googletest/googletest/scripts/ |
pump.py | 140 def __init__(self, start=None, end=None, value=None, token_type=None): 150 self.token_type = token_type 154 self.start, self.value, self.token_type) 160 self.token_type) 171 for (regex, token_type) in token_table: 178 best_match_token_type = token_type 197 (start_column, length, token_type) = m 202 return MakeToken(lines, found_start, found_end, token_type) 237 def MakeToken(lines, start, end, token_type) [all...] |
/external/v8/testing/gtest/scripts/ |
pump.py | 140 def __init__(self, start=None, end=None, value=None, token_type=None): 150 self.token_type = token_type 154 self.start, self.value, self.token_type) 160 self.token_type) 171 for (regex, token_type) in token_table: 178 best_match_token_type = token_type 197 (start_column, length, token_type) = m 202 return MakeToken(lines, found_start, found_end, token_type) 237 def MakeToken(lines, start, end, token_type) [all...] |
/external/vulkan-validation-layers/tests/gtest-1.7.0/scripts/ |
pump.py | 140 def __init__(self, start=None, end=None, value=None, token_type=None): 150 self.token_type = token_type 154 self.start, self.value, self.token_type) 160 self.token_type) 171 for (regex, token_type) in token_table: 178 best_match_token_type = token_type 197 (start_column, length, token_type) = m 202 return MakeToken(lines, found_start, found_end, token_type) 237 def MakeToken(lines, start, end, token_type) [all...] |
/external/autotest/client/common_lib/cros/ |
dbus_send.py | 47 token_type = token_stream.pop(0) 48 if token_type == 'variant': 49 token_type = token_stream.pop(0) 50 if token_type == 'object': 51 token_type = token_stream.pop(0) # Should be 'path' 55 if token_type in INT_TYPES: 57 if token_type == 'string' or token_type == 'path': 59 if token_type == 'boolean': 61 if token_type == 'double' [all...] |
/external/bison/examples/calc++/ |
calc++-driver.hh | 10 yy::calcxx_parser::token_type \
|
calc++-parser.hh | 106 typedef token::yytokentype token_type; typedef in class:yy::calcxx_parser
|
/external/antlr/antlr-3.4/runtime/Ruby/lib/antlr3/tree/ |
wizard.rb | 197 @token_type = tokenizer.next_token 201 case @token_type 205 @token_type == EOF and return node 214 @token_type != :open and return nil 215 @token_type = @tokenizer.next_token 219 case @token_type 229 @token_type == :close or return nil 230 @token_type = @tokenizer.next_token 236 if @token_type == :percent 237 ( @token_type = @tokenizer.next_token ) == :identifier or return ni [all...] |
debug.rb | 27 def create_from_token( token_type, from_token, text = nil ) 33 def create_from_type( token_type, text )
|
/external/v8/src/parsing/ |
token.cc | 45 const char Token::token_type[] = { member in class:v8::internal::Token
|
/device/linaro/bootloader/edk2/BaseTools/Source/C/VfrCompile/Pccts/h/ |
AParser.h | 55 #define NLA (token_type[lap&(LLk-1)])/* --> next LA */
176 ANTLRTokenType *token_type; // fast reference cache of token.getType()
member in class:ANTLRParser 237 //// return demand_look ? token_type[(labase+(i)-1)&(LLk-1)] :
238 //// token_type[(lap+(i)-1)&(LLk-1)];
245 return token_type[(lap+(i)-1)&(LLk-1)];
|
/device/linaro/bootloader/edk2/EdkCompatibilityPkg/Other/Maintained/Tools/Pccts/h/ |
AParser.h | 55 #define NLA (token_type[lap&(LLk-1)])/* --> next LA */
176 ANTLRTokenType *token_type; // fast reference cache of token.getType()
member in class:ANTLRParser 236 //// return demand_look ? token_type[(labase+(i)-1)&(LLk-1)] :
237 //// token_type[(lap+(i)-1)&(LLk-1)];
244 return token_type[(lap+(i)-1)&(LLk-1)];
|
/external/libmojo/mojo/public/tools/bindings/pylib/mojom_tests/parse/ |
lexer_unittest.py | 42 def _MakeLexToken(token_type, value, lineno=1, lexpos=0): 46 rv.type, rv.value, rv.lineno, rv.lexpos = token_type, value, lineno, lexpos
|
/external/libmojo/third_party/jinja2/ |
parser.py | 431 token_type = self.stream.current.type 432 if token_type in _compare_operators: 434 ops.append(nodes.Operand(token_type, self.parse_add())) 529 token_type = self.stream.current.type 531 if token_type == 'sub': 534 elif token_type == 'add': 662 token_type = self.stream.current.type 663 if token_type == 'dot' or token_type == 'lbracket': 667 elif token_type == 'lparen' [all...] |