/external/chromium_org/third_party/closure_linter/closure_linter/ |
error_fixer_test.py | 35 start_token = testutil.TokenizeSourceAndRunEcmaPass(_TEST_SCRIPT) 36 second_token = start_token.next 37 self.error_fixer.HandleFile('test_file', start_token) 39 self.error_fixer._DeleteToken(start_token) 44 start_token = testutil.TokenizeSourceAndRunEcmaPass(_TEST_SCRIPT) 45 fourth_token = start_token.next.next.next 46 self.error_fixer.HandleFile('test_file', start_token) 48 self.error_fixer._DeleteTokens(start_token, 3)
|
javascriptstatetracker_test.py | 70 start_token = function.start_token 75 function.start_token.type) 77 self.assertEquals('function', start_token.string) 78 self.assertEquals(3, start_token.line_number) 79 self.assertEquals(0, start_token.start_index) 93 start_token = function.start_token 98 function.start_token.type) 100 self.assertEquals('function', start_token.string [all...] |
aliaspass_test.py | 31 def _GetTokenByLineAndString(start_token, string, line_number): 32 for token in start_token: 40 start_token = testutil.TokenizeSourceAndRunEcmaPass(_TEST_SCOPE_SCRIPT) 45 alias_pass.Process(start_token) 56 start_token = testutil.TokenizeSourceAndRunEcmaPass(_TEST_ALIAS_SCRIPT) 58 alias_pass.Process(start_token) 60 alias_token = _GetTokenByLineAndString(start_token, 'Event', 4) 63 my_class_token = _GetTokenByLineAndString(start_token, 'myClass', 9) 66 component_token = _GetTokenByLineAndString(start_token, 'Component', 17) 70 event_token = _GetTokenByLineAndString(start_token, 'Event.Something', 17 [all...] |
testutil.py | 57 start_token = TokenizeSource(source) 59 ecma_pass.Process(start_token) 60 return start_token 72 start_token = TokenizeSourceAndRunEcmaPass(source) 78 for token in start_token:
|
scopeutil_test.py | 31 def _FindContexts(start_token): 38 start_token: First token in the token stream. 47 for token in start_token: 81 start_token = testutil.TokenizeSourceAndRunEcmaPass(script) 83 start_token, ecmametadatapass.EcmaContext.VAR) 117 start_token = testutil.TokenizeSourceAndRunEcmaPass(source) 118 for context in _FindContexts(start_token): 159 self.start_token = testutil.TokenizeSourceAndRunEcmaPass(_TEST_SCRIPT) 163 for context in _FindContexts(self.start_token): 181 for context in _FindContexts(self.start_token) [all...] |
checker.py | 66 def Check(self, start_token, limited_doc_checks=False, is_html=False, 74 start_token: The first token in the token stream. 82 self._alias_pass.Process(start_token) 88 self._ExecutePass(start_token, self._DependencyPass, stop_token) 90 self._ExecutePass(start_token, self._LintPass, stop_token)
|
aliaspass.py | 88 def Process(self, start_token): 92 start_token: The first token in the stream. 95 if start_token is None: 99 self._CheckGoogScopeCalls(start_token) 103 context = start_token.metadata.context 107 def _CheckGoogScopeCalls(self, start_token): 115 scope_tokens = [t for t in start_token if IsScopeToken(t)] 190 token = scope_block.start_token
|
scopeutil.py | 58 last_code_token = parent.start_token.metadata.last_code 82 reversed(block_context.start_token)) 142 if context.start_token and context.end_token: 143 statement_tokens = tokenutil.GetTokenRange(context.start_token,
|
tokenutil.py | 99 def CustomSearch(start_token, func, end_func=None, distance=None, 104 start_token: The token to start searching from 118 token = start_token 148 def Search(start_token, token_types, distance=None, reverse=False): 152 start_token: The token to start searching from 164 return CustomSearch(start_token, lambda token: token.IsAnyType(token_types), 168 def SearchExcept(start_token, token_types, distance=None, reverse=False): 172 start_token: The token to start searching from 184 return CustomSearch(start_token, 189 def SearchUntil(start_token, token_types, end_types, distance=None [all...] |
javascriptstatetracker.py | 153 start_token = self._block_stack.pop() 154 if tokenutil.GoogScopeOrNoneFromStartBlock(start_token):
|
closurizednamespacesinfo.py | 368 start_token = tokenutil.GetIdentifierStart(token) 369 if start_token and start_token != token: 371 identifier = tokenutil.GetIdentifierForToken(start_token) 373 start_token = token 374 # If an alias is defined on the start_token, use it instead. 375 if (start_token and 376 start_token.metadata and 377 start_token.metadata.aliased_symbol and 378 not start_token.metadata.is_alias_definition) [all...] |
checkerbase.py | 129 def Check(self, start_token, limited_doc_checks=False, is_html=False, 134 start_token: First token in token stream. 142 self._ExecutePass(start_token, self._LintPass, stop_token=stop_token)
|
error_fixer.py | 241 start_token = token.attached_object.type_start_token 243 if start_token.type != Type.DOC_START_BRACE: 245 len(start_token.string) - len(start_token.string.lstrip())) 247 start_token = tokenutil.SplitToken(start_token, leading_space) 249 if token.attached_object.type_end_token == start_token.previous: 250 token.attached_object.type_end_token = start_token 252 new_token = Token('{', Type.DOC_START_BRACE, start_token.line, 253 start_token.line_number [all...] |
ecmametadatapass.py | 51 start_token: The token where this context starts. 118 def __init__(self, context_type, start_token, parent=None): 123 start_token: The token where this context starts. 128 start_token: The token where this context starts. 134 self.start_token = start_token 176 return tokenutil.Compare(context1.start_token, context2.start_token) 337 keyword_token = result.start_token.metadata.last_code 349 start_block_token = pre_keyword_token.metadata.context.start_token [all...] |
statetracker.py | 247 start_token: The token that starts the doc comment. 251 def __init__(self, start_token): 255 start_token: The first token in the doc comment. 258 self.start_token = start_token 358 for token in self.start_token: 563 def _GetNextPartialIdentifierToken(start_token): 571 start_token: The token to start searching after. 576 token = start_token.next 589 def _GetEndTokenAndContents(start_token) [all...] |
indentation.py | 155 start_token = self._PopTo(Type.START_BLOCK) 157 if start_token: 158 goog_scope = tokenutil.GoogScopeOrNoneFromStartBlock(start_token.token) 177 (start_token.line_number)), 452 start_token = token 455 if not token or token.line_number != start_token.line_number:
|
/external/chromium-trace/trace-viewer/third_party/closure_linter/closure_linter/ |
tokenutil.py | 98 def CustomSearch(start_token, func, end_func=None, distance=None, 103 start_token: The token to start searching from 117 token = start_token 147 def Search(start_token, token_types, distance=None, reverse=False): 151 start_token: The token to start searching from 163 return CustomSearch(start_token, lambda token: token.IsAnyType(token_types), 167 def SearchExcept(start_token, token_types, distance=None, reverse=False): 171 start_token: The token to start searching from 183 return CustomSearch(start_token, 188 def SearchUntil(start_token, token_types, end_types, distance=None [all...] |
error_fixer.py | 225 start_token = token.attached_object.type_start_token 227 if start_token.type != Type.DOC_START_BRACE: 229 len(start_token.string) - len(start_token.string.lstrip())) 231 start_token = tokenutil.SplitToken(start_token, leading_space) 233 if token.attached_object.type_end_token == start_token.previous: 234 token.attached_object.type_end_token = start_token 236 new_token = Token('{', Type.DOC_START_BRACE, start_token.line, 237 start_token.line_number [all...] |
ecmametadatapass.py | 51 start_token: The token where this context starts. 118 def __init__(self, type, start_token, parent): 123 start_token: The token where this context starts. 127 self.start_token = start_token 290 keyword_token = result.start_token.metadata.last_code 302 start_block_token = pre_keyword_token.metadata.context.start_token
|
indentation.py | 155 start_token = self._PopTo(Type.START_BLOCK) 157 if start_token: 158 goog_scope = self._GoogScopeOrNone(start_token.token) 177 (start_token.line_number)), 442 start_token = token 445 if not token or token.line_number != start_token.line_number:
|
statetracker.py | 221 start_token: The token that starts the doc comment. 225 def __init__(self, start_token): 229 start_token: The first token in the doc comment. 234 self.start_token = start_token 438 def _GetNextIdentifierToken(start_token): 448 start_token: The token to start searching after. 453 token = start_token.next 467 def _GetEndTokenAndContents(start_token): 474 start_token: The token immediately before the first content token [all...] |
/external/linux-tools-perf/perf-3.12.0/tools/perf/util/ |
parse-events.l | 93 int start_token; 95 start_token = parse_events_get_extra(yyscanner); 97 if (start_token == PE_START_TERMS) 99 else if (start_token == PE_START_EVENTS) 102 if (start_token) { 104 return start_token;
|
/external/antlr/antlr-3.4/runtime/Ruby/lib/antlr3/tree/ |
debug.rb | 96 def set_token_boundaries( tree, start_token, stop_token ) 97 super( tree, start_token, stop_token ) 98 return unless tree && start_token && stop_token 100 start_token.token_index, stop_token.token_index )
|
/external/chromium_org/storage/common/fileapi/ |
file_system_util.cc | 417 std::string start_token(":"); 418 start_token = start_token.append( 424 start_token = StringToUpperASCII(start_token); 426 size_t pos = filesystem_name_upper.find(start_token); 432 *filesystem_id = filesystem_name.substr(pos + start_token.length(),
|
/external/pdfium/core/src/fpdfapi/fpdf_parser/ |
fpdf_parser_utility.cpp | 199 FX_BOOL CPDF_SimpleParser::FindTagPair(FX_BSTR start_token, FX_BSTR end_token,
202 if (!start_token.IsEmpty()) {
203 if (!SkipWord(start_token)) {
|