HomeSort by relevance Sort by last modified time
    Searched refs:start_token (Results 1 - 25 of 36) sorted by null

1 2

  /external/chromium_org/third_party/closure_linter/closure_linter/
error_fixer_test.py 35 start_token = testutil.TokenizeSourceAndRunEcmaPass(_TEST_SCRIPT)
36 second_token = start_token.next
37 self.error_fixer.HandleFile('test_file', start_token)
39 self.error_fixer._DeleteToken(start_token)
44 start_token = testutil.TokenizeSourceAndRunEcmaPass(_TEST_SCRIPT)
45 fourth_token = start_token.next.next.next
46 self.error_fixer.HandleFile('test_file', start_token)
48 self.error_fixer._DeleteTokens(start_token, 3)
javascriptstatetracker_test.py 70 start_token = function.start_token
75 function.start_token.type)
77 self.assertEquals('function', start_token.string)
78 self.assertEquals(3, start_token.line_number)
79 self.assertEquals(0, start_token.start_index)
93 start_token = function.start_token
98 function.start_token.type)
100 self.assertEquals('function', start_token.string
    [all...]
aliaspass_test.py 31 def _GetTokenByLineAndString(start_token, string, line_number):
32 for token in start_token:
40 start_token = testutil.TokenizeSourceAndRunEcmaPass(_TEST_SCOPE_SCRIPT)
45 alias_pass.Process(start_token)
56 start_token = testutil.TokenizeSourceAndRunEcmaPass(_TEST_ALIAS_SCRIPT)
58 alias_pass.Process(start_token)
60 alias_token = _GetTokenByLineAndString(start_token, 'Event', 4)
63 my_class_token = _GetTokenByLineAndString(start_token, 'myClass', 9)
66 component_token = _GetTokenByLineAndString(start_token, 'Component', 17)
70 event_token = _GetTokenByLineAndString(start_token, 'Event.Something', 17
    [all...]
testutil.py 57 start_token = TokenizeSource(source)
59 ecma_pass.Process(start_token)
60 return start_token
72 start_token = TokenizeSourceAndRunEcmaPass(source)
78 for token in start_token:
scopeutil_test.py 31 def _FindContexts(start_token):
38 start_token: First token in the token stream.
47 for token in start_token:
81 start_token = testutil.TokenizeSourceAndRunEcmaPass(script)
83 start_token, ecmametadatapass.EcmaContext.VAR)
117 start_token = testutil.TokenizeSourceAndRunEcmaPass(source)
118 for context in _FindContexts(start_token):
159 self.start_token = testutil.TokenizeSourceAndRunEcmaPass(_TEST_SCRIPT)
163 for context in _FindContexts(self.start_token):
181 for context in _FindContexts(self.start_token)
    [all...]
checker.py 66 def Check(self, start_token, limited_doc_checks=False, is_html=False,
74 start_token: The first token in the token stream.
82 self._alias_pass.Process(start_token)
88 self._ExecutePass(start_token, self._DependencyPass, stop_token)
90 self._ExecutePass(start_token, self._LintPass, stop_token)
aliaspass.py 88 def Process(self, start_token):
92 start_token: The first token in the stream.
95 if start_token is None:
99 self._CheckGoogScopeCalls(start_token)
103 context = start_token.metadata.context
107 def _CheckGoogScopeCalls(self, start_token):
115 scope_tokens = [t for t in start_token if IsScopeToken(t)]
190 token = scope_block.start_token
scopeutil.py 58 last_code_token = parent.start_token.metadata.last_code
82 reversed(block_context.start_token))
142 if context.start_token and context.end_token:
143 statement_tokens = tokenutil.GetTokenRange(context.start_token,
tokenutil.py 99 def CustomSearch(start_token, func, end_func=None, distance=None,
104 start_token: The token to start searching from
118 token = start_token
148 def Search(start_token, token_types, distance=None, reverse=False):
152 start_token: The token to start searching from
164 return CustomSearch(start_token, lambda token: token.IsAnyType(token_types),
168 def SearchExcept(start_token, token_types, distance=None, reverse=False):
172 start_token: The token to start searching from
184 return CustomSearch(start_token,
189 def SearchUntil(start_token, token_types, end_types, distance=None
    [all...]
javascriptstatetracker.py 153 start_token = self._block_stack.pop()
154 if tokenutil.GoogScopeOrNoneFromStartBlock(start_token):
closurizednamespacesinfo.py 368 start_token = tokenutil.GetIdentifierStart(token)
369 if start_token and start_token != token:
371 identifier = tokenutil.GetIdentifierForToken(start_token)
373 start_token = token
374 # If an alias is defined on the start_token, use it instead.
375 if (start_token and
376 start_token.metadata and
377 start_token.metadata.aliased_symbol and
378 not start_token.metadata.is_alias_definition)
    [all...]
checkerbase.py 129 def Check(self, start_token, limited_doc_checks=False, is_html=False,
134 start_token: First token in token stream.
142 self._ExecutePass(start_token, self._LintPass, stop_token=stop_token)
error_fixer.py 241 start_token = token.attached_object.type_start_token
243 if start_token.type != Type.DOC_START_BRACE:
245 len(start_token.string) - len(start_token.string.lstrip()))
247 start_token = tokenutil.SplitToken(start_token, leading_space)
249 if token.attached_object.type_end_token == start_token.previous:
250 token.attached_object.type_end_token = start_token
252 new_token = Token('{', Type.DOC_START_BRACE, start_token.line,
253 start_token.line_number
    [all...]
ecmametadatapass.py 51 start_token: The token where this context starts.
118 def __init__(self, context_type, start_token, parent=None):
123 start_token: The token where this context starts.
128 start_token: The token where this context starts.
134 self.start_token = start_token
176 return tokenutil.Compare(context1.start_token, context2.start_token)
337 keyword_token = result.start_token.metadata.last_code
349 start_block_token = pre_keyword_token.metadata.context.start_token
    [all...]
statetracker.py 247 start_token: The token that starts the doc comment.
251 def __init__(self, start_token):
255 start_token: The first token in the doc comment.
258 self.start_token = start_token
358 for token in self.start_token:
563 def _GetNextPartialIdentifierToken(start_token):
571 start_token: The token to start searching after.
576 token = start_token.next
589 def _GetEndTokenAndContents(start_token)
    [all...]
indentation.py 155 start_token = self._PopTo(Type.START_BLOCK)
157 if start_token:
158 goog_scope = tokenutil.GoogScopeOrNoneFromStartBlock(start_token.token)
177 (start_token.line_number)),
452 start_token = token
455 if not token or token.line_number != start_token.line_number:
  /external/chromium-trace/trace-viewer/third_party/closure_linter/closure_linter/
tokenutil.py 98 def CustomSearch(start_token, func, end_func=None, distance=None,
103 start_token: The token to start searching from
117 token = start_token
147 def Search(start_token, token_types, distance=None, reverse=False):
151 start_token: The token to start searching from
163 return CustomSearch(start_token, lambda token: token.IsAnyType(token_types),
167 def SearchExcept(start_token, token_types, distance=None, reverse=False):
171 start_token: The token to start searching from
183 return CustomSearch(start_token,
188 def SearchUntil(start_token, token_types, end_types, distance=None
    [all...]
error_fixer.py 225 start_token = token.attached_object.type_start_token
227 if start_token.type != Type.DOC_START_BRACE:
229 len(start_token.string) - len(start_token.string.lstrip()))
231 start_token = tokenutil.SplitToken(start_token, leading_space)
233 if token.attached_object.type_end_token == start_token.previous:
234 token.attached_object.type_end_token = start_token
236 new_token = Token('{', Type.DOC_START_BRACE, start_token.line,
237 start_token.line_number
    [all...]
ecmametadatapass.py 51 start_token: The token where this context starts.
118 def __init__(self, type, start_token, parent):
123 start_token: The token where this context starts.
127 self.start_token = start_token
290 keyword_token = result.start_token.metadata.last_code
302 start_block_token = pre_keyword_token.metadata.context.start_token
indentation.py 155 start_token = self._PopTo(Type.START_BLOCK)
157 if start_token:
158 goog_scope = self._GoogScopeOrNone(start_token.token)
177 (start_token.line_number)),
442 start_token = token
445 if not token or token.line_number != start_token.line_number:
statetracker.py 221 start_token: The token that starts the doc comment.
225 def __init__(self, start_token):
229 start_token: The first token in the doc comment.
234 self.start_token = start_token
438 def _GetNextIdentifierToken(start_token):
448 start_token: The token to start searching after.
453 token = start_token.next
467 def _GetEndTokenAndContents(start_token):
474 start_token: The token immediately before the first content token
    [all...]
  /external/linux-tools-perf/perf-3.12.0/tools/perf/util/
parse-events.l 93 int start_token;
95 start_token = parse_events_get_extra(yyscanner);
97 if (start_token == PE_START_TERMS)
99 else if (start_token == PE_START_EVENTS)
102 if (start_token) {
104 return start_token;
  /external/antlr/antlr-3.4/runtime/Ruby/lib/antlr3/tree/
debug.rb 96 def set_token_boundaries( tree, start_token, stop_token )
97 super( tree, start_token, stop_token )
98 return unless tree && start_token && stop_token
100 start_token.token_index, stop_token.token_index )
  /external/chromium_org/storage/common/fileapi/
file_system_util.cc 417 std::string start_token(":");
418 start_token = start_token.append(
424 start_token = StringToUpperASCII(start_token);
426 size_t pos = filesystem_name_upper.find(start_token);
432 *filesystem_id = filesystem_name.substr(pos + start_token.length(),
  /external/pdfium/core/src/fpdfapi/fpdf_parser/
fpdf_parser_utility.cpp 199 FX_BOOL CPDF_SimpleParser::FindTagPair(FX_BSTR start_token, FX_BSTR end_token,
202 if (!start_token.IsEmpty()) {
203 if (!SkipWord(start_token)) {

Completed in 226 milliseconds

1 2