/external/chromium-trace/trace-viewer/third_party/closure_linter/closure_linter/ |
tokenutil.py | 98 def CustomSearch(start_token, func, end_func=None, distance=None, 103 start_token: The token to start searching from 117 token = start_token 147 def Search(start_token, token_types, distance=None, reverse=False): 151 start_token: The token to start searching from 163 return CustomSearch(start_token, lambda token: token.IsAnyType(token_types), 167 def SearchExcept(start_token, token_types, distance=None, reverse=False): 171 start_token: The token to start searching from 183 return CustomSearch(start_token, 188 def SearchUntil(start_token, token_types, end_types, distance=None [all...] |
error_fixer.py | 225 start_token = token.attached_object.type_start_token 227 if start_token.type != Type.DOC_START_BRACE: 229 len(start_token.string) - len(start_token.string.lstrip())) 231 start_token = tokenutil.SplitToken(start_token, leading_space) 233 if token.attached_object.type_end_token == start_token.previous: 234 token.attached_object.type_end_token = start_token 236 new_token = Token('{', Type.DOC_START_BRACE, start_token.line, 237 start_token.line_number [all...] |
ecmametadatapass.py | 51 start_token: The token where this context starts. 118 def __init__(self, type, start_token, parent): 123 start_token: The token where this context starts. 127 self.start_token = start_token 290 keyword_token = result.start_token.metadata.last_code 302 start_block_token = pre_keyword_token.metadata.context.start_token
|
indentation.py | 155 start_token = self._PopTo(Type.START_BLOCK) 157 if start_token: 158 goog_scope = self._GoogScopeOrNone(start_token.token) 177 (start_token.line_number)), 442 start_token = token 445 if not token or token.line_number != start_token.line_number:
|
statetracker.py | 221 start_token: The token that starts the doc comment. 225 def __init__(self, start_token): 229 start_token: The first token in the doc comment. 234 self.start_token = start_token 438 def _GetNextIdentifierToken(start_token): 448 start_token: The token to start searching after. 453 token = start_token.next 467 def _GetEndTokenAndContents(start_token): 474 start_token: The token immediately before the first content token [all...] |
javascriptlintrules.py | 216 block_start = doc_comment.start_token
|
/external/chromium_org/third_party/closure_linter/closure_linter/ |
tokenutil.py | 98 def CustomSearch(start_token, func, end_func=None, distance=None, 103 start_token: The token to start searching from 117 token = start_token 147 def Search(start_token, token_types, distance=None, reverse=False): 151 start_token: The token to start searching from 163 return CustomSearch(start_token, lambda token: token.IsAnyType(token_types), 167 def SearchExcept(start_token, token_types, distance=None, reverse=False): 171 start_token: The token to start searching from 183 return CustomSearch(start_token, 188 def SearchUntil(start_token, token_types, end_types, distance=None [all...] |
error_fixer.py | 225 start_token = token.attached_object.type_start_token 227 if start_token.type != Type.DOC_START_BRACE: 229 len(start_token.string) - len(start_token.string.lstrip())) 231 start_token = tokenutil.SplitToken(start_token, leading_space) 233 if token.attached_object.type_end_token == start_token.previous: 234 token.attached_object.type_end_token = start_token 236 new_token = Token('{', Type.DOC_START_BRACE, start_token.line, 237 start_token.line_number [all...] |
ecmametadatapass.py | 51 start_token: The token where this context starts. 118 def __init__(self, type, start_token, parent): 123 start_token: The token where this context starts. 127 self.start_token = start_token 290 keyword_token = result.start_token.metadata.last_code 302 start_block_token = pre_keyword_token.metadata.context.start_token
|
indentation.py | 155 start_token = self._PopTo(Type.START_BLOCK) 157 if start_token: 158 goog_scope = self._GoogScopeOrNone(start_token.token) 177 (start_token.line_number)), 442 start_token = token 445 if not token or token.line_number != start_token.line_number:
|
statetracker.py | 221 start_token: The token that starts the doc comment. 225 def __init__(self, start_token): 229 start_token: The first token in the doc comment. 234 self.start_token = start_token 438 def _GetNextIdentifierToken(start_token): 448 start_token: The token to start searching after. 453 token = start_token.next 467 def _GetEndTokenAndContents(start_token): 474 start_token: The token immediately before the first content token [all...] |
javascriptlintrules.py | 216 block_start = doc_comment.start_token
|
/external/linux-tools-perf/perf-3.12.0/tools/perf/util/ |
parse-events.l | 93 int start_token; 95 start_token = parse_events_get_extra(yyscanner); 97 if (start_token == PE_START_TERMS) 99 else if (start_token == PE_START_EVENTS) 102 if (start_token) { 104 return start_token;
|
parse-events.c | 846 static int parse_events__scanner(const char *str, void *data, int start_token); 849 int start_token) 869 return parse_events__scanner(s, data, start_token); 872 static int parse_events__scanner(const char *str, void *data, int start_token) 878 ret = parse_events_lex_init_extra(start_token, &scanner); 893 ret = parse_events_fixup(ret, str, data, start_token);
|
parse-events-flex.c | 1353 int start_token; variable [all...] |
/external/antlr/antlr-3.4/runtime/Ruby/lib/antlr3/tree/ |
debug.rb | 96 def set_token_boundaries( tree, start_token, stop_token ) 97 super( tree, start_token, stop_token ) 98 return unless tree && start_token && stop_token 100 start_token.token_index, stop_token.token_index )
|
/external/chromium_org/webkit/common/fileapi/ |
file_system_util.cc | 418 std::string start_token(":"); 419 start_token = start_token.append( 425 start_token = StringToUpperASCII(start_token); 427 size_t pos = filesystem_name_upper.find(start_token); 433 *filesystem_id = filesystem_name.substr(pos + start_token.length(),
|
/external/pdfium/core/src/fpdfapi/fpdf_parser/ |
fpdf_parser_utility.cpp | 199 FX_BOOL CPDF_SimpleParser::FindTagPair(FX_BSTR start_token, FX_BSTR end_token,
202 if (!start_token.IsEmpty()) {
203 if (!SkipWord(start_token)) {
|
/external/mksh/src/ |
syn.c | 29 int start_token; /* token than began nesting (eg, FOR) */ member in struct:nesting_state 838 if (nesting.start_token) { 839 c = nesting.start_token; 879 nesting.start_token = tok; 908 nesting.start_token = 0;
|
/external/chromium_org/third_party/protobuf/src/google/protobuf/compiler/ |
parser.cc | 1109 io::Tokenizer::Token start_token; local [all...] |
/external/chromium_org/third_party/WebKit/Source/devtools/front_end/UglifyJS/ |
parse-js.js | 296 function start_token() { 527 start_token(); [all...] |
/external/pdfium/core/include/fpdfapi/ |
fpdf_parser.h | 220 FX_BOOL FindTagPair(FX_BSTR start_token, FX_BSTR end_token,
|
/ndk/sources/host-tools/sed-4.2.1/lib/ |
regcomp.c | 2475 re_token_t start_token = *token; local [all...] |