/external/chromium-trace/trace-viewer/third_party/closure_linter/closure_linter/ |
tokenutil.py | 98 def CustomSearch(start_token, func, end_func=None, distance=None, 103 start_token: The token to start searching from 117 token = start_token 147 def Search(start_token, token_types, distance=None, reverse=False): 151 start_token: The token to start searching from 163 return CustomSearch(start_token, lambda token: token.IsAnyType(token_types), 167 def SearchExcept(start_token, token_types, distance=None, reverse=False): 171 start_token: The token to start searching from 183 return CustomSearch(start_token, 188 def SearchUntil(start_token, token_types, end_types, distance=None [all...] |
error_fixer.py | 225 start_token = token.attached_object.type_start_token 227 if start_token.type != Type.DOC_START_BRACE: 229 len(start_token.string) - len(start_token.string.lstrip())) 231 start_token = tokenutil.SplitToken(start_token, leading_space) 233 if token.attached_object.type_end_token == start_token.previous: 234 token.attached_object.type_end_token = start_token 236 new_token = Token('{', Type.DOC_START_BRACE, start_token.line, 237 start_token.line_number [all...] |
ecmametadatapass.py | 51 start_token: The token where this context starts. 118 def __init__(self, type, start_token, parent): 123 start_token: The token where this context starts. 127 self.start_token = start_token 290 keyword_token = result.start_token.metadata.last_code 302 start_block_token = pre_keyword_token.metadata.context.start_token
|
indentation.py | 155 start_token = self._PopTo(Type.START_BLOCK) 157 if start_token: 158 goog_scope = self._GoogScopeOrNone(start_token.token) 177 (start_token.line_number)), 442 start_token = token 445 if not token or token.line_number != start_token.line_number:
|
statetracker.py | 221 start_token: The token that starts the doc comment. 225 def __init__(self, start_token): 229 start_token: The first token in the doc comment. 234 self.start_token = start_token 438 def _GetNextIdentifierToken(start_token): 448 start_token: The token to start searching after. 453 token = start_token.next 467 def _GetEndTokenAndContents(start_token): 474 start_token: The token immediately before the first content token [all...] |
javascriptlintrules.py | 216 block_start = doc_comment.start_token
|
/external/chromium_org/third_party/closure_linter/closure_linter/ |
tokenutil.py | 98 def CustomSearch(start_token, func, end_func=None, distance=None, 103 start_token: The token to start searching from 117 token = start_token 147 def Search(start_token, token_types, distance=None, reverse=False): 151 start_token: The token to start searching from 163 return CustomSearch(start_token, lambda token: token.IsAnyType(token_types), 167 def SearchExcept(start_token, token_types, distance=None, reverse=False): 171 start_token: The token to start searching from 183 return CustomSearch(start_token, 188 def SearchUntil(start_token, token_types, end_types, distance=None [all...] |
error_fixer.py | 225 start_token = token.attached_object.type_start_token 227 if start_token.type != Type.DOC_START_BRACE: 229 len(start_token.string) - len(start_token.string.lstrip())) 231 start_token = tokenutil.SplitToken(start_token, leading_space) 233 if token.attached_object.type_end_token == start_token.previous: 234 token.attached_object.type_end_token = start_token 236 new_token = Token('{', Type.DOC_START_BRACE, start_token.line, 237 start_token.line_number [all...] |
ecmametadatapass.py | 51 start_token: The token where this context starts. 118 def __init__(self, type, start_token, parent): 123 start_token: The token where this context starts. 127 self.start_token = start_token 290 keyword_token = result.start_token.metadata.last_code 302 start_block_token = pre_keyword_token.metadata.context.start_token
|
indentation.py | 155 start_token = self._PopTo(Type.START_BLOCK) 157 if start_token: 158 goog_scope = self._GoogScopeOrNone(start_token.token) 177 (start_token.line_number)), 442 start_token = token 445 if not token or token.line_number != start_token.line_number:
|
statetracker.py | 221 start_token: The token that starts the doc comment. 225 def __init__(self, start_token): 229 start_token: The first token in the doc comment. 234 self.start_token = start_token 438 def _GetNextIdentifierToken(start_token): 448 start_token: The token to start searching after. 453 token = start_token.next 467 def _GetEndTokenAndContents(start_token): 474 start_token: The token immediately before the first content token [all...] |
javascriptlintrules.py | 216 block_start = doc_comment.start_token
|
/external/antlr/antlr-3.4/runtime/Ruby/lib/antlr3/tree/ |
debug.rb | 96 def set_token_boundaries( tree, start_token, stop_token ) 97 super( tree, start_token, stop_token ) 98 return unless tree && start_token && stop_token 100 start_token.token_index, stop_token.token_index )
|
/external/chromium_org/webkit/common/fileapi/ |
file_system_util.cc | 352 std::string start_token(":"); 353 start_token = start_token.append( 359 start_token = StringToUpperASCII(start_token); 361 size_t pos = filesystem_name_upper.find(start_token); 367 *filesystem_id = filesystem_name.substr(pos + start_token.length(),
|
/external/mksh/src/ |
syn.c | 29 int start_token; /* token than began nesting (eg, FOR) */ member in struct:nesting_state 845 if (nesting.start_token) { 846 c = nesting.start_token; 886 nesting.start_token = tok; 915 nesting.start_token = 0;
|
/external/chromium_org/third_party/protobuf/src/google/protobuf/compiler/ |
parser.cc | 1109 io::Tokenizer::Token start_token; local [all...] |
/external/chromium_org/third_party/WebKit/Source/devtools/front_end/UglifyJS/ |
parse-js.js | 296 function start_token() { 527 start_token(); [all...] |
/ndk/sources/host-tools/sed-4.2.1/lib/ |
regcomp.c | 2475 re_token_t start_token = *token; local [all...] |