/external/antlr/antlr-3.4/runtime/Perl5/lib/ANTLR/Runtime/ |
UnwantedTokenException.pm | 12 return $self->token; 19 if ($self->expecting == ANTLR::Runtime::Token->INVALID_TOKEN_TYPE) { 26 if (defined $self->token) { 27 return "UnwantedTokenException(found=" . $self->token->get_text() . "$exp)";
|
/external/antlr/antlr-3.4/runtime/Python/tests/ |
t003lexer.py | 27 token = lexer.nextToken() 28 self.failUnlessEqual(token.type, self.lexerModule.ZERO) 30 token = lexer.nextToken() 31 self.failUnlessEqual(token.type, self.lexerModule.FOOZE) 33 token = lexer.nextToken() 34 self.failUnlessEqual(token.type, self.lexerModule.ONE) 36 token = lexer.nextToken() 37 self.failUnlessEqual(token.type, self.lexerModule.EOF) 45 token = lexer.nextToken()
|
/external/chromium_org/chrome/renderer/resources/extensions/enterprise_platform_keys/ |
token.js | 9 * Implementation of enterprise.platformKeys.Token. 10 * @param {string} id The id of the new Token. 18 exports.Token = 19 utils.expose('Token', TokenImpl, {readonly:['id', 'subtleCrypto']});
|
/external/chromium_org/tools/gn/ |
tokenizer.cc | 34 Token::Type GetSpecificOperatorType(base::StringPiece value) { 36 return Token::EQUAL; 38 return Token::PLUS; 40 return Token::MINUS; 42 return Token::PLUS_EQUALS; 44 return Token::MINUS_EQUALS; 46 return Token::EQUAL_EQUAL; 48 return Token::NOT_EQUAL; 50 return Token::LESS_EQUAL; 52 return Token::GREATER_EQUAL [all...] |
/external/antlr/antlr-3.4/runtime/ActionScript/project/src/org/antlr/runtime/ |
RecognizerSharedState.as | 11 /** Track the set of token types that can follow any rule invocation. 19 * matched a token. Prevents generation of more than one error message 26 * but no token is consumed during recovery...another error is found, 28 * one token/tree node is consumed for two errors. 32 /** In lieu of a return value, this indicates that a rule or token 33 * has failed to match. Reset to false upon valid token match. 46 * the stop token index for each rule. ruleMemo[ruleIndex] is 48 * get back the stop token for associated rule or MEMO_RULE_FAILED. 59 /** The goal of all lexer rules/methods is to create a token object. 61 * create a single token. nextToken will return this object afte 67 public var token:Token; variable [all...] |
/external/antlr/antlr-3.4/runtime/C/include/ |
antlr3commontoken.h | 2 * \brief Defines the interface for a common token. 4 * All token streams should provide their tokens using an instance 5 * of this common token. A custom pointer is provided, wher you may attach 6 * a further structure to enhance the common token if you feel the need 7 * to do so. The C runtime will assume that a token provides implementations 45 /** How many tokens to allocate at once in the token factory 49 /* Base token types, which all lexer/parser tokens come after in sequence. 52 /** Indicator of an invalid token 58 /** Imaginary token type to cause a traversal of child nodes in a tree parser 62 /** Imaginary token type to signal the end of a stream of child nodes [all...] |
/external/jsilver/src/com/google/clearsilver/jsilver/syntax/lexer/ |
Lexer.java | 11 protected Token token; field in class:Lexer 32 public Token peek() throws LexerException, IOException 34 while(this.token == null) 36 this.token = getToken(); 40 return this.token; 43 public Token next() throws LexerException, IOException 45 while(this.token == null) 47 this.token = getToken(); 51 Token result = this.token 160 @SuppressWarnings("hiding") Token token = new0( local 175 @SuppressWarnings("hiding") Token token = new1( local 190 @SuppressWarnings("hiding") Token token = new2( local 204 @SuppressWarnings("hiding") Token token = new3( local 218 @SuppressWarnings("hiding") Token token = new4( local 232 @SuppressWarnings("hiding") Token token = new5( local 246 @SuppressWarnings("hiding") Token token = new6( local 260 @SuppressWarnings("hiding") Token token = new7( local 274 @SuppressWarnings("hiding") Token token = new8( local 289 @SuppressWarnings("hiding") Token token = new9( local 303 @SuppressWarnings("hiding") Token token = new10( local 317 @SuppressWarnings("hiding") Token token = new11( local 331 @SuppressWarnings("hiding") Token token = new12( local 345 @SuppressWarnings("hiding") Token token = new13( local 359 @SuppressWarnings("hiding") Token token = new14( local 373 @SuppressWarnings("hiding") Token token = new15( local 387 @SuppressWarnings("hiding") Token token = new16( local 401 @SuppressWarnings("hiding") Token token = new17( local 415 @SuppressWarnings("hiding") Token token = new18( local 429 @SuppressWarnings("hiding") Token token = new19( local 443 @SuppressWarnings("hiding") Token token = new20( local 457 @SuppressWarnings("hiding") Token token = new21( local 471 @SuppressWarnings("hiding") Token token = new22( local 485 @SuppressWarnings("hiding") Token token = new23( local 499 @SuppressWarnings("hiding") Token token = new24( local 513 @SuppressWarnings("hiding") Token token = new25( local 527 @SuppressWarnings("hiding") Token token = new26( local 541 @SuppressWarnings("hiding") Token token = new27( local 555 @SuppressWarnings("hiding") Token token = new28( local 569 @SuppressWarnings("hiding") Token token = new29( local 583 @SuppressWarnings("hiding") Token token = new30( local 597 @SuppressWarnings("hiding") Token token = new31( local 611 @SuppressWarnings("hiding") Token token = new32( local 625 @SuppressWarnings("hiding") Token token = new33( local 639 @SuppressWarnings("hiding") Token token = new34( local 654 @SuppressWarnings("hiding") Token token = new35( local 668 @SuppressWarnings("hiding") Token token = new36( local 682 @SuppressWarnings("hiding") Token token = new37( local 696 @SuppressWarnings("hiding") Token token = new38( local 710 @SuppressWarnings("hiding") Token token = new39( local 724 @SuppressWarnings("hiding") Token token = new40( local 738 @SuppressWarnings("hiding") Token token = new41( local 752 @SuppressWarnings("hiding") Token token = new42( local 766 @SuppressWarnings("hiding") Token token = new43( local 780 @SuppressWarnings("hiding") Token token = new44( local 794 @SuppressWarnings("hiding") Token token = new45( local 808 @SuppressWarnings("hiding") Token token = new46( local 822 @SuppressWarnings("hiding") Token token = new47( local 837 @SuppressWarnings("hiding") Token token = new48( local 852 @SuppressWarnings("hiding") Token token = new49( local 867 @SuppressWarnings("hiding") Token token = new50( local 882 @SuppressWarnings("hiding") Token token = new51( local 897 @SuppressWarnings("hiding") Token token = new52( local 912 @SuppressWarnings("hiding") Token token = new53( local 926 @SuppressWarnings("hiding") Token token = new54( local 941 @SuppressWarnings("hiding") Token token = new55( local 956 @SuppressWarnings("hiding") Token token = new56( local 982 @SuppressWarnings("hiding") EOF token = new EOF( local [all...] |
/external/chromium-trace/trace-viewer/third_party/closure_linter/closure_linter/common/ |
tokens.py | 24 """Token types common to all languages.""" 30 class Token(object): 31 """Token class for intelligent text splitting. 33 The token class represents a string of characters and an identifying type. 36 type: The type of token. 37 string: The characters the token comprises. 38 length: The length of the token. 39 line: The text of the line the token is found in. 40 line_number: The number of the line the token is found in. 42 previous: The token before this one [all...] |
/external/chromium_org/third_party/closure_linter/closure_linter/common/ |
tokens.py | 24 """Token types common to all languages.""" 30 class Token(object): 31 """Token class for intelligent text splitting. 33 The token class represents a string of characters and an identifying type. 36 type: The type of token. 37 string: The characters the token comprises. 38 length: The length of the token. 39 line: The text of the line the token is found in. 40 line_number: The number of the line the token is found in. 42 previous: The token before this one [all...] |
/external/owasp/sanitizer/src/main/org/owasp/html/ |
CssGrammar.java | 63 // dropping spaces and comments to identify property names and token runs 79 if (!(it.hasTokenAfterSpace() && ":".equals(it.token()))) { 96 String token = it.token(); local 103 handler.startFunction(token); 105 handler.endFunction(token); 106 continue; // Skip the advance over token. 108 handler.identifier(token); 111 if (token.length() == 4 || token.length() == 7) [all...] |
StylingPolicy.java | 74 private void emitToken(String token) { 77 sanitizedCss.append(token); 88 public void url(String token) { 106 public void startFunction(String token) { 110 token = Strings.toLowerCase(token); 111 String key = cssProperty.fnKeys.get(token); 116 emitToken(token); 120 public void quotedString(String token) { 132 && token.length() > [all...] |
/external/chromium_org/third_party/WebKit/Source/core/css/parser/ |
MediaQueryParser.cpp | 23 PassRefPtrWillBeRawPtr<MediaQuerySet> MediaQueryParser::parseMediaCondition(MediaQueryTokenIterator token, MediaQueryTokenIterator endToken) 25 return MediaQueryParser(MediaConditionParser).parseImpl(token, endToken); 59 void MediaQueryParser::readRestrictor(MediaQueryTokenType type, const MediaQueryToken& token) 61 readMediaType(type, token); 64 void MediaQueryParser::readMediaType(MediaQueryTokenType type, const MediaQueryToken& token) 69 if (m_state == ReadRestrictor && equalIgnoringCase(token.value(), "not")) { 71 } else if (m_state == ReadRestrictor && equalIgnoringCase(token.value(), "only")) { 74 m_mediaQueryData.setMediaType(token.value()); 82 skipUntilComma(type, token); 86 void MediaQueryParser::readAnd(MediaQueryTokenType type, const MediaQueryToken& token) [all...] |
/external/antlr/antlr-3.4/runtime/JavaScript/src/org/antlr/runtime/ |
RecognizerSharedState.js | 9 /** Track the set of token types that can follow any rule invocation. 18 * matched a token. Prevents generation of more than one error message 25 * but no token is consumed during recovery...another error is found, 27 * one token/tree node is consumed for two errors. 31 /** In lieu of a return value, this indicates that a rule or token 32 * has failed to match. Reset to false upon valid token match. 45 * the stop token index for each rule. ruleMemo[ruleIndex] is 47 * get back the stop token for associated rule or MEMO_RULE_FAILED. 58 /** The goal of all lexer rules/methods is to create a token object. 60 * create a single token. nextToken will return this object afte [all...] |
/external/chromium-trace/trace-viewer/third_party/closure_linter/closure_linter/ |
closurizednamespacesinfo_test.py | 76 token = self._GetRequireTokens('package.Something') 81 self.assertFalse(namespaces_info.IsExtraRequire(token), 87 self.assertTrue(namespaces_info.IsExtraRequire(token), 96 token = self._tokenizer.TokenizeFile(input_lines) 97 namespaces_info = self._GetInitializedNamespacesInfo(token, ['package'], []) 99 self.assertFalse(namespaces_info.IsExtraProvide(token), 108 token = self._tokenizer.TokenizeFile(input_lines) 109 namespaces_info = self._GetInitializedNamespacesInfo(token, ['package'], []) 111 self.assertFalse(namespaces_info.IsExtraProvide(token), 117 token = self._tokenizer.TokenizeFile(input_lines [all...] |
requireprovidesorter_test.py | 52 token = self._tokenizer.TokenizeFile(input_lines) 54 self._metadata_pass.Process(token) 57 sorter.FixRequires(token) 59 self.assertEquals(expected_lines, self._GetLines(token)) 61 def _GetLines(self, token): 62 """Returns an array of lines based on the specified token stream.""" 65 while token: 66 line += token.string 67 if token.IsLastInLine(): 70 token = token.nex [all...] |
checkerbase.py | 60 def _HandleError(self, code, message, token, position=None, 64 self.__checker.HandleError(code, message, token, position, fix_data) 74 def CheckToken(self, token, parser_state): 75 """Checks a token, given the current parser_state, for warnings and errors. 78 token: The current token under consideration. 108 lint_rules: LintRules object defining lint errors given a token 110 state_tracker: Object that tracks the current state in the token stream. 113 metadata_pass: Object that builds metadata about the token stream. 122 # tokenizer, only the token stream [all...] |
/external/chromium_org/third_party/closure_linter/closure_linter/ |
closurizednamespacesinfo_test.py | 76 token = self._GetRequireTokens('package.Something') 81 self.assertFalse(namespaces_info.IsExtraRequire(token), 87 self.assertTrue(namespaces_info.IsExtraRequire(token), 96 token = self._tokenizer.TokenizeFile(input_lines) 97 namespaces_info = self._GetInitializedNamespacesInfo(token, ['package'], []) 99 self.assertFalse(namespaces_info.IsExtraProvide(token), 108 token = self._tokenizer.TokenizeFile(input_lines) 109 namespaces_info = self._GetInitializedNamespacesInfo(token, ['package'], []) 111 self.assertFalse(namespaces_info.IsExtraProvide(token), 117 token = self._tokenizer.TokenizeFile(input_lines [all...] |
requireprovidesorter_test.py | 52 token = self._tokenizer.TokenizeFile(input_lines) 54 self._metadata_pass.Process(token) 57 sorter.FixRequires(token) 59 self.assertEquals(expected_lines, self._GetLines(token)) 61 def _GetLines(self, token): 62 """Returns an array of lines based on the specified token stream.""" 65 while token: 66 line += token.string 67 if token.IsLastInLine(): 70 token = token.nex [all...] |
checkerbase.py | 60 def _HandleError(self, code, message, token, position=None, 64 self.__checker.HandleError(code, message, token, position, fix_data) 74 def CheckToken(self, token, parser_state): 75 """Checks a token, given the current parser_state, for warnings and errors. 78 token: The current token under consideration. 108 lint_rules: LintRules object defining lint errors given a token 110 state_tracker: Object that tracks the current state in the token stream. 113 metadata_pass: Object that builds metadata about the token stream. 122 # tokenizer, only the token stream [all...] |
/external/nist-sip/java/gov/nist/javax/sip/header/ims/ |
PMediaAuthorization.java | 51 * P-Media-Authorization-Token 52 * *(COMMA P-Media-Authorization-Token) 53 * P-Media-Authorization-Token = 1*HEXDIG 70 * P-Media-Authorization Token 72 private String token; field in class:PMediaAuthorization 85 * Get the media authorization token. 87 * @return token 91 return token; 96 * Set the media authorization token. 98 * @param token - media authorization token to se [all...] |
/external/antlr/antlr-3.4/runtime/Java/src/main/java/org/antlr/runtime/tree/ |
CommonTree.java | 30 import org.antlr.runtime.Token; 32 /** A tree node that is wrapper for a Token object. After 3.0 release 39 /** A single token is the payload */ 40 public Token token; field in class:CommonTree 42 /** What token indexes bracket all tokens associated with this node 57 this.token = node.token; 62 public CommonTree(Token t) { 63 this.token = t [all...] |
/external/chromium_org/v8/src/ |
preparser.cc | 59 if (scanner->current_token() == Token::FUTURE_RESERVED_WORD) { 62 Token::FUTURE_STRICT_RESERVED_WORD) { 64 } else if (scanner->current_token() == Token::YIELD) { 116 ASSERT_EQ(Token::LBRACE, scanner()->current_token()); 124 ASSERT_EQ(Token::RBRACE, scanner()->peek()); 168 case Token::FUNCTION: 170 case Token::LET: 171 case Token::CONST: 186 if (directive_prologue && peek() != Token::STRING) { 237 case Token::LBRACE 604 Token::Value token = peek(); local [all...] |
/external/antlr/antlr-3.4/runtime/JavaScript/tests/functional/ |
t003lexer.html | 29 token; 31 token = lexer.nextToken(); 32 assertEquals(token.getType(), lexer.ZERO); 34 token = lexer.nextToken(); 35 assertEquals(token.getType(), lexer.FOOZE); 37 token = lexer.nextToken(); 38 assertEquals(token.getType(), lexer.ONE); 40 token = lexer.nextToken(); 41 assertEquals(token.getType(), lexer.EOF); 47 token; [all...] |
/external/chromium_org/google_apis/drive/ |
auth_service_interface.h | 17 // Called when fetching of access token is complete. 33 // Starts fetching OAuth2 access token from the refresh token. 37 // True if an OAuth2 access token is retrieved and believed to be fresh. 38 // The access token is used to access the Drive server. 41 // True if an OAuth2 refresh token is present. Its absence means that user 43 // The refresh token is used to get the access token. 46 // Returns OAuth2 access token. 49 // Clears OAuth2 access token [all...] |
/external/jsmn/ |
jsmn.c | 6 * Allocates a fresh unused token from the token pull. 24 * Fills token type and boundaries. 26 static void jsmn_fill_token(jsmntok_t *token, jsmntype_t type, 28 token->type = type; 29 token->start = start; 30 token->end = end; 31 token->size = 0; 35 * Fills next available token with JSON primitive. 39 jsmntok_t *token; local 88 jsmntok_t *token; local 157 jsmntok_t *token; local [all...] |