/external/chromium_org/third_party/ply/ |
lex.py | 57 # This regular expression is used to match valid token names 60 # Exception thrown when invalid token encountered and no default error 68 # Token class. This class is used to represent the tokens produced. 107 # token() - Get the next token 298 # opttoken() - Return the next token from the Lexer 304 def token(self): member in class:Lexer 322 # Create a token for return 332 # If no token type was set, it's an ignored token [all...] |
/external/chromium_org/third_party/protobuf/src/google/protobuf/io/ |
tokenizer.h | 102 // is NOT included in the token; it's up to the parser to 115 // Structure representing a token read from the token stream. 116 struct Token { 118 string text; // The exact text of the token as it appeared in 123 // the token within the input stream. They are zero-based. 129 // Get the current token. This is updated when Next() is called. Before 131 const Token& current(); 133 // Return the previous token -- i.e. what current() returned before the 135 const Token& previous() [all...] |
/external/chromium_org/third_party/skia/third_party/lua/src/ |
llex.h | 43 typedef struct Token { 44 int token; member in struct:Token 46 } Token; 54 int lastline; /* line of last token `consumed' */ 55 Token t; /* current token */ 56 Token lookahead; /* look ahead token */ 75 LUAI_FUNC const char *luaX_token2str (LexState *ls, int token);
|
/external/chromium_org/tools/gn/ |
token.h | 11 class Token { 55 Token(); 56 Token(const Location& location, Type t, const base::StringPiece& v); 68 // Helper functions for comparing this token to something.
|
/external/clang/include/clang/Lex/ |
Token.h | 1 //===--- Token.h - Token interface ------------------------------*- C++ -*-===// 10 // This file defines the Token interface. 28 /// Token - This structure provides full information about a lexed token. 30 /// information as possible about each returned token. This is expected to be 33 /// The parser can create a special "annotation token" representing a stream of 35 /// can be represented by a single typename annotation token that carries 37 class Token { 38 /// The location of the token [all...] |
/external/javassist/src/main/javassist/compiler/ |
Lex.java | 18 class Token { 19 public Token next = null; 30 private Token currentToken; 31 private Token lookAheadTokens; 42 currentToken = new Token(); 55 Token t; 63 * Looks at the next token. 70 Token tk = lookAheadTokens; 79 Token tk2; 80 tk.next = tk2 = new Token(); [all...] |
/external/llvm/examples/Kaleidoscope/Chapter2/ |
toy.cpp | 14 enum Token { 27 /// gettok - Return the next token from standard input. 136 /// CurTok/getNextToken - Provide a simple token buffer. CurTok is the current 137 /// token the parser is looking at. getNextToken reads another token from the 148 /// GetTokPrecedence - Get the precedence of the pending binary operator token. 224 default: return Error("unknown token when expecting an expression"); 333 // Skip token for error recovery. 342 // Skip token for error recovery. 352 // Skip token for error recovery [all...] |
/external/llvm/lib/MC/MCParser/ |
AsmLexer.cpp | 449 AsmToken Token = LexToken(); 458 return Token; 477 // If this comment starts with a '#', then return the Hash token and let 492 // EndOfStatement token before the Eof token.
|
/external/mesa3d/src/gallium/auxiliary/tgsi/ |
tgsi_parse.h | 106 struct tgsi_token Token;
|
/external/protobuf/src/google/protobuf/io/ |
tokenizer.h | 101 // is NOT included in the token; it's up to the parser to 114 // Structure representing a token read from the token stream. 115 struct Token { 117 string text; // The exact text of the token as it appeared in 122 // the token within the input stream. They are zero-based. 127 // Get the current token. This is updated when Next() is called. Before 129 const Token& current(); 131 // Advance to the next token. Returns false if the end of the input is 137 // Parses a TYPE_FLOAT token. This never fails, so long as the text actuall [all...] |
/packages/apps/QuickSearchBox/src/com/android/quicksearchbox/util/ |
LevenshteinDistance.java | 33 private final Token[] mSource; 34 private final Token[] mTarget; 38 public LevenshteinDistance(Token[] source, Token[] target) { 65 final Token[] src = mSource; 66 final Token[] trg = mTarget; 72 Token sourceToken = src[s-1]; 74 Token targetToken = trg[t-1]; 99 * Gets the list of operations which were applied to each target token; {@link #calculate} must 101 * @return A list of {@link EditOperation}s indicating the origin of each token in the targe [all...] |
/prebuilts/python/darwin-x86/2.7.5/lib/python2.7/lib2to3/pgen2/ |
tokenize.py | 11 the token type (see token.py) 12 the token (a string) 13 the starting (row, column) indices of the token (a 2-tuple of ints) 14 the ending (row, column) indices of the token (a 2-tuple of ints) 26 each time a new token is found.""" 34 from lib2to3.pgen2.token import * 36 from . import token namespace 37 __all__ = [x for x in dir(token) if x[0] != '_'] + ["tokenize", 39 del token [all...] |
/prebuilts/python/darwin-x86/2.7.5/lib/python2.7/ |
tokenize.py | 8 the token type (see token.py) 9 the token (a string) 10 the starting (row, column) indices of the token (a 2-tuple of ints) 11 the ending (row, column) indices of the token (a 2-tuple of ints) 23 each time a new token is found.""" 30 from token import * 32 import token namespace 33 __all__ = [x for x in dir(token) if not x.startswith("_")] 36 del token [all...] |
/prebuilts/python/linux-x86/2.7.5/lib/python2.7/lib2to3/pgen2/ |
tokenize.py | 11 the token type (see token.py) 12 the token (a string) 13 the starting (row, column) indices of the token (a 2-tuple of ints) 14 the ending (row, column) indices of the token (a 2-tuple of ints) 26 each time a new token is found.""" 34 from lib2to3.pgen2.token import * 36 from . import token namespace 37 __all__ = [x for x in dir(token) if x[0] != '_'] + ["tokenize", 39 del token [all...] |
/prebuilts/python/linux-x86/2.7.5/lib/python2.7/ |
tokenize.py | 8 the token type (see token.py) 9 the token (a string) 10 the starting (row, column) indices of the token (a 2-tuple of ints) 11 the ending (row, column) indices of the token (a 2-tuple of ints) 23 each time a new token is found.""" 30 from token import * 32 import token namespace 33 __all__ = [x for x in dir(token) if not x.startswith("_")] 36 del token [all...] |
/external/vixl/src/a64/ |
debugger-a64.cc | 42 // Debugger command lines are broken up in token of different type to make 44 class Token { 46 virtual ~Token() {} 48 // Token type. 56 // Token properties. 61 static Token* Tokenize(const char* arg); 65 template<typename T> class ValueToken : public Token { 89 static Token* Tokenize(const char* arg); 90 static RegisterToken* Cast(Token* tok) { 111 static Token* Tokenize(const char* arg) 897 Token* token = RegisterToken::Tokenize(arg); local [all...] |
/external/antlr/antlr-3.4/runtime/Python/antlr3/ |
streams.py | 37 from antlr3.tokens import Token, CommonToken 56 or token type sequence (such as interpretation). 66 Negative indexes are allowed. LA(-1) is previous token (token 67 just matched). LA(-i) where i is before first token should 257 Get Token at current input pointer + i ahead where i=1 is next Token. 258 i<0 indicates tokens in the past. So -1 is previous token and -2 is 259 two tokens ago. LT(0) is undefined. For i>=n, return Token.EOFToken. 278 Get a token at an absolute index i; 0..n-1. This is really onl [all...] |
/external/chromium_org/chrome/common/extensions/docs/examples/apps/hello-python/oauth2/ |
__init__.py | 68 def build_xoauth_string(url, consumer, token=None): 70 request = Request.from_consumer_and_token(consumer, token, 74 request.sign_request(signing_method, consumer, token) 120 the service provider for a request token, kicking off the OAuth 141 class Token(object): 146 requests to identify the token being used, but the secret is used only in 148 token to. 151 token* that the live user authorizes with the service provider. The 152 consumer then exchanges the request token for an *access token* that ca [all...] |
/external/chromium_org/gpu/config/ |
gpu_test_expectations_parser.cc | 28 enum Token { 131 Token ParseToken(const std::string& word) { 139 return static_cast<Token>(i); 226 Token token = ParseToken(tokens[i]); local 227 switch (token) { 249 if (token == kConfigGPUDeviceID) { 253 if (!UpdateTestConfig(config, token, 0)) 274 Token token = ParseToken(tokens[i]) local [all...] |
/external/clang/include/clang/AST/ |
CommentLexer.h | 10 // This file defines lexer for structured comments and supporting token class. 56 /// \brief Comment token. 57 class Token { 61 /// The location of the token. 64 /// The actual kind of the token. 67 /// Length of the token spelling in comment. Can be 0 for synthenized 71 /// Contains text value associated with a token. 74 /// Integer value associated with a token. 76 /// If the token is a konwn command, contains command ID and TextPtr is 295 void formTokenWithChars(Token &Result, const char *TokEnd [all...] |
/external/clang/lib/ASTMatchers/Dynamic/ |
Parser.cpp | 27 /// \brief Simple structure to hold information for one token from the parser. 72 /// \brief Returns but doesn't consume the next token. 75 /// \brief Consumes and returns the next token. 144 // location to become a code completion token. 347 // We must find a , token to continue. 428 // If the prefix of this completion matches the completion token, add it to 489 const TokenInfo Token = Tokenizer->consumeNextToken(); 490 Error->addError(Token.Range, Error->ET_ParserInvalidToken) << Token.Text; 494 llvm_unreachable("Unknown token kind.") [all...] |
/external/deqp/executor/ |
xeXMLParser.hpp | 43 enum Token 45 TOKEN_INCOMPLETE = 0, //!< Not enough data to determine token. 74 const char* getTokenName (Token token); 94 Token getToken (void) const { return m_curToken; } 128 Token m_curToken; //!< Current token. 129 int m_curTokenLen; //!< Length of current token.
|
/external/deqp/framework/opengl/ |
gluVarTypeUtil.hpp | 41 enum Token 56 Token getToken (void) const { return m_token; } 66 Token m_token;
|
/external/llvm/examples/Kaleidoscope/Chapter3/ |
toy.cpp | 19 enum Token { 32 /// gettok - Return the next token from standard input. 156 /// CurTok/getNextToken - Provide a simple token buffer. CurTok is the current 157 /// token the parser is looking at. getNextToken reads another token from the 168 /// GetTokPrecedence - Get the precedence of the pending binary operator token. 245 default: return Error("unknown token when expecting an expression"); 482 // Skip token for error recovery. 494 // Skip token for error recovery. 507 // Skip token for error recovery [all...] |
/external/llvm/examples/Kaleidoscope/Chapter4/ |
toy.cpp | 26 enum Token { 39 /// gettok - Return the next token from standard input. 163 /// CurTok/getNextToken - Provide a simple token buffer. CurTok is the current 164 /// token the parser is looking at. getNextToken reads another token from the 175 /// GetTokPrecedence - Get the precedence of the pending binary operator token. 252 default: return Error("unknown token when expecting an expression"); 495 // Skip token for error recovery. 507 // Skip token for error recovery. 525 // Skip token for error recovery [all...] |