|   /external/chromium_org/third_party/ply/ | 
| lex.py  | 57 # This regular expression is used to match valid token names 60 # Exception thrown when invalid token encountered and no default error 68 # Token class.  This class is used to represent the tokens produced. 107 #    token()          -  Get the next token 298     # opttoken() - Return the next token from the Lexer 304     def token(self):  member in class:Lexer  322                 # Create a token for return 332                    # If no token type was set, it's an ignored token     [all...] | 
|   /external/chromium_org/third_party/protobuf/src/google/protobuf/io/ | 
| tokenizer.h  | 102                       // is NOT included in the token; it's up to the parser to 115   // Structure representing a token read from the token stream. 116   struct Token { 118     string text;       // The exact text of the token as it appeared in 123     // the token within the input stream.  They are zero-based. 129   // Get the current token.  This is updated when Next() is called.  Before 131   const Token& current(); 133   // Return the previous token -- i.e. what current() returned before the 135   const Token& previous()     [all...] | 
|   /external/chromium_org/third_party/skia/third_party/lua/src/ | 
| llex.h  | 43 typedef struct Token { 44   int token;  member in struct:Token  46 } Token; 54   int lastline;  /* line of last token `consumed' */ 55   Token t;  /* current token */ 56   Token lookahead;  /* look ahead token */ 75 LUAI_FUNC const char *luaX_token2str (LexState *ls, int token);
  | 
|   /external/chromium_org/tools/gn/ | 
| token.h  | 11 class Token { 55   Token(); 56   Token(const Location& location, Type t, const base::StringPiece& v); 68   // Helper functions for comparing this token to something.
  | 
|   /external/clang/include/clang/Lex/ | 
| Token.h  | 1 //===--- Token.h - Token interface ------------------------------*- C++ -*-===// 10 //  This file defines the Token interface. 28 /// Token - This structure provides full information about a lexed token. 30 /// information as possible about each returned token.  This is expected to be 33 /// The parser can create a special "annotation token" representing a stream of 35 /// can be represented by a single typename annotation token that carries 37 class Token { 38   /// The location of the token     [all...] | 
|   /external/javassist/src/main/javassist/compiler/ | 
| Lex.java  | 18 class Token { 19     public Token next = null; 30     private Token currentToken; 31     private Token lookAheadTokens; 42         currentToken = new Token(); 55             Token t; 63      * Looks at the next token. 70         Token tk = lookAheadTokens; 79                 Token tk2; 80                 tk.next = tk2 = new Token();     [all...] | 
|   /external/llvm/examples/Kaleidoscope/Chapter2/ | 
| toy.cpp  | 14 enum Token { 27 /// gettok - Return the next token from standard input. 136 /// CurTok/getNextToken - Provide a simple token buffer.  CurTok is the current 137 /// token the parser is looking at.  getNextToken reads another token from the 148 /// GetTokPrecedence - Get the precedence of the pending binary operator token. 224   default: return Error("unknown token when expecting an expression"); 333     // Skip token for error recovery. 342     // Skip token for error recovery. 352     // Skip token for error recovery     [all...] | 
|   /external/llvm/lib/MC/MCParser/ | 
| AsmLexer.cpp  | 449   AsmToken Token = LexToken(); 458   return Token; 477     // If this comment starts with a '#', then return the Hash token and let 492   // EndOfStatement token before the Eof token.
  | 
|   /external/mesa3d/src/gallium/auxiliary/tgsi/ | 
| tgsi_parse.h  | 106    struct tgsi_token             Token;
  | 
|   /external/protobuf/src/google/protobuf/io/ | 
| tokenizer.h  | 101                       // is NOT included in the token; it's up to the parser to 114   // Structure representing a token read from the token stream. 115   struct Token { 117     string text;       // The exact text of the token as it appeared in 122     // the token within the input stream.  They are zero-based. 127   // Get the current token.  This is updated when Next() is called.  Before 129   const Token& current(); 131   // Advance to the next token.  Returns false if the end of the input is 137   // Parses a TYPE_FLOAT token.  This never fails, so long as the text actuall     [all...] | 
|   /packages/apps/QuickSearchBox/src/com/android/quicksearchbox/util/ | 
| LevenshteinDistance.java  | 33     private final Token[] mSource; 34     private final Token[] mTarget; 38     public LevenshteinDistance(Token[] source, Token[] target) { 65         final Token[] src = mSource; 66         final Token[] trg = mTarget; 72             Token sourceToken = src[s-1]; 74                 Token targetToken = trg[t-1]; 99      * Gets the list of operations which were applied to each target token; {@link #calculate} must 101      * @return A list of {@link EditOperation}s indicating the origin of each token in the targe     [all...] | 
|   /prebuilts/python/darwin-x86/2.7.5/lib/python2.7/lib2to3/pgen2/ | 
| tokenize.py  | 11     the token type (see token.py) 12     the token (a string) 13     the starting (row, column) indices of the token (a 2-tuple of ints) 14     the ending (row, column) indices of the token (a 2-tuple of ints) 26 each time a new token is found.""" 34 from lib2to3.pgen2.token import * 36 from . import token  namespace  37 __all__ = [x for x in dir(token) if x[0] != '_'] + ["tokenize", 39 del token     [all...] | 
|   /prebuilts/python/darwin-x86/2.7.5/lib/python2.7/ | 
| tokenize.py  | 8     the token type (see token.py) 9     the token (a string) 10     the starting (row, column) indices of the token (a 2-tuple of ints) 11     the ending (row, column) indices of the token (a 2-tuple of ints) 23 each time a new token is found.""" 30 from token import * 32 import token  namespace  33 __all__ = [x for x in dir(token) if not x.startswith("_")] 36 del token     [all...] | 
|   /prebuilts/python/linux-x86/2.7.5/lib/python2.7/lib2to3/pgen2/ | 
| tokenize.py  | 11     the token type (see token.py) 12     the token (a string) 13     the starting (row, column) indices of the token (a 2-tuple of ints) 14     the ending (row, column) indices of the token (a 2-tuple of ints) 26 each time a new token is found.""" 34 from lib2to3.pgen2.token import * 36 from . import token  namespace  37 __all__ = [x for x in dir(token) if x[0] != '_'] + ["tokenize", 39 del token     [all...] | 
|   /prebuilts/python/linux-x86/2.7.5/lib/python2.7/ | 
| tokenize.py  | 8     the token type (see token.py) 9     the token (a string) 10     the starting (row, column) indices of the token (a 2-tuple of ints) 11     the ending (row, column) indices of the token (a 2-tuple of ints) 23 each time a new token is found.""" 30 from token import * 32 import token  namespace  33 __all__ = [x for x in dir(token) if not x.startswith("_")] 36 del token     [all...] | 
|   /external/vixl/src/a64/ | 
| debugger-a64.cc  | 42 // Debugger command lines are broken up in token of different type to make 44 class Token { 46   virtual ~Token() {} 48   // Token type. 56   // Token properties. 61   static Token* Tokenize(const char* arg); 65 template<typename T> class ValueToken : public Token { 89   static Token* Tokenize(const char* arg); 90   static RegisterToken* Cast(Token* tok) { 111   static Token* Tokenize(const char* arg) 897  Token* token = RegisterToken::Tokenize(arg);  local      [all...] | 
|   /external/antlr/antlr-3.4/runtime/Python/antlr3/ | 
| streams.py  | 37 from antlr3.tokens import Token, CommonToken 56     or token type sequence (such as interpretation). 66         Negative indexes are allowed.  LA(-1) is previous token (token 67 	just matched).  LA(-i) where i is before first token should 257         Get Token at current input pointer + i ahead where i=1 is next Token. 258         i<0 indicates tokens in the past.  So -1 is previous token and -2 is 259         two tokens ago. LT(0) is undefined.  For i>=n, return Token.EOFToken. 278         Get a token at an absolute index i; 0..n-1.  This is really onl     [all...] | 
|   /external/chromium_org/chrome/common/extensions/docs/examples/apps/hello-python/oauth2/ | 
| __init__.py  | 68 def build_xoauth_string(url, consumer, token=None): 70     request = Request.from_consumer_and_token(consumer, token, 74     request.sign_request(signing_method, consumer, token) 120     the service provider for a request token, kicking off the OAuth 141 class Token(object): 146     requests to identify the token being used, but the secret is used only in 148     token to. 151     token* that the live user authorizes with the service provider. The 152     consumer then exchanges the request token for an *access token* that ca     [all...] | 
|   /external/chromium_org/gpu/config/ | 
| gpu_test_expectations_parser.cc  | 28 enum Token { 131 Token ParseToken(const std::string& word) { 139       return static_cast<Token>(i); 226     Token token = ParseToken(tokens[i]);  local  227     switch (token) { 249         if (token == kConfigGPUDeviceID) { 253           if (!UpdateTestConfig(config, token, 0)) 274     Token token = ParseToken(tokens[i])  local      [all...] | 
|   /external/clang/include/clang/AST/ | 
| CommentLexer.h  | 10 //  This file defines lexer for structured comments and supporting token class. 56 /// \brief Comment token. 57 class Token { 61   /// The location of the token. 64   /// The actual kind of the token. 67   /// Length of the token spelling in comment.  Can be 0 for synthenized 71   /// Contains text value associated with a token. 74   /// Integer value associated with a token. 76   /// If the token is a konwn command, contains command ID and TextPtr is 295   void formTokenWithChars(Token &Result, const char *TokEnd     [all...] | 
|   /external/clang/lib/ASTMatchers/Dynamic/ | 
| Parser.cpp  | 27 /// \brief Simple structure to hold information for one token from the parser. 72   /// \brief Returns but doesn't consume the next token. 75   /// \brief Consumes and returns the next token. 144           // location to become a code completion token. 347         // We must find a , token to continue. 428 // If the prefix of this completion matches the completion token, add it to 489     const TokenInfo Token = Tokenizer->consumeNextToken(); 490     Error->addError(Token.Range, Error->ET_ParserInvalidToken) << Token.Text; 494   llvm_unreachable("Unknown token kind.")     [all...] | 
|   /external/deqp/executor/ | 
| xeXMLParser.hpp  | 43 enum Token 45 	TOKEN_INCOMPLETE = 0,					//!< Not enough data to determine token. 74 const char* getTokenName (Token token); 94 	Token				getToken			(void) const		{ return m_curToken;	} 128 	Token						m_curToken;			//!< Current token. 129 	int							m_curTokenLen;		//!< Length of current token.
  | 
|   /external/deqp/framework/opengl/ | 
| gluVarTypeUtil.hpp  | 41 	enum Token 56 	Token			getToken						(void) const { return m_token;															} 66 	Token			m_token;
  | 
|   /external/llvm/examples/Kaleidoscope/Chapter3/ | 
| toy.cpp  | 19 enum Token { 32 /// gettok - Return the next token from standard input. 156 /// CurTok/getNextToken - Provide a simple token buffer.  CurTok is the current 157 /// token the parser is looking at.  getNextToken reads another token from the 168 /// GetTokPrecedence - Get the precedence of the pending binary operator token. 245   default: return Error("unknown token when expecting an expression"); 482     // Skip token for error recovery. 494     // Skip token for error recovery. 507     // Skip token for error recovery     [all...] | 
|   /external/llvm/examples/Kaleidoscope/Chapter4/ | 
| toy.cpp  | 26 enum Token { 39 /// gettok - Return the next token from standard input. 163 /// CurTok/getNextToken - Provide a simple token buffer.  CurTok is the current 164 /// token the parser is looking at.  getNextToken reads another token from the 175 /// GetTokPrecedence - Get the precedence of the pending binary operator token. 252   default: return Error("unknown token when expecting an expression"); 495     // Skip token for error recovery. 507     // Skip token for error recovery. 525     // Skip token for error recovery     [all...] |