/external/clang/lib/Lex/ |
TokenLexer.cpp | 562 // a lexer and other overhead. 583 // Make a lexer to lex this string from. Lex just this one token. 584 // Make a lexer object so that we lex and expand the paste result. 585 Lexer TL(SourceMgr.getLocForStartOfFile(LocFileID), [all...] |
Pragma.cpp | 288 // Make and enter a lexer object so that we lex and expand the tokens just 290 Lexer *TL = Lexer::Create_PragmaLexer(TokLoc, PragmaLoc, RParenLoc, 359 // Get the current file lexer we're looking at. Ignore _Pragma 'files' etc. 365 assert(CurPPLexer && "No current lexer?"); 423 // Get the current file lexer we're looking at. Ignore _Pragma 'files' etc. [all...] |
ModuleMap.cpp | 23 #include "clang/Lex/Lexer.h" [all...] |
HeaderSearch.cpp | |
PPMacroExpansion.cpp | [all...] |
/external/chromium_org/third_party/ply/ |
lex.py | 103 # The following Lexer class implements the lexer runtime. There are only 106 # input() - Store a new string in the lexer 108 # clone() - Clone the lexer 114 class Lexer: 121 self.lexstatere = {} # Dictionary mapping lexer states to master regexs 122 self.lexstateretext = {} # Dictionary mapping lexer states to regex strings 123 self.lexstaterenames = {} # Dictionary mapping lexer states to symbol names 124 self.lexstate = "INITIAL" # Current lexer state 125 self.lexstatestack = [] # Stack of lexer state [all...] |
/prebuilts/tools/common/m2/repository/com/thoughtworks/qdox/qdox/1.12/ |
qdox-1.12.jar | |
/external/antlr/antlr-3.4/runtime/Python/antlr3/ |
recognizers.py | 91 # LEXER FIELDS (must be in same state object to avoid casting 92 # constantly in generated code and Lexer object) :( 95 ## The goal of all lexer rules/methods is to create a token object. 98 # matching lexer rule(s). If you subclass to allow multiple token 131 lexer, parser, and tree grammars. This is all the parsing 156 ## State of a lexer, parser, or tree parser are collected into a state 403 Get number of recognition errors (lexer, parser, tree parser). Each 404 recognizer tracks its own number. So parser and lexer each have 431 your token objects because you don't have to go modify your lexer 777 If you change what tokens must be created by the lexer, [all...] |
/external/antlr/antlr-3.4/runtime/Ruby/lib/antlr3/ |
streams.rb | 71 The job of a TokenStream is to read lexer output and then provide ANTLR 280 usually a lexer. They provide the operations required by higher-level 290 to them for some other purpose, you have the lexer assign the comments and 311 # expected to return the token source object (such as a lexer) from which 346 feed text input into an ANTLR Lexer for token processing. 532 # This is an extra utility method for use inside lexer actions if needed. 540 # This is an extra utility method for use inside lexer actions if needed. 548 # This is an extra utility method for use inside lexer actions if needed. 556 # This is an extra utility method for use inside lexer actions if needed. 726 Using some TokenSource (such as a lexer), the stream collects a token sequence [all...] |
/external/chromium_org/tools/idl_parser/ |
idl_parser.py | 129 # The Parser inherits the from the Lexer to provide PLY with the tokenizing 943 return self.lexer.last 945 def __init__(self, lexer, verbose=False, debug=False, mute_error=False): 946 self.lexer = lexer 947 self.tokens = lexer.KnownTokens() 975 filename = self.lexer.Lexer().filename 1059 # Access lexer errors, despite being private 1061 return self._parse_errors + self.lexer._lex_error [all...] |
/external/clang/lib/AST/ |
CommentParser.cpp | 35 /// This flag is set when there are no more tokens we can fetch from lexer. 95 /// fetch from lexer. 257 Parser::Parser(Lexer &L, Sema &S, llvm::BumpPtrAllocator &Allocator,
|
Expr.cpp | 30 #include "clang/Lex/Lexer.h" [all...] |
/external/clang/lib/Frontend/ |
SerializedDiagnosticPrinter.cpp | 17 #include "clang/Lex/Lexer.h" 308 TokSize = Lexer::MeasureTokenLength(Range.getEnd(),
|
VerifyDiagnosticConsumer.cpp | 580 /// any expected-* directives. As a Lexer is used rather than a full-blown 586 // Create a raw lexer to pull all the comments out of FID. 590 // Create a lexer to lex all the tokens of the main file in raw mode. 592 Lexer RawLex(FID, FromFile, SM, LangOpts);
|
ASTUnit.cpp | [all...] |
/external/antlr/antlr-3.4/tool/src/main/resources/org/antlr/codegen/templates/Python/ |
Python.stg | 47 outputFile(LEXER,PARSER,TREE_PARSER, actionScope, actions, 82 <if(LEXER)> 88 main = ParserMain("<recognizer.grammar.name>Lexer", <recognizer.name>)<\n> 107 lexer(grammar, name, tokens, scopes, rules, numRules, filterMode, 108 labelType="CommonToken", superClass="Lexer") ::= << 139 <actions.lexer.init> 142 <actions.lexer.members> 157 /** A override of Lexer.nextToken() that backtracks over mTokens() looking 203 # is Lexer always superclass? 559 /** How to generate a rule in the lexer; naked blocks are used fo [all...] |
/external/clang/lib/Edit/ |
RewriteObjCFoundationAPI.cpp | 21 #include "clang/Lex/Lexer.h" 690 StringRef text = Lexer::getSourceText( [all...] |
/external/clang/lib/Sema/ |
Sema.cpp | 48 return Lexer::getLocForEndOfToken(Loc, Offset, SourceMgr, LangOpts); [all...] |
AnalysisBasedWarnings.cpp | 37 #include "clang/Lex/Lexer.h" 651 SourceLocation ElseKwLoc = Lexer::getLocForEndOfToken( [all...] |
SemaExprObjC.cpp | [all...] |
/external/clang/unittests/AST/ |
CommentParser.cpp | 61 Lexer L(Allocator, Diags, Traits, Begin, Source, Source + strlen(Source)); [all...] |
/external/clang/tools/libclang/ |
CIndex.cpp | 35 #include "clang/Lex/Lexer.h" 146 unsigned Length = Lexer::MeasureTokenLength(SM.getSpellingLoc(EndLoc), [all...] |
/external/clang/lib/Rewrite/Frontend/ |
RewriteModernObjC.cpp | 24 #include "clang/Lex/Lexer.h" [all...] |
/external/antlr/antlr-3.4/tool/src/main/resources/org/antlr/codegen/templates/CSharp2/ |
CSharp2.stg | 43 outputFile( LEXER,PARSER,TREE_PARSER, actionScope, actions, 102 lexer(grammar, name, tokens, scopes, rules, numRules, filterMode, labelType="CommonToken", 103 superClass={<if(actions.(actionScope).superClass)><actions.(actionScope).superClass><else>Antlr.Runtime.Lexer<endif>}) ::= << 110 <actions.lexer.members> 199 /** A override of Lexer.nextToken() that backtracks over mTokens() looking 675 /** How to generate a rule in the lexer; naked blocks are used for 726 /** How to generate code for the implicitly-defined lexer grammar rule 727 * that chooses between lexer rules. 949 <if(!LEXER)>state.errorRecovery=false;<endif><if(backtracking)>state.failed=false;<endif> 957 <if(LEXER)> [all...] |
/external/antlr/antlr-3.4/tool/src/main/resources/org/antlr/codegen/templates/CSharp3/ |
CSharp3.stg | 44 outputFile( LEXER,PARSER,TREE_PARSER, actionScope, actions, 100 lexer(grammar, name, tokens, scopes, rules, numRules, filterMode, labelType="CommonToken", 101 superClass={<if(actions.(actionScope).superClass)><actions.(actionScope).superClass><else>Antlr.Runtime.Lexer<endif>}, 109 <actions.lexer.members> 194 /** A override of Lexer.nextToken() that backtracks over mTokens() looking 653 /** How to generate a rule in the lexer; naked blocks are used for 702 /** How to generate code for the implicitly-defined lexer grammar rule 703 * that chooses between lexer rules. 929 <if(!LEXER)>state.errorRecovery=false;<endif><if(backtracking)>state.failed=false;<endif> 937 <if(LEXER)> [all...] |