/external/antlr/antlr-3.4/runtime/JavaScript/tests/functional/ |
t039labels.html | 41 lexer = new TLexer(cstream), 42 tstream = new org.antlr.runtime.CommonTokenStream(lexer),
|
t041parameters.html | 41 lexer = new TLexer(cstream), 42 tstream = new org.antlr.runtime.CommonTokenStream(lexer),
|
t045dfabug.html | 41 lexer = new TLexer(cstream), 42 tstream = new org.antlr.runtime.CommonTokenStream(lexer),
|
t046rewrite.html | 70 lexer = new TLexer(cstream), 71 tstream = new org.antlr.runtime.TokenRewriteStream(lexer),
|
/external/antlr/antlr-3.4/runtime/Perl5/ |
MANIFEST | 34 lib/ANTLR/Runtime/Lexer.pm 54 t/lexer.t
|
/external/antlr/antlr-3.4/runtime/Python/antlr3/ |
main.py | 191 lexer = self.lexerClass(inStream) 192 for token in lexer: 207 "--lexer", 233 lexer = self.lexerClass(inStream) 234 tokenStream = antlr3.CommonTokenStream(lexer) 255 "--lexer", 291 lexer = self.lexerClass(inStream) 292 tokenStream = antlr3.CommonTokenStream(lexer)
|
/external/antlr/antlr-3.4/runtime/Ruby/lib/antlr3/ |
template.rb | 77 | does not appear to have an associated lexer class 91 autoload :Lexer, 'antlr3/template/group-file' 96 lexer = Lexer.new( source, options ) 97 parser = Parser.new( lexer, options ) 110 lexer = Lexer.new( input, options ) 111 parser = Parser.new( lexer, options )
|
main.rb | 279 if name =~ /(\w+)::(Lexer|Parser|TreeParser)$/ 315 generated lexer file is run directly from the command line. 325 lexer = @lexer_class.new( in_stream ) 329 token = lexer.next_token 394 opt.on( '--lexer-name CLASS_NAME', "name of the lexer class to use" ) { |val| 399 opt.on( '--lexer-file PATH_TO_LIBRARY', "path to library defining the lexer class" ) { |val| 403 warn( "unable to load the library specified by --lexer-file: #{ $! }" ) 431 fail( "unable to locate the lexer class ``#@lexer_class_name''" [all...] |
/external/chromium_org/third_party/angle/src/compiler/preprocessor/ |
ExpressionParser.y | 46 #include "Lexer.h" 64 pp::Lexer* lexer; 263 context->lexer->lex(token); 277 ExpressionParser::ExpressionParser(Lexer *lexer, Diagnostics *diagnostics) 278 : mLexer(lexer), 287 context.lexer = mLexer;
|
/external/clang/docs/ |
LibFormat.rst | 25 tooling::Replacements reformat(const FormatStyle &Style, Lexer &Lex, 29 This reads a token stream out of the lexer ``Lex`` and reformats all the code
|
/external/nist-sip/java/gov/nist/javax/sip/parser/ |
FromParser.java | 47 protected FromParser(Lexer lexer) { 48 super(lexer); 55 this.lexer.match(TokenTypes.FROM); 56 this.lexer.SPorHT(); 57 this.lexer.match(':'); 58 this.lexer.SPorHT(); 60 this.lexer.match('\n');
|
RSeqParser.java | 54 * param lexer the lexer to use to parse the header 56 protected RSeqParser(Lexer lexer) { 57 super(lexer); 75 String number = this.lexer.number(); 81 this.lexer.SPorHT(); 83 this.lexer.match('\n');
|
ContentLengthParser.java | 48 protected ContentLengthParser(Lexer lexer) { 49 super(lexer); 58 String number = this.lexer.number(); 60 this.lexer.SPorHT(); 61 this.lexer.match('\n');
|
MaxForwardsParser.java | 47 protected MaxForwardsParser(Lexer lexer) { 48 super(lexer); 57 String number = this.lexer.number(); 59 this.lexer.SPorHT(); 60 this.lexer.match('\n');
|
Parser.java | 46 lexer.getBuffer() + ":" + exceptionString, 47 lexer.getPtr()); 50 protected Lexer getLexer() { 51 return (Lexer) this.lexer; 58 Token tok = lexer.match(SIP); 61 lexer.match('/'); 62 tok = lexer.match(ID); 80 Token[] tokens = this.lexer.peekNextToken(1); 93 lexer.consume() [all...] |
/external/antlr/antlr-3.4/tool/src/test/java/org/antlr/test/ |
TestInterpretedLexing.java | 81 "lexer grammar t;\n"+ 95 "lexer grammar t;\n"+ 106 "lexer grammar t;\n"+ 120 "lexer grammar t;\n"+ 147 "lexer grammar t;\n"+ 164 "lexer grammar t;\n"+
|
/external/owasp/sanitizer/src/tests/org/owasp/html/ |
HtmlLexerTest.java | 125 HtmlLexer lexer = new HtmlLexer(input); local 131 while (lexer.hasNext()) { 132 HtmlToken t = lexer.next(); 149 HtmlLexer lexer = new HtmlLexer(markup); local 151 while (lexer.hasNext()) { 152 HtmlToken t = lexer.next();
|
/external/antlr/antlr-3.4/runtime/Python/tests/ |
testbase.py | 200 m = re.match(r'\s*((lexer|parser|tree)\s+|)grammar\s+(\S+);', grammar, re.MULTILINE) 207 assert self.grammarType in ('lexer', 'parser', 'tree', 'combined'), self.grammarType 285 """Optionally build a subclass of generated lexer class""" 312 """Build lexer instance. Arguments are passed to lexer.__init__().""" 314 if self.grammarType == 'lexer': 318 self.lexerModule = self.__load_module(self.grammarName + 'Lexer') 319 cls = getattr(self.lexerModule, self.grammarName + 'Lexer') 323 lexer = cls(*args, **kwargs) 325 return lexer [all...] |
/external/llvm/lib/MC/MCParser/ |
AsmParser.cpp | 125 AsmLexer Lexer; 200 MCAsmLexer &getLexer() override { return Lexer; } 309 /// \brief Reset the current lexer position to that given by \p Loc. The 493 : Lexer(_MAI), Ctx(_Ctx), Out(_Out), MAI(_MAI), SrcMgr(_SM), 502 Lexer.setBuffer(SrcMgr.getMemoryBuffer(CurBuffer)->getBuffer()); 570 SrcMgr.AddIncludeFile(Filename, Lexer.getLoc(), IncludedFile); 575 Lexer.setBuffer(SrcMgr.getMemoryBuffer(CurBuffer)->getBuffer()); 585 SrcMgr.AddIncludeFile(Filename, Lexer.getLoc(), IncludedFile); 596 Lexer.setBuffer(SrcMgr.getMemoryBuffer(CurBuffer)->getBuffer(), 601 const AsmToken *tok = &Lexer.Lex() [all...] |
/external/llvm/docs/tutorial/ |
LangImpl1.rst | 2 Kaleidoscope: Tutorial Introduction and the Lexer 39 language, and the definition of its Lexer - This shows where we are 42 to implement everything in C++ instead of using lexer and parser 46 With the lexer in place, we can talk about parsing techniques and 86 language including a hand-written lexer, parser, AST, as well as code 146 The Lexer 152 "`lexer <http://en.wikipedia.org/wiki/Lexical_analysis>`_" (aka 154 the lexer includes a token code and potentially some metadata (e.g. the 159 // The lexer returns tokens [0-255] if it is an unknown character, otherwise one 174 Each token returned by our lexer will either be one of the Token enu [all...] |
/external/antlr/antlr-3.4/runtime/C/include/ |
antlr3recognizersharedstate.h | 8 * (either lexer, parser or tree parser) that are need to 53 * of any recognizer (lexer, parser, tree parser). 136 /** The goal of all lexer rules/methods is to create a token object. 144 /** The goal of all lexer rules being to create a token, then a lexer 149 /** A lexer is a source of tokens, produced by all the generated (or 181 * lexer rules.
|
/external/owasp/sanitizer/src/main/org/owasp/html/ |
HtmlSanitizer.java | 130 HtmlLexer lexer = new HtmlLexer(html); local 134 while (lexer.hasNext()) { 135 HtmlToken token = lexer.next(); 149 while (lexer.hasNext() 150 && lexer.next().type != HtmlTokenType.TAGEND) { 158 while (lexer.hasNext()) { 159 HtmlToken tagBodyToken = lexer.next();
|
/external/clang/lib/ARCMigrate/ |
Transforms.cpp | 18 #include "clang/Lex/Lexer.h" 141 if (!Lexer::isAtEndOfMacroExpansion(loc, SM, Ctx.getLangOpts(), &loc)) 144 loc = Lexer::getLocForEndOfToken(loc, /*Offset=*/0, SM, Ctx.getLangOpts()); 158 Lexer lexer(SM.getLocForStartOfFile(locInfo.first), 162 lexer.LexFromRawLexer(tok); 408 Lexer lexer(SM.getLocForStartOfFile(locInfo.first), 412 lexer.LexFromRawLexer(tok); 414 lexer.LexFromRawLexer(tok) [all...] |
/external/clang/lib/AST/ |
CommentLexer.cpp | 13 void Token::dump(const Lexer &L, const SourceManager &SM) const { 49 StringRef Lexer::resolveHTMLNamedCharacterReference(StringRef Name) const { 61 StringRef Lexer::resolveHTMLDecimalCharacterReference(StringRef Name) const { 71 StringRef Lexer::resolveHTMLHexCharacterReference(StringRef Name) const { 82 void Lexer::skipLineStartingDecorations() { 271 void Lexer::formTokenWithChars(Token &Result, const char *TokEnd, 284 void Lexer::lexCommentText(Token &T) { 441 void Lexer::setupAndLexVerbatimBlock(Token &T, 466 void Lexer::lexVerbatimBlockFirstLine(Token &T) { 511 void Lexer::lexVerbatimBlockBody(Token &T) [all...] |
/external/antlr/antlr-3.4/runtime/CSharp3/Sources/Antlr3.Runtime.Test/SlimParsing/ |
SlimLexer.cs | 39 /** <summary>Where is the lexer drawing characters from?</summary> */ 105 // wack Lexer state variables 180 * Instruct the lexer to skip creating a token for current lexer rule 182 * a lexer rule finishes with token set to SKIP_TOKEN. Recall that 193 /** <summary>This is the lexer entry point that sets instance var 'token'</summary> */ 202 /** <summary>Set the char stream and reset the lexer</summary> */ 303 Recover( mte ); // don't really recover; just consume in lexer 339 /** TODO: not thought about recovery in lexer yet.
|