/external/llvm/examples/OCaml-Kaleidoscope/Chapter2/ |
lexer.ml | 5 let rec lex = parser var 7 | [< ' (' ' | '\n' | '\r' | '\t'); stream >] -> lex stream 27 [< 'Token.Kwd c; lex stream >] 36 | [< stream=lex >] -> 43 | [< stream=lex >] -> 50 | [< ' ('\n'); stream=lex >] -> stream
|
/external/llvm/examples/OCaml-Kaleidoscope/Chapter3/ |
lexer.ml | 5 let rec lex = parser var 7 | [< ' (' ' | '\n' | '\r' | '\t'); stream >] -> lex stream 27 [< 'Token.Kwd c; lex stream >] 36 | [< stream=lex >] -> 43 | [< stream=lex >] -> 50 | [< ' ('\n'); stream=lex >] -> stream
|
/external/llvm/examples/OCaml-Kaleidoscope/Chapter4/ |
lexer.ml | 5 let rec lex = parser var 7 | [< ' (' ' | '\n' | '\r' | '\t'); stream >] -> lex stream 27 [< 'Token.Kwd c; lex stream >] 36 | [< stream=lex >] -> 43 | [< stream=lex >] -> 50 | [< ' ('\n'); stream=lex >] -> stream
|
/external/llvm/examples/OCaml-Kaleidoscope/Chapter5/ |
lexer.ml | 5 let rec lex = parser var 7 | [< ' (' ' | '\n' | '\r' | '\t'); stream >] -> lex stream 27 [< 'Token.Kwd c; lex stream >] 36 | [< stream=lex >] -> 43 | [< stream=lex >] -> 55 | [< ' ('\n'); stream=lex >] -> stream
|
/external/llvm/examples/OCaml-Kaleidoscope/Chapter6/ |
lexer.ml | 5 let rec lex = parser var 7 | [< ' (' ' | '\n' | '\r' | '\t'); stream >] -> lex stream 27 [< 'Token.Kwd c; lex stream >] 36 | [< stream=lex >] -> 43 | [< stream=lex >] -> 57 | [< ' ('\n'); stream=lex >] -> stream
|
/external/llvm/examples/OCaml-Kaleidoscope/Chapter7/ |
lexer.ml | 5 let rec lex = parser var 7 | [< ' (' ' | '\n' | '\r' | '\t'); stream >] -> lex stream 27 [< 'Token.Kwd c; lex stream >] 36 | [< stream=lex >] -> 43 | [< stream=lex >] -> 58 | [< ' ('\n'); stream=lex >] -> stream
|
/external/javassist/src/main/javassist/compiler/ |
CompileError.java | 22 private Lex lex; field in class:CompileError 25 public CompileError(String s, Lex l) { 27 lex = l; 32 lex = null; 43 public Lex getLex() { return lex; }
|
Parser.java | 21 private Lex lex; field in class:Parser 23 public Parser(Lex lex) { 24 this.lex = lex; 27 public boolean hasMore() { return lex.lookAhead() >= 0; } 46 if (lex.lookAhead() == Identifier && lex.lookAhead(1) == '(') { 53 if (lex.get() != Identifier [all...] |
/external/antlr/antlr-3.4/runtime/ObjC/Framework/examples/treeparser/ |
Main.java | 7 LangLexer lex = new LangLexer(input); local 8 CommonTokenStream tokens = new CommonTokenStream(lex);
|
/external/owasp/sanitizer/src/tests/org/owasp/html/ |
HtmlLexerTest.java | 51 lex(input, actual); method 124 private static void lex(String input, Appendable out) throws Exception { method in class:HtmlLexerTest
|
/external/selinux/sepolgen/src/sepolgen/ |
classperms.py | 55 from . import lex namespace 56 lex.lex() 103 #lex.input(txt) 105 # tok = lex.token()
|
lex.py | 2 # ply: lex.py 483 # lex(module) 487 def lex(module=None,object=None,debug=0,optimize=0,lextab="lextab",reflags=0,nowarn=0): function 548 raise SyntaxError("lex: module does not define 'tokens'") 550 raise SyntaxError("lex: tokens must be a list or tuple.") 557 print("lex: Bad token name '%s'" % n) 560 print("lex: Warning. Token '%s' multiply defined." % n) 566 print("lex: tokens = '%s'" % list(lexobj.lextokens.keys())) 571 print("lex: Invalid literal %s. Must be a single character" % repr(c)) 576 print("lex: Invalid literals specification. literals must be a sequence of characters." [all...] |
yacc.py | 95 # .lexpos = Starting lex position 96 # .endlexpos = Ending lex position (optional, set automatically) 197 # If no lexer was given, we will try to use the lex module 199 from . import lex namespace 200 lexer = lex.lexer [all...] |
/prebuilts/python/linux-x86/2.7.5/lib/python2.7/site-packages/sepolgen/ |
classperms.py | 55 from . import lex namespace 56 lex.lex() 103 #lex.input(txt) 105 # tok = lex.token()
|
lex.py | 2 # ply: lex.py 483 # lex(module) 487 def lex(module=None,object=None,debug=0,optimize=0,lextab="lextab",reflags=0,nowarn=0): function 548 raise SyntaxError("lex: module does not define 'tokens'") 550 raise SyntaxError("lex: tokens must be a list or tuple.") 557 print("lex: Bad token name '%s'" % n) 560 print("lex: Warning. Token '%s' multiply defined." % n) 566 print("lex: tokens = '%s'" % list(lexobj.lextokens.keys())) 571 print("lex: Invalid literal %s. Must be a single character" % repr(c)) 576 print("lex: Invalid literals specification. literals must be a sequence of characters." [all...] |
yacc.py | 95 # .lexpos = Starting lex position 96 # .endlexpos = Ending lex position (optional, set automatically) 197 # If no lexer was given, we will try to use the lex module 199 from . import lex namespace 200 lexer = lex.lexer [all...] |
/external/llvm/utils/lit/lit/ |
ShUtil.py | 62 # Otherwise, lex the operator and convert to a redirection 132 lex_one_token - Lex a single 'sh' token. """ 162 def lex(self): member in class:ShLexer 175 self.tokens = ShLexer(data, win32Escapes = win32Escapes).lex() 177 def lex(self): member in class:ShParser 183 token = self.lex() 189 tok = self.lex() 206 args.append(self.lex()) 215 op = self.lex() 216 arg = self.lex() 253 def lex(self, str, *args, **kwargs): member in class:TestShLexer [all...] |
/external/svox/pico/lib/ |
picowa.c | 66 /* lex knowledge base */ 67 picoklex_Lex lex; member in struct:wa_subobj 105 /* kb lex */ 106 wa->lex = picoklex_getLex(this->voice->kbArray[PICOKNOW_KBID_LEX_MAIN]); 107 if (wa->lex == NULL) { 111 PICODBG_DEBUG(("got lex")); 290 if (!found && !picoklex_lexLookup(wa->lex, content, head->len, &lexres)) { 291 /* no lex entry found, WORDGRAPH(POS,NA)graph */ 309 ensured for user lex too */ 344 /* set lex pos in info1 * [all...] |
picosa.c | 86 2. lex-index lookup and G2P (both directions possible, left-to-right done): 229 /* lex knowledge base */ 230 picoklex_Lex lex; member in struct:sa_subobj 402 /* kb lex */ 403 sa->lex = picoklex_getLex(this->voice->kbArray[PICOKNOW_KBID_LEX_MAIN]); 404 if (sa->lex == NULL) { 408 PICODBG_DEBUG(("got lex")); 735 picoklex_Lex lex, 742 if (picoklex_lexIndLookup(lex, &(sa->cbuf1[sa->headx[ind].cind + 1]), 1197 picoklex_Lex lex; local [all...] |
/system/core/libcutils/ |
config_utils.c | 232 static int lex(cstate *cs, int value) 240 #define lex(cs,v) _lex(cs,v) macro 248 switch(lex(cs, 0)){ 272 switch(lex(cs, 1)) { 274 if(lex(cs, 0) != T_TEXT) 300 switch(lex(&cs, 0)) {
|
/external/owasp/sanitizer/src/main/org/owasp/html/ |
CssTokens.java | 85 public static CssTokens lex(String css) { method in class:CssTokens 87 lexer.lex(); 422 void lex() { method in class:CssTokens.Lexer [all...] |
/external/antlr/antlr-3.4/tool/src/main/java/org/antlr/tool/ |
Strip.java | 63 ANTLRv3Lexer lex = new ANTLRv3Lexer(input); local 64 tokens = new TokenRewriteStream(lex);
|
/external/chromium-trace/catapult/third_party/Paste/paste/util/ |
template.py | 378 def lex(s, name=None, trim_whitespace=True): function 380 Lex a string into chunks: 382 >>> lex('hey') 384 >>> lex('hey {{you}}') 386 >>> lex('hey {{') 390 >>> lex('hey }}') 394 >>> lex('hey {{ {{') 443 >>> tokens = lex('{{if x}}\nx\n{{endif}}\ny', trim_whitespace=False) 537 tokens = lex(s, name=name)
|
/toolchain/binutils/binutils-2.25/gas/ |
app.c | 59 static char lex[256]; variable 86 #define IS_SYMBOL_COMPONENT(c) (lex[c] == LEX_IS_SYMBOL_COMPONENT) 87 #define IS_WHITESPACE(c) (lex[c] == LEX_IS_WHITESPACE) 88 #define IS_LINE_SEPARATOR(c) (lex[c] == LEX_IS_LINE_SEPARATOR) 89 #define IS_PARALLEL_SEPARATOR(c) (lex[c] == LEX_IS_PARALLEL_SEPARATOR) 90 #define IS_COMMENT(c) (lex[c] == LEX_IS_COMMENT_START) 91 #define IS_LINE_COMMENT(c) (lex[c] == LEX_IS_LINE_COMMENT_START) 92 #define IS_NEWLINE(c) (lex[c] == LEX_IS_NEWLINE) 106 lex[' '] = LEX_IS_WHITESPACE; 107 lex['\t'] = LEX_IS_WHITESPACE [all...] |
/external/clang/lib/Format/ |
Format.cpp | 25 #include "clang/Lex/Lexer.h" 770 Lex.reset(new Lexer(ID, SourceMgr.getBuffer(ID), SourceMgr, 779 ArrayRef<FormatToken *> lex() { function in class:clang::format::__anon6466::FormatTokenLexer [all...] |