/external/antlr/antlr-3.4/runtime/Python/tests/ |
t048rewrite.py | 25 tokens = self._parse("abc") 26 tokens.insertBefore(0, "0") 28 result = tokens.toString() 34 tokens = self._parse("abc") 35 tokens.insertAfter(2, "x") 37 result = tokens.toString() 43 tokens = self._parse("abc") 44 tokens.insertBefore(1, "x") 45 tokens.insertAfter(1, "x") 47 result = tokens.toString( [all...] |
/external/apache-harmony/regex/src/test/java/org/apache/harmony/tests/java/util/regex/ |
SplitTest.java | 45 String tokens[]; local 47 tokens = p.split(input, 1); 48 assertEquals(1, tokens.length); 49 assertTrue(tokens[0].equals(input)); 50 tokens = p.split(input, 2); 51 assertEquals(2, tokens.length); 52 assertEquals("poodle", tokens[0]); 53 assertEquals("zoo", tokens[1]); 54 tokens = p.split(input, 5); 55 assertEquals(2, tokens.length) [all...] |
/libcore/luni/src/test/java/org/apache/harmony/regex/tests/java/util/regex/ |
SplitTest.java | 67 String tokens[]; local 69 tokens = p.split(input, 1); 70 assertEquals(1, tokens.length); 71 assertTrue(tokens[0].equals(input)); 72 tokens = p.split(input, 2); 73 assertEquals(2, tokens.length); 74 assertEquals("poodle", tokens[0]); 75 assertEquals("zoo", tokens[1]); 76 tokens = p.split(input, 5); 77 assertEquals(2, tokens.length) [all...] |
/external/qemu/ |
json-parser.h | 20 QObject *json_parser_parse(QList *tokens, va_list *ap);
|
/external/antlr/antlr-3.4/tool/src/test/java/org/antlr/test/ |
TestCommonTokenStream.java | 48 // Tokens: 012345678901234567 52 BufferedTokenStream tokens = new BufferedTokenStream(lexEngine); local 54 String result = tokens.LT(1).getText(); 69 // Tokens: 012345678901234567 73 BufferedTokenStream tokens = new BufferedTokenStream(lexEngine); local 75 String result = tokens.LT(2).getText(); 90 // Tokens: 012345678901234567 94 BufferedTokenStream tokens = new BufferedTokenStream(lexEngine); local 97 Token t = tokens.LT(i); 100 t = tokens.LT(i) 124 BufferedTokenStream tokens = new BufferedTokenStream(lexEngine); local 155 BufferedTokenStream tokens = new BufferedTokenStream(lexEngine); local 200 CommonTokenStream tokens = new CommonTokenStream(lexer); local [all...] |
TestTokenRewriteStream.java | 51 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local 52 tokens.insertBefore(0, "0"); 53 String result = tokens.toString(); 66 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local 67 tokens.insertAfter(2, "x"); 68 String result = tokens.toString(); 81 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local 82 tokens.fill(); 83 tokens.insertBefore(1, "x"); 84 tokens.insertAfter(1, "x") 98 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local 114 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local 130 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local 151 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local 186 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local 233 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local 250 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local 268 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local 285 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local 309 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local 326 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local 343 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local 361 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local 378 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local 395 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local 412 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local 429 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local 453 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local 470 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local 486 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local 502 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local 526 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local 550 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local 569 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local 586 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local 604 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local 621 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local 638 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local 656 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local 673 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local 697 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local 714 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local 731 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local 748 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local 765 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local 782 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local 799 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local [all...] |
/external/chromium/net/tools/flip_server/ |
balsa_headers_token_utils.cc | 13 BalsaHeaders::HeaderTokenList* tokens) { 14 CHECK(tokens); 24 ParseTokenList(value_begin, line_end, tokens); 46 BalsaHeaders::HeaderTokenList tokens; local 47 TokenizeHeaderLine(*headers, *header_line, &tokens); 49 if (tokens.empty()) { 53 } else if (tokens.size() == 1) { 57 const base::StringPiece& new_last_token = tokens[tokens.size() - 2]; 84 BalsaHeaders::HeaderTokenList tokens; local [all...] |
balsa_headers_token_utils.h | 5 // Utility class that performs basic operations on header value tokens: parsing 6 // them out, checking for presense of certain tokens, and removing them. 30 // with that key, all of them will be tokenized and tokens will be added to 34 BalsaHeaders::HeaderTokenList* tokens); 43 // in some buffer, populates tokens list with beginning and end indices 44 // of all tokens present in the value string. 47 BalsaHeaders::HeaderTokenList* tokens); 54 BalsaHeaders::HeaderTokenList* tokens);
|
/dalvik/vm/compiler/template/ |
gen-template.py | 49 # Set handler_size_bytes to the value of tokens[1], and handler_size_bits to 53 def setHandlerSize(tokens): 55 if len(tokens) != 2: 61 handler_size_bytes = bytes = int(tokens[1]) 76 def importFile(tokens): 77 if len(tokens) != 2: 79 source = tokens[1] 81 appendSourceFile(tokens[1], getGlobalSubDict(), asm_fp, None) 90 def setAsmStub(tokens): 92 if len(tokens) != 2 385 tokens = line.split(' ') # tokenize variable [all...] |
/dalvik/vm/mterp/ |
gen-mterp.py | 59 def setHandlerStyle(tokens): 61 if len(tokens) != 2: 63 style = tokens[1] 69 # Set handler_size_bytes to the value of tokens[1], and handler_size_bits to 73 def setHandlerSize(tokens): 77 if len(tokens) != 2: 83 handler_size_bytes = bytes = int(tokens[1]) 98 def importFile(tokens): 99 if len(tokens) != 2: 101 source = tokens[1 580 tokens = line.split(' ') # tokenize variable [all...] |
/external/clang/bindings/python/tests/cindex/ |
test_tokens.py | 15 tokens = list(tu.get_tokens(extent=r)) 17 assert len(tokens) == 5 18 assert tokens[1].spelling == 'i' 19 assert tokens[1].kind == TokenKind.IDENTIFIER 21 cursor = tokens[1].cursor 23 assert tokens[1].cursor == tokens[2].cursor 31 tokens = list(tu.get_tokens(extent=r)) 32 eq_(len(tokens), 4) 34 loc = tokens[1].locatio [all...] |
/external/webkit/Source/WebCore/loader/ |
FTPDirectoryParser.cpp | 119 const char *tokens[16]; /* 16 is more than enough */ local 120 unsigned int toklen[WTF_ARRAY_LENGTH(tokens)]; 130 tokens[0] = state.carryBuffer; 136 while (pos < linelen && numtoks < WTF_ARRAY_LENGTH(tokens)) 143 tokens[numtoks] = &line[pos]; 147 if (tokens[numtoks] != &line[pos]) 149 toklen[numtoks] = (&line[pos] - tokens[numtoks]); 158 linelen_sans_wsp = &(tokens[numtoks-1][toklen[numtoks-1]]) - tokens[0]; 159 if (numtoks == WTF_ARRAY_LENGTH(tokens)) [all...] |
/cts/suite/audio_quality/lib/src/task/ |
TaskSound.cpp | 64 UniquePtr<std::vector<android::String8> > tokens(StringUtil::split(type, ':')); 65 if (tokens.get() == NULL) { 70 if (StringUtil::compare(tokens->at(0), "file") == 0) { 71 if (tokens->size() != 2) { 72 LOGE("Wrong number of parameters %d", tokens->size()); 74 buffer = Buffer::loadFromFile(tokens->at(1)); 75 } else if (StringUtil::compare(tokens->at(0), "sin") == 0) { 76 if (tokens->size() != 4) { 77 LOGE("Wrong number of parameters %d", tokens->size()); 79 int amplitude = atoi(tokens->at(1).string()) [all...] |
/external/jmonkeyengine/engine/src/core-plugins/com/jme3/font/plugins/ |
BitmapFontLoader.java | 64 String[] tokens = line.split(regex); local 65 if (tokens[0].equals("info")){ 67 for (int i = 1; i < tokens.length; i++){ 68 if (tokens[i].equals("size")){ 69 charSet.setRenderedSize(Integer.parseInt(tokens[i + 1])); 72 }else if (tokens[0].equals("common")){ 74 for (int i = 1; i < tokens.length; i++){ 75 String token = tokens[i]; 77 charSet.setLineHeight(Integer.parseInt(tokens[i + 1])); 79 charSet.setBase(Integer.parseInt(tokens[i + 1])) [all...] |
/external/antlr/antlr-3.4/runtime/Perl5/examples/tweak/ |
tweak.pl | 13 my $tokens = ANTLR::Runtime::TokenRewriteStream({ token_source => $lexer }); 14 my $parser = TParser->new({ input => $tokens }); 16 print "$tokens\n";
|
/external/clang/test/Preprocessor/ |
macro_paste_msextensions.c | 12 #define comment /##/ dead tokens live here 19 #define nested(x) int x comment cute little dead tokens... 21 nested(baz) rise of the dead tokens
|
/cts/tools/dasm/src/dasm/tokens/ |
relative_number_token.java | 17 package dasm.tokens;
|
/frameworks/base/wifi/java/android/net/wifi/p2p/ |
WifiP2pProvDiscEvent.java | 62 String[] tokens = string.split(" "); local 64 if (tokens.length < 2) { 68 if (tokens[0].endsWith("PBC-REQ")) event = PBC_REQ; 69 else if (tokens[0].endsWith("PBC-RESP")) event = PBC_RSP; 70 else if (tokens[0].endsWith("ENTER-PIN")) event = ENTER_PIN; 71 else if (tokens[0].endsWith("SHOW-PIN")) event = SHOW_PIN; 76 device.deviceAddress = tokens[1]; 79 pin = tokens[2];
|
/external/antlr/antlr-3.4/runtime/JavaScript/tests/functional/ |
rhino-python.extensions | 33 tokens should be sent to the parser sometimes without a corresponding 48 This TokenStream normally just passes tokens through to the parser. 63 A queue of tokens is built up to hold multiple DEDENT tokens that 77 /** The queue of tokens */ 78 this.tokens = []; 117 if (this.tokens.length>0 ) { 118 var t = this.tokens[0]; 119 this.tokens.splice(0,1); 137 this.tokens = this.tokens.concat(hiddenTokens) [all...] |
/external/webp/src/utils/ |
huffman_encode.c | 281 HuffmanTreeToken* tokens, 285 tokens->code = value; 286 tokens->extra_bits = 0; 287 ++tokens; 294 tokens->code = value; 295 tokens->extra_bits = 0; 296 ++tokens; 300 tokens->code = 16; 301 tokens->extra_bits = repetitions - 3; 302 ++tokens; [all...] |
/external/antlr/antlr-3.4/runtime/CSharp2/Sources/Antlr3.Runtime/Antlr.Runtime/ |
LegacyCommonTokenStream.cs | 43 * The most common stream of tokens is one where every token is buffered up 44 * and tokens are prefiltered for a certain channel (the parser will only 45 * see these tokens and cannot change the filter channel number during the 49 * <remarks>TODO: how to access the full token stream? How to track all tokens matched per rule?</remarks> 61 protected List<IToken> tokens; field in class:Antlr.Runtime.LegacyCommonTokenStream 63 /** <summary>Map<tokentype, channel> to override some Tokens' channel numbers</summary> */ 66 /** <summary>Set<tokentype>; discard any tokens with this type</summary> */ 69 /** <summary>Skip tokens on any channel but this one; this is how we skip whitespace...</summary> */ 72 /** <summary>By default, track all incoming tokens</summary> */ 79 * The index into the tokens list of the current token (next toke [all...] |
/external/antlr/antlr-3.4/runtime/CSharp3/Sources/Antlr3.Runtime/ |
LegacyCommonTokenStream.cs | 41 * The most common stream of tokens is one where every token is buffered up 42 * and tokens are prefiltered for a certain channel (the parser will only 43 * see these tokens and cannot change the filter channel number during the 47 * <remarks>TODO: how to access the full token stream? How to track all tokens matched per rule?</remarks> 60 protected List<IToken> tokens; field in class:Antlr.Runtime.LegacyCommonTokenStream 62 /** <summary>Map<tokentype, channel> to override some Tokens' channel numbers</summary> */ 65 /** <summary>Set<tokentype>; discard any tokens with this type</summary> */ 68 /** <summary>Skip tokens on any channel but this one; this is how we skip whitespace...</summary> */ 71 /** <summary>By default, track all incoming tokens</summary> */ 78 * The index into the tokens list of the current token (next toke [all...] |
/external/antlr/antlr-3.4/runtime/Java/src/main/java/org/antlr/runtime/ |
BufferedTokenStream.java | 35 /** Buffer all input tokens but do on-demand fetching of new tokens from 37 * proper lexing of future tokens. The ST template parser needs this, 43 * tokens to the parser. The stream can't ignore off-channel tokens. 47 * to confuse small moving window of tokens it uses for the full buffer. 57 protected List<Token> tokens = new ArrayList<Token>(100); field in class:BufferedTokenStream 62 /** The index into the tokens list of the current token (next token 63 * to consume). tokens[p] should be LT(1). p=-1 indicates need 108 public int size() { return tokens.size(); [all...] |
LegacyCommonTokenStream.java | 32 /** The most common stream of tokens is one where every token is buffered up 33 * and tokens are prefiltered for a certain channel (the parser will only 34 * see these tokens and cannot change the filter channel number during the 37 * TODO: how to access the full token stream? How to track all tokens matched per rule? 45 protected List tokens; field in class:LegacyCommonTokenStream 47 /** Map<tokentype, channel> to override some Tokens' channel numbers */ 50 /** Set<tokentype>; discard any tokens with this type */ 53 /** Skip tokens on any channel but this one; this is how we skip whitespace... */ 56 /** By default, track all incoming tokens */ 64 /** The index into the tokens list of the current token (next toke [all...] |
/packages/providers/ContactsProvider/src/com/android/providers/contacts/ |
NameSplitter.java | 234 // Iterate over tokens, skipping over empty ones and marking tokens that 319 * Parses a full name and returns components as a list of tokens. 321 public int tokenize(String[] tokens, String fullName) { 335 tokens[count++] = tokenizer.mTokens[i]; 395 NameTokenizer tokens = new NameTokenizer(fullName); local 396 parsePrefix(name, tokens); 398 // If the name consists of just one or two tokens, treat them as first/last name, 400 if (tokens.mEndPointer > 2) { 401 parseSuffix(name, tokens); [all...] |