/external/mesa3d/src/glsl/glcpp/tests/ |
000-content-with-spaces.c | 1 this is four tokens
|
/external/apache-harmony/regex/src/test/java/org/apache/harmony/tests/java/util/regex/ |
SplitTest.java | 45 String tokens[]; local 47 tokens = p.split(input, 1); 48 assertEquals(1, tokens.length); 49 assertTrue(tokens[0].equals(input)); 50 tokens = p.split(input, 2); 51 assertEquals(2, tokens.length); 52 assertEquals("poodle", tokens[0]); 53 assertEquals("zoo", tokens[1]); 54 tokens = p.split(input, 5); 55 assertEquals(2, tokens.length) [all...] |
/libcore/luni/src/test/java/org/apache/harmony/regex/tests/java/util/regex/ |
SplitTest.java | 67 String tokens[]; local 69 tokens = p.split(input, 1); 70 assertEquals(1, tokens.length); 71 assertTrue(tokens[0].equals(input)); 72 tokens = p.split(input, 2); 73 assertEquals(2, tokens.length); 74 assertEquals("poodle", tokens[0]); 75 assertEquals("zoo", tokens[1]); 76 tokens = p.split(input, 5); 77 assertEquals(2, tokens.length) [all...] |
/external/qemu/ |
json-parser.h | 20 QObject *json_parser_parse(QList *tokens, va_list *ap);
|
json-streamer.h | 22 void (*emit)(struct JSONMessageParser *parser, QList *tokens); 26 QList *tokens; member in struct:JSONMessageParser
|
json-streamer.c | 52 qlist_append(parser->tokens, dict); 56 parser->emit(parser, parser->tokens); 57 QDECREF(parser->tokens); 58 parser->tokens = qlist_new(); 68 parser->tokens = qlist_new(); 87 QDECREF(parser->tokens);
|
/packages/providers/ContactsProvider/tests/src/com/android/providers/contacts/ |
HanziToPinyinTest.java | 43 ArrayList<Token> tokens = HanziToPinyin.getInstance().get(ONE_HANZI); local 44 assertEquals(tokens.size(), 1); 45 assertEquals(tokens.get(0).type, Token.PINYIN); 46 assertTrue(tokens.get(0).target.equalsIgnoreCase("DU")); 48 tokens = HanziToPinyin.getInstance().get(TWO_HANZI); 49 assertEquals(tokens.size(), 2); 50 assertEquals(tokens.get(0).type, Token.PINYIN); 51 assertEquals(tokens.get(1).type, Token.PINYIN); 52 assertTrue(tokens.get(0).target.equalsIgnoreCase("DU")); 53 assertTrue(tokens.get(1).target.equalsIgnoreCase("JUAN")) [all...] |
/dalvik/vm/compiler/template/ |
gen-template.py | 49 # Set handler_size_bytes to the value of tokens[1], and handler_size_bits to 53 def setHandlerSize(tokens): 55 if len(tokens) != 2: 61 handler_size_bytes = bytes = int(tokens[1]) 76 def importFile(tokens): 77 if len(tokens) != 2: 79 source = tokens[1] 81 appendSourceFile(tokens[1], getGlobalSubDict(), asm_fp, None) 90 def setAsmStub(tokens): 92 if len(tokens) != 2 385 tokens = line.split(' ') # tokenize variable [all...] |
/external/chromium/net/tools/flip_server/ |
balsa_headers_token_utils.cc | 13 BalsaHeaders::HeaderTokenList* tokens) { 14 CHECK(tokens); 24 ParseTokenList(value_begin, line_end, tokens); 46 BalsaHeaders::HeaderTokenList tokens; local 47 TokenizeHeaderLine(*headers, *header_line, &tokens); 49 if (tokens.empty()) { 53 } else if (tokens.size() == 1) { 57 const base::StringPiece& new_last_token = tokens[tokens.size() - 2]; 84 BalsaHeaders::HeaderTokenList tokens; local [all...] |
balsa_headers_token_utils.h | 5 // Utility class that performs basic operations on header value tokens: parsing 6 // them out, checking for presense of certain tokens, and removing them. 30 // with that key, all of them will be tokenized and tokens will be added to 34 BalsaHeaders::HeaderTokenList* tokens); 43 // in some buffer, populates tokens list with beginning and end indices 44 // of all tokens present in the value string. 47 BalsaHeaders::HeaderTokenList* tokens); 54 BalsaHeaders::HeaderTokenList* tokens);
|
/dalvik/vm/mterp/ |
gen-mterp.py | 58 def setHandlerStyle(tokens): 60 if len(tokens) != 2: 62 style = tokens[1] 68 # Set handler_size_bytes to the value of tokens[1], and handler_size_bits to 72 def setHandlerSize(tokens): 76 if len(tokens) != 2: 82 handler_size_bytes = bytes = int(tokens[1]) 97 def importFile(tokens): 98 if len(tokens) != 2: 100 source = tokens[1 569 tokens = line.split(' ') # tokenize variable [all...] |
/external/webkit/Source/WebCore/loader/ |
FTPDirectoryParser.cpp | 119 const char *tokens[16]; /* 16 is more than enough */ local 120 unsigned int toklen[WTF_ARRAY_LENGTH(tokens)]; 130 tokens[0] = state.carryBuffer; 136 while (pos < linelen && numtoks < WTF_ARRAY_LENGTH(tokens)) 143 tokens[numtoks] = &line[pos]; 147 if (tokens[numtoks] != &line[pos]) 149 toklen[numtoks] = (&line[pos] - tokens[numtoks]); 158 linelen_sans_wsp = &(tokens[numtoks-1][toklen[numtoks-1]]) - tokens[0]; 159 if (numtoks == WTF_ARRAY_LENGTH(tokens)) [all...] |
/external/clang/test/Preprocessor/ |
macro_paste_msextensions.c | 12 #define comment /##/ dead tokens live here 19 #define nested(x) int x comment cute little dead tokens... 21 nested(baz) rise of the dead tokens
|
/cts/tools/dasm/src/dasm/tokens/ |
relative_number_token.java | 17 package dasm.tokens;
|
number_token.java | 17 package dasm.tokens;
|
/external/antlr/src/org/antlr/runtime/ |
BufferedTokenStream.java | 35 /** Buffer all input tokens but do on-demand fetching of new tokens from 37 * proper lexing of future tokens. The ST template parser needs this, 43 * tokens to the parser. The stream can't ignore off-channel tokens. 47 * to confuse small moving window of tokens it uses for the full buffer. 57 protected List<Token> tokens = new ArrayList<Token>(100); field in class:BufferedTokenStream 62 /** The index into the tokens list of the current token (next token 63 * to consume). tokens[p] should be LT(1). p=-1 indicates need 108 public int size() { return tokens.size(); [all...] |
LegacyCommonTokenStream.java | 32 /** The most common stream of tokens is one where every token is buffered up 33 * and tokens are prefiltered for a certain channel (the parser will only 34 * see these tokens and cannot change the filter channel number during the 37 * TODO: how to access the full token stream? How to track all tokens matched per rule? 45 protected List tokens; field in class:LegacyCommonTokenStream 47 /** Map<tokentype, channel> to override some Tokens' channel numbers */ 50 /** Set<tokentype>; discard any tokens with this type */ 53 /** Skip tokens on any channel but this one; this is how we skip whitespace... */ 56 /** By default, track all incoming tokens */ 64 /** The index into the tokens list of the current token (next toke [all...] |
CommonTokenStream.java | 31 /** The most common stream of tokens where every token is buffered up 32 * and tokens are filtered for a certain channel (the parser will only 33 * see these tokens). 35 * Even though it buffers all of the tokens, this token stream pulls tokens 36 * from the tokens source on demand. In other words, until you ask for a 40 * is that this stream knows how to ignore off channel tokens. There may be 49 /** Skip tokens on any channel but this one; this is how we skip whitespace... */ 68 while ( tokens.get(p).getChannel()!=channel ) { 79 // find k good tokens looking backward [all...] |
/external/nist-sip/java/gov/nist/javax/sip/header/ |
Server.java | 53 /** Product tokens. 62 StringBuffer tokens = new StringBuffer(); local 66 tokens.append((String) it.next()); 68 tokens.append('/'); 72 return tokens.toString();
|
/frameworks/base/opengl/libagl/ |
TokenManager.cpp | 35 status_t TokenManager::getToken(GLsizei n, GLuint *tokens) 39 *tokens++ = mTokenizer.acquire(); 43 void TokenManager::recycleTokens(GLsizei n, const GLuint *tokens) 47 const GLuint token = *tokens++;
|
TokenManager.h | 40 status_t getToken(GLsizei n, GLuint *tokens); 41 void recycleTokens(GLsizei n, const GLuint *tokens);
|
/packages/providers/ContactsProvider/src/com/android/providers/contacts/ |
NameSplitter.java | 230 // Iterate over tokens, skipping over empty ones and marking tokens that 315 * Parses a full name and returns components as a list of tokens. 317 public int tokenize(String[] tokens, String fullName) { 334 tokens[count++] = tokenizer.mTokens[i]; 394 NameTokenizer tokens = new NameTokenizer(fullName); local 395 parsePrefix(name, tokens); 397 // If the name consists of just one or two tokens, treat them as first/last name, 399 if (tokens.mEndPointer > 2) { 400 parseSuffix(name, tokens); [all...] |
/frameworks/ex/common/java/com/android/common/ |
Rfc822Validator.java | 66 Rfc822Token[] tokens = Rfc822Tokenizer.tokenize(text); local 68 return tokens.length == 1 && 70 matcher(tokens[0].getAddress()).matches(); 74 * Specify if the validator should remove invalid tokens instead of trying 76 * tokens. 78 * @param remove true to remove tokens with the wrong format, false to 124 Rfc822Token[] tokens = Rfc822Tokenizer.tokenize(cs); 127 for (int i = 0; i < tokens.length; i++) { 128 String text = tokens[i].getAddress(); 137 tokens[i].setAddress(removeIllegalCharacters(text) + "@" + mDomain) [all...] |
/external/bluetooth/glib/gio/ |
gicon.c | 166 GPtrArray *tokens; local 180 tokens = g_ptr_array_new (); 181 if (!icon_iface->to_tokens (icon, tokens, &version)) 183 g_ptr_array_free (tokens, TRUE); 189 all the tokens are url escaped to ensure they have no spaces in them */ 195 for (i = 0; i < tokens->len; i++) 199 token = g_ptr_array_index (tokens, i); 209 g_ptr_array_free (tokens, TRUE); 301 g_icon_new_from_tokens (char **tokens, 317 num_tokens = g_strv_length (tokens); 458 gchar **tokens; local [all...] |
/frameworks/base/opengl/tools/glgen/src/ |
CFunc.java | 112 String[] tokens = s.split("\\s"); local 116 String ftypeName = tokens[i++]; 119 ftypeName = tokens[i++]; 123 String fname = tokens[i++]; 126 fname = tokens[i++]; 132 while (i < tokens.length) { 133 String tok = tokens[i++]; 149 argTypeName = tokens[i++]; 157 argName = tokens[i++];
|