HomeSort by relevance Sort by last modified time
    Searched refs:tokens (Results 1 - 25 of 429) sorted by null

1 2 3 4 5 6 7 8 91011>>

  /external/mesa3d/src/glsl/glcpp/tests/
000-content-with-spaces.c 1 this is four tokens
  /external/antlr/antlr-3.4/runtime/Python/tests/
t048rewrite.py 25 tokens = self._parse("abc")
26 tokens.insertBefore(0, "0")
28 result = tokens.toString()
34 tokens = self._parse("abc")
35 tokens.insertAfter(2, "x")
37 result = tokens.toString()
43 tokens = self._parse("abc")
44 tokens.insertBefore(1, "x")
45 tokens.insertAfter(1, "x")
47 result = tokens.toString(
    [all...]
  /external/apache-harmony/regex/src/test/java/org/apache/harmony/tests/java/util/regex/
SplitTest.java 45 String tokens[]; local
47 tokens = p.split(input, 1);
48 assertEquals(1, tokens.length);
49 assertTrue(tokens[0].equals(input));
50 tokens = p.split(input, 2);
51 assertEquals(2, tokens.length);
52 assertEquals("poodle", tokens[0]);
53 assertEquals("zoo", tokens[1]);
54 tokens = p.split(input, 5);
55 assertEquals(2, tokens.length)
    [all...]
  /libcore/luni/src/test/java/org/apache/harmony/regex/tests/java/util/regex/
SplitTest.java 67 String tokens[]; local
69 tokens = p.split(input, 1);
70 assertEquals(1, tokens.length);
71 assertTrue(tokens[0].equals(input));
72 tokens = p.split(input, 2);
73 assertEquals(2, tokens.length);
74 assertEquals("poodle", tokens[0]);
75 assertEquals("zoo", tokens[1]);
76 tokens = p.split(input, 5);
77 assertEquals(2, tokens.length)
    [all...]
  /external/qemu/
json-parser.h 20 QObject *json_parser_parse(QList *tokens, va_list *ap);
  /packages/providers/ContactsProvider/tests/src/com/android/providers/contacts/
HanziToPinyinTest.java 43 ArrayList<Token> tokens = HanziToPinyin.getInstance().get(ONE_HANZI); local
44 assertEquals(tokens.size(), 1);
45 assertEquals(tokens.get(0).type, Token.PINYIN);
46 assertTrue(tokens.get(0).target.equalsIgnoreCase("DU"));
48 tokens = HanziToPinyin.getInstance().get(TWO_HANZI);
49 assertEquals(tokens.size(), 2);
50 assertEquals(tokens.get(0).type, Token.PINYIN);
51 assertEquals(tokens.get(1).type, Token.PINYIN);
52 assertTrue(tokens.get(0).target.equalsIgnoreCase("DU"));
53 assertTrue(tokens.get(1).target.equalsIgnoreCase("JUAN"))
    [all...]
  /external/antlr/antlr-3.4/tool/src/test/java/org/antlr/test/
TestCommonTokenStream.java 48 // Tokens: 012345678901234567
52 BufferedTokenStream tokens = new BufferedTokenStream(lexEngine); local
54 String result = tokens.LT(1).getText();
69 // Tokens: 012345678901234567
73 BufferedTokenStream tokens = new BufferedTokenStream(lexEngine); local
75 String result = tokens.LT(2).getText();
90 // Tokens: 012345678901234567
94 BufferedTokenStream tokens = new BufferedTokenStream(lexEngine); local
97 Token t = tokens.LT(i);
100 t = tokens.LT(i)
124 BufferedTokenStream tokens = new BufferedTokenStream(lexEngine); local
155 BufferedTokenStream tokens = new BufferedTokenStream(lexEngine); local
200 CommonTokenStream tokens = new CommonTokenStream(lexer); local
    [all...]
TestTokenRewriteStream.java 51 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
52 tokens.insertBefore(0, "0");
53 String result = tokens.toString();
66 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
67 tokens.insertAfter(2, "x");
68 String result = tokens.toString();
81 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
82 tokens.fill();
83 tokens.insertBefore(1, "x");
84 tokens.insertAfter(1, "x")
98 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
114 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
130 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
151 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
186 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
233 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
250 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
268 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
285 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
309 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
326 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
343 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
361 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
378 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
395 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
412 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
429 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
453 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
470 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
486 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
502 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
526 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
550 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
569 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
586 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
604 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
621 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
638 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
656 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
673 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
697 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
714 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
731 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
748 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
765 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
782 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
799 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); local
    [all...]
  /dalvik/vm/compiler/template/
gen-template.py 49 # Set handler_size_bytes to the value of tokens[1], and handler_size_bits to
53 def setHandlerSize(tokens):
55 if len(tokens) != 2:
61 handler_size_bytes = bytes = int(tokens[1])
76 def importFile(tokens):
77 if len(tokens) != 2:
79 source = tokens[1]
81 appendSourceFile(tokens[1], getGlobalSubDict(), asm_fp, None)
90 def setAsmStub(tokens):
92 if len(tokens) != 2
385 tokens = line.split(' ') # tokenize variable
    [all...]
  /external/chromium/net/tools/flip_server/
balsa_headers_token_utils.cc 13 BalsaHeaders::HeaderTokenList* tokens) {
14 CHECK(tokens);
24 ParseTokenList(value_begin, line_end, tokens);
46 BalsaHeaders::HeaderTokenList tokens; local
47 TokenizeHeaderLine(*headers, *header_line, &tokens);
49 if (tokens.empty()) {
53 } else if (tokens.size() == 1) {
57 const base::StringPiece& new_last_token = tokens[tokens.size() - 2];
84 BalsaHeaders::HeaderTokenList tokens; local
    [all...]
balsa_headers_token_utils.h 5 // Utility class that performs basic operations on header value tokens: parsing
6 // them out, checking for presense of certain tokens, and removing them.
30 // with that key, all of them will be tokenized and tokens will be added to
34 BalsaHeaders::HeaderTokenList* tokens);
43 // in some buffer, populates tokens list with beginning and end indices
44 // of all tokens present in the value string.
47 BalsaHeaders::HeaderTokenList* tokens);
54 BalsaHeaders::HeaderTokenList* tokens);
  /dalvik/vm/mterp/
gen-mterp.py 58 def setHandlerStyle(tokens):
60 if len(tokens) != 2:
62 style = tokens[1]
68 # Set handler_size_bytes to the value of tokens[1], and handler_size_bits to
72 def setHandlerSize(tokens):
76 if len(tokens) != 2:
82 handler_size_bytes = bytes = int(tokens[1])
97 def importFile(tokens):
98 if len(tokens) != 2:
100 source = tokens[1
569 tokens = line.split(' ') # tokenize variable
    [all...]
  /external/webkit/Source/WebCore/loader/
FTPDirectoryParser.cpp 119 const char *tokens[16]; /* 16 is more than enough */ local
120 unsigned int toklen[WTF_ARRAY_LENGTH(tokens)];
130 tokens[0] = state.carryBuffer;
136 while (pos < linelen && numtoks < WTF_ARRAY_LENGTH(tokens))
143 tokens[numtoks] = &line[pos];
147 if (tokens[numtoks] != &line[pos])
149 toklen[numtoks] = (&line[pos] - tokens[numtoks]);
158 linelen_sans_wsp = &(tokens[numtoks-1][toklen[numtoks-1]]) - tokens[0];
159 if (numtoks == WTF_ARRAY_LENGTH(tokens))
    [all...]
  /cts/suite/audio_quality/lib/src/task/
TaskSound.cpp 64 UniquePtr<std::vector<android::String8> > tokens(StringUtil::split(type, ':'));
65 if (tokens.get() == NULL) {
70 if (StringUtil::compare(tokens->at(0), "file") == 0) {
71 if (tokens->size() != 2) {
72 LOGE("Wrong number of parameters %d", tokens->size());
74 buffer = Buffer::loadFromFile(tokens->at(1));
75 } else if (StringUtil::compare(tokens->at(0), "sin") == 0) {
76 if (tokens->size() != 4) {
77 LOGE("Wrong number of parameters %d", tokens->size());
79 int amplitude = atoi(tokens->at(1).string())
    [all...]
  /external/jmonkeyengine/engine/src/core-plugins/com/jme3/font/plugins/
BitmapFontLoader.java 64 String[] tokens = line.split(regex); local
65 if (tokens[0].equals("info")){
67 for (int i = 1; i < tokens.length; i++){
68 if (tokens[i].equals("size")){
69 charSet.setRenderedSize(Integer.parseInt(tokens[i + 1]));
72 }else if (tokens[0].equals("common")){
74 for (int i = 1; i < tokens.length; i++){
75 String token = tokens[i];
77 charSet.setLineHeight(Integer.parseInt(tokens[i + 1]));
79 charSet.setBase(Integer.parseInt(tokens[i + 1]))
    [all...]
  /external/antlr/antlr-3.4/runtime/Perl5/examples/tweak/
tweak.pl 13 my $tokens = ANTLR::Runtime::TokenRewriteStream({ token_source => $lexer });
14 my $parser = TParser->new({ input => $tokens });
16 print "$tokens\n";
  /external/clang/test/Preprocessor/
macro_paste_msextensions.c 12 #define comment /##/ dead tokens live here
19 #define nested(x) int x comment cute little dead tokens...
21 nested(baz) rise of the dead tokens
  /cts/tools/dasm/src/dasm/tokens/
relative_number_token.java 17 package dasm.tokens;
number_token.java 17 package dasm.tokens;
  /frameworks/base/wifi/java/android/net/wifi/p2p/
WifiP2pProvDiscEvent.java 62 String[] tokens = string.split(" "); local
64 if (tokens.length < 2) {
68 if (tokens[0].endsWith("PBC-REQ")) event = PBC_REQ;
69 else if (tokens[0].endsWith("PBC-RESP")) event = PBC_RSP;
70 else if (tokens[0].endsWith("ENTER-PIN")) event = ENTER_PIN;
71 else if (tokens[0].endsWith("SHOW-PIN")) event = SHOW_PIN;
76 device.deviceAddress = tokens[1];
79 pin = tokens[2];
  /external/antlr/antlr-3.4/runtime/JavaScript/tests/functional/
rhino-python.extensions 33 tokens should be sent to the parser sometimes without a corresponding
48 This TokenStream normally just passes tokens through to the parser.
63 A queue of tokens is built up to hold multiple DEDENT tokens that
77 /** The queue of tokens */
78 this.tokens = [];
117 if (this.tokens.length>0 ) {
118 var t = this.tokens[0];
119 this.tokens.splice(0,1);
137 this.tokens = this.tokens.concat(hiddenTokens)
    [all...]
  /external/antlr/antlr-3.4/runtime/CSharp2/Sources/Antlr3.Runtime/Antlr.Runtime/
LegacyCommonTokenStream.cs 43 * The most common stream of tokens is one where every token is buffered up
44 * and tokens are prefiltered for a certain channel (the parser will only
45 * see these tokens and cannot change the filter channel number during the
49 * <remarks>TODO: how to access the full token stream? How to track all tokens matched per rule?</remarks>
61 protected List<IToken> tokens; field in class:Antlr.Runtime.LegacyCommonTokenStream
63 /** <summary>Map<tokentype, channel> to override some Tokens' channel numbers</summary> */
66 /** <summary>Set<tokentype>; discard any tokens with this type</summary> */
69 /** <summary>Skip tokens on any channel but this one; this is how we skip whitespace...</summary> */
72 /** <summary>By default, track all incoming tokens</summary> */
79 * The index into the tokens list of the current token (next toke
    [all...]
  /external/antlr/antlr-3.4/runtime/CSharp3/Sources/Antlr3.Runtime/
LegacyCommonTokenStream.cs 41 * The most common stream of tokens is one where every token is buffered up
42 * and tokens are prefiltered for a certain channel (the parser will only
43 * see these tokens and cannot change the filter channel number during the
47 * <remarks>TODO: how to access the full token stream? How to track all tokens matched per rule?</remarks>
60 protected List<IToken> tokens; field in class:Antlr.Runtime.LegacyCommonTokenStream
62 /** <summary>Map<tokentype, channel> to override some Tokens' channel numbers</summary> */
65 /** <summary>Set<tokentype>; discard any tokens with this type</summary> */
68 /** <summary>Skip tokens on any channel but this one; this is how we skip whitespace...</summary> */
71 /** <summary>By default, track all incoming tokens</summary> */
78 * The index into the tokens list of the current token (next toke
    [all...]
  /external/antlr/antlr-3.4/runtime/Java/src/main/java/org/antlr/runtime/
BufferedTokenStream.java 35 /** Buffer all input tokens but do on-demand fetching of new tokens from
37 * proper lexing of future tokens. The ST template parser needs this,
43 * tokens to the parser. The stream can't ignore off-channel tokens.
47 * to confuse small moving window of tokens it uses for the full buffer.
57 protected List<Token> tokens = new ArrayList<Token>(100); field in class:BufferedTokenStream
62 /** The index into the tokens list of the current token (next token
63 * to consume). tokens[p] should be LT(1). p=-1 indicates need
108 public int size() { return tokens.size();
    [all...]
LegacyCommonTokenStream.java 32 /** The most common stream of tokens is one where every token is buffered up
33 * and tokens are prefiltered for a certain channel (the parser will only
34 * see these tokens and cannot change the filter channel number during the
37 * TODO: how to access the full token stream? How to track all tokens matched per rule?
45 protected List tokens; field in class:LegacyCommonTokenStream
47 /** Map<tokentype, channel> to override some Tokens' channel numbers */
50 /** Set<tokentype>; discard any tokens with this type */
53 /** Skip tokens on any channel but this one; this is how we skip whitespace... */
56 /** By default, track all incoming tokens */
64 /** The index into the tokens list of the current token (next toke
    [all...]

Completed in 759 milliseconds

1 2 3 4 5 6 7 8 91011>>