/external/smack/src/org/xbill/DNS/ |
HINFORecord.java | 51 rdataFromString(Tokenizer st, Name origin) throws IOException {
|
LOCRecord.java | 94 parsePosition(Tokenizer st, String type) throws IOException { 135 parseDouble(Tokenizer st, String type, boolean required, long min, long max, 139 Tokenizer.Token token = st.get(); 161 rdataFromString(Tokenizer st, Name origin) throws IOException {
|
MINFORecord.java | 54 rdataFromString(Tokenizer st, Name origin) throws IOException {
|
NSAPRecord.java | 81 rdataFromString(Tokenizer st, Name origin) throws IOException {
|
PXRecord.java | 53 rdataFromString(Tokenizer st, Name origin) throws IOException {
|
SSHFPRecord.java | 66 rdataFromString(Tokenizer st, Name origin) throws IOException {
|
X25Record.java | 61 rdataFromString(Tokenizer st, Name origin) throws IOException {
|
/cts/tests/tests/widget/src/android/widget/cts/ |
MultiAutoCompleteTextViewTest.java | 38 import android.widget.MultiAutoCompleteTextView.Tokenizer; 132 // There will be an endless loop when using CommaTokenizer as the Tokenizer 199 private class MockTokenizer implements Tokenizer {
|
/external/chromium/net/ftp/ |
ftp_directory_listing_parser_unittest.cc | 100 StringTokenizer tokenizer(expected_listing, "\r\n"); 101 while (tokenizer.GetNext()) 102 lines.push_back(tokenizer.token());
|
ftp_util.cc | 30 StringTokenizer tokenizer(unix_path, "/"); 32 while (tokenizer.GetNext()) 33 tokens.push_back(tokenizer.token());
|
/external/chromium_org/third_party/angle_dx11/src/compiler/preprocessor/ |
preprocessor.vcxproj | 147 <ClCompile Include="Tokenizer.cpp" />
164 <ClInclude Include="Tokenizer.h" />
167 <None Include="Tokenizer.l" />
|
/external/chromium_org/tools/gn/ |
err.cc | 13 #include "tools/gn/tokenizer.h" 19 size_t line_off = Tokenizer::ByteOffsetOfNthLine(data, n); 21 while (end < data.size() && !Tokenizer::IsNewline(data, end))
|
string_utils.cc | 10 #include "tools/gn/tokenizer.h" 63 if (!Tokenizer::IsIdentifierFirstChar(input[*i])) { 74 while (*i < size && Tokenizer::IsIdentifierContinuingChar(input[*i]))
|
/external/emma/core/java12/com/vladium/emma/filter/ |
IInclExclFilter.java | 54 for (StringTokenizer tokenizer = new StringTokenizer (specs1, separators); 55 tokenizer.hasMoreTokens (); ) 57 _specs.add (tokenizer.nextToken ());
|
/external/emma/core/java12/com/vladium/emma/rt/ |
RTExitHook.java | 84 final StringTokenizer tokenizer = new StringTokenizer (closureList, ","); local 85 while (tokenizer.hasMoreTokens ()) 87 final String className = tokenizer.nextToken ();
|
/external/proguard/src/proguard/ant/ |
ClassSpecificationElement.java | 210 StringTokenizer tokenizer = new StringTokenizer(access, " ,"); local 211 while (tokenizer.hasMoreTokens()) 213 String token = tokenizer.nextToken();
|
MemberSpecificationElement.java | 178 StringTokenizer tokenizer = new StringTokenizer(access, " ,"); local 179 while (tokenizer.hasMoreTokens()) 181 String token = tokenizer.nextToken();
|
/libcore/luni/src/main/java/java/io/ |
StreamTokenizer.java | 177 * The tokenizer's initial state is as follows: 217 * as {@code TT_EOF} in {@code ttype} by this tokenizer. 229 * @return this tokenizer's current line number. 248 * Parses the next token from this tokenizer's source stream or reader. The 485 * character by this tokenizer. That is, it has no special meaning as a 500 * shall be treated as an ordinary character by this tokenizer. That is, 522 * Specifies that this tokenizer shall parse numbers. 596 * Returns the state of this tokenizer in a readable format. 598 * @return the current state of this tokenizer. 635 * shall be treated as whitespace characters by this tokenizer [all...] |
/external/chromium-trace/trace-viewer/third_party/closure_linter/closure_linter/ |
javascripttokenizer.py | 27 from closure_linter.common import tokenizer namespace 47 class JavaScriptTokenizer(tokenizer.Tokenizer): 48 """JavaScript tokenizer. 334 """Create a tokenizer object. 347 tokenizer.Tokenizer.__init__(self, JavaScriptModes.TEXT_MODE, matchers,
|
/external/chromium_org/third_party/closure_linter/closure_linter/ |
javascripttokenizer.py | 27 from closure_linter.common import tokenizer namespace 47 class JavaScriptTokenizer(tokenizer.Tokenizer): 48 """JavaScript tokenizer. 334 """Create a tokenizer object. 347 tokenizer.Tokenizer.__init__(self, JavaScriptModes.TEXT_MODE, matchers,
|
/external/antlr/antlr-3.4/runtime/ObjC/Framework/ |
ANTLRTreeWizard.m | 369 ANTLRTreePatternLexer *tokenizer = [ANTLRTreePatternLexer newANTLRTreePatternLexer:pattern]; 370 ANTLRTreePatternParser *parser = [ANTLRTreePatternParser newANTLRTreePatternParser:tokenizer 444 ANTLRTreePatternLexer *tokenizer = [ANTLRTreePatternLexer newANTLRTreePatternLexer:pattern]; 446 [ANTLRTreePatternParser newANTLRTreePatternParser:tokenizer Wizard:self Adaptor:[ANTLRTreePatternTreeAdaptor newTreeAdaptor]]; 487 TreePatternLexer tokenizer = new TreePatternLexer(pattern); 489 new TreePatternParser(tokenizer, this, new TreePatternTreeAdaptor()); 498 ANTLRTreePatternLexer *tokenizer = [ANTLRTreePatternLexer newANTLRTreePatternLexer:pattern]; 499 ANTLRTreePatternParser *parser = [ANTLRTreePatternParser newANTLRTreePatternParser:tokenizer 576 ANTLRTreePatternLexer *tokenizer = [ANTLRTreePatternLexer newANTLRTreePatternLexer:pattern]; 577 ANTLRTreePatternParser *parser = [ANTLRTreePatternParser newANTLRTreePatternParser:tokenizer Wizard:self Adaptor:adaptor] [all...] |
/external/chromium_org/net/http/ |
http_util.cc | 117 base::StringTokenizer tokenizer(begin + param_start, content_type_str.end(), 119 tokenizer.set_quote_chars("\""); 120 while (tokenizer.GetNext()) { 122 std::find(tokenizer.token_begin(), tokenizer.token_end(), '='); 123 if (equals_sign == tokenizer.token_end()) 126 string::const_iterator param_name_begin = tokenizer.token_begin(); 131 string::const_iterator param_value_end = tokenizer.token_end(); 132 DCHECK(param_value_begin <= tokenizer.token_end()); 643 base::StringTokenizer tokenizer(str, std::string(1, '\0')) [all...] |
/external/chromium_org/third_party/protobuf/src/google/protobuf/compiler/ |
importer.cc | 50 #include <google/protobuf/io/tokenizer.h> 82 // - It implements the ErrorCollector interface (used by Tokenizer and Parser) 132 // Set up the tokenizer and parser. 134 io::Tokenizer tokenizer(input.get(), &file_error_collector); 146 return parser.Parse(&tokenizer, output) &&
|
/external/protobuf/src/google/protobuf/compiler/ |
importer.cc | 50 #include <google/protobuf/io/tokenizer.h> 82 // - It implements the ErrorCollector interface (used by Tokenizer and Parser) 132 // Set up the tokenizer and parser. 134 io::Tokenizer tokenizer(input.get(), &file_error_collector); 146 return parser.Parse(&tokenizer, output) &&
|
/development/samples/HoneycombGallery/src/com/example/android/hcgallery/ |
ContentFragment.java | 247 StringTokenizer tokenizer = new StringTokenizer(textData, "||"); local 248 if (tokenizer.countTokens() != 2) { 254 category = Integer.parseInt(tokenizer.nextToken()); 255 entryId = Integer.parseInt(tokenizer.nextToken());
|