/external/chromium/net/base/ |
pem_tokenizer_unittest.cc | 20 PEMTokenizer tokenizer(string_piece, accepted_types); 21 EXPECT_TRUE(tokenizer.GetNext()); 23 EXPECT_EQ("EXPECTED-BLOCK", tokenizer.block_type()); 24 EXPECT_EQ("MatchesAcceptedBlockType", tokenizer.data()); 26 EXPECT_FALSE(tokenizer.GetNext()); 38 PEMTokenizer tokenizer(string_piece, accepted_types); 39 EXPECT_TRUE(tokenizer.GetNext()); 41 EXPECT_EQ("EXPECTED-BLOCK", tokenizer.block_type()); 42 EXPECT_EQ("MatchesAcceptedBlockType", tokenizer.data()); 44 EXPECT_FALSE(tokenizer.GetNext()) [all...] |
/external/protobuf/src/google/protobuf/io/ |
tokenizer_unittest.cc | 39 #include <google/protobuf/io/tokenizer.h> 182 EXPECT_TRUE(Tokenizer::ParseInteger(text, kuint64max, &result)); 197 Tokenizer::TokenType type; 207 { "hello", Tokenizer::TYPE_IDENTIFIER }, 210 { "123", Tokenizer::TYPE_INTEGER }, 211 { "0xab6", Tokenizer::TYPE_INTEGER }, 212 { "0XAB6", Tokenizer::TYPE_INTEGER }, 213 { "0X1234567", Tokenizer::TYPE_INTEGER }, 214 { "0x89abcdef", Tokenizer::TYPE_INTEGER }, 215 { "0x89ABCDEF", Tokenizer::TYPE_INTEGER } [all...] |
/external/doclava/src/com/google/doclava/apicheck/ |
ApiFile.java | 67 final Tokenizer tokenizer = new Tokenizer(filename, (new String(buf, 0, size)).toCharArray()); local 71 String token = tokenizer.getToken(); 76 parsePackage(api, tokenizer); 78 throw new ApiParseException("expected package got " + token, tokenizer.getLine()); 88 private static void parsePackage(ApiInfo api, Tokenizer tokenizer) 94 token = tokenizer.requireToken(); 95 assertIdent(tokenizer, token) [all...] |
/frameworks/native/libs/utils/ |
Tokenizer.cpp | 17 #define LOG_TAG "Tokenizer" 26 #include <utils/Tokenizer.h> 28 // Enables debug output for the tokenizer. 38 Tokenizer::Tokenizer(const String8& filename, FileMap* fileMap, char* buffer, 45 Tokenizer::~Tokenizer() { 54 status_t Tokenizer::open(const String8& filename, Tokenizer** outTokenizer) { 97 *outTokenizer = new Tokenizer(filename, fileMap, buffer, ownBuffer, length) [all...] |
PropertyMap.cpp | 121 Tokenizer* tokenizer; local 122 status_t status = Tokenizer::open(filename, &tokenizer); 134 Parser parser(map, tokenizer); 139 tokenizer->getFilename().string(), tokenizer->getLineNumber(), 148 delete tokenizer; 156 PropertyMap::Parser::Parser(PropertyMap* map, Tokenizer* tokenizer) [all...] |
/external/protobuf/python/google/protobuf/internal/ |
text_format_test.py | 301 tokenizer = text_format._Tokenizer(text) 302 methods = [(tokenizer.ConsumeIdentifier, 'identifier1'), 304 (tokenizer.ConsumeString, 'string1'), 305 (tokenizer.ConsumeIdentifier, 'identifier2'), 307 (tokenizer.ConsumeInt32, 123), 308 (tokenizer.ConsumeIdentifier, 'identifier3'), 310 (tokenizer.ConsumeString, 'string'), 311 (tokenizer.ConsumeIdentifier, 'identifiER_4'), 313 (tokenizer.ConsumeFloat, 1.1e+2), 314 (tokenizer.ConsumeIdentifier, 'ID5') [all...] |