/frameworks/native/opengl/libagl/ |
Tokenizer.h | 1 /* libs/opengles/Tokenizer.h 29 class Tokenizer 32 Tokenizer(); 33 Tokenizer(const Tokenizer& other); 34 ~Tokenizer();
|
Tokenizer.cpp | 1 /* libs/opengles/Tokenizer.cpp 20 #include "Tokenizer.h" 26 ANDROID_BASIC_TYPES_TRAITS(Tokenizer::run_t) 28 Tokenizer::Tokenizer() 32 Tokenizer::Tokenizer(const Tokenizer& other) 37 Tokenizer::~Tokenizer() [all...] |
/external/chromium-trace/trace-viewer/third_party/closure_linter/closure_linter/common/ |
tokenizer.py | 28 class Tokenizer(object): 29 """General purpose tokenizer. 32 mode: The latest mode of the tokenizer. This allows patterns to distinguish 41 """Initialize the tokenizer.
|
/external/chromium_org/third_party/closure_linter/closure_linter/common/ |
tokenizer.py | 28 class Tokenizer(object): 29 """General purpose tokenizer. 32 mode: The latest mode of the tokenizer. This allows patterns to distinguish 41 """Initialize the tokenizer.
|
/external/chromium_org/tools/gn/ |
tokenizer.cc | 5 #include "tools/gn/tokenizer.h" 70 Tokenizer::Tokenizer(const InputFile* input_file, Err* err) 79 Tokenizer::~Tokenizer() { 83 std::vector<Token> Tokenizer::Tokenize(const InputFile* input_file, Err* err) { 84 Tokenizer t(input_file, err); 88 std::vector<Token> Tokenizer::Run() { 136 size_t Tokenizer::ByteOffsetOfNthLine(const base::StringPiece& buf, int n) { 157 bool Tokenizer::IsNewline(const base::StringPiece& buffer, size_t offset) [all...] |
tokenizer.h | 18 class Tokenizer { 27 // This is a helper function for error output so that the tokenizer's 45 // InputFile must outlive the tokenizer and all generated tokens. 46 explicit Tokenizer(const InputFile* input_file, Err* err); 47 ~Tokenizer(); 86 DISALLOW_COPY_AND_ASSIGN(Tokenizer);
|
/system/core/libutils/ |
Tokenizer.cpp | 17 #define LOG_TAG "Tokenizer" 26 #include <utils/Tokenizer.h> 28 // Enables debug output for the tokenizer. 38 Tokenizer::Tokenizer(const String8& filename, FileMap* fileMap, char* buffer, 45 Tokenizer::~Tokenizer() { 54 status_t Tokenizer::open(const String8& filename, Tokenizer** outTokenizer) { 97 *outTokenizer = new Tokenizer(filename, fileMap, buffer, ownBuffer, length) [all...] |
/external/chromium_org/third_party/angle/src/compiler/preprocessor/ |
Tokenizer.h | 19 class Tokenizer : public Lexer 36 Tokenizer(Diagnostics *diagnostics); 37 ~Tokenizer(); 48 PP_DISALLOW_COPY_AND_ASSIGN(Tokenizer);
|
Tokenizer.cpp | 1 #line 16 "./Tokenizer.l" 12 #line 13 "./Tokenizer.cpp" 542 #include "Tokenizer.h" 588 #define YY_EXTRA_TYPE pp::Tokenizer::Context* [all...] |
/external/clang/include/clang/ASTMatchers/Dynamic/ |
Parser.h | 182 Parser(CodeTokenizer *Tokenizer, Sema *S, 194 CodeTokenizer *const Tokenizer;
|
/external/deqp/executor/ |
xeXMLParser.cpp | 57 Tokenizer::Tokenizer (void) 65 Tokenizer::~Tokenizer (void) 69 void Tokenizer::clear (void) 77 void Tokenizer::error (const std::string& what) 82 void Tokenizer::feed (const deUint8* bytes, int numBytes) 98 int Tokenizer::getChar (int offset) const 108 void Tokenizer::advance (void) 338 void Tokenizer::getString (std::string& dst) cons [all...] |
xeXMLParser.hpp | 83 class Tokenizer 86 Tokenizer (void); 87 ~Tokenizer (void); 89 void clear (void); //!< Resets tokenizer to initial state. 103 Tokenizer (const Tokenizer& other); 104 Tokenizer& operator= (const Tokenizer& other); 190 Tokenizer m_tokenizer; 203 inline void Tokenizer::getTokenStr (std::string& dst) cons [all...] |
/system/core/include/utils/ |
Tokenizer.h | 28 * A simple tokenizer for loading and parsing ASCII text files line by line. 30 class Tokenizer { 31 Tokenizer(const String8& filename, FileMap* fileMap, char* buffer, 35 ~Tokenizer(); 40 * Returns NO_ERROR and a tokenizer for the file, if successful. 43 static status_t open(const String8& filename, Tokenizer** outTokenizer); 48 * Returns NO_ERROR and a tokenizer for the string, if successful. 52 const char* contents, Tokenizer** outTokenizer); 119 Tokenizer(const Tokenizer& other); // not copyabl [all...] |
/external/protobuf/src/google/protobuf/io/ |
tokenizer.cc | 91 #include <google/protobuf/io/tokenizer.h> 105 // For instance, Tokenizer::ConsumeZeroOrMore<Whitespace>() will eat 180 Tokenizer::Tokenizer(ZeroCopyInputStream* input, 201 Tokenizer::~Tokenizer() { 212 void Tokenizer::NextChar() { 233 void Tokenizer::Refresh() { 263 inline void Tokenizer::StartToken() { 271 inline void Tokenizer::EndToken() [all...] |
tokenizer.h | 51 class Tokenizer; 81 class LIBPROTOBUF_EXPORT Tokenizer { 83 // Construct a Tokenizer that reads and tokenizes text from the given 86 Tokenizer(ZeroCopyInputStream* input, ErrorCollector* error_collector); 87 ~Tokenizer(); 138 // comes from a TYPE_FLOAT token parsed by Tokenizer. If it doesn't, the 143 // comes from a TYPE_STRING token parsed by Tokenizer. If it doesn't, the 153 // parsed by a Tokenizer, the result is undefined (possibly an assert 180 GOOGLE_DISALLOW_EVIL_CONSTRUCTORS(Tokenizer); 290 inline const Tokenizer::Token& Tokenizer::current() [all...] |
/external/chromium_org/third_party/protobuf/src/google/protobuf/io/ |
tokenizer.cc | 91 #include <google/protobuf/io/tokenizer.h> 108 // For instance, Tokenizer::ConsumeZeroOrMore<Whitespace>() will eat 185 Tokenizer::Tokenizer(ZeroCopyInputStream* input, 208 Tokenizer::~Tokenizer() { 219 void Tokenizer::NextChar() { 240 void Tokenizer::Refresh() { 270 inline void Tokenizer::RecordTo(string* target) { 275 inline void Tokenizer::StopRecording() [all...] |
tokenizer.h | 52 class Tokenizer; 82 class LIBPROTOBUF_EXPORT Tokenizer { 84 // Construct a Tokenizer that reads and tokenizes text from the given 87 Tokenizer(ZeroCopyInputStream* input, ErrorCollector* error_collector); 88 ~Tokenizer(); 191 // comes from a TYPE_FLOAT token parsed by Tokenizer. If it doesn't, the 196 // comes from a TYPE_STRING token parsed by Tokenizer. If it doesn't, the 206 // parsed by a Tokenizer, the result is undefined (possibly an assert 233 GOOGLE_DISALLOW_EVIL_CONSTRUCTORS(Tokenizer); 367 inline const Tokenizer::Token& Tokenizer::current() [all...] |
/frameworks/base/cmds/appwidget/src/com/android/commands/appwidget/ |
AppWidget.java | 62 private final Tokenizer mTokenizer; 65 mTokenizer = new Tokenizer(args); 119 private static class Tokenizer { 123 public Tokenizer(String[] args) {
|
/frameworks/base/core/java/android/widget/ |
MultiAutoCompleteTextView.java | 34 * You must provide a {@link Tokenizer} to distinguish the 60 private Tokenizer mTokenizer; 82 * Sets the Tokenizer that will be used to determine the relevant 85 public void setTokenizer(Tokenizer t) { 92 * {@link Tokenizer#findTokenStart} to {@link #getSelectionEnd} 116 * {@link Tokenizer#findTokenStart} to {@link #getSelectionEnd} 180 * {@link Tokenizer#findTokenStart} to {@link #getSelectionEnd} by the 182 * {@link Tokenizer#terminateToken}. 217 public static interface Tokenizer { 238 * This simple Tokenizer can be used for lists where the items ar [all...] |
/prebuilts/python/darwin-x86/2.7.5/lib/python2.7/ |
sre_parse.py | 178 class Tokenizer: 679 source = Tokenizer(str) 707 s = Tokenizer(source)
|
/prebuilts/python/linux-x86/2.7.5/lib/python2.7/ |
sre_parse.py | 178 class Tokenizer: 679 source = Tokenizer(str) 707 s = Tokenizer(source)
|
/external/doclava/src/com/google/doclava/apicheck/ |
ApiFile.java | 67 final Tokenizer tokenizer = new Tokenizer(filename, (new String(buf, 0, size)).toCharArray()); local 71 String token = tokenizer.getToken(); 76 parsePackage(api, tokenizer); 78 throw new ApiParseException("expected package got " + token, tokenizer.getLine()); 88 private static void parsePackage(ApiInfo api, Tokenizer tokenizer) 94 token = tokenizer.requireToken(); 95 assertIdent(tokenizer, token) [all...] |
/external/libedit/src/ |
histedit.h | 234 typedef struct tokenizer Tokenizer; 239 Tokenizer *tok_init(const char *); 240 void tok_end(Tokenizer *); 241 void tok_reset(Tokenizer *); 242 int tok_line(Tokenizer *, const LineInfo *, 244 int tok_str(Tokenizer *, const char *, 307 /* Wide character tokenizer support */
|
/external/smack/src/org/xbill/DNS/ |
Tokenizer.java | 26 * Tokenizer is used to parse DNS records and zones from text format, 32 public class Tokenizer { 142 * Creates a Tokenizer from an arbitrary input stream. 146 Tokenizer(InputStream is) { 161 * Creates a Tokenizer from a string. 165 Tokenizer(String s) { 170 * Creates a Tokenizer from a file. 174 Tokenizer(File f) throws FileNotFoundException { 225 * Gets the next token from a tokenizer. 339 * Gets the next token from a tokenizer, ignoring whitespace and comments [all...] |
/prebuilts/libs/libedit/include/ |
histedit.h | 234 typedef struct tokenizer Tokenizer; 239 Tokenizer *tok_init(const char *); 240 void tok_end(Tokenizer *); 241 void tok_reset(Tokenizer *); 242 int tok_line(Tokenizer *, const LineInfo *, 244 int tok_str(Tokenizer *, const char *, 307 /* Wide character tokenizer support */
|