/frameworks/opt/chips/src/com/android/ex/chips/ |
RecipientEntry.java | 160 final Rfc822Token[] tokens = Rfc822Tokenizer.tokenize(address);
|
DropdownChipLayouter.java | 169 destination = Rfc822Tokenizer.tokenize(entry.getDestination())[0].getAddress();
|
/external/google-breakpad/ |
Makefile.am | 271 src/processor/tokenize.cc \ 272 src/processor/tokenize.h 660 src/processor/tokenize.o \ 728 src/processor/tokenize.o \ 767 src/processor/tokenize.o \ 820 src/processor/tokenize.o \ 863 src/processor/tokenize.o \ [all...] |
/frameworks/base/media/java/android/media/ |
WebVttRenderer.java | 180 void tokenize(); method in interface:Tokenizer.TokenizerPhase 204 public void tokenize() { method in class:Tokenizer.DataTokenizer 241 public void tokenize() { method in class:Tokenizer.TagTokenizer 317 void tokenize(String s) { method in class:Tokenizer 321 mPhase.tokenize(); [all...] |
/external/minijail/ |
syscall_filter.c | 389 while ((group = tokenize(&arg_filter_str, "||")) != NULL) { 392 while ((comp = tokenize(&group_str, "&&")) != NULL) {
|
/external/python/cpython3/Lib/idlelib/ |
editor.py | 9 import tokenize [all...] |
/prebuilts/ndk/r16/sources/third_party/shaderc/third_party/glslang/hlsl/ |
hlslScanContext.cpp | 457 void HlslScanContext::tokenize(HlslToken& token) function in class:glslang::HlslScanContext 482 int token = ppContext.tokenize(ppToken); [all...] |
/external/deqp/framework/randomshaders/ |
rsgShaderGenerator.cpp | 340 // Tokenize shader and write source 343 shader.tokenize(m_state, tokenStr);
|
rsgBinaryOps.cpp | 93 void BinaryOp<Precedence, Assoc>::tokenize (GeneratorState& state, TokenStream& str) const function in class:rsg::BinaryOp 95 m_leftValueExpr->tokenize(state, str); 97 m_rightValueExpr->tokenize(state, str);
|
/hardware/libhardware/modules/sensors/dynamic_sensor/ |
HidRawDevice.cpp | 166 std::vector<HidItem> tokens = HidItem::tokenize(mDeviceInfo.descriptor);
|
/hardware/libhardware/modules/sensors/dynamic_sensor/HidUtils/ |
HidParser.cpp | 133 std::vector<HidItem> hidItemVector = HidItem::tokenize(begin, size);
|
/libcore/ojluni/src/main/java/java/io/ |
StreamTokenizer.java | 218 * @deprecated As of JDK version 1.1, the preferred way to tokenize an
|
/packages/apps/Email/provider_src/com/android/email/provider/ |
EmailConversationCursor.java | 186 final Rfc822Token[] tokens = Rfc822Tokenizer.tokenize(fromString);
|
/packages/apps/Messaging/src/com/android/messaging/ui/contact/ |
ContactRecipientAdapter.java | 348 final Rfc822Token[] tokens = Rfc822Tokenizer.tokenize(inAddresses.get(i).toLowerCase());
|
/prebuilts/ndk/r16/sources/third_party/shaderc/third_party/glslang/glslang/MachineIndependent/ |
Scan.cpp | 294 return parseContext.getScanContext()->tokenize(parseContext.getPpContext(), token); 645 int TScanContext::tokenize(TPpContext* pp, TParserToken& token) function in class:glslang::TScanContext 650 int token = pp->tokenize(ppToken); [all...] |
/external/python/cpython3/Lib/ |
inspect.py | 44 import tokenize [all...] |
tokenize.py | 3 tokenize(readline) is a generator that breaks a stream of bytes into 41 __all__ = token.__all__ + ["COMMENT", "tokenize", "detect_encoding", 320 token, which is the first token sequence output by tokenize. 330 # Output bytes will tokenize back to the input 331 t1 = [tok[:2] for tok in tokenize(f.readline)] 334 t2 = [tok[:2] for tok in tokenize(readline)] 359 in the same way as the tokenize() generator. 464 def tokenize(readline): function 466 The tokenize() generator requires one argument, readline, which 484 # built yet and tokenize is imported [all...] |
/external/webrtc/webrtc/base/ |
logging.cc | 274 tokenize(params, ' ', &tokens);
|
/prebuilts/ndk/r16/sources/third_party/shaderc/third_party/glslang/glslang/MachineIndependent/preprocessor/ |
PpContext.h | 190 int tokenize(TPpToken& ppToken);
|
/device/linaro/bootloader/edk2/AppPkg/Applications/Python/Python-2.7.2/Lib/ |
trace.py | 57 import tokenize
427 for ttype, tstr, start, end, line in tokenize.generate_tokens(f.readline):
|
/external/apache-xml/src/main/java/org/apache/xpath/compiler/ |
XPathParser.java | 112 lexer.tokenize(expression); 188 lexer.tokenize(expression); [all...] |
/external/ply/ply/ply/ |
cpp.py | 177 # tokenize() 179 # Utility function. Given a string of text, tokenize into a list of tokens 182 def tokenize(self,text): member in class:Preprocessor 555 # tokens = tokenize(line) 795 tokens = self.tokenize(tokens) [all...] |
/external/python/cpython2/Lib/ |
trace.py | 57 import tokenize 427 for ttype, tstr, start, end, line in tokenize.generate_tokens(f.readline):
|
/external/python/cpython3/Tools/parser/ |
unparse.py | 4 import tokenize 671 encoding = tokenize.detect_encoding(pyfile.readline)[0]
|
/frameworks/base/core/tests/coretests/src/android/text/ |
TextUtilsTest.java | 251 Rfc822Token[] tokens = Rfc822Tokenizer.tokenize("Foo Bar (something) <foo@google.com>"); 261 Rfc822Token[] tokens = Rfc822Tokenizer.tokenize("\"Foo Bar\\");
|