/external/chromium_org/third_party/sqlite/src/test/ |
tokenize.test | 14 # $Id: tokenize.test,v 1.1 2008/07/08 00:06:51 drh Exp $ 20 do_test tokenize-1.1 { 23 do_test tokenize-1.2 { 26 do_test tokenize-1.3 { 29 do_test tokenize-1.4 { 32 do_test tokenize-1.5 { 35 do_test tokenize-1.6 { 38 do_test tokenize-1.7 { 41 do_test tokenize-1.8 { 44 do_test tokenize-1.9 [all...] |
fts3ad.test | 29 CREATE VIRTUAL TABLE t1 USING fts3(content, tokenize porter); 67 CREATE VIRTUAL TABLE t1 USING fts3(content, tokenize porter); 75 CREATE VIRTUAL TABLE t1 USING fts3(content, tokenize= porter); 83 CREATE VIRTUAL TABLE t1 USING fts3(content, tokenize= simple); 91 CREATE VIRTUAL TABLE t1 USING fts3(content, tokenize= porter); 99 CREATE VIRTUAL TABLE t1 USING fts3(content, tokenize = porter);
|
/external/chromium_org/remoting/base/ |
capabilities.cc | 17 Tokenize(capabilities, " ", &caps); 24 Tokenize(client_capabilities, " ", &client_caps); 28 Tokenize(host_capabilities, " ", &host_caps);
|
/external/chromium_org/third_party/WebKit/Source/devtools/front_end/cm/ |
xml.js | 68 state.tokenize = parser; 88 state.tokenize = inBlock("meta", "?>"); 92 state.tokenize = inTag; 116 state.tokenize = inText; 123 state.tokenize = inText; 126 var next = state.tokenize(stream, state); 129 state.tokenize = inAttribute(ch); 131 return state.tokenize(stream, state); 142 state.tokenize = inTag; 156 state.tokenize = inText [all...] |
shell.js | 50 return tokenize(stream, state); 62 return tokenize(stream, state); 113 return tokenize(stream, state); 123 function tokenize(stream, state) { 130 return tokenize(stream, state);
|
coffeescript.js | 83 state.tokenize = longComment; 84 return state.tokenize(stream, state); 135 state.tokenize = tokenFactory(stream.current(), false, "string"); 136 return state.tokenize(stream, state); 141 state.tokenize = tokenFactory(stream.current(), true, "string-2"); 142 return state.tokenize(stream, state); 187 state.tokenize = tokenBase; 197 state.tokenize = tokenBase; 208 state.tokenize = tokenBase; 265 var style = state.tokenize(stream, state) [all...] |
clike.js | 35 state.tokenize = tokenString(ch); 36 return state.tokenize(stream, state); 48 state.tokenize = tokenComment; 82 state.tokenize = null; 91 state.tokenize = null; 124 tokenize: null, 140 var style = (state.tokenize || tokenBase)(stream, state); 161 if (state.tokenize != tokenBase && state.tokenize != null) return CodeMirror.Pass; 194 state.tokenize = cppHook [all...] |
/external/deqp/framework/randomshaders/ |
rsgShader.cpp | 93 void Shader::tokenize (GeneratorState& state, TokenStream& str) const function in class:rsg::Shader 99 // Tokenize global declaration statements 101 m_globalStatements[ndx]->tokenize(state, str); 103 // Tokenize all functions 107 m_functions[ndx]->tokenize(state, str); 110 // Tokenize main 112 m_mainFunction.tokenize(state, str); 125 void Function::tokenize (GeneratorState& state, TokenStream& str) const function in class:rsg::Function 146 // Tokenize body 147 m_functionBlock.tokenize(state, str) [all...] |
rsgExpression.hpp | 57 virtual void tokenize (GeneratorState& state, TokenStream& str) const = DE_NULL; 74 void tokenize (GeneratorState& state, TokenStream& str) const { DE_UNREF(state); str << Token(m_variable->getName()); } function in class:rsg::VariableAccess 113 void tokenize (GeneratorState& state, TokenStream& str) const; 131 void tokenize (GeneratorState& state, TokenStream& str) const; 149 void tokenize (GeneratorState& state, TokenStream& str) const; 167 void tokenize (GeneratorState& state, TokenStream& str) const; 189 void tokenize (GeneratorState& state, TokenStream& str) const; 214 void tokenize (GeneratorState& state, TokenStream& str) const; 233 void tokenize (GeneratorState& state, TokenStream& str) const; 255 void tokenize (GeneratorState& state, TokenStream& str) const [all...] |
rsgStatement.hpp | 44 virtual void tokenize (GeneratorState& state, TokenStream& str) const = DE_NULL; 62 void tokenize (GeneratorState& state, TokenStream& str) const; 78 void tokenize (GeneratorState& state, TokenStream& str) const; 98 void tokenize (GeneratorState& state, TokenStream& str) const; 119 void tokenize (GeneratorState& state, TokenStream& str) const; 145 void tokenize (GeneratorState& state, TokenStream& str) const;
|
rsgStatement.cpp | 203 void BlockStatement::tokenize (GeneratorState& state, TokenStream& str) const function in class:rsg::BlockStatement 208 (*i)->tokenize(state, str); 219 void ExpressionStatement::tokenize (GeneratorState& state, TokenStream& str) const function in class:rsg::ExpressionStatement 222 m_expression->tokenize(state, str); 333 void DeclarationStatement::tokenize (GeneratorState& state, TokenStream& str) const function in class:rsg::DeclarationStatement 340 m_expression->tokenize(state, str); 456 void ConditionalStatement::tokenize (GeneratorState& state, TokenStream& str) const function in class:rsg::ConditionalStatement 462 m_condition->tokenize(state, str); 469 m_trueStatement->tokenize(state, str); 473 m_trueStatement->tokenize(state, str) 556 void AssignStatement::tokenize (GeneratorState& state, TokenStream& str) const function in class:rsg::AssignStatement [all...] |
/external/chromium_org/third_party/WebKit/Tools/Scripts/webkitpy/thirdparty/coverage/ |
phystokens.py | 3 import keyword, re, token, tokenize namespace 9 tokenize.generate_tokens() doesn't return a token for the backslash that 38 if last_ttype == tokenize.COMMENT: 77 ws_tokens = [token.INDENT, token.DEDENT, token.NEWLINE, tokenize.NL] 81 tokgen = tokenize.generate_tokens(StringIO(source).readline) 98 tok_class = tokenize.tok_name.get(ttype, 'xx').lower()[:3]
|
/external/chromium_org/ui/app_list/search/ |
tokenized_string.cc | 18 Tokenize(); 23 void TokenizedString::Tokenize() {
|
/external/chromium_org/google_apis/drive/ |
time_util.cc | 28 int num_of_token = Tokenize(timezone, ":", &parts); 56 if (Tokenize(raw_value, "T", &parts) != 2) 71 } else if (Tokenize(time_and_tz, "+", &parts) == 2) { 77 } else if (Tokenize(time_and_tz, "-", &parts) == 2) { 92 if (Tokenize(date, "-", &parts) != 3) 105 int num_of_token = Tokenize(time, ":", &parts); 115 int num_of_seconds_token = Tokenize(parts[2], ".", &seconds_parts);
|
/external/chromium_org/net/tools/balsa/ |
balsa_headers_token_utils.cc | 23 // Tokenize 45 // Tokenize just that line. 83 // Tokenize just that line 99 // We may have more then 1 line with the same header key. Tokenize them all
|
/packages/apps/ContactsCommon/res/values-land/ |
integers.xml | 20 <!-- The number of characters in the snippet before we need to tokenize and ellipse. -->
|
/packages/apps/ContactsCommon/res/values-sw600dp-land/ |
integers.xml | 20 <!-- The number of characters in the snippet before we need to tokenize and ellipse. -->
|
/packages/apps/ContactsCommon/res/values-sw720dp/ |
integers.xml | 20 <!-- The number of characters in the snippet before we need to tokenize and ellipse. -->
|
/packages/apps/ContactsCommon/res/values-sw720dp-land/ |
integers.xml | 20 <!-- The number of characters in the snippet before we need to tokenize and ellipse. -->
|
/prebuilts/python/darwin-x86/2.7.5/lib/python2.7/lib2to3/pgen2/ |
driver.py | 26 from . import grammar, parse, token, tokenize, pgen namespace 40 # XXX Move the prefix computation into a wrapper around tokenize. 59 if type in (tokenize.COMMENT, tokenize.NL): 88 tokens = tokenize.generate_tokens(stream.readline) 105 tokens = tokenize.generate_tokens(StringIO.StringIO(text).readline)
|
/prebuilts/python/linux-x86/2.7.5/lib/python2.7/lib2to3/pgen2/ |
driver.py | 26 from . import grammar, parse, token, tokenize, pgen namespace 40 # XXX Move the prefix computation into a wrapper around tokenize. 59 if type in (tokenize.COMMENT, tokenize.NL): 88 tokens = tokenize.generate_tokens(stream.readline) 105 tokens = tokenize.generate_tokens(StringIO.StringIO(text).readline)
|
/external/chromium_org/third_party/skia/experimental/webtry/res/js/cm/ |
clike.js | 32 state.tokenize = tokenString(ch); 33 return state.tokenize(stream, state); 45 state.tokenize = tokenComment; 79 state.tokenize = null; 88 state.tokenize = null; 121 tokenize: null, 137 var style = (state.tokenize || tokenBase)(stream, state); 158 if (state.tokenize != tokenBase && state.tokenize != null) return CodeMirror.Pass; 191 state.tokenize = cppHook [all...] |
/external/chromium_org/third_party/webrtc/base/ |
stringencode_unittest.cc | 219 EXPECT_EQ(5ul, tokenize("one two three four five", ' ', &fields)); 221 EXPECT_EQ(1ul, tokenize("one", ' ', &fields)); 225 EXPECT_EQ(5ul, tokenize(" one two three four five ", ' ', &fields)); 227 EXPECT_EQ(1ul, tokenize(" one ", ' ', &fields)); 229 EXPECT_EQ(0ul, tokenize(" ", ' ', &fields)); 236 tokenize("find middle one", ' ', &fields); 242 tokenize(" find middle one ", ' ', &fields); 246 tokenize(" ", ' ', &fields); 266 ASSERT_EQ(0ul, tokenize("D \"A B", ' ', '(', ')', NULL)); 269 tokenize("A B C", ' ', '"', '"', &fields) [all...] |
/external/skia/experimental/webtry/res/js/cm/ |
clike.js | 32 state.tokenize = tokenString(ch); 33 return state.tokenize(stream, state); 45 state.tokenize = tokenComment; 79 state.tokenize = null; 88 state.tokenize = null; 121 tokenize: null, 137 var style = (state.tokenize || tokenBase)(stream, state); 158 if (state.tokenize != tokenBase && state.tokenize != null) return CodeMirror.Pass; 191 state.tokenize = cppHook [all...] |
/external/chromium_org/chrome/browser/devtools/device/adb/ |
adb_device_info_query.cc | 95 Tokenize(response, "\n", &entries); 98 Tokenize(entries[i], " \r", &fields); 121 Tokenize(response, "\n", &entries); 124 Tokenize(entries[i], " \r", &fields); 232 Tokenize(response, "\r", &lines); 246 Tokenize(str, "-", &pairs); 253 Tokenize(pairs[1].substr(1, pairs[1].size() - 2), ",", &numbers);
|