Home
last modified time | relevance | path

Searched refs:tokenize (Results 1 – 25 of 170) sorted by relevance

1234567

/external/minijail/
Dutil_unittest.cc25 TEST(tokenize, null_stringp) { in TEST() argument
26 ASSERT_EQ(nullptr, tokenize(nullptr, nullptr)); in TEST()
27 ASSERT_EQ(nullptr, tokenize(nullptr, "")); in TEST()
28 ASSERT_EQ(nullptr, tokenize(nullptr, ",")); in TEST()
31 ASSERT_EQ(nullptr, tokenize(&p, nullptr)); in TEST()
35 TEST(tokenize, null_delim) { in TEST() argument
38 ASSERT_EQ(str, tokenize(&p, nullptr)); in TEST()
43 ASSERT_EQ(str, tokenize(&p, "")); in TEST()
49 TEST(tokenize, basic) { in TEST() argument
52 ASSERT_EQ("a", std::string(tokenize(&p, ","))); in TEST()
[all …]
Dminijail0_cli.c97 char *src = tokenize(&arg, ","); in add_binding()
98 char *dest = tokenize(&arg, ","); in add_binding()
99 char *flags = tokenize(&arg, ","); in add_binding()
116 char *type = tokenize(&arg, ","); in add_rlimit()
117 char *cur = tokenize(&arg, ","); in add_rlimit()
118 char *max = tokenize(&arg, ","); in add_rlimit()
156 char *src = tokenize(&arg, ","); in add_mount()
157 char *dest = tokenize(&arg, ","); in add_mount()
158 char *type = tokenize(&arg, ","); in add_mount()
159 char *flags = tokenize(&arg, ","); in add_mount()
[all …]
/external/google-breakpad/src/testing/scripts/generator/cpp/
Dast.py46 from cpp import tokenize
549 if parts[-1].token_type == tokenize.NAME:
579 if (type_name and type_name[-1].token_type == tokenize.NAME and
580 p.token_type == tokenize.NAME):
581 type_name.append(tokenize.Token(tokenize.SYNTAX, ' ', 0, 0))
738 if token.token_type == tokenize.NAME:
749 if next.token_type == tokenize.SYNTAX and next.name == '(':
754 syntax = tokenize.SYNTAX
763 new_temp = self._GetTokensUpTo(tokenize.SYNTAX, ';')
766 last_token = tokenize.Token(tokenize.SYNTAX, ';', 0, 0)
[all …]
/external/v8/testing/gmock/scripts/generator/cpp/
Dast.py46 from cpp import tokenize
551 if parts[-1].token_type == tokenize.NAME:
581 if (type_name and type_name[-1].token_type == tokenize.NAME and
582 p.token_type == tokenize.NAME):
583 type_name.append(tokenize.Token(tokenize.SYNTAX, ' ', 0, 0))
739 if token.token_type == tokenize.NAME:
750 if next.token_type == tokenize.SYNTAX and next.name == '(':
755 syntax = tokenize.SYNTAX
764 new_temp = self._GetTokensUpTo(tokenize.SYNTAX, ';')
767 last_token = tokenize.Token(tokenize.SYNTAX, ';', 0, 0)
[all …]
/external/googletest/googlemock/scripts/generator/cpp/
Dast.py46 from cpp import tokenize
551 if parts[-1].token_type == tokenize.NAME:
581 if (type_name and type_name[-1].token_type == tokenize.NAME and
582 p.token_type == tokenize.NAME):
583 type_name.append(tokenize.Token(tokenize.SYNTAX, ' ', 0, 0))
739 if token.token_type == tokenize.NAME:
750 if next.token_type == tokenize.SYNTAX and next.name == '(':
755 syntax = tokenize.SYNTAX
764 new_temp = self._GetTokensUpTo(tokenize.SYNTAX, ';')
767 last_token = tokenize.Token(tokenize.SYNTAX, ';', 0, 0)
[all …]
/external/python/cpython3/Tools/i18n/
Dpygettext.py166 import tokenize
337 if ttype == tokenize.STRING:
340 elif ttype not in (tokenize.COMMENT, tokenize.NL):
344 if ttype == tokenize.NAME and tstring in ('class', 'def'):
347 if ttype == tokenize.NAME and tstring in opts.keywords:
352 if ttype == tokenize.OP and tstring == ':':
357 if ttype == tokenize.STRING:
360 elif ttype not in (tokenize.NEWLINE, tokenize.INDENT,
361 tokenize.COMMENT):
366 if ttype == tokenize.OP and tstring == '(':
[all …]
/external/python/cpython2/Tools/i18n/
Dpygettext.py165 import tokenize
375 if ttype == tokenize.STRING:
378 elif ttype not in (tokenize.COMMENT, tokenize.NL):
382 if ttype == tokenize.NAME and tstring in ('class', 'def'):
385 if ttype == tokenize.NAME and tstring in opts.keywords:
390 if ttype == tokenize.OP and tstring == ':':
395 if ttype == tokenize.STRING:
398 elif ttype not in (tokenize.NEWLINE, tokenize.INDENT,
399 tokenize.COMMENT):
404 if ttype == tokenize.OP and tstring == '(':
[all …]
/external/python/cpython3/Doc/library/
Dtokenize.rst1 :mod:`tokenize` --- Tokenizer for Python source
4 .. module:: tokenize
10 **Source code:** :source:`Lib/tokenize.py`
14 The :mod:`tokenize` module provides a lexical scanner for Python source code,
22 :term:`named tuple` returned from :func:`tokenize.tokenize`.
29 .. function:: tokenize(readline)
31 The :func:`.tokenize` generator requires one argument, *readline*, which
56 :func:`.tokenize` determines the source encoding of the file by looking for a
61 :mod:`tokenize`, as are three additional token type values:
78 into text. The first token returned by :func:`.tokenize` will always be an
[all …]
/external/python/cpython2/Tools/scripts/
Dcheckappend.py39 import tokenize
106 tokenize.tokenize(self.file.readline, self.tokeneater)
107 except tokenize.TokenError, msg:
113 NEWLINE=tokenize.NEWLINE,
114 JUNK=(tokenize.COMMENT, tokenize.NL),
115 OP=tokenize.OP,
116 NAME=tokenize.NAME):
Dcleanfuture.py42 import tokenize
157 STRING = tokenize.STRING
158 NL = tokenize.NL
159 NEWLINE = tokenize.NEWLINE
160 COMMENT = tokenize.COMMENT
161 NAME = tokenize.NAME
162 OP = tokenize.OP
165 get = tokenize.generate_tokens(self.getline).next
Dreindent.py44 import tokenize
188 tokenize.tokenize(self.getline, self.tokeneater)
270 INDENT=tokenize.INDENT,
271 DEDENT=tokenize.DEDENT,
272 NEWLINE=tokenize.NEWLINE,
273 COMMENT=tokenize.COMMENT,
274 NL=tokenize.NL):
/external/deqp/framework/randomshaders/
DrsgExpression.hpp57 virtual void tokenize (GeneratorState& state, TokenStream& str) const = DE_NULL;
74 …void tokenize (GeneratorState& state, TokenStream& str) const { DE_UNREF(state); str << Tok… in tokenize() function in rsg::VariableAccess
113 void tokenize (GeneratorState& state, TokenStream& str) const;
131 void tokenize (GeneratorState& state, TokenStream& str) const;
149 void tokenize (GeneratorState& state, TokenStream& str) const;
167 void tokenize (GeneratorState& state, TokenStream& str) const;
189 void tokenize (GeneratorState& state, TokenStream& str) const;
214 void tokenize (GeneratorState& state, TokenStream& str) const;
233 void tokenize (GeneratorState& state, TokenStream& str) const;
255 void tokenize (GeneratorState& state, TokenStream& str) const;
DrsgStatement.cpp203 void BlockStatement::tokenize (GeneratorState& state, TokenStream& str) const in tokenize() function in rsg::BlockStatement
208 (*i)->tokenize(state, str); in tokenize()
219 void ExpressionStatement::tokenize (GeneratorState& state, TokenStream& str) const in tokenize() function in rsg::ExpressionStatement
222 m_expression->tokenize(state, str); in tokenize()
333 void DeclarationStatement::tokenize (GeneratorState& state, TokenStream& str) const in tokenize() function in rsg::DeclarationStatement
340 m_expression->tokenize(state, str); in tokenize()
456 void ConditionalStatement::tokenize (GeneratorState& state, TokenStream& str) const in tokenize() function in rsg::ConditionalStatement
462 m_condition->tokenize(state, str); in tokenize()
469 m_trueStatement->tokenize(state, str); in tokenize()
473 m_trueStatement->tokenize(state, str); in tokenize()
[all …]
DrsgStatement.hpp44 virtual void tokenize (GeneratorState& state, TokenStream& str) const = DE_NULL;
62 void tokenize (GeneratorState& state, TokenStream& str) const;
78 void tokenize (GeneratorState& state, TokenStream& str) const;
98 void tokenize (GeneratorState& state, TokenStream& str) const;
119 void tokenize (GeneratorState& state, TokenStream& str) const;
145 void tokenize (GeneratorState& state, TokenStream& str) const;
DrsgShader.cpp93 void Shader::tokenize (GeneratorState& state, TokenStream& str) const in tokenize() function in rsg::Shader
101 m_globalStatements[ndx]->tokenize(state, str); in tokenize()
107 m_functions[ndx]->tokenize(state, str); in tokenize()
112 m_mainFunction.tokenize(state, str); in tokenize()
125 void Function::tokenize (GeneratorState& state, TokenStream& str) const in tokenize() function in rsg::Function
147 m_functionBlock.tokenize(state, str); in tokenize()
/external/python/cpython3/Lib/
Dtabnanny.py26 import tokenize
27 if not hasattr(tokenize, 'NL'):
97 f = tokenize.open(file)
106 process_tokens(tokenize.generate_tokens(f.readline))
108 except tokenize.TokenError as msg:
277 INDENT = tokenize.INDENT
278 DEDENT = tokenize.DEDENT
279 NEWLINE = tokenize.NEWLINE
280 JUNK = tokenize.COMMENT, tokenize.NL
/external/webrtc/webrtc/base/
Dstringencode_unittest.cc220 EXPECT_EQ(5ul, tokenize("one two three four five", ' ', &fields)); in TEST()
222 EXPECT_EQ(1ul, tokenize("one", ' ', &fields)); in TEST()
226 EXPECT_EQ(5ul, tokenize(" one two three four five ", ' ', &fields)); in TEST()
228 EXPECT_EQ(1ul, tokenize(" one ", ' ', &fields)); in TEST()
230 EXPECT_EQ(0ul, tokenize(" ", ' ', &fields)); in TEST()
237 tokenize("find middle one", ' ', &fields); in TEST()
243 tokenize(" find middle one ", ' ', &fields); in TEST()
247 tokenize(" ", ' ', &fields); in TEST()
267 ASSERT_EQ(0ul, tokenize("D \"A B", ' ', '(', ')', NULL)); in TEST()
270 tokenize("A B C", ' ', '"', '"', &fields); in TEST()
[all …]
/external/python/cpython2/Doc/library/
Dtokenize.rst1 :mod:`tokenize` --- Tokenizer for Python source
4 .. module:: tokenize
9 **Source code:** :source:`Lib/tokenize.py`
13 The :mod:`tokenize` module provides a lexical scanner for Python source code,
22 :func:`tokenize.generate_tokens` for the character sequence that identifies a
48 .. function:: tokenize(readline[, tokeneater])
50 The :func:`.tokenize` function accepts two parameters: one representing the input
51 stream, and one providing an output mechanism for :func:`.tokenize`.
67 :mod:`tokenize`, as are two additional token type values that might be passed to
68 the *tokeneater* function by :func:`.tokenize`:
[all …]
/external/python/cpython2/Lib/
Dtabnanny.py26 import tokenize
27 if not hasattr(tokenize, 'NL'):
106 process_tokens(tokenize.generate_tokens(f.readline))
108 except tokenize.TokenError, msg:
274 INDENT = tokenize.INDENT
275 DEDENT = tokenize.DEDENT
276 NEWLINE = tokenize.NEWLINE
277 JUNK = tokenize.COMMENT, tokenize.NL
/external/python/cpython3/Tools/scripts/
Dhighlight.py11 import tokenize
35 tok_type = tokenize.COMMENT
37 for tok in tokenize.generate_tokens(readline):
41 if tok_type == tokenize.COMMENT:
43 elif tok_type == tokenize.OP and tok_str[:1] not in '{}[](),.:;@':
45 elif tok_type == tokenize.STRING:
47 if prev_tok_type == tokenize.INDENT or scol==0:
49 elif tok_type == tokenize.NAME:
Dcleanfuture.py42 import tokenize
157 STRING = tokenize.STRING
158 NL = tokenize.NL
159 NEWLINE = tokenize.NEWLINE
160 COMMENT = tokenize.COMMENT
161 NAME = tokenize.NAME
162 OP = tokenize.OP
165 get = tokenize.generate_tokens(self.getline).__next__
Dreindent.py46 import tokenize
121 encoding, _ = tokenize.detect_encoding(f.readline)
198 tokens = tokenize.generate_tokens(self.getline)
282 INDENT=tokenize.INDENT,
283 DEDENT=tokenize.DEDENT,
284 NEWLINE=tokenize.NEWLINE,
285 COMMENT=tokenize.COMMENT,
286 NL=tokenize.NL):
/external/autotest/utils/
Dreindent.py44 import tokenize
176 tokenize.tokenize(self.getline, self.tokeneater)
258 INDENT=tokenize.INDENT,
259 DEDENT=tokenize.DEDENT,
260 NEWLINE=tokenize.NEWLINE,
261 COMMENT=tokenize.COMMENT,
262 NL=tokenize.NL):
/external/swiftshader/third_party/LLVM/lib/MC/MCDisassembler/
DEDInst.cpp165 int EDInst::tokenize() { in tokenize() function in EDInst
175 return TokenizeResult.setResult(EDToken::tokenize(Tokens, in tokenize()
183 if (tokenize()) in numTokens()
189 if (tokenize()) in getToken()
197 if (tokenize()) in visitTokens()
/external/chromium-trace/catapult/common/py_utils/py_utils/refactor/
Doffset_token.py8 import tokenize
66 tokenize_tokens = tokenize.generate_tokens(f.readline)
90 while offset_tokens[0].type == tokenize.NL:
115 return tokenize.untokenize(tokenize_tokens).replace('\\\n', ' \\\n')

1234567