/external/autotest/client/common_lib/cros/ |
D | dbus_send.py | 47 token_type = token_stream.pop(0) 48 if token_type == 'variant': 49 token_type = token_stream.pop(0) 50 if token_type == 'object': 51 token_type = token_stream.pop(0) # Should be 'path' 55 if token_type in INT_TYPES: 57 if token_type == 'string' or token_type == 'path': 59 if token_type == 'boolean': 61 if token_type == 'double': 63 if token_type == 'array': [all …]
|
/external/v8/testing/gmock/scripts/generator/cpp/ |
D | tokenize.py | 74 def __init__(self, token_type, name, start, end): argument 75 self.token_type = token_type 147 token_type = UNKNOWN 151 token_type = NAME 159 token_type = CONSTANT 162 token_type = CONSTANT 173 token_type = SYNTAX 183 token_type = SYNTAX 186 token_type = CONSTANT 196 token_type = CONSTANT [all …]
|
D | ast.py | 551 if parts[-1].token_type == tokenize.NAME: 581 if (type_name and type_name[-1].token_type == tokenize.NAME and 582 p.token_type == tokenize.NAME): 714 if token.token_type == _INTERNAL_TOKEN: 739 if token.token_type == tokenize.NAME: 750 if next.token_type == tokenize.SYNTAX and next.name == '(': 806 elif token.token_type == tokenize.SYNTAX: 812 if (token.token_type == tokenize.NAME and 816 elif token.token_type == tokenize.PREPROCESSOR: 855 while (last_token.token_type != expected_token_type or [all …]
|
/external/googletest/googlemock/scripts/generator/cpp/ |
D | tokenize.py | 74 def __init__(self, token_type, name, start, end): argument 75 self.token_type = token_type 147 token_type = UNKNOWN 151 token_type = NAME 159 token_type = CONSTANT 162 token_type = CONSTANT 173 token_type = SYNTAX 183 token_type = SYNTAX 186 token_type = CONSTANT 196 token_type = CONSTANT [all …]
|
D | ast.py | 551 if parts[-1].token_type == tokenize.NAME: 581 if (type_name and type_name[-1].token_type == tokenize.NAME and 582 p.token_type == tokenize.NAME): 714 if token.token_type == _INTERNAL_TOKEN: 739 if token.token_type == tokenize.NAME: 750 if next.token_type == tokenize.SYNTAX and next.name == '(': 806 elif token.token_type == tokenize.SYNTAX: 812 if (token.token_type == tokenize.NAME and 816 elif token.token_type == tokenize.PREPROCESSOR: 855 while (last_token.token_type != expected_token_type or [all …]
|
/external/google-breakpad/src/testing/scripts/generator/cpp/ |
D | tokenize.py | 74 def __init__(self, token_type, name, start, end): argument 75 self.token_type = token_type 147 token_type = UNKNOWN 151 token_type = NAME 159 token_type = CONSTANT 162 token_type = CONSTANT 173 token_type = SYNTAX 183 token_type = SYNTAX 186 token_type = CONSTANT 196 token_type = CONSTANT [all …]
|
D | ast.py | 549 if parts[-1].token_type == tokenize.NAME: 579 if (type_name and type_name[-1].token_type == tokenize.NAME and 580 p.token_type == tokenize.NAME): 713 if token.token_type == _INTERNAL_TOKEN: 738 if token.token_type == tokenize.NAME: 749 if next.token_type == tokenize.SYNTAX and next.name == '(': 805 elif token.token_type == tokenize.SYNTAX: 811 if (token.token_type == tokenize.NAME and 815 elif token.token_type == tokenize.PREPROCESSOR: 854 while (last_token.token_type != expected_token_type or [all …]
|
/external/protobuf/gtest/scripts/ |
D | pump.py | 141 def __init__(self, start=None, end=None, value=None, token_type=None): argument 151 self.token_type = token_type 155 self.start, self.value, self.token_type) 161 self.token_type) 172 for (regex, token_type) in token_table: 179 best_match_token_type = token_type 198 (start_column, length, token_type) = m 203 return MakeToken(lines, found_start, found_end, token_type) 227 def MakeToken(lines, start, end, token_type): argument 230 return Token(start, end, SubString(lines, start, end), token_type) [all …]
|
/external/vulkan-validation-layers/tests/gtest-1.7.0/scripts/ |
D | pump.py | 140 def __init__(self, start=None, end=None, value=None, token_type=None): argument 150 self.token_type = token_type 154 self.start, self.value, self.token_type) 160 self.token_type) 171 for (regex, token_type) in token_table: 178 best_match_token_type = token_type 197 (start_column, length, token_type) = m 202 return MakeToken(lines, found_start, found_end, token_type) 237 def MakeToken(lines, start, end, token_type): argument 240 return Token(start, end, SubString(lines, start, end), token_type) [all …]
|
/external/v8/testing/gtest/scripts/ |
D | pump.py | 140 def __init__(self, start=None, end=None, value=None, token_type=None): argument 150 self.token_type = token_type 154 self.start, self.value, self.token_type) 160 self.token_type) 171 for (regex, token_type) in token_table: 178 best_match_token_type = token_type 197 (start_column, length, token_type) = m 202 return MakeToken(lines, found_start, found_end, token_type) 237 def MakeToken(lines, start, end, token_type): argument 240 return Token(start, end, SubString(lines, start, end), token_type) [all …]
|
/external/googletest/googletest/scripts/ |
D | pump.py | 140 def __init__(self, start=None, end=None, value=None, token_type=None): argument 150 self.token_type = token_type 154 self.start, self.value, self.token_type) 160 self.token_type) 171 for (regex, token_type) in token_table: 178 best_match_token_type = token_type 197 (start_column, length, token_type) = m 202 return MakeToken(lines, found_start, found_end, token_type) 237 def MakeToken(lines, start, end, token_type): argument 240 return Token(start, end, SubString(lines, start, end), token_type) [all …]
|
/external/google-breakpad/src/testing/gtest/scripts/ |
D | pump.py | 140 def __init__(self, start=None, end=None, value=None, token_type=None): argument 150 self.token_type = token_type 154 self.start, self.value, self.token_type) 160 self.token_type) 171 for (regex, token_type) in token_table: 178 best_match_token_type = token_type 197 (start_column, length, token_type) = m 202 return MakeToken(lines, found_start, found_end, token_type) 237 def MakeToken(lines, start, end, token_type): argument 240 return Token(start, end, SubString(lines, start, end), token_type) [all …]
|
/external/chromium-trace/catapult/common/py_utils/py_utils/refactor/ |
D | offset_token.py | 26 def __init__(self, token_type, string, offset): argument 27 self._type = token_type 70 token_type, string, (srow, scol), _, _ = next_token 72 offset_tokens.append(OffsetToken(token_type, string, (0, 0))) 76 offset_tokens.append(OffsetToken(token_type, string, (0, scol - ecol))) 79 token_type, string, (srow - erow, scol)))
|
D | snippet.py | 97 def __init__(self, token_type, tokens): argument 102 self._type = token_type 107 def Create(cls, token_type, string, offset=(0, 0)): argument 108 return cls(token_type, 109 [offset_token.OffsetToken(token_type, string, offset)])
|
/external/bison/examples/calc++/ |
D | calc++-scanner.ll | 16 /* By default yylex returns int, we use token_type. 18 not of token_type. */ 44 [-+*/] return yy::calcxx_parser::token_type (yytext[0]);
|
D | calc++-driver.hh | 10 yy::calcxx_parser::token_type \
|
D | calc++-parser.hh | 106 typedef token::yytokentype token_type; typedef in yy::calcxx_parser
|
/external/libmojo/third_party/jinja2/ |
D | parser.py | 431 token_type = self.stream.current.type 432 if token_type in _compare_operators: 434 ops.append(nodes.Operand(token_type, self.parse_add())) 529 token_type = self.stream.current.type 531 if token_type == 'sub': 534 elif token_type == 'add': 662 token_type = self.stream.current.type 663 if token_type == 'dot' or token_type == 'lbracket': 667 elif token_type == 'lparen': 675 token_type = self.stream.current.type [all …]
|
D | lexer.py | 145 def _describe_token_type(token_type): argument 146 if token_type in reverse_operators: 147 return reverse_operators[token_type] 161 }.get(token_type, token_type)
|
/external/antlr/antlr-3.4/runtime/Ruby/lib/antlr3/tree/ |
D | debug.rb | 27 def create_from_token( token_type, from_token, text = nil ) argument 33 def create_from_type( token_type, text ) argument
|
/external/libmojo/mojo/public/tools/bindings/pylib/mojom_tests/parse/ |
D | lexer_unittest.py | 42 def _MakeLexToken(token_type, value, lineno=1, lexpos=0): argument 46 rv.type, rv.value, rv.lineno, rv.lexpos = token_type, value, lineno, lexpos
|
/external/v8/src/parsing/ |
D | token.cc | 45 const char Token::token_type[] = { member in v8::internal::Token
|
D | token.h | 199 return token_type[tok] == 'K'; in IsKeyword() 346 static const char token_type[NUM_TOKENS]; variable
|
/external/antlr/antlr-3.4/runtime/Ruby/lib/antlr3/ |
D | tree.rb | 858 def create_from_token( token_type, from_token, text = nil ) argument 860 from_token.type = token_type 866 def create_from_type( token_type, text ) argument 867 from_token = create_token( token_type, DEFAULT_CHANNEL, text )
|
/external/antlr/antlr-3.4/runtime/Perl5/lib/ANTLR/Runtime/ |
D | BaseRecognizer.pm | 430 my ($self, $input, $token_type) = @_; 433 while ($ttype != ANTLR::Runtime::Token->EOF && $ttype != $token_type) {
|