Home
last modified time | relevance | path

Searched refs:token_type (Results 1 – 25 of 36) sorted by relevance

12

/external/autotest/client/common_lib/cros/
Ddbus_send.py47 token_type = token_stream.pop(0)
48 if token_type == 'variant':
49 token_type = token_stream.pop(0)
50 if token_type == 'object':
51 token_type = token_stream.pop(0) # Should be 'path'
55 if token_type in INT_TYPES:
57 if token_type == 'string' or token_type == 'path':
59 if token_type == 'boolean':
61 if token_type == 'double':
63 if token_type == 'array':
[all …]
/external/v8/testing/gmock/scripts/generator/cpp/
Dtokenize.py74 def __init__(self, token_type, name, start, end): argument
75 self.token_type = token_type
147 token_type = UNKNOWN
151 token_type = NAME
159 token_type = CONSTANT
162 token_type = CONSTANT
173 token_type = SYNTAX
183 token_type = SYNTAX
186 token_type = CONSTANT
196 token_type = CONSTANT
[all …]
Dast.py551 if parts[-1].token_type == tokenize.NAME:
581 if (type_name and type_name[-1].token_type == tokenize.NAME and
582 p.token_type == tokenize.NAME):
714 if token.token_type == _INTERNAL_TOKEN:
739 if token.token_type == tokenize.NAME:
750 if next.token_type == tokenize.SYNTAX and next.name == '(':
806 elif token.token_type == tokenize.SYNTAX:
812 if (token.token_type == tokenize.NAME and
816 elif token.token_type == tokenize.PREPROCESSOR:
855 while (last_token.token_type != expected_token_type or
[all …]
/external/googletest/googlemock/scripts/generator/cpp/
Dtokenize.py74 def __init__(self, token_type, name, start, end): argument
75 self.token_type = token_type
147 token_type = UNKNOWN
151 token_type = NAME
159 token_type = CONSTANT
162 token_type = CONSTANT
173 token_type = SYNTAX
183 token_type = SYNTAX
186 token_type = CONSTANT
196 token_type = CONSTANT
[all …]
Dast.py551 if parts[-1].token_type == tokenize.NAME:
581 if (type_name and type_name[-1].token_type == tokenize.NAME and
582 p.token_type == tokenize.NAME):
714 if token.token_type == _INTERNAL_TOKEN:
739 if token.token_type == tokenize.NAME:
750 if next.token_type == tokenize.SYNTAX and next.name == '(':
806 elif token.token_type == tokenize.SYNTAX:
812 if (token.token_type == tokenize.NAME and
816 elif token.token_type == tokenize.PREPROCESSOR:
855 while (last_token.token_type != expected_token_type or
[all …]
/external/google-breakpad/src/testing/scripts/generator/cpp/
Dtokenize.py74 def __init__(self, token_type, name, start, end): argument
75 self.token_type = token_type
147 token_type = UNKNOWN
151 token_type = NAME
159 token_type = CONSTANT
162 token_type = CONSTANT
173 token_type = SYNTAX
183 token_type = SYNTAX
186 token_type = CONSTANT
196 token_type = CONSTANT
[all …]
Dast.py549 if parts[-1].token_type == tokenize.NAME:
579 if (type_name and type_name[-1].token_type == tokenize.NAME and
580 p.token_type == tokenize.NAME):
713 if token.token_type == _INTERNAL_TOKEN:
738 if token.token_type == tokenize.NAME:
749 if next.token_type == tokenize.SYNTAX and next.name == '(':
805 elif token.token_type == tokenize.SYNTAX:
811 if (token.token_type == tokenize.NAME and
815 elif token.token_type == tokenize.PREPROCESSOR:
854 while (last_token.token_type != expected_token_type or
[all …]
/external/protobuf/gtest/scripts/
Dpump.py141 def __init__(self, start=None, end=None, value=None, token_type=None): argument
151 self.token_type = token_type
155 self.start, self.value, self.token_type)
161 self.token_type)
172 for (regex, token_type) in token_table:
179 best_match_token_type = token_type
198 (start_column, length, token_type) = m
203 return MakeToken(lines, found_start, found_end, token_type)
227 def MakeToken(lines, start, end, token_type): argument
230 return Token(start, end, SubString(lines, start, end), token_type)
[all …]
/external/vulkan-validation-layers/tests/gtest-1.7.0/scripts/
Dpump.py140 def __init__(self, start=None, end=None, value=None, token_type=None): argument
150 self.token_type = token_type
154 self.start, self.value, self.token_type)
160 self.token_type)
171 for (regex, token_type) in token_table:
178 best_match_token_type = token_type
197 (start_column, length, token_type) = m
202 return MakeToken(lines, found_start, found_end, token_type)
237 def MakeToken(lines, start, end, token_type): argument
240 return Token(start, end, SubString(lines, start, end), token_type)
[all …]
/external/v8/testing/gtest/scripts/
Dpump.py140 def __init__(self, start=None, end=None, value=None, token_type=None): argument
150 self.token_type = token_type
154 self.start, self.value, self.token_type)
160 self.token_type)
171 for (regex, token_type) in token_table:
178 best_match_token_type = token_type
197 (start_column, length, token_type) = m
202 return MakeToken(lines, found_start, found_end, token_type)
237 def MakeToken(lines, start, end, token_type): argument
240 return Token(start, end, SubString(lines, start, end), token_type)
[all …]
/external/googletest/googletest/scripts/
Dpump.py140 def __init__(self, start=None, end=None, value=None, token_type=None): argument
150 self.token_type = token_type
154 self.start, self.value, self.token_type)
160 self.token_type)
171 for (regex, token_type) in token_table:
178 best_match_token_type = token_type
197 (start_column, length, token_type) = m
202 return MakeToken(lines, found_start, found_end, token_type)
237 def MakeToken(lines, start, end, token_type): argument
240 return Token(start, end, SubString(lines, start, end), token_type)
[all …]
/external/google-breakpad/src/testing/gtest/scripts/
Dpump.py140 def __init__(self, start=None, end=None, value=None, token_type=None): argument
150 self.token_type = token_type
154 self.start, self.value, self.token_type)
160 self.token_type)
171 for (regex, token_type) in token_table:
178 best_match_token_type = token_type
197 (start_column, length, token_type) = m
202 return MakeToken(lines, found_start, found_end, token_type)
237 def MakeToken(lines, start, end, token_type): argument
240 return Token(start, end, SubString(lines, start, end), token_type)
[all …]
/external/chromium-trace/catapult/common/py_utils/py_utils/refactor/
Doffset_token.py26 def __init__(self, token_type, string, offset): argument
27 self._type = token_type
70 token_type, string, (srow, scol), _, _ = next_token
72 offset_tokens.append(OffsetToken(token_type, string, (0, 0)))
76 offset_tokens.append(OffsetToken(token_type, string, (0, scol - ecol)))
79 token_type, string, (srow - erow, scol)))
Dsnippet.py97 def __init__(self, token_type, tokens): argument
102 self._type = token_type
107 def Create(cls, token_type, string, offset=(0, 0)): argument
108 return cls(token_type,
109 [offset_token.OffsetToken(token_type, string, offset)])
/external/bison/examples/calc++/
Dcalc++-scanner.ll16 /* By default yylex returns int, we use token_type.
18 not of token_type. */
44 [-+*/] return yy::calcxx_parser::token_type (yytext[0]);
Dcalc++-driver.hh10 yy::calcxx_parser::token_type \
Dcalc++-parser.hh106 typedef token::yytokentype token_type; typedef in yy::calcxx_parser
/external/libmojo/third_party/jinja2/
Dparser.py431 token_type = self.stream.current.type
432 if token_type in _compare_operators:
434 ops.append(nodes.Operand(token_type, self.parse_add()))
529 token_type = self.stream.current.type
531 if token_type == 'sub':
534 elif token_type == 'add':
662 token_type = self.stream.current.type
663 if token_type == 'dot' or token_type == 'lbracket':
667 elif token_type == 'lparen':
675 token_type = self.stream.current.type
[all …]
Dlexer.py145 def _describe_token_type(token_type): argument
146 if token_type in reverse_operators:
147 return reverse_operators[token_type]
161 }.get(token_type, token_type)
/external/antlr/antlr-3.4/runtime/Ruby/lib/antlr3/tree/
Ddebug.rb27 def create_from_token( token_type, from_token, text = nil ) argument
33 def create_from_type( token_type, text ) argument
/external/libmojo/mojo/public/tools/bindings/pylib/mojom_tests/parse/
Dlexer_unittest.py42 def _MakeLexToken(token_type, value, lineno=1, lexpos=0): argument
46 rv.type, rv.value, rv.lineno, rv.lexpos = token_type, value, lineno, lexpos
/external/v8/src/parsing/
Dtoken.cc45 const char Token::token_type[] = { member in v8::internal::Token
Dtoken.h199 return token_type[tok] == 'K'; in IsKeyword()
346 static const char token_type[NUM_TOKENS]; variable
/external/antlr/antlr-3.4/runtime/Ruby/lib/antlr3/
Dtree.rb858 def create_from_token( token_type, from_token, text = nil ) argument
860 from_token.type = token_type
866 def create_from_type( token_type, text ) argument
867 from_token = create_token( token_type, DEFAULT_CHANNEL, text )
/external/antlr/antlr-3.4/runtime/Perl5/lib/ANTLR/Runtime/
DBaseRecognizer.pm430 my ($self, $input, $token_type) = @_;
433 while ($ttype != ANTLR::Runtime::Token->EOF && $ttype != $token_type) {

12