/external/rust/crates/syn/tests/ |
D | test_item.rs | 12 let tokens = TokenStream::from_iter(vec![ in test_macro_variable_attr() localVariable 53 let tokens = quote! { in test_negative_impl() localVariable 66 let tokens = quote! { in test_negative_impl() localVariable 78 let tokens = quote! { in test_negative_impl() localVariable 111 let tokens = quote! { in test_negative_impl() localVariable 125 let tokens = TokenStream::from_iter(vec![ in test_macro_variable_impl() localVariable 168 let tokens = quote!(trait Trait where {}); in test_supertraits() localVariable 180 let tokens = quote!(trait Trait: where {}); in test_supertraits() localVariable 193 let tokens = quote!(trait Trait: Sized where {}); in test_supertraits() localVariable 219 let tokens = quote!(trait Trait: Sized + where {}); in test_supertraits() localVariable [all …]
|
D | test_expr.rs | 11 let tokens = quote!(..100u32); in test_expr_parse() localVariable 21 let tokens = quote!(..100u32); in test_expr_parse() localVariable 35 let tokens = quote!(fut.await); in test_await() localVariable 124 let tokens = TokenStream::from_iter(vec![ in test_macro_variable_func() localVariable 146 let tokens = TokenStream::from_iter(vec![ in test_macro_variable_func() localVariable 202 let tokens = TokenStream::from_iter(vec![ in test_macro_variable_macro() localVariable 229 let tokens = TokenStream::from_iter(vec![ in test_macro_variable_struct() localVariable 251 let tokens = TokenStream::from_iter(vec![ in test_macro_variable_match_arm() localVariable
|
D | test_ty.rs | 21 let tokens = TokenStream::from_iter(vec![ in test_macro_variable_type() localVariable 55 let tokens = TokenStream::from_iter(vec![ in test_macro_variable_type() localVariable 95 let tokens = TokenStream::from_iter(vec![ in test_group_angle_brackets() localVariable 147 let tokens = TokenStream::from_iter(vec![ in test_group_colons() localVariable 184 let tokens = TokenStream::from_iter(vec![ in test_group_colons() localVariable
|
D | test_generics.rs | 127 let tokens = quote!('a); in test_ty_param_bound() localVariable 134 let tokens = quote!('_); in test_ty_param_bound() localVariable 141 let tokens = quote!(Debug); in test_ty_param_bound() localVariable 156 let tokens = quote!(?Sized); in test_ty_param_bound() localVariable
|
/external/rust/crates/quote/tests/ |
D | test.rs | 19 let tokens = quote! { in test_quote_impl() localVariable 41 let tokens = quote_spanned! {span=> in test_quote_spanned_impl() localVariable 63 let tokens = quote!(#x <#x> (#x) [#x] {#x}); in test_substitution() localVariable 95 let tokens = quote! { in test_advanced() localVariable 151 let tokens = quote! { in test_integer() localVariable 165 let tokens = quote! { in test_floating() localVariable 182 let tokens = quote! { in test_char() localVariable 192 let tokens = quote!(#s); in test_str() localVariable 200 let tokens = quote!(#s); in test_string() localVariable 209 let tokens = quote!(struct #foo; enum #bar {}); in test_ident() localVariable [all …]
|
/external/antlr/tool/src/test/java/org/antlr/test/ |
D | TestTokenRewriteStream.java | 53 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); in testInsertBeforeIndex0() local 68 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); in testInsertAfterLastIndex() local 83 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); in test2InsertBeforeAfterMiddleIndex() local 100 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); in testReplaceIndex0() local 116 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); in testReplaceLastIndex() local 132 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); in testReplaceMiddleIndex() local 153 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); in testToStringStartStop() local 188 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); in testToStringStartStop2() local 235 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); in test2ReplaceMiddleIndex() local 252 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); in test2ReplaceMiddleIndex1InsertBefore() local [all …]
|
D | TestCommonTokenStream.java | 54 BufferedTokenStream tokens = new BufferedTokenStream(lexEngine); in testFirstToken() local 75 BufferedTokenStream tokens = new BufferedTokenStream(lexEngine); in test2ndToken() local 96 BufferedTokenStream tokens = new BufferedTokenStream(lexEngine); in testCompleteBuffer() local 126 BufferedTokenStream tokens = new BufferedTokenStream(lexEngine); in testCompleteBufferAfterConsuming() local 157 BufferedTokenStream tokens = new BufferedTokenStream(lexEngine); in testLookback() local 204 CommonTokenStream tokens = new CommonTokenStream(lexer); in testOffChannel() local
|
D | TestInterpretedParsing.java | 68 FilteringTokenStream tokens = new FilteringTokenStream(lexEngine); in testSimpleParse() local 103 FilteringTokenStream tokens = new FilteringTokenStream(lexEngine); in testMismatchedTokenError() local 138 FilteringTokenStream tokens = new FilteringTokenStream(lexEngine); in testMismatchedSetError() local 173 FilteringTokenStream tokens = new FilteringTokenStream(lexEngine); in testNoViableAltError() local
|
/external/minijail/tools/ |
D | parser.py | 263 def _parse_constant(self, tokens): argument 288 def parse_value(self, tokens): argument 317 def _parse_atom(self, tokens): argument 338 def _parse_clause(self, tokens): argument 351 def parse_argument_expression(self, tokens): argument 375 def _parse_default_action(self, tokens): argument 405 def parse_action(self, tokens): argument 439 def _parse_single_filter(self, tokens): argument 457 def parse_filter(self, tokens): argument 478 def _parse_key_value_pair(self, tokens): argument [all …]
|
/external/libtextclassifier/native/utils/ |
D | tokenizer-utils_test.cc | 26 std::vector<Token> tokens = in TEST() local 69 std::vector<Token> tokens = TokenizeOnDelimiters( in TEST() local 100 std::vector<Token> tokens = TokenizeOnDelimiters( in TEST() local 146 std::vector<Token> tokens = in TEST() local 158 std::vector<Token> tokens = in TEST() local 169 std::vector<Token> tokens = in TEST() local 178 std::vector<Token> tokens = in TEST() local 192 std::vector<Token> tokens = in TEST() local
|
D | tokenizer_test.cc | 178 std::vector<Token> tokens = tokenizer.Tokenize("Hello world!"); in TEST() local 349 std::vector<Token> tokens; in TEST() local 379 std::vector<Token> tokens = tokenizer.Tokenize("พระบาท สมเด็จ พระ ปร มิ"); in TEST() local 399 std::vector<Token> tokens = in TEST() local 427 std::vector<Token> tokens = tokenizer.Tokenize("3.1 3﹒2 3.3"); in TEST() local 445 std::vector<Token> tokens = tokenizer.Tokenize("พระบาทสมเด็จพระปรมิ"); in TEST() local 495 std::vector<Token> tokens = tokenizer.Tokenize( in TEST() local 548 std::vector<Token> tokens = tokenizer.Tokenize("7% -3.14 68.9#? 7% $99 .18."); in TEST() local 565 std::vector<Token> tokens = tokenizer.Tokenize("2 pércént 3パーセント"); in TEST() local 577 std::vector<Token> tokens = tokenizer.Tokenize("3 3﹒2 3.3%"); in TEST() local [all …]
|
D | bert_tokenizer.cc | 63 std::vector<std::string> tokens = {token}; in TokenizeSingleToken() local 68 std::vector<std::string> tokens = PreTokenize(input); in Tokenize() local 73 const std::vector<std::string>& tokens) { in Tokenize() 106 const std::vector<Token> tokens = in PreTokenize() local
|
/external/libtextclassifier/native/annotator/duration/ |
D | duration_test.cc | 152 std::vector<Token> tokens = Tokenize(text); in TEST_F() local 171 std::vector<Token> tokens = Tokenize(text); in TEST_F() local 190 std::vector<Token> tokens = Tokenize(text); in TEST_F() local 209 std::vector<Token> tokens = Tokenize(text); in TEST_F() local 228 std::vector<Token> tokens = Tokenize(text); in TEST_F() local 247 std::vector<Token> tokens = Tokenize(text); in TEST_F() local 266 std::vector<Token> tokens = Tokenize(text); in TEST_F() local 286 std::vector<Token> tokens = Tokenize(text); in TEST_F() local 305 std::vector<Token> tokens = Tokenize(text); in TEST_F() local 323 std::vector<Token> tokens = Tokenize(text); in TEST_F() local [all …]
|
/external/libtextclassifier/native/utils/grammar/parsing/ |
D | lexer_test.cc | 88 std::vector<Token> tokens = tokenizer_.Tokenize("This is a word"); in TEST_F() local 97 std::vector<Token> tokens = tokenizer_.Tokenize("1234This a4321cde"); in TEST_F() local 107 std::vector<Token> tokens = tokenizer_.Tokenize("10/18/2014"); in TEST_F() local 117 std::vector<Token> tokens = tokenizer_.Tokenize("电话:0871—6857(曹"); in TEST_F() local 130 std::vector<Token> tokens = tokenizer_.Tokenize("电话 :0871—6857(曹"); in TEST_F() local 143 std::vector<Token> tokens = in TEST_F() local 158 std::vector<Token> tokens = tokenizer_.Tokenize("The+2345++the +"); in TEST_F() local
|
/external/pdfium/samples/ |
D | pdfium_test_event_helper.cc | 19 const std::vector<std::string>& tokens) { in SendCharCodeEvent() 31 const std::vector<std::string>& tokens) { in SendKeyCodeEvent() 56 const std::vector<std::string>& tokens) { in SendMouseDownEvent() 76 const std::vector<std::string>& tokens) { in SendMouseUpEvent() 95 const std::vector<std::string>& tokens) { in SendMouseDoubleClickEvent() 113 const std::vector<std::string>& tokens) { in SendMouseMoveEvent() 126 const std::vector<std::string>& tokens) { in SendFocusEvent() 146 auto tokens = StringSplit(command[0], ','); in SendPageEvents() local
|
/external/llvm-project/flang/lib/Parser/ |
D | token-sequence.cpp | 51 std::size_t tokens{start_.size()}; in SkipBlanks() local 65 std::size_t tokens{start_.size()}; in IsAnythingLeft() local 110 std::size_t tokens{that.SizeInTokens()}; in Put() local 120 const TokenSequence &that, std::size_t at, std::size_t tokens) { in Put() 158 std::size_t tokens{start_.size()}; in ToLowerCase() local 206 std::size_t tokens{SizeInTokens()}; in HasBlanks() local 216 std::size_t tokens{SizeInTokens()}; in HasRedundantBlanks() local 229 std::size_t tokens{SizeInTokens()}; in RemoveBlanks() local 241 std::size_t tokens{SizeInTokens()}; in RemoveRedundantBlanks() local 256 std::size_t tokens{SizeInTokens()}; in ClipComment() local [all …]
|
/external/libtextclassifier/native/annotator/ |
D | feature-processor_test.cc | 87 std::vector<Token> tokens{Token("Hělló", 0, 5), in TEST_F() local 104 std::vector<Token> tokens{Token("Hělló", 0, 5), in TEST_F() local 120 std::vector<Token> tokens{Token("Hělló", 0, 5), in TEST_F() local 136 std::vector<Token> tokens{Token("Hělló", 0, 5), in TEST_F() local 152 std::vector<Token> tokens{Token("Hělló", 0, 5), in TEST_F() local 179 std::vector<Token> tokens = {Token("Fiřst", 0, 5), in TEST_F() local 204 std::vector<Token> tokens = {Token("Fiřst", 0, 5), in TEST_F() local 229 std::vector<Token> tokens = {Token("Fiřst", 0, 5), in TEST_F() local 254 std::vector<Token> tokens = {Token("Fiřst", 0, 5), in TEST_F() local 281 std::vector<Token> tokens = {Token("Fiřst", 0, 5), in TEST_F() local [all …]
|
/external/jackson-databind/src/main/java/com/fasterxml/jackson/databind/type/ |
D | TypeParser.java | 32 MyTokenizer tokens = new MyTokenizer(canonical.trim()); in parse() local 41 protected JavaType parseType(MyTokenizer tokens) in parseType() 63 protected List<JavaType> parseTypes(MyTokenizer tokens) in parseTypes() 79 protected Class<?> findClass(String className, MyTokenizer tokens) in findClass() 89 protected IllegalArgumentException _problem(MyTokenizer tokens, String msg) in _problem()
|
/external/mesa3d/src/gallium/auxiliary/tgsi/ |
D | tgsi_parse.c | 36 const struct tgsi_token *tokens ) in tgsi_parse_init() 279 tgsi_dup_tokens(const struct tgsi_token *tokens) in tgsi_dup_tokens() 305 tgsi_free_tokens(const struct tgsi_token *tokens) in tgsi_free_tokens() 312 tgsi_dump_tokens(const struct tgsi_token *tokens) in tgsi_dump_tokens() 327 tgsi_get_processor_type(const struct tgsi_token *tokens) in tgsi_get_processor_type()
|
/external/mesa3d/src/gallium/drivers/nouveau/ |
D | nouveau_compiler.c | 38 nv30_fp(int chipset, struct tgsi_token tokens[], in nv30_fp() 51 nv30_vp(int chipset, struct tgsi_token tokens[], in nv30_vp() 65 nv30_codegen(int chipset, int type, struct tgsi_token tokens[], in nv30_codegen() 105 nouveau_codegen(int chipset, int type, struct tgsi_token tokens[], in nouveau_codegen() 142 struct tgsi_token tokens[4096]; in main() local
|
/external/python/cpython3/Parser/pgen/ |
D | token.py | 4 def generate_tokens(tokens): argument 19 def generate_opmap(tokens): argument
|
/external/libtextclassifier/native/annotator/number/ |
D | number.cc | 77 bool NumberAnnotator::TokensAreValidStart(const std::vector<Token>& tokens, in TokensAreValidStart() 86 const std::vector<Token>& tokens, const int prefix_end_index) const { in TokensAreValidNumberPrefix() argument 120 bool NumberAnnotator::TokensAreValidEnding(const std::vector<Token>& tokens, in TokensAreValidEnding() 144 const std::vector<Token>& tokens, const int suffix_start_index) const { in TokensAreValidNumberSuffix() argument 179 const std::vector<Token>& tokens, in FindPercentSuffixEndCodepoint() 224 const std::vector<Token> tokens = tokenizer_.Tokenize(context); in FindAll() local
|
/external/ply/ply/ply/ |
D | cpp.py | 26 tokens = ( variable 314 def tokenstrip(self,tokens): argument 489 def expand_macros(self,tokens,expanded=None): argument 561 def evalexpr(self,tokens): argument 751 def include(self,tokens): argument 800 def define(self,tokens): argument 870 def undef(self,tokens): argument
|
/external/python/pycparser/pycparser/ply/ |
D | cpp.py | 24 tokens = ( variable 312 def tokenstrip(self,tokens): argument 486 def expand_macros(self,tokens,expanded=None): argument 552 def evalexpr(self,tokens): argument 742 def include(self,tokens): argument 791 def define(self,tokens): argument 861 def undef(self,tokens): argument
|
/external/mesa3d/src/gallium/targets/graw-null/ |
D | graw_util.c | 19 struct tgsi_token tokens[1024]; in graw_parse_geometry_shader() local 34 struct tgsi_token tokens[1024]; in graw_parse_vertex_shader() local 49 struct tgsi_token tokens[1024]; in graw_parse_fragment_shader() local
|