Home
last modified time | relevance | path

Searched defs:tokens (Results 1 – 25 of 562) sorted by relevance

12345678910>>...23

/external/rust/crates/syn/tests/
Dtest_item.rs12 let tokens = TokenStream::from_iter(vec![ in test_macro_variable_attr() localVariable
53 let tokens = quote! { in test_negative_impl() localVariable
66 let tokens = quote! { in test_negative_impl() localVariable
78 let tokens = quote! { in test_negative_impl() localVariable
111 let tokens = quote! { in test_negative_impl() localVariable
125 let tokens = TokenStream::from_iter(vec![ in test_macro_variable_impl() localVariable
168 let tokens = quote!(trait Trait where {}); in test_supertraits() localVariable
180 let tokens = quote!(trait Trait: where {}); in test_supertraits() localVariable
193 let tokens = quote!(trait Trait: Sized where {}); in test_supertraits() localVariable
219 let tokens = quote!(trait Trait: Sized + where {}); in test_supertraits() localVariable
[all …]
Dtest_expr.rs11 let tokens = quote!(..100u32); in test_expr_parse() localVariable
21 let tokens = quote!(..100u32); in test_expr_parse() localVariable
35 let tokens = quote!(fut.await); in test_await() localVariable
124 let tokens = TokenStream::from_iter(vec![ in test_macro_variable_func() localVariable
146 let tokens = TokenStream::from_iter(vec![ in test_macro_variable_func() localVariable
202 let tokens = TokenStream::from_iter(vec![ in test_macro_variable_macro() localVariable
229 let tokens = TokenStream::from_iter(vec![ in test_macro_variable_struct() localVariable
251 let tokens = TokenStream::from_iter(vec![ in test_macro_variable_match_arm() localVariable
Dtest_ty.rs21 let tokens = TokenStream::from_iter(vec![ in test_macro_variable_type() localVariable
55 let tokens = TokenStream::from_iter(vec![ in test_macro_variable_type() localVariable
95 let tokens = TokenStream::from_iter(vec![ in test_group_angle_brackets() localVariable
147 let tokens = TokenStream::from_iter(vec![ in test_group_colons() localVariable
184 let tokens = TokenStream::from_iter(vec![ in test_group_colons() localVariable
Dtest_generics.rs127 let tokens = quote!('a); in test_ty_param_bound() localVariable
134 let tokens = quote!('_); in test_ty_param_bound() localVariable
141 let tokens = quote!(Debug); in test_ty_param_bound() localVariable
156 let tokens = quote!(?Sized); in test_ty_param_bound() localVariable
/external/rust/crates/quote/tests/
Dtest.rs19 let tokens = quote! { in test_quote_impl() localVariable
41 let tokens = quote_spanned! {span=> in test_quote_spanned_impl() localVariable
63 let tokens = quote!(#x <#x> (#x) [#x] {#x}); in test_substitution() localVariable
95 let tokens = quote! { in test_advanced() localVariable
151 let tokens = quote! { in test_integer() localVariable
165 let tokens = quote! { in test_floating() localVariable
182 let tokens = quote! { in test_char() localVariable
192 let tokens = quote!(#s); in test_str() localVariable
200 let tokens = quote!(#s); in test_string() localVariable
209 let tokens = quote!(struct #foo; enum #bar {}); in test_ident() localVariable
[all …]
/external/antlr/tool/src/test/java/org/antlr/test/
DTestTokenRewriteStream.java53 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); in testInsertBeforeIndex0() local
68 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); in testInsertAfterLastIndex() local
83 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); in test2InsertBeforeAfterMiddleIndex() local
100 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); in testReplaceIndex0() local
116 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); in testReplaceLastIndex() local
132 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); in testReplaceMiddleIndex() local
153 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); in testToStringStartStop() local
188 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); in testToStringStartStop2() local
235 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); in test2ReplaceMiddleIndex() local
252 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); in test2ReplaceMiddleIndex1InsertBefore() local
[all …]
DTestCommonTokenStream.java54 BufferedTokenStream tokens = new BufferedTokenStream(lexEngine); in testFirstToken() local
75 BufferedTokenStream tokens = new BufferedTokenStream(lexEngine); in test2ndToken() local
96 BufferedTokenStream tokens = new BufferedTokenStream(lexEngine); in testCompleteBuffer() local
126 BufferedTokenStream tokens = new BufferedTokenStream(lexEngine); in testCompleteBufferAfterConsuming() local
157 BufferedTokenStream tokens = new BufferedTokenStream(lexEngine); in testLookback() local
204 CommonTokenStream tokens = new CommonTokenStream(lexer); in testOffChannel() local
DTestInterpretedParsing.java68 FilteringTokenStream tokens = new FilteringTokenStream(lexEngine); in testSimpleParse() local
103 FilteringTokenStream tokens = new FilteringTokenStream(lexEngine); in testMismatchedTokenError() local
138 FilteringTokenStream tokens = new FilteringTokenStream(lexEngine); in testMismatchedSetError() local
173 FilteringTokenStream tokens = new FilteringTokenStream(lexEngine); in testNoViableAltError() local
/external/minijail/tools/
Dparser.py263 def _parse_constant(self, tokens): argument
288 def parse_value(self, tokens): argument
317 def _parse_atom(self, tokens): argument
338 def _parse_clause(self, tokens): argument
351 def parse_argument_expression(self, tokens): argument
375 def _parse_default_action(self, tokens): argument
405 def parse_action(self, tokens): argument
439 def _parse_single_filter(self, tokens): argument
457 def parse_filter(self, tokens): argument
478 def _parse_key_value_pair(self, tokens): argument
[all …]
/external/libtextclassifier/native/utils/
Dtokenizer-utils_test.cc26 std::vector<Token> tokens = in TEST() local
69 std::vector<Token> tokens = TokenizeOnDelimiters( in TEST() local
100 std::vector<Token> tokens = TokenizeOnDelimiters( in TEST() local
146 std::vector<Token> tokens = in TEST() local
158 std::vector<Token> tokens = in TEST() local
169 std::vector<Token> tokens = in TEST() local
178 std::vector<Token> tokens = in TEST() local
192 std::vector<Token> tokens = in TEST() local
Dtokenizer_test.cc178 std::vector<Token> tokens = tokenizer.Tokenize("Hello world!"); in TEST() local
349 std::vector<Token> tokens; in TEST() local
379 std::vector<Token> tokens = tokenizer.Tokenize("พระบาท สมเด็จ พระ ปร มิ"); in TEST() local
399 std::vector<Token> tokens = in TEST() local
427 std::vector<Token> tokens = tokenizer.Tokenize("3.1 3﹒2 3.3"); in TEST() local
445 std::vector<Token> tokens = tokenizer.Tokenize("พระบาทสมเด็จพระปรมิ"); in TEST() local
495 std::vector<Token> tokens = tokenizer.Tokenize( in TEST() local
548 std::vector<Token> tokens = tokenizer.Tokenize("7% -3.14 68.9#? 7% $99 .18."); in TEST() local
565 std::vector<Token> tokens = tokenizer.Tokenize("2 pércént 3パーセント"); in TEST() local
577 std::vector<Token> tokens = tokenizer.Tokenize("3 3﹒2 3.3%"); in TEST() local
[all …]
Dbert_tokenizer.cc63 std::vector<std::string> tokens = {token}; in TokenizeSingleToken() local
68 std::vector<std::string> tokens = PreTokenize(input); in Tokenize() local
73 const std::vector<std::string>& tokens) { in Tokenize()
106 const std::vector<Token> tokens = in PreTokenize() local
/external/libtextclassifier/native/annotator/duration/
Dduration_test.cc152 std::vector<Token> tokens = Tokenize(text); in TEST_F() local
171 std::vector<Token> tokens = Tokenize(text); in TEST_F() local
190 std::vector<Token> tokens = Tokenize(text); in TEST_F() local
209 std::vector<Token> tokens = Tokenize(text); in TEST_F() local
228 std::vector<Token> tokens = Tokenize(text); in TEST_F() local
247 std::vector<Token> tokens = Tokenize(text); in TEST_F() local
266 std::vector<Token> tokens = Tokenize(text); in TEST_F() local
286 std::vector<Token> tokens = Tokenize(text); in TEST_F() local
305 std::vector<Token> tokens = Tokenize(text); in TEST_F() local
323 std::vector<Token> tokens = Tokenize(text); in TEST_F() local
[all …]
/external/libtextclassifier/native/utils/grammar/parsing/
Dlexer_test.cc88 std::vector<Token> tokens = tokenizer_.Tokenize("This is a word"); in TEST_F() local
97 std::vector<Token> tokens = tokenizer_.Tokenize("1234This a4321cde"); in TEST_F() local
107 std::vector<Token> tokens = tokenizer_.Tokenize("10/18/2014"); in TEST_F() local
117 std::vector<Token> tokens = tokenizer_.Tokenize("电话:0871—6857(曹"); in TEST_F() local
130 std::vector<Token> tokens = tokenizer_.Tokenize("电话 :0871—6857(曹"); in TEST_F() local
143 std::vector<Token> tokens = in TEST_F() local
158 std::vector<Token> tokens = tokenizer_.Tokenize("The+2345++the +"); in TEST_F() local
/external/pdfium/samples/
Dpdfium_test_event_helper.cc19 const std::vector<std::string>& tokens) { in SendCharCodeEvent()
31 const std::vector<std::string>& tokens) { in SendKeyCodeEvent()
56 const std::vector<std::string>& tokens) { in SendMouseDownEvent()
76 const std::vector<std::string>& tokens) { in SendMouseUpEvent()
95 const std::vector<std::string>& tokens) { in SendMouseDoubleClickEvent()
113 const std::vector<std::string>& tokens) { in SendMouseMoveEvent()
126 const std::vector<std::string>& tokens) { in SendFocusEvent()
146 auto tokens = StringSplit(command[0], ','); in SendPageEvents() local
/external/llvm-project/flang/lib/Parser/
Dtoken-sequence.cpp51 std::size_t tokens{start_.size()}; in SkipBlanks() local
65 std::size_t tokens{start_.size()}; in IsAnythingLeft() local
110 std::size_t tokens{that.SizeInTokens()}; in Put() local
120 const TokenSequence &that, std::size_t at, std::size_t tokens) { in Put()
158 std::size_t tokens{start_.size()}; in ToLowerCase() local
206 std::size_t tokens{SizeInTokens()}; in HasBlanks() local
216 std::size_t tokens{SizeInTokens()}; in HasRedundantBlanks() local
229 std::size_t tokens{SizeInTokens()}; in RemoveBlanks() local
241 std::size_t tokens{SizeInTokens()}; in RemoveRedundantBlanks() local
256 std::size_t tokens{SizeInTokens()}; in ClipComment() local
[all …]
/external/libtextclassifier/native/annotator/
Dfeature-processor_test.cc87 std::vector<Token> tokens{Token("Hělló", 0, 5), in TEST_F() local
104 std::vector<Token> tokens{Token("Hělló", 0, 5), in TEST_F() local
120 std::vector<Token> tokens{Token("Hělló", 0, 5), in TEST_F() local
136 std::vector<Token> tokens{Token("Hělló", 0, 5), in TEST_F() local
152 std::vector<Token> tokens{Token("Hělló", 0, 5), in TEST_F() local
179 std::vector<Token> tokens = {Token("Fiřst", 0, 5), in TEST_F() local
204 std::vector<Token> tokens = {Token("Fiřst", 0, 5), in TEST_F() local
229 std::vector<Token> tokens = {Token("Fiřst", 0, 5), in TEST_F() local
254 std::vector<Token> tokens = {Token("Fiřst", 0, 5), in TEST_F() local
281 std::vector<Token> tokens = {Token("Fiřst", 0, 5), in TEST_F() local
[all …]
/external/jackson-databind/src/main/java/com/fasterxml/jackson/databind/type/
DTypeParser.java32 MyTokenizer tokens = new MyTokenizer(canonical.trim()); in parse() local
41 protected JavaType parseType(MyTokenizer tokens) in parseType()
63 protected List<JavaType> parseTypes(MyTokenizer tokens) in parseTypes()
79 protected Class<?> findClass(String className, MyTokenizer tokens) in findClass()
89 protected IllegalArgumentException _problem(MyTokenizer tokens, String msg) in _problem()
/external/mesa3d/src/gallium/auxiliary/tgsi/
Dtgsi_parse.c36 const struct tgsi_token *tokens ) in tgsi_parse_init()
279 tgsi_dup_tokens(const struct tgsi_token *tokens) in tgsi_dup_tokens()
305 tgsi_free_tokens(const struct tgsi_token *tokens) in tgsi_free_tokens()
312 tgsi_dump_tokens(const struct tgsi_token *tokens) in tgsi_dump_tokens()
327 tgsi_get_processor_type(const struct tgsi_token *tokens) in tgsi_get_processor_type()
/external/mesa3d/src/gallium/drivers/nouveau/
Dnouveau_compiler.c38 nv30_fp(int chipset, struct tgsi_token tokens[], in nv30_fp()
51 nv30_vp(int chipset, struct tgsi_token tokens[], in nv30_vp()
65 nv30_codegen(int chipset, int type, struct tgsi_token tokens[], in nv30_codegen()
105 nouveau_codegen(int chipset, int type, struct tgsi_token tokens[], in nouveau_codegen()
142 struct tgsi_token tokens[4096]; in main() local
/external/python/cpython3/Parser/pgen/
Dtoken.py4 def generate_tokens(tokens): argument
19 def generate_opmap(tokens): argument
/external/libtextclassifier/native/annotator/number/
Dnumber.cc77 bool NumberAnnotator::TokensAreValidStart(const std::vector<Token>& tokens, in TokensAreValidStart()
86 const std::vector<Token>& tokens, const int prefix_end_index) const { in TokensAreValidNumberPrefix() argument
120 bool NumberAnnotator::TokensAreValidEnding(const std::vector<Token>& tokens, in TokensAreValidEnding()
144 const std::vector<Token>& tokens, const int suffix_start_index) const { in TokensAreValidNumberSuffix() argument
179 const std::vector<Token>& tokens, in FindPercentSuffixEndCodepoint()
224 const std::vector<Token> tokens = tokenizer_.Tokenize(context); in FindAll() local
/external/ply/ply/ply/
Dcpp.py26 tokens = ( variable
314 def tokenstrip(self,tokens): argument
489 def expand_macros(self,tokens,expanded=None): argument
561 def evalexpr(self,tokens): argument
751 def include(self,tokens): argument
800 def define(self,tokens): argument
870 def undef(self,tokens): argument
/external/python/pycparser/pycparser/ply/
Dcpp.py24 tokens = ( variable
312 def tokenstrip(self,tokens): argument
486 def expand_macros(self,tokens,expanded=None): argument
552 def evalexpr(self,tokens): argument
742 def include(self,tokens): argument
791 def define(self,tokens): argument
861 def undef(self,tokens): argument
/external/mesa3d/src/gallium/targets/graw-null/
Dgraw_util.c19 struct tgsi_token tokens[1024]; in graw_parse_geometry_shader() local
34 struct tgsi_token tokens[1024]; in graw_parse_vertex_shader() local
49 struct tgsi_token tokens[1024]; in graw_parse_fragment_shader() local

12345678910>>...23