/external/vixl/src/aarch64/ |
D | debugger-aarch64.cc | 63 static Token* Tokenize(const char* arg); 100 static Token* Tokenize(const char* arg); 122 static Token* Tokenize(const char* arg); 148 static Token* Tokenize(const char* arg); 166 static Token* Tokenize(const char* arg); 183 static Token* Tokenize(const char* arg); 216 static Token* Tokenize(const char* arg); 785 Token* Token::Tokenize(const char* arg) { in Tokenize() function in vixl::aarch64::Token 793 Token* token = RegisterToken::Tokenize(arg); in Tokenize() 798 token = FPRegisterToken::Tokenize(arg); in Tokenize() [all …]
|
/external/libtextclassifier/ |
D | tokenizer_test.cc | 70 std::vector<Token> Tokenize(const std::string& utf8_text) const { in Tokenize() function in libtextclassifier2::__anona3edc44c0111::TestingTokenizerProxy 71 return tokenizer_->Tokenize(utf8_text); in Tokenize() 146 std::vector<Token> tokens = tokenizer.Tokenize("Hello world!"); in TEST() 177 EXPECT_THAT(tokenizer.Tokenize("앨라배마 주 전화(123) 456-789웹사이트"), in TEST() 311 tokens = tokenizer.Tokenize( in TEST() 315 tokens = tokenizer.Tokenize("問少目 hello 木輸ยามきゃ"); in TEST()
|
D | tokenizer.cc | 79 std::vector<Token> Tokenizer::Tokenize(const std::string& text) const { in Tokenize() function in libtextclassifier2::Tokenizer 81 return Tokenize(text_unicode); in Tokenize() 84 std::vector<Token> Tokenizer::Tokenize(const UnicodeText& text_unicode) const { in Tokenize() function in libtextclassifier2::Tokenizer
|
D | tokenizer.h | 42 std::vector<Token> Tokenize(const std::string& text) const; 45 std::vector<Token> Tokenize(const UnicodeText& text_unicode) const;
|
D | feature-processor_test.cc | 313 std::vector<Token> tokens = feature_processor.Tokenize("one, two, three"); in TEST() 354 tokens = feature_processor3.Tokenize("zero, one, two, three, four"); in TEST() 388 std::vector<Token> tokens = feature_processor.Tokenize("one, two, three"); in TEST() 429 tokens = feature_processor3.Tokenize("zero, one, two, three, four"); in TEST() 564 {0, 3}, feature_processor.Tokenize("aaa bbb ccc")), in TEST() 567 {0, 3}, feature_processor.Tokenize("aaa bbb ěěě")), in TEST() 570 {0, 3}, feature_processor.Tokenize("ěěě řřř ěěě")), in TEST() 860 EXPECT_EQ(feature_processor.Tokenize("앨라배마123웹사이트"), in TEST() 870 EXPECT_EQ(feature_processor2.Tokenize("앨라배마123웹사이트"), in TEST() 883 std::vector<Token> tokens = feature_processor.Tokenize("พระบาทสมเด็จพระปรมิ"); in TEST() [all …]
|
D | feature-processor.h | 127 std::vector<Token> Tokenize(const std::string& text) const; 130 std::vector<Token> Tokenize(const UnicodeText& text_unicode) const;
|
D | feature-processor.cc | 178 std::vector<Token> FeatureProcessor::Tokenize(const std::string& text) const { in Tokenize() function in libtextclassifier2::FeatureProcessor 180 return Tokenize(text_unicode); in Tokenize() 183 std::vector<Token> FeatureProcessor::Tokenize( in Tokenize() function in libtextclassifier2::FeatureProcessor 187 return tokenizer_.Tokenize(text_unicode); in Tokenize() 204 return tokenizer_.Tokenize(text_unicode); in Tokenize() 923 std::vector<Token> tokens = tokenizer_.Tokenize(text); in TokenizeSubstring()
|
D | text-classifier.cc | 658 *tokens = selection_feature_processor_->Tokenize(context_unicode); in ModelSuggestSelection() 825 tokens = classification_feature_processor_->Tokenize(context); in ModelClassifyText() 1122 *tokens = selection_feature_processor_->Tokenize(line_str); in ModelAnnotate()
|
/external/google-breakpad/src/processor/ |
D | basic_source_line_resolver.cc | 471 if (!Tokenize(file_line, kWhitespace, 2, &tokens)) { in ParseFile() 499 if (!Tokenize(function_line, kWhitespace, 4, &tokens)) { in ParseFunction() 531 if (!Tokenize(line_line, kWhitespace, 4, &tokens)) { in ParseLine() 580 if (!Tokenize(public_line, kWhitespace, 3, &tokens)) { in ParsePublicSymbol()
|
D | tokenize.h | 53 bool Tokenize(char *line,
|
D | tokenize.cc | 45 bool Tokenize(char *line, in Tokenize() function
|
D | windows_frame_info.h | 126 if (!Tokenize(&buffer[0], " \r\n", 11, &tokens)) in ParseFromString()
|
/external/tensorflow/tensorflow/contrib/lite/testing/ |
D | tokenize.h | 37 void Tokenize(std::istream* input, TokenProcessor* processor);
|
D | message.cc | 91 Tokenize(input, &stack); in Read()
|
D | tokenize.cc | 23 void Tokenize(std::istream* input, TokenProcessor* processor) { in Tokenize() function
|
D | tokenize_test.cc | 39 Tokenize(&ss, &collector); in TokenizeString()
|
/external/chromium-trace/catapult/common/py_utils/py_utils/refactor/ |
D | offset_token.py | 55 def Tokenize(f): function
|
D | snippet.py | 206 tokens = offset_token.Tokenize(f)
|
/external/google-breakpad/src/testing/gtest/scripts/ |
D | pump.py | 382 def Tokenize(s): function 579 tokens = list(Tokenize(pump_src_text))
|
/external/protobuf/gtest/scripts/ |
D | pump.py | 376 def Tokenize(s): function 571 for token in Tokenize(s):
|
/external/v8/testing/gtest/scripts/ |
D | pump.py | 382 def Tokenize(s): function 579 tokens = list(Tokenize(pump_src_text))
|
/external/vulkan-validation-layers/tests/gtest-1.7.0/scripts/ |
D | pump.py | 382 def Tokenize(s): function 579 tokens = list(Tokenize(pump_src_text))
|
/external/googletest/googletest/scripts/ |
D | pump.py | 382 def Tokenize(s): function 579 tokens = list(Tokenize(pump_src_text))
|