Home
last modified time | relevance | path

Searched refs:Tokenize (Results 1 – 23 of 23) sorted by relevance

/external/vixl/src/aarch64/
Ddebugger-aarch64.cc63 static Token* Tokenize(const char* arg);
100 static Token* Tokenize(const char* arg);
122 static Token* Tokenize(const char* arg);
148 static Token* Tokenize(const char* arg);
166 static Token* Tokenize(const char* arg);
183 static Token* Tokenize(const char* arg);
216 static Token* Tokenize(const char* arg);
785 Token* Token::Tokenize(const char* arg) { in Tokenize() function in vixl::aarch64::Token
793 Token* token = RegisterToken::Tokenize(arg); in Tokenize()
798 token = FPRegisterToken::Tokenize(arg); in Tokenize()
[all …]
/external/libtextclassifier/
Dtokenizer_test.cc70 std::vector<Token> Tokenize(const std::string& utf8_text) const { in Tokenize() function in libtextclassifier2::__anona3edc44c0111::TestingTokenizerProxy
71 return tokenizer_->Tokenize(utf8_text); in Tokenize()
146 std::vector<Token> tokens = tokenizer.Tokenize("Hello world!"); in TEST()
177 EXPECT_THAT(tokenizer.Tokenize("앨라배마 주 전화(123) 456-789웹사이트"), in TEST()
311 tokens = tokenizer.Tokenize( in TEST()
315 tokens = tokenizer.Tokenize("問少目 hello 木輸ยามきゃ"); in TEST()
Dtokenizer.cc79 std::vector<Token> Tokenizer::Tokenize(const std::string& text) const { in Tokenize() function in libtextclassifier2::Tokenizer
81 return Tokenize(text_unicode); in Tokenize()
84 std::vector<Token> Tokenizer::Tokenize(const UnicodeText& text_unicode) const { in Tokenize() function in libtextclassifier2::Tokenizer
Dtokenizer.h42 std::vector<Token> Tokenize(const std::string& text) const;
45 std::vector<Token> Tokenize(const UnicodeText& text_unicode) const;
Dfeature-processor_test.cc313 std::vector<Token> tokens = feature_processor.Tokenize("one, two, three"); in TEST()
354 tokens = feature_processor3.Tokenize("zero, one, two, three, four"); in TEST()
388 std::vector<Token> tokens = feature_processor.Tokenize("one, two, three"); in TEST()
429 tokens = feature_processor3.Tokenize("zero, one, two, three, four"); in TEST()
564 {0, 3}, feature_processor.Tokenize("aaa bbb ccc")), in TEST()
567 {0, 3}, feature_processor.Tokenize("aaa bbb ěěě")), in TEST()
570 {0, 3}, feature_processor.Tokenize("ěěě řřř ěěě")), in TEST()
860 EXPECT_EQ(feature_processor.Tokenize("앨라배마123웹사이트"), in TEST()
870 EXPECT_EQ(feature_processor2.Tokenize("앨라배마123웹사이트"), in TEST()
883 std::vector<Token> tokens = feature_processor.Tokenize("พระบาทสมเด็จพระปรมิ"); in TEST()
[all …]
Dfeature-processor.h127 std::vector<Token> Tokenize(const std::string& text) const;
130 std::vector<Token> Tokenize(const UnicodeText& text_unicode) const;
Dfeature-processor.cc178 std::vector<Token> FeatureProcessor::Tokenize(const std::string& text) const { in Tokenize() function in libtextclassifier2::FeatureProcessor
180 return Tokenize(text_unicode); in Tokenize()
183 std::vector<Token> FeatureProcessor::Tokenize( in Tokenize() function in libtextclassifier2::FeatureProcessor
187 return tokenizer_.Tokenize(text_unicode); in Tokenize()
204 return tokenizer_.Tokenize(text_unicode); in Tokenize()
923 std::vector<Token> tokens = tokenizer_.Tokenize(text); in TokenizeSubstring()
Dtext-classifier.cc658 *tokens = selection_feature_processor_->Tokenize(context_unicode); in ModelSuggestSelection()
825 tokens = classification_feature_processor_->Tokenize(context); in ModelClassifyText()
1122 *tokens = selection_feature_processor_->Tokenize(line_str); in ModelAnnotate()
/external/google-breakpad/src/processor/
Dbasic_source_line_resolver.cc471 if (!Tokenize(file_line, kWhitespace, 2, &tokens)) { in ParseFile()
499 if (!Tokenize(function_line, kWhitespace, 4, &tokens)) { in ParseFunction()
531 if (!Tokenize(line_line, kWhitespace, 4, &tokens)) { in ParseLine()
580 if (!Tokenize(public_line, kWhitespace, 3, &tokens)) { in ParsePublicSymbol()
Dtokenize.h53 bool Tokenize(char *line,
Dtokenize.cc45 bool Tokenize(char *line, in Tokenize() function
Dwindows_frame_info.h126 if (!Tokenize(&buffer[0], " \r\n", 11, &tokens)) in ParseFromString()
/external/tensorflow/tensorflow/contrib/lite/testing/
Dtokenize.h37 void Tokenize(std::istream* input, TokenProcessor* processor);
Dmessage.cc91 Tokenize(input, &stack); in Read()
Dtokenize.cc23 void Tokenize(std::istream* input, TokenProcessor* processor) { in Tokenize() function
Dtokenize_test.cc39 Tokenize(&ss, &collector); in TokenizeString()
/external/chromium-trace/catapult/common/py_utils/py_utils/refactor/
Doffset_token.py55 def Tokenize(f): function
Dsnippet.py206 tokens = offset_token.Tokenize(f)
/external/google-breakpad/src/testing/gtest/scripts/
Dpump.py382 def Tokenize(s): function
579 tokens = list(Tokenize(pump_src_text))
/external/protobuf/gtest/scripts/
Dpump.py376 def Tokenize(s): function
571 for token in Tokenize(s):
/external/v8/testing/gtest/scripts/
Dpump.py382 def Tokenize(s): function
579 tokens = list(Tokenize(pump_src_text))
/external/vulkan-validation-layers/tests/gtest-1.7.0/scripts/
Dpump.py382 def Tokenize(s): function
579 tokens = list(Tokenize(pump_src_text))
/external/googletest/googletest/scripts/
Dpump.py382 def Tokenize(s): function
579 tokens = list(Tokenize(pump_src_text))