• Home
  • Raw
  • Download

Lines Matching refs:Tokenize

112   std::vector<Token> Tokenize(const UnicodeText& text) {  in Tokenize()  function in libtextclassifier3::__anon8aa146150111::DurationAnnotatorTest
113 return feature_processor_->Tokenize(text); in Tokenize()
152 std::vector<Token> tokens = Tokenize(text); in TEST_F()
171 std::vector<Token> tokens = Tokenize(text); in TEST_F()
190 std::vector<Token> tokens = Tokenize(text); in TEST_F()
209 std::vector<Token> tokens = Tokenize(text); in TEST_F()
228 std::vector<Token> tokens = Tokenize(text); in TEST_F()
247 std::vector<Token> tokens = Tokenize(text); in TEST_F()
266 std::vector<Token> tokens = Tokenize(text); in TEST_F()
286 std::vector<Token> tokens = Tokenize(text); in TEST_F()
305 std::vector<Token> tokens = Tokenize(text); in TEST_F()
323 std::vector<Token> tokens = Tokenize(text); in TEST_F()
334 std::vector<Token> tokens = Tokenize(text); in TEST_F()
352 std::vector<Token> tokens = Tokenize(text); in TEST_F()
383 std::vector<Token> tokens = Tokenize(text); in TEST_F()
402 std::vector<Token> tokens = Tokenize(text); in TEST_F()
422 std::vector<Token> tokens = Tokenize(text); in TEST_F()
440 std::vector<Token> tokens = Tokenize(text); in TEST_F()
493 std::vector<Token> Tokenize(const UnicodeText& text) { in Tokenize() function in libtextclassifier3::__anon8aa146150111::JapaneseDurationAnnotatorTest
494 return feature_processor_->Tokenize(text); in Tokenize()
504 std::vector<Token> tokens = Tokenize(text); in TEST_F()
522 std::vector<Token> tokens = Tokenize(text); in TEST_F()
540 std::vector<Token> tokens = Tokenize(text); in TEST_F()
550 std::vector<Token> tokens = Tokenize(text); in TEST_F()