Lines Matching refs:Tokenize
112 std::vector<Token> Tokenize(const UnicodeText& text) { in Tokenize() function in libtextclassifier3::__anoncdfecbf10111::DurationAnnotatorTest
113 return feature_processor_->Tokenize(text); in Tokenize()
152 std::vector<Token> tokens = Tokenize(text); in TEST_F()
171 std::vector<Token> tokens = Tokenize(text); in TEST_F()
190 std::vector<Token> tokens = Tokenize(text); in TEST_F()
209 std::vector<Token> tokens = Tokenize(text); in TEST_F()
228 std::vector<Token> tokens = Tokenize(text); in TEST_F()
247 std::vector<Token> tokens = Tokenize(text); in TEST_F()
266 std::vector<Token> tokens = Tokenize(text); in TEST_F()
286 std::vector<Token> tokens = Tokenize(text); in TEST_F()
305 std::vector<Token> tokens = Tokenize(text); in TEST_F()
323 std::vector<Token> tokens = Tokenize(text); in TEST_F()
334 std::vector<Token> tokens = Tokenize(text); in TEST_F()
352 std::vector<Token> tokens = Tokenize(text); in TEST_F()
383 std::vector<Token> tokens = Tokenize(text); in TEST_F()
402 std::vector<Token> tokens = Tokenize(text); in TEST_F()
422 std::vector<Token> tokens = Tokenize(text); in TEST_F()
440 std::vector<Token> tokens = Tokenize(text); in TEST_F()
458 std::vector<Token> tokens = Tokenize(text); in TEST_F()
475 std::vector<Token> tokens = Tokenize(text); in TEST_F()
529 std::vector<Token> Tokenize(const UnicodeText& text) { in Tokenize() function in libtextclassifier3::__anoncdfecbf10111::JapaneseDurationAnnotatorTest
530 return feature_processor_->Tokenize(text); in Tokenize()
540 std::vector<Token> tokens = Tokenize(text); in TEST_F()
558 std::vector<Token> tokens = Tokenize(text); in TEST_F()
576 std::vector<Token> tokens = Tokenize(text); in TEST_F()
586 std::vector<Token> tokens = Tokenize(text); in TEST_F()