Lines Matching refs:Tokenize
94 std::vector<Token> Tokenize(const std::string& utf8_text) const { in Tokenize() function in libtextclassifier3::__anon5b5dd84c0111::TestingTokenizerProxy
95 return tokenizer_->Tokenize(utf8_text); in Tokenize()
178 std::vector<Token> tokens = tokenizer.Tokenize("Hello world!"); in TEST()
213 EXPECT_THAT(tokenizer.Tokenize("앨라배마 주 전화(123) 456-789웹사이트"), in TEST()
351 tokens = tokenizer.Tokenize( in TEST()
355 tokens = tokenizer.Tokenize("問少目 hello 木輸ยามきゃ"); in TEST()
379 std::vector<Token> tokens = tokenizer.Tokenize("พระบาท สมเด็จ พระ ปร มิ"); in TEST()
400 tokenizer.Tokenize("The interval is: -(12, 138*)"); in TEST()
427 std::vector<Token> tokens = tokenizer.Tokenize("3.1 3﹒2 3.3"); in TEST()
445 std::vector<Token> tokens = tokenizer.Tokenize("พระบาทสมเด็จพระปรมิ"); in TEST()
495 std::vector<Token> tokens = tokenizer.Tokenize( in TEST()
526 EXPECT_EQ(tokenizer.Tokenize("앨라배마123웹사이트"), in TEST()
536 EXPECT_EQ(tokenizer.Tokenize("앨라배마123웹사이트"), in TEST()
548 std::vector<Token> tokens = tokenizer.Tokenize("7% -3.14 68.9#? 7% $99 .18."); in TEST()
565 std::vector<Token> tokens = tokenizer.Tokenize("2 pércént 3パーセント"); in TEST()
577 std::vector<Token> tokens = tokenizer.Tokenize("3 3﹒2 3.3%"); in TEST()
589 std::vector<Token> tokens = tokenizer.Tokenize("15.12.2019 january's 3.2"); in TEST()
604 std::vector<Token> tokens = tokenizer.Tokenize("The+2345++the +íí+"); in TEST()
618 std::vector<Token> tokens = tokenizer.Tokenize("2 3 4 5"); in TEST()