Home
last modified time | relevance | path

Searched refs:Tokenize (Results 1 – 24 of 24) sorted by relevance

/third_party/gn/src/gn/
Dtokenizer_unittest.cc25 std::vector<Token> results = Tokenizer::Tokenize(&input_file, &err); in CheckTokenizer()
45 std::vector<Token> results = Tokenizer::Tokenize(&empty_string_input, &err); in TEST()
51 results = Tokenizer::Tokenize(&whitespace_input, &err); in TEST()
127 std::vector<Token> results = Tokenizer::Tokenize(&input, &err); in TEST()
212 std::vector<Token> results = Tokenizer::Tokenize( in TEST()
224 Tokenizer::Tokenize(&input, &err, WhitespaceTransform::kInvalidToSpace); in TEST()
Dparser_unittest.cc17 *result = Tokenizer::Tokenize(input, &err); in GetTokens()
62 std::vector<Token> tokens = Tokenizer::Tokenize(&input_file, &err); in DoParserErrorTest()
80 std::vector<Token> tokens = Tokenizer::Tokenize(&input_file, &err); in DoExpressionErrorTest()
Dtokenizer.h29 static std::vector<Token> Tokenize(
Dtest_with_scope.cc75 std::vector<Token> tokens = Tokenizer::Tokenize(&input_file, err); in ExecuteExpression()
333 tokens_ = Tokenizer::Tokenize(&input_file_, &parse_err_); in TestParseInput()
Dinput_file_manager.cc87 *tokens = Tokenizer::Tokenize(file, err); in DoLoadFile()
Dtokenizer.cc82 std::vector<Token> Tokenizer::Tokenize( in Tokenize() function in Tokenizer
Dstring_utils.cc69 std::vector<Token> tokens = Tokenizer::Tokenize(&input_file, err); in AppendInterpolatedExpression()
Dsetup.cc564 args_tokens_ = Tokenizer::Tokenize(args_input_file_.get(), err); in FillArgsFromArgsInputFile()
829 dotfile_tokens_ = Tokenizer::Tokenize(dotfile_input_file_.get(), err); in RunConfigFile()
Dloader_unittest.cc132 canned->tokens = Tokenizer::Tokenize(canned->input_file.get(), &err); in AddCannedResponse()
Dinput_conversion.cc61 *tokens = Tokenizer::Tokenize(input_file, err); in ParseValueOrScope()
Dcommand_format.cc1256 Tokenizer::Tokenize(&file, &err, WhitespaceTransform::kInvalidToSpace); in FormatStringToString()
/third_party/flutter/skia/third_party/externals/wuffs/lang/generate/
Dgenerate.go106 tokens, _, err := t.Tokenize(tm, filename, src)
125 tokens, _, err := t.Tokenize(tm, filename, src)
/third_party/flutter/skia/third_party/externals/wuffs/lang/ast/
Dstring_test.go76 tokens, _, err := t.Tokenize(tm, filename, []byte(tc))
/third_party/flutter/skia/third_party/externals/wuffs/lang/check/
Dcheck_test.go99 tokens, comments, err := t.Tokenize(tm, filename, []byte(src))
215 tokens, _, err := t.Tokenize(tm, filename, []byte(src))
Dresolve.go128 tokens, _, err := t.Tokenize(c.tm, filename, buf)
Dcheck.go205 tokens, _, err := t.Tokenize(c.tm, filename, src)
/third_party/flutter/skia/third_party/externals/wuffs/cmd/wuffsfmt/
Dmain.go123 tokens, comments, err := t.Tokenize(tm, filename, src)
/third_party/flutter/skia/third_party/externals/wuffs/lang/token/
Dtoken.go116 func Tokenize(m *Map, filename string, src []byte) (tokens []Token, comments []string, retErr error… func
/third_party/python/Doc/library/
Dtokenize.rst62 Tokenize a source reading unicode strings instead of bytes.
/third_party/node/tools/
Dtest.py1084 def Tokenize(self): member in Tokenizer
1211 tokens = Tokenizer(expr).Tokenize()
Dlint-md.mjs2862 * Tokenize input until we reach end-of-string
38082 * Tokenize subcontent.
38198 * Tokenize embedded tokens.
/third_party/mindspore/patches/
D0022-support-cross-compile-with-ohos-ndk.patch340 @@ -141,7 +141,7 @@ std::vector<std::string> Tokenize(const std::string &src, const std::string &del
/third_party/python/Misc/NEWS.d/
D3.8.0a1.rst4077 Tokenize module now implicitly emits a NEWLINE when provided with input that
/third_party/chromium/patch/
D0004-ohos-3.2-Beta5.patch42679 +-std::vector<std::pair<const char*, int>> Tokenize(StringRef str) {
42814 +- list_of_tokens.emplace_back(Tokenize(GetString(input, i)));
43433 -std::vector<std::pair<const char*, int>> Tokenize(StringRef str) {
43568 - list_of_tokens.emplace_back(Tokenize(GetString(input, i)));