/third_party/gn/src/gn/ |
D | tokenizer_unittest.cc | 25 std::vector<Token> results = Tokenizer::Tokenize(&input_file, &err); in CheckTokenizer() 45 std::vector<Token> results = Tokenizer::Tokenize(&empty_string_input, &err); in TEST() 51 results = Tokenizer::Tokenize(&whitespace_input, &err); in TEST() 127 std::vector<Token> results = Tokenizer::Tokenize(&input, &err); in TEST() 212 std::vector<Token> results = Tokenizer::Tokenize( in TEST() 224 Tokenizer::Tokenize(&input, &err, WhitespaceTransform::kInvalidToSpace); in TEST()
|
D | parser_unittest.cc | 17 *result = Tokenizer::Tokenize(input, &err); in GetTokens() 62 std::vector<Token> tokens = Tokenizer::Tokenize(&input_file, &err); in DoParserErrorTest() 80 std::vector<Token> tokens = Tokenizer::Tokenize(&input_file, &err); in DoExpressionErrorTest()
|
D | tokenizer.h | 29 static std::vector<Token> Tokenize(
|
D | test_with_scope.cc | 75 std::vector<Token> tokens = Tokenizer::Tokenize(&input_file, err); in ExecuteExpression() 333 tokens_ = Tokenizer::Tokenize(&input_file_, &parse_err_); in TestParseInput()
|
D | input_file_manager.cc | 87 *tokens = Tokenizer::Tokenize(file, err); in DoLoadFile()
|
D | tokenizer.cc | 82 std::vector<Token> Tokenizer::Tokenize( in Tokenize() function in Tokenizer
|
D | string_utils.cc | 69 std::vector<Token> tokens = Tokenizer::Tokenize(&input_file, err); in AppendInterpolatedExpression()
|
D | setup.cc | 564 args_tokens_ = Tokenizer::Tokenize(args_input_file_.get(), err); in FillArgsFromArgsInputFile() 829 dotfile_tokens_ = Tokenizer::Tokenize(dotfile_input_file_.get(), err); in RunConfigFile()
|
D | loader_unittest.cc | 132 canned->tokens = Tokenizer::Tokenize(canned->input_file.get(), &err); in AddCannedResponse()
|
D | input_conversion.cc | 61 *tokens = Tokenizer::Tokenize(input_file, err); in ParseValueOrScope()
|
D | command_format.cc | 1256 Tokenizer::Tokenize(&file, &err, WhitespaceTransform::kInvalidToSpace); in FormatStringToString()
|
/third_party/flutter/skia/third_party/externals/wuffs/lang/generate/ |
D | generate.go | 106 tokens, _, err := t.Tokenize(tm, filename, src) 125 tokens, _, err := t.Tokenize(tm, filename, src)
|
/third_party/flutter/skia/third_party/externals/wuffs/lang/ast/ |
D | string_test.go | 76 tokens, _, err := t.Tokenize(tm, filename, []byte(tc))
|
/third_party/flutter/skia/third_party/externals/wuffs/lang/check/ |
D | check_test.go | 99 tokens, comments, err := t.Tokenize(tm, filename, []byte(src)) 215 tokens, _, err := t.Tokenize(tm, filename, []byte(src))
|
D | resolve.go | 128 tokens, _, err := t.Tokenize(c.tm, filename, buf)
|
D | check.go | 205 tokens, _, err := t.Tokenize(c.tm, filename, src)
|
/third_party/flutter/skia/third_party/externals/wuffs/cmd/wuffsfmt/ |
D | main.go | 123 tokens, comments, err := t.Tokenize(tm, filename, src)
|
/third_party/flutter/skia/third_party/externals/wuffs/lang/token/ |
D | token.go | 116 func Tokenize(m *Map, filename string, src []byte) (tokens []Token, comments []string, retErr error… func
|
/third_party/python/Doc/library/ |
D | tokenize.rst | 62 Tokenize a source reading unicode strings instead of bytes.
|
/third_party/node/tools/ |
D | test.py | 1084 def Tokenize(self): member in Tokenizer 1211 tokens = Tokenizer(expr).Tokenize()
|
D | lint-md.mjs | 2862 * Tokenize input until we reach end-of-string 38082 * Tokenize subcontent. 38198 * Tokenize embedded tokens.
|
/third_party/mindspore/patches/ |
D | 0022-support-cross-compile-with-ohos-ndk.patch | 340 @@ -141,7 +141,7 @@ std::vector<std::string> Tokenize(const std::string &src, const std::string &del
|
/third_party/python/Misc/NEWS.d/ |
D | 3.8.0a1.rst | 4077 Tokenize module now implicitly emits a NEWLINE when provided with input that
|
/third_party/chromium/patch/ |
D | 0004-ohos-3.2-Beta5.patch | 42679 +-std::vector<std::pair<const char*, int>> Tokenize(StringRef str) { 42814 +- list_of_tokens.emplace_back(Tokenize(GetString(input, i))); 43433 -std::vector<std::pair<const char*, int>> Tokenize(StringRef str) { 43568 - list_of_tokens.emplace_back(Tokenize(GetString(input, i)));
|