/external/mesa3d/src/freedreno/ir3/ |
D | ir3_lexer.l | 31 #define TOKEN(t) (ir3_yylval.tok = t) macro 76 "@localsize" return TOKEN(T_A_LOCALSIZE); 77 "@const" return TOKEN(T_A_CONST); 78 "@buf" return TOKEN(T_A_BUF); 79 "@invocationid" return TOKEN(T_A_INVOCATIONID); 80 "@wgid" return TOKEN(T_A_WGID); 81 "@numwg" return TOKEN(T_A_NUMWG); 82 "@in" return TOKEN(T_A_IN); 83 "@out" return TOKEN(T_A_OUT); 84 "@tex" return TOKEN(T_A_TEX); [all …]
|
D | ir3_parser.y | 242 %token <num> T_INT 243 %token <unum> T_HEX 244 %token <flt> T_FLOAT 245 %token <str> T_IDENTIFIER 246 %token <num> T_REGISTER 247 %token <num> T_CONSTANT 250 %token <tok> T_A_LOCALSIZE 251 %token <tok> T_A_CONST 252 %token <tok> T_A_BUF 253 %token <tok> T_A_INVOCATIONID [all …]
|
/external/swiftshader/src/OpenGL/compiler/preprocessor/ |
D | DirectiveParser.cpp | 26 #include "Token.h" 49 static DirectiveType getDirective(const pp::Token *token) in getDirective() argument 65 if (token->type != pp::Token::IDENTIFIER) in getDirective() 68 if (token->text == kDirectiveDefine) in getDirective() 70 else if (token->text == kDirectiveUndef) in getDirective() 72 else if (token->text == kDirectiveIf) in getDirective() 74 else if (token->text == kDirectiveIfdef) in getDirective() 76 else if (token->text == kDirectiveIfndef) in getDirective() 78 else if (token->text == kDirectiveElse) in getDirective() 80 else if (token->text == kDirectiveElif) in getDirective() [all …]
|
/external/angle/src/compiler/preprocessor/ |
D | DirectiveParser.cpp | 19 #include "compiler/preprocessor/Token.h" 45 DirectiveType getDirective(const pp::Token *token) in getDirective() argument 61 if (token->type != pp::Token::IDENTIFIER) in getDirective() 64 if (token->text == kDirectiveDefine) in getDirective() 66 if (token->text == kDirectiveUndef) in getDirective() 68 if (token->text == kDirectiveIf) in getDirective() 70 if (token->text == kDirectiveIfdef) in getDirective() 72 if (token->text == kDirectiveIfndef) in getDirective() 74 if (token->text == kDirectiveElse) in getDirective() 76 if (token->text == kDirectiveElif) in getDirective() [all …]
|
/external/antlr/runtime/Ruby/test/functional/lexer/ |
D | basic.rb | 20 token = lexer.next_token 21 token.name.should == 'ZERO' 23 token = lexer.next_token 24 token.name.should == '<EOF>' 30 token_types = lexer.map { |token| token.name } 34 example "mismatched token" do 38 token = lexer.next_token 62 token = lexer.next_token 63 token.name.should == 'ZERO' 65 token = lexer.next_token [all …]
|
/external/deqp-deps/amber/src/amberscript/ |
D | parser.cc | 207 for (auto token = tokenizer->NextToken();; token = tokenizer->NextToken()) { in ParseBufferData() local 208 if (token->IsEOL()) in ParseBufferData() 210 if (token->IsEOS()) { in ParseBufferData() 217 if (token->IsIdentifier() && token->AsString() == "END") in ParseBufferData() 219 if (!token->IsInteger() && !token->IsDouble() && !token->IsHex()) in ParseBufferData() 220 return Result("invalid BUFFER data value: " + token->ToOriginalString()); in ParseBufferData() 230 token->ConvertToDouble(); in ParseBufferData() 232 double val = token->IsHex() ? static_cast<double>(token->AsHex()) in ParseBufferData() 233 : token->AsDouble(); in ParseBufferData() 237 if (token->IsDouble()) { in ParseBufferData() [all …]
|
/external/angle/src/tests/preprocessor_tests/ |
D | token_test.cpp | 9 #include "compiler/preprocessor/Token.h" 16 pp::Token token; in TEST() local 17 EXPECT_EQ(0, token.type); in TEST() 18 EXPECT_EQ(0u, token.flags); in TEST() 19 EXPECT_EQ(0, token.location.line); in TEST() 20 EXPECT_EQ(0, token.location.file); in TEST() 21 EXPECT_EQ("", token.text); in TEST() 26 pp::Token token; in TEST() local 27 token.type = 1; in TEST() 28 token.flags = 1; in TEST() [all …]
|
/external/antlr/runtime/ObjC/Framework/test/runtime/token/ |
D | CommonTokenTest.m | 17 CommonToken *token = [[CommonToken newToken] retain]; 18 STAssertNotNil(token, @"Token was nil"); 19 [token release]; 24 CommonToken *token = [[CommonToken eofToken] retain]; 25 STAssertNotNil(token, @"Token was nil"); 26 STAssertEquals(token.type, (NSInteger)TokenTypeEOF, @"Token was not of type TokenTypeEOF"); 27 [token release]; 32 CommonToken *token = [[CommonToken newToken:TokenTypeUP] retain]; 33 token.text = @"<UP>"; 34 STAssertNotNil(token, @"Token was nil"); [all …]
|
/external/antlr/runtime/C/src/ |
D | antlr3commontoken.c | 2 * Contains the default implementation of the common token used within 38 /* Token API 40 static pANTLR3_STRING getText (pANTLR3_COMMON_TOKEN token); 41 static void setText (pANTLR3_COMMON_TOKEN token, pANTLR3_STRING text); 42 static void setText8 (pANTLR3_COMMON_TOKEN token, pANTLR3_UINT8 text); 43 static ANTLR3_UINT32 getType (pANTLR3_COMMON_TOKEN token); 44 static void setType (pANTLR3_COMMON_TOKEN token, ANTLR3_UINT32 type); 45 static ANTLR3_UINT32 getLine (pANTLR3_COMMON_TOKEN token); 46 static void setLine (pANTLR3_COMMON_TOKEN token, ANTLR3_UINT32 line); 47 static ANTLR3_INT32 getCharPositionInLine (pANTLR3_COMMON_TOKEN token); [all …]
|
/external/icing/icing/tokenization/ |
D | raw-query-tokenizer_test.cc | 63 IsOkAndHolds(ElementsAre(EqualsToken(Token::REGULAR, "Hello"), in TEST_F() 64 EqualsToken(Token::REGULAR, "World")))); in TEST_F() 67 IsOkAndHolds(ElementsAre(EqualsToken(Token::REGULAR, "hElLo"), in TEST_F() 68 EqualsToken(Token::REGULAR, "WORLD")))); in TEST_F() 83 EqualsToken(Token::QUERY_LEFT_PARENTHESES, ""), in TEST_F() 84 EqualsToken(Token::QUERY_RIGHT_PARENTHESES, "")))); in TEST_F() 88 EqualsToken(Token::QUERY_LEFT_PARENTHESES, ""), in TEST_F() 89 EqualsToken(Token::QUERY_RIGHT_PARENTHESES, "")))); in TEST_F() 93 EqualsToken(Token::QUERY_LEFT_PARENTHESES, ""), in TEST_F() 94 EqualsToken(Token::REGULAR, "term1"), in TEST_F() [all …]
|
/external/libxml2/doc/ |
D | apibuild.py | 406 def push(self, token): argument 407 self.tokens.insert(0, token); 410 print("Last token: ", self.last) 411 print("Token queue: ", self.tokens) 414 def token(self): member in CLexer 633 def error(self, msg, token=-1): argument 638 if token != -1: 639 print("Got token ", token) 643 def debug(self, msg, token=-1): argument 645 if token != -1: [all …]
|
/external/pdfium/xfa/fxfa/fm2js/ |
D | cxfa_fmlexer_unittest.cpp | 15 CXFA_FMToken token = lexer.NextToken(); in TEST() local 16 EXPECT_EQ(TOKeof, token.m_type); in TEST() 22 CXFA_FMToken token = lexer.NextToken(); in TEST() local 23 EXPECT_EQ(TOKeof, token.m_type); in TEST() 29 CXFA_FMToken token = lexer->NextToken(); in TEST() local 31 EXPECT_EQ(TOKminus, token.m_type); in TEST() 32 token = lexer->NextToken(); in TEST() 33 EXPECT_EQ(L"12", token.m_string); in TEST() 34 token = lexer->NextToken(); in TEST() 35 EXPECT_EQ(TOKeof, token.m_type); in TEST() [all …]
|
/external/deqp-deps/amber/src/vkscript/ |
D | command_parser.cc | 87 for (auto token = tokenizer_->NextToken(); !token->IsEOS(); in Parse() local 88 token = tokenizer_->NextToken()) { in Parse() 89 if (token->IsEOL()) in Parse() 92 if (!token->IsIdentifier()) { in Parse() 95 token->ToOriginalString())); in Parse() 98 std::string cmd_name = token->AsString(); in Parse() 101 token = tokenizer_->NextToken(); in Parse() 102 if (!token->IsIdentifier()) in Parse() 104 token->ToOriginalString())); in Parse() 106 cmd_name = token->AsString(); in Parse() [all …]
|
/external/doclava/src/com/google/doclava/apicheck/ |
D | ApiFile.java | 74 String token = tokenizer.getToken(); in parseApi() local 75 if (token == null) { in parseApi() 78 if ("package".equals(token)) { in parseApi() 81 throw new ApiParseException("expected package got " + token, tokenizer.getLine()); in parseApi() 93 String token; in parsePackage() local 97 token = tokenizer.requireToken(); in parsePackage() 98 assertIdent(tokenizer, token); in parsePackage() 99 name = token; in parsePackage() 101 token = tokenizer.requireToken(); in parsePackage() 102 if (!"{".equals(token)) { in parsePackage() [all …]
|
/external/snakeyaml/src/main/java/org/yaml/snakeyaml/parser/ |
D | ParserImpl.java | 50 import org.yaml.snakeyaml.tokens.Token; 183 StreamStartToken token = (StreamStartToken) scanner.getToken(); in produce() local 184 Event event = new StreamStartEvent(token.getStartMark(), token.getEndMark()); in produce() 194 … if (!scanner.checkToken(Token.ID.Directive, Token.ID.DocumentStart, Token.ID.StreamEnd)) { in produce() 196 Token token = scanner.peekToken(); in produce() local 197 Mark startMark = token.getStartMark(); in produce() 214 while (scanner.checkToken(Token.ID.DocumentEnd)) { in produce() 219 if (!scanner.checkToken(Token.ID.StreamEnd)) { in produce() 220 Token token = scanner.peekToken(); in produce() local 221 Mark startMark = token.getStartMark(); in produce() [all …]
|
/external/llvm-project/mlir/test/Dialect/Async/ |
D | async-ref-counting.mlir | 7 func @token_arg_no_uses(%arg0: !async.token) { 13 func @token_arg_conditional_await(%arg0: !async.token, %arg1: i1) { 21 async.await %arg0 : !async.token 27 // CHECK: %[[TOKEN:.*]] = async.execute 28 // CHECK: async.drop_ref %[[TOKEN]] {count = 1 : i32} 29 %token = async.execute { 36 func @token_return() -> !async.token { 37 // CHECK: %[[TOKEN:.*]] = async.execute 38 %token = async.execute { 41 // CHECK: return %[[TOKEN]] [all …]
|
/external/antlr/runtime/Python/tests/ |
D | t010lexer.py | 27 token = lexer.nextToken() 28 assert token.type == self.lexerModule.IDENTIFIER 29 assert token.start == 0, token.start 30 assert token.stop == 5, token.stop 31 assert token.text == 'foobar', token.text 33 token = lexer.nextToken() 34 assert token.type == self.lexerModule.WS 35 assert token.start == 6, token.start 36 assert token.stop == 6, token.stop 37 assert token.text == ' ', token.text [all …]
|
D | t011lexer.py | 27 token = lexer.nextToken() 28 assert token.type == self.lexerModule.IDENTIFIER 29 assert token.start == 0, token.start 30 assert token.stop == 5, token.stop 31 assert token.text == 'foobar', token.text 33 token = lexer.nextToken() 34 assert token.type == self.lexerModule.WS 35 assert token.start == 6, token.start 36 assert token.stop == 6, token.stop 37 assert token.text == ' ', token.text [all …]
|
/external/rust/crates/serde_test/src/ |
D | de.rs | 8 use token::Token; 12 tokens: &'de [Token], 18 Some(token) if token == $expected => {} 20 "expected Token::{} but deserialization wants Token::{}", 24 "end of tokens but deserialization wants Token::{}", 32 ($token:expr) => { 33 panic!("deserialization did not expect this token: {}", $token) 44 pub fn new(tokens: &'de [Token]) -> Self { in new() 48 fn peek_token_opt(&self) -> Option<Token> { in peek_token_opt() argument 52 fn peek_token(&self) -> Token { in peek_token() argument [all …]
|
/external/python/cpython3/Doc/library/ |
D | token-list.inc | 18 Token value for ``"("``. 22 Token value for ``")"``. 26 Token value for ``"["``. 30 Token value for ``"]"``. 34 Token value for ``":"``. 38 Token value for ``","``. 42 Token value for ``";"``. 46 Token value for ``"+"``. 50 Token value for ``"-"``. 54 Token value for ``"*"``. [all …]
|
/external/libxml2/ |
D | rngparser.c | 59 const xmlChar *token; member 92 tokenPtr token; member 244 * Scan the schema to get the next token 252 tokenPtr token; in xmlCRNGNextToken() local 256 token = &(ctxt->tokens[(ctxt->firstToken + ctxt->nbTokens) % MAX_TOKEN]); in xmlCRNGNextToken() 257 token->toktype = CRNG_NONE; in xmlCRNGNextToken() 291 token->toklen = cur - ctxt->cur; in xmlCRNGNextToken() 292 token->token = xmlDictLookup(ctxt->dict, ctxt->cur, token->toklen); in xmlCRNGNextToken() 293 token->toktype = CRNG_LITERAL_SEGMENT; in xmlCRNGNextToken() 294 token->prefix = NULL; in xmlCRNGNextToken() [all …]
|
/external/turbine/java/com/google/turbine/parse/ |
D | StreamLexer.java | 34 /** The start position of the current token. */ 37 /** The start position of the current numeric literal or identifier token. */ 40 /** The value of the current string or character literal token. */ 97 public Token next() { in next() 124 return Token.EOF; in next() 180 return Token.DIVEQ; in next() 182 return Token.DIV; in next() 246 return Token.EOF; in next() 265 return Token.LPAREN; in next() 268 return Token.RPAREN; in next() [all …]
|
/external/libtextclassifier/native/utils/ |
D | tokenizer_test.cc | 94 std::vector<Token> Tokenize(const std::string& utf8_text) const { in Tokenize() 178 std::vector<Token> tokens = tokenizer.Tokenize("Hello world!"); in TEST() 181 ElementsAreArray({Token("Hello", 0, 5), Token("world!", 6, 12)})); in TEST() 214 std::vector<Token>({Token("앨라배마", 0, 4), Token("주", 5, 6), in TEST() 215 Token("전화", 7, 10), Token("(123)", 10, 15), in TEST() 216 Token("456-789", 16, 23), in TEST() 217 Token("웹사이트", 23, 28)})); in TEST() 349 std::vector<Token> tokens; in TEST() 359 ElementsAreArray({Token("問", 0, 1), in TEST() 360 Token("少", 1, 2), in TEST() [all …]
|
/external/libtextclassifier/native/annotator/ |
D | feature-processor_test.cc | 87 std::vector<Token> tokens{Token("Hělló", 0, 5), in TEST_F() 88 Token("fěěbař@google.com", 6, 23), in TEST_F() 89 Token("heře!", 24, 29)}; in TEST_F() 95 {Token("Hělló", 0, 5), in TEST_F() 96 Token("fěě", 6, 9), in TEST_F() 97 Token("bař", 9, 12), in TEST_F() 98 Token("@google.com", 12, 23), in TEST_F() 99 Token("heře!", 24, 29)})); in TEST_F() 104 std::vector<Token> tokens{Token("Hělló", 0, 5), in TEST_F() 105 Token("fěěbař@google.com", 6, 23), in TEST_F() [all …]
|
/external/ImageMagick/MagickWand/tests/ |
D | script-token-test-results.txt | 1 l=8, c=1, stat=0, len=64, token="-option" 2 l=8, c=9, stat=0, len=64, token="key" 3 l=9, c=1, stat=0, len=64, token="+reset" 4 l=9, c=10, stat=0, len=64, token="imbedded#hash" 5 l=11, c=1, stat=0, len=64, token="This is a single token" 6 l=13, c=1, stat=0, len=64, token="And\ 'even '"more "complex" 7 l=15, c=1, stat=0, len=64, token="Backslash chars \n are returned as is" 8 l=16, c=1, stat=0, len=64, token="regardless \n of quoting" 9 l=18, c=1, stat=0, len=64, token="Single quote escapes" 10 l=19, c=2, stat=0, len=64, token="'" [all …]
|