/device/generic/opengl-transport/host/commands/emugen/ |
D | Parser.cpp | 95 static std::string buildTypeString(const std::vector<std::string>& tokens, in buildTypeString() argument 100 const std::string& token = tokens[n]; in buildTypeString() 111 std::vector<std::string> tokens; in normalizeTypeDeclaration() local 112 if (!parseTypeTokens(input, &tokens, NULL)) { in normalizeTypeDeclaration() 115 return buildTypeString(tokens, tokens.size()); in normalizeTypeDeclaration() 124 std::vector<std::string> tokens; in parseTypeDeclaration() local 126 if (!parseTypeTokens(input, &tokens, error)) { in parseTypeDeclaration() 132 for (size_t n = 0; n < tokens.size(); ++n) { in parseTypeDeclaration() 133 if (tokens[n] != "*" && tokens[n] != "const") { in parseTypeDeclaration() 142 *typeName = buildTypeString(tokens, tokens.size()); in parseTypeDeclaration() [all …]
|
/device/linaro/bootloader/edk2/AppPkg/Applications/Python/Python-2.7.2/Lib/lib2to3/pgen2/ |
D | driver.py | 38 def parse_tokens(self, tokens, debug=False): argument 47 for quintuple in tokens: 88 tokens = tokenize.generate_tokens(stream.readline) 89 return self.parse_tokens(tokens, debug) 105 tokens = tokenize.generate_tokens(StringIO.StringIO(text).readline) 106 return self.parse_tokens(tokens, debug)
|
D | pgen.py | 79 if itoken in c.tokens: 80 return c.tokens[itoken] 83 c.tokens[itoken] = ilabel 100 if itoken in c.tokens: 101 return c.tokens[itoken] 104 c.tokens[itoken] = ilabel
|
D | tokenize.py | 186 self.tokens = [] 195 self.tokens.append(" " * col_offset) 204 self.tokens.append(token) 209 return "".join(self.tokens) 214 toks_append = self.tokens.append
|
D | grammar.py | 83 self.tokens = {}
|
/device/linaro/bootloader/edk2/AppPkg/Applications/Python/Python-2.7.10/Lib/ |
D | token.py | 103 tokens = {} 109 tokens[val] = name # reverse so we can sort them... 110 keys = tokens.keys() 128 lines.append("%s = %d" % (tokens[val], val))
|
D | tokenize.py | 182 self.tokens = [] 193 self.tokens.append("\\\n" * row_offset) 197 self.tokens.append(" " * col_offset) 209 self.tokens.append(token) 214 return "".join(self.tokens) 218 toks_append = self.tokens.append
|
/device/linaro/bootloader/edk2/AppPkg/Applications/Python/Python-2.7.2/Lib/ |
D | token.py | 105 tokens = {} 111 tokens[val] = name # reverse so we can sort them... 112 keys = tokens.keys() 130 lines.append("%s = %d" % (tokens[val], val))
|
D | tokenize.py | 181 self.tokens = [] 190 self.tokens.append(" " * col_offset) 199 self.tokens.append(token) 204 return "".join(self.tokens) 209 toks_append = self.tokens.append
|
D | tabnanny.py | 273 def process_tokens(tokens): argument 281 for (type, token, start, end, line) in tokens:
|
/device/linaro/bootloader/edk2/BaseTools/Source/C/VfrCompile/Pccts/antlr/ |
D | makefile.cygwin | 186 PCCTS_GEN=antlr.c scan.c err.c tokens.h mode.h parser.dlg stdpccts.h remap.h 198 antlr.o : antlr.c mode.h tokens.h 200 scan.o : scan.c mode.h tokens.h
|
D | makefile | 186 PCCTS_GEN=antlr.c scan.c err.c tokens.h mode.h parser.dlg stdpccts.h remap.h 198 antlr.o : antlr.c mode.h tokens.h 200 scan.o : scan.c mode.h tokens.h
|
D | makefile1 | 63 PCCTS_GEN=antlr.c scan.c err.c tokens.h mode.h parser.dlg stdpccts.h remap.h 72 antlr.c stdpccts.h parser.dlg tokens.h err.c : antlr.g 75 antlr.o : antlr.c mode.h tokens.h 77 scan.o : scan.c mode.h tokens.h
|
D | AntlrDDK.mak | 45 $(ANTLR_SRC)\tokens.h \ 60 $(ANTLR_SRC)\tokens.h \ 74 $(ANTLR_SRC)\tokens.h \ 170 $(ANTLR_SRC)\tokens.h \
|
D | AntlrMS.mak | 46 $(ANTLR_SRC)\tokens.h \ 61 $(ANTLR_SRC)\tokens.h \ 75 $(ANTLR_SRC)\tokens.h \ 171 $(ANTLR_SRC)\tokens.h \
|
/device/linaro/bootloader/edk2/AppPkg/Applications/Python/Python-2.7.2/Parser/ |
D | spark.py | 251 def finalState(self, tokens): argument 255 if len(self.newrules[self._START]) == 2 and len(tokens) == 0: 301 def parse(self, tokens): argument 315 for i in xrange(len(tokens)): 320 self.makeSet(tokens[i], sets, i) 323 self.makeSet(None, sets, len(tokens)) 327 finalitem = (self.finalState(tokens), 0) 329 if len(tokens) > 0: 330 self.error(tokens[i-1]) 335 tokens, len(sets)-2) [all …]
|
/device/linaro/bootloader/edk2/BaseTools/Source/C/VfrCompile/Pccts/dlg/ |
D | makefile1 | 40 dlg_p.c parser.dlg err.c tokens.h : dlg_p.g 46 dlg_p.$(OBJ_EXT) : dlg_p.c dlg.h tokens.h mode.h 49 dlg_a.$(OBJ_EXT) : dlg_a.c dlg.h tokens.h mode.h
|
D | DlgDDK.mak | 42 $(DLG_SRC)\tokens.h \ 54 $(DLG_SRC)\tokens.h \ 66 $(DLG_SRC)\tokens.h \ 77 $(DLG_SRC)\tokens.h \
|
D | DlgMS.mak | 41 $(DLG_SRC)\tokens.h \ 53 $(DLG_SRC)\tokens.h \ 65 $(DLG_SRC)\tokens.h \ 76 $(DLG_SRC)\tokens.h \
|
D | makefile.cygwin | 140 dlg_p.$(OBJ_EXT) : dlg_p.c dlg.h tokens.h mode.h 143 dlg_a.$(OBJ_EXT) : dlg_a.c dlg.h tokens.h mode.h
|
D | makefile | 140 dlg_p.$(OBJ_EXT) : dlg_p.c dlg.h tokens.h mode.h 143 dlg_a.$(OBJ_EXT) : dlg_a.c dlg.h tokens.h mode.h
|
/device/linaro/bootloader/edk2/AppPkg/Applications/Python/Python-2.7.2/Lib/lib2to3/ |
D | patcomp.py | 36 tokens = tokenize.generate_tokens(StringIO.StringIO(input).readline) 37 for quintuple in tokens: 58 tokens = tokenize_wrapper(input) 60 root = self.driver.parse_tokens(tokens, debug=debug)
|
D | btm_utils.py | 9 tokens = grammar.opmap variable 206 if name in tokens: 207 new_node = MinNode(type=tokens[name])
|
/device/linaro/bootloader/edk2/EdkCompatibilityPkg/Other/Maintained/Tools/Pccts/ |
D | KNOWN_PROBLEMS.txt | 120 When antlr uses #tokdefs to define tokens the fields of 123 the set of tokens (using ~Token or .) it can refer to 124 tokens which don't have names, generating a fatal error.
|
/device/linaro/bootloader/edk2/BaseTools/Source/C/VfrCompile/Pccts/ |
D | NOTES.bcc | 100 #antlr.c stdpccts.h parser.dlg tokens.h err.c : antlr.g 103 antlr.$(OBJ_EXT): antlr.c mode.h tokens.h 105 scan.$(OBJ_EXT): scan.c mode.h tokens.h
|