/system/tools/hidl/docs/src/parser/elements/declarations/ |
D | AbstractDeclarationParser.kt | 36 if (token.category != TokenCategory.Annotation && token.identifier != TokenGrammar.DOC_END) in scanTokens() 53 if (iter.hasPrevious() && peekPreviousToken(iter)!!.identifier == openDelimiter) { in scanDelimitedList() 57 if (token.identifier != openDelimiter) in scanDelimitedList() 65 if (token.identifier == closeDelimiter) { in scanDelimitedList() 68 } else if (token.identifier == delimiter) { in scanDelimitedList() 80 if (token.identifier == TokenGrammar.DOC_START) { in scanDelimitedList() 82 } else if (token.identifier == TokenGrammar.DOC_END) { in scanDelimitedList() 87 …if ((token.identifier == delimiter || token.identifier == closeDelimiter) && nestLevel == 0 && !in… in scanDelimitedList() 93 if (token.identifier == openDelimiter) { in scanDelimitedList() 95 } else if (token.identifier == closeDelimiter && nestLevel > 0) { in scanDelimitedList() [all …]
|
D | EnumDeclarationParser.kt | 41 assert(token.identifier == TokenGrammar.ENUM) in parseTokens() 42 assert(tokens.last().identifier == TokenGrammar.SEMICOLON) in parseTokens() 51 if (token.identifier != TokenGrammar.COLON) in parseTokens() 57 while (iter.hasNext() && peekToken(iter)!!.identifier != TokenGrammar.BRACE_OPEN) { in parseTokens() 71 if (statementTokens.first().identifier == TokenGrammar.DOC_START) { in parseTokens() 72 … val idx = statementTokens.indexOfFirst { it.identifier == TokenGrammar.DOC_END } in parseTokens() 99 if (tokens.any { it.identifier == TokenGrammar.EQUAL }) { in <lambda>() 100 this.value = tokens.takeLastWhile { it.identifier != TokenGrammar.EQUAL } in <lambda>()
|
D | CompoundDeclarationParser.kt | 42 assert(token.identifier == TokenGrammar.STRUCT || token.identifier == TokenGrammar.UNION) in parseTokens() 43 assert(tokens.last().identifier == TokenGrammar.SEMICOLON) in parseTokens() 46 this.type = token.identifier in parseTokens() 66 if (statementTokens.first().identifier == TokenGrammar.DOC_START) { in parseTokens() 67 … val idx = statementTokens.indexOfFirst { it.identifier == TokenGrammar.DOC_END } in parseTokens() 77 when(statementTokens.first().identifier) { in parseTokens() 81 typeDef = statementTokens.first().identifier, in parseTokens()
|
D | InterfaceDeclarationParser.kt | 37 assert(tokens.first().identifier == TokenGrammar.INTERFACE) in parseTokens() 38 assert(tokens.last().identifier == TokenGrammar.SEMICOLON) in parseTokens() 41 val sigToks = tokens.takeWhile { it.identifier != TokenGrammar.BRACE_OPEN } in parseTokens() 48 if (sigToks.any { it.identifier == TokenGrammar.EXTENDS }) { in parseTokens()
|
D | MethodDeclarationParser.kt | 41 assert(tokens.last().identifier == TokenGrammar.SEMICOLON) in parseTokens() 62 if (token.identifier == TokenGrammar.PAREN_OPEN) { in parseTokens() 69 if (token.identifier == TokenGrammar.GENERATES) { in parseTokens() 75 assert(token.identifier == TokenGrammar.SEMICOLON) in parseTokens()
|
D | TypedefDeclarationParser.kt | 33 assert(tokens.first().identifier == TokenGrammar.TYPEDEF) in parseTokens() 34 assert(tokens.last().identifier == TokenGrammar.SEMICOLON) in parseTokens()
|
/system/tools/hidl/docs/src/parser/elements/ |
D | AbstractParser.kt | 34 while (iter.hasNext() && peekToken(iter)?.identifier == TokenGrammar.EMPTY_LINE) { 56 while (peekPreviousToken(iter)?.identifier == TokenGrammar.EMPTY_LINE) iter.previous() in parseTokens() 59 if (peekPreviousToken(iter)?.identifier == TokenGrammar.DOC_START) iter.previous() in parseTokens() 61 if (peekToken(iter)!!.identifier != TokenGrammar.DOC_START) in parseTokens() 69 if (token.identifier == TokenGrammar.DOC_END) { in parseTokens() 71 } else if (token.identifier == TokenGrammar.DOC_START) { in parseTokens() 76 if (peekPreviousToken(iter)?.identifier != TokenGrammar.DOC_END) { in parseTokens() 95 if (token.identifier == TokenGrammar.DOC_START) { in scanDeclarationTokens() 97 } else if (token.identifier == TokenGrammar.DOC_END) { in scanDeclarationTokens() 103 } else if (token.identifier == TokenGrammar.BRACE_OPEN) { in scanDeclarationTokens() [all …]
|
D | DocAnnotationParser.kt | 44 when (it.identifier) { in formatValue() 62 if (peekToken(iter)?.identifier == TokenGrammar.AT) iter.next() in scanTokens() 73 token.identifier == TokenGrammar.EMPTY_LINE -> break@loop in scanTokens() 76 token.identifier == TokenGrammar.DOC_END || in scanTokens() 77 …token.identifier == TokenGrammar.AT && peekToken(iter)?.category == TokenCategory.DocAnnotation ->… in scanTokens() 97 this.tag = token.identifier in parseTokens() 101 if (token.identifier == TokenGrammar.PARAM || token.identifier == TokenGrammar.RETURN) { in parseTokens() 114 descTokens.map { if (it.identifier == TokenGrammar.EMPTY_LINE) "\n\n" else it.value } in parseTokens()
|
D | EntryParser.kt | 76 while (peekPreviousToken(iter)?.identifier == TokenGrammar.EMPTY_LINE) iter.previous() in scanTokens() 78 if (peekPreviousToken(iter)?.identifier == TokenGrammar.DOC_START) iter.previous() in scanTokens() 95 assert(peekToken(iter)!!.identifier == TokenGrammar.DOC_START) in parseTokens() 97 assert(peekPreviousToken(iter)!!.identifier == TokenGrammar.DOC_END) in parseTokens() 100 if (peekToken(iter)?.identifier != TokenGrammar.DOC_START) { in parseTokens() 115 while (iter.hasNext() && peekToken(iter)!!.identifier == TokenGrammar.AT) { in parseTokens() 130 assert(token.identifier != TokenGrammar.AT && token.category != TokenCategory.Annotation) in parseTokens() 136 if (token.identifier == TokenGrammar.DOC_START) { in parseTokens() 138 } else if (token.identifier == TokenGrammar.PACKAGE) { in parseTokens() 150 this.declarationParser = when (token.identifier) { in parseTokens()
|
D | DocParser.kt | 39 .filterNot { it.identifier == TokenGrammar.DOC_START } in formatDescription() 40 .filterNot { it.identifier == TokenGrammar.DOC_END } in formatDescription() 42 when (it.identifier) { in formatDescription() 60 assert(token.identifier == TokenGrammar.DOC_START) in parseTokens() 61 assert(tokens.last().identifier == TokenGrammar.DOC_END) in parseTokens() 67 …token.identifier == TokenGrammar.AT && peekToken(iter)?.category == TokenCategory.DocAnnotation ->… in parseTokens() 70 token.identifier == TokenGrammar.DOC_END -> break@loop in parseTokens()
|
D | AnnotationParser.kt | 39 if (peekToken(iter)?.identifier == TokenGrammar.AT) iter.next() in scanTokens() 55 this.name = token.identifier in parseTokens()
|
D | EntryCollectionParser.kt | 58 if (token.identifier == TokenGrammar.DOC_START) { in parseTokens()
|
/system/tools/hidl/docs/src/parser/files/ |
D | package.kt | 37 if (token.identifier == TokenGrammar.PACKAGE) { in parsePackageInfo() 42 … if (token.identifier != TokenGrammar.AT && token.identifier != TokenGrammar.SEMICOLON) { in parsePackageInfo() 51 if (token.identifier != TokenGrammar.AT) break in parsePackageInfo() 56 if (token.identifier != TokenGrammar.SEMICOLON) break in parsePackageInfo()
|
D | AbstractFileParser.kt | 59 private fun getEntriesByCompoundDeclarationParser(identifier: TokenGrammar): List<EntryParser> { in getEntriesByCompoundDeclarationParser() 61 .filter { (it.declarationParser as CompoundDeclarationParser).type == identifier } in getEntriesByCompoundDeclarationParser() 84 if (token.identifier == TokenGrammar.DOC_START) { in insertDocsForRequiredTypes() 87 } else if (token.identifier == TokenGrammar.DOC_END) { in insertDocsForRequiredTypes() 91 } else if (!inDoc && token.identifier in REQUIRED_DOC_TYPES) { in insertDocsForRequiredTypes() 93 if (peekToken(iter)?.identifier == TokenGrammar.CHEVRON_CLOSE) { in insertDocsForRequiredTypes() 124 if (token.identifier == TokenGrammar.AT || token.category == TokenCategory.Annotation) { in indexInsertionPointforDocTokens() 128 return if (token.identifier == TokenGrammar.DOC_END) -1 else idx+1 in indexInsertionPointforDocTokens()
|
D | InterfaceFileParser.kt | 64 if (token.identifier == TokenGrammar.DOC_START) { in isInterface() 67 } else if (token.identifier == TokenGrammar.DOC_END) { in isInterface() 71 } else if (!inDoc && token.identifier == TokenGrammar.INTERFACE) { in isInterface()
|
/system/tools/aidl/ |
D | aidl_language_y.yy | 71 %token<token> IDENTIFIER "identifier" 115 %type<token> identifier error 121 * occurring in the identifier position. Therefore identifier is a 125 identifier 148 : identifier { 152 | qualified_name '.' identifier 195 | PARCELABLE identifier '{' variable_decls '}' { 216 : type identifier ';' { 219 | type identifier '=' constant_value ';' { 228 : INTERFACE identifier '{' interface_members '}' { [all …]
|
D | aidl_language_l.ll | 21 identifier [_a-zA-Z][_a-zA-Z0-9]* 70 @{identifier} { yylval->token = new AidlToken(yytext + 1, extra_text); 96 {identifier} { yylval->token = new AidlToken(yytext, extra_text);
|
/system/tools/hidl/ |
D | hidl-gen_y.yy | 62 bool isValidInterfaceField(const std::string& identifier, std::string *errorMsg) { 90 if (std::find(reserved.begin(), reserved.end(), identifier) != reserved.end()) { 91 *errorMsg = identifier + " cannot be a name inside an interface"; 97 bool isValidStructField(const std::string& identifier, std::string *errorMsg) { 103 if (std::find(reserved.begin(), reserved.end(), identifier) != reserved.end()) { 104 *errorMsg = identifier + " cannot be a name inside an struct or union"; 110 bool isValidCompoundTypeField(CompoundType::Style style, const std::string& identifier, 116 return isValidStructField(identifier, errorMsg);; 119 bool isValidIdentifier(const std::string& identifier, std::string *errorMsg) { 154 if (std::find(keywords.begin(), keywords.end(), identifier) != keywords.end()) { [all …]
|
/system/core/demangle/ |
D | demangle.cpp | 89 std::string identifier(name, p); in Filter() local 90 std::string demangled_name = demangle(identifier.c_str()); in Filter() 93 if (compare) Compare(identifier.c_str(), demangled_name); in Filter()
|
/system/tools/hidl/docs/src/lexer/ |
D | Token.kt | 19 data class Token(val identifier: TokenGrammar, val value: String, val category: TokenCategory) constant in lexer.Token 166 fun newToken(value: String, identifier: TokenGrammar): Token { in newToken() 167 return Token(identifier, value, identifier.category) in newToken()
|
D | HidlLexer.kt | 96 … tokens.add(TokenGrammar.newToken(identifier = annotation, value = annotationArgs.toString())) in <lambda>()
|
/system/ca-certificates/ |
D | README.cacerts | 6 and the "n" is a unique integer identifier starting at 0 to deal
|
/system/bt/service/doc/ |
D | IBluetoothGattClient.txt | 46 * case of an error. |client_id| is the identifier obtained via 57 * an error. |client_id| is the identifier obtained via registerClient. 62 * Initiate a read request for the remote characteristic with identifier 77 * Initiate a write request for the remote characteristic with identifier 100 * Initiate a read request for the remote descriptor with identifier 115 * Initiate a write request for the remote descriptor with identifier
|
/system/hardware/interfaces/wifi/keystore/1.0/ |
D | IKeystore.hal | 51 * @param keyId the key identifier associated with the credential. 67 * @param keyId the key identifier associated with the credential.
|
/system/chre/external/flatbuffers/include/flatbuffers/ |
D | flatbuffers.h | 1449 inline bool BufferHasIdentifier(const void *buf, const char *identifier) { 1451 identifier, FlatBufferBuilder::kFileIdentifierLength) == 0; 1562 template<typename T> bool VerifyBufferFromStart(const char *identifier, 1564 if (identifier && 1566 !BufferHasIdentifier(start, identifier))) { 1581 template<typename T> bool VerifyBuffer(const char *identifier) { 1582 return VerifyBufferFromStart<T>(identifier, buf_); 1585 template<typename T> bool VerifySizePrefixedBuffer(const char *identifier) { 1588 VerifyBufferFromStart<T>(identifier, buf_ + sizeof(uoffset_t));
|