/external/antlr/runtime/C/include/ |
D | antlr3convertutf.h | 106 typedef ANTLR3_UINT32 UTF32; /* at least 32 bits */ typedef 111 #define UNI_REPLACEMENT_CHAR (UTF32)0x0000FFFD 112 #define UNI_MAX_BMP (UTF32)0x0000FFFF 113 #define UNI_MAX_UTF16 (UTF32)0x0010FFFF 114 #define UNI_MAX_UTF32 (UTF32)0x7FFFFFFF 115 #define UNI_MAX_LEGAL_UTF32 (UTF32)0x0010FFFF 117 #define UNI_SUR_HIGH_START (UTF32)0xD800 118 #define UNI_SUR_HIGH_END (UTF32)0xDBFF 119 #define UNI_SUR_LOW_START (UTF32)0xDC00 120 #define UNI_SUR_LOW_END (UTF32)0xDFFF [all …]
|
/external/unicode/ |
D | ConvertUTF.c | 49 static const UTF32 halfBase = 0x0010000UL; 50 static const UTF32 halfMask = 0x3FFUL; 52 #define UNI_SUR_HIGH_START (UTF32)0xD800 53 #define UNI_SUR_HIGH_END (UTF32)0xDBFF 54 #define UNI_SUR_LOW_START (UTF32)0xDC00 55 #define UNI_SUR_LOW_END (UTF32)0xDFFF 62 const UTF32** sourceStart, const UTF32* sourceEnd, in ConvertUTF32toUTF16() 65 const UTF32* source = *sourceStart; in ConvertUTF32toUTF16() 68 UTF32 ch; in ConvertUTF32toUTF16() 112 UTF32** targetStart, UTF32* targetEnd, ConversionFlags flags) { in ConvertUTF16toUTF32() [all …]
|
D | ConvertUTF.h | 90 typedef unsigned long UTF32; /* at least 32 bits */ typedef 96 #define UNI_REPLACEMENT_CHAR (UTF32)0x0000FFFD 97 #define UNI_MAX_BMP (UTF32)0x0000FFFF 98 #define UNI_MAX_UTF16 (UTF32)0x0010FFFF 99 #define UNI_MAX_UTF32 (UTF32)0x7FFFFFFF 100 #define UNI_MAX_LEGAL_UTF32 (UTF32)0x0010FFFF 129 UTF32** targetStart, UTF32* targetEnd, ConversionFlags flags); 132 const UTF32** sourceStart, const UTF32* sourceEnd, 137 UTF32** targetStart, UTF32* targetEnd, ConversionFlags flags); 140 const UTF32** sourceStart, const UTF32* sourceEnd,
|
D | harness.c | 151 UTF32 utf32_buf[2], utf32_result[2]; in test02() 154 UTF32 *utf32SourceStart, *utf32TargetStart; in test02() 174 …result = ConvertUTF32toUTF16((const UTF32 **) &utf32SourceStart, &(utf32_buf[1]), &utf16TargetStar… in test02() 313 UTF32 utf32_buf[2], utf32_result[2]; in test03() 315 UTF32 *utf32SourceStart, *utf32TargetStart; in test03() 334 …result = ConvertUTF32toUTF8((const UTF32 **) &utf32SourceStart, &(utf32_buf[1]), & utf8TargetStart… in test03() 403 UTF32 utf32_buf[2]; in test04() 405 UTF32 *utf32SourceStart, *utf32TargetStart; in test04() 421 …result = ConvertUTF32toUTF8((const UTF32 **) &utf32SourceStart, &(utf32_buf[1]), & utf8TargetStart… in test04()
|
/external/google-breakpad/src/common/ |
D | convert_UTF.c | 65 static const UTF32 halfBase = 0x0010000UL; 66 static const UTF32 halfMask = 0x3FFUL; 68 #define UNI_SUR_HIGH_START (UTF32)0xD800 69 #define UNI_SUR_HIGH_END (UTF32)0xDBFF 70 #define UNI_SUR_LOW_START (UTF32)0xDC00 71 #define UNI_SUR_LOW_END (UTF32)0xDFFF 82 ConversionResult ConvertUTF32toUTF16 (const UTF32** sourceStart, const UTF32* sourceEnd, in ConvertUTF32toUTF16() 85 const UTF32* source = *sourceStart; in ConvertUTF32toUTF16() 88 UTF32 ch; in ConvertUTF32toUTF16() 131 … UTF32** targetStart, UTF32* targetEnd, ConversionFlags flags) { in ConvertUTF16toUTF32() [all …]
|
D | convert_UTF.h | 109 typedef unsigned long UTF32; /* at least 32 bits */ typedef 115 #define UNI_REPLACEMENT_CHAR (UTF32)0x0000FFFD 116 #define UNI_MAX_BMP (UTF32)0x0000FFFF 117 #define UNI_MAX_UTF16 (UTF32)0x0010FFFF 118 #define UNI_MAX_UTF32 (UTF32)0x7FFFFFFF 119 #define UNI_MAX_LEGAL_UTF32 (UTF32)0x0010FFFF 145 UTF32** targetStart, UTF32* targetEnd, ConversionFlags flags); 147 ConversionResult ConvertUTF32toUTF8 (const UTF32** sourceStart, const UTF32* sourceEnd, 151 UTF32** targetStart, UTF32* targetEnd, ConversionFlags flags); 153 ConversionResult ConvertUTF32toUTF16 (const UTF32** sourceStart, const UTF32* sourceEnd,
|
D | string_conversion.cc | 87 const UTF32 *source_ptr = reinterpret_cast<const UTF32 *>(in); in UTF32ToUTF16() 88 const UTF32 *source_end_ptr = source_ptr + source_length; in UTF32ToUTF16() 103 const UTF32 *source_ptr = reinterpret_cast<const UTF32 *>(&in); in UTF32ToUTF16Char() 104 const UTF32 *source_end_ptr = source_ptr + 1; in UTF32ToUTF16Char()
|
/external/antlr/runtime/Cpp/include/ |
D | antlr3convertutf.hpp | 106 typedef ANTLR_UINT32 UTF32; /* at least 32 bits */ typedef 111 #define UNI_REPLACEMENT_CHAR (UTF32)0x0000FFFD 112 #define UNI_MAX_BMP (UTF32)0x0000FFFF 113 #define UNI_MAX_UTF16 (UTF32)0x0010FFFF 114 #define UNI_MAX_UTF32 (UTF32)0x7FFFFFFF 115 #define UNI_MAX_LEGAL_UTF32 (UTF32)0x0010FFFF 117 #define UNI_SUR_HIGH_START (UTF32)0xD800 118 #define UNI_SUR_HIGH_END (UTF32)0xDBFF 119 #define UNI_SUR_LOW_START (UTF32)0xDC00 120 #define UNI_SUR_LOW_END (UTF32)0xDFFF [all …]
|
D | antlr3intstream.inl | 443 UTF32 ch; 444 UTF32 ch2; 557 // Construct the UTF32 code point 578 UTF32 ch; 579 UTF32 ch2; 695 // Construct the UTF32 code point 716 UTF32 ch; 717 UTF32 ch2; 833 // Construct the UTF32 code point 854 UTF32 ch; [all …]
|
/external/llvm/lib/Support/ |
D | ConvertUTF.c | 58 static const UTF32 halfBase = 0x0010000UL; 59 static const UTF32 halfMask = 0x3FFUL; 61 #define UNI_SUR_HIGH_START (UTF32)0xD800 62 #define UNI_SUR_HIGH_END (UTF32)0xDBFF 63 #define UNI_SUR_LOW_START (UTF32)0xDC00 64 #define UNI_SUR_LOW_END (UTF32)0xDFFF 93 static const UTF32 offsetsFromUTF8[6] = { 0x00000000UL, 0x00003080UL, 0x000E2080UL, 119 const UTF32** sourceStart, const UTF32* sourceEnd, in ConvertUTF32toUTF16() 122 const UTF32* source = *sourceStart; in ConvertUTF32toUTF16() 125 UTF32 ch; in ConvertUTF32toUTF16() [all …]
|
D | ConvertUTFWrapper.cpp | 51 UTF32 *targetStart = reinterpret_cast<UTF32*>(ResultPtr); in ConvertUTF8toWide() 67 const UTF32 *SourceStart = &Source; in ConvertCodePointToUTF8() 68 const UTF32 *SourceEnd = SourceStart + 1; in ConvertCodePointToUTF8() 230 const UTF32 *Start = reinterpret_cast<const UTF32 *>(Source.data()); in convertWideToUTF8() 231 const UTF32 *End = in convertWideToUTF8() 232 reinterpret_cast<const UTF32 *>(Source.data() + Source.size()); in convertWideToUTF8()
|
D | Unicode.cpp | 350 UTF32 buf[1]; in columnWidthUTF8() 352 UTF32 *Target = &buf[0]; in columnWidthUTF8()
|
/external/swiftshader/third_party/llvm-7.0/llvm/lib/Support/ |
D | ConvertUTF.cpp | 87 static const UTF32 halfBase = 0x0010000UL; 88 static const UTF32 halfMask = 0x3FFUL; 90 #define UNI_SUR_HIGH_START (UTF32)0xD800 91 #define UNI_SUR_HIGH_END (UTF32)0xDBFF 92 #define UNI_SUR_LOW_START (UTF32)0xDC00 93 #define UNI_SUR_LOW_END (UTF32)0xDFFF 120 static const UTF32 offsetsFromUTF8[6] = { 0x00000000UL, 0x00003080UL, 0x000E2080UL, 146 const UTF32** sourceStart, const UTF32* sourceEnd, in ConvertUTF32toUTF16() 149 const UTF32* source = *sourceStart; in ConvertUTF32toUTF16() 152 UTF32 ch; in ConvertUTF32toUTF16() [all …]
|
D | DJB.cpp | 22 static UTF32 chopOneUTF32(StringRef &Buffer) { in chopOneUTF32() 23 UTF32 C; in chopOneUTF32() 27 UTF32 *Begin32 = &C; in chopOneUTF32() 38 static StringRef toUTF8(UTF32 C, MutableArrayRef<UTF8> Storage) { in toUTF8() 39 const UTF32 *Begin32 = &C; in toUTF8() 52 static UTF32 foldCharDwarf(UTF32 C) { in foldCharDwarf() 62 UTF32 C = chopOneUTF32(Buffer); in caseFoldingDjbHashCharSlow()
|
D | ConvertUTFWrapper.cpp | 51 UTF32 *targetStart = reinterpret_cast<UTF32*>(ResultPtr); in ConvertUTF8toWide() 67 const UTF32 *SourceStart = &Source; in ConvertCodePointToUTF8() 68 const UTF32 *SourceEnd = SourceStart + 1; in ConvertCodePointToUTF8() 230 const UTF32 *Start = reinterpret_cast<const UTF32 *>(Source.data()); in convertWideToUTF8() 231 const UTF32 *End = in convertWideToUTF8() 232 reinterpret_cast<const UTF32 *>(Source.data() + Source.size()); in convertWideToUTF8()
|
D | Unicode.cpp | 350 UTF32 buf[1]; in columnWidthUTF8() 352 UTF32 *Target = &buf[0]; in columnWidthUTF8()
|
/external/swiftshader/third_party/llvm-subzero/lib/Support/ |
D | ConvertUTF.cpp | 60 static const UTF32 halfBase = 0x0010000UL; 61 static const UTF32 halfMask = 0x3FFUL; 63 #define UNI_SUR_HIGH_START (UTF32)0xD800 64 #define UNI_SUR_HIGH_END (UTF32)0xDBFF 65 #define UNI_SUR_LOW_START (UTF32)0xDC00 66 #define UNI_SUR_LOW_END (UTF32)0xDFFF 93 static const UTF32 offsetsFromUTF8[6] = { 0x00000000UL, 0x00003080UL, 0x000E2080UL, 119 const UTF32** sourceStart, const UTF32* sourceEnd, in ConvertUTF32toUTF16() 122 const UTF32* source = *sourceStart; in ConvertUTF32toUTF16() 125 UTF32 ch; in ConvertUTF32toUTF16() [all …]
|
D | ConvertUTFWrapper.cpp | 51 UTF32 *targetStart = reinterpret_cast<UTF32*>(ResultPtr); in ConvertUTF8toWide() 67 const UTF32 *SourceStart = &Source; in ConvertCodePointToUTF8() 68 const UTF32 *SourceEnd = SourceStart + 1; in ConvertCodePointToUTF8() 230 const UTF32 *Start = reinterpret_cast<const UTF32 *>(Source.data()); in convertWideToUTF8() 231 const UTF32 *End = in convertWideToUTF8() 232 reinterpret_cast<const UTF32 *>(Source.data() + Source.size()); in convertWideToUTF8()
|
/external/antlr/runtime/C/src/ |
D | antlr3convertutf.c | 53 const UTF32** sourceStart, const UTF32* sourceEnd, in ConvertUTF32toUTF16() 56 const UTF32* source = *sourceStart; in ConvertUTF32toUTF16() 59 UTF32 ch; in ConvertUTF32toUTF16() 103 UTF32** targetStart, UTF32* targetEnd, ConversionFlags flags) { in ConvertUTF16toUTF32() 106 UTF32* target = *targetStart; in ConvertUTF16toUTF32() 107 UTF32 ch, ch2; in ConvertUTF16toUTF32() 181 static const UTF32 offsetsFromUTF8[6] = { 0x00000000UL, 0x00003080UL, 0x000E2080UL, 212 UTF32 ch; in ConvertUTF16toUTF8() 214 const UTF32 byteMask = 0xBF; in ConvertUTF16toUTF8() 215 const UTF32 byteMark = 0x80; in ConvertUTF16toUTF8() [all …]
|
D | antlr3inputstream.c | 788 UTF32 ch; in antlr3UTF16Consume() 789 UTF32 ch2; in antlr3UTF16Consume() 867 UTF32 ch; in antlr3UTF16LA() 868 UTF32 ch2; in antlr3UTF16LA() 1070 UTF32 ch; in antlr3UTF16ConsumeLE() 1071 UTF32 ch2; in antlr3UTF16ConsumeLE() 1147 UTF32 ch; in antlr3UTF16LALE() 1148 UTF32 ch2; in antlr3UTF16LALE() 1289 UTF32 ch; in antlr3UTF16ConsumeBE() 1290 UTF32 ch2; in antlr3UTF16ConsumeBE() [all …]
|
/external/llvm/include/llvm/Support/ |
D | ConvertUTF.h | 101 typedef unsigned int UTF32; /* at least 32 bits */ typedef 107 #define UNI_REPLACEMENT_CHAR (UTF32)0x0000FFFD 108 #define UNI_MAX_BMP (UTF32)0x0000FFFF 109 #define UNI_MAX_UTF16 (UTF32)0x0010FFFF 110 #define UNI_MAX_UTF32 (UTF32)0x7FFFFFFF 111 #define UNI_MAX_LEGAL_UTF32 (UTF32)0x0010FFFF 145 UTF32** targetStart, UTF32* targetEnd, ConversionFlags flags); 153 UTF32** targetStart, UTF32* targetEnd, ConversionFlags flags); 160 const UTF32** sourceStart, const UTF32* sourceEnd, 165 UTF32** targetStart, UTF32* targetEnd, ConversionFlags flags); [all …]
|
/external/swiftshader/third_party/llvm-subzero/include/llvm/Support/ |
D | ConvertUTF.h | 109 typedef unsigned int UTF32; /* at least 32 bits */ typedef 115 #define UNI_REPLACEMENT_CHAR (UTF32)0x0000FFFD 116 #define UNI_MAX_BMP (UTF32)0x0000FFFF 117 #define UNI_MAX_UTF16 (UTF32)0x0010FFFF 118 #define UNI_MAX_UTF32 (UTF32)0x7FFFFFFF 119 #define UNI_MAX_LEGAL_UTF32 (UTF32)0x0010FFFF 148 UTF32** targetStart, UTF32* targetEnd, ConversionFlags flags); 156 UTF32** targetStart, UTF32* targetEnd, ConversionFlags flags); 163 const UTF32** sourceStart, const UTF32* sourceEnd, 168 UTF32** targetStart, UTF32* targetEnd, ConversionFlags flags); [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/include/llvm/Support/ |
D | ConvertUTF.h | 110 typedef unsigned int UTF32; /* at least 32 bits */ typedef 116 #define UNI_REPLACEMENT_CHAR (UTF32)0x0000FFFD 117 #define UNI_MAX_BMP (UTF32)0x0000FFFF 118 #define UNI_MAX_UTF16 (UTF32)0x0010FFFF 119 #define UNI_MAX_UTF32 (UTF32)0x7FFFFFFF 120 #define UNI_MAX_LEGAL_UTF32 (UTF32)0x0010FFFF 149 UTF32** targetStart, UTF32* targetEnd, ConversionFlags flags); 157 UTF32** targetStart, UTF32* targetEnd, ConversionFlags flags); 164 const UTF32** sourceStart, const UTF32* sourceEnd, 169 UTF32** targetStart, UTF32* targetEnd, ConversionFlags flags); [all …]
|
/external/icu/icu4j/main/classes/charset/src/com/ibm/icu/charset/ |
D | UConverterSharedData.java | 384 static final int UTF32 = UTF16 + 1; // 30 field in UConverterSharedData.UConverterType 385 static final int CESU8 = UTF32 + 1; // 31
|
/external/clang/lib/Lex/ |
D | LiteralSupport.cpp | 374 typedef uint32_t UTF32; in EncodeUCNEscape() typedef 375 UTF32 UcnVal = 0; in EncodeUCNEscape() 392 UTF32 *ResultPtr = reinterpret_cast<UTF32*>(ResultBuf); in EncodeUCNEscape() 403 if (UcnVal <= (UTF32)0xFFFF) { in EncodeUCNEscape() 426 if (UcnVal < (UTF32)0x80) in EncodeUCNEscape() 428 else if (UcnVal < (UTF32)0x800) in EncodeUCNEscape() 430 else if (UcnVal < (UTF32)0x10000) in EncodeUCNEscape() 1502 UTF32 *ResultWidePtr = reinterpret_cast<UTF32*>(ResultPtr); in init() 1523 UTF32 *ResultWidePtr = reinterpret_cast<UTF32*>(ResultBuf.data()); in init()
|