/external/mesa3d/src/gallium/auxiliary/draw/ |
D | draw_pipe_wide_point.c | 98 static void set_texcoords(const struct widepoint_stage *wide, in set_texcoords() argument 101 const struct draw_context *draw = wide->stage.draw; in set_texcoords() 106 for (i = 0; i < wide->num_texcoord_gen; i++) { in set_texcoords() 107 const uint slot = wide->texcoord_gen_slot[i]; in set_texcoords() 127 const struct widepoint_stage *wide = widepoint_stage(stage); in widepoint_point() local 147 if (wide->psize_slot >= 0) { in widepoint_point() 148 half_size = header->v[0]->data[wide->psize_slot][0]; in widepoint_point() 152 half_size = wide->half_point_size; in widepoint_point() 155 left_adj = -half_size + wide->xbias; in widepoint_point() 156 right_adj = half_size + wide->xbias; in widepoint_point() [all …]
|
D | draw_pipe_wide_line.c | 193 struct wideline_stage *wide = CALLOC_STRUCT(wideline_stage); in draw_wide_line_stage() local 194 if (!wide) in draw_wide_line_stage() 197 wide->stage.draw = draw; in draw_wide_line_stage() 198 wide->stage.name = "wide-line"; in draw_wide_line_stage() 199 wide->stage.next = NULL; in draw_wide_line_stage() 200 wide->stage.point = draw_pipe_passthrough_point; in draw_wide_line_stage() 201 wide->stage.line = wideline_first_line; in draw_wide_line_stage() 202 wide->stage.tri = draw_pipe_passthrough_tri; in draw_wide_line_stage() 203 wide->stage.flush = wideline_flush; in draw_wide_line_stage() 204 wide->stage.reset_stipple_counter = wideline_reset_stipple_counter; in draw_wide_line_stage() [all …]
|
/external/llvm/test/CodeGen/PowerPC/ |
D | vsx-infl-copy1.ll | 29 %wide.load32 = load <4 x i32>, <4 x i32>* null, align 4 33 %wide.load36 = load <4 x i32>, <4 x i32>* %1, align 4 34 %wide.load37 = load <4 x i32>, <4 x i32>* %addr1, align 4 38 %wide.load38 = load <4 x i32>, <4 x i32>* %3, align 4 42 %wide.load39 = load <4 x i32>, <4 x i32>* %5, align 4 45 %wide.load40 = load <4 x i32>, <4 x i32>* %7, align 4 49 %wide.load41 = load <4 x i32>, <4 x i32>* %9, align 4 52 %wide.load42 = load <4 x i32>, <4 x i32>* %11, align 4 56 %wide.load43 = load <4 x i32>, <4 x i32>* %13, align 4 60 %wide.load44 = load <4 x i32>, <4 x i32>* %15, align 4 [all …]
|
D | swaps-le-1.ll | 33 %wide.load = load <4 x i32>, <4 x i32>* %1, align 16 36 %wide.load13 = load <4 x i32>, <4 x i32>* %3, align 16 37 %4 = add nsw <4 x i32> %wide.load13, %wide.load 40 %wide.load14 = load <4 x i32>, <4 x i32>* %6, align 16 41 %7 = mul nsw <4 x i32> %4, %wide.load14 48 %wide.load.1 = load <4 x i32>, <4 x i32>* %11, align 16 51 %wide.load13.1 = load <4 x i32>, <4 x i32>* %13, align 16 52 %14 = add nsw <4 x i32> %wide.load13.1, %wide.load.1 55 %wide.load14.1 = load <4 x i32>, <4 x i32>* %16, align 16 56 %17 = mul nsw <4 x i32> %14, %wide.load14.1 [all …]
|
D | unal-altivec2.ll | 17 %wide.load = load <4 x float>, <4 x float>* %1, align 4 18 %2 = call <4 x float> @llvm_cos_v4f32(<4 x float> %wide.load) 25 %wide.load.1 = load <4 x float>, <4 x float>* %6, align 4 26 %7 = call <4 x float> @llvm_cos_v4f32(<4 x float> %wide.load.1) 33 %wide.load.2 = load <4 x float>, <4 x float>* %11, align 4 34 %12 = call <4 x float> @llvm_cos_v4f32(<4 x float> %wide.load.2) 41 %wide.load.3 = load <4 x float>, <4 x float>* %16, align 4 42 %17 = call <4 x float> @llvm_cos_v4f32(<4 x float> %wide.load.3) 49 %wide.load.4 = load <4 x float>, <4 x float>* %21, align 4 50 %22 = call <4 x float> @llvm_cos_v4f32(<4 x float> %wide.load.4) [all …]
|
D | qpx-unal-cons-lds.ll | 52 %wide.vec = load <8 x double>, <8 x double>* %2, align 8 53 …%strided.vec = shufflevector <8 x double> %wide.vec, <8 x double> undef, <4 x i32> <i32 0, i32 2, … 62 %wide.vec.1 = load <8 x double>, <8 x double>* %8, align 8 63 …%strided.vec.1 = shufflevector <8 x double> %wide.vec.1, <8 x double> undef, <4 x i32> <i32 0, i32… 72 %wide.vec.2 = load <8 x double>, <8 x double>* %14, align 8 73 …%strided.vec.2 = shufflevector <8 x double> %wide.vec.2, <8 x double> undef, <4 x i32> <i32 0, i32… 82 %wide.vec.3 = load <8 x double>, <8 x double>* %20, align 8 83 …%strided.vec.3 = shufflevector <8 x double> %wide.vec.3, <8 x double> undef, <4 x i32> <i32 0, i32… 92 %wide.vec.4 = load <8 x double>, <8 x double>* %26, align 8 93 …%strided.vec.4 = shufflevector <8 x double> %wide.vec.4, <8 x double> undef, <4 x i32> <i32 0, i32… [all …]
|
D | vsx-infl-copy2.ll | 32 %wide.load72 = load <4 x i32>, <4 x i32>* null, align 4 36 %wide.load73 = load <4 x i32>, <4 x i32>* %1, align 4 40 %wide.load74 = load <4 x i32>, <4 x i32>* %3, align 4 44 %wide.load76 = load <4 x i32>, <4 x i32>* %5, align 4 48 %wide.load78 = load <4 x i32>, <4 x i32>* %7, align 4 52 %wide.load79 = load <4 x i32>, <4 x i32>* %9, align 4 56 %wide.load80 = load <4 x i32>, <4 x i32>* %11, align 4 60 %wide.load81 = load <4 x i32>, <4 x i32>* %13, align 4 64 %wide.load82 = load <4 x i32>, <4 x i32>* %15, align 4 65 %16 = mul <4 x i32> %wide.load72, <i32 269850533, i32 269850533, i32 269850533, i32 269850533> [all …]
|
/external/llvm/test/CodeGen/Generic/ |
D | vector-redux.ll | 29 %wide.load = load <4 x i32>, <4 x i32>* %1, align 16 32 %wide.load5 = load <4 x i32>, <4 x i32>* %3, align 16 33 %4 = add nsw <4 x i32> %wide.load, %vec.phi 34 %5 = add nsw <4 x i32> %wide.load5, %vec.phi4 38 %wide.load.1 = load <4 x i32>, <4 x i32>* %7, align 16 41 %wide.load5.1 = load <4 x i32>, <4 x i32>* %9, align 16 42 %10 = add nsw <4 x i32> %wide.load.1, %4 43 %11 = add nsw <4 x i32> %wide.load5.1, %5 47 %wide.load.2 = load <4 x i32>, <4 x i32>* %13, align 16 50 %wide.load5.2 = load <4 x i32>, <4 x i32>* %15, align 16 [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/Generic/ |
D | vector-redux.ll | 29 %wide.load = load <4 x i32>, <4 x i32>* %1, align 16 32 %wide.load5 = load <4 x i32>, <4 x i32>* %3, align 16 33 %4 = add nsw <4 x i32> %wide.load, %vec.phi 34 %5 = add nsw <4 x i32> %wide.load5, %vec.phi4 38 %wide.load.1 = load <4 x i32>, <4 x i32>* %7, align 16 41 %wide.load5.1 = load <4 x i32>, <4 x i32>* %9, align 16 42 %10 = add nsw <4 x i32> %wide.load.1, %4 43 %11 = add nsw <4 x i32> %wide.load5.1, %5 47 %wide.load.2 = load <4 x i32>, <4 x i32>* %13, align 16 50 %wide.load5.2 = load <4 x i32>, <4 x i32>* %15, align 16 [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/PowerPC/ |
D | vsx-infl-copy1.ll | 37 %wide.load32 = load <4 x i32>, <4 x i32>* null, align 4 41 %wide.load36 = load <4 x i32>, <4 x i32>* %1, align 4 42 %wide.load37 = load <4 x i32>, <4 x i32>* %addr1, align 4 46 %wide.load38 = load <4 x i32>, <4 x i32>* %3, align 4 50 %wide.load39 = load <4 x i32>, <4 x i32>* %5, align 4 53 %wide.load40 = load <4 x i32>, <4 x i32>* %7, align 4 57 %wide.load41 = load <4 x i32>, <4 x i32>* %9, align 4 60 %wide.load42 = load <4 x i32>, <4 x i32>* %11, align 4 64 %wide.load43 = load <4 x i32>, <4 x i32>* %13, align 4 68 %wide.load44 = load <4 x i32>, <4 x i32>* %15, align 4 [all …]
|
D | unal-altivec2.ll | 17 %wide.load = load <4 x float>, <4 x float>* %1, align 4 18 %2 = call <4 x float> @llvm_cos_v4f32(<4 x float> %wide.load) 25 %wide.load.1 = load <4 x float>, <4 x float>* %6, align 4 26 %7 = call <4 x float> @llvm_cos_v4f32(<4 x float> %wide.load.1) 33 %wide.load.2 = load <4 x float>, <4 x float>* %11, align 4 34 %12 = call <4 x float> @llvm_cos_v4f32(<4 x float> %wide.load.2) 41 %wide.load.3 = load <4 x float>, <4 x float>* %16, align 4 42 %17 = call <4 x float> @llvm_cos_v4f32(<4 x float> %wide.load.3) 49 %wide.load.4 = load <4 x float>, <4 x float>* %21, align 4 50 %22 = call <4 x float> @llvm_cos_v4f32(<4 x float> %wide.load.4) [all …]
|
D | qpx-unal-cons-lds.ll | 52 %wide.vec = load <8 x double>, <8 x double>* %2, align 8 53 …%strided.vec = shufflevector <8 x double> %wide.vec, <8 x double> undef, <4 x i32> <i32 0, i32 2, … 62 %wide.vec.1 = load <8 x double>, <8 x double>* %8, align 8 63 …%strided.vec.1 = shufflevector <8 x double> %wide.vec.1, <8 x double> undef, <4 x i32> <i32 0, i32… 72 %wide.vec.2 = load <8 x double>, <8 x double>* %14, align 8 73 …%strided.vec.2 = shufflevector <8 x double> %wide.vec.2, <8 x double> undef, <4 x i32> <i32 0, i32… 82 %wide.vec.3 = load <8 x double>, <8 x double>* %20, align 8 83 …%strided.vec.3 = shufflevector <8 x double> %wide.vec.3, <8 x double> undef, <4 x i32> <i32 0, i32… 92 %wide.vec.4 = load <8 x double>, <8 x double>* %26, align 8 93 …%strided.vec.4 = shufflevector <8 x double> %wide.vec.4, <8 x double> undef, <4 x i32> <i32 0, i32… [all …]
|
D | swaps-le-1.ll | 51 %wide.load = load <4 x i32>, <4 x i32>* %1, align 8 54 %wide.load13 = load <4 x i32>, <4 x i32>* %3, align 8 55 %4 = add nsw <4 x i32> %wide.load13, %wide.load 58 %wide.load14 = load <4 x i32>, <4 x i32>* %6, align 8 59 %7 = mul nsw <4 x i32> %4, %wide.load14 66 %wide.load.1 = load <4 x i32>, <4 x i32>* %11, align 8 69 %wide.load13.1 = load <4 x i32>, <4 x i32>* %13, align 8 70 %14 = add nsw <4 x i32> %wide.load13.1, %wide.load.1 73 %wide.load14.1 = load <4 x i32>, <4 x i32>* %16, align 8 74 %17 = mul nsw <4 x i32> %14, %wide.load14.1 [all …]
|
D | vsx-infl-copy2.ll | 32 %wide.load72 = load <4 x i32>, <4 x i32>* null, align 4 36 %wide.load73 = load <4 x i32>, <4 x i32>* %1, align 4 40 %wide.load74 = load <4 x i32>, <4 x i32>* %3, align 4 44 %wide.load76 = load <4 x i32>, <4 x i32>* %5, align 4 48 %wide.load78 = load <4 x i32>, <4 x i32>* %7, align 4 52 %wide.load79 = load <4 x i32>, <4 x i32>* %9, align 4 56 %wide.load80 = load <4 x i32>, <4 x i32>* %11, align 4 60 %wide.load81 = load <4 x i32>, <4 x i32>* %13, align 4 64 %wide.load82 = load <4 x i32>, <4 x i32>* %15, align 4 65 %16 = mul <4 x i32> %wide.load72, <i32 269850533, i32 269850533, i32 269850533, i32 269850533> [all …]
|
/external/apache-commons-bcel/src/main/java/org/apache/bcel/generic/ |
D | IINC.java | 32 private boolean wide; field in IINC 63 if (wide) { in dump() 67 if (wide) { in dump() 78 wide = super.getIndex() > org.apache.bcel.Const.MAX_BYTE; in setWide() 80 wide = wide || (c > Byte.MAX_VALUE); in setWide() 82 wide = wide || (c < Byte.MIN_VALUE); in setWide() 84 if (wide) { in setWide() 96 protected void initFromFile( final ByteSequence bytes, final boolean wide ) throws IOException { in initFromFile() argument 97 this.wide = wide; in initFromFile() 98 if (wide) { in initFromFile()
|
D | RET.java | 34 private boolean wide; field in RET 58 if (wide) { in dump() 62 if (wide) { in dump() 71 wide = index > org.apache.bcel.Const.MAX_BYTE; in setWide() 72 if (wide) { in setWide() 84 protected void initFromFile( final ByteSequence bytes, final boolean wide ) throws IOException { in initFromFile() argument 85 this.wide = wide; in initFromFile() 86 if (wide) { in initFromFile()
|
/external/smali/smali/src/test/resources/LexerTest/ |
D | InstructionTest.smali | 8 move-result-wide 12 return-wide 14 move-wide 65 sget-wide 72 sput-wide 79 sget-wide-volatile 82 sput-wide-volatile 89 const-wide/high16 91 const-wide/16 105 iget-wide [all …]
|
D | InstructionTest.tokens | 8 INSTRUCTION_FORMAT11x("move-result-wide") 12 INSTRUCTION_FORMAT11x("return-wide") 14 INSTRUCTION_FORMAT12x_OR_ID("move-wide") 65 INSTRUCTION_FORMAT21c_FIELD("sget-wide") 72 INSTRUCTION_FORMAT21c_FIELD("sput-wide") 79 INSTRUCTION_FORMAT21c_FIELD_ODEX("sget-wide-volatile") 82 INSTRUCTION_FORMAT21c_FIELD_ODEX("sput-wide-volatile") 89 INSTRUCTION_FORMAT21lh("const-wide/high16") 91 INSTRUCTION_FORMAT21s("const-wide/16") 105 INSTRUCTION_FORMAT22c_FIELD("iget-wide") [all …]
|
/external/llvm/test/CodeGen/X86/ |
D | pmovsx-inreg.ll | 9 %wide.load35 = load <2 x i8>, <2 x i8>* %in, align 1 10 %sext = sext <2 x i8> %wide.load35 to <2 x i64> 26 %wide.load35 = load <4 x i8>, <4 x i8>* %in, align 1 27 %sext = sext <4 x i8> %wide.load35 to <4 x i64> 37 %wide.load35 = load <4 x i8>, <4 x i8>* %in, align 1 38 %sext = sext <4 x i8> %wide.load35 to <4 x i32> 54 %wide.load35 = load <8 x i8>, <8 x i8>* %in, align 1 55 %sext = sext <8 x i8> %wide.load35 to <8 x i32> 65 %wide.load35 = load <8 x i8>, <8 x i8>* %in, align 1 66 %sext = sext <8 x i8> %wide.load35 to <8 x i16> [all …]
|
/external/libchrome/base/strings/ |
D | sys_string_conversions_unittest.cc | 173 std::wstring wide = kConvertRoundtripCases[i]; in TEST() local 174 std::wstring trip = SysNativeMBToWide(SysWideToNativeMB(wide)); in TEST() 175 EXPECT_EQ(wide.size(), trip.size()); in TEST() 176 EXPECT_EQ(wide, trip); in TEST() 181 std::wstring wide = kConvertRoundtripCases[i]; in TEST() local 182 std::wstring trip = SysNativeMBToWide(WideToUTF8(wide)); in TEST() 183 EXPECT_EQ(wide.size(), trip.size()); in TEST() 184 EXPECT_EQ(wide, trip); in TEST() 188 std::wstring wide = kConvertRoundtripCases[i]; in TEST() local 189 std::wstring trip = UTF8ToWide(SysWideToNativeMB(wide)); in TEST() [all …]
|
D | old_utf_string_conversions.cc | 93 std::string WideToUTF8(const std::wstring& wide) { in WideToUTF8() argument 94 if (IsStringASCII(wide)) { in WideToUTF8() 95 return std::string(wide.data(), wide.data() + wide.length()); in WideToUTF8() 99 PrepareForUTF8Output(wide.data(), wide.length(), &ret); in WideToUTF8() 100 ConvertUnicode(wide.data(), wide.length(), &ret); in WideToUTF8() 135 string16 WideToUTF16(const std::wstring& wide) { in WideToUTF16() argument 136 return wide; in WideToUTF16() 158 string16 WideToUTF16(const std::wstring& wide) { in WideToUTF16() argument 160 WideToUTF16(wide.data(), wide.length(), &ret); in WideToUTF16()
|
D | sys_string_conversions_posix.cc | 16 std::string SysWideToUTF8(const std::wstring& wide) { in SysWideToUTF8() argument 19 return WideToUTF8(wide); in SysWideToUTF8() 33 std::string SysWideToNativeMB(const std::wstring& wide) { in SysWideToNativeMB() argument 34 return WideToUTF8(wide); in SysWideToNativeMB() 43 std::string SysWideToNativeMB(const std::wstring& wide) { in SysWideToNativeMB() argument 50 for (size_t i = 0; i < wide.size(); ++i) { in SysWideToNativeMB() 51 const wchar_t src = wide[i]; in SysWideToNativeMB() 81 for (size_t i = 0, j = 0; i < wide.size(); ++i) { in SysWideToNativeMB() 82 const wchar_t src = wide[i]; in SysWideToNativeMB()
|
/external/proguard/src/proguard/classfile/instruction/ |
D | VariableInstruction.java | 35 public boolean wide; field in VariableInstruction 46 public VariableInstruction(boolean wide) in VariableInstruction() argument 48 this.wide = wide; in VariableInstruction() 72 this.wide = requiredVariableIndexSize() > 1 || in VariableInstruction() 87 this.wide = variableInstruction.wide; in copy() 243 wide = requiredVariableIndexSize() > 1 || in shrink() 252 return wide; in isWide() 300 return (wide ? 2 : 1) + variableIndexSize() + constantSize(); in length() 315 (wide ? "_w" : "") + in toString() 332 wide ? 2 : in variableIndexSize() [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/MC/ARM/ |
D | thumb2-narrow-dp.ll | 106 ANDS r0, r2, r1 // Must be wide - 3 distinct registers 109 ANDS.W r0, r0, r1 // Explicitly wide 111 AND r0, r1, r0 // Must use wide encoding as not flag-setting 114 ANDS r8, r1, r8 // high registers so must use wide encoding 118 ANDS r2, r2, r1, lsl #1 // Must use wide - shifted register 136 ANDEQ r0, r2, r1 // Must be wide - 3 distinct registers 142 ANDEQ.W r0, r0, r1 // Explicitly wide 146 ANDSEQ r0, r1, r0 // Must use wide encoding as flag-setting 152 ANDEQ r8, r1, r8 // high registers so must use wide encoding 160 ANDEQ r0, r0, r1, lsl #1 // Must use wide - shifted register [all …]
|
/external/llvm/test/MC/ARM/ |
D | thumb2-narrow-dp.ll | 106 ANDS r0, r2, r1 // Must be wide - 3 distinct registers 109 ANDS.W r0, r0, r1 // Explicitly wide 111 AND r0, r1, r0 // Must use wide encoding as not flag-setting 114 ANDS r8, r1, r8 // high registers so must use wide encoding 118 ANDS r2, r2, r1, lsl #1 // Must use wide - shifted register 136 ANDEQ r0, r2, r1 // Must be wide - 3 distinct registers 142 ANDEQ.W r0, r0, r1 // Explicitly wide 146 ANDSEQ r0, r1, r0 // Must use wide encoding as flag-setting 152 ANDEQ r8, r1, r8 // high registers so must use wide encoding 160 ANDEQ r0, r0, r1, lsl #1 // Must use wide - shifted register [all …]
|