/external/libvpx/libvpx/vpx_dsp/mips/ |
D | idct32x32_msa.c | 44 v8i16 reg0, reg1, reg2, reg3, reg4, reg5, reg6, reg7; in idct32x8_row_even_process_store() local 48 LD_SH8(tmp_buf, 32, reg0, reg1, reg2, reg3, reg4, reg5, reg6, reg7); in idct32x8_row_even_process_store() 51 DOTP_CONST_PAIR(reg5, reg3, cospi_12_64, cospi_20_64, reg5, reg3); in idct32x8_row_even_process_store() 52 BUTTERFLY_4(reg1, reg7, reg3, reg5, vec1, vec3, vec2, vec0); in idct32x8_row_even_process_store() 65 LD_SH8((tmp_buf + 16), 32, reg0, reg1, reg2, reg3, reg4, reg5, reg6, reg7); in idct32x8_row_even_process_store() 67 DOTP_CONST_PAIR(reg4, reg3, cospi_14_64, cospi_18_64, reg4, reg3); in idct32x8_row_even_process_store() 77 reg5 = reg7 + reg3; in idct32x8_row_even_process_store() 78 reg7 = reg7 - reg3; in idct32x8_row_even_process_store() 79 reg3 = vec0; in idct32x8_row_even_process_store() 82 reg2 = reg3 + reg4; in idct32x8_row_even_process_store() [all …]
|
D | idct16x16_msa.c | 16 v8i16 reg3, reg13, reg11, reg5, reg7, reg9, reg1, reg15; in vpx_idct16_1d_rows_msa() local 19 LD_SH8(input, 16, reg0, reg1, reg2, reg3, reg4, reg5, reg6, reg7); in vpx_idct16_1d_rows_msa() 23 TRANSPOSE8x8_SH_SH(reg0, reg1, reg2, reg3, reg4, reg5, reg6, reg7, reg0, reg1, in vpx_idct16_1d_rows_msa() 24 reg2, reg3, reg4, reg5, reg6, reg7); in vpx_idct16_1d_rows_msa() 49 DOTP_CONST_PAIR(reg13, reg3, cospi_6_64, cospi_26_64, loc0, loc1); in vpx_idct16_1d_rows_msa() 50 BUTTERFLY_4(loc0, loc1, reg11, reg5, reg13, reg3, reg11, reg5); in vpx_idct16_1d_rows_msa() 52 loc1 = reg15 + reg3; in vpx_idct16_1d_rows_msa() 53 reg3 = reg15 - reg3; in vpx_idct16_1d_rows_msa() 85 DOTP_CONST_PAIR(reg3, reg13, cospi_16_64, cospi_16_64, reg3, reg13); in vpx_idct16_1d_rows_msa() 86 BUTTERFLY_4(reg12, reg14, reg13, reg3, reg8, reg6, reg7, reg5); in vpx_idct16_1d_rows_msa() [all …]
|
/external/u-boot/drivers/mtd/nand/ |
D | nand_ecc.c | 68 uint8_t idx, reg1, reg2, reg3, tmp1, tmp2; in nand_calculate_ecc() local 72 reg1 = reg2 = reg3 = 0; in nand_calculate_ecc() 82 reg3 ^= (uint8_t) i; in nand_calculate_ecc() 88 tmp1 = (reg3 & 0x80) >> 0; /* B7 -> B7 */ in nand_calculate_ecc() 90 tmp1 |= (reg3 & 0x40) >> 1; /* B6 -> B5 */ in nand_calculate_ecc() 92 tmp1 |= (reg3 & 0x20) >> 2; /* B5 -> B3 */ in nand_calculate_ecc() 94 tmp1 |= (reg3 & 0x10) >> 3; /* B4 -> B1 */ in nand_calculate_ecc() 97 tmp2 = (reg3 & 0x08) << 4; /* B3 -> B7 */ in nand_calculate_ecc() 99 tmp2 |= (reg3 & 0x04) << 3; /* B2 -> B5 */ in nand_calculate_ecc() 101 tmp2 |= (reg3 & 0x02) << 2; /* B1 -> B3 */ in nand_calculate_ecc() [all …]
|
/external/libvpx/libvpx/vpx_ports/ |
D | asmdefs_mmi.h | 21 #define MMI_ADDU(reg1, reg2, reg3) \ argument 22 "daddu " #reg1 ", " #reg2 ", " #reg3 " \n\t" 30 #define MMI_SUBU(reg1, reg2, reg3) \ argument 31 "dsubu " #reg1 ", " #reg2 ", " #reg3 " \n\t" 50 #define MMI_ADDU(reg1, reg2, reg3) \ argument 51 "addu " #reg1 ", " #reg2 ", " #reg3 " \n\t" 59 #define MMI_SUBU(reg1, reg2, reg3) \ argument 60 "subu " #reg1 ", " #reg2 ", " #reg3 " \n\t"
|
/external/llvm/test/CodeGen/ARM/ |
D | fast-isel-pic.ll | 39 ; THUMB: movw r[[reg3:[0-9]+]], 40 ; THUMB: movt r[[reg3]], 41 ; THUMB: add r[[reg3]], pc 42 ; THUMB: ldr r[[reg3]], [r[[reg3]]] 44 ; THUMB-ELF: ldr r[[reg3:[0-9]+]], 45 ; THUMB-ELF: ldr r[[reg4:[0-9]+]], [r[[reg3]]]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/ARM/ |
D | fast-isel-pic.ll | 39 ; THUMB: movw r[[reg3:[0-9]+]], 40 ; THUMB: movt r[[reg3]], 41 ; THUMB: add r[[reg3]], pc 42 ; THUMB: ldr r[[reg3]], [r[[reg3]]] 44 ; THUMB-ELF: ldr r[[reg3:[0-9]+]], 45 ; THUMB-ELF: ldr r[[reg4:[0-9]+]], [r[[reg3]]]
|
/external/elfutils/tests/ |
D | run-dwarfcfi.sh | 41 reg3: same_value 58 reg3: undefined 75 reg3: undefined 92 reg3: undefined 109 reg3: undefined 126 reg3: undefined
|
D | run-varlocs.sh | 68 [40051c,40052a) {reg3} 107 [400408,400421) {reg3} 179 [40118e,40119c) {reg3} 181 [4011a7,4011b5) {reg3} 257 [40118e,40119c) {reg3} 259 [4011a7,4011b5) {reg3} 331 [40050e,40051c) {reg3} 333 [400527,400535) {reg3}
|
D | run-readelf-zdebug-rel.sh | 237 [ 0] reg3 243 [ 0] reg3 246 [ 0] reg3
|
/external/libyuv/files/source/ |
D | row_msa.cc | 481 v16u8 reg0, reg1, reg2, reg3; in I422ToRGB24Row_MSA() local 508 reg3 = (v16u8)__msa_pckev_b((v16i8)vec5, (v16i8)vec2); in I422ToRGB24Row_MSA() 510 dst0 = (v16u8)__msa_vshf_b(shuffler0, (v16i8)reg3, (v16i8)reg0); in I422ToRGB24Row_MSA() 511 dst1 = (v16u8)__msa_vshf_b(shuffler1, (v16i8)reg3, (v16i8)reg1); in I422ToRGB24Row_MSA() 512 dst2 = (v16u8)__msa_vshf_b(shuffler2, (v16i8)reg3, (v16i8)reg2); in I422ToRGB24Row_MSA() 774 v8u16 reg0, reg1, reg2, reg3, reg4, reg5; in ARGBToYRow_MSA() local 793 reg3 = (v8u16)__msa_ilvev_b(zero, (v16i8)vec3); in ARGBToYRow_MSA() 799 reg3 *= const_0x81; in ARGBToYRow_MSA() 803 reg1 += reg3; in ARGBToYRow_MSA() 826 v8u16 reg0, reg1, reg2, reg3, reg4, reg5, reg6, reg7, reg8, reg9; in ARGBToUVRow_MSA() local [all …]
|
D | rotate_msa.cc | 85 v16u8 reg0, reg1, reg2, reg3, reg4, reg5, reg6, reg7; in TransposeWx16_MSA() local 99 ILVRL_H(vec0, vec2, vec1, vec3, reg0, reg1, reg2, reg3); in TransposeWx16_MSA() 111 ILVRL_W(reg2, reg6, reg3, reg7, res4, res5, res6, res7); in TransposeWx16_MSA() 121 ILVRL_H(vec0, vec2, vec1, vec3, reg0, reg1, reg2, reg3); in TransposeWx16_MSA() 147 res8 = (v16u8)__msa_ilvr_w((v4i32)reg7, (v4i32)reg3); in TransposeWx16_MSA() 148 res9 = (v16u8)__msa_ilvl_w((v4i32)reg7, (v4i32)reg3); in TransposeWx16_MSA() 166 v16u8 reg0, reg1, reg2, reg3, reg4, reg5, reg6, reg7; in TransposeUVWx16_MSA() local 180 ILVRL_H(vec0, vec2, vec1, vec3, reg0, reg1, reg2, reg3); in TransposeUVWx16_MSA() 192 ILVRL_W(reg2, reg6, reg3, reg7, res4, res5, res6, res7); in TransposeUVWx16_MSA() 202 ILVRL_H(vec0, vec2, vec1, vec3, reg0, reg1, reg2, reg3); in TransposeUVWx16_MSA() [all …]
|
D | scale_msa.cc | 70 v8u16 reg0, reg1, reg2, reg3; in ScaleARGBRowDown2Box_MSA() local 85 reg3 = __msa_hadd_u_h(vec3, vec3); in ScaleARGBRowDown2Box_MSA() 87 reg1 += reg3; in ScaleARGBRowDown2Box_MSA() 133 v8u16 reg0, reg1, reg2, reg3, reg4, reg5, reg6, reg7; in ScaleARGBRowDownEvenBox_MSA() local 160 reg3 = __msa_hadd_u_h(vec3, vec3); in ScaleARGBRowDownEvenBox_MSA() 162 reg5 = (v8u16)__msa_pckev_d((v2i64)reg3, (v2i64)reg1); in ScaleARGBRowDownEvenBox_MSA() 164 reg7 = (v8u16)__msa_pckod_d((v2i64)reg3, (v2i64)reg1); in ScaleARGBRowDownEvenBox_MSA() 296 v4u32 reg0, reg1, reg2, reg3; in ScaleRowDown4Box_MSA() local 334 reg3 = __msa_hadd_u_w(vec3, vec3); in ScaleRowDown4Box_MSA() 338 reg3 = (v4u32)__msa_srari_w((v4i32)reg3, 4); in ScaleRowDown4Box_MSA() [all …]
|
/external/libvpx/libvpx/third_party/libyuv/source/ |
D | scale_msa.cc | 78 v8u16 reg0, reg1, reg2, reg3; in ScaleARGBRowDown2Box_MSA() local 93 reg3 = __msa_hadd_u_h(vec3, vec3); in ScaleARGBRowDown2Box_MSA() 95 reg1 += reg3; in ScaleARGBRowDown2Box_MSA() 141 v8u16 reg0, reg1, reg2, reg3, reg4, reg5, reg6, reg7; in ScaleARGBRowDownEvenBox_MSA() local 168 reg3 = __msa_hadd_u_h(vec3, vec3); in ScaleARGBRowDownEvenBox_MSA() 170 reg5 = (v8u16)__msa_pckev_d((v2i64)reg3, (v2i64)reg1); in ScaleARGBRowDownEvenBox_MSA() 172 reg7 = (v8u16)__msa_pckod_d((v2i64)reg3, (v2i64)reg1); in ScaleARGBRowDownEvenBox_MSA() 304 v4u32 reg0, reg1, reg2, reg3; in ScaleRowDown4Box_MSA() local 342 reg3 = __msa_hadd_u_w(vec3, vec3); in ScaleRowDown4Box_MSA() 346 reg3 = (v4u32)__msa_srari_w((v4i32)reg3, 4); in ScaleRowDown4Box_MSA() [all …]
|
D | rotate_msa.cc | 85 v16u8 reg0, reg1, reg2, reg3, reg4, reg5, reg6, reg7; in TransposeWx16_MSA() local 99 ILVRL_H(vec0, vec2, vec1, vec3, reg0, reg1, reg2, reg3); in TransposeWx16_MSA() 111 ILVRL_W(reg2, reg6, reg3, reg7, res4, res5, res6, res7); in TransposeWx16_MSA() 121 ILVRL_H(vec0, vec2, vec1, vec3, reg0, reg1, reg2, reg3); in TransposeWx16_MSA() 147 res8 = (v16u8)__msa_ilvr_w((v4i32)reg7, (v4i32)reg3); in TransposeWx16_MSA() 148 res9 = (v16u8)__msa_ilvl_w((v4i32)reg7, (v4i32)reg3); in TransposeWx16_MSA() 166 v16u8 reg0, reg1, reg2, reg3, reg4, reg5, reg6, reg7; in TransposeUVWx16_MSA() local 180 ILVRL_H(vec0, vec2, vec1, vec3, reg0, reg1, reg2, reg3); in TransposeUVWx16_MSA() 192 ILVRL_W(reg2, reg6, reg3, reg7, res4, res5, res6, res7); in TransposeUVWx16_MSA() 202 ILVRL_H(vec0, vec2, vec1, vec3, reg0, reg1, reg2, reg3); in TransposeUVWx16_MSA() [all …]
|
D | row_msa.cc | 481 v16u8 reg0, reg1, reg2, reg3; in I422ToRGB24Row_MSA() local 508 reg3 = (v16u8)__msa_pckev_b((v16i8)vec5, (v16i8)vec2); in I422ToRGB24Row_MSA() 510 dst0 = (v16u8)__msa_vshf_b(shuffler0, (v16i8)reg3, (v16i8)reg0); in I422ToRGB24Row_MSA() 511 dst1 = (v16u8)__msa_vshf_b(shuffler1, (v16i8)reg3, (v16i8)reg1); in I422ToRGB24Row_MSA() 512 dst2 = (v16u8)__msa_vshf_b(shuffler2, (v16i8)reg3, (v16i8)reg2); in I422ToRGB24Row_MSA() 774 v8u16 reg0, reg1, reg2, reg3, reg4, reg5; in ARGBToYRow_MSA() local 793 reg3 = (v8u16)__msa_ilvev_b(zero, (v16i8)vec3); in ARGBToYRow_MSA() 799 reg3 *= const_0x81; in ARGBToYRow_MSA() 803 reg1 += reg3; in ARGBToYRow_MSA() 826 v8u16 reg0, reg1, reg2, reg3, reg4, reg5, reg6, reg7, reg8, reg9; in ARGBToUVRow_MSA() local [all …]
|
D | compare_msa.cc | 59 v4i32 reg0 = {0}, reg1 = {0}, reg2 = {0}, reg3 = {0}; in SumSquareError_MSA() local 78 reg3 = __msa_dpadd_s_w(reg3, vec3, vec3); in SumSquareError_MSA() 84 reg2 += reg3; in SumSquareError_MSA()
|
/external/v8/src/interpreter/ |
D | bytecode-register.cc | 97 bool Register::AreContiguous(Register reg1, Register reg2, Register reg3, in AreContiguous() argument 102 if (reg3.is_valid() && reg2.index() + 1 != reg3.index()) { in AreContiguous() 105 if (reg4.is_valid() && reg3.index() + 1 != reg4.index()) { in AreContiguous()
|
/external/llvm/test/CodeGen/AMDGPU/ |
D | pv-packing.ll | 6 … x float> inreg %reg0, <4 x float> inreg %reg1, <4 x float> inreg %reg2, <4 x float> inreg %reg3) { 14 %6 = extractelement <4 x float> %reg3, i32 0 15 %7 = extractelement <4 x float> %reg3, i32 1 16 %8 = extractelement <4 x float> %reg3, i32 2
|
D | load-input-fold.ll | 3 … x float> inreg %reg0, <4 x float> inreg %reg1, <4 x float> inreg %reg2, <4 x float> inreg %reg3) { 13 %8 = extractelement <4 x float> %reg3, i32 0 14 %9 = extractelement <4 x float> %reg3, i32 1 15 %10 = extractelement <4 x float> %reg3, i32 2 16 %11 = extractelement <4 x float> %reg3, i32 3
|
D | pv.ll | 6 …eg0, <4 x float> inreg %reg1, <4 x float> inreg %reg2, <4 x float> inreg %reg3, <4 x float> inreg … 16 %8 = extractelement <4 x float> %reg3, i32 0 17 %9 = extractelement <4 x float> %reg3, i32 1 18 %10 = extractelement <4 x float> %reg3, i32 2 19 %11 = extractelement <4 x float> %reg3, i32 3
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/AMDGPU/ |
D | pv-packing.ll | 6 … x float> inreg %reg0, <4 x float> inreg %reg1, <4 x float> inreg %reg2, <4 x float> inreg %reg3) { 14 %6 = extractelement <4 x float> %reg3, i32 0 15 %7 = extractelement <4 x float> %reg3, i32 1 16 %8 = extractelement <4 x float> %reg3, i32 2
|
D | load-input-fold.ll | 3 … x float> inreg %reg0, <4 x float> inreg %reg1, <4 x float> inreg %reg2, <4 x float> inreg %reg3) { 13 %8 = extractelement <4 x float> %reg3, i32 0 14 %9 = extractelement <4 x float> %reg3, i32 1 15 %10 = extractelement <4 x float> %reg3, i32 2 16 %11 = extractelement <4 x float> %reg3, i32 3
|
/external/u-boot/arch/arm/lib/ |
D | memcpy.S | 20 .macro ldr4w ptr reg1 reg2 reg3 reg4 abort 21 ldmia \ptr!, {\reg1, \reg2, \reg3, \reg4} 24 .macro ldr8w ptr reg1 reg2 reg3 reg4 reg5 reg6 reg7 reg8 abort 25 ldmia \ptr!, {\reg1, \reg2, \reg3, \reg4, \reg5, \reg6, \reg7, \reg8} 36 .macro str8w ptr reg1 reg2 reg3 reg4 reg5 reg6 reg7 reg8 abort 37 stmia \ptr!, {\reg1, \reg2, \reg3, \reg4, \reg5, \reg6, \reg7, \reg8}
|
/external/vixl/src/aarch64/ |
D | operands-aarch64.h | 486 const CPURegister& reg3 = NoReg, 500 const CPURegister& reg3 = NoCPUReg, 512 const CPURegister& reg3 = NoReg, 526 const CPURegister& reg3 = NoCPUReg, 535 const VRegister& reg3 = NoVReg, 545 const VRegister& reg3 = NoVReg, 554 CPURegister reg3 = NoCPUReg, 556 : list_(reg1.GetBit() | reg2.GetBit() | reg3.GetBit() | reg4.GetBit()), 559 VIXL_ASSERT(AreSameSizeAndType(reg1, reg2, reg3, reg4));
|
/external/vixl/src/aarch32/ |
D | instructions-aarch32.h | 465 RegisterList(Register reg1, Register reg2, Register reg3) 467 RegisterToList(reg3)) {} 468 RegisterList(Register reg1, Register reg2, Register reg3, Register reg4) 470 RegisterToList(reg3) | RegisterToList(reg4)) {} 554 VRegisterList(VRegister reg1, VRegister reg2, VRegister reg3) in VRegisterList() argument 556 RegisterToList(reg3)) {} in VRegisterList() 557 VRegisterList(VRegister reg1, VRegister reg2, VRegister reg3, VRegister reg4) in VRegisterList() argument 559 RegisterToList(reg3) | RegisterToList(reg4)) {} in VRegisterList()
|