/external/swiftshader/third_party/llvm-7.0/llvm/test/tools/llvm-mca/X86/BtVer2/ |
D | dependent-pmuld-paddd.s | 5 vpaddd %xmm1, %xmm1, %xmm0 label 6 vpaddd %xmm0, %xmm0, %xmm3 label 25 # CHECK-NEXT: 1 1 0.50 vpaddd %xmm1, %xmm1, %xmm0 26 # CHECK-NEXT: 1 1 0.50 vpaddd %xmm0, %xmm0, %xmm3 51 … - 0.50 0.50 - - - - - 1.00 - vpaddd %xmm1, %xmm1, %xmm0 52 … - - 1.00 - - - - 1.00 - - vpaddd %xmm0, %xmm0, %xmm3 59 # CHECK-NEXT: [0,1] D==eER . . . . . . vpaddd %xmm1, %xmm1, %xmm0 60 # CHECK-NEXT: [0,2] .D==eER . . . . . . vpaddd %xmm0, %xmm0, %xmm3 62 # CHECK-NEXT: [1,1] . D===eER . . . . . . vpaddd %xmm1, %xmm1, %xmm0 63 # CHECK-NEXT: [1,2] . D====eER. . . . . . vpaddd %xmm0, %xmm0, %xmm3 [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/X86/ |
D | x86-32-vector-calling-conv.ll | 5 ; DARWIN: vpaddd %xmm3, %xmm2, %xmm2 6 ; DARWIN: vpaddd %xmm2, %xmm1, %xmm1 7 ; DARWIN: vpaddd %xmm1, %xmm0, %xmm0 8 ; LINUX: vpaddd {{[0-9]+}}(%e{{s|b}}p), %xmm2, %xmm2 9 ; LINUX: vpaddd %xmm2, %xmm1, %xmm1 10 ; LINUX: vpaddd %xmm1, %xmm0, %xmm0 19 ; DARWIN: vpaddd %ymm3, %ymm2, %ymm2 20 ; DARWIN: vpaddd %ymm2, %ymm1, %ymm1 21 ; DARWIN: vpaddd %ymm1, %ymm0, %ymm0 22 ; LINUX: vpaddd {{[0-9]+}}(%e{{s|b}}p), %ymm2, %ymm2 [all …]
|
D | avx-logic.ll | 313 ; AVX1-NEXT: vpaddd %xmm1, %xmm0, %xmm3 316 ; AVX1-NEXT: vpaddd %xmm1, %xmm0, %xmm0 320 ; AVX1-NEXT: vpaddd %xmm4, %xmm0, %xmm0 322 ; AVX1-NEXT: vpaddd %xmm2, %xmm1, %xmm1 328 ; INT256-NEXT: vpaddd %ymm1, %ymm0, %ymm0 330 ; INT256-NEXT: vpaddd %ymm2, %ymm0, %ymm0 341 ; AVX1-NEXT: vpaddd %xmm1, %xmm0, %xmm3 344 ; AVX1-NEXT: vpaddd %xmm1, %xmm0, %xmm0 348 ; AVX1-NEXT: vpaddd %xmm4, %xmm0, %xmm0 350 ; AVX1-NEXT: vpaddd %xmm2, %xmm1, %xmm1 [all …]
|
D | madd.ll | 46 ; AVX-NEXT: vpaddd %xmm0, %xmm1, %xmm0 52 ; AVX-NEXT: vpaddd %xmm1, %xmm0, %xmm0 122 ; AVX1-NEXT: vpaddd %xmm0, %xmm1, %xmm1 129 ; AVX1-NEXT: vpaddd %xmm1, %xmm0, %xmm0 131 ; AVX1-NEXT: vpaddd %xmm1, %xmm0, %xmm0 147 ; AVX256-NEXT: vpaddd %ymm0, %ymm1, %ymm0 153 ; AVX256-NEXT: vpaddd %ymm1, %ymm0, %ymm0 155 ; AVX256-NEXT: vpaddd %ymm1, %ymm0, %ymm0 240 ; AVX1-NEXT: vpaddd %xmm5, %xmm4, %xmm4 242 ; AVX1-NEXT: vpaddd %xmm1, %xmm2, %xmm1 [all …]
|
D | avx512vl_vnni-intrinsics-upgrade.ll | 19 ; X86-NEXT: vpaddd %ymm0, %ymm4, %ymm0 # EVEX TO VEX Compression encoding: [0xc5,0xdd,0xfe,0xc0] 20 ; X86-NEXT: vpaddd %ymm0, %ymm3, %ymm0 # EVEX TO VEX Compression encoding: [0xc5,0xe5,0xfe,0xc0] 31 ; X64-NEXT: vpaddd %ymm0, %ymm4, %ymm0 # EVEX TO VEX Compression encoding: [0xc5,0xdd,0xfe,0xc0] 32 ; X64-NEXT: vpaddd %ymm0, %ymm3, %ymm0 # EVEX TO VEX Compression encoding: [0xc5,0xe5,0xfe,0xc0] 57 ; X86-NEXT: vpaddd %xmm0, %xmm4, %xmm0 # EVEX TO VEX Compression encoding: [0xc5,0xd9,0xfe,0xc0] 58 ; X86-NEXT: vpaddd %xmm0, %xmm3, %xmm0 # EVEX TO VEX Compression encoding: [0xc5,0xe1,0xfe,0xc0] 69 ; X64-NEXT: vpaddd %xmm0, %xmm4, %xmm0 # EVEX TO VEX Compression encoding: [0xc5,0xd9,0xfe,0xc0] 70 ; X64-NEXT: vpaddd %xmm0, %xmm3, %xmm0 # EVEX TO VEX Compression encoding: [0xc5,0xe1,0xfe,0xc0] 95 ; X86-NEXT: vpaddd %ymm0, %ymm4, %ymm0 # EVEX TO VEX Compression encoding: [0xc5,0xdd,0xfe,0xc0] 96 ; X86-NEXT: vpaddd %ymm0, %ymm3, %ymm0 # EVEX TO VEX Compression encoding: [0xc5,0xe5,0xfe,0xc0] [all …]
|
D | avx512vl_vnni-intrinsics.ll | 18 ; X86-NEXT: vpaddd %ymm0, %ymm4, %ymm0 # EVEX TO VEX Compression encoding: [0xc5,0xdd,0xfe,0xc0] 19 ; X86-NEXT: vpaddd %ymm0, %ymm3, %ymm0 # EVEX TO VEX Compression encoding: [0xc5,0xe5,0xfe,0xc0] 30 ; X64-NEXT: vpaddd %ymm0, %ymm4, %ymm0 # EVEX TO VEX Compression encoding: [0xc5,0xdd,0xfe,0xc0] 31 ; X64-NEXT: vpaddd %ymm0, %ymm3, %ymm0 # EVEX TO VEX Compression encoding: [0xc5,0xe5,0xfe,0xc0] 59 ; X86-NEXT: vpaddd %xmm0, %xmm4, %xmm0 # EVEX TO VEX Compression encoding: [0xc5,0xd9,0xfe,0xc0] 60 ; X86-NEXT: vpaddd %xmm0, %xmm3, %xmm0 # EVEX TO VEX Compression encoding: [0xc5,0xe1,0xfe,0xc0] 71 ; X64-NEXT: vpaddd %xmm0, %xmm4, %xmm0 # EVEX TO VEX Compression encoding: [0xc5,0xd9,0xfe,0xc0] 72 ; X64-NEXT: vpaddd %xmm0, %xmm3, %xmm0 # EVEX TO VEX Compression encoding: [0xc5,0xe1,0xfe,0xc0] 102 ; X86-NEXT: vpaddd %ymm0, %ymm4, %ymm0 # EVEX TO VEX Compression encoding: [0xc5,0xdd,0xfe,0xc0] 103 ; X86-NEXT: vpaddd %ymm0, %ymm3, %ymm0 # EVEX TO VEX Compression encoding: [0xc5,0xe5,0xfe,0xc0] [all …]
|
D | avx512vnni-intrinsics-upgrade.ll | 18 ; X86-NEXT: vpaddd %zmm0, %zmm4, %zmm0 # encoding: [0x62,0xf1,0x5d,0x48,0xfe,0xc0] 19 ; X86-NEXT: vpaddd %zmm0, %zmm3, %zmm0 # encoding: [0x62,0xf1,0x65,0x48,0xfe,0xc0] 30 ; X64-NEXT: vpaddd %zmm0, %zmm4, %zmm0 # encoding: [0x62,0xf1,0x5d,0x48,0xfe,0xc0] 31 ; X64-NEXT: vpaddd %zmm0, %zmm3, %zmm0 # encoding: [0x62,0xf1,0x65,0x48,0xfe,0xc0] 55 ; X86-NEXT: vpaddd %zmm0, %zmm4, %zmm0 # encoding: [0x62,0xf1,0x5d,0x48,0xfe,0xc0] 56 ; X86-NEXT: vpaddd %zmm0, %zmm3, %zmm0 # encoding: [0x62,0xf1,0x65,0x48,0xfe,0xc0] 67 ; X64-NEXT: vpaddd %zmm0, %zmm4, %zmm0 # encoding: [0x62,0xf1,0x5d,0x48,0xfe,0xc0] 68 ; X64-NEXT: vpaddd %zmm0, %zmm3, %zmm0 # encoding: [0x62,0xf1,0x65,0x48,0xfe,0xc0] 92 ; X86-NEXT: vpaddd %zmm0, %zmm4, %zmm0 # encoding: [0x62,0xf1,0x5d,0x48,0xfe,0xc0] 93 ; X86-NEXT: vpaddd %zmm0, %zmm3, %zmm0 # encoding: [0x62,0xf1,0x65,0x48,0xfe,0xc0] [all …]
|
D | avx512vnni-intrinsics.ll | 17 ; X86-NEXT: vpaddd %zmm0, %zmm4, %zmm0 # encoding: [0x62,0xf1,0x5d,0x48,0xfe,0xc0] 18 ; X86-NEXT: vpaddd %zmm0, %zmm3, %zmm0 # encoding: [0x62,0xf1,0x65,0x48,0xfe,0xc0] 29 ; X64-NEXT: vpaddd %zmm0, %zmm4, %zmm0 # encoding: [0x62,0xf1,0x5d,0x48,0xfe,0xc0] 30 ; X64-NEXT: vpaddd %zmm0, %zmm3, %zmm0 # encoding: [0x62,0xf1,0x65,0x48,0xfe,0xc0] 57 ; X86-NEXT: vpaddd %zmm0, %zmm4, %zmm0 # encoding: [0x62,0xf1,0x5d,0x48,0xfe,0xc0] 58 ; X86-NEXT: vpaddd %zmm0, %zmm3, %zmm0 # encoding: [0x62,0xf1,0x65,0x48,0xfe,0xc0] 69 ; X64-NEXT: vpaddd %zmm0, %zmm4, %zmm0 # encoding: [0x62,0xf1,0x5d,0x48,0xfe,0xc0] 70 ; X64-NEXT: vpaddd %zmm0, %zmm3, %zmm0 # encoding: [0x62,0xf1,0x65,0x48,0xfe,0xc0] 97 ; X86-NEXT: vpaddd %zmm0, %zmm4, %zmm0 # encoding: [0x62,0xf1,0x5d,0x48,0xfe,0xc0] 98 ; X86-NEXT: vpaddd %zmm0, %zmm3, %zmm0 # encoding: [0x62,0xf1,0x65,0x48,0xfe,0xc0] [all …]
|
D | sad.ll | 47 ; AVX1-NEXT: vpaddd %xmm1, %xmm2, %xmm2 54 ; AVX1-NEXT: vpaddd %xmm3, %xmm2, %xmm2 55 ; AVX1-NEXT: vpaddd %xmm2, %xmm0, %xmm0 56 ; AVX1-NEXT: vpaddd %xmm0, %xmm1, %xmm0 58 ; AVX1-NEXT: vpaddd %xmm1, %xmm0, %xmm0 74 ; AVX2-NEXT: vpaddd %ymm1, %ymm2, %ymm1 78 ; AVX2-NEXT: vpaddd %ymm0, %ymm1, %ymm0 80 ; AVX2-NEXT: vpaddd %ymm1, %ymm0, %ymm0 82 ; AVX2-NEXT: vpaddd %ymm1, %ymm0, %ymm0 97 ; AVX512-NEXT: vpaddd %zmm0, %zmm1, %zmm0 [all …]
|
D | avx512cdvl-intrinsics.ll | 13 ; X86-NEXT: vpaddd %xmm0, %xmm2, %xmm0 14 ; X86-NEXT: vpaddd %xmm0, %xmm1, %xmm0 23 ; X64-NEXT: vpaddd %xmm0, %xmm2, %xmm0 24 ; X64-NEXT: vpaddd %xmm0, %xmm1, %xmm0 48 ; X86-NEXT: vpaddd %ymm2, %ymm1, %ymm0 56 ; X64-NEXT: vpaddd %ymm2, %ymm1, %ymm0 131 ; X86-NEXT: vpaddd %xmm2, %xmm0, %xmm0 132 ; X86-NEXT: vpaddd %xmm0, %xmm1, %xmm0 141 ; X64-NEXT: vpaddd %xmm2, %xmm0, %xmm0 142 ; X64-NEXT: vpaddd %xmm0, %xmm1, %xmm0 [all …]
|
/external/llvm/test/CodeGen/X86/ |
D | x86-32-vector-calling-conv.ll | 5 ; DARWIN-DAG: vpaddd %xmm1, %xmm0, %xmm0 6 ; DARWIN-DAG: vpaddd %xmm3, %xmm2, %xmm1 7 ; DARWIN: vpaddd %xmm1, %xmm0, %xmm0 8 ; LINUX-DAG: vpaddd %xmm1, %xmm0, %xmm0 9 ; LINUX-DAG: vpaddd {{[0-9]+}}(%e{{s|b}}p), %xmm2, %xmm1 10 ; LINUX: vpaddd %xmm1, %xmm0, %xmm0 19 ; DARWIN-DAG: vpaddd %ymm1, %ymm0, %ymm0 20 ; DARWIN-DAG: vpaddd %ymm3, %ymm2, %ymm1 21 ; DARWIN: vpaddd %ymm1, %ymm0, %ymm0 22 ; LINUX-DAG: vpaddd %ymm1, %ymm0, %ymm0 [all …]
|
/external/boringssl/linux-x86_64/crypto/cipher_extra/ |
D | chacha20_poly1305_x86_64.S | 4119 vpaddd .avx2_init(%rip),%ymm12,%ymm12 4130 vpaddd %ymm4,%ymm0,%ymm0 4133 vpaddd %ymm12,%ymm8,%ymm8 4138 vpaddd %ymm4,%ymm0,%ymm0 4141 vpaddd %ymm12,%ymm8,%ymm8 4149 vpaddd %ymm4,%ymm0,%ymm0 4152 vpaddd %ymm12,%ymm8,%ymm8 4157 vpaddd %ymm4,%ymm0,%ymm0 4160 vpaddd %ymm12,%ymm8,%ymm8 4171 vpaddd .chacha20_consts(%rip),%ymm0,%ymm0 [all …]
|
D | aes128gcmsiv-x86_64.S | 634 vpaddd %xmm13,%xmm9,%xmm10 635 vpaddd %xmm13,%xmm10,%xmm11 636 vpaddd %xmm13,%xmm11,%xmm12 740 vpaddd one(%rip),%xmm15,%xmm1 741 vpaddd two(%rip),%xmm15,%xmm2 742 vpaddd three(%rip),%xmm15,%xmm3 770 vpaddd %xmm4,%xmm0,%xmm0 777 vpaddd %xmm4,%xmm1,%xmm1 784 vpaddd %xmm4,%xmm2,%xmm2 791 vpaddd %xmm4,%xmm3,%xmm3 [all …]
|
/external/boringssl/mac-x86_64/crypto/cipher_extra/ |
D | chacha20_poly1305_x86_64.S | 4118 vpaddd .avx2_init(%rip),%ymm12,%ymm12 4129 vpaddd %ymm4,%ymm0,%ymm0 4132 vpaddd %ymm12,%ymm8,%ymm8 4137 vpaddd %ymm4,%ymm0,%ymm0 4140 vpaddd %ymm12,%ymm8,%ymm8 4148 vpaddd %ymm4,%ymm0,%ymm0 4151 vpaddd %ymm12,%ymm8,%ymm8 4156 vpaddd %ymm4,%ymm0,%ymm0 4159 vpaddd %ymm12,%ymm8,%ymm8 4170 vpaddd .chacha20_consts(%rip),%ymm0,%ymm0 [all …]
|
D | aes128gcmsiv-x86_64.S | 634 vpaddd %xmm13,%xmm9,%xmm10 635 vpaddd %xmm13,%xmm10,%xmm11 636 vpaddd %xmm13,%xmm11,%xmm12 738 vpaddd one(%rip),%xmm15,%xmm1 739 vpaddd two(%rip),%xmm15,%xmm2 740 vpaddd three(%rip),%xmm15,%xmm3 768 vpaddd %xmm4,%xmm0,%xmm0 775 vpaddd %xmm4,%xmm1,%xmm1 782 vpaddd %xmm4,%xmm2,%xmm2 789 vpaddd %xmm4,%xmm3,%xmm3 [all …]
|
/external/libyuv/files/source/ |
D | compare_win.cc | 86 vpaddd ymm0, ymm0, ymm1 in SumSquareError_AVX2() 87 vpaddd ymm0, ymm0, ymm2 in SumSquareError_AVX2() 92 vpaddd ymm0, ymm0, ymm1 in SumSquareError_AVX2() 94 vpaddd ymm0, ymm0, ymm1 in SumSquareError_AVX2() 96 vpaddd ymm0, ymm0, ymm1 in SumSquareError_AVX2() 199 vpaddd xmm3, xmm3, xmm4 // add 16 results in HashDjb2_AVX2() 200 vpaddd xmm1, xmm1, xmm2 in HashDjb2_AVX2() 201 vpaddd xmm1, xmm1, xmm3 in HashDjb2_AVX2() 203 vpaddd xmm1, xmm1,xmm2 in HashDjb2_AVX2() 205 vpaddd xmm1, xmm1, xmm2 in HashDjb2_AVX2() [all …]
|
/external/libvpx/libvpx/third_party/libyuv/source/ |
D | compare_win.cc | 105 vpaddd ymm0, ymm0, ymm1 in SumSquareError_AVX2() 106 vpaddd ymm0, ymm0, ymm2 in SumSquareError_AVX2() 111 vpaddd ymm0, ymm0, ymm1 in SumSquareError_AVX2() 113 vpaddd ymm0, ymm0, ymm1 in SumSquareError_AVX2() 115 vpaddd ymm0, ymm0, ymm1 in SumSquareError_AVX2() 218 vpaddd xmm3, xmm3, xmm4 // add 16 results in HashDjb2_AVX2() 219 vpaddd xmm1, xmm1, xmm2 in HashDjb2_AVX2() 220 vpaddd xmm1, xmm1, xmm3 in HashDjb2_AVX2() 222 vpaddd xmm1, xmm1,xmm2 in HashDjb2_AVX2() 224 vpaddd xmm1, xmm1, xmm2 in HashDjb2_AVX2() [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/AArch64/ |
D | arm64-addp.ll | 9 %vpaddd.i = fadd double %lane0.i, %lane1.i 10 ret double %vpaddd.i 20 %vpaddd.i = add i64 %lane0.i, %lane1.i 21 ret i64 %vpaddd.i 30 %vpaddd.i = fadd float %lane0.i, %lane1.i 31 ret float %vpaddd.i
|
/external/llvm/test/CodeGen/AArch64/ |
D | arm64-addp.ll | 9 %vpaddd.i = fadd double %lane0.i, %lane1.i 10 ret double %vpaddd.i 20 %vpaddd.i = add i64 %lane0.i, %lane1.i 21 ret i64 %vpaddd.i 30 %vpaddd.i = fadd float %lane0.i, %lane1.i 31 ret float %vpaddd.i
|
/external/boringssl/linux-x86_64/crypto/chacha/ |
D | chacha-x86_64.S | 1084 vpaddd .Lincy(%rip),%ymm4,%ymm4 1111 vpaddd .Leight(%rip),%ymm4,%ymm4 1123 vpaddd %ymm0,%ymm8,%ymm8 1126 vpaddd %ymm1,%ymm9,%ymm9 1129 vpaddd %ymm4,%ymm12,%ymm12 1135 vpaddd %ymm5,%ymm13,%ymm13 1140 vpaddd %ymm0,%ymm8,%ymm8 1143 vpaddd %ymm1,%ymm9,%ymm9 1146 vpaddd %ymm4,%ymm12,%ymm12 1152 vpaddd %ymm5,%ymm13,%ymm13 [all …]
|
/external/boringssl/mac-x86_64/crypto/chacha/ |
D | chacha-x86_64.S | 1077 vpaddd L$incy(%rip),%ymm4,%ymm4 1104 vpaddd L$eight(%rip),%ymm4,%ymm4 1116 vpaddd %ymm0,%ymm8,%ymm8 1119 vpaddd %ymm1,%ymm9,%ymm9 1122 vpaddd %ymm4,%ymm12,%ymm12 1128 vpaddd %ymm5,%ymm13,%ymm13 1133 vpaddd %ymm0,%ymm8,%ymm8 1136 vpaddd %ymm1,%ymm9,%ymm9 1139 vpaddd %ymm4,%ymm12,%ymm12 1145 vpaddd %ymm5,%ymm13,%ymm13 [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/X86/GlobalISel/ |
D | add-vec.ll | 28 ; ALL-NEXT: vpaddd %xmm1, %xmm0, %xmm0 92 ; SKX-NEXT: vpaddd %ymm1, %ymm0, %ymm0 97 ; AVX2-NEXT: vpaddd %ymm1, %ymm0, %ymm0 104 ; AVX1-NEXT: vpaddd %xmm3, %xmm2, %xmm2 105 ; AVX1-NEXT: vpaddd %xmm1, %xmm0, %xmm0 196 ; SKX-NEXT: vpaddd %zmm1, %zmm0, %zmm0 201 ; AVX2-NEXT: vpaddd %ymm2, %ymm0, %ymm0 202 ; AVX2-NEXT: vpaddd %ymm3, %ymm1, %ymm1 210 ; AVX1-NEXT: vpaddd %xmm6, %xmm4, %xmm4 212 ; AVX1-NEXT: vpaddd %xmm6, %xmm5, %xmm5 [all …]
|
/external/boringssl/win-x86_64/crypto/chacha/ |
D | chacha-x86_64.asm | 1166 vpaddd ymm4,ymm4,YMMWORD[$L$incy] 1193 vpaddd ymm4,ymm4,YMMWORD[$L$eight] 1205 vpaddd ymm8,ymm8,ymm0 1208 vpaddd ymm9,ymm9,ymm1 1211 vpaddd ymm12,ymm12,ymm4 1217 vpaddd ymm13,ymm13,ymm5 1222 vpaddd ymm8,ymm8,ymm0 1225 vpaddd ymm9,ymm9,ymm1 1228 vpaddd ymm12,ymm12,ymm4 1234 vpaddd ymm13,ymm13,ymm5 [all …]
|
/external/boringssl/win-x86_64/crypto/cipher_extra/ |
D | aes128gcmsiv-x86_64.asm | 715 vpaddd xmm10,xmm9,xmm13 716 vpaddd xmm11,xmm10,xmm13 717 vpaddd xmm12,xmm11,xmm13 833 vpaddd xmm1,xmm15,XMMWORD[one] 834 vpaddd xmm2,xmm15,XMMWORD[two] 835 vpaddd xmm3,xmm15,XMMWORD[three] 863 vpaddd xmm0,xmm0,xmm4 870 vpaddd xmm1,xmm1,xmm4 877 vpaddd xmm2,xmm2,xmm4 884 vpaddd xmm3,xmm3,xmm4 [all …]
|
/external/libjpeg-turbo/simd/x86_64/ |
D | jccolext-avx2.asm | 394 vpaddd ymm7, ymm7, ymm1 395 vpaddd ymm4, ymm4, ymm6 396 vpaddd ymm7, ymm7, ymm5 397 vpaddd ymm4, ymm4, ymm5 426 vpaddd ymm5, ymm5, ymm0 427 vpaddd ymm4, ymm4, ymm6 428 vpaddd ymm5, ymm5, ymm1 429 vpaddd ymm4, ymm4, ymm1 454 vpaddd ymm0, ymm0, YMMWORD [wk(4)] 455 vpaddd ymm4, ymm4, YMMWORD [wk(5)] [all …]
|