/external/libaom/libaom/aom_dsp/x86/ |
D | common_avx2.h | 57 __m256i tr1_0 = _mm256_unpacklo_epi32(tr0_0, tr0_2); in mm256_transpose_16x16() 59 __m256i tr1_2 = _mm256_unpacklo_epi32(tr0_1, tr0_3); in mm256_transpose_16x16() 61 __m256i tr1_4 = _mm256_unpacklo_epi32(tr0_4, tr0_6); in mm256_transpose_16x16() 63 __m256i tr1_6 = _mm256_unpacklo_epi32(tr0_5, tr0_7); in mm256_transpose_16x16() 66 __m256i tr1_8 = _mm256_unpacklo_epi32(tr0_8, tr0_a); in mm256_transpose_16x16() 68 __m256i tr1_a = _mm256_unpacklo_epi32(tr0_9, tr0_b); in mm256_transpose_16x16() 70 __m256i tr1_c = _mm256_unpacklo_epi32(tr0_c, tr0_e); in mm256_transpose_16x16() 72 __m256i tr1_e = _mm256_unpacklo_epi32(tr0_d, tr0_f); in mm256_transpose_16x16()
|
D | txfm_common_avx2.h | 142 b[i / 2 + 0] = _mm256_unpacklo_epi32(a[i], a[i + 1]); in transpose_16bit_16x16_avx2() 182 const __m256i b0 = _mm256_unpacklo_epi32(a0, a1); in transpose_16bit_16x8_avx2() 183 const __m256i b1 = _mm256_unpacklo_epi32(a2, a3); in transpose_16bit_16x8_avx2() 184 const __m256i b2 = _mm256_unpacklo_epi32(a4, a5); in transpose_16bit_16x8_avx2() 185 const __m256i b3 = _mm256_unpacklo_epi32(a6, a7); in transpose_16bit_16x8_avx2()
|
D | avg_intrin_avx2.c | 66 b0 = _mm256_unpacklo_epi32(a0, a1); in hadamard_col8x2_avx2() 67 b1 = _mm256_unpacklo_epi32(a4, a5); in hadamard_col8x2_avx2() 70 b4 = _mm256_unpacklo_epi32(a2, a3); in hadamard_col8x2_avx2() 71 b5 = _mm256_unpacklo_epi32(a6, a7); in hadamard_col8x2_avx2()
|
D | highbd_variance_avx2.c | 45 __m256i v_d_l = _mm256_unpacklo_epi32(v_sum0, v_sse_d); in aom_highbd_calc8x8var_avx2() 73 __m256i v_d_l = _mm256_unpacklo_epi32(v_sum0, v_sse_d); in aom_highbd_calc16x16var_avx2()
|
D | intrapred_avx2.c | 119 ww0 = _mm256_unpacklo_epi32(w0, w1); // 00 10 20 30 01 11 21 31 in highbd_transpose4x16_avx2() 120 ww1 = _mm256_unpacklo_epi32(w2, w3); // 40 50 60 70 41 51 61 71 in highbd_transpose4x16_avx2() 140 ww0 = _mm256_unpacklo_epi32(w0, w1); // 00 10 20 30 01 11 21 31 in highbd_transpose8x16_16x8_avx2() 141 ww1 = _mm256_unpacklo_epi32(w2, w3); // 40 50 60 70 41 51 61 71 in highbd_transpose8x16_16x8_avx2() 157 ww0 = _mm256_unpacklo_epi32(w0, w1); // 04 14 24 34 05 15 25 35 in highbd_transpose8x16_16x8_avx2() 158 ww1 = _mm256_unpacklo_epi32(w2, w3); // 44 54 64 74 45 55 65 75 in highbd_transpose8x16_16x8_avx2() 178 ww0 = _mm256_unpacklo_epi32(w0, w1); // in highbd_transpose16x16_avx2() 179 ww1 = _mm256_unpacklo_epi32(w2, w3); // in highbd_transpose16x16_avx2() 195 ww0 = _mm256_unpacklo_epi32(w0, w1); // in highbd_transpose16x16_avx2() 196 ww1 = _mm256_unpacklo_epi32(w2, w3); // in highbd_transpose16x16_avx2() [all …]
|
D | sse_avx2.c | 38 const __m256i sum0_4x64 = _mm256_unpacklo_epi32(*sum_all, zero); in summary_all_avx2()
|
/external/libvpx/libvpx/vp9/encoder/x86/ |
D | vp9_error_avx2.c | 47 sse_256 = _mm256_unpacklo_epi32(dqcoeff_256, zero); in vp9_block_error_avx2() 48 ssz_256 = _mm256_unpacklo_epi32(coeff_256, zero); in vp9_block_error_avx2() 76 exp_dqcoeff_lo = _mm256_unpacklo_epi32(dqcoeff_0, zero); in vp9_block_error_avx2() 79 exp_coeff_lo = _mm256_unpacklo_epi32(coeff_0, zero); in vp9_block_error_avx2() 131 sse_256 = _mm256_unpacklo_epi32(error, zero); in vp9_block_error_fp_avx2() 140 const __m256i exp_error_lo = _mm256_unpacklo_epi32(error, zero); in vp9_block_error_fp_avx2()
|
/external/libvpx/libvpx/vpx_dsp/x86/ |
D | fwd_dct32x32_impl_avx2.h | 1553 u[0] = _mm256_unpacklo_epi32(lstep3[12], lstep3[10]); in FDCT32x32_2D_AVX2() 1555 u[2] = _mm256_unpacklo_epi32(lstep3[13], lstep3[11]); in FDCT32x32_2D_AVX2() 1592 u[0] = _mm256_unpacklo_epi32(lstep3[36], lstep3[58]); in FDCT32x32_2D_AVX2() 1594 u[2] = _mm256_unpacklo_epi32(lstep3[37], lstep3[59]); in FDCT32x32_2D_AVX2() 1596 u[4] = _mm256_unpacklo_epi32(lstep3[38], lstep3[56]); in FDCT32x32_2D_AVX2() 1598 u[6] = _mm256_unpacklo_epi32(lstep3[39], lstep3[57]); in FDCT32x32_2D_AVX2() 1600 u[8] = _mm256_unpacklo_epi32(lstep3[40], lstep3[54]); in FDCT32x32_2D_AVX2() 1602 u[10] = _mm256_unpacklo_epi32(lstep3[41], lstep3[55]); in FDCT32x32_2D_AVX2() 1604 u[12] = _mm256_unpacklo_epi32(lstep3[42], lstep3[52]); in FDCT32x32_2D_AVX2() 1606 u[14] = _mm256_unpacklo_epi32(lstep3[43], lstep3[53]); in FDCT32x32_2D_AVX2() [all …]
|
D | avg_intrin_avx2.c | 57 a0 = _mm256_unpacklo_epi32(b0, b1); in highbd_hadamard_col8_avx2() 58 a1 = _mm256_unpacklo_epi32(b2, b3); in highbd_hadamard_col8_avx2() 61 a4 = _mm256_unpacklo_epi32(b4, b5); in highbd_hadamard_col8_avx2() 62 a5 = _mm256_unpacklo_epi32(b6, b7); in highbd_hadamard_col8_avx2() 268 b0 = _mm256_unpacklo_epi32(a0, a1); in hadamard_col8x2_avx2() 269 b1 = _mm256_unpacklo_epi32(a4, a5); in hadamard_col8x2_avx2() 272 b4 = _mm256_unpacklo_epi32(a2, a3); in hadamard_col8x2_avx2() 273 b5 = _mm256_unpacklo_epi32(a6, a7); in hadamard_col8x2_avx2()
|
/external/libaom/libaom/av1/encoder/x86/ |
D | error_intrin_avx2.c | 57 exp_dqcoeff_lo = _mm256_unpacklo_epi32(dqcoeff_reg, zero_reg); in av1_block_error_avx2() 60 exp_coeff_lo = _mm256_unpacklo_epi32(coeff_reg, zero_reg); in av1_block_error_avx2()
|
D | wedge_utils_avx2.c | 135 v_acc0_d = _mm256_add_epi64(_mm256_unpacklo_epi32(v_acc0_d, v_sign_d), in av1_wedge_sign_from_residuals_avx2()
|
D | av1_fwd_txfm2d_avx2.c | 1435 __m256i temp0 = _mm256_unpacklo_epi32(inputA[0], inputA[2]); in transpose_32_8x8_avx2() 1437 __m256i temp2 = _mm256_unpacklo_epi32(inputA[1], inputA[3]); in transpose_32_8x8_avx2() 1439 __m256i temp4 = _mm256_unpacklo_epi32(inputA[4], inputA[6]); in transpose_32_8x8_avx2() 1441 __m256i temp6 = _mm256_unpacklo_epi32(inputA[5], inputA[7]); in transpose_32_8x8_avx2() 1444 __m256i t0 = _mm256_unpacklo_epi32(temp0, temp2); in transpose_32_8x8_avx2() 1446 __m256i t2 = _mm256_unpacklo_epi32(temp1, temp3); in transpose_32_8x8_avx2() 1448 __m256i t4 = _mm256_unpacklo_epi32(temp4, temp6); in transpose_32_8x8_avx2() 1450 __m256i t6 = _mm256_unpacklo_epi32(temp5, temp7); in transpose_32_8x8_avx2()
|
D | highbd_fwd_txfm_avx2.c | 151 u0 = _mm256_unpacklo_epi32(in[0 * instride], in[1 * instride]); in av1_fwd_txfm_transpose_8x8_avx2() 154 u2 = _mm256_unpacklo_epi32(in[2 * instride], in[3 * instride]); in av1_fwd_txfm_transpose_8x8_avx2() 157 u4 = _mm256_unpacklo_epi32(in[4 * instride], in[5 * instride]); in av1_fwd_txfm_transpose_8x8_avx2() 160 u6 = _mm256_unpacklo_epi32(in[6 * instride], in[7 * instride]); in av1_fwd_txfm_transpose_8x8_avx2()
|
/external/eigen/Eigen/src/Core/arch/CUDA/ |
D | PacketMathHalf.h | 568 __m256i abcd_03 = _mm256_unpacklo_epi32(ab_07, cd_07); 570 __m256i efgh_03 = _mm256_unpacklo_epi32(ef_07, gh_07); 572 __m256i ijkl_03 = _mm256_unpacklo_epi32(ij_07, kl_07); 574 __m256i mnop_03 = _mm256_unpacklo_epi32(mn_07, op_07); 577 __m256i abcd_8b = _mm256_unpacklo_epi32(ab_8f, cd_8f); 579 __m256i efgh_8b = _mm256_unpacklo_epi32(ef_8f, gh_8f); 581 __m256i ijkl_8b = _mm256_unpacklo_epi32(ij_8f, kl_8f); 583 __m256i mnop_8b = _mm256_unpacklo_epi32(mn_8f, op_8f);
|
/external/libaom/libaom/av1/common/x86/ |
D | highbd_wiener_convolve_avx2.c | 225 const __m256i res_lo = _mm256_unpacklo_epi32(res_even, res_odd); in av1_highbd_wiener_convolve_add_src_avx2()
|
D | wiener_convolve_avx2.c | 231 const __m256i res_lo = _mm256_unpacklo_epi32(res_even, res_odd); in av1_wiener_convolve_add_src_avx2()
|
D | highbd_inv_txfm_avx2.c | 123 u0 = _mm256_unpacklo_epi32(in[0], in[1]); in transpose_8x8_avx2() 126 u2 = _mm256_unpacklo_epi32(in[2], in[3]); in transpose_8x8_avx2() 129 u4 = _mm256_unpacklo_epi32(in[4], in[5]); in transpose_8x8_avx2() 132 u6 = _mm256_unpacklo_epi32(in[6], in[7]); in transpose_8x8_avx2() 160 u0 = _mm256_unpacklo_epi32(in[7], in[6]); in transpose_8x8_flip_avx2() 163 u2 = _mm256_unpacklo_epi32(in[5], in[4]); in transpose_8x8_flip_avx2() 166 u4 = _mm256_unpacklo_epi32(in[3], in[2]); in transpose_8x8_flip_avx2() 169 u6 = _mm256_unpacklo_epi32(in[1], in[0]); in transpose_8x8_flip_avx2()
|
D | highbd_jnt_convolve_avx2.c | 543 __m256i res1 = _mm256_unpacklo_epi32(res_even, res_odd); in av1_highbd_dist_wtd_convolve_x_avx2()
|
/external/libaom/libaom/aom_dsp/simd/ |
D | v256_intrinsics_x86.h | 166 return _mm256_unpacklo_epi32( in v256_ziplo_32() 740 _mm256_unpacklo_epi32(d, v256_zero()))); in v256_ssd_s16()
|
/external/clang/test/CodeGen/ |
D | avx2-builtins.c | 1213 return _mm256_unpacklo_epi32(a, b); in test_mm256_unpacklo_epi32()
|
/external/clang/lib/Headers/ |
D | avx2intrin.h | 815 _mm256_unpacklo_epi32(__m256i __a, __m256i __b) in _mm256_unpacklo_epi32() function
|
D | avx512vlintrin.h | 6943 (__v8si)_mm256_unpacklo_epi32(__A, __B), in _mm256_mask_unpacklo_epi32() 6951 (__v8si)_mm256_unpacklo_epi32(__A, __B), in _mm256_maskz_unpacklo_epi32()
|