Home
last modified time | relevance | path

Searched refs:_mm256_and_si256 (Results 1 – 15 of 15) sorted by relevance

/external/libaom/libaom/aom_dsp/x86/
Dhighbd_quantize_intrin_avx2.c59 prod_lo = _mm256_and_si256(prod_lo, mask); in mm256_mul_shift_epi32()
86 q = _mm256_and_si256(q, flag2); in quantize()
87 dq = _mm256_and_si256(dq, flag2); in quantize()
103 cur_eob = _mm256_and_si256(cur_eob, nz); in quantize()
Dintrapred_avx2.c680 ptl = _mm256_and_si256(mask2, *topleft); in paeth_pred()
683 pt = _mm256_and_si256(mask1, pt); in paeth_pred()
1040 _mm256_and_si256( in highbd_dr_prediction_z1_4xN_internal_avx2()
1047 _mm256_and_si256(_mm256_set1_epi32(x), _mm256_set1_epi32(0x3f)), 1); in highbd_dr_prediction_z1_4xN_internal_avx2()
1133 _mm256_and_si256( in highbd_dr_prediction_z1_8xN_internal_avx2()
1141 _mm256_and_si256(_mm256_set1_epi32(x), _mm256_set1_epi32(0x3f)), 1); in highbd_dr_prediction_z1_8xN_internal_avx2()
1211 _mm256_and_si256(_mm256_set1_epi32(x), _mm256_set1_epi32(0x3f)), 1); in highbd_dr_prediction_z1_16xN_internal_avx2()
1305 _mm256_and_si256(_mm256_set1_epi32(x), _mm256_set1_epi32(0x3f)), 1); in highbd_dr_prediction_z1_32xN_internal_avx2()
1421 _mm256_and_si256(_mm256_set1_epi32(x), _mm256_set1_epi32(0x3f)), 1); in highbd_dr_prediction_z1_64xN_avx2()
1851 _mm256_and_si256( in highbd_dr_prediction_32bit_z2_Nx8_avx2()
[all …]
Dsum_squares_avx2.c50 _mm256_add_epi64(v_acc_q, _mm256_and_si256(v_acc_d, v_zext_mask_q)); in aom_sum_squares_2d_i16_nxn_avx2()
Dblend_a64_mask_avx2.c453 const __m256i v_rvsal_w = _mm256_and_si256(v_rvsl_b, v_zmask_b); in blend_a64_mask_sx_sy_w16_avx2()
455 _mm256_and_si256(_mm256_srli_si256(v_rvsl_b, 1), v_zmask_b); in blend_a64_mask_sx_sy_w16_avx2()
488 const __m256i v_rvsal_w = _mm256_and_si256(v_rvsl_b, v_zmask_b); in blend_a64_mask_sx_sy_w32n_avx2()
489 const __m256i v_rvsah_w = _mm256_and_si256(v_rvsh_b, v_zmask_b); in blend_a64_mask_sx_sy_w32n_avx2()
491 _mm256_and_si256(_mm256_srli_si256(v_rvsl_b, 1), v_zmask_b); in blend_a64_mask_sx_sy_w32n_avx2()
493 _mm256_and_si256(_mm256_srli_si256(v_rvsh_b, 1), v_zmask_b); in blend_a64_mask_sx_sy_w32n_avx2()
591 const __m256i v_m0_w = _mm256_and_si256(v_al_b, v_zmask_b); in blend_a64_mask_sx_w16_avx2()
Dsad_highbd_avx2.c718 u0 = _mm256_and_si256(u0, mask); in get_4d_sad_from_mm256_epi32()
719 u1 = _mm256_and_si256(u1, mask); in get_4d_sad_from_mm256_epi32()
720 u2 = _mm256_and_si256(u2, mask); in get_4d_sad_from_mm256_epi32()
721 u3 = _mm256_and_si256(u3, mask); in get_4d_sad_from_mm256_epi32()
/external/libaom/libaom/aom_dsp/simd/
Dv256_intrinsics_x86.h495 SIMD_INLINE v256 v256_and(v256 a, v256 b) { return _mm256_and_si256(a, b); } in v256_and()
531 _mm256_and_si256(_mm256_xor_si256(a, b), v256_dup_8(1))); in v256_rdavg_u8()
537 _mm256_and_si256(_mm256_xor_si256(a, b), v256_dup_16(1))); in v256_rdavg_u16()
601 return _mm256_and_si256(_mm256_set1_epi8((uint8_t)(0xff << c)), in v256_shl_8()
606 return _mm256_and_si256(_mm256_set1_epi8(0xff >> c), in v256_shr_u8()
682 _mm256_and_si256(_mm256_set1_epi8((uint8_t)(0xff << (c))), \
685 _mm256_and_si256(_mm256_set1_epi8(0xff >> (c)), _mm256_srli_epi16(a, c))
/external/libaom/libaom/av1/encoder/x86/
Dav1_quantize_avx2.c127 cur_eob = _mm256_and_si256(cur_eob, nzero_coeff); in quantize()
203 cur_eob = _mm256_and_si256(cur_eob, nzero_coeff); in quantize_32x32()
282 cur_eob = _mm256_and_si256(cur_eob, nzero_coeff); in quantize_64x64()
Dav1_highbd_quantize_avx2.c86 cur_eob = _mm256_and_si256(cur_eob, nz); in quantize()
Dwedge_utils_avx2.c64 _mm256_and_si256(v_sq0_d, v_zext_q), _mm256_srli_epi64(v_sq0_d, 32)); in av1_wedge_sse_from_residuals_avx2()
/external/libaom/libaom/av1/common/x86/
Dselfguided_avx2.c255 sum1 = _mm256_and_si256(mask[idx], sum1); in calc_ab()
256 sum2 = _mm256_and_si256(mask[idx], sum2); in calc_ab()
392 sum1 = _mm256_and_si256(mask[idx], sum1); in calc_ab_fast()
393 sum2 = _mm256_and_si256(mask[idx], sum2); in calc_ab_fast()
Dhighbd_inv_txfm_avx2.c42 mask = _mm256_and_si256(mask, max); in highbd_clamp_epi16_avx2()
45 clamped = _mm256_and_si256(clamped, mask); in highbd_clamp_epi16_avx2()
/external/libvpx/libvpx/vp9/encoder/x86/
Dvp9_quantize_avx2.c46 return _mm256_and_si256(iscan_plus_one, nzero_coeff0); in scan_eob_256()
/external/clang/test/CodeGen/
Davx2-builtins.c92 return _mm256_and_si256(a, b); in test_mm256_and_si256()
/external/clang/lib/Headers/
Davx2intrin.h134 _mm256_and_si256(__m256i __a, __m256i __b) in _mm256_and_si256() function
Davx512vlintrin.h896 (__v8si)_mm256_and_si256(__A, __B), in _mm256_mask_and_epi32()
1010 (__v4di)_mm256_and_si256(__A, __B), in _mm256_mask_and_epi64()