/external/oboe/samples/RhythmGame/third_party/glm/simd/ |
D | integer.h | 28 Reg1 = _mm_and_si128(Reg1, Mask4); in glm_i128_interleave() 34 Reg1 = _mm_and_si128(Reg1, Mask3); in glm_i128_interleave() 40 Reg1 = _mm_and_si128(Reg1, Mask2); in glm_i128_interleave() 46 Reg1 = _mm_and_si128(Reg1, Mask1); in glm_i128_interleave() 52 Reg1 = _mm_and_si128(Reg1, Mask0); in glm_i128_interleave() 81 Reg1 = _mm_and_si128(Reg1, Mask4); in glm_i128_interleave2() 87 Reg1 = _mm_and_si128(Reg1, Mask3); in glm_i128_interleave2() 93 Reg1 = _mm_and_si128(Reg1, Mask2); in glm_i128_interleave2() 99 Reg1 = _mm_and_si128(Reg1, Mask1); in glm_i128_interleave2() 105 Reg1 = _mm_and_si128(Reg1, Mask0); in glm_i128_interleave2()
|
/external/XNNPACK/src/f32-argmaxpool/ |
D | 9p8x-sse2-c4.c | 80 vidx = _mm_or_si128(_mm_andnot_si128(vm1, vidx), _mm_and_si128(vm1, _mm_set1_epi32(1))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() 84 vidx = _mm_or_si128(_mm_andnot_si128(vm2, vidx), _mm_and_si128(vm2, _mm_set1_epi32(2))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() 88 vidx = _mm_or_si128(_mm_andnot_si128(vm3, vidx), _mm_and_si128(vm3, _mm_set1_epi32(3))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() 92 vidx = _mm_or_si128(_mm_andnot_si128(vm4, vidx), _mm_and_si128(vm4, _mm_set1_epi32(4))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() 96 vidx = _mm_or_si128(_mm_andnot_si128(vm5, vidx), _mm_and_si128(vm5, _mm_set1_epi32(5))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() 100 vidx = _mm_or_si128(_mm_andnot_si128(vm6, vidx), _mm_and_si128(vm6, _mm_set1_epi32(6))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() 104 vidx = _mm_or_si128(_mm_andnot_si128(vm7, vidx), _mm_and_si128(vm7, _mm_set1_epi32(7))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() 108 vidx = _mm_or_si128(_mm_andnot_si128(vm8, vidx), _mm_and_si128(vm8, _mm_set1_epi32(8))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() 165 vidx = _mm_or_si128(_mm_andnot_si128(vm0, vidx), _mm_and_si128(vm0, vidx0)); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() 170 vidx = _mm_or_si128(_mm_andnot_si128(vm1, vidx), _mm_and_si128(vm1, vidx1)); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() [all …]
|
D | 9x-sse2-c4.c | 99 vidx = _mm_or_si128(_mm_andnot_si128(vm1, vidx), _mm_and_si128(vm1, _mm_set1_epi32(1))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() 103 vidx = _mm_or_si128(_mm_andnot_si128(vm2, vidx), _mm_and_si128(vm2, _mm_set1_epi32(2))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() 107 vidx = _mm_or_si128(_mm_andnot_si128(vm3, vidx), _mm_and_si128(vm3, _mm_set1_epi32(3))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() 111 vidx = _mm_or_si128(_mm_andnot_si128(vm4, vidx), _mm_and_si128(vm4, _mm_set1_epi32(4))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() 115 vidx = _mm_or_si128(_mm_andnot_si128(vm5, vidx), _mm_and_si128(vm5, _mm_set1_epi32(5))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() 119 vidx = _mm_or_si128(_mm_andnot_si128(vm6, vidx), _mm_and_si128(vm6, _mm_set1_epi32(6))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() 123 vidx = _mm_or_si128(_mm_andnot_si128(vm7, vidx), _mm_and_si128(vm7, _mm_set1_epi32(7))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() 127 vidx = _mm_or_si128(_mm_andnot_si128(vm8, vidx), _mm_and_si128(vm8, _mm_set1_epi32(8))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() 150 vidx = _mm_or_si128(_mm_andnot_si128(vm1, vidx), _mm_and_si128(vm1, _mm_set1_epi32(1))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() 154 vidx = _mm_or_si128(_mm_andnot_si128(vm2, vidx), _mm_and_si128(vm2, _mm_set1_epi32(2))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() [all …]
|
/external/libavc/common/x86/ |
D | ih264_iquant_itrans_recon_dc_ssse3.c | 153 temp4 = _mm_and_si128(temp4, sign_reg); in ih264_iquant_itrans_recon_4x4_dc_ssse3() 155 temp5 = _mm_and_si128(temp5, sign_reg); in ih264_iquant_itrans_recon_4x4_dc_ssse3() 306 temp1 = _mm_and_si128(temp1, sign_reg); in ih264_iquant_itrans_recon_8x8_dc_ssse3() 308 temp2 = _mm_and_si128(temp2, sign_reg); in ih264_iquant_itrans_recon_8x8_dc_ssse3() 310 temp3 = _mm_and_si128(temp3, sign_reg); in ih264_iquant_itrans_recon_8x8_dc_ssse3() 312 temp4 = _mm_and_si128(temp4, sign_reg); in ih264_iquant_itrans_recon_8x8_dc_ssse3() 314 temp5 = _mm_and_si128(temp5, sign_reg); in ih264_iquant_itrans_recon_8x8_dc_ssse3() 316 temp6 = _mm_and_si128(temp6, sign_reg); in ih264_iquant_itrans_recon_8x8_dc_ssse3() 318 temp7 = _mm_and_si128(temp7, sign_reg); in ih264_iquant_itrans_recon_8x8_dc_ssse3() 320 temp8 = _mm_and_si128(temp8, sign_reg); in ih264_iquant_itrans_recon_8x8_dc_ssse3() [all …]
|
D | ih264_deblk_luma_ssse3.c | 196 flag1_16x8 = _mm_and_si128(flag1_16x8, flag2_16x8); in ih264_deblk_luma_vert_bs4_ssse3() 212 flag1_16x8 = _mm_and_si128(flag1_16x8, flag2_16x8); in ih264_deblk_luma_vert_bs4_ssse3() 227 flag2_16x8 = _mm_and_si128(flag1_16x8, flag2_16x8); in ih264_deblk_luma_vert_bs4_ssse3() 240 flag3_16x8 = _mm_and_si128(flag3_16x8, flag2_16x8); in ih264_deblk_luma_vert_bs4_ssse3() 253 flag4_16x8 = _mm_and_si128(flag4_16x8, flag2_16x8); in ih264_deblk_luma_vert_bs4_ssse3() 384 p0_16x8 = _mm_and_si128(p0_16x8, in ih264_deblk_luma_vert_bs4_ssse3() 386 p0_16x8_1 = _mm_and_si128(p0_16x8_1, flag1_16x8); in ih264_deblk_luma_vert_bs4_ssse3() 388 q0_16x8 = _mm_and_si128(q0_16x8, in ih264_deblk_luma_vert_bs4_ssse3() 390 q0_16x8_1 = _mm_and_si128(q0_16x8_1, flag1_16x8); in ih264_deblk_luma_vert_bs4_ssse3() 394 p0_16x8 = _mm_and_si128(p0_16x8, in ih264_deblk_luma_vert_bs4_ssse3() [all …]
|
D | ih264_deblk_chroma_ssse3.c | 153 flag1 = _mm_and_si128(flag1, _mm_cmpgt_epi16(beta_cbcr_16x8, diff)); in ih264_deblk_chroma_vert_bs4_ssse3() 157 flag1 = _mm_and_si128(flag1, _mm_cmpgt_epi16(beta_cbcr_16x8, diff)); in ih264_deblk_chroma_vert_bs4_ssse3() 184 flag2 = _mm_and_si128(flag2, _mm_cmpgt_epi16(beta_cbcr_16x8, diff)); in ih264_deblk_chroma_vert_bs4_ssse3() 188 flag2 = _mm_and_si128(flag2, _mm_cmpgt_epi16(beta_cbcr_16x8, diff)); in ih264_deblk_chroma_vert_bs4_ssse3() 207 p0_uv_8x16_1 = _mm_and_si128(p0_uv_16x8, in ih264_deblk_chroma_vert_bs4_ssse3() 209 p0_uv_8x16_2 = _mm_and_si128(p0_uv_8x16_2, flag1); in ih264_deblk_chroma_vert_bs4_ssse3() 212 q0_uv_8x16_1 = _mm_and_si128(q0_uv_16x8, in ih264_deblk_chroma_vert_bs4_ssse3() 214 q0_uv_8x16_2 = _mm_and_si128(q0_uv_8x16_2, flag1); in ih264_deblk_chroma_vert_bs4_ssse3() 322 flag1 = _mm_and_si128(flag1, _mm_cmpgt_epi16(beta_cbcr_16x8, diff)); in ih264_deblk_chroma_horz_bs4_ssse3() 326 flag1 = _mm_and_si128(flag1, _mm_cmpgt_epi16(beta_cbcr_16x8, diff)); in ih264_deblk_chroma_horz_bs4_ssse3() [all …]
|
/external/XNNPACK/src/qs8-vadd/gen/ |
D | minmax-sse41-mul16-ld64-x32.c | 75 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32() 76 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32() 77 …vxprod89ABCDEFhi = _mm_sub_epi16(vxprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vx89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32() 78 …vyprod89ABCDEFhi = _mm_sub_epi16(vyprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vy89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32() 79 …vxprodGHIJKLMNhi = _mm_sub_epi16(vxprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vxGHIJKLMN, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32() 80 …vyprodGHIJKLMNhi = _mm_sub_epi16(vyprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vyGHIJKLMN, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32() 81 …vxprodOPQRSTUVhi = _mm_sub_epi16(vxprodOPQRSTUVhi, _mm_and_si128(_mm_srai_epi16(vxOPQRSTUV, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32() 82 …vyprodOPQRSTUVhi = _mm_sub_epi16(vyprodOPQRSTUVhi, _mm_and_si128(_mm_srai_epi16(vyOPQRSTUV, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32() 102 …const __m128i vrem0123 = _mm_add_epi32(_mm_and_si128(vacc0123, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32() 103 …const __m128i vrem4567 = _mm_add_epi32(_mm_and_si128(vacc4567, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32() [all …]
|
D | minmax-sse2-mul16-ld64-x32.c | 83 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32() 84 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32() 85 …vxprod89ABCDEFhi = _mm_sub_epi16(vxprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vx89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32() 86 …vyprod89ABCDEFhi = _mm_sub_epi16(vyprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vy89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32() 87 …vxprodGHIJKLMNhi = _mm_sub_epi16(vxprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vxGHIJKLMN, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32() 88 …vyprodGHIJKLMNhi = _mm_sub_epi16(vyprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vyGHIJKLMN, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32() 89 …vxprodOPQRSTUVhi = _mm_sub_epi16(vxprodOPQRSTUVhi, _mm_and_si128(_mm_srai_epi16(vxOPQRSTUV, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32() 90 …vyprodOPQRSTUVhi = _mm_sub_epi16(vyprodOPQRSTUVhi, _mm_and_si128(_mm_srai_epi16(vyOPQRSTUV, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32() 110 …const __m128i vrem0123 = _mm_add_epi32(_mm_and_si128(vacc0123, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32() 111 …const __m128i vrem4567 = _mm_add_epi32(_mm_and_si128(vacc4567, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32() [all …]
|
D | minmax-sse41-mul16-ld64-x24.c | 67 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x24() 68 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x24() 69 …vxprod89ABCDEFhi = _mm_sub_epi16(vxprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vx89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x24() 70 …vyprod89ABCDEFhi = _mm_sub_epi16(vyprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vy89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x24() 71 …vxprodGHIJKLMNhi = _mm_sub_epi16(vxprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vxGHIJKLMN, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x24() 72 …vyprodGHIJKLMNhi = _mm_sub_epi16(vyprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vyGHIJKLMN, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x24() 88 …const __m128i vrem0123 = _mm_add_epi32(_mm_and_si128(vacc0123, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x24() 89 …const __m128i vrem4567 = _mm_add_epi32(_mm_and_si128(vacc4567, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x24() 90 …const __m128i vrem89AB = _mm_add_epi32(_mm_and_si128(vacc89AB, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x24() 91 …const __m128i vremCDEF = _mm_add_epi32(_mm_and_si128(vaccCDEF, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x24() [all …]
|
D | minmax-sse2-mul16-ld64-x24.c | 73 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x24() 74 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x24() 75 …vxprod89ABCDEFhi = _mm_sub_epi16(vxprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vx89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x24() 76 …vyprod89ABCDEFhi = _mm_sub_epi16(vyprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vy89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x24() 77 …vxprodGHIJKLMNhi = _mm_sub_epi16(vxprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vxGHIJKLMN, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x24() 78 …vyprodGHIJKLMNhi = _mm_sub_epi16(vyprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vyGHIJKLMN, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x24() 94 …const __m128i vrem0123 = _mm_add_epi32(_mm_and_si128(vacc0123, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x24() 95 …const __m128i vrem4567 = _mm_add_epi32(_mm_and_si128(vacc4567, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x24() 96 …const __m128i vrem89AB = _mm_add_epi32(_mm_and_si128(vacc89AB, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x24() 97 …const __m128i vremCDEF = _mm_add_epi32(_mm_and_si128(vaccCDEF, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x24() [all …]
|
D | minmax-sse2-mul16-ld64-x16.c | 63 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x16() 64 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x16() 65 …vxprod89ABCDEFhi = _mm_sub_epi16(vxprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vx89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x16() 66 …vyprod89ABCDEFhi = _mm_sub_epi16(vyprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vy89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x16() 78 …const __m128i vrem0123 = _mm_add_epi32(_mm_and_si128(vacc0123, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x16() 79 …const __m128i vrem4567 = _mm_add_epi32(_mm_and_si128(vacc4567, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x16() 80 …const __m128i vrem89AB = _mm_add_epi32(_mm_and_si128(vacc89AB, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x16() 81 …const __m128i vremCDEF = _mm_add_epi32(_mm_and_si128(vaccCDEF, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x16() 120 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x16() 121 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x16() [all …]
|
D | minmax-sse41-mul16-ld64-x16.c | 59 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x16() 60 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x16() 61 …vxprod89ABCDEFhi = _mm_sub_epi16(vxprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vx89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x16() 62 …vyprod89ABCDEFhi = _mm_sub_epi16(vyprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vy89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x16() 74 …const __m128i vrem0123 = _mm_add_epi32(_mm_and_si128(vacc0123, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x16() 75 …const __m128i vrem4567 = _mm_add_epi32(_mm_and_si128(vacc4567, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x16() 76 …const __m128i vrem89AB = _mm_add_epi32(_mm_and_si128(vacc89AB, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x16() 77 …const __m128i vremCDEF = _mm_add_epi32(_mm_and_si128(vaccCDEF, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x16() 114 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x16() 115 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x16() [all …]
|
/external/XNNPACK/src/qs8-vaddc/gen/ |
D | minmax-sse2-mul16-ld64-x32.c | 62 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x32() 63 …vxprod89ABCDEFhi = _mm_sub_epi16(vxprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vx89ABCDEF, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x32() 64 …vxprodGHIJKLMNhi = _mm_sub_epi16(vxprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vxGHIJKLMN, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x32() 65 …vxprodOPQRSTUVhi = _mm_sub_epi16(vxprodOPQRSTUVhi, _mm_and_si128(_mm_srai_epi16(vxOPQRSTUV, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x32() 76 …const __m128i vrem0123 = _mm_add_epi32(_mm_and_si128(vacc0123, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x32() 77 …const __m128i vrem4567 = _mm_add_epi32(_mm_and_si128(vacc4567, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x32() 78 …const __m128i vrem89AB = _mm_add_epi32(_mm_and_si128(vacc89AB, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x32() 79 …const __m128i vremCDEF = _mm_add_epi32(_mm_and_si128(vaccCDEF, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x32() 80 …const __m128i vremGHIJ = _mm_add_epi32(_mm_and_si128(vaccGHIJ, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x32() 81 …const __m128i vremKLMN = _mm_add_epi32(_mm_and_si128(vaccKLMN, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x32() [all …]
|
D | minmax-sse41-mul16-ld64-x32.c | 58 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x32() 59 …vxprod89ABCDEFhi = _mm_sub_epi16(vxprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vx89ABCDEF, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x32() 60 …vxprodGHIJKLMNhi = _mm_sub_epi16(vxprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vxGHIJKLMN, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x32() 61 …vxprodOPQRSTUVhi = _mm_sub_epi16(vxprodOPQRSTUVhi, _mm_and_si128(_mm_srai_epi16(vxOPQRSTUV, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x32() 72 …const __m128i vrem0123 = _mm_add_epi32(_mm_and_si128(vacc0123, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x32() 73 …const __m128i vrem4567 = _mm_add_epi32(_mm_and_si128(vacc4567, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x32() 74 …const __m128i vrem89AB = _mm_add_epi32(_mm_and_si128(vacc89AB, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x32() 75 …const __m128i vremCDEF = _mm_add_epi32(_mm_and_si128(vaccCDEF, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x32() 76 …const __m128i vremGHIJ = _mm_add_epi32(_mm_and_si128(vaccGHIJ, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x32() 77 …const __m128i vremKLMN = _mm_add_epi32(_mm_and_si128(vaccKLMN, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x32() [all …]
|
D | minmax-sse41-mul16-ld64-x24.c | 54 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x24() 55 …vxprod89ABCDEFhi = _mm_sub_epi16(vxprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vx89ABCDEF, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x24() 56 …vxprodGHIJKLMNhi = _mm_sub_epi16(vxprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vxGHIJKLMN, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x24() 65 …const __m128i vrem0123 = _mm_add_epi32(_mm_and_si128(vacc0123, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x24() 66 …const __m128i vrem4567 = _mm_add_epi32(_mm_and_si128(vacc4567, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x24() 67 …const __m128i vrem89AB = _mm_add_epi32(_mm_and_si128(vacc89AB, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x24() 68 …const __m128i vremCDEF = _mm_add_epi32(_mm_and_si128(vaccCDEF, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x24() 69 …const __m128i vremGHIJ = _mm_add_epi32(_mm_and_si128(vaccGHIJ, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x24() 70 …const __m128i vremKLMN = _mm_add_epi32(_mm_and_si128(vaccKLMN, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x24() 109 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x24() [all …]
|
D | minmax-sse2-mul16-ld64-x24.c | 57 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x24() 58 …vxprod89ABCDEFhi = _mm_sub_epi16(vxprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vx89ABCDEF, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x24() 59 …vxprodGHIJKLMNhi = _mm_sub_epi16(vxprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vxGHIJKLMN, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x24() 68 …const __m128i vrem0123 = _mm_add_epi32(_mm_and_si128(vacc0123, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x24() 69 …const __m128i vrem4567 = _mm_add_epi32(_mm_and_si128(vacc4567, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x24() 70 …const __m128i vrem89AB = _mm_add_epi32(_mm_and_si128(vacc89AB, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x24() 71 …const __m128i vremCDEF = _mm_add_epi32(_mm_and_si128(vaccCDEF, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x24() 72 …const __m128i vremGHIJ = _mm_add_epi32(_mm_and_si128(vaccGHIJ, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x24() 73 …const __m128i vremKLMN = _mm_add_epi32(_mm_and_si128(vaccKLMN, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x24() 113 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x24() [all …]
|
D | minmax-sse41-mul16-ld64-x16.c | 50 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x16() 51 …vxprod89ABCDEFhi = _mm_sub_epi16(vxprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vx89ABCDEF, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x16() 58 …const __m128i vrem0123 = _mm_add_epi32(_mm_and_si128(vacc0123, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x16() 59 …const __m128i vrem4567 = _mm_add_epi32(_mm_and_si128(vacc4567, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x16() 60 …const __m128i vrem89AB = _mm_add_epi32(_mm_and_si128(vacc89AB, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x16() 61 …const __m128i vremCDEF = _mm_add_epi32(_mm_and_si128(vaccCDEF, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x16() 93 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x16() 98 …const __m128i vrem0123 = _mm_add_epi32(_mm_and_si128(vacc0123, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x16() 99 …const __m128i vrem4567 = _mm_add_epi32(_mm_and_si128(vacc4567, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x16()
|
D | minmax-sse41-mul32-ld32-x32.c | 57 …const __m128i vrem0123 = _mm_add_epi32(_mm_and_si128(vacc0123, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul32_ld32_x32() 58 …const __m128i vrem4567 = _mm_add_epi32(_mm_and_si128(vacc4567, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul32_ld32_x32() 59 …const __m128i vrem89AB = _mm_add_epi32(_mm_and_si128(vacc89AB, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul32_ld32_x32() 60 …const __m128i vremCDEF = _mm_add_epi32(_mm_and_si128(vaccCDEF, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul32_ld32_x32() 61 …const __m128i vremGHIJ = _mm_add_epi32(_mm_and_si128(vaccGHIJ, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul32_ld32_x32() 62 …const __m128i vremKLMN = _mm_add_epi32(_mm_and_si128(vaccKLMN, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul32_ld32_x32() 63 …const __m128i vremOPQR = _mm_add_epi32(_mm_and_si128(vaccOPQR, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul32_ld32_x32() 64 …const __m128i vremSTUV = _mm_add_epi32(_mm_and_si128(vaccSTUV, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul32_ld32_x32() 106 …const __m128i vrem0123 = _mm_add_epi32(_mm_and_si128(vacc0123, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul32_ld32_x32() 107 …const __m128i vrem4567 = _mm_add_epi32(_mm_and_si128(vacc4567, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul32_ld32_x32()
|
D | minmax-xop-mul32-ld32-x32.c | 62 …const __m128i vrem0123 = _mm_add_epi32(_mm_and_si128(vacc0123, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__xop_mul32_ld32_x32() 63 …const __m128i vrem4567 = _mm_add_epi32(_mm_and_si128(vacc4567, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__xop_mul32_ld32_x32() 64 …const __m128i vrem89AB = _mm_add_epi32(_mm_and_si128(vacc89AB, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__xop_mul32_ld32_x32() 65 …const __m128i vremCDEF = _mm_add_epi32(_mm_and_si128(vaccCDEF, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__xop_mul32_ld32_x32() 66 …const __m128i vremGHIJ = _mm_add_epi32(_mm_and_si128(vaccGHIJ, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__xop_mul32_ld32_x32() 67 …const __m128i vremKLMN = _mm_add_epi32(_mm_and_si128(vaccKLMN, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__xop_mul32_ld32_x32() 68 …const __m128i vremOPQR = _mm_add_epi32(_mm_and_si128(vaccOPQR, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__xop_mul32_ld32_x32() 69 …const __m128i vremSTUV = _mm_add_epi32(_mm_and_si128(vaccSTUV, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__xop_mul32_ld32_x32() 111 …const __m128i vrem0123 = _mm_add_epi32(_mm_and_si128(vacc0123, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__xop_mul32_ld32_x32() 112 …const __m128i vrem4567 = _mm_add_epi32(_mm_and_si128(vacc4567, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__xop_mul32_ld32_x32()
|
D | minmax-sse2-mul16-ld64-x16.c | 52 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x16() 53 …vxprod89ABCDEFhi = _mm_sub_epi16(vxprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vx89ABCDEF, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x16() 60 …const __m128i vrem0123 = _mm_add_epi32(_mm_and_si128(vacc0123, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x16() 61 …const __m128i vrem4567 = _mm_add_epi32(_mm_and_si128(vacc4567, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x16() 62 …const __m128i vrem89AB = _mm_add_epi32(_mm_and_si128(vacc89AB, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x16() 63 …const __m128i vremCDEF = _mm_add_epi32(_mm_and_si128(vaccCDEF, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x16() 96 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x16() 101 …const __m128i vrem0123 = _mm_add_epi32(_mm_and_si128(vacc0123, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x16() 102 …const __m128i vrem4567 = _mm_add_epi32(_mm_and_si128(vacc4567, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x16()
|
/external/libvpx/libvpx/vpx_dsp/x86/ |
D | loopfilter_avx2.c | 66 abs_p1q1 = _mm_srli_epi16(_mm_and_si128(abs_p1q1, fe), 1); in vpx_lpf_horizontal_16_avx2() 99 filt = _mm_and_si128(_mm_subs_epi8(qs1ps1, qs1), hev); in vpx_lpf_horizontal_16_avx2() 105 filt = _mm_and_si128(filt, mask); in vpx_lpf_horizontal_16_avx2() 137 flat = _mm_and_si128(flat, mask); in vpx_lpf_horizontal_16_avx2() 163 flat2 = _mm_and_si128(flat2, flat); // flat2 & flat & mask in vpx_lpf_horizontal_16_avx2() 310 flat_q2p2 = _mm_and_si128(flat, flat_q2p2); in vpx_lpf_horizontal_16_avx2() 314 flat_q1p1 = _mm_and_si128(flat, flat_q1p1); in vpx_lpf_horizontal_16_avx2() 318 flat_q0p0 = _mm_and_si128(flat, flat_q0p0); in vpx_lpf_horizontal_16_avx2() 322 flat2_q6p6 = _mm_and_si128(flat2, flat2_q6p6); in vpx_lpf_horizontal_16_avx2() 328 flat2_q5p5 = _mm_and_si128(flat2, flat2_q5p5); in vpx_lpf_horizontal_16_avx2() [all …]
|
D | highbd_loopfilter_sse2.c | 41 ubounded = _mm_and_si128(ubounded, max); in signed_char_clamp_bd_sse2() 42 lbounded = _mm_and_si128(lbounded, min); in signed_char_clamp_bd_sse2() 121 mask = _mm_and_si128(mask, _mm_adds_epu16(limit_v, one)); in vpx_highbd_lpf_horizontal_16_sse2() 156 filt = _mm_and_si128(signed_char_clamp_bd_sse2(_mm_subs_epi16(ps1, qs1), bd), in vpx_highbd_lpf_horizontal_16_sse2() 162 filt = _mm_and_si128(filt, mask); in vpx_highbd_lpf_horizontal_16_sse2() 208 flat = _mm_and_si128(flat, mask); in vpx_highbd_lpf_horizontal_16_sse2() 246 flat2 = _mm_and_si128(flat2, flat); // flat2 & flat & mask in vpx_highbd_lpf_horizontal_16_sse2() 355 flat_p2 = _mm_and_si128(flat, flat_p2); in vpx_highbd_lpf_horizontal_16_sse2() 359 flat_q2 = _mm_and_si128(flat, flat_q2); in vpx_highbd_lpf_horizontal_16_sse2() 364 flat_p1 = _mm_and_si128(flat, flat_p1); in vpx_highbd_lpf_horizontal_16_sse2() [all …]
|
/external/OpenCL-CTS/test_common/harness/ |
D | mt19937.cpp | 148 _mm_and_si128(_mm_load_si128((__m128i *)(mt + kk)), in genrand_int32() 150 _mm_and_si128(_mm_loadu_si128((__m128i *)(mt + kk + 1)), in genrand_int32() 154 __m128i mask = _mm_cmpeq_epi32(_mm_and_si128(vy, one.v), one.v); in genrand_int32() 156 __m128i vmag01 = _mm_and_si128(mask, matrix_a.v); in genrand_int32() 184 _mm_and_si128(_mm_load_si128((__m128i *)(mt + kk)), in genrand_int32() 187 _mm_and_si128(_mm_loadu_si128((__m128i *)(mt + kk + 1)), in genrand_int32() 191 __m128i mask = _mm_cmpeq_epi32(_mm_and_si128(vy, one.v), one.v); in genrand_int32() 193 __m128i vmag01 = _mm_and_si128(mask, matrix_a.v); in genrand_int32() 221 vy = _mm_xor_si128(vy, _mm_and_si128(_mm_slli_epi32(vy, 7), c0.v)); in genrand_int32() 223 vy = _mm_xor_si128(vy, _mm_and_si128(_mm_slli_epi32(vy, 15), c1.v)); in genrand_int32()
|
/external/XNNPACK/src/x8-zip/ |
D | x3-sse2.c | 39 const __m128i vxeye = _mm_or_si128(_mm_and_si128(vx, vmask0x00FF00FF), _mm_slli_epi16(vy, 8)); in xnn_x8_zip_x3_ukernel__sse2() 43 …const __m128i vzexo = _mm_or_si128(_mm_and_si128(vz, vmask0x00FF00FF), _mm_andnot_si128(vmask0x00F… in xnn_x8_zip_x3_ukernel__sse2() 46 …const __m128i vxeyezexo = _mm_or_si128(_mm_and_si128(vxeye, vmask0x0000FFFF), _mm_slli_epi32(vzexo… in xnn_x8_zip_x3_ukernel__sse2() 48 …const __m128i vyozoxeye = _mm_or_si128(_mm_and_si128(vyozo, vmask0x0000FFFF), _mm_andnot_si128(vma… in xnn_x8_zip_x3_ukernel__sse2() 88 const __m128i vxeye = _mm_or_si128(_mm_and_si128(vx, vmask0x00FF00FF), _mm_slli_epi16(vy, 8)); in xnn_x8_zip_x3_ukernel__sse2() 92 …const __m128i vzexo = _mm_or_si128(_mm_and_si128(vz, vmask0x00FF00FF), _mm_andnot_si128(vmask0x00F… in xnn_x8_zip_x3_ukernel__sse2() 95 …const __m128i vxeyezexo = _mm_or_si128(_mm_and_si128(vxeye, vmask0x0000FFFF), _mm_slli_epi32(vzexo… in xnn_x8_zip_x3_ukernel__sse2() 97 …const __m128i vyozoxeye = _mm_or_si128(_mm_and_si128(vyozo, vmask0x0000FFFF), _mm_andnot_si128(vma… in xnn_x8_zip_x3_ukernel__sse2()
|
/external/skqp/src/opts/ |
D | SkBlitRow_opts.h | 23 __m128i rb = _mm_and_si128(mask, c); in SkPMSrcOver_SSE2() 111 auto ANDed = _mm_and_si128(s3, _mm_and_si128(s2, _mm_and_si128(s1, s0))); in blit_row_s32a_opaque() 146 if (0xffff == _mm_movemask_epi8(_mm_cmpeq_epi8(_mm_and_si128(ORed, alphaMask), in blit_row_s32a_opaque() 160 auto ANDed = _mm_and_si128(s3, _mm_and_si128(s2, _mm_and_si128(s1, s0))); in blit_row_s32a_opaque() 161 if (0xffff == _mm_movemask_epi8(_mm_cmpeq_epi8(_mm_and_si128(ANDed, alphaMask), in blit_row_s32a_opaque()
|