Home
last modified time | relevance | path

Searched refs:_mm_and_si128 (Results 1 – 25 of 352) sorted by relevance

12345678910>>...15

/external/oboe/samples/RhythmGame/third_party/glm/simd/
Dinteger.h28 Reg1 = _mm_and_si128(Reg1, Mask4); in glm_i128_interleave()
34 Reg1 = _mm_and_si128(Reg1, Mask3); in glm_i128_interleave()
40 Reg1 = _mm_and_si128(Reg1, Mask2); in glm_i128_interleave()
46 Reg1 = _mm_and_si128(Reg1, Mask1); in glm_i128_interleave()
52 Reg1 = _mm_and_si128(Reg1, Mask0); in glm_i128_interleave()
81 Reg1 = _mm_and_si128(Reg1, Mask4); in glm_i128_interleave2()
87 Reg1 = _mm_and_si128(Reg1, Mask3); in glm_i128_interleave2()
93 Reg1 = _mm_and_si128(Reg1, Mask2); in glm_i128_interleave2()
99 Reg1 = _mm_and_si128(Reg1, Mask1); in glm_i128_interleave2()
105 Reg1 = _mm_and_si128(Reg1, Mask0); in glm_i128_interleave2()
/external/XNNPACK/src/f32-argmaxpool/
D9p8x-sse2-c4.c80 vidx = _mm_or_si128(_mm_andnot_si128(vm1, vidx), _mm_and_si128(vm1, _mm_set1_epi32(1))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4()
84 vidx = _mm_or_si128(_mm_andnot_si128(vm2, vidx), _mm_and_si128(vm2, _mm_set1_epi32(2))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4()
88 vidx = _mm_or_si128(_mm_andnot_si128(vm3, vidx), _mm_and_si128(vm3, _mm_set1_epi32(3))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4()
92 vidx = _mm_or_si128(_mm_andnot_si128(vm4, vidx), _mm_and_si128(vm4, _mm_set1_epi32(4))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4()
96 vidx = _mm_or_si128(_mm_andnot_si128(vm5, vidx), _mm_and_si128(vm5, _mm_set1_epi32(5))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4()
100 vidx = _mm_or_si128(_mm_andnot_si128(vm6, vidx), _mm_and_si128(vm6, _mm_set1_epi32(6))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4()
104 vidx = _mm_or_si128(_mm_andnot_si128(vm7, vidx), _mm_and_si128(vm7, _mm_set1_epi32(7))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4()
108 vidx = _mm_or_si128(_mm_andnot_si128(vm8, vidx), _mm_and_si128(vm8, _mm_set1_epi32(8))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4()
165 vidx = _mm_or_si128(_mm_andnot_si128(vm0, vidx), _mm_and_si128(vm0, vidx0)); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4()
170 vidx = _mm_or_si128(_mm_andnot_si128(vm1, vidx), _mm_and_si128(vm1, vidx1)); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4()
[all …]
D9x-sse2-c4.c99 vidx = _mm_or_si128(_mm_andnot_si128(vm1, vidx), _mm_and_si128(vm1, _mm_set1_epi32(1))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4()
103 vidx = _mm_or_si128(_mm_andnot_si128(vm2, vidx), _mm_and_si128(vm2, _mm_set1_epi32(2))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4()
107 vidx = _mm_or_si128(_mm_andnot_si128(vm3, vidx), _mm_and_si128(vm3, _mm_set1_epi32(3))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4()
111 vidx = _mm_or_si128(_mm_andnot_si128(vm4, vidx), _mm_and_si128(vm4, _mm_set1_epi32(4))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4()
115 vidx = _mm_or_si128(_mm_andnot_si128(vm5, vidx), _mm_and_si128(vm5, _mm_set1_epi32(5))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4()
119 vidx = _mm_or_si128(_mm_andnot_si128(vm6, vidx), _mm_and_si128(vm6, _mm_set1_epi32(6))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4()
123 vidx = _mm_or_si128(_mm_andnot_si128(vm7, vidx), _mm_and_si128(vm7, _mm_set1_epi32(7))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4()
127 vidx = _mm_or_si128(_mm_andnot_si128(vm8, vidx), _mm_and_si128(vm8, _mm_set1_epi32(8))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4()
150 vidx = _mm_or_si128(_mm_andnot_si128(vm1, vidx), _mm_and_si128(vm1, _mm_set1_epi32(1))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4()
154 vidx = _mm_or_si128(_mm_andnot_si128(vm2, vidx), _mm_and_si128(vm2, _mm_set1_epi32(2))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4()
[all …]
/external/libavc/common/x86/
Dih264_iquant_itrans_recon_dc_ssse3.c153 temp4 = _mm_and_si128(temp4, sign_reg); in ih264_iquant_itrans_recon_4x4_dc_ssse3()
155 temp5 = _mm_and_si128(temp5, sign_reg); in ih264_iquant_itrans_recon_4x4_dc_ssse3()
306 temp1 = _mm_and_si128(temp1, sign_reg); in ih264_iquant_itrans_recon_8x8_dc_ssse3()
308 temp2 = _mm_and_si128(temp2, sign_reg); in ih264_iquant_itrans_recon_8x8_dc_ssse3()
310 temp3 = _mm_and_si128(temp3, sign_reg); in ih264_iquant_itrans_recon_8x8_dc_ssse3()
312 temp4 = _mm_and_si128(temp4, sign_reg); in ih264_iquant_itrans_recon_8x8_dc_ssse3()
314 temp5 = _mm_and_si128(temp5, sign_reg); in ih264_iquant_itrans_recon_8x8_dc_ssse3()
316 temp6 = _mm_and_si128(temp6, sign_reg); in ih264_iquant_itrans_recon_8x8_dc_ssse3()
318 temp7 = _mm_and_si128(temp7, sign_reg); in ih264_iquant_itrans_recon_8x8_dc_ssse3()
320 temp8 = _mm_and_si128(temp8, sign_reg); in ih264_iquant_itrans_recon_8x8_dc_ssse3()
[all …]
Dih264_deblk_luma_ssse3.c196 flag1_16x8 = _mm_and_si128(flag1_16x8, flag2_16x8); in ih264_deblk_luma_vert_bs4_ssse3()
212 flag1_16x8 = _mm_and_si128(flag1_16x8, flag2_16x8); in ih264_deblk_luma_vert_bs4_ssse3()
227 flag2_16x8 = _mm_and_si128(flag1_16x8, flag2_16x8); in ih264_deblk_luma_vert_bs4_ssse3()
240 flag3_16x8 = _mm_and_si128(flag3_16x8, flag2_16x8); in ih264_deblk_luma_vert_bs4_ssse3()
253 flag4_16x8 = _mm_and_si128(flag4_16x8, flag2_16x8); in ih264_deblk_luma_vert_bs4_ssse3()
384 p0_16x8 = _mm_and_si128(p0_16x8, in ih264_deblk_luma_vert_bs4_ssse3()
386 p0_16x8_1 = _mm_and_si128(p0_16x8_1, flag1_16x8); in ih264_deblk_luma_vert_bs4_ssse3()
388 q0_16x8 = _mm_and_si128(q0_16x8, in ih264_deblk_luma_vert_bs4_ssse3()
390 q0_16x8_1 = _mm_and_si128(q0_16x8_1, flag1_16x8); in ih264_deblk_luma_vert_bs4_ssse3()
394 p0_16x8 = _mm_and_si128(p0_16x8, in ih264_deblk_luma_vert_bs4_ssse3()
[all …]
Dih264_deblk_chroma_ssse3.c153 flag1 = _mm_and_si128(flag1, _mm_cmpgt_epi16(beta_cbcr_16x8, diff)); in ih264_deblk_chroma_vert_bs4_ssse3()
157 flag1 = _mm_and_si128(flag1, _mm_cmpgt_epi16(beta_cbcr_16x8, diff)); in ih264_deblk_chroma_vert_bs4_ssse3()
184 flag2 = _mm_and_si128(flag2, _mm_cmpgt_epi16(beta_cbcr_16x8, diff)); in ih264_deblk_chroma_vert_bs4_ssse3()
188 flag2 = _mm_and_si128(flag2, _mm_cmpgt_epi16(beta_cbcr_16x8, diff)); in ih264_deblk_chroma_vert_bs4_ssse3()
207 p0_uv_8x16_1 = _mm_and_si128(p0_uv_16x8, in ih264_deblk_chroma_vert_bs4_ssse3()
209 p0_uv_8x16_2 = _mm_and_si128(p0_uv_8x16_2, flag1); in ih264_deblk_chroma_vert_bs4_ssse3()
212 q0_uv_8x16_1 = _mm_and_si128(q0_uv_16x8, in ih264_deblk_chroma_vert_bs4_ssse3()
214 q0_uv_8x16_2 = _mm_and_si128(q0_uv_8x16_2, flag1); in ih264_deblk_chroma_vert_bs4_ssse3()
322 flag1 = _mm_and_si128(flag1, _mm_cmpgt_epi16(beta_cbcr_16x8, diff)); in ih264_deblk_chroma_horz_bs4_ssse3()
326 flag1 = _mm_and_si128(flag1, _mm_cmpgt_epi16(beta_cbcr_16x8, diff)); in ih264_deblk_chroma_horz_bs4_ssse3()
[all …]
/external/XNNPACK/src/qs8-vadd/gen/
Dminmax-sse41-mul16-ld64-x32.c75 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32()
76 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32()
77 …vxprod89ABCDEFhi = _mm_sub_epi16(vxprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vx89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32()
78 …vyprod89ABCDEFhi = _mm_sub_epi16(vyprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vy89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32()
79 …vxprodGHIJKLMNhi = _mm_sub_epi16(vxprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vxGHIJKLMN, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32()
80 …vyprodGHIJKLMNhi = _mm_sub_epi16(vyprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vyGHIJKLMN, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32()
81 …vxprodOPQRSTUVhi = _mm_sub_epi16(vxprodOPQRSTUVhi, _mm_and_si128(_mm_srai_epi16(vxOPQRSTUV, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32()
82 …vyprodOPQRSTUVhi = _mm_sub_epi16(vyprodOPQRSTUVhi, _mm_and_si128(_mm_srai_epi16(vyOPQRSTUV, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32()
102 …const __m128i vrem0123 = _mm_add_epi32(_mm_and_si128(vacc0123, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32()
103 …const __m128i vrem4567 = _mm_add_epi32(_mm_and_si128(vacc4567, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x32()
[all …]
Dminmax-sse2-mul16-ld64-x32.c83 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32()
84 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32()
85 …vxprod89ABCDEFhi = _mm_sub_epi16(vxprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vx89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32()
86 …vyprod89ABCDEFhi = _mm_sub_epi16(vyprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vy89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32()
87 …vxprodGHIJKLMNhi = _mm_sub_epi16(vxprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vxGHIJKLMN, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32()
88 …vyprodGHIJKLMNhi = _mm_sub_epi16(vyprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vyGHIJKLMN, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32()
89 …vxprodOPQRSTUVhi = _mm_sub_epi16(vxprodOPQRSTUVhi, _mm_and_si128(_mm_srai_epi16(vxOPQRSTUV, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32()
90 …vyprodOPQRSTUVhi = _mm_sub_epi16(vyprodOPQRSTUVhi, _mm_and_si128(_mm_srai_epi16(vyOPQRSTUV, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32()
110 …const __m128i vrem0123 = _mm_add_epi32(_mm_and_si128(vacc0123, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32()
111 …const __m128i vrem4567 = _mm_add_epi32(_mm_and_si128(vacc4567, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x32()
[all …]
Dminmax-sse41-mul16-ld64-x24.c67 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x24()
68 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x24()
69 …vxprod89ABCDEFhi = _mm_sub_epi16(vxprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vx89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x24()
70 …vyprod89ABCDEFhi = _mm_sub_epi16(vyprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vy89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x24()
71 …vxprodGHIJKLMNhi = _mm_sub_epi16(vxprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vxGHIJKLMN, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x24()
72 …vyprodGHIJKLMNhi = _mm_sub_epi16(vyprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vyGHIJKLMN, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x24()
88 …const __m128i vrem0123 = _mm_add_epi32(_mm_and_si128(vacc0123, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x24()
89 …const __m128i vrem4567 = _mm_add_epi32(_mm_and_si128(vacc4567, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x24()
90 …const __m128i vrem89AB = _mm_add_epi32(_mm_and_si128(vacc89AB, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x24()
91 …const __m128i vremCDEF = _mm_add_epi32(_mm_and_si128(vaccCDEF, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x24()
[all …]
Dminmax-sse2-mul16-ld64-x24.c73 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x24()
74 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x24()
75 …vxprod89ABCDEFhi = _mm_sub_epi16(vxprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vx89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x24()
76 …vyprod89ABCDEFhi = _mm_sub_epi16(vyprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vy89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x24()
77 …vxprodGHIJKLMNhi = _mm_sub_epi16(vxprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vxGHIJKLMN, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x24()
78 …vyprodGHIJKLMNhi = _mm_sub_epi16(vyprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vyGHIJKLMN, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x24()
94 …const __m128i vrem0123 = _mm_add_epi32(_mm_and_si128(vacc0123, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x24()
95 …const __m128i vrem4567 = _mm_add_epi32(_mm_and_si128(vacc4567, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x24()
96 …const __m128i vrem89AB = _mm_add_epi32(_mm_and_si128(vacc89AB, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x24()
97 …const __m128i vremCDEF = _mm_add_epi32(_mm_and_si128(vaccCDEF, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x24()
[all …]
Dminmax-sse2-mul16-ld64-x16.c63 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x16()
64 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x16()
65 …vxprod89ABCDEFhi = _mm_sub_epi16(vxprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vx89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x16()
66 …vyprod89ABCDEFhi = _mm_sub_epi16(vyprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vy89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x16()
78 …const __m128i vrem0123 = _mm_add_epi32(_mm_and_si128(vacc0123, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x16()
79 …const __m128i vrem4567 = _mm_add_epi32(_mm_and_si128(vacc4567, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x16()
80 …const __m128i vrem89AB = _mm_add_epi32(_mm_and_si128(vacc89AB, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x16()
81 …const __m128i vremCDEF = _mm_add_epi32(_mm_and_si128(vaccCDEF, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x16()
120 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x16()
121 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse2_mul16_ld64_x16()
[all …]
Dminmax-sse41-mul16-ld64-x16.c59 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x16()
60 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x16()
61 …vxprod89ABCDEFhi = _mm_sub_epi16(vxprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vx89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x16()
62 …vyprod89ABCDEFhi = _mm_sub_epi16(vyprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vy89ABCDEF, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x16()
74 …const __m128i vrem0123 = _mm_add_epi32(_mm_and_si128(vacc0123, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x16()
75 …const __m128i vrem4567 = _mm_add_epi32(_mm_and_si128(vacc4567, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x16()
76 …const __m128i vrem89AB = _mm_add_epi32(_mm_and_si128(vacc89AB, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x16()
77 …const __m128i vremCDEF = _mm_add_epi32(_mm_and_si128(vaccCDEF, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x16()
114 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x16()
115 …vyprod01234567hi = _mm_sub_epi16(vyprod01234567hi, _mm_and_si128(_mm_srai_epi16(vy01234567, 15), v… in xnn_qs8_vadd_minmax_ukernel__sse41_mul16_ld64_x16()
[all …]
/external/XNNPACK/src/qs8-vaddc/gen/
Dminmax-sse2-mul16-ld64-x32.c62 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x32()
63 …vxprod89ABCDEFhi = _mm_sub_epi16(vxprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vx89ABCDEF, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x32()
64 …vxprodGHIJKLMNhi = _mm_sub_epi16(vxprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vxGHIJKLMN, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x32()
65 …vxprodOPQRSTUVhi = _mm_sub_epi16(vxprodOPQRSTUVhi, _mm_and_si128(_mm_srai_epi16(vxOPQRSTUV, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x32()
76 …const __m128i vrem0123 = _mm_add_epi32(_mm_and_si128(vacc0123, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x32()
77 …const __m128i vrem4567 = _mm_add_epi32(_mm_and_si128(vacc4567, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x32()
78 …const __m128i vrem89AB = _mm_add_epi32(_mm_and_si128(vacc89AB, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x32()
79 …const __m128i vremCDEF = _mm_add_epi32(_mm_and_si128(vaccCDEF, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x32()
80 …const __m128i vremGHIJ = _mm_add_epi32(_mm_and_si128(vaccGHIJ, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x32()
81 …const __m128i vremKLMN = _mm_add_epi32(_mm_and_si128(vaccKLMN, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x32()
[all …]
Dminmax-sse41-mul16-ld64-x32.c58 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x32()
59 …vxprod89ABCDEFhi = _mm_sub_epi16(vxprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vx89ABCDEF, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x32()
60 …vxprodGHIJKLMNhi = _mm_sub_epi16(vxprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vxGHIJKLMN, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x32()
61 …vxprodOPQRSTUVhi = _mm_sub_epi16(vxprodOPQRSTUVhi, _mm_and_si128(_mm_srai_epi16(vxOPQRSTUV, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x32()
72 …const __m128i vrem0123 = _mm_add_epi32(_mm_and_si128(vacc0123, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x32()
73 …const __m128i vrem4567 = _mm_add_epi32(_mm_and_si128(vacc4567, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x32()
74 …const __m128i vrem89AB = _mm_add_epi32(_mm_and_si128(vacc89AB, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x32()
75 …const __m128i vremCDEF = _mm_add_epi32(_mm_and_si128(vaccCDEF, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x32()
76 …const __m128i vremGHIJ = _mm_add_epi32(_mm_and_si128(vaccGHIJ, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x32()
77 …const __m128i vremKLMN = _mm_add_epi32(_mm_and_si128(vaccKLMN, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x32()
[all …]
Dminmax-sse41-mul16-ld64-x24.c54 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x24()
55 …vxprod89ABCDEFhi = _mm_sub_epi16(vxprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vx89ABCDEF, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x24()
56 …vxprodGHIJKLMNhi = _mm_sub_epi16(vxprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vxGHIJKLMN, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x24()
65 …const __m128i vrem0123 = _mm_add_epi32(_mm_and_si128(vacc0123, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x24()
66 …const __m128i vrem4567 = _mm_add_epi32(_mm_and_si128(vacc4567, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x24()
67 …const __m128i vrem89AB = _mm_add_epi32(_mm_and_si128(vacc89AB, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x24()
68 …const __m128i vremCDEF = _mm_add_epi32(_mm_and_si128(vaccCDEF, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x24()
69 …const __m128i vremGHIJ = _mm_add_epi32(_mm_and_si128(vaccGHIJ, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x24()
70 …const __m128i vremKLMN = _mm_add_epi32(_mm_and_si128(vaccKLMN, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x24()
109 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x24()
[all …]
Dminmax-sse2-mul16-ld64-x24.c57 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x24()
58 …vxprod89ABCDEFhi = _mm_sub_epi16(vxprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vx89ABCDEF, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x24()
59 …vxprodGHIJKLMNhi = _mm_sub_epi16(vxprodGHIJKLMNhi, _mm_and_si128(_mm_srai_epi16(vxGHIJKLMN, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x24()
68 …const __m128i vrem0123 = _mm_add_epi32(_mm_and_si128(vacc0123, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x24()
69 …const __m128i vrem4567 = _mm_add_epi32(_mm_and_si128(vacc4567, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x24()
70 …const __m128i vrem89AB = _mm_add_epi32(_mm_and_si128(vacc89AB, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x24()
71 …const __m128i vremCDEF = _mm_add_epi32(_mm_and_si128(vaccCDEF, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x24()
72 …const __m128i vremGHIJ = _mm_add_epi32(_mm_and_si128(vaccGHIJ, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x24()
73 …const __m128i vremKLMN = _mm_add_epi32(_mm_and_si128(vaccKLMN, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x24()
113 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x24()
[all …]
Dminmax-sse41-mul16-ld64-x16.c50 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x16()
51 …vxprod89ABCDEFhi = _mm_sub_epi16(vxprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vx89ABCDEF, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x16()
58 …const __m128i vrem0123 = _mm_add_epi32(_mm_and_si128(vacc0123, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x16()
59 …const __m128i vrem4567 = _mm_add_epi32(_mm_and_si128(vacc4567, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x16()
60 …const __m128i vrem89AB = _mm_add_epi32(_mm_and_si128(vacc89AB, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x16()
61 …const __m128i vremCDEF = _mm_add_epi32(_mm_and_si128(vaccCDEF, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x16()
93 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x16()
98 …const __m128i vrem0123 = _mm_add_epi32(_mm_and_si128(vacc0123, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x16()
99 …const __m128i vrem4567 = _mm_add_epi32(_mm_and_si128(vacc4567, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x16()
Dminmax-sse41-mul32-ld32-x32.c57 …const __m128i vrem0123 = _mm_add_epi32(_mm_and_si128(vacc0123, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul32_ld32_x32()
58 …const __m128i vrem4567 = _mm_add_epi32(_mm_and_si128(vacc4567, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul32_ld32_x32()
59 …const __m128i vrem89AB = _mm_add_epi32(_mm_and_si128(vacc89AB, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul32_ld32_x32()
60 …const __m128i vremCDEF = _mm_add_epi32(_mm_and_si128(vaccCDEF, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul32_ld32_x32()
61 …const __m128i vremGHIJ = _mm_add_epi32(_mm_and_si128(vaccGHIJ, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul32_ld32_x32()
62 …const __m128i vremKLMN = _mm_add_epi32(_mm_and_si128(vaccKLMN, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul32_ld32_x32()
63 …const __m128i vremOPQR = _mm_add_epi32(_mm_and_si128(vaccOPQR, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul32_ld32_x32()
64 …const __m128i vremSTUV = _mm_add_epi32(_mm_and_si128(vaccSTUV, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul32_ld32_x32()
106 …const __m128i vrem0123 = _mm_add_epi32(_mm_and_si128(vacc0123, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul32_ld32_x32()
107 …const __m128i vrem4567 = _mm_add_epi32(_mm_and_si128(vacc4567, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse41_mul32_ld32_x32()
Dminmax-xop-mul32-ld32-x32.c62 …const __m128i vrem0123 = _mm_add_epi32(_mm_and_si128(vacc0123, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__xop_mul32_ld32_x32()
63 …const __m128i vrem4567 = _mm_add_epi32(_mm_and_si128(vacc4567, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__xop_mul32_ld32_x32()
64 …const __m128i vrem89AB = _mm_add_epi32(_mm_and_si128(vacc89AB, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__xop_mul32_ld32_x32()
65 …const __m128i vremCDEF = _mm_add_epi32(_mm_and_si128(vaccCDEF, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__xop_mul32_ld32_x32()
66 …const __m128i vremGHIJ = _mm_add_epi32(_mm_and_si128(vaccGHIJ, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__xop_mul32_ld32_x32()
67 …const __m128i vremKLMN = _mm_add_epi32(_mm_and_si128(vaccKLMN, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__xop_mul32_ld32_x32()
68 …const __m128i vremOPQR = _mm_add_epi32(_mm_and_si128(vaccOPQR, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__xop_mul32_ld32_x32()
69 …const __m128i vremSTUV = _mm_add_epi32(_mm_and_si128(vaccSTUV, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__xop_mul32_ld32_x32()
111 …const __m128i vrem0123 = _mm_add_epi32(_mm_and_si128(vacc0123, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__xop_mul32_ld32_x32()
112 …const __m128i vrem4567 = _mm_add_epi32(_mm_and_si128(vacc4567, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__xop_mul32_ld32_x32()
Dminmax-sse2-mul16-ld64-x16.c52 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x16()
53 …vxprod89ABCDEFhi = _mm_sub_epi16(vxprod89ABCDEFhi, _mm_and_si128(_mm_srai_epi16(vx89ABCDEF, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x16()
60 …const __m128i vrem0123 = _mm_add_epi32(_mm_and_si128(vacc0123, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x16()
61 …const __m128i vrem4567 = _mm_add_epi32(_mm_and_si128(vacc4567, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x16()
62 …const __m128i vrem89AB = _mm_add_epi32(_mm_and_si128(vacc89AB, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x16()
63 …const __m128i vremCDEF = _mm_add_epi32(_mm_and_si128(vaccCDEF, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x16()
96 …vxprod01234567hi = _mm_sub_epi16(vxprod01234567hi, _mm_and_si128(_mm_srai_epi16(vx01234567, 15), v… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x16()
101 …const __m128i vrem0123 = _mm_add_epi32(_mm_and_si128(vacc0123, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x16()
102 …const __m128i vrem4567 = _mm_add_epi32(_mm_and_si128(vacc4567, vremainder_mask), _mm_cmpgt_epi32(_… in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x16()
/external/libvpx/libvpx/vpx_dsp/x86/
Dloopfilter_avx2.c66 abs_p1q1 = _mm_srli_epi16(_mm_and_si128(abs_p1q1, fe), 1); in vpx_lpf_horizontal_16_avx2()
99 filt = _mm_and_si128(_mm_subs_epi8(qs1ps1, qs1), hev); in vpx_lpf_horizontal_16_avx2()
105 filt = _mm_and_si128(filt, mask); in vpx_lpf_horizontal_16_avx2()
137 flat = _mm_and_si128(flat, mask); in vpx_lpf_horizontal_16_avx2()
163 flat2 = _mm_and_si128(flat2, flat); // flat2 & flat & mask in vpx_lpf_horizontal_16_avx2()
310 flat_q2p2 = _mm_and_si128(flat, flat_q2p2); in vpx_lpf_horizontal_16_avx2()
314 flat_q1p1 = _mm_and_si128(flat, flat_q1p1); in vpx_lpf_horizontal_16_avx2()
318 flat_q0p0 = _mm_and_si128(flat, flat_q0p0); in vpx_lpf_horizontal_16_avx2()
322 flat2_q6p6 = _mm_and_si128(flat2, flat2_q6p6); in vpx_lpf_horizontal_16_avx2()
328 flat2_q5p5 = _mm_and_si128(flat2, flat2_q5p5); in vpx_lpf_horizontal_16_avx2()
[all …]
Dhighbd_loopfilter_sse2.c41 ubounded = _mm_and_si128(ubounded, max); in signed_char_clamp_bd_sse2()
42 lbounded = _mm_and_si128(lbounded, min); in signed_char_clamp_bd_sse2()
121 mask = _mm_and_si128(mask, _mm_adds_epu16(limit_v, one)); in vpx_highbd_lpf_horizontal_16_sse2()
156 filt = _mm_and_si128(signed_char_clamp_bd_sse2(_mm_subs_epi16(ps1, qs1), bd), in vpx_highbd_lpf_horizontal_16_sse2()
162 filt = _mm_and_si128(filt, mask); in vpx_highbd_lpf_horizontal_16_sse2()
208 flat = _mm_and_si128(flat, mask); in vpx_highbd_lpf_horizontal_16_sse2()
246 flat2 = _mm_and_si128(flat2, flat); // flat2 & flat & mask in vpx_highbd_lpf_horizontal_16_sse2()
355 flat_p2 = _mm_and_si128(flat, flat_p2); in vpx_highbd_lpf_horizontal_16_sse2()
359 flat_q2 = _mm_and_si128(flat, flat_q2); in vpx_highbd_lpf_horizontal_16_sse2()
364 flat_p1 = _mm_and_si128(flat, flat_p1); in vpx_highbd_lpf_horizontal_16_sse2()
[all …]
/external/OpenCL-CTS/test_common/harness/
Dmt19937.cpp148 _mm_and_si128(_mm_load_si128((__m128i *)(mt + kk)), in genrand_int32()
150 _mm_and_si128(_mm_loadu_si128((__m128i *)(mt + kk + 1)), in genrand_int32()
154 __m128i mask = _mm_cmpeq_epi32(_mm_and_si128(vy, one.v), one.v); in genrand_int32()
156 __m128i vmag01 = _mm_and_si128(mask, matrix_a.v); in genrand_int32()
184 _mm_and_si128(_mm_load_si128((__m128i *)(mt + kk)), in genrand_int32()
187 _mm_and_si128(_mm_loadu_si128((__m128i *)(mt + kk + 1)), in genrand_int32()
191 __m128i mask = _mm_cmpeq_epi32(_mm_and_si128(vy, one.v), one.v); in genrand_int32()
193 __m128i vmag01 = _mm_and_si128(mask, matrix_a.v); in genrand_int32()
221 vy = _mm_xor_si128(vy, _mm_and_si128(_mm_slli_epi32(vy, 7), c0.v)); in genrand_int32()
223 vy = _mm_xor_si128(vy, _mm_and_si128(_mm_slli_epi32(vy, 15), c1.v)); in genrand_int32()
/external/XNNPACK/src/x8-zip/
Dx3-sse2.c39 const __m128i vxeye = _mm_or_si128(_mm_and_si128(vx, vmask0x00FF00FF), _mm_slli_epi16(vy, 8)); in xnn_x8_zip_x3_ukernel__sse2()
43 …const __m128i vzexo = _mm_or_si128(_mm_and_si128(vz, vmask0x00FF00FF), _mm_andnot_si128(vmask0x00F… in xnn_x8_zip_x3_ukernel__sse2()
46 …const __m128i vxeyezexo = _mm_or_si128(_mm_and_si128(vxeye, vmask0x0000FFFF), _mm_slli_epi32(vzexo… in xnn_x8_zip_x3_ukernel__sse2()
48 …const __m128i vyozoxeye = _mm_or_si128(_mm_and_si128(vyozo, vmask0x0000FFFF), _mm_andnot_si128(vma… in xnn_x8_zip_x3_ukernel__sse2()
88 const __m128i vxeye = _mm_or_si128(_mm_and_si128(vx, vmask0x00FF00FF), _mm_slli_epi16(vy, 8)); in xnn_x8_zip_x3_ukernel__sse2()
92 …const __m128i vzexo = _mm_or_si128(_mm_and_si128(vz, vmask0x00FF00FF), _mm_andnot_si128(vmask0x00F… in xnn_x8_zip_x3_ukernel__sse2()
95 …const __m128i vxeyezexo = _mm_or_si128(_mm_and_si128(vxeye, vmask0x0000FFFF), _mm_slli_epi32(vzexo… in xnn_x8_zip_x3_ukernel__sse2()
97 …const __m128i vyozoxeye = _mm_or_si128(_mm_and_si128(vyozo, vmask0x0000FFFF), _mm_andnot_si128(vma… in xnn_x8_zip_x3_ukernel__sse2()
/external/skqp/src/opts/
DSkBlitRow_opts.h23 __m128i rb = _mm_and_si128(mask, c); in SkPMSrcOver_SSE2()
111 auto ANDed = _mm_and_si128(s3, _mm_and_si128(s2, _mm_and_si128(s1, s0))); in blit_row_s32a_opaque()
146 if (0xffff == _mm_movemask_epi8(_mm_cmpeq_epi8(_mm_and_si128(ORed, alphaMask), in blit_row_s32a_opaque()
160 auto ANDed = _mm_and_si128(s3, _mm_and_si128(s2, _mm_and_si128(s1, s0))); in blit_row_s32a_opaque()
161 if (0xffff == _mm_movemask_epi8(_mm_cmpeq_epi8(_mm_and_si128(ANDed, alphaMask), in blit_row_s32a_opaque()

12345678910>>...15