Home
last modified time | relevance | path

Searched refs:_mm_andnot_si128 (Results 1 – 25 of 64) sorted by relevance

123

/external/XNNPACK/src/f32-argmaxpool/
D9p8x-sse2-c4.c80 vidx = _mm_or_si128(_mm_andnot_si128(vm1, vidx), _mm_and_si128(vm1, _mm_set1_epi32(1))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4()
84 vidx = _mm_or_si128(_mm_andnot_si128(vm2, vidx), _mm_and_si128(vm2, _mm_set1_epi32(2))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4()
88 vidx = _mm_or_si128(_mm_andnot_si128(vm3, vidx), _mm_and_si128(vm3, _mm_set1_epi32(3))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4()
92 vidx = _mm_or_si128(_mm_andnot_si128(vm4, vidx), _mm_and_si128(vm4, _mm_set1_epi32(4))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4()
96 vidx = _mm_or_si128(_mm_andnot_si128(vm5, vidx), _mm_and_si128(vm5, _mm_set1_epi32(5))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4()
100 vidx = _mm_or_si128(_mm_andnot_si128(vm6, vidx), _mm_and_si128(vm6, _mm_set1_epi32(6))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4()
104 vidx = _mm_or_si128(_mm_andnot_si128(vm7, vidx), _mm_and_si128(vm7, _mm_set1_epi32(7))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4()
108 vidx = _mm_or_si128(_mm_andnot_si128(vm8, vidx), _mm_and_si128(vm8, _mm_set1_epi32(8))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4()
165 vidx = _mm_or_si128(_mm_andnot_si128(vm0, vidx), _mm_and_si128(vm0, vidx0)); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4()
170 vidx = _mm_or_si128(_mm_andnot_si128(vm1, vidx), _mm_and_si128(vm1, vidx1)); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4()
[all …]
D9x-sse2-c4.c99 vidx = _mm_or_si128(_mm_andnot_si128(vm1, vidx), _mm_and_si128(vm1, _mm_set1_epi32(1))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4()
103 vidx = _mm_or_si128(_mm_andnot_si128(vm2, vidx), _mm_and_si128(vm2, _mm_set1_epi32(2))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4()
107 vidx = _mm_or_si128(_mm_andnot_si128(vm3, vidx), _mm_and_si128(vm3, _mm_set1_epi32(3))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4()
111 vidx = _mm_or_si128(_mm_andnot_si128(vm4, vidx), _mm_and_si128(vm4, _mm_set1_epi32(4))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4()
115 vidx = _mm_or_si128(_mm_andnot_si128(vm5, vidx), _mm_and_si128(vm5, _mm_set1_epi32(5))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4()
119 vidx = _mm_or_si128(_mm_andnot_si128(vm6, vidx), _mm_and_si128(vm6, _mm_set1_epi32(6))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4()
123 vidx = _mm_or_si128(_mm_andnot_si128(vm7, vidx), _mm_and_si128(vm7, _mm_set1_epi32(7))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4()
127 vidx = _mm_or_si128(_mm_andnot_si128(vm8, vidx), _mm_and_si128(vm8, _mm_set1_epi32(8))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4()
150 vidx = _mm_or_si128(_mm_andnot_si128(vm1, vidx), _mm_and_si128(vm1, _mm_set1_epi32(1))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4()
154 vidx = _mm_or_si128(_mm_andnot_si128(vm2, vidx), _mm_and_si128(vm2, _mm_set1_epi32(2))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4()
[all …]
D4x-sse2-c4.c64 vidx = _mm_or_si128(_mm_andnot_si128(vm1, vidx), _mm_and_si128(vm1, _mm_set1_epi32(1))); in xnn_f32_argmaxpool_ukernel_4x__sse2_c4()
68 vidx = _mm_or_si128(_mm_andnot_si128(vm2, vidx), _mm_and_si128(vm2, _mm_set1_epi32(2))); in xnn_f32_argmaxpool_ukernel_4x__sse2_c4()
72 vidx = _mm_or_si128(_mm_andnot_si128(vm3, vidx), _mm_and_si128(vm3, _mm_set1_epi32(3))); in xnn_f32_argmaxpool_ukernel_4x__sse2_c4()
90 vidx = _mm_or_si128(_mm_andnot_si128(vm1, vidx), _mm_and_si128(vm1, _mm_set1_epi32(1))); in xnn_f32_argmaxpool_ukernel_4x__sse2_c4()
94 vidx = _mm_or_si128(_mm_andnot_si128(vm2, vidx), _mm_and_si128(vm2, _mm_set1_epi32(2))); in xnn_f32_argmaxpool_ukernel_4x__sse2_c4()
98 vidx = _mm_or_si128(_mm_andnot_si128(vm3, vidx), _mm_and_si128(vm3, _mm_set1_epi32(3))); in xnn_f32_argmaxpool_ukernel_4x__sse2_c4()
/external/XNNPACK/src/x8-zip/
Dx3-sse2.c41 … const __m128i vyozo = _mm_or_si128(_mm_andnot_si128(vmask0x00FF00FF, vz), _mm_srli_epi16(vy, 8)); in xnn_x8_zip_x3_ukernel__sse2()
43 …const __m128i vzexo = _mm_or_si128(_mm_and_si128(vz, vmask0x00FF00FF), _mm_andnot_si128(vmask0x00F… in xnn_x8_zip_x3_ukernel__sse2()
48 …28i vyozoxeye = _mm_or_si128(_mm_and_si128(vyozo, vmask0x0000FFFF), _mm_andnot_si128(vmask0x0000FF… in xnn_x8_zip_x3_ukernel__sse2()
50 …const __m128i vzexoyozo = _mm_or_si128(_mm_andnot_si128(vmask0x0000FFFF, vyozo), _mm_srli_epi32(vz… in xnn_x8_zip_x3_ukernel__sse2()
90 … const __m128i vyozo = _mm_or_si128(_mm_andnot_si128(vmask0x00FF00FF, vz), _mm_srli_epi16(vy, 8)); in xnn_x8_zip_x3_ukernel__sse2()
92 …const __m128i vzexo = _mm_or_si128(_mm_and_si128(vz, vmask0x00FF00FF), _mm_andnot_si128(vmask0x00F… in xnn_x8_zip_x3_ukernel__sse2()
97 …28i vyozoxeye = _mm_or_si128(_mm_and_si128(vyozo, vmask0x0000FFFF), _mm_andnot_si128(vmask0x0000FF… in xnn_x8_zip_x3_ukernel__sse2()
99 …const __m128i vzexoyozo = _mm_or_si128(_mm_andnot_si128(vmask0x0000FFFF, vyozo), _mm_srli_epi32(vz… in xnn_x8_zip_x3_ukernel__sse2()
/external/libvpx/libvpx/vpx_dsp/x86/
Dloopfilter_avx2.c122 filt = _mm_andnot_si128(_mm_srai_epi16(_mm_unpacklo_epi8(zero, hev), 0x8), in vpx_lpf_horizontal_16_avx2()
309 q2p2 = _mm_andnot_si128(flat, q2p2); in vpx_lpf_horizontal_16_avx2()
313 qs1ps1 = _mm_andnot_si128(flat, qs1ps1); in vpx_lpf_horizontal_16_avx2()
317 qs0ps0 = _mm_andnot_si128(flat, qs0ps0); in vpx_lpf_horizontal_16_avx2()
321 q6p6 = _mm_andnot_si128(flat2, q6p6); in vpx_lpf_horizontal_16_avx2()
327 q5p5 = _mm_andnot_si128(flat2, q5p5); in vpx_lpf_horizontal_16_avx2()
333 q4p4 = _mm_andnot_si128(flat2, q4p4); in vpx_lpf_horizontal_16_avx2()
339 q3p3 = _mm_andnot_si128(flat2, q3p3); in vpx_lpf_horizontal_16_avx2()
345 q2p2 = _mm_andnot_si128(flat2, q2p2); in vpx_lpf_horizontal_16_avx2()
351 q1p1 = _mm_andnot_si128(flat2, q1p1); in vpx_lpf_horizontal_16_avx2()
[all …]
Dhighbd_loopfilter_sse2.c40 retval = _mm_andnot_si128(_mm_or_si128(ubounded, lbounded), value); in signed_char_clamp_bd_sse2()
176 filt = _mm_andnot_si128(hev, filt); in vpx_highbd_lpf_horizontal_16_sse2()
353 p2 = _mm_andnot_si128(flat, p2); in vpx_highbd_lpf_horizontal_16_sse2()
358 q2 = _mm_andnot_si128(flat, q2); in vpx_highbd_lpf_horizontal_16_sse2()
362 ps1 = _mm_andnot_si128(flat, ps1); in vpx_highbd_lpf_horizontal_16_sse2()
367 qs1 = _mm_andnot_si128(flat, qs1); in vpx_highbd_lpf_horizontal_16_sse2()
371 ps0 = _mm_andnot_si128(flat, ps0); in vpx_highbd_lpf_horizontal_16_sse2()
376 qs0 = _mm_andnot_si128(flat, qs0); in vpx_highbd_lpf_horizontal_16_sse2()
382 p6 = _mm_andnot_si128(flat2, p6); in vpx_highbd_lpf_horizontal_16_sse2()
387 q6 = _mm_andnot_si128(flat2, q6); in vpx_highbd_lpf_horizontal_16_sse2()
[all …]
Dloopfilter_sse2.c94 filter = _mm_andnot_si128(hev, filter); \
329 filt = _mm_andnot_si128(_mm_srai_epi16(_mm_unpacklo_epi8(zero, hev), 0x8), in vpx_lpf_horizontal_16_sse2()
507 q2p2 = _mm_andnot_si128(flat, q2p2); in vpx_lpf_horizontal_16_sse2()
511 qs1ps1 = _mm_andnot_si128(flat, qs1ps1); in vpx_lpf_horizontal_16_sse2()
515 qs0ps0 = _mm_andnot_si128(flat, qs0ps0); in vpx_lpf_horizontal_16_sse2()
519 q6p6 = _mm_andnot_si128(flat2, q6p6); in vpx_lpf_horizontal_16_sse2()
525 q5p5 = _mm_andnot_si128(flat2, q5p5); in vpx_lpf_horizontal_16_sse2()
531 q4p4 = _mm_andnot_si128(flat2, q4p4); in vpx_lpf_horizontal_16_sse2()
537 q3p3 = _mm_andnot_si128(flat2, q3p3); in vpx_lpf_horizontal_16_sse2()
543 q2p2 = _mm_andnot_si128(flat2, q2p2); in vpx_lpf_horizontal_16_sse2()
[all …]
Dquantize_sse2.h80 eob0 = _mm_andnot_si128(zero_coeff0, scan0); in scan_for_eob()
81 eob1 = _mm_andnot_si128(zero_coeff1, scan1); in scan_for_eob()
Dpost_proc_sse2.c130 mask_0 = _mm_andnot_si128(mask_0, this_row); in vpx_mbpost_proc_down_sse2()
/external/libvpx/libvpx/vp8/encoder/x86/
Ddenoising_sse2.c91 adj = _mm_andnot_si128(mask0, adj); in vp8_denoiser_filter_sse2()
95 padj = _mm_andnot_si128(diff_sign, adj); in vp8_denoiser_filter_sse2()
153 padj = _mm_andnot_si128(diff_sign, adj); in vp8_denoiser_filter_sse2()
267 adj = _mm_andnot_si128(mask0, adj); in vp8_denoiser_filter_uv_sse2()
271 padj = _mm_andnot_si128(diff_sign, adj); in vp8_denoiser_filter_uv_sse2()
340 padj = _mm_andnot_si128(diff_sign, adj); in vp8_denoiser_filter_uv_sse2()
/external/libaom/libaom/aom_dsp/x86/
Dhighbd_loopfilter_sse2.c252 filt = _mm_andnot_si128(*hev, filt); in highbd_filter4_sse2()
317 filter = _mm_andnot_si128(hev, filter); in highbd_filter4_dual_sse2()
479 pq[i] = _mm_andnot_si128(flat, pq[i]); in highbd_lpf_internal_14_sse2()
488 pq[i] = _mm_andnot_si128(flat2, pq[i]); in highbd_lpf_internal_14_sse2()
674 ps[i] = _mm_andnot_si128(flat, ps[i]); in highbd_lpf_internal_14_dual_sse2()
677 qs[i] = _mm_andnot_si128(flat, qs[i]); in highbd_lpf_internal_14_dual_sse2()
681 p[2] = _mm_andnot_si128(flat, p[2]); in highbd_lpf_internal_14_dual_sse2()
686 q[2] = _mm_andnot_si128(flat, q[2]); in highbd_lpf_internal_14_dual_sse2()
691 ps[i] = _mm_andnot_si128(flat, ps[i]); in highbd_lpf_internal_14_dual_sse2()
694 qs[i] = _mm_andnot_si128(flat, qs[i]); in highbd_lpf_internal_14_dual_sse2()
[all …]
Dloopfilter_sse2.c178 filter = _mm_andnot_si128(*hev, filter); in filter4_sse2()
231 filter = _mm_andnot_si128(*hev, filter); in filter4_dual_sse2()
571 *q2p2 = _mm_andnot_si128(flat, *q2p2); in lpf_internal_14_dual_sse2()
575 qs1ps1 = _mm_andnot_si128(flat, qs1ps1); in lpf_internal_14_dual_sse2()
579 qs0ps0 = _mm_andnot_si128(flat, qs0ps0); in lpf_internal_14_dual_sse2()
667 *q5p5 = _mm_andnot_si128(flat2, *q5p5); in lpf_internal_14_dual_sse2()
671 *q4p4 = _mm_andnot_si128(flat2, *q4p4); in lpf_internal_14_dual_sse2()
675 *q3p3 = _mm_andnot_si128(flat2, *q3p3); in lpf_internal_14_dual_sse2()
679 *q2p2 = _mm_andnot_si128(flat2, *q2p2); in lpf_internal_14_dual_sse2()
683 *q1p1 = _mm_andnot_si128(flat2, *q1p1); in lpf_internal_14_dual_sse2()
[all …]
Dquantize_x86.h111 eob0 = _mm_andnot_si128(zero_coeff0, scan0); in scan_for_eob()
112 eob1 = _mm_andnot_si128(zero_coeff1, scan1); in scan_for_eob()
/external/fec/
Dviterbi27_sse2.c161 survivor0 = _mm_or_si128(_mm_and_si128(decision0,m1),_mm_andnot_si128(decision0,m0));
162 survivor1 = _mm_or_si128(_mm_and_si128(decision1,m3),_mm_andnot_si128(decision1,m2));
/external/libvpx/libvpx/vp9/encoder/x86/
Dvp9_denoiser_sse2.c70 adj = _mm_andnot_si128(mask0, adj); in vp9_denoiser_16x1_sse2()
74 padj = _mm_andnot_si128(diff_sign, adj); in vp9_denoiser_16x1_sse2()
106 padj = _mm_andnot_si128(diff_sign, adj); in vp9_denoiser_adj_16x1_sse2()
/external/gemmlowp/fixedpoint/
Dfixedpoint_sse.h90 return _mm_andnot_si128(a, _mm_set1_epi32(-1));
95 return to_int16x8_m128i(_mm_andnot_si128(a.v, _mm_set1_epi16(-1)));
163 _mm_andnot_si128(if_mask, else_val));
/external/webrtc/modules/video_processing/util/
Ddenoiser_filter_sse2.cc179 adj = _mm_andnot_si128(mask0, adj); in MbDenoise()
183 padj = _mm_andnot_si128(diff_sign, adj); in MbDenoise()
/external/libaom/libaom/av1/encoder/x86/
Dav1_highbd_quantize_sse4.c77 qcoeff[0] = _mm_andnot_si128(qcoeff[2], qcoeff[0]); in quantize_coeff_phase2()
78 dquan[0] = _mm_andnot_si128(qcoeff[2], dquan[0]); in quantize_coeff_phase2()
/external/libaom/libaom/aom_dsp/simd/
Dv128_intrinsics_x86.h387 SIMD_INLINE v128 v128_andn(v128 a, v128 b) { return _mm_andnot_si128(b, a); } in v128_andn()
464 return _mm_or_si128(_mm_andnot_si128(mask, b), _mm_and_si128(mask, a)); in v128_min_s8()
484 return _mm_or_si128(_mm_andnot_si128(mask, b), _mm_and_si128(mask, a)); in v128_max_s8()
497 return _mm_or_si128(_mm_andnot_si128(mask, b), _mm_and_si128(mask, a)); in v128_min_s32()
506 return _mm_or_si128(_mm_andnot_si128(mask, b), _mm_and_si128(mask, a)); in v128_max_s32()
Dv64_intrinsics_x86.h355 SIMD_INLINE v64 v64_andn(v64 a, v64 b) { return _mm_andnot_si128(b, a); } in v64_andn()
406 return _mm_or_si128(_mm_andnot_si128(mask, b), _mm_and_si128(mask, a)); in v64_min_s8()
415 return _mm_or_si128(_mm_andnot_si128(mask, b), _mm_and_si128(mask, a)); in v64_max_s8()
/external/oboe/samples/RhythmGame/third_party/glm/detail/
Dfunc_integer_simd.inl22 __m128i const set2 = _mm_andnot_si128(set0, _mm_set1_epi32(-1));
/external/skqp/src/core/
DSkBlitRow_D32.cpp52 diff_ag = _mm_andnot_si128(mask, diff_ag); in SkPMLerp_SSE2()
121 dst_ag = _mm_andnot_si128(mask, dst_ag); in SkBlendARGB32_SSE2()
/external/skia/src/core/
DSkBlitRow_D32.cpp51 diff_ag = _mm_andnot_si128(mask, diff_ag); in SkPMLerp_SSE2()
120 dst_ag = _mm_andnot_si128(mask, dst_ag); in SkBlendARGB32_SSE2()
/external/skia/include/private/
DSkNx_sse.h277 _mm_andnot_si128(fVec, e.fVec)); in thenElse()
347 _mm_andnot_si128(fVec, e.fVec)); in thenElse()
395 _mm_andnot_si128(fVec, e.fVec)); in thenElse()
573 _mm_andnot_si128(fVec, e.fVec)); in thenElse()
641 _mm_andnot_si128(fVec, e.fVec)); in thenElse()
684 _mm_andnot_si128(fVec, e.fVec)); in thenElse()
/external/XNNPACK/src/math/
Dexp-sse2-rr2-lut64-p2.c62 __m128i veo = _mm_slli_epi32(_mm_andnot_si128(vindex_mask, _mm_castps_si128(vn)), 17); in xnn_math_f32_exp__sse2_rr2_lut64_p2()

123