/external/XNNPACK/src/f32-argmaxpool/ |
D | 9p8x-sse2-c4.c | 80 vidx = _mm_or_si128(_mm_andnot_si128(vm1, vidx), _mm_and_si128(vm1, _mm_set1_epi32(1))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() 84 vidx = _mm_or_si128(_mm_andnot_si128(vm2, vidx), _mm_and_si128(vm2, _mm_set1_epi32(2))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() 88 vidx = _mm_or_si128(_mm_andnot_si128(vm3, vidx), _mm_and_si128(vm3, _mm_set1_epi32(3))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() 92 vidx = _mm_or_si128(_mm_andnot_si128(vm4, vidx), _mm_and_si128(vm4, _mm_set1_epi32(4))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() 96 vidx = _mm_or_si128(_mm_andnot_si128(vm5, vidx), _mm_and_si128(vm5, _mm_set1_epi32(5))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() 100 vidx = _mm_or_si128(_mm_andnot_si128(vm6, vidx), _mm_and_si128(vm6, _mm_set1_epi32(6))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() 104 vidx = _mm_or_si128(_mm_andnot_si128(vm7, vidx), _mm_and_si128(vm7, _mm_set1_epi32(7))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() 108 vidx = _mm_or_si128(_mm_andnot_si128(vm8, vidx), _mm_and_si128(vm8, _mm_set1_epi32(8))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() 165 vidx = _mm_or_si128(_mm_andnot_si128(vm0, vidx), _mm_and_si128(vm0, vidx0)); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() 170 vidx = _mm_or_si128(_mm_andnot_si128(vm1, vidx), _mm_and_si128(vm1, vidx1)); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() [all …]
|
D | 9x-sse2-c4.c | 99 vidx = _mm_or_si128(_mm_andnot_si128(vm1, vidx), _mm_and_si128(vm1, _mm_set1_epi32(1))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() 103 vidx = _mm_or_si128(_mm_andnot_si128(vm2, vidx), _mm_and_si128(vm2, _mm_set1_epi32(2))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() 107 vidx = _mm_or_si128(_mm_andnot_si128(vm3, vidx), _mm_and_si128(vm3, _mm_set1_epi32(3))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() 111 vidx = _mm_or_si128(_mm_andnot_si128(vm4, vidx), _mm_and_si128(vm4, _mm_set1_epi32(4))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() 115 vidx = _mm_or_si128(_mm_andnot_si128(vm5, vidx), _mm_and_si128(vm5, _mm_set1_epi32(5))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() 119 vidx = _mm_or_si128(_mm_andnot_si128(vm6, vidx), _mm_and_si128(vm6, _mm_set1_epi32(6))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() 123 vidx = _mm_or_si128(_mm_andnot_si128(vm7, vidx), _mm_and_si128(vm7, _mm_set1_epi32(7))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() 127 vidx = _mm_or_si128(_mm_andnot_si128(vm8, vidx), _mm_and_si128(vm8, _mm_set1_epi32(8))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() 150 vidx = _mm_or_si128(_mm_andnot_si128(vm1, vidx), _mm_and_si128(vm1, _mm_set1_epi32(1))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() 154 vidx = _mm_or_si128(_mm_andnot_si128(vm2, vidx), _mm_and_si128(vm2, _mm_set1_epi32(2))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() [all …]
|
D | 4x-sse2-c4.c | 64 vidx = _mm_or_si128(_mm_andnot_si128(vm1, vidx), _mm_and_si128(vm1, _mm_set1_epi32(1))); in xnn_f32_argmaxpool_ukernel_4x__sse2_c4() 68 vidx = _mm_or_si128(_mm_andnot_si128(vm2, vidx), _mm_and_si128(vm2, _mm_set1_epi32(2))); in xnn_f32_argmaxpool_ukernel_4x__sse2_c4() 72 vidx = _mm_or_si128(_mm_andnot_si128(vm3, vidx), _mm_and_si128(vm3, _mm_set1_epi32(3))); in xnn_f32_argmaxpool_ukernel_4x__sse2_c4() 90 vidx = _mm_or_si128(_mm_andnot_si128(vm1, vidx), _mm_and_si128(vm1, _mm_set1_epi32(1))); in xnn_f32_argmaxpool_ukernel_4x__sse2_c4() 94 vidx = _mm_or_si128(_mm_andnot_si128(vm2, vidx), _mm_and_si128(vm2, _mm_set1_epi32(2))); in xnn_f32_argmaxpool_ukernel_4x__sse2_c4() 98 vidx = _mm_or_si128(_mm_andnot_si128(vm3, vidx), _mm_and_si128(vm3, _mm_set1_epi32(3))); in xnn_f32_argmaxpool_ukernel_4x__sse2_c4()
|
/external/XNNPACK/src/x8-zip/ |
D | x3-sse2.c | 41 … const __m128i vyozo = _mm_or_si128(_mm_andnot_si128(vmask0x00FF00FF, vz), _mm_srli_epi16(vy, 8)); in xnn_x8_zip_x3_ukernel__sse2() 43 …const __m128i vzexo = _mm_or_si128(_mm_and_si128(vz, vmask0x00FF00FF), _mm_andnot_si128(vmask0x00F… in xnn_x8_zip_x3_ukernel__sse2() 48 …28i vyozoxeye = _mm_or_si128(_mm_and_si128(vyozo, vmask0x0000FFFF), _mm_andnot_si128(vmask0x0000FF… in xnn_x8_zip_x3_ukernel__sse2() 50 …const __m128i vzexoyozo = _mm_or_si128(_mm_andnot_si128(vmask0x0000FFFF, vyozo), _mm_srli_epi32(vz… in xnn_x8_zip_x3_ukernel__sse2() 90 … const __m128i vyozo = _mm_or_si128(_mm_andnot_si128(vmask0x00FF00FF, vz), _mm_srli_epi16(vy, 8)); in xnn_x8_zip_x3_ukernel__sse2() 92 …const __m128i vzexo = _mm_or_si128(_mm_and_si128(vz, vmask0x00FF00FF), _mm_andnot_si128(vmask0x00F… in xnn_x8_zip_x3_ukernel__sse2() 97 …28i vyozoxeye = _mm_or_si128(_mm_and_si128(vyozo, vmask0x0000FFFF), _mm_andnot_si128(vmask0x0000FF… in xnn_x8_zip_x3_ukernel__sse2() 99 …const __m128i vzexoyozo = _mm_or_si128(_mm_andnot_si128(vmask0x0000FFFF, vyozo), _mm_srli_epi32(vz… in xnn_x8_zip_x3_ukernel__sse2()
|
/external/libvpx/libvpx/vpx_dsp/x86/ |
D | loopfilter_avx2.c | 122 filt = _mm_andnot_si128(_mm_srai_epi16(_mm_unpacklo_epi8(zero, hev), 0x8), in vpx_lpf_horizontal_16_avx2() 309 q2p2 = _mm_andnot_si128(flat, q2p2); in vpx_lpf_horizontal_16_avx2() 313 qs1ps1 = _mm_andnot_si128(flat, qs1ps1); in vpx_lpf_horizontal_16_avx2() 317 qs0ps0 = _mm_andnot_si128(flat, qs0ps0); in vpx_lpf_horizontal_16_avx2() 321 q6p6 = _mm_andnot_si128(flat2, q6p6); in vpx_lpf_horizontal_16_avx2() 327 q5p5 = _mm_andnot_si128(flat2, q5p5); in vpx_lpf_horizontal_16_avx2() 333 q4p4 = _mm_andnot_si128(flat2, q4p4); in vpx_lpf_horizontal_16_avx2() 339 q3p3 = _mm_andnot_si128(flat2, q3p3); in vpx_lpf_horizontal_16_avx2() 345 q2p2 = _mm_andnot_si128(flat2, q2p2); in vpx_lpf_horizontal_16_avx2() 351 q1p1 = _mm_andnot_si128(flat2, q1p1); in vpx_lpf_horizontal_16_avx2() [all …]
|
D | highbd_loopfilter_sse2.c | 40 retval = _mm_andnot_si128(_mm_or_si128(ubounded, lbounded), value); in signed_char_clamp_bd_sse2() 176 filt = _mm_andnot_si128(hev, filt); in vpx_highbd_lpf_horizontal_16_sse2() 353 p2 = _mm_andnot_si128(flat, p2); in vpx_highbd_lpf_horizontal_16_sse2() 358 q2 = _mm_andnot_si128(flat, q2); in vpx_highbd_lpf_horizontal_16_sse2() 362 ps1 = _mm_andnot_si128(flat, ps1); in vpx_highbd_lpf_horizontal_16_sse2() 367 qs1 = _mm_andnot_si128(flat, qs1); in vpx_highbd_lpf_horizontal_16_sse2() 371 ps0 = _mm_andnot_si128(flat, ps0); in vpx_highbd_lpf_horizontal_16_sse2() 376 qs0 = _mm_andnot_si128(flat, qs0); in vpx_highbd_lpf_horizontal_16_sse2() 382 p6 = _mm_andnot_si128(flat2, p6); in vpx_highbd_lpf_horizontal_16_sse2() 387 q6 = _mm_andnot_si128(flat2, q6); in vpx_highbd_lpf_horizontal_16_sse2() [all …]
|
D | loopfilter_sse2.c | 94 filter = _mm_andnot_si128(hev, filter); \ 329 filt = _mm_andnot_si128(_mm_srai_epi16(_mm_unpacklo_epi8(zero, hev), 0x8), in vpx_lpf_horizontal_16_sse2() 507 q2p2 = _mm_andnot_si128(flat, q2p2); in vpx_lpf_horizontal_16_sse2() 511 qs1ps1 = _mm_andnot_si128(flat, qs1ps1); in vpx_lpf_horizontal_16_sse2() 515 qs0ps0 = _mm_andnot_si128(flat, qs0ps0); in vpx_lpf_horizontal_16_sse2() 519 q6p6 = _mm_andnot_si128(flat2, q6p6); in vpx_lpf_horizontal_16_sse2() 525 q5p5 = _mm_andnot_si128(flat2, q5p5); in vpx_lpf_horizontal_16_sse2() 531 q4p4 = _mm_andnot_si128(flat2, q4p4); in vpx_lpf_horizontal_16_sse2() 537 q3p3 = _mm_andnot_si128(flat2, q3p3); in vpx_lpf_horizontal_16_sse2() 543 q2p2 = _mm_andnot_si128(flat2, q2p2); in vpx_lpf_horizontal_16_sse2() [all …]
|
D | quantize_sse2.h | 80 eob0 = _mm_andnot_si128(zero_coeff0, scan0); in scan_for_eob() 81 eob1 = _mm_andnot_si128(zero_coeff1, scan1); in scan_for_eob()
|
D | post_proc_sse2.c | 130 mask_0 = _mm_andnot_si128(mask_0, this_row); in vpx_mbpost_proc_down_sse2()
|
/external/libvpx/libvpx/vp8/encoder/x86/ |
D | denoising_sse2.c | 91 adj = _mm_andnot_si128(mask0, adj); in vp8_denoiser_filter_sse2() 95 padj = _mm_andnot_si128(diff_sign, adj); in vp8_denoiser_filter_sse2() 153 padj = _mm_andnot_si128(diff_sign, adj); in vp8_denoiser_filter_sse2() 267 adj = _mm_andnot_si128(mask0, adj); in vp8_denoiser_filter_uv_sse2() 271 padj = _mm_andnot_si128(diff_sign, adj); in vp8_denoiser_filter_uv_sse2() 340 padj = _mm_andnot_si128(diff_sign, adj); in vp8_denoiser_filter_uv_sse2()
|
/external/libaom/libaom/aom_dsp/x86/ |
D | highbd_loopfilter_sse2.c | 252 filt = _mm_andnot_si128(*hev, filt); in highbd_filter4_sse2() 317 filter = _mm_andnot_si128(hev, filter); in highbd_filter4_dual_sse2() 479 pq[i] = _mm_andnot_si128(flat, pq[i]); in highbd_lpf_internal_14_sse2() 488 pq[i] = _mm_andnot_si128(flat2, pq[i]); in highbd_lpf_internal_14_sse2() 674 ps[i] = _mm_andnot_si128(flat, ps[i]); in highbd_lpf_internal_14_dual_sse2() 677 qs[i] = _mm_andnot_si128(flat, qs[i]); in highbd_lpf_internal_14_dual_sse2() 681 p[2] = _mm_andnot_si128(flat, p[2]); in highbd_lpf_internal_14_dual_sse2() 686 q[2] = _mm_andnot_si128(flat, q[2]); in highbd_lpf_internal_14_dual_sse2() 691 ps[i] = _mm_andnot_si128(flat, ps[i]); in highbd_lpf_internal_14_dual_sse2() 694 qs[i] = _mm_andnot_si128(flat, qs[i]); in highbd_lpf_internal_14_dual_sse2() [all …]
|
D | loopfilter_sse2.c | 178 filter = _mm_andnot_si128(*hev, filter); in filter4_sse2() 231 filter = _mm_andnot_si128(*hev, filter); in filter4_dual_sse2() 571 *q2p2 = _mm_andnot_si128(flat, *q2p2); in lpf_internal_14_dual_sse2() 575 qs1ps1 = _mm_andnot_si128(flat, qs1ps1); in lpf_internal_14_dual_sse2() 579 qs0ps0 = _mm_andnot_si128(flat, qs0ps0); in lpf_internal_14_dual_sse2() 667 *q5p5 = _mm_andnot_si128(flat2, *q5p5); in lpf_internal_14_dual_sse2() 671 *q4p4 = _mm_andnot_si128(flat2, *q4p4); in lpf_internal_14_dual_sse2() 675 *q3p3 = _mm_andnot_si128(flat2, *q3p3); in lpf_internal_14_dual_sse2() 679 *q2p2 = _mm_andnot_si128(flat2, *q2p2); in lpf_internal_14_dual_sse2() 683 *q1p1 = _mm_andnot_si128(flat2, *q1p1); in lpf_internal_14_dual_sse2() [all …]
|
D | quantize_x86.h | 111 eob0 = _mm_andnot_si128(zero_coeff0, scan0); in scan_for_eob() 112 eob1 = _mm_andnot_si128(zero_coeff1, scan1); in scan_for_eob()
|
/external/fec/ |
D | viterbi27_sse2.c | 161 survivor0 = _mm_or_si128(_mm_and_si128(decision0,m1),_mm_andnot_si128(decision0,m0)); 162 survivor1 = _mm_or_si128(_mm_and_si128(decision1,m3),_mm_andnot_si128(decision1,m2));
|
/external/libvpx/libvpx/vp9/encoder/x86/ |
D | vp9_denoiser_sse2.c | 70 adj = _mm_andnot_si128(mask0, adj); in vp9_denoiser_16x1_sse2() 74 padj = _mm_andnot_si128(diff_sign, adj); in vp9_denoiser_16x1_sse2() 106 padj = _mm_andnot_si128(diff_sign, adj); in vp9_denoiser_adj_16x1_sse2()
|
/external/gemmlowp/fixedpoint/ |
D | fixedpoint_sse.h | 90 return _mm_andnot_si128(a, _mm_set1_epi32(-1)); 95 return to_int16x8_m128i(_mm_andnot_si128(a.v, _mm_set1_epi16(-1))); 163 _mm_andnot_si128(if_mask, else_val));
|
/external/webrtc/modules/video_processing/util/ |
D | denoiser_filter_sse2.cc | 179 adj = _mm_andnot_si128(mask0, adj); in MbDenoise() 183 padj = _mm_andnot_si128(diff_sign, adj); in MbDenoise()
|
/external/libaom/libaom/av1/encoder/x86/ |
D | av1_highbd_quantize_sse4.c | 77 qcoeff[0] = _mm_andnot_si128(qcoeff[2], qcoeff[0]); in quantize_coeff_phase2() 78 dquan[0] = _mm_andnot_si128(qcoeff[2], dquan[0]); in quantize_coeff_phase2()
|
/external/libaom/libaom/aom_dsp/simd/ |
D | v128_intrinsics_x86.h | 387 SIMD_INLINE v128 v128_andn(v128 a, v128 b) { return _mm_andnot_si128(b, a); } in v128_andn() 464 return _mm_or_si128(_mm_andnot_si128(mask, b), _mm_and_si128(mask, a)); in v128_min_s8() 484 return _mm_or_si128(_mm_andnot_si128(mask, b), _mm_and_si128(mask, a)); in v128_max_s8() 497 return _mm_or_si128(_mm_andnot_si128(mask, b), _mm_and_si128(mask, a)); in v128_min_s32() 506 return _mm_or_si128(_mm_andnot_si128(mask, b), _mm_and_si128(mask, a)); in v128_max_s32()
|
D | v64_intrinsics_x86.h | 355 SIMD_INLINE v64 v64_andn(v64 a, v64 b) { return _mm_andnot_si128(b, a); } in v64_andn() 406 return _mm_or_si128(_mm_andnot_si128(mask, b), _mm_and_si128(mask, a)); in v64_min_s8() 415 return _mm_or_si128(_mm_andnot_si128(mask, b), _mm_and_si128(mask, a)); in v64_max_s8()
|
/external/oboe/samples/RhythmGame/third_party/glm/detail/ |
D | func_integer_simd.inl | 22 __m128i const set2 = _mm_andnot_si128(set0, _mm_set1_epi32(-1));
|
/external/skqp/src/core/ |
D | SkBlitRow_D32.cpp | 52 diff_ag = _mm_andnot_si128(mask, diff_ag); in SkPMLerp_SSE2() 121 dst_ag = _mm_andnot_si128(mask, dst_ag); in SkBlendARGB32_SSE2()
|
/external/skia/src/core/ |
D | SkBlitRow_D32.cpp | 51 diff_ag = _mm_andnot_si128(mask, diff_ag); in SkPMLerp_SSE2() 120 dst_ag = _mm_andnot_si128(mask, dst_ag); in SkBlendARGB32_SSE2()
|
/external/skia/include/private/ |
D | SkNx_sse.h | 277 _mm_andnot_si128(fVec, e.fVec)); in thenElse() 347 _mm_andnot_si128(fVec, e.fVec)); in thenElse() 395 _mm_andnot_si128(fVec, e.fVec)); in thenElse() 573 _mm_andnot_si128(fVec, e.fVec)); in thenElse() 641 _mm_andnot_si128(fVec, e.fVec)); in thenElse() 684 _mm_andnot_si128(fVec, e.fVec)); in thenElse()
|
/external/XNNPACK/src/math/ |
D | exp-sse2-rr2-lut64-p2.c | 62 __m128i veo = _mm_slli_epi32(_mm_andnot_si128(vindex_mask, _mm_castps_si128(vn)), 17); in xnn_math_f32_exp__sse2_rr2_lut64_p2()
|