Home
last modified time | relevance | path

Searched refs:_mm_blend_epi16 (Results 1 – 25 of 88) sorted by relevance

1234

/external/python/cpython3/Modules/_blake2/impl/
Dblake2s-load-sse41.h30 t0 = _mm_blend_epi16(m1, m2, 0x0C); \
32 t2 = _mm_blend_epi16(t0, t1, 0xF0); \
37 t1 = _mm_blend_epi16(m1,m3,0xC0); \
38 t2 = _mm_blend_epi16(t0, t1, 0xF0); \
43 t1 = _mm_blend_epi16(m2, t0, 0x30); \
44 t2 = _mm_blend_epi16(m0, t1, 0xF0); \
50 t2 = _mm_blend_epi16(t0, t1, 0x0C); \
55 t1 = _mm_blend_epi16(m3,m1,0x0C); \
56 t2 = _mm_blend_epi16(t0, t1, 0x0F); \
61 t1 = _mm_blend_epi16(t0, m0, 0xF0); \
[all …]
Dblake2b-load-sse41.h93 b1 = _mm_blend_epi16(m1, m6, 0xF0); \
100 b0 = _mm_blend_epi16(m5, m1, 0xF0); \
132 b0 = _mm_blend_epi16(m1, m2, 0xF0); \
133 b1 = _mm_blend_epi16(m2, m7, 0xF0); \
156 b0 = _mm_blend_epi16(m0, m3, 0xF0); \
157 b1 = _mm_blend_epi16(m2, m7, 0xF0); \
164 b0 = _mm_blend_epi16(m7, m5, 0xF0); \
165 b1 = _mm_blend_epi16(m3, m1, 0xF0); \
173 b1 = _mm_blend_epi16(m4, m6, 0xF0); \
196 b0 = _mm_blend_epi16(m2, m3, 0xF0); \
[all …]
/external/XNNPACK/src/qs8-requantization/
Dprecise-sse4.c83 const __m128i x_abs_scaled = _mm_blend_epi16(x_abs_scaled02, x_abs_scaled13, 0xCC); in xnn_qs8_requantize_precise__sse4()
84 const __m128i y_abs_scaled = _mm_blend_epi16(y_abs_scaled02, y_abs_scaled13, 0xCC); in xnn_qs8_requantize_precise__sse4()
85 const __m128i z_abs_scaled = _mm_blend_epi16(z_abs_scaled02, z_abs_scaled13, 0xCC); in xnn_qs8_requantize_precise__sse4()
86 const __m128i w_abs_scaled = _mm_blend_epi16(w_abs_scaled02, w_abs_scaled13, 0xCC); in xnn_qs8_requantize_precise__sse4()
Dq31-sse4.c86 const __m128i x_q31product = _mm_blend_epi16(x_q31product_even, x_q31product_odd, 0xCC); in xnn_qs8_requantize_q31__sse4()
87 const __m128i y_q31product = _mm_blend_epi16(y_q31product_even, y_q31product_odd, 0xCC); in xnn_qs8_requantize_q31__sse4()
88 const __m128i z_q31product = _mm_blend_epi16(z_q31product_even, z_q31product_odd, 0xCC); in xnn_qs8_requantize_q31__sse4()
89 const __m128i w_q31product = _mm_blend_epi16(w_q31product_even, w_q31product_odd, 0xCC); in xnn_qs8_requantize_q31__sse4()
/external/XNNPACK/src/qu8-requantization/
Dprecise-sse4.c83 const __m128i x_abs_scaled = _mm_blend_epi16(x_abs_scaled02, x_abs_scaled13, 0xCC); in xnn_qu8_requantize_precise__sse4()
84 const __m128i y_abs_scaled = _mm_blend_epi16(y_abs_scaled02, y_abs_scaled13, 0xCC); in xnn_qu8_requantize_precise__sse4()
85 const __m128i z_abs_scaled = _mm_blend_epi16(z_abs_scaled02, z_abs_scaled13, 0xCC); in xnn_qu8_requantize_precise__sse4()
86 const __m128i w_abs_scaled = _mm_blend_epi16(w_abs_scaled02, w_abs_scaled13, 0xCC); in xnn_qu8_requantize_precise__sse4()
Dq31-sse4.c86 const __m128i x_q31product = _mm_blend_epi16(x_q31product_even, x_q31product_odd, 0xCC); in xnn_qu8_requantize_q31__sse4()
87 const __m128i y_q31product = _mm_blend_epi16(y_q31product_even, y_q31product_odd, 0xCC); in xnn_qu8_requantize_q31__sse4()
88 const __m128i z_q31product = _mm_blend_epi16(z_q31product_even, z_q31product_odd, 0xCC); in xnn_qu8_requantize_q31__sse4()
89 const __m128i w_q31product = _mm_blend_epi16(w_q31product_even, w_q31product_odd, 0xCC); in xnn_qu8_requantize_q31__sse4()
/external/XNNPACK/src/qs8-gavgpool/gen/
D7x-minmax-sse41-c24-acc2.c160 …const __m128i vabsout0123 = _mm_blend_epi16(vabsout02, _mm_shuffle_epi32(vabsout13, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7x__sse41_c24_acc2()
161 …const __m128i vabsout4567 = _mm_blend_epi16(vabsout46, _mm_shuffle_epi32(vabsout57, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7x__sse41_c24_acc2()
162 …const __m128i vabsout89AB = _mm_blend_epi16(vabsout8A, _mm_shuffle_epi32(vabsout9B, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7x__sse41_c24_acc2()
163 …const __m128i vabsoutCDEF = _mm_blend_epi16(vabsoutCE, _mm_shuffle_epi32(vabsoutDF, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7x__sse41_c24_acc2()
164 …const __m128i vabsoutGHIJ = _mm_blend_epi16(vabsoutGI, _mm_shuffle_epi32(vabsoutHJ, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7x__sse41_c24_acc2()
165 …const __m128i vabsoutKLMN = _mm_blend_epi16(vabsoutKM, _mm_shuffle_epi32(vabsoutLN, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7x__sse41_c24_acc2()
241 …const __m128i vabsout0123 = _mm_blend_epi16(vabsout02, _mm_shuffle_epi32(vabsout13, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7x__sse41_c24_acc2()
242 …const __m128i vabsout4567 = _mm_blend_epi16(vabsout46, _mm_shuffle_epi32(vabsout57, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7x__sse41_c24_acc2()
D7x-minmax-sse41-c16-acc2.c133 …const __m128i vabsout0123 = _mm_blend_epi16(vabsout02, _mm_shuffle_epi32(vabsout13, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7x__sse41_c16_acc2()
134 …const __m128i vabsout4567 = _mm_blend_epi16(vabsout46, _mm_shuffle_epi32(vabsout57, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7x__sse41_c16_acc2()
135 …const __m128i vabsout89AB = _mm_blend_epi16(vabsout8A, _mm_shuffle_epi32(vabsout9B, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7x__sse41_c16_acc2()
136 …const __m128i vabsoutCDEF = _mm_blend_epi16(vabsoutCE, _mm_shuffle_epi32(vabsoutDF, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7x__sse41_c16_acc2()
206 …const __m128i vabsout0123 = _mm_blend_epi16(vabsout02, _mm_shuffle_epi32(vabsout13, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7x__sse41_c16_acc2()
207 …const __m128i vabsout4567 = _mm_blend_epi16(vabsout46, _mm_shuffle_epi32(vabsout57, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7x__sse41_c16_acc2()
D7x-minmax-sse41-c8-acc2.c106 …const __m128i vabsout0123 = _mm_blend_epi16(vabsout02, _mm_shuffle_epi32(vabsout13, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7x__sse41_c8_acc2()
107 …const __m128i vabsout4567 = _mm_blend_epi16(vabsout46, _mm_shuffle_epi32(vabsout57, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7x__sse41_c8_acc2()
173 …const __m128i vabsout0123 = _mm_blend_epi16(vabsout02, _mm_shuffle_epi32(vabsout13, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7x__sse41_c8_acc2()
174 …const __m128i vabsout4567 = _mm_blend_epi16(vabsout46, _mm_shuffle_epi32(vabsout57, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7x__sse41_c8_acc2()
D7p7x-minmax-sse41-c16-acc2.c262 …const __m128i vabsout0123 = _mm_blend_epi16(vabsout02, _mm_shuffle_epi32(vabsout13, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7p7x__sse41_c16_acc2()
263 …const __m128i vabsout4567 = _mm_blend_epi16(vabsout46, _mm_shuffle_epi32(vabsout57, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7p7x__sse41_c16_acc2()
264 …const __m128i vabsout89AB = _mm_blend_epi16(vabsout8A, _mm_shuffle_epi32(vabsout9B, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7p7x__sse41_c16_acc2()
265 …const __m128i vabsoutCDEF = _mm_blend_epi16(vabsoutCE, _mm_shuffle_epi32(vabsoutDF, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7p7x__sse41_c16_acc2()
336 …const __m128i vabsout0123 = _mm_blend_epi16(vabsout02, _mm_shuffle_epi32(vabsout13, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7p7x__sse41_c16_acc2()
337 …const __m128i vabsout4567 = _mm_blend_epi16(vabsout46, _mm_shuffle_epi32(vabsout57, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7p7x__sse41_c16_acc2()
D7p7x-minmax-sse41-c24-acc2.c399 …const __m128i vabsout0123 = _mm_blend_epi16(vabsout02, _mm_shuffle_epi32(vabsout13, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7p7x__sse41_c24_acc2()
400 …const __m128i vabsout4567 = _mm_blend_epi16(vabsout46, _mm_shuffle_epi32(vabsout57, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7p7x__sse41_c24_acc2()
401 …const __m128i vabsout89AB = _mm_blend_epi16(vabsout8A, _mm_shuffle_epi32(vabsout9B, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7p7x__sse41_c24_acc2()
402 …const __m128i vabsoutCDEF = _mm_blend_epi16(vabsoutCE, _mm_shuffle_epi32(vabsoutDF, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7p7x__sse41_c24_acc2()
403 …const __m128i vabsoutGHIJ = _mm_blend_epi16(vabsoutGI, _mm_shuffle_epi32(vabsoutHJ, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7p7x__sse41_c24_acc2()
404 …const __m128i vabsoutKLMN = _mm_blend_epi16(vabsoutKM, _mm_shuffle_epi32(vabsoutLN, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7p7x__sse41_c24_acc2()
481 …const __m128i vabsout0123 = _mm_blend_epi16(vabsout02, _mm_shuffle_epi32(vabsout13, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7p7x__sse41_c24_acc2()
482 …const __m128i vabsout4567 = _mm_blend_epi16(vabsout46, _mm_shuffle_epi32(vabsout57, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7p7x__sse41_c24_acc2()
D7p7x-minmax-sse41-c8-acc2.c201 …const __m128i vabsout0123 = _mm_blend_epi16(vabsout02, _mm_shuffle_epi32(vabsout13, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7p7x__sse41_c8_acc2()
202 …const __m128i vabsout4567 = _mm_blend_epi16(vabsout46, _mm_shuffle_epi32(vabsout57, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7p7x__sse41_c8_acc2()
269 …const __m128i vabsout0123 = _mm_blend_epi16(vabsout02, _mm_shuffle_epi32(vabsout13, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7p7x__sse41_c8_acc2()
270 …const __m128i vabsout4567 = _mm_blend_epi16(vabsout46, _mm_shuffle_epi32(vabsout57, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7p7x__sse41_c8_acc2()
/external/pdfium/third_party/libopenjpeg20/
Dmct.c252 y = _mm_blend_epi16(lo, hi, 0xCC); in opj_mct_encode_real()
262 y = _mm_add_epi32(y, _mm_blend_epi16(lo, hi, 0xCC)); in opj_mct_encode_real()
272 y = _mm_add_epi32(y, _mm_blend_epi16(lo, hi, 0xCC)); in opj_mct_encode_real()
287 u = _mm_blend_epi16(lo, hi, 0xCC); in opj_mct_encode_real()
297 u = _mm_sub_epi32(u, _mm_blend_epi16(lo, hi, 0xCC)); in opj_mct_encode_real()
307 u = _mm_sub_epi32(u, _mm_blend_epi16(lo, hi, 0xCC)); in opj_mct_encode_real()
322 v = _mm_blend_epi16(lo, hi, 0xCC); in opj_mct_encode_real()
332 v = _mm_sub_epi32(v, _mm_blend_epi16(lo, hi, 0xCC)); in opj_mct_encode_real()
342 v = _mm_sub_epi32(v, _mm_blend_epi16(lo, hi, 0xCC)); in opj_mct_encode_real()
/external/gemmlowp/internal/
Dpack_sse.h80 __m128i xmm9 = _mm_blend_epi16(xmm5, xmm7, 0xcc); in Pack()
81 __m128i xmm10 = _mm_blend_epi16(xmm8, xmm6, 0xcc); in Pack()
/external/llvm-project/clang/test/CodeGen/
Dppc-smmintrin.c66 _mm_blend_epi16(m1, m2, 0); in test_blend()
/external/XNNPACK/src/qs8-gemm/gen/
D4x4c2-minmax-xop-ld128.c221 const __m128i vq31prod0x0123 = _mm_blend_epi16(vq31prod0x02, vq31prod0x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_4x4c2__xop_ld128()
222 const __m128i vq31prod1x0123 = _mm_blend_epi16(vq31prod1x02, vq31prod1x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_4x4c2__xop_ld128()
223 const __m128i vq31prod2x0123 = _mm_blend_epi16(vq31prod2x02, vq31prod2x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_4x4c2__xop_ld128()
224 const __m128i vq31prod3x0123 = _mm_blend_epi16(vq31prod3x02, vq31prod3x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_4x4c2__xop_ld128()
D4x4c2-minmax-xop-ld64.c221 const __m128i vq31prod0x0123 = _mm_blend_epi16(vq31prod0x02, vq31prod0x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_4x4c2__xop_ld64()
222 const __m128i vq31prod1x0123 = _mm_blend_epi16(vq31prod1x02, vq31prod1x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_4x4c2__xop_ld64()
223 const __m128i vq31prod2x0123 = _mm_blend_epi16(vq31prod2x02, vq31prod2x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_4x4c2__xop_ld64()
224 const __m128i vq31prod3x0123 = _mm_blend_epi16(vq31prod3x02, vq31prod3x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_4x4c2__xop_ld64()
D4x4c2-xw-minmax-xop.c214 const __m128i vq31prod0x0123 = _mm_blend_epi16(vq31prod0x02, vq31prod0x13, 0xCC); in xnn_qs8_gemm_xw_minmax_ukernel_4x4c2__xop()
215 const __m128i vq31prod1x0123 = _mm_blend_epi16(vq31prod1x02, vq31prod1x13, 0xCC); in xnn_qs8_gemm_xw_minmax_ukernel_4x4c2__xop()
216 const __m128i vq31prod2x0123 = _mm_blend_epi16(vq31prod2x02, vq31prod2x13, 0xCC); in xnn_qs8_gemm_xw_minmax_ukernel_4x4c2__xop()
217 const __m128i vq31prod3x0123 = _mm_blend_epi16(vq31prod3x02, vq31prod3x13, 0xCC); in xnn_qs8_gemm_xw_minmax_ukernel_4x4c2__xop()
D4x4c2-xw-minmax-sse41.c209 const __m128i vq31prod0x0123 = _mm_blend_epi16(vq31prod0x02, vq31prod0x13, 0xCC); in xnn_qs8_gemm_xw_minmax_ukernel_4x4c2__sse41()
210 const __m128i vq31prod1x0123 = _mm_blend_epi16(vq31prod1x02, vq31prod1x13, 0xCC); in xnn_qs8_gemm_xw_minmax_ukernel_4x4c2__sse41()
211 const __m128i vq31prod2x0123 = _mm_blend_epi16(vq31prod2x02, vq31prod2x13, 0xCC); in xnn_qs8_gemm_xw_minmax_ukernel_4x4c2__sse41()
212 const __m128i vq31prod3x0123 = _mm_blend_epi16(vq31prod3x02, vq31prod3x13, 0xCC); in xnn_qs8_gemm_xw_minmax_ukernel_4x4c2__sse41()
D3x4c8-minmax-xop-ld128.c147 const __m128i vq31prod0x0123 = _mm_blend_epi16(vq31prod0x02, vq31prod0x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_3x4c8__xop_ld128()
148 const __m128i vq31prod1x0123 = _mm_blend_epi16(vq31prod1x02, vq31prod1x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_3x4c8__xop_ld128()
149 const __m128i vq31prod2x0123 = _mm_blend_epi16(vq31prod2x02, vq31prod2x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_3x4c8__xop_ld128()
D3x4c8-minmax-sse41-ld128.c142 const __m128i vq31prod0x0123 = _mm_blend_epi16(vq31prod0x02, vq31prod0x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_3x4c8__sse41_ld128()
143 const __m128i vq31prod1x0123 = _mm_blend_epi16(vq31prod1x02, vq31prod1x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_3x4c8__sse41_ld128()
144 const __m128i vq31prod2x0123 = _mm_blend_epi16(vq31prod2x02, vq31prod2x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_3x4c8__sse41_ld128()
D3x4c8-minmax-xop-ld64.c149 const __m128i vq31prod0x0123 = _mm_blend_epi16(vq31prod0x02, vq31prod0x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_3x4c8__xop_ld64()
150 const __m128i vq31prod1x0123 = _mm_blend_epi16(vq31prod1x02, vq31prod1x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_3x4c8__xop_ld64()
151 const __m128i vq31prod2x0123 = _mm_blend_epi16(vq31prod2x02, vq31prod2x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_3x4c8__xop_ld64()
D3x4c8-xw-minmax-xop.c145 const __m128i vq31prod0x0123 = _mm_blend_epi16(vq31prod0x02, vq31prod0x13, 0xCC); in xnn_qs8_gemm_xw_minmax_ukernel_3x4c8__xop()
146 const __m128i vq31prod1x0123 = _mm_blend_epi16(vq31prod1x02, vq31prod1x13, 0xCC); in xnn_qs8_gemm_xw_minmax_ukernel_3x4c8__xop()
147 const __m128i vq31prod2x0123 = _mm_blend_epi16(vq31prod2x02, vq31prod2x13, 0xCC); in xnn_qs8_gemm_xw_minmax_ukernel_3x4c8__xop()
D3x4c8-minmax-sse41-ld64.c144 const __m128i vq31prod0x0123 = _mm_blend_epi16(vq31prod0x02, vq31prod0x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_3x4c8__sse41_ld64()
145 const __m128i vq31prod1x0123 = _mm_blend_epi16(vq31prod1x02, vq31prod1x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_3x4c8__sse41_ld64()
146 const __m128i vq31prod2x0123 = _mm_blend_epi16(vq31prod2x02, vq31prod2x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_3x4c8__sse41_ld64()
/external/XNNPACK/src/qs8-igemm/gen/
D3x4c8-minmax-xop-ld64.c166 const __m128i vq31prod0x0123 = _mm_blend_epi16(vq31prod0x02, vq31prod0x13, 0xCC); in xnn_qs8_igemm_minmax_ukernel_3x4c8__xop_ld64()
167 const __m128i vq31prod1x0123 = _mm_blend_epi16(vq31prod1x02, vq31prod1x13, 0xCC); in xnn_qs8_igemm_minmax_ukernel_3x4c8__xop_ld64()
168 const __m128i vq31prod2x0123 = _mm_blend_epi16(vq31prod2x02, vq31prod2x13, 0xCC); in xnn_qs8_igemm_minmax_ukernel_3x4c8__xop_ld64()

1234