/external/python/cpython3/Modules/_blake2/impl/ |
D | blake2s-load-sse41.h | 30 t0 = _mm_blend_epi16(m1, m2, 0x0C); \ 32 t2 = _mm_blend_epi16(t0, t1, 0xF0); \ 37 t1 = _mm_blend_epi16(m1,m3,0xC0); \ 38 t2 = _mm_blend_epi16(t0, t1, 0xF0); \ 43 t1 = _mm_blend_epi16(m2, t0, 0x30); \ 44 t2 = _mm_blend_epi16(m0, t1, 0xF0); \ 50 t2 = _mm_blend_epi16(t0, t1, 0x0C); \ 55 t1 = _mm_blend_epi16(m3,m1,0x0C); \ 56 t2 = _mm_blend_epi16(t0, t1, 0x0F); \ 61 t1 = _mm_blend_epi16(t0, m0, 0xF0); \ [all …]
|
D | blake2b-load-sse41.h | 93 b1 = _mm_blend_epi16(m1, m6, 0xF0); \ 100 b0 = _mm_blend_epi16(m5, m1, 0xF0); \ 132 b0 = _mm_blend_epi16(m1, m2, 0xF0); \ 133 b1 = _mm_blend_epi16(m2, m7, 0xF0); \ 156 b0 = _mm_blend_epi16(m0, m3, 0xF0); \ 157 b1 = _mm_blend_epi16(m2, m7, 0xF0); \ 164 b0 = _mm_blend_epi16(m7, m5, 0xF0); \ 165 b1 = _mm_blend_epi16(m3, m1, 0xF0); \ 173 b1 = _mm_blend_epi16(m4, m6, 0xF0); \ 196 b0 = _mm_blend_epi16(m2, m3, 0xF0); \ [all …]
|
/external/XNNPACK/src/qs8-requantization/ |
D | precise-sse4.c | 83 const __m128i x_abs_scaled = _mm_blend_epi16(x_abs_scaled02, x_abs_scaled13, 0xCC); in xnn_qs8_requantize_precise__sse4() 84 const __m128i y_abs_scaled = _mm_blend_epi16(y_abs_scaled02, y_abs_scaled13, 0xCC); in xnn_qs8_requantize_precise__sse4() 85 const __m128i z_abs_scaled = _mm_blend_epi16(z_abs_scaled02, z_abs_scaled13, 0xCC); in xnn_qs8_requantize_precise__sse4() 86 const __m128i w_abs_scaled = _mm_blend_epi16(w_abs_scaled02, w_abs_scaled13, 0xCC); in xnn_qs8_requantize_precise__sse4()
|
D | q31-sse4.c | 86 const __m128i x_q31product = _mm_blend_epi16(x_q31product_even, x_q31product_odd, 0xCC); in xnn_qs8_requantize_q31__sse4() 87 const __m128i y_q31product = _mm_blend_epi16(y_q31product_even, y_q31product_odd, 0xCC); in xnn_qs8_requantize_q31__sse4() 88 const __m128i z_q31product = _mm_blend_epi16(z_q31product_even, z_q31product_odd, 0xCC); in xnn_qs8_requantize_q31__sse4() 89 const __m128i w_q31product = _mm_blend_epi16(w_q31product_even, w_q31product_odd, 0xCC); in xnn_qs8_requantize_q31__sse4()
|
/external/XNNPACK/src/qu8-requantization/ |
D | precise-sse4.c | 83 const __m128i x_abs_scaled = _mm_blend_epi16(x_abs_scaled02, x_abs_scaled13, 0xCC); in xnn_qu8_requantize_precise__sse4() 84 const __m128i y_abs_scaled = _mm_blend_epi16(y_abs_scaled02, y_abs_scaled13, 0xCC); in xnn_qu8_requantize_precise__sse4() 85 const __m128i z_abs_scaled = _mm_blend_epi16(z_abs_scaled02, z_abs_scaled13, 0xCC); in xnn_qu8_requantize_precise__sse4() 86 const __m128i w_abs_scaled = _mm_blend_epi16(w_abs_scaled02, w_abs_scaled13, 0xCC); in xnn_qu8_requantize_precise__sse4()
|
D | q31-sse4.c | 86 const __m128i x_q31product = _mm_blend_epi16(x_q31product_even, x_q31product_odd, 0xCC); in xnn_qu8_requantize_q31__sse4() 87 const __m128i y_q31product = _mm_blend_epi16(y_q31product_even, y_q31product_odd, 0xCC); in xnn_qu8_requantize_q31__sse4() 88 const __m128i z_q31product = _mm_blend_epi16(z_q31product_even, z_q31product_odd, 0xCC); in xnn_qu8_requantize_q31__sse4() 89 const __m128i w_q31product = _mm_blend_epi16(w_q31product_even, w_q31product_odd, 0xCC); in xnn_qu8_requantize_q31__sse4()
|
/external/XNNPACK/src/qs8-gavgpool/gen/ |
D | 7x-minmax-sse41-c24-acc2.c | 160 …const __m128i vabsout0123 = _mm_blend_epi16(vabsout02, _mm_shuffle_epi32(vabsout13, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7x__sse41_c24_acc2() 161 …const __m128i vabsout4567 = _mm_blend_epi16(vabsout46, _mm_shuffle_epi32(vabsout57, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7x__sse41_c24_acc2() 162 …const __m128i vabsout89AB = _mm_blend_epi16(vabsout8A, _mm_shuffle_epi32(vabsout9B, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7x__sse41_c24_acc2() 163 …const __m128i vabsoutCDEF = _mm_blend_epi16(vabsoutCE, _mm_shuffle_epi32(vabsoutDF, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7x__sse41_c24_acc2() 164 …const __m128i vabsoutGHIJ = _mm_blend_epi16(vabsoutGI, _mm_shuffle_epi32(vabsoutHJ, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7x__sse41_c24_acc2() 165 …const __m128i vabsoutKLMN = _mm_blend_epi16(vabsoutKM, _mm_shuffle_epi32(vabsoutLN, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7x__sse41_c24_acc2() 241 …const __m128i vabsout0123 = _mm_blend_epi16(vabsout02, _mm_shuffle_epi32(vabsout13, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7x__sse41_c24_acc2() 242 …const __m128i vabsout4567 = _mm_blend_epi16(vabsout46, _mm_shuffle_epi32(vabsout57, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7x__sse41_c24_acc2()
|
D | 7x-minmax-sse41-c16-acc2.c | 133 …const __m128i vabsout0123 = _mm_blend_epi16(vabsout02, _mm_shuffle_epi32(vabsout13, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7x__sse41_c16_acc2() 134 …const __m128i vabsout4567 = _mm_blend_epi16(vabsout46, _mm_shuffle_epi32(vabsout57, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7x__sse41_c16_acc2() 135 …const __m128i vabsout89AB = _mm_blend_epi16(vabsout8A, _mm_shuffle_epi32(vabsout9B, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7x__sse41_c16_acc2() 136 …const __m128i vabsoutCDEF = _mm_blend_epi16(vabsoutCE, _mm_shuffle_epi32(vabsoutDF, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7x__sse41_c16_acc2() 206 …const __m128i vabsout0123 = _mm_blend_epi16(vabsout02, _mm_shuffle_epi32(vabsout13, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7x__sse41_c16_acc2() 207 …const __m128i vabsout4567 = _mm_blend_epi16(vabsout46, _mm_shuffle_epi32(vabsout57, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7x__sse41_c16_acc2()
|
D | 7x-minmax-sse41-c8-acc2.c | 106 …const __m128i vabsout0123 = _mm_blend_epi16(vabsout02, _mm_shuffle_epi32(vabsout13, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7x__sse41_c8_acc2() 107 …const __m128i vabsout4567 = _mm_blend_epi16(vabsout46, _mm_shuffle_epi32(vabsout57, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7x__sse41_c8_acc2() 173 …const __m128i vabsout0123 = _mm_blend_epi16(vabsout02, _mm_shuffle_epi32(vabsout13, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7x__sse41_c8_acc2() 174 …const __m128i vabsout4567 = _mm_blend_epi16(vabsout46, _mm_shuffle_epi32(vabsout57, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7x__sse41_c8_acc2()
|
D | 7p7x-minmax-sse41-c16-acc2.c | 262 …const __m128i vabsout0123 = _mm_blend_epi16(vabsout02, _mm_shuffle_epi32(vabsout13, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7p7x__sse41_c16_acc2() 263 …const __m128i vabsout4567 = _mm_blend_epi16(vabsout46, _mm_shuffle_epi32(vabsout57, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7p7x__sse41_c16_acc2() 264 …const __m128i vabsout89AB = _mm_blend_epi16(vabsout8A, _mm_shuffle_epi32(vabsout9B, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7p7x__sse41_c16_acc2() 265 …const __m128i vabsoutCDEF = _mm_blend_epi16(vabsoutCE, _mm_shuffle_epi32(vabsoutDF, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7p7x__sse41_c16_acc2() 336 …const __m128i vabsout0123 = _mm_blend_epi16(vabsout02, _mm_shuffle_epi32(vabsout13, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7p7x__sse41_c16_acc2() 337 …const __m128i vabsout4567 = _mm_blend_epi16(vabsout46, _mm_shuffle_epi32(vabsout57, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7p7x__sse41_c16_acc2()
|
D | 7p7x-minmax-sse41-c24-acc2.c | 399 …const __m128i vabsout0123 = _mm_blend_epi16(vabsout02, _mm_shuffle_epi32(vabsout13, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7p7x__sse41_c24_acc2() 400 …const __m128i vabsout4567 = _mm_blend_epi16(vabsout46, _mm_shuffle_epi32(vabsout57, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7p7x__sse41_c24_acc2() 401 …const __m128i vabsout89AB = _mm_blend_epi16(vabsout8A, _mm_shuffle_epi32(vabsout9B, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7p7x__sse41_c24_acc2() 402 …const __m128i vabsoutCDEF = _mm_blend_epi16(vabsoutCE, _mm_shuffle_epi32(vabsoutDF, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7p7x__sse41_c24_acc2() 403 …const __m128i vabsoutGHIJ = _mm_blend_epi16(vabsoutGI, _mm_shuffle_epi32(vabsoutHJ, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7p7x__sse41_c24_acc2() 404 …const __m128i vabsoutKLMN = _mm_blend_epi16(vabsoutKM, _mm_shuffle_epi32(vabsoutLN, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7p7x__sse41_c24_acc2() 481 …const __m128i vabsout0123 = _mm_blend_epi16(vabsout02, _mm_shuffle_epi32(vabsout13, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7p7x__sse41_c24_acc2() 482 …const __m128i vabsout4567 = _mm_blend_epi16(vabsout46, _mm_shuffle_epi32(vabsout57, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7p7x__sse41_c24_acc2()
|
D | 7p7x-minmax-sse41-c8-acc2.c | 201 …const __m128i vabsout0123 = _mm_blend_epi16(vabsout02, _mm_shuffle_epi32(vabsout13, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7p7x__sse41_c8_acc2() 202 …const __m128i vabsout4567 = _mm_blend_epi16(vabsout46, _mm_shuffle_epi32(vabsout57, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7p7x__sse41_c8_acc2() 269 …const __m128i vabsout0123 = _mm_blend_epi16(vabsout02, _mm_shuffle_epi32(vabsout13, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7p7x__sse41_c8_acc2() 270 …const __m128i vabsout4567 = _mm_blend_epi16(vabsout46, _mm_shuffle_epi32(vabsout57, _MM_SHUFFLE(2,… in xnn_qs8_gavgpool_minmax_ukernel_7p7x__sse41_c8_acc2()
|
/external/pdfium/third_party/libopenjpeg20/ |
D | mct.c | 252 y = _mm_blend_epi16(lo, hi, 0xCC); in opj_mct_encode_real() 262 y = _mm_add_epi32(y, _mm_blend_epi16(lo, hi, 0xCC)); in opj_mct_encode_real() 272 y = _mm_add_epi32(y, _mm_blend_epi16(lo, hi, 0xCC)); in opj_mct_encode_real() 287 u = _mm_blend_epi16(lo, hi, 0xCC); in opj_mct_encode_real() 297 u = _mm_sub_epi32(u, _mm_blend_epi16(lo, hi, 0xCC)); in opj_mct_encode_real() 307 u = _mm_sub_epi32(u, _mm_blend_epi16(lo, hi, 0xCC)); in opj_mct_encode_real() 322 v = _mm_blend_epi16(lo, hi, 0xCC); in opj_mct_encode_real() 332 v = _mm_sub_epi32(v, _mm_blend_epi16(lo, hi, 0xCC)); in opj_mct_encode_real() 342 v = _mm_sub_epi32(v, _mm_blend_epi16(lo, hi, 0xCC)); in opj_mct_encode_real()
|
/external/gemmlowp/internal/ |
D | pack_sse.h | 80 __m128i xmm9 = _mm_blend_epi16(xmm5, xmm7, 0xcc); in Pack() 81 __m128i xmm10 = _mm_blend_epi16(xmm8, xmm6, 0xcc); in Pack()
|
/external/llvm-project/clang/test/CodeGen/ |
D | ppc-smmintrin.c | 66 _mm_blend_epi16(m1, m2, 0); in test_blend()
|
/external/XNNPACK/src/qs8-gemm/gen/ |
D | 4x4c2-minmax-xop-ld128.c | 221 const __m128i vq31prod0x0123 = _mm_blend_epi16(vq31prod0x02, vq31prod0x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_4x4c2__xop_ld128() 222 const __m128i vq31prod1x0123 = _mm_blend_epi16(vq31prod1x02, vq31prod1x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_4x4c2__xop_ld128() 223 const __m128i vq31prod2x0123 = _mm_blend_epi16(vq31prod2x02, vq31prod2x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_4x4c2__xop_ld128() 224 const __m128i vq31prod3x0123 = _mm_blend_epi16(vq31prod3x02, vq31prod3x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_4x4c2__xop_ld128()
|
D | 4x4c2-minmax-xop-ld64.c | 221 const __m128i vq31prod0x0123 = _mm_blend_epi16(vq31prod0x02, vq31prod0x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_4x4c2__xop_ld64() 222 const __m128i vq31prod1x0123 = _mm_blend_epi16(vq31prod1x02, vq31prod1x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_4x4c2__xop_ld64() 223 const __m128i vq31prod2x0123 = _mm_blend_epi16(vq31prod2x02, vq31prod2x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_4x4c2__xop_ld64() 224 const __m128i vq31prod3x0123 = _mm_blend_epi16(vq31prod3x02, vq31prod3x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_4x4c2__xop_ld64()
|
D | 4x4c2-xw-minmax-xop.c | 214 const __m128i vq31prod0x0123 = _mm_blend_epi16(vq31prod0x02, vq31prod0x13, 0xCC); in xnn_qs8_gemm_xw_minmax_ukernel_4x4c2__xop() 215 const __m128i vq31prod1x0123 = _mm_blend_epi16(vq31prod1x02, vq31prod1x13, 0xCC); in xnn_qs8_gemm_xw_minmax_ukernel_4x4c2__xop() 216 const __m128i vq31prod2x0123 = _mm_blend_epi16(vq31prod2x02, vq31prod2x13, 0xCC); in xnn_qs8_gemm_xw_minmax_ukernel_4x4c2__xop() 217 const __m128i vq31prod3x0123 = _mm_blend_epi16(vq31prod3x02, vq31prod3x13, 0xCC); in xnn_qs8_gemm_xw_minmax_ukernel_4x4c2__xop()
|
D | 4x4c2-xw-minmax-sse41.c | 209 const __m128i vq31prod0x0123 = _mm_blend_epi16(vq31prod0x02, vq31prod0x13, 0xCC); in xnn_qs8_gemm_xw_minmax_ukernel_4x4c2__sse41() 210 const __m128i vq31prod1x0123 = _mm_blend_epi16(vq31prod1x02, vq31prod1x13, 0xCC); in xnn_qs8_gemm_xw_minmax_ukernel_4x4c2__sse41() 211 const __m128i vq31prod2x0123 = _mm_blend_epi16(vq31prod2x02, vq31prod2x13, 0xCC); in xnn_qs8_gemm_xw_minmax_ukernel_4x4c2__sse41() 212 const __m128i vq31prod3x0123 = _mm_blend_epi16(vq31prod3x02, vq31prod3x13, 0xCC); in xnn_qs8_gemm_xw_minmax_ukernel_4x4c2__sse41()
|
D | 3x4c8-minmax-xop-ld128.c | 147 const __m128i vq31prod0x0123 = _mm_blend_epi16(vq31prod0x02, vq31prod0x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_3x4c8__xop_ld128() 148 const __m128i vq31prod1x0123 = _mm_blend_epi16(vq31prod1x02, vq31prod1x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_3x4c8__xop_ld128() 149 const __m128i vq31prod2x0123 = _mm_blend_epi16(vq31prod2x02, vq31prod2x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_3x4c8__xop_ld128()
|
D | 3x4c8-minmax-sse41-ld128.c | 142 const __m128i vq31prod0x0123 = _mm_blend_epi16(vq31prod0x02, vq31prod0x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_3x4c8__sse41_ld128() 143 const __m128i vq31prod1x0123 = _mm_blend_epi16(vq31prod1x02, vq31prod1x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_3x4c8__sse41_ld128() 144 const __m128i vq31prod2x0123 = _mm_blend_epi16(vq31prod2x02, vq31prod2x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_3x4c8__sse41_ld128()
|
D | 3x4c8-minmax-xop-ld64.c | 149 const __m128i vq31prod0x0123 = _mm_blend_epi16(vq31prod0x02, vq31prod0x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_3x4c8__xop_ld64() 150 const __m128i vq31prod1x0123 = _mm_blend_epi16(vq31prod1x02, vq31prod1x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_3x4c8__xop_ld64() 151 const __m128i vq31prod2x0123 = _mm_blend_epi16(vq31prod2x02, vq31prod2x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_3x4c8__xop_ld64()
|
D | 3x4c8-xw-minmax-xop.c | 145 const __m128i vq31prod0x0123 = _mm_blend_epi16(vq31prod0x02, vq31prod0x13, 0xCC); in xnn_qs8_gemm_xw_minmax_ukernel_3x4c8__xop() 146 const __m128i vq31prod1x0123 = _mm_blend_epi16(vq31prod1x02, vq31prod1x13, 0xCC); in xnn_qs8_gemm_xw_minmax_ukernel_3x4c8__xop() 147 const __m128i vq31prod2x0123 = _mm_blend_epi16(vq31prod2x02, vq31prod2x13, 0xCC); in xnn_qs8_gemm_xw_minmax_ukernel_3x4c8__xop()
|
D | 3x4c8-minmax-sse41-ld64.c | 144 const __m128i vq31prod0x0123 = _mm_blend_epi16(vq31prod0x02, vq31prod0x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_3x4c8__sse41_ld64() 145 const __m128i vq31prod1x0123 = _mm_blend_epi16(vq31prod1x02, vq31prod1x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_3x4c8__sse41_ld64() 146 const __m128i vq31prod2x0123 = _mm_blend_epi16(vq31prod2x02, vq31prod2x13, 0xCC); in xnn_qs8_gemm_minmax_ukernel_3x4c8__sse41_ld64()
|
/external/XNNPACK/src/qs8-igemm/gen/ |
D | 3x4c8-minmax-xop-ld64.c | 166 const __m128i vq31prod0x0123 = _mm_blend_epi16(vq31prod0x02, vq31prod0x13, 0xCC); in xnn_qs8_igemm_minmax_ukernel_3x4c8__xop_ld64() 167 const __m128i vq31prod1x0123 = _mm_blend_epi16(vq31prod1x02, vq31prod1x13, 0xCC); in xnn_qs8_igemm_minmax_ukernel_3x4c8__xop_ld64() 168 const __m128i vq31prod2x0123 = _mm_blend_epi16(vq31prod2x02, vq31prod2x13, 0xCC); in xnn_qs8_igemm_minmax_ukernel_3x4c8__xop_ld64()
|