Home
last modified time | relevance | path

Searched refs:_mm_cvtsi128_si32 (Results 1 – 25 of 399) sorted by relevance

12345678910>>...16

/external/XNNPACK/src/x8-zip/
Dxm-sse2.c54 *((uint32_t*) output) = _mm_cvtsi128_si32(vxyzw0); in xnn_x8_zip_xm_ukernel__sse2()
57 *((uint32_t*) output) = _mm_cvtsi128_si32(vxyzw0); in xnn_x8_zip_xm_ukernel__sse2()
60 *((uint32_t*) output) = _mm_cvtsi128_si32(vxyzw0); in xnn_x8_zip_xm_ukernel__sse2()
63 *((uint32_t*) output) = _mm_cvtsi128_si32(vxyzw0); in xnn_x8_zip_xm_ukernel__sse2()
66 *((uint32_t*) output) = _mm_cvtsi128_si32(vxyzw1); in xnn_x8_zip_xm_ukernel__sse2()
69 *((uint32_t*) output) = _mm_cvtsi128_si32(vxyzw1); in xnn_x8_zip_xm_ukernel__sse2()
72 *((uint32_t*) output) = _mm_cvtsi128_si32(vxyzw1); in xnn_x8_zip_xm_ukernel__sse2()
75 *((uint32_t*) output) = _mm_cvtsi128_si32(vxyzw1); in xnn_x8_zip_xm_ukernel__sse2()
78 *((uint32_t*) output) = _mm_cvtsi128_si32(vxyzw2); in xnn_x8_zip_xm_ukernel__sse2()
81 *((uint32_t*) output) = _mm_cvtsi128_si32(vxyzw2); in xnn_x8_zip_xm_ukernel__sse2()
[all …]
/external/libmpeg2/common/x86/
Dideint_cac_ssse3.c136 sum_1 = _mm_cvtsi128_si32(sum_t[i + 0]); in ideint_cac_8x8_ssse3()
137 sum_2 = _mm_cvtsi128_si32(sum_b[i + 0]); in ideint_cac_8x8_ssse3()
142 sum_3 = _mm_cvtsi128_si32(sum_t[i + 1]); in ideint_cac_8x8_ssse3()
143 sum_4 = _mm_cvtsi128_si32(sum_b[i + 1]); in ideint_cac_8x8_ssse3()
151 sum_1 = _mm_cvtsi128_si32(_mm_srli_si128(sum_t[i + 0], 8)); in ideint_cac_8x8_ssse3()
152 sum_2 = _mm_cvtsi128_si32(_mm_srli_si128(sum_b[i + 0], 8)); in ideint_cac_8x8_ssse3()
157 sum_3 = _mm_cvtsi128_si32(_mm_srli_si128(sum_t[i + 1], 8)); in ideint_cac_8x8_ssse3()
158 sum_4 = _mm_cvtsi128_si32(_mm_srli_si128(sum_b[i + 1], 8)); in ideint_cac_8x8_ssse3()
221 diff_sum = _mm_cvtsi128_si32(diff); in ideint_cac_8x8_ssse3()
224 diff_sum = _mm_cvtsi128_si32(_mm_srli_si128(diff, 8)); in ideint_cac_8x8_ssse3()
Dideint_spatial_filter_ssse3.c161 adiff[0] = _mm_cvtsi128_si32(diffs[0]); in ideint_spatial_filter_ssse3()
162 adiff[1] = _mm_cvtsi128_si32(diffs[1]); in ideint_spatial_filter_ssse3()
163 adiff[2] = _mm_cvtsi128_si32(diffs[2]); in ideint_spatial_filter_ssse3()
164 adiff[3] = _mm_cvtsi128_si32(_mm_srli_si128(diffs[0], 8)); in ideint_spatial_filter_ssse3()
165 adiff[4] = _mm_cvtsi128_si32(_mm_srli_si128(diffs[1], 8)); in ideint_spatial_filter_ssse3()
166 adiff[5] = _mm_cvtsi128_si32(_mm_srli_si128(diffs[2], 8)); in ideint_spatial_filter_ssse3()
/external/libaom/libaom/aom_dsp/x86/
Djnt_sad_ssse3.c57 _mm_cvtsi128_si32(sad) + _mm_cvtsi128_si32(_mm_srli_si128(sad, 8)); in aom_sad4xh_sse2()
88 _mm_cvtsi128_si32(sad) + _mm_cvtsi128_si32(_mm_srli_si128(sad, 8)); in aom_sad8xh_sse2()
112 _mm_cvtsi128_si32(sad) + _mm_cvtsi128_si32(_mm_srli_si128(sad, 8)); in aom_sad16xh_sse2()
138 _mm_cvtsi128_si32(sad) + _mm_cvtsi128_si32(_mm_srli_si128(sad, 8)); in aom_sad32xh_sse2()
164 _mm_cvtsi128_si32(sad) + _mm_cvtsi128_si32(_mm_srli_si128(sad, 8)); in aom_sad64xh_sse2()
190 _mm_cvtsi128_si32(sad) + _mm_cvtsi128_si32(_mm_srli_si128(sad, 8)); in aom_sad128xh_sse2()
Dintrapred_sse2.c115 uint32_t sum = _mm_cvtsi128_si32(sum_above); in aom_dc_predictor_4x8_sse2()
120 const uint32_t pred = _mm_cvtsi128_si32(row); in aom_dc_predictor_4x8_sse2()
130 uint32_t sum = _mm_cvtsi128_si32(sum_above); in aom_dc_predictor_4x16_sse2()
135 const uint32_t pred = _mm_cvtsi128_si32(row); in aom_dc_predictor_4x16_sse2()
145 uint32_t sum = _mm_cvtsi128_si32(sum_above); in aom_dc_predictor_8x4_sse2()
159 uint32_t sum = _mm_cvtsi128_si32(sum_above); in aom_dc_predictor_8x16_sse2()
172 uint32_t sum = _mm_cvtsi128_si32(sum_above); in aom_dc_predictor_8x32_sse2()
185 uint32_t sum = _mm_cvtsi128_si32(sum_above); in aom_dc_predictor_16x4_sse2()
198 uint32_t sum = _mm_cvtsi128_si32(sum_above); in aom_dc_predictor_16x8_sse2()
211 uint32_t sum = _mm_cvtsi128_si32(sum_above); in aom_dc_predictor_16x32_sse2()
[all …]
Dsad_avx2.c41 res = _mm_cvtsi128_si32(sum_sad128); \
74 res = _mm_cvtsi128_si32(sum_sad128); \
128 res = _mm_cvtsi128_si32(sum_sad128); \
167 res = _mm_cvtsi128_si32(sum_sad128); \
Dobmc_variance_avx2.c79 *sum = _mm_cvtsi128_si32(v_d); in obmc_variance_w8n()
80 *sse = _mm_cvtsi128_si32(_mm_srli_si128(v_d, 4)); in obmc_variance_w8n()
149 *sum = _mm_cvtsi128_si32(res0); in obmc_variance_w16n()
150 *sse = _mm_cvtsi128_si32(_mm_srli_si128(res0, 4)); in obmc_variance_w16n()
Dmasked_sad_intrin_ssse3.c136 _mm_cvtsi128_si32(res) + _mm_cvtsi128_si32(_mm_srli_si128(res, 8)); in masked_sad_ssse3()
181 _mm_cvtsi128_si32(res) + _mm_cvtsi128_si32(_mm_srli_si128(res, 8)); in aom_masked_sad8xh_ssse3()
225 int32_t sad = _mm_cvtsi128_si32(res); in aom_masked_sad4xh_ssse3()
340 int sad = _mm_cvtsi128_si32(res); in highbd_masked_sad_ssse3()
400 int sad = _mm_cvtsi128_si32(res); in aom_highbd_masked_sad4xh_ssse3()
/external/neven/Embedded/common/src/b_BasicEm/
DMathSSE2.c99 resultL = _mm_cvtsi128_si32( m_XMM7 ); in bbs_dotProduct_64SSE2()
190 resultL = _mm_cvtsi128_si32( m_XMM0 ); /* 1st 32bits */ in bbs_dotProduct_128SSE2()
194 resultL += _mm_cvtsi128_si32( m_XMM0 ); /* 2nd 32bits */ in bbs_dotProduct_128SSE2()
198 resultL += _mm_cvtsi128_si32( m_XMM0 ); /* 3rd 32bits */ in bbs_dotProduct_128SSE2()
202 resultL += _mm_cvtsi128_si32( m_XMM0 ); /* 4th 32bits */ in bbs_dotProduct_128SSE2()
294 resultL = _mm_cvtsi128_si32( m_XMM0 ); /* 1st 32bits */ in bbs_dotProduct_u128SSE2()
298 resultL += _mm_cvtsi128_si32( m_XMM0 ); /* 2nd 32bits */ in bbs_dotProduct_u128SSE2()
302 resultL += _mm_cvtsi128_si32( m_XMM0 ); /* 3rd 32bits */ in bbs_dotProduct_u128SSE2()
306 resultL += _mm_cvtsi128_si32( m_XMM0 ); /* 4th 32bits */ in bbs_dotProduct_u128SSE2()
/external/webp/src/dsp/
Ddec_sse2.c190 WebPUint32ToMem(dst + 0 * BPS, _mm_cvtsi128_si32(dst0)); in Transform_SSE2()
191 WebPUint32ToMem(dst + 1 * BPS, _mm_cvtsi128_si32(dst1)); in Transform_SSE2()
192 WebPUint32ToMem(dst + 2 * BPS, _mm_cvtsi128_si32(dst2)); in Transform_SSE2()
193 WebPUint32ToMem(dst + 3 * BPS, _mm_cvtsi128_si32(dst3)); in Transform_SSE2()
236 WebPUint32ToMem(dst + 0 * BPS, _mm_cvtsi128_si32(dst0)); in TransformAC3()
237 WebPUint32ToMem(dst + 1 * BPS, _mm_cvtsi128_si32(dst1)); in TransformAC3()
238 WebPUint32ToMem(dst + 2 * BPS, _mm_cvtsi128_si32(dst2)); in TransformAC3()
239 WebPUint32ToMem(dst + 3 * BPS, _mm_cvtsi128_si32(dst3)); in TransformAC3()
543 WebPUint32ToMem(dst, _mm_cvtsi128_si32(*x)); in Store4x4_SSE2()
911 const uint32_t vals = _mm_cvtsi128_si32(avg); in VE4_SSE2()
[all …]
Denc_sse2.c188 WebPUint32ToMem(&dst[0 * BPS], _mm_cvtsi128_si32(ref0)); in ITransform_SSE2()
189 WebPUint32ToMem(&dst[1 * BPS], _mm_cvtsi128_si32(ref1)); in ITransform_SSE2()
190 WebPUint32ToMem(&dst[2 * BPS], _mm_cvtsi128_si32(ref2)); in ITransform_SSE2()
191 WebPUint32ToMem(&dst[3 * BPS], _mm_cvtsi128_si32(ref3)); in ITransform_SSE2()
633 const int DC = _mm_cvtsi128_si32(sum) + 4; in DC8uvNoLeft_SSE2()
725 const uint32_t vals = _mm_cvtsi128_si32(avg); in VE4_SSE2()
763 WebPUint32ToMem(dst + 0 * BPS, _mm_cvtsi128_si32( abcdefg )); in LD4_SSE2()
764 WebPUint32ToMem(dst + 1 * BPS, _mm_cvtsi128_si32(_mm_srli_si128(abcdefg, 1))); in LD4_SSE2()
765 WebPUint32ToMem(dst + 2 * BPS, _mm_cvtsi128_si32(_mm_srli_si128(abcdefg, 2))); in LD4_SSE2()
766 WebPUint32ToMem(dst + 3 * BPS, _mm_cvtsi128_si32(_mm_srli_si128(abcdefg, 3))); in LD4_SSE2()
[all …]
/external/libvpx/libvpx/vpx_dsp/x86/
Dmem_sse2.h36 const uint32_t val = _mm_cvtsi128_si32(v); in store_unaligned_u32()
108 *(int *)(d + 0 * stride) = _mm_cvtsi128_si32(s[0]); in store_8bit_4x4()
109 *(int *)(d + 1 * stride) = _mm_cvtsi128_si32(s[1]); in store_8bit_4x4()
110 *(int *)(d + 2 * stride) = _mm_cvtsi128_si32(s[2]); in store_8bit_4x4()
111 *(int *)(d + 3 * stride) = _mm_cvtsi128_si32(s[3]); in store_8bit_4x4()
Dsad_avx2.c38 res = _mm_cvtsi128_si32(sum_sad128); \
70 res = _mm_cvtsi128_si32(sum_sad128); \
121 res = _mm_cvtsi128_si32(sum_sad128); \
159 res = _mm_cvtsi128_si32(sum_sad128); \
/external/libaom/libaom/aom_dsp/simd/
Dv256_intrinsics_x86.h35 return (uint32_t)_mm_cvtsi128_si32(_mm256_extracti128_si256(a, 0)); in v256_low_u32()
399 return (int64_t)_mm_cvtsi128_si32(l) + in v256_dotp_s16()
400 (int64_t)_mm_cvtsi128_si32(_mm_srli_si128(l, 4)) + in v256_dotp_s16()
401 (int64_t)_mm_cvtsi128_si32(_mm_srli_si128(l, 8)) + in v256_dotp_s16()
402 (int64_t)_mm_cvtsi128_si32(_mm_srli_si128(l, 12)) + in v256_dotp_s16()
403 (int64_t)_mm_cvtsi128_si32(h) + in v256_dotp_s16()
404 (int64_t)_mm_cvtsi128_si32(_mm_srli_si128(h, 4)) + in v256_dotp_s16()
405 (int64_t)_mm_cvtsi128_si32(_mm_srli_si128(h, 8)) + in v256_dotp_s16()
406 (int64_t)_mm_cvtsi128_si32(_mm_srli_si128(h, 12)); in v256_dotp_s16()
422 return (int64_t)_mm_cvtsi128_si32(l) + in v256_dotp_s32()
[all …]
/external/flac/src/libFLAC/
Dfixed_intrin_ssse3.c108 total_error_0 = _mm_cvtsi128_si32(total_err0); in FLAC__fixed_compute_best_predictor_intrin_ssse3()
111 total_error_4 = _mm_cvtsi128_si32(total_err2); in FLAC__fixed_compute_best_predictor_intrin_ssse3()
112 total_error_2 = _mm_cvtsi128_si32(total_err1); in FLAC__fixed_compute_best_predictor_intrin_ssse3()
115 total_error_3 = _mm_cvtsi128_si32(total_err2); in FLAC__fixed_compute_best_predictor_intrin_ssse3()
116 total_error_1 = _mm_cvtsi128_si32(total_err1); in FLAC__fixed_compute_best_predictor_intrin_ssse3()
Dfixed_intrin_sse2.c114 total_error_0 = _mm_cvtsi128_si32(total_err0); in FLAC__fixed_compute_best_predictor_intrin_sse2()
117 total_error_4 = _mm_cvtsi128_si32(total_err2); in FLAC__fixed_compute_best_predictor_intrin_sse2()
118 total_error_2 = _mm_cvtsi128_si32(total_err1); in FLAC__fixed_compute_best_predictor_intrin_sse2()
121 total_error_3 = _mm_cvtsi128_si32(total_err2); in FLAC__fixed_compute_best_predictor_intrin_sse2()
122 total_error_1 = _mm_cvtsi128_si32(total_err1); in FLAC__fixed_compute_best_predictor_intrin_sse2()
/external/libhevc/common/x86/
Dihevc_chroma_intra_pred_filters_sse42_intr.c345 acc_dc_u = _mm_cvtsi128_si32(src_temp4); in ihevc_intra_pred_chroma_dc_sse42()
346 acc_dc_v = _mm_cvtsi128_si32(temp_sad); in ihevc_intra_pred_chroma_dc_sse42()
374 acc_dc_u = _mm_cvtsi128_si32(src_temp4); in ihevc_intra_pred_chroma_dc_sse42()
375 acc_dc_v = _mm_cvtsi128_si32(temp_sad); in ihevc_intra_pred_chroma_dc_sse42()
395 acc_dc_u = _mm_cvtsi128_si32(src_temp4); in ihevc_intra_pred_chroma_dc_sse42()
396 acc_dc_v = _mm_cvtsi128_si32(temp_sad); in ihevc_intra_pred_chroma_dc_sse42()
Dihevc_deblk_ssse3_intr.c187 d0 = _mm_cvtsi128_si32(mask_d_result_4x32b); in ihevc_deblk_luma_vert_ssse3()
188 d3 = _mm_cvtsi128_si32(temp_coef0_8x16b); in ihevc_deblk_luma_vert_ssse3()
189 dp = _mm_cvtsi128_si32(temp_coef1_8x16b); in ihevc_deblk_luma_vert_ssse3()
190 dq = _mm_cvtsi128_si32(mask_16x8b); in ihevc_deblk_luma_vert_ssse3()
199 de_0 = _mm_cvtsi128_si32(mask_de_result_8x16b); in ihevc_deblk_luma_vert_ssse3()
200 de_1 = _mm_cvtsi128_si32(temp_coef0_8x16b); in ihevc_deblk_luma_vert_ssse3()
201 de_2 = _mm_cvtsi128_si32(temp_coef1_8x16b); in ihevc_deblk_luma_vert_ssse3()
202 de_3 = _mm_cvtsi128_si32(mask_16x8b); in ihevc_deblk_luma_vert_ssse3()
647 d0 = _mm_cvtsi128_si32(mask_d_result_4x32b); in ihevc_deblk_luma_horz_ssse3()
648 d3 = _mm_cvtsi128_si32(temp_coef0_8x16b); in ihevc_deblk_luma_horz_ssse3()
[all …]
/external/libavc/common/x86/
Dih264_luma_intra_pred_filters_ssse3.c357 row1 = _mm_cvtsi128_si32(res_16x8b); in ih264_intra_pred_luma_4x4_mode_diag_dl_ssse3()
359 row2 = _mm_cvtsi128_si32(res_16x8b); in ih264_intra_pred_luma_4x4_mode_diag_dl_ssse3()
361 row3 = _mm_cvtsi128_si32(res_16x8b); in ih264_intra_pred_luma_4x4_mode_diag_dl_ssse3()
363 row4 = _mm_cvtsi128_si32(res_16x8b); in ih264_intra_pred_luma_4x4_mode_diag_dl_ssse3()
445 row1 = _mm_cvtsi128_si32(res2_16x8b); in ih264_intra_pred_luma_4x4_mode_diag_dr_ssse3()
447 row2 = _mm_cvtsi128_si32(res2_16x8b); in ih264_intra_pred_luma_4x4_mode_diag_dr_ssse3()
449 row3 = _mm_cvtsi128_si32(res2_16x8b); in ih264_intra_pred_luma_4x4_mode_diag_dr_ssse3()
450 row4 = _mm_cvtsi128_si32(res1_16x8b); in ih264_intra_pred_luma_4x4_mode_diag_dr_ssse3()
539 row1 = _mm_cvtsi128_si32(row1_16x8b); in ih264_intra_pred_luma_4x4_mode_vert_r_ssse3()
540 row2 = _mm_cvtsi128_si32(row2_16x8b); in ih264_intra_pred_luma_4x4_mode_vert_r_ssse3()
[all …]
Dih264_weighted_pred_sse42.c118 *((WORD32 *)(pu1_dst)) = _mm_cvtsi128_si32(y0_0_16x8b); in ih264_default_weighted_pred_luma_sse42()
119 *((WORD32 *)(pu1_dst + dst_strd)) = _mm_cvtsi128_si32(y0_1_16x8b); in ih264_default_weighted_pred_luma_sse42()
120 *((WORD32 *)(pu1_dst + (dst_strd << 1))) = _mm_cvtsi128_si32(y0_2_16x8b); in ih264_default_weighted_pred_luma_sse42()
121 *((WORD32 *)(pu1_dst + dst_strd * 3)) = _mm_cvtsi128_si32(y0_3_16x8b); in ih264_default_weighted_pred_luma_sse42()
273 *((WORD32 *)(pu1_dst)) = _mm_cvtsi128_si32(uv0_0_16x8b); in ih264_default_weighted_pred_chroma_sse42()
274 *((WORD32 *)(pu1_dst + dst_strd)) = _mm_cvtsi128_si32(uv0_1_16x8b); in ih264_default_weighted_pred_chroma_sse42()
436 *((WORD32 *)(pu1_dst)) = _mm_cvtsi128_si32(y_0_16x8b); in ih264_weighted_pred_luma_sse42()
437 *((WORD32 *)(pu1_dst + dst_strd)) = _mm_cvtsi128_si32(y_1_16x8b); in ih264_weighted_pred_luma_sse42()
438 *((WORD32 *)(pu1_dst + (dst_strd << 1))) = _mm_cvtsi128_si32(y_2_16x8b); in ih264_weighted_pred_luma_sse42()
439 *((WORD32 *)(pu1_dst + dst_strd * 3)) = _mm_cvtsi128_si32(y_3_16x8b); in ih264_weighted_pred_luma_sse42()
[all …]
/external/libaom/libaom/av1/common/x86/
Dconvolve_avx2.c134 *(uint16_t *)p_0 = (uint16_t)_mm_cvtsi128_si32(res_0); in av1_convolve_y_sr_avx2()
135 *(uint16_t *)p_1 = (uint16_t)_mm_cvtsi128_si32(res_1); in av1_convolve_y_sr_avx2()
247 *(uint16_t *)p_0 = (uint16_t)_mm_cvtsi128_si32(res_0); in av1_convolve_y_sr_avx2()
248 *(uint16_t *)p_1 = (uint16_t)_mm_cvtsi128_si32(res_1); in av1_convolve_y_sr_avx2()
332 *(uint16_t *)p_0 = (uint16_t)_mm_cvtsi128_si32(res_0); in av1_convolve_x_sr_avx2()
333 *(uint16_t *)p_1 = (uint16_t)_mm_cvtsi128_si32(res_1); in av1_convolve_x_sr_avx2()
404 *(uint16_t *)p_0 = _mm_cvtsi128_si32(res_0); in av1_convolve_x_sr_avx2()
405 *(uint16_t *)p_1 = _mm_cvtsi128_si32(res_1); in av1_convolve_x_sr_avx2()
/external/XNNPACK/src/f32-velu/gen/
Dvelu-avx-rr2-lut16-p3-x40.c83 …const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx0_lo))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x40()
85 …const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx0_hi))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x40()
112 …const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx1_lo))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x40()
114 …const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx1_hi))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x40()
141 …const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx2_lo))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x40()
143 …const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx2_hi))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x40()
170 …const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx3_lo))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x40()
172 …const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx3_hi))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x40()
199 …const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx4_lo))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x40()
201 …const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx4_hi))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x40()
[all …]
Dvelu-avx-rr2-lut16-p3-x32.c80 …const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx0_lo))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x32()
82 …const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx0_hi))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x32()
109 …const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx1_lo))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x32()
111 …const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx1_hi))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x32()
138 …const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx2_lo))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x32()
140 …const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx2_hi))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x32()
167 …const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx3_lo))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x32()
169 …const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx3_hi))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x32()
284 …(const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx_lo))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x32()
286 …(const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx_hi))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x32()
[all …]
Dvelu-avx-rr2-lut16-p3-x48.c86 …const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx0_lo))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x48()
88 …const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx0_hi))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x48()
115 …const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx1_lo))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x48()
117 …const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx1_hi))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x48()
144 …const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx2_lo))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x48()
146 …const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx2_hi))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x48()
173 …const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx3_lo))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x48()
175 …const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx3_hi))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x48()
202 …const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx4_lo))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x48()
204 …const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx4_hi))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x48()
[all …]
Dvelu-avx-rr2-lut16-p3-x24.c77 …const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx0_lo))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x24()
79 …const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx0_hi))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x24()
106 …const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx1_lo))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x24()
108 …const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx1_hi))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x24()
135 …const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx2_lo))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x24()
137 …const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx2_hi))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x24()
235 …(const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx_lo))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x24()
237 …(const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx_hi))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x24()
301 …(const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx_lo))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x24()
303 …(const void*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) _mm_cvtsi128_si32(vidx_hi))); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x24()

12345678910>>...16