Home
last modified time | relevance | path

Searched refs:_mm_set1_epi32 (Results 1 – 25 of 233) sorted by relevance

12345678910

/external/libaom/libaom/av1/common/x86/
Dhighbd_inv_txfm_sse4.c42 __m128i rnding = _mm_set1_epi32(1 << (shift - 1)); in round_shift_4x4()
87 __m128i max_clip_val = _mm_set1_epi32((1 << bd) - 1); in highbd_get_recon_8x8_sse4_1()
166 __m128i offset = _mm_set1_epi32((1 << shift) >> 1); in shift_and_clamp_sse4_1()
356 const __m128i clamp_lo_out = _mm_set1_epi32(-(1 << (log_range_out - 1))); in idct32_stage9_sse4_1()
357 const __m128i clamp_hi_out = _mm_set1_epi32((1 << (log_range_out - 1)) - 1); in idct32_stage9_sse4_1()
370 __m128i offset = _mm_set1_epi32((1 << shift) >> 1); in neg_shift_sse4_1()
389 const __m128i cospi32 = _mm_set1_epi32(cospi[32]); in idct4x4_sse4_1()
390 const __m128i cospi48 = _mm_set1_epi32(cospi[48]); in idct4x4_sse4_1()
391 const __m128i cospi16 = _mm_set1_epi32(cospi[16]); in idct4x4_sse4_1()
392 const __m128i cospim16 = _mm_set1_epi32(-cospi[16]); in idct4x4_sse4_1()
[all …]
Dhighbd_convolve_2d_sse4.c43 const __m128i wt0 = _mm_set1_epi32(w0); in av1_highbd_dist_wtd_convolve_2d_copy_sse4_1()
44 const __m128i wt1 = _mm_set1_epi32(w1); in av1_highbd_dist_wtd_convolve_2d_copy_sse4_1()
51 const __m128i offset_const = _mm_set1_epi32(offset); in av1_highbd_dist_wtd_convolve_2d_copy_sse4_1()
55 const __m128i rounding_const = _mm_set1_epi32((1 << rounding_shift) >> 1); in av1_highbd_dist_wtd_convolve_2d_copy_sse4_1()
191 const __m128i wt0 = _mm_set1_epi32(w0); in av1_highbd_dist_wtd_convolve_2d_sse4_1()
192 const __m128i wt1 = _mm_set1_epi32(w1); in av1_highbd_dist_wtd_convolve_2d_sse4_1()
197 const __m128i offset_const = _mm_set1_epi32(offset); in av1_highbd_dist_wtd_convolve_2d_sse4_1()
200 const __m128i rounding_const = _mm_set1_epi32((1 << rounding_shift) >> 1); in av1_highbd_dist_wtd_convolve_2d_sse4_1()
228 const __m128i round_const = _mm_set1_epi32( in av1_highbd_dist_wtd_convolve_2d_sse4_1()
295 const __m128i round_const = _mm_set1_epi32( in av1_highbd_dist_wtd_convolve_2d_sse4_1()
Dhighbd_jnt_convolve_sse4.c40 const __m128i wt0 = _mm_set1_epi32(w0); in av1_highbd_dist_wtd_convolve_y_sse4_1()
41 const __m128i wt1 = _mm_set1_epi32(w1); in av1_highbd_dist_wtd_convolve_y_sse4_1()
43 _mm_set1_epi32(((1 << conv_params->round_1) >> 1)); in av1_highbd_dist_wtd_convolve_y_sse4_1()
50 const __m128i offset_const = _mm_set1_epi32(offset); in av1_highbd_dist_wtd_convolve_y_sse4_1()
53 const __m128i rounding_const = _mm_set1_epi32((1 << rounding_shift) >> 1); in av1_highbd_dist_wtd_convolve_y_sse4_1()
282 const __m128i wt0 = _mm_set1_epi32(w0); in av1_highbd_dist_wtd_convolve_x_sse4_1()
283 const __m128i wt1 = _mm_set1_epi32(w1); in av1_highbd_dist_wtd_convolve_x_sse4_1()
287 _mm_set1_epi32(((1 << conv_params->round_0) >> 1)); in av1_highbd_dist_wtd_convolve_x_sse4_1()
294 const __m128i offset_const = _mm_set1_epi32(offset); in av1_highbd_dist_wtd_convolve_x_sse4_1()
297 const __m128i rounding_const = _mm_set1_epi32((1 << rounding_shift) >> 1); in av1_highbd_dist_wtd_convolve_x_sse4_1()
Dselfguided_sse4.c144 return _mm_set1_epi32((1 << shift) >> 1); in round_for_shift()
159 an = _mm_max_epi32(_mm_mullo_epi32(a, _mm_set1_epi32(n)), bb); in compute_p()
162 an = _mm_mullo_epi32(sum2, _mm_set1_epi32(n)); in compute_p()
176 const __m128i s = _mm_set1_epi32(params->s[radius_idx]); in calc_ab()
178 const __m128i one_over_n = _mm_set1_epi32(av1_one_by_x[n - 1]); in calc_ab()
215 _mm_set1_epi32(255)); in calc_ab()
228 _mm_sub_epi32(_mm_set1_epi32(SGRPROJ_SGR), a_res); in calc_ab()
317 const __m128i s = _mm_set1_epi32(params->s[radius_idx]); in calc_ab_fast()
319 const __m128i one_over_n = _mm_set1_epi32(av1_one_by_x[n - 1]); in calc_ab_fast()
356 _mm_set1_epi32(255)); in calc_ab_fast()
[all …]
Dav1_convolve_scale_sse4.c32 const __m128i round_add = _mm_set1_epi32(round_add32); in hfilter8()
127 _mm_set1_epi32(((1 << conv_params->round_1) >> 1)); in vfilter8()
128 const __m128i res_add_const = _mm_set1_epi32(1 << offset_bits); in vfilter8()
270 const __m128i round_add = _mm_set1_epi32(round_add32); in highbd_hfilter8()
344 const __m128i sub = _mm_set1_epi32(sub32); in highbd_vfilter8()
353 const __m128i bits_const = _mm_set1_epi32(((1 << bits) >> 1)); in highbd_vfilter8()
355 _mm_set1_epi32(((1 << conv_params->round_1) >> 1)); in highbd_vfilter8()
356 const __m128i res_add_const = _mm_set1_epi32(1 << offset_bits); in highbd_vfilter8()
360 __m128i round_bits_const = _mm_set1_epi32(((1 << round_bits) >> 1)); in highbd_vfilter8()
364 const __m128i wt0 = _mm_set1_epi32(w0); in highbd_vfilter8()
[all …]
/external/oboe/samples/RhythmGame/third_party/glm/simd/
Dinteger.h10 glm_uvec4 const Mask4 = _mm_set1_epi32(0x0000FFFF); in glm_i128_interleave()
11 glm_uvec4 const Mask3 = _mm_set1_epi32(0x00FF00FF); in glm_i128_interleave()
12 glm_uvec4 const Mask2 = _mm_set1_epi32(0x0F0F0F0F); in glm_i128_interleave()
13 glm_uvec4 const Mask1 = _mm_set1_epi32(0x33333333); in glm_i128_interleave()
14 glm_uvec4 const Mask0 = _mm_set1_epi32(0x55555555); in glm_i128_interleave()
64 glm_uvec4 const Mask4 = _mm_set1_epi32(0x0000FFFF); in glm_i128_interleave2()
65 glm_uvec4 const Mask3 = _mm_set1_epi32(0x00FF00FF); in glm_i128_interleave2()
66 glm_uvec4 const Mask2 = _mm_set1_epi32(0x0F0F0F0F); in glm_i128_interleave2()
67 glm_uvec4 const Mask1 = _mm_set1_epi32(0x33333333); in glm_i128_interleave2()
68 glm_uvec4 const Mask0 = _mm_set1_epi32(0x55555555); in glm_i128_interleave2()
Dcommon.h84 return _mm_and_ps(x, _mm_castsi128_ps(_mm_set1_epi32(0x7FFFFFFF))); in glm_vec4_abs()
115 glm_vec4 const sgn0 = _mm_castsi128_ps(_mm_set1_epi32(0x80000000)); in glm_vec4_round()
147 glm_vec4 const sgn0 = _mm_castsi128_ps(_mm_set1_epi32(0x80000000)); in glm_vec4_roundEven()
223 glm_ivec4 const t3 = _mm_set1_epi32(0xFF000000); // exponent mask in glm_vec4_nan()
237 …return _mm_castsi128_ps(_mm_cmpeq_epi32(t2, _mm_set1_epi32(0xFF000000))); // exponent is all 1s, … in glm_vec4_inf()
/external/XNNPACK/src/f32-argmaxpool/
D9x-sse2-c4.c99 vidx = _mm_or_si128(_mm_andnot_si128(vm1, vidx), _mm_and_si128(vm1, _mm_set1_epi32(1))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4()
103 vidx = _mm_or_si128(_mm_andnot_si128(vm2, vidx), _mm_and_si128(vm2, _mm_set1_epi32(2))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4()
107 vidx = _mm_or_si128(_mm_andnot_si128(vm3, vidx), _mm_and_si128(vm3, _mm_set1_epi32(3))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4()
111 vidx = _mm_or_si128(_mm_andnot_si128(vm4, vidx), _mm_and_si128(vm4, _mm_set1_epi32(4))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4()
115 vidx = _mm_or_si128(_mm_andnot_si128(vm5, vidx), _mm_and_si128(vm5, _mm_set1_epi32(5))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4()
119 vidx = _mm_or_si128(_mm_andnot_si128(vm6, vidx), _mm_and_si128(vm6, _mm_set1_epi32(6))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4()
123 vidx = _mm_or_si128(_mm_andnot_si128(vm7, vidx), _mm_and_si128(vm7, _mm_set1_epi32(7))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4()
127 vidx = _mm_or_si128(_mm_andnot_si128(vm8, vidx), _mm_and_si128(vm8, _mm_set1_epi32(8))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4()
150 vidx = _mm_or_si128(_mm_andnot_si128(vm1, vidx), _mm_and_si128(vm1, _mm_set1_epi32(1))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4()
154 vidx = _mm_or_si128(_mm_andnot_si128(vm2, vidx), _mm_and_si128(vm2, _mm_set1_epi32(2))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4()
[all …]
D4x-sse2-c4.c64 vidx = _mm_or_si128(_mm_andnot_si128(vm1, vidx), _mm_and_si128(vm1, _mm_set1_epi32(1))); in xnn_f32_argmaxpool_ukernel_4x__sse2_c4()
68 vidx = _mm_or_si128(_mm_andnot_si128(vm2, vidx), _mm_and_si128(vm2, _mm_set1_epi32(2))); in xnn_f32_argmaxpool_ukernel_4x__sse2_c4()
72 vidx = _mm_or_si128(_mm_andnot_si128(vm3, vidx), _mm_and_si128(vm3, _mm_set1_epi32(3))); in xnn_f32_argmaxpool_ukernel_4x__sse2_c4()
90 vidx = _mm_or_si128(_mm_andnot_si128(vm1, vidx), _mm_and_si128(vm1, _mm_set1_epi32(1))); in xnn_f32_argmaxpool_ukernel_4x__sse2_c4()
94 vidx = _mm_or_si128(_mm_andnot_si128(vm2, vidx), _mm_and_si128(vm2, _mm_set1_epi32(2))); in xnn_f32_argmaxpool_ukernel_4x__sse2_c4()
98 vidx = _mm_or_si128(_mm_andnot_si128(vm3, vidx), _mm_and_si128(vm3, _mm_set1_epi32(3))); in xnn_f32_argmaxpool_ukernel_4x__sse2_c4()
D9p8x-sse2-c4.c80 vidx = _mm_or_si128(_mm_andnot_si128(vm1, vidx), _mm_and_si128(vm1, _mm_set1_epi32(1))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4()
84 vidx = _mm_or_si128(_mm_andnot_si128(vm2, vidx), _mm_and_si128(vm2, _mm_set1_epi32(2))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4()
88 vidx = _mm_or_si128(_mm_andnot_si128(vm3, vidx), _mm_and_si128(vm3, _mm_set1_epi32(3))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4()
92 vidx = _mm_or_si128(_mm_andnot_si128(vm4, vidx), _mm_and_si128(vm4, _mm_set1_epi32(4))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4()
96 vidx = _mm_or_si128(_mm_andnot_si128(vm5, vidx), _mm_and_si128(vm5, _mm_set1_epi32(5))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4()
100 vidx = _mm_or_si128(_mm_andnot_si128(vm6, vidx), _mm_and_si128(vm6, _mm_set1_epi32(6))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4()
104 vidx = _mm_or_si128(_mm_andnot_si128(vm7, vidx), _mm_and_si128(vm7, _mm_set1_epi32(7))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4()
108 vidx = _mm_or_si128(_mm_andnot_si128(vm8, vidx), _mm_and_si128(vm8, _mm_set1_epi32(8))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4()
116 const __m128i v1 = _mm_set1_epi32(1); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4()
117 const __m128i v8 = _mm_set1_epi32(8); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4()
/external/libaom/libaom/av1/encoder/x86/
Dav1_fwd_txfm1d_sse4.c465 const __m128i __rounding = _mm_set1_epi32(1 << (cos_bit - 1)); in av1_fdct64_sse4_1()
467 __m128i cospi_m32 = _mm_set1_epi32(-cospi[32]); in av1_fdct64_sse4_1()
468 __m128i cospi_p32 = _mm_set1_epi32(cospi[32]); in av1_fdct64_sse4_1()
469 __m128i cospi_m16 = _mm_set1_epi32(-cospi[16]); in av1_fdct64_sse4_1()
470 __m128i cospi_p48 = _mm_set1_epi32(cospi[48]); in av1_fdct64_sse4_1()
471 __m128i cospi_m48 = _mm_set1_epi32(-cospi[48]); in av1_fdct64_sse4_1()
472 __m128i cospi_p16 = _mm_set1_epi32(cospi[16]); in av1_fdct64_sse4_1()
473 __m128i cospi_m08 = _mm_set1_epi32(-cospi[8]); in av1_fdct64_sse4_1()
474 __m128i cospi_p56 = _mm_set1_epi32(cospi[56]); in av1_fdct64_sse4_1()
475 __m128i cospi_m56 = _mm_set1_epi32(-cospi[56]); in av1_fdct64_sse4_1()
[all …]
Dhighbd_fwd_txfm_sse4.c65 const __m128i cospi32 = _mm_set1_epi32(cospi[32]); in fdct4x4_sse4_1()
66 const __m128i cospi48 = _mm_set1_epi32(cospi[48]); in fdct4x4_sse4_1()
67 const __m128i cospi16 = _mm_set1_epi32(cospi[16]); in fdct4x4_sse4_1()
68 const __m128i rnding = _mm_set1_epi32(1 << (bit - 1)); in fdct4x4_sse4_1()
131 const __m128i rnding = _mm_set1_epi32(1 << (bit - 1)); in fadst4x4_sse4_1()
132 const __m128i sinpi1 = _mm_set1_epi32((int)sinpi[1]); in fadst4x4_sse4_1()
133 const __m128i sinpi2 = _mm_set1_epi32((int)sinpi[2]); in fadst4x4_sse4_1()
134 const __m128i sinpi3 = _mm_set1_epi32((int)sinpi[3]); in fadst4x4_sse4_1()
135 const __m128i sinpi4 = _mm_set1_epi32((int)sinpi[4]); in fadst4x4_sse4_1()
193 __m128i fact = _mm_set1_epi32(NewSqrt2); in idtx4x4_sse4_1()
[all …]
/external/flac/src/libFLAC/
Dlpc_intrin_avx2.c779 q0 = _mm256_cvtepu32_epi64(_mm_set1_epi32(qlp_coeff[0 ])); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2()
780 q1 = _mm256_cvtepu32_epi64(_mm_set1_epi32(qlp_coeff[1 ])); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2()
781 q2 = _mm256_cvtepu32_epi64(_mm_set1_epi32(qlp_coeff[2 ])); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2()
782 q3 = _mm256_cvtepu32_epi64(_mm_set1_epi32(qlp_coeff[3 ])); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2()
783 q4 = _mm256_cvtepu32_epi64(_mm_set1_epi32(qlp_coeff[4 ])); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2()
784 q5 = _mm256_cvtepu32_epi64(_mm_set1_epi32(qlp_coeff[5 ])); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2()
785 q6 = _mm256_cvtepu32_epi64(_mm_set1_epi32(qlp_coeff[6 ])); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2()
786 q7 = _mm256_cvtepu32_epi64(_mm_set1_epi32(qlp_coeff[7 ])); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2()
787 q8 = _mm256_cvtepu32_epi64(_mm_set1_epi32(qlp_coeff[8 ])); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2()
788 q9 = _mm256_cvtepu32_epi64(_mm_set1_epi32(qlp_coeff[9 ])); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2()
[all …]
/external/libvpx/libvpx/vp9/encoder/x86/
Dvp9_diamond_search_sad_avx.c79 const __m128i v_max_mv_w = _mm_set1_epi32(maxmv.as_int); in vp9_diamond_search_sad_avx()
81 const __m128i v_min_mv_w = _mm_set1_epi32(minmv.as_int); in vp9_diamond_search_sad_avx()
83 const __m128i v_spb_d = _mm_set1_epi32(sad_per_bit); in vp9_diamond_search_sad_avx()
85 const __m128i v_joint_cost_0_d = _mm_set1_epi32(x->nmvjointsadcost[0]); in vp9_diamond_search_sad_avx()
86 const __m128i v_joint_cost_1_d = _mm_set1_epi32(x->nmvjointsadcost[1]); in vp9_diamond_search_sad_avx()
99 const __m128i vfcmv = _mm_set1_epi32(fcenter_mv.as_int); in vp9_diamond_search_sad_avx()
106 __m128i v_bmv_w = _mm_set1_epi32(bmv.as_int); in vp9_diamond_search_sad_avx()
120 __m128i v_ba_d = _mm_set1_epi32((intptr_t)best_address); in vp9_diamond_search_sad_avx()
237 _mm_set1_epi32(1 << (VP9_PROB_COST_SHIFT - 1))); in vp9_diamond_search_sad_avx()
296 v_bmv_w = _mm_set1_epi32(bmv.as_int); in vp9_diamond_search_sad_avx()
[all …]
/external/libgav1/libgav1/src/dsp/x86/
Dintrapred_smooth_sse4.cc66 const __m128i cvtepi32_epi8 = _mm_set1_epi32(0x0C080400); in WriteSmoothHorizontalSum4()
131 const __m128i cvtepi32_epi8 = _mm_set1_epi32(0x0C080400); in WriteSmoothPredSum4()
185 const __m128i round = _mm_set1_epi32(256); in WriteSmoothPred4x8()
187 const __m128i cvtepi32_epi8 = _mm_set1_epi32(0x0C080400); in WriteSmoothPred4x8()
223 const __m128i scale = _mm_set1_epi32(256); in Smooth4x4_SSE4_1()
365 const __m128i round = _mm_set1_epi32(256); in WriteSmoothPred8xH()
486 const __m128i round = _mm_set1_epi32(256); in SmoothWxH()
538 const __m128i top_right = _mm_set1_epi32(top_ptr[3]); in SmoothHorizontal4x4_SSE4_1()
542 __m128i scale = _mm_set1_epi32(256); in SmoothHorizontal4x4_SSE4_1()
545 scale = _mm_set1_epi32(128); in SmoothHorizontal4x4_SSE4_1()
[all …]
Ddistance_weighted_blend_sse4.cc60 const __m128i weights = _mm_set1_epi32(weight_0 | (weight_1 << 16)); in DistanceWeightedBlend4xH_SSE4_1()
104 const __m128i weights = _mm_set1_epi32(weight_0 | (weight_1 << 16)); in DistanceWeightedBlend8xH_SSE4_1()
132 const __m128i weights = _mm_set1_epi32(weight_0 | (weight_1 << 16)); in DistanceWeightedBlendLarge_SSE4_1()
235 const __m128i bias = _mm_set1_epi32(offset); in ComputeWeightedAverage8()
263 const __m128i weight0 = _mm_set1_epi32(weight_0); in DistanceWeightedBlend4xH_SSE4_1()
264 const __m128i weight1 = _mm_set1_epi32(weight_1); in DistanceWeightedBlend4xH_SSE4_1()
307 const __m128i weight0 = _mm_set1_epi32(weight_0); in DistanceWeightedBlend8xH_SSE4_1()
308 const __m128i weight1 = _mm_set1_epi32(weight_1); in DistanceWeightedBlend8xH_SSE4_1()
339 const __m128i weight0 = _mm_set1_epi32(weight_0); in DistanceWeightedBlendLarge_SSE4_1()
340 const __m128i weight1 = _mm_set1_epi32(weight_1); in DistanceWeightedBlendLarge_SSE4_1()
/external/mesa3d/src/gallium/drivers/swr/rasterizer/core/
Dstate_funcs.h36 __m128i vMin = _mm_set1_epi32(*min); in expandThenBlend4()
37 __m128i vMax = _mm_set1_epi32(*max); in expandThenBlend4()
45 _vXi[i] = _mm_set1_epi32(_xi[i]); in PrecalcSampleData()
46 _vYi[i] = _mm_set1_epi32(_yi[i]); in PrecalcSampleData()
/external/libavc/encoder/x86/
Dih264e_half_pel_ssse3.c120 coeff0_1_16x8b = _mm_set1_epi32(0xFB01FB01); //c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 in ih264e_sixtapfilter_horz_ssse3()
121 coeff2_3_16x8b = _mm_set1_epi32(0x14141414); //c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 in ih264e_sixtapfilter_horz_ssse3()
122 coeff4_5_16x8b = _mm_set1_epi32(0x01FB01FB); //c4 c5 c4 c5 c4 c5 c4 c5 c4 c5 c4 c5 c4 c5 c4 c5 in ih264e_sixtapfilter_horz_ssse3()
273 … coeff0_1_16x8b = _mm_set1_epi32(0xFB01FB01); //c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 in ih264e_sixtap_filter_2dvh_vert_ssse3()
274 … coeff2_3_16x8b = _mm_set1_epi32(0x14141414); //c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 in ih264e_sixtap_filter_2dvh_vert_ssse3()
275 … coeff4_5_16x8b = _mm_set1_epi32(0x01FB01FB); //c4 c5 c4 c5 c4 c5 c4 c5 c4 c5 c4 c5 c4 c5 c4 c5 in ih264e_sixtap_filter_2dvh_vert_ssse3()
384 coeff0_1_8x16b = _mm_set1_epi32(0xFFFB0001); //c0 c1 c0 c1 c0 c1 c0 c1 in ih264e_sixtap_filter_2dvh_vert_ssse3()
385 coeff2_3_8x16b = _mm_set1_epi32(0x00140014); //c2 c3 c2 c3 c2 c3 c2 c3 in ih264e_sixtap_filter_2dvh_vert_ssse3()
386 coeff4_5_8x16b = _mm_set1_epi32(0x0001FFFB); //c4 c5 c4 c5 c4 c5 c4 c5 in ih264e_sixtap_filter_2dvh_vert_ssse3()
388 const_val512_4x32b = _mm_set1_epi32(512); in ih264e_sixtap_filter_2dvh_vert_ssse3()
/external/libavc/common/x86/
Dih264_inter_pred_filters_ssse3.c233 coeff0_1_16x8b = _mm_set1_epi32(0xFB01FB01); //c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 in ih264_inter_pred_luma_horz_ssse3()
234 coeff2_3_16x8b = _mm_set1_epi32(0x14141414); //c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 in ih264_inter_pred_luma_horz_ssse3()
235 coeff4_5_16x8b = _mm_set1_epi32(0x01FB01FB); //c4 c5 c4 c5 c4 c5 c4 c5 c4 c5 c4 c5 c4 c5 c4 c5 in ih264_inter_pred_luma_horz_ssse3()
502 coeff0_1_16x8b = _mm_set1_epi32(0xFB01FB01); //c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 in ih264_inter_pred_luma_vert_ssse3()
503 coeff2_3_16x8b = _mm_set1_epi32(0x14141414); //c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 in ih264_inter_pred_luma_vert_ssse3()
504 coeff4_5_16x8b = _mm_set1_epi32(0x01FB01FB); //c4 c5 c4 c5 c4 c5 c4 c5 c4 c5 c4 c5 c4 c5 c4 c5 in ih264_inter_pred_luma_vert_ssse3()
797 … coeff0_1_16x8b = _mm_set1_epi32(0xFB01FB01); //c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 in ih264_inter_pred_luma_horz_hpel_vert_hpel_ssse3()
798 … coeff2_3_16x8b = _mm_set1_epi32(0x14141414); //c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 in ih264_inter_pred_luma_horz_hpel_vert_hpel_ssse3()
799 … coeff4_5_16x8b = _mm_set1_epi32(0x01FB01FB); //c4 c5 c5 c5 c4 c5 c5 c5 c4 c5 c5 c5 c4 c5 c5 c5 in ih264_inter_pred_luma_horz_hpel_vert_hpel_ssse3()
876 coeff0_1_8x16b = _mm_set1_epi32(0xFFFB0001); in ih264_inter_pred_luma_horz_hpel_vert_hpel_ssse3()
[all …]
Dih264_deblk_chroma_ssse3.c147 alpha_cbcr_16x8 = _mm_set1_epi32(alpha_cbcr); in ih264_deblk_chroma_vert_bs4_ssse3()
152 beta_cbcr_16x8 = _mm_set1_epi32(beta_cbcr); in ih264_deblk_chroma_vert_bs4_ssse3()
178 alpha_cbcr_16x8 = _mm_set1_epi32(alpha_cbcr); in ih264_deblk_chroma_vert_bs4_ssse3()
183 beta_cbcr_16x8 = _mm_set1_epi32(beta_cbcr); in ih264_deblk_chroma_vert_bs4_ssse3()
316 alpha_cbcr_16x8 = _mm_set1_epi32(alpha_cbcr); in ih264_deblk_chroma_horz_bs4_ssse3()
321 beta_cbcr_16x8 = _mm_set1_epi32(beta_cbcr); in ih264_deblk_chroma_horz_bs4_ssse3()
347 alpha_cbcr_16x8 = _mm_set1_epi32(alpha_cbcr); in ih264_deblk_chroma_horz_bs4_ssse3()
352 beta_cbcr_16x8 = _mm_set1_epi32(beta_cbcr); in ih264_deblk_chroma_horz_bs4_ssse3()
496 alpha_cbcr_16x8 = _mm_set1_epi32(alpha_cbcr); in ih264_deblk_chroma_vert_bslt4_ssse3()
501 beta_cbcr_16x8 = _mm_set1_epi32(beta_cbcr); in ih264_deblk_chroma_vert_bslt4_ssse3()
[all …]
/external/gemmlowp/fixedpoint/
Dfixedpoint_sse.h90 return _mm_andnot_si128(a, _mm_set1_epi32(-1));
130 return _mm_sign_epi32(a, _mm_set1_epi32(-1));
196 return MaskIfEqual(a, _mm_set1_epi32(0));
206 return MaskIfNotEqual(a, _mm_set1_epi32(0));
297 one = _mm_set1_epi32(1);
298 sign_bit_mask = _mm_set1_epi32(0x80000000);
328 min = _mm_set1_epi32(std::numeric_limits<std::int32_t>::min());
372 return _mm_set1_epi32(x);
/external/oboe/samples/RhythmGame/third_party/glm/detail/
Dfunc_integer_simd.inl18 __m128i const set1 = _mm_set1_epi32(Mask);
22 __m128i const set2 = _mm_andnot_si128(set0, _mm_set1_epi32(-1));
39 __m128i const set1 = _mm_set1_epi32(Mask);
/external/libvpx/libvpx/vpx_dsp/x86/
Dhighbd_inv_txfm_sse2.h220 io[0] = wraplow_16bit_shift5(io[0], io[8], _mm_set1_epi32(16)); in highbd_idct8x8_final_round()
221 io[1] = wraplow_16bit_shift5(io[1], io[9], _mm_set1_epi32(16)); in highbd_idct8x8_final_round()
222 io[2] = wraplow_16bit_shift5(io[2], io[10], _mm_set1_epi32(16)); in highbd_idct8x8_final_round()
223 io[3] = wraplow_16bit_shift5(io[3], io[11], _mm_set1_epi32(16)); in highbd_idct8x8_final_round()
224 io[4] = wraplow_16bit_shift5(io[4], io[12], _mm_set1_epi32(16)); in highbd_idct8x8_final_round()
225 io[5] = wraplow_16bit_shift5(io[5], io[13], _mm_set1_epi32(16)); in highbd_idct8x8_final_round()
226 io[6] = wraplow_16bit_shift5(io[6], io[14], _mm_set1_epi32(16)); in highbd_idct8x8_final_round()
227 io[7] = wraplow_16bit_shift5(io[7], io[15], _mm_set1_epi32(16)); in highbd_idct8x8_final_round()
395 const __m128i final_rounding = _mm_set1_epi32(1 << 5); in highbd_write_buffer_4()
Dhighbd_idct4x4_add_sse4.c42 io[0] = wraplow_16bit_shift4(io[0], io[1], _mm_set1_epi32(8)); in vpx_highbd_idct4x4_16_add_sse4_1()
43 io[1] = wraplow_16bit_shift4(io[2], io[3], _mm_set1_epi32(8)); in vpx_highbd_idct4x4_16_add_sse4_1()
/external/XNNPACK/src/math/
Dexp-sse2-rr2-lut64-p2.c39 const __m128i vmin_exponent = _mm_set1_epi32(0xC1000000); in xnn_math_f32_exp__sse2_rr2_lut64_p2()
40 const __m128i vmax_exponent = _mm_set1_epi32(0x3F800000); in xnn_math_f32_exp__sse2_rr2_lut64_p2()
42 const __m128i vindex_mask = _mm_set1_epi32(0x3F); in xnn_math_f32_exp__sse2_rr2_lut64_p2()

12345678910