/external/libaom/libaom/av1/common/x86/ |
D | highbd_inv_txfm_sse4.c | 42 __m128i rnding = _mm_set1_epi32(1 << (shift - 1)); in round_shift_4x4() 87 __m128i max_clip_val = _mm_set1_epi32((1 << bd) - 1); in highbd_get_recon_8x8_sse4_1() 166 __m128i offset = _mm_set1_epi32((1 << shift) >> 1); in shift_and_clamp_sse4_1() 356 const __m128i clamp_lo_out = _mm_set1_epi32(-(1 << (log_range_out - 1))); in idct32_stage9_sse4_1() 357 const __m128i clamp_hi_out = _mm_set1_epi32((1 << (log_range_out - 1)) - 1); in idct32_stage9_sse4_1() 370 __m128i offset = _mm_set1_epi32((1 << shift) >> 1); in neg_shift_sse4_1() 389 const __m128i cospi32 = _mm_set1_epi32(cospi[32]); in idct4x4_sse4_1() 390 const __m128i cospi48 = _mm_set1_epi32(cospi[48]); in idct4x4_sse4_1() 391 const __m128i cospi16 = _mm_set1_epi32(cospi[16]); in idct4x4_sse4_1() 392 const __m128i cospim16 = _mm_set1_epi32(-cospi[16]); in idct4x4_sse4_1() [all …]
|
D | highbd_convolve_2d_sse4.c | 43 const __m128i wt0 = _mm_set1_epi32(w0); in av1_highbd_dist_wtd_convolve_2d_copy_sse4_1() 44 const __m128i wt1 = _mm_set1_epi32(w1); in av1_highbd_dist_wtd_convolve_2d_copy_sse4_1() 51 const __m128i offset_const = _mm_set1_epi32(offset); in av1_highbd_dist_wtd_convolve_2d_copy_sse4_1() 55 const __m128i rounding_const = _mm_set1_epi32((1 << rounding_shift) >> 1); in av1_highbd_dist_wtd_convolve_2d_copy_sse4_1() 191 const __m128i wt0 = _mm_set1_epi32(w0); in av1_highbd_dist_wtd_convolve_2d_sse4_1() 192 const __m128i wt1 = _mm_set1_epi32(w1); in av1_highbd_dist_wtd_convolve_2d_sse4_1() 197 const __m128i offset_const = _mm_set1_epi32(offset); in av1_highbd_dist_wtd_convolve_2d_sse4_1() 200 const __m128i rounding_const = _mm_set1_epi32((1 << rounding_shift) >> 1); in av1_highbd_dist_wtd_convolve_2d_sse4_1() 228 const __m128i round_const = _mm_set1_epi32( in av1_highbd_dist_wtd_convolve_2d_sse4_1() 295 const __m128i round_const = _mm_set1_epi32( in av1_highbd_dist_wtd_convolve_2d_sse4_1()
|
D | highbd_jnt_convolve_sse4.c | 40 const __m128i wt0 = _mm_set1_epi32(w0); in av1_highbd_dist_wtd_convolve_y_sse4_1() 41 const __m128i wt1 = _mm_set1_epi32(w1); in av1_highbd_dist_wtd_convolve_y_sse4_1() 43 _mm_set1_epi32(((1 << conv_params->round_1) >> 1)); in av1_highbd_dist_wtd_convolve_y_sse4_1() 50 const __m128i offset_const = _mm_set1_epi32(offset); in av1_highbd_dist_wtd_convolve_y_sse4_1() 53 const __m128i rounding_const = _mm_set1_epi32((1 << rounding_shift) >> 1); in av1_highbd_dist_wtd_convolve_y_sse4_1() 282 const __m128i wt0 = _mm_set1_epi32(w0); in av1_highbd_dist_wtd_convolve_x_sse4_1() 283 const __m128i wt1 = _mm_set1_epi32(w1); in av1_highbd_dist_wtd_convolve_x_sse4_1() 287 _mm_set1_epi32(((1 << conv_params->round_0) >> 1)); in av1_highbd_dist_wtd_convolve_x_sse4_1() 294 const __m128i offset_const = _mm_set1_epi32(offset); in av1_highbd_dist_wtd_convolve_x_sse4_1() 297 const __m128i rounding_const = _mm_set1_epi32((1 << rounding_shift) >> 1); in av1_highbd_dist_wtd_convolve_x_sse4_1()
|
D | selfguided_sse4.c | 144 return _mm_set1_epi32((1 << shift) >> 1); in round_for_shift() 159 an = _mm_max_epi32(_mm_mullo_epi32(a, _mm_set1_epi32(n)), bb); in compute_p() 162 an = _mm_mullo_epi32(sum2, _mm_set1_epi32(n)); in compute_p() 176 const __m128i s = _mm_set1_epi32(params->s[radius_idx]); in calc_ab() 178 const __m128i one_over_n = _mm_set1_epi32(av1_one_by_x[n - 1]); in calc_ab() 215 _mm_set1_epi32(255)); in calc_ab() 228 _mm_sub_epi32(_mm_set1_epi32(SGRPROJ_SGR), a_res); in calc_ab() 317 const __m128i s = _mm_set1_epi32(params->s[radius_idx]); in calc_ab_fast() 319 const __m128i one_over_n = _mm_set1_epi32(av1_one_by_x[n - 1]); in calc_ab_fast() 356 _mm_set1_epi32(255)); in calc_ab_fast() [all …]
|
D | av1_convolve_scale_sse4.c | 32 const __m128i round_add = _mm_set1_epi32(round_add32); in hfilter8() 127 _mm_set1_epi32(((1 << conv_params->round_1) >> 1)); in vfilter8() 128 const __m128i res_add_const = _mm_set1_epi32(1 << offset_bits); in vfilter8() 270 const __m128i round_add = _mm_set1_epi32(round_add32); in highbd_hfilter8() 344 const __m128i sub = _mm_set1_epi32(sub32); in highbd_vfilter8() 353 const __m128i bits_const = _mm_set1_epi32(((1 << bits) >> 1)); in highbd_vfilter8() 355 _mm_set1_epi32(((1 << conv_params->round_1) >> 1)); in highbd_vfilter8() 356 const __m128i res_add_const = _mm_set1_epi32(1 << offset_bits); in highbd_vfilter8() 360 __m128i round_bits_const = _mm_set1_epi32(((1 << round_bits) >> 1)); in highbd_vfilter8() 364 const __m128i wt0 = _mm_set1_epi32(w0); in highbd_vfilter8() [all …]
|
/external/oboe/samples/RhythmGame/third_party/glm/simd/ |
D | integer.h | 10 glm_uvec4 const Mask4 = _mm_set1_epi32(0x0000FFFF); in glm_i128_interleave() 11 glm_uvec4 const Mask3 = _mm_set1_epi32(0x00FF00FF); in glm_i128_interleave() 12 glm_uvec4 const Mask2 = _mm_set1_epi32(0x0F0F0F0F); in glm_i128_interleave() 13 glm_uvec4 const Mask1 = _mm_set1_epi32(0x33333333); in glm_i128_interleave() 14 glm_uvec4 const Mask0 = _mm_set1_epi32(0x55555555); in glm_i128_interleave() 64 glm_uvec4 const Mask4 = _mm_set1_epi32(0x0000FFFF); in glm_i128_interleave2() 65 glm_uvec4 const Mask3 = _mm_set1_epi32(0x00FF00FF); in glm_i128_interleave2() 66 glm_uvec4 const Mask2 = _mm_set1_epi32(0x0F0F0F0F); in glm_i128_interleave2() 67 glm_uvec4 const Mask1 = _mm_set1_epi32(0x33333333); in glm_i128_interleave2() 68 glm_uvec4 const Mask0 = _mm_set1_epi32(0x55555555); in glm_i128_interleave2()
|
D | common.h | 84 return _mm_and_ps(x, _mm_castsi128_ps(_mm_set1_epi32(0x7FFFFFFF))); in glm_vec4_abs() 115 glm_vec4 const sgn0 = _mm_castsi128_ps(_mm_set1_epi32(0x80000000)); in glm_vec4_round() 147 glm_vec4 const sgn0 = _mm_castsi128_ps(_mm_set1_epi32(0x80000000)); in glm_vec4_roundEven() 223 glm_ivec4 const t3 = _mm_set1_epi32(0xFF000000); // exponent mask in glm_vec4_nan() 237 …return _mm_castsi128_ps(_mm_cmpeq_epi32(t2, _mm_set1_epi32(0xFF000000))); // exponent is all 1s, … in glm_vec4_inf()
|
/external/XNNPACK/src/f32-argmaxpool/ |
D | 9x-sse2-c4.c | 99 vidx = _mm_or_si128(_mm_andnot_si128(vm1, vidx), _mm_and_si128(vm1, _mm_set1_epi32(1))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() 103 vidx = _mm_or_si128(_mm_andnot_si128(vm2, vidx), _mm_and_si128(vm2, _mm_set1_epi32(2))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() 107 vidx = _mm_or_si128(_mm_andnot_si128(vm3, vidx), _mm_and_si128(vm3, _mm_set1_epi32(3))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() 111 vidx = _mm_or_si128(_mm_andnot_si128(vm4, vidx), _mm_and_si128(vm4, _mm_set1_epi32(4))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() 115 vidx = _mm_or_si128(_mm_andnot_si128(vm5, vidx), _mm_and_si128(vm5, _mm_set1_epi32(5))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() 119 vidx = _mm_or_si128(_mm_andnot_si128(vm6, vidx), _mm_and_si128(vm6, _mm_set1_epi32(6))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() 123 vidx = _mm_or_si128(_mm_andnot_si128(vm7, vidx), _mm_and_si128(vm7, _mm_set1_epi32(7))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() 127 vidx = _mm_or_si128(_mm_andnot_si128(vm8, vidx), _mm_and_si128(vm8, _mm_set1_epi32(8))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() 150 vidx = _mm_or_si128(_mm_andnot_si128(vm1, vidx), _mm_and_si128(vm1, _mm_set1_epi32(1))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() 154 vidx = _mm_or_si128(_mm_andnot_si128(vm2, vidx), _mm_and_si128(vm2, _mm_set1_epi32(2))); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() [all …]
|
D | 4x-sse2-c4.c | 64 vidx = _mm_or_si128(_mm_andnot_si128(vm1, vidx), _mm_and_si128(vm1, _mm_set1_epi32(1))); in xnn_f32_argmaxpool_ukernel_4x__sse2_c4() 68 vidx = _mm_or_si128(_mm_andnot_si128(vm2, vidx), _mm_and_si128(vm2, _mm_set1_epi32(2))); in xnn_f32_argmaxpool_ukernel_4x__sse2_c4() 72 vidx = _mm_or_si128(_mm_andnot_si128(vm3, vidx), _mm_and_si128(vm3, _mm_set1_epi32(3))); in xnn_f32_argmaxpool_ukernel_4x__sse2_c4() 90 vidx = _mm_or_si128(_mm_andnot_si128(vm1, vidx), _mm_and_si128(vm1, _mm_set1_epi32(1))); in xnn_f32_argmaxpool_ukernel_4x__sse2_c4() 94 vidx = _mm_or_si128(_mm_andnot_si128(vm2, vidx), _mm_and_si128(vm2, _mm_set1_epi32(2))); in xnn_f32_argmaxpool_ukernel_4x__sse2_c4() 98 vidx = _mm_or_si128(_mm_andnot_si128(vm3, vidx), _mm_and_si128(vm3, _mm_set1_epi32(3))); in xnn_f32_argmaxpool_ukernel_4x__sse2_c4()
|
D | 9p8x-sse2-c4.c | 80 vidx = _mm_or_si128(_mm_andnot_si128(vm1, vidx), _mm_and_si128(vm1, _mm_set1_epi32(1))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() 84 vidx = _mm_or_si128(_mm_andnot_si128(vm2, vidx), _mm_and_si128(vm2, _mm_set1_epi32(2))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() 88 vidx = _mm_or_si128(_mm_andnot_si128(vm3, vidx), _mm_and_si128(vm3, _mm_set1_epi32(3))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() 92 vidx = _mm_or_si128(_mm_andnot_si128(vm4, vidx), _mm_and_si128(vm4, _mm_set1_epi32(4))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() 96 vidx = _mm_or_si128(_mm_andnot_si128(vm5, vidx), _mm_and_si128(vm5, _mm_set1_epi32(5))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() 100 vidx = _mm_or_si128(_mm_andnot_si128(vm6, vidx), _mm_and_si128(vm6, _mm_set1_epi32(6))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() 104 vidx = _mm_or_si128(_mm_andnot_si128(vm7, vidx), _mm_and_si128(vm7, _mm_set1_epi32(7))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() 108 vidx = _mm_or_si128(_mm_andnot_si128(vm8, vidx), _mm_and_si128(vm8, _mm_set1_epi32(8))); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() 116 const __m128i v1 = _mm_set1_epi32(1); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() 117 const __m128i v8 = _mm_set1_epi32(8); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4()
|
/external/libaom/libaom/av1/encoder/x86/ |
D | av1_fwd_txfm1d_sse4.c | 465 const __m128i __rounding = _mm_set1_epi32(1 << (cos_bit - 1)); in av1_fdct64_sse4_1() 467 __m128i cospi_m32 = _mm_set1_epi32(-cospi[32]); in av1_fdct64_sse4_1() 468 __m128i cospi_p32 = _mm_set1_epi32(cospi[32]); in av1_fdct64_sse4_1() 469 __m128i cospi_m16 = _mm_set1_epi32(-cospi[16]); in av1_fdct64_sse4_1() 470 __m128i cospi_p48 = _mm_set1_epi32(cospi[48]); in av1_fdct64_sse4_1() 471 __m128i cospi_m48 = _mm_set1_epi32(-cospi[48]); in av1_fdct64_sse4_1() 472 __m128i cospi_p16 = _mm_set1_epi32(cospi[16]); in av1_fdct64_sse4_1() 473 __m128i cospi_m08 = _mm_set1_epi32(-cospi[8]); in av1_fdct64_sse4_1() 474 __m128i cospi_p56 = _mm_set1_epi32(cospi[56]); in av1_fdct64_sse4_1() 475 __m128i cospi_m56 = _mm_set1_epi32(-cospi[56]); in av1_fdct64_sse4_1() [all …]
|
D | highbd_fwd_txfm_sse4.c | 65 const __m128i cospi32 = _mm_set1_epi32(cospi[32]); in fdct4x4_sse4_1() 66 const __m128i cospi48 = _mm_set1_epi32(cospi[48]); in fdct4x4_sse4_1() 67 const __m128i cospi16 = _mm_set1_epi32(cospi[16]); in fdct4x4_sse4_1() 68 const __m128i rnding = _mm_set1_epi32(1 << (bit - 1)); in fdct4x4_sse4_1() 131 const __m128i rnding = _mm_set1_epi32(1 << (bit - 1)); in fadst4x4_sse4_1() 132 const __m128i sinpi1 = _mm_set1_epi32((int)sinpi[1]); in fadst4x4_sse4_1() 133 const __m128i sinpi2 = _mm_set1_epi32((int)sinpi[2]); in fadst4x4_sse4_1() 134 const __m128i sinpi3 = _mm_set1_epi32((int)sinpi[3]); in fadst4x4_sse4_1() 135 const __m128i sinpi4 = _mm_set1_epi32((int)sinpi[4]); in fadst4x4_sse4_1() 193 __m128i fact = _mm_set1_epi32(NewSqrt2); in idtx4x4_sse4_1() [all …]
|
/external/flac/src/libFLAC/ |
D | lpc_intrin_avx2.c | 779 q0 = _mm256_cvtepu32_epi64(_mm_set1_epi32(qlp_coeff[0 ])); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2() 780 q1 = _mm256_cvtepu32_epi64(_mm_set1_epi32(qlp_coeff[1 ])); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2() 781 q2 = _mm256_cvtepu32_epi64(_mm_set1_epi32(qlp_coeff[2 ])); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2() 782 q3 = _mm256_cvtepu32_epi64(_mm_set1_epi32(qlp_coeff[3 ])); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2() 783 q4 = _mm256_cvtepu32_epi64(_mm_set1_epi32(qlp_coeff[4 ])); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2() 784 q5 = _mm256_cvtepu32_epi64(_mm_set1_epi32(qlp_coeff[5 ])); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2() 785 q6 = _mm256_cvtepu32_epi64(_mm_set1_epi32(qlp_coeff[6 ])); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2() 786 q7 = _mm256_cvtepu32_epi64(_mm_set1_epi32(qlp_coeff[7 ])); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2() 787 q8 = _mm256_cvtepu32_epi64(_mm_set1_epi32(qlp_coeff[8 ])); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2() 788 q9 = _mm256_cvtepu32_epi64(_mm_set1_epi32(qlp_coeff[9 ])); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_avx2() [all …]
|
/external/libvpx/libvpx/vp9/encoder/x86/ |
D | vp9_diamond_search_sad_avx.c | 79 const __m128i v_max_mv_w = _mm_set1_epi32(maxmv.as_int); in vp9_diamond_search_sad_avx() 81 const __m128i v_min_mv_w = _mm_set1_epi32(minmv.as_int); in vp9_diamond_search_sad_avx() 83 const __m128i v_spb_d = _mm_set1_epi32(sad_per_bit); in vp9_diamond_search_sad_avx() 85 const __m128i v_joint_cost_0_d = _mm_set1_epi32(x->nmvjointsadcost[0]); in vp9_diamond_search_sad_avx() 86 const __m128i v_joint_cost_1_d = _mm_set1_epi32(x->nmvjointsadcost[1]); in vp9_diamond_search_sad_avx() 99 const __m128i vfcmv = _mm_set1_epi32(fcenter_mv.as_int); in vp9_diamond_search_sad_avx() 106 __m128i v_bmv_w = _mm_set1_epi32(bmv.as_int); in vp9_diamond_search_sad_avx() 120 __m128i v_ba_d = _mm_set1_epi32((intptr_t)best_address); in vp9_diamond_search_sad_avx() 237 _mm_set1_epi32(1 << (VP9_PROB_COST_SHIFT - 1))); in vp9_diamond_search_sad_avx() 296 v_bmv_w = _mm_set1_epi32(bmv.as_int); in vp9_diamond_search_sad_avx() [all …]
|
/external/libgav1/libgav1/src/dsp/x86/ |
D | intrapred_smooth_sse4.cc | 66 const __m128i cvtepi32_epi8 = _mm_set1_epi32(0x0C080400); in WriteSmoothHorizontalSum4() 131 const __m128i cvtepi32_epi8 = _mm_set1_epi32(0x0C080400); in WriteSmoothPredSum4() 185 const __m128i round = _mm_set1_epi32(256); in WriteSmoothPred4x8() 187 const __m128i cvtepi32_epi8 = _mm_set1_epi32(0x0C080400); in WriteSmoothPred4x8() 223 const __m128i scale = _mm_set1_epi32(256); in Smooth4x4_SSE4_1() 365 const __m128i round = _mm_set1_epi32(256); in WriteSmoothPred8xH() 486 const __m128i round = _mm_set1_epi32(256); in SmoothWxH() 538 const __m128i top_right = _mm_set1_epi32(top_ptr[3]); in SmoothHorizontal4x4_SSE4_1() 542 __m128i scale = _mm_set1_epi32(256); in SmoothHorizontal4x4_SSE4_1() 545 scale = _mm_set1_epi32(128); in SmoothHorizontal4x4_SSE4_1() [all …]
|
D | distance_weighted_blend_sse4.cc | 60 const __m128i weights = _mm_set1_epi32(weight_0 | (weight_1 << 16)); in DistanceWeightedBlend4xH_SSE4_1() 104 const __m128i weights = _mm_set1_epi32(weight_0 | (weight_1 << 16)); in DistanceWeightedBlend8xH_SSE4_1() 132 const __m128i weights = _mm_set1_epi32(weight_0 | (weight_1 << 16)); in DistanceWeightedBlendLarge_SSE4_1() 235 const __m128i bias = _mm_set1_epi32(offset); in ComputeWeightedAverage8() 263 const __m128i weight0 = _mm_set1_epi32(weight_0); in DistanceWeightedBlend4xH_SSE4_1() 264 const __m128i weight1 = _mm_set1_epi32(weight_1); in DistanceWeightedBlend4xH_SSE4_1() 307 const __m128i weight0 = _mm_set1_epi32(weight_0); in DistanceWeightedBlend8xH_SSE4_1() 308 const __m128i weight1 = _mm_set1_epi32(weight_1); in DistanceWeightedBlend8xH_SSE4_1() 339 const __m128i weight0 = _mm_set1_epi32(weight_0); in DistanceWeightedBlendLarge_SSE4_1() 340 const __m128i weight1 = _mm_set1_epi32(weight_1); in DistanceWeightedBlendLarge_SSE4_1()
|
/external/mesa3d/src/gallium/drivers/swr/rasterizer/core/ |
D | state_funcs.h | 36 __m128i vMin = _mm_set1_epi32(*min); in expandThenBlend4() 37 __m128i vMax = _mm_set1_epi32(*max); in expandThenBlend4() 45 _vXi[i] = _mm_set1_epi32(_xi[i]); in PrecalcSampleData() 46 _vYi[i] = _mm_set1_epi32(_yi[i]); in PrecalcSampleData()
|
/external/libavc/encoder/x86/ |
D | ih264e_half_pel_ssse3.c | 120 coeff0_1_16x8b = _mm_set1_epi32(0xFB01FB01); //c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 in ih264e_sixtapfilter_horz_ssse3() 121 coeff2_3_16x8b = _mm_set1_epi32(0x14141414); //c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 in ih264e_sixtapfilter_horz_ssse3() 122 coeff4_5_16x8b = _mm_set1_epi32(0x01FB01FB); //c4 c5 c4 c5 c4 c5 c4 c5 c4 c5 c4 c5 c4 c5 c4 c5 in ih264e_sixtapfilter_horz_ssse3() 273 … coeff0_1_16x8b = _mm_set1_epi32(0xFB01FB01); //c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 in ih264e_sixtap_filter_2dvh_vert_ssse3() 274 … coeff2_3_16x8b = _mm_set1_epi32(0x14141414); //c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 in ih264e_sixtap_filter_2dvh_vert_ssse3() 275 … coeff4_5_16x8b = _mm_set1_epi32(0x01FB01FB); //c4 c5 c4 c5 c4 c5 c4 c5 c4 c5 c4 c5 c4 c5 c4 c5 in ih264e_sixtap_filter_2dvh_vert_ssse3() 384 coeff0_1_8x16b = _mm_set1_epi32(0xFFFB0001); //c0 c1 c0 c1 c0 c1 c0 c1 in ih264e_sixtap_filter_2dvh_vert_ssse3() 385 coeff2_3_8x16b = _mm_set1_epi32(0x00140014); //c2 c3 c2 c3 c2 c3 c2 c3 in ih264e_sixtap_filter_2dvh_vert_ssse3() 386 coeff4_5_8x16b = _mm_set1_epi32(0x0001FFFB); //c4 c5 c4 c5 c4 c5 c4 c5 in ih264e_sixtap_filter_2dvh_vert_ssse3() 388 const_val512_4x32b = _mm_set1_epi32(512); in ih264e_sixtap_filter_2dvh_vert_ssse3()
|
/external/libavc/common/x86/ |
D | ih264_inter_pred_filters_ssse3.c | 233 coeff0_1_16x8b = _mm_set1_epi32(0xFB01FB01); //c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 in ih264_inter_pred_luma_horz_ssse3() 234 coeff2_3_16x8b = _mm_set1_epi32(0x14141414); //c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 in ih264_inter_pred_luma_horz_ssse3() 235 coeff4_5_16x8b = _mm_set1_epi32(0x01FB01FB); //c4 c5 c4 c5 c4 c5 c4 c5 c4 c5 c4 c5 c4 c5 c4 c5 in ih264_inter_pred_luma_horz_ssse3() 502 coeff0_1_16x8b = _mm_set1_epi32(0xFB01FB01); //c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 in ih264_inter_pred_luma_vert_ssse3() 503 coeff2_3_16x8b = _mm_set1_epi32(0x14141414); //c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 in ih264_inter_pred_luma_vert_ssse3() 504 coeff4_5_16x8b = _mm_set1_epi32(0x01FB01FB); //c4 c5 c4 c5 c4 c5 c4 c5 c4 c5 c4 c5 c4 c5 c4 c5 in ih264_inter_pred_luma_vert_ssse3() 797 … coeff0_1_16x8b = _mm_set1_epi32(0xFB01FB01); //c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 c0 c1 in ih264_inter_pred_luma_horz_hpel_vert_hpel_ssse3() 798 … coeff2_3_16x8b = _mm_set1_epi32(0x14141414); //c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 c2 c3 in ih264_inter_pred_luma_horz_hpel_vert_hpel_ssse3() 799 … coeff4_5_16x8b = _mm_set1_epi32(0x01FB01FB); //c4 c5 c5 c5 c4 c5 c5 c5 c4 c5 c5 c5 c4 c5 c5 c5 in ih264_inter_pred_luma_horz_hpel_vert_hpel_ssse3() 876 coeff0_1_8x16b = _mm_set1_epi32(0xFFFB0001); in ih264_inter_pred_luma_horz_hpel_vert_hpel_ssse3() [all …]
|
D | ih264_deblk_chroma_ssse3.c | 147 alpha_cbcr_16x8 = _mm_set1_epi32(alpha_cbcr); in ih264_deblk_chroma_vert_bs4_ssse3() 152 beta_cbcr_16x8 = _mm_set1_epi32(beta_cbcr); in ih264_deblk_chroma_vert_bs4_ssse3() 178 alpha_cbcr_16x8 = _mm_set1_epi32(alpha_cbcr); in ih264_deblk_chroma_vert_bs4_ssse3() 183 beta_cbcr_16x8 = _mm_set1_epi32(beta_cbcr); in ih264_deblk_chroma_vert_bs4_ssse3() 316 alpha_cbcr_16x8 = _mm_set1_epi32(alpha_cbcr); in ih264_deblk_chroma_horz_bs4_ssse3() 321 beta_cbcr_16x8 = _mm_set1_epi32(beta_cbcr); in ih264_deblk_chroma_horz_bs4_ssse3() 347 alpha_cbcr_16x8 = _mm_set1_epi32(alpha_cbcr); in ih264_deblk_chroma_horz_bs4_ssse3() 352 beta_cbcr_16x8 = _mm_set1_epi32(beta_cbcr); in ih264_deblk_chroma_horz_bs4_ssse3() 496 alpha_cbcr_16x8 = _mm_set1_epi32(alpha_cbcr); in ih264_deblk_chroma_vert_bslt4_ssse3() 501 beta_cbcr_16x8 = _mm_set1_epi32(beta_cbcr); in ih264_deblk_chroma_vert_bslt4_ssse3() [all …]
|
/external/gemmlowp/fixedpoint/ |
D | fixedpoint_sse.h | 90 return _mm_andnot_si128(a, _mm_set1_epi32(-1)); 130 return _mm_sign_epi32(a, _mm_set1_epi32(-1)); 196 return MaskIfEqual(a, _mm_set1_epi32(0)); 206 return MaskIfNotEqual(a, _mm_set1_epi32(0)); 297 one = _mm_set1_epi32(1); 298 sign_bit_mask = _mm_set1_epi32(0x80000000); 328 min = _mm_set1_epi32(std::numeric_limits<std::int32_t>::min()); 372 return _mm_set1_epi32(x);
|
/external/oboe/samples/RhythmGame/third_party/glm/detail/ |
D | func_integer_simd.inl | 18 __m128i const set1 = _mm_set1_epi32(Mask); 22 __m128i const set2 = _mm_andnot_si128(set0, _mm_set1_epi32(-1)); 39 __m128i const set1 = _mm_set1_epi32(Mask);
|
/external/libvpx/libvpx/vpx_dsp/x86/ |
D | highbd_inv_txfm_sse2.h | 220 io[0] = wraplow_16bit_shift5(io[0], io[8], _mm_set1_epi32(16)); in highbd_idct8x8_final_round() 221 io[1] = wraplow_16bit_shift5(io[1], io[9], _mm_set1_epi32(16)); in highbd_idct8x8_final_round() 222 io[2] = wraplow_16bit_shift5(io[2], io[10], _mm_set1_epi32(16)); in highbd_idct8x8_final_round() 223 io[3] = wraplow_16bit_shift5(io[3], io[11], _mm_set1_epi32(16)); in highbd_idct8x8_final_round() 224 io[4] = wraplow_16bit_shift5(io[4], io[12], _mm_set1_epi32(16)); in highbd_idct8x8_final_round() 225 io[5] = wraplow_16bit_shift5(io[5], io[13], _mm_set1_epi32(16)); in highbd_idct8x8_final_round() 226 io[6] = wraplow_16bit_shift5(io[6], io[14], _mm_set1_epi32(16)); in highbd_idct8x8_final_round() 227 io[7] = wraplow_16bit_shift5(io[7], io[15], _mm_set1_epi32(16)); in highbd_idct8x8_final_round() 395 const __m128i final_rounding = _mm_set1_epi32(1 << 5); in highbd_write_buffer_4()
|
D | highbd_idct4x4_add_sse4.c | 42 io[0] = wraplow_16bit_shift4(io[0], io[1], _mm_set1_epi32(8)); in vpx_highbd_idct4x4_16_add_sse4_1() 43 io[1] = wraplow_16bit_shift4(io[2], io[3], _mm_set1_epi32(8)); in vpx_highbd_idct4x4_16_add_sse4_1()
|
/external/XNNPACK/src/math/ |
D | exp-sse2-rr2-lut64-p2.c | 39 const __m128i vmin_exponent = _mm_set1_epi32(0xC1000000); in xnn_math_f32_exp__sse2_rr2_lut64_p2() 40 const __m128i vmax_exponent = _mm_set1_epi32(0x3F800000); in xnn_math_f32_exp__sse2_rr2_lut64_p2() 42 const __m128i vindex_mask = _mm_set1_epi32(0x3F); in xnn_math_f32_exp__sse2_rr2_lut64_p2()
|