/external/libjpeg-turbo/simd/arm/aarch64/ |
D | jccolext-neon.c | 87 uint32x4_t y_lh = vmull_laneq_u16(vget_high_u16(r_l), consts, 0); in jsimd_rgb_ycc_convert_neon() 88 y_lh = vmlal_laneq_u16(y_lh, vget_high_u16(g_l), consts, 1); in jsimd_rgb_ycc_convert_neon() 89 y_lh = vmlal_laneq_u16(y_lh, vget_high_u16(b_l), consts, 2); in jsimd_rgb_ycc_convert_neon() 93 uint32x4_t y_hh = vmull_laneq_u16(vget_high_u16(r_h), consts, 0); in jsimd_rgb_ycc_convert_neon() 94 y_hh = vmlal_laneq_u16(y_hh, vget_high_u16(g_h), consts, 1); in jsimd_rgb_ycc_convert_neon() 95 y_hh = vmlal_laneq_u16(y_hh, vget_high_u16(b_h), consts, 2); in jsimd_rgb_ycc_convert_neon() 103 cb_lh = vmlsl_laneq_u16(cb_lh, vget_high_u16(r_l), consts, 3); in jsimd_rgb_ycc_convert_neon() 104 cb_lh = vmlsl_laneq_u16(cb_lh, vget_high_u16(g_l), consts, 4); in jsimd_rgb_ycc_convert_neon() 105 cb_lh = vmlal_laneq_u16(cb_lh, vget_high_u16(b_l), consts, 5); in jsimd_rgb_ycc_convert_neon() 111 cb_hh = vmlsl_laneq_u16(cb_hh, vget_high_u16(r_h), consts, 3); in jsimd_rgb_ycc_convert_neon() [all …]
|
/external/libgav1/libgav1/src/dsp/arm/ |
D | super_res_neon.cc | 217 uint32x4_t res_hi = vmull_u16(vget_high_u16(src[1]), vget_high_u16(f[1])); in SuperRes() 218 res_hi = vmlal_u16(res_hi, vget_high_u16(src[3]), vget_high_u16(f[3])); in SuperRes() 219 res_hi = vmlal_u16(res_hi, vget_high_u16(src[4]), vget_high_u16(f[4])); in SuperRes() 220 res_hi = vmlal_u16(res_hi, vget_high_u16(src[6]), vget_high_u16(f[6])); in SuperRes() 222 uint32x4_t temp_hi = vmull_u16(vget_high_u16(src[0]), vget_high_u16(f[0])); in SuperRes() 223 temp_hi = vmlal_u16(temp_hi, vget_high_u16(src[2]), vget_high_u16(f[2])); in SuperRes() 224 temp_hi = vmlal_u16(temp_hi, vget_high_u16(src[5]), vget_high_u16(f[5])); in SuperRes() 225 temp_hi = vmlal_u16(temp_hi, vget_high_u16(src[7]), vget_high_u16(f[7])); in SuperRes()
|
D | intrapred_smooth_neon.cc | 111 vget_high_u16(weighted_top), vget_high_u16(weighted_left), in Smooth4Or8xN_NEON() 112 vget_high_u16(weighted_tr), vget_high_u16(weighted_bl)); in Smooth4Or8xN_NEON() 131 vget_high_u16(weighted_top_low), vget_high_u16(weighted_left_low), in CalculateWeightsAndPred() 132 vget_high_u16(weighted_tr_low), vget_high_u16(weighted_bl)); in CalculateWeightsAndPred() 143 vget_high_u16(weighted_top_high), vget_high_u16(weighted_left_high), in CalculateWeightsAndPred() 144 vget_high_u16(weighted_tr_high), vget_high_u16(weighted_bl)); in CalculateWeightsAndPred()
|
D | average_blend_neon.cc | 151 vaddl_u16(vget_high_u16(pred0), vget_high_u16(pred1)); in AverageBlend8Row() 205 vst1_u16(dst, vget_high_u16(result)); in AverageBlend_NEON()
|
D | intrapred_cfl_neon.cc | 198 vst1_s16(luma[y + 1], vreinterpret_s16_u16(vget_high_u16(row_shifted))); in CflSubsampler444_NEON() 203 vcombine_u16(vget_high_u16(row_shifted), vget_high_u16(row_shifted)); in CflSubsampler444_NEON() 207 vst1_s16(luma[y + 1], vreinterpret_s16_u16(vget_high_u16(row_shifted))); in CflSubsampler444_NEON() 457 return vcombine_u16(vpadd_u16(vget_low_u16(a), vget_high_u16(a)), in vpaddq_u16() 458 vpadd_u16(vget_low_u16(b), vget_high_u16(b))); in vpaddq_u16() 471 const uint16x4_t a = vget_high_u16(row); in LastRowResult() 492 vreinterpret_s16_u16(vget_high_u16(result_shifted))); in StoreLumaResults4_420() 923 vaddl_u16(vget_low_u16(final_fill), vget_high_u16(final_fill)); in CflSubsampler420Impl_8xH_NEON() 1024 vget_low_u16(final_inner_sum), vget_high_u16(final_inner_sum)); in CflSubsampler420Impl_WxH_NEON() 1129 vst1_u16(dst, vget_high_u16(sum)); in CflIntraPredictor4xN_NEON()
|
/external/libhevc/encoder/arm/ |
D | ihevce_ssd_calculator_neon.c | 78 ssd_low = vaddl_u16(vget_low_u16(sqabs_low), vget_high_u16(sqabs_low)); in ihevce_4x4_ssd_computer_neon() 79 ssd_high = vaddl_u16(vget_low_u16(sqabs_high), vget_high_u16(sqabs_high)); in ihevce_4x4_ssd_computer_neon() 104 ssd_val = vaddl_u16(vget_low_u16(sqabs), vget_high_u16(sqabs)); in ihevce_1x8_ssd_computer_neon() 130 ssd_low = vaddl_u16(vget_low_u16(sqabs_low), vget_high_u16(sqabs_low)); in ihevce_1x16_ssd_computer_neon() 131 ssd_high = vaddl_u16(vget_low_u16(sqabs_high), vget_high_u16(sqabs_high)); in ihevce_1x16_ssd_computer_neon() 164 ssd_0 = vaddl_u16(vget_low_u16(sqabs_0), vget_high_u16(sqabs_0)); in ihevce_1x32_ssd_computer_neon() 165 ssd_1 = vaddl_u16(vget_low_u16(sqabs_1), vget_high_u16(sqabs_1)); in ihevce_1x32_ssd_computer_neon() 166 ssd_2 = vaddl_u16(vget_low_u16(sqabs_2), vget_high_u16(sqabs_2)); in ihevce_1x32_ssd_computer_neon() 167 ssd_3 = vaddl_u16(vget_low_u16(sqabs_3), vget_high_u16(sqabs_3)); in ihevce_1x32_ssd_computer_neon() 220 ssd_0 = vaddl_u16(vget_low_u16(sqabs_0), vget_high_u16(sqabs_0)); in ihevce_1x64_ssd_computer_neon() [all …]
|
D | ihevce_ssd_and_sad_calculator_neon.c | 84 b = vaddl_u16(vget_low_u16(sq_abs_l), vget_high_u16(sq_abs_l)); in ihevce_ssd_and_sad_calculator_neon() 85 d = vaddl_u16(vget_low_u16(sq_abs_h), vget_high_u16(sq_abs_h)); in ihevce_ssd_and_sad_calculator_neon() 108 tmp_a = vaddl_u16(vget_low_u16(sqabs), vget_high_u16(sqabs)); in ihevce_ssd_and_sad_calculator_neon() 150 tmp_a = vaddl_u16(vget_low_u16(sqabs_l), vget_high_u16(sqabs_l)); in ihevce_ssd_and_sad_calculator_neon() 151 tmp_c = vaddl_u16(vget_low_u16(sqabs_h), vget_high_u16(sqabs_h)); in ihevce_ssd_and_sad_calculator_neon() 196 tmp_a = vaddl_u16(vget_low_u16(sqabs_l), vget_high_u16(sqabs_l)); in ihevce_ssd_and_sad_calculator_neon() 197 tmp_c = vaddl_u16(vget_low_u16(sqabs_h), vget_high_u16(sqabs_h)); in ihevce_ssd_and_sad_calculator_neon() 207 tmp_a = vaddl_u16(vget_low_u16(sqabs_l), vget_high_u16(sqabs_l)); in ihevce_ssd_and_sad_calculator_neon() 208 tmp_c = vaddl_u16(vget_low_u16(sqabs_h), vget_high_u16(sqabs_h)); in ihevce_ssd_and_sad_calculator_neon() 259 tmp_a = vaddl_u16(vget_low_u16(sqabs_l), vget_high_u16(sqabs_l)); in ihevce_ssd_and_sad_calculator_neon() [all …]
|
D | ihevce_coarse_layer_sad_neon.c | 202 tmp_a0 = vpadd_u16(vget_low_u16(abs_01), vget_high_u16(abs_01)); in hme_store_4x4_sads_high_speed_neon() 203 tmp_a1 = vpadd_u16(vget_low_u16(abs_23), vget_high_u16(abs_23)); in hme_store_4x4_sads_high_speed_neon() 205 tmp_a0 = vpadd_u16(vget_low_u16(abs_01), vget_high_u16(abs_01)); in hme_store_4x4_sads_high_speed_neon() 222 tmp_a = vpadd_u16(vget_low_u16(abs_01), vget_high_u16(abs_01)); in hme_store_4x4_sads_high_speed_neon() 359 tmp_a.val[0] = vpadd_u16(vget_low_u16(abs_a_01), vget_high_u16(abs_a_01)); in hme_store_4x4_sads_high_quality_neon() 360 tmp_a.val[1] = vpadd_u16(vget_low_u16(abs_a_23), vget_high_u16(abs_a_23)); in hme_store_4x4_sads_high_quality_neon() 362 tmp_a.val[0] = vpadd_u16(vget_low_u16(abs_a_01), vget_high_u16(abs_a_01)); in hme_store_4x4_sads_high_quality_neon() 363 tmp_b0 = vpadd_u16(vget_low_u16(abs_b_01), vget_high_u16(abs_b_01)); in hme_store_4x4_sads_high_quality_neon() 364 tmp_b1 = vpadd_u16(vget_low_u16(abs_b_23), vget_high_u16(abs_b_23)); in hme_store_4x4_sads_high_quality_neon() 366 tmp_a.val[1] = vpadd_u16(vget_low_u16(abs_b_01), vget_high_u16(abs_b_01)); in hme_store_4x4_sads_high_quality_neon() [all …]
|
D | ihevce_me_neon.c | 197 tmp_a0 = vpadd_u16(vget_low_u16(abs_01), vget_high_u16(abs_01)); in ihevce_sad4_4x4_neon() 198 tmp_a1 = vpadd_u16(vget_low_u16(abs_23), vget_high_u16(abs_23)); in ihevce_sad4_4x4_neon() 200 tmp_a0 = vpadd_u16(vget_low_u16(abs_01), vget_high_u16(abs_01)); in ihevce_sad4_4x4_neon() 230 tmp_a0 = vpadd_u16(vget_low_u16(abs_0), vget_high_u16(abs_0)); in ihevce_sad4_8x8_neon() 231 tmp_a1 = vpadd_u16(vget_low_u16(abs_1), vget_high_u16(abs_1)); in ihevce_sad4_8x8_neon() 233 tmp_a0 = vpadd_u16(vget_low_u16(abs_2), vget_high_u16(abs_2)); in ihevce_sad4_8x8_neon() 234 tmp_a1 = vpadd_u16(vget_low_u16(abs_3), vget_high_u16(abs_3)); in ihevce_sad4_8x8_neon() 236 tmp_a0 = vpadd_u16(vget_low_u16(abs_0), vget_high_u16(abs_0)); in ihevce_sad4_8x8_neon() 237 tmp_a1 = vpadd_u16(vget_low_u16(abs_1), vget_high_u16(abs_1)); in ihevce_sad4_8x8_neon()
|
D | ihevce_had_compute_neon.c | 218 vget_lane_s64(vreinterpret_s64_u16(vget_high_u16(cmp)), 0)) \ in hadamard_sad8x8_using4x4() 599 uint32x4_t s0 = vaddl_u16(vget_low_u16(r0), vget_high_u16(r0)); in ihevce_HAD_16x16_8bit_neon() 600 uint32x4_t s1 = vaddl_u16(vget_low_u16(r1), vget_high_u16(r1)); in ihevce_HAD_16x16_8bit_neon() 741 uint32x4_t s0 = vaddl_u16(vget_low_u16(r0), vget_high_u16(r0)); in ihevce_HAD_32x32_8bit_neon() 742 uint32x4_t s1 = vaddl_u16(vget_low_u16(r1), vget_high_u16(r1)); in ihevce_HAD_32x32_8bit_neon() 934 vget_lane_s64(vreinterpret_s64_u16(vget_high_u16(cmp)), 0)) \ in ihevce_compute_16x16HAD_using_8x8_neon() 963 uint32x4_t s0 = vaddl_u16(vget_low_u16(r0), vget_high_u16(r0)); in ihevce_compute_16x16HAD_using_8x8_neon() 964 uint32x4_t s1 = vaddl_u16(vget_low_u16(r1), vget_high_u16(r1)); in ihevce_compute_16x16HAD_using_8x8_neon() 1173 uint32x4_t s0 = vaddl_u16(vget_low_u16(r0), vget_high_u16(r0)); in ihevce_compute_32x32HAD_using_16x16_neon() 1174 uint32x4_t s1 = vaddl_u16(vget_low_u16(r1), vget_high_u16(r1)); in ihevce_compute_32x32HAD_using_16x16_neon()
|
/external/libjpeg-turbo/simd/arm/aarch32/ |
D | jccolext-neon.c | 101 uint32x4_t y_high = vmull_lane_u16(vget_high_u16(r), consts.val[0], 0); in jsimd_rgb_ycc_convert_neon() 102 y_high = vmlal_lane_u16(y_high, vget_high_u16(g), consts.val[0], 1); in jsimd_rgb_ycc_convert_neon() 103 y_high = vmlal_lane_u16(y_high, vget_high_u16(b), consts.val[0], 2); in jsimd_rgb_ycc_convert_neon() 111 cb_high = vmlsl_lane_u16(cb_high, vget_high_u16(r), consts.val[0], 3); in jsimd_rgb_ycc_convert_neon() 112 cb_high = vmlsl_lane_u16(cb_high, vget_high_u16(g), consts.val[1], 0); in jsimd_rgb_ycc_convert_neon() 113 cb_high = vmlal_lane_u16(cb_high, vget_high_u16(b), consts.val[1], 1); in jsimd_rgb_ycc_convert_neon() 121 cr_high = vmlal_lane_u16(cr_high, vget_high_u16(r), consts.val[1], 1); in jsimd_rgb_ycc_convert_neon() 122 cr_high = vmlsl_lane_u16(cr_high, vget_high_u16(g), consts.val[1], 2); in jsimd_rgb_ycc_convert_neon() 123 cr_high = vmlsl_lane_u16(cr_high, vget_high_u16(b), consts.val[1], 3); in jsimd_rgb_ycc_convert_neon()
|
/external/libjpeg-turbo/simd/arm/ |
D | jcgryext-neon.c | 78 uint32x4_t y_lh = vmull_n_u16(vget_high_u16(r_l), F_0_298); in jsimd_rgb_gray_convert_neon() 80 uint32x4_t y_hh = vmull_n_u16(vget_high_u16(r_h), F_0_298); in jsimd_rgb_gray_convert_neon() 82 y_lh = vmlal_n_u16(y_lh, vget_high_u16(g_l), F_0_587); in jsimd_rgb_gray_convert_neon() 84 y_hh = vmlal_n_u16(y_hh, vget_high_u16(g_h), F_0_587); in jsimd_rgb_gray_convert_neon() 86 y_lh = vmlal_n_u16(y_lh, vget_high_u16(b_l), F_0_113); in jsimd_rgb_gray_convert_neon() 88 y_hh = vmlal_n_u16(y_hh, vget_high_u16(b_h), F_0_113); in jsimd_rgb_gray_convert_neon()
|
D | jquanti-neon.c | 142 int32x4_t row0_h = vreinterpretq_s32_u32(vmull_u16(vget_high_u16(abs_row0), in jsimd_quantize_neon() 143 vget_high_u16(recip0))); in jsimd_quantize_neon() 146 int32x4_t row1_h = vreinterpretq_s32_u32(vmull_u16(vget_high_u16(abs_row1), in jsimd_quantize_neon() 147 vget_high_u16(recip1))); in jsimd_quantize_neon() 150 int32x4_t row2_h = vreinterpretq_s32_u32(vmull_u16(vget_high_u16(abs_row2), in jsimd_quantize_neon() 151 vget_high_u16(recip2))); in jsimd_quantize_neon() 154 int32x4_t row3_h = vreinterpretq_s32_u32(vmull_u16(vget_high_u16(abs_row3), in jsimd_quantize_neon() 155 vget_high_u16(recip3))); in jsimd_quantize_neon()
|
/external/webrtc/modules/audio_processing/aecm/ |
D | aecm_core_neon.cc | 73 far_energy_v = vaddw_u16(far_energy_v, vget_high_u16(spectrum_v)); in WebRtcAecm_CalcLinearEnergiesNeon() 78 vget_high_u16(spectrum_v)); in WebRtcAecm_CalcLinearEnergiesNeon() 90 vget_high_u16(spectrum_v)); in WebRtcAecm_CalcLinearEnergiesNeon() 152 echo_est_v_high = vmull_u16(vget_high_u16(far_spectrum_v), in WebRtcAecm_StoreAdaptiveChannelNeon() 153 vget_high_u16(vreinterpretq_u16_s16(adapt_v))); in WebRtcAecm_StoreAdaptiveChannelNeon()
|
/external/libvpx/libvpx/vpx_dsp/arm/ |
D | sad4d_neon.c | 54 a[0] = vpadd_u16(vget_low_u16(abs[0]), vget_high_u16(abs[0])); in sad4x_4d() 55 a[1] = vpadd_u16(vget_low_u16(abs[1]), vget_high_u16(abs[1])); in sad4x_4d() 77 const uint16x4_t a0 = vadd_u16(vget_low_u16(sum[0]), vget_high_u16(sum[0])); in sad_512_pel_final_neon() 78 const uint16x4_t a1 = vadd_u16(vget_low_u16(sum[1]), vget_high_u16(sum[1])); in sad_512_pel_final_neon() 79 const uint16x4_t a2 = vadd_u16(vget_low_u16(sum[2]), vget_high_u16(sum[2])); in sad_512_pel_final_neon() 80 const uint16x4_t a3 = vadd_u16(vget_low_u16(sum[3]), vget_high_u16(sum[3])); in sad_512_pel_final_neon() 90 const uint16x4_t a0 = vpadd_u16(vget_low_u16(sum[0]), vget_high_u16(sum[0])); in sad_1024_pel_final_neon() 91 const uint16x4_t a1 = vpadd_u16(vget_low_u16(sum[1]), vget_high_u16(sum[1])); in sad_1024_pel_final_neon() 92 const uint16x4_t a2 = vpadd_u16(vget_low_u16(sum[2]), vget_high_u16(sum[2])); in sad_1024_pel_final_neon() 93 const uint16x4_t a3 = vpadd_u16(vget_low_u16(sum[3]), vget_high_u16(sum[3])); in sad_1024_pel_final_neon()
|
D | highbd_vpx_convolve8_neon.c | 171 s4 = vreinterpret_s16_u16(vget_high_u16(t0)); in vpx_highbd_convolve8_horiz_neon() 172 s5 = vreinterpret_s16_u16(vget_high_u16(t1)); in vpx_highbd_convolve8_horiz_neon() 173 s6 = vreinterpret_s16_u16(vget_high_u16(t2)); in vpx_highbd_convolve8_horiz_neon() 197 vst1_u16(dst + 2 * dst_stride, vget_high_u16(d01)); in vpx_highbd_convolve8_horiz_neon() 198 vst1_u16(dst + 3 * dst_stride, vget_high_u16(d23)); in vpx_highbd_convolve8_horiz_neon() 258 vst1_u16(dst, vget_high_u16(d0)); in vpx_highbd_convolve8_horiz_neon() 260 vst1_u16(dst, vget_high_u16(d1)); in vpx_highbd_convolve8_horiz_neon() 262 vst1_u16(dst, vget_high_u16(d2)); in vpx_highbd_convolve8_horiz_neon() 264 vst1_u16(dst, vget_high_u16(d3)); in vpx_highbd_convolve8_horiz_neon() 380 s4 = vreinterpret_s16_u16(vget_high_u16(t0)); in vpx_highbd_convolve8_avg_horiz_neon() [all …]
|
D | highbd_intrapred_neon.c | 84 uint16x4_t sum = vadd_u16(vget_low_u16(ref_u16), vget_high_u16(ref_u16)); in dc_sum_8() 104 uint16x4_t sum = vadd_u16(vget_low_u16(p0), vget_high_u16(p0)); in vpx_highbd_dc_predictor_8x8_neon() 148 uint16x4_t sum = vadd_u16(vget_low_u16(p0), vget_high_u16(p0)); in dc_sum_16() 171 uint16x4_t pal1 = vadd_u16(vget_low_u16(pal0), vget_high_u16(pal0)); in vpx_highbd_dc_predictor_16x16_neon() 218 uint16x4_t sum = vadd_u16(vget_low_u16(p2), vget_high_u16(p2)); in dc_sum_32() 249 const uint16x4_t pal1 = vadd_u16(vget_low_u16(pal0), vget_high_u16(pal0)); in vpx_highbd_dc_predictor_32x32_neon() 298 const uint16x4_t avg2_high = vget_high_u16(avg2); in vpx_highbd_d45_predictor_4x4_neon() 325 const uint16x8_t above_right = vdupq_lane_u16(vget_high_u16(A0), 3); in vpx_highbd_d45_predictor_8x8_neon() 360 const uint16x8_t above_right = vdupq_lane_u16(vget_high_u16(A0_1), 3); in vpx_highbd_d45_predictor_16x16_neon() 400 const uint16x8_t above_right = vdupq_lane_u16(vget_high_u16(A0_3), 3); in vpx_highbd_d45_predictor_32x32_neon() [all …]
|
D | variance_neon.c | 299 d23s16 = vreinterpret_s16_u16(vget_high_u16(q11u16)); in vpx_mse16x16_neon() 304 d25s16 = vreinterpret_s16_u16(vget_high_u16(q12u16)); in vpx_mse16x16_neon() 309 d27s16 = vreinterpret_s16_u16(vget_high_u16(q13u16)); in vpx_mse16x16_neon() 314 d29s16 = vreinterpret_s16_u16(vget_high_u16(q14u16)); in vpx_mse16x16_neon()
|
/external/libaom/libaom/aom_dsp/arm/ |
D | variance_neon.c | 187 d23s16 = vreinterpret_s16_u16(vget_high_u16(q11u16)); in aom_variance16x8_neon() 193 d25s16 = vreinterpret_s16_u16(vget_high_u16(q12u16)); in aom_variance16x8_neon() 199 d27s16 = vreinterpret_s16_u16(vget_high_u16(q13u16)); in aom_variance16x8_neon() 205 d29s16 = vreinterpret_s16_u16(vget_high_u16(q14u16)); in aom_variance16x8_neon() 261 d23s16 = vreinterpret_s16_u16(vget_high_u16(q11u16)); in aom_variance8x16_neon() 267 d25s16 = vreinterpret_s16_u16(vget_high_u16(q12u16)); in aom_variance8x16_neon() 321 d23s16 = vreinterpret_s16_u16(vget_high_u16(q11u16)); in aom_mse16x16_neon() 326 d25s16 = vreinterpret_s16_u16(vget_high_u16(q12u16)); in aom_mse16x16_neon() 331 d27s16 = vreinterpret_s16_u16(vget_high_u16(q13u16)); in aom_mse16x16_neon() 336 d29s16 = vreinterpret_s16_u16(vget_high_u16(q14u16)); in aom_mse16x16_neon()
|
/external/libaom/libaom/av1/common/arm/ |
D | transpose_neon.h | 367 *a4 = vcombine_u16(vget_high_u16(vreinterpretq_u16_u32(c0.val[0])), in transpose_u16_8x8() 368 vget_high_u16(vreinterpretq_u16_u32(c2.val[0]))); in transpose_u16_8x8() 372 *a6 = vcombine_u16(vget_high_u16(vreinterpretq_u16_u32(c0.val[1])), in transpose_u16_8x8() 373 vget_high_u16(vreinterpretq_u16_u32(c2.val[1]))); in transpose_u16_8x8() 377 *a5 = vcombine_u16(vget_high_u16(vreinterpretq_u16_u32(c1.val[0])), in transpose_u16_8x8() 378 vget_high_u16(vreinterpretq_u16_u32(c3.val[0]))); in transpose_u16_8x8() 382 *a7 = vcombine_u16(vget_high_u16(vreinterpretq_u16_u32(c1.val[1])), in transpose_u16_8x8() 383 vget_high_u16(vreinterpretq_u16_u32(c3.val[1]))); in transpose_u16_8x8()
|
D | selfguided_neon.c | 143 d4 = vget_high_u16(s16_4); in calc_ab_internal_common() 144 d5 = vget_high_u16(s16_5); in calc_ab_internal_common() 145 d6 = vget_high_u16(s16_6); in calc_ab_internal_common() 146 d7 = vget_high_u16(s16_7); in calc_ab_internal_common() 232 s4 = vmull_u16(vget_high_u16(s16_0), one_by_n_minus_1_vec); in calc_ab_internal_common() 233 s5 = vmull_u16(vget_high_u16(s16_1), one_by_n_minus_1_vec); in calc_ab_internal_common() 234 s6 = vmull_u16(vget_high_u16(s16_2), one_by_n_minus_1_vec); in calc_ab_internal_common() 235 s7 = vmull_u16(vget_high_u16(s16_3), one_by_n_minus_1_vec); in calc_ab_internal_common() 241 s4 = vmulq_u32(s4, vmovl_u16(vget_high_u16(s16_4))); in calc_ab_internal_common() 242 s5 = vmulq_u32(s5, vmovl_u16(vget_high_u16(s16_5))); in calc_ab_internal_common() [all …]
|
/external/webp/src/dsp/ |
D | yuv_neon.c | 32 const uint16x4_t r_hi = vget_high_u16(r); in ConvertRGBToY_NEON() 34 const uint16x4_t g_hi = vget_high_u16(g); in ConvertRGBToY_NEON() 36 const uint16x4_t b_hi = vget_high_u16(b); in ConvertRGBToY_NEON() 92 const int16x4_t r_hi = vreinterpret_s16_u16(vget_high_u16(r)); \ 94 const int16x4_t g_hi = vreinterpret_s16_u16(vget_high_u16(g)); \ 96 const int16x4_t b_hi = vreinterpret_s16_u16(vget_high_u16(b))
|
/external/zlib/ |
D | adler32_simd.c | 291 v_s2 = vmlal_u16(v_s2, vget_high_u16(v_column_sum_1), in adler32_simd_() 295 v_s2 = vmlal_u16(v_s2, vget_high_u16(v_column_sum_2), in adler32_simd_() 299 v_s2 = vmlal_u16(v_s2, vget_high_u16(v_column_sum_3), in adler32_simd_() 303 v_s2 = vmlal_u16(v_s2, vget_high_u16(v_column_sum_4), in adler32_simd_()
|
/external/angle/third_party/zlib/ |
D | adler32_simd.c | 291 v_s2 = vmlal_u16(v_s2, vget_high_u16(v_column_sum_1), in adler32_simd_() 295 v_s2 = vmlal_u16(v_s2, vget_high_u16(v_column_sum_2), in adler32_simd_() 299 v_s2 = vmlal_u16(v_s2, vget_high_u16(v_column_sum_3), in adler32_simd_() 303 v_s2 = vmlal_u16(v_s2, vget_high_u16(v_column_sum_4), in adler32_simd_()
|
/external/tensorflow/tensorflow/core/kernels/ |
D | quantized_instance_norm.cc | 74 const uint16x4_t v_high_high = vget_high_u16(v_high_u16); in ColMeanAndVariance() 76 const uint16x4_t v_low_high = vget_high_u16(v_low_u16); in ColMeanAndVariance() 169 vcvtq_f32_u32(vmovl_u16(vget_high_u16(v_high))), in MinAndMax() 171 vcvtq_f32_u32(vmovl_u16(vget_high_u16(v_low))), in MinAndMax() 225 vcvtq_f32_u32(vmovl_u16(vget_high_u16(v_high))), in InstanceNorm() 227 vcvtq_f32_u32(vmovl_u16(vget_high_u16(v_low))), in InstanceNorm()
|