/external/libgav1/src/dsp/arm/ |
D | loop_filter_10bit_neon.cc | 265 const uint16x8_t p0q0_output = vbslq_u16(needs_filter4_mask_8, f_p0q0, p0q0); in Horizontal4_NEON() 270 const uint16x8_t p1q1_output = vbslq_u16(p1q1_mask, f_p1q1, p1q1); in Horizontal4_NEON() 322 const uint16x8_t p0q0_output = vbslq_u16(needs_filter4_mask_8, f_p0q0, p0q0); in Vertical4_NEON() 327 const uint16x8_t p1q1_output = vbslq_u16(p1q1_mask, f_p1q1, p1q1); in Vertical4_NEON() 436 f4_p1q1 = vbslq_u16(hev_mask_8, p1q1, f4_p1q1); in Horizontal6_NEON() 448 p1q1_output = vbslq_u16(needs_filter_mask_8, f4_p1q1, p1q1); in Horizontal6_NEON() 449 p0q0_output = vbslq_u16(needs_filter_mask_8, f4_p0q0, p0q0); in Horizontal6_NEON() 452 p1q1_output = vbslq_u16(is_flat3_mask_8, f6_p1q1, f4_p1q1); in Horizontal6_NEON() 453 p1q1_output = vbslq_u16(needs_filter_mask_8, p1q1_output, p1q1); in Horizontal6_NEON() 454 p0q0_output = vbslq_u16(is_flat3_mask_8, f6_p0q0, f4_p0q0); in Horizontal6_NEON() [all …]
|
D | intra_edge_neon.cc | 368 const uint16x8_t dst_remainder = vbslq_u16(mask, result, src_1); in IntraEdgeFilter_NEON() 435 const uint16x8_t dst_remainder = vbslq_u16(mask, result, src_2); in IntraEdgeFilter_NEON()
|
D | intrapred_neon.cc | 1291 uint16x8_t result = vbslq_u16(left_mask, left, top); in Paeth8xH_NEON() 1298 result = vbslq_u16(left_or_top_mask, result, top_left); in Paeth8xH_NEON() 1341 uint16x8_t result = vbslq_u16(left_mask, left, top[i]); in PaethWxH_NEON() 1348 result = vbslq_u16(left_or_top_mask, result, top_left); in PaethWxH_NEON()
|
D | intrapred_cfl_neon.cc | 120 vbslq_u16(x_mask, sum_row_shifted, max_luma_sum); in CflSubsampler420_NEON() 154 vbslq_u16(x_mask, sum_row_shifted, max_luma_sum); in CflSubsampler420_NEON()
|
D | intrapred_directional_neon.cc | 1062 vbslq_u16(max_base_mask, combined, final_top_val); in DirectionalZone1_WxH() 1155 vbslq_u16(max_base_mask, combined, final_top_val); in DirectionalZone1_Large() 1866 const uint16x8_t output = vbslq_u16(blend, val, dst_blend); in DirectionalZone1Blend_8x8()
|
/external/libaom/aom_dsp/arm/ |
D | highbd_loopfilter_neon.c | 268 const uint16x8_t p0q0_output = vbslq_u16(needs_filter4_mask_8, f_p0q0, p0q0); in aom_highbd_lpf_horizontal_4_neon() 273 const uint16x8_t p1q1_output = vbslq_u16(p1q1_mask, f_p1q1, p1q1); in aom_highbd_lpf_horizontal_4_neon() 334 const uint16x8_t p0q0_output = vbslq_u16(needs_filter4_mask_8, f_p0q0, p0q0); in aom_highbd_lpf_vertical_4_neon() 339 const uint16x8_t p1q1_output = vbslq_u16(p1q1_mask, f_p1q1, p1q1); in aom_highbd_lpf_vertical_4_neon() 458 f4_p1q1 = vbslq_u16(hev_mask_8, p1q1, f4_p1q1); in aom_highbd_lpf_horizontal_6_neon() 470 p1q1_output = vbslq_u16(needs_filter_mask_8, f4_p1q1, p1q1); in aom_highbd_lpf_horizontal_6_neon() 471 p0q0_output = vbslq_u16(needs_filter_mask_8, f4_p0q0, p0q0); in aom_highbd_lpf_horizontal_6_neon() 474 p1q1_output = vbslq_u16(is_flat3_mask_8, f6_p1q1, f4_p1q1); in aom_highbd_lpf_horizontal_6_neon() 475 p1q1_output = vbslq_u16(needs_filter_mask_8, p1q1_output, p1q1); in aom_highbd_lpf_horizontal_6_neon() 476 p0q0_output = vbslq_u16(is_flat3_mask_8, f6_p0q0, f4_p0q0); in aom_highbd_lpf_horizontal_6_neon() [all …]
|
D | highbd_intrapred_neon.c | 248 uint16x8_t result = vbslq_u16(left_mask, left, top); in highbd_paeth_4or8_x_h_neon() 255 result = vbslq_u16(left_or_top_mask, result, top_left); in highbd_paeth_4or8_x_h_neon() 294 const uint16x8_t result = vbslq_u16(left_mask, left, top); in select_paeth() 301 return vbslq_u16(left_or_top_mask, result, top_left); in select_paeth()
|
/external/XNNPACK/src/f32-f16-vcvt/gen/ |
D | vcvt-neon-x32.c | 155 vh0 = vbslq_u16(vnanmaskh0, vnanh, vh0); in xnn_f32_f16_vcvt_ukernel__neon_x32() 156 vh1 = vbslq_u16(vnanmaskh1, vnanh, vh1); in xnn_f32_f16_vcvt_ukernel__neon_x32() 157 vh2 = vbslq_u16(vnanmaskh2, vnanh, vh2); in xnn_f32_f16_vcvt_ukernel__neon_x32() 158 vh3 = vbslq_u16(vnanmaskh3, vnanh, vh3); in xnn_f32_f16_vcvt_ukernel__neon_x32()
|
D | vcvt-neon-x24.c | 129 vh0 = vbslq_u16(vnanmaskh0, vnanh, vh0); in xnn_f32_f16_vcvt_ukernel__neon_x24() 130 vh1 = vbslq_u16(vnanmaskh1, vnanh, vh1); in xnn_f32_f16_vcvt_ukernel__neon_x24() 131 vh2 = vbslq_u16(vnanmaskh2, vnanh, vh2); in xnn_f32_f16_vcvt_ukernel__neon_x24()
|
D | vcvt-neon-x16.c | 103 vh0 = vbslq_u16(vnanmaskh0, vnanh, vh0); in xnn_f32_f16_vcvt_ukernel__neon_x16() 104 vh1 = vbslq_u16(vnanmaskh1, vnanh, vh1); in xnn_f32_f16_vcvt_ukernel__neon_x16()
|
D | vcvt-neon-x8.c | 77 vh0 = vbslq_u16(vnanmaskh0, vnanh, vh0); in xnn_f32_f16_vcvt_ukernel__neon_x8()
|
/external/libvpx/vpx_dsp/arm/ |
D | highbd_loopfilter_neon.c | 144 return vbslq_u16(flat, vrshrq_n_u16(*sum, 4), in); in apply_15_tap_filter_kernel() 182 *op2 = vbslq_u16(flat, *op2, p2); in apply_7_tap_filter() 183 *op1 = vbslq_u16(flat, tp1, *op1); in apply_7_tap_filter() 184 *op0 = vbslq_u16(flat, tp0, *op0); in apply_7_tap_filter() 185 *oq0 = vbslq_u16(flat, tq0, *oq0); in apply_7_tap_filter() 186 *oq1 = vbslq_u16(flat, tq1, *oq1); in apply_7_tap_filter() 187 *oq2 = vbslq_u16(flat, *oq2, q2); in apply_7_tap_filter() 213 *op6 = vbslq_u16(flat2, vrshrq_n_u16(sum, 4), p6); in apply_15_tap_filter()
|
/external/ComputeLibrary/src/cpu/kernels/pool2d/neon/ |
D | fp16.cpp | 115 … const uint16x8_t tmp_indices0 = vbslq_u16(vcgeq_f16(v_x0, v_x1), voffset_x0, voffset_x1); in pooling2_f16_maxpool_indices() 116 … const uint16x8_t tmp_indices1 = vbslq_u16(vcgeq_f16(v_x2, v_x3), voffset_x2, voffset_x3); in pooling2_f16_maxpool_indices() 117 …const uint16x8_t tmp_indices2 = vbslq_u16(vcgeq_f16(vmaxq_f16(v_x0, v_x1), vmaxq_f16(v_x2, v_x3)… in pooling2_f16_maxpool_indices()
|
/external/XNNPACK/src/math/ |
D | cvt-f32-f16-neon.c | 69 vh = vbslq_u16(vnanmaskh, vnanh, vh); in xnn_math_f32_f16_cvt__neon()
|
/external/XNNPACK/src/f32-f16-vcvt/ |
D | neon.c.in | 85 vh${N} = vbslq_u16(vnanmaskh${N}, vnanh, vh${N});
|
/external/llvm/test/CodeGen/AArch64/ |
D | fp16-vector-shuffle.ll | 18 ; float16x8_t select_128(float16x8_t a, float16x8_t b, uint16x8_t c) { return vbslq_u16(c, a, b); }
|
/external/psimd/include/ |
D | psimd.h | 724 return (psimd_u16) vbslq_u16((uint16x8_t) mask, (uint16x8_t) a, (uint16x8_t) b); in psimd_blend_u16()
|
/external/eigen/Eigen/src/Core/arch/NEON/ |
D | PacketMath.h | 3079 { return vbslq_u16(mask, a, b); } 3246 res = vbslq_u16(vcgeq_u16(a, vmulq_u16(temp, temp)), temp, res);
|
/external/pffft/ |
D | sse2neon.h | 1844 vreinterpretq_m128i_u16(vbslq_u16(_mask_vec, _b, _a)); \
|
/external/clang/test/CodeGen/ |
D | aarch64-neon-intrinsics.c | 1276 return vbslq_u16(v1, v2, v3); in test_vbslq_u16()
|
D | arm_neon_intrinsics.c | 1290 return vbslq_u16(a, b, c); in test_vbslq_u16()
|
/external/neon_2_sse/ |
D | NEON_2_SSE.h | 2210 _NEON2SSESTORAGE uint16x8_t vbslq_u16(uint16x8_t a, uint16x8_t b, uint16x8_t c); // VBSL q0,q0,q0 15626 _NEON2SSESTORAGE uint16x8_t vbslq_u16(uint16x8_t a, uint16x8_t b, uint16x8_t c); // VBSL q0,q0,q0 15627 #define vbslq_u16 vbslq_s8 macro
|