/external/XNNPACK/src/u8-rmax/ |
D | neon.c | 36 uint8x8_t vmax8 = vmax_u8(vget_low_u8(vmax), vget_high_u8(vmax)); in xnn_u8_rmax_ukernel__neon() 45 vmax = vmax_u8(vmax, vx); in xnn_u8_rmax_ukernel__neon()
|
/external/XNNPACK/src/u8-clamp/ |
D | neon-x64.c | 46 vout = vmax_u8(vout, vget_low_u8(voutput_min)); in xnn_u8_clamp_ukernel__neon_x64() 52 vout = vmax_u8(vout, vget_low_u8(voutput_min)); in xnn_u8_clamp_ukernel__neon_x64()
|
/external/XNNPACK/src/qu8-gavgpool/ |
D | 7x-minmax-neon-c8.c | 129 vout = vmax_u8(vout, voutput_min); in xnn_qu8_gavgpool_minmax_ukernel_7x__neon_c8() 199 vout = vmax_u8(vout, voutput_min); in xnn_qu8_gavgpool_minmax_ukernel_7x__neon_c8()
|
D | 7p7x-minmax-neon-c8.c | 205 vout = vmax_u8(vout, voutput_min); in xnn_qu8_gavgpool_minmax_ukernel_7p7x__neon_c8() 277 vout = vmax_u8(vout, voutput_min); in xnn_qu8_gavgpool_minmax_ukernel_7p7x__neon_c8()
|
/external/webp/src/dsp/ |
D | cost_neon.c | 35 const uint8x8_t eob_8x8 = vmax_u8(vget_low_u8(masked), vget_high_u8(masked)); in SetResidualCoeffs_NEON()
|
/external/libvpx/libvpx/vpx_dsp/arm/ |
D | deblock_neon.c | 37 uint8x8_t max = vmax_u8(a2_v0, a1_v0); in generate_mask() 38 max = vmax_u8(b1_v0, max); in generate_mask() 39 max = vmax_u8(b2_v0, max); in generate_mask()
|
D | avg_neon.c | 223 uint8x8_t ab_max = vmax_u8(vget_high_u8(ab07_max), vget_low_u8(ab07_max)); in vpx_minmax_8x8_neon()
|
/external/libaom/libaom/aom_dsp/arm/ |
D | loopfilter_neon.c | 33 mask_8x8 = vmax_u8(mask_8x8, vabd_u8(p2q2, p1q1)); in lpf_mask() 34 mask_8x8 = vmax_u8(mask_8x8, vabd_u8(p1q1, p0q0)); in lpf_mask() 91 flat_8x8 = vmax_u8(flat_8x8, vabd_u8(p2q2, p0q0)); in lpf_flat_mask4() 92 flat_8x8 = vmax_u8(flat_8x8, vabd_u8(p3q3, p0q0)); in lpf_flat_mask4() 107 flat_8x8 = vmax_u8(flat_8x8, vabd_u8(p2q2, p0q0)); in lpf_flat_mask3() 128 mask_8x8 = vmax_u8(mask_8x8, vabd_u8(p1q1, p0q0)); in lpf_mask3_chroma()
|
/external/libgav1/libgav1/src/dsp/arm/ |
D | loop_filter_neon.cc | 252 const uint8x8_t a = vmax_u8(abd_p0p1_q0q1, abd_p0p2_q0q2); in IsFlat3() 265 const uint8x8_t a = vmax_u8(abd_p0p1_q0q1, abd_p1p2_q1q2); in NeedsFilter6() 482 const uint8x8_t a = vmax_u8(abd_p0n0_q0n0, abd_p0n1_q0n1); in IsFlat4() 483 const uint8x8_t b = vmax_u8(a, abd_p0n2_q0n2); in IsFlat4() 498 const uint8x8_t a = vmax_u8(abd_p0p1_q0q1, abd_p1p2_q1q2); in NeedsFilter8() 499 const uint8x8_t b = vmax_u8(a, abd_p2p3_q2q3); in NeedsFilter8()
|
D | loop_restoration_neon.cc | 1039 val = vmax_u8(val, vdup_n_u8(5)); in CalculateIntermediate()
|
/external/XNNPACK/src/qu8-vadd/ |
D | minmax-neon.c | 188 vy = vmax_u8(vy, vget_low_u8(vy_min)); in xnn_qu8_vadd_minmax_ukernel__neon() 231 vy = vmax_u8(vy, vget_low_u8(vy_min)); in xnn_qu8_vadd_minmax_ukernel__neon()
|
/external/XNNPACK/src/qu8-avgpool/ |
D | 9x-minmax-neon-c8.c | 184 vout = vmax_u8(vout, voutput_min); in xnn_qu8_avgpool_minmax_ukernel_9x__neon_c8() 257 vout = vmax_u8(vout, voutput_min); in xnn_qu8_avgpool_minmax_ukernel_9x__neon_c8()
|
D | 9p8x-minmax-neon-c8.c | 328 vout = vmax_u8(vout, voutput_min); in xnn_qu8_avgpool_minmax_ukernel_9p8x__neon_c8() 402 vout = vmax_u8(vout, voutput_min); in xnn_qu8_avgpool_minmax_ukernel_9p8x__neon_c8()
|
/external/XNNPACK/src/qu8-dwconv/ |
D | up8x9-minmax-neon.c | 162 vout = vmax_u8(vout, voutput_min); in xnn_qu8_dwconv_minmax_ukernel_up8x9__neon() 253 vout = vmax_u8(vout, voutput_min); in xnn_qu8_dwconv_minmax_ukernel_up8x9__neon()
|
/external/libaom/libaom/aom_dsp/simd/ |
D | v64_intrinsics_arm.h | 341 vmax_u8(vreinterpret_u8_s64(x), vreinterpret_u8_s64(y))); in v64_max_u8()
|
/external/tensorflow/tensorflow/lite/kernels/ |
D | cpu_backend_gemm_custom_gemv.h | 243 res8 = vmax_u8(res8, vdup_n_u8(clamp_min)); in ClampAndStore()
|
/external/skia/src/effects/imagefilters/ |
D | SkMorphologyImageFilter.cpp | 565 extreme = (type == MorphType::kDilate) ? vmax_u8(src_pixel, extreme)
|
/external/skqp/src/effects/imagefilters/ |
D | SkMorphologyImageFilter.cpp | 582 extreme = (type == kDilate) ? vmax_u8(src_pixel, extreme) in morph()
|
/external/tensorflow/tensorflow/lite/kernels/internal/optimized/ |
D | optimized_ops.h | 1920 vmax_u8(output_activation_min_vector, in AddElementwise() 2018 vmax_u8(output_activation_min_vector, in AddScalarBroadcast() 2431 vmax_u8(output_activation_min_vector, in MulElementwise() 2501 vmax_u8(output_activation_min_vector, in MulSimpleBroadcast() 3342 buf8 = vmax_u8(buf8, vdup_n_u8(params.quantized_activation_min)); \ in AveragePool() 3356 buf8 = vmax_u8(buf8, vdup_n_u8(params.quantized_activation_min)); in AveragePool() 3496 acc_reg = vmax_u8(acc_reg, input_reg); in MaxPool() 3519 a = vmax_u8(a, vdup_n_u8(params.quantized_activation_min)); in MaxPool()
|
D | legacy_optimized_ops.h | 978 res8 = vmax_u8(res8, vdup_n_u8(output_activation_min)); in LegacyFullyConnectedAsGEMVWorkerImpl() 4035 uint8x8_t max8 = vmax_u8(vget_low_u8(max16), vget_high_u8(max16)); in Softmax() 4037 max8 = vmax_u8(max8, vld1_u8(input_data_ptr + c)); in Softmax() 4040 uint8x8_t max4 = vmax_u8(max8, vext_u8(max8, max8, 4)); in Softmax() 4041 uint8x8_t max2 = vmax_u8(max4, vext_u8(max4, max4, 2)); in Softmax()
|
D | depthwiseconv_uint8_transitional.h | 3243 return vmax_u8(a, b); 3895 return vmax_u8(a, b); 4347 return vmax_u8(a, b); 5021 return vmax_u8(a, b);
|
/external/pffft/ |
D | sse2neon.h | 3372 vmax_u8(vreinterpret_u8_m64(a), vreinterpret_u8_m64(b))); in _mm_max_pu8()
|
/external/llvm-project/clang/test/CodeGen/ |
D | aarch64-neon-intrinsics.c | 3770 return vmax_u8(a, b); in test_vmax_u8()
|
D | arm_neon_intrinsics.c | 6718 return vmax_u8(a, b); in test_vmax_u8()
|
/external/clang/test/CodeGen/ |
D | arm_neon_intrinsics.c | 8527 return vmax_u8(a, b); in test_vmax_u8()
|