Home
last modified time | relevance | path

Searched refs:vmaxq_s32 (Results 1 – 25 of 31) sorted by relevance

12

/external/XNNPACK/src/f32-hswish/gen/
Dhswish-neon-x16.c47 vacc0123 = vreinterpretq_f32_s32(vmaxq_s32(vreinterpretq_s32_f32(vacc0123), vzero)); in xnn_f32_hswish_ukernel__neon_x16()
48 vacc4567 = vreinterpretq_f32_s32(vmaxq_s32(vreinterpretq_s32_f32(vacc4567), vzero)); in xnn_f32_hswish_ukernel__neon_x16()
49 vacc89AB = vreinterpretq_f32_s32(vmaxq_s32(vreinterpretq_s32_f32(vacc89AB), vzero)); in xnn_f32_hswish_ukernel__neon_x16()
50 vaccCDEF = vreinterpretq_f32_s32(vmaxq_s32(vreinterpretq_s32_f32(vaccCDEF), vzero)); in xnn_f32_hswish_ukernel__neon_x16()
71 vacc = vreinterpretq_f32_s32(vmaxq_s32(vreinterpretq_s32_f32(vacc), vzero)); in xnn_f32_hswish_ukernel__neon_x16()
80 vacc = vreinterpretq_f32_s32(vmaxq_s32(vreinterpretq_s32_f32(vacc), vzero)); in xnn_f32_hswish_ukernel__neon_x16()
Dhswish-neon-x8.c41 vacc0123 = vreinterpretq_f32_s32(vmaxq_s32(vreinterpretq_s32_f32(vacc0123), vzero)); in xnn_f32_hswish_ukernel__neon_x8()
42 vacc4567 = vreinterpretq_f32_s32(vmaxq_s32(vreinterpretq_s32_f32(vacc4567), vzero)); in xnn_f32_hswish_ukernel__neon_x8()
57 vacc = vreinterpretq_f32_s32(vmaxq_s32(vreinterpretq_s32_f32(vacc), vzero)); in xnn_f32_hswish_ukernel__neon_x8()
66 vacc = vreinterpretq_f32_s32(vmaxq_s32(vreinterpretq_s32_f32(vacc), vzero)); in xnn_f32_hswish_ukernel__neon_x8()
Dhswish-neon-x4.c36 vacc = vreinterpretq_f32_s32(vmaxq_s32(vreinterpretq_s32_f32(vacc), vzero)); in xnn_f32_hswish_ukernel__neon_x4()
45 vacc = vreinterpretq_f32_s32(vmaxq_s32(vreinterpretq_s32_f32(vacc), vzero)); in xnn_f32_hswish_ukernel__neon_x4()
/external/XNNPACK/src/f32-hswish/
Dneon.c.in41 …vacc${ABC[N:N+4]} = vreinterpretq_f32_s32(vmaxq_s32(vreinterpretq_s32_f32(vacc${ABC[N:N+4]}), vzer…
56 vacc = vreinterpretq_f32_s32(vmaxq_s32(vreinterpretq_s32_f32(vacc), vzero));
65 vacc = vreinterpretq_f32_s32(vmaxq_s32(vreinterpretq_s32_f32(vacc), vzero));
/external/tensorflow/tensorflow/lite/kernels/internal/optimized/integer_ops/
Dmean.h100 temp_sum.val[0] = vminq_s32(vmaxq_s32(temp_sum.val[0], min_dup), max_dup); in MeanImpl()
101 temp_sum.val[1] = vminq_s32(vmaxq_s32(temp_sum.val[1], min_dup), max_dup); in MeanImpl()
102 temp_sum.val[2] = vminq_s32(vmaxq_s32(temp_sum.val[2], min_dup), max_dup); in MeanImpl()
103 temp_sum.val[3] = vminq_s32(vmaxq_s32(temp_sum.val[3], min_dup), max_dup); in MeanImpl()
/external/libaom/libaom/av1/common/arm/
Dconvolve_neon.h55 sum0 = vmaxq_s32(sum0, zero); in wiener_convolve8_vert_4x8()
56 sum1 = vmaxq_s32(sum1, zero); in wiener_convolve8_vert_4x8()
152 sum_0 = vmaxq_s32(sum_0, zero); in wiener_convolve8_horiz_4x8()
222 sum0 = vmaxq_s32(sum0, zero); in convolve8_4x4_s32()
Dconvolve_neon.c143 sum0 = vmaxq_s32(sum0, zero); in convolve8_vert_4x4_s32()
184 sum0 = vmaxq_s32(sum0, zero); in convolve8_vert_8x4_s32()
185 sum1 = vmaxq_s32(sum1, zero); in convolve8_vert_8x4_s32()
/external/webrtc/common_audio/signal_processing/
Dmin_max_operations_neon.c181 max32x4_0 = vmaxq_s32(max32x4_0, in32x4_0); in WebRtcSpl_MaxValueW32Neon()
182 max32x4_1 = vmaxq_s32(max32x4_1, in32x4_1); in WebRtcSpl_MaxValueW32Neon()
185 int32x4_t max32x4 = vmaxq_s32(max32x4_0, max32x4_1); in WebRtcSpl_MaxValueW32Neon()
/external/tensorflow/tensorflow/lite/kernels/internal/optimized/
Dneon_tensor_utils.cc884 vmaxq_s32(vminq_s32(temp_val.val[0], max_val_dup), min_val_dup); in NeonMatrixBatchVectorAccumulateImpl()
886 vmaxq_s32(vminq_s32(temp_val.val[1], max_val_dup), min_val_dup); in NeonMatrixBatchVectorAccumulateImpl()
949 vmaxq_s32(vminq_s32(temp_val.val[0], max_val_dup), min_val_dup); in NeonMatrixBatchVectorAccumulateImpl()
951 vmaxq_s32(vminq_s32(temp_val.val[1], max_val_dup), min_val_dup); in NeonMatrixBatchVectorAccumulateImpl()
953 vmaxq_s32(vminq_s32(temp_val.val[2], max_val_dup), min_val_dup); in NeonMatrixBatchVectorAccumulateImpl()
955 vmaxq_s32(vminq_s32(temp_val.val[3], max_val_dup), min_val_dup); in NeonMatrixBatchVectorAccumulateImpl()
1787 vmaxq_s32(vminq_s32(temp_val.val[0], max_val_dup), min_val_dup); in NeonCwiseMul()
1789 vmaxq_s32(vminq_s32(temp_val.val[1], max_val_dup), min_val_dup); in NeonCwiseMul()
2338 int32x4_t max0_i32x4 = vmaxq_s32(f2i0_i32x4, neg_scale_i32x4); in NeonSymmetricQuantizeFloats()
2339 int32x4_t max1_i32x4 = vmaxq_s32(f2i1_i32x4, neg_scale_i32x4); in NeonSymmetricQuantizeFloats()
[all …]
Doptimized_ops.h1046 temp_sum.val[0] = vminq_s32(vmaxq_s32(temp_sum.val[0], min_dup), max_dup); in MeanImpl()
1047 temp_sum.val[1] = vminq_s32(vmaxq_s32(temp_sum.val[1], min_dup), max_dup); in MeanImpl()
1048 temp_sum.val[2] = vminq_s32(vmaxq_s32(temp_sum.val[2], min_dup), max_dup); in MeanImpl()
1049 temp_sum.val[3] = vminq_s32(vmaxq_s32(temp_sum.val[3], min_dup), max_dup); in MeanImpl()
4041 vmaxq_s32(vminq_s32(temp_val.val[0], max_val_dup), min_val_dup); in SoftmaxInt8LUT()
4043 vmaxq_s32(vminq_s32(temp_val.val[1], max_val_dup), min_val_dup); in SoftmaxInt8LUT()
4045 vmaxq_s32(vminq_s32(temp_val.val[2], max_val_dup), min_val_dup); in SoftmaxInt8LUT()
4047 vmaxq_s32(vminq_s32(temp_val.val[3], max_val_dup), min_val_dup); in SoftmaxInt8LUT()
5813 vmaxq_s32(vminq_s32(temp_val.val[0], max_val_dup), min_val_dup); in Quantize()
5815 vmaxq_s32(vminq_s32(temp_val.val[1], max_val_dup), min_val_dup); in Quantize()
[all …]
Ddepthwiseconv_uint8.h1885 acc[j] = vmaxq_s32(acc[j], output_activation_min_vec);
1928 acc0 = vmaxq_s32(acc0, output_activation_min_vec);
1929 acc1 = vmaxq_s32(acc1, output_activation_min_vec);
1960 acc = vmaxq_s32(acc, output_activation_min_vec);
/external/llvm-project/clang/test/CodeGen/arm-mve-intrinsics/
Dvmaxq.c48 return vmaxq_s32(a, b); in test_vmaxq_s32()
/external/XNNPACK/src/math/
Dexp-neonfma-rr2-p5.c61 int32x4_t ven = vmaxq_s32(veo, vmin_exponent); in xnn_math_f32_exp__neonfma_rr2_p5()
Dexp-neonfma-rr2-lut64-p2.c63 int32x4_t ven = vmaxq_s32(veo, vmin_exponent); in xnn_math_f32_exp__neonfma_rr2_lut64_p2()
/external/libopus/silk/arm/
DLPC_inv_pred_gain_neon_intr.c128 max_s32x4 = vmaxq_s32( max_s32x4, s0_s32x4 ); in LPC_inverse_pred_gain_QA_neon()
130 max_s32x4 = vmaxq_s32( max_s32x4, s1_s32x4 ); in LPC_inverse_pred_gain_QA_neon()
DNSQ_del_dec_neon_intr.c697 tmp1_s32x4 = vmaxq_s32( tmp1_s32x4, vdupq_n_s32( -( 31 << 10 ) ) ); in silk_noise_shape_quantizer_del_dec_neon()
755 tmp2_s32x4 = vaddq_s32( tmp2_s32x4, vmaxq_s32( rd1_Q10_s32x4, rd2_Q10_s32x4 ) ); in silk_noise_shape_quantizer_del_dec_neon()
/external/gemmlowp/internal/
Dsimd_wrappers_neon.h102 inline Int32x4 Max(Int32x4 a, Int32x4 b) { return vmaxq_s32(a, b); }
105 return vmaxq_s32(a, vdupq_n_s32(b));
/external/skqp/src/core/
DSkBitmapProcState_matrixProcs.cpp325 res = vmaxq_s32(res, vdupq_n_s32(0)); in clamp4()
/external/tensorflow/tensorflow/lite/kernels/
Dcpu_backend_gemm_custom_gemv.h571 vmaxq_s32(multiplier_exponent, vdupq_n_s32(0));
/external/eigen/Eigen/src/Core/arch/NEON/
DPacketMath.h219 …ONG_INLINE Packet4i pmax<Packet4i>(const Packet4i& a, const Packet4i& b) { return vmaxq_s32(a,b); }
/external/skia/include/private/
DSkNx_neon.h541 AI static SkNx Max(const SkNx& a, const SkNx& b) { return vmaxq_s32(a.fVec, b.fVec); } in Max()
/external/skqp/include/private/
DSkNx_neon.h563 AI static SkNx Max(const SkNx& a, const SkNx& b) { return vmaxq_s32(a.fVec, b.fVec); } in Max()
/external/libaom/libaom/aom_dsp/simd/
Dv128_intrinsics_arm.h412 vmaxq_s32(vreinterpretq_s32_s64(x), vreinterpretq_s32_s64(y))); in v128_max_s32()
/external/libgav1/libgav1/src/dsp/arm/
Dinverse_transform_10bit_neon.cc181 *a = vmaxq_s32(vminq_s32(x, *max), *min); in HadamardRotation()
182 *b = vmaxq_s32(vminq_s32(y, *max), *min); in HadamardRotation()
/external/psimd/include/
Dpsimd.h839 return (psimd_s32) vmaxq_s32((int32x4_t) a, (int32x4_t) b); in psimd_max_s32()

12