| /external/XNNPACK/src/f32-vbinary/gen/ |
| D | vminc-sse-x8.c | 39 __m128 vy0123 = _mm_min_ps(va0123, vb); in xnn_f32_vminc_ukernel__sse_x8() local 52 __m128 vy0123 = _mm_min_ps(va0123, vb); in xnn_f32_vminc_ukernel__sse_x8() local 59 __m128 vy0123 = _mm_min_ps(va0123, vb); in xnn_f32_vminc_ukernel__sse_x8() local
|
| D | vmaxc-sse-x8.c | 39 __m128 vy0123 = _mm_max_ps(va0123, vb); in xnn_f32_vmaxc_ukernel__sse_x8() local 52 __m128 vy0123 = _mm_max_ps(va0123, vb); in xnn_f32_vmaxc_ukernel__sse_x8() local 59 __m128 vy0123 = _mm_max_ps(va0123, vb); in xnn_f32_vmaxc_ukernel__sse_x8() local
|
| D | vsqrdiffc-sse-x8.c | 39 __m128 vy0123 = _mm_sub_ps(va0123, vb); in xnn_f32_vsqrdiffc_ukernel__sse_x8() local 54 __m128 vy0123 = _mm_sub_ps(va0123, vb); in xnn_f32_vsqrdiffc_ukernel__sse_x8() local 62 __m128 vy0123 = _mm_sub_ps(va0123, vb); in xnn_f32_vsqrdiffc_ukernel__sse_x8() local
|
| D | vminc-neon-x8.c | 37 float32x4_t vy0123 = vminq_f32(va0123, vb); in xnn_f32_vminc_ukernel__neon_x8() local 48 float32x4_t vy0123 = vminq_f32(va0123, vb); in xnn_f32_vminc_ukernel__neon_x8() local 54 float32x4_t vy0123 = vminq_f32(va0123, vb); in xnn_f32_vminc_ukernel__neon_x8() local
|
| D | vsqrdiffc-neon-x8.c | 37 float32x4_t vy0123 = vsubq_f32(va0123, vb); in xnn_f32_vsqrdiffc_ukernel__neon_x8() local 50 float32x4_t vy0123 = vsubq_f32(va0123, vb); in xnn_f32_vsqrdiffc_ukernel__neon_x8() local 57 float32x4_t vy0123 = vsubq_f32(va0123, vb); in xnn_f32_vsqrdiffc_ukernel__neon_x8() local
|
| D | vmaxc-neon-x8.c | 37 float32x4_t vy0123 = vmaxq_f32(va0123, vb); in xnn_f32_vmaxc_ukernel__neon_x8() local 48 float32x4_t vy0123 = vmaxq_f32(va0123, vb); in xnn_f32_vmaxc_ukernel__neon_x8() local 54 float32x4_t vy0123 = vmaxq_f32(va0123, vb); in xnn_f32_vmaxc_ukernel__neon_x8() local
|
| D | vrsubc-minmax-sse-x8.c | 41 __m128 vy0123 = _mm_sub_ps(vb, va0123); in xnn_f32_vrsubc_minmax_ukernel__sse_x8() local 59 __m128 vy0123 = _mm_sub_ps(vb, va0123); in xnn_f32_vrsubc_minmax_ukernel__sse_x8() local 68 __m128 vy0123 = _mm_sub_ps(vb, va0123); in xnn_f32_vrsubc_minmax_ukernel__sse_x8() local
|
| D | vaddc-minmax-sse-x8.c | 41 __m128 vy0123 = _mm_add_ps(va0123, vb); in xnn_f32_vaddc_minmax_ukernel__sse_x8() local 59 __m128 vy0123 = _mm_add_ps(va0123, vb); in xnn_f32_vaddc_minmax_ukernel__sse_x8() local 68 __m128 vy0123 = _mm_add_ps(va0123, vb); in xnn_f32_vaddc_minmax_ukernel__sse_x8() local
|
| D | vsubc-minmax-sse-x8.c | 41 __m128 vy0123 = _mm_sub_ps(va0123, vb); in xnn_f32_vsubc_minmax_ukernel__sse_x8() local 59 __m128 vy0123 = _mm_sub_ps(va0123, vb); in xnn_f32_vsubc_minmax_ukernel__sse_x8() local 68 __m128 vy0123 = _mm_sub_ps(va0123, vb); in xnn_f32_vsubc_minmax_ukernel__sse_x8() local
|
| D | vmulc-minmax-sse-x8.c | 41 __m128 vy0123 = _mm_mul_ps(va0123, vb); in xnn_f32_vmulc_minmax_ukernel__sse_x8() local 59 __m128 vy0123 = _mm_mul_ps(va0123, vb); in xnn_f32_vmulc_minmax_ukernel__sse_x8() local 68 __m128 vy0123 = _mm_mul_ps(va0123, vb); in xnn_f32_vmulc_minmax_ukernel__sse_x8() local
|
| D | vrdivc-minmax-sse-x8.c | 41 __m128 vy0123 = _mm_div_ps(vb, va0123); in xnn_f32_vrdivc_minmax_ukernel__sse_x8() local 59 __m128 vy0123 = _mm_div_ps(vb, va0123); in xnn_f32_vrdivc_minmax_ukernel__sse_x8() local 68 __m128 vy0123 = _mm_div_ps(vb, va0123); in xnn_f32_vrdivc_minmax_ukernel__sse_x8() local
|
| D | vdivc-minmax-sse-x8.c | 41 __m128 vy0123 = _mm_div_ps(va0123, vb); in xnn_f32_vdivc_minmax_ukernel__sse_x8() local 59 __m128 vy0123 = _mm_div_ps(va0123, vb); in xnn_f32_vdivc_minmax_ukernel__sse_x8() local 68 __m128 vy0123 = _mm_div_ps(va0123, vb); in xnn_f32_vdivc_minmax_ukernel__sse_x8() local
|
| D | vsubc-minmax-neon-x8.c | 39 float32x4_t vy0123 = vsubq_f32(va0123, vb); in xnn_f32_vsubc_minmax_ukernel__neon_x8() local 55 float32x4_t vy0123 = vsubq_f32(va0123, vb); in xnn_f32_vsubc_minmax_ukernel__neon_x8() local 63 float32x4_t vy0123 = vsubq_f32(va0123, vb); in xnn_f32_vsubc_minmax_ukernel__neon_x8() local
|
| D | vmin-sse-x8.c | 42 __m128 vy0123 = _mm_min_ps(va0123, vb0123); in xnn_f32_vmin_ukernel__sse_x8() local 58 __m128 vy0123 = _mm_min_ps(va0123, vb0123); in xnn_f32_vmin_ukernel__sse_x8() local 66 __m128 vy0123 = _mm_min_ps(va0123, vb0123); in xnn_f32_vmin_ukernel__sse_x8() local
|
| D | vrsubc-minmax-neon-x8.c | 39 float32x4_t vy0123 = vsubq_f32(vb, va0123); in xnn_f32_vrsubc_minmax_ukernel__neon_x8() local 55 float32x4_t vy0123 = vsubq_f32(vb, va0123); in xnn_f32_vrsubc_minmax_ukernel__neon_x8() local 63 float32x4_t vy0123 = vsubq_f32(vb, va0123); in xnn_f32_vrsubc_minmax_ukernel__neon_x8() local
|
| D | vmax-sse-x8.c | 42 __m128 vy0123 = _mm_max_ps(va0123, vb0123); in xnn_f32_vmax_ukernel__sse_x8() local 58 __m128 vy0123 = _mm_max_ps(va0123, vb0123); in xnn_f32_vmax_ukernel__sse_x8() local 66 __m128 vy0123 = _mm_max_ps(va0123, vb0123); in xnn_f32_vmax_ukernel__sse_x8() local
|
| D | vrdivc-minmax-neon-x8.c | 39 float32x4_t vy0123 = vdivq_f32(vb, va0123); in xnn_f32_vrdivc_minmax_ukernel__neon_x8() local 55 float32x4_t vy0123 = vdivq_f32(vb, va0123); in xnn_f32_vrdivc_minmax_ukernel__neon_x8() local 63 float32x4_t vy0123 = vdivq_f32(vb, va0123); in xnn_f32_vrdivc_minmax_ukernel__neon_x8() local
|
| D | vsqrdiff-sse-x8.c | 42 __m128 vy0123 = _mm_sub_ps(va0123, vb0123); in xnn_f32_vsqrdiff_ukernel__sse_x8() local 60 __m128 vy0123 = _mm_sub_ps(va0123, vb0123); in xnn_f32_vsqrdiff_ukernel__sse_x8() local 69 __m128 vy0123 = _mm_sub_ps(va0123, vb0123); in xnn_f32_vsqrdiff_ukernel__sse_x8() local
|
| D | vdivc-minmax-neon-x8.c | 39 float32x4_t vy0123 = vdivq_f32(va0123, vb); in xnn_f32_vdivc_minmax_ukernel__neon_x8() local 55 float32x4_t vy0123 = vdivq_f32(va0123, vb); in xnn_f32_vdivc_minmax_ukernel__neon_x8() local 63 float32x4_t vy0123 = vdivq_f32(va0123, vb); in xnn_f32_vdivc_minmax_ukernel__neon_x8() local
|
| D | vmulc-minmax-neon-x8.c | 39 float32x4_t vy0123 = vmulq_f32(va0123, vb); in xnn_f32_vmulc_minmax_ukernel__neon_x8() local 55 float32x4_t vy0123 = vmulq_f32(va0123, vb); in xnn_f32_vmulc_minmax_ukernel__neon_x8() local 63 float32x4_t vy0123 = vmulq_f32(va0123, vb); in xnn_f32_vmulc_minmax_ukernel__neon_x8() local
|
| D | vaddc-minmax-neon-x8.c | 39 float32x4_t vy0123 = vaddq_f32(va0123, vb); in xnn_f32_vaddc_minmax_ukernel__neon_x8() local 55 float32x4_t vy0123 = vaddq_f32(va0123, vb); in xnn_f32_vaddc_minmax_ukernel__neon_x8() local 63 float32x4_t vy0123 = vaddq_f32(va0123, vb); in xnn_f32_vaddc_minmax_ukernel__neon_x8() local
|
| D | vmin-neon-x8.c | 38 float32x4_t vy0123 = vminq_f32(va0123, vb0123); in xnn_f32_vmin_ukernel__neon_x8() local 50 float32x4_t vy0123 = vminq_f32(va0123, vb0123); in xnn_f32_vmin_ukernel__neon_x8() local 57 float32x4_t vy0123 = vminq_f32(va0123, vb0123); in xnn_f32_vmin_ukernel__neon_x8() local
|
| D | vsqrdiff-neon-x8.c | 38 float32x4_t vy0123 = vsubq_f32(va0123, vb0123); in xnn_f32_vsqrdiff_ukernel__neon_x8() local 52 float32x4_t vy0123 = vsubq_f32(va0123, vb0123); in xnn_f32_vsqrdiff_ukernel__neon_x8() local 60 float32x4_t vy0123 = vsubq_f32(va0123, vb0123); in xnn_f32_vsqrdiff_ukernel__neon_x8() local
|
| D | vmax-neon-x8.c | 38 float32x4_t vy0123 = vmaxq_f32(va0123, vb0123); in xnn_f32_vmax_ukernel__neon_x8() local 50 float32x4_t vy0123 = vmaxq_f32(va0123, vb0123); in xnn_f32_vmax_ukernel__neon_x8() local 57 float32x4_t vy0123 = vmaxq_f32(va0123, vb0123); in xnn_f32_vmax_ukernel__neon_x8() local
|
| D | vadd-minmax-sse-x8.c | 44 __m128 vy0123 = _mm_add_ps(va0123, vb0123); in xnn_f32_vadd_minmax_ukernel__sse_x8() local 65 __m128 vy0123 = _mm_add_ps(va0123, vb0123); in xnn_f32_vadd_minmax_ukernel__sse_x8() local 75 __m128 vy0123 = _mm_add_ps(va0123, vb0123); in xnn_f32_vadd_minmax_ukernel__sse_x8() local
|