| /external/XNNPACK/src/f16-vbinary/gen/ |
| D | vmaxc-neonfp16arith-x16.c | 41 float16x8_t vy01234567 = vmaxq_f16(va01234567, vb); in xnn_f16_vmaxc_ukernel__neonfp16arith_x16() local 52 float16x8_t vy01234567 = vmaxq_f16(va01234567, vb); in xnn_f16_vmaxc_ukernel__neonfp16arith_x16() local 58 float16x8_t vy01234567 = vmaxq_f16(va01234567, vb); in xnn_f16_vmaxc_ukernel__neonfp16arith_x16() local
|
| D | vsqrdiffc-neonfp16arith-x16.c | 41 float16x8_t vy01234567 = vsubq_f16(va01234567, vb); in xnn_f16_vsqrdiffc_ukernel__neonfp16arith_x16() local 54 float16x8_t vy01234567 = vsubq_f16(va01234567, vb); in xnn_f16_vsqrdiffc_ukernel__neonfp16arith_x16() local 61 float16x8_t vy01234567 = vsubq_f16(va01234567, vb); in xnn_f16_vsqrdiffc_ukernel__neonfp16arith_x16() local
|
| D | vminc-neonfp16arith-x16.c | 41 float16x8_t vy01234567 = vminq_f16(va01234567, vb); in xnn_f16_vminc_ukernel__neonfp16arith_x16() local 52 float16x8_t vy01234567 = vminq_f16(va01234567, vb); in xnn_f16_vminc_ukernel__neonfp16arith_x16() local 58 float16x8_t vy01234567 = vminq_f16(va01234567, vb); in xnn_f16_vminc_ukernel__neonfp16arith_x16() local
|
| D | vdivc-minmax-neonfp16arith-x16.c | 43 float16x8_t vy01234567 = vdivq_f16(va01234567, vb); in xnn_f16_vdivc_minmax_ukernel__neonfp16arith_x16() local 59 float16x8_t vy01234567 = vdivq_f16(va01234567, vb); in xnn_f16_vdivc_minmax_ukernel__neonfp16arith_x16() local 67 float16x8_t vy01234567 = vdivq_f16(va01234567, vb); in xnn_f16_vdivc_minmax_ukernel__neonfp16arith_x16() local
|
| D | vrsubc-minmax-neonfp16arith-x16.c | 43 float16x8_t vy01234567 = vsubq_f16(vb, va01234567); in xnn_f16_vrsubc_minmax_ukernel__neonfp16arith_x16() local 59 float16x8_t vy01234567 = vsubq_f16(vb, va01234567); in xnn_f16_vrsubc_minmax_ukernel__neonfp16arith_x16() local 67 float16x8_t vy01234567 = vsubq_f16(vb, va01234567); in xnn_f16_vrsubc_minmax_ukernel__neonfp16arith_x16() local
|
| D | vsubc-minmax-neonfp16arith-x16.c | 43 float16x8_t vy01234567 = vsubq_f16(va01234567, vb); in xnn_f16_vsubc_minmax_ukernel__neonfp16arith_x16() local 59 float16x8_t vy01234567 = vsubq_f16(va01234567, vb); in xnn_f16_vsubc_minmax_ukernel__neonfp16arith_x16() local 67 float16x8_t vy01234567 = vsubq_f16(va01234567, vb); in xnn_f16_vsubc_minmax_ukernel__neonfp16arith_x16() local
|
| D | vmulc-minmax-neonfp16arith-x16.c | 43 float16x8_t vy01234567 = vmulq_f16(va01234567, vb); in xnn_f16_vmulc_minmax_ukernel__neonfp16arith_x16() local 59 float16x8_t vy01234567 = vmulq_f16(va01234567, vb); in xnn_f16_vmulc_minmax_ukernel__neonfp16arith_x16() local 67 float16x8_t vy01234567 = vmulq_f16(va01234567, vb); in xnn_f16_vmulc_minmax_ukernel__neonfp16arith_x16() local
|
| D | vaddc-minmax-neonfp16arith-x16.c | 43 float16x8_t vy01234567 = vaddq_f16(va01234567, vb); in xnn_f16_vaddc_minmax_ukernel__neonfp16arith_x16() local 59 float16x8_t vy01234567 = vaddq_f16(va01234567, vb); in xnn_f16_vaddc_minmax_ukernel__neonfp16arith_x16() local 67 float16x8_t vy01234567 = vaddq_f16(va01234567, vb); in xnn_f16_vaddc_minmax_ukernel__neonfp16arith_x16() local
|
| D | vrdivc-minmax-neonfp16arith-x16.c | 43 float16x8_t vy01234567 = vdivq_f16(vb, va01234567); in xnn_f16_vrdivc_minmax_ukernel__neonfp16arith_x16() local 59 float16x8_t vy01234567 = vdivq_f16(vb, va01234567); in xnn_f16_vrdivc_minmax_ukernel__neonfp16arith_x16() local 67 float16x8_t vy01234567 = vdivq_f16(vb, va01234567); in xnn_f16_vrdivc_minmax_ukernel__neonfp16arith_x16() local
|
| D | vmin-neonfp16arith-x16.c | 42 float16x8_t vy01234567 = vminq_f16(va01234567, vb01234567); in xnn_f16_vmin_ukernel__neonfp16arith_x16() local 54 float16x8_t vy01234567 = vminq_f16(va01234567, vb01234567); in xnn_f16_vmin_ukernel__neonfp16arith_x16() local 61 float16x8_t vy01234567 = vminq_f16(va01234567, vb01234567); in xnn_f16_vmin_ukernel__neonfp16arith_x16() local
|
| D | vsqrdiff-neonfp16arith-x16.c | 42 float16x8_t vy01234567 = vsubq_f16(va01234567, vb01234567); in xnn_f16_vsqrdiff_ukernel__neonfp16arith_x16() local 56 float16x8_t vy01234567 = vsubq_f16(va01234567, vb01234567); in xnn_f16_vsqrdiff_ukernel__neonfp16arith_x16() local 64 float16x8_t vy01234567 = vsubq_f16(va01234567, vb01234567); in xnn_f16_vsqrdiff_ukernel__neonfp16arith_x16() local
|
| D | vmax-neonfp16arith-x16.c | 42 float16x8_t vy01234567 = vmaxq_f16(va01234567, vb01234567); in xnn_f16_vmax_ukernel__neonfp16arith_x16() local 54 float16x8_t vy01234567 = vmaxq_f16(va01234567, vb01234567); in xnn_f16_vmax_ukernel__neonfp16arith_x16() local 61 float16x8_t vy01234567 = vmaxq_f16(va01234567, vb01234567); in xnn_f16_vmax_ukernel__neonfp16arith_x16() local
|
| D | vadd-minmax-neonfp16arith-x16.c | 44 float16x8_t vy01234567 = vaddq_f16(va01234567, vb01234567); in xnn_f16_vadd_minmax_ukernel__neonfp16arith_x16() local 61 float16x8_t vy01234567 = vaddq_f16(va01234567, vb01234567); in xnn_f16_vadd_minmax_ukernel__neonfp16arith_x16() local 70 float16x8_t vy01234567 = vaddq_f16(va01234567, vb01234567); in xnn_f16_vadd_minmax_ukernel__neonfp16arith_x16() local
|
| D | vmul-minmax-neonfp16arith-x16.c | 44 float16x8_t vy01234567 = vmulq_f16(va01234567, vb01234567); in xnn_f16_vmul_minmax_ukernel__neonfp16arith_x16() local 61 float16x8_t vy01234567 = vmulq_f16(va01234567, vb01234567); in xnn_f16_vmul_minmax_ukernel__neonfp16arith_x16() local 70 float16x8_t vy01234567 = vmulq_f16(va01234567, vb01234567); in xnn_f16_vmul_minmax_ukernel__neonfp16arith_x16() local
|
| D | vsub-minmax-neonfp16arith-x16.c | 44 float16x8_t vy01234567 = vsubq_f16(va01234567, vb01234567); in xnn_f16_vsub_minmax_ukernel__neonfp16arith_x16() local 61 float16x8_t vy01234567 = vsubq_f16(va01234567, vb01234567); in xnn_f16_vsub_minmax_ukernel__neonfp16arith_x16() local 70 float16x8_t vy01234567 = vsubq_f16(va01234567, vb01234567); in xnn_f16_vsub_minmax_ukernel__neonfp16arith_x16() local
|
| D | vdiv-minmax-neonfp16arith-x16.c | 44 float16x8_t vy01234567 = vdivq_f16(va01234567, vb01234567); in xnn_f16_vdiv_minmax_ukernel__neonfp16arith_x16() local 61 float16x8_t vy01234567 = vdivq_f16(va01234567, vb01234567); in xnn_f16_vdiv_minmax_ukernel__neonfp16arith_x16() local 70 float16x8_t vy01234567 = vdivq_f16(va01234567, vb01234567); in xnn_f16_vdiv_minmax_ukernel__neonfp16arith_x16() local
|
| D | vminc-neonfp16arith-x8.c | 40 float16x8_t vy01234567 = vminq_f16(va01234567, vb); in xnn_f16_vminc_ukernel__neonfp16arith_x8() local 46 float16x8_t vy01234567 = vminq_f16(va01234567, vb); in xnn_f16_vminc_ukernel__neonfp16arith_x8() local
|
| D | vmaxc-neonfp16arith-x8.c | 40 float16x8_t vy01234567 = vmaxq_f16(va01234567, vb); in xnn_f16_vmaxc_ukernel__neonfp16arith_x8() local 46 float16x8_t vy01234567 = vmaxq_f16(va01234567, vb); in xnn_f16_vmaxc_ukernel__neonfp16arith_x8() local
|
| D | vsqrdiffc-neonfp16arith-x8.c | 40 float16x8_t vy01234567 = vsubq_f16(va01234567, vb); in xnn_f16_vsqrdiffc_ukernel__neonfp16arith_x8() local 47 float16x8_t vy01234567 = vsubq_f16(va01234567, vb); in xnn_f16_vsqrdiffc_ukernel__neonfp16arith_x8() local
|
| D | vsqrdiff-neonfp16arith-x8.c | 40 float16x8_t vy01234567 = vsubq_f16(va01234567, vb01234567); in xnn_f16_vsqrdiff_ukernel__neonfp16arith_x8() local 48 float16x8_t vy01234567 = vsubq_f16(va01234567, vb01234567); in xnn_f16_vsqrdiff_ukernel__neonfp16arith_x8() local
|
| D | vmax-neonfp16arith-x8.c | 40 float16x8_t vy01234567 = vmaxq_f16(va01234567, vb01234567); in xnn_f16_vmax_ukernel__neonfp16arith_x8() local 47 float16x8_t vy01234567 = vmaxq_f16(va01234567, vb01234567); in xnn_f16_vmax_ukernel__neonfp16arith_x8() local
|
| D | vmin-neonfp16arith-x8.c | 40 float16x8_t vy01234567 = vminq_f16(va01234567, vb01234567); in xnn_f16_vmin_ukernel__neonfp16arith_x8() local 47 float16x8_t vy01234567 = vminq_f16(va01234567, vb01234567); in xnn_f16_vmin_ukernel__neonfp16arith_x8() local
|
| D | vmulc-minmax-neonfp16arith-x8.c | 42 float16x8_t vy01234567 = vmulq_f16(va01234567, vb); in xnn_f16_vmulc_minmax_ukernel__neonfp16arith_x8() local 50 float16x8_t vy01234567 = vmulq_f16(va01234567, vb); in xnn_f16_vmulc_minmax_ukernel__neonfp16arith_x8() local
|
| D | vsubc-minmax-neonfp16arith-x8.c | 42 float16x8_t vy01234567 = vsubq_f16(va01234567, vb); in xnn_f16_vsubc_minmax_ukernel__neonfp16arith_x8() local 50 float16x8_t vy01234567 = vsubq_f16(va01234567, vb); in xnn_f16_vsubc_minmax_ukernel__neonfp16arith_x8() local
|
| D | vrsubc-minmax-neonfp16arith-x8.c | 42 float16x8_t vy01234567 = vsubq_f16(vb, va01234567); in xnn_f16_vrsubc_minmax_ukernel__neonfp16arith_x8() local 50 float16x8_t vy01234567 = vsubq_f16(vb, va01234567); in xnn_f16_vrsubc_minmax_ukernel__neonfp16arith_x8() local
|