Searched refs:vacc0n3 (Results 1 – 5 of 5) sorted by relevance
/external/XNNPACK/src/f32-spmm/gen/ |
D | 4x4-minmax-neonfma.c | 193 float32x2_t vacc0n3 = vld1_dup_f32(w); w += 1; in xnn_f32_spmm_minmax_ukernel_4x4__neonfma() local 204 vacc0n3 = vfma_laneq_f32(vacc0n3, vi0, vw, 3); in xnn_f32_spmm_minmax_ukernel_4x4__neonfma() 210 float32x2_t vout0n3 = vmin_f32(vacc0n3, vget_low_f32(vmax)); in xnn_f32_spmm_minmax_ukernel_4x4__neonfma()
|
D | 8x4-minmax-neonfma.c | 292 float32x2_t vacc0n3 = vld1_dup_f32(w); w += 1; in xnn_f32_spmm_minmax_ukernel_8x4__neonfma() local 303 vacc0n3 = vfma_laneq_f32(vacc0n3, vi0, vw, 3); in xnn_f32_spmm_minmax_ukernel_8x4__neonfma() 309 float32x2_t vout0n3 = vmin_f32(vacc0n3, vget_low_f32(vmax)); in xnn_f32_spmm_minmax_ukernel_8x4__neonfma()
|
D | 12x4-minmax-neonfma.c | 418 float32x2_t vacc0n3 = vld1_dup_f32(w); w += 1; in xnn_f32_spmm_minmax_ukernel_12x4__neonfma() local 429 vacc0n3 = vfma_laneq_f32(vacc0n3, vi0, vw, 3); in xnn_f32_spmm_minmax_ukernel_12x4__neonfma() 435 float32x2_t vout0n3 = vmin_f32(vacc0n3, vget_low_f32(vmax)); in xnn_f32_spmm_minmax_ukernel_12x4__neonfma()
|
D | 16x4-minmax-neonfma.c | 445 float32x2_t vacc0n3 = vld1_dup_f32(w); w += 1; in xnn_f32_spmm_minmax_ukernel_16x4__neonfma() local 456 vacc0n3 = vfma_laneq_f32(vacc0n3, vi0, vw, 3); in xnn_f32_spmm_minmax_ukernel_16x4__neonfma() 462 float32x2_t vout0n3 = vmin_f32(vacc0n3, vget_low_f32(vmax)); in xnn_f32_spmm_minmax_ukernel_16x4__neonfma()
|
D | 32x4-minmax-neonfma.c | 708 float32x2_t vacc0n3 = vld1_dup_f32(w); w += 1; in xnn_f32_spmm_minmax_ukernel_32x4__neonfma() local 719 vacc0n3 = vfma_laneq_f32(vacc0n3, vi0, vw, 3); in xnn_f32_spmm_minmax_ukernel_32x4__neonfma() 725 float32x2_t vout0n3 = vmin_f32(vacc0n3, vget_low_f32(vmax)); in xnn_f32_spmm_minmax_ukernel_32x4__neonfma()
|