Lines Matching refs:va0
48 float32x4_t va0 = vld1q_f32(a0); a0 += 4; in xnn_f32_gemm_minmax_ukernel_1x8s4__neonfma() local
54 vacc0x0123 = vfmaq_f32(vacc0x0123, va0, vb0123c0); in xnn_f32_gemm_minmax_ukernel_1x8s4__neonfma()
55 vacc0x4567 = vfmaq_f32(vacc0x4567, va0, vb4567c0); in xnn_f32_gemm_minmax_ukernel_1x8s4__neonfma()
57 va0 = vextq_f32(va0, va0, 1); in xnn_f32_gemm_minmax_ukernel_1x8s4__neonfma()
62 vacc0x0123 = vfmaq_f32(vacc0x0123, va0, vb0123c1); in xnn_f32_gemm_minmax_ukernel_1x8s4__neonfma()
63 vacc0x4567 = vfmaq_f32(vacc0x4567, va0, vb4567c1); in xnn_f32_gemm_minmax_ukernel_1x8s4__neonfma()
65 va0 = vextq_f32(va0, va0, 1); in xnn_f32_gemm_minmax_ukernel_1x8s4__neonfma()
70 vacc0x0123 = vfmaq_f32(vacc0x0123, va0, vb0123c2); in xnn_f32_gemm_minmax_ukernel_1x8s4__neonfma()
71 vacc0x4567 = vfmaq_f32(vacc0x4567, va0, vb4567c2); in xnn_f32_gemm_minmax_ukernel_1x8s4__neonfma()
73 va0 = vextq_f32(va0, va0, 1); in xnn_f32_gemm_minmax_ukernel_1x8s4__neonfma()
78 vacc0x0123 = vfmaq_f32(vacc0x0123, va0, vb0123c3); in xnn_f32_gemm_minmax_ukernel_1x8s4__neonfma()
79 vacc0x4567 = vfmaq_f32(vacc0x4567, va0, vb4567c3); in xnn_f32_gemm_minmax_ukernel_1x8s4__neonfma()
87 const float32x4_t va0 = vld1q_dup_f32(a0); a0 += 1; in xnn_f32_gemm_minmax_ukernel_1x8s4__neonfma() local
92 vacc0x0123 = vfmaq_f32(vacc0x0123, va0, vb0123); in xnn_f32_gemm_minmax_ukernel_1x8s4__neonfma()
93 vacc0x4567 = vfmaq_f32(vacc0x4567, va0, vb4567); in xnn_f32_gemm_minmax_ukernel_1x8s4__neonfma()