Lines Matching refs:va0
60 float32x4_t va0 = vld1q_f32(a0); a0 += 4; in xnn_f32_igemm_minmax_ukernel_1x8s4__neonfma() local
66 vacc0x0123 = vfmaq_f32(vacc0x0123, va0, vb0123c0); in xnn_f32_igemm_minmax_ukernel_1x8s4__neonfma()
67 vacc0x4567 = vfmaq_f32(vacc0x4567, va0, vb4567c0); in xnn_f32_igemm_minmax_ukernel_1x8s4__neonfma()
69 va0 = vextq_f32(va0, va0, 1); in xnn_f32_igemm_minmax_ukernel_1x8s4__neonfma()
74 vacc0x0123 = vfmaq_f32(vacc0x0123, va0, vb0123c1); in xnn_f32_igemm_minmax_ukernel_1x8s4__neonfma()
75 vacc0x4567 = vfmaq_f32(vacc0x4567, va0, vb4567c1); in xnn_f32_igemm_minmax_ukernel_1x8s4__neonfma()
77 va0 = vextq_f32(va0, va0, 1); in xnn_f32_igemm_minmax_ukernel_1x8s4__neonfma()
82 vacc0x0123 = vfmaq_f32(vacc0x0123, va0, vb0123c2); in xnn_f32_igemm_minmax_ukernel_1x8s4__neonfma()
83 vacc0x4567 = vfmaq_f32(vacc0x4567, va0, vb4567c2); in xnn_f32_igemm_minmax_ukernel_1x8s4__neonfma()
85 va0 = vextq_f32(va0, va0, 1); in xnn_f32_igemm_minmax_ukernel_1x8s4__neonfma()
90 vacc0x0123 = vfmaq_f32(vacc0x0123, va0, vb0123c3); in xnn_f32_igemm_minmax_ukernel_1x8s4__neonfma()
91 vacc0x4567 = vfmaq_f32(vacc0x4567, va0, vb4567c3); in xnn_f32_igemm_minmax_ukernel_1x8s4__neonfma()
99 const float32x4_t va0 = vld1q_dup_f32(a0); a0 += 1; in xnn_f32_igemm_minmax_ukernel_1x8s4__neonfma() local
104 vacc0x0123 = vfmaq_f32(vacc0x0123, va0, vb0123); in xnn_f32_igemm_minmax_ukernel_1x8s4__neonfma()
105 vacc0x4567 = vfmaq_f32(vacc0x4567, va0, vb4567); in xnn_f32_igemm_minmax_ukernel_1x8s4__neonfma()