Lines Matching refs:va0
60 float32x4_t va0 = vld1q_f32(a0); a0 += 4; in xnn_f32_igemm_minmax_ukernel_1x8s4__neon() local
66 vacc0x0123 = vmlaq_f32(vacc0x0123, va0, vb0123c0); in xnn_f32_igemm_minmax_ukernel_1x8s4__neon()
67 vacc0x4567 = vmlaq_f32(vacc0x4567, va0, vb4567c0); in xnn_f32_igemm_minmax_ukernel_1x8s4__neon()
69 va0 = vextq_f32(va0, va0, 1); in xnn_f32_igemm_minmax_ukernel_1x8s4__neon()
74 vacc0x0123 = vmlaq_f32(vacc0x0123, va0, vb0123c1); in xnn_f32_igemm_minmax_ukernel_1x8s4__neon()
75 vacc0x4567 = vmlaq_f32(vacc0x4567, va0, vb4567c1); in xnn_f32_igemm_minmax_ukernel_1x8s4__neon()
77 va0 = vextq_f32(va0, va0, 1); in xnn_f32_igemm_minmax_ukernel_1x8s4__neon()
82 vacc0x0123 = vmlaq_f32(vacc0x0123, va0, vb0123c2); in xnn_f32_igemm_minmax_ukernel_1x8s4__neon()
83 vacc0x4567 = vmlaq_f32(vacc0x4567, va0, vb4567c2); in xnn_f32_igemm_minmax_ukernel_1x8s4__neon()
85 va0 = vextq_f32(va0, va0, 1); in xnn_f32_igemm_minmax_ukernel_1x8s4__neon()
90 vacc0x0123 = vmlaq_f32(vacc0x0123, va0, vb0123c3); in xnn_f32_igemm_minmax_ukernel_1x8s4__neon()
91 vacc0x4567 = vmlaq_f32(vacc0x4567, va0, vb4567c3); in xnn_f32_igemm_minmax_ukernel_1x8s4__neon()
99 const float32x4_t va0 = vld1q_dup_f32(a0); a0 += 1; in xnn_f32_igemm_minmax_ukernel_1x8s4__neon() local
104 vacc0x0123 = vmlaq_f32(vacc0x0123, va0, vb0123); in xnn_f32_igemm_minmax_ukernel_1x8s4__neon()
105 vacc0x4567 = vmlaq_f32(vacc0x4567, va0, vb4567); in xnn_f32_igemm_minmax_ukernel_1x8s4__neon()