Lines Matching refs:vb
58 const float32x4_t vb = vld1q_f32(w); w += 4; in xnn_f32_spmm_ukernel_12x4__neonfma() local
60 vacc0123c0 = vfmaq_laneq_f32(vacc0123c0, va0123, vb, 0); in xnn_f32_spmm_ukernel_12x4__neonfma()
61 vacc4567c0 = vfmaq_laneq_f32(vacc4567c0, va4567, vb, 0); in xnn_f32_spmm_ukernel_12x4__neonfma()
62 vacc89ABc0 = vfmaq_laneq_f32(vacc89ABc0, va89AB, vb, 0); in xnn_f32_spmm_ukernel_12x4__neonfma()
63 vacc0123c1 = vfmaq_laneq_f32(vacc0123c1, va0123, vb, 1); in xnn_f32_spmm_ukernel_12x4__neonfma()
64 vacc4567c1 = vfmaq_laneq_f32(vacc4567c1, va4567, vb, 1); in xnn_f32_spmm_ukernel_12x4__neonfma()
65 vacc89ABc1 = vfmaq_laneq_f32(vacc89ABc1, va89AB, vb, 1); in xnn_f32_spmm_ukernel_12x4__neonfma()
66 vacc0123c2 = vfmaq_laneq_f32(vacc0123c2, va0123, vb, 2); in xnn_f32_spmm_ukernel_12x4__neonfma()
67 vacc4567c2 = vfmaq_laneq_f32(vacc4567c2, va4567, vb, 2); in xnn_f32_spmm_ukernel_12x4__neonfma()
68 vacc89ABc2 = vfmaq_laneq_f32(vacc89ABc2, va89AB, vb, 2); in xnn_f32_spmm_ukernel_12x4__neonfma()
69 vacc0123c3 = vfmaq_laneq_f32(vacc0123c3, va0123, vb, 3); in xnn_f32_spmm_ukernel_12x4__neonfma()
70 vacc4567c3 = vfmaq_laneq_f32(vacc4567c3, va4567, vb, 3); in xnn_f32_spmm_ukernel_12x4__neonfma()
71 vacc89ABc3 = vfmaq_laneq_f32(vacc89ABc3, va89AB, vb, 3); in xnn_f32_spmm_ukernel_12x4__neonfma()
130 const float32x4_t vb = vld1q_dup_f32(w); w += 1; in xnn_f32_spmm_ukernel_12x4__neonfma() local
131 vacc0123 = vfmaq_f32(vacc0123, va0123, vb); in xnn_f32_spmm_ukernel_12x4__neonfma()
132 vacc4567 = vfmaq_f32(vacc4567, va4567, vb); in xnn_f32_spmm_ukernel_12x4__neonfma()
133 vacc89AB = vfmaq_f32(vacc89AB, va89AB, vb); in xnn_f32_spmm_ukernel_12x4__neonfma()
178 const float32x4_t vb = vld1q_f32(w); w += 4; in xnn_f32_spmm_ukernel_12x4__neonfma() local
180 vacc0123c0 = vfmaq_laneq_f32(vacc0123c0, va0123, vb, 0); in xnn_f32_spmm_ukernel_12x4__neonfma()
181 vacc4567c0 = vfmaq_laneq_f32(vacc4567c0, va4567, vb, 0); in xnn_f32_spmm_ukernel_12x4__neonfma()
182 vacc0123c1 = vfmaq_laneq_f32(vacc0123c1, va0123, vb, 1); in xnn_f32_spmm_ukernel_12x4__neonfma()
183 vacc4567c1 = vfmaq_laneq_f32(vacc4567c1, va4567, vb, 1); in xnn_f32_spmm_ukernel_12x4__neonfma()
184 vacc0123c2 = vfmaq_laneq_f32(vacc0123c2, va0123, vb, 2); in xnn_f32_spmm_ukernel_12x4__neonfma()
185 vacc4567c2 = vfmaq_laneq_f32(vacc4567c2, va4567, vb, 2); in xnn_f32_spmm_ukernel_12x4__neonfma()
186 vacc0123c3 = vfmaq_laneq_f32(vacc0123c3, va0123, vb, 3); in xnn_f32_spmm_ukernel_12x4__neonfma()
187 vacc4567c3 = vfmaq_laneq_f32(vacc4567c3, va4567, vb, 3); in xnn_f32_spmm_ukernel_12x4__neonfma()
232 const float32x4_t vb = vld1q_dup_f32(w); w += 1; in xnn_f32_spmm_ukernel_12x4__neonfma() local
233 vacc0123 = vfmaq_f32(vacc0123, va0123, vb); in xnn_f32_spmm_ukernel_12x4__neonfma()
234 vacc4567 = vfmaq_f32(vacc4567, va4567, vb); in xnn_f32_spmm_ukernel_12x4__neonfma()
269 const float32x4_t vb = vld1q_f32(w); w += 4; in xnn_f32_spmm_ukernel_12x4__neonfma() local
271 vacc0123c0 = vfmaq_laneq_f32(vacc0123c0, va0123, vb, 0); in xnn_f32_spmm_ukernel_12x4__neonfma()
272 vacc0123c1 = vfmaq_laneq_f32(vacc0123c1, va0123, vb, 1); in xnn_f32_spmm_ukernel_12x4__neonfma()
273 vacc0123c2 = vfmaq_laneq_f32(vacc0123c2, va0123, vb, 2); in xnn_f32_spmm_ukernel_12x4__neonfma()
274 vacc0123c3 = vfmaq_laneq_f32(vacc0123c3, va0123, vb, 3); in xnn_f32_spmm_ukernel_12x4__neonfma()
305 const float32x4_t vb = vld1q_dup_f32(w); w += 1; in xnn_f32_spmm_ukernel_12x4__neonfma() local
306 vacc0123 = vfmaq_f32(vacc0123, va0123, vb); in xnn_f32_spmm_ukernel_12x4__neonfma()
338 const float32x4_t vb = vld1q_f32(w); w += 4; in xnn_f32_spmm_ukernel_12x4__neonfma() local
340 vacc01c0 = vfma_laneq_f32(vacc01c0, va01, vb, 0); in xnn_f32_spmm_ukernel_12x4__neonfma()
341 vacc01c1 = vfma_laneq_f32(vacc01c1, va01, vb, 1); in xnn_f32_spmm_ukernel_12x4__neonfma()
342 vacc01c2 = vfma_laneq_f32(vacc01c2, va01, vb, 2); in xnn_f32_spmm_ukernel_12x4__neonfma()
343 vacc01c3 = vfma_laneq_f32(vacc01c3, va01, vb, 3); in xnn_f32_spmm_ukernel_12x4__neonfma()
374 const float32x2_t vb = vld1_dup_f32(w); w += 1; in xnn_f32_spmm_ukernel_12x4__neonfma() local
375 vacc01 = vfma_f32(vacc01, va01, vb); in xnn_f32_spmm_ukernel_12x4__neonfma()
406 const float32x4_t vb = vld1q_f32(w); w += 4; in xnn_f32_spmm_ukernel_12x4__neonfma() local
408 vacc0c0 = vfma_laneq_f32(vacc0c0, va0, vb, 0); in xnn_f32_spmm_ukernel_12x4__neonfma()
409 vacc0c1 = vfma_laneq_f32(vacc0c1, va0, vb, 1); in xnn_f32_spmm_ukernel_12x4__neonfma()
410 vacc0c2 = vfma_laneq_f32(vacc0c2, va0, vb, 2); in xnn_f32_spmm_ukernel_12x4__neonfma()
411 vacc0c3 = vfma_laneq_f32(vacc0c3, va0, vb, 3); in xnn_f32_spmm_ukernel_12x4__neonfma()
442 const float32x2_t vb = vld1_dup_f32(w); w += 1; in xnn_f32_spmm_ukernel_12x4__neonfma() local
443 vacc0 = vfma_f32(vacc0, va0, vb); in xnn_f32_spmm_ukernel_12x4__neonfma()