Lines Matching refs:vacc0p1
182 float vacc0p1 = vi1x0 * vk1x0; in xnn_f32_dwconv_minmax_ukernel_up2x25__scalar_acc2() local
200 vacc0p1 = math_muladd_f32(vi3x0, vk3x0, vacc0p1); in xnn_f32_dwconv_minmax_ukernel_up2x25__scalar_acc2()
218 vacc0p1 = math_muladd_f32(vi5x0, vk5x0, vacc0p1); in xnn_f32_dwconv_minmax_ukernel_up2x25__scalar_acc2()
236 vacc0p1 = math_muladd_f32(vi7x0, vk7x0, vacc0p1); in xnn_f32_dwconv_minmax_ukernel_up2x25__scalar_acc2()
254 vacc0p1 = math_muladd_f32(vi9x0, vk9x0, vacc0p1); in xnn_f32_dwconv_minmax_ukernel_up2x25__scalar_acc2()
272 vacc0p1 = math_muladd_f32(vi11x0, vk11x0, vacc0p1); in xnn_f32_dwconv_minmax_ukernel_up2x25__scalar_acc2()
290 vacc0p1 = math_muladd_f32(vi13x0, vk13x0, vacc0p1); in xnn_f32_dwconv_minmax_ukernel_up2x25__scalar_acc2()
308 vacc0p1 = math_muladd_f32(vi15x0, vk15x0, vacc0p1); in xnn_f32_dwconv_minmax_ukernel_up2x25__scalar_acc2()
326 vacc0p1 = math_muladd_f32(vi17x0, vk17x0, vacc0p1); in xnn_f32_dwconv_minmax_ukernel_up2x25__scalar_acc2()
344 vacc0p1 = math_muladd_f32(vi19x0, vk19x0, vacc0p1); in xnn_f32_dwconv_minmax_ukernel_up2x25__scalar_acc2()
362 vacc0p1 = math_muladd_f32(vi21x0, vk21x0, vacc0p1); in xnn_f32_dwconv_minmax_ukernel_up2x25__scalar_acc2()
380 vacc0p1 = math_muladd_f32(vi23x0, vk23x0, vacc0p1); in xnn_f32_dwconv_minmax_ukernel_up2x25__scalar_acc2()
396 vacc0p0 = vacc0p0 + vacc0p1; in xnn_f32_dwconv_minmax_ukernel_up2x25__scalar_acc2()
417 float vacc0p1 = vi1 * vk1; in xnn_f32_dwconv_minmax_ukernel_up2x25__scalar_acc2() local
423 vacc0p1 = math_muladd_f32(vi3, vk3, vacc0p1); in xnn_f32_dwconv_minmax_ukernel_up2x25__scalar_acc2()
429 vacc0p1 = math_muladd_f32(vi5, vk5, vacc0p1); in xnn_f32_dwconv_minmax_ukernel_up2x25__scalar_acc2()
435 vacc0p1 = math_muladd_f32(vi7, vk7, vacc0p1); in xnn_f32_dwconv_minmax_ukernel_up2x25__scalar_acc2()
441 vacc0p1 = math_muladd_f32(vi9, vk9, vacc0p1); in xnn_f32_dwconv_minmax_ukernel_up2x25__scalar_acc2()
447 vacc0p1 = math_muladd_f32(vi11, vk11, vacc0p1); in xnn_f32_dwconv_minmax_ukernel_up2x25__scalar_acc2()
453 vacc0p1 = math_muladd_f32(vi13, vk13, vacc0p1); in xnn_f32_dwconv_minmax_ukernel_up2x25__scalar_acc2()
459 vacc0p1 = math_muladd_f32(vi15, vk15, vacc0p1); in xnn_f32_dwconv_minmax_ukernel_up2x25__scalar_acc2()
465 vacc0p1 = math_muladd_f32(vi17, vk17, vacc0p1); in xnn_f32_dwconv_minmax_ukernel_up2x25__scalar_acc2()
471 vacc0p1 = math_muladd_f32(vi19, vk19, vacc0p1); in xnn_f32_dwconv_minmax_ukernel_up2x25__scalar_acc2()
477 vacc0p1 = math_muladd_f32(vi21, vk21, vacc0p1); in xnn_f32_dwconv_minmax_ukernel_up2x25__scalar_acc2()
483 vacc0p1 = math_muladd_f32(vi23, vk23, vacc0p1); in xnn_f32_dwconv_minmax_ukernel_up2x25__scalar_acc2()
489 vacc0p0 = vacc0p0 + vacc0p1; in xnn_f32_dwconv_minmax_ukernel_up2x25__scalar_acc2()