Lines Matching refs:vacc0123p0
166 float32x4_t vacc0123p0 = vld1q_f32(w); w += 4; in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma() local
171 vacc0123p0 = vfmaq_f32(vacc0123p0, vi0x0123, vk0x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
175 vacc0123p0 = vfmaq_f32(vacc0123p0, vi1x0123, vk1x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
179 vacc0123p0 = vfmaq_f32(vacc0123p0, vi2x0123, vk2x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
183 vacc0123p0 = vfmaq_f32(vacc0123p0, vi3x0123, vk3x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
187 vacc0123p0 = vfmaq_f32(vacc0123p0, vi4x0123, vk4x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
191 vacc0123p0 = vfmaq_f32(vacc0123p0, vi5x0123, vk5x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
195 vacc0123p0 = vfmaq_f32(vacc0123p0, vi6x0123, vk6x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
199 vacc0123p0 = vfmaq_f32(vacc0123p0, vi7x0123, vk7x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
203 vacc0123p0 = vfmaq_f32(vacc0123p0, vi8x0123, vk8x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
207 vacc0123p0 = vfmaq_f32(vacc0123p0, vi9x0123, vk9x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
211 vacc0123p0 = vfmaq_f32(vacc0123p0, vi10x0123, vk10x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
215 vacc0123p0 = vfmaq_f32(vacc0123p0, vi11x0123, vk11x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
219 vacc0123p0 = vfmaq_f32(vacc0123p0, vi12x0123, vk12x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
223 vacc0123p0 = vfmaq_f32(vacc0123p0, vi13x0123, vk13x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
227 vacc0123p0 = vfmaq_f32(vacc0123p0, vi14x0123, vk14x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
231 vacc0123p0 = vfmaq_f32(vacc0123p0, vi15x0123, vk15x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
235 vacc0123p0 = vfmaq_f32(vacc0123p0, vi16x0123, vk16x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
239 vacc0123p0 = vfmaq_f32(vacc0123p0, vi17x0123, vk17x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
243 vacc0123p0 = vfmaq_f32(vacc0123p0, vi18x0123, vk18x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
247 vacc0123p0 = vfmaq_f32(vacc0123p0, vi19x0123, vk19x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
251 vacc0123p0 = vfmaq_f32(vacc0123p0, vi20x0123, vk20x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
255 vacc0123p0 = vfmaq_f32(vacc0123p0, vi21x0123, vk21x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
259 vacc0123p0 = vfmaq_f32(vacc0123p0, vi22x0123, vk22x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
263 vacc0123p0 = vfmaq_f32(vacc0123p0, vi23x0123, vk23x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
267 vacc0123p0 = vfmaq_f32(vacc0123p0, vi24x0123, vk24x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
270 float32x4_t vacc0123 = vmaxq_f32(vacc0123p0, vmin); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
276 float32x4_t vacc0123p0 = vld1q_f32(w); w += 4; in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma() local
281 vacc0123p0 = vfmaq_f32(vacc0123p0, vi0x0123, vk0x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
285 vacc0123p0 = vfmaq_f32(vacc0123p0, vi1x0123, vk1x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
289 vacc0123p0 = vfmaq_f32(vacc0123p0, vi2x0123, vk2x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
293 vacc0123p0 = vfmaq_f32(vacc0123p0, vi3x0123, vk3x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
297 vacc0123p0 = vfmaq_f32(vacc0123p0, vi4x0123, vk4x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
301 vacc0123p0 = vfmaq_f32(vacc0123p0, vi5x0123, vk5x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
305 vacc0123p0 = vfmaq_f32(vacc0123p0, vi6x0123, vk6x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
309 vacc0123p0 = vfmaq_f32(vacc0123p0, vi7x0123, vk7x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
313 vacc0123p0 = vfmaq_f32(vacc0123p0, vi8x0123, vk8x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
317 vacc0123p0 = vfmaq_f32(vacc0123p0, vi9x0123, vk9x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
321 vacc0123p0 = vfmaq_f32(vacc0123p0, vi10x0123, vk10x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
325 vacc0123p0 = vfmaq_f32(vacc0123p0, vi11x0123, vk11x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
329 vacc0123p0 = vfmaq_f32(vacc0123p0, vi12x0123, vk12x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
333 vacc0123p0 = vfmaq_f32(vacc0123p0, vi13x0123, vk13x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
337 vacc0123p0 = vfmaq_f32(vacc0123p0, vi14x0123, vk14x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
341 vacc0123p0 = vfmaq_f32(vacc0123p0, vi15x0123, vk15x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
345 vacc0123p0 = vfmaq_f32(vacc0123p0, vi16x0123, vk16x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
349 vacc0123p0 = vfmaq_f32(vacc0123p0, vi17x0123, vk17x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
353 vacc0123p0 = vfmaq_f32(vacc0123p0, vi18x0123, vk18x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
357 vacc0123p0 = vfmaq_f32(vacc0123p0, vi19x0123, vk19x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
361 vacc0123p0 = vfmaq_f32(vacc0123p0, vi20x0123, vk20x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
365 vacc0123p0 = vfmaq_f32(vacc0123p0, vi21x0123, vk21x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
369 vacc0123p0 = vfmaq_f32(vacc0123p0, vi22x0123, vk22x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
373 vacc0123p0 = vfmaq_f32(vacc0123p0, vi23x0123, vk23x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
377 vacc0123p0 = vfmaq_f32(vacc0123p0, vi24x0123, vk24x0123); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()
380 float32x4_t vacc0123 = vmaxq_f32(vacc0123p0, vmin); in xnn_f32_dwconv_minmax_ukernel_up4x25__neonfma()