/external/XNNPACK/src/x32-pad/ |
D | neon.c | 44 vst1q_lane_u32(output, vfill, 0); output += 1; in xnn_x32_pad_ukernel__neon() 78 vst1q_lane_u32(output, vfill, 0); output += 1; in xnn_x32_pad_ukernel__neon()
|
/external/XNNPACK/src/qs8-gemm/gen/ |
D | 8x8c4-minmax-neondot.c | 328 …vst1q_lane_u32(__builtin_assume_aligned(c0, 1), vreinterpretq_u32_s8(vout0x01234567_1x01234567), 0… in xnn_qs8_gemm_minmax_ukernel_8x8c4__neondot() 329 …vst1q_lane_u32(__builtin_assume_aligned(c1, 1), vreinterpretq_u32_s8(vout0x01234567_1x01234567), 2… in xnn_qs8_gemm_minmax_ukernel_8x8c4__neondot() 330 …vst1q_lane_u32(__builtin_assume_aligned(c2, 1), vreinterpretq_u32_s8(vout2x01234567_3x01234567), 0… in xnn_qs8_gemm_minmax_ukernel_8x8c4__neondot() 331 …vst1q_lane_u32(__builtin_assume_aligned(c3, 1), vreinterpretq_u32_s8(vout2x01234567_3x01234567), 2… in xnn_qs8_gemm_minmax_ukernel_8x8c4__neondot() 332 …vst1q_lane_u32(__builtin_assume_aligned(c4, 1), vreinterpretq_u32_s8(vout4x01234567_5x01234567), 0… in xnn_qs8_gemm_minmax_ukernel_8x8c4__neondot() 333 …vst1q_lane_u32(__builtin_assume_aligned(c5, 1), vreinterpretq_u32_s8(vout4x01234567_5x01234567), 2… in xnn_qs8_gemm_minmax_ukernel_8x8c4__neondot() 334 …vst1q_lane_u32(__builtin_assume_aligned(c6, 1), vreinterpretq_u32_s8(vout6x01234567_7x01234567), 0… in xnn_qs8_gemm_minmax_ukernel_8x8c4__neondot() 335 …vst1q_lane_u32(__builtin_assume_aligned(c7, 1), vreinterpretq_u32_s8(vout6x01234567_7x01234567), 2… in xnn_qs8_gemm_minmax_ukernel_8x8c4__neondot()
|
D | 6x8c4-minmax-neondot.c | 270 …vst1q_lane_u32(__builtin_assume_aligned(c0, 1), vreinterpretq_u32_s8(vout0x01234567_1x01234567), 0… in xnn_qs8_gemm_minmax_ukernel_6x8c4__neondot() 271 …vst1q_lane_u32(__builtin_assume_aligned(c1, 1), vreinterpretq_u32_s8(vout0x01234567_1x01234567), 2… in xnn_qs8_gemm_minmax_ukernel_6x8c4__neondot() 272 …vst1q_lane_u32(__builtin_assume_aligned(c2, 1), vreinterpretq_u32_s8(vout2x01234567_3x01234567), 0… in xnn_qs8_gemm_minmax_ukernel_6x8c4__neondot() 273 …vst1q_lane_u32(__builtin_assume_aligned(c3, 1), vreinterpretq_u32_s8(vout2x01234567_3x01234567), 2… in xnn_qs8_gemm_minmax_ukernel_6x8c4__neondot() 274 …vst1q_lane_u32(__builtin_assume_aligned(c4, 1), vreinterpretq_u32_s8(vout4x01234567_5x01234567), 0… in xnn_qs8_gemm_minmax_ukernel_6x8c4__neondot() 275 …vst1q_lane_u32(__builtin_assume_aligned(c5, 1), vreinterpretq_u32_s8(vout4x01234567_5x01234567), 2… in xnn_qs8_gemm_minmax_ukernel_6x8c4__neondot()
|
D | 4x8c4-minmax-neondot.c | 212 …vst1q_lane_u32(__builtin_assume_aligned(c0, 1), vreinterpretq_u32_s8(vout0x01234567_1x01234567), 0… in xnn_qs8_gemm_minmax_ukernel_4x8c4__neondot() 213 …vst1q_lane_u32(__builtin_assume_aligned(c1, 1), vreinterpretq_u32_s8(vout0x01234567_1x01234567), 2… in xnn_qs8_gemm_minmax_ukernel_4x8c4__neondot() 214 …vst1q_lane_u32(__builtin_assume_aligned(c2, 1), vreinterpretq_u32_s8(vout2x01234567_3x01234567), 0… in xnn_qs8_gemm_minmax_ukernel_4x8c4__neondot() 215 …vst1q_lane_u32(__builtin_assume_aligned(c3, 1), vreinterpretq_u32_s8(vout2x01234567_3x01234567), 2… in xnn_qs8_gemm_minmax_ukernel_4x8c4__neondot()
|
D | 8x16c4-minmax-neondot.c | 496 …vst1q_lane_u32(__builtin_assume_aligned(c0, 1), vreinterpretq_u32_s8(vout0x01234567_1x01234567), 0… in xnn_qs8_gemm_minmax_ukernel_8x16c4__neondot() 497 …vst1q_lane_u32(__builtin_assume_aligned(c1, 1), vreinterpretq_u32_s8(vout0x01234567_1x01234567), 2… in xnn_qs8_gemm_minmax_ukernel_8x16c4__neondot() 498 …vst1q_lane_u32(__builtin_assume_aligned(c2, 1), vreinterpretq_u32_s8(vout2x01234567_3x01234567), 0… in xnn_qs8_gemm_minmax_ukernel_8x16c4__neondot() 499 …vst1q_lane_u32(__builtin_assume_aligned(c3, 1), vreinterpretq_u32_s8(vout2x01234567_3x01234567), 2… in xnn_qs8_gemm_minmax_ukernel_8x16c4__neondot() 500 …vst1q_lane_u32(__builtin_assume_aligned(c4, 1), vreinterpretq_u32_s8(vout4x01234567_5x01234567), 0… in xnn_qs8_gemm_minmax_ukernel_8x16c4__neondot() 501 …vst1q_lane_u32(__builtin_assume_aligned(c5, 1), vreinterpretq_u32_s8(vout4x01234567_5x01234567), 2… in xnn_qs8_gemm_minmax_ukernel_8x16c4__neondot() 502 …vst1q_lane_u32(__builtin_assume_aligned(c6, 1), vreinterpretq_u32_s8(vout6x01234567_7x01234567), 0… in xnn_qs8_gemm_minmax_ukernel_8x16c4__neondot() 503 …vst1q_lane_u32(__builtin_assume_aligned(c7, 1), vreinterpretq_u32_s8(vout6x01234567_7x01234567), 2… in xnn_qs8_gemm_minmax_ukernel_8x16c4__neondot()
|
D | 6x16c4-minmax-neondot.c | 398 …vst1q_lane_u32(__builtin_assume_aligned(c0, 1), vreinterpretq_u32_s8(vout0x01234567_1x01234567), 0… in xnn_qs8_gemm_minmax_ukernel_6x16c4__neondot() 399 …vst1q_lane_u32(__builtin_assume_aligned(c1, 1), vreinterpretq_u32_s8(vout0x01234567_1x01234567), 2… in xnn_qs8_gemm_minmax_ukernel_6x16c4__neondot() 400 …vst1q_lane_u32(__builtin_assume_aligned(c2, 1), vreinterpretq_u32_s8(vout2x01234567_3x01234567), 0… in xnn_qs8_gemm_minmax_ukernel_6x16c4__neondot() 401 …vst1q_lane_u32(__builtin_assume_aligned(c3, 1), vreinterpretq_u32_s8(vout2x01234567_3x01234567), 2… in xnn_qs8_gemm_minmax_ukernel_6x16c4__neondot() 402 …vst1q_lane_u32(__builtin_assume_aligned(c4, 1), vreinterpretq_u32_s8(vout4x01234567_5x01234567), 0… in xnn_qs8_gemm_minmax_ukernel_6x16c4__neondot() 403 …vst1q_lane_u32(__builtin_assume_aligned(c5, 1), vreinterpretq_u32_s8(vout4x01234567_5x01234567), 2… in xnn_qs8_gemm_minmax_ukernel_6x16c4__neondot()
|
D | 4x16c4-minmax-neondot.c | 300 …vst1q_lane_u32(__builtin_assume_aligned(c0, 1), vreinterpretq_u32_s8(vout0x01234567_1x01234567), 0… in xnn_qs8_gemm_minmax_ukernel_4x16c4__neondot() 301 …vst1q_lane_u32(__builtin_assume_aligned(c1, 1), vreinterpretq_u32_s8(vout0x01234567_1x01234567), 2… in xnn_qs8_gemm_minmax_ukernel_4x16c4__neondot() 302 …vst1q_lane_u32(__builtin_assume_aligned(c2, 1), vreinterpretq_u32_s8(vout2x01234567_3x01234567), 0… in xnn_qs8_gemm_minmax_ukernel_4x16c4__neondot() 303 …vst1q_lane_u32(__builtin_assume_aligned(c3, 1), vreinterpretq_u32_s8(vout2x01234567_3x01234567), 2… in xnn_qs8_gemm_minmax_ukernel_4x16c4__neondot()
|
D | 4x8c2-minmax-neon-mull-padal-dup.c | 303 …vst1q_lane_u32(__builtin_assume_aligned(c0, 1), vreinterpretq_u32_s8(vout0x01234567_1x01234567), 0… in xnn_qs8_gemm_minmax_ukernel_4x8c2__neon_mull_padal_dup() 304 …vst1q_lane_u32(__builtin_assume_aligned(c1, 1), vreinterpretq_u32_s8(vout0x01234567_1x01234567), 2… in xnn_qs8_gemm_minmax_ukernel_4x8c2__neon_mull_padal_dup() 305 …vst1q_lane_u32(__builtin_assume_aligned(c2, 1), vreinterpretq_u32_s8(vout2x01234567_3x01234567), 0… in xnn_qs8_gemm_minmax_ukernel_4x8c2__neon_mull_padal_dup() 306 …vst1q_lane_u32(__builtin_assume_aligned(c3, 1), vreinterpretq_u32_s8(vout2x01234567_3x01234567), 2… in xnn_qs8_gemm_minmax_ukernel_4x8c2__neon_mull_padal_dup()
|
D | 4x8c8-minmax-neon-mull-padal.c | 339 …vst1q_lane_u32(__builtin_assume_aligned(c0, 1), vreinterpretq_u32_s8(vout0x01234567_1x01234567), 0… in xnn_qs8_gemm_minmax_ukernel_4x8c8__neon_mull_padal() 340 …vst1q_lane_u32(__builtin_assume_aligned(c1, 1), vreinterpretq_u32_s8(vout0x01234567_1x01234567), 2… in xnn_qs8_gemm_minmax_ukernel_4x8c8__neon_mull_padal() 341 …vst1q_lane_u32(__builtin_assume_aligned(c2, 1), vreinterpretq_u32_s8(vout2x01234567_3x01234567), 0… in xnn_qs8_gemm_minmax_ukernel_4x8c8__neon_mull_padal() 342 …vst1q_lane_u32(__builtin_assume_aligned(c3, 1), vreinterpretq_u32_s8(vout2x01234567_3x01234567), 2… in xnn_qs8_gemm_minmax_ukernel_4x8c8__neon_mull_padal()
|
D | 4x8-minmax-neon-mlal-lane.c | 356 …vst1q_lane_u32(__builtin_assume_aligned(c0, 1), vreinterpretq_u32_s8(vout0x01234567_1x01234567), 0… in xnn_qs8_gemm_minmax_ukernel_4x8__neon_mlal_lane() 357 …vst1q_lane_u32(__builtin_assume_aligned(c1, 1), vreinterpretq_u32_s8(vout0x01234567_1x01234567), 2… in xnn_qs8_gemm_minmax_ukernel_4x8__neon_mlal_lane() 358 …vst1q_lane_u32(__builtin_assume_aligned(c2, 1), vreinterpretq_u32_s8(vout2x01234567_3x01234567), 0… in xnn_qs8_gemm_minmax_ukernel_4x8__neon_mlal_lane() 359 …vst1q_lane_u32(__builtin_assume_aligned(c3, 1), vreinterpretq_u32_s8(vout2x01234567_3x01234567), 2… in xnn_qs8_gemm_minmax_ukernel_4x8__neon_mlal_lane()
|
/external/XNNPACK/src/qs8-igemm/gen/ |
D | 8x8c4-minmax-neondot.c | 342 …vst1q_lane_u32(__builtin_assume_aligned(c7, 1), vreinterpretq_u32_s8(vout6x01234567_7x01234567), 2… in xnn_qs8_igemm_minmax_ukernel_8x8c4__neondot() 343 …vst1q_lane_u32(__builtin_assume_aligned(c6, 1), vreinterpretq_u32_s8(vout6x01234567_7x01234567), 0… in xnn_qs8_igemm_minmax_ukernel_8x8c4__neondot() 344 …vst1q_lane_u32(__builtin_assume_aligned(c5, 1), vreinterpretq_u32_s8(vout4x01234567_5x01234567), 2… in xnn_qs8_igemm_minmax_ukernel_8x8c4__neondot() 345 …vst1q_lane_u32(__builtin_assume_aligned(c4, 1), vreinterpretq_u32_s8(vout4x01234567_5x01234567), 0… in xnn_qs8_igemm_minmax_ukernel_8x8c4__neondot() 346 …vst1q_lane_u32(__builtin_assume_aligned(c3, 1), vreinterpretq_u32_s8(vout2x01234567_3x01234567), 2… in xnn_qs8_igemm_minmax_ukernel_8x8c4__neondot() 347 …vst1q_lane_u32(__builtin_assume_aligned(c2, 1), vreinterpretq_u32_s8(vout2x01234567_3x01234567), 0… in xnn_qs8_igemm_minmax_ukernel_8x8c4__neondot() 348 …vst1q_lane_u32(__builtin_assume_aligned(c1, 1), vreinterpretq_u32_s8(vout0x01234567_1x01234567), 2… in xnn_qs8_igemm_minmax_ukernel_8x8c4__neondot() 349 …vst1q_lane_u32(__builtin_assume_aligned(c0, 1), vreinterpretq_u32_s8(vout0x01234567_1x01234567), 0… in xnn_qs8_igemm_minmax_ukernel_8x8c4__neondot()
|
D | 6x8c4-minmax-neondot.c | 282 …vst1q_lane_u32(__builtin_assume_aligned(c5, 1), vreinterpretq_u32_s8(vout4x01234567_5x01234567), 2… in xnn_qs8_igemm_minmax_ukernel_6x8c4__neondot() 283 …vst1q_lane_u32(__builtin_assume_aligned(c4, 1), vreinterpretq_u32_s8(vout4x01234567_5x01234567), 0… in xnn_qs8_igemm_minmax_ukernel_6x8c4__neondot() 284 …vst1q_lane_u32(__builtin_assume_aligned(c3, 1), vreinterpretq_u32_s8(vout2x01234567_3x01234567), 2… in xnn_qs8_igemm_minmax_ukernel_6x8c4__neondot() 285 …vst1q_lane_u32(__builtin_assume_aligned(c2, 1), vreinterpretq_u32_s8(vout2x01234567_3x01234567), 0… in xnn_qs8_igemm_minmax_ukernel_6x8c4__neondot() 286 …vst1q_lane_u32(__builtin_assume_aligned(c1, 1), vreinterpretq_u32_s8(vout0x01234567_1x01234567), 2… in xnn_qs8_igemm_minmax_ukernel_6x8c4__neondot() 287 …vst1q_lane_u32(__builtin_assume_aligned(c0, 1), vreinterpretq_u32_s8(vout0x01234567_1x01234567), 0… in xnn_qs8_igemm_minmax_ukernel_6x8c4__neondot()
|
D | 4x8c4-minmax-neondot.c | 222 …vst1q_lane_u32(__builtin_assume_aligned(c3, 1), vreinterpretq_u32_s8(vout2x01234567_3x01234567), 2… in xnn_qs8_igemm_minmax_ukernel_4x8c4__neondot() 223 …vst1q_lane_u32(__builtin_assume_aligned(c2, 1), vreinterpretq_u32_s8(vout2x01234567_3x01234567), 0… in xnn_qs8_igemm_minmax_ukernel_4x8c4__neondot() 224 …vst1q_lane_u32(__builtin_assume_aligned(c1, 1), vreinterpretq_u32_s8(vout0x01234567_1x01234567), 2… in xnn_qs8_igemm_minmax_ukernel_4x8c4__neondot() 225 …vst1q_lane_u32(__builtin_assume_aligned(c0, 1), vreinterpretq_u32_s8(vout0x01234567_1x01234567), 0… in xnn_qs8_igemm_minmax_ukernel_4x8c4__neondot()
|
D | 8x16c4-minmax-neondot.c | 510 …vst1q_lane_u32(__builtin_assume_aligned(c7, 1), vreinterpretq_u32_s8(vout6x01234567_7x01234567), 2… in xnn_qs8_igemm_minmax_ukernel_8x16c4__neondot() 511 …vst1q_lane_u32(__builtin_assume_aligned(c6, 1), vreinterpretq_u32_s8(vout6x01234567_7x01234567), 0… in xnn_qs8_igemm_minmax_ukernel_8x16c4__neondot() 512 …vst1q_lane_u32(__builtin_assume_aligned(c5, 1), vreinterpretq_u32_s8(vout4x01234567_5x01234567), 2… in xnn_qs8_igemm_minmax_ukernel_8x16c4__neondot() 513 …vst1q_lane_u32(__builtin_assume_aligned(c4, 1), vreinterpretq_u32_s8(vout4x01234567_5x01234567), 0… in xnn_qs8_igemm_minmax_ukernel_8x16c4__neondot() 514 …vst1q_lane_u32(__builtin_assume_aligned(c3, 1), vreinterpretq_u32_s8(vout2x01234567_3x01234567), 2… in xnn_qs8_igemm_minmax_ukernel_8x16c4__neondot() 515 …vst1q_lane_u32(__builtin_assume_aligned(c2, 1), vreinterpretq_u32_s8(vout2x01234567_3x01234567), 0… in xnn_qs8_igemm_minmax_ukernel_8x16c4__neondot() 516 …vst1q_lane_u32(__builtin_assume_aligned(c1, 1), vreinterpretq_u32_s8(vout0x01234567_1x01234567), 2… in xnn_qs8_igemm_minmax_ukernel_8x16c4__neondot() 517 …vst1q_lane_u32(__builtin_assume_aligned(c0, 1), vreinterpretq_u32_s8(vout0x01234567_1x01234567), 0… in xnn_qs8_igemm_minmax_ukernel_8x16c4__neondot()
|
D | 6x16c4-minmax-neondot.c | 410 …vst1q_lane_u32(__builtin_assume_aligned(c5, 1), vreinterpretq_u32_s8(vout4x01234567_5x01234567), 2… in xnn_qs8_igemm_minmax_ukernel_6x16c4__neondot() 411 …vst1q_lane_u32(__builtin_assume_aligned(c4, 1), vreinterpretq_u32_s8(vout4x01234567_5x01234567), 0… in xnn_qs8_igemm_minmax_ukernel_6x16c4__neondot() 412 …vst1q_lane_u32(__builtin_assume_aligned(c3, 1), vreinterpretq_u32_s8(vout2x01234567_3x01234567), 2… in xnn_qs8_igemm_minmax_ukernel_6x16c4__neondot() 413 …vst1q_lane_u32(__builtin_assume_aligned(c2, 1), vreinterpretq_u32_s8(vout2x01234567_3x01234567), 0… in xnn_qs8_igemm_minmax_ukernel_6x16c4__neondot() 414 …vst1q_lane_u32(__builtin_assume_aligned(c1, 1), vreinterpretq_u32_s8(vout0x01234567_1x01234567), 2… in xnn_qs8_igemm_minmax_ukernel_6x16c4__neondot() 415 …vst1q_lane_u32(__builtin_assume_aligned(c0, 1), vreinterpretq_u32_s8(vout0x01234567_1x01234567), 0… in xnn_qs8_igemm_minmax_ukernel_6x16c4__neondot()
|
D | 4x16c4-minmax-neondot.c | 310 …vst1q_lane_u32(__builtin_assume_aligned(c3, 1), vreinterpretq_u32_s8(vout2x01234567_3x01234567), 2… in xnn_qs8_igemm_minmax_ukernel_4x16c4__neondot() 311 …vst1q_lane_u32(__builtin_assume_aligned(c2, 1), vreinterpretq_u32_s8(vout2x01234567_3x01234567), 0… in xnn_qs8_igemm_minmax_ukernel_4x16c4__neondot() 312 …vst1q_lane_u32(__builtin_assume_aligned(c1, 1), vreinterpretq_u32_s8(vout0x01234567_1x01234567), 2… in xnn_qs8_igemm_minmax_ukernel_4x16c4__neondot() 313 …vst1q_lane_u32(__builtin_assume_aligned(c0, 1), vreinterpretq_u32_s8(vout0x01234567_1x01234567), 0… in xnn_qs8_igemm_minmax_ukernel_4x16c4__neondot()
|
D | 4x8c2-minmax-neon-mull-padal-dup.c | 320 …vst1q_lane_u32(__builtin_assume_aligned(c3, 1), vreinterpretq_u32_s8(vout2x01234567_3x01234567), 2… in xnn_qs8_igemm_minmax_ukernel_4x8c2__neon_mull_padal_dup() 321 …vst1q_lane_u32(__builtin_assume_aligned(c2, 1), vreinterpretq_u32_s8(vout2x01234567_3x01234567), 0… in xnn_qs8_igemm_minmax_ukernel_4x8c2__neon_mull_padal_dup() 322 …vst1q_lane_u32(__builtin_assume_aligned(c1, 1), vreinterpretq_u32_s8(vout0x01234567_1x01234567), 2… in xnn_qs8_igemm_minmax_ukernel_4x8c2__neon_mull_padal_dup() 323 …vst1q_lane_u32(__builtin_assume_aligned(c0, 1), vreinterpretq_u32_s8(vout0x01234567_1x01234567), 0… in xnn_qs8_igemm_minmax_ukernel_4x8c2__neon_mull_padal_dup()
|
D | 4x8c8-minmax-neon-mull-padal.c | 356 …vst1q_lane_u32(__builtin_assume_aligned(c3, 1), vreinterpretq_u32_s8(vout2x01234567_3x01234567), 2… in xnn_qs8_igemm_minmax_ukernel_4x8c8__neon_mull_padal() 357 …vst1q_lane_u32(__builtin_assume_aligned(c2, 1), vreinterpretq_u32_s8(vout2x01234567_3x01234567), 0… in xnn_qs8_igemm_minmax_ukernel_4x8c8__neon_mull_padal() 358 …vst1q_lane_u32(__builtin_assume_aligned(c1, 1), vreinterpretq_u32_s8(vout0x01234567_1x01234567), 2… in xnn_qs8_igemm_minmax_ukernel_4x8c8__neon_mull_padal() 359 …vst1q_lane_u32(__builtin_assume_aligned(c0, 1), vreinterpretq_u32_s8(vout0x01234567_1x01234567), 0… in xnn_qs8_igemm_minmax_ukernel_4x8c8__neon_mull_padal()
|
/external/libvpx/libvpx/vpx_dsp/arm/ |
D | vpx_convolve8_neon.c | 390 vst1q_lane_u32((uint32_t *)(dst + 0 * dst_stride), d0123, 0); in vpx_convolve8_avg_horiz_neon() 391 vst1q_lane_u32((uint32_t *)(dst + 1 * dst_stride), d0123, 2); in vpx_convolve8_avg_horiz_neon() 392 vst1q_lane_u32((uint32_t *)(dst + 2 * dst_stride), d0123, 1); in vpx_convolve8_avg_horiz_neon() 393 vst1q_lane_u32((uint32_t *)(dst + 3 * dst_stride), d0123, 3); in vpx_convolve8_avg_horiz_neon() 477 vst1q_lane_u32((uint32_t *)dst, d0415, 0); in vpx_convolve8_avg_horiz_neon() 479 vst1q_lane_u32((uint32_t *)dst, d0415, 2); in vpx_convolve8_avg_horiz_neon() 481 vst1q_lane_u32((uint32_t *)dst, d2637, 0); in vpx_convolve8_avg_horiz_neon() 483 vst1q_lane_u32((uint32_t *)dst, d2637, 2); in vpx_convolve8_avg_horiz_neon() 485 vst1q_lane_u32((uint32_t *)dst, d0415, 1); in vpx_convolve8_avg_horiz_neon() 487 vst1q_lane_u32((uint32_t *)dst, d0415, 3); in vpx_convolve8_avg_horiz_neon() [all …]
|
/external/XNNPACK/src/x32-unpool/ |
D | neon.c | 35 vst1q_lane_u32(o, vfill, 0); in xnn_x32_unpool_ukernel__neon()
|
/external/XNNPACK/src/x32-fill/ |
D | neon.c | 44 vst1q_lane_u32(output, vfill, 0); output += 1; in xnn_x32_fill_ukernel__neon()
|
/external/XNNPACK/src/qu8-gemm/ |
D | 8x8-minmax-neon.c | 577 …vst1q_lane_u32(__builtin_assume_aligned(c0, 1), vreinterpretq_u32_u8(vout0x01234567_1x01234567), 0… in xnn_qu8_gemm_minmax_ukernel_8x8__neon() 578 …vst1q_lane_u32(__builtin_assume_aligned(c1, 1), vreinterpretq_u32_u8(vout0x01234567_1x01234567), 2… in xnn_qu8_gemm_minmax_ukernel_8x8__neon() 579 …vst1q_lane_u32(__builtin_assume_aligned(c2, 1), vreinterpretq_u32_u8(vout2x01234567_3x01234567), 0… in xnn_qu8_gemm_minmax_ukernel_8x8__neon() 580 …vst1q_lane_u32(__builtin_assume_aligned(c3, 1), vreinterpretq_u32_u8(vout2x01234567_3x01234567), 2… in xnn_qu8_gemm_minmax_ukernel_8x8__neon() 581 …vst1q_lane_u32(__builtin_assume_aligned(c4, 1), vreinterpretq_u32_u8(vout4x01234567_5x01234567), 0… in xnn_qu8_gemm_minmax_ukernel_8x8__neon() 582 …vst1q_lane_u32(__builtin_assume_aligned(c5, 1), vreinterpretq_u32_u8(vout4x01234567_5x01234567), 2… in xnn_qu8_gemm_minmax_ukernel_8x8__neon() 583 …vst1q_lane_u32(__builtin_assume_aligned(c6, 1), vreinterpretq_u32_u8(vout6x01234567_7x01234567), 0… in xnn_qu8_gemm_minmax_ukernel_8x8__neon() 584 …vst1q_lane_u32(__builtin_assume_aligned(c7, 1), vreinterpretq_u32_u8(vout6x01234567_7x01234567), 2… in xnn_qu8_gemm_minmax_ukernel_8x8__neon()
|
/external/XNNPACK/src/qu8-igemm/ |
D | 8x8-minmax-neon.c | 615 …vst1q_lane_u32(__builtin_assume_aligned(c7, 1), vreinterpretq_u32_u8(vout6x01234567_7x01234567), 2… in xnn_qu8_igemm_minmax_ukernel_8x8__neon() 616 …vst1q_lane_u32(__builtin_assume_aligned(c6, 1), vreinterpretq_u32_u8(vout6x01234567_7x01234567), 0… in xnn_qu8_igemm_minmax_ukernel_8x8__neon() 617 …vst1q_lane_u32(__builtin_assume_aligned(c5, 1), vreinterpretq_u32_u8(vout4x01234567_5x01234567), 2… in xnn_qu8_igemm_minmax_ukernel_8x8__neon() 618 …vst1q_lane_u32(__builtin_assume_aligned(c4, 1), vreinterpretq_u32_u8(vout4x01234567_5x01234567), 0… in xnn_qu8_igemm_minmax_ukernel_8x8__neon() 619 …vst1q_lane_u32(__builtin_assume_aligned(c3, 1), vreinterpretq_u32_u8(vout2x01234567_3x01234567), 2… in xnn_qu8_igemm_minmax_ukernel_8x8__neon() 620 …vst1q_lane_u32(__builtin_assume_aligned(c2, 1), vreinterpretq_u32_u8(vout2x01234567_3x01234567), 0… in xnn_qu8_igemm_minmax_ukernel_8x8__neon() 621 …vst1q_lane_u32(__builtin_assume_aligned(c1, 1), vreinterpretq_u32_u8(vout0x01234567_1x01234567), 2… in xnn_qu8_igemm_minmax_ukernel_8x8__neon() 622 …vst1q_lane_u32(__builtin_assume_aligned(c0, 1), vreinterpretq_u32_u8(vout0x01234567_1x01234567), 0… in xnn_qu8_igemm_minmax_ukernel_8x8__neon()
|
D | 4x8-minmax-neon.c | 386 …vst1q_lane_u32(__builtin_assume_aligned(c3, 1), vreinterpretq_u32_u8(vout2x01234567_3x01234567), 2… in xnn_qu8_igemm_minmax_ukernel_4x8__neon() 387 …vst1q_lane_u32(__builtin_assume_aligned(c2, 1), vreinterpretq_u32_u8(vout2x01234567_3x01234567), 0… in xnn_qu8_igemm_minmax_ukernel_4x8__neon() 388 …vst1q_lane_u32(__builtin_assume_aligned(c1, 1), vreinterpretq_u32_u8(vout0x01234567_1x01234567), 2… in xnn_qu8_igemm_minmax_ukernel_4x8__neon() 389 …vst1q_lane_u32(__builtin_assume_aligned(c0, 1), vreinterpretq_u32_u8(vout0x01234567_1x01234567), 0… in xnn_qu8_igemm_minmax_ukernel_4x8__neon()
|
/external/webp/src/dsp/ |
D | lossless_neon.c | 270 vst1q_lane_u32(&out[i + (LANE)], vreinterpretq_u32_u8(res), (LANE)); \ 294 vst1q_lane_u32(&out[i + (LANE)], vreinterpretq_u32_u8(res), (LANE)); \ 357 vst1q_lane_u32(&out[i + (LANE)], vreinterpretq_u32_u8(res), (LANE)); \ 388 vst1q_lane_u32(&out[i + (LANE)], vreinterpretq_u32_u8(res), (LANE)); \
|