/external/XNNPACK/src/qs8-gemm/gen/ |
D | 6x8-minmax-rndnu-neon-mlal-lane.c | 97 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qs8_gemm_minmax_rndnu_ukernel_6x8__neon_mlal_lane() local 236 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qs8_gemm_minmax_rndnu_ukernel_6x8__neon_mlal_lane() local
|
D | 6x8-minmax-rndnu-neon-mlal-lane-prfm.c | 97 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qs8_gemm_minmax_rndnu_ukernel_6x8__neon_mlal_lane_prfm() local 237 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qs8_gemm_minmax_rndnu_ukernel_6x8__neon_mlal_lane_prfm() local
|
D | 6x16-minmax-rndnu-neon-mlal-lane-prfm.c | 109 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qs8_gemm_minmax_rndnu_ukernel_6x16__neon_mlal_lane_prfm() local 370 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qs8_gemm_minmax_rndnu_ukernel_6x16__neon_mlal_lane_prfm() local
|
D | 6x16-minmax-rndnu-neon-mlal-lane.c | 109 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qs8_gemm_minmax_rndnu_ukernel_6x16__neon_mlal_lane() local 368 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qs8_gemm_minmax_rndnu_ukernel_6x16__neon_mlal_lane() local
|
/external/XNNPACK/src/qu8-igemm/gen/ |
D | 6x8-minmax-rndnu-neon-mlal-lane.c | 119 const int16x8_t vxa4 = vreinterpretq_s16_u16(vmovl_u8(va4)); in xnn_qu8_igemm_minmax_rndnu_ukernel_6x8__neon_mlal_lane() local 258 const int16x8_t vxa4 = vreinterpretq_s16_u16(vmovl_u8(va4)); in xnn_qu8_igemm_minmax_rndnu_ukernel_6x8__neon_mlal_lane() local
|
D | 6x16-minmax-rndnu-neon-mlal-lane.c | 131 const int16x8_t vxa4 = vreinterpretq_s16_u16(vmovl_u8(va4)); in xnn_qu8_igemm_minmax_rndnu_ukernel_6x16__neon_mlal_lane() local 390 const int16x8_t vxa4 = vreinterpretq_s16_u16(vmovl_u8(va4)); in xnn_qu8_igemm_minmax_rndnu_ukernel_6x16__neon_mlal_lane() local
|
/external/XNNPACK/src/qs8-igemm/gen/ |
D | 6x8-minmax-rndnu-neon-mlal-lane-prfm.c | 118 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qs8_igemm_minmax_rndnu_ukernel_6x8__neon_mlal_lane_prfm() local 258 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qs8_igemm_minmax_rndnu_ukernel_6x8__neon_mlal_lane_prfm() local
|
D | 6x8-minmax-rndnu-neon-mlal-lane.c | 118 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qs8_igemm_minmax_rndnu_ukernel_6x8__neon_mlal_lane() local 257 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qs8_igemm_minmax_rndnu_ukernel_6x8__neon_mlal_lane() local
|
D | 6x16-minmax-rndnu-neon-mlal-lane-prfm.c | 130 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qs8_igemm_minmax_rndnu_ukernel_6x16__neon_mlal_lane_prfm() local 391 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qs8_igemm_minmax_rndnu_ukernel_6x16__neon_mlal_lane_prfm() local
|
D | 6x16-minmax-rndnu-neon-mlal-lane.c | 130 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qs8_igemm_minmax_rndnu_ukernel_6x16__neon_mlal_lane() local 389 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qs8_igemm_minmax_rndnu_ukernel_6x16__neon_mlal_lane() local
|
/external/XNNPACK/src/qu8-gemm/gen/ |
D | 6x8-minmax-rndnu-neon-mlal-lane.c | 98 const int16x8_t vxa4 = vreinterpretq_s16_u16(vmovl_u8(va4)); in xnn_qu8_gemm_minmax_rndnu_ukernel_6x8__neon_mlal_lane() local 237 const int16x8_t vxa4 = vreinterpretq_s16_u16(vmovl_u8(va4)); in xnn_qu8_gemm_minmax_rndnu_ukernel_6x8__neon_mlal_lane() local
|
D | 6x16-minmax-rndnu-neon-mlal-lane.c | 110 const int16x8_t vxa4 = vreinterpretq_s16_u16(vmovl_u8(va4)); in xnn_qu8_gemm_minmax_rndnu_ukernel_6x16__neon_mlal_lane() local 369 const int16x8_t vxa4 = vreinterpretq_s16_u16(vmovl_u8(va4)); in xnn_qu8_gemm_minmax_rndnu_ukernel_6x16__neon_mlal_lane() local
|
/external/XNNPACK/src/qc8-gemm/gen/ |
D | 6x8-minmax-fp32-neon-mlal-lane.c | 97 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qc8_gemm_minmax_fp32_ukernel_6x8__neon_mlal_lane() local 236 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qc8_gemm_minmax_fp32_ukernel_6x8__neon_mlal_lane() local
|
D | 6x8-minmax-fp32-neon-mlal-lane-prfm.c | 97 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qc8_gemm_minmax_fp32_ukernel_6x8__neon_mlal_lane_prfm() local 237 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qc8_gemm_minmax_fp32_ukernel_6x8__neon_mlal_lane_prfm() local
|
D | 6x8-minmax-fp32-neonv8-mlal-lane-prfm.c | 98 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qc8_gemm_minmax_fp32_ukernel_6x8__neonv8_mlal_lane_prfm() local 238 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qc8_gemm_minmax_fp32_ukernel_6x8__neonv8_mlal_lane_prfm() local
|
D | 6x8-minmax-fp32-neonv8-mlal-lane.c | 98 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qc8_gemm_minmax_fp32_ukernel_6x8__neonv8_mlal_lane() local 237 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qc8_gemm_minmax_fp32_ukernel_6x8__neonv8_mlal_lane() local
|
D | 6x16-minmax-fp32-neon-mlal-lane-prfm.c | 109 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qc8_gemm_minmax_fp32_ukernel_6x16__neon_mlal_lane_prfm() local 370 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qc8_gemm_minmax_fp32_ukernel_6x16__neon_mlal_lane_prfm() local
|
/external/XNNPACK/src/qc8-igemm/gen/ |
D | 6x8-minmax-fp32-neon-mlal-lane.c | 118 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qc8_igemm_minmax_fp32_ukernel_6x8__neon_mlal_lane() local 257 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qc8_igemm_minmax_fp32_ukernel_6x8__neon_mlal_lane() local
|
D | 6x8-minmax-fp32-neonv8-mlal-lane.c | 119 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qc8_igemm_minmax_fp32_ukernel_6x8__neonv8_mlal_lane() local 258 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qc8_igemm_minmax_fp32_ukernel_6x8__neonv8_mlal_lane() local
|
D | 6x8-minmax-fp32-neon-mlal-lane-prfm.c | 118 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qc8_igemm_minmax_fp32_ukernel_6x8__neon_mlal_lane_prfm() local 258 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qc8_igemm_minmax_fp32_ukernel_6x8__neon_mlal_lane_prfm() local
|
D | 6x8-minmax-fp32-neonv8-mlal-lane-prfm.c | 119 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qc8_igemm_minmax_fp32_ukernel_6x8__neonv8_mlal_lane_prfm() local 259 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qc8_igemm_minmax_fp32_ukernel_6x8__neonv8_mlal_lane_prfm() local
|
D | 6x16-minmax-fp32-neon-mlal-lane-prfm.c | 130 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qc8_igemm_minmax_fp32_ukernel_6x16__neon_mlal_lane_prfm() local 391 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qc8_igemm_minmax_fp32_ukernel_6x16__neon_mlal_lane_prfm() local
|
D | 6x16-minmax-fp32-neonv8-mlal-lane.c | 131 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qc8_igemm_minmax_fp32_ukernel_6x16__neonv8_mlal_lane() local 390 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qc8_igemm_minmax_fp32_ukernel_6x16__neonv8_mlal_lane() local
|
D | 6x16-minmax-fp32-neonv8-mlal-lane-prfm.c | 131 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qc8_igemm_minmax_fp32_ukernel_6x16__neonv8_mlal_lane_prfm() local 392 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qc8_igemm_minmax_fp32_ukernel_6x16__neonv8_mlal_lane_prfm() local
|
D | 6x16-minmax-fp32-neon-mlal-lane.c | 130 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qc8_igemm_minmax_fp32_ukernel_6x16__neon_mlal_lane() local 389 const int16x8_t vxa4 = vmovl_s8(va4); in xnn_qc8_igemm_minmax_fp32_ukernel_6x16__neon_mlal_lane() local
|