| /external/XNNPACK/src/f32-spmm/gen/ |
| D | 8x1-minmax-scalar-pipelined.c | 85 float vout1 = math_min_f32(vacc1, vmax); in xnn_f32_spmm_minmax_ukernel_8x1__scalar_pipelined() local 151 float vout1 = math_min_f32(vacc1, vmax); in xnn_f32_spmm_minmax_ukernel_8x1__scalar_pipelined() local 195 float vout1 = math_min_f32(vacc1, vmax); in xnn_f32_spmm_minmax_ukernel_8x1__scalar_pipelined() local
|
| D | 4x1-minmax-scalar-pipelined.c | 69 float vout1 = math_min_f32(vacc1, vmax); in xnn_f32_spmm_minmax_ukernel_4x1__scalar_pipelined() local 115 float vout1 = math_min_f32(vacc1, vmax); in xnn_f32_spmm_minmax_ukernel_4x1__scalar_pipelined() local
|
| D | 8x1-minmax-scalar.c | 142 float vout1 = math_min_f32(vacc1, vmax); in xnn_f32_spmm_minmax_ukernel_8x1__scalar() local 239 float vout1 = math_min_f32(vacc1, vmax); in xnn_f32_spmm_minmax_ukernel_8x1__scalar() local 304 float vout1 = math_min_f32(vacc1, vmax); in xnn_f32_spmm_minmax_ukernel_8x1__scalar() local
|
| /external/XNNPACK/src/s16-vlshift/gen/ |
| D | scalar-x2.c | 35 const uint16_t vout1 = vi1 << shift; in xnn_s16_vlshift_ukernel__scalar_x2() local
|
| D | scalar-x3.c | 36 const uint16_t vout1 = vi1 << shift; in xnn_s16_vlshift_ukernel__scalar_x3() local
|
| D | scalar-x4.c | 37 const uint16_t vout1 = vi1 << shift; in xnn_s16_vlshift_ukernel__scalar_x4() local
|
| D | neon-x16.c | 38 const int16x8_t vout1 = vshlq_s16(vi1, vshift); in xnn_s16_vlshift_ukernel__neon_x16() local
|
| D | neon-x24.c | 39 const int16x8_t vout1 = vshlq_s16(vi1, vshift); in xnn_s16_vlshift_ukernel__neon_x24() local
|
| D | neon-x32.c | 40 const int16x8_t vout1 = vshlq_s16(vi1, vshift); in xnn_s16_vlshift_ukernel__neon_x32() local
|
| /external/XNNPACK/src/qu8-vcvt/gen/ |
| D | vcvt-scalar-x2.c | 33 int32_t vout1 = math_asr_s32(vacc1, 8); in xnn_qu8_vcvt_ukernel__scalar_x2() local
|
| D | vcvt-scalar-x4.c | 37 int32_t vout1 = math_asr_s32(vacc1, 8); in xnn_qu8_vcvt_ukernel__scalar_x4() local
|
| /external/XNNPACK/src/qs8-vcvt/gen/ |
| D | vcvt-scalar-x2.c | 33 int32_t vout1 = math_asr_s32(vacc1, 8); in xnn_qs8_vcvt_ukernel__scalar_x2() local
|
| D | vcvt-scalar-x4.c | 37 int32_t vout1 = math_asr_s32(vacc1, 8); in xnn_qs8_vcvt_ukernel__scalar_x4() local
|
| /external/XNNPACK/src/qs8-vlrelu/gen/ |
| D | vlrelu-scalar-andxor-x2.c | 47 int32_t vout1 = math_asr_s32(vacc1, 8); in xnn_qs8_vlrelu_ukernel__scalar_andxor_x2() local
|
| D | vlrelu-scalar-select-x2.c | 41 int32_t vout1 = math_asr_s32(vacc1, 8); in xnn_qs8_vlrelu_ukernel__scalar_select_x2() local
|
| D | vlrelu-scalar-andxor-x4.c | 59 int32_t vout1 = math_asr_s32(vacc1, 8); in xnn_qs8_vlrelu_ukernel__scalar_andxor_x4() local
|
| /external/XNNPACK/src/qu8-vlrelu/gen/ |
| D | vlrelu-scalar-select-x2.c | 41 int32_t vout1 = math_asr_s32(vacc1, 8); in xnn_qu8_vlrelu_ukernel__scalar_select_x2() local
|
| D | vlrelu-scalar-andxor-x2.c | 47 int32_t vout1 = math_asr_s32(vacc1, 8); in xnn_qu8_vlrelu_ukernel__scalar_andxor_x2() local
|
| D | vlrelu-scalar-select-x4.c | 49 int32_t vout1 = math_asr_s32(vacc1, 8); in xnn_qu8_vlrelu_ukernel__scalar_select_x4() local
|
| /external/XNNPACK/src/s16-window/gen/ |
| D | scalar-x2.c | 46 int32_t vout1 = (int32_t) vi1 * (int32_t) w1; in xnn_s16_window_ukernel__scalar_x2() local
|
| D | scalar-x3.c | 48 int32_t vout1 = (int32_t) vi1 * (int32_t) w1; in xnn_s16_window_ukernel__scalar_x3() local
|
| D | neon-shift15-x16.c | 47 const int16x8_t vout1 = vqdmulhq_s16(vi1, vw1); in xnn_s16_window_shift15_ukernel__neon_x16() local
|
| /external/XNNPACK/src/cs16-vsquareabs/gen/ |
| D | scalar-x2.c | 40 const uint32_t vout1 = vrsquare1 + visquare1; in xnn_cs16_vsquareabs_ukernel__scalar_x2() local
|
| /external/XNNPACK/src/qu8-vaddc/gen/ |
| D | minmax-scalar-x2.c | 40 int32_t vout1 = math_asr_s32(vacc1, vshift); in xnn_qu8_vaddc_minmax_ukernel__scalar_x2() local
|
| /external/XNNPACK/src/qs8-vaddc/gen/ |
| D | minmax-scalar-x2.c | 40 int32_t vout1 = math_asr_s32(vacc1, vshift); in xnn_qs8_vaddc_minmax_ukernel__scalar_x2() local
|