| /external/XNNPACK/src/f32-spmm/gen/ |
| D | 4x1-minmax-scalar-pipelined.c | 68 float vout0 = math_min_f32(vacc0, vmax); in xnn_f32_spmm_minmax_ukernel_4x1__scalar_pipelined() local 114 float vout0 = math_min_f32(vacc0, vmax); in xnn_f32_spmm_minmax_ukernel_4x1__scalar_pipelined() local 148 float vout0 = math_min_f32(vacc0, vmax); in xnn_f32_spmm_minmax_ukernel_4x1__scalar_pipelined() local
|
| D | 8x1-minmax-scalar-pipelined.c | 84 float vout0 = math_min_f32(vacc0, vmax); in xnn_f32_spmm_minmax_ukernel_8x1__scalar_pipelined() local 150 float vout0 = math_min_f32(vacc0, vmax); in xnn_f32_spmm_minmax_ukernel_8x1__scalar_pipelined() local 194 float vout0 = math_min_f32(vacc0, vmax); in xnn_f32_spmm_minmax_ukernel_8x1__scalar_pipelined() local 228 float vout0 = math_min_f32(vacc0, vmax); in xnn_f32_spmm_minmax_ukernel_8x1__scalar_pipelined() local
|
| D | 2x1-minmax-scalar-pipelined.c | 60 float vout0 = math_min_f32(vacc0, vmax); in xnn_f32_spmm_minmax_ukernel_2x1__scalar_pipelined() local 96 float vout0 = math_min_f32(vacc0, vmax); in xnn_f32_spmm_minmax_ukernel_2x1__scalar_pipelined() local
|
| D | 4x1-minmax-scalar.c | 101 float vout0 = math_min_f32(vacc0, vmax); in xnn_f32_spmm_minmax_ukernel_4x1__scalar() local 168 float vout0 = math_min_f32(vacc0, vmax); in xnn_f32_spmm_minmax_ukernel_4x1__scalar() local 218 float vout0 = math_min_f32(vacc0, vmax); in xnn_f32_spmm_minmax_ukernel_4x1__scalar() local
|
| D | 2x1-minmax-scalar.c | 81 float vout0 = math_min_f32(vacc0, vmax); in xnn_f32_spmm_minmax_ukernel_2x1__scalar() local 133 float vout0 = math_min_f32(vacc0, vmax); in xnn_f32_spmm_minmax_ukernel_2x1__scalar() local
|
| D | 8x1-minmax-scalar.c | 141 float vout0 = math_min_f32(vacc0, vmax); in xnn_f32_spmm_minmax_ukernel_8x1__scalar() local 238 float vout0 = math_min_f32(vacc0, vmax); in xnn_f32_spmm_minmax_ukernel_8x1__scalar() local 303 float vout0 = math_min_f32(vacc0, vmax); in xnn_f32_spmm_minmax_ukernel_8x1__scalar() local 353 float vout0 = math_min_f32(vacc0, vmax); in xnn_f32_spmm_minmax_ukernel_8x1__scalar() local
|
| D | 8x2-minmax-scalar.c | 183 float vout0 = math_min_f32(vacc0, vmax); in xnn_f32_spmm_minmax_ukernel_8x2__scalar() local 302 float vout0 = math_min_f32(vacc0, vmax); in xnn_f32_spmm_minmax_ukernel_8x2__scalar() local 379 float vout0 = math_min_f32(vacc0, vmax); in xnn_f32_spmm_minmax_ukernel_8x2__scalar() local 436 float vout0 = math_min_f32(vacc0, vmax); in xnn_f32_spmm_minmax_ukernel_8x2__scalar() local
|
| /external/XNNPACK/src/s16-vlshift/gen/ |
| D | scalar-x2.c | 34 const uint16_t vout0 = vi0 << shift; in xnn_s16_vlshift_ukernel__scalar_x2() local
|
| D | scalar-x3.c | 35 const uint16_t vout0 = vi0 << shift; in xnn_s16_vlshift_ukernel__scalar_x3() local
|
| D | scalar-x4.c | 36 const uint16_t vout0 = vi0 << shift; in xnn_s16_vlshift_ukernel__scalar_x4() local
|
| D | neon-x16.c | 37 const int16x8_t vout0 = vshlq_s16(vi0, vshift); in xnn_s16_vlshift_ukernel__neon_x16() local
|
| D | neon-x24.c | 38 const int16x8_t vout0 = vshlq_s16(vi0, vshift); in xnn_s16_vlshift_ukernel__neon_x24() local
|
| D | neon-x32.c | 39 const int16x8_t vout0 = vshlq_s16(vi0, vshift); in xnn_s16_vlshift_ukernel__neon_x32() local
|
| /external/XNNPACK/src/qu8-vcvt/gen/ |
| D | vcvt-scalar-x2.c | 32 int32_t vout0 = math_asr_s32(vacc0, 8); in xnn_qu8_vcvt_ukernel__scalar_x2() local
|
| D | vcvt-scalar-x4.c | 36 int32_t vout0 = math_asr_s32(vacc0, 8); in xnn_qu8_vcvt_ukernel__scalar_x4() local
|
| /external/XNNPACK/src/qs8-vcvt/gen/ |
| D | vcvt-scalar-x2.c | 32 int32_t vout0 = math_asr_s32(vacc0, 8); in xnn_qs8_vcvt_ukernel__scalar_x2() local
|
| D | vcvt-scalar-x4.c | 36 int32_t vout0 = math_asr_s32(vacc0, 8); in xnn_qs8_vcvt_ukernel__scalar_x4() local
|
| /external/XNNPACK/src/cs16-bfly4/ |
| D | samples1-neon.c | 32 int16x4_t vout0 = vadd_s16(vget_low_s16(vout), vget_high_s16(vout)); in xnn_cs16_bfly4_samples1_ukernel__neon() local
|
| /external/XNNPACK/src/qs8-vlrelu/gen/ |
| D | vlrelu-scalar-andxor-x2.c | 46 int32_t vout0 = math_asr_s32(vacc0, 8); in xnn_qs8_vlrelu_ukernel__scalar_andxor_x2() local
|
| D | vlrelu-scalar-select-x2.c | 40 int32_t vout0 = math_asr_s32(vacc0, 8); in xnn_qs8_vlrelu_ukernel__scalar_select_x2() local
|
| /external/XNNPACK/src/qu8-vlrelu/gen/ |
| D | vlrelu-scalar-select-x2.c | 40 int32_t vout0 = math_asr_s32(vacc0, 8); in xnn_qu8_vlrelu_ukernel__scalar_select_x2() local
|
| D | vlrelu-scalar-andxor-x2.c | 46 int32_t vout0 = math_asr_s32(vacc0, 8); in xnn_qu8_vlrelu_ukernel__scalar_andxor_x2() local
|
| /external/XNNPACK/src/s16-window/gen/ |
| D | scalar-x2.c | 45 int32_t vout0 = (int32_t) vi0 * (int32_t) w0; in xnn_s16_window_ukernel__scalar_x2() local
|
| /external/XNNPACK/src/cs16-vsquareabs/gen/ |
| D | scalar-x2.c | 39 const uint32_t vout0 = vrsquare0 + visquare0; in xnn_cs16_vsquareabs_ukernel__scalar_x2() local
|
| /external/XNNPACK/src/qu8-vaddc/gen/ |
| D | minmax-scalar-x2.c | 39 int32_t vout0 = math_asr_s32(vacc0, vshift); in xnn_qu8_vaddc_minmax_ukernel__scalar_x2() local
|