| /external/XNNPACK/src/f32-raddstoreexpminusmax/gen/ |
| D | sse2-rr2-p5-x20.c | 47 const __m128 viGHIJ = _mm_loadu_ps(input + 16); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_rr2_p5_x20() local
|
| D | sse2-rr2-p5-x20-acc2.c | 48 const __m128 viGHIJ = _mm_loadu_ps(input + 16); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_rr2_p5_x20_acc2() local
|
| D | neonfma-rr1-p5-x20.c | 45 const float32x4_t viGHIJ = vld1q_f32(input); input += 4; in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_rr1_p5_x20() local
|
| D | neon-rr2-p5-x20.c | 46 const float32x4_t viGHIJ = vld1q_f32(input); input += 4; in xnn_f32_raddstoreexpminusmax_ukernel__neon_rr2_p5_x20() local
|
| D | neonfma-rr1-p5-x20-acc5.c | 49 const float32x4_t viGHIJ = vld1q_f32(input); input += 4; in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_rr1_p5_x20_acc5() local
|
| D | neon-rr2-p5-x20-acc2.c | 47 const float32x4_t viGHIJ = vld1q_f32(input); input += 4; in xnn_f32_raddstoreexpminusmax_ukernel__neon_rr2_p5_x20_acc2() local
|
| D | wasmsimd-rr2-p5-x20-acc2.c | 48 const v128_t viGHIJ = wasm_v128_load(input + 16); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_rr2_p5_x20_acc2() local
|
| D | wasmsimd-rr2-p5-x20.c | 47 const v128_t viGHIJ = wasm_v128_load(input + 16); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_rr2_p5_x20() local
|
| D | neonfma-rr1-p5-x20-acc2.c | 46 const float32x4_t viGHIJ = vld1q_f32(input); input += 4; in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_rr1_p5_x20_acc2() local
|
| D | sse2-rr2-p5-x20-acc5.c | 51 const __m128 viGHIJ = _mm_loadu_ps(input + 16); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_rr2_p5_x20_acc5() local
|
| D | wasmsimd-rr2-p5-x20-acc5.c | 51 const v128_t viGHIJ = wasm_v128_load(input + 16); in xnn_f32_raddstoreexpminusmax_ukernel__wasmsimd_rr2_p5_x20_acc5() local
|
| D | neon-rr2-p5-x20-acc5.c | 50 const float32x4_t viGHIJ = vld1q_f32(input); input += 4; in xnn_f32_raddstoreexpminusmax_ukernel__neon_rr2_p5_x20_acc5() local
|
| D | neon-rr2-lut64-p2-x20-acc2.c | 46 const float32x4_t viGHIJ = vld1q_f32(input); input += 4; in xnn_f32_raddstoreexpminusmax_ukernel__neon_rr2_lut64_p2_x20_acc2() local
|
| /external/XNNPACK/src/f32-spmm/gen/ |
| D | 32x2-minmax-neonfma.c | 65 const float32x4_t viGHIJ = vld1q_f32(input + 16); in xnn_f32_spmm_minmax_ukernel_32x2__neonfma() local 166 const float32x4_t viGHIJ = vld1q_f32(input + 16); in xnn_f32_spmm_minmax_ukernel_32x2__neonfma() local
|
| D | 32x1-minmax-wasmsimd-arm.c | 57 const v128_t viGHIJ = wasm_v128_load(input + 16); in xnn_f32_spmm_minmax_ukernel_32x1__wasmsimd_arm() local
|
| D | 32x1-minmax-wasmsimd-x86-pipelined.c | 45 v128_t viGHIJ = wasm_v128_load(input + 16); in xnn_f32_spmm_minmax_ukernel_32x1__wasmsimd_x86_pipelined() local
|
| D | 32x1-minmax-neon.c | 57 const float32x4_t viGHIJ = vld1q_f32(input + 16); in xnn_f32_spmm_minmax_ukernel_32x1__neon() local
|
| D | 32x1-minmax-neonfma-pipelined.c | 45 float32x4_t viGHIJ = vld1q_f32(input + 16); in xnn_f32_spmm_minmax_ukernel_32x1__neonfma_pipelined() local
|
| D | 32x1-minmax-wasmsimd-arm-pipelined-x2.c | 45 v128_t viGHIJ = wasm_v128_load(input + 16); in xnn_f32_spmm_minmax_ukernel_32x1__wasmsimd_arm_pipelined_x2() local
|
| D | 32x1-minmax-wasmsimd-x86.c | 57 const v128_t viGHIJ = wasm_v128_load(input + 16); in xnn_f32_spmm_minmax_ukernel_32x1__wasmsimd_x86() local
|
| D | 32x1-minmax-neonfma.c | 57 const float32x4_t viGHIJ = vld1q_f32(input + 16); in xnn_f32_spmm_minmax_ukernel_32x1__neonfma() local
|
| D | 32x1-minmax-wasmsimd-arm-pipelined.c | 45 v128_t viGHIJ = wasm_v128_load(input + 16); in xnn_f32_spmm_minmax_ukernel_32x1__wasmsimd_arm_pipelined() local
|
| D | 32x1-minmax-wasmsimd-x86-pipelined-x2.c | 45 v128_t viGHIJ = wasm_v128_load(input + 16); in xnn_f32_spmm_minmax_ukernel_32x1__wasmsimd_x86_pipelined_x2() local
|
| D | 32x1-minmax-sse.c | 57 const __m128 viGHIJ = _mm_loadu_ps(input + 16); in xnn_f32_spmm_minmax_ukernel_32x1__sse() local
|
| D | 32x1-minmax-neon-pipelined.c | 45 float32x4_t viGHIJ = vld1q_f32(input + 16); in xnn_f32_spmm_minmax_ukernel_32x1__neon_pipelined() local
|