/external/XNNPACK/src/f32-sigmoid/ |
D | wasmsimd-lut64-p2-div.c.in | 112 const uint64_t vidx_lo = wasm_i64x2_extract_lane(vidx, 0); variable 148 const uint64_t vidx_lo = wasm_i64x2_extract_lane(vidx, 0); variable
|
/external/XNNPACK/src/f32-raddstoreexpminusmax/gen/ |
D | neonfma-lut64-p2-x4.c | 73 const uint64_t vidx_lo = vgetq_lane_u64(vidx, 0); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x4() local 149 const uint64_t vidx_lo = vgetq_lane_u64(vidx, 0); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x4() local
|
D | neon-lut64-p2-x4.c | 74 const uint64_t vidx_lo = vgetq_lane_u64(vidx, 0); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x4() local 150 const uint64_t vidx_lo = vgetq_lane_u64(vidx, 0); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x4() local
|
D | neon-lut64-p2-x8.c | 172 const uint64_t vidx_lo = vgetq_lane_u64(vidx, 0); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x8() local 248 const uint64_t vidx_lo = vgetq_lane_u64(vidx, 0); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x8() local
|
D | neonfma-lut64-p2-x8.c | 171 const uint64_t vidx_lo = vgetq_lane_u64(vidx, 0); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x8() local 247 const uint64_t vidx_lo = vgetq_lane_u64(vidx, 0); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x8() local
|
D | neonfma-lut64-p2-x8-acc2.c | 174 const uint64_t vidx_lo = vgetq_lane_u64(vidx, 0); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x8_acc2() local 250 const uint64_t vidx_lo = vgetq_lane_u64(vidx, 0); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x8_acc2() local
|
/external/XNNPACK/src/f32-sigmoid/gen/ |
D | neonfma-rr1-lut64-p2-div-x4.c | 45 const uint64_t vidx_lo = vgetq_lane_u64(vidx, 0); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_div_x4() local 79 const uint64_t vidx_lo = vgetq_lane_u64(vidx, 0); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_div_x4() local
|
D | neonfma-rr1-lut2048-p1-div-x4.c | 45 const uint64_t vidx_lo = vgetq_lane_u64(vidx, 0); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_div_x4() local 78 const uint64_t vidx_lo = vgetq_lane_u64(vidx, 0); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_div_x4() local
|
D | neonfma-rr1-lut64-p2-nr2fma-x4.c | 45 const uint64_t vidx_lo = vgetq_lane_u64(vidx, 0); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr2fma_x4() local 83 const uint64_t vidx_lo = vgetq_lane_u64(vidx, 0); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr2fma_x4() local
|
D | neonfma-rr1-lut2048-p1-nr1recps1fma-x4.c | 45 const uint64_t vidx_lo = vgetq_lane_u64(vidx, 0); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_nr1recps1fma_x4() local 82 const uint64_t vidx_lo = vgetq_lane_u64(vidx, 0); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_nr1recps1fma_x4() local
|
D | neonfma-rr1-lut2048-p1-nr2recps-x4.c | 45 const uint64_t vidx_lo = vgetq_lane_u64(vidx, 0); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_nr2recps_x4() local 82 const uint64_t vidx_lo = vgetq_lane_u64(vidx, 0); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_nr2recps_x4() local
|
D | wasmsimd-lut64-p2-div-x4.c | 47 const uint64_t vidx_lo = wasm_i64x2_extract_lane(vidx, 0); in xnn_f32_sigmoid_ukernel__wasmsimd_lut64_p2_div_x4() local 83 const uint64_t vidx_lo = wasm_i64x2_extract_lane(vidx, 0); in xnn_f32_sigmoid_ukernel__wasmsimd_lut64_p2_div_x4() local
|
D | neonfma-rr1-lut2048-p1-nr2fma-x4.c | 45 const uint64_t vidx_lo = vgetq_lane_u64(vidx, 0); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_nr2fma_x4() local 82 const uint64_t vidx_lo = vgetq_lane_u64(vidx, 0); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_nr2fma_x4() local
|
D | neonfma-rr1-lut64-p2-nr2recps-x4.c | 45 const uint64_t vidx_lo = vgetq_lane_u64(vidx, 0); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr2recps_x4() local 83 const uint64_t vidx_lo = vgetq_lane_u64(vidx, 0); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr2recps_x4() local
|
D | neonfma-rr1-lut64-p2-nr1recps1fma-x4.c | 45 const uint64_t vidx_lo = vgetq_lane_u64(vidx, 0); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr1recps1fma_x4() local 83 const uint64_t vidx_lo = vgetq_lane_u64(vidx, 0); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr1recps1fma_x4() local
|
D | neon-rr2-lut2048-p1-nr2recps-x4.c | 46 const uint64_t vidx_lo = vgetq_lane_u64(vidx, 0); in xnn_f32_sigmoid_ukernel__neon_rr2_lut2048_p1_nr2recps_x4() local 84 const uint64_t vidx_lo = vgetq_lane_u64(vidx, 0); in xnn_f32_sigmoid_ukernel__neon_rr2_lut2048_p1_nr2recps_x4() local
|
D | neon-rr2-lut64-p2-nr2recps-x4.c | 46 const uint64_t vidx_lo = vgetq_lane_u64(vidx, 0); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x4() local 85 const uint64_t vidx_lo = vgetq_lane_u64(vidx, 0); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x4() local
|
D | sse41-lut64-p2-div-x4.c | 49 const uint64_t vidx_lo = (uint64_t) _mm_cvtsi128_si64(vidx); in xnn_f32_sigmoid_ukernel__sse41_lut64_p2_div_x4() local 91 const uint64_t vidx_lo = (uint64_t) _mm_cvtsi128_si64(vidx); in xnn_f32_sigmoid_ukernel__sse41_lut64_p2_div_x4() local
|
D | sse2-lut64-p2-div-x4.c | 49 const uint64_t vidx_lo = (uint64_t) _mm_cvtsi128_si64(vidx); in xnn_f32_sigmoid_ukernel__sse2_lut64_p2_div_x4() local 96 const uint64_t vidx_lo = (uint64_t) _mm_cvtsi128_si64(vidx); in xnn_f32_sigmoid_ukernel__sse2_lut64_p2_div_x4() local
|
/external/XNNPACK/src/f32-velu/gen/ |
D | velu-neonfma-rr1-lut16-p3-x4.c | 53 const uint64_t vidx_lo = vgetq_lane_u64(vidx, 0); in xnn_f32_velu_ukernel__neonfma_rr1_lut16_p3_x4() local 89 const uint64_t vidx_lo = vgetq_lane_u64(vidx, 0); in xnn_f32_velu_ukernel__neonfma_rr1_lut16_p3_x4() local
|
D | velu-neon-rr2-lut16-p3-x4.c | 54 const uint64_t vidx_lo = vgetq_lane_u64(vidx, 0); in xnn_f32_velu_ukernel__neon_rr2_lut16_p3_x4() local 91 const uint64_t vidx_lo = vgetq_lane_u64(vidx, 0); in xnn_f32_velu_ukernel__neon_rr2_lut16_p3_x4() local
|
D | velu-wasmsimd-arm-rr2-lut16-p3-x4.c | 55 const uint64_t vidx_lo = wasm_i64x2_extract_lane(vidx, 0); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_lut16_p3_x4() local 93 const uint64_t vidx_lo = wasm_i64x2_extract_lane(vidx, 0); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_lut16_p3_x4() local
|
D | velu-wasmsimd-x86-rr2-lut16-p3-x4.c | 55 const uint64_t vidx_lo = wasm_i64x2_extract_lane(vidx, 0); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_lut16_p3_x4() local 96 const uint64_t vidx_lo = wasm_i64x2_extract_lane(vidx, 0); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_lut16_p3_x4() local
|
D | velu-sse41-rr2-lut16-p3-x4.c | 56 const uint64_t vidx_lo = (uint64_t) _mm_cvtsi128_si64(vidx); in xnn_f32_velu_ukernel__sse41_rr2_lut16_p3_x4() local 99 const uint64_t vidx_lo = (uint64_t) _mm_cvtsi128_si64(vidx); in xnn_f32_velu_ukernel__sse41_rr2_lut16_p3_x4() local
|
D | velu-sse2-rr2-lut16-p3-x4.c | 56 const uint64_t vidx_lo = (uint64_t) _mm_cvtsi128_si64(vidx); in xnn_f32_velu_ukernel__sse2_rr2_lut16_p3_x4() local 104 const uint64_t vidx_lo = (uint64_t) _mm_cvtsi128_si64(vidx); in xnn_f32_velu_ukernel__sse2_rr2_lut16_p3_x4() local
|