Home
last modified time | relevance | path

Searched refs:vidx67 (Results 1 – 25 of 117) sorted by relevance

12345

/external/XNNPACK/src/f32-raddstoreexpminusmax/gen/
Dneon-lut64-p2-x8.c82 const uint64_t vidx67 = vgetq_lane_u64(vidx4567, 1); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x8() local
87 float32x2_t vl67 = vld1_dup_f32(&xnn_table_exp2_k_over_64[(uint32_t) vidx67]); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x8()
93 vl67 = vld1_lane_f32(&xnn_table_exp2_k_over_64[(uint32_t) (vidx67 >> 32)], vl67, 1); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x8()
Dneonfma-lut64-p2-x8-acc2.c82 const uint64_t vidx67 = vgetq_lane_u64(vidx4567, 1); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x8_acc2() local
87 float32x2_t vl67 = vld1_dup_f32(&xnn_table_exp2_k_over_64[(uint32_t) vidx67]); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x8_acc2()
93 vl67 = vld1_lane_f32(&xnn_table_exp2_k_over_64[(uint32_t) (vidx67 >> 32)], vl67, 1); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x8_acc2()
Dneon-lut64-p2-x8-acc2.c83 const uint64_t vidx67 = vgetq_lane_u64(vidx4567, 1); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x8_acc2() local
88 float32x2_t vl67 = vld1_dup_f32(&xnn_table_exp2_k_over_64[(uint32_t) vidx67]); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x8_acc2()
94 vl67 = vld1_lane_f32(&xnn_table_exp2_k_over_64[(uint32_t) (vidx67 >> 32)], vl67, 1); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x8_acc2()
Dneonfma-lut64-p2-x8.c81 const uint64_t vidx67 = vgetq_lane_u64(vidx4567, 1); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x8() local
86 float32x2_t vl67 = vld1_dup_f32(&xnn_table_exp2_k_over_64[(uint32_t) vidx67]); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x8()
92 vl67 = vld1_lane_f32(&xnn_table_exp2_k_over_64[(uint32_t) (vidx67 >> 32)], vl67, 1); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x8()
Dneonfma-lut64-p2-x12-acc2.c86 const uint64_t vidx67 = vgetq_lane_u64(vidx4567, 1); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x12_acc2() local
94 float32x2_t vl67 = vld1_dup_f32(&xnn_table_exp2_k_over_64[(uint32_t) vidx67]); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x12_acc2()
102 vl67 = vld1_lane_f32(&xnn_table_exp2_k_over_64[(uint32_t) (vidx67 >> 32)], vl67, 1); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x12_acc2()
Dneon-lut64-p2-x12.c86 const uint64_t vidx67 = vgetq_lane_u64(vidx4567, 1); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x12() local
94 float32x2_t vl67 = vld1_dup_f32(&xnn_table_exp2_k_over_64[(uint32_t) vidx67]); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x12()
102 vl67 = vld1_lane_f32(&xnn_table_exp2_k_over_64[(uint32_t) (vidx67 >> 32)], vl67, 1); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x12()
Dneonfma-lut64-p2-x12.c85 const uint64_t vidx67 = vgetq_lane_u64(vidx4567, 1); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x12() local
93 float32x2_t vl67 = vld1_dup_f32(&xnn_table_exp2_k_over_64[(uint32_t) vidx67]); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x12()
101 vl67 = vld1_lane_f32(&xnn_table_exp2_k_over_64[(uint32_t) (vidx67 >> 32)], vl67, 1); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x12()
Dneon-lut64-p2-x12-acc3.c88 const uint64_t vidx67 = vgetq_lane_u64(vidx4567, 1); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x12_acc3() local
96 float32x2_t vl67 = vld1_dup_f32(&xnn_table_exp2_k_over_64[(uint32_t) vidx67]); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x12_acc3()
104 vl67 = vld1_lane_f32(&xnn_table_exp2_k_over_64[(uint32_t) (vidx67 >> 32)], vl67, 1); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x12_acc3()
/external/XNNPACK/src/f32-sigmoid/gen/
Dneonfma-rr1-lut64-p2-div-x8.c58 const uint64_t vidx67 = vgetq_lane_u64(vidx4567, 1); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_div_x8() local
60 float32x2_t vl67 = vld1_dup_f32(&xnn_table_exp2minus_k_over_64[(uint32_t) vidx67]); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_div_x8()
66 vl67 = vld1_lane_f32(&xnn_table_exp2minus_k_over_64[(uint32_t) (vidx67 >> 32)], vl67, 1); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_div_x8()
Dwasmsimd-lut64-p2-div-x8.c62 const uint64_t vidx67 = wasm_i64x2_extract_lane(vidx4567, 1); in xnn_f32_sigmoid_ukernel__wasmsimd_lut64_p2_div_x8() local
65 …onst float vl6 = *((const float*) ((uintptr_t) xnn_table_exp2minus_k_over_64 + (uint32_t) vidx67)); in xnn_f32_sigmoid_ukernel__wasmsimd_lut64_p2_div_x8()
66 …at vl7 = *((const float*) ((uintptr_t) xnn_table_exp2minus_k_over_64 + (uint32_t) (vidx67 >> 32))); in xnn_f32_sigmoid_ukernel__wasmsimd_lut64_p2_div_x8()
Dneonfma-rr1-lut2048-p1-div-x8.c57 const uint64_t vidx67 = vgetq_lane_u64(vidx4567, 1); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_div_x8() local
59 float32x2_t vl67 = vld1_dup_f32(&xnn_table_exp2minus_k_over_2048[(uint32_t) vidx67]); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_div_x8()
65 vl67 = vld1_lane_f32(&xnn_table_exp2minus_k_over_2048[(uint32_t) (vidx67 >> 32)], vl67, 1); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_div_x8()
Dneonfma-rr1-lut64-p2-nr2recps-x8.c58 const uint64_t vidx67 = vgetq_lane_u64(vidx4567, 1); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr2recps_x8() local
60 float32x2_t vl67 = vld1_dup_f32(&xnn_table_exp2minus_k_over_64[(uint32_t) vidx67]); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr2recps_x8()
66 vl67 = vld1_lane_f32(&xnn_table_exp2minus_k_over_64[(uint32_t) (vidx67 >> 32)], vl67, 1); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr2recps_x8()
Dneon-rr2-lut64-p2-nr2recps-x8.c59 const uint64_t vidx67 = vgetq_lane_u64(vidx4567, 1); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x8() local
61 float32x2_t vl67 = vld1_dup_f32(&xnn_table_exp2minus_k_over_64[(uint32_t) vidx67]); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x8()
67 vl67 = vld1_lane_f32(&xnn_table_exp2minus_k_over_64[(uint32_t) (vidx67 >> 32)], vl67, 1); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x8()
Dneon-rr2-lut2048-p1-nr2recps-x8.c58 const uint64_t vidx67 = vgetq_lane_u64(vidx4567, 1); in xnn_f32_sigmoid_ukernel__neon_rr2_lut2048_p1_nr2recps_x8() local
60 float32x2_t vl67 = vld1_dup_f32(&xnn_table_exp2minus_k_over_2048[(uint32_t) vidx67]); in xnn_f32_sigmoid_ukernel__neon_rr2_lut2048_p1_nr2recps_x8()
66 vl67 = vld1_lane_f32(&xnn_table_exp2minus_k_over_2048[(uint32_t) (vidx67 >> 32)], vl67, 1); in xnn_f32_sigmoid_ukernel__neon_rr2_lut2048_p1_nr2recps_x8()
Dneonfma-rr1-lut64-p2-nr1recps1fma-x8.c58 const uint64_t vidx67 = vgetq_lane_u64(vidx4567, 1); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr1recps1fma_x8() local
60 float32x2_t vl67 = vld1_dup_f32(&xnn_table_exp2minus_k_over_64[(uint32_t) vidx67]); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr1recps1fma_x8()
66 vl67 = vld1_lane_f32(&xnn_table_exp2minus_k_over_64[(uint32_t) (vidx67 >> 32)], vl67, 1); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr1recps1fma_x8()
Dneonfma-rr1-lut2048-p1-nr2recps-x8.c57 const uint64_t vidx67 = vgetq_lane_u64(vidx4567, 1); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_nr2recps_x8() local
59 float32x2_t vl67 = vld1_dup_f32(&xnn_table_exp2minus_k_over_2048[(uint32_t) vidx67]); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_nr2recps_x8()
65 vl67 = vld1_lane_f32(&xnn_table_exp2minus_k_over_2048[(uint32_t) (vidx67 >> 32)], vl67, 1); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_nr2recps_x8()
Dneonfma-rr1-lut2048-p1-nr2fma-x8.c57 const uint64_t vidx67 = vgetq_lane_u64(vidx4567, 1); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_nr2fma_x8() local
59 float32x2_t vl67 = vld1_dup_f32(&xnn_table_exp2minus_k_over_2048[(uint32_t) vidx67]); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_nr2fma_x8()
65 vl67 = vld1_lane_f32(&xnn_table_exp2minus_k_over_2048[(uint32_t) (vidx67 >> 32)], vl67, 1); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_nr2fma_x8()
Dneonfma-rr1-lut2048-p1-nr1recps1fma-x8.c57 const uint64_t vidx67 = vgetq_lane_u64(vidx4567, 1); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_nr1recps1fma_x8() local
59 float32x2_t vl67 = vld1_dup_f32(&xnn_table_exp2minus_k_over_2048[(uint32_t) vidx67]); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_nr1recps1fma_x8()
65 vl67 = vld1_lane_f32(&xnn_table_exp2minus_k_over_2048[(uint32_t) (vidx67 >> 32)], vl67, 1); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_nr1recps1fma_x8()
Dneonfma-rr1-lut64-p2-nr2fma-x8.c58 const uint64_t vidx67 = vgetq_lane_u64(vidx4567, 1); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr2fma_x8() local
60 float32x2_t vl67 = vld1_dup_f32(&xnn_table_exp2minus_k_over_64[(uint32_t) vidx67]); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr2fma_x8()
66 vl67 = vld1_lane_f32(&xnn_table_exp2minus_k_over_64[(uint32_t) (vidx67 >> 32)], vl67, 1); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr2fma_x8()
Dwasmsimd-lut64-p2-div-x12.c67 const uint64_t vidx67 = wasm_i64x2_extract_lane(vidx4567, 1); in xnn_f32_sigmoid_ukernel__wasmsimd_lut64_p2_div_x12() local
70 …onst float vl6 = *((const float*) ((uintptr_t) xnn_table_exp2minus_k_over_64 + (uint32_t) vidx67)); in xnn_f32_sigmoid_ukernel__wasmsimd_lut64_p2_div_x12()
71 …at vl7 = *((const float*) ((uintptr_t) xnn_table_exp2minus_k_over_64 + (uint32_t) (vidx67 >> 32))); in xnn_f32_sigmoid_ukernel__wasmsimd_lut64_p2_div_x12()
Dneonfma-rr1-lut2048-p1-div-x12.c62 const uint64_t vidx67 = vgetq_lane_u64(vidx4567, 1); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_div_x12() local
64 float32x2_t vl67 = vld1_dup_f32(&xnn_table_exp2minus_k_over_2048[(uint32_t) vidx67]); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_div_x12()
74 vl67 = vld1_lane_f32(&xnn_table_exp2minus_k_over_2048[(uint32_t) (vidx67 >> 32)], vl67, 1); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_div_x12()
/external/XNNPACK/src/f32-velu/gen/
Dvelu-wasmsimd-arm-rr2-lut16-p3-x8.c69 const uint64_t vidx67 = wasm_i64x2_extract_lane(vidx4567, 1); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_lut16_p3_x8() local
72 …onst float vl6 = *((const float*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) vidx67)); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_lut16_p3_x8()
73 …at vl7 = *((const float*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) (vidx67 >> 32))); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_lut16_p3_x8()
Dvelu-neonfma-rr1-lut16-p3-x8.c67 const uint64_t vidx67 = vgetq_lane_u64(vidx4567, 1); in xnn_f32_velu_ukernel__neonfma_rr1_lut16_p3_x8() local
69 …7 = vld1_dup_s32((const int32_t*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) vidx67)); in xnn_f32_velu_ukernel__neonfma_rr1_lut16_p3_x8()
71 …(const int32_t*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) (vidx67 >> 32)), vl67, 1); in xnn_f32_velu_ukernel__neonfma_rr1_lut16_p3_x8()
Dvelu-wasmsimd-x86-rr2-lut16-p3-x8.c69 const uint64_t vidx67 = wasm_i64x2_extract_lane(vidx4567, 1); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_lut16_p3_x8() local
72 …onst float vl6 = *((const float*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) vidx67)); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_lut16_p3_x8()
73 …at vl7 = *((const float*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) (vidx67 >> 32))); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_lut16_p3_x8()
Dvelu-neon-rr2-lut16-p3-x8.c68 const uint64_t vidx67 = vgetq_lane_u64(vidx4567, 1); in xnn_f32_velu_ukernel__neon_rr2_lut16_p3_x8() local
70 …7 = vld1_dup_s32((const int32_t*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) vidx67)); in xnn_f32_velu_ukernel__neon_rr2_lut16_p3_x8()
72 …(const int32_t*) ((uintptr_t) xnn_table_exp2minus_k_over_16 + (uint32_t) (vidx67 >> 32)), vl67, 1); in xnn_f32_velu_ukernel__neon_rr2_lut16_p3_x8()

12345