Lines Matching refs:vn
51 v128_t vn = wasm_f32x4_add(wasm_f32x4_mul(vz, vlog2e), vmagic_bias); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_lut16_p3_x4() local
52 const v128_t vidx = wasm_i32x4_shl(wasm_v128_and(vn, vindex_mask), 2); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_lut16_p3_x4()
53 const v128_t ven = wasm_i32x4_shl(vn, 19); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_lut16_p3_x4()
64 vn = wasm_f32x4_sub(vn, vmagic_bias); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_lut16_p3_x4()
66 v128_t vt = wasm_f32x4_add(wasm_f32x4_mul(vn, vminus_ln2_hi), vz); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_lut16_p3_x4()
67 vt = wasm_f32x4_add(wasm_f32x4_mul(vn, vminus_ln2_lo), vt); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_lut16_p3_x4()
89 v128_t vn = wasm_f32x4_add(wasm_f32x4_mul(vz, vlog2e), vmagic_bias); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_lut16_p3_x4() local
90 const v128_t vidx = wasm_i32x4_shl(wasm_v128_and(vn, vindex_mask), 2); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_lut16_p3_x4()
91 const v128_t ven = wasm_i32x4_shl(vn, 19); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_lut16_p3_x4()
102 vn = wasm_f32x4_sub(vn, vmagic_bias); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_lut16_p3_x4()
104 v128_t vt = wasm_f32x4_add(wasm_f32x4_mul(vn, vminus_ln2_hi), vz); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_lut16_p3_x4()
105 vt = wasm_f32x4_add(wasm_f32x4_mul(vn, vminus_ln2_lo), vt); in xnn_f32_velu_ukernel__wasmsimd_arm_rr2_lut16_p3_x4()