Lines Matching refs:vn
51 v128_t vn = wasm_f32x4_add(wasm_f32x4_mul(vz, vlog2e), vmagic_bias); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_lut16_p3_x4() local
52 const v128_t vidx = wasm_i32x4_shl(wasm_v128_and(vn, vindex_mask), 2); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_lut16_p3_x4()
53 const v128_t ven = wasm_i32x4_shl(vn, 19); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_lut16_p3_x4()
64 vn = wasm_f32x4_sub(vn, vmagic_bias); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_lut16_p3_x4()
66 v128_t vt = wasm_f32x4_add(wasm_f32x4_mul(vn, vminus_ln2_hi), vz); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_lut16_p3_x4()
68 vt = wasm_f32x4_add(wasm_f32x4_mul(vn, vminus_ln2_lo), vt); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_lut16_p3_x4()
92 v128_t vn = wasm_f32x4_add(wasm_f32x4_mul(vz, vlog2e), vmagic_bias); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_lut16_p3_x4() local
93 const v128_t vidx = wasm_i32x4_shl(wasm_v128_and(vn, vindex_mask), 2); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_lut16_p3_x4()
94 const v128_t ven = wasm_i32x4_shl(vn, 19); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_lut16_p3_x4()
105 vn = wasm_f32x4_sub(vn, vmagic_bias); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_lut16_p3_x4()
107 v128_t vt = wasm_f32x4_add(wasm_f32x4_mul(vn, vminus_ln2_hi), vz); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_lut16_p3_x4()
109 vt = wasm_f32x4_add(wasm_f32x4_mul(vn, vminus_ln2_lo), vt); in xnn_f32_velu_ukernel__wasmsimd_x86_rr2_lut16_p3_x4()