Lines Matching refs:vn
43 v128_t vn = wasm_f32x4_add(vmagic_bias, wasm_f32x4_mul(vz, vminus_log2e)); in xnn_f32_sigmoid_ukernel__wasmsimd_lut64_p2_div_x4() local
44 const v128_t ve = wasm_i32x4_shl(vn, 17); in xnn_f32_sigmoid_ukernel__wasmsimd_lut64_p2_div_x4()
46 const v128_t vidx = wasm_i32x4_shl(wasm_v128_and(vn, vindex_mask), 2); in xnn_f32_sigmoid_ukernel__wasmsimd_lut64_p2_div_x4()
56 vn = wasm_f32x4_sub(vn, vmagic_bias); in xnn_f32_sigmoid_ukernel__wasmsimd_lut64_p2_div_x4()
58 v128_t vt = wasm_f32x4_add(vz, wasm_f32x4_mul(vn, vln2_hi)); in xnn_f32_sigmoid_ukernel__wasmsimd_lut64_p2_div_x4()
59 vt = wasm_f32x4_add(vt, wasm_f32x4_mul(vn, vln2_lo)); in xnn_f32_sigmoid_ukernel__wasmsimd_lut64_p2_div_x4()
79 v128_t vn = wasm_f32x4_add(vmagic_bias, wasm_f32x4_mul(vz, vminus_log2e)); in xnn_f32_sigmoid_ukernel__wasmsimd_lut64_p2_div_x4() local
80 const v128_t ve = wasm_i32x4_shl(vn, 17); in xnn_f32_sigmoid_ukernel__wasmsimd_lut64_p2_div_x4()
82 const v128_t vidx = wasm_i32x4_shl(wasm_v128_and(vn, vindex_mask), 2); in xnn_f32_sigmoid_ukernel__wasmsimd_lut64_p2_div_x4()
92 vn = wasm_f32x4_sub(vn, vmagic_bias); in xnn_f32_sigmoid_ukernel__wasmsimd_lut64_p2_div_x4()
94 v128_t vt = wasm_f32x4_add(vz, wasm_f32x4_mul(vn, vln2_hi)); in xnn_f32_sigmoid_ukernel__wasmsimd_lut64_p2_div_x4()
95 vt = wasm_f32x4_add(vt, wasm_f32x4_mul(vn, vln2_lo)); in xnn_f32_sigmoid_ukernel__wasmsimd_lut64_p2_div_x4()