Home
last modified time | relevance | path

Searched defs:__m256d (Results 1 – 25 of 50) sorted by relevance

12

/prebuilts/rust/linux-x86/1.82.0/lib/rustlib/src/rust/library/stdarch/crates/core_arch/src/x86/
Dfma.rs48 pub unsafe fn _mm256_fmadd_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fmadd_pd()
136 pub unsafe fn _mm256_fmaddsub_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fmaddsub_pd()
192 pub unsafe fn _mm256_fmsub_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fmsub_pd()
280 pub unsafe fn _mm256_fmsubadd_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fmsubadd_pd()
336 pub unsafe fn _mm256_fnmadd_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fnmadd_pd()
422 pub unsafe fn _mm256_fnmsub_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fnmsub_pd()
Davx.rs33 pub unsafe fn _mm256_add_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_add_pd()
58 pub unsafe fn _mm256_and_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_and_pd()
87 pub unsafe fn _mm256_or_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_or_pd()
116 pub unsafe fn _mm256_shuffle_pd<const MASK: i32>(a: __m256d, b: __m256d) -> __m256d { in _mm256_shuffle_pd()
165 pub unsafe fn _mm256_andnot_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_andnot_pd()
194 pub unsafe fn _mm256_max_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_max_pd()
218 pub unsafe fn _mm256_min_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_min_pd()
242 pub unsafe fn _mm256_mul_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_mul_pd()
266 pub unsafe fn _mm256_addsub_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_addsub_pd()
298 pub unsafe fn _mm256_sub_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_sub_pd()
[all …]
Dtest.rs63 pub unsafe fn assert_eq_m256d(a: __m256d, b: __m256d) { in assert_eq_m256d()
71 pub unsafe fn get_m256d(a: __m256d, idx: usize) -> f64 { in get_m256d()
Davx512dq.rs46 pub unsafe fn _mm256_mask_and_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_and_pd()
59 pub unsafe fn _mm256_maskz_and_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_and_pd()
244 pub unsafe fn _mm256_mask_andnot_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_andnot_pd()
258 pub unsafe fn _mm256_maskz_andnot_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_andnot_pd()
443 pub unsafe fn _mm256_mask_or_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_or_pd()
456 pub unsafe fn _mm256_maskz_or_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_or_pd()
640 pub unsafe fn _mm256_mask_xor_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_xor_pd()
653 pub unsafe fn _mm256_maskz_xor_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_xor_pd()
922 pub unsafe fn _mm256_broadcast_f64x2(a: __m128d) -> __m256d { in _mm256_broadcast_f64x2()
934 pub unsafe fn _mm256_mask_broadcast_f64x2(src: __m256d, k: __mmask8, a: __m128d) -> __m256d { in _mm256_mask_broadcast_f64x2()
[all …]
Davx512f.rs523 pub unsafe fn _mm256_mask_mov_pd(src: __m256d, k: __mmask8, a: __m256d) -> __m256d { in _mm256_mask_mov_pd()
535 pub unsafe fn _mm256_maskz_mov_pd(k: __mmask8, a: __m256d) -> __m256d { in _mm256_maskz_mov_pd()
867 pub unsafe fn _mm256_mask_add_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_add_pd()
879 pub unsafe fn _mm256_maskz_add_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_add_pd()
1211 pub unsafe fn _mm256_mask_sub_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_sub_pd()
1223 pub unsafe fn _mm256_maskz_sub_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_sub_pd()
1686 pub unsafe fn _mm256_mask_mul_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_mul_pd()
1698 pub unsafe fn _mm256_maskz_mul_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_mul_pd()
1858 pub unsafe fn _mm256_mask_div_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_div_pd()
1870 pub unsafe fn _mm256_maskz_div_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_div_pd()
[all …]
Davx2.rs586 pub unsafe fn _mm256_broadcastsd_pd(a: __m128d) -> __m256d { in _mm256_broadcastsd_pd()
1328 ) -> __m256d { in _mm256_i32gather_pd()
1348 pub unsafe fn _mm256_mask_i32gather_pd<const SCALE: i32>( in _mm256_mask_i32gather_pd()
1352 mask: __m256d, in _mm256_mask_i32gather_pd()
1353 ) -> __m256d { in _mm256_mask_i32gather_pd()
1695 ) -> __m256d { in _mm256_i64gather_pd()
1715 pub unsafe fn _mm256_mask_i64gather_pd<const SCALE: i32>( in _mm256_mask_i64gather_pd()
1719 mask: __m256d, in _mm256_mask_i64gather_pd()
1720 ) -> __m256d { in _mm256_mask_i64gather_pd()
2304 pub unsafe fn _mm256_permute4x64_pd<const IMM8: i32>(a: __m256d) -> __m256d { in _mm256_permute4x64_pd()
[all …]
/prebuilts/rust/linux-musl-x86/1.83.0/lib/rustlib/src/rust/library/stdarch/crates/core_arch/src/x86/
Dfma.rs48 pub unsafe fn _mm256_fmadd_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fmadd_pd()
136 pub unsafe fn _mm256_fmaddsub_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fmaddsub_pd()
192 pub unsafe fn _mm256_fmsub_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fmsub_pd()
280 pub unsafe fn _mm256_fmsubadd_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fmsubadd_pd()
336 pub unsafe fn _mm256_fnmadd_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fnmadd_pd()
422 pub unsafe fn _mm256_fnmsub_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fnmsub_pd()
Davx.rs33 pub unsafe fn _mm256_add_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_add_pd()
58 pub unsafe fn _mm256_and_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_and_pd()
87 pub unsafe fn _mm256_or_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_or_pd()
116 pub unsafe fn _mm256_shuffle_pd<const MASK: i32>(a: __m256d, b: __m256d) -> __m256d { in _mm256_shuffle_pd()
165 pub unsafe fn _mm256_andnot_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_andnot_pd()
194 pub unsafe fn _mm256_max_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_max_pd()
218 pub unsafe fn _mm256_min_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_min_pd()
242 pub unsafe fn _mm256_mul_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_mul_pd()
266 pub unsafe fn _mm256_addsub_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_addsub_pd()
298 pub unsafe fn _mm256_sub_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_sub_pd()
[all …]
Dtest.rs63 pub unsafe fn assert_eq_m256d(a: __m256d, b: __m256d) { in assert_eq_m256d()
71 pub unsafe fn get_m256d(a: __m256d, idx: usize) -> f64 { in get_m256d()
Davx512dq.rs46 pub unsafe fn _mm256_mask_and_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_and_pd()
59 pub unsafe fn _mm256_maskz_and_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_and_pd()
244 pub unsafe fn _mm256_mask_andnot_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_andnot_pd()
258 pub unsafe fn _mm256_maskz_andnot_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_andnot_pd()
443 pub unsafe fn _mm256_mask_or_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_or_pd()
456 pub unsafe fn _mm256_maskz_or_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_or_pd()
640 pub unsafe fn _mm256_mask_xor_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_xor_pd()
653 pub unsafe fn _mm256_maskz_xor_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_xor_pd()
922 pub unsafe fn _mm256_broadcast_f64x2(a: __m128d) -> __m256d { in _mm256_broadcast_f64x2()
934 pub unsafe fn _mm256_mask_broadcast_f64x2(src: __m256d, k: __mmask8, a: __m128d) -> __m256d { in _mm256_mask_broadcast_f64x2()
[all …]
Davx512f.rs523 pub unsafe fn _mm256_mask_mov_pd(src: __m256d, k: __mmask8, a: __m256d) -> __m256d { in _mm256_mask_mov_pd()
535 pub unsafe fn _mm256_maskz_mov_pd(k: __mmask8, a: __m256d) -> __m256d { in _mm256_maskz_mov_pd()
867 pub unsafe fn _mm256_mask_add_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_add_pd()
879 pub unsafe fn _mm256_maskz_add_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_add_pd()
1211 pub unsafe fn _mm256_mask_sub_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_sub_pd()
1223 pub unsafe fn _mm256_maskz_sub_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_sub_pd()
1686 pub unsafe fn _mm256_mask_mul_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_mul_pd()
1698 pub unsafe fn _mm256_maskz_mul_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_mul_pd()
1858 pub unsafe fn _mm256_mask_div_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_div_pd()
1870 pub unsafe fn _mm256_maskz_div_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_div_pd()
[all …]
Davx2.rs586 pub unsafe fn _mm256_broadcastsd_pd(a: __m128d) -> __m256d { in _mm256_broadcastsd_pd()
1328 ) -> __m256d { in _mm256_i32gather_pd()
1348 pub unsafe fn _mm256_mask_i32gather_pd<const SCALE: i32>( in _mm256_mask_i32gather_pd()
1352 mask: __m256d, in _mm256_mask_i32gather_pd()
1353 ) -> __m256d { in _mm256_mask_i32gather_pd()
1695 ) -> __m256d { in _mm256_i64gather_pd()
1715 pub unsafe fn _mm256_mask_i64gather_pd<const SCALE: i32>( in _mm256_mask_i64gather_pd()
1719 mask: __m256d, in _mm256_mask_i64gather_pd()
1720 ) -> __m256d { in _mm256_mask_i64gather_pd()
2304 pub unsafe fn _mm256_permute4x64_pd<const IMM8: i32>(a: __m256d) -> __m256d { in _mm256_permute4x64_pd()
[all …]
/prebuilts/rust/linux-x86/1.83.0/lib/rustlib/src/rust/library/stdarch/crates/core_arch/src/x86/
Dfma.rs48 pub unsafe fn _mm256_fmadd_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fmadd_pd()
136 pub unsafe fn _mm256_fmaddsub_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fmaddsub_pd()
192 pub unsafe fn _mm256_fmsub_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fmsub_pd()
280 pub unsafe fn _mm256_fmsubadd_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fmsubadd_pd()
336 pub unsafe fn _mm256_fnmadd_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fnmadd_pd()
422 pub unsafe fn _mm256_fnmsub_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fnmsub_pd()
Davx.rs33 pub unsafe fn _mm256_add_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_add_pd()
58 pub unsafe fn _mm256_and_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_and_pd()
87 pub unsafe fn _mm256_or_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_or_pd()
116 pub unsafe fn _mm256_shuffle_pd<const MASK: i32>(a: __m256d, b: __m256d) -> __m256d { in _mm256_shuffle_pd()
165 pub unsafe fn _mm256_andnot_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_andnot_pd()
194 pub unsafe fn _mm256_max_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_max_pd()
218 pub unsafe fn _mm256_min_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_min_pd()
242 pub unsafe fn _mm256_mul_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_mul_pd()
266 pub unsafe fn _mm256_addsub_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_addsub_pd()
298 pub unsafe fn _mm256_sub_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_sub_pd()
[all …]
Dtest.rs63 pub unsafe fn assert_eq_m256d(a: __m256d, b: __m256d) { in assert_eq_m256d()
71 pub unsafe fn get_m256d(a: __m256d, idx: usize) -> f64 { in get_m256d()
Davx512dq.rs46 pub unsafe fn _mm256_mask_and_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_and_pd()
59 pub unsafe fn _mm256_maskz_and_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_and_pd()
244 pub unsafe fn _mm256_mask_andnot_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_andnot_pd()
258 pub unsafe fn _mm256_maskz_andnot_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_andnot_pd()
443 pub unsafe fn _mm256_mask_or_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_or_pd()
456 pub unsafe fn _mm256_maskz_or_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_or_pd()
640 pub unsafe fn _mm256_mask_xor_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_xor_pd()
653 pub unsafe fn _mm256_maskz_xor_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_xor_pd()
922 pub unsafe fn _mm256_broadcast_f64x2(a: __m128d) -> __m256d { in _mm256_broadcast_f64x2()
934 pub unsafe fn _mm256_mask_broadcast_f64x2(src: __m256d, k: __mmask8, a: __m128d) -> __m256d { in _mm256_mask_broadcast_f64x2()
[all …]
Davx512f.rs523 pub unsafe fn _mm256_mask_mov_pd(src: __m256d, k: __mmask8, a: __m256d) -> __m256d { in _mm256_mask_mov_pd()
535 pub unsafe fn _mm256_maskz_mov_pd(k: __mmask8, a: __m256d) -> __m256d { in _mm256_maskz_mov_pd()
867 pub unsafe fn _mm256_mask_add_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_add_pd()
879 pub unsafe fn _mm256_maskz_add_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_add_pd()
1211 pub unsafe fn _mm256_mask_sub_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_sub_pd()
1223 pub unsafe fn _mm256_maskz_sub_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_sub_pd()
1686 pub unsafe fn _mm256_mask_mul_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_mul_pd()
1698 pub unsafe fn _mm256_maskz_mul_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_mul_pd()
1858 pub unsafe fn _mm256_mask_div_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_div_pd()
1870 pub unsafe fn _mm256_maskz_div_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_div_pd()
[all …]
Davx2.rs586 pub unsafe fn _mm256_broadcastsd_pd(a: __m128d) -> __m256d { in _mm256_broadcastsd_pd()
1328 ) -> __m256d { in _mm256_i32gather_pd()
1348 pub unsafe fn _mm256_mask_i32gather_pd<const SCALE: i32>( in _mm256_mask_i32gather_pd()
1352 mask: __m256d, in _mm256_mask_i32gather_pd()
1353 ) -> __m256d { in _mm256_mask_i32gather_pd()
1695 ) -> __m256d { in _mm256_i64gather_pd()
1715 pub unsafe fn _mm256_mask_i64gather_pd<const SCALE: i32>( in _mm256_mask_i64gather_pd()
1719 mask: __m256d, in _mm256_mask_i64gather_pd()
1720 ) -> __m256d { in _mm256_mask_i64gather_pd()
2304 pub unsafe fn _mm256_permute4x64_pd<const IMM8: i32>(a: __m256d) -> __m256d { in _mm256_permute4x64_pd()
[all …]
Dmod.rs672 fn as_m256d(self) -> __m256d; in as_m256d()
680 impl m256dExt for __m256d { implementation
/prebuilts/rust/linux-musl-x86/1.82.0/lib/rustlib/src/rust/library/stdarch/crates/core_arch/src/x86/
Dfma.rs48 pub unsafe fn _mm256_fmadd_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fmadd_pd()
136 pub unsafe fn _mm256_fmaddsub_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fmaddsub_pd()
192 pub unsafe fn _mm256_fmsub_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fmsub_pd()
280 pub unsafe fn _mm256_fmsubadd_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fmsubadd_pd()
336 pub unsafe fn _mm256_fnmadd_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fnmadd_pd()
422 pub unsafe fn _mm256_fnmsub_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fnmsub_pd()
Davx.rs33 pub unsafe fn _mm256_add_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_add_pd()
58 pub unsafe fn _mm256_and_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_and_pd()
87 pub unsafe fn _mm256_or_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_or_pd()
116 pub unsafe fn _mm256_shuffle_pd<const MASK: i32>(a: __m256d, b: __m256d) -> __m256d { in _mm256_shuffle_pd()
165 pub unsafe fn _mm256_andnot_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_andnot_pd()
194 pub unsafe fn _mm256_max_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_max_pd()
218 pub unsafe fn _mm256_min_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_min_pd()
242 pub unsafe fn _mm256_mul_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_mul_pd()
266 pub unsafe fn _mm256_addsub_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_addsub_pd()
298 pub unsafe fn _mm256_sub_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_sub_pd()
[all …]
Dtest.rs63 pub unsafe fn assert_eq_m256d(a: __m256d, b: __m256d) { in assert_eq_m256d()
71 pub unsafe fn get_m256d(a: __m256d, idx: usize) -> f64 { in get_m256d()
Davx512dq.rs46 pub unsafe fn _mm256_mask_and_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_and_pd()
59 pub unsafe fn _mm256_maskz_and_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_and_pd()
244 pub unsafe fn _mm256_mask_andnot_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_andnot_pd()
258 pub unsafe fn _mm256_maskz_andnot_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_andnot_pd()
443 pub unsafe fn _mm256_mask_or_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_or_pd()
456 pub unsafe fn _mm256_maskz_or_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_or_pd()
640 pub unsafe fn _mm256_mask_xor_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_xor_pd()
653 pub unsafe fn _mm256_maskz_xor_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_xor_pd()
922 pub unsafe fn _mm256_broadcast_f64x2(a: __m128d) -> __m256d { in _mm256_broadcast_f64x2()
934 pub unsafe fn _mm256_mask_broadcast_f64x2(src: __m256d, k: __mmask8, a: __m128d) -> __m256d { in _mm256_mask_broadcast_f64x2()
[all …]
Davx512f.rs523 pub unsafe fn _mm256_mask_mov_pd(src: __m256d, k: __mmask8, a: __m256d) -> __m256d { in _mm256_mask_mov_pd()
535 pub unsafe fn _mm256_maskz_mov_pd(k: __mmask8, a: __m256d) -> __m256d { in _mm256_maskz_mov_pd()
867 pub unsafe fn _mm256_mask_add_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_add_pd()
879 pub unsafe fn _mm256_maskz_add_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_add_pd()
1211 pub unsafe fn _mm256_mask_sub_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_sub_pd()
1223 pub unsafe fn _mm256_maskz_sub_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_sub_pd()
1686 pub unsafe fn _mm256_mask_mul_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_mul_pd()
1698 pub unsafe fn _mm256_maskz_mul_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_mul_pd()
1858 pub unsafe fn _mm256_mask_div_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_div_pd()
1870 pub unsafe fn _mm256_maskz_div_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_div_pd()
[all …]
Davx2.rs586 pub unsafe fn _mm256_broadcastsd_pd(a: __m128d) -> __m256d { in _mm256_broadcastsd_pd()
1328 ) -> __m256d { in _mm256_i32gather_pd()
1348 pub unsafe fn _mm256_mask_i32gather_pd<const SCALE: i32>( in _mm256_mask_i32gather_pd()
1352 mask: __m256d, in _mm256_mask_i32gather_pd()
1353 ) -> __m256d { in _mm256_mask_i32gather_pd()
1695 ) -> __m256d { in _mm256_i64gather_pd()
1715 pub unsafe fn _mm256_mask_i64gather_pd<const SCALE: i32>( in _mm256_mask_i64gather_pd()
1719 mask: __m256d, in _mm256_mask_i64gather_pd()
1720 ) -> __m256d { in _mm256_mask_i64gather_pd()
2304 pub unsafe fn _mm256_permute4x64_pd<const IMM8: i32>(a: __m256d) -> __m256d { in _mm256_permute4x64_pd()
[all …]

12