| /prebuilts/rust/linux-x86/1.82.0/lib/rustlib/src/rust/library/stdarch/crates/core_arch/src/x86/ |
| D | fma.rs | 48 pub unsafe fn _mm256_fmadd_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fmadd_pd() 136 pub unsafe fn _mm256_fmaddsub_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fmaddsub_pd() 192 pub unsafe fn _mm256_fmsub_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fmsub_pd() 280 pub unsafe fn _mm256_fmsubadd_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fmsubadd_pd() 336 pub unsafe fn _mm256_fnmadd_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fnmadd_pd() 422 pub unsafe fn _mm256_fnmsub_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fnmsub_pd()
|
| D | avx.rs | 33 pub unsafe fn _mm256_add_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_add_pd() 58 pub unsafe fn _mm256_and_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_and_pd() 87 pub unsafe fn _mm256_or_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_or_pd() 116 pub unsafe fn _mm256_shuffle_pd<const MASK: i32>(a: __m256d, b: __m256d) -> __m256d { in _mm256_shuffle_pd() 165 pub unsafe fn _mm256_andnot_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_andnot_pd() 194 pub unsafe fn _mm256_max_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_max_pd() 218 pub unsafe fn _mm256_min_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_min_pd() 242 pub unsafe fn _mm256_mul_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_mul_pd() 266 pub unsafe fn _mm256_addsub_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_addsub_pd() 298 pub unsafe fn _mm256_sub_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_sub_pd() [all …]
|
| D | test.rs | 63 pub unsafe fn assert_eq_m256d(a: __m256d, b: __m256d) { in assert_eq_m256d() 71 pub unsafe fn get_m256d(a: __m256d, idx: usize) -> f64 { in get_m256d()
|
| D | avx512dq.rs | 46 pub unsafe fn _mm256_mask_and_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_and_pd() 59 pub unsafe fn _mm256_maskz_and_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_and_pd() 244 pub unsafe fn _mm256_mask_andnot_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_andnot_pd() 258 pub unsafe fn _mm256_maskz_andnot_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_andnot_pd() 443 pub unsafe fn _mm256_mask_or_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_or_pd() 456 pub unsafe fn _mm256_maskz_or_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_or_pd() 640 pub unsafe fn _mm256_mask_xor_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_xor_pd() 653 pub unsafe fn _mm256_maskz_xor_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_xor_pd() 922 pub unsafe fn _mm256_broadcast_f64x2(a: __m128d) -> __m256d { in _mm256_broadcast_f64x2() 934 pub unsafe fn _mm256_mask_broadcast_f64x2(src: __m256d, k: __mmask8, a: __m128d) -> __m256d { in _mm256_mask_broadcast_f64x2() [all …]
|
| D | avx512f.rs | 523 pub unsafe fn _mm256_mask_mov_pd(src: __m256d, k: __mmask8, a: __m256d) -> __m256d { in _mm256_mask_mov_pd() 535 pub unsafe fn _mm256_maskz_mov_pd(k: __mmask8, a: __m256d) -> __m256d { in _mm256_maskz_mov_pd() 867 pub unsafe fn _mm256_mask_add_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_add_pd() 879 pub unsafe fn _mm256_maskz_add_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_add_pd() 1211 pub unsafe fn _mm256_mask_sub_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_sub_pd() 1223 pub unsafe fn _mm256_maskz_sub_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_sub_pd() 1686 pub unsafe fn _mm256_mask_mul_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_mul_pd() 1698 pub unsafe fn _mm256_maskz_mul_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_mul_pd() 1858 pub unsafe fn _mm256_mask_div_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_div_pd() 1870 pub unsafe fn _mm256_maskz_div_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_div_pd() [all …]
|
| D | avx2.rs | 586 pub unsafe fn _mm256_broadcastsd_pd(a: __m128d) -> __m256d { in _mm256_broadcastsd_pd() 1328 ) -> __m256d { in _mm256_i32gather_pd() 1348 pub unsafe fn _mm256_mask_i32gather_pd<const SCALE: i32>( in _mm256_mask_i32gather_pd() 1352 mask: __m256d, in _mm256_mask_i32gather_pd() 1353 ) -> __m256d { in _mm256_mask_i32gather_pd() 1695 ) -> __m256d { in _mm256_i64gather_pd() 1715 pub unsafe fn _mm256_mask_i64gather_pd<const SCALE: i32>( in _mm256_mask_i64gather_pd() 1719 mask: __m256d, in _mm256_mask_i64gather_pd() 1720 ) -> __m256d { in _mm256_mask_i64gather_pd() 2304 pub unsafe fn _mm256_permute4x64_pd<const IMM8: i32>(a: __m256d) -> __m256d { in _mm256_permute4x64_pd() [all …]
|
| /prebuilts/rust/linux-musl-x86/1.83.0/lib/rustlib/src/rust/library/stdarch/crates/core_arch/src/x86/ |
| D | fma.rs | 48 pub unsafe fn _mm256_fmadd_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fmadd_pd() 136 pub unsafe fn _mm256_fmaddsub_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fmaddsub_pd() 192 pub unsafe fn _mm256_fmsub_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fmsub_pd() 280 pub unsafe fn _mm256_fmsubadd_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fmsubadd_pd() 336 pub unsafe fn _mm256_fnmadd_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fnmadd_pd() 422 pub unsafe fn _mm256_fnmsub_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fnmsub_pd()
|
| D | avx.rs | 33 pub unsafe fn _mm256_add_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_add_pd() 58 pub unsafe fn _mm256_and_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_and_pd() 87 pub unsafe fn _mm256_or_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_or_pd() 116 pub unsafe fn _mm256_shuffle_pd<const MASK: i32>(a: __m256d, b: __m256d) -> __m256d { in _mm256_shuffle_pd() 165 pub unsafe fn _mm256_andnot_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_andnot_pd() 194 pub unsafe fn _mm256_max_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_max_pd() 218 pub unsafe fn _mm256_min_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_min_pd() 242 pub unsafe fn _mm256_mul_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_mul_pd() 266 pub unsafe fn _mm256_addsub_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_addsub_pd() 298 pub unsafe fn _mm256_sub_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_sub_pd() [all …]
|
| D | test.rs | 63 pub unsafe fn assert_eq_m256d(a: __m256d, b: __m256d) { in assert_eq_m256d() 71 pub unsafe fn get_m256d(a: __m256d, idx: usize) -> f64 { in get_m256d()
|
| D | avx512dq.rs | 46 pub unsafe fn _mm256_mask_and_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_and_pd() 59 pub unsafe fn _mm256_maskz_and_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_and_pd() 244 pub unsafe fn _mm256_mask_andnot_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_andnot_pd() 258 pub unsafe fn _mm256_maskz_andnot_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_andnot_pd() 443 pub unsafe fn _mm256_mask_or_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_or_pd() 456 pub unsafe fn _mm256_maskz_or_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_or_pd() 640 pub unsafe fn _mm256_mask_xor_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_xor_pd() 653 pub unsafe fn _mm256_maskz_xor_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_xor_pd() 922 pub unsafe fn _mm256_broadcast_f64x2(a: __m128d) -> __m256d { in _mm256_broadcast_f64x2() 934 pub unsafe fn _mm256_mask_broadcast_f64x2(src: __m256d, k: __mmask8, a: __m128d) -> __m256d { in _mm256_mask_broadcast_f64x2() [all …]
|
| D | avx512f.rs | 523 pub unsafe fn _mm256_mask_mov_pd(src: __m256d, k: __mmask8, a: __m256d) -> __m256d { in _mm256_mask_mov_pd() 535 pub unsafe fn _mm256_maskz_mov_pd(k: __mmask8, a: __m256d) -> __m256d { in _mm256_maskz_mov_pd() 867 pub unsafe fn _mm256_mask_add_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_add_pd() 879 pub unsafe fn _mm256_maskz_add_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_add_pd() 1211 pub unsafe fn _mm256_mask_sub_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_sub_pd() 1223 pub unsafe fn _mm256_maskz_sub_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_sub_pd() 1686 pub unsafe fn _mm256_mask_mul_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_mul_pd() 1698 pub unsafe fn _mm256_maskz_mul_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_mul_pd() 1858 pub unsafe fn _mm256_mask_div_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_div_pd() 1870 pub unsafe fn _mm256_maskz_div_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_div_pd() [all …]
|
| D | avx2.rs | 586 pub unsafe fn _mm256_broadcastsd_pd(a: __m128d) -> __m256d { in _mm256_broadcastsd_pd() 1328 ) -> __m256d { in _mm256_i32gather_pd() 1348 pub unsafe fn _mm256_mask_i32gather_pd<const SCALE: i32>( in _mm256_mask_i32gather_pd() 1352 mask: __m256d, in _mm256_mask_i32gather_pd() 1353 ) -> __m256d { in _mm256_mask_i32gather_pd() 1695 ) -> __m256d { in _mm256_i64gather_pd() 1715 pub unsafe fn _mm256_mask_i64gather_pd<const SCALE: i32>( in _mm256_mask_i64gather_pd() 1719 mask: __m256d, in _mm256_mask_i64gather_pd() 1720 ) -> __m256d { in _mm256_mask_i64gather_pd() 2304 pub unsafe fn _mm256_permute4x64_pd<const IMM8: i32>(a: __m256d) -> __m256d { in _mm256_permute4x64_pd() [all …]
|
| /prebuilts/rust/linux-x86/1.83.0/lib/rustlib/src/rust/library/stdarch/crates/core_arch/src/x86/ |
| D | fma.rs | 48 pub unsafe fn _mm256_fmadd_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fmadd_pd() 136 pub unsafe fn _mm256_fmaddsub_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fmaddsub_pd() 192 pub unsafe fn _mm256_fmsub_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fmsub_pd() 280 pub unsafe fn _mm256_fmsubadd_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fmsubadd_pd() 336 pub unsafe fn _mm256_fnmadd_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fnmadd_pd() 422 pub unsafe fn _mm256_fnmsub_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fnmsub_pd()
|
| D | avx.rs | 33 pub unsafe fn _mm256_add_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_add_pd() 58 pub unsafe fn _mm256_and_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_and_pd() 87 pub unsafe fn _mm256_or_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_or_pd() 116 pub unsafe fn _mm256_shuffle_pd<const MASK: i32>(a: __m256d, b: __m256d) -> __m256d { in _mm256_shuffle_pd() 165 pub unsafe fn _mm256_andnot_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_andnot_pd() 194 pub unsafe fn _mm256_max_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_max_pd() 218 pub unsafe fn _mm256_min_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_min_pd() 242 pub unsafe fn _mm256_mul_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_mul_pd() 266 pub unsafe fn _mm256_addsub_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_addsub_pd() 298 pub unsafe fn _mm256_sub_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_sub_pd() [all …]
|
| D | test.rs | 63 pub unsafe fn assert_eq_m256d(a: __m256d, b: __m256d) { in assert_eq_m256d() 71 pub unsafe fn get_m256d(a: __m256d, idx: usize) -> f64 { in get_m256d()
|
| D | avx512dq.rs | 46 pub unsafe fn _mm256_mask_and_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_and_pd() 59 pub unsafe fn _mm256_maskz_and_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_and_pd() 244 pub unsafe fn _mm256_mask_andnot_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_andnot_pd() 258 pub unsafe fn _mm256_maskz_andnot_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_andnot_pd() 443 pub unsafe fn _mm256_mask_or_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_or_pd() 456 pub unsafe fn _mm256_maskz_or_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_or_pd() 640 pub unsafe fn _mm256_mask_xor_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_xor_pd() 653 pub unsafe fn _mm256_maskz_xor_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_xor_pd() 922 pub unsafe fn _mm256_broadcast_f64x2(a: __m128d) -> __m256d { in _mm256_broadcast_f64x2() 934 pub unsafe fn _mm256_mask_broadcast_f64x2(src: __m256d, k: __mmask8, a: __m128d) -> __m256d { in _mm256_mask_broadcast_f64x2() [all …]
|
| D | avx512f.rs | 523 pub unsafe fn _mm256_mask_mov_pd(src: __m256d, k: __mmask8, a: __m256d) -> __m256d { in _mm256_mask_mov_pd() 535 pub unsafe fn _mm256_maskz_mov_pd(k: __mmask8, a: __m256d) -> __m256d { in _mm256_maskz_mov_pd() 867 pub unsafe fn _mm256_mask_add_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_add_pd() 879 pub unsafe fn _mm256_maskz_add_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_add_pd() 1211 pub unsafe fn _mm256_mask_sub_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_sub_pd() 1223 pub unsafe fn _mm256_maskz_sub_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_sub_pd() 1686 pub unsafe fn _mm256_mask_mul_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_mul_pd() 1698 pub unsafe fn _mm256_maskz_mul_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_mul_pd() 1858 pub unsafe fn _mm256_mask_div_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_div_pd() 1870 pub unsafe fn _mm256_maskz_div_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_div_pd() [all …]
|
| D | avx2.rs | 586 pub unsafe fn _mm256_broadcastsd_pd(a: __m128d) -> __m256d { in _mm256_broadcastsd_pd() 1328 ) -> __m256d { in _mm256_i32gather_pd() 1348 pub unsafe fn _mm256_mask_i32gather_pd<const SCALE: i32>( in _mm256_mask_i32gather_pd() 1352 mask: __m256d, in _mm256_mask_i32gather_pd() 1353 ) -> __m256d { in _mm256_mask_i32gather_pd() 1695 ) -> __m256d { in _mm256_i64gather_pd() 1715 pub unsafe fn _mm256_mask_i64gather_pd<const SCALE: i32>( in _mm256_mask_i64gather_pd() 1719 mask: __m256d, in _mm256_mask_i64gather_pd() 1720 ) -> __m256d { in _mm256_mask_i64gather_pd() 2304 pub unsafe fn _mm256_permute4x64_pd<const IMM8: i32>(a: __m256d) -> __m256d { in _mm256_permute4x64_pd() [all …]
|
| D | mod.rs | 672 fn as_m256d(self) -> __m256d; in as_m256d() 680 impl m256dExt for __m256d { implementation
|
| /prebuilts/rust/linux-musl-x86/1.82.0/lib/rustlib/src/rust/library/stdarch/crates/core_arch/src/x86/ |
| D | fma.rs | 48 pub unsafe fn _mm256_fmadd_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fmadd_pd() 136 pub unsafe fn _mm256_fmaddsub_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fmaddsub_pd() 192 pub unsafe fn _mm256_fmsub_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fmsub_pd() 280 pub unsafe fn _mm256_fmsubadd_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fmsubadd_pd() 336 pub unsafe fn _mm256_fnmadd_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fnmadd_pd() 422 pub unsafe fn _mm256_fnmsub_pd(a: __m256d, b: __m256d, c: __m256d) -> __m256d { in _mm256_fnmsub_pd()
|
| D | avx.rs | 33 pub unsafe fn _mm256_add_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_add_pd() 58 pub unsafe fn _mm256_and_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_and_pd() 87 pub unsafe fn _mm256_or_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_or_pd() 116 pub unsafe fn _mm256_shuffle_pd<const MASK: i32>(a: __m256d, b: __m256d) -> __m256d { in _mm256_shuffle_pd() 165 pub unsafe fn _mm256_andnot_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_andnot_pd() 194 pub unsafe fn _mm256_max_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_max_pd() 218 pub unsafe fn _mm256_min_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_min_pd() 242 pub unsafe fn _mm256_mul_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_mul_pd() 266 pub unsafe fn _mm256_addsub_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_addsub_pd() 298 pub unsafe fn _mm256_sub_pd(a: __m256d, b: __m256d) -> __m256d { in _mm256_sub_pd() [all …]
|
| D | test.rs | 63 pub unsafe fn assert_eq_m256d(a: __m256d, b: __m256d) { in assert_eq_m256d() 71 pub unsafe fn get_m256d(a: __m256d, idx: usize) -> f64 { in get_m256d()
|
| D | avx512dq.rs | 46 pub unsafe fn _mm256_mask_and_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_and_pd() 59 pub unsafe fn _mm256_maskz_and_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_and_pd() 244 pub unsafe fn _mm256_mask_andnot_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_andnot_pd() 258 pub unsafe fn _mm256_maskz_andnot_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_andnot_pd() 443 pub unsafe fn _mm256_mask_or_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_or_pd() 456 pub unsafe fn _mm256_maskz_or_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_or_pd() 640 pub unsafe fn _mm256_mask_xor_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_xor_pd() 653 pub unsafe fn _mm256_maskz_xor_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_xor_pd() 922 pub unsafe fn _mm256_broadcast_f64x2(a: __m128d) -> __m256d { in _mm256_broadcast_f64x2() 934 pub unsafe fn _mm256_mask_broadcast_f64x2(src: __m256d, k: __mmask8, a: __m128d) -> __m256d { in _mm256_mask_broadcast_f64x2() [all …]
|
| D | avx512f.rs | 523 pub unsafe fn _mm256_mask_mov_pd(src: __m256d, k: __mmask8, a: __m256d) -> __m256d { in _mm256_mask_mov_pd() 535 pub unsafe fn _mm256_maskz_mov_pd(k: __mmask8, a: __m256d) -> __m256d { in _mm256_maskz_mov_pd() 867 pub unsafe fn _mm256_mask_add_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_add_pd() 879 pub unsafe fn _mm256_maskz_add_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_add_pd() 1211 pub unsafe fn _mm256_mask_sub_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_sub_pd() 1223 pub unsafe fn _mm256_maskz_sub_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_sub_pd() 1686 pub unsafe fn _mm256_mask_mul_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_mul_pd() 1698 pub unsafe fn _mm256_maskz_mul_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_mul_pd() 1858 pub unsafe fn _mm256_mask_div_pd(src: __m256d, k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_mask_div_pd() 1870 pub unsafe fn _mm256_maskz_div_pd(k: __mmask8, a: __m256d, b: __m256d) -> __m256d { in _mm256_maskz_div_pd() [all …]
|
| D | avx2.rs | 586 pub unsafe fn _mm256_broadcastsd_pd(a: __m128d) -> __m256d { in _mm256_broadcastsd_pd() 1328 ) -> __m256d { in _mm256_i32gather_pd() 1348 pub unsafe fn _mm256_mask_i32gather_pd<const SCALE: i32>( in _mm256_mask_i32gather_pd() 1352 mask: __m256d, in _mm256_mask_i32gather_pd() 1353 ) -> __m256d { in _mm256_mask_i32gather_pd() 1695 ) -> __m256d { in _mm256_i64gather_pd() 1715 pub unsafe fn _mm256_mask_i64gather_pd<const SCALE: i32>( in _mm256_mask_i64gather_pd() 1719 mask: __m256d, in _mm256_mask_i64gather_pd() 1720 ) -> __m256d { in _mm256_mask_i64gather_pd() 2304 pub unsafe fn _mm256_permute4x64_pd<const IMM8: i32>(a: __m256d) -> __m256d { in _mm256_permute4x64_pd() [all …]
|