/external/llvm-project/llvm/test/CodeGen/AArch64/ |
D | llvm-masked-gather-legal-for-sve.ll | 12 …d_gather_nxv4i32(<vscale x 4 x i32*> %ld, <vscale x 4 x i1> %masks, <vscale x 4 x i32> %passthru) { 13 …her.nxv4i32(<vscale x 4 x i32*> %ld, i32 0, <vscale x 4 x i1> %masks, <vscale x 4 x i32> %passthru) 22 …er_nxv2f64(<vscale x 2 x double*> %ld, <vscale x 2 x i1> %masks, <vscale x 2 x double> %passthru) { 23 …v2f64(<vscale x 2 x double*> %ld, i32 0, <vscale x 2 x i1> %masks, <vscale x 2 x double> %passthru) 32 …gather_nxv2f16(<vscale x 2 x half*> %ld, <vscale x 2 x i1> %masks, <vscale x 2 x half> %passthru) { 33 …r.nxv2f16(<vscale x 2 x half*> %ld, i32 0, <vscale x 2 x i1> %masks, <vscale x 2 x half> %passthru) 43 define <2 x float> @masked_gather_v2f32(<2 x float*> %ld, <2 x i1> %masks, <2 x float> %passthru) { 44 …x float> @llvm.masked.gather.v2f32(<2 x float*> %ld, i32 0, <2 x i1> %masks, <2 x float> %passthru) 54 define <4 x i32> @masked_gather_v4i32(<4 x i32*> %ld, <4 x i1> %masks, <4 x i32> %passthru) { 55 …ll <4 x i32> @llvm.masked.gather.v4i32(<4 x i32*> %ld, i32 0, <4 x i1> %masks, <4 x i32> %passthru) [all …]
|
/external/llvm/test/Transforms/InstCombine/ |
D | masked_intrinsics.ll | 5 …double> @llvm.masked.gather.v2f64(<2 x double*> %ptrs, i32, <2 x i1> %mask, <2 x double> %passthru) 8 define <2 x double> @load_zeromask(<2 x double>* %ptr, <2 x double> %passthru) { 9 …ked.load.v2f64.p0v2f64(<2 x double>* %ptr, i32 1, <2 x i1> zeroinitializer, <2 x double> %passthru) 13 ; CHECK-NEXT: ret <2 x double> %passthru 16 define <2 x double> @load_onemask(<2 x double>* %ptr, <2 x double> %passthru) { 17 …masked.load.v2f64.p0v2f64(<2 x double>* %ptr, i32 2, <2 x i1> <i1 1, i1 1>, <2 x double> %passthru) 25 define <2 x double> @load_undefmask(<2 x double>* %ptr, <2 x double> %passthru) { 26 …ed.load.v2f64.p0v2f64(<2 x double>* %ptr, i32 2, <2 x i1> <i1 1, i1 undef>, <2 x double> %passthru) 51 define <2 x double> @gather_zeromask(<2 x double*> %ptrs, <2 x double> %passthru) { 52 …m.masked.gather.v2f64(<2 x double*> %ptrs, i32 5, <2 x i1> zeroinitializer, <2 x double> %passthru) [all …]
|
/external/llvm-project/llvm/test/Verifier/ |
D | scatter_gather.ll | 5 define <16 x float> @gather2(<16 x float*> %ptrs, <16 x i1>* %mask, <16 x float> %passthru) { 6 …masked.gather.v16f32.v16p0f32(<16 x float*> %ptrs, i32 4, <16 x i1>* %mask, <16 x float> %passthru) 13 define <8 x float> @gather3(<8 x float*> %ptrs, <16 x i1> %mask, <8 x float> %passthru) { 14 …llvm.masked.gather.v8f32.v8p0f32(<8 x float*> %ptrs, i32 4, <16 x i1> %mask, <8 x float> %passthru) 21 define <8 x float>* @gather4(<8 x float*> %ptrs, <8 x i1> %mask, <8 x float> %passthru) { 22 …lvm.masked.gather.p0v8f32.v8p0f32(<8 x float*> %ptrs, i32 4, <8 x i1> %mask, <8 x float> %passthru) 29 define <8 x float> @gather5(<8 x float*>* %ptrs, <8 x i1> %mask, <8 x float> %passthru) { 30 …vm.masked.gather.v8f32.p0v8p0f32(<8 x float*>* %ptrs, i32 4, <8 x i1> %mask, <8 x float> %passthru) 37 define <8 x float> @gather6(<8 x float> %ptrs, <8 x i1> %mask, <8 x float> %passthru) { 38 …t> @llvm.masked.gather.v8f32.v8f32(<8 x float> %ptrs, i32 4, <8 x i1> %mask, <8 x float> %passthru) [all …]
|
/external/llvm-project/llvm/test/CodeGen/X86/ |
D | vector-shuffle-masked.ll | 4 define <4 x i32> @mask_shuffle_v4i32_1234(<4 x i32> %a, <4 x i32> %b, <4 x i32> %passthru, i8 %mask… 14 %res = select <4 x i1> %mask.extract, <4 x i32> %shuffle, <4 x i32> %passthru 31 define <4 x i32> @mask_shuffle_v4i32_2345(<4 x i32> %a, <4 x i32> %b, <4 x i32> %passthru, i8 %mask… 41 %res = select <4 x i1> %mask.extract, <4 x i32> %shuffle, <4 x i32> %passthru 58 define <2 x i64> @mask_shuffle_v2i64_12(<2 x i64> %a, <2 x i64> %b, <2 x i64> %passthru, i8 %mask) { 68 %res = select <2 x i1> %mask.extract, <2 x i64> %shuffle, <2 x i64> %passthru 85 define <4 x i64> @mask_shuffle_v4i64_1234(<4 x i64> %a, <4 x i64> %b, <4 x i64> %passthru, i8 %mask… 95 %res = select <4 x i1> %mask.extract, <4 x i64> %shuffle, <4 x i64> %passthru 112 define <4 x i64> @mask_shuffle_v4i64_1230(<4 x i64> %a, <4 x i64> %passthru, i8 %mask) { 122 %res = select <4 x i1> %mask.extract, <4 x i64> %shuffle, <4 x i64> %passthru [all …]
|
D | avx512-rndscale.ll | 149 define <2 x double> @floor_v2f64_mask(<2 x double> %p, <2 x double> %passthru, <2 x i64> %cmp) { 158 %s = select <2 x i1> %c, <2 x double> %t, <2 x double> %passthru 162 define <4 x float> @floor_v4f32_mask(<4 x float> %p, <4 x float> %passthru, <4 x i32> %cmp) { 171 %s = select <4 x i1> %c, <4 x float> %t, <4 x float> %passthru 175 define <4 x double> @floor_v4f64_mask(<4 x double> %p, <4 x double> %passthru, <4 x i64> %cmp) { 184 %s = select <4 x i1> %c, <4 x double> %t, <4 x double> %passthru 188 define <8 x float> @floor_v8f32_mask(<8 x float> %p, <8 x float> %passthru, <8 x i32> %cmp) { 197 %s = select <8 x i1> %c, <8 x float> %t, <8 x float> %passthru 201 define <8 x double> @floor_v8f64_mask(<8 x double> %p, <8 x double> %passthru, <8 x i64> %cmp) { 210 %s = select <8 x i1> %c, <8 x double> %t, <8 x double> %passthru [all …]
|
D | avx512-gfni-intrinsics.ll | 6 define <16 x i8> @test_vgf2p8affineinvqb_128(<16 x i8> %src1, <16 x i8> %src2, <16 x i8> %passthru,… 31 %6 = select <16 x i1> %1, <16 x i8> %4, <16 x i8> %passthru 38 define <32 x i8> @test_vgf2p8affineinvqb_256(<32 x i8> %src1, <32 x i8> %src2, <32 x i8> %passthru,… 63 %6 = select <32 x i1> %1, <32 x i8> %4, <32 x i8> %passthru 70 define <64 x i8> @test_vgf2p8affineinvqb_512(<64 x i8> %src1, <64 x i8> %src2, <64 x i8> %passthru,… 95 %6 = select <64 x i1> %1, <64 x i8> %4, <64 x i8> %passthru 102 define <16 x i8> @test_vgf2p8affineqb_128(<16 x i8> %src1, <16 x i8> %src2, <16 x i8> %passthru, i1… 127 %6 = select <16 x i1> %1, <16 x i8> %4, <16 x i8> %passthru 134 define <32 x i8> @test_vgf2p8affineqb_256(<32 x i8> %src1, <32 x i8> %src2, <32 x i8> %passthru, i3… 159 %6 = select <32 x i1> %1, <32 x i8> %4, <32 x i8> %passthru [all …]
|
D | stack-folding-avx512bf16.ll | 27 …@stack_fold_cvtne2ps2bf16_mask(<16 x float> %a0, <16 x float> %a1, <32 x i16>* %passthru, i32 %U) { 43 %4 = load <32 x i16>, <32 x i16>* %passthru 80 define <16 x i16> @stack_fold_cvtneps2bf16_mask(<16 x float> %a0, <16 x i16>* %passthru, i16 %U) { 96 %4 = load <16 x i16>, <16 x i16>* %passthru 133 …dpbf16ps_mask(<16 x float>* %a0, <16 x i32> %a1, <16 x i32> %a2, <16 x float>* %passthru, i16 %U) { 189 …tack_fold_cvtne2ps2bf16_mask_ymm(<8 x float> %a0, <8 x float> %a1, <16 x i16>* %passthru, i16 %U) { 205 %4 = load <16 x i16>, <16 x i16>* %passthru 243 define <8 x i16> @stack_fold_cvtneps2bf16_mask_ymm(<8 x float> %a0, <8 x i16>* %passthru, i8 %U) { 260 %4 = load <8 x i16>, <8 x i16>* %passthru 298 …vdpbf16ps_mask_ymm(<8 x float>* %a0, <8 x i32> %a1, <8 x i32> %a2, <8 x float>* %passthru, i8 %U) { [all …]
|
D | stack-folding-fp-avx512.ll | 26 …e> @stack_fold_addpd_zmm_k(<8 x double> %a0, <8 x double> %a1, i8 %mask, <8 x double>* %passthru) { 41 %4 = load <8 x double>, <8 x double>* %passthru 46 …_fold_addpd_zmm_k_commuted(<8 x double> %a0, <8 x double> %a1, i8 %mask, <8 x double>* %passthru) { 61 %4 = load <8 x double>, <8 x double>* %passthru 97 …> @stack_fold_addps_zmm_k(<16 x float> %a0, <16 x float> %a1, i16 %mask, <16 x float>* %passthru) { 112 %4 = load <16 x float>, <16 x float>* %passthru 117 …fold_addps_zmm_k_commuted(<16 x float> %a0, <16 x float> %a1, i16 %mask, <16 x float>* %passthru) { 132 %4 = load <16 x float>, <16 x float>* %passthru 637 …_maxpd_zmm_commutable_k(<8 x double> %a0, <8 x double> %a1, i8 %mask, <8 x double>* %passthru) #1 { 652 %4 = load <8 x double>, <8 x double>* %passthru [all …]
|
D | avx512-cvt.ll | 776 …@f32to4f64_mask_load(<4 x float>* %p, <4 x double> %b1, <4 x double> %a1, <4 x double> %passthru) { 797 %c = select <4 x i1> %mask, <4 x double> %a, <4 x double> %passthru 2038 define <2 x i64> @test_2f64toub(<2 x double> %a, <2 x i64> %passthru) { 2079 %select = select <2 x i1> %mask, <2 x i64> %passthru, <2 x i64> zeroinitializer 2083 define <4 x i64> @test_4f64toub(<4 x double> %a, <4 x i64> %passthru) { 2120 %select = select <4 x i1> %mask, <4 x i64> %passthru, <4 x i64> zeroinitializer 2124 define <8 x i64> @test_8f64toub(<8 x double> %a, <8 x i64> %passthru) { 2157 %select = select <8 x i1> %mask, <8 x i64> %passthru, <8 x i64> zeroinitializer 2161 define <2 x i64> @test_2f32toub(<2 x float> %a, <2 x i64> %passthru) { 2200 %select = select <2 x i1> %mask, <2 x i64> %passthru, <2 x i64> zeroinitializer [all …]
|
/external/llvm-project/llvm/test/Transforms/InstCombine/ |
D | pr44541.ll | 6 define i16 @passthru(i16 returned %x) { 7 ; CHECK-LABEL: @passthru( 15 ; CHECK-NEXT: [[ZERO:%.*]] = call i16 @passthru(i16 0) 20 %zero = call i16 @passthru(i16 0)
|
D | masked_intrinsics.ll | 6 …@llvm.masked.gather.v2f64.v2p0f64(<2 x double*> %ptrs, i32, <2 x i1> %mask, <2 x double> %passthru) 7 …@llvm.masked.gather.v4f64.v4p0f64(<4 x double*> %ptrs, i32, <4 x i1> %mask, <4 x double> %passthru) 10 define <2 x double> @load_zeromask(<2 x double>* %ptr, <2 x double> %passthru) { 14 …ked.load.v2f64.p0v2f64(<2 x double>* %ptr, i32 1, <2 x i1> zeroinitializer, <2 x double> %passthru) 18 define <2 x double> @load_onemask(<2 x double>* %ptr, <2 x double> %passthru) { 23 …masked.load.v2f64.p0v2f64(<2 x double>* %ptr, i32 2, <2 x i1> <i1 1, i1 1>, <2 x double> %passthru) 27 define <2 x double> @load_undefmask(<2 x double>* %ptr, <2 x double> %passthru) { 32 …ed.load.v2f64.p0v2f64(<2 x double>* %ptr, i32 2, <2 x i1> <i1 1, i1 undef>, <2 x double> %passthru) 38 define <2 x double> @load_cemask(<2 x double>* %ptr, <2 x double> %passthru) { 43 …f64(<2 x double>* %ptr, i32 2, <2 x i1> <i1 1, i1 ptrtoint (i8* @G to i1)>, <2 x double> %passthru) [all …]
|
D | debuginfo.ll | 16 define i8* @passthru(i8* %a, i32 %b, i64 %c) !dbg !1 { 36 ; CHECK-LABEL: define i8* @passthru(i8* %a, i32 %b, i64 %c) 87 !1 = distinct !DISubprogram(name: "passthru", line: 79, isLocal: true, isDefinition: true, virtualI…
|
/external/llvm-project/llvm/test/Transforms/ScalarizeMaskedMemIntrin/X86/ |
D | expand-masked-gather.ll | 4 define <2 x i64> @scalarize_v2i64(<2 x i64*> %p, <2 x i1> %mask, <2 x i64> %passthru) { 29 …x i64> @llvm.masked.gather.v2i64.v2p0i64(<2 x i64*> %p, i32 8, <2 x i1> %mask, <2 x i64> %passthru) 33 define <2 x i64> @scalarize_v2i64_ones_mask(<2 x i64*> %p, <2 x i64> %passthru) { 43 …masked.gather.v2i64.v2p0i64(<2 x i64*> %p, i32 8, <2 x i1> <i1 true, i1 true>, <2 x i64> %passthru) 47 define <2 x i64> @scalarize_v2i64_zero_mask(<2 x i64*> %p, <2 x i64> %passthru) { 51 …sked.gather.v2i64.v2p0i64(<2 x i64*> %p, i32 8, <2 x i1> <i1 false, i1 false>, <2 x i64> %passthru) 55 define <2 x i64> @scalarize_v2i64_const_mask(<2 x i64*> %p, <2 x i64> %passthru) { 62 …asked.gather.v2i64.v2p0i64(<2 x i64*> %p, i32 8, <2 x i1> <i1 false, i1 true>, <2 x i64> %passthru)
|
D | expand-masked-expandload.ll | 4 define <2 x i64> @scalarize_v2i64(i64* %p, <2 x i1> %mask, <2 x i64> %passthru) { 29 … call <2 x i64> @llvm.masked.expandload.v2i64.p0v2i64(i64* %p, <2 x i1> %mask, <2 x i64> %passthru) 33 define <2 x i64> @scalarize_v2i64_ones_mask(i64* %p, <2 x i64> %passthru) { 44 …4> @llvm.masked.expandload.v2i64.p0v2i64(i64* %p, <2 x i1> <i1 true, i1 true>, <2 x i64> %passthru) 48 define <2 x i64> @scalarize_v2i64_zero_mask(i64* %p, <2 x i64> %passthru) { 53 … @llvm.masked.expandload.v2i64.p0v2i64(i64* %p, <2 x i1> <i1 false, i1 false>, <2 x i64> %passthru) 57 define <2 x i64> @scalarize_v2i64_const_mask(i64* %p, <2 x i64> %passthru) { 65 …> @llvm.masked.expandload.v2i64.p0v2i64(i64* %p, <2 x i1> <i1 false, i1 true>, <2 x i64> %passthru)
|
D | expand-masked-load.ll | 4 define <2 x i64> @scalarize_v2i64(<2 x i64>* %p, <2 x i1> %mask, <2 x i64> %passthru) { 30 …x i64> @llvm.masked.load.v2i64.p0v2i64(<2 x i64>* %p, i32 128, <2 x i1> %mask, <2 x i64> %passthru) 34 define <2 x i64> @scalarize_v2i64_ones_mask(<2 x i64>* %p, <2 x i64> %passthru) { 39 …m.masked.load.v2i64.p0v2i64(<2 x i64>* %p, i32 8, <2 x i1> <i1 true, i1 true>, <2 x i64> %passthru) 43 define <2 x i64> @scalarize_v2i64_zero_mask(<2 x i64>* %p, <2 x i64> %passthru) { 48 …masked.load.v2i64.p0v2i64(<2 x i64>* %p, i32 8, <2 x i1> <i1 false, i1 false>, <2 x i64> %passthru) 52 define <2 x i64> @scalarize_v2i64_const_mask(<2 x i64>* %p, <2 x i64> %passthru) { 60 ….masked.load.v2i64.p0v2i64(<2 x i64>* %p, i32 8, <2 x i1> <i1 false, i1 true>, <2 x i64> %passthru) 65 define <2 x i24> @scalarize_v2i24(<2 x i24>* %p, <2 x i1> %mask, <2 x i24> %passthru) { 91 …2 x i24> @llvm.masked.load.v2i24.p0v2i24(<2 x i24>* %p, i32 8, <2 x i1> %mask, <2 x i24> %passthru) [all …]
|
/external/llvm-project/llvm/test/Transforms/ScalarizeMaskedMemIntrin/AArch64/ |
D | expand-masked-load.ll | 5 define <2 x i64> @scalarize_v2i64(<2 x i64>* %p, <2 x i1> %mask, <2 x i64> %passthru) { 31 …x i64> @llvm.masked.load.v2i64.p0v2i64(<2 x i64>* %p, i32 128, <2 x i1> %mask, <2 x i64> %passthru) 35 define <2 x i64> @scalarize_v2i64_ones_mask(<2 x i64>* %p, <2 x i64> %passthru) { 40 …m.masked.load.v2i64.p0v2i64(<2 x i64>* %p, i32 8, <2 x i1> <i1 true, i1 true>, <2 x i64> %passthru) 44 define <2 x i64> @scalarize_v2i64_zero_mask(<2 x i64>* %p, <2 x i64> %passthru) { 49 …masked.load.v2i64.p0v2i64(<2 x i64>* %p, i32 8, <2 x i1> <i1 false, i1 false>, <2 x i64> %passthru) 53 define <2 x i64> @scalarize_v2i64_const_mask(<2 x i64>* %p, <2 x i64> %passthru) { 61 ….masked.load.v2i64.p0v2i64(<2 x i64>* %p, i32 8, <2 x i1> <i1 false, i1 true>, <2 x i64> %passthru) 66 define <2 x i24> @scalarize_v2i24(<2 x i24>* %p, <2 x i1> %mask, <2 x i24> %passthru) { 92 …2 x i24> @llvm.masked.load.v2i24.p0v2i24(<2 x i24>* %p, i32 8, <2 x i1> %mask, <2 x i24> %passthru) [all …]
|
/external/libwebsockets/plugins/generic-sessions/ |
D | protocol_lws_messageboard.c | 219 goto passthru; in callback_messageboard() 300 goto passthru; in callback_messageboard() 304 goto passthru; in callback_messageboard() 324 goto passthru; in callback_messageboard() 336 goto passthru; in callback_messageboard() 390 passthru: in callback_messageboard()
|
/external/llvm-project/llvm/test/Transforms/InstCombine/X86/ |
D | x86-avx512.ll | 1765 define <8 x i32> @identity_test_permvar_si_256_mask(<8 x i32> %a0, <8 x i32> %passthru, i8 %mask) { 1773 %3 = select <8 x i1> %2, <8 x i32> %1, <8 x i32> %passthru 1786 define <8 x i32> @zero_test_permvar_si_256_mask(<8 x i32> %a0, <8 x i32> %passthru, i8 %mask) { 1795 %3 = select <8 x i1> %2, <8 x i32> %1, <8 x i32> %passthru 1808 define <8 x i32> @shuffle_test_permvar_si_256_mask(<8 x i32> %a0, <8 x i32> %passthru, i8 %mask) { 1817 %3 = select <8 x i1> %2, <8 x i32> %1, <8 x i32> %passthru 1830 define <8 x i32> @undef_test_permvar_si_256_mask(<8 x i32> %a0, <8 x i32> %passthru, i8 %mask) { 1839 %3 = select <8 x i1> %2, <8 x i32> %1, <8 x i32> %passthru 1853 define <8 x float> @identity_test_permvar_sf_256_mask(<8 x float> %a0, <8 x float> %passthru, i8 %m… 1861 %3 = select <8 x i1> %2, <8 x float> %1, <8 x float> %passthru [all …]
|
/external/mesa3d/src/gallium/auxiliary/gallivm/ |
D | lp_bld_gather.c | 334 LLVMValueRef passthru = LLVMGetUndef(src_vec_type); in lp_build_gather_avx2() local 336 LLVMValueRef args[] = { src_ptr, alignment, mask, passthru }; in lp_build_gather_avx2() 370 LLVMValueRef passthru = LLVMGetUndef(src_vec_type); in lp_build_gather_avx2() local 375 LLVMValueRef args[] = { passthru, base_ptr, offsets, mask, scale }; in lp_build_gather_avx2()
|
/external/llvm-project/llvm/test/Assembler/ |
D | auto_upgrade_intrinsics.ll | 71 define <2 x double> @tests.masked.load(<2 x double>* %ptr, <2 x i1> %mask, <2 x double> %passthru) … 74 … double> @llvm.masked.load.v2f64(<2 x double>* %ptr, i32 1, <2 x i1> %mask, <2 x double> %passthru) 89 define <2 x double> @tests.masked.gather(<2 x double*> %ptr, <2 x i1> %mask, <2 x double> %passthru… 92 …ouble> @llvm.masked.gather.v2f64(<2 x double*> %ptr, i32 1, <2 x i1> %mask, <2 x double> %passthru)
|
/external/curl/lib/ |
D | http_chunks.h | 97 CURLcode *passthru);
|
/external/kernel-headers/original/uapi/sound/ |
D | hdsp.h | 68 unsigned char passthru; member
|
D | hdspm.h | 75 unsigned int passthru; member
|
/external/llvm-project/llvm/test/Transforms/HardwareLoops/ARM/ |
D | calls.ll | 319 …test_masked_i32(<4 x i1> %mask, <4 x i32>* %a, <4 x i32>* %b, <4 x i32>* %c, <4 x i32> %passthru) { 327 …32> @llvm.masked.load.v4i32.p0v4i32(<4 x i32>* %addr.a, i32 4, <4 x i1> %mask, <4 x i32> %passthru) 328 …32> @llvm.masked.load.v4i32.p0v4i32(<4 x i32>* %addr.b, i32 4, <4 x i1> %mask, <4 x i32> %passthru) 346 …ked_f32(<4 x i1> %mask, <4 x float>* %a, <4 x float>* %b, <4 x float>* %c, <4 x float> %passthru) { 354 …@llvm.masked.load.v4f32.p0v4f32(<4 x float>* %addr.a, i32 4, <4 x i1> %mask, <4 x float> %passthru) 355 …@llvm.masked.load.v4f32.p0v4f32(<4 x float>* %addr.b, i32 4, <4 x i1> %mask, <4 x float> %passthru) 373 …scatter(<4 x i1> %mask, <4 x float*> %a, <4 x float*> %b, <4 x float*> %c, <4 x float> %passthru) { 378 …t> @llvm.masked.gather.v4f32.p0v4f32(<4 x float*> %a, i32 4, <4 x i1> %mask, <4 x float> %passthru) 379 …t> @llvm.masked.gather.v4f32.p0v4f32(<4 x float*> %b, i32 4, <4 x i1> %mask, <4 x float> %passthru)
|
/external/llvm/test/Assembler/ |
D | auto_upgrade_intrinsics.ll | 63 define <2 x double> @tests.masked.load(<2 x double>* %ptr, <2 x i1> %mask, <2 x double> %passthru) … 66 … double> @llvm.masked.load.v2f64(<2 x double>* %ptr, i32 1, <2 x i1> %mask, <2 x double> %passthru)
|