/external/llvm/test/Assembler/ |
D | 2007-12-11-AddressSpaces.ll | 1 ; RUN: llvm-as < %s | llvm-dis | grep "addrspace(33)" | count 7 2 ; RUN: llvm-as < %s | llvm-dis | grep "addrspace(42)" | count 2 3 ; RUN: llvm-as < %s | llvm-dis | grep "addrspace(66)" | count 2 4 ; RUN: llvm-as < %s | llvm-dis | grep "addrspace(11)" | count 6 5 ; RUN: llvm-as < %s | llvm-dis | grep "addrspace(22)" | count 5 7 %struct.mystruct = type { i32, i32 addrspace(33)*, i32, i32 addrspace(33)* } 8 @input = weak addrspace(42) global %struct.mystruct zeroinitializer ; <%struct.mystruct addrspac… 9 @output = addrspace(66) global %struct.mystruct zeroinitializer ; <%struct.mystruct addrspace(66)… 10 @y = external addrspace(33) global i32 addrspace(11)* addrspace(22)* ; <i32 addrspace(11)* addrsp… 14 …p1 = load i32 addrspace(33)* addrspace(42)* getelementptr (%struct.mystruct addrspace(42)* @input,… [all …]
|
/external/llvm/test/CodeGen/XCore/ |
D | resources.ll | 3 declare i8 addrspace(1)* @llvm.xcore.getr.p1i8(i32 %type) 4 declare void @llvm.xcore.freer.p1i8(i8 addrspace(1)* %r) 5 declare i32 @llvm.xcore.in.p1i8(i8 addrspace(1)* %r) 6 declare i32 @llvm.xcore.int.p1i8(i8 addrspace(1)* %r) 7 declare i32 @llvm.xcore.inct.p1i8(i8 addrspace(1)* %r) 8 declare void @llvm.xcore.out.p1i8(i8 addrspace(1)* %r, i32 %value) 9 declare void @llvm.xcore.outt.p1i8(i8 addrspace(1)* %r, i32 %value) 10 declare void @llvm.xcore.outct.p1i8(i8 addrspace(1)* %r, i32 %value) 11 declare void @llvm.xcore.chkct.p1i8(i8 addrspace(1)* %r, i32 %value) 12 declare i32 @llvm.xcore.testct.p1i8(i8 addrspace(1)* %r) [all …]
|
D | threads.ll | 3 declare i8 addrspace(1)* @llvm.xcore.getst.p1i8.p1i8(i8 addrspace(1)* %r) 4 declare void @llvm.xcore.msync.p1i8(i8 addrspace(1)* %r) 6 declare void @llvm.xcore.mjoin.p1i8(i8 addrspace(1)* %r) 7 declare void @llvm.xcore.initsp.p1i8(i8 addrspace(1)* %r, i8* %value) 8 declare void @llvm.xcore.initpc.p1i8(i8 addrspace(1)* %r, i8* %value) 9 declare void @llvm.xcore.initlr.p1i8(i8 addrspace(1)* %r, i8* %value) 10 declare void @llvm.xcore.initcp.p1i8(i8 addrspace(1)* %r, i8* %value) 11 declare void @llvm.xcore.initdp.p1i8(i8 addrspace(1)* %r, i8* %value) 13 define i8 addrspace(1)* @getst(i8 addrspace(1)* %r) { 16 %result = call i8 addrspace(1)* @llvm.xcore.getst.p1i8.p1i8(i8 addrspace(1)* %r) [all …]
|
/external/llvm/test/CodeGen/R600/ |
D | kcache-fold.ll | 7 %0 = load <4 x float> addrspace(8)* null 9 …%2 = load <4 x float> addrspace(8)* getelementptr ([1024 x <4 x float>] addrspace(8)* null, i64 0,… 11 …%4 = load <4 x float> addrspace(8)* getelementptr ([1024 x <4 x float>] addrspace(8)* null, i64 0,… 15 %8 = load <4 x float> addrspace(8)* null 17 …%10 = load <4 x float> addrspace(8)* getelementptr ([1024 x <4 x float>] addrspace(8)* null, i64 0… 19 …%12 = load <4 x float> addrspace(8)* getelementptr ([1024 x <4 x float>] addrspace(8)* null, i64 0… 23 %16 = load <4 x float> addrspace(8)* null 25 …%18 = load <4 x float> addrspace(8)* getelementptr ([1024 x <4 x float>] addrspace(8)* null, i64 0… 27 …%20 = load <4 x float> addrspace(8)* getelementptr ([1024 x <4 x float>] addrspace(8)* null, i64 0… 31 %24 = load <4 x float> addrspace(8)* null [all …]
|
D | vec4-expand.ll | 9 define void @fp_to_sint(<4 x i32> addrspace(1)* %out, <4 x float> addrspace(1)* %in) { 10 %value = load <4 x float> addrspace(1) * %in 12 store <4 x i32> %result, <4 x i32> addrspace(1)* %out 22 define void @fp_to_uint(<4 x i32> addrspace(1)* %out, <4 x float> addrspace(1)* %in) { 23 %value = load <4 x float> addrspace(1) * %in 25 store <4 x i32> %result, <4 x i32> addrspace(1)* %out 35 define void @sint_to_fp(<4 x float> addrspace(1)* %out, <4 x i32> addrspace(1)* %in) { 36 %value = load <4 x i32> addrspace(1) * %in 38 store <4 x float> %result, <4 x float> addrspace(1)* %out 48 define void @uint_to_fp(<4 x float> addrspace(1)* %out, <4 x i32> addrspace(1)* %in) { [all …]
|
D | schedule-vs-if-nested-loop.ll | 24 %11 = load <4 x float> addrspace(9)* null 27 %14 = load <4 x float> addrspace(9)* null 30 %17 = load <4 x float> addrspace(9)* null 33 %20 = load <4 x float> addrspace(9)* null 36 …%23 = load <4 x float> addrspace(9)* getelementptr ([1024 x <4 x float>] addrspace(9)* null, i64 0… 40 …%27 = load <4 x float> addrspace(9)* getelementptr ([1024 x <4 x float>] addrspace(9)* null, i64 0… 44 …%31 = load <4 x float> addrspace(9)* getelementptr ([1024 x <4 x float>] addrspace(9)* null, i64 0… 48 …%35 = load <4 x float> addrspace(9)* getelementptr ([1024 x <4 x float>] addrspace(9)* null, i64 0… 52 …%39 = load <4 x float> addrspace(9)* getelementptr ([1024 x <4 x float>] addrspace(9)* null, i64 0… 56 …%43 = load <4 x float> addrspace(9)* getelementptr ([1024 x <4 x float>] addrspace(9)* null, i64 0… [all …]
|
D | set-dx10.ll | 9 define void @fcmp_une_select_fptosi(i32 addrspace(1)* %out, float %in) { 15 store i32 %3, i32 addrspace(1)* %out 21 define void @fcmp_une_select_i32(i32 addrspace(1)* %out, float %in) { 25 store i32 %1, i32 addrspace(1)* %out 31 define void @fcmp_ueq_select_fptosi(i32 addrspace(1)* %out, float %in) { 37 store i32 %3, i32 addrspace(1)* %out 43 define void @fcmp_ueq_select_i32(i32 addrspace(1)* %out, float %in) { 47 store i32 %1, i32 addrspace(1)* %out 53 define void @fcmp_ugt_select_fptosi(i32 addrspace(1)* %out, float %in) { 59 store i32 %3, i32 addrspace(1)* %out [all …]
|
D | unsupported-cc.ll | 7 define void @slt(i32 addrspace(1)* %out, i32 %in) { 11 store i32 %1, i32 addrspace(1)* %out 17 define void @ult_i32(i32 addrspace(1)* %out, i32 %in) { 21 store i32 %1, i32 addrspace(1)* %out 27 define void @ult_float(float addrspace(1)* %out, float %in) { 31 store float %1, float addrspace(1)* %out 37 define void @olt(float addrspace(1)* %out, float %in) { 41 store float %1, float addrspace(1)* %out 47 define void @sle(i32 addrspace(1)* %out, i32 %in) { 51 store i32 %1, i32 addrspace(1)* %out [all …]
|
D | fcmp.ll | 6 define void @fcmp_sext(i32 addrspace(1)* %out, float addrspace(1)* %in) { 8 %0 = load float addrspace(1)* %in 9 %arrayidx1 = getelementptr inbounds float addrspace(1)* %in, i32 1 10 %1 = load float addrspace(1)* %arrayidx1 13 store i32 %sext, i32 addrspace(1)* %out 24 define void @fcmp_br(i32 addrspace(1)* %out, float %in) { 30 %1 = getelementptr i32 addrspace(1)* %out, i32 1 31 store i32 0, i32 addrspace(1)* %1 35 store i32 0, i32 addrspace(1)* %out
|
D | dagcombiner-bug-illegal-vec4-int-to-fp.ll | 13 define void @sint(<4 x float> addrspace(1)* %out, i32 addrspace(1)* %in) { 15 %ptr = getelementptr i32 addrspace(1)* %in, i32 1 16 %sint = load i32 addrspace(1) * %in 20 store <4 x float> %splat, <4 x float> addrspace(1)* %out 27 define void @uint(<4 x float> addrspace(1)* %out, i32 addrspace(1)* %in) { 29 %ptr = getelementptr i32 addrspace(1)* %in, i32 1 30 %uint = load i32 addrspace(1) * %in 34 store <4 x float> %splat, <4 x float> addrspace(1)* %out
|
D | selectcc-opt.ll | 7 define void @test_a(i32 addrspace(1)* %out, float %in) { 19 %7 = getelementptr i32 addrspace(1)* %out, i32 1 20 store i32 0, i32 addrspace(1)* %7 24 store i32 0, i32 addrspace(1)* %out 34 define void @test_b(i32 addrspace(1)* %out, float %in) { 46 %7 = getelementptr i32 addrspace(1)* %out, i32 1 47 store i32 0, i32 addrspace(1)* %7 51 store i32 0, i32 addrspace(1)* %out 58 define void @test_c(float addrspace(1)* %out, i32 %in) { 62 store float %1, float addrspace(1)* %out
|
D | short-args.ll | 6 define void @i8_arg(i32 addrspace(1)* nocapture %out, i8 %in) nounwind { 9 store i32 %0, i32 addrspace(1)* %out, align 4 16 define void @i8_zext_arg(i32 addrspace(1)* nocapture %out, i8 zeroext %in) nounwind { 19 store i32 %0, i32 addrspace(1)* %out, align 4 26 define void @i16_arg(i32 addrspace(1)* nocapture %out, i16 %in) nounwind { 29 store i32 %0, i32 addrspace(1)* %out, align 4 36 define void @i16_zext_arg(i32 addrspace(1)* nocapture %out, i16 zeroext %in) nounwind { 39 store i32 %0, i32 addrspace(1)* %out, align 4
|
/external/llvm/test/CodeGen/NVPTX/ |
D | ld-addrspace.ll | 8 define i8 @ld_global_i8(i8 addrspace(1)* %ptr) { 13 %a = load i8 addrspace(1)* %ptr 17 define i8 @ld_shared_i8(i8 addrspace(3)* %ptr) { 22 %a = load i8 addrspace(3)* %ptr 26 define i8 @ld_local_i8(i8 addrspace(5)* %ptr) { 31 %a = load i8 addrspace(5)* %ptr 36 define i16 @ld_global_i16(i16 addrspace(1)* %ptr) { 41 %a = load i16 addrspace(1)* %ptr 45 define i16 @ld_shared_i16(i16 addrspace(3)* %ptr) { 50 %a = load i16 addrspace(3)* %ptr [all …]
|
D | st-addrspace.ll | 9 define void @st_global_i8(i8 addrspace(1)* %ptr, i8 %a) { 14 store i8 %a, i8 addrspace(1)* %ptr 18 define void @st_shared_i8(i8 addrspace(3)* %ptr, i8 %a) { 23 store i8 %a, i8 addrspace(3)* %ptr 27 define void @st_local_i8(i8 addrspace(5)* %ptr, i8 %a) { 32 store i8 %a, i8 addrspace(5)* %ptr 38 define void @st_global_i16(i16 addrspace(1)* %ptr, i16 %a) { 43 store i16 %a, i16 addrspace(1)* %ptr 47 define void @st_shared_i16(i16 addrspace(3)* %ptr, i16 %a) { 52 store i16 %a, i16 addrspace(3)* %ptr [all …]
|
D | ld-generic.ll | 6 define i8 @ld_global_i8(i8 addrspace(0)* %ptr) { 11 %a = load i8 addrspace(0)* %ptr 16 define i16 @ld_global_i16(i16 addrspace(0)* %ptr) { 21 %a = load i16 addrspace(0)* %ptr 26 define i32 @ld_global_i32(i32 addrspace(0)* %ptr) { 31 %a = load i32 addrspace(0)* %ptr 36 define i64 @ld_global_i64(i64 addrspace(0)* %ptr) { 41 %a = load i64 addrspace(0)* %ptr 46 define float @ld_global_f32(float addrspace(0)* %ptr) { 51 %a = load float addrspace(0)* %ptr [all …]
|
D | st-generic.ll | 7 define void @st_global_i8(i8 addrspace(0)* %ptr, i8 %a) { 12 store i8 %a, i8 addrspace(0)* %ptr 18 define void @st_global_i16(i16 addrspace(0)* %ptr, i16 %a) { 23 store i16 %a, i16 addrspace(0)* %ptr 29 define void @st_global_i32(i32 addrspace(0)* %ptr, i32 %a) { 34 store i32 %a, i32 addrspace(0)* %ptr 40 define void @st_global_i64(i64 addrspace(0)* %ptr, i64 %a) { 45 store i64 %a, i64 addrspace(0)* %ptr 51 define void @st_global_f32(float addrspace(0)* %ptr, float %a) { 56 store float %a, float addrspace(0)* %ptr [all …]
|
D | global-ordering.ll | 11 @a2 = addrspace(1) global i8 addrspace(1)* @a 12 @a = addrspace(1) global i8 2 19 @b2 = addrspace(1) global [2 x i8 addrspace(1)*] [i8 addrspace(1)* @b, i8 addrspace(1)* @b] 20 @b = addrspace(1) global i8 1
|
D | vector-select.ll | 7 define void @foo(<2 x i32> addrspace(1)* %def_a, <2 x i32> addrspace(1)* %def_b, <2 x i32> addrspac… 9 %tmp4 = load <2 x i32> addrspace(1)* %def_a 10 %tmp6 = load <2 x i32> addrspace(1)* %def_c 11 %tmp8 = load <2 x i32> addrspace(1)* %def_b 14 store <2 x i32> %cond, <2 x i32> addrspace(1)* %def_c
|
/external/llvm/test/Transforms/ScalarRepl/ |
D | address-space.ll | 10 ; CHECK: load float addrspace(2)* 12 ; CHECK: store float {{.*}}, float addrspace(2)* 13 define void @Test(%struct.anon addrspace(2)* %pPtr) nounwind { 16 …%arrayidx = getelementptr inbounds %struct.anon addrspace(2)* %pPtr, i64 0 ; <%struct.anon addrspa… 18 …%tmp2 = bitcast %struct.anon addrspace(2)* %arrayidx to i8 addrspace(2)* ; <i8 addrspace(2)*> [#us… 19 call void @llvm.memcpy.p0i8.p2i8.i64(i8* %tmp1, i8 addrspace(2)* %tmp2, i64 4, i32 4, i1 false) 25 …%arrayidx7 = getelementptr inbounds %struct.anon addrspace(2)* %pPtr, i64 0 ; <%struct.anon addrsp… 26 …%tmp8 = bitcast %struct.anon addrspace(2)* %arrayidx7 to i8 addrspace(2)* ; <i8 addrspace(2)*> [#u… 28 call void @llvm.memcpy.p2i8.p0i8.i64(i8 addrspace(2)* %tmp8, i8* %tmp9, i64 4, i32 4, i1 false) 32 declare void @llvm.memcpy.p0i8.p2i8.i64(i8* nocapture, i8 addrspace(2)* nocapture, i64, i32, i1) no… [all …]
|
/external/llvm/test/Transforms/GlobalOpt/ |
D | 2008-07-17-addrspace.ll | 7 ; RUN: cat %t | grep addrspace.*global 9 @struct = internal addrspace(1) global { i32, i32 } zeroinitializer 10 @array = internal addrspace(1) global [ 2 x i32 ] zeroinitializer 13 %A = load i32 addrspace(1) * getelementptr ({ i32, i32 } addrspace(1) * @struct, i32 0, i32 0) 14 %B = load i32 addrspace(1) * getelementptr ([ 2 x i32 ] addrspace(1) * @array, i32 0, i32 0) 23 store i32 %R, i32 addrspace(1) * getelementptr ([ 2 x i32 ] addrspace(1) * @array, i32 0, i32 0) 24 store i32 %R, i32 addrspace(1) * getelementptr ({ i32, i32 } addrspace(1) * @struct, i32 0, i32 0)
|
/external/llvm/test/CodeGen/SI/ |
D | sanity.ll | 8 …%0 = load <4 x i32> addrspace(2)* addrspace(8)* inttoptr (i32 6 to <4 x i32> addrspace(2)* addrspa… 9 %1 = getelementptr <4 x i32> addrspace(2)* %0, i32 0 10 %2 = load <4 x i32> addrspace(2)* %1 17 …%9 = load <4 x i32> addrspace(2)* addrspace(8)* inttoptr (i32 6 to <4 x i32> addrspace(2)* addrspa… 18 %10 = getelementptr <4 x i32> addrspace(2)* %9, i32 1 19 %11 = load <4 x i32> addrspace(2)* %10
|
/external/llvm/test/CodeGen/X86/ |
D | atomic-dagsched.ll | 21 %4 = load i64* addrspace(256)* inttoptr (i64 264 to i64* addrspace(256)*), align 8, !tbaa !0 37 %8 = bitcast i8* %ptrtoarg4 to i32 addrspace(1)* 38 %asr.iv911 = bitcast i8* %asr.iv9 to <8 x i32> addrspace(1)* 39 %9 = load <8 x i32> addrspace(1)* %asr.iv911, align 4 48 %10 = atomicrmw min i32 addrspace(1)* %8, i32 %extract8vector_func.i seq_cst 49 %11 = atomicrmw min i32 addrspace(1)* %8, i32 %extract9vector_func.i seq_cst 50 %12 = atomicrmw min i32 addrspace(1)* %8, i32 %extract10vector_func.i seq_cst 51 %13 = atomicrmw min i32 addrspace(1)* %8, i32 %extract11vector_func.i seq_cst 52 %14 = atomicrmw min i32 addrspace(1)* %8, i32 %extract12vector_func.i seq_cst 53 %15 = atomicrmw min i32 addrspace(1)* %8, i32 %extract13vector_func.i seq_cst [all …]
|
D | movgs.ll | 7 …%tmp = load i32* addrspace(256)* getelementptr (i32* addrspace(256)* inttoptr (i32 72 to i32* addr… 21 define i64 @test2(void (i8*)* addrspace(256)* %tmp8) nounwind { 23 %tmp9 = load void (i8*)* addrspace(256)* %tmp8, align 8 39 define <2 x i64> @pmovsxwd_1(i64 addrspace(256)* %p) nounwind readonly { 41 %0 = load i64 addrspace(256)* %p 62 …%tmp = load i32* addrspace(256)* getelementptr (i32* addrspace(256)* inttoptr (i32 72 to i32* addr… 64 …%tmp2 = load i32* addrspace(257)* getelementptr (i32* addrspace(257)* inttoptr (i32 72 to i32* add…
|
D | scalar_widen_div.ll | 7 define void @vectorDiv (<2 x i32> addrspace(1)* %nsource, <2 x i32> addrspace(1)* %dsource, <2 x i3… 13 %nsource.addr = alloca <2 x i32> addrspace(1)*, align 4 14 %dsource.addr = alloca <2 x i32> addrspace(1)*, align 4 15 %qdest.addr = alloca <2 x i32> addrspace(1)*, align 4 17 store <2 x i32> addrspace(1)* %nsource, <2 x i32> addrspace(1)** %nsource.addr 18 store <2 x i32> addrspace(1)* %dsource, <2 x i32> addrspace(1)** %dsource.addr 19 store <2 x i32> addrspace(1)* %qdest, <2 x i32> addrspace(1)** %qdest.addr 20 %tmp = load <2 x i32> addrspace(1)** %qdest.addr 22 %arrayidx = getelementptr <2 x i32> addrspace(1)* %tmp, i32 %tmp1 23 %tmp2 = load <2 x i32> addrspace(1)** %nsource.addr [all …]
|
/external/llvm/test/CodeGen/ARM/ |
D | fast-isel-crash.ll | 7 define internal void @stretch(<4 x i8> addrspace(1)* %src, <4 x i8> addrspace(1)* %dst, i32 %width,… 17 define void @wrap(i8 addrspace(1)* addrspace(1)* %arglist, i32 addrspace(1)* %gtid) nounwind ssp { 19 …call void @stretch(<4 x i8> addrspace(1)* undef, <4 x i8> addrspace(1)* undef, i32 undef, i32 unde…
|