/external/llvm/test/Assembler/ |
D | 2007-12-11-AddressSpaces.ll | 1 ; RUN: llvm-as < %s | llvm-dis | grep "addrspace(33)" | count 7 2 ; RUN: llvm-as < %s | llvm-dis | grep "addrspace(42)" | count 2 3 ; RUN: llvm-as < %s | llvm-dis | grep "addrspace(66)" | count 2 4 ; RUN: llvm-as < %s | llvm-dis | grep "addrspace(11)" | count 6 5 ; RUN: llvm-as < %s | llvm-dis | grep "addrspace(22)" | count 5 7 %struct.mystruct = type { i32, i32 addrspace(33)*, i32, i32 addrspace(33)* } 8 @input = weak addrspace(42) global %struct.mystruct zeroinitializer ; <%struct.mystruct addrspac… 9 @output = addrspace(66) global %struct.mystruct zeroinitializer ; <%struct.mystruct addrspace(66)… 10 @y = external addrspace(33) global i32 addrspace(11)* addrspace(22)* ; <i32 addrspace(11)* addrsp… 14 …p1 = load i32 addrspace(33)* addrspace(42)* getelementptr (%struct.mystruct addrspace(42)* @input,… [all …]
|
/external/llvm/test/Transforms/InstCombine/ |
D | constant-fold-address-space-pointer.ll | 4 @g = addrspace(3) global i32 89 6 @const_zero_i8_as1 = addrspace(1) constant i8 0 7 @const_zero_i32_as1 = addrspace(1) constant i32 0 9 @const_zero_i8_as2 = addrspace(2) constant i8 0 10 @const_zero_i32_as2 = addrspace(2) constant i32 0 12 @const_zero_i8_as3 = addrspace(3) constant i8 0 13 @const_zero_i32_as3 = addrspace(3) constant i32 0 17 define i32 addrspace(3)* @test_constant_fold_inttoptr_as_pointer_same_size() { 19 ; CHECK-NEXT: ret i32 addrspace(3)* @const_zero_i32_as3 20 %x = ptrtoint i32 addrspace(3)* @const_zero_i32_as3 to i32 [all …]
|
D | addrspacecast.ll | 7 declare void @llvm.memcpy.p0i8.p1i8.i32(i8*, i8 addrspace(1)*, i32, i32, i1) nounwind 8 declare void @llvm.memcpy.p0i8.p2i8.i32(i8*, i8 addrspace(2)*, i32, i32, i1) nounwind 11 define i32* @combine_redundant_addrspacecast(i32 addrspace(1)* %x) nounwind { 13 ; CHECK: addrspacecast i32 addrspace(1)* %x to i32* 15 %y = addrspacecast i32 addrspace(1)* %x to i32 addrspace(3)* 16 %z = addrspacecast i32 addrspace(3)* %y to i32* 20 define <4 x i32*> @combine_redundant_addrspacecast_vector(<4 x i32 addrspace(1)*> %x) nounwind { 22 ; CHECK: addrspacecast <4 x i32 addrspace(1)*> %x to <4 x i32*> 24 %y = addrspacecast <4 x i32 addrspace(1)*> %x to <4 x i32 addrspace(3)*> 25 %z = addrspacecast <4 x i32 addrspace(3)*> %y to <4 x i32*> [all …]
|
D | multi-size-address-space-pointer.ll | 5 define i32 @test_as0(i32 addrspace(0)* %a) { 8 %arrayidx = getelementptr i32 addrspace(0)* %a, i64 1 9 %y = load i32 addrspace(0)* %arrayidx, align 4 13 define i32 @test_as1(i32 addrspace(1)* %a) { 15 ; CHECK: %arrayidx = getelementptr i32 addrspace(1)* %a, i64 1 16 %arrayidx = getelementptr i32 addrspace(1)* %a, i32 1 17 %y = load i32 addrspace(1)* %arrayidx, align 4 21 define i32 @test_as2(i32 addrspace(2)* %a) { 23 ; CHECK: %arrayidx = getelementptr i32 addrspace(2)* %a, i8 1 24 %arrayidx = getelementptr i32 addrspace(2)* %a, i32 1 [all …]
|
/external/llvm/test/Other/ |
D | constant-fold-gep-address-spaces.ll | 10 ; PLAIN: @G8 = global i8 addrspace(1)* getelementptr (i8 addrspace(1)* inttoptr (i32 1 to i8 addrsp… 11 @G8 = global i8 addrspace(1)* getelementptr (i8 addrspace(1)* inttoptr (i32 1 to i8 addrspace(1)*),… 12 ; PLAIN: @G1 = global i1 addrspace(2)* getelementptr (i1 addrspace(2)* inttoptr (i8 1 to i1 addrspa… 13 @G1 = global i1 addrspace(2)* getelementptr (i1 addrspace(2)* inttoptr (i8 1 to i1 addrspace(2)*), … 14 ; PLAIN: @F8 = global i8 addrspace(1)* getelementptr (i8 addrspace(1)* inttoptr (i32 1 to i8 addrsp… 15 @F8 = global i8 addrspace(1)* getelementptr (i8 addrspace(1)* inttoptr (i32 1 to i8 addrspace(1)*),… 16 ; PLAIN: @F1 = global i1 addrspace(2)* getelementptr (i1 addrspace(2)* inttoptr (i8 1 to i1 addrspa… 17 @F1 = global i1 addrspace(2)* getelementptr (i1 addrspace(2)* inttoptr (i8 1 to i1 addrspace(2)*), … 18 ; PLAIN: @H8 = global i8 addrspace(1)* getelementptr (i8 addrspace(1)* null, i32 -1) 19 @H8 = global i8 addrspace(1)* getelementptr (i8 addrspace(1)* inttoptr (i32 0 to i8 addrspace(1)*),… [all …]
|
/external/llvm/test/CodeGen/R600/ |
D | local-atomics.ll | 10 define void @lds_atomic_xchg_ret_i32(i32 addrspace(1)* %out, i32 addrspace(3)* %ptr) nounwind { 11 %result = atomicrmw xchg i32 addrspace(3)* %ptr, i32 4 seq_cst 12 store i32 %result, i32 addrspace(1)* %out, align 4 19 define void @lds_atomic_xchg_ret_i32_offset(i32 addrspace(1)* %out, i32 addrspace(3)* %ptr) nounwin… 20 %gep = getelementptr i32 addrspace(3)* %ptr, i32 4 21 %result = atomicrmw xchg i32 addrspace(3)* %gep, i32 4 seq_cst 22 store i32 %result, i32 addrspace(1)* %out, align 4 34 define void @lds_atomic_add_ret_i32(i32 addrspace(1)* %out, i32 addrspace(3)* %ptr) nounwind { 35 %result = atomicrmw add i32 addrspace(3)* %ptr, i32 4 seq_cst 36 store i32 %result, i32 addrspace(1)* %out, align 4 [all …]
|
D | local-atomics64.ll | 6 define void @lds_atomic_xchg_ret_i64(i64 addrspace(1)* %out, i64 addrspace(3)* %ptr) nounwind { 7 %result = atomicrmw xchg i64 addrspace(3)* %ptr, i64 4 seq_cst 8 store i64 %result, i64 addrspace(1)* %out, align 8 15 define void @lds_atomic_xchg_ret_i64_offset(i64 addrspace(1)* %out, i64 addrspace(3)* %ptr) nounwin… 16 %gep = getelementptr i64 addrspace(3)* %ptr, i32 4 17 %result = atomicrmw xchg i64 addrspace(3)* %gep, i64 4 seq_cst 18 store i64 %result, i64 addrspace(1)* %out, align 8 25 define void @lds_atomic_add_ret_i64(i64 addrspace(1)* %out, i64 addrspace(3)* %ptr) nounwind { 26 %result = atomicrmw add i64 addrspace(3)* %ptr, i64 4 seq_cst 27 store i64 %result, i64 addrspace(1)* %out, align 8 [all …]
|
D | unhandled-loop-condition-assertion.ll | 9 define void @branch_true(i8 addrspace(1)* nocapture %main, i32 %main_stride) #0 { 20 %main.addr.011 = phi i8 addrspace(1)* [ %main, %for.body.lr.ph ], [ %add.ptr6, %for.body ] 21 %0 = bitcast i8 addrspace(1)* %main.addr.011 to i32 addrspace(1)* 22 %1 = load i32 addrspace(1)* %0, align 4 23 %add.ptr = getelementptr inbounds i8 addrspace(1)* %main.addr.011, i32 %main_stride 24 %2 = bitcast i8 addrspace(1)* %add.ptr to i32 addrspace(1)* 25 %3 = load i32 addrspace(1)* %2, align 4 26 %add.ptr1 = getelementptr inbounds i8 addrspace(1)* %main.addr.011, i32 %add.ptr.sum 27 %4 = bitcast i8 addrspace(1)* %add.ptr1 to i32 addrspace(1)* 28 %5 = load i32 addrspace(1)* %4, align 4 [all …]
|
D | extload.ll | 7 define void @anyext_load_i8(i8 addrspace(1)* nocapture noalias %out, i8 addrspace(1)* nocapture noa… 8 %cast = bitcast i8 addrspace(1)* %src to i32 addrspace(1)* 9 %load = load i32 addrspace(1)* %cast, align 1 11 %castOut = bitcast i8 addrspace(1)* %out to <4 x i8> addrspace(1)* 12 store <4 x i8> %x, <4 x i8> addrspace(1)* %castOut, align 1 21 define void @anyext_load_i16(i16 addrspace(1)* nocapture noalias %out, i16 addrspace(1)* nocapture … 22 %cast = bitcast i16 addrspace(1)* %src to i32 addrspace(1)* 23 %load = load i32 addrspace(1)* %cast, align 1 25 %castOut = bitcast i16 addrspace(1)* %out to <2 x i16> addrspace(1)* 26 store <2 x i16> %x, <2 x i16> addrspace(1)* %castOut, align 1 [all …]
|
D | fcmp64.ll | 6 define void @flt_f64(double addrspace(1)* %out, double addrspace(1)* %in1, 7 double addrspace(1)* %in2) { 8 %r0 = load double addrspace(1)* %in1 9 %r1 = load double addrspace(1)* %in2 12 store double %r3, double addrspace(1)* %out 19 define void @fle_f64(double addrspace(1)* %out, double addrspace(1)* %in1, 20 double addrspace(1)* %in2) { 21 %r0 = load double addrspace(1)* %in1 22 %r1 = load double addrspace(1)* %in2 25 store double %r3, double addrspace(1)* %out [all …]
|
D | srem.ll | 4 define void @srem_i32(i32 addrspace(1)* %out, i32 addrspace(1)* %in) { 5 %den_ptr = getelementptr i32 addrspace(1)* %in, i32 1 6 %num = load i32 addrspace(1) * %in 7 %den = load i32 addrspace(1) * %den_ptr 9 store i32 %result, i32 addrspace(1)* %out 13 define void @srem_i32_4(i32 addrspace(1)* %out, i32 addrspace(1)* %in) { 14 %num = load i32 addrspace(1) * %in 16 store i32 %result, i32 addrspace(1)* %out 20 define void @srem_v2i32(<2 x i32> addrspace(1)* %out, <2 x i32> addrspace(1)* %in) { 21 %den_ptr = getelementptr <2 x i32> addrspace(1)* %in, i32 1 [all …]
|
D | fma.ll | 13 define void @fma_f32(float addrspace(1)* %out, float addrspace(1)* %in1, 14 float addrspace(1)* %in2, float addrspace(1)* %in3) { 15 %r0 = load float addrspace(1)* %in1 16 %r1 = load float addrspace(1)* %in2 17 %r2 = load float addrspace(1)* %in3 19 store float %r3, float addrspace(1)* %out 26 define void @fma_v2f32(<2 x float> addrspace(1)* %out, <2 x float> addrspace(1)* %in1, 27 <2 x float> addrspace(1)* %in2, <2 x float> addrspace(1)* %in3) { 28 %r0 = load <2 x float> addrspace(1)* %in1 29 %r1 = load <2 x float> addrspace(1)* %in2 [all …]
|
D | 32-bit-local-address-space.ll | 15 define void @local_address_load(i32 addrspace(1)* %out, i32 addrspace(3)* %in) { 17 %0 = load i32 addrspace(3)* %in 18 store i32 %0, i32 addrspace(1)* %out 26 define void @local_address_gep(i32 addrspace(1)* %out, i32 addrspace(3)* %in, i32 %offset) { 28 %0 = getelementptr i32 addrspace(3)* %in, i32 %offset 29 %1 = load i32 addrspace(3)* %0 30 store i32 %1, i32 addrspace(1)* %out 37 define void @local_address_gep_const_offset(i32 addrspace(1)* %out, i32 addrspace(3)* %in) { 39 %0 = getelementptr i32 addrspace(3)* %in, i32 1 40 %1 = load i32 addrspace(3)* %0 [all …]
|
D | local-64.ll | 6 define void @local_i32_load(i32 addrspace(1)* %out, i32 addrspace(3)* %in) nounwind { 7 %gep = getelementptr i32 addrspace(3)* %in, i32 7 8 %val = load i32 addrspace(3)* %gep, align 4 9 store i32 %val, i32 addrspace(1)* %out, align 4 16 define void @local_i32_load_0_offset(i32 addrspace(1)* %out, i32 addrspace(3)* %in) nounwind { 17 %val = load i32 addrspace(3)* %in, align 4 18 store i32 %val, i32 addrspace(1)* %out, align 4 26 define void @local_i8_load_i16_max_offset(i8 addrspace(1)* %out, i8 addrspace(3)* %in) nounwind { 27 %gep = getelementptr i8 addrspace(3)* %in, i32 65535 28 %val = load i8 addrspace(3)* %gep, align 4 [all …]
|
D | load.ll | 14 define void @load_i8(i32 addrspace(1)* %out, i8 addrspace(1)* %in) { 15 %1 = load i8 addrspace(1)* %in 17 store i32 %2, i32 addrspace(1)* %out 28 define void @load_i8_sext(i32 addrspace(1)* %out, i8 addrspace(1)* %in) { 30 %0 = load i8 addrspace(1)* %in 32 store i32 %1, i32 addrspace(1)* %out 41 define void @load_v2i8(<2 x i32> addrspace(1)* %out, <2 x i8> addrspace(1)* %in) { 43 %0 = load <2 x i8> addrspace(1)* %in 45 store <2 x i32> %1, <2 x i32> addrspace(1)* %out 62 define void @load_v2i8_sext(<2 x i32> addrspace(1)* %out, <2 x i8> addrspace(1)* %in) { [all …]
|
D | xor.ll | 13 define void @xor_v2i32(<2 x i32> addrspace(1)* %out, <2 x i32> addrspace(1)* %in0, <2 x i32> addrsp… 14 %a = load <2 x i32> addrspace(1) * %in0 15 %b = load <2 x i32> addrspace(1) * %in1 17 store <2 x i32> %result, <2 x i32> addrspace(1)* %out 33 define void @xor_v4i32(<4 x i32> addrspace(1)* %out, <4 x i32> addrspace(1)* %in0, <4 x i32> addrsp… 34 %a = load <4 x i32> addrspace(1) * %in0 35 %b = load <4 x i32> addrspace(1) * %in1 37 store <4 x i32> %result, <4 x i32> addrspace(1)* %out 47 define void @xor_i1(float addrspace(1)* %out, float addrspace(1)* %in0, float addrspace(1)* %in1) { 48 %a = load float addrspace(1) * %in0 [all …]
|
D | llvm.AMDGPU.bfe.u32.ll | 9 define void @bfe_u32_arg_arg_arg(i32 addrspace(1)* %out, i32 %src0, i32 %src1, i32 %src2) nounwind { 11 store i32 %bfe_u32, i32 addrspace(1)* %out, align 4 18 define void @bfe_u32_arg_arg_imm(i32 addrspace(1)* %out, i32 %src0, i32 %src1) nounwind { 20 store i32 %bfe_u32, i32 addrspace(1)* %out, align 4 27 define void @bfe_u32_arg_imm_arg(i32 addrspace(1)* %out, i32 %src0, i32 %src2) nounwind { 29 store i32 %bfe_u32, i32 addrspace(1)* %out, align 4 36 define void @bfe_u32_imm_arg_arg(i32 addrspace(1)* %out, i32 %src1, i32 %src2) nounwind { 38 store i32 %bfe_u32, i32 addrspace(1)* %out, align 4 46 define void @bfe_u32_arg_0_width_reg_offset(i32 addrspace(1)* %out, i32 %src0, i32 %src1) nounwind { 48 store i32 %bfe_u32, i32 addrspace(1)* %out, align 4 [all …]
|
D | and.ll | 11 define void @test2(<2 x i32> addrspace(1)* %out, <2 x i32> addrspace(1)* %in) { 12 %b_ptr = getelementptr <2 x i32> addrspace(1)* %in, i32 1 13 %a = load <2 x i32> addrspace(1) * %in 14 %b = load <2 x i32> addrspace(1) * %b_ptr 16 store <2 x i32> %result, <2 x i32> addrspace(1)* %out 31 define void @test4(<4 x i32> addrspace(1)* %out, <4 x i32> addrspace(1)* %in) { 32 %b_ptr = getelementptr <4 x i32> addrspace(1)* %in, i32 1 33 %a = load <4 x i32> addrspace(1) * %in 34 %b = load <4 x i32> addrspace(1) * %b_ptr 36 store <4 x i32> %result, <4 x i32> addrspace(1)* %out [all …]
|
D | sdiv.ll | 15 define void @sdiv_i32(i32 addrspace(1)* %out, i32 addrspace(1)* %in) { 16 %den_ptr = getelementptr i32 addrspace(1)* %in, i32 1 17 %num = load i32 addrspace(1) * %in 18 %den = load i32 addrspace(1) * %den_ptr 20 store i32 %result, i32 addrspace(1)* %out 25 define void @sdiv_i32_4(i32 addrspace(1)* %out, i32 addrspace(1)* %in) { 26 %num = load i32 addrspace(1) * %in 28 store i32 %result, i32 addrspace(1)* %out 45 define void @slow_sdiv_i32_3435(i32 addrspace(1)* %out, i32 addrspace(1)* %in) { 46 %num = load i32 addrspace(1) * %in [all …]
|
D | si-vector-hang.ll | 17 define void @test_8_min_char(i8 addrspace(1)* nocapture %out, i8 addrspace(1)* nocapture readonly %… 19 %0 = load i8 addrspace(1)* %in0, align 1 21 %arrayidx2.i.i = getelementptr inbounds i8 addrspace(1)* %in0, i64 1 22 %2 = load i8 addrspace(1)* %arrayidx2.i.i, align 1 24 %arrayidx6.i.i = getelementptr inbounds i8 addrspace(1)* %in0, i64 2 25 %4 = load i8 addrspace(1)* %arrayidx6.i.i, align 1 27 %arrayidx10.i.i = getelementptr inbounds i8 addrspace(1)* %in0, i64 3 28 %6 = load i8 addrspace(1)* %arrayidx10.i.i, align 1 30 %arrayidx.i.i = getelementptr inbounds i8 addrspace(1)* %in0, i64 4 31 %8 = load i8 addrspace(1)* %arrayidx.i.i, align 1 [all …]
|
D | mubuf.ll | 10 define void @mubuf_load0(i32 addrspace(1)* %out, i32 addrspace(1)* %in) { 12 %0 = getelementptr i32 addrspace(1)* %in, i64 1 13 %1 = load i32 addrspace(1)* %0 14 store i32 %1, i32 addrspace(1)* %out 21 define void @mubuf_load1(i8 addrspace(1)* %out, i8 addrspace(1)* %in) { 23 %0 = getelementptr i8 addrspace(1)* %in, i64 4095 24 %1 = load i8 addrspace(1)* %0 25 store i8 %1, i8 addrspace(1)* %out 32 define void @mubuf_load2(i32 addrspace(1)* %out, i32 addrspace(1)* %in) { 34 %0 = getelementptr i32 addrspace(1)* %in, i64 1024 [all …]
|
D | bitcast.ll | 9 define void @v32i8_to_v8i32(<32 x i8> addrspace(2)* inreg) #0 { 11 %1 = load <32 x i8> addrspace(2)* %0 22 define void @i8ptr_v16i8ptr(<16 x i8> addrspace(1)* %out, i8 addrspace(1)* %in) { 24 %0 = bitcast i8 addrspace(1)* %in to <16 x i8> addrspace(1)* 25 %1 = load <16 x i8> addrspace(1)* %0 26 store <16 x i8> %1, <16 x i8> addrspace(1)* %out 30 define void @f32_to_v2i16(<2 x i16> addrspace(1)* %out, float addrspace(1)* %in) nounwind { 31 %load = load float addrspace(1)* %in, align 4 33 store <2 x i16> %bc, <2 x i16> addrspace(1)* %out, align 4 37 define void @v2i16_to_f32(float addrspace(1)* %out, <2 x i16> addrspace(1)* %in) nounwind { [all …]
|
D | or.ll | 12 define void @or_v2i32(<2 x i32> addrspace(1)* %out, <2 x i32> addrspace(1)* %in) { 13 %b_ptr = getelementptr <2 x i32> addrspace(1)* %in, i32 1 14 %a = load <2 x i32> addrspace(1) * %in 15 %b = load <2 x i32> addrspace(1) * %b_ptr 17 store <2 x i32> %result, <2 x i32> addrspace(1)* %out 33 define void @or_v4i32(<4 x i32> addrspace(1)* %out, <4 x i32> addrspace(1)* %in) { 34 %b_ptr = getelementptr <4 x i32> addrspace(1)* %in, i32 1 35 %a = load <4 x i32> addrspace(1) * %in 36 %b = load <4 x i32> addrspace(1) * %b_ptr 38 store <4 x i32> %result, <4 x i32> addrspace(1)* %out [all …]
|
D | kcache-fold.ll | 7 %0 = load <4 x float> addrspace(8)* null 9 …%2 = load <4 x float> addrspace(8)* getelementptr ([1024 x <4 x float>] addrspace(8)* null, i64 0,… 11 …%4 = load <4 x float> addrspace(8)* getelementptr ([1024 x <4 x float>] addrspace(8)* null, i64 0,… 15 %8 = load <4 x float> addrspace(8)* null 17 …%10 = load <4 x float> addrspace(8)* getelementptr ([1024 x <4 x float>] addrspace(8)* null, i64 0… 19 …%12 = load <4 x float> addrspace(8)* getelementptr ([1024 x <4 x float>] addrspace(8)* null, i64 0… 23 %16 = load <4 x float> addrspace(8)* null 25 …%18 = load <4 x float> addrspace(8)* getelementptr ([1024 x <4 x float>] addrspace(8)* null, i64 0… 27 …%20 = load <4 x float> addrspace(8)* getelementptr ([1024 x <4 x float>] addrspace(8)* null, i64 0… 31 %24 = load <4 x float> addrspace(8)* null [all …]
|
/external/llvm/test/CodeGen/XCore/ |
D | resources.ll | 3 declare i8 addrspace(1)* @llvm.xcore.getr.p1i8(i32 %type) 4 declare void @llvm.xcore.freer.p1i8(i8 addrspace(1)* %r) 5 declare i32 @llvm.xcore.in.p1i8(i8 addrspace(1)* %r) 6 declare i32 @llvm.xcore.int.p1i8(i8 addrspace(1)* %r) 7 declare i32 @llvm.xcore.inct.p1i8(i8 addrspace(1)* %r) 8 declare void @llvm.xcore.out.p1i8(i8 addrspace(1)* %r, i32 %value) 9 declare void @llvm.xcore.outt.p1i8(i8 addrspace(1)* %r, i32 %value) 10 declare void @llvm.xcore.outct.p1i8(i8 addrspace(1)* %r, i32 %value) 11 declare void @llvm.xcore.chkct.p1i8(i8 addrspace(1)* %r, i32 %value) 12 declare i32 @llvm.xcore.testct.p1i8(i8 addrspace(1)* %r) [all …]
|