/external/llvm/test/Transforms/LoadStoreVectorizer/AMDGPU/ |
D | pointer-elements.ll | 9 ; CHECK: inttoptr i64 %{{[0-9]+}} to i8 addrspace(1)* 10 ; CHECK: inttoptr i64 %{{[0-9]+}} to i8 addrspace(1)* 12 define void @merge_v2p1i8(i8 addrspace(1)* addrspace(1)* nocapture %a, i8 addrspace(1)* addrspace(1… 14 %a.1 = getelementptr inbounds i8 addrspace(1)*, i8 addrspace(1)* addrspace(1)* %a, i64 1 15 %b.1 = getelementptr inbounds i8 addrspace(1)*, i8 addrspace(1)* addrspace(1)* %b, i64 1 17 %ld.c = load i8 addrspace(1)*, i8 addrspace(1)* addrspace(1)* %b, align 4 18 %ld.c.idx.1 = load i8 addrspace(1)*, i8 addrspace(1)* addrspace(1)* %b.1, align 4 20 store i8 addrspace(1)* null, i8 addrspace(1)* addrspace(1)* %a, align 4 21 store i8 addrspace(1)* null, i8 addrspace(1)* addrspace(1)* %a.1, align 4 28 ; CHECK: inttoptr i32 %{{[0-9]+}} to i8 addrspace(3)* [all …]
|
D | merge-stores.ll | 12 ; CHECK: store <2 x i8> <i8 -56, i8 123>, <2 x i8> addrspace(1)* %{{[0-9]+}}, align 2 13 define void @merge_global_store_2_constants_i8(i8 addrspace(1)* %out) #0 { 14 %out.gep.1 = getelementptr i8, i8 addrspace(1)* %out, i32 1 16 store i8 123, i8 addrspace(1)* %out.gep.1 17 store i8 456, i8 addrspace(1)* %out, align 2 23 define void @merge_global_store_2_constants_i8_natural_align(i8 addrspace(1)* %out) #0 { 24 %out.gep.1 = getelementptr i8, i8 addrspace(1)* %out, i32 1 26 store i8 123, i8 addrspace(1)* %out.gep.1 27 store i8 456, i8 addrspace(1)* %out 32 ; CHECK: store <2 x i16> <i16 456, i16 123>, <2 x i16> addrspace(1)* %{{[0-9]+}}, align 4 [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/LoadStoreVectorizer/AMDGPU/ |
D | pointer-elements.ll | 9 ; CHECK: inttoptr i64 %{{[^ ]+}} to i8 addrspace(1)* 10 ; CHECK: inttoptr i64 %{{[^ ]+}} to i8 addrspace(1)* 12 …efine amdgpu_kernel void @merge_v2p1i8(i8 addrspace(1)* addrspace(1)* nocapture %a, i8 addrspace(1… 14 %a.1 = getelementptr inbounds i8 addrspace(1)*, i8 addrspace(1)* addrspace(1)* %a, i64 1 15 %b.1 = getelementptr inbounds i8 addrspace(1)*, i8 addrspace(1)* addrspace(1)* %b, i64 1 17 %ld.c = load i8 addrspace(1)*, i8 addrspace(1)* addrspace(1)* %b, align 4 18 %ld.c.idx.1 = load i8 addrspace(1)*, i8 addrspace(1)* addrspace(1)* %b.1, align 4 20 store i8 addrspace(1)* null, i8 addrspace(1)* addrspace(1)* %a, align 4 21 store i8 addrspace(1)* null, i8 addrspace(1)* addrspace(1)* %a.1, align 4 28 ; CHECK: inttoptr i32 %{{[^ ]+}} to i8 addrspace(3)* [all …]
|
D | merge-stores.ll | 12 ; CHECK: store <2 x i8> <i8 -56, i8 123>, <2 x i8> addrspace(1)* %{{[0-9]+}}, align 2 13 define amdgpu_kernel void @merge_global_store_2_constants_i8(i8 addrspace(1)* %out) #0 { 14 %out.gep.1 = getelementptr i8, i8 addrspace(1)* %out, i32 1 16 store i8 123, i8 addrspace(1)* %out.gep.1 17 store i8 456, i8 addrspace(1)* %out, align 2 23 define amdgpu_kernel void @merge_global_store_2_constants_i8_natural_align(i8 addrspace(1)* %out) #… 24 %out.gep.1 = getelementptr i8, i8 addrspace(1)* %out, i32 1 26 store i8 123, i8 addrspace(1)* %out.gep.1 27 store i8 456, i8 addrspace(1)* %out 32 ; CHECK: store <2 x i16> <i16 456, i16 123>, <2 x i16> addrspace(1)* %{{[0-9]+}}, align 4 [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/InstSimplify/ |
D | simplify-nested-bitcast.ll | 2 %0 = type { i64, i64, i8 addrspace(1)*, i8 addrspace(1)* } 3 …8*, %struct.__block_descriptor addrspace(1)* }, %0, [17 x i8], { i8**, i32, i32, i8*, %struct.__bl… 5 %struct.__block_literal_generic = type { i8*, i32, i32, i8*, %struct.__block_descriptor addrspace(1… 7 @__aaa_struct_ptr = external addrspace(1) global %__aaa_struct 8 …addrspace(1)* } { i8** null, i32 1342177280, i32 0, i8* bitcast (i32 (i8 addrspace(4)*, i32 addrsp… 11 define i32 @bl0_block_invoke(i8 addrspace(4)* nocapture readnone, i32 addrspace(1)* nocapture reado… 13 %2 = load i32, i32 addrspace(1)* %1, align 4 19 define i32 @f0(i32 addrspace(1)*, i32 (i32 addrspace(1)*) addrspace(4)*) #1 { 21 …%block.literal = bitcast i32 (i32 addrspace(1)*) addrspace(4)* %1 to %struct.__block_literal_gener… 22 …bounds %struct.__block_literal_generic, %struct.__block_literal_generic addrspace(4)* %block.liter… [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/AMDGPU/ |
D | captured-frame-index.ll | 6 define amdgpu_kernel void @store_fi_lifetime(i32 addrspace(1)* %out, i32 %in) #0 { 8 %b = alloca i8, addrspace(5) 9 call void @llvm.lifetime.start.p5i8(i64 1, i8 addrspace(5)* %b) 10 store volatile i8 addrspace(5)* %b, i8 addrspace(5)* addrspace(1)* undef 11 call void @llvm.lifetime.end.p5i8(i64 1, i8 addrspace(5)* %b) 21 define amdgpu_kernel void @stored_fi_to_lds(float addrspace(5)* addrspace(3)* %ptr) #0 { 22 %tmp = alloca float, addrspace(5) 23 store float 4.0, float addrspace(5)*%tmp 24 store float addrspace(5)* %tmp, float addrspace(5)* addrspace(3)* %ptr 41 define amdgpu_kernel void @stored_fi_to_lds_2_small_objects(float addrspace(5)* addrspace(3)* %ptr)… [all …]
|
D | schedule-regpressure-limit.ll | 8 define amdgpu_kernel void @load_fma_store(float addrspace(3)* nocapture readonly %arg, float addrsp… 10 %tmp = getelementptr inbounds float, float addrspace(3)* %arg, i32 1 11 %tmp2 = load float, float addrspace(3)* %tmp, align 4 12 %tmp3 = getelementptr inbounds float, float addrspace(3)* %arg, i32 2 13 %tmp4 = load float, float addrspace(3)* %tmp3, align 4 14 %tmp5 = getelementptr inbounds float, float addrspace(3)* %arg, i32 3 15 %tmp6 = load float, float addrspace(3)* %tmp5, align 4 17 %tmp8 = getelementptr inbounds float, float addrspace(3)* %arg, i32 5 18 %tmp9 = load float, float addrspace(3)* %tmp8, align 4 19 %tmp10 = getelementptr inbounds float, float addrspace(3)* %arg, i32 6 [all …]
|
D | schedule-ilp.ll | 5 define amdgpu_kernel void @load_fma_store(float addrspace(3)* nocapture readonly %arg, float addrsp… 7 %tmp = getelementptr inbounds float, float addrspace(3)* %arg, i32 1 8 %tmp2 = load float, float addrspace(3)* %tmp, align 4 9 %tmp3 = getelementptr inbounds float, float addrspace(3)* %arg, i32 2 10 %tmp4 = load float, float addrspace(3)* %tmp3, align 4 11 %tmp5 = getelementptr inbounds float, float addrspace(3)* %arg, i32 3 12 %tmp6 = load float, float addrspace(3)* %tmp5, align 4 14 %tmp8 = getelementptr inbounds float, float addrspace(3)* %arg, i32 5 15 %tmp9 = load float, float addrspace(3)* %tmp8, align 4 16 %tmp10 = getelementptr inbounds float, float addrspace(3)* %arg, i32 6 [all …]
|
D | schedule-regpressure-limit3.ll | 7 define amdgpu_kernel void @load_fma_store(float addrspace(3)* nocapture readonly %arg, float addrsp… 9 %tmp = getelementptr inbounds float, float addrspace(3)* %arg, i32 1 10 %tmp2 = load float, float addrspace(3)* %tmp, align 4 11 %tmp3 = getelementptr inbounds float, float addrspace(3)* %arg, i32 2 12 %tmp4 = load float, float addrspace(3)* %tmp3, align 4 13 %tmp5 = getelementptr inbounds float, float addrspace(3)* %arg, i32 3 14 %tmp6 = load float, float addrspace(3)* %tmp5, align 4 16 %tmp8 = getelementptr inbounds float, float addrspace(3)* %arg, i32 5 17 %tmp9 = load float, float addrspace(3)* %tmp8, align 4 18 %tmp10 = getelementptr inbounds float, float addrspace(3)* %arg, i32 6 [all …]
|
D | addrspacecast-constantexpr.ll | 3 declare void @llvm.memcpy.p1i32.p4i32.i32(i32 addrspace(1)* nocapture, i32 addrspace(4)* nocapture,… 5 @lds.i32 = unnamed_addr addrspace(3) global i32 undef, align 4 6 @lds.arr = unnamed_addr addrspace(3) global [256 x i32] undef, align 4 8 @global.i32 = unnamed_addr addrspace(1) global i32 undef, align 4 9 @global.arr = unnamed_addr addrspace(1) global [256 x i32] undef, align 4 13 store i32 7, i32 addrspace(3)* addrspacecast (i32 addrspace(4)* null to i32 addrspace(3)*) 19 store i32 7, i32 addrspace(4)* addrspacecast (i32 addrspace(3)* null to i32 addrspace(4)*) 25 store i32 7, i32 addrspace(4)* addrspacecast (i32 addrspace(3)* @lds.i32 to i32 addrspace(4)*) 31 … i32 addrspace(4)* getelementptr ([256 x i32], [256 x i32] addrspace(4)* addrspacecast ([256 x i32… 37 store i32 7, i32 addrspace(4)* addrspacecast (i32 addrspace(1)* @global.i32 to i32 addrspace(4)*) [all …]
|
D | merge-stores.ll | 15 define amdgpu_kernel void @merge_global_store_2_constants_i8(i8 addrspace(1)* %out) #0 { 16 %out.gep.1 = getelementptr i8, i8 addrspace(1)* %out, i32 1 18 store i8 123, i8 addrspace(1)* %out.gep.1 19 store i8 456, i8 addrspace(1)* %out, align 2 27 define amdgpu_kernel void @merge_global_store_2_constants_i8_natural_align(i8 addrspace(1)* %out) #… 28 %out.gep.1 = getelementptr i8, i8 addrspace(1)* %out, i32 1 30 store i8 123, i8 addrspace(1)* %out.gep.1 31 store i8 456, i8 addrspace(1)* %out 37 define amdgpu_kernel void @merge_global_store_2_constants_i16(i16 addrspace(1)* %out) #0 { 38 %out.gep.1 = getelementptr i16, i16 addrspace(1)* %out, i32 1 [all …]
|
D | image-resource-id.ll | 10 define amdgpu_kernel void @test_2d_rd_1_0(%opencl.image2d_t addrspace(1)* %in, ; read_only 11 i32 addrspace(1)* %out) { 14 %opencl.image2d_t addrspace(1)* %in) #0 15 store i32 %0, i32 addrspace(1)* %out 24 define amdgpu_kernel void @test_3d_rd_1_0(%opencl.image3d_t addrspace(1)* %in, ; read_only 25 i32 addrspace(1)* %out) { 28 %opencl.image3d_t addrspace(1)* %in) #0 29 store i32 %0, i32 addrspace(1)* %out 40 define amdgpu_kernel void @test_2d_wr_1_0(%opencl.image2d_t addrspace(1)* %in, ; write_only 41 i32 addrspace(1)* %out) { [all …]
|
D | enqueue-kernel.ll | 3 ; CHECK: @__test_block_invoke_kernel.runtime_handle = addrspace(1) global [2 x i64] zeroinitializer 4 ; CHECK: @__test_block_invoke_2_kernel.runtime_handle = addrspace(1) global [2 x i64] zeroinitializ… 5 ; CHECK: @__amdgpu_enqueued_kernel.runtime_handle = addrspace(1) global [2 x i64] zeroinitializer 6 ; CHECK: @__amdgpu_enqueued_kernel.1.runtime_handle = addrspace(1) global [2 x i64] zeroinitializer 13 define amdgpu_kernel void @non_caller(i8 addrspace(1)* %a, i8 %b, i64 addrspace(1)* %c, i64 %d) loc… 20 define amdgpu_kernel void @caller_indirect(i8 addrspace(1)* %a, i8 %b, i64 addrspace(1)* %c, i64 %d… 22 call void @caller(i8 addrspace(1)* %a, i8 %b, i64 addrspace(1)* %c, i64 %d) 39 define amdgpu_kernel void @caller(i8 addrspace(1)* %a, i8 %b, i64 addrspace(1)* %c, i64 %d) local_u… 42 %block = alloca <{ i32, i32, i8 addrspace(1)*, i8 }>, align 8, addrspace(5) 43 %tmp = alloca %struct.ndrange_t, align 4, addrspace(5) [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/InstCombine/ |
D | constant-fold-address-space-pointer.ll | 4 @g = addrspace(3) global i32 89 6 @const_zero_i8_as1 = addrspace(1) constant i8 0 7 @const_zero_i32_as1 = addrspace(1) constant i32 0 9 @const_zero_i8_as2 = addrspace(2) constant i8 0 10 @const_zero_i32_as2 = addrspace(2) constant i32 0 12 @const_zero_i8_as3 = addrspace(3) constant i8 0 13 @const_zero_i32_as3 = addrspace(3) constant i32 0 17 define i32 addrspace(3)* @test_constant_fold_inttoptr_as_pointer_same_size() { 19 ; CHECK-NEXT: ret i32 addrspace(3)* @const_zero_i32_as3 20 %x = ptrtoint i32 addrspace(3)* @const_zero_i32_as3 to i32 [all …]
|
D | addrspacecast.ll | 7 declare void @llvm.memcpy.p0i8.p1i8.i32(i8*, i8 addrspace(1)*, i32, i1) nounwind 8 declare void @llvm.memcpy.p0i8.p2i8.i32(i8*, i8 addrspace(2)*, i32, i1) nounwind 11 define i32* @combine_redundant_addrspacecast(i32 addrspace(1)* %x) nounwind { 13 ; CHECK: addrspacecast i32 addrspace(1)* %x to i32* 15 %y = addrspacecast i32 addrspace(1)* %x to i32 addrspace(3)* 16 %z = addrspacecast i32 addrspace(3)* %y to i32* 20 define <4 x i32*> @combine_redundant_addrspacecast_vector(<4 x i32 addrspace(1)*> %x) nounwind { 22 ; CHECK: addrspacecast <4 x i32 addrspace(1)*> %x to <4 x i32*> 24 %y = addrspacecast <4 x i32 addrspace(1)*> %x to <4 x i32 addrspace(3)*> 25 %z = addrspacecast <4 x i32 addrspace(3)*> %y to <4 x i32*> [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/LoopVectorize/X86/ |
D | pr23997.ll | 8 define void @foo(i8 addrspace(1)* align 8 dereferenceable_or_null(16), i8 addrspace(1)* align 8 der… 13 ; CHECK-NEXT: [[DOT10:%.*]] = getelementptr inbounds i8, i8 addrspace(1)* [[TMP0:%.*]], i64 16 14 ; CHECK-NEXT: [[DOT11:%.*]] = bitcast i8 addrspace(1)* [[DOT10]] to i8 addrspace(1)* addrspace(1… 15 ; CHECK-NEXT: [[DOT12:%.*]] = getelementptr inbounds i8, i8 addrspace(1)* [[TMP1:%.*]], i64 16 16 ; CHECK-NEXT: [[DOT13:%.*]] = bitcast i8 addrspace(1)* [[DOT12]] to i8 addrspace(1)* addrspace(1… 26 ; CHECK-NEXT: [[SCEVGEP:%.*]] = getelementptr i8, i8 addrspace(1)* [[TMP0]], i64 [[TMP6]] 27 ; CHECK-NEXT: [[SCEVGEP2:%.*]] = getelementptr i8, i8 addrspace(1)* [[TMP1]], i64 [[TMP6]] 28 ; CHECK-NEXT: [[BOUND0:%.*]] = icmp ult i8 addrspace(1)* [[DOT10]], [[SCEVGEP2]] 29 ; CHECK-NEXT: [[BOUND1:%.*]] = icmp ult i8 addrspace(1)* [[DOT12]], [[SCEVGEP]] 37 ; CHECK-NEXT: [[TMP7:%.*]] = getelementptr inbounds i8 addrspace(1)*, i8 addrspace(1)* addrspace… [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/SROA/ |
D | alloca-address-space.ll | 4 declare void @llvm.memcpy.p2i8.p2i8.i32(i8 addrspace(2)* nocapture, i8 addrspace(2)* nocapture read… 5 declare void @llvm.memcpy.p1i8.p2i8.i32(i8 addrspace(1)* nocapture, i8 addrspace(2)* nocapture read… 6 declare void @llvm.memcpy.p2i8.p1i8.i32(i8 addrspace(2)* nocapture, i8 addrspace(1)* nocapture read… 7 declare void @llvm.memcpy.p1i8.p1i8.i32(i8 addrspace(1)* nocapture, i8 addrspace(1)* nocapture read… 12 ; CHECK: load <2 x i64>, <2 x i64> addrspace(1)* %a, align 2 13 ; CHECK: store <2 x i64> {{.*}}, <2 x i64> addrspace(1)* {{.*}}, align 2 15 define void @test_address_space_1_1(<2 x i64> addrspace(1)* %a, i16 addrspace(1)* %b) { 16 %aa = alloca <2 x i64>, align 16, addrspace(2) 17 %aptr = bitcast <2 x i64> addrspace(1)* %a to i8 addrspace(1)* 18 %aaptr = bitcast <2 x i64> addrspace(2)* %aa to i8 addrspace(2)* [all …]
|
/external/llvm/test/Transforms/InstCombine/ |
D | addrspacecast.ll | 7 declare void @llvm.memcpy.p0i8.p1i8.i32(i8*, i8 addrspace(1)*, i32, i32, i1) nounwind 8 declare void @llvm.memcpy.p0i8.p2i8.i32(i8*, i8 addrspace(2)*, i32, i32, i1) nounwind 11 define i32* @combine_redundant_addrspacecast(i32 addrspace(1)* %x) nounwind { 13 ; CHECK: addrspacecast i32 addrspace(1)* %x to i32* 15 %y = addrspacecast i32 addrspace(1)* %x to i32 addrspace(3)* 16 %z = addrspacecast i32 addrspace(3)* %y to i32* 20 define <4 x i32*> @combine_redundant_addrspacecast_vector(<4 x i32 addrspace(1)*> %x) nounwind { 22 ; CHECK: addrspacecast <4 x i32 addrspace(1)*> %x to <4 x i32*> 24 %y = addrspacecast <4 x i32 addrspace(1)*> %x to <4 x i32 addrspace(3)*> 25 %z = addrspacecast <4 x i32 addrspace(3)*> %y to <4 x i32*> [all …]
|
D | constant-fold-address-space-pointer.ll | 4 @g = addrspace(3) global i32 89 6 @const_zero_i8_as1 = addrspace(1) constant i8 0 7 @const_zero_i32_as1 = addrspace(1) constant i32 0 9 @const_zero_i8_as2 = addrspace(2) constant i8 0 10 @const_zero_i32_as2 = addrspace(2) constant i32 0 12 @const_zero_i8_as3 = addrspace(3) constant i8 0 13 @const_zero_i32_as3 = addrspace(3) constant i32 0 17 define i32 addrspace(3)* @test_constant_fold_inttoptr_as_pointer_same_size() { 19 ; CHECK-NEXT: ret i32 addrspace(3)* @const_zero_i32_as3 20 %x = ptrtoint i32 addrspace(3)* @const_zero_i32_as3 to i32 [all …]
|
/external/llvm/test/CodeGen/AMDGPU/ |
D | addrspacecast-constantexpr.ll | 3 declare void @llvm.memcpy.p1i32.p4i32.i32(i32 addrspace(1)* nocapture, i32 addrspace(4)* nocapture,… 5 @lds.i32 = unnamed_addr addrspace(3) global i32 undef, align 4 6 @lds.arr = unnamed_addr addrspace(3) global [256 x i32] undef, align 4 8 @global.i32 = unnamed_addr addrspace(1) global i32 undef, align 4 9 @global.arr = unnamed_addr addrspace(1) global [256 x i32] undef, align 4 13 store i32 7, i32 addrspace(3)* addrspacecast (i32 addrspace(4)* null to i32 addrspace(3)*) 19 store i32 7, i32 addrspace(4)* addrspacecast (i32 addrspace(3)* null to i32 addrspace(4)*) 25 store i32 7, i32 addrspace(4)* addrspacecast (i32 addrspace(3)* @lds.i32 to i32 addrspace(4)*) 31 … i32 addrspace(4)* getelementptr ([256 x i32], [256 x i32] addrspace(4)* addrspacecast ([256 x i32… 37 store i32 7, i32 addrspace(4)* addrspacecast (i32 addrspace(1)* @global.i32 to i32 addrspace(4)*) [all …]
|
D | merge-stores.ll | 18 define void @merge_global_store_2_constants_i8(i8 addrspace(1)* %out) #0 { 19 %out.gep.1 = getelementptr i8, i8 addrspace(1)* %out, i32 1 21 store i8 123, i8 addrspace(1)* %out.gep.1 22 store i8 456, i8 addrspace(1)* %out, align 2 30 define void @merge_global_store_2_constants_i8_natural_align(i8 addrspace(1)* %out) #0 { 31 %out.gep.1 = getelementptr i8, i8 addrspace(1)* %out, i32 1 33 store i8 123, i8 addrspace(1)* %out.gep.1 34 store i8 456, i8 addrspace(1)* %out 40 define void @merge_global_store_2_constants_i16(i16 addrspace(1)* %out) #0 { 41 %out.gep.1 = getelementptr i16, i16 addrspace(1)* %out, i32 1 [all …]
|
D | lds-alignment.ll | 3 @lds.align16.0 = internal unnamed_addr addrspace(3) global [38 x i8] undef, align 16 4 @lds.align16.1 = internal unnamed_addr addrspace(3) global [38 x i8] undef, align 16 6 @lds.align8.0 = internal unnamed_addr addrspace(3) global [38 x i8] undef, align 8 7 @lds.align32.0 = internal unnamed_addr addrspace(3) global [38 x i8] undef, align 32 9 @lds.missing.align.0 = internal unnamed_addr addrspace(3) global [39 x i32] undef 10 @lds.missing.align.1 = internal unnamed_addr addrspace(3) global [7 x i64] undef 12 declare void @llvm.memcpy.p3i8.p1i8.i32(i8 addrspace(3)* nocapture, i8 addrspace(1)* nocapture read… 13 declare void @llvm.memcpy.p1i8.p3i8.i32(i8 addrspace(1)* nocapture, i8 addrspace(3)* nocapture read… 18 define void @test_no_round_size_1(i8 addrspace(1)* %out, i8 addrspace(1)* %in) #1 { 19 %lds.align16.0.bc = bitcast [38 x i8] addrspace(3)* @lds.align16.0 to i8 addrspace(3)* [all …]
|
D | image-resource-id.ll | 10 define void @test_2d_rd_1_0(%opencl.image2d_t addrspace(1)* %in, ; read_only 11 i32 addrspace(1)* %out) { 14 %opencl.image2d_t addrspace(1)* %in) #0 15 store i32 %0, i32 addrspace(1)* %out 24 define void @test_3d_rd_1_0(%opencl.image3d_t addrspace(1)* %in, ; read_only 25 i32 addrspace(1)* %out) { 28 %opencl.image3d_t addrspace(1)* %in) #0 29 store i32 %0, i32 addrspace(1)* %out 40 define void @test_2d_wr_1_0(%opencl.image2d_t addrspace(1)* %in, ; write_only 41 i32 addrspace(1)* %out) { [all …]
|
/external/llvm/test/Assembler/ |
D | 2007-12-11-AddressSpaces.ll | 1 ; RUN: llvm-as < %s | llvm-dis | grep "addrspace(33)" | count 7 2 ; RUN: llvm-as < %s | llvm-dis | grep "addrspace(42)" | count 2 3 ; RUN: llvm-as < %s | llvm-dis | grep "addrspace(66)" | count 2 4 ; RUN: llvm-as < %s | llvm-dis | grep "addrspace(11)" | count 6 5 ; RUN: llvm-as < %s | llvm-dis | grep "addrspace(22)" | count 5 8 %struct.mystruct = type { i32, i32 addrspace(33)*, i32, i32 addrspace(33)* } 9 @input = weak addrspace(42) global %struct.mystruct zeroinitializer ; <%struct.mystruct addrspac… 10 @output = addrspace(66) global %struct.mystruct zeroinitializer ; <%struct.mystruct addrspace(66)… 11 @y = external addrspace(33) global i32 addrspace(11)* addrspace(22)* ; <i32 addrspace(11)* addrsp… 15 …addrspace(33)*, i32 addrspace(33)* addrspace(42)* getelementptr (%struct.mystruct, %struct.mystruc… [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Assembler/ |
D | 2007-12-11-AddressSpaces.ll | 1 ; RUN: llvm-as < %s | llvm-dis | grep "addrspace(33)" | count 7 2 ; RUN: llvm-as < %s | llvm-dis | grep "addrspace(42)" | count 2 3 ; RUN: llvm-as < %s | llvm-dis | grep "addrspace(66)" | count 2 4 ; RUN: llvm-as < %s | llvm-dis | grep "addrspace(11)" | count 6 5 ; RUN: llvm-as < %s | llvm-dis | grep "addrspace(22)" | count 5 8 %struct.mystruct = type { i32, i32 addrspace(33)*, i32, i32 addrspace(33)* } 9 @input = weak addrspace(42) global %struct.mystruct zeroinitializer ; <%struct.mystruct addrspac… 10 @output = addrspace(66) global %struct.mystruct zeroinitializer ; <%struct.mystruct addrspace(66)… 11 @y = external addrspace(33) global i32 addrspace(11)* addrspace(22)* ; <i32 addrspace(11)* addrsp… 15 …addrspace(33)*, i32 addrspace(33)* addrspace(42)* getelementptr (%struct.mystruct, %struct.mystruc… [all …]
|