Home
last modified time | relevance | path

Searched refs:addrspace (Results 1 – 25 of 2245) sorted by relevance

12345678910>>...90

/external/llvm/test/Transforms/LoadStoreVectorizer/AMDGPU/
Dpointer-elements.ll9 ; CHECK: inttoptr i64 %{{[0-9]+}} to i8 addrspace(1)*
10 ; CHECK: inttoptr i64 %{{[0-9]+}} to i8 addrspace(1)*
12 define void @merge_v2p1i8(i8 addrspace(1)* addrspace(1)* nocapture %a, i8 addrspace(1)* addrspace(1…
14 %a.1 = getelementptr inbounds i8 addrspace(1)*, i8 addrspace(1)* addrspace(1)* %a, i64 1
15 %b.1 = getelementptr inbounds i8 addrspace(1)*, i8 addrspace(1)* addrspace(1)* %b, i64 1
17 %ld.c = load i8 addrspace(1)*, i8 addrspace(1)* addrspace(1)* %b, align 4
18 %ld.c.idx.1 = load i8 addrspace(1)*, i8 addrspace(1)* addrspace(1)* %b.1, align 4
20 store i8 addrspace(1)* null, i8 addrspace(1)* addrspace(1)* %a, align 4
21 store i8 addrspace(1)* null, i8 addrspace(1)* addrspace(1)* %a.1, align 4
28 ; CHECK: inttoptr i32 %{{[0-9]+}} to i8 addrspace(3)*
[all …]
Dmerge-stores.ll12 ; CHECK: store <2 x i8> <i8 -56, i8 123>, <2 x i8> addrspace(1)* %{{[0-9]+}}, align 2
13 define void @merge_global_store_2_constants_i8(i8 addrspace(1)* %out) #0 {
14 %out.gep.1 = getelementptr i8, i8 addrspace(1)* %out, i32 1
16 store i8 123, i8 addrspace(1)* %out.gep.1
17 store i8 456, i8 addrspace(1)* %out, align 2
23 define void @merge_global_store_2_constants_i8_natural_align(i8 addrspace(1)* %out) #0 {
24 %out.gep.1 = getelementptr i8, i8 addrspace(1)* %out, i32 1
26 store i8 123, i8 addrspace(1)* %out.gep.1
27 store i8 456, i8 addrspace(1)* %out
32 ; CHECK: store <2 x i16> <i16 456, i16 123>, <2 x i16> addrspace(1)* %{{[0-9]+}}, align 4
[all …]
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/LoadStoreVectorizer/AMDGPU/
Dpointer-elements.ll9 ; CHECK: inttoptr i64 %{{[^ ]+}} to i8 addrspace(1)*
10 ; CHECK: inttoptr i64 %{{[^ ]+}} to i8 addrspace(1)*
12 …efine amdgpu_kernel void @merge_v2p1i8(i8 addrspace(1)* addrspace(1)* nocapture %a, i8 addrspace(1…
14 %a.1 = getelementptr inbounds i8 addrspace(1)*, i8 addrspace(1)* addrspace(1)* %a, i64 1
15 %b.1 = getelementptr inbounds i8 addrspace(1)*, i8 addrspace(1)* addrspace(1)* %b, i64 1
17 %ld.c = load i8 addrspace(1)*, i8 addrspace(1)* addrspace(1)* %b, align 4
18 %ld.c.idx.1 = load i8 addrspace(1)*, i8 addrspace(1)* addrspace(1)* %b.1, align 4
20 store i8 addrspace(1)* null, i8 addrspace(1)* addrspace(1)* %a, align 4
21 store i8 addrspace(1)* null, i8 addrspace(1)* addrspace(1)* %a.1, align 4
28 ; CHECK: inttoptr i32 %{{[^ ]+}} to i8 addrspace(3)*
[all …]
Dmerge-stores.ll12 ; CHECK: store <2 x i8> <i8 -56, i8 123>, <2 x i8> addrspace(1)* %{{[0-9]+}}, align 2
13 define amdgpu_kernel void @merge_global_store_2_constants_i8(i8 addrspace(1)* %out) #0 {
14 %out.gep.1 = getelementptr i8, i8 addrspace(1)* %out, i32 1
16 store i8 123, i8 addrspace(1)* %out.gep.1
17 store i8 456, i8 addrspace(1)* %out, align 2
23 define amdgpu_kernel void @merge_global_store_2_constants_i8_natural_align(i8 addrspace(1)* %out) #…
24 %out.gep.1 = getelementptr i8, i8 addrspace(1)* %out, i32 1
26 store i8 123, i8 addrspace(1)* %out.gep.1
27 store i8 456, i8 addrspace(1)* %out
32 ; CHECK: store <2 x i16> <i16 456, i16 123>, <2 x i16> addrspace(1)* %{{[0-9]+}}, align 4
[all …]
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/InstSimplify/
Dsimplify-nested-bitcast.ll2 %0 = type { i64, i64, i8 addrspace(1)*, i8 addrspace(1)* }
3 …8*, %struct.__block_descriptor addrspace(1)* }, %0, [17 x i8], { i8**, i32, i32, i8*, %struct.__bl…
5 %struct.__block_literal_generic = type { i8*, i32, i32, i8*, %struct.__block_descriptor addrspace(1…
7 @__aaa_struct_ptr = external addrspace(1) global %__aaa_struct
8addrspace(1)* } { i8** null, i32 1342177280, i32 0, i8* bitcast (i32 (i8 addrspace(4)*, i32 addrsp…
11 define i32 @bl0_block_invoke(i8 addrspace(4)* nocapture readnone, i32 addrspace(1)* nocapture reado…
13 %2 = load i32, i32 addrspace(1)* %1, align 4
19 define i32 @f0(i32 addrspace(1)*, i32 (i32 addrspace(1)*) addrspace(4)*) #1 {
21 …%block.literal = bitcast i32 (i32 addrspace(1)*) addrspace(4)* %1 to %struct.__block_literal_gener…
22 …bounds %struct.__block_literal_generic, %struct.__block_literal_generic addrspace(4)* %block.liter…
[all …]
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/AMDGPU/
Dcaptured-frame-index.ll6 define amdgpu_kernel void @store_fi_lifetime(i32 addrspace(1)* %out, i32 %in) #0 {
8 %b = alloca i8, addrspace(5)
9 call void @llvm.lifetime.start.p5i8(i64 1, i8 addrspace(5)* %b)
10 store volatile i8 addrspace(5)* %b, i8 addrspace(5)* addrspace(1)* undef
11 call void @llvm.lifetime.end.p5i8(i64 1, i8 addrspace(5)* %b)
21 define amdgpu_kernel void @stored_fi_to_lds(float addrspace(5)* addrspace(3)* %ptr) #0 {
22 %tmp = alloca float, addrspace(5)
23 store float 4.0, float addrspace(5)*%tmp
24 store float addrspace(5)* %tmp, float addrspace(5)* addrspace(3)* %ptr
41 define amdgpu_kernel void @stored_fi_to_lds_2_small_objects(float addrspace(5)* addrspace(3)* %ptr)…
[all …]
Dschedule-regpressure-limit.ll8 define amdgpu_kernel void @load_fma_store(float addrspace(3)* nocapture readonly %arg, float addrsp…
10 %tmp = getelementptr inbounds float, float addrspace(3)* %arg, i32 1
11 %tmp2 = load float, float addrspace(3)* %tmp, align 4
12 %tmp3 = getelementptr inbounds float, float addrspace(3)* %arg, i32 2
13 %tmp4 = load float, float addrspace(3)* %tmp3, align 4
14 %tmp5 = getelementptr inbounds float, float addrspace(3)* %arg, i32 3
15 %tmp6 = load float, float addrspace(3)* %tmp5, align 4
17 %tmp8 = getelementptr inbounds float, float addrspace(3)* %arg, i32 5
18 %tmp9 = load float, float addrspace(3)* %tmp8, align 4
19 %tmp10 = getelementptr inbounds float, float addrspace(3)* %arg, i32 6
[all …]
Dschedule-ilp.ll5 define amdgpu_kernel void @load_fma_store(float addrspace(3)* nocapture readonly %arg, float addrsp…
7 %tmp = getelementptr inbounds float, float addrspace(3)* %arg, i32 1
8 %tmp2 = load float, float addrspace(3)* %tmp, align 4
9 %tmp3 = getelementptr inbounds float, float addrspace(3)* %arg, i32 2
10 %tmp4 = load float, float addrspace(3)* %tmp3, align 4
11 %tmp5 = getelementptr inbounds float, float addrspace(3)* %arg, i32 3
12 %tmp6 = load float, float addrspace(3)* %tmp5, align 4
14 %tmp8 = getelementptr inbounds float, float addrspace(3)* %arg, i32 5
15 %tmp9 = load float, float addrspace(3)* %tmp8, align 4
16 %tmp10 = getelementptr inbounds float, float addrspace(3)* %arg, i32 6
[all …]
Dschedule-regpressure-limit3.ll7 define amdgpu_kernel void @load_fma_store(float addrspace(3)* nocapture readonly %arg, float addrsp…
9 %tmp = getelementptr inbounds float, float addrspace(3)* %arg, i32 1
10 %tmp2 = load float, float addrspace(3)* %tmp, align 4
11 %tmp3 = getelementptr inbounds float, float addrspace(3)* %arg, i32 2
12 %tmp4 = load float, float addrspace(3)* %tmp3, align 4
13 %tmp5 = getelementptr inbounds float, float addrspace(3)* %arg, i32 3
14 %tmp6 = load float, float addrspace(3)* %tmp5, align 4
16 %tmp8 = getelementptr inbounds float, float addrspace(3)* %arg, i32 5
17 %tmp9 = load float, float addrspace(3)* %tmp8, align 4
18 %tmp10 = getelementptr inbounds float, float addrspace(3)* %arg, i32 6
[all …]
Daddrspacecast-constantexpr.ll3 declare void @llvm.memcpy.p1i32.p4i32.i32(i32 addrspace(1)* nocapture, i32 addrspace(4)* nocapture,…
5 @lds.i32 = unnamed_addr addrspace(3) global i32 undef, align 4
6 @lds.arr = unnamed_addr addrspace(3) global [256 x i32] undef, align 4
8 @global.i32 = unnamed_addr addrspace(1) global i32 undef, align 4
9 @global.arr = unnamed_addr addrspace(1) global [256 x i32] undef, align 4
13 store i32 7, i32 addrspace(3)* addrspacecast (i32 addrspace(4)* null to i32 addrspace(3)*)
19 store i32 7, i32 addrspace(4)* addrspacecast (i32 addrspace(3)* null to i32 addrspace(4)*)
25 store i32 7, i32 addrspace(4)* addrspacecast (i32 addrspace(3)* @lds.i32 to i32 addrspace(4)*)
31 … i32 addrspace(4)* getelementptr ([256 x i32], [256 x i32] addrspace(4)* addrspacecast ([256 x i32…
37 store i32 7, i32 addrspace(4)* addrspacecast (i32 addrspace(1)* @global.i32 to i32 addrspace(4)*)
[all …]
Dmerge-stores.ll15 define amdgpu_kernel void @merge_global_store_2_constants_i8(i8 addrspace(1)* %out) #0 {
16 %out.gep.1 = getelementptr i8, i8 addrspace(1)* %out, i32 1
18 store i8 123, i8 addrspace(1)* %out.gep.1
19 store i8 456, i8 addrspace(1)* %out, align 2
27 define amdgpu_kernel void @merge_global_store_2_constants_i8_natural_align(i8 addrspace(1)* %out) #…
28 %out.gep.1 = getelementptr i8, i8 addrspace(1)* %out, i32 1
30 store i8 123, i8 addrspace(1)* %out.gep.1
31 store i8 456, i8 addrspace(1)* %out
37 define amdgpu_kernel void @merge_global_store_2_constants_i16(i16 addrspace(1)* %out) #0 {
38 %out.gep.1 = getelementptr i16, i16 addrspace(1)* %out, i32 1
[all …]
Dimage-resource-id.ll10 define amdgpu_kernel void @test_2d_rd_1_0(%opencl.image2d_t addrspace(1)* %in, ; read_only
11 i32 addrspace(1)* %out) {
14 %opencl.image2d_t addrspace(1)* %in) #0
15 store i32 %0, i32 addrspace(1)* %out
24 define amdgpu_kernel void @test_3d_rd_1_0(%opencl.image3d_t addrspace(1)* %in, ; read_only
25 i32 addrspace(1)* %out) {
28 %opencl.image3d_t addrspace(1)* %in) #0
29 store i32 %0, i32 addrspace(1)* %out
40 define amdgpu_kernel void @test_2d_wr_1_0(%opencl.image2d_t addrspace(1)* %in, ; write_only
41 i32 addrspace(1)* %out) {
[all …]
Denqueue-kernel.ll3 ; CHECK: @__test_block_invoke_kernel.runtime_handle = addrspace(1) global [2 x i64] zeroinitializer
4 ; CHECK: @__test_block_invoke_2_kernel.runtime_handle = addrspace(1) global [2 x i64] zeroinitializ…
5 ; CHECK: @__amdgpu_enqueued_kernel.runtime_handle = addrspace(1) global [2 x i64] zeroinitializer
6 ; CHECK: @__amdgpu_enqueued_kernel.1.runtime_handle = addrspace(1) global [2 x i64] zeroinitializer
13 define amdgpu_kernel void @non_caller(i8 addrspace(1)* %a, i8 %b, i64 addrspace(1)* %c, i64 %d) loc…
20 define amdgpu_kernel void @caller_indirect(i8 addrspace(1)* %a, i8 %b, i64 addrspace(1)* %c, i64 %d…
22 call void @caller(i8 addrspace(1)* %a, i8 %b, i64 addrspace(1)* %c, i64 %d)
39 define amdgpu_kernel void @caller(i8 addrspace(1)* %a, i8 %b, i64 addrspace(1)* %c, i64 %d) local_u…
42 %block = alloca <{ i32, i32, i8 addrspace(1)*, i8 }>, align 8, addrspace(5)
43 %tmp = alloca %struct.ndrange_t, align 4, addrspace(5)
[all …]
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/InstCombine/
Dconstant-fold-address-space-pointer.ll4 @g = addrspace(3) global i32 89
6 @const_zero_i8_as1 = addrspace(1) constant i8 0
7 @const_zero_i32_as1 = addrspace(1) constant i32 0
9 @const_zero_i8_as2 = addrspace(2) constant i8 0
10 @const_zero_i32_as2 = addrspace(2) constant i32 0
12 @const_zero_i8_as3 = addrspace(3) constant i8 0
13 @const_zero_i32_as3 = addrspace(3) constant i32 0
17 define i32 addrspace(3)* @test_constant_fold_inttoptr_as_pointer_same_size() {
19 ; CHECK-NEXT: ret i32 addrspace(3)* @const_zero_i32_as3
20 %x = ptrtoint i32 addrspace(3)* @const_zero_i32_as3 to i32
[all …]
Daddrspacecast.ll7 declare void @llvm.memcpy.p0i8.p1i8.i32(i8*, i8 addrspace(1)*, i32, i1) nounwind
8 declare void @llvm.memcpy.p0i8.p2i8.i32(i8*, i8 addrspace(2)*, i32, i1) nounwind
11 define i32* @combine_redundant_addrspacecast(i32 addrspace(1)* %x) nounwind {
13 ; CHECK: addrspacecast i32 addrspace(1)* %x to i32*
15 %y = addrspacecast i32 addrspace(1)* %x to i32 addrspace(3)*
16 %z = addrspacecast i32 addrspace(3)* %y to i32*
20 define <4 x i32*> @combine_redundant_addrspacecast_vector(<4 x i32 addrspace(1)*> %x) nounwind {
22 ; CHECK: addrspacecast <4 x i32 addrspace(1)*> %x to <4 x i32*>
24 %y = addrspacecast <4 x i32 addrspace(1)*> %x to <4 x i32 addrspace(3)*>
25 %z = addrspacecast <4 x i32 addrspace(3)*> %y to <4 x i32*>
[all …]
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/LoopVectorize/X86/
Dpr23997.ll8 define void @foo(i8 addrspace(1)* align 8 dereferenceable_or_null(16), i8 addrspace(1)* align 8 der…
13 ; CHECK-NEXT: [[DOT10:%.*]] = getelementptr inbounds i8, i8 addrspace(1)* [[TMP0:%.*]], i64 16
14 ; CHECK-NEXT: [[DOT11:%.*]] = bitcast i8 addrspace(1)* [[DOT10]] to i8 addrspace(1)* addrspace(1…
15 ; CHECK-NEXT: [[DOT12:%.*]] = getelementptr inbounds i8, i8 addrspace(1)* [[TMP1:%.*]], i64 16
16 ; CHECK-NEXT: [[DOT13:%.*]] = bitcast i8 addrspace(1)* [[DOT12]] to i8 addrspace(1)* addrspace(1…
26 ; CHECK-NEXT: [[SCEVGEP:%.*]] = getelementptr i8, i8 addrspace(1)* [[TMP0]], i64 [[TMP6]]
27 ; CHECK-NEXT: [[SCEVGEP2:%.*]] = getelementptr i8, i8 addrspace(1)* [[TMP1]], i64 [[TMP6]]
28 ; CHECK-NEXT: [[BOUND0:%.*]] = icmp ult i8 addrspace(1)* [[DOT10]], [[SCEVGEP2]]
29 ; CHECK-NEXT: [[BOUND1:%.*]] = icmp ult i8 addrspace(1)* [[DOT12]], [[SCEVGEP]]
37 ; CHECK-NEXT: [[TMP7:%.*]] = getelementptr inbounds i8 addrspace(1)*, i8 addrspace(1)* addrspace
[all …]
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/SROA/
Dalloca-address-space.ll4 declare void @llvm.memcpy.p2i8.p2i8.i32(i8 addrspace(2)* nocapture, i8 addrspace(2)* nocapture read…
5 declare void @llvm.memcpy.p1i8.p2i8.i32(i8 addrspace(1)* nocapture, i8 addrspace(2)* nocapture read…
6 declare void @llvm.memcpy.p2i8.p1i8.i32(i8 addrspace(2)* nocapture, i8 addrspace(1)* nocapture read…
7 declare void @llvm.memcpy.p1i8.p1i8.i32(i8 addrspace(1)* nocapture, i8 addrspace(1)* nocapture read…
12 ; CHECK: load <2 x i64>, <2 x i64> addrspace(1)* %a, align 2
13 ; CHECK: store <2 x i64> {{.*}}, <2 x i64> addrspace(1)* {{.*}}, align 2
15 define void @test_address_space_1_1(<2 x i64> addrspace(1)* %a, i16 addrspace(1)* %b) {
16 %aa = alloca <2 x i64>, align 16, addrspace(2)
17 %aptr = bitcast <2 x i64> addrspace(1)* %a to i8 addrspace(1)*
18 %aaptr = bitcast <2 x i64> addrspace(2)* %aa to i8 addrspace(2)*
[all …]
/external/llvm/test/Transforms/InstCombine/
Daddrspacecast.ll7 declare void @llvm.memcpy.p0i8.p1i8.i32(i8*, i8 addrspace(1)*, i32, i32, i1) nounwind
8 declare void @llvm.memcpy.p0i8.p2i8.i32(i8*, i8 addrspace(2)*, i32, i32, i1) nounwind
11 define i32* @combine_redundant_addrspacecast(i32 addrspace(1)* %x) nounwind {
13 ; CHECK: addrspacecast i32 addrspace(1)* %x to i32*
15 %y = addrspacecast i32 addrspace(1)* %x to i32 addrspace(3)*
16 %z = addrspacecast i32 addrspace(3)* %y to i32*
20 define <4 x i32*> @combine_redundant_addrspacecast_vector(<4 x i32 addrspace(1)*> %x) nounwind {
22 ; CHECK: addrspacecast <4 x i32 addrspace(1)*> %x to <4 x i32*>
24 %y = addrspacecast <4 x i32 addrspace(1)*> %x to <4 x i32 addrspace(3)*>
25 %z = addrspacecast <4 x i32 addrspace(3)*> %y to <4 x i32*>
[all …]
Dconstant-fold-address-space-pointer.ll4 @g = addrspace(3) global i32 89
6 @const_zero_i8_as1 = addrspace(1) constant i8 0
7 @const_zero_i32_as1 = addrspace(1) constant i32 0
9 @const_zero_i8_as2 = addrspace(2) constant i8 0
10 @const_zero_i32_as2 = addrspace(2) constant i32 0
12 @const_zero_i8_as3 = addrspace(3) constant i8 0
13 @const_zero_i32_as3 = addrspace(3) constant i32 0
17 define i32 addrspace(3)* @test_constant_fold_inttoptr_as_pointer_same_size() {
19 ; CHECK-NEXT: ret i32 addrspace(3)* @const_zero_i32_as3
20 %x = ptrtoint i32 addrspace(3)* @const_zero_i32_as3 to i32
[all …]
/external/llvm/test/CodeGen/AMDGPU/
Daddrspacecast-constantexpr.ll3 declare void @llvm.memcpy.p1i32.p4i32.i32(i32 addrspace(1)* nocapture, i32 addrspace(4)* nocapture,…
5 @lds.i32 = unnamed_addr addrspace(3) global i32 undef, align 4
6 @lds.arr = unnamed_addr addrspace(3) global [256 x i32] undef, align 4
8 @global.i32 = unnamed_addr addrspace(1) global i32 undef, align 4
9 @global.arr = unnamed_addr addrspace(1) global [256 x i32] undef, align 4
13 store i32 7, i32 addrspace(3)* addrspacecast (i32 addrspace(4)* null to i32 addrspace(3)*)
19 store i32 7, i32 addrspace(4)* addrspacecast (i32 addrspace(3)* null to i32 addrspace(4)*)
25 store i32 7, i32 addrspace(4)* addrspacecast (i32 addrspace(3)* @lds.i32 to i32 addrspace(4)*)
31 … i32 addrspace(4)* getelementptr ([256 x i32], [256 x i32] addrspace(4)* addrspacecast ([256 x i32…
37 store i32 7, i32 addrspace(4)* addrspacecast (i32 addrspace(1)* @global.i32 to i32 addrspace(4)*)
[all …]
Dmerge-stores.ll18 define void @merge_global_store_2_constants_i8(i8 addrspace(1)* %out) #0 {
19 %out.gep.1 = getelementptr i8, i8 addrspace(1)* %out, i32 1
21 store i8 123, i8 addrspace(1)* %out.gep.1
22 store i8 456, i8 addrspace(1)* %out, align 2
30 define void @merge_global_store_2_constants_i8_natural_align(i8 addrspace(1)* %out) #0 {
31 %out.gep.1 = getelementptr i8, i8 addrspace(1)* %out, i32 1
33 store i8 123, i8 addrspace(1)* %out.gep.1
34 store i8 456, i8 addrspace(1)* %out
40 define void @merge_global_store_2_constants_i16(i16 addrspace(1)* %out) #0 {
41 %out.gep.1 = getelementptr i16, i16 addrspace(1)* %out, i32 1
[all …]
Dlds-alignment.ll3 @lds.align16.0 = internal unnamed_addr addrspace(3) global [38 x i8] undef, align 16
4 @lds.align16.1 = internal unnamed_addr addrspace(3) global [38 x i8] undef, align 16
6 @lds.align8.0 = internal unnamed_addr addrspace(3) global [38 x i8] undef, align 8
7 @lds.align32.0 = internal unnamed_addr addrspace(3) global [38 x i8] undef, align 32
9 @lds.missing.align.0 = internal unnamed_addr addrspace(3) global [39 x i32] undef
10 @lds.missing.align.1 = internal unnamed_addr addrspace(3) global [7 x i64] undef
12 declare void @llvm.memcpy.p3i8.p1i8.i32(i8 addrspace(3)* nocapture, i8 addrspace(1)* nocapture read…
13 declare void @llvm.memcpy.p1i8.p3i8.i32(i8 addrspace(1)* nocapture, i8 addrspace(3)* nocapture read…
18 define void @test_no_round_size_1(i8 addrspace(1)* %out, i8 addrspace(1)* %in) #1 {
19 %lds.align16.0.bc = bitcast [38 x i8] addrspace(3)* @lds.align16.0 to i8 addrspace(3)*
[all …]
Dimage-resource-id.ll10 define void @test_2d_rd_1_0(%opencl.image2d_t addrspace(1)* %in, ; read_only
11 i32 addrspace(1)* %out) {
14 %opencl.image2d_t addrspace(1)* %in) #0
15 store i32 %0, i32 addrspace(1)* %out
24 define void @test_3d_rd_1_0(%opencl.image3d_t addrspace(1)* %in, ; read_only
25 i32 addrspace(1)* %out) {
28 %opencl.image3d_t addrspace(1)* %in) #0
29 store i32 %0, i32 addrspace(1)* %out
40 define void @test_2d_wr_1_0(%opencl.image2d_t addrspace(1)* %in, ; write_only
41 i32 addrspace(1)* %out) {
[all …]
/external/llvm/test/Assembler/
D2007-12-11-AddressSpaces.ll1 ; RUN: llvm-as < %s | llvm-dis | grep "addrspace(33)" | count 7
2 ; RUN: llvm-as < %s | llvm-dis | grep "addrspace(42)" | count 2
3 ; RUN: llvm-as < %s | llvm-dis | grep "addrspace(66)" | count 2
4 ; RUN: llvm-as < %s | llvm-dis | grep "addrspace(11)" | count 6
5 ; RUN: llvm-as < %s | llvm-dis | grep "addrspace(22)" | count 5
8 %struct.mystruct = type { i32, i32 addrspace(33)*, i32, i32 addrspace(33)* }
9 @input = weak addrspace(42) global %struct.mystruct zeroinitializer ; <%struct.mystruct addrspac…
10 @output = addrspace(66) global %struct.mystruct zeroinitializer ; <%struct.mystruct addrspace(66)…
11 @y = external addrspace(33) global i32 addrspace(11)* addrspace(22)* ; <i32 addrspace(11)* addrsp…
15addrspace(33)*, i32 addrspace(33)* addrspace(42)* getelementptr (%struct.mystruct, %struct.mystruc…
[all …]
/external/swiftshader/third_party/llvm-7.0/llvm/test/Assembler/
D2007-12-11-AddressSpaces.ll1 ; RUN: llvm-as < %s | llvm-dis | grep "addrspace(33)" | count 7
2 ; RUN: llvm-as < %s | llvm-dis | grep "addrspace(42)" | count 2
3 ; RUN: llvm-as < %s | llvm-dis | grep "addrspace(66)" | count 2
4 ; RUN: llvm-as < %s | llvm-dis | grep "addrspace(11)" | count 6
5 ; RUN: llvm-as < %s | llvm-dis | grep "addrspace(22)" | count 5
8 %struct.mystruct = type { i32, i32 addrspace(33)*, i32, i32 addrspace(33)* }
9 @input = weak addrspace(42) global %struct.mystruct zeroinitializer ; <%struct.mystruct addrspac…
10 @output = addrspace(66) global %struct.mystruct zeroinitializer ; <%struct.mystruct addrspace(66)…
11 @y = external addrspace(33) global i32 addrspace(11)* addrspace(22)* ; <i32 addrspace(11)* addrsp…
15addrspace(33)*, i32 addrspace(33)* addrspace(42)* getelementptr (%struct.mystruct, %struct.mystruc…
[all …]

12345678910>>...90