/external/llvm/test/Transforms/LoadStoreVectorizer/AMDGPU/ |
D | merge-stores.ll | 14 %out.gep.1 = getelementptr i8, i8 addrspace(1)* %out, i32 1 16 store i8 123, i8 addrspace(1)* %out.gep.1 24 %out.gep.1 = getelementptr i8, i8 addrspace(1)* %out, i32 1 26 store i8 123, i8 addrspace(1)* %out.gep.1 34 %out.gep.1 = getelementptr i16, i16 addrspace(1)* %out, i32 1 36 store i16 123, i16 addrspace(1)* %out.gep.1 44 %out.gep.1 = getelementptr i16, i16 addrspace(1)* %out, i32 1 46 store i16 0, i16 addrspace(1)* %out.gep.1 54 %out.gep.1 = getelementptr i16, i16 addrspace(1)* %out, i32 1 56 store i16 123, i16 addrspace(1)* %out.gep.1 [all …]
|
/external/llvm/test/CodeGen/AMDGPU/ |
D | merge-stores.ll | 19 %out.gep.1 = getelementptr i8, i8 addrspace(1)* %out, i32 1 21 store i8 123, i8 addrspace(1)* %out.gep.1 31 %out.gep.1 = getelementptr i8, i8 addrspace(1)* %out, i32 1 33 store i8 123, i8 addrspace(1)* %out.gep.1 41 %out.gep.1 = getelementptr i16, i16 addrspace(1)* %out, i32 1 43 store i16 123, i16 addrspace(1)* %out.gep.1 51 %out.gep.1 = getelementptr i16, i16 addrspace(1)* %out, i32 1 53 store i16 0, i16 addrspace(1)* %out.gep.1 63 %out.gep.1 = getelementptr i16, i16 addrspace(1)* %out, i32 1 65 store i16 123, i16 addrspace(1)* %out.gep.1 [all …]
|
D | mad-combine.ll | 36 %gep.0 = getelementptr float, float addrspace(1)* %in, i32 %tid 37 %gep.1 = getelementptr float, float addrspace(1)* %gep.0, i32 1 38 %gep.2 = getelementptr float, float addrspace(1)* %gep.0, i32 2 39 %gep.out = getelementptr float, float addrspace(1)* %out, i32 %tid 41 %a = load volatile float, float addrspace(1)* %gep.0 42 %b = load volatile float, float addrspace(1)* %gep.1 43 %c = load volatile float, float addrspace(1)* %gep.2 47 store float %fma, float addrspace(1)* %gep.out 75 %gep.0 = getelementptr float, float addrspace(1)* %in, i32 %tid 76 %gep.1 = getelementptr float, float addrspace(1)* %gep.0, i32 1 [all …]
|
D | commute-compares.ll | 13 %gep.in = getelementptr i32, i32 addrspace(1)* %in, i32 %tid 14 %gep.out = getelementptr i32, i32 addrspace(1)* %out, i32 %tid 15 %val = load i32, i32 addrspace(1)* %gep.in 18 store i32 %ext, i32 addrspace(1)* %gep.out 26 %gep.in = getelementptr i32, i32 addrspace(1)* %in, i32 %tid 27 %gep.out = getelementptr i32, i32 addrspace(1)* %out, i32 %tid 28 %val = load i32, i32 addrspace(1)* %gep.in 31 store i32 %ext, i32 addrspace(1)* %gep.out 41 %gep.in = getelementptr i32, i32 addrspace(1)* %in, i32 %tid 42 %gep.out = getelementptr i32, i32 addrspace(1)* %out, i32 %tid [all …]
|
D | fma-combine.ll | 18 %gep.0 = getelementptr double, double addrspace(1)* %in, i32 %tid 19 %gep.1 = getelementptr double, double addrspace(1)* %gep.0, i32 1 20 %gep.2 = getelementptr double, double addrspace(1)* %gep.0, i32 2 21 %gep.out = getelementptr double, double addrspace(1)* %out, i32 %tid 23 %a = load volatile double, double addrspace(1)* %gep.0 24 %b = load volatile double, double addrspace(1)* %gep.1 25 %c = load volatile double, double addrspace(1)* %gep.2 29 store double %fma, double addrspace(1)* %gep.out 46 %gep.0 = getelementptr double, double addrspace(1)* %in, i32 %tid 47 %gep.1 = getelementptr double, double addrspace(1)* %gep.0, i32 1 [all …]
|
D | fmuladd.ll | 41 %gep.0 = getelementptr float, float addrspace(1)* %out, i32 %tid 42 %gep.1 = getelementptr float, float addrspace(1)* %gep.0, i32 1 43 %gep.out = getelementptr float, float addrspace(1)* %out, i32 %tid 45 %r1 = load volatile float, float addrspace(1)* %gep.0 46 %r2 = load volatile float, float addrspace(1)* %gep.1 49 store float %r3, float addrspace(1)* %gep.out 60 %gep.0 = getelementptr float, float addrspace(1)* %out, i32 %tid 61 %gep.1 = getelementptr float, float addrspace(1)* %gep.0, i32 1 62 %gep.out = getelementptr float, float addrspace(1)* %out, i32 %tid 64 %r1 = load volatile float, float addrspace(1)* %gep.0 [all …]
|
D | madmk.ll | 17 %gep.0 = getelementptr float, float addrspace(1)* %in, i32 %tid 18 %gep.1 = getelementptr float, float addrspace(1)* %gep.0, i32 1 19 %out.gep = getelementptr float, float addrspace(1)* %out, i32 %tid 21 %a = load volatile float, float addrspace(1)* %gep.0, align 4 22 %b = load volatile float, float addrspace(1)* %gep.1, align 4 26 store float %madmk, float addrspace(1)* %out.gep, align 4 41 %in.gep.0 = getelementptr float, float addrspace(1)* %in, i32 %tid 42 %in.gep.1 = getelementptr float, float addrspace(1)* %in.gep.0, i32 1 43 %in.gep.2 = getelementptr float, float addrspace(1)* %in.gep.0, i32 2 45 %out.gep.0 = getelementptr float, float addrspace(1)* %out, i32 %tid [all …]
|
D | shift-and-i64-ubfe.ll | 13 %in.gep = getelementptr i64, i64 addrspace(1)* %in, i32 %id.x 14 %out.gep = getelementptr i64, i64 addrspace(1)* %out, i32 %id.x 15 %ld.64 = load i64, i64 addrspace(1)* %in.gep 18 store i64 %bit, i64 addrspace(1)* %out.gep 30 %in.gep = getelementptr i64, i64 addrspace(1)* %in, i32 %id.x 31 %out.gep = getelementptr i64, i64 addrspace(1)* %out, i32 %id.x 32 %ld.64 = load i64, i64 addrspace(1)* %in.gep 35 store i64 %bit, i64 addrspace(1)* %out.gep 46 %in.gep = getelementptr i64, i64 addrspace(1)* %in, i32 %id.x 47 %out.gep = getelementptr i64, i64 addrspace(1)* %out, i32 %id.x [all …]
|
D | fmax_legacy.f64.ll | 9 %gep.0 = getelementptr double, double addrspace(1)* %in, i32 %tid 10 %gep.1 = getelementptr double, double addrspace(1)* %gep.0, i32 1 12 %a = load double, double addrspace(1)* %gep.0, align 8 13 %b = load double, double addrspace(1)* %gep.1, align 8 24 %gep.0 = getelementptr double, double addrspace(1)* %in, i32 %tid 25 %gep.1 = getelementptr double, double addrspace(1)* %gep.0, i32 1 27 %a = load double, double addrspace(1)* %gep.0, align 8 28 %b = load double, double addrspace(1)* %gep.1, align 8 39 %gep.0 = getelementptr double, double addrspace(1)* %in, i32 %tid 40 %gep.1 = getelementptr double, double addrspace(1)* %gep.0, i32 1 [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/AMDGPU/ |
D | merge-stores.ll | 16 %out.gep.1 = getelementptr i8, i8 addrspace(1)* %out, i32 1 18 store i8 123, i8 addrspace(1)* %out.gep.1 28 %out.gep.1 = getelementptr i8, i8 addrspace(1)* %out, i32 1 30 store i8 123, i8 addrspace(1)* %out.gep.1 38 %out.gep.1 = getelementptr i16, i16 addrspace(1)* %out, i32 1 40 store i16 123, i16 addrspace(1)* %out.gep.1 48 %out.gep.1 = getelementptr i16, i16 addrspace(1)* %out, i32 1 50 store i16 0, i16 addrspace(1)* %out.gep.1 60 %out.gep.1 = getelementptr i16, i16 addrspace(1)* %out, i32 1 62 store i16 123, i16 addrspace(1)* %out.gep.1 [all …]
|
D | mad-combine.ll | 36 %gep.0 = getelementptr float, float addrspace(1)* %in, i32 %tid 37 %gep.1 = getelementptr float, float addrspace(1)* %gep.0, i32 1 38 %gep.2 = getelementptr float, float addrspace(1)* %gep.0, i32 2 39 %gep.out = getelementptr float, float addrspace(1)* %out, i32 %tid 41 %a = load volatile float, float addrspace(1)* %gep.0 42 %b = load volatile float, float addrspace(1)* %gep.1 43 %c = load volatile float, float addrspace(1)* %gep.2 47 store float %fma, float addrspace(1)* %gep.out 75 %gep.0 = getelementptr float, float addrspace(1)* %in, i32 %tid 76 %gep.1 = getelementptr float, float addrspace(1)* %gep.0, i32 1 [all …]
|
D | commute-compares.ll | 13 %gep.in = getelementptr i32, i32 addrspace(1)* %in, i32 %tid 14 %gep.out = getelementptr i32, i32 addrspace(1)* %out, i32 %tid 15 %val = load i32, i32 addrspace(1)* %gep.in 18 store i32 %ext, i32 addrspace(1)* %gep.out 26 %gep.in = getelementptr i32, i32 addrspace(1)* %in, i32 %tid 27 %gep.out = getelementptr i32, i32 addrspace(1)* %out, i32 %tid 28 %val = load i32, i32 addrspace(1)* %gep.in 31 store i32 %ext, i32 addrspace(1)* %gep.out 41 %gep.in = getelementptr i32, i32 addrspace(1)* %in, i32 %tid 42 %gep.out = getelementptr i32, i32 addrspace(1)* %out, i32 %tid [all …]
|
D | fma-combine.ll | 23 %gep.0 = getelementptr double, double addrspace(1)* %in, i32 %tid 24 %gep.1 = getelementptr double, double addrspace(1)* %gep.0, i32 1 25 %gep.2 = getelementptr double, double addrspace(1)* %gep.0, i32 2 26 %gep.out = getelementptr double, double addrspace(1)* %out, i32 %tid 28 %a = load volatile double, double addrspace(1)* %gep.0 29 %b = load volatile double, double addrspace(1)* %gep.1 30 %c = load volatile double, double addrspace(1)* %gep.2 34 store double %fma, double addrspace(1)* %gep.out 51 %gep.0 = getelementptr double, double addrspace(1)* %in, i32 %tid 52 %gep.1 = getelementptr double, double addrspace(1)* %gep.0, i32 1 [all …]
|
D | icmp.i16.ll | 15 %a.gep = getelementptr inbounds i16, i16 addrspace(1)* %a.ptr, i64 %tid.ext 16 %b.gep = getelementptr inbounds i16, i16 addrspace(1)* %b.ptr, i64 %tid.ext 17 %out.gep = getelementptr inbounds i32, i32 addrspace(1)* %out, i64 %tid.ext 18 %a = load i16, i16 addrspace(1)* %a.gep 19 %b = load i16, i16 addrspace(1)* %b.gep 22 store i32 %tmp1, i32 addrspace(1)* %out.gep 33 %a.gep = getelementptr inbounds i16, i16 addrspace(1)* %a.ptr, i64 %tid.ext 34 %b.gep = getelementptr inbounds i16, i16 addrspace(1)* %b.ptr, i64 %tid.ext 35 %out.gep = getelementptr inbounds i32, i32 addrspace(1)* %out, i64 %tid.ext 36 %a = load i16, i16 addrspace(1)* %a.gep [all …]
|
D | madmk.ll | 17 %gep.0 = getelementptr float, float addrspace(1)* %in, i32 %tid 18 %gep.1 = getelementptr float, float addrspace(1)* %gep.0, i32 1 19 %out.gep = getelementptr float, float addrspace(1)* %out, i32 %tid 21 %a = load volatile float, float addrspace(1)* %gep.0, align 4 22 %b = load volatile float, float addrspace(1)* %gep.1, align 4 26 store float %madmk, float addrspace(1)* %out.gep, align 4 41 %in.gep.0 = getelementptr float, float addrspace(1)* %in, i32 %tid 42 %in.gep.1 = getelementptr float, float addrspace(1)* %in.gep.0, i32 1 43 %in.gep.2 = getelementptr float, float addrspace(1)* %in.gep.0, i32 2 45 %out.gep.0 = getelementptr float, float addrspace(1)* %out, i32 %tid [all …]
|
D | memory-legalizer-atomic-cmpxchg.ll | 12 %gep = getelementptr i32, i32* %out, i32 4 13 %val = cmpxchg volatile i32* %gep, i32 %old, i32 %in monotonic monotonic 25 %gep = getelementptr i32, i32* %out, i32 4 26 %val = cmpxchg volatile i32* %gep, i32 %old, i32 %in acquire monotonic 38 %gep = getelementptr i32, i32* %out, i32 4 39 %val = cmpxchg volatile i32* %gep, i32 %old, i32 %in release monotonic 51 %gep = getelementptr i32, i32* %out, i32 4 52 %val = cmpxchg volatile i32* %gep, i32 %old, i32 %in acq_rel monotonic 64 %gep = getelementptr i32, i32* %out, i32 4 65 %val = cmpxchg volatile i32* %gep, i32 %old, i32 %in seq_cst monotonic [all …]
|
D | shift-and-i64-ubfe.ll | 14 %in.gep = getelementptr i64, i64 addrspace(1)* %in, i32 %id.x 15 %out.gep = getelementptr i64, i64 addrspace(1)* %out, i32 %id.x 16 %ld.64 = load i64, i64 addrspace(1)* %in.gep 19 store i64 %bit, i64 addrspace(1)* %out.gep 32 %in.gep = getelementptr i64, i64 addrspace(1)* %in, i32 %id.x 33 %out.gep = getelementptr i64, i64 addrspace(1)* %out, i32 %id.x 34 %ld.64 = load i64, i64 addrspace(1)* %in.gep 37 store i64 %bit, i64 addrspace(1)* %out.gep 48 %in.gep = getelementptr i64, i64 addrspace(1)* %in, i32 %id.x 49 %out.gep = getelementptr i64, i64 addrspace(1)* %out, i32 %id.x [all …]
|
D | add.i16.ll | 3 ; FIXME: Need to handle non-uniform case for function below (load without gep). 11 %gep.out = getelementptr inbounds i16, i16 addrspace(1)* %out, i32 %tid 12 %gep.in0 = getelementptr inbounds i16, i16 addrspace(1)* %in0, i32 %tid 13 %gep.in1 = getelementptr inbounds i16, i16 addrspace(1)* %in1, i32 %tid 14 %a = load volatile i16, i16 addrspace(1)* %gep.in0 15 %b = load volatile i16, i16 addrspace(1)* %gep.in1 21 ; FIXME: Need to handle non-uniform case for function below (load without gep). 28 %gep.out = getelementptr inbounds i16, i16 addrspace(1)* %out, i32 %tid 29 %gep.in0 = getelementptr inbounds i16, i16 addrspace(1)* %in0, i32 %tid 30 %a = load volatile i16, i16 addrspace(1)* %gep.in0 [all …]
|
D | fneg-combines.ll | 20 %a.gep = getelementptr inbounds float, float addrspace(1)* %a.ptr, i64 %tid.ext 21 %b.gep = getelementptr inbounds float, float addrspace(1)* %b.ptr, i64 %tid.ext 22 %out.gep = getelementptr inbounds float, float addrspace(1)* %out, i64 %tid.ext 23 %a = load volatile float, float addrspace(1)* %a.gep 24 %b = load volatile float, float addrspace(1)* %b.gep 27 store float %fneg, float addrspace(1)* %out.gep 41 %a.gep = getelementptr inbounds float, float addrspace(1)* %a.ptr, i64 %tid.ext 42 %b.gep = getelementptr inbounds float, float addrspace(1)* %b.ptr, i64 %tid.ext 43 %out.gep = getelementptr inbounds float, float addrspace(1)* %out, i64 %tid.ext 44 %a = load volatile float, float addrspace(1)* %a.gep [all …]
|
D | shrink-add-sub-constant.ll | 13 %gep = getelementptr inbounds i32, i32 addrspace(1)* %in, i64 %tid.ext 14 %gep.out = getelementptr inbounds i32, i32 addrspace(1)* %out, i64 %tid.ext 15 %x = load i32, i32 addrspace(1)* %gep 17 store i32 %result, i32 addrspace(1)* %gep.out 29 %gep = getelementptr inbounds i32, i32 addrspace(1)* %in, i64 %tid.ext 30 %gep.out = getelementptr inbounds i32, i32 addrspace(1)* %out, i64 %tid.ext 31 %x = load volatile i32, i32 addrspace(1)* %gep 32 %y = load volatile i32, i32 addrspace(1)* %gep 35 store volatile i32 %result0, i32 addrspace(1)* %gep.out 36 store volatile i32 %result1, i32 addrspace(1)* %gep.out [all …]
|
D | sub.i16.ll | 4 ; FIXME: Need to handle non-uniform case for function below (load without gep). 12 %gep.out = getelementptr inbounds i16, i16 addrspace(1)* %out, i32 %tid 13 %gep.in0 = getelementptr inbounds i16, i16 addrspace(1)* %in0, i32 %tid 14 %gep.in1 = getelementptr inbounds i16, i16 addrspace(1)* %in1, i32 %tid 15 %a = load volatile i16, i16 addrspace(1)* %gep.in0 16 %b = load volatile i16, i16 addrspace(1)* %gep.in1 22 ; FIXME: Need to handle non-uniform case for function below (load without gep). 29 %gep.out = getelementptr inbounds i16, i16 addrspace(1)* %out, i32 %tid 30 %gep.in0 = getelementptr inbounds i16, i16 addrspace(1)* %in0, i32 %tid 31 %a = load volatile i16, i16 addrspace(1)* %gep.in0 [all …]
|
D | v_cndmask.ll | 14 %f.gep = getelementptr float, float addrspace(1)* %fptr, i32 %idx 15 %f = load float, float addrspace(1)* %f.gep 53 %out.gep = getelementptr inbounds float, float addrspace(1)* %out, i64 %tid.ext 56 store float %select, float addrspace(1)* %out.gep 68 %out.gep = getelementptr inbounds float, float addrspace(1)* %out, i64 %tid.ext 71 store float %select, float addrspace(1)* %out.gep 83 %out.gep = getelementptr inbounds float, float addrspace(1)* %out, i64 %tid.ext 86 store float %select, float addrspace(1)* %out.gep 98 %out.gep = getelementptr inbounds float, float addrspace(1)* %out, i64 %tid.ext 101 store float %select, float addrspace(1)* %out.gep [all …]
|
D | fmax_legacy.f64.ll | 9 %gep.0 = getelementptr double, double addrspace(1)* %in, i32 %tid 10 %gep.1 = getelementptr double, double addrspace(1)* %gep.0, i32 1 12 %a = load double, double addrspace(1)* %gep.0, align 8 13 %b = load double, double addrspace(1)* %gep.1, align 8 24 %gep.0 = getelementptr double, double addrspace(1)* %in, i32 %tid 25 %gep.1 = getelementptr double, double addrspace(1)* %gep.0, i32 1 27 %a = load double, double addrspace(1)* %gep.0, align 8 28 %b = load double, double addrspace(1)* %gep.1, align 8 39 %gep.0 = getelementptr double, double addrspace(1)* %in, i32 %tid 40 %gep.1 = getelementptr double, double addrspace(1)* %gep.0, i32 1 [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/LoadStoreVectorizer/AMDGPU/ |
D | merge-stores.ll | 14 %out.gep.1 = getelementptr i8, i8 addrspace(1)* %out, i32 1 16 store i8 123, i8 addrspace(1)* %out.gep.1 24 %out.gep.1 = getelementptr i8, i8 addrspace(1)* %out, i32 1 26 store i8 123, i8 addrspace(1)* %out.gep.1 34 %out.gep.1 = getelementptr i16, i16 addrspace(1)* %out, i32 1 36 store i16 123, i16 addrspace(1)* %out.gep.1 44 %out.gep.1 = getelementptr i16, i16 addrspace(1)* %out, i32 1 46 store i16 0, i16 addrspace(1)* %out.gep.1 54 %out.gep.1 = getelementptr i16, i16 addrspace(1)* %out, i32 1 56 store i16 123, i16 addrspace(1)* %out.gep.1 [all …]
|
D | merge-stores-private.ll | 20 %out.gep.1 = getelementptr i32, i32 addrspace(5)* %out, i32 1 21 %out.gep.2 = getelementptr i32, i32 addrspace(5)* %out, i32 2 22 %out.gep.3 = getelementptr i32, i32 addrspace(5)* %out, i32 3 25 store i32 1, i32 addrspace(5)* %out.gep.1 26 store i32 23, i32 addrspace(5)* %out.gep.2 27 store i32 19, i32 addrspace(5)* %out.gep.3 33 ; ALIGNED: store i32 1, i32 addrspace(5)* %out.gep.1, align 1 34 ; ALIGNED: store i32 23, i32 addrspace(5)* %out.gep.2, align 1 35 ; ALIGNED: store i32 19, i32 addrspace(5)* %out.gep.3, align 1 47 %out.gep.1 = getelementptr i32, i32 addrspace(5)* %out, i32 1 [all …]
|