; RUN: llc -march=amdgcn -verify-machineinstrs< %s | FileCheck -check-prefix=SI -check-prefix=ALIGNED %s ; RUN: llc -march=amdgcn -mcpu=bonaire -mattr=+unaligned-buffer-access -verify-machineinstrs< %s | FileCheck -check-prefix=SI -check-prefix=UNALIGNED %s ; RUN: llc -march=amdgcn -mcpu=tonga -verify-machineinstrs< %s | FileCheck -check-prefix=SI -check-prefix=ALIGNED %s ; SI-LABEL: {{^}}local_unaligned_load_store_i16: ; SI: ds_read_u8 ; SI: ds_read_u8 ; SI: ds_write_b8 ; SI: ds_write_b8 ; SI: s_endpgm define void @local_unaligned_load_store_i16(i16 addrspace(3)* %p, i16 addrspace(3)* %r) #0 { %v = load i16, i16 addrspace(3)* %p, align 1 store i16 %v, i16 addrspace(3)* %r, align 1 ret void } ; SI-LABEL: {{^}}global_unaligned_load_store_i16: ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_store_byte ; ALIGNED: buffer_store_byte ; UNALIGNED: buffer_load_ushort ; UNALIGNED: buffer_store_short ; SI: s_endpgm define void @global_unaligned_load_store_i16(i16 addrspace(1)* %p, i16 addrspace(1)* %r) #0 { %v = load i16, i16 addrspace(1)* %p, align 1 store i16 %v, i16 addrspace(1)* %r, align 1 ret void } ; FUNC-LABEL: {{^}}local_unaligned_load_store_i32: ; SI: ds_read_u8 ; SI: ds_read_u8 ; SI: ds_read_u8 ; SI: ds_read_u8 ; SI-NOT: v_or ; SI-NOT: v_lshl ; SI: ds_write_b8 ; SI: ds_write_b8 ; SI: ds_write_b8 ; SI: ds_write_b8 ; SI: s_endpgm define void @local_unaligned_load_store_i32(i32 addrspace(3)* %p, i32 addrspace(3)* %r) #0 { %v = load i32, i32 addrspace(3)* %p, align 1 store i32 %v, i32 addrspace(3)* %r, align 1 ret void } ; SI-LABEL: {{^}}global_unaligned_load_store_i32: ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_store_byte ; ALIGNED: buffer_store_byte ; ALIGNED: buffer_store_byte ; ALIGNED: buffer_store_byte ; UNALIGNED: buffer_load_dword ; UNALIGNED: buffer_store_dword define void @global_unaligned_load_store_i32(i32 addrspace(1)* %p, i32 addrspace(1)* %r) #0 { %v = load i32, i32 addrspace(1)* %p, align 1 store i32 %v, i32 addrspace(1)* %r, align 1 ret void } ; SI-LABEL: {{^}}global_align2_load_store_i32: ; ALIGNED: buffer_load_ushort ; ALIGNED: buffer_load_ushort ; ALIGNED: buffer_store_short ; ALIGNED: buffer_store_short ; UNALIGNED: buffer_load_dword ; UNALIGNED: buffer_store_dword define void @global_align2_load_store_i32(i32 addrspace(1)* %p, i32 addrspace(1)* %r) #0 { %v = load i32, i32 addrspace(1)* %p, align 2 store i32 %v, i32 addrspace(1)* %r, align 2 ret void } ; FUNC-LABEL: {{^}}local_align2_load_store_i32: ; GCN: ds_read_u16 ; GCN: ds_read_u16 ; GCN: ds_write_b16 ; GCN: ds_write_b16 define void @local_align2_load_store_i32(i32 addrspace(3)* %p, i32 addrspace(3)* %r) #0 { %v = load i32, i32 addrspace(3)* %p, align 2 store i32 %v, i32 addrspace(3)* %r, align 2 ret void } ; FUNC-LABEL: {{^}}local_unaligned_load_store_i64: ; SI: ds_read_u8 ; SI: ds_read_u8 ; SI: ds_read_u8 ; SI: ds_read_u8 ; SI: ds_read_u8 ; SI: ds_read_u8 ; SI: ds_read_u8 ; SI: ds_read_u8 ; SI-NOT: v_or_b32 ; SI-NOT: v_lshl ; SI: ds_write_b8 ; SI-NOT: v_or_b32 ; SI-NOT: v_lshl ; SI: ds_write_b8 ; SI-NOT: v_or_b32 ; SI-NOT: v_lshl ; SI: ds_write_b8 ; SI-NOT: v_or_b32 ; SI-NOT: v_lshl ; SI: ds_write_b8 ; SI-NOT: v_or_b32 ; SI-NOT: v_lshl ; SI: ds_write_b8 ; SI-NOT: v_or_b32 ; SI-NOT: v_lshl ; SI: ds_write_b8 ; SI-NOT: v_or_b32 ; SI-NOT: v_lshl ; SI: ds_write_b8 ; SI-NOT: v_or_b32 ; SI-NOT: v_lshl ; SI: ds_write_b8 ; SI: s_endpgm define void @local_unaligned_load_store_i64(i64 addrspace(3)* %p, i64 addrspace(3)* %r) #0 { %v = load i64, i64 addrspace(3)* %p, align 1 store i64 %v, i64 addrspace(3)* %r, align 1 ret void } ; SI-LABEL: {{^}}local_unaligned_load_store_v2i32: ; SI: ds_read_u8 ; SI: ds_read_u8 ; SI: ds_read_u8 ; SI: ds_read_u8 ; SI: ds_read_u8 ; SI: ds_read_u8 ; SI: ds_read_u8 ; SI: ds_read_u8 ; SI-NOT: v_or_b32 ; SI-NOT: v_lshl ; SI: ds_write_b8 ; SI-NOT: v_or_b32 ; SI-NOT: v_lshl ; SI: ds_write_b8 ; SI-NOT: v_or_b32 ; SI-NOT: v_lshl ; SI: ds_write_b8 ; SI-NOT: v_or_b32 ; SI-NOT: v_lshl ; SI: ds_write_b8 ; SI-NOT: v_or_b32 ; SI-NOT: v_lshl ; SI: ds_write_b8 ; SI-NOT: v_or_b32 ; SI-NOT: v_lshl ; SI: ds_write_b8 ; SI-NOT: v_or_b32 ; SI-NOT: v_lshl ; SI: ds_write_b8 ; SI-NOT: v_or_b32 ; SI-NOT: v_lshl ; SI: ds_write_b8 ; SI: s_endpgm define void @local_unaligned_load_store_v2i32(<2 x i32> addrspace(3)* %p, <2 x i32> addrspace(3)* %r) #0 { %v = load <2 x i32>, <2 x i32> addrspace(3)* %p, align 1 store <2 x i32> %v, <2 x i32> addrspace(3)* %r, align 1 ret void } ; SI-LABEL: {{^}}global_align2_load_store_i64: ; ALIGNED: buffer_load_ushort ; ALIGNED: buffer_load_ushort ; ALIGNED-NOT: v_or_ ; ALIGNED-NOT: v_lshl ; ALIGNED: buffer_load_ushort ; ALIGNED-NOT: v_or_ ; ALIGNED-NOT: v_lshl ; ALIGNED: buffer_load_ushort ; ALIGNED-NOT: v_or_ ; ALIGNED-NOT: v_lshl ; ALIGNED: buffer_store_short ; ALIGNED: buffer_store_short ; ALIGNED: buffer_store_short ; ALIGNED: buffer_store_short ; UNALIGNED: buffer_load_dwordx2 ; UNALIGNED: buffer_store_dwordx2 define void @global_align2_load_store_i64(i64 addrspace(1)* %p, i64 addrspace(1)* %r) #0 { %v = load i64, i64 addrspace(1)* %p, align 2 store i64 %v, i64 addrspace(1)* %r, align 2 ret void } ; SI-LABEL: {{^}}unaligned_load_store_i64_global: ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED-NOT: v_or_ ; ALIGNED-NOT: v_lshl ; ALIGNED: buffer_store_byte ; ALIGNED: buffer_store_byte ; ALIGNED: buffer_store_byte ; ALIGNED: buffer_store_byte ; ALIGNED: buffer_store_byte ; ALIGNED: buffer_store_byte ; ALIGNED: buffer_store_byte ; ALIGNED: buffer_store_byte ; UNALIGNED: buffer_load_dwordx2 ; UNALIGNED: buffer_store_dwordx2 define void @unaligned_load_store_i64_global(i64 addrspace(1)* %p, i64 addrspace(1)* %r) #0 { %v = load i64, i64 addrspace(1)* %p, align 1 store i64 %v, i64 addrspace(1)* %r, align 1 ret void } ; FUNC-LABEL: {{^}}local_unaligned_load_store_v4i32: ; GCN: ds_read_u8 ; GCN: ds_read_u8 ; GCN: ds_read_u8 ; GCN: ds_read_u8 ; GCN: ds_read_u8 ; GCN: ds_read_u8 ; GCN: ds_read_u8 ; GCN: ds_read_u8 ; GCN: ds_read_u8 ; GCN: ds_read_u8 ; GCN: ds_read_u8 ; GCN: ds_read_u8 ; GCN: ds_read_u8 ; GCN: ds_read_u8 ; GCN: ds_read_u8 ; GCN: ds_read_u8 ; GCN: ds_write_b8 ; GCN: ds_write_b8 ; GCN: ds_write_b8 ; GCN: ds_write_b8 ; GCN: ds_write_b8 ; GCN: ds_write_b8 ; GCN: ds_write_b8 ; GCN: ds_write_b8 ; GCN: ds_write_b8 ; GCN: ds_write_b8 ; GCN: ds_write_b8 ; GCN: ds_write_b8 ; GCN: ds_write_b8 ; GCN: ds_write_b8 ; GCN: ds_write_b8 ; GCN: ds_write_b8 ; GCN: s_endpgm define void @local_unaligned_load_store_v4i32(<4 x i32> addrspace(3)* %p, <4 x i32> addrspace(3)* %r) #0 { %v = load <4 x i32>, <4 x i32> addrspace(3)* %p, align 1 store <4 x i32> %v, <4 x i32> addrspace(3)* %r, align 1 ret void } ; SI-LABEL: {{^}}global_unaligned_load_store_v4i32 ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_store_byte ; ALIGNED: buffer_store_byte ; ALIGNED: buffer_store_byte ; ALIGNED: buffer_store_byte ; ALIGNED: buffer_store_byte ; ALIGNED: buffer_store_byte ; ALIGNED: buffer_store_byte ; ALIGNED: buffer_store_byte ; ALIGNED: buffer_store_byte ; ALIGNED: buffer_store_byte ; ALIGNED: buffer_store_byte ; ALIGNED: buffer_store_byte ; ALIGNED: buffer_store_byte ; ALIGNED: buffer_store_byte ; ALIGNED: buffer_store_byte ; ALIGNED: buffer_store_byte ; UNALIGNED: buffer_load_dwordx4 ; UNALIGNED: buffer_store_dwordx4 define void @global_unaligned_load_store_v4i32(<4 x i32> addrspace(1)* %p, <4 x i32> addrspace(1)* %r) #0 { %v = load <4 x i32>, <4 x i32> addrspace(1)* %p, align 1 store <4 x i32> %v, <4 x i32> addrspace(1)* %r, align 1 ret void } ; FUNC-LABEL: {{^}}local_load_i64_align_4: ; GCN: ds_read2_b32 define void @local_load_i64_align_4(i64 addrspace(1)* nocapture %out, i64 addrspace(3)* %in) #0 { %val = load i64, i64 addrspace(3)* %in, align 4 store i64 %val, i64 addrspace(1)* %out, align 8 ret void } ; FUNC-LABEL: {{^}}local_load_i64_align_4_with_offset ; GCN: ds_read2_b32 v[{{[0-9]+}}:{{[0-9]+}}], v{{[0-9]}} offset0:8 offset1:9 define void @local_load_i64_align_4_with_offset(i64 addrspace(1)* nocapture %out, i64 addrspace(3)* %in) #0 { %ptr = getelementptr i64, i64 addrspace(3)* %in, i32 4 %val = load i64, i64 addrspace(3)* %ptr, align 4 store i64 %val, i64 addrspace(1)* %out, align 8 ret void } ; FUNC-LABEL: {{^}}local_load_i64_align_4_with_split_offset: ; The tests for the case where the lo offset is 8-bits, but the hi offset is 9-bits ; GCN: ds_read2_b32 v[{{[0-9]+}}:{{[0-9]+}}], v{{[0-9]}} offset1:1 ; GCN: s_endpgm define void @local_load_i64_align_4_with_split_offset(i64 addrspace(1)* nocapture %out, i64 addrspace(3)* %in) #0 { %ptr = bitcast i64 addrspace(3)* %in to i32 addrspace(3)* %ptr255 = getelementptr i32, i32 addrspace(3)* %ptr, i32 255 %ptri64 = bitcast i32 addrspace(3)* %ptr255 to i64 addrspace(3)* %val = load i64, i64 addrspace(3)* %ptri64, align 4 store i64 %val, i64 addrspace(1)* %out, align 8 ret void } ; FUNC-LABEL: {{^}}local_load_i64_align_1: ; GCN: ds_read_u8 ; GCN: ds_read_u8 ; GCN: ds_read_u8 ; GCN: ds_read_u8 ; GCN: ds_read_u8 ; GCN: ds_read_u8 ; GCN: ds_read_u8 ; GCN: ds_read_u8 ; GCN: store_dwordx2 define void @local_load_i64_align_1(i64 addrspace(1)* nocapture %out, i64 addrspace(3)* %in) #0 { %val = load i64, i64 addrspace(3)* %in, align 1 store i64 %val, i64 addrspace(1)* %out, align 8 ret void } ; FUNC-LABEL: {{^}}local_store_i64_align_4: ; GCN: ds_write2_b32 define void @local_store_i64_align_4(i64 addrspace(3)* %out, i64 %val) #0 { store i64 %val, i64 addrspace(3)* %out, align 4 ret void } ; FUNC-LABEL: {{^}}local_store_i64_align_4_with_offset ; GCN: ds_write2_b32 v{{[0-9]+}}, v{{[0-9]+}}, v{{[0-9]+}} offset0:8 offset1:9 ; GCN: s_endpgm define void @local_store_i64_align_4_with_offset(i64 addrspace(3)* %out) #0 { %ptr = getelementptr i64, i64 addrspace(3)* %out, i32 4 store i64 0, i64 addrspace(3)* %ptr, align 4 ret void } ; FUNC-LABEL: {{^}}local_store_i64_align_4_with_split_offset: ; The tests for the case where the lo offset is 8-bits, but the hi offset is 9-bits ; GCN: ds_write2_b32 v{{[0-9]+}}, v{{[0-9]+}}, v{{[0-9]+}} offset1:1 ; GCN: s_endpgm define void @local_store_i64_align_4_with_split_offset(i64 addrspace(3)* %out) #0 { %ptr = bitcast i64 addrspace(3)* %out to i32 addrspace(3)* %ptr255 = getelementptr i32, i32 addrspace(3)* %ptr, i32 255 %ptri64 = bitcast i32 addrspace(3)* %ptr255 to i64 addrspace(3)* store i64 0, i64 addrspace(3)* %out, align 4 ret void } ; SI-LABEL: {{^}}constant_unaligned_load_i32: ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; UNALIGNED: s_load_dword ; SI: buffer_store_dword define void @constant_unaligned_load_i32(i32 addrspace(2)* %p, i32 addrspace(1)* %r) #0 { %v = load i32, i32 addrspace(2)* %p, align 1 store i32 %v, i32 addrspace(1)* %r, align 4 ret void } ; SI-LABEL: {{^}}constant_align2_load_i32: ; ALIGNED: buffer_load_ushort ; ALIGNED: buffer_load_ushort ; UNALIGNED: s_load_dword ; UNALIGNED: buffer_store_dword define void @constant_align2_load_i32(i32 addrspace(2)* %p, i32 addrspace(1)* %r) #0 { %v = load i32, i32 addrspace(2)* %p, align 2 store i32 %v, i32 addrspace(1)* %r, align 4 ret void } ; SI-LABEL: {{^}}constant_align2_load_i64: ; ALIGNED: buffer_load_ushort ; ALIGNED: buffer_load_ushort ; ALIGNED: buffer_load_ushort ; ALIGNED: buffer_load_ushort ; UNALIGNED: s_load_dwordx2 ; UNALIGNED: buffer_store_dwordx2 define void @constant_align2_load_i64(i64 addrspace(2)* %p, i64 addrspace(1)* %r) #0 { %v = load i64, i64 addrspace(2)* %p, align 2 store i64 %v, i64 addrspace(1)* %r, align 4 ret void } ; SI-LABEL: {{^}}constant_align4_load_i64: ; SI: s_load_dwordx2 ; SI: buffer_store_dwordx2 define void @constant_align4_load_i64(i64 addrspace(2)* %p, i64 addrspace(1)* %r) #0 { %v = load i64, i64 addrspace(2)* %p, align 4 store i64 %v, i64 addrspace(1)* %r, align 4 ret void } ; SI-LABEL: {{^}}constant_align4_load_v4i32: ; SI: s_load_dwordx4 ; SI: buffer_store_dwordx4 define void @constant_align4_load_v4i32(<4 x i32> addrspace(2)* %p, <4 x i32> addrspace(1)* %r) #0 { %v = load <4 x i32>, <4 x i32> addrspace(2)* %p, align 4 store <4 x i32> %v, <4 x i32> addrspace(1)* %r, align 4 ret void } ; SI-LABEL: {{^}}constant_unaligned_load_v2i32: ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; UNALIGNED: buffer_load_dwordx2 ; SI: buffer_store_dwordx2 define void @constant_unaligned_load_v2i32(<2 x i32> addrspace(2)* %p, <2 x i32> addrspace(1)* %r) #0 { %v = load <2 x i32>, <2 x i32> addrspace(2)* %p, align 1 store <2 x i32> %v, <2 x i32> addrspace(1)* %r, align 4 ret void } ; SI-LABEL: {{^}}constant_unaligned_load_v4i32: ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; ALIGNED: buffer_load_ubyte ; UNALIGNED: buffer_load_dwordx4 ; SI: buffer_store_dwordx4 define void @constant_unaligned_load_v4i32(<4 x i32> addrspace(2)* %p, <4 x i32> addrspace(1)* %r) #0 { %v = load <4 x i32>, <4 x i32> addrspace(2)* %p, align 1 store <4 x i32> %v, <4 x i32> addrspace(1)* %r, align 4 ret void } ; SI-LABEL: {{^}}constant_align4_load_i8: ; SI: buffer_load_ubyte ; SI: buffer_store_byte define void @constant_align4_load_i8(i8 addrspace(2)* %p, i8 addrspace(1)* %r) #0 { %v = load i8, i8 addrspace(2)* %p, align 4 store i8 %v, i8 addrspace(1)* %r, align 4 ret void } ; SI-LABEL: {{^}}constant_align2_load_i8: ; SI: buffer_load_ubyte ; SI: buffer_store_byte define void @constant_align2_load_i8(i8 addrspace(2)* %p, i8 addrspace(1)* %r) #0 { %v = load i8, i8 addrspace(2)* %p, align 2 store i8 %v, i8 addrspace(1)* %r, align 2 ret void } ; SI-LABEL: {{^}}constant_align4_merge_load_2_i32: ; SI: s_load_dwordx2 s{{\[}}[[LO:[0-9]+]]:[[HI:[0-9]+]]{{\]}}, s{{\[[0-9]+:[0-9]+\]}}, 0x0{{$}} ; SI-DAG: v_mov_b32_e32 v[[VLO:[0-9]+]], s[[LO]] ; SI-DAG: v_mov_b32_e32 v[[VHI:[0-9]+]], s[[HI]] ; SI: buffer_store_dwordx2 v{{\[}}[[VLO]]:[[VHI]]{{\]}} define void @constant_align4_merge_load_2_i32(i32 addrspace(2)* %p, i32 addrspace(1)* %r) #0 { %gep0 = getelementptr i32, i32 addrspace(2)* %p, i64 1 %v0 = load i32, i32 addrspace(2)* %p, align 4 %v1 = load i32, i32 addrspace(2)* %gep0, align 4 %gep1 = getelementptr i32, i32 addrspace(1)* %r, i64 1 store i32 %v0, i32 addrspace(1)* %r, align 4 store i32 %v1, i32 addrspace(1)* %gep1, align 4 ret void } attributes #0 = { nounwind }