1; RUN: llc < %s -march=xcore > %t1.s 2; RUN: grep "bl __misaligned_load" %t1.s | count 1 3; RUN: grep ld16s %t1.s | count 2 4; RUN: grep ldw %t1.s | count 2 5; RUN: grep shl %t1.s | count 2 6; RUN: grep shr %t1.s | count 1 7; RUN: grep zext %t1.s | count 1 8; RUN: grep "or " %t1.s | count 2 9 10; Byte aligned load. Expands to call to __misaligned_load. 11define i32 @align1(i32* %p) nounwind { 12entry: 13 %0 = load i32* %p, align 1 ; <i32> [#uses=1] 14 ret i32 %0 15} 16 17; Half word aligned load. Expands to two 16bit loads. 18define i32 @align2(i32* %p) nounwind { 19entry: 20 %0 = load i32* %p, align 2 ; <i32> [#uses=1] 21 ret i32 %0 22} 23 24@a = global [5 x i8] zeroinitializer, align 4 25 26; Constant offset from word aligned base. Expands to two 32bit loads. 27define i32 @align3() nounwind { 28entry: 29 %0 = load i32* bitcast (i8* getelementptr ([5 x i8]* @a, i32 0, i32 1) to i32*), align 1 30 ret i32 %0 31} 32