Searched refs:a52 (Results 1 – 25 of 42) sorted by relevance
12
117 %a52 = and i1 %a51, %a51118 %a53 = and i1 %a52, %a52
178 a56, a55, a54, a53, a52, a51, a50, a49, \ argument
156 %a52 = load volatile i32, i32* @a157 %x52 = add i32 %x51, %a52
166 %a52 = load volatile i32, i32* @a167 %x52 = add i32 %x51, %a52
168 a52[0] = 0
271 …8 %a44, i8 %a45, i8 %a46, i8 %a47, i8 %a48, i8 %a49, i8 %a50, i8 %a51, i8 %a52, i8 %a53, i8 %a54, …465 %ins52 = insertelement <64 x i8> %ins51, i8 %a52, i32 52
45 …a43 /a44 /a45 /a46 /a47 /a48 /a49 /a4a /a4b /a4c /a4d /a4e /a4f /a50 /a51 /a52 /a53 /a54 /a55 /a56…
441 %a52 = load i8, i8* getelementptr inbounds ([64 x i8], [64 x i8]* @a8, i32 0, i64 52), align 1505 %r52 = call i8 @llvm.abs.i8(i8 %a52, i1 false)
538 %a52 = load i8, i8* getelementptr inbounds ([64 x i8], [64 x i8]* @a8, i32 0, i64 52), align 1666 %r52 = add i8 %a52, %b52
538 %a52 = load i8, i8* getelementptr inbounds ([64 x i8], [64 x i8]* @a8, i32 0, i64 52), align 1666 %r52 = sub i8 %a52, %b52
481 %a52 = load i8, i8* getelementptr inbounds ([64 x i8], [64 x i8]* @a8, i32 0, i64 52), align 1609 %r52 = call i8 @llvm.uadd.sat.i8(i8 %a52, i8 %b52)
481 %a52 = load i8, i8* getelementptr inbounds ([64 x i8], [64 x i8]* @a8, i32 0, i64 52), align 1609 %r52 = call i8 @llvm.usub.sat.i8(i8 %a52, i8 %b52)
557 %a52 = load i8, i8* getelementptr inbounds ([64 x i8], [64 x i8]* @a8, i32 0, i64 52), align 1685 %r52 = call i8 @llvm.umin.i8(i8 %a52, i8 %b52)
635 %a52 = load i8, i8* getelementptr inbounds ([64 x i8], [64 x i8]* @a8, i32 0, i64 52), align 1763 %r52 = shl i8 %a52, %b52
605 %a52 = load i8, i8* getelementptr inbounds ([64 x i8], [64 x i8]* @a8, i32 0, i64 52), align 1733 %r52 = mul i8 %a52, %b52
557 %a52 = load i8, i8* getelementptr inbounds ([64 x i8], [64 x i8]* @a8, i32 0, i64 52), align 1685 %r52 = call i8 @llvm.smin.i8(i8 %a52, i8 %b52)
557 %a52 = load i8, i8* getelementptr inbounds ([64 x i8], [64 x i8]* @a8, i32 0, i64 52), align 1685 %r52 = call i8 @llvm.umax.i8(i8 %a52, i8 %b52)
557 %a52 = load i8, i8* getelementptr inbounds ([64 x i8], [64 x i8]* @a8, i32 0, i64 52), align 1685 %r52 = call i8 @llvm.smax.i8(i8 %a52, i8 %b52)
603 %a52 = load i8, i8* getelementptr inbounds ([64 x i8], [64 x i8]* @a8, i32 0, i64 52), align 1731 %r52 = call i8 @llvm.sadd.sat.i8(i8 %a52, i8 %b52)
683 %a52 = load i8, i8* getelementptr inbounds ([64 x i8], [64 x i8]* @a8, i32 0, i64 52), align 1811 %r52 = lshr i8 %a52, %b52
603 %a52 = load i8, i8* getelementptr inbounds ([64 x i8], [64 x i8]* @a8, i32 0, i64 52), align 1731 %r52 = call i8 @llvm.ssub.sat.i8(i8 %a52, i8 %b52)
734 %a52 = load i8, i8* getelementptr inbounds ([64 x i8], [64 x i8]* @a8, i32 0, i64 52), align 1862 %r52 = ashr i8 %a52, %b52
745 %a52 = load i8, i8* getelementptr inbounds ([64 x i8], [64 x i8]* @a8, i32 0, i64 52), align 1873 %r52 = call i8 @llvm.smul.fix.i8(i8 %a52, i8 %b52, i32 3)1674 %a52 = load i8, i8* getelementptr inbounds ([64 x i8], [64 x i8]* @a8, i32 0, i64 52), align 11802 %r52 = call i8 @llvm.umul.fix.i8(i8 %a52, i8 %b52, i32 3)