Searched refs:a56 (Results 1 – 25 of 41) sorted by relevance
12
121 %a56 = and i1 %a55, %a55122 %a57 = and i1 %a56, %a56
178 a56, a55, a54, a53, a52, a51, a50, a49, \ argument
172 a56[0] = 0
271 …8 %a48, i8 %a49, i8 %a50, i8 %a51, i8 %a52, i8 %a53, i8 %a54, i8 %a55, i8 %a56, i8 %a57, i8 %a58, …469 %ins56 = insertelement <64 x i8> %ins55, i8 %a56, i32 56
45 …a47 /a48 /a49 /a4a /a4b /a4c /a4d /a4e /a4f /a50 /a51 /a52 /a53 /a54 /a55 /a56 /a57 /a58 /a59 /a5a…
445 %a56 = load i8, i8* getelementptr inbounds ([64 x i8], [64 x i8]* @a8, i32 0, i64 56), align 1509 %r56 = call i8 @llvm.abs.i8(i8 %a56, i1 false)
542 %a56 = load i8, i8* getelementptr inbounds ([64 x i8], [64 x i8]* @a8, i32 0, i64 56), align 1670 %r56 = add i8 %a56, %b56
542 %a56 = load i8, i8* getelementptr inbounds ([64 x i8], [64 x i8]* @a8, i32 0, i64 56), align 1670 %r56 = sub i8 %a56, %b56
485 %a56 = load i8, i8* getelementptr inbounds ([64 x i8], [64 x i8]* @a8, i32 0, i64 56), align 1613 %r56 = call i8 @llvm.uadd.sat.i8(i8 %a56, i8 %b56)
485 %a56 = load i8, i8* getelementptr inbounds ([64 x i8], [64 x i8]* @a8, i32 0, i64 56), align 1613 %r56 = call i8 @llvm.usub.sat.i8(i8 %a56, i8 %b56)
561 %a56 = load i8, i8* getelementptr inbounds ([64 x i8], [64 x i8]* @a8, i32 0, i64 56), align 1689 %r56 = call i8 @llvm.umin.i8(i8 %a56, i8 %b56)
639 %a56 = load i8, i8* getelementptr inbounds ([64 x i8], [64 x i8]* @a8, i32 0, i64 56), align 1767 %r56 = shl i8 %a56, %b56
609 %a56 = load i8, i8* getelementptr inbounds ([64 x i8], [64 x i8]* @a8, i32 0, i64 56), align 1737 %r56 = mul i8 %a56, %b56
561 %a56 = load i8, i8* getelementptr inbounds ([64 x i8], [64 x i8]* @a8, i32 0, i64 56), align 1689 %r56 = call i8 @llvm.smin.i8(i8 %a56, i8 %b56)
561 %a56 = load i8, i8* getelementptr inbounds ([64 x i8], [64 x i8]* @a8, i32 0, i64 56), align 1689 %r56 = call i8 @llvm.umax.i8(i8 %a56, i8 %b56)
561 %a56 = load i8, i8* getelementptr inbounds ([64 x i8], [64 x i8]* @a8, i32 0, i64 56), align 1689 %r56 = call i8 @llvm.smax.i8(i8 %a56, i8 %b56)
607 %a56 = load i8, i8* getelementptr inbounds ([64 x i8], [64 x i8]* @a8, i32 0, i64 56), align 1735 %r56 = call i8 @llvm.sadd.sat.i8(i8 %a56, i8 %b56)
687 %a56 = load i8, i8* getelementptr inbounds ([64 x i8], [64 x i8]* @a8, i32 0, i64 56), align 1815 %r56 = lshr i8 %a56, %b56
607 %a56 = load i8, i8* getelementptr inbounds ([64 x i8], [64 x i8]* @a8, i32 0, i64 56), align 1735 %r56 = call i8 @llvm.ssub.sat.i8(i8 %a56, i8 %b56)
738 %a56 = load i8, i8* getelementptr inbounds ([64 x i8], [64 x i8]* @a8, i32 0, i64 56), align 1866 %r56 = ashr i8 %a56, %b56
749 %a56 = load i8, i8* getelementptr inbounds ([64 x i8], [64 x i8]* @a8, i32 0, i64 56), align 1877 %r56 = call i8 @llvm.smul.fix.i8(i8 %a56, i8 %b56, i32 3)1678 %a56 = load i8, i8* getelementptr inbounds ([64 x i8], [64 x i8]* @a8, i32 0, i64 56), align 11806 %r56 = call i8 @llvm.umul.fix.i8(i8 %a56, i8 %b56, i32 3)
1391 3a56 8 108 20792664 PUBLIC 6a56 10 RtlUnwind
1442 3a56 8 108 20792787 PUBLIC 6a56 10 RtlUnwind