; RUN: llc < %s -march=mips -mcpu=mips2 | FileCheck %s \ ; RUN: -check-prefixes=NOT-R2-R6,GP32,GP32-NOT-MM,NOT-MM ; RUN: llc < %s -march=mips -mcpu=mips32 | FileCheck %s \ ; RUN: -check-prefixes=NOT-R2-R6,GP32,GP32-NOT-MM,NOT-MM ; RUN: llc < %s -march=mips -mcpu=mips32r2 | FileCheck %s \ ; RUN: -check-prefixes=R2-R6,GP32,GP32-NOT-MM,NOT-MM ; RUN: llc < %s -march=mips -mcpu=mips32r3 | FileCheck %s \ ; RUN: -check-prefixes=R2-R6,GP32,GP32-NOT-MM,NOT-MM ; RUN: llc < %s -march=mips -mcpu=mips32r5 | FileCheck %s \ ; RUN: -check-prefixes=R2-R6,GP32,GP32-NOT-MM,NOT-MM ; RUN: llc < %s -march=mips -mcpu=mips32r6 | FileCheck %s \ ; RUN: -check-prefixes=R2-R6,GP32,GP32-NOT-MM,NOT-MM ; RUN: llc < %s -march=mips -mcpu=mips32r3 -mattr=+micromips | FileCheck %s \ ; RUN: -check-prefixes=GP32-MM,GP32,MM ; RUN: llc < %s -march=mips -mcpu=mips32r6 -mattr=+micromips | FileCheck %s \ ; RUN: -check-prefixes=GP32-MM,GP32,MM ; RUN: llc < %s -march=mips64 -mcpu=mips3 | FileCheck %s \ ; RUN: -check-prefixes=NOT-R2-R6,GP64,NOT-MM ; RUN: llc < %s -march=mips64 -mcpu=mips4 | FileCheck %s \ ; RUN: -check-prefixes=NOT-R2-R6,GP64,NOT-MM ; RUN: llc < %s -march=mips64 -mcpu=mips64 | FileCheck %s \ ; RUN: -check-prefixes=NOT-R2-R6,GP64,NOT-MM ; RUN: llc < %s -march=mips64 -mcpu=mips64r2 | FileCheck %s \ ; RUN: -check-prefixes=R2-R6,GP64,NOT-MM ; RUN: llc < %s -march=mips64 -mcpu=mips64r3 | FileCheck %s \ ; RUN: -check-prefixes=R2-R6,GP64,NOT-MM ; RUN: llc < %s -march=mips64 -mcpu=mips64r5 | FileCheck %s \ ; RUN: -check-prefixes=R2-R6,GP64,NOT-MM ; RUN: llc < %s -march=mips64 -mcpu=mips64r6 | FileCheck %s \ ; RUN: -check-prefixes=R2-R6,GP64,NOT-MM ; RUN: llc < %s -march=mips64 -mcpu=mips64r6 -mattr=+micromips | FileCheck %s \ ; RUN: -check-prefixes=GP64,MM define signext i1 @sub_i1(i1 signext %a, i1 signext %b) { entry: ; ALL-LABEL: sub_i1: ; NOT-MM: subu $[[T0:[0-9]+]], $4, $5 ; NOT-MM: sll $[[T0]], $[[T0]], 31 ; NOT-MM: sra $2, $[[T0]], 31 ; MM: subu16 $[[T0:[0-9]+]], $4, $5 ; MM: sll $[[T1:[0-9]+]], $[[T0]], 31 ; MM: sra $[[T0]], $[[T1]], 31 %r = sub i1 %a, %b ret i1 %r } define signext i8 @sub_i8(i8 signext %a, i8 signext %b) { entry: ; ALL-LABEL: sub_i8: ; NOT-R2-R6: subu $[[T0:[0-9]+]], $4, $5 ; NOT-R2-R6: sll $[[T0]], $[[T0]], 24 ; NOT-R2-R6: sra $2, $[[T0]], 24 ; R2-R6: subu $[[T0:[0-9]+]], $4, $5 ; R2-R6: seb $2, $[[T0:[0-9]+]] ; MM: subu16 $[[T0:[0-9]+]], $4, $5 ; MM: seb $[[T0]], $[[T0]] %r = sub i8 %a, %b ret i8 %r } define signext i16 @sub_i16(i16 signext %a, i16 signext %b) { entry: ; ALL-LABEL: sub_i16: ; NOT-R2-R6: subu $[[T0:[0-9]+]], $4, $5 ; NOT-R2-R6: sll $[[T0]], $[[T0]], 16 ; NOT-R2-R6: sra $2, $[[T0]], 16 ; R2-R6: subu $[[T0:[0-9]+]], $4, $5 ; R2-R6: seh $2, $[[T0:[0-9]+]] ; MM: subu16 $[[T0:[0-9]+]], $4, $5 ; MM: seh $[[T0]], $[[T0]] %r = sub i16 %a, %b ret i16 %r } define signext i32 @sub_i32(i32 signext %a, i32 signext %b) { entry: ; ALL-LABEL: sub_i32: ; NOT-MM: subu $2, $4, $5 ; MM: subu16 $2, $4, $5 %r = sub i32 %a, %b ret i32 %r } define signext i64 @sub_i64(i64 signext %a, i64 signext %b) { entry: ; ALL-LABEL: sub_i64: ; GP32: subu $3, $5, $7 ; GP32: sltu $[[T0:[0-9]+]], $5, $7 ; GP32: addu $[[T1:[0-9]+]], $[[T0]], $6 ; GP32: subu $2, $4, $[[T1]] ; GP64: dsubu $2, $4, $5 %r = sub i64 %a, %b ret i64 %r } define signext i128 @sub_i128(i128 signext %a, i128 signext %b) { entry: ; ALL-LABEL: sub_i128: ; GP32-NOT-MM: lw $[[T0:[0-9]+]], 20($sp) ; GP32-NOT-MM: sltu $[[T1:[0-9]+]], $5, $[[T0]] ; GP32-NOT-MM: lw $[[T2:[0-9]+]], 16($sp) ; GP32-NOT-MM: addu $[[T3:[0-9]+]], $[[T1]], $[[T2]] ; GP32-NOT-MM: lw $[[T4:[0-9]+]], 24($sp) ; GP32-NOT-MM: lw $[[T5:[0-9]+]], 28($sp) ; GP32-NOT-MM: subu $[[T6:[0-9]+]], $7, $[[T5]] ; GP32-NOT-MM: subu $2, $4, $[[T3]] ; GP32-NOT-MM: sltu $[[T8:[0-9]+]], $6, $[[T4]] ; GP32-NOT-MM: addu $[[T9:[0-9]+]], $[[T8]], $[[T0]] ; GP32-NOT-MM: subu $3, $5, $[[T9]] ; GP32-NOT-MM: sltu $[[T10:[0-9]+]], $7, $[[T5]] ; GP32-NOT-MM: addu $[[T11:[0-9]+]], $[[T10]], $[[T4]] ; GP32-NOT-MM: subu $4, $6, $[[T11]] ; GP32-NOT-MM: move $5, $[[T6]] ; GP32-MM: lw $[[T0:[0-9]+]], 20($sp) ; GP32-MM: sltu $[[T1:[0-9]+]], $[[T2:[0-9]+]], $[[T0]] ; GP32-MM: lw $[[T3:[0-9]+]], 16($sp) ; GP32-MM: addu $[[T3]], $[[T1]], $[[T3]] ; GP32-MM: lw $[[T4:[0-9]+]], 28($sp) ; GP32-MM: subu $[[T1]], $7, $[[T4]] ; GP32-MM: subu $[[T3]], $[[T5:[0-9]+]], $[[T3]] ; GP32-MM: lw $[[T5]], 24($sp) ; GP32-MM: sltu $[[T6:[0-9]+]], $6, $[[T5]] ; GP32-MM: addu $[[T0]], $[[T6]], $[[T0]] ; GP32-MM: subu $[[T0]], $5, $[[T0]] ; GP32-MM: sltu $[[T2]], $7, $[[T4]] ; GP32-MM: addu $[[T5]], $[[T2]], $[[T5]] ; GP32-MM: subu $[[T5]], $6, $[[T5]] ; GP32-MM: move $[[T2]], $[[T1]] ; GP64: dsubu $3, $5, $7 ; GP64: sltu $[[T0:[0-9]+]], $5, $7 ; GP64: daddu $[[T1:[0-9]+]], $[[T0]], $6 ; GP64: dsubu $2, $4, $[[T1]] %r = sub i128 %a, %b ret i128 %r }