; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py ; RUN: llc < %s -mtriple=i686-unknown-unknown -mattr=+bmi | FileCheck %s --check-prefixes=CHECK,X86,BMI1,X86-BMI1 ; RUN: llc < %s -mtriple=i686-unknown-unknown -mattr=+bmi,+bmi2 | FileCheck %s --check-prefixes=CHECK,X86,BMI2,X86-BMI2 ; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mattr=+bmi | FileCheck %s --check-prefixes=CHECK,X64,BMI1,X64-BMI1 ; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mattr=+bmi,+bmi2 | FileCheck %s --check-prefixes=CHECK,X64,BMI2,X64-BMI2 define i32 @andn32(i32 %x, i32 %y) { ; X86-LABEL: andn32: ; X86: # %bb.0: ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: andnl {{[0-9]+}}(%esp), %eax, %eax ; X86-NEXT: retl ; ; X64-LABEL: andn32: ; X64: # %bb.0: ; X64-NEXT: andnl %esi, %edi, %eax ; X64-NEXT: retq %tmp1 = xor i32 %x, -1 %tmp2 = and i32 %y, %tmp1 ret i32 %tmp2 } define i32 @andn32_load(i32 %x, i32* %y) { ; X86-LABEL: andn32_load: ; X86: # %bb.0: ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx ; X86-NEXT: andnl (%eax), %ecx, %eax ; X86-NEXT: retl ; ; X64-LABEL: andn32_load: ; X64: # %bb.0: ; X64-NEXT: andnl (%rsi), %edi, %eax ; X64-NEXT: retq %y1 = load i32, i32* %y %tmp1 = xor i32 %x, -1 %tmp2 = and i32 %y1, %tmp1 ret i32 %tmp2 } define i64 @andn64(i64 %x, i64 %y) { ; X86-LABEL: andn64: ; X86: # %bb.0: ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx ; X86-NEXT: andnl {{[0-9]+}}(%esp), %eax, %eax ; X86-NEXT: andnl {{[0-9]+}}(%esp), %ecx, %edx ; X86-NEXT: retl ; ; X64-LABEL: andn64: ; X64: # %bb.0: ; X64-NEXT: andnq %rsi, %rdi, %rax ; X64-NEXT: retq %tmp1 = xor i64 %x, -1 %tmp2 = and i64 %tmp1, %y ret i64 %tmp2 } ; Don't choose a 'test' if an 'andn' can be used. define i1 @andn_cmp(i32 %x, i32 %y) { ; X86-LABEL: andn_cmp: ; X86: # %bb.0: ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: andnl {{[0-9]+}}(%esp), %eax, %eax ; X86-NEXT: sete %al ; X86-NEXT: retl ; ; X64-LABEL: andn_cmp: ; X64: # %bb.0: ; X64-NEXT: andnl %esi, %edi, %eax ; X64-NEXT: sete %al ; X64-NEXT: retq %notx = xor i32 %x, -1 %and = and i32 %notx, %y %cmp = icmp eq i32 %and, 0 ret i1 %cmp } ; Recognize a disguised andn in the following 4 tests. define i1 @and_cmp1(i32 %x, i32 %y) { ; X86-LABEL: and_cmp1: ; X86: # %bb.0: ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: andnl {{[0-9]+}}(%esp), %eax, %eax ; X86-NEXT: sete %al ; X86-NEXT: retl ; ; X64-LABEL: and_cmp1: ; X64: # %bb.0: ; X64-NEXT: andnl %esi, %edi, %eax ; X64-NEXT: sete %al ; X64-NEXT: retq %and = and i32 %x, %y %cmp = icmp eq i32 %and, %y ret i1 %cmp } define i1 @and_cmp2(i32 %x, i32 %y) { ; X86-LABEL: and_cmp2: ; X86: # %bb.0: ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: andnl {{[0-9]+}}(%esp), %eax, %eax ; X86-NEXT: setne %al ; X86-NEXT: retl ; ; X64-LABEL: and_cmp2: ; X64: # %bb.0: ; X64-NEXT: andnl %esi, %edi, %eax ; X64-NEXT: setne %al ; X64-NEXT: retq %and = and i32 %y, %x %cmp = icmp ne i32 %and, %y ret i1 %cmp } define i1 @and_cmp3(i32 %x, i32 %y) { ; X86-LABEL: and_cmp3: ; X86: # %bb.0: ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: andnl {{[0-9]+}}(%esp), %eax, %eax ; X86-NEXT: sete %al ; X86-NEXT: retl ; ; X64-LABEL: and_cmp3: ; X64: # %bb.0: ; X64-NEXT: andnl %esi, %edi, %eax ; X64-NEXT: sete %al ; X64-NEXT: retq %and = and i32 %x, %y %cmp = icmp eq i32 %y, %and ret i1 %cmp } define i1 @and_cmp4(i32 %x, i32 %y) { ; X86-LABEL: and_cmp4: ; X86: # %bb.0: ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: andnl {{[0-9]+}}(%esp), %eax, %eax ; X86-NEXT: setne %al ; X86-NEXT: retl ; ; X64-LABEL: and_cmp4: ; X64: # %bb.0: ; X64-NEXT: andnl %esi, %edi, %eax ; X64-NEXT: setne %al ; X64-NEXT: retq %and = and i32 %y, %x %cmp = icmp ne i32 %y, %and ret i1 %cmp } ; A mask and compare against constant is ok for an 'andn' too ; even though the BMI instruction doesn't have an immediate form. define i1 @and_cmp_const(i32 %x) { ; X86-LABEL: and_cmp_const: ; X86: # %bb.0: ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: notl %eax ; X86-NEXT: andl $43, %eax ; X86-NEXT: sete %al ; X86-NEXT: retl ; ; X64-LABEL: and_cmp_const: ; X64: # %bb.0: ; X64-NEXT: notl %edi ; X64-NEXT: andl $43, %edi ; X64-NEXT: sete %al ; X64-NEXT: retq %and = and i32 %x, 43 %cmp = icmp eq i32 %and, 43 ret i1 %cmp } ; But don't use 'andn' if the mask is a power-of-two. define i1 @and_cmp_const_power_of_two(i32 %x, i32 %y) { ; X86-LABEL: and_cmp_const_power_of_two: ; X86: # %bb.0: ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx ; X86-NEXT: btl %ecx, %eax ; X86-NEXT: setae %al ; X86-NEXT: retl ; ; X64-LABEL: and_cmp_const_power_of_two: ; X64: # %bb.0: ; X64-NEXT: btl %esi, %edi ; X64-NEXT: setae %al ; X64-NEXT: retq %shl = shl i32 1, %y %and = and i32 %x, %shl %cmp = icmp ne i32 %and, %shl ret i1 %cmp } ; Don't transform to 'andn' if there's another use of the 'and'. define i32 @and_cmp_not_one_use(i32 %x) { ; X86-LABEL: and_cmp_not_one_use: ; X86: # %bb.0: ; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx ; X86-NEXT: andl $37, %ecx ; X86-NEXT: xorl %eax, %eax ; X86-NEXT: cmpl $37, %ecx ; X86-NEXT: sete %al ; X86-NEXT: addl %ecx, %eax ; X86-NEXT: retl ; ; X64-LABEL: and_cmp_not_one_use: ; X64: # %bb.0: ; X64-NEXT: andl $37, %edi ; X64-NEXT: xorl %eax, %eax ; X64-NEXT: cmpl $37, %edi ; X64-NEXT: sete %al ; X64-NEXT: addl %edi, %eax ; X64-NEXT: retq %and = and i32 %x, 37 %cmp = icmp eq i32 %and, 37 %ext = zext i1 %cmp to i32 %add = add i32 %and, %ext ret i32 %add } ; Verify that we're not transforming invalid comparison predicates. define i1 @not_an_andn1(i32 %x, i32 %y) { ; X86-LABEL: not_an_andn1: ; X86: # %bb.0: ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx ; X86-NEXT: andl %eax, %ecx ; X86-NEXT: cmpl %ecx, %eax ; X86-NEXT: setg %al ; X86-NEXT: retl ; ; X64-LABEL: not_an_andn1: ; X64: # %bb.0: ; X64-NEXT: andl %esi, %edi ; X64-NEXT: cmpl %edi, %esi ; X64-NEXT: setg %al ; X64-NEXT: retq %and = and i32 %x, %y %cmp = icmp sgt i32 %y, %and ret i1 %cmp } define i1 @not_an_andn2(i32 %x, i32 %y) { ; X86-LABEL: not_an_andn2: ; X86: # %bb.0: ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx ; X86-NEXT: andl %eax, %ecx ; X86-NEXT: cmpl %ecx, %eax ; X86-NEXT: setbe %al ; X86-NEXT: retl ; ; X64-LABEL: not_an_andn2: ; X64: # %bb.0: ; X64-NEXT: andl %esi, %edi ; X64-NEXT: cmpl %edi, %esi ; X64-NEXT: setbe %al ; X64-NEXT: retq %and = and i32 %y, %x %cmp = icmp ule i32 %y, %and ret i1 %cmp } ; Don't choose a 'test' if an 'andn' can be used. define i1 @andn_cmp_swap_ops(i64 %x, i64 %y) { ; X86-LABEL: andn_cmp_swap_ops: ; X86: # %bb.0: ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx ; X86-NEXT: andnl {{[0-9]+}}(%esp), %ecx, %ecx ; X86-NEXT: andnl {{[0-9]+}}(%esp), %eax, %eax ; X86-NEXT: orl %ecx, %eax ; X86-NEXT: sete %al ; X86-NEXT: retl ; ; X64-LABEL: andn_cmp_swap_ops: ; X64: # %bb.0: ; X64-NEXT: andnq %rsi, %rdi, %rax ; X64-NEXT: sete %al ; X64-NEXT: retq %notx = xor i64 %x, -1 %and = and i64 %y, %notx %cmp = icmp eq i64 %and, 0 ret i1 %cmp } ; Use a 'test' (not an 'and') because 'andn' only works for i32/i64. define i1 @andn_cmp_i8(i8 %x, i8 %y) { ; X86-LABEL: andn_cmp_i8: ; X86: # %bb.0: ; X86-NEXT: movb {{[0-9]+}}(%esp), %al ; X86-NEXT: notb %al ; X86-NEXT: testb %al, {{[0-9]+}}(%esp) ; X86-NEXT: sete %al ; X86-NEXT: retl ; ; X64-LABEL: andn_cmp_i8: ; X64: # %bb.0: ; X64-NEXT: notb %sil ; X64-NEXT: testb %sil, %dil ; X64-NEXT: sete %al ; X64-NEXT: retq %noty = xor i8 %y, -1 %and = and i8 %x, %noty %cmp = icmp eq i8 %and, 0 ret i1 %cmp } declare i32 @llvm.x86.bmi.bextr.32(i32, i32) define i32 @bextr32(i32 %x, i32 %y) { ; X86-LABEL: bextr32: ; X86: # %bb.0: ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: bextrl %eax, {{[0-9]+}}(%esp), %eax ; X86-NEXT: retl ; ; X64-LABEL: bextr32: ; X64: # %bb.0: ; X64-NEXT: bextrl %esi, %edi, %eax ; X64-NEXT: retq %tmp = tail call i32 @llvm.x86.bmi.bextr.32(i32 %x, i32 %y) ret i32 %tmp } define i32 @bextr32_load(i32* %x, i32 %y) { ; X86-LABEL: bextr32_load: ; X86: # %bb.0: ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx ; X86-NEXT: bextrl %eax, (%ecx), %eax ; X86-NEXT: retl ; ; X64-LABEL: bextr32_load: ; X64: # %bb.0: ; X64-NEXT: bextrl %esi, (%rdi), %eax ; X64-NEXT: retq %x1 = load i32, i32* %x %tmp = tail call i32 @llvm.x86.bmi.bextr.32(i32 %x1, i32 %y) ret i32 %tmp } define i32 @bextr32b(i32 %x) uwtable ssp { ; X86-LABEL: bextr32b: ; X86: # %bb.0: ; X86-NEXT: movl $3076, %eax # imm = 0xC04 ; X86-NEXT: bextrl %eax, {{[0-9]+}}(%esp), %eax ; X86-NEXT: retl ; ; X64-LABEL: bextr32b: ; X64: # %bb.0: ; X64-NEXT: movl $3076, %eax # imm = 0xC04 ; X64-NEXT: bextrl %eax, %edi, %eax ; X64-NEXT: retq %1 = lshr i32 %x, 4 %2 = and i32 %1, 4095 ret i32 %2 } ; Make sure we still use AH subreg trick to extract 15:8 define i32 @bextr32_subreg(i32 %x) uwtable ssp { ; X86-LABEL: bextr32_subreg: ; X86: # %bb.0: ; X86-NEXT: movzbl {{[0-9]+}}(%esp), %eax ; X86-NEXT: retl ; ; X64-LABEL: bextr32_subreg: ; X64: # %bb.0: ; X64-NEXT: movl %edi, %eax ; X64-NEXT: movzbl %ah, %eax ; X64-NEXT: retq %1 = lshr i32 %x, 8 %2 = and i32 %1, 255 ret i32 %2 } define i32 @bextr32b_load(i32* %x) uwtable ssp { ; X86-LABEL: bextr32b_load: ; X86: # %bb.0: ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: movl $3076, %ecx # imm = 0xC04 ; X86-NEXT: bextrl %ecx, (%eax), %eax ; X86-NEXT: retl ; ; X64-LABEL: bextr32b_load: ; X64: # %bb.0: ; X64-NEXT: movl $3076, %eax # imm = 0xC04 ; X64-NEXT: bextrl %eax, (%rdi), %eax ; X64-NEXT: retq %1 = load i32, i32* %x %2 = lshr i32 %1, 4 %3 = and i32 %2, 4095 ret i32 %3 } ; PR34042 define i32 @bextr32c(i32 %x, i16 zeroext %y) { ; X86-LABEL: bextr32c: ; X86: # %bb.0: ; X86-NEXT: movswl {{[0-9]+}}(%esp), %eax ; X86-NEXT: bextrl %eax, {{[0-9]+}}(%esp), %eax ; X86-NEXT: retl ; ; X64-LABEL: bextr32c: ; X64: # %bb.0: ; X64-NEXT: bextrl %esi, %edi, %eax ; X64-NEXT: retq %tmp0 = sext i16 %y to i32 %tmp1 = tail call i32 @llvm.x86.bmi.bextr.32(i32 %x, i32 %tmp0) ret i32 %tmp1 } define i32 @non_bextr32(i32 %x) { ; X86-LABEL: non_bextr32: ; X86: # %bb.0: # %entry ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: shrl $2, %eax ; X86-NEXT: andl $111, %eax ; X86-NEXT: retl ; ; X64-LABEL: non_bextr32: ; X64: # %bb.0: # %entry ; X64-NEXT: shrl $2, %edi ; X64-NEXT: andl $111, %edi ; X64-NEXT: movl %edi, %eax ; X64-NEXT: retq entry: %shr = lshr i32 %x, 2 %and = and i32 %shr, 111 ret i32 %and } define i32 @blsi32(i32 %x) { ; X86-LABEL: blsi32: ; X86: # %bb.0: ; X86-NEXT: blsil {{[0-9]+}}(%esp), %eax ; X86-NEXT: retl ; ; X64-LABEL: blsi32: ; X64: # %bb.0: ; X64-NEXT: blsil %edi, %eax ; X64-NEXT: retq %tmp = sub i32 0, %x %tmp2 = and i32 %x, %tmp ret i32 %tmp2 } define i32 @blsi32_load(i32* %x) { ; X86-LABEL: blsi32_load: ; X86: # %bb.0: ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: blsil (%eax), %eax ; X86-NEXT: retl ; ; X64-LABEL: blsi32_load: ; X64: # %bb.0: ; X64-NEXT: blsil (%rdi), %eax ; X64-NEXT: retq %x1 = load i32, i32* %x %tmp = sub i32 0, %x1 %tmp2 = and i32 %x1, %tmp ret i32 %tmp2 } define i64 @blsi64(i64 %x) { ; X86-LABEL: blsi64: ; X86: # %bb.0: ; X86-NEXT: pushl %esi ; X86-NEXT: .cfi_def_cfa_offset 8 ; X86-NEXT: .cfi_offset %esi, -8 ; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx ; X86-NEXT: movl {{[0-9]+}}(%esp), %esi ; X86-NEXT: xorl %edx, %edx ; X86-NEXT: movl %ecx, %eax ; X86-NEXT: negl %eax ; X86-NEXT: sbbl %esi, %edx ; X86-NEXT: andl %esi, %edx ; X86-NEXT: andl %ecx, %eax ; X86-NEXT: popl %esi ; X86-NEXT: .cfi_def_cfa_offset 4 ; X86-NEXT: retl ; ; X64-LABEL: blsi64: ; X64: # %bb.0: ; X64-NEXT: blsiq %rdi, %rax ; X64-NEXT: retq %tmp = sub i64 0, %x %tmp2 = and i64 %tmp, %x ret i64 %tmp2 } define i32 @blsmsk32(i32 %x) { ; X86-LABEL: blsmsk32: ; X86: # %bb.0: ; X86-NEXT: blsmskl {{[0-9]+}}(%esp), %eax ; X86-NEXT: retl ; ; X64-LABEL: blsmsk32: ; X64: # %bb.0: ; X64-NEXT: blsmskl %edi, %eax ; X64-NEXT: retq %tmp = sub i32 %x, 1 %tmp2 = xor i32 %x, %tmp ret i32 %tmp2 } define i32 @blsmsk32_load(i32* %x) { ; X86-LABEL: blsmsk32_load: ; X86: # %bb.0: ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: blsmskl (%eax), %eax ; X86-NEXT: retl ; ; X64-LABEL: blsmsk32_load: ; X64: # %bb.0: ; X64-NEXT: blsmskl (%rdi), %eax ; X64-NEXT: retq %x1 = load i32, i32* %x %tmp = sub i32 %x1, 1 %tmp2 = xor i32 %x1, %tmp ret i32 %tmp2 } define i64 @blsmsk64(i64 %x) { ; X86-LABEL: blsmsk64: ; X86: # %bb.0: ; X86-NEXT: pushl %esi ; X86-NEXT: .cfi_def_cfa_offset 8 ; X86-NEXT: .cfi_offset %esi, -8 ; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx ; X86-NEXT: movl {{[0-9]+}}(%esp), %esi ; X86-NEXT: movl %ecx, %eax ; X86-NEXT: addl $-1, %eax ; X86-NEXT: movl %esi, %edx ; X86-NEXT: adcl $-1, %edx ; X86-NEXT: xorl %ecx, %eax ; X86-NEXT: xorl %esi, %edx ; X86-NEXT: popl %esi ; X86-NEXT: .cfi_def_cfa_offset 4 ; X86-NEXT: retl ; ; X64-LABEL: blsmsk64: ; X64: # %bb.0: ; X64-NEXT: blsmskq %rdi, %rax ; X64-NEXT: retq %tmp = sub i64 %x, 1 %tmp2 = xor i64 %tmp, %x ret i64 %tmp2 } define i32 @blsr32(i32 %x) { ; X86-LABEL: blsr32: ; X86: # %bb.0: ; X86-NEXT: blsrl {{[0-9]+}}(%esp), %eax ; X86-NEXT: retl ; ; X64-LABEL: blsr32: ; X64: # %bb.0: ; X64-NEXT: blsrl %edi, %eax ; X64-NEXT: retq %tmp = sub i32 %x, 1 %tmp2 = and i32 %x, %tmp ret i32 %tmp2 } define i32 @blsr32_load(i32* %x) { ; X86-LABEL: blsr32_load: ; X86: # %bb.0: ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: blsrl (%eax), %eax ; X86-NEXT: retl ; ; X64-LABEL: blsr32_load: ; X64: # %bb.0: ; X64-NEXT: blsrl (%rdi), %eax ; X64-NEXT: retq %x1 = load i32, i32* %x %tmp = sub i32 %x1, 1 %tmp2 = and i32 %x1, %tmp ret i32 %tmp2 } define i64 @blsr64(i64 %x) { ; X86-LABEL: blsr64: ; X86: # %bb.0: ; X86-NEXT: pushl %esi ; X86-NEXT: .cfi_def_cfa_offset 8 ; X86-NEXT: .cfi_offset %esi, -8 ; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx ; X86-NEXT: movl {{[0-9]+}}(%esp), %esi ; X86-NEXT: movl %ecx, %eax ; X86-NEXT: addl $-1, %eax ; X86-NEXT: movl %esi, %edx ; X86-NEXT: adcl $-1, %edx ; X86-NEXT: andl %ecx, %eax ; X86-NEXT: andl %esi, %edx ; X86-NEXT: popl %esi ; X86-NEXT: .cfi_def_cfa_offset 4 ; X86-NEXT: retl ; ; X64-LABEL: blsr64: ; X64: # %bb.0: ; X64-NEXT: blsrq %rdi, %rax ; X64-NEXT: retq %tmp = sub i64 %x, 1 %tmp2 = and i64 %tmp, %x ret i64 %tmp2 } ; PR35792 - https://bugs.llvm.org/show_bug.cgi?id=35792 define i64 @blsr_disguised_constant(i64 %x) { ; X86-LABEL: blsr_disguised_constant: ; X86: # %bb.0: ; X86-NEXT: blsrl {{[0-9]+}}(%esp), %eax ; X86-NEXT: movzwl %ax, %eax ; X86-NEXT: xorl %edx, %edx ; X86-NEXT: retl ; ; X64-LABEL: blsr_disguised_constant: ; X64: # %bb.0: ; X64-NEXT: blsrl %edi, %eax ; X64-NEXT: movzwl %ax, %eax ; X64-NEXT: retq %a1 = and i64 %x, 65535 %a2 = add i64 %x, 65535 %r = and i64 %a1, %a2 ret i64 %r } ; The add here used to get shrunk, but the and did not thus hiding the blsr pattern. ; We now use the knowledge that upper bits of the shift guarantee the and result has 0s in the upper bits to reduce it too. define i64 @blsr_disguised_shrunk_add(i64 %x) { ; X86-LABEL: blsr_disguised_shrunk_add: ; X86: # %bb.0: ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: shrl $16, %eax ; X86-NEXT: blsrl %eax, %eax ; X86-NEXT: xorl %edx, %edx ; X86-NEXT: retl ; ; X64-LABEL: blsr_disguised_shrunk_add: ; X64: # %bb.0: ; X64-NEXT: shrq $48, %rdi ; X64-NEXT: blsrl %edi, %eax ; X64-NEXT: retq %a = lshr i64 %x, 48 %b = add i64 %a, -1 %c = and i64 %b, %a ret i64 %c }