SimplifySelectCC(DL, N00, N01, ExtTrueVal, Zero, CC, true))
return SCC;
- if (!VT.isVector() && !TLI.convertSelectOfConstantsToMath()) {
+ if (!VT.isVector()) {
EVT SetCCVT = getSetCCResultType(N00VT);
// Don't do this transform for i1 because there's a select transform
// that would reverse it.
return SDValue();
// Don't do this for crazy integer types.
- EVT VT = N->getValueType(0);
- if (!DAG.getTargetLoweringInfo().isTypeLegal(VT))
+ if (!DAG.getTargetLoweringInfo().isTypeLegal(LHS.getValueType()))
return SDValue();
- // We're going to use the condition bit in math or logic ops. We could allow
- // this with a wider condition value (post-legalization it becomes an i8),
- // but if nothing is creating selects that late, it doesn't matter.
- if (Cond.getValueType() != MVT::i1)
- return SDValue();
+ // If this is efficiently invertible, canonicalize the LHSC/RHSC values
+ // so that TrueC (the true value) is larger than FalseC.
+ bool NeedsCondInvert = false;
+ if (TrueC->getAPIntValue().ult(FalseC->getAPIntValue()) &&
+ // Efficiently invertible.
+ (Cond.getOpcode() == ISD::SETCC || // setcc -> invertible.
+ (Cond.getOpcode() == ISD::XOR && // xor(X, C) -> invertible.
+ isa<ConstantSDNode>(Cond.getOperand(1))))) {
+ NeedsCondInvert = true;
+ std::swap(TrueC, FalseC);
+ }
- // A power-of-2 multiply is just a shift. LEA also cheaply handles multiply by
- // 3, 5, or 9 with i32/i64, so those get transformed too.
- // TODO: For constants that do not differ by power-of-2 or small multiplier,
- // convert to 'and' + 'add'.
- APInt AbsDiff = (TrueC->getAPIntValue() - FalseC->getAPIntValue()).abs();
- if (AbsDiff.isPowerOf2() ||
- ((VT == MVT::i32 || VT == MVT::i64) &&
- (AbsDiff == 3 || AbsDiff == 5 || AbsDiff == 9))) {
+ // Optimize C ? 8 : 0 -> zext(C) << 3. Likewise for any pow2/0.
+ if (FalseC->getAPIntValue() == 0 && TrueC->getAPIntValue().isPowerOf2()) {
+ if (NeedsCondInvert) // Invert the condition if needed.
+ Cond = DAG.getNode(ISD::XOR, DL, Cond.getValueType(), Cond,
+ DAG.getConstant(1, DL, Cond.getValueType()));
- // We need a positive multiplier constant for shift/LEA codegen. The 'not'
- // of the condition can usually be folded into a compare predicate, but even
- // without that, the sequence should be cheaper than a CMOV alternative.
- if (TrueC->getAPIntValue().slt(FalseC->getAPIntValue())) {
- Cond = DAG.getNOT(DL, Cond, MVT::i1);
- std::swap(TrueC, FalseC);
- }
+ // Zero extend the condition if needed.
+ Cond = DAG.getNode(ISD::ZERO_EXTEND, DL, LHS.getValueType(), Cond);
- // select Cond, TC, FC --> (zext(Cond) * (TC - FC)) + FC
- SDValue R = DAG.getNode(ISD::ZERO_EXTEND, DL, VT, Cond);
+ unsigned ShAmt = TrueC->getAPIntValue().logBase2();
+ return DAG.getNode(ISD::SHL, DL, LHS.getValueType(), Cond,
+ DAG.getConstant(ShAmt, DL, MVT::i8));
+ }
- // Multiply condition by the difference if non-one.
- if (!AbsDiff.isOneValue())
- R = DAG.getNode(ISD::MUL, DL, VT, R, DAG.getConstant(AbsDiff, DL, VT));
+ // Optimize cases that will turn into an LEA instruction. This requires
+ // an i32 or i64 and an efficient multiplier (1, 2, 3, 4, 5, 8, 9).
+ if (N->getValueType(0) == MVT::i32 || N->getValueType(0) == MVT::i64) {
+ uint64_t Diff = TrueC->getZExtValue() - FalseC->getZExtValue();
+ if (N->getValueType(0) == MVT::i32)
+ Diff = (unsigned)Diff;
- // Add the base if non-zero.
- if (!FalseC->isNullValue())
- R = DAG.getNode(ISD::ADD, DL, VT, R, SDValue(FalseC, 0));
+ bool IsFastMultiplier = false;
+ if (Diff < 10) {
+ switch ((unsigned char)Diff) {
+ default:
+ break;
+ case 1: // result = add base, cond
+ case 2: // result = lea base( , cond*2)
+ case 3: // result = lea base(cond, cond*2)
+ case 4: // result = lea base( , cond*4)
+ case 5: // result = lea base(cond, cond*4)
+ case 8: // result = lea base( , cond*8)
+ case 9: // result = lea base(cond, cond*8)
+ IsFastMultiplier = true;
+ break;
+ }
+ }
- return R;
+ if (IsFastMultiplier) {
+ APInt Diff = TrueC->getAPIntValue() - FalseC->getAPIntValue();
+ if (NeedsCondInvert) // Invert the condition if needed.
+ Cond = DAG.getNode(ISD::XOR, DL, Cond.getValueType(), Cond,
+ DAG.getConstant(1, DL, Cond.getValueType()));
+
+ // Zero extend the condition if needed.
+ Cond = DAG.getNode(ISD::ZERO_EXTEND, DL, FalseC->getValueType(0), Cond);
+ // Scale the condition by the difference.
+ if (Diff != 1)
+ Cond = DAG.getNode(ISD::MUL, DL, Cond.getValueType(), Cond,
+ DAG.getConstant(Diff, DL, Cond.getValueType()));
+
+ // Add the base if non-zero.
+ if (FalseC->getAPIntValue() != 0)
+ Cond = DAG.getNode(ISD::ADD, DL, Cond.getValueType(), Cond,
+ SDValue(FalseC, 0));
+ return Cond;
+ }
}
return SDValue();
define i32 @main() nounwind {
; CHECK-LABEL: main:
; CHECK: # BB#0: # %entry
+; CHECK-NEXT: cmpq $0, {{.*}}(%rip)
+; CHECK-NEXT: movb $-106, %al
+; CHECK-NEXT: jne .LBB0_2
+; CHECK-NEXT: # BB#1: # %entry
; CHECK-NEXT: xorl %eax, %eax
-; CHECK-NEXT: cmpq {{.*}}(%rip), %rax
-; CHECK-NEXT: sbbl %eax, %eax
-; CHECK-NEXT: andl $150, %eax
+; CHECK-NEXT: .LBB0_2: # %entry
; CHECK-NEXT: testb %al, %al
-; CHECK-NEXT: jle .LBB0_1
-; CHECK-NEXT: # BB#2: # %if.then
+; CHECK-NEXT: jle .LBB0_3
+; CHECK-NEXT: # BB#4: # %if.then
; CHECK-NEXT: movl $1, {{.*}}(%rip)
; CHECK-NEXT: movl $1, %esi
-; CHECK-NEXT: jmp .LBB0_3
-; CHECK-NEXT: .LBB0_1: # %entry.if.end_crit_edge
+; CHECK-NEXT: jmp .LBB0_5
+; CHECK-NEXT: .LBB0_3: # %entry.if.end_crit_edge
; CHECK-NEXT: movl {{.*}}(%rip), %esi
-; CHECK-NEXT: .LBB0_3: # %if.end
+; CHECK-NEXT: .LBB0_5: # %if.end
; CHECK-NEXT: pushq %rax
; CHECK-NEXT: movl $.L.str, %edi
; CHECK-NEXT: xorl %eax, %eax
; X86-NEXT: subl %ecx, %eax
; X86-NEXT: jmp .LBB4_3
; X86-NEXT: .LBB4_1: # %res_block
-; X86-NEXT: setae %al
-; X86-NEXT: movzbl %al, %eax
-; X86-NEXT: leal -1(%eax,%eax), %eax
+; X86-NEXT: xorl %ecx, %ecx
+; X86-NEXT: incl %ecx
+; X86-NEXT: xorl %eax, %eax
+; X86-NEXT: decl %eax
+; X86-NEXT: cmpw %si, %dx
+; X86-NEXT: cmovael %ecx, %eax
; X86-NEXT: .LBB4_3: # %endblock
; X86-NEXT: popl %esi
; X86-NEXT: retl
; X64-NEXT: subl %ecx, %eax
; X64-NEXT: retq
; X64-NEXT: .LBB4_1: # %res_block
-; X64-NEXT: setae %al
-; X64-NEXT: movzbl %al, %eax
-; X64-NEXT: leal -1(%rax,%rax), %eax
+; X64-NEXT: movl $-1, %ecx
+; X64-NEXT: movl $1, %eax
+; X64-NEXT: cmovbl %ecx, %eax
; X64-NEXT: retq
%m = tail call i32 @memcmp(i8* %X, i8* %Y, i64 3) nounwind
ret i32 %m
; X86-NEXT: subl %ecx, %eax
; X86-NEXT: jmp .LBB9_3
; X86-NEXT: .LBB9_1: # %res_block
-; X86-NEXT: setae %al
-; X86-NEXT: movzbl %al, %eax
-; X86-NEXT: leal -1(%eax,%eax), %eax
+; X86-NEXT: xorl %ecx, %ecx
+; X86-NEXT: incl %ecx
+; X86-NEXT: xorl %eax, %eax
+; X86-NEXT: decl %eax
+; X86-NEXT: cmpl %esi, %edx
+; X86-NEXT: cmovael %ecx, %eax
; X86-NEXT: .LBB9_3: # %endblock
; X86-NEXT: popl %esi
; X86-NEXT: retl
; X64-NEXT: subl %ecx, %eax
; X64-NEXT: retq
; X64-NEXT: .LBB9_1: # %res_block
-; X64-NEXT: setae %al
-; X64-NEXT: movzbl %al, %eax
-; X64-NEXT: leal -1(%rax,%rax), %eax
+; X64-NEXT: movl $-1, %ecx
+; X64-NEXT: movl $1, %eax
+; X64-NEXT: cmovbl %ecx, %eax
; X64-NEXT: retq
%m = tail call i32 @memcmp(i8* %X, i8* %Y, i64 5) nounwind
ret i32 %m
; X86-NEXT: cmpl %edx, %ecx
; X86-NEXT: je .LBB11_3
; X86-NEXT: .LBB11_1: # %res_block
+; X86-NEXT: xorl %esi, %esi
+; X86-NEXT: incl %esi
; X86-NEXT: xorl %eax, %eax
+; X86-NEXT: decl %eax
; X86-NEXT: cmpl %edx, %ecx
-; X86-NEXT: setae %al
-; X86-NEXT: leal -1(%eax,%eax), %eax
+; X86-NEXT: cmovael %esi, %eax
; X86-NEXT: .LBB11_3: # %endblock
; X86-NEXT: popl %esi
; X86-NEXT: retl
; X64-NEXT: # BB#3: # %endblock
; X64-NEXT: retq
; X64-NEXT: .LBB15_1: # %res_block
-; X64-NEXT: xorl %eax, %eax
; X64-NEXT: cmpq %rdx, %rcx
-; X64-NEXT: setae %al
-; X64-NEXT: leal -1(%rax,%rax), %eax
+; X64-NEXT: movl $-1, %ecx
+; X64-NEXT: movl $1, %eax
+; X64-NEXT: cmovbl %ecx, %eax
; X64-NEXT: retq
%m = tail call i32 @memcmp(i8* %X, i8* %Y, i64 12) nounwind
ret i32 %m
; X64-NEXT: # BB#3: # %endblock
; X64-NEXT: retq
; X64-NEXT: .LBB16_1: # %res_block
-; X64-NEXT: xorl %eax, %eax
; X64-NEXT: cmpq %rdx, %rcx
-; X64-NEXT: setae %al
-; X64-NEXT: leal -1(%rax,%rax), %eax
+; X64-NEXT: movl $-1, %ecx
+; X64-NEXT: movl $1, %eax
+; X64-NEXT: cmovbl %ecx, %eax
; X64-NEXT: retq
%m = tail call i32 @memcmp(i8* %X, i8* %Y, i64 16) nounwind
ret i32 %m
; X86-NEXT: popl %esi
; X86-NEXT: retl
; X86-NEXT: .LBB4_1: # %res_block
-; X86-NEXT: setae %al
-; X86-NEXT: movzbl %al, %eax
-; X86-NEXT: leal -1(%eax,%eax), %eax
+; X86-NEXT: movl $-1, %ecx
+; X86-NEXT: movl $1, %eax
+; X86-NEXT: cmovbl %ecx, %eax
; X86-NEXT: popl %esi
; X86-NEXT: retl
;
; X64-NEXT: subl %ecx, %eax
; X64-NEXT: retq
; X64-NEXT: .LBB4_1: # %res_block
-; X64-NEXT: setae %al
-; X64-NEXT: movzbl %al, %eax
-; X64-NEXT: leal -1(%rax,%rax), %eax
+; X64-NEXT: movl $-1, %ecx
+; X64-NEXT: movl $1, %eax
+; X64-NEXT: cmovbl %ecx, %eax
; X64-NEXT: retq
%m = tail call i32 @memcmp(i8* %X, i8* %Y, i64 3) nounwind
ret i32 %m
; X86-NEXT: popl %esi
; X86-NEXT: retl
; X86-NEXT: .LBB9_1: # %res_block
-; X86-NEXT: setae %al
-; X86-NEXT: movzbl %al, %eax
-; X86-NEXT: leal -1(%eax,%eax), %eax
+; X86-NEXT: movl $-1, %ecx
+; X86-NEXT: movl $1, %eax
+; X86-NEXT: cmovbl %ecx, %eax
; X86-NEXT: popl %esi
; X86-NEXT: retl
;
; X64-NEXT: subl %ecx, %eax
; X64-NEXT: retq
; X64-NEXT: .LBB9_1: # %res_block
-; X64-NEXT: setae %al
-; X64-NEXT: movzbl %al, %eax
-; X64-NEXT: leal -1(%rax,%rax), %eax
+; X64-NEXT: movl $-1, %ecx
+; X64-NEXT: movl $1, %eax
+; X64-NEXT: cmovbl %ecx, %eax
; X64-NEXT: retq
%m = tail call i32 @memcmp(i8* %X, i8* %Y, i64 5) nounwind
ret i32 %m
; X86-NEXT: popl %esi
; X86-NEXT: retl
; X86-NEXT: .LBB11_1: # %res_block
-; X86-NEXT: xorl %eax, %eax
; X86-NEXT: cmpl %edx, %ecx
-; X86-NEXT: setae %al
-; X86-NEXT: leal -1(%eax,%eax), %eax
+; X86-NEXT: movl $-1, %ecx
+; X86-NEXT: movl $1, %eax
+; X86-NEXT: cmovbl %ecx, %eax
; X86-NEXT: popl %esi
; X86-NEXT: retl
;
; X64-NEXT: # BB#3: # %endblock
; X64-NEXT: retq
; X64-NEXT: .LBB15_1: # %res_block
-; X64-NEXT: xorl %eax, %eax
; X64-NEXT: cmpq %rdx, %rcx
-; X64-NEXT: setae %al
-; X64-NEXT: leal -1(%rax,%rax), %eax
+; X64-NEXT: movl $-1, %ecx
+; X64-NEXT: movl $1, %eax
+; X64-NEXT: cmovbl %ecx, %eax
; X64-NEXT: retq
%m = tail call i32 @memcmp(i8* %X, i8* %Y, i64 12) nounwind
ret i32 %m
; X64-NEXT: # BB#3: # %endblock
; X64-NEXT: retq
; X64-NEXT: .LBB16_1: # %res_block
-; X64-NEXT: xorl %eax, %eax
; X64-NEXT: cmpq %rdx, %rcx
-; X64-NEXT: setae %al
-; X64-NEXT: leal -1(%rax,%rax), %eax
+; X64-NEXT: movl $-1, %ecx
+; X64-NEXT: movl $1, %eax
+; X64-NEXT: cmovbl %ecx, %eax
; X64-NEXT: retq
%m = tail call i32 @memcmp(i8* %X, i8* %Y, i64 16) nounwind
ret i32 %m
; CHECK-NEXT: cmpl 16(%eax), %edx
; CHECK-NEXT: movl $0, 16(%eax)
; CHECK-NEXT: sbbl %ecx, %edx
-; CHECK-NEXT: setl %al
-; CHECK-NEXT: movzbl %al, %eax
-; CHECK-NEXT: negl %eax
+; CHECK-NEXT: movl $-1, %eax
+; CHECK-NEXT: jl .LBB0_2
+; CHECK-NEXT: # BB#1:
+; CHECK-NEXT: xorl %eax, %eax
+; CHECK-NEXT: .LBB0_2:
; CHECK-NEXT: retl
%used = getelementptr inbounds i64, i64* %so, i32 3
store i64 0, i64* %used, align 8
; X86-NEXT: .Lcfi103:
; X86-NEXT: .cfi_adjust_cfa_offset -8
; X86-NEXT: xorl $32, %eax
-; X86-NEXT: xorl %ecx, %ecx
; X86-NEXT: orl %ebx, %eax
-; X86-NEXT: setne %cl
-; X86-NEXT: negl %ecx
-; X86-NEXT: movl %ecx, %eax
+; X86-NEXT: movl $-1, %eax
+; X86-NEXT: jne .LBB1_2
+; X86-NEXT: # BB#1:
+; X86-NEXT: xorl %eax, %eax
+; X86-NEXT: .LBB1_2:
; X86-NEXT: popl %esi
; X86-NEXT: popl %edi
; X86-NEXT: popl %ebx
; X64-HSW-NEXT: pushq %r14
; X64-HSW-NEXT: .Lcfi2:
; X64-HSW-NEXT: .cfi_def_cfa_offset 32
-; X64-HSW-NEXT: pushq %rbx
+; X64-HSW-NEXT: pushq %r12
; X64-HSW-NEXT: .Lcfi3:
; X64-HSW-NEXT: .cfi_def_cfa_offset 40
-; X64-HSW-NEXT: pushq %rax
+; X64-HSW-NEXT: pushq %rbx
; X64-HSW-NEXT: .Lcfi4:
; X64-HSW-NEXT: .cfi_def_cfa_offset 48
; X64-HSW-NEXT: .Lcfi5:
-; X64-HSW-NEXT: .cfi_offset %rbx, -40
+; X64-HSW-NEXT: .cfi_offset %rbx, -48
; X64-HSW-NEXT: .Lcfi6:
-; X64-HSW-NEXT: .cfi_offset %r14, -32
+; X64-HSW-NEXT: .cfi_offset %r12, -40
; X64-HSW-NEXT: .Lcfi7:
-; X64-HSW-NEXT: .cfi_offset %r15, -24
+; X64-HSW-NEXT: .cfi_offset %r14, -32
; X64-HSW-NEXT: .Lcfi8:
+; X64-HSW-NEXT: .cfi_offset %r15, -24
+; X64-HSW-NEXT: .Lcfi9:
; X64-HSW-NEXT: .cfi_offset %rbp, -16
+; X64-HSW-NEXT: xorl %r12d, %r12d
; X64-HSW-NEXT: movl $1, %edi
; X64-HSW-NEXT: xorl %esi, %esi
; X64-HSW-NEXT: callq mult
; X64-HSW-NEXT: movl $16, %esi
; X64-HSW-NEXT: callq mult
; X64-HSW-NEXT: xorl $32, %eax
-; X64-HSW-NEXT: xorl %ecx, %ecx
; X64-HSW-NEXT: orl %ebx, %eax
-; X64-HSW-NEXT: setne %cl
-; X64-HSW-NEXT: negl %ecx
-; X64-HSW-NEXT: movl %ecx, %eax
-; X64-HSW-NEXT: addq $8, %rsp
+; X64-HSW-NEXT: movl $-1, %eax
+; X64-HSW-NEXT: cmovel %r12d, %eax
; X64-HSW-NEXT: popq %rbx
+; X64-HSW-NEXT: popq %r12
; X64-HSW-NEXT: popq %r14
; X64-HSW-NEXT: popq %r15
; X64-HSW-NEXT: popq %rbp
define i32 @fn() {
; X86-LABEL: fn:
; X86: # BB#0: # %entry
-; X86-NEXT: xorl %eax, %eax
; X86-NEXT: cmpl $1, %eax
-; X86-NEXT: setne %al
; X86-NEXT: sete %cl
-; X86-NEXT: negl %eax
+; X86-NEXT: movl $-1, %eax
+; X86-NEXT: jne .LBB0_2
+; X86-NEXT: # BB#1: # %entry
+; X86-NEXT: xorl %eax, %eax
+; X86-NEXT: .LBB0_2: # %entry
; X86-NEXT: addb %cl, %cl
; X86-NEXT: shll %cl, %eax
; X86-NEXT: .p2align 4, 0x90
-; X86-NEXT: .LBB0_1: # %bb1
+; X86-NEXT: .LBB0_3: # %bb1
; X86-NEXT: # =>This Inner Loop Header: Depth=1
; X86-NEXT: testl %eax, %eax
-; X86-NEXT: je .LBB0_1
-; X86-NEXT: # BB#2: # %bb2
+; X86-NEXT: je .LBB0_3
+; X86-NEXT: # BB#4: # %bb2
; X86-NEXT: retl
;
; X64-LABEL: fn:
; X64: # BB#0: # %entry
-; X64-NEXT: xorl %eax, %eax
+; X64-NEXT: xorl %edx, %edx
; X64-NEXT: cmpl $1, %eax
-; X64-NEXT: setne %al
; X64-NEXT: sete %cl
-; X64-NEXT: negl %eax
+; X64-NEXT: movl $-1, %eax
+; X64-NEXT: cmovel %edx, %eax
; X64-NEXT: addb %cl, %cl
; X64-NEXT: shll %cl, %eax
; X64-NEXT: .p2align 4, 0x90
define i32 @ugt_select_neg1_or_0(i32 %x, i32 %y) nounwind {
; CHECK-LABEL: ugt_select_neg1_or_0:
; CHECK: # BB#0:
-; CHECK-NEXT: cmpl %esi, %edi
-; CHECK-NEXT: sbbl %eax, %eax
+; CHECK-NEXT: xorl %ecx, %ecx
+; CHECK-NEXT: cmpl %edi, %esi
+; CHECK-NEXT: movl $-1, %eax
+; CHECK-NEXT: cmovbel %ecx, %eax
; CHECK-NEXT: retq
%cmp = icmp ugt i32 %y, %x
%ext = sext i1 %cmp to i32
;
; MCU-LABEL: test9b:
; MCU: # BB#0:
-; MCU-NEXT: movl %edx, %ecx
+; MCU-NEXT: orl %edx, %eax
+; MCU-NEXT: movl $-1, %edx
+; MCU-NEXT: je .LBB10_2
+; MCU-NEXT: # BB#1:
; MCU-NEXT: xorl %edx, %edx
-; MCU-NEXT: orl %ecx, %eax
-; MCU-NEXT: sete %dl
-; MCU-NEXT: negl %edx
+; MCU-NEXT: .LBB10_2:
; MCU-NEXT: movl {{[0-9]+}}(%esp), %eax
; MCU-NEXT: orl %edx, %eax
; MCU-NEXT: orl {{[0-9]+}}(%esp), %edx
;; Select between -1 and 1.
define i64 @test10(i64 %x, i64 %y) nounwind readnone ssp noredzone {
-; CHECK-LABEL: test10:
-; CHECK: ## BB#0:
-; CHECK-NEXT: xorl %eax, %eax
-; CHECK-NEXT: testq %rdi, %rdi
-; CHECK-NEXT: setne %al
-; CHECK-NEXT: leaq -1(%rax,%rax), %rax
-; CHECK-NEXT: retq
-; CHECK-NEXT: ## -- End function
+; GENERIC-LABEL: test10:
+; GENERIC: ## BB#0:
+; GENERIC-NEXT: cmpq $1, %rdi
+; GENERIC-NEXT: sbbq %rax, %rax
+; GENERIC-NEXT: orq $1, %rax
+; GENERIC-NEXT: retq
+; GENERIC-NEXT: ## -- End function
+;
+; ATOM-LABEL: test10:
+; ATOM: ## BB#0:
+; ATOM-NEXT: cmpq $1, %rdi
+; ATOM-NEXT: sbbq %rax, %rax
+; ATOM-NEXT: orq $1, %rax
+; ATOM-NEXT: nop
+; ATOM-NEXT: nop
+; ATOM-NEXT: retq
+; ATOM-NEXT: ## -- End function
;
; MCU-LABEL: test10:
; MCU: # BB#0:
}
define i32 @test14(i32 %a, i32 %b) nounwind {
-; CHECK-LABEL: test14:
-; CHECK: ## BB#0:
-; CHECK-NEXT: xorl %eax, %eax
-; CHECK-NEXT: cmpl %esi, %edi
-; CHECK-NEXT: setae %al
-; CHECK-NEXT: negl %eax
-; CHECK-NEXT: retq
-; CHECK-NEXT: ## -- End function
+; GENERIC-LABEL: test14:
+; GENERIC: ## BB#0:
+; GENERIC-NEXT: cmpl %esi, %edi
+; GENERIC-NEXT: sbbl %eax, %eax
+; GENERIC-NEXT: notl %eax
+; GENERIC-NEXT: retq
+; GENERIC-NEXT: ## -- End function
+;
+; ATOM-LABEL: test14:
+; ATOM: ## BB#0:
+; ATOM-NEXT: cmpl %esi, %edi
+; ATOM-NEXT: sbbl %eax, %eax
+; ATOM-NEXT: notl %eax
+; ATOM-NEXT: nop
+; ATOM-NEXT: nop
+; ATOM-NEXT: retq
+; ATOM-NEXT: ## -- End function
;
; MCU-LABEL: test14:
; MCU: # BB#0:
-; MCU-NEXT: xorl %ecx, %ecx
; MCU-NEXT: cmpl %edx, %eax
-; MCU-NEXT: setae %cl
-; MCU-NEXT: negl %ecx
-; MCU-NEXT: movl %ecx, %eax
+; MCU-NEXT: sbbl %eax, %eax
+; MCU-NEXT: notl %eax
; MCU-NEXT: retl
%c = icmp uge i32 %a, %b
%d = sext i1 %c to i32
;
; MCU-LABEL: test16:
; MCU: # BB#0: # %entry
-; MCU-NEXT: movl %eax, %ecx
+; MCU-NEXT: orl %edx, %eax
+; MCU-NEXT: movl $-1, %eax
+; MCU-NEXT: jne .LBB18_2
+; MCU-NEXT: # BB#1: # %entry
; MCU-NEXT: xorl %eax, %eax
-; MCU-NEXT: orl %edx, %ecx
-; MCU-NEXT: setne %al
-; MCU-NEXT: negl %eax
+; MCU-NEXT: .LBB18_2: # %entry
; MCU-NEXT: movl %eax, %edx
; MCU-NEXT: retl
entry:
; GENERIC-LABEL: test17:
; GENERIC: ## BB#0: ## %entry
; GENERIC-NEXT: negw %di
-; GENERIC-NEXT: sbbl %eax, %eax
-; GENERIC-NEXT: ## kill: %AX<def> %AX<kill> %EAX<kill>
+; GENERIC-NEXT: sbbw %ax, %ax
; GENERIC-NEXT: retq
; GENERIC-NEXT: ## -- End function
;
; ATOM-LABEL: test17:
; ATOM: ## BB#0: ## %entry
; ATOM-NEXT: negw %di
-; ATOM-NEXT: sbbl %eax, %eax
-; ATOM-NEXT: ## kill: %AX<def> %AX<kill> %EAX<kill>
+; ATOM-NEXT: sbbw %ax, %ax
; ATOM-NEXT: nop
; ATOM-NEXT: nop
; ATOM-NEXT: nop
; MCU-LABEL: test17:
; MCU: # BB#0: # %entry
; MCU-NEXT: negw %ax
-; MCU-NEXT: sbbl %eax, %eax
-; MCU-NEXT: # kill: %AX<def> %AX<kill> %EAX<kill>
+; MCU-NEXT: sbbw %ax, %ax
; MCU-NEXT: retl
entry:
%cmp = icmp ne i16 %x, 0
define i32 @select_lea_2(i1 zeroext %cond) {
; CHECK-LABEL: select_lea_2:
; CHECK: # BB#0:
-; CHECK-NEXT: xorb $1, %dil
-; CHECK-NEXT: movzbl %dil, %eax
-; CHECK-NEXT: leal -1(%rax,%rax), %eax
+; CHECK-NEXT: testb %dil, %dil
+; CHECK-NEXT: movl $-1, %ecx
+; CHECK-NEXT: movl $1, %eax
+; CHECK-NEXT: cmovnel %ecx, %eax
; CHECK-NEXT: retq
%sel = select i1 %cond, i32 -1, i32 1
ret i32 %sel
define i64 @select_lea_3(i1 zeroext %cond) {
; CHECK-LABEL: select_lea_3:
; CHECK: # BB#0:
-; CHECK-NEXT: xorb $1, %dil
-; CHECK-NEXT: movzbl %dil, %eax
-; CHECK-NEXT: leaq -2(%rax,%rax,2), %rax
+; CHECK-NEXT: testb %dil, %dil
+; CHECK-NEXT: movl $1, %ecx
+; CHECK-NEXT: movq $-2, %rax
+; CHECK-NEXT: cmoveq %rcx, %rax
; CHECK-NEXT: retq
%sel = select i1 %cond, i64 -2, i64 1
ret i64 %sel
define i32 @select_lea_5(i1 zeroext %cond) {
; CHECK-LABEL: select_lea_5:
; CHECK: # BB#0:
-; CHECK-NEXT: xorb $1, %dil
-; CHECK-NEXT: movzbl %dil, %eax
-; CHECK-NEXT: leal -2(%rax,%rax,4), %eax
+; CHECK-NEXT: testb %dil, %dil
+; CHECK-NEXT: movl $-2, %ecx
+; CHECK-NEXT: movl $3, %eax
+; CHECK-NEXT: cmovnel %ecx, %eax
; CHECK-NEXT: retq
%sel = select i1 %cond, i32 -2, i32 3
ret i32 %sel
define i64 @select_lea_9(i1 zeroext %cond) {
; CHECK-LABEL: select_lea_9:
; CHECK: # BB#0:
-; CHECK-NEXT: xorb $1, %dil
-; CHECK-NEXT: movzbl %dil, %eax
-; CHECK-NEXT: leaq -7(%rax,%rax,8), %rax
+; CHECK-NEXT: testb %dil, %dil
+; CHECK-NEXT: movl $2, %ecx
+; CHECK-NEXT: movq $-7, %rax
+; CHECK-NEXT: cmoveq %rcx, %rax
; CHECK-NEXT: retq
%sel = select i1 %cond, i64 -7, i64 2
ret i64 %sel
define i8 @select_pow2_diff(i1 zeroext %cond) {
; CHECK-LABEL: select_pow2_diff:
; CHECK: # BB#0:
-; CHECK-NEXT: shlb $4, %dil
-; CHECK-NEXT: orb $3, %dil
-; CHECK-NEXT: movl %edi, %eax
+; CHECK-NEXT: testb %dil, %dil
+; CHECK-NEXT: movb $19, %al
+; CHECK-NEXT: jne .LBB22_2
+; CHECK-NEXT: # BB#1:
+; CHECK-NEXT: movb $3, %al
+; CHECK-NEXT: .LBB22_2:
; CHECK-NEXT: retq
%sel = select i1 %cond, i8 19, i8 3
ret i8 %sel
define i16 @select_pow2_diff_invert(i1 zeroext %cond) {
; CHECK-LABEL: select_pow2_diff_invert:
; CHECK: # BB#0:
-; CHECK-NEXT: xorb $1, %dil
-; CHECK-NEXT: movzbl %dil, %eax
-; CHECK-NEXT: shll $6, %eax
-; CHECK-NEXT: orl $7, %eax
-; CHECK-NEXT: # kill: %AX<def> %AX<kill> %EAX<kill>
+; CHECK-NEXT: testb %dil, %dil
+; CHECK-NEXT: movw $7, %cx
+; CHECK-NEXT: movw $71, %ax
+; CHECK-NEXT: cmovnew %cx, %ax
; CHECK-NEXT: retq
%sel = select i1 %cond, i16 7, i16 71
ret i16 %sel
define i32 @select_pow2_diff_neg(i1 zeroext %cond) {
; CHECK-LABEL: select_pow2_diff_neg:
; CHECK: # BB#0:
-; CHECK-NEXT: shlb $4, %dil
-; CHECK-NEXT: movzbl %dil, %eax
-; CHECK-NEXT: orl $-25, %eax
+; CHECK-NEXT: testb %dil, %dil
+; CHECK-NEXT: movl $-9, %ecx
+; CHECK-NEXT: movl $-25, %eax
+; CHECK-NEXT: cmovnel %ecx, %eax
; CHECK-NEXT: retq
%sel = select i1 %cond, i32 -9, i32 -25
ret i32 %sel
define i64 @select_pow2_diff_neg_invert(i1 zeroext %cond) {
; CHECK-LABEL: select_pow2_diff_neg_invert:
; CHECK: # BB#0:
-; CHECK-NEXT: xorb $1, %dil
-; CHECK-NEXT: movzbl %dil, %eax
-; CHECK-NEXT: shlq $7, %rax
-; CHECK-NEXT: addq $-99, %rax
+; CHECK-NEXT: testb %dil, %dil
+; CHECK-NEXT: movl $29, %ecx
+; CHECK-NEXT: movq $-99, %rax
+; CHECK-NEXT: cmoveq %rcx, %rax
; CHECK-NEXT: retq
%sel = select i1 %cond, i64 -99, i64 29
ret i64 %sel
;
; X64-LABEL: t3:
; X64: # BB#0: # %entry
-; X64-NEXT: xorl %eax, %eax
-; X64-NEXT: testl %eax, %eax
-; X64-NEXT: sete %al
-; X64-NEXT: negq %rax
+; X64-NEXT: cmpl $1, %eax
+; X64-NEXT: sbbq %rax, %rax
; X64-NEXT: cmpq %rax, %rax
; X64-NEXT: xorl %eax, %eax
; X64-NEXT: retq
define i32 @t4(i64 %x) nounwind readnone ssp {
; X32-LABEL: t4:
; X32: # BB#0:
-; X32-NEXT: movl {{[0-9]+}}(%esp), %ecx
+; X32-NEXT: movl {{[0-9]+}}(%esp), %eax
+; X32-NEXT: orl {{[0-9]+}}(%esp), %eax
+; X32-NEXT: movl $-1, %eax
+; X32-NEXT: je .LBB3_2
+; X32-NEXT: # BB#1:
; X32-NEXT: xorl %eax, %eax
-; X32-NEXT: orl {{[0-9]+}}(%esp), %ecx
-; X32-NEXT: sete %al
-; X32-NEXT: negl %eax
+; X32-NEXT: .LBB3_2:
; X32-NEXT: retl
;
; X64-LABEL: t4: