From 237e208f16dfd316d12e7dbf6000be614b73422b Mon Sep 17 00:00:00 2001 From: Sanjay Patel Date: Tue, 5 Feb 2019 15:26:42 +0000 Subject: [PATCH] [AArch64][x86] add tests for unsigned subtract with overflow; NFC llvm-svn: 353178 --- llvm/test/CodeGen/AArch64/cgp-usubo.ll | 197 +++++++++++++++++++++++++++++++ llvm/test/CodeGen/X86/cgp-usubo.ll | 205 +++++++++++++++++++++++++++++++++ 2 files changed, 402 insertions(+) create mode 100644 llvm/test/CodeGen/AArch64/cgp-usubo.ll create mode 100644 llvm/test/CodeGen/X86/cgp-usubo.ll diff --git a/llvm/test/CodeGen/AArch64/cgp-usubo.ll b/llvm/test/CodeGen/AArch64/cgp-usubo.ll new file mode 100644 index 0000000..1afa585 --- /dev/null +++ b/llvm/test/CodeGen/AArch64/cgp-usubo.ll @@ -0,0 +1,197 @@ +; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py +; RUN: llc < %s -mtriple=aarch64-- | FileCheck %s + +; CodeGenPrepare is expected to form overflow intrinsics to improve DAG/isel. + +define i1 @usubo_ult_i64(i64 %x, i64 %y, i64* %p) nounwind { +; CHECK-LABEL: usubo_ult_i64: +; CHECK: // %bb.0: +; CHECK-NEXT: subs x8, x0, x1 +; CHECK-NEXT: cset w0, lo +; CHECK-NEXT: str x8, [x2] +; CHECK-NEXT: ret + %s = sub i64 %x, %y + store i64 %s, i64* %p + %ov = icmp ult i64 %x, %y + ret i1 %ov +} + +; Verify insertion point for single-BB. Toggle predicate. + +define i1 @usubo_ugt_i32(i32 %x, i32 %y, i32* %p) nounwind { +; CHECK-LABEL: usubo_ugt_i32: +; CHECK: // %bb.0: +; CHECK-NEXT: cmp w1, w0 +; CHECK-NEXT: cset w8, hi +; CHECK-NEXT: sub w9, w0, w1 +; CHECK-NEXT: mov w0, w8 +; CHECK-NEXT: str w9, [x2] +; CHECK-NEXT: ret + %ov = icmp ugt i32 %y, %x + %s = sub i32 %x, %y + store i32 %s, i32* %p + ret i1 %ov +} + +; Constant operand should match. + +define i1 @usubo_ugt_constant_op0_i8(i8 %x, i8* %p) nounwind { +; CHECK-LABEL: usubo_ugt_constant_op0_i8: +; CHECK: // %bb.0: +; CHECK-NEXT: and w8, w0, #0xff +; CHECK-NEXT: mov w9, #42 +; CHECK-NEXT: cmp w8, #42 // =42 +; CHECK-NEXT: sub w9, w9, w0 +; CHECK-NEXT: cset w0, hi +; CHECK-NEXT: strb w9, [x1] +; CHECK-NEXT: ret + %s = sub i8 42, %x + %ov = icmp ugt i8 %x, 42 + store i8 %s, i8* %p + ret i1 %ov +} + +; Compare with constant operand 0 is canonicalized by commuting, but verify match for non-canonical form. + +define i1 @usubo_ult_constant_op0_i16(i16 %x, i16* %p) nounwind { +; CHECK-LABEL: usubo_ult_constant_op0_i16: +; CHECK: // %bb.0: +; CHECK-NEXT: and w8, w0, #0xffff +; CHECK-NEXT: mov w9, #43 +; CHECK-NEXT: cmp w8, #43 // =43 +; CHECK-NEXT: sub w9, w9, w0 +; CHECK-NEXT: cset w0, hi +; CHECK-NEXT: strh w9, [x1] +; CHECK-NEXT: ret + %s = sub i16 43, %x + %ov = icmp ult i16 43, %x + store i16 %s, i16* %p + ret i1 %ov +} + +; Subtract with constant operand 1 is canonicalized to add. + +define i1 @usubo_ult_constant_op1_i16(i16 %x, i16* %p) nounwind { +; CHECK-LABEL: usubo_ult_constant_op1_i16: +; CHECK: // %bb.0: +; CHECK-NEXT: and w8, w0, #0xffff +; CHECK-NEXT: cmp w8, #44 // =44 +; CHECK-NEXT: sub w9, w0, #44 // =44 +; CHECK-NEXT: cset w0, lo +; CHECK-NEXT: strh w9, [x1] +; CHECK-NEXT: ret + %s = add i16 %x, -44 + %ov = icmp ult i16 %x, 44 + store i16 %s, i16* %p + ret i1 %ov +} + +define i1 @usubo_ugt_constant_op1_i8(i8 %x, i8* %p) nounwind { +; CHECK-LABEL: usubo_ugt_constant_op1_i8: +; CHECK: // %bb.0: +; CHECK-NEXT: and w8, w0, #0xff +; CHECK-NEXT: cmp w8, #45 // =45 +; CHECK-NEXT: cset w8, lo +; CHECK-NEXT: sub w9, w0, #45 // =45 +; CHECK-NEXT: mov w0, w8 +; CHECK-NEXT: strb w9, [x1] +; CHECK-NEXT: ret + %ov = icmp ugt i8 45, %x + %s = add i8 %x, -45 + store i8 %s, i8* %p + ret i1 %ov +} + +; Special-case: subtract 1 changes the compare predicate and constant. + +define i1 @usubo_eq_constant1_op1_i32(i32 %x, i32* %p) nounwind { +; CHECK-LABEL: usubo_eq_constant1_op1_i32: +; CHECK: // %bb.0: +; CHECK-NEXT: cmp w0, #0 // =0 +; CHECK-NEXT: sub w8, w0, #1 // =1 +; CHECK-NEXT: cset w0, eq +; CHECK-NEXT: str w8, [x1] +; CHECK-NEXT: ret + %s = add i32 %x, -1 + %ov = icmp eq i32 %x, 0 + store i32 %s, i32* %p + ret i1 %ov +} + +; Verify insertion point for multi-BB. + +declare void @call(i1) + +define i1 @usubo_ult_sub_dominates_i64(i64 %x, i64 %y, i64* %p, i1 %cond) nounwind { +; CHECK-LABEL: usubo_ult_sub_dominates_i64: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: tbz w3, #0, .LBB7_2 +; CHECK-NEXT: // %bb.1: // %t +; CHECK-NEXT: subs x8, x0, x1 +; CHECK-NEXT: cset w0, lo +; CHECK-NEXT: str x8, [x2] +; CHECK-NEXT: ret +; CHECK-NEXT: .LBB7_2: // %f +; CHECK-NEXT: and w0, w3, #0x1 +; CHECK-NEXT: ret +entry: + br i1 %cond, label %t, label %f + +t: + %s = sub i64 %x, %y + store i64 %s, i64* %p + br i1 %cond, label %end, label %f + +f: + ret i1 %cond + +end: + %ov = icmp ult i64 %x, %y + ret i1 %ov +} + +define i1 @usubo_ult_cmp_dominates_i64(i64 %x, i64 %y, i64* %p, i1 %cond) nounwind { +; CHECK-LABEL: usubo_ult_cmp_dominates_i64: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: str x22, [sp, #-48]! // 8-byte Folded Spill +; CHECK-NEXT: stp x21, x20, [sp, #16] // 16-byte Folded Spill +; CHECK-NEXT: mov w20, w3 +; CHECK-NEXT: stp x19, x30, [sp, #32] // 16-byte Folded Spill +; CHECK-NEXT: tbz w3, #0, .LBB8_3 +; CHECK-NEXT: // %bb.1: // %t +; CHECK-NEXT: cmp x0, x1 +; CHECK-NEXT: mov x22, x0 +; CHECK-NEXT: cset w0, lo +; CHECK-NEXT: mov x19, x2 +; CHECK-NEXT: mov x21, x1 +; CHECK-NEXT: bl call +; CHECK-NEXT: subs x8, x22, x21 +; CHECK-NEXT: b.hs .LBB8_3 +; CHECK-NEXT: // %bb.2: // %end +; CHECK-NEXT: cset w0, lo +; CHECK-NEXT: str x8, [x19] +; CHECK-NEXT: b .LBB8_4 +; CHECK-NEXT: .LBB8_3: // %f +; CHECK-NEXT: and w0, w20, #0x1 +; CHECK-NEXT: .LBB8_4: // %f +; CHECK-NEXT: ldp x19, x30, [sp, #32] // 16-byte Folded Reload +; CHECK-NEXT: ldp x21, x20, [sp, #16] // 16-byte Folded Reload +; CHECK-NEXT: ldr x22, [sp], #48 // 8-byte Folded Reload +; CHECK-NEXT: ret +entry: + br i1 %cond, label %t, label %f + +t: + %ov = icmp ult i64 %x, %y + call void @call(i1 %ov) + br i1 %ov, label %end, label %f + +f: + ret i1 %cond + +end: + %s = sub i64 %x, %y + store i64 %s, i64* %p + ret i1 %ov +} + diff --git a/llvm/test/CodeGen/X86/cgp-usubo.ll b/llvm/test/CodeGen/X86/cgp-usubo.ll new file mode 100644 index 0000000..39361d1 --- /dev/null +++ b/llvm/test/CodeGen/X86/cgp-usubo.ll @@ -0,0 +1,205 @@ +; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py +; RUN: llc < %s -mtriple=x86_64-- | FileCheck %s + +; CodeGenPrepare is expected to form overflow intrinsics to improve DAG/isel. + +define i1 @usubo_ult_i64(i64 %x, i64 %y, i64* %p) nounwind { +; CHECK-LABEL: usubo_ult_i64: +; CHECK: # %bb.0: +; CHECK-NEXT: subq %rsi, %rdi +; CHECK-NEXT: movq %rdi, (%rdx) +; CHECK-NEXT: setb %al +; CHECK-NEXT: retq + %s = sub i64 %x, %y + store i64 %s, i64* %p + %ov = icmp ult i64 %x, %y + ret i1 %ov +} + +; Verify insertion point for single-BB. Toggle predicate. + +define i1 @usubo_ugt_i32(i32 %x, i32 %y, i32* %p) nounwind { +; CHECK-LABEL: usubo_ugt_i32: +; CHECK: # %bb.0: +; CHECK-NEXT: cmpl %edi, %esi +; CHECK-NEXT: seta %al +; CHECK-NEXT: subl %esi, %edi +; CHECK-NEXT: movl %edi, (%rdx) +; CHECK-NEXT: retq + %ov = icmp ugt i32 %y, %x + %s = sub i32 %x, %y + store i32 %s, i32* %p + ret i1 %ov +} + +; Constant operand should match. + +define i1 @usubo_ugt_constant_op0_i8(i8 %x, i8* %p) nounwind { +; CHECK-LABEL: usubo_ugt_constant_op0_i8: +; CHECK: # %bb.0: +; CHECK-NEXT: movb $42, %cl +; CHECK-NEXT: subb %dil, %cl +; CHECK-NEXT: cmpb $42, %dil +; CHECK-NEXT: seta %al +; CHECK-NEXT: movb %cl, (%rsi) +; CHECK-NEXT: retq + %s = sub i8 42, %x + %ov = icmp ugt i8 %x, 42 + store i8 %s, i8* %p + ret i1 %ov +} + +; Compare with constant operand 0 is canonicalized by commuting, but verify match for non-canonical form. + +define i1 @usubo_ult_constant_op0_i16(i16 %x, i16* %p) nounwind { +; CHECK-LABEL: usubo_ult_constant_op0_i16: +; CHECK: # %bb.0: +; CHECK-NEXT: movl $43, %ecx +; CHECK-NEXT: subl %edi, %ecx +; CHECK-NEXT: cmpw $43, %di +; CHECK-NEXT: seta %al +; CHECK-NEXT: movw %cx, (%rsi) +; CHECK-NEXT: retq + %s = sub i16 43, %x + %ov = icmp ult i16 43, %x + store i16 %s, i16* %p + ret i1 %ov +} + +; Subtract with constant operand 1 is canonicalized to add. + +define i1 @usubo_ult_constant_op1_i16(i16 %x, i16* %p) nounwind { +; CHECK-LABEL: usubo_ult_constant_op1_i16: +; CHECK: # %bb.0: +; CHECK-NEXT: movl %edi, %ecx +; CHECK-NEXT: addl $-44, %ecx +; CHECK-NEXT: cmpw $44, %di +; CHECK-NEXT: setb %al +; CHECK-NEXT: movw %cx, (%rsi) +; CHECK-NEXT: retq + %s = add i16 %x, -44 + %ov = icmp ult i16 %x, 44 + store i16 %s, i16* %p + ret i1 %ov +} + +define i1 @usubo_ugt_constant_op1_i8(i8 %x, i8* %p) nounwind { +; CHECK-LABEL: usubo_ugt_constant_op1_i8: +; CHECK: # %bb.0: +; CHECK-NEXT: cmpb $45, %dil +; CHECK-NEXT: setb %al +; CHECK-NEXT: addb $-45, %dil +; CHECK-NEXT: movb %dil, (%rsi) +; CHECK-NEXT: retq + %ov = icmp ugt i8 45, %x + %s = add i8 %x, -45 + store i8 %s, i8* %p + ret i1 %ov +} + +; Special-case: subtract 1 changes the compare predicate and constant. + +define i1 @usubo_eq_constant1_op1_i32(i32 %x, i32* %p) nounwind { +; CHECK-LABEL: usubo_eq_constant1_op1_i32: +; CHECK: # %bb.0: +; CHECK-NEXT: # kill: def $edi killed $edi def $rdi +; CHECK-NEXT: leal -1(%rdi), %ecx +; CHECK-NEXT: testl %edi, %edi +; CHECK-NEXT: sete %al +; CHECK-NEXT: movl %ecx, (%rsi) +; CHECK-NEXT: retq + %s = add i32 %x, -1 + %ov = icmp eq i32 %x, 0 + store i32 %s, i32* %p + ret i1 %ov +} + +; Verify insertion point for multi-BB. + +declare void @call(i1) + +define i1 @usubo_ult_sub_dominates_i64(i64 %x, i64 %y, i64* %p, i1 %cond) nounwind { +; CHECK-LABEL: usubo_ult_sub_dominates_i64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: testb $1, %cl +; CHECK-NEXT: je .LBB7_2 +; CHECK-NEXT: # %bb.1: # %t +; CHECK-NEXT: movq %rdi, %rax +; CHECK-NEXT: subq %rsi, %rax +; CHECK-NEXT: movq %rax, (%rdx) +; CHECK-NEXT: testb $1, %cl +; CHECK-NEXT: je .LBB7_2 +; CHECK-NEXT: # %bb.3: # %end +; CHECK-NEXT: cmpq %rsi, %rdi +; CHECK-NEXT: setb %al +; CHECK-NEXT: retq +; CHECK-NEXT: .LBB7_2: # %f +; CHECK-NEXT: movl %ecx, %eax +; CHECK-NEXT: retq +entry: + br i1 %cond, label %t, label %f + +t: + %s = sub i64 %x, %y + store i64 %s, i64* %p + br i1 %cond, label %end, label %f + +f: + ret i1 %cond + +end: + %ov = icmp ult i64 %x, %y + ret i1 %ov +} + +define i1 @usubo_ult_cmp_dominates_i64(i64 %x, i64 %y, i64* %p, i1 %cond) nounwind { +; CHECK-LABEL: usubo_ult_cmp_dominates_i64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: pushq %rbp +; CHECK-NEXT: pushq %r15 +; CHECK-NEXT: pushq %r14 +; CHECK-NEXT: pushq %rbx +; CHECK-NEXT: pushq %rax +; CHECK-NEXT: movl %ecx, %ebp +; CHECK-NEXT: testb $1, %bpl +; CHECK-NEXT: je .LBB8_2 +; CHECK-NEXT: # %bb.1: # %t +; CHECK-NEXT: movq %rdx, %r14 +; CHECK-NEXT: movq %rsi, %r15 +; CHECK-NEXT: movq %rdi, %rbx +; CHECK-NEXT: xorl %edi, %edi +; CHECK-NEXT: cmpq %rsi, %rbx +; CHECK-NEXT: setb %dil +; CHECK-NEXT: callq call +; CHECK-NEXT: subq %r15, %rbx +; CHECK-NEXT: jae .LBB8_2 +; CHECK-NEXT: # %bb.4: # %end +; CHECK-NEXT: setb %al +; CHECK-NEXT: movq %rbx, (%r14) +; CHECK-NEXT: jmp .LBB8_3 +; CHECK-NEXT: .LBB8_2: # %f +; CHECK-NEXT: movl %ebp, %eax +; CHECK-NEXT: .LBB8_3: # %f +; CHECK-NEXT: addq $8, %rsp +; CHECK-NEXT: popq %rbx +; CHECK-NEXT: popq %r14 +; CHECK-NEXT: popq %r15 +; CHECK-NEXT: popq %rbp +; CHECK-NEXT: retq +entry: + br i1 %cond, label %t, label %f + +t: + %ov = icmp ult i64 %x, %y + call void @call(i1 %ov) + br i1 %ov, label %end, label %f + +f: + ret i1 %cond + +end: + %s = sub i64 %x, %y + store i64 %s, i64* %p + ret i1 %ov +} + -- 2.7.4