From f6cd43c0987a5ecec2bdb812a80792c5b90281f9 Mon Sep 17 00:00:00 2001 From: Craig Topper Date: Fri, 15 Oct 2021 17:39:00 -0700 Subject: [PATCH] [X86] Add more tests for D111858. NFC Add tests with sub instead of neg. --- llvm/test/CodeGen/X86/neg-abs.ll | 167 +++++++++++++++++++++++++++++++++++++++ 1 file changed, 167 insertions(+) diff --git a/llvm/test/CodeGen/X86/neg-abs.ll b/llvm/test/CodeGen/X86/neg-abs.ll index 1eaa559..a8320cb 100644 --- a/llvm/test/CodeGen/X86/neg-abs.ll +++ b/llvm/test/CodeGen/X86/neg-abs.ll @@ -154,3 +154,170 @@ define i128 @neg_abs_i128(i128 %x) nounwind { %neg = sub nsw i128 0, %abs ret i128 %neg } + +define i8 @sub_abs_i8(i8 %x, i8 %y) nounwind { +; X86-LABEL: sub_abs_i8: +; X86: # %bb.0: +; X86-NEXT: movb {{[0-9]+}}(%esp), %al +; X86-NEXT: movb {{[0-9]+}}(%esp), %cl +; X86-NEXT: movl %ecx, %edx +; X86-NEXT: sarb $7, %dl +; X86-NEXT: addb %dl, %cl +; X86-NEXT: xorb %dl, %cl +; X86-NEXT: subb %cl, %al +; X86-NEXT: retl +; +; X64-LABEL: sub_abs_i8: +; X64: # %bb.0: +; X64-NEXT: movl %esi, %eax +; X64-NEXT: movl %edi, %ecx +; X64-NEXT: sarb $7, %cl +; X64-NEXT: addb %cl, %dil +; X64-NEXT: xorb %cl, %dil +; X64-NEXT: subb %dil, %al +; X64-NEXT: # kill: def $al killed $al killed $eax +; X64-NEXT: retq + %abs = tail call i8 @llvm.abs.i8(i8 %x, i1 false) + %neg = sub nsw i8 %y, %abs + ret i8 %neg +} + +define i16 @sub_abs_i16(i16 %x, i16 %y) nounwind { +; X86-LABEL: sub_abs_i16: +; X86: # %bb.0: +; X86-NEXT: movl {{[0-9]+}}(%esp), %eax +; X86-NEXT: movswl {{[0-9]+}}(%esp), %ecx +; X86-NEXT: movl %ecx, %edx +; X86-NEXT: sarl $15, %edx +; X86-NEXT: addl %edx, %ecx +; X86-NEXT: xorl %edx, %ecx +; X86-NEXT: subl %ecx, %eax +; X86-NEXT: # kill: def $ax killed $ax killed $eax +; X86-NEXT: retl +; +; X64-LABEL: sub_abs_i16: +; X64: # %bb.0: +; X64-NEXT: movl %esi, %eax +; X64-NEXT: movl %edi, %ecx +; X64-NEXT: negw %cx +; X64-NEXT: cmovsw %di, %cx +; X64-NEXT: subl %ecx, %eax +; X64-NEXT: # kill: def $ax killed $ax killed $eax +; X64-NEXT: retq + %abs = tail call i16 @llvm.abs.i16(i16 %x, i1 false) + %neg = sub i16 %y, %abs + ret i16 %neg +} + +define i32 @sub_abs_i32(i32 %x, i32 %y) nounwind { +; X86-LABEL: sub_abs_i32: +; X86: # %bb.0: +; X86-NEXT: movl {{[0-9]+}}(%esp), %eax +; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx +; X86-NEXT: movl %ecx, %edx +; X86-NEXT: sarl $31, %edx +; X86-NEXT: addl %edx, %ecx +; X86-NEXT: xorl %edx, %ecx +; X86-NEXT: subl %ecx, %eax +; X86-NEXT: retl +; +; X64-LABEL: sub_abs_i32: +; X64: # %bb.0: +; X64-NEXT: movl %esi, %eax +; X64-NEXT: movl %edi, %ecx +; X64-NEXT: negl %ecx +; X64-NEXT: cmovsl %edi, %ecx +; X64-NEXT: subl %ecx, %eax +; X64-NEXT: retq + %abs = tail call i32 @llvm.abs.i32(i32 %x, i1 false) + %neg = sub i32 %y, %abs + ret i32 %neg +} + +define i64 @sub_abs_i64(i64 %x, i64 %y) nounwind { +; X86-LABEL: sub_abs_i64: +; X86: # %bb.0: +; X86-NEXT: pushl %edi +; X86-NEXT: pushl %esi +; X86-NEXT: movl {{[0-9]+}}(%esp), %eax +; X86-NEXT: movl {{[0-9]+}}(%esp), %edx +; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx +; X86-NEXT: movl %ecx, %esi +; X86-NEXT: sarl $31, %esi +; X86-NEXT: movl {{[0-9]+}}(%esp), %edi +; X86-NEXT: addl %esi, %edi +; X86-NEXT: adcl %esi, %ecx +; X86-NEXT: xorl %esi, %ecx +; X86-NEXT: xorl %esi, %edi +; X86-NEXT: subl %edi, %eax +; X86-NEXT: sbbl %ecx, %edx +; X86-NEXT: popl %esi +; X86-NEXT: popl %edi +; X86-NEXT: retl +; +; X64-LABEL: sub_abs_i64: +; X64: # %bb.0: +; X64-NEXT: movq %rsi, %rax +; X64-NEXT: movq %rdi, %rcx +; X64-NEXT: negq %rcx +; X64-NEXT: cmovsq %rdi, %rcx +; X64-NEXT: subq %rcx, %rax +; X64-NEXT: retq + %abs = tail call i64 @llvm.abs.i64(i64 %x, i1 false) + %neg = sub i64 %y, %abs + ret i64 %neg +} + +define i128 @sub_abs_i128(i128 %x, i128 %y) nounwind { +; X86-LABEL: sub_abs_i128: +; X86: # %bb.0: +; X86-NEXT: pushl %edi +; X86-NEXT: pushl %esi +; X86-NEXT: movl {{[0-9]+}}(%esp), %eax +; X86-NEXT: movl %eax, %ecx +; X86-NEXT: sarl $31, %ecx +; X86-NEXT: movl {{[0-9]+}}(%esp), %edx +; X86-NEXT: addl %ecx, %edx +; X86-NEXT: movl {{[0-9]+}}(%esp), %esi +; X86-NEXT: adcl %ecx, %esi +; X86-NEXT: movl {{[0-9]+}}(%esp), %edi +; X86-NEXT: adcl %ecx, %edi +; X86-NEXT: adcl %ecx, %eax +; X86-NEXT: xorl %ecx, %eax +; X86-NEXT: xorl %ecx, %edi +; X86-NEXT: xorl %ecx, %esi +; X86-NEXT: xorl %ecx, %edx +; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx +; X86-NEXT: subl %edx, %ecx +; X86-NEXT: movl {{[0-9]+}}(%esp), %edx +; X86-NEXT: sbbl %esi, %edx +; X86-NEXT: movl {{[0-9]+}}(%esp), %esi +; X86-NEXT: sbbl %edi, %esi +; X86-NEXT: movl {{[0-9]+}}(%esp), %edi +; X86-NEXT: sbbl %eax, %edi +; X86-NEXT: movl {{[0-9]+}}(%esp), %eax +; X86-NEXT: movl %ecx, (%eax) +; X86-NEXT: movl %edx, 4(%eax) +; X86-NEXT: movl %esi, 8(%eax) +; X86-NEXT: movl %edi, 12(%eax) +; X86-NEXT: popl %esi +; X86-NEXT: popl %edi +; X86-NEXT: retl $4 +; +; X64-LABEL: sub_abs_i128: +; X64: # %bb.0: +; X64-NEXT: movq %rdx, %rax +; X64-NEXT: movq %rsi, %rdx +; X64-NEXT: sarq $63, %rdx +; X64-NEXT: addq %rdx, %rdi +; X64-NEXT: adcq %rdx, %rsi +; X64-NEXT: xorq %rdx, %rsi +; X64-NEXT: xorq %rdx, %rdi +; X64-NEXT: subq %rdi, %rax +; X64-NEXT: sbbq %rsi, %rcx +; X64-NEXT: movq %rcx, %rdx +; X64-NEXT: retq + %abs = tail call i128 @llvm.abs.i128(i128 %x, i1 false) + %neg = sub i128 %y, %abs + ret i128 %neg +} -- 2.7.4