From b8cb73dd38664b6e4ff6bbb0a4143c6e209038f0 Mon Sep 17 00:00:00 2001 From: Craig Topper Date: Tue, 26 Nov 2019 10:32:40 -0800 Subject: [PATCH] [X86] Pre-commit test modifications for D68857. NFC Patch by Chen Liu(LiuChen3) Differential Revision: https://reviews.llvm.org/D70706 --- llvm/test/CodeGen/X86/fp-strict-scalar.ll | 4 ++-- llvm/test/CodeGen/X86/fp80-strict-scalar.ll | 13 ++++--------- 2 files changed, 6 insertions(+), 11 deletions(-) diff --git a/llvm/test/CodeGen/X86/fp-strict-scalar.ll b/llvm/test/CodeGen/X86/fp-strict-scalar.ll index 8813047..8d2e710 100644 --- a/llvm/test/CodeGen/X86/fp-strict-scalar.ll +++ b/llvm/test/CodeGen/X86/fp-strict-scalar.ll @@ -16,7 +16,7 @@ declare float @llvm.experimental.constrained.fmul.f32(float, float, metadata, me declare double @llvm.experimental.constrained.fdiv.f64(double, double, metadata, metadata) declare float @llvm.experimental.constrained.fdiv.f32(float, float, metadata, metadata) declare double @llvm.experimental.constrained.fpext.f64.f32(float, metadata) -declare float @llvm.experimental.constrained.fptrunc.f64.f32(double, metadata, metadata) +declare float @llvm.experimental.constrained.fptrunc.f32.f64(double, metadata, metadata) declare float @llvm.experimental.constrained.sqrt.f32(float, metadata, metadata) declare double @llvm.experimental.constrained.sqrt.f64(double, metadata, metadata) @@ -480,7 +480,7 @@ define void @fptrunc_double_to_f32(double* %val, float *%ret) nounwind strictfp ; X87-NEXT: popl %eax ; X87-NEXT: retl %1 = load double, double* %val, align 8 - %res = call float @llvm.experimental.constrained.fptrunc.f64.f32(double %1, + %res = call float @llvm.experimental.constrained.fptrunc.f32.f64(double %1, metadata !"round.dynamic", metadata !"fpexcept.strict") #0 store float %res, float* %ret, align 4 diff --git a/llvm/test/CodeGen/X86/fp80-strict-scalar.ll b/llvm/test/CodeGen/X86/fp80-strict-scalar.ll index 2795008..8b80c015 100644 --- a/llvm/test/CodeGen/X86/fp80-strict-scalar.ll +++ b/llvm/test/CodeGen/X86/fp80-strict-scalar.ll @@ -9,8 +9,8 @@ declare x86_fp80 @llvm.experimental.constrained.fdiv.x86_fp80(x86_fp80, x86_fp80 declare x86_fp80 @llvm.experimental.constrained.fpext.x86_fp80.f32(float, metadata) declare x86_fp80 @llvm.experimental.constrained.fpext.x86_fp80.f64(double, metadata) declare x86_fp80 @llvm.experimental.constrained.sqrt.x86_fp80(x86_fp80, metadata, metadata) -declare float @llvm.experimental.constrained.fptrunc.x86_fp80.f32(x86_fp80, metadata, metadata) -declare double @llvm.experimental.constrained.fptrunc.x86_fp80.f64(x86_fp80, metadata, metadata) +declare float @llvm.experimental.constrained.fptrunc.f32.x86_fp80(x86_fp80, metadata, metadata) +declare double @llvm.experimental.constrained.fptrunc.f64.x86_fp80(x86_fp80, metadata, metadata) define x86_fp80 @fadd_fp80(x86_fp80 %a, x86_fp80 %b) nounwind strictfp { ; X86-LABEL: fadd_fp80: @@ -106,7 +106,6 @@ define x86_fp80 @fpext_f32_to_fp80(float %a) nounwind strictfp { %ret = call x86_fp80 @llvm.experimental.constrained.fpext.x86_fp80.f32(float %a, metadata !"fpexcept.strict") #0 ret x86_fp80 %ret - } define x86_fp80 @fpext_f64_to_fp80(double %a) nounwind strictfp { @@ -123,7 +122,6 @@ define x86_fp80 @fpext_f64_to_fp80(double %a) nounwind strictfp { %ret = call x86_fp80 @llvm.experimental.constrained.fpext.x86_fp80.f64(double %a, metadata !"fpexcept.strict") #0 ret x86_fp80 %ret - } define float @fptrunc_fp80_to_f32(x86_fp80 %a) nounwind strictfp { @@ -142,11 +140,10 @@ define float @fptrunc_fp80_to_f32(x86_fp80 %a) nounwind strictfp { ; X64-NEXT: fstps -{{[0-9]+}}(%rsp) ; X64-NEXT: movss {{.*#+}} xmm0 = mem[0],zero,zero,zero ; X64-NEXT: retq - %ret = call float @llvm.experimental.constrained.fptrunc.x86_fp80.f32(x86_fp80 %a, + %ret = call float @llvm.experimental.constrained.fptrunc.f32.x86_fp80(x86_fp80 %a, metadata !"round.dynamic", metadata !"fpexcept.strict") #0 ret float %ret - } define double @fptrunc_fp80_to_f64(x86_fp80 %a) nounwind strictfp { @@ -169,11 +166,10 @@ define double @fptrunc_fp80_to_f64(x86_fp80 %a) nounwind strictfp { ; X64-NEXT: fstpl -{{[0-9]+}}(%rsp) ; X64-NEXT: movsd {{.*#+}} xmm0 = mem[0],zero ; X64-NEXT: retq - %ret = call double @llvm.experimental.constrained.fptrunc.x86_fp80.f64(x86_fp80 %a, + %ret = call double @llvm.experimental.constrained.fptrunc.f64.x86_fp80(x86_fp80 %a, metadata !"round.dynamic", metadata !"fpexcept.strict") #0 ret double %ret - } define x86_fp80 @fsqrt_fp80(x86_fp80 %a) nounwind strictfp { @@ -192,7 +188,6 @@ define x86_fp80 @fsqrt_fp80(x86_fp80 %a) nounwind strictfp { metadata !"round.dynamic", metadata !"fpexcept.strict") #0 ret x86_fp80 %ret - } attributes #0 = { strictfp } -- 2.7.4