From df02beb4165d3c07842878c6ed104622925be35b Mon Sep 17 00:00:00 2001 From: Craig Topper Date: Sun, 21 Apr 2019 07:12:56 +0000 Subject: [PATCH] [X86] Add the rounding control operand to the printing for some scalar FMA instructions. llvm-svn: 358844 --- llvm/lib/Target/X86/X86InstrAVX512.td | 2 +- llvm/test/CodeGen/X86/avx512-fma-intrinsics.ll | 4 ++-- llvm/test/CodeGen/X86/avx512-intrinsics-fast-isel.ll | 4 ++-- 3 files changed, 5 insertions(+), 5 deletions(-) diff --git a/llvm/lib/Target/X86/X86InstrAVX512.td b/llvm/lib/Target/X86/X86InstrAVX512.td index d3adc06..420a70a 100644 --- a/llvm/lib/Target/X86/X86InstrAVX512.td +++ b/llvm/lib/Target/X86/X86InstrAVX512.td @@ -6825,7 +6825,7 @@ let Constraints = "$src1 = $dst", hasSideEffects = 0 in { def rb : AVX512FMA3S, EVEX_B, EVEX_RC, Sched<[SchedWriteFMA.Scl]>; }// isCodeGenOnly = 1 diff --git a/llvm/test/CodeGen/X86/avx512-fma-intrinsics.ll b/llvm/test/CodeGen/X86/avx512-fma-intrinsics.ll index 9943806..ff737b9 100644 --- a/llvm/test/CodeGen/X86/avx512-fma-intrinsics.ll +++ b/llvm/test/CodeGen/X86/avx512-fma-intrinsics.ll @@ -1154,14 +1154,14 @@ define <4 x float> @foo() { ; X86: # %bb.0: # %entry ; X86-NEXT: vmovss (%eax), %xmm0 # EVEX TO VEX Compression encoding: [0xc5,0xfa,0x10,0x00] ; X86-NEXT: # xmm0 = mem[0],zero,zero,zero -; X86-NEXT: vfmsub213ss %xmm0, %xmm0, %xmm0 # encoding: [0x62,0xf2,0x7d,0x38,0xab,0xc0] +; X86-NEXT: vfmsub213ss {rd-sae}, %xmm0, %xmm0, %xmm0 # encoding: [0x62,0xf2,0x7d,0x38,0xab,0xc0] ; X86-NEXT: retl # encoding: [0xc3] ; ; X64-LABEL: foo: ; X64: # %bb.0: # %entry ; X64-NEXT: vmovss (%rax), %xmm0 # EVEX TO VEX Compression encoding: [0xc5,0xfa,0x10,0x00] ; X64-NEXT: # xmm0 = mem[0],zero,zero,zero -; X64-NEXT: vfmsub213ss %xmm0, %xmm0, %xmm0 # encoding: [0x62,0xf2,0x7d,0x38,0xab,0xc0] +; X64-NEXT: vfmsub213ss {rd-sae}, %xmm0, %xmm0, %xmm0 # encoding: [0x62,0xf2,0x7d,0x38,0xab,0xc0] ; X64-NEXT: retq # encoding: [0xc3] entry: %0 = load <4 x float>, <4 x float>* undef, align 16 diff --git a/llvm/test/CodeGen/X86/avx512-intrinsics-fast-isel.ll b/llvm/test/CodeGen/X86/avx512-intrinsics-fast-isel.ll index e222c0e..80d5466 100644 --- a/llvm/test/CodeGen/X86/avx512-intrinsics-fast-isel.ll +++ b/llvm/test/CodeGen/X86/avx512-intrinsics-fast-isel.ll @@ -5875,7 +5875,7 @@ define <2 x double> @test_mm_mask3_fmsub_round_sd(<2 x double> %__W, <2 x double ; X86: # %bb.0: # %entry ; X86-NEXT: movb {{[0-9]+}}(%esp), %al ; X86-NEXT: vxorpd {{\.LCPI.*}}, %xmm2, %xmm3 -; X86-NEXT: vfmadd213sd %xmm3, %xmm0, %xmm1 +; X86-NEXT: vfmadd213sd {rn-sae}, %xmm3, %xmm0, %xmm1 ; X86-NEXT: kmovw %eax, %k1 ; X86-NEXT: vmovsd %xmm1, %xmm2, %xmm2 {%k1} ; X86-NEXT: vmovapd %xmm2, %xmm0 @@ -5884,7 +5884,7 @@ define <2 x double> @test_mm_mask3_fmsub_round_sd(<2 x double> %__W, <2 x double ; X64-LABEL: test_mm_mask3_fmsub_round_sd: ; X64: # %bb.0: # %entry ; X64-NEXT: vxorpd {{.*}}(%rip), %xmm2, %xmm3 -; X64-NEXT: vfmadd213sd %xmm3, %xmm0, %xmm1 +; X64-NEXT: vfmadd213sd {rn-sae}, %xmm3, %xmm0, %xmm1 ; X64-NEXT: kmovw %edi, %k1 ; X64-NEXT: vmovsd %xmm1, %xmm2, %xmm2 {%k1} ; X64-NEXT: vmovapd %xmm2, %xmm0 -- 2.7.4