From: jacquesguan Date: Fri, 16 Sep 2022 02:31:18 +0000 (+0800) Subject: [RISCV][test] Add precommit test for D132923. X-Git-Tag: upstream/17.0.6~33394 X-Git-Url: http://review.tizen.org/git/?a=commitdiff_plain;h=461d49909c2c1bd1d72fcfd53799c6e0267aee49;p=platform%2Fupstream%2Fllvm.git [RISCV][test] Add precommit test for D132923. --- diff --git a/llvm/test/CodeGen/RISCV/rvv/fixed-vectors-peephole-vmerge-vops.ll b/llvm/test/CodeGen/RISCV/rvv/fixed-vectors-peephole-vmerge-vops.ll index 4274b0e..bbd6b8e 100644 --- a/llvm/test/CodeGen/RISCV/rvv/fixed-vectors-peephole-vmerge-vops.ll +++ b/llvm/test/CodeGen/RISCV/rvv/fixed-vectors-peephole-vmerge-vops.ll @@ -189,3 +189,195 @@ define <8 x i32> @vpmerge_vpload2(<8 x i32> %passthru, <8 x i32> * %p, <8 x i32> %b = call <8 x i32> @llvm.vp.merge.nxv2i32(<8 x i1> %m, <8 x i32> %a, <8 x i32> %passthru, i32 %vl) ret <8 x i32> %b } + +declare <8 x i16> @llvm.vp.select.nxv2i16(<8 x i1>, <8 x i16>, <8 x i16>, i32) +declare <8 x i32> @llvm.vp.select.nxv2i32(<8 x i1>, <8 x i32>, <8 x i32>, i32) +declare <8 x float> @llvm.vp.select.nxv2f32(<8 x i1>, <8 x float>, <8 x float>, i32) +declare <8 x double> @llvm.vp.select.nxv2f64(<8 x i1>, <8 x double>, <8 x double>, i32) + +; Test binary operator with vp.select and vp.add. +define <8 x i32> @vpselect_vpadd(<8 x i32> %passthru, <8 x i32> %x, <8 x i32> %y, <8 x i1> %m, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_vpadd: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e32, m1, ta, mu +; CHECK-NEXT: vadd.vv v9, v9, v10 +; CHECK-NEXT: vmerge.vvm v8, v8, v9, v0 +; CHECK-NEXT: ret + %splat = insertelement <8 x i1> poison, i1 -1, i32 0 + %mask = shufflevector <8 x i1> %splat, <8 x i1> poison, <8 x i32> zeroinitializer + %a = call <8 x i32> @llvm.vp.add.nxv2i32(<8 x i32> %x, <8 x i32> %y, <8 x i1> %mask, i32 %vl) + %b = call <8 x i32> @llvm.vp.select.nxv2i32(<8 x i1> %m, <8 x i32> %a, <8 x i32> %passthru, i32 %vl) + ret <8 x i32> %b +} + +; Test glued node of select should not be deleted. +define <8 x i32> @vpselect_vpadd2(<8 x i32> %passthru, <8 x i32> %x, <8 x i32> %y, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_vpadd2: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e32, m1, ta, mu +; CHECK-NEXT: vmseq.vv v0, v9, v10 +; CHECK-NEXT: vadd.vv v9, v9, v10 +; CHECK-NEXT: vmerge.vvm v8, v8, v9, v0 +; CHECK-NEXT: ret + %splat = insertelement <8 x i1> poison, i1 -1, i32 0 + %mask = shufflevector <8 x i1> %splat, <8 x i1> poison, <8 x i32> zeroinitializer + %a = call <8 x i32> @llvm.vp.add.nxv2i32(<8 x i32> %x, <8 x i32> %y, <8 x i1> %mask, i32 %vl) + %m = call <8 x i1> @llvm.vp.icmp.nxv2i32(<8 x i32> %x, <8 x i32> %y, metadata !"eq", <8 x i1> %mask, i32 %vl) + %b = call <8 x i32> @llvm.vp.select.nxv2i32(<8 x i1> %m, <8 x i32> %a, <8 x i32> %passthru, i32 %vl) + ret <8 x i32> %b +} + +; Test vp.select have all-ones mask. +define <8 x i32> @vpselect_vpadd3(<8 x i32> %passthru, <8 x i32> %x, <8 x i32> %y, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_vpadd3: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetivli zero, 8, e8, mf4, ta, mu +; CHECK-NEXT: vmset.m v0 +; CHECK-NEXT: vsetvli zero, a0, e32, m1, ta, mu +; CHECK-NEXT: vadd.vv v9, v9, v10 +; CHECK-NEXT: vmerge.vvm v8, v8, v9, v0 +; CHECK-NEXT: ret + %splat = insertelement <8 x i1> poison, i1 -1, i32 0 + %mask = shufflevector <8 x i1> %splat, <8 x i1> poison, <8 x i32> zeroinitializer + %a = call <8 x i32> @llvm.vp.add.nxv2i32(<8 x i32> %x, <8 x i32> %y, <8 x i1> %mask, i32 %vl) + %b = call <8 x i32> @llvm.vp.select.nxv2i32(<8 x i1> %mask, <8 x i32> %a, <8 x i32> %passthru, i32 %vl) + ret <8 x i32> %b +} + +; Test float binary operator with vp.select and vp.fadd. +define <8 x float> @vpselect_vpfadd(<8 x float> %passthru, <8 x float> %x, <8 x float> %y, <8 x i1> %m, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_vpfadd: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e32, m1, ta, mu +; CHECK-NEXT: vfadd.vv v9, v9, v10 +; CHECK-NEXT: vmerge.vvm v8, v8, v9, v0 +; CHECK-NEXT: ret + %splat = insertelement <8 x i1> poison, i1 -1, i32 0 + %mask = shufflevector <8 x i1> %splat, <8 x i1> poison, <8 x i32> zeroinitializer + %a = call <8 x float> @llvm.vp.fadd.nxv2f32(<8 x float> %x, <8 x float> %y, <8 x i1> %mask, i32 %vl) + %b = call <8 x float> @llvm.vp.select.nxv2f32(<8 x i1> %m, <8 x float> %a, <8 x float> %passthru, i32 %vl) + ret <8 x float> %b +} + +; Test conversion by fptosi. +define <8 x i16> @vpselect_vpfptosi(<8 x i16> %passthru, <8 x float> %x, <8 x i1> %m, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_vpfptosi: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e16, mf2, ta, mu +; CHECK-NEXT: vfncvt.rtz.x.f.w v10, v9 +; CHECK-NEXT: vmerge.vvm v8, v8, v10, v0 +; CHECK-NEXT: ret + %splat = insertelement <8 x i1> poison, i1 -1, i32 0 + %mask = shufflevector <8 x i1> %splat, <8 x i1> poison, <8 x i32> zeroinitializer + %a = call <8 x i16> @llvm.vp.fptosi.nxv2i16.nxv2f32(<8 x float> %x, <8 x i1> %mask, i32 %vl) + %b = call <8 x i16> @llvm.vp.select.nxv2i16(<8 x i1> %m, <8 x i16> %a, <8 x i16> %passthru, i32 %vl) + ret <8 x i16> %b +} + +; Test conversion by sitofp. +define <8 x float> @vpselect_vpsitofp(<8 x float> %passthru, <8 x i64> %x, <8 x i1> %m, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_vpsitofp: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e32, m1, ta, mu +; CHECK-NEXT: vfncvt.f.x.w v9, v10 +; CHECK-NEXT: vmerge.vvm v8, v8, v9, v0 +; CHECK-NEXT: ret + %splat = insertelement <8 x i1> poison, i1 -1, i32 0 + %mask = shufflevector <8 x i1> %splat, <8 x i1> poison, <8 x i32> zeroinitializer + %a = call <8 x float> @llvm.vp.sitofp.nxv2f32.nxv2i64(<8 x i64> %x, <8 x i1> %mask, i32 %vl) + %b = call <8 x float> @llvm.vp.select.nxv2f32(<8 x i1> %m, <8 x float> %a, <8 x float> %passthru, i32 %vl) + ret <8 x float> %b +} + +; Test integer extension by vp.zext. +define <8 x i32> @vpselect_vpzext(<8 x i32> %passthru, <8 x i8> %x, <8 x i1> %m, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_vpzext: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e32, m1, ta, mu +; CHECK-NEXT: vzext.vf4 v10, v9 +; CHECK-NEXT: vmerge.vvm v8, v8, v10, v0 +; CHECK-NEXT: ret + %splat = insertelement <8 x i1> poison, i1 -1, i32 0 + %mask = shufflevector <8 x i1> %splat, <8 x i1> poison, <8 x i32> zeroinitializer + %a = call <8 x i32> @llvm.vp.zext.nxv2i32.nxv2i8(<8 x i8> %x, <8 x i1> %mask, i32 %vl) + %b = call <8 x i32> @llvm.vp.select.nxv2i32(<8 x i1> %m, <8 x i32> %a, <8 x i32> %passthru, i32 %vl) + ret <8 x i32> %b +} + +; Test integer truncation by vp.trunc. +define <8 x i32> @vpselect_vptrunc(<8 x i32> %passthru, <8 x i64> %x, <8 x i1> %m, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_vptrunc: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e32, m1, ta, mu +; CHECK-NEXT: vnsrl.wi v9, v10, 0 +; CHECK-NEXT: vmerge.vvm v8, v8, v9, v0 +; CHECK-NEXT: ret + %splat = insertelement <8 x i1> poison, i1 -1, i32 0 + %mask = shufflevector <8 x i1> %splat, <8 x i1> poison, <8 x i32> zeroinitializer + %a = call <8 x i32> @llvm.vp.trunc.nxv2i32.nxv2i64(<8 x i64> %x, <8 x i1> %mask, i32 %vl) + %b = call <8 x i32> @llvm.vp.select.nxv2i32(<8 x i1> %m, <8 x i32> %a, <8 x i32> %passthru, i32 %vl) + ret <8 x i32> %b +} + +; Test integer extension by vp.fpext. +define <8 x double> @vpselect_vpfpext(<8 x double> %passthru, <8 x float> %x, <8 x i1> %m, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_vpfpext: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e32, m1, ta, mu +; CHECK-NEXT: vfwcvt.f.f.v v12, v10 +; CHECK-NEXT: vsetvli zero, zero, e64, m2, ta, mu +; CHECK-NEXT: vmerge.vvm v8, v8, v12, v0 +; CHECK-NEXT: ret + %splat = insertelement <8 x i1> poison, i1 -1, i32 0 + %mask = shufflevector <8 x i1> %splat, <8 x i1> poison, <8 x i32> zeroinitializer + %a = call <8 x double> @llvm.vp.fpext.nxv2f64.nxv2f32(<8 x float> %x, <8 x i1> %mask, i32 %vl) + %b = call <8 x double> @llvm.vp.select.nxv2f64(<8 x i1> %m, <8 x double> %a, <8 x double> %passthru, i32 %vl) + ret <8 x double> %b +} + +; Test integer truncation by vp.trunc. +define <8 x float> @vpselect_vpfptrunc(<8 x float> %passthru, <8 x double> %x, <8 x i1> %m, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_vpfptrunc: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e32, m1, ta, mu +; CHECK-NEXT: vfncvt.f.f.w v9, v10 +; CHECK-NEXT: vmerge.vvm v8, v8, v9, v0 +; CHECK-NEXT: ret + %splat = insertelement <8 x i1> poison, i1 -1, i32 0 + %mask = shufflevector <8 x i1> %splat, <8 x i1> poison, <8 x i32> zeroinitializer + %a = call <8 x float> @llvm.vp.fptrunc.nxv2f32.nxv2f64(<8 x double> %x, <8 x i1> %mask, i32 %vl) + %b = call <8 x float> @llvm.vp.select.nxv2f32(<8 x i1> %m, <8 x float> %a, <8 x float> %passthru, i32 %vl) + ret <8 x float> %b +} + +; Test load operation by vp.load. +define <8 x i32> @vpselect_vpload(<8 x i32> %passthru, <8 x i32> * %p, <8 x i1> %m, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_vpload: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a1, e32, m1, ta, mu +; CHECK-NEXT: vle32.v v9, (a0) +; CHECK-NEXT: vmerge.vvm v8, v8, v9, v0 +; CHECK-NEXT: ret + %splat = insertelement <8 x i1> poison, i1 -1, i32 0 + %mask = shufflevector <8 x i1> %splat, <8 x i1> poison, <8 x i32> zeroinitializer + %a = call <8 x i32> @llvm.vp.load.nxv2i32.p0nxv2i32(<8 x i32> * %p, <8 x i1> %mask, i32 %vl) + %b = call <8 x i32> @llvm.vp.select.nxv2i32(<8 x i1> %m, <8 x i32> %a, <8 x i32> %passthru, i32 %vl) + ret <8 x i32> %b +} + +; Test result have chain and glued node. +define <8 x i32> @vpselect_vpload2(<8 x i32> %passthru, <8 x i32> * %p, <8 x i32> %x, <8 x i32> %y, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_vpload2: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a1, e32, m1, ta, mu +; CHECK-NEXT: vle32.v v11, (a0) +; CHECK-NEXT: vmseq.vv v0, v9, v10 +; CHECK-NEXT: vmerge.vvm v8, v8, v11, v0 +; CHECK-NEXT: ret + %splat = insertelement <8 x i1> poison, i1 -1, i32 0 + %mask = shufflevector <8 x i1> %splat, <8 x i1> poison, <8 x i32> zeroinitializer + %a = call <8 x i32> @llvm.vp.load.nxv2i32.p0nxv2i32(<8 x i32> * %p, <8 x i1> %mask, i32 %vl) + %m = call <8 x i1> @llvm.vp.icmp.nxv2i32(<8 x i32> %x, <8 x i32> %y, metadata !"eq", <8 x i1> %mask, i32 %vl) + %b = call <8 x i32> @llvm.vp.select.nxv2i32(<8 x i1> %m, <8 x i32> %a, <8 x i32> %passthru, i32 %vl) + ret <8 x i32> %b +} diff --git a/llvm/test/CodeGen/RISCV/rvv/rvv-peephole-vmerge-vops-mir.ll b/llvm/test/CodeGen/RISCV/rvv/rvv-peephole-vmerge-vops-mir.ll index 8898e31..242df34 100644 --- a/llvm/test/CodeGen/RISCV/rvv/rvv-peephole-vmerge-vops-mir.ll +++ b/llvm/test/CodeGen/RISCV/rvv/rvv-peephole-vmerge-vops-mir.ll @@ -2,6 +2,7 @@ ; RUN: llc < %s -mtriple=riscv64 -mattr=+v -stop-after=finalize-isel | FileCheck %s declare @llvm.vp.merge.nxv2i32(, , , i32) +declare @llvm.vp.select.nxv2i32(, , , i32) declare @llvm.vp.load.nxv2i32.p0nxv2i32( *, , i32) ; Test result has chain output of true operand of merge.vvm. @@ -15,7 +16,7 @@ define void @vpmerge_vpload_store( %passthru, poison, i1 -1, i32 0 @@ -25,3 +26,25 @@ define void @vpmerge_vpload_store( %passthru, %b, * %p ret void } + +define void @vpselect_vpload_store( %passthru, * %p, %m, i32 zeroext %vl) { + ; CHECK-LABEL: name: vpselect_vpload_store + ; CHECK: bb.0 (%ir-block.0): + ; CHECK-NEXT: liveins: $v8, $x10, $v0, $x11 + ; CHECK-NEXT: {{ $}} + ; CHECK-NEXT: [[COPY:%[0-9]+]]:gprnox0 = COPY $x11 + ; CHECK-NEXT: [[COPY1:%[0-9]+]]:vr = COPY $v0 + ; CHECK-NEXT: [[COPY2:%[0-9]+]]:gpr = COPY $x10 + ; CHECK-NEXT: [[COPY3:%[0-9]+]]:vr = COPY $v8 + ; CHECK-NEXT: [[PseudoVLE32_V_M1_:%[0-9]+]]:vr = PseudoVLE32_V_M1 [[COPY2]], [[COPY]], 5 /* e32 */ :: (load unknown-size from %ir.p, align 8) + ; CHECK-NEXT: $v0 = COPY [[COPY1]] + ; CHECK-NEXT: [[PseudoVMERGE_VVM_M1_:%[0-9]+]]:vrnov0 = PseudoVMERGE_VVM_M1 [[COPY3]], killed [[PseudoVLE32_V_M1_]], $v0, [[COPY]], 5 /* e32 */ + ; CHECK-NEXT: VS1R_V killed [[PseudoVMERGE_VVM_M1_]], [[COPY2]] :: (store unknown-size into %ir.p, align 8) + ; CHECK-NEXT: PseudoRET + %splat = insertelement poison, i1 -1, i32 0 + %mask = shufflevector %splat, poison, zeroinitializer + %a = call @llvm.vp.load.nxv2i32.p0nxv2i32( * %p, %mask, i32 %vl) + %b = call @llvm.vp.select.nxv2i32( %m, %a, %passthru, i32 %vl) + store %b, * %p + ret void +} diff --git a/llvm/test/CodeGen/RISCV/rvv/rvv-peephole-vmerge-vops.ll b/llvm/test/CodeGen/RISCV/rvv/rvv-peephole-vmerge-vops.ll index 49d240a..ae099eb 100644 --- a/llvm/test/CodeGen/RISCV/rvv/rvv-peephole-vmerge-vops.ll +++ b/llvm/test/CodeGen/RISCV/rvv/rvv-peephole-vmerge-vops.ll @@ -432,3 +432,453 @@ define @vpmerge_trunc( %passthru, @llvm.vp.merge.nxv2i32( %m, %a, %passthru, i32 %vl) ret %b } + +declare @llvm.vp.select.nxv2i16(, , , i32) +declare @llvm.vp.select.nxv2i32(, , , i32) +declare @llvm.vp.select.nxv2f32(, , , i32) +declare @llvm.vp.select.nxv2f64(, , , i32) + +; Test binary operator with vp.select and vp.smax. +define @vpselect_vpadd( %passthru, %x, %y, %m, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_vpadd: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e32, m1, ta, mu +; CHECK-NEXT: vadd.vv v9, v9, v10 +; CHECK-NEXT: vmerge.vvm v8, v8, v9, v0 +; CHECK-NEXT: ret + %splat = insertelement poison, i1 -1, i32 0 + %mask = shufflevector %splat, poison, zeroinitializer + %a = call @llvm.vp.add.nxv2i32( %x, %y, %mask, i32 %vl) + %b = call @llvm.vp.select.nxv2i32( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test glued node of select should not be deleted. +define @vpselect_vpadd2( %passthru, %x, %y, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_vpadd2: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e32, m1, ta, mu +; CHECK-NEXT: vmseq.vv v0, v9, v10 +; CHECK-NEXT: vadd.vv v9, v9, v10 +; CHECK-NEXT: vmerge.vvm v8, v8, v9, v0 +; CHECK-NEXT: ret + %splat = insertelement poison, i1 -1, i32 0 + %mask = shufflevector %splat, poison, zeroinitializer + %a = call @llvm.vp.add.nxv2i32( %x, %y, %mask, i32 %vl) + %m = call @llvm.vp.icmp.nxv2i32( %x, %y, metadata !"eq", %mask, i32 %vl) + %b = call @llvm.vp.select.nxv2i32( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test vp.select has all-ones mask. +define @vpselect_vpadd3( %passthru, %x, %y, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_vpadd3: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a1, zero, e8, mf4, ta, mu +; CHECK-NEXT: vmset.m v0 +; CHECK-NEXT: vsetvli zero, a0, e32, m1, ta, mu +; CHECK-NEXT: vadd.vv v9, v9, v10 +; CHECK-NEXT: vmerge.vvm v8, v8, v9, v0 +; CHECK-NEXT: ret + %splat = insertelement poison, i1 -1, i32 0 + %mask = shufflevector %splat, poison, zeroinitializer + %a = call @llvm.vp.add.nxv2i32( %x, %y, %mask, i32 %vl) + %b = call @llvm.vp.select.nxv2i32( %mask, %a, %passthru, i32 %vl) + ret %b +} + +; Test float binary operator with vp.select and vp.fadd. +define @vpselect_vpfadd( %passthru, %x, %y, %m, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_vpfadd: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e32, m1, ta, mu +; CHECK-NEXT: vfadd.vv v9, v9, v10 +; CHECK-NEXT: vmerge.vvm v8, v8, v9, v0 +; CHECK-NEXT: ret + %splat = insertelement poison, i1 -1, i32 0 + %mask = shufflevector %splat, poison, zeroinitializer + %a = call @llvm.vp.fadd.nxv2f32( %x, %y, %mask, i32 %vl) + %b = call @llvm.vp.select.nxv2f32( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test for binary operator with specific EEW by riscv.vrgatherei16. +define @vpselect_vrgatherei16( %passthru, %x, %y, %m, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_vrgatherei16: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e32, m1, ta, mu +; CHECK-NEXT: vrgatherei16.vv v8, v9, v10 +; CHECK-NEXT: ret + %1 = zext i32 %vl to i64 + %2 = tail call @llvm.riscv.vrgatherei16.vv.nxv2i32.i64( undef, %x, %y, i64 %1) + %3 = tail call @llvm.vp.select.nxv2i32( %m, %2, %passthru, i32 %vl) + ret %2 +} + +; Test conversion by fptosi. +define @vpselect_vpfptosi( %passthru, %x, %m, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_vpfptosi: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e16, mf2, ta, mu +; CHECK-NEXT: vfncvt.rtz.x.f.w v10, v9 +; CHECK-NEXT: vmerge.vvm v8, v8, v10, v0 +; CHECK-NEXT: ret + %splat = insertelement poison, i1 -1, i32 0 + %mask = shufflevector %splat, poison, zeroinitializer + %a = call @llvm.vp.fptosi.nxv2i16.nxv2f32( %x, %mask, i32 %vl) + %b = call @llvm.vp.select.nxv2i16( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test conversion by sitofp. +define @vpselect_vpsitofp( %passthru, %x, %m, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_vpsitofp: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e32, m1, ta, mu +; CHECK-NEXT: vfncvt.f.x.w v9, v10 +; CHECK-NEXT: vmerge.vvm v8, v8, v9, v0 +; CHECK-NEXT: ret + %splat = insertelement poison, i1 -1, i32 0 + %mask = shufflevector %splat, poison, zeroinitializer + %a = call @llvm.vp.sitofp.nxv2f32.nxv2i64( %x, %mask, i32 %vl) + %b = call @llvm.vp.select.nxv2f32( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test integer extension by vp.zext. +define @vpselect_vpzext( %passthru, %x, %m, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_vpzext: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e32, m1, ta, mu +; CHECK-NEXT: vzext.vf4 v10, v9 +; CHECK-NEXT: vmerge.vvm v8, v8, v10, v0 +; CHECK-NEXT: ret + %splat = insertelement poison, i1 -1, i32 0 + %mask = shufflevector %splat, poison, zeroinitializer + %a = call @llvm.vp.zext.nxv2i32.nxv2i8( %x, %mask, i32 %vl) + %b = call @llvm.vp.select.nxv2i32( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test integer truncation by vp.trunc. +define @vpselect_vptrunc( %passthru, %x, %m, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_vptrunc: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e32, m1, ta, mu +; CHECK-NEXT: vnsrl.wi v9, v10, 0 +; CHECK-NEXT: vmerge.vvm v8, v8, v9, v0 +; CHECK-NEXT: ret + %splat = insertelement poison, i1 -1, i32 0 + %mask = shufflevector %splat, poison, zeroinitializer + %a = call @llvm.vp.trunc.nxv2i32.nxv2i64( %x, %mask, i32 %vl) + %b = call @llvm.vp.select.nxv2i32( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test integer extension by vp.fpext. +define @vpselect_vpfpext( %passthru, %x, %m, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_vpfpext: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e32, m1, ta, mu +; CHECK-NEXT: vfwcvt.f.f.v v12, v10 +; CHECK-NEXT: vsetvli zero, zero, e64, m2, ta, mu +; CHECK-NEXT: vmerge.vvm v8, v8, v12, v0 +; CHECK-NEXT: ret + %splat = insertelement poison, i1 -1, i32 0 + %mask = shufflevector %splat, poison, zeroinitializer + %a = call @llvm.vp.fpext.nxv2f64.nxv2f32( %x, %mask, i32 %vl) + %b = call @llvm.vp.select.nxv2f64( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test integer truncation by vp.trunc. +define @vpselect_vpfptrunc( %passthru, %x, %m, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_vpfptrunc: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e32, m1, ta, mu +; CHECK-NEXT: vfncvt.f.f.w v9, v10 +; CHECK-NEXT: vmerge.vvm v8, v8, v9, v0 +; CHECK-NEXT: ret + %splat = insertelement poison, i1 -1, i32 0 + %mask = shufflevector %splat, poison, zeroinitializer + %a = call @llvm.vp.fptrunc.nxv2f32.nxv2f64( %x, %mask, i32 %vl) + %b = call @llvm.vp.select.nxv2f32( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test load operation by vp.load. +define @vpselect_vpload( %passthru, * %p, %m, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_vpload: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a1, e32, m1, ta, mu +; CHECK-NEXT: vle32.v v9, (a0) +; CHECK-NEXT: vmerge.vvm v8, v8, v9, v0 +; CHECK-NEXT: ret + %splat = insertelement poison, i1 -1, i32 0 + %mask = shufflevector %splat, poison, zeroinitializer + %a = call @llvm.vp.load.nxv2i32.p0nxv2i32( * %p, %mask, i32 %vl) + %b = call @llvm.vp.select.nxv2i32( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test result has chain and glued node. +define @vpselect_vpload2( %passthru, * %p, %x, %y, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_vpload2: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a1, e32, m1, ta, mu +; CHECK-NEXT: vle32.v v11, (a0) +; CHECK-NEXT: vmseq.vv v0, v9, v10 +; CHECK-NEXT: vmerge.vvm v8, v8, v11, v0 +; CHECK-NEXT: ret + %splat = insertelement poison, i1 -1, i32 0 + %mask = shufflevector %splat, poison, zeroinitializer + %a = call @llvm.vp.load.nxv2i32.p0nxv2i32( * %p, %mask, i32 %vl) + %m = call @llvm.vp.icmp.nxv2i32( %x, %y, metadata !"eq", %mask, i32 %vl) + %b = call @llvm.vp.select.nxv2i32( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test result has chain output of true operand of select.vvm. +define void @vpselect_vpload_store( %passthru, * %p, %m, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_vpload_store: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a1, e32, m1, ta, mu +; CHECK-NEXT: vle32.v v9, (a0) +; CHECK-NEXT: vmerge.vvm v8, v8, v9, v0 +; CHECK-NEXT: vs1r.v v8, (a0) +; CHECK-NEXT: ret + %splat = insertelement poison, i1 -1, i32 0 + %mask = shufflevector %splat, poison, zeroinitializer + %a = call @llvm.vp.load.nxv2i32.p0nxv2i32( * %p, %mask, i32 %vl) + %b = call @llvm.vp.select.nxv2i32( %m, %a, %passthru, i32 %vl) + store %b, * %p + ret void +} + +; FIXME: select vselect.vvm and vleffN.v +define @vpselect_vleff( %passthru, * %p, %m, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_vleff: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a1, e32, m1, ta, mu +; CHECK-NEXT: vle32ff.v v9, (a0) +; CHECK-NEXT: vsetvli zero, a1, e32, m1, ta, mu +; CHECK-NEXT: vmerge.vvm v8, v8, v9, v0 +; CHECK-NEXT: ret + %1 = zext i32 %vl to i64 + %a = call { , i64 } @llvm.riscv.vleff.nxv2i32( undef, * %p, i64 %1) + %b = extractvalue { , i64 } %a, 0 + %c = call @llvm.vp.select.nxv2i32( %m, %b, %passthru, i32 %vl) + ret %c +} + +; Test strided load by riscv.vlse +define @vpselect_vlse( %passthru, * %p, %m, i64 %s, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_vlse: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a2, e32, m1, ta, mu +; CHECK-NEXT: vlse32.v v9, (a0), a1 +; CHECK-NEXT: vmerge.vvm v8, v8, v9, v0 +; CHECK-NEXT: ret + %1 = zext i32 %vl to i64 + %a = call @llvm.riscv.vlse.nxv2i32( undef, * %p, i64 %s, i64 %1) + %b = call @llvm.vp.select.nxv2i32( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test indexed load by riscv.vluxei +define @vpselect_vluxei( %passthru, * %p, %idx, %m, i64 %s, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_vluxei: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a2, e32, m1, ta, mu +; CHECK-NEXT: vluxei64.v v9, (a0), v10 +; CHECK-NEXT: vmerge.vvm v8, v8, v9, v0 +; CHECK-NEXT: ret + %1 = zext i32 %vl to i64 + %a = call @llvm.riscv.vluxei.nxv2i32.nxv2i64( undef, * %p, %idx, i64 %1) + %b = call @llvm.vp.select.nxv2i32( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test vector index by riscv.vid +define @vpselect_vid( %passthru, %m, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_vid: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e32, m1, ta, mu +; CHECK-NEXT: vid.v v9 +; CHECK-NEXT: vmerge.vvm v8, v8, v9, v0 +; CHECK-NEXT: ret + %1 = zext i32 %vl to i64 + %a = call @llvm.riscv.vid.nxv2i32( undef, i64 %1) + %b = call @llvm.vp.select.nxv2i32( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test riscv.viota +define @vpselect_viota( %passthru, %m, %vm, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_viota: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e32, m1, ta, mu +; CHECK-NEXT: viota.m v10, v9 +; CHECK-NEXT: vmerge.vvm v8, v8, v10, v0 +; CHECK-NEXT: ret + %1 = zext i32 %vl to i64 + %a = call @llvm.riscv.viota.nxv2i32( undef, %vm, i64 %1) + %b = call @llvm.vp.select.nxv2i32( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test riscv.vfclass +define @vpselect_vflcass( %passthru, %vf, %m, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_vflcass: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e32, m1, ta, mu +; CHECK-NEXT: vfclass.v v9, v9 +; CHECK-NEXT: vmerge.vvm v8, v8, v9, v0 +; CHECK-NEXT: ret + %1 = zext i32 %vl to i64 + %a = call @llvm.riscv.vfclass.nxv2i32( undef, %vf, i64 %1) + %b = call @llvm.vp.select.nxv2i32( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test riscv.vfsqrt +define @vpselect_vfsqrt( %passthru, %vf, %m, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_vfsqrt: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e32, m1, ta, mu +; CHECK-NEXT: vfsqrt.v v9, v9 +; CHECK-NEXT: vmerge.vvm v8, v8, v9, v0 +; CHECK-NEXT: ret + %1 = zext i32 %vl to i64 + %a = call @llvm.riscv.vfsqrt.nxv2f32( undef, %vf, i64 %1) + %b = call @llvm.vp.select.nxv2f32( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test reciprocal operation by riscv.vfrec7 +define @vpselect_vfrec7( %passthru, %vf, %m, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_vfrec7: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e32, m1, ta, mu +; CHECK-NEXT: vfrec7.v v9, v9 +; CHECK-NEXT: vmerge.vvm v8, v8, v9, v0 +; CHECK-NEXT: ret + %1 = zext i32 %vl to i64 + %a = call @llvm.riscv.vfrec7.nxv2f32( undef, %vf, i64 %1) + %b = call @llvm.vp.select.nxv2f32( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test vector operations with VLMAX vector length. + +; Test binary operator with vp.select and add. +define @vpselect_add( %passthru, %x, %y, %m, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_add: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a1, zero, e32, m1, ta, mu +; CHECK-NEXT: vadd.vv v9, v9, v10 +; CHECK-NEXT: vsetvli zero, a0, e32, m1, ta, mu +; CHECK-NEXT: vmerge.vvm v8, v8, v9, v0 +; CHECK-NEXT: ret + %a = add %x, %y + %b = call @llvm.vp.select.nxv2i32( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test binary operator with vp.select and fadd. +define @vpselect_fadd( %passthru, %x, %y, %m, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_fadd: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a1, zero, e32, m1, ta, mu +; CHECK-NEXT: vfadd.vv v9, v9, v10 +; CHECK-NEXT: vsetvli zero, a0, e32, m1, ta, mu +; CHECK-NEXT: vmerge.vvm v8, v8, v9, v0 +; CHECK-NEXT: ret + %a = fadd %x, %y + %b = call @llvm.vp.select.nxv2f32( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test conversion by fptosi. +define @vpselect_fptosi( %passthru, %x, %m, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_fptosi: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a1, zero, e16, mf2, ta, mu +; CHECK-NEXT: vfncvt.rtz.x.f.w v10, v9 +; CHECK-NEXT: vsetvli zero, a0, e16, mf2, ta, mu +; CHECK-NEXT: vmerge.vvm v8, v8, v10, v0 +; CHECK-NEXT: ret + %a = fptosi %x to + %b = call @llvm.vp.select.nxv2i16( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test conversion by sitofp. +define @vpselect_sitofp( %passthru, %x, %m, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_sitofp: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a1, zero, e32, m1, ta, mu +; CHECK-NEXT: vfncvt.f.x.w v9, v10 +; CHECK-NEXT: vsetvli zero, a0, e32, m1, ta, mu +; CHECK-NEXT: vmerge.vvm v8, v8, v9, v0 +; CHECK-NEXT: ret + %a = sitofp %x to + %b = call @llvm.vp.select.nxv2f32( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test float extension by fpext. +define @vpselect_fpext( %passthru, %x, %m, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_fpext: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a1, zero, e32, m1, ta, mu +; CHECK-NEXT: vfwcvt.f.f.v v12, v10 +; CHECK-NEXT: vsetvli zero, a0, e64, m2, ta, mu +; CHECK-NEXT: vmerge.vvm v8, v8, v12, v0 +; CHECK-NEXT: ret + %a = fpext %x to + %b = call @llvm.vp.select.nxv2f64( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test float truncation by fptrunc. +define @vpselect_fptrunc( %passthru, %x, %m, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_fptrunc: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a1, zero, e32, m1, ta, mu +; CHECK-NEXT: vfncvt.f.f.w v9, v10 +; CHECK-NEXT: vsetvli zero, a0, e32, m1, ta, mu +; CHECK-NEXT: vmerge.vvm v8, v8, v9, v0 +; CHECK-NEXT: ret + %a = fptrunc %x to + %b = call @llvm.vp.select.nxv2f32( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test integer extension by zext. +define @vpselect_zext( %passthru, %x, %m, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_zext: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a1, zero, e32, m1, ta, mu +; CHECK-NEXT: vzext.vf4 v10, v9 +; CHECK-NEXT: vsetvli zero, a0, e32, m1, ta, mu +; CHECK-NEXT: vmerge.vvm v8, v8, v10, v0 +; CHECK-NEXT: ret + %a = zext %x to + %b = call @llvm.vp.select.nxv2i32( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test integer truncation by trunc. +define @vpselect_trunc( %passthru, %x, %m, i32 zeroext %vl) { +; CHECK-LABEL: vpselect_trunc: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a1, zero, e32, m1, ta, mu +; CHECK-NEXT: vnsrl.wi v9, v10, 0 +; CHECK-NEXT: vsetvli zero, a0, e32, m1, ta, mu +; CHECK-NEXT: vmerge.vvm v8, v8, v9, v0 +; CHECK-NEXT: ret + %a = trunc %x to + %b = call @llvm.vp.select.nxv2i32( %m, %a, %passthru, i32 %vl) + ret %b +}