; CHECK: # %bb.0:
; CHECK-NEXT: vsetvli zero, zero, e8,mf8,ta,mu
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: slli a0, a0, 24
-; CHECK-NEXT: srai a0, a0, 24
; CHECK-NEXT: ret
%r = extractelement <vscale x 1 x i8> %v, i32 0
ret i8 %r
; CHECK-NEXT: vsetvli a0, zero, e8,mf8,ta,mu
; CHECK-NEXT: vslidedown.vi v25, v8, 2
; CHECK-NEXT: vmv.x.s a0, v25
-; CHECK-NEXT: slli a0, a0, 24
-; CHECK-NEXT: srai a0, a0, 24
; CHECK-NEXT: ret
%r = extractelement <vscale x 1 x i8> %v, i32 2
ret i8 %r
; CHECK-NEXT: vsetvli a1, zero, e8,mf8,ta,mu
; CHECK-NEXT: vslidedown.vx v25, v8, a0
; CHECK-NEXT: vmv.x.s a0, v25
-; CHECK-NEXT: slli a0, a0, 24
-; CHECK-NEXT: srai a0, a0, 24
; CHECK-NEXT: ret
%r = extractelement <vscale x 1 x i8> %v, i32 %idx
ret i8 %r
; CHECK: # %bb.0:
; CHECK-NEXT: vsetvli zero, zero, e8,mf4,ta,mu
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: slli a0, a0, 24
-; CHECK-NEXT: srai a0, a0, 24
; CHECK-NEXT: ret
%r = extractelement <vscale x 2 x i8> %v, i32 0
ret i8 %r
; CHECK-NEXT: vsetvli a0, zero, e8,mf4,ta,mu
; CHECK-NEXT: vslidedown.vi v25, v8, 2
; CHECK-NEXT: vmv.x.s a0, v25
-; CHECK-NEXT: slli a0, a0, 24
-; CHECK-NEXT: srai a0, a0, 24
; CHECK-NEXT: ret
%r = extractelement <vscale x 2 x i8> %v, i32 2
ret i8 %r
; CHECK-NEXT: vsetvli a1, zero, e8,mf4,ta,mu
; CHECK-NEXT: vslidedown.vx v25, v8, a0
; CHECK-NEXT: vmv.x.s a0, v25
-; CHECK-NEXT: slli a0, a0, 24
-; CHECK-NEXT: srai a0, a0, 24
; CHECK-NEXT: ret
%r = extractelement <vscale x 2 x i8> %v, i32 %idx
ret i8 %r
; CHECK: # %bb.0:
; CHECK-NEXT: vsetvli zero, zero, e8,mf2,ta,mu
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: slli a0, a0, 24
-; CHECK-NEXT: srai a0, a0, 24
; CHECK-NEXT: ret
%r = extractelement <vscale x 4 x i8> %v, i32 0
ret i8 %r
; CHECK-NEXT: vsetvli a0, zero, e8,mf2,ta,mu
; CHECK-NEXT: vslidedown.vi v25, v8, 2
; CHECK-NEXT: vmv.x.s a0, v25
-; CHECK-NEXT: slli a0, a0, 24
-; CHECK-NEXT: srai a0, a0, 24
; CHECK-NEXT: ret
%r = extractelement <vscale x 4 x i8> %v, i32 2
ret i8 %r
; CHECK-NEXT: vsetvli a1, zero, e8,mf2,ta,mu
; CHECK-NEXT: vslidedown.vx v25, v8, a0
; CHECK-NEXT: vmv.x.s a0, v25
-; CHECK-NEXT: slli a0, a0, 24
-; CHECK-NEXT: srai a0, a0, 24
; CHECK-NEXT: ret
%r = extractelement <vscale x 4 x i8> %v, i32 %idx
ret i8 %r
; CHECK: # %bb.0:
; CHECK-NEXT: vsetvli zero, zero, e8,m1,ta,mu
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: slli a0, a0, 24
-; CHECK-NEXT: srai a0, a0, 24
; CHECK-NEXT: ret
%r = extractelement <vscale x 8 x i8> %v, i32 0
ret i8 %r
; CHECK-NEXT: vsetvli a0, zero, e8,m1,ta,mu
; CHECK-NEXT: vslidedown.vi v25, v8, 2
; CHECK-NEXT: vmv.x.s a0, v25
-; CHECK-NEXT: slli a0, a0, 24
-; CHECK-NEXT: srai a0, a0, 24
; CHECK-NEXT: ret
%r = extractelement <vscale x 8 x i8> %v, i32 2
ret i8 %r
; CHECK-NEXT: vsetvli a1, zero, e8,m1,ta,mu
; CHECK-NEXT: vslidedown.vx v25, v8, a0
; CHECK-NEXT: vmv.x.s a0, v25
-; CHECK-NEXT: slli a0, a0, 24
-; CHECK-NEXT: srai a0, a0, 24
; CHECK-NEXT: ret
%r = extractelement <vscale x 8 x i8> %v, i32 %idx
ret i8 %r
; CHECK: # %bb.0:
; CHECK-NEXT: vsetvli zero, zero, e8,m2,ta,mu
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: slli a0, a0, 24
-; CHECK-NEXT: srai a0, a0, 24
; CHECK-NEXT: ret
%r = extractelement <vscale x 16 x i8> %v, i32 0
ret i8 %r
; CHECK-NEXT: vsetvli a0, zero, e8,m2,ta,mu
; CHECK-NEXT: vslidedown.vi v26, v8, 2
; CHECK-NEXT: vmv.x.s a0, v26
-; CHECK-NEXT: slli a0, a0, 24
-; CHECK-NEXT: srai a0, a0, 24
; CHECK-NEXT: ret
%r = extractelement <vscale x 16 x i8> %v, i32 2
ret i8 %r
; CHECK-NEXT: vsetvli a1, zero, e8,m2,ta,mu
; CHECK-NEXT: vslidedown.vx v26, v8, a0
; CHECK-NEXT: vmv.x.s a0, v26
-; CHECK-NEXT: slli a0, a0, 24
-; CHECK-NEXT: srai a0, a0, 24
; CHECK-NEXT: ret
%r = extractelement <vscale x 16 x i8> %v, i32 %idx
ret i8 %r
; CHECK: # %bb.0:
; CHECK-NEXT: vsetvli zero, zero, e8,m4,ta,mu
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: slli a0, a0, 24
-; CHECK-NEXT: srai a0, a0, 24
; CHECK-NEXT: ret
%r = extractelement <vscale x 32 x i8> %v, i32 0
ret i8 %r
; CHECK-NEXT: vsetvli a0, zero, e8,m4,ta,mu
; CHECK-NEXT: vslidedown.vi v28, v8, 2
; CHECK-NEXT: vmv.x.s a0, v28
-; CHECK-NEXT: slli a0, a0, 24
-; CHECK-NEXT: srai a0, a0, 24
; CHECK-NEXT: ret
%r = extractelement <vscale x 32 x i8> %v, i32 2
ret i8 %r
; CHECK-NEXT: vsetvli a1, zero, e8,m4,ta,mu
; CHECK-NEXT: vslidedown.vx v28, v8, a0
; CHECK-NEXT: vmv.x.s a0, v28
-; CHECK-NEXT: slli a0, a0, 24
-; CHECK-NEXT: srai a0, a0, 24
; CHECK-NEXT: ret
%r = extractelement <vscale x 32 x i8> %v, i32 %idx
ret i8 %r
; CHECK: # %bb.0:
; CHECK-NEXT: vsetvli zero, zero, e8,m8,ta,mu
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: slli a0, a0, 24
-; CHECK-NEXT: srai a0, a0, 24
; CHECK-NEXT: ret
%r = extractelement <vscale x 64 x i8> %v, i32 0
ret i8 %r
; CHECK-NEXT: vsetvli a0, zero, e8,m8,ta,mu
; CHECK-NEXT: vslidedown.vi v8, v8, 2
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: slli a0, a0, 24
-; CHECK-NEXT: srai a0, a0, 24
; CHECK-NEXT: ret
%r = extractelement <vscale x 64 x i8> %v, i32 2
ret i8 %r
; CHECK-NEXT: vsetvli a1, zero, e8,m8,ta,mu
; CHECK-NEXT: vslidedown.vx v8, v8, a0
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: slli a0, a0, 24
-; CHECK-NEXT: srai a0, a0, 24
; CHECK-NEXT: ret
%r = extractelement <vscale x 64 x i8> %v, i32 %idx
ret i8 %r
; CHECK: # %bb.0:
; CHECK-NEXT: vsetvli zero, zero, e16,mf4,ta,mu
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: slli a0, a0, 16
-; CHECK-NEXT: srai a0, a0, 16
; CHECK-NEXT: ret
%r = extractelement <vscale x 1 x i16> %v, i32 0
ret i16 %r
; CHECK-NEXT: vsetvli a0, zero, e16,mf4,ta,mu
; CHECK-NEXT: vslidedown.vi v25, v8, 2
; CHECK-NEXT: vmv.x.s a0, v25
-; CHECK-NEXT: slli a0, a0, 16
-; CHECK-NEXT: srai a0, a0, 16
; CHECK-NEXT: ret
%r = extractelement <vscale x 1 x i16> %v, i32 2
ret i16 %r
; CHECK-NEXT: vsetvli a1, zero, e16,mf4,ta,mu
; CHECK-NEXT: vslidedown.vx v25, v8, a0
; CHECK-NEXT: vmv.x.s a0, v25
-; CHECK-NEXT: slli a0, a0, 16
-; CHECK-NEXT: srai a0, a0, 16
; CHECK-NEXT: ret
%r = extractelement <vscale x 1 x i16> %v, i32 %idx
ret i16 %r
; CHECK: # %bb.0:
; CHECK-NEXT: vsetvli zero, zero, e16,mf2,ta,mu
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: slli a0, a0, 16
-; CHECK-NEXT: srai a0, a0, 16
; CHECK-NEXT: ret
%r = extractelement <vscale x 2 x i16> %v, i32 0
ret i16 %r
; CHECK-NEXT: vsetvli a0, zero, e16,mf2,ta,mu
; CHECK-NEXT: vslidedown.vi v25, v8, 2
; CHECK-NEXT: vmv.x.s a0, v25
-; CHECK-NEXT: slli a0, a0, 16
-; CHECK-NEXT: srai a0, a0, 16
; CHECK-NEXT: ret
%r = extractelement <vscale x 2 x i16> %v, i32 2
ret i16 %r
; CHECK-NEXT: vsetvli a1, zero, e16,mf2,ta,mu
; CHECK-NEXT: vslidedown.vx v25, v8, a0
; CHECK-NEXT: vmv.x.s a0, v25
-; CHECK-NEXT: slli a0, a0, 16
-; CHECK-NEXT: srai a0, a0, 16
; CHECK-NEXT: ret
%r = extractelement <vscale x 2 x i16> %v, i32 %idx
ret i16 %r
; CHECK: # %bb.0:
; CHECK-NEXT: vsetvli zero, zero, e16,m1,ta,mu
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: slli a0, a0, 16
-; CHECK-NEXT: srai a0, a0, 16
; CHECK-NEXT: ret
%r = extractelement <vscale x 4 x i16> %v, i32 0
ret i16 %r
; CHECK-NEXT: vsetvli a0, zero, e16,m1,ta,mu
; CHECK-NEXT: vslidedown.vi v25, v8, 2
; CHECK-NEXT: vmv.x.s a0, v25
-; CHECK-NEXT: slli a0, a0, 16
-; CHECK-NEXT: srai a0, a0, 16
; CHECK-NEXT: ret
%r = extractelement <vscale x 4 x i16> %v, i32 2
ret i16 %r
; CHECK-NEXT: vsetvli a1, zero, e16,m1,ta,mu
; CHECK-NEXT: vslidedown.vx v25, v8, a0
; CHECK-NEXT: vmv.x.s a0, v25
-; CHECK-NEXT: slli a0, a0, 16
-; CHECK-NEXT: srai a0, a0, 16
; CHECK-NEXT: ret
%r = extractelement <vscale x 4 x i16> %v, i32 %idx
ret i16 %r
; CHECK: # %bb.0:
; CHECK-NEXT: vsetvli zero, zero, e16,m2,ta,mu
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: slli a0, a0, 16
-; CHECK-NEXT: srai a0, a0, 16
; CHECK-NEXT: ret
%r = extractelement <vscale x 8 x i16> %v, i32 0
ret i16 %r
; CHECK-NEXT: vsetvli a0, zero, e16,m2,ta,mu
; CHECK-NEXT: vslidedown.vi v26, v8, 2
; CHECK-NEXT: vmv.x.s a0, v26
-; CHECK-NEXT: slli a0, a0, 16
-; CHECK-NEXT: srai a0, a0, 16
; CHECK-NEXT: ret
%r = extractelement <vscale x 8 x i16> %v, i32 2
ret i16 %r
; CHECK-NEXT: vsetvli a1, zero, e16,m2,ta,mu
; CHECK-NEXT: vslidedown.vx v26, v8, a0
; CHECK-NEXT: vmv.x.s a0, v26
-; CHECK-NEXT: slli a0, a0, 16
-; CHECK-NEXT: srai a0, a0, 16
; CHECK-NEXT: ret
%r = extractelement <vscale x 8 x i16> %v, i32 %idx
ret i16 %r
; CHECK: # %bb.0:
; CHECK-NEXT: vsetvli zero, zero, e16,m4,ta,mu
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: slli a0, a0, 16
-; CHECK-NEXT: srai a0, a0, 16
; CHECK-NEXT: ret
%r = extractelement <vscale x 16 x i16> %v, i32 0
ret i16 %r
; CHECK-NEXT: vsetvli a0, zero, e16,m4,ta,mu
; CHECK-NEXT: vslidedown.vi v28, v8, 2
; CHECK-NEXT: vmv.x.s a0, v28
-; CHECK-NEXT: slli a0, a0, 16
-; CHECK-NEXT: srai a0, a0, 16
; CHECK-NEXT: ret
%r = extractelement <vscale x 16 x i16> %v, i32 2
ret i16 %r
; CHECK-NEXT: vsetvli a1, zero, e16,m4,ta,mu
; CHECK-NEXT: vslidedown.vx v28, v8, a0
; CHECK-NEXT: vmv.x.s a0, v28
-; CHECK-NEXT: slli a0, a0, 16
-; CHECK-NEXT: srai a0, a0, 16
; CHECK-NEXT: ret
%r = extractelement <vscale x 16 x i16> %v, i32 %idx
ret i16 %r
; CHECK: # %bb.0:
; CHECK-NEXT: vsetvli zero, zero, e16,m8,ta,mu
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: slli a0, a0, 16
-; CHECK-NEXT: srai a0, a0, 16
; CHECK-NEXT: ret
%r = extractelement <vscale x 32 x i16> %v, i32 0
ret i16 %r
; CHECK-NEXT: vsetvli a0, zero, e16,m8,ta,mu
; CHECK-NEXT: vslidedown.vi v8, v8, 2
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: slli a0, a0, 16
-; CHECK-NEXT: srai a0, a0, 16
; CHECK-NEXT: ret
%r = extractelement <vscale x 32 x i16> %v, i32 2
ret i16 %r
; CHECK-NEXT: vsetvli a1, zero, e16,m8,ta,mu
; CHECK-NEXT: vslidedown.vx v8, v8, a0
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: slli a0, a0, 16
-; CHECK-NEXT: srai a0, a0, 16
; CHECK-NEXT: ret
%r = extractelement <vscale x 32 x i16> %v, i32 %idx
ret i16 %r
; CHECK: # %bb.0:
; CHECK-NEXT: vsetvli zero, zero, e8,mf8,ta,mu
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: slli a0, a0, 56
-; CHECK-NEXT: srai a0, a0, 56
; CHECK-NEXT: ret
%r = extractelement <vscale x 1 x i8> %v, i32 0
ret i8 %r
; CHECK-NEXT: vsetvli a0, zero, e8,mf8,ta,mu
; CHECK-NEXT: vslidedown.vi v25, v8, 2
; CHECK-NEXT: vmv.x.s a0, v25
-; CHECK-NEXT: slli a0, a0, 56
-; CHECK-NEXT: srai a0, a0, 56
; CHECK-NEXT: ret
%r = extractelement <vscale x 1 x i8> %v, i32 2
ret i8 %r
; CHECK-NEXT: vsetvli a1, zero, e8,mf8,ta,mu
; CHECK-NEXT: vslidedown.vx v25, v8, a0
; CHECK-NEXT: vmv.x.s a0, v25
-; CHECK-NEXT: slli a0, a0, 56
-; CHECK-NEXT: srai a0, a0, 56
; CHECK-NEXT: ret
%r = extractelement <vscale x 1 x i8> %v, i32 %idx
ret i8 %r
; CHECK: # %bb.0:
; CHECK-NEXT: vsetvli zero, zero, e8,mf4,ta,mu
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: slli a0, a0, 56
-; CHECK-NEXT: srai a0, a0, 56
; CHECK-NEXT: ret
%r = extractelement <vscale x 2 x i8> %v, i32 0
ret i8 %r
; CHECK-NEXT: vsetvli a0, zero, e8,mf4,ta,mu
; CHECK-NEXT: vslidedown.vi v25, v8, 2
; CHECK-NEXT: vmv.x.s a0, v25
-; CHECK-NEXT: slli a0, a0, 56
-; CHECK-NEXT: srai a0, a0, 56
; CHECK-NEXT: ret
%r = extractelement <vscale x 2 x i8> %v, i32 2
ret i8 %r
; CHECK-NEXT: vsetvli a1, zero, e8,mf4,ta,mu
; CHECK-NEXT: vslidedown.vx v25, v8, a0
; CHECK-NEXT: vmv.x.s a0, v25
-; CHECK-NEXT: slli a0, a0, 56
-; CHECK-NEXT: srai a0, a0, 56
; CHECK-NEXT: ret
%r = extractelement <vscale x 2 x i8> %v, i32 %idx
ret i8 %r
; CHECK: # %bb.0:
; CHECK-NEXT: vsetvli zero, zero, e8,mf2,ta,mu
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: slli a0, a0, 56
-; CHECK-NEXT: srai a0, a0, 56
; CHECK-NEXT: ret
%r = extractelement <vscale x 4 x i8> %v, i32 0
ret i8 %r
; CHECK-NEXT: vsetvli a0, zero, e8,mf2,ta,mu
; CHECK-NEXT: vslidedown.vi v25, v8, 2
; CHECK-NEXT: vmv.x.s a0, v25
-; CHECK-NEXT: slli a0, a0, 56
-; CHECK-NEXT: srai a0, a0, 56
; CHECK-NEXT: ret
%r = extractelement <vscale x 4 x i8> %v, i32 2
ret i8 %r
; CHECK-NEXT: vsetvli a1, zero, e8,mf2,ta,mu
; CHECK-NEXT: vslidedown.vx v25, v8, a0
; CHECK-NEXT: vmv.x.s a0, v25
-; CHECK-NEXT: slli a0, a0, 56
-; CHECK-NEXT: srai a0, a0, 56
; CHECK-NEXT: ret
%r = extractelement <vscale x 4 x i8> %v, i32 %idx
ret i8 %r
; CHECK: # %bb.0:
; CHECK-NEXT: vsetvli zero, zero, e8,m1,ta,mu
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: slli a0, a0, 56
-; CHECK-NEXT: srai a0, a0, 56
; CHECK-NEXT: ret
%r = extractelement <vscale x 8 x i8> %v, i32 0
ret i8 %r
; CHECK-NEXT: vsetvli a0, zero, e8,m1,ta,mu
; CHECK-NEXT: vslidedown.vi v25, v8, 2
; CHECK-NEXT: vmv.x.s a0, v25
-; CHECK-NEXT: slli a0, a0, 56
-; CHECK-NEXT: srai a0, a0, 56
; CHECK-NEXT: ret
%r = extractelement <vscale x 8 x i8> %v, i32 2
ret i8 %r
; CHECK-NEXT: vsetvli a1, zero, e8,m1,ta,mu
; CHECK-NEXT: vslidedown.vx v25, v8, a0
; CHECK-NEXT: vmv.x.s a0, v25
-; CHECK-NEXT: slli a0, a0, 56
-; CHECK-NEXT: srai a0, a0, 56
; CHECK-NEXT: ret
%r = extractelement <vscale x 8 x i8> %v, i32 %idx
ret i8 %r
; CHECK: # %bb.0:
; CHECK-NEXT: vsetvli zero, zero, e8,m2,ta,mu
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: slli a0, a0, 56
-; CHECK-NEXT: srai a0, a0, 56
; CHECK-NEXT: ret
%r = extractelement <vscale x 16 x i8> %v, i32 0
ret i8 %r
; CHECK-NEXT: vsetvli a0, zero, e8,m2,ta,mu
; CHECK-NEXT: vslidedown.vi v26, v8, 2
; CHECK-NEXT: vmv.x.s a0, v26
-; CHECK-NEXT: slli a0, a0, 56
-; CHECK-NEXT: srai a0, a0, 56
; CHECK-NEXT: ret
%r = extractelement <vscale x 16 x i8> %v, i32 2
ret i8 %r
; CHECK-NEXT: vsetvli a1, zero, e8,m2,ta,mu
; CHECK-NEXT: vslidedown.vx v26, v8, a0
; CHECK-NEXT: vmv.x.s a0, v26
-; CHECK-NEXT: slli a0, a0, 56
-; CHECK-NEXT: srai a0, a0, 56
; CHECK-NEXT: ret
%r = extractelement <vscale x 16 x i8> %v, i32 %idx
ret i8 %r
; CHECK: # %bb.0:
; CHECK-NEXT: vsetvli zero, zero, e8,m4,ta,mu
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: slli a0, a0, 56
-; CHECK-NEXT: srai a0, a0, 56
; CHECK-NEXT: ret
%r = extractelement <vscale x 32 x i8> %v, i32 0
ret i8 %r
; CHECK-NEXT: vsetvli a0, zero, e8,m4,ta,mu
; CHECK-NEXT: vslidedown.vi v28, v8, 2
; CHECK-NEXT: vmv.x.s a0, v28
-; CHECK-NEXT: slli a0, a0, 56
-; CHECK-NEXT: srai a0, a0, 56
; CHECK-NEXT: ret
%r = extractelement <vscale x 32 x i8> %v, i32 2
ret i8 %r
; CHECK-NEXT: vsetvli a1, zero, e8,m4,ta,mu
; CHECK-NEXT: vslidedown.vx v28, v8, a0
; CHECK-NEXT: vmv.x.s a0, v28
-; CHECK-NEXT: slli a0, a0, 56
-; CHECK-NEXT: srai a0, a0, 56
; CHECK-NEXT: ret
%r = extractelement <vscale x 32 x i8> %v, i32 %idx
ret i8 %r
; CHECK: # %bb.0:
; CHECK-NEXT: vsetvli zero, zero, e8,m8,ta,mu
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: slli a0, a0, 56
-; CHECK-NEXT: srai a0, a0, 56
; CHECK-NEXT: ret
%r = extractelement <vscale x 64 x i8> %v, i32 0
ret i8 %r
; CHECK-NEXT: vsetvli a0, zero, e8,m8,ta,mu
; CHECK-NEXT: vslidedown.vi v8, v8, 2
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: slli a0, a0, 56
-; CHECK-NEXT: srai a0, a0, 56
; CHECK-NEXT: ret
%r = extractelement <vscale x 64 x i8> %v, i32 2
ret i8 %r
; CHECK-NEXT: vsetvli a1, zero, e8,m8,ta,mu
; CHECK-NEXT: vslidedown.vx v8, v8, a0
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: slli a0, a0, 56
-; CHECK-NEXT: srai a0, a0, 56
; CHECK-NEXT: ret
%r = extractelement <vscale x 64 x i8> %v, i32 %idx
ret i8 %r
; CHECK: # %bb.0:
; CHECK-NEXT: vsetvli zero, zero, e16,mf4,ta,mu
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: slli a0, a0, 48
-; CHECK-NEXT: srai a0, a0, 48
; CHECK-NEXT: ret
%r = extractelement <vscale x 1 x i16> %v, i32 0
ret i16 %r
; CHECK-NEXT: vsetvli a0, zero, e16,mf4,ta,mu
; CHECK-NEXT: vslidedown.vi v25, v8, 2
; CHECK-NEXT: vmv.x.s a0, v25
-; CHECK-NEXT: slli a0, a0, 48
-; CHECK-NEXT: srai a0, a0, 48
; CHECK-NEXT: ret
%r = extractelement <vscale x 1 x i16> %v, i32 2
ret i16 %r
; CHECK-NEXT: vsetvli a1, zero, e16,mf4,ta,mu
; CHECK-NEXT: vslidedown.vx v25, v8, a0
; CHECK-NEXT: vmv.x.s a0, v25
-; CHECK-NEXT: slli a0, a0, 48
-; CHECK-NEXT: srai a0, a0, 48
; CHECK-NEXT: ret
%r = extractelement <vscale x 1 x i16> %v, i32 %idx
ret i16 %r
; CHECK: # %bb.0:
; CHECK-NEXT: vsetvli zero, zero, e16,mf2,ta,mu
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: slli a0, a0, 48
-; CHECK-NEXT: srai a0, a0, 48
; CHECK-NEXT: ret
%r = extractelement <vscale x 2 x i16> %v, i32 0
ret i16 %r
; CHECK-NEXT: vsetvli a0, zero, e16,mf2,ta,mu
; CHECK-NEXT: vslidedown.vi v25, v8, 2
; CHECK-NEXT: vmv.x.s a0, v25
-; CHECK-NEXT: slli a0, a0, 48
-; CHECK-NEXT: srai a0, a0, 48
; CHECK-NEXT: ret
%r = extractelement <vscale x 2 x i16> %v, i32 2
ret i16 %r
; CHECK-NEXT: vsetvli a1, zero, e16,mf2,ta,mu
; CHECK-NEXT: vslidedown.vx v25, v8, a0
; CHECK-NEXT: vmv.x.s a0, v25
-; CHECK-NEXT: slli a0, a0, 48
-; CHECK-NEXT: srai a0, a0, 48
; CHECK-NEXT: ret
%r = extractelement <vscale x 2 x i16> %v, i32 %idx
ret i16 %r
; CHECK: # %bb.0:
; CHECK-NEXT: vsetvli zero, zero, e16,m1,ta,mu
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: slli a0, a0, 48
-; CHECK-NEXT: srai a0, a0, 48
; CHECK-NEXT: ret
%r = extractelement <vscale x 4 x i16> %v, i32 0
ret i16 %r
; CHECK-NEXT: vsetvli a0, zero, e16,m1,ta,mu
; CHECK-NEXT: vslidedown.vi v25, v8, 2
; CHECK-NEXT: vmv.x.s a0, v25
-; CHECK-NEXT: slli a0, a0, 48
-; CHECK-NEXT: srai a0, a0, 48
; CHECK-NEXT: ret
%r = extractelement <vscale x 4 x i16> %v, i32 2
ret i16 %r
; CHECK-NEXT: vsetvli a1, zero, e16,m1,ta,mu
; CHECK-NEXT: vslidedown.vx v25, v8, a0
; CHECK-NEXT: vmv.x.s a0, v25
-; CHECK-NEXT: slli a0, a0, 48
-; CHECK-NEXT: srai a0, a0, 48
; CHECK-NEXT: ret
%r = extractelement <vscale x 4 x i16> %v, i32 %idx
ret i16 %r
; CHECK: # %bb.0:
; CHECK-NEXT: vsetvli zero, zero, e16,m2,ta,mu
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: slli a0, a0, 48
-; CHECK-NEXT: srai a0, a0, 48
; CHECK-NEXT: ret
%r = extractelement <vscale x 8 x i16> %v, i32 0
ret i16 %r
; CHECK-NEXT: vsetvli a0, zero, e16,m2,ta,mu
; CHECK-NEXT: vslidedown.vi v26, v8, 2
; CHECK-NEXT: vmv.x.s a0, v26
-; CHECK-NEXT: slli a0, a0, 48
-; CHECK-NEXT: srai a0, a0, 48
; CHECK-NEXT: ret
%r = extractelement <vscale x 8 x i16> %v, i32 2
ret i16 %r
; CHECK-NEXT: vsetvli a1, zero, e16,m2,ta,mu
; CHECK-NEXT: vslidedown.vx v26, v8, a0
; CHECK-NEXT: vmv.x.s a0, v26
-; CHECK-NEXT: slli a0, a0, 48
-; CHECK-NEXT: srai a0, a0, 48
; CHECK-NEXT: ret
%r = extractelement <vscale x 8 x i16> %v, i32 %idx
ret i16 %r
; CHECK: # %bb.0:
; CHECK-NEXT: vsetvli zero, zero, e16,m4,ta,mu
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: slli a0, a0, 48
-; CHECK-NEXT: srai a0, a0, 48
; CHECK-NEXT: ret
%r = extractelement <vscale x 16 x i16> %v, i32 0
ret i16 %r
; CHECK-NEXT: vsetvli a0, zero, e16,m4,ta,mu
; CHECK-NEXT: vslidedown.vi v28, v8, 2
; CHECK-NEXT: vmv.x.s a0, v28
-; CHECK-NEXT: slli a0, a0, 48
-; CHECK-NEXT: srai a0, a0, 48
; CHECK-NEXT: ret
%r = extractelement <vscale x 16 x i16> %v, i32 2
ret i16 %r
; CHECK-NEXT: vsetvli a1, zero, e16,m4,ta,mu
; CHECK-NEXT: vslidedown.vx v28, v8, a0
; CHECK-NEXT: vmv.x.s a0, v28
-; CHECK-NEXT: slli a0, a0, 48
-; CHECK-NEXT: srai a0, a0, 48
; CHECK-NEXT: ret
%r = extractelement <vscale x 16 x i16> %v, i32 %idx
ret i16 %r
; CHECK: # %bb.0:
; CHECK-NEXT: vsetvli zero, zero, e16,m8,ta,mu
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: slli a0, a0, 48
-; CHECK-NEXT: srai a0, a0, 48
; CHECK-NEXT: ret
%r = extractelement <vscale x 32 x i16> %v, i32 0
ret i16 %r
; CHECK-NEXT: vsetvli a0, zero, e16,m8,ta,mu
; CHECK-NEXT: vslidedown.vi v8, v8, 2
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: slli a0, a0, 48
-; CHECK-NEXT: srai a0, a0, 48
; CHECK-NEXT: ret
%r = extractelement <vscale x 32 x i16> %v, i32 2
ret i16 %r
; CHECK-NEXT: vsetvli a1, zero, e16,m8,ta,mu
; CHECK-NEXT: vslidedown.vx v8, v8, a0
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: slli a0, a0, 48
-; CHECK-NEXT: srai a0, a0, 48
; CHECK-NEXT: ret
%r = extractelement <vscale x 32 x i16> %v, i32 %idx
ret i16 %r
; CHECK: # %bb.0:
; CHECK-NEXT: vsetvli zero, zero, e32,mf2,ta,mu
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: sext.w a0, a0
; CHECK-NEXT: ret
%r = extractelement <vscale x 1 x i32> %v, i32 0
ret i32 %r
; CHECK-NEXT: vsetvli a0, zero, e32,mf2,ta,mu
; CHECK-NEXT: vslidedown.vi v25, v8, 2
; CHECK-NEXT: vmv.x.s a0, v25
-; CHECK-NEXT: sext.w a0, a0
; CHECK-NEXT: ret
%r = extractelement <vscale x 1 x i32> %v, i32 2
ret i32 %r
; CHECK-NEXT: vsetvli a1, zero, e32,mf2,ta,mu
; CHECK-NEXT: vslidedown.vx v25, v8, a0
; CHECK-NEXT: vmv.x.s a0, v25
-; CHECK-NEXT: sext.w a0, a0
; CHECK-NEXT: ret
%r = extractelement <vscale x 1 x i32> %v, i32 %idx
ret i32 %r
; CHECK: # %bb.0:
; CHECK-NEXT: vsetvli zero, zero, e32,m1,ta,mu
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: sext.w a0, a0
; CHECK-NEXT: ret
%r = extractelement <vscale x 2 x i32> %v, i32 0
ret i32 %r
; CHECK-NEXT: vsetvli a0, zero, e32,m1,ta,mu
; CHECK-NEXT: vslidedown.vi v25, v8, 2
; CHECK-NEXT: vmv.x.s a0, v25
-; CHECK-NEXT: sext.w a0, a0
; CHECK-NEXT: ret
%r = extractelement <vscale x 2 x i32> %v, i32 2
ret i32 %r
; CHECK-NEXT: vsetvli a1, zero, e32,m1,ta,mu
; CHECK-NEXT: vslidedown.vx v25, v8, a0
; CHECK-NEXT: vmv.x.s a0, v25
-; CHECK-NEXT: sext.w a0, a0
; CHECK-NEXT: ret
%r = extractelement <vscale x 2 x i32> %v, i32 %idx
ret i32 %r
; CHECK: # %bb.0:
; CHECK-NEXT: vsetvli zero, zero, e32,m2,ta,mu
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: sext.w a0, a0
; CHECK-NEXT: ret
%r = extractelement <vscale x 4 x i32> %v, i32 0
ret i32 %r
; CHECK-NEXT: vsetvli a0, zero, e32,m2,ta,mu
; CHECK-NEXT: vslidedown.vi v26, v8, 2
; CHECK-NEXT: vmv.x.s a0, v26
-; CHECK-NEXT: sext.w a0, a0
; CHECK-NEXT: ret
%r = extractelement <vscale x 4 x i32> %v, i32 2
ret i32 %r
; CHECK-NEXT: vsetvli a1, zero, e32,m2,ta,mu
; CHECK-NEXT: vslidedown.vx v26, v8, a0
; CHECK-NEXT: vmv.x.s a0, v26
-; CHECK-NEXT: sext.w a0, a0
; CHECK-NEXT: ret
%r = extractelement <vscale x 4 x i32> %v, i32 %idx
ret i32 %r
; CHECK: # %bb.0:
; CHECK-NEXT: vsetvli zero, zero, e32,m4,ta,mu
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: sext.w a0, a0
; CHECK-NEXT: ret
%r = extractelement <vscale x 8 x i32> %v, i32 0
ret i32 %r
; CHECK-NEXT: vsetvli a0, zero, e32,m4,ta,mu
; CHECK-NEXT: vslidedown.vi v28, v8, 2
; CHECK-NEXT: vmv.x.s a0, v28
-; CHECK-NEXT: sext.w a0, a0
; CHECK-NEXT: ret
%r = extractelement <vscale x 8 x i32> %v, i32 2
ret i32 %r
; CHECK-NEXT: vsetvli a1, zero, e32,m4,ta,mu
; CHECK-NEXT: vslidedown.vx v28, v8, a0
; CHECK-NEXT: vmv.x.s a0, v28
-; CHECK-NEXT: sext.w a0, a0
; CHECK-NEXT: ret
%r = extractelement <vscale x 8 x i32> %v, i32 %idx
ret i32 %r
; CHECK: # %bb.0:
; CHECK-NEXT: vsetvli zero, zero, e32,m8,ta,mu
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: sext.w a0, a0
; CHECK-NEXT: ret
%r = extractelement <vscale x 16 x i32> %v, i32 0
ret i32 %r
; CHECK-NEXT: vsetvli a0, zero, e32,m8,ta,mu
; CHECK-NEXT: vslidedown.vi v8, v8, 2
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: sext.w a0, a0
; CHECK-NEXT: ret
%r = extractelement <vscale x 16 x i32> %v, i32 2
ret i32 %r
; CHECK-NEXT: vsetvli a1, zero, e32,m8,ta,mu
; CHECK-NEXT: vslidedown.vx v8, v8, a0
; CHECK-NEXT: vmv.x.s a0, v8
-; CHECK-NEXT: sext.w a0, a0
; CHECK-NEXT: ret
%r = extractelement <vscale x 16 x i32> %v, i32 %idx
ret i32 %r