; SSE: # %bb.0:
; SSE-NEXT: psllw $7, %xmm0
; SSE-NEXT: pmovmskb %xmm0, %eax
-; SSE-NEXT: movl %eax, %ecx
-; SSE-NEXT: shrl $8, %ecx
-; SSE-NEXT: xorb %al, %cl
+; SSE-NEXT: xorb %ah, %al
; SSE-NEXT: setnp %al
; SSE-NEXT: retq
;
; AVX: # %bb.0:
; AVX-NEXT: vpsllw $7, %xmm0, %xmm0
; AVX-NEXT: vpmovmskb %xmm0, %eax
-; AVX-NEXT: movl %eax, %ecx
-; AVX-NEXT: shrl $8, %ecx
-; AVX-NEXT: xorb %al, %cl
+; AVX-NEXT: xorb %ah, %al
; AVX-NEXT: setnp %al
; AVX-NEXT: retq
;
; AVX512: # %bb.0:
; AVX512-NEXT: vpsllw $7, %xmm0, %xmm0
; AVX512-NEXT: vpmovmskb %xmm0, %eax
-; AVX512-NEXT: movl %eax, %ecx
-; AVX512-NEXT: shrl $8, %ecx
-; AVX512-NEXT: xorb %al, %cl
+; AVX512-NEXT: xorb %ah, %al
; AVX512-NEXT: setnp %al
; AVX512-NEXT: retq
%a = trunc <16 x i8> %0 to <16 x i1>
; SSE-NEXT: packuswb %xmm1, %xmm0
; SSE-NEXT: psllw $7, %xmm0
; SSE-NEXT: pmovmskb %xmm0, %eax
-; SSE-NEXT: movl %eax, %ecx
-; SSE-NEXT: shrl $8, %ecx
-; SSE-NEXT: xorb %al, %cl
+; SSE-NEXT: xorb %ah, %al
; SSE-NEXT: setnp %al
; SSE-NEXT: retq
;
; AVX1-NEXT: vpackuswb %xmm1, %xmm0, %xmm0
; AVX1-NEXT: vpsllw $7, %xmm0, %xmm0
; AVX1-NEXT: vpmovmskb %xmm0, %eax
-; AVX1-NEXT: movl %eax, %ecx
-; AVX1-NEXT: shrl $8, %ecx
-; AVX1-NEXT: xorb %al, %cl
+; AVX1-NEXT: xorb %ah, %al
; AVX1-NEXT: setnp %al
; AVX1-NEXT: vzeroupper
; AVX1-NEXT: retq
; AVX2-NEXT: vpackuswb %xmm1, %xmm0, %xmm0
; AVX2-NEXT: vpsllw $7, %xmm0, %xmm0
; AVX2-NEXT: vpmovmskb %xmm0, %eax
-; AVX2-NEXT: movl %eax, %ecx
-; AVX2-NEXT: shrl $8, %ecx
-; AVX2-NEXT: xorb %al, %cl
+; AVX2-NEXT: xorb %ah, %al
; AVX2-NEXT: setnp %al
; AVX2-NEXT: vzeroupper
; AVX2-NEXT: retq
; SSE-NEXT: pxor %xmm1, %xmm0
; SSE-NEXT: psllw $7, %xmm0
; SSE-NEXT: pmovmskb %xmm0, %eax
-; SSE-NEXT: movl %eax, %ecx
-; SSE-NEXT: shrl $8, %ecx
-; SSE-NEXT: xorb %al, %cl
+; SSE-NEXT: xorb %ah, %al
; SSE-NEXT: setnp %al
; SSE-NEXT: retq
;
; AVX1-NEXT: vpxor %xmm1, %xmm0, %xmm0
; AVX1-NEXT: vpsllw $7, %xmm0, %xmm0
; AVX1-NEXT: vpmovmskb %xmm0, %eax
-; AVX1-NEXT: movl %eax, %ecx
-; AVX1-NEXT: shrl $8, %ecx
-; AVX1-NEXT: xorb %al, %cl
+; AVX1-NEXT: xorb %ah, %al
; AVX1-NEXT: setnp %al
; AVX1-NEXT: vzeroupper
; AVX1-NEXT: retq
; AVX2-NEXT: movl %eax, %ecx
; AVX2-NEXT: shrl $16, %ecx
; AVX2-NEXT: xorl %eax, %ecx
-; AVX2-NEXT: movl %ecx, %eax
-; AVX2-NEXT: shrl $8, %eax
-; AVX2-NEXT: xorb %cl, %al
+; AVX2-NEXT: xorb %ch, %cl
; AVX2-NEXT: setnp %al
; AVX2-NEXT: vzeroupper
; AVX2-NEXT: retq
; AVX512BW-NEXT: movl %eax, %ecx
; AVX512BW-NEXT: shrl $16, %ecx
; AVX512BW-NEXT: xorl %eax, %ecx
-; AVX512BW-NEXT: movl %ecx, %eax
-; AVX512BW-NEXT: shrl $8, %eax
-; AVX512BW-NEXT: xorb %cl, %al
+; AVX512BW-NEXT: xorb %ch, %cl
; AVX512BW-NEXT: setnp %al
; AVX512BW-NEXT: vzeroupper
; AVX512BW-NEXT: retq
; AVX512VL-NEXT: movl %eax, %ecx
; AVX512VL-NEXT: shrl $16, %ecx
; AVX512VL-NEXT: xorl %eax, %ecx
-; AVX512VL-NEXT: movl %ecx, %eax
-; AVX512VL-NEXT: shrl $8, %eax
-; AVX512VL-NEXT: xorb %cl, %al
+; AVX512VL-NEXT: xorb %ch, %cl
; AVX512VL-NEXT: setnp %al
; AVX512VL-NEXT: vzeroupper
; AVX512VL-NEXT: retq
; SSE2-NEXT: packuswb %xmm2, %xmm0
; SSE2-NEXT: psllw $7, %xmm0
; SSE2-NEXT: pmovmskb %xmm0, %eax
-; SSE2-NEXT: movl %eax, %ecx
-; SSE2-NEXT: shrl $8, %ecx
-; SSE2-NEXT: xorb %al, %cl
+; SSE2-NEXT: xorb %ah, %al
; SSE2-NEXT: setnp %al
; SSE2-NEXT: retq
;
; SSE41-NEXT: packuswb %xmm2, %xmm0
; SSE41-NEXT: psllw $7, %xmm0
; SSE41-NEXT: pmovmskb %xmm0, %eax
-; SSE41-NEXT: movl %eax, %ecx
-; SSE41-NEXT: shrl $8, %ecx
-; SSE41-NEXT: xorb %al, %cl
+; SSE41-NEXT: xorb %ah, %al
; SSE41-NEXT: setnp %al
; SSE41-NEXT: retq
;
; AVX1-NEXT: vpackuswb %xmm1, %xmm0, %xmm0
; AVX1-NEXT: vpsllw $7, %xmm0, %xmm0
; AVX1-NEXT: vpmovmskb %xmm0, %eax
-; AVX1-NEXT: movl %eax, %ecx
-; AVX1-NEXT: shrl $8, %ecx
-; AVX1-NEXT: xorb %al, %cl
+; AVX1-NEXT: xorb %ah, %al
; AVX1-NEXT: setnp %al
; AVX1-NEXT: vzeroupper
; AVX1-NEXT: retq
; AVX2-NEXT: vpackuswb %xmm1, %xmm0, %xmm0
; AVX2-NEXT: vpsllw $7, %xmm0, %xmm0
; AVX2-NEXT: vpmovmskb %xmm0, %eax
-; AVX2-NEXT: movl %eax, %ecx
-; AVX2-NEXT: shrl $8, %ecx
-; AVX2-NEXT: xorb %al, %cl
+; AVX2-NEXT: xorb %ah, %al
; AVX2-NEXT: setnp %al
; AVX2-NEXT: vzeroupper
; AVX2-NEXT: retq
; SSE-NEXT: pxor %xmm2, %xmm0
; SSE-NEXT: psllw $7, %xmm0
; SSE-NEXT: pmovmskb %xmm0, %eax
-; SSE-NEXT: movl %eax, %ecx
-; SSE-NEXT: shrl $8, %ecx
-; SSE-NEXT: xorb %al, %cl
+; SSE-NEXT: xorb %ah, %al
; SSE-NEXT: setnp %al
; SSE-NEXT: retq
;
; AVX1-NEXT: vpackuswb %xmm1, %xmm0, %xmm0
; AVX1-NEXT: vpsllw $7, %xmm0, %xmm0
; AVX1-NEXT: vpmovmskb %xmm0, %eax
-; AVX1-NEXT: movl %eax, %ecx
-; AVX1-NEXT: shrl $8, %ecx
-; AVX1-NEXT: xorb %al, %cl
+; AVX1-NEXT: xorb %ah, %al
; AVX1-NEXT: setnp %al
; AVX1-NEXT: vzeroupper
; AVX1-NEXT: retq
; AVX2-NEXT: movl %eax, %ecx
; AVX2-NEXT: shrl $16, %ecx
; AVX2-NEXT: xorl %eax, %ecx
-; AVX2-NEXT: movl %ecx, %eax
-; AVX2-NEXT: shrl $8, %eax
-; AVX2-NEXT: xorb %cl, %al
+; AVX2-NEXT: xorb %ch, %cl
; AVX2-NEXT: setnp %al
; AVX2-NEXT: vzeroupper
; AVX2-NEXT: retq
; AVX512BW-NEXT: movl %eax, %ecx
; AVX512BW-NEXT: shrl $16, %ecx
; AVX512BW-NEXT: xorl %eax, %ecx
-; AVX512BW-NEXT: movl %ecx, %eax
-; AVX512BW-NEXT: shrl $8, %eax
-; AVX512BW-NEXT: xorb %cl, %al
+; AVX512BW-NEXT: xorb %ch, %cl
; AVX512BW-NEXT: setnp %al
; AVX512BW-NEXT: vzeroupper
; AVX512BW-NEXT: retq
; AVX512VL-NEXT: movl %eax, %ecx
; AVX512VL-NEXT: shrl $16, %ecx
; AVX512VL-NEXT: xorl %eax, %ecx
-; AVX512VL-NEXT: movl %ecx, %eax
-; AVX512VL-NEXT: shrl $8, %eax
-; AVX512VL-NEXT: xorb %cl, %al
+; AVX512VL-NEXT: xorb %ch, %cl
; AVX512VL-NEXT: setnp %al
; AVX512VL-NEXT: vzeroupper
; AVX512VL-NEXT: retq
; SSE-NEXT: pxor %xmm0, %xmm1
; SSE-NEXT: psllw $7, %xmm1
; SSE-NEXT: pmovmskb %xmm1, %eax
-; SSE-NEXT: movl %eax, %ecx
-; SSE-NEXT: shrl $8, %ecx
-; SSE-NEXT: xorb %al, %cl
+; SSE-NEXT: xorb %ah, %al
; SSE-NEXT: setnp %al
; SSE-NEXT: retq
;
; AVX1-NEXT: vxorps %xmm1, %xmm0, %xmm0
; AVX1-NEXT: vpsllw $7, %xmm0, %xmm0
; AVX1-NEXT: vpmovmskb %xmm0, %eax
-; AVX1-NEXT: movl %eax, %ecx
-; AVX1-NEXT: shrl $8, %ecx
-; AVX1-NEXT: xorb %al, %cl
+; AVX1-NEXT: xorb %ah, %al
; AVX1-NEXT: setnp %al
; AVX1-NEXT: vzeroupper
; AVX1-NEXT: retq
; AVX2-NEXT: movl %eax, %ecx
; AVX2-NEXT: shrl $16, %ecx
; AVX2-NEXT: xorl %eax, %ecx
-; AVX2-NEXT: movl %ecx, %eax
-; AVX2-NEXT: shrl $8, %eax
-; AVX2-NEXT: xorb %cl, %al
+; AVX2-NEXT: xorb %ch, %cl
; AVX2-NEXT: setnp %al
; AVX2-NEXT: vzeroupper
; AVX2-NEXT: retq
; AVX512BW-NEXT: movl %ecx, %eax
; AVX512BW-NEXT: shrl $16, %eax
; AVX512BW-NEXT: xorl %ecx, %eax
-; AVX512BW-NEXT: movl %eax, %ecx
-; AVX512BW-NEXT: shrl $8, %ecx
-; AVX512BW-NEXT: xorb %al, %cl
+; AVX512BW-NEXT: xorb %ah, %al
; AVX512BW-NEXT: setnp %al
; AVX512BW-NEXT: vzeroupper
; AVX512BW-NEXT: retq
; AVX512VL-NEXT: movl %ecx, %eax
; AVX512VL-NEXT: shrl $16, %eax
; AVX512VL-NEXT: xorl %ecx, %eax
-; AVX512VL-NEXT: movl %eax, %ecx
-; AVX512VL-NEXT: shrl $8, %ecx
-; AVX512VL-NEXT: xorb %al, %cl
+; AVX512VL-NEXT: xorb %ah, %al
; AVX512VL-NEXT: setnp %al
; AVX512VL-NEXT: vzeroupper
; AVX512VL-NEXT: retq
; SSE-NEXT: pxor %xmm1, %xmm1
; SSE-NEXT: pcmpeqb %xmm0, %xmm1
; SSE-NEXT: pmovmskb %xmm1, %eax
-; SSE-NEXT: movl %eax, %ecx
-; SSE-NEXT: shrl $8, %ecx
-; SSE-NEXT: xorb %al, %cl
+; SSE-NEXT: xorb %ah, %al
; SSE-NEXT: setnp %al
; SSE-NEXT: retq
;
; AVX-NEXT: vpxor %xmm1, %xmm1, %xmm1
; AVX-NEXT: vpcmpeqb %xmm1, %xmm0, %xmm0
; AVX-NEXT: vpmovmskb %xmm0, %eax
-; AVX-NEXT: movl %eax, %ecx
-; AVX-NEXT: shrl $8, %ecx
-; AVX-NEXT: xorb %al, %cl
+; AVX-NEXT: xorb %ah, %al
; AVX-NEXT: setnp %al
; AVX-NEXT: retq
;
; AVX512F-NEXT: vpxor %xmm1, %xmm1, %xmm1
; AVX512F-NEXT: vpcmpeqb %xmm1, %xmm0, %xmm0
; AVX512F-NEXT: vpmovmskb %xmm0, %eax
-; AVX512F-NEXT: movl %eax, %ecx
-; AVX512F-NEXT: shrl $8, %ecx
-; AVX512F-NEXT: xorb %al, %cl
+; AVX512F-NEXT: xorb %ah, %al
; AVX512F-NEXT: setnp %al
; AVX512F-NEXT: retq
;
; SSE-NEXT: pcmpeqw %xmm2, %xmm0
; SSE-NEXT: packsswb %xmm1, %xmm0
; SSE-NEXT: pmovmskb %xmm0, %eax
-; SSE-NEXT: movl %eax, %ecx
-; SSE-NEXT: shrl $8, %ecx
-; SSE-NEXT: xorb %al, %cl
+; SSE-NEXT: xorb %ah, %al
; SSE-NEXT: setnp %al
; SSE-NEXT: retq
;
; AVX1-NEXT: vpcmpeqw %xmm2, %xmm0, %xmm0
; AVX1-NEXT: vpacksswb %xmm1, %xmm0, %xmm0
; AVX1-NEXT: vpmovmskb %xmm0, %eax
-; AVX1-NEXT: movl %eax, %ecx
-; AVX1-NEXT: shrl $8, %ecx
-; AVX1-NEXT: xorb %al, %cl
+; AVX1-NEXT: xorb %ah, %al
; AVX1-NEXT: setnp %al
; AVX1-NEXT: vzeroupper
; AVX1-NEXT: retq
; AVX2-NEXT: vextracti128 $1, %ymm0, %xmm1
; AVX2-NEXT: vpacksswb %xmm1, %xmm0, %xmm0
; AVX2-NEXT: vpmovmskb %xmm0, %eax
-; AVX2-NEXT: movl %eax, %ecx
-; AVX2-NEXT: shrl $8, %ecx
-; AVX2-NEXT: xorb %al, %cl
+; AVX2-NEXT: xorb %ah, %al
; AVX2-NEXT: setnp %al
; AVX2-NEXT: vzeroupper
; AVX2-NEXT: retq
; SSE-NEXT: pcmpeqb %xmm2, %xmm0
; SSE-NEXT: pxor %xmm1, %xmm0
; SSE-NEXT: pmovmskb %xmm0, %eax
-; SSE-NEXT: movl %eax, %ecx
-; SSE-NEXT: shrl $8, %ecx
-; SSE-NEXT: xorb %al, %cl
+; SSE-NEXT: xorb %ah, %al
; SSE-NEXT: setnp %al
; SSE-NEXT: retq
;
; AVX1-NEXT: vpcmpeqb %xmm2, %xmm0, %xmm0
; AVX1-NEXT: vpxor %xmm1, %xmm0, %xmm0
; AVX1-NEXT: vpmovmskb %xmm0, %eax
-; AVX1-NEXT: movl %eax, %ecx
-; AVX1-NEXT: shrl $8, %ecx
-; AVX1-NEXT: xorb %al, %cl
+; AVX1-NEXT: xorb %ah, %al
; AVX1-NEXT: setnp %al
; AVX1-NEXT: vzeroupper
; AVX1-NEXT: retq
; AVX2-NEXT: movl %eax, %ecx
; AVX2-NEXT: shrl $16, %ecx
; AVX2-NEXT: xorl %eax, %ecx
-; AVX2-NEXT: movl %ecx, %eax
-; AVX2-NEXT: shrl $8, %eax
-; AVX2-NEXT: xorb %cl, %al
+; AVX2-NEXT: xorb %ch, %cl
; AVX2-NEXT: setnp %al
; AVX2-NEXT: vzeroupper
; AVX2-NEXT: retq
; AVX512BW-NEXT: movl %eax, %ecx
; AVX512BW-NEXT: shrl $16, %ecx
; AVX512BW-NEXT: xorl %eax, %ecx
-; AVX512BW-NEXT: movl %ecx, %eax
-; AVX512BW-NEXT: shrl $8, %eax
-; AVX512BW-NEXT: xorb %cl, %al
+; AVX512BW-NEXT: xorb %ch, %cl
; AVX512BW-NEXT: setnp %al
; AVX512BW-NEXT: vzeroupper
; AVX512BW-NEXT: retq
; AVX512VL-NEXT: movl %eax, %ecx
; AVX512VL-NEXT: shrl $16, %ecx
; AVX512VL-NEXT: xorl %eax, %ecx
-; AVX512VL-NEXT: movl %ecx, %eax
-; AVX512VL-NEXT: shrl $8, %eax
-; AVX512VL-NEXT: xorb %cl, %al
+; AVX512VL-NEXT: xorb %ch, %cl
; AVX512VL-NEXT: setnp %al
; AVX512VL-NEXT: vzeroupper
; AVX512VL-NEXT: retq
; SSE-NEXT: packssdw %xmm1, %xmm0
; SSE-NEXT: packsswb %xmm2, %xmm0
; SSE-NEXT: pmovmskb %xmm0, %eax
-; SSE-NEXT: movl %eax, %ecx
-; SSE-NEXT: shrl $8, %ecx
-; SSE-NEXT: xorb %al, %cl
+; SSE-NEXT: xorb %ah, %al
; SSE-NEXT: setnp %al
; SSE-NEXT: retq
;
; AVX1-NEXT: vpackssdw %xmm2, %xmm0, %xmm0
; AVX1-NEXT: vpacksswb %xmm1, %xmm0, %xmm0
; AVX1-NEXT: vpmovmskb %xmm0, %eax
-; AVX1-NEXT: movl %eax, %ecx
-; AVX1-NEXT: shrl $8, %ecx
-; AVX1-NEXT: xorb %al, %cl
+; AVX1-NEXT: xorb %ah, %al
; AVX1-NEXT: setnp %al
; AVX1-NEXT: vzeroupper
; AVX1-NEXT: retq
; AVX2-NEXT: vpacksswb %xmm1, %xmm0, %xmm0
; AVX2-NEXT: vpshufd {{.*#+}} xmm0 = xmm0[0,2,1,3]
; AVX2-NEXT: vpmovmskb %xmm0, %eax
-; AVX2-NEXT: movl %eax, %ecx
-; AVX2-NEXT: shrl $8, %ecx
-; AVX2-NEXT: xorb %al, %cl
+; AVX2-NEXT: xorb %ah, %al
; AVX2-NEXT: setnp %al
; AVX2-NEXT: vzeroupper
; AVX2-NEXT: retq
; SSE-NEXT: packsswb %xmm3, %xmm2
; SSE-NEXT: pxor %xmm0, %xmm2
; SSE-NEXT: pmovmskb %xmm2, %eax
-; SSE-NEXT: movl %eax, %ecx
-; SSE-NEXT: shrl $8, %ecx
-; SSE-NEXT: xorb %al, %cl
+; SSE-NEXT: xorb %ah, %al
; SSE-NEXT: setnp %al
; SSE-NEXT: retq
;
; AVX1-NEXT: vpacksswb %xmm2, %xmm0, %xmm0
; AVX1-NEXT: vpxor %xmm1, %xmm0, %xmm0
; AVX1-NEXT: vpmovmskb %xmm0, %eax
-; AVX1-NEXT: movl %eax, %ecx
-; AVX1-NEXT: shrl $8, %ecx
-; AVX1-NEXT: xorb %al, %cl
+; AVX1-NEXT: xorb %ah, %al
; AVX1-NEXT: setnp %al
; AVX1-NEXT: vzeroupper
; AVX1-NEXT: retq
; AVX2-NEXT: movl %eax, %ecx
; AVX2-NEXT: shrl $16, %ecx
; AVX2-NEXT: xorl %eax, %ecx
-; AVX2-NEXT: movl %ecx, %eax
-; AVX2-NEXT: shrl $8, %eax
-; AVX2-NEXT: xorb %cl, %al
+; AVX2-NEXT: xorb %ch, %cl
; AVX2-NEXT: setnp %al
; AVX2-NEXT: vzeroupper
; AVX2-NEXT: retq
; AVX512BW-NEXT: movl %eax, %ecx
; AVX512BW-NEXT: shrl $16, %ecx
; AVX512BW-NEXT: xorl %eax, %ecx
-; AVX512BW-NEXT: movl %ecx, %eax
-; AVX512BW-NEXT: shrl $8, %eax
-; AVX512BW-NEXT: xorb %cl, %al
+; AVX512BW-NEXT: xorb %ch, %cl
; AVX512BW-NEXT: setnp %al
; AVX512BW-NEXT: vzeroupper
; AVX512BW-NEXT: retq
; AVX512VL-NEXT: movl %eax, %ecx
; AVX512VL-NEXT: shrl $16, %ecx
; AVX512VL-NEXT: xorl %eax, %ecx
-; AVX512VL-NEXT: movl %ecx, %eax
-; AVX512VL-NEXT: shrl $8, %eax
-; AVX512VL-NEXT: xorb %cl, %al
+; AVX512VL-NEXT: xorb %ch, %cl
; AVX512VL-NEXT: setnp %al
; AVX512VL-NEXT: vzeroupper
; AVX512VL-NEXT: retq
; SSE-NEXT: pxor %xmm2, %xmm1
; SSE-NEXT: pxor %xmm0, %xmm1
; SSE-NEXT: pmovmskb %xmm1, %eax
-; SSE-NEXT: movl %eax, %ecx
-; SSE-NEXT: shrl $8, %ecx
-; SSE-NEXT: xorb %al, %cl
+; SSE-NEXT: xorb %ah, %al
; SSE-NEXT: setnp %al
; SSE-NEXT: retq
;
; AVX1-NEXT: vpxor %xmm0, %xmm3, %xmm0
; AVX1-NEXT: vpxor %xmm0, %xmm4, %xmm0
; AVX1-NEXT: vpmovmskb %xmm0, %eax
-; AVX1-NEXT: movl %eax, %ecx
-; AVX1-NEXT: shrl $8, %ecx
-; AVX1-NEXT: xorb %al, %cl
+; AVX1-NEXT: xorb %ah, %al
; AVX1-NEXT: setnp %al
; AVX1-NEXT: vzeroupper
; AVX1-NEXT: retq
; AVX2-NEXT: movl %eax, %ecx
; AVX2-NEXT: shrl $16, %ecx
; AVX2-NEXT: xorl %eax, %ecx
-; AVX2-NEXT: movl %ecx, %eax
-; AVX2-NEXT: shrl $8, %eax
-; AVX2-NEXT: xorb %cl, %al
+; AVX2-NEXT: xorb %ch, %cl
; AVX2-NEXT: setnp %al
; AVX2-NEXT: vzeroupper
; AVX2-NEXT: retq
; AVX512BW-NEXT: movl %ecx, %eax
; AVX512BW-NEXT: shrl $16, %eax
; AVX512BW-NEXT: xorl %ecx, %eax
-; AVX512BW-NEXT: movl %eax, %ecx
-; AVX512BW-NEXT: shrl $8, %ecx
-; AVX512BW-NEXT: xorb %al, %cl
+; AVX512BW-NEXT: xorb %ah, %al
; AVX512BW-NEXT: setnp %al
; AVX512BW-NEXT: vzeroupper
; AVX512BW-NEXT: retq
; AVX512VL-NEXT: movl %ecx, %eax
; AVX512VL-NEXT: shrl $16, %eax
; AVX512VL-NEXT: xorl %ecx, %eax
-; AVX512VL-NEXT: movl %eax, %ecx
-; AVX512VL-NEXT: shrl $8, %ecx
-; AVX512VL-NEXT: xorb %al, %cl
+; AVX512VL-NEXT: xorb %ah, %al
; AVX512VL-NEXT: setnp %al
; AVX512VL-NEXT: vzeroupper
; AVX512VL-NEXT: retq