define dso_local i32* @lea_global_data() #0 {
; SMALL-STATIC-LABEL: lea_global_data:
; SMALL-STATIC: # %bb.0:
-; SMALL-STATIC-NEXT: movl $global_data, %eax
+; SMALL-STATIC-NEXT: movl $.Lglobal_data$local, %eax
; SMALL-STATIC-NEXT: retq
;
; MEDIUM-STATIC-LABEL: lea_global_data:
; MEDIUM-STATIC: # %bb.0:
-; MEDIUM-STATIC-NEXT: movabsq $global_data, %rax
+; MEDIUM-STATIC-NEXT: movabsq $.Lglobal_data$local, %rax
; MEDIUM-STATIC-NEXT: retq
;
; LARGE-STATIC-LABEL: lea_global_data:
; LARGE-STATIC: # %bb.0:
-; LARGE-STATIC-NEXT: movabsq $global_data, %rax
+; LARGE-STATIC-NEXT: movabsq $.Lglobal_data$local, %rax
; LARGE-STATIC-NEXT: retq
;
; SMALL-PIC-LABEL: lea_global_data:
; SMALL-PIC: # %bb.0:
-; SMALL-PIC-NEXT: leaq global_data(%rip), %rax
+; SMALL-PIC-NEXT: leaq .Lglobal_data$local(%rip), %rax
; SMALL-PIC-NEXT: retq
;
; MEDIUM-PIC-LABEL: lea_global_data:
; MEDIUM-PIC: # %bb.0:
; MEDIUM-PIC-NEXT: leaq _GLOBAL_OFFSET_TABLE_(%rip), %rcx
-; MEDIUM-PIC-NEXT: movabsq $global_data@GOTOFF, %rax
+; MEDIUM-PIC-NEXT: movabsq $.Lglobal_data$local@GOTOFF, %rax
; MEDIUM-PIC-NEXT: addq %rcx, %rax
; MEDIUM-PIC-NEXT: retq
;
; LARGE-PIC-NEXT: leaq .L1$pb(%rip), %rax
; LARGE-PIC-NEXT: movabsq $_GLOBAL_OFFSET_TABLE_-.L1$pb, %rcx
; LARGE-PIC-NEXT: addq %rax, %rcx
-; LARGE-PIC-NEXT: movabsq $global_data@GOTOFF, %rax
+; LARGE-PIC-NEXT: movabsq $.Lglobal_data$local@GOTOFF, %rax
; LARGE-PIC-NEXT: addq %rcx, %rax
; LARGE-PIC-NEXT: retq
ret i32* getelementptr inbounds ([10 x i32], [10 x i32]* @global_data, i64 0, i64 0)
define dso_local i32 @load_global_data() #0 {
; SMALL-STATIC-LABEL: load_global_data:
; SMALL-STATIC: # %bb.0:
-; SMALL-STATIC-NEXT: movl global_data+8(%rip), %eax
+; SMALL-STATIC-NEXT: movl .Lglobal_data$local+8(%rip), %eax
; SMALL-STATIC-NEXT: retq
;
; MEDIUM-STATIC-LABEL: load_global_data:
; MEDIUM-STATIC: # %bb.0:
-; MEDIUM-STATIC-NEXT: movabsq $global_data, %rax
+; MEDIUM-STATIC-NEXT: movabsq $.Lglobal_data$local, %rax
; MEDIUM-STATIC-NEXT: movl 8(%rax), %eax
; MEDIUM-STATIC-NEXT: retq
;
; LARGE-STATIC-LABEL: load_global_data:
; LARGE-STATIC: # %bb.0:
-; LARGE-STATIC-NEXT: movabsq $global_data, %rax
+; LARGE-STATIC-NEXT: movabsq $.Lglobal_data$local, %rax
; LARGE-STATIC-NEXT: movl 8(%rax), %eax
; LARGE-STATIC-NEXT: retq
;
; SMALL-PIC-LABEL: load_global_data:
; SMALL-PIC: # %bb.0:
-; SMALL-PIC-NEXT: movl global_data+8(%rip), %eax
+; SMALL-PIC-NEXT: movl .Lglobal_data$local+8(%rip), %eax
; SMALL-PIC-NEXT: retq
;
; MEDIUM-PIC-LABEL: load_global_data:
; MEDIUM-PIC: # %bb.0:
; MEDIUM-PIC-NEXT: leaq _GLOBAL_OFFSET_TABLE_(%rip), %rax
-; MEDIUM-PIC-NEXT: movabsq $global_data@GOTOFF, %rcx
+; MEDIUM-PIC-NEXT: movabsq $.Lglobal_data$local@GOTOFF, %rcx
; MEDIUM-PIC-NEXT: movl 8(%rax,%rcx), %eax
; MEDIUM-PIC-NEXT: retq
;
; LARGE-PIC-NEXT: leaq .L3$pb(%rip), %rax
; LARGE-PIC-NEXT: movabsq $_GLOBAL_OFFSET_TABLE_-.L3$pb, %rcx
; LARGE-PIC-NEXT: addq %rax, %rcx
-; LARGE-PIC-NEXT: movabsq $global_data@GOTOFF, %rax
+; LARGE-PIC-NEXT: movabsq $.Lglobal_data$local@GOTOFF, %rax
; LARGE-PIC-NEXT: movl 8(%rcx,%rax), %eax
; LARGE-PIC-NEXT: retq
%rv = load i32, i32* getelementptr inbounds ([10 x i32], [10 x i32]* @global_data, i64 0, i64 2)
define dso_local void ()* @lea_global_fn() #0 {
; SMALL-STATIC-LABEL: lea_global_fn:
; SMALL-STATIC: # %bb.0:
-; SMALL-STATIC-NEXT: movl $global_fn, %eax
+; SMALL-STATIC-NEXT: movl $.Lglobal_fn$local, %eax
; SMALL-STATIC-NEXT: retq
;
; MEDIUM-STATIC-LABEL: lea_global_fn:
; MEDIUM-STATIC: # %bb.0:
-; MEDIUM-STATIC-NEXT: movabsq $global_fn, %rax
+; MEDIUM-STATIC-NEXT: movabsq $.Lglobal_fn$local, %rax
; MEDIUM-STATIC-NEXT: retq
;
; LARGE-STATIC-LABEL: lea_global_fn:
; LARGE-STATIC: # %bb.0:
-; LARGE-STATIC-NEXT: movabsq $global_fn, %rax
+; LARGE-STATIC-NEXT: movabsq $.Lglobal_fn$local, %rax
; LARGE-STATIC-NEXT: retq
;
; SMALL-PIC-LABEL: lea_global_fn:
; SMALL-PIC: # %bb.0:
-; SMALL-PIC-NEXT: leaq global_fn(%rip), %rax
+; SMALL-PIC-NEXT: leaq .Lglobal_fn$local(%rip), %rax
; SMALL-PIC-NEXT: retq
;
; MEDIUM-PIC-LABEL: lea_global_fn:
; MEDIUM-PIC: # %bb.0:
-; MEDIUM-PIC-NEXT: movabsq $global_fn, %rax
+; MEDIUM-PIC-NEXT: movabsq $.Lglobal_fn$local, %rax
; MEDIUM-PIC-NEXT: retq
;
; LARGE-PIC-LABEL: lea_global_fn:
; LARGE-PIC-NEXT: leaq .L8$pb(%rip), %rax
; LARGE-PIC-NEXT: movabsq $_GLOBAL_OFFSET_TABLE_-.L8$pb, %rcx
; LARGE-PIC-NEXT: addq %rax, %rcx
-; LARGE-PIC-NEXT: movabsq $global_fn@GOTOFF, %rax
+; LARGE-PIC-NEXT: movabsq $.Lglobal_fn$local@GOTOFF, %rax
; LARGE-PIC-NEXT: addq %rcx, %rax
; LARGE-PIC-NEXT: retq
ret void ()* @global_fn
define void @PR42833() {
; SSE2-LABEL: PR42833:
; SSE2: # %bb.0:
-; SSE2-NEXT: movdqa c+{{.*}}(%rip), %xmm1
-; SSE2-NEXT: movdqa c+{{.*}}(%rip), %xmm0
+; SSE2-NEXT: movdqa .Lc$local+{{.*}}(%rip), %xmm1
+; SSE2-NEXT: movdqa .Lc$local+{{.*}}(%rip), %xmm0
; SSE2-NEXT: movd %xmm0, %eax
; SSE2-NEXT: addl {{.*}}(%rip), %eax
; SSE2-NEXT: movd %eax, %xmm2
; SSE2-NEXT: pshufd {{.*#+}} xmm3 = xmm6[0,2,2,3]
; SSE2-NEXT: punpckldq {{.*#+}} xmm5 = xmm5[0],xmm3[0],xmm5[1],xmm3[1]
; SSE2-NEXT: movss {{.*#+}} xmm5 = xmm4[0],xmm5[1,2,3]
-; SSE2-NEXT: movdqa d+{{.*}}(%rip), %xmm3
+; SSE2-NEXT: movdqa .Ld$local+{{.*}}(%rip), %xmm3
; SSE2-NEXT: psubd %xmm1, %xmm3
; SSE2-NEXT: paddd %xmm1, %xmm1
-; SSE2-NEXT: movdqa %xmm1, c+{{.*}}(%rip)
-; SSE2-NEXT: movaps %xmm5, c+{{.*}}(%rip)
-; SSE2-NEXT: movdqa c+{{.*}}(%rip), %xmm1
-; SSE2-NEXT: movdqa c+{{.*}}(%rip), %xmm4
-; SSE2-NEXT: movdqa d+{{.*}}(%rip), %xmm5
-; SSE2-NEXT: movdqa d+{{.*}}(%rip), %xmm6
-; SSE2-NEXT: movdqa d+{{.*}}(%rip), %xmm7
+; SSE2-NEXT: movdqa %xmm1, .Lc$local+{{.*}}(%rip)
+; SSE2-NEXT: movaps %xmm5, .Lc$local+{{.*}}(%rip)
+; SSE2-NEXT: movdqa .Lc$local+{{.*}}(%rip), %xmm1
+; SSE2-NEXT: movdqa .Lc$local+{{.*}}(%rip), %xmm4
+; SSE2-NEXT: movdqa .Ld$local+{{.*}}(%rip), %xmm5
+; SSE2-NEXT: movdqa .Ld$local+{{.*}}(%rip), %xmm6
+; SSE2-NEXT: movdqa .Ld$local+{{.*}}(%rip), %xmm7
; SSE2-NEXT: movss {{.*#+}} xmm0 = xmm2[0],xmm0[1,2,3]
; SSE2-NEXT: psubd %xmm0, %xmm7
; SSE2-NEXT: psubd %xmm4, %xmm6
; SSE2-NEXT: psubd %xmm1, %xmm5
-; SSE2-NEXT: movdqa %xmm5, d+{{.*}}(%rip)
-; SSE2-NEXT: movdqa %xmm6, d+{{.*}}(%rip)
-; SSE2-NEXT: movdqa %xmm3, d+{{.*}}(%rip)
-; SSE2-NEXT: movdqa %xmm7, d+{{.*}}(%rip)
+; SSE2-NEXT: movdqa %xmm5, .Ld$local+{{.*}}(%rip)
+; SSE2-NEXT: movdqa %xmm6, .Ld$local+{{.*}}(%rip)
+; SSE2-NEXT: movdqa %xmm3, .Ld$local+{{.*}}(%rip)
+; SSE2-NEXT: movdqa %xmm7, .Ld$local+{{.*}}(%rip)
; SSE2-NEXT: paddd %xmm4, %xmm4
; SSE2-NEXT: paddd %xmm1, %xmm1
-; SSE2-NEXT: movdqa %xmm1, c+{{.*}}(%rip)
-; SSE2-NEXT: movdqa %xmm4, c+{{.*}}(%rip)
+; SSE2-NEXT: movdqa %xmm1, .Lc$local+{{.*}}(%rip)
+; SSE2-NEXT: movdqa %xmm4, .Lc$local+{{.*}}(%rip)
; SSE2-NEXT: retq
;
; SSE42-LABEL: PR42833:
; SSE42: # %bb.0:
-; SSE42-NEXT: movdqa c+{{.*}}(%rip), %xmm1
-; SSE42-NEXT: movdqa c+{{.*}}(%rip), %xmm0
+; SSE42-NEXT: movdqa .Lc$local+{{.*}}(%rip), %xmm1
+; SSE42-NEXT: movdqa .Lc$local+{{.*}}(%rip), %xmm0
; SSE42-NEXT: movd %xmm0, %eax
; SSE42-NEXT: addl {{.*}}(%rip), %eax
; SSE42-NEXT: movdqa {{.*#+}} xmm2 = <u,1,1,1>
; SSE42-NEXT: cvttps2dq %xmm2, %xmm2
; SSE42-NEXT: pmulld %xmm0, %xmm2
; SSE42-NEXT: pblendw {{.*#+}} xmm2 = xmm3[0,1],xmm2[2,3,4,5,6,7]
-; SSE42-NEXT: movdqa d+{{.*}}(%rip), %xmm3
+; SSE42-NEXT: movdqa .Ld$local+{{.*}}(%rip), %xmm3
; SSE42-NEXT: psubd %xmm1, %xmm3
; SSE42-NEXT: paddd %xmm1, %xmm1
-; SSE42-NEXT: movdqa %xmm1, c+{{.*}}(%rip)
-; SSE42-NEXT: movdqa %xmm2, c+{{.*}}(%rip)
-; SSE42-NEXT: movdqa c+{{.*}}(%rip), %xmm1
-; SSE42-NEXT: movdqa c+{{.*}}(%rip), %xmm2
-; SSE42-NEXT: movdqa d+{{.*}}(%rip), %xmm4
-; SSE42-NEXT: movdqa d+{{.*}}(%rip), %xmm5
-; SSE42-NEXT: movdqa d+{{.*}}(%rip), %xmm6
+; SSE42-NEXT: movdqa %xmm1, .Lc$local+{{.*}}(%rip)
+; SSE42-NEXT: movdqa %xmm2, .Lc$local+{{.*}}(%rip)
+; SSE42-NEXT: movdqa .Lc$local+{{.*}}(%rip), %xmm1
+; SSE42-NEXT: movdqa .Lc$local+{{.*}}(%rip), %xmm2
+; SSE42-NEXT: movdqa .Ld$local+{{.*}}(%rip), %xmm4
+; SSE42-NEXT: movdqa .Ld$local+{{.*}}(%rip), %xmm5
+; SSE42-NEXT: movdqa .Ld$local+{{.*}}(%rip), %xmm6
; SSE42-NEXT: pinsrd $0, %eax, %xmm0
; SSE42-NEXT: psubd %xmm0, %xmm6
; SSE42-NEXT: psubd %xmm2, %xmm5
; SSE42-NEXT: psubd %xmm1, %xmm4
-; SSE42-NEXT: movdqa %xmm4, d+{{.*}}(%rip)
-; SSE42-NEXT: movdqa %xmm5, d+{{.*}}(%rip)
-; SSE42-NEXT: movdqa %xmm3, d+{{.*}}(%rip)
-; SSE42-NEXT: movdqa %xmm6, d+{{.*}}(%rip)
+; SSE42-NEXT: movdqa %xmm4, .Ld$local+{{.*}}(%rip)
+; SSE42-NEXT: movdqa %xmm5, .Ld$local+{{.*}}(%rip)
+; SSE42-NEXT: movdqa %xmm3, .Ld$local+{{.*}}(%rip)
+; SSE42-NEXT: movdqa %xmm6, .Ld$local+{{.*}}(%rip)
; SSE42-NEXT: paddd %xmm2, %xmm2
; SSE42-NEXT: paddd %xmm1, %xmm1
-; SSE42-NEXT: movdqa %xmm1, c+{{.*}}(%rip)
-; SSE42-NEXT: movdqa %xmm2, c+{{.*}}(%rip)
+; SSE42-NEXT: movdqa %xmm1, .Lc$local+{{.*}}(%rip)
+; SSE42-NEXT: movdqa %xmm2, .Lc$local+{{.*}}(%rip)
; SSE42-NEXT: retq
;
; AVX1-LABEL: PR42833:
; AVX1: # %bb.0:
-; AVX1-NEXT: vmovdqa c+{{.*}}(%rip), %xmm0
+; AVX1-NEXT: vmovdqa .Lc$local+{{.*}}(%rip), %xmm0
; AVX1-NEXT: vmovd %xmm0, %eax
; AVX1-NEXT: addl {{.*}}(%rip), %eax
; AVX1-NEXT: vmovdqa {{.*#+}} xmm1 = <u,1,1,1>
; AVX1-NEXT: vpinsrd $0, %eax, %xmm1, %xmm1
; AVX1-NEXT: vpaddd %xmm1, %xmm0, %xmm2
-; AVX1-NEXT: vmovdqa c+{{.*}}(%rip), %xmm3
+; AVX1-NEXT: vmovdqa .Lc$local+{{.*}}(%rip), %xmm3
; AVX1-NEXT: vpslld $23, %xmm1, %xmm1
; AVX1-NEXT: vpaddd {{.*}}(%rip), %xmm1, %xmm1
; AVX1-NEXT: vcvttps2dq %xmm1, %xmm1
; AVX1-NEXT: vpslld $1, %xmm3, %xmm3
; AVX1-NEXT: vinsertf128 $1, %xmm3, %ymm1, %ymm1
; AVX1-NEXT: vblendps {{.*#+}} ymm1 = ymm2[0],ymm1[1,2,3,4,5,6,7]
-; AVX1-NEXT: vmovdqa d+{{.*}}(%rip), %xmm2
-; AVX1-NEXT: vpsubd c+{{.*}}(%rip), %xmm2, %xmm2
-; AVX1-NEXT: vmovups %ymm1, c+{{.*}}(%rip)
+; AVX1-NEXT: vmovdqa .Ld$local+{{.*}}(%rip), %xmm2
+; AVX1-NEXT: vpsubd .Lc$local+{{.*}}(%rip), %xmm2, %xmm2
+; AVX1-NEXT: vmovups %ymm1, .Lc$local+{{.*}}(%rip)
; AVX1-NEXT: vpinsrd $0, %eax, %xmm0, %xmm0
-; AVX1-NEXT: vmovdqa d+{{.*}}(%rip), %xmm1
+; AVX1-NEXT: vmovdqa .Ld$local+{{.*}}(%rip), %xmm1
; AVX1-NEXT: vpsubd %xmm0, %xmm1, %xmm0
-; AVX1-NEXT: vmovdqa d+{{.*}}(%rip), %xmm1
-; AVX1-NEXT: vmovdqa c+{{.*}}(%rip), %xmm3
+; AVX1-NEXT: vmovdqa .Ld$local+{{.*}}(%rip), %xmm1
+; AVX1-NEXT: vmovdqa .Lc$local+{{.*}}(%rip), %xmm3
; AVX1-NEXT: vpsubd %xmm3, %xmm1, %xmm1
-; AVX1-NEXT: vmovdqa d+{{.*}}(%rip), %xmm4
-; AVX1-NEXT: vmovdqa c+{{.*}}(%rip), %xmm5
+; AVX1-NEXT: vmovdqa .Ld$local+{{.*}}(%rip), %xmm4
+; AVX1-NEXT: vmovdqa .Lc$local+{{.*}}(%rip), %xmm5
; AVX1-NEXT: vpsubd %xmm5, %xmm4, %xmm4
-; AVX1-NEXT: vmovdqa %xmm2, d+{{.*}}(%rip)
-; AVX1-NEXT: vmovdqa %xmm4, d+{{.*}}(%rip)
-; AVX1-NEXT: vmovdqa %xmm1, d+{{.*}}(%rip)
-; AVX1-NEXT: vmovdqa %xmm0, d+{{.*}}(%rip)
+; AVX1-NEXT: vmovdqa %xmm2, .Ld$local+{{.*}}(%rip)
+; AVX1-NEXT: vmovdqa %xmm4, .Ld$local+{{.*}}(%rip)
+; AVX1-NEXT: vmovdqa %xmm1, .Ld$local+{{.*}}(%rip)
+; AVX1-NEXT: vmovdqa %xmm0, .Ld$local+{{.*}}(%rip)
; AVX1-NEXT: vpaddd %xmm3, %xmm3, %xmm0
; AVX1-NEXT: vpaddd %xmm5, %xmm5, %xmm1
-; AVX1-NEXT: vmovdqa %xmm1, c+{{.*}}(%rip)
-; AVX1-NEXT: vmovdqa %xmm0, c+{{.*}}(%rip)
+; AVX1-NEXT: vmovdqa %xmm1, .Lc$local+{{.*}}(%rip)
+; AVX1-NEXT: vmovdqa %xmm0, .Lc$local+{{.*}}(%rip)
; AVX1-NEXT: vzeroupper
; AVX1-NEXT: retq
;
; AVX2-LABEL: PR42833:
; AVX2: # %bb.0:
; AVX2-NEXT: movl {{.*}}(%rip), %eax
-; AVX2-NEXT: vmovdqu c+{{.*}}(%rip), %ymm0
-; AVX2-NEXT: addl c+{{.*}}(%rip), %eax
+; AVX2-NEXT: vmovdqu .Lc$local+{{.*}}(%rip), %ymm0
+; AVX2-NEXT: addl .Lc$local+{{.*}}(%rip), %eax
; AVX2-NEXT: vmovd %eax, %xmm1
; AVX2-NEXT: vpblendd {{.*#+}} ymm2 = ymm1[0],mem[1,2,3,4,5,6,7]
; AVX2-NEXT: vpaddd %ymm2, %ymm0, %ymm3
; AVX2-NEXT: vpsllvd %ymm2, %ymm0, %ymm2
; AVX2-NEXT: vpblendd {{.*#+}} ymm2 = ymm3[0],ymm2[1,2,3,4,5,6,7]
-; AVX2-NEXT: vmovdqu %ymm2, c+{{.*}}(%rip)
-; AVX2-NEXT: vmovdqu c+{{.*}}(%rip), %ymm2
-; AVX2-NEXT: vmovdqu d+{{.*}}(%rip), %ymm3
-; AVX2-NEXT: vmovdqu d+{{.*}}(%rip), %ymm4
+; AVX2-NEXT: vmovdqu %ymm2, .Lc$local+{{.*}}(%rip)
+; AVX2-NEXT: vmovdqu .Lc$local+{{.*}}(%rip), %ymm2
+; AVX2-NEXT: vmovdqu .Ld$local+{{.*}}(%rip), %ymm3
+; AVX2-NEXT: vmovdqu .Ld$local+{{.*}}(%rip), %ymm4
; AVX2-NEXT: vpblendd {{.*#+}} ymm0 = ymm1[0],ymm0[1,2,3,4,5,6,7]
; AVX2-NEXT: vpsubd %ymm0, %ymm4, %ymm0
; AVX2-NEXT: vpsubd %ymm2, %ymm3, %ymm1
-; AVX2-NEXT: vmovdqu %ymm1, d+{{.*}}(%rip)
-; AVX2-NEXT: vmovdqu %ymm0, d+{{.*}}(%rip)
+; AVX2-NEXT: vmovdqu %ymm1, .Ld$local+{{.*}}(%rip)
+; AVX2-NEXT: vmovdqu %ymm0, .Ld$local+{{.*}}(%rip)
; AVX2-NEXT: vpaddd %ymm2, %ymm2, %ymm0
-; AVX2-NEXT: vmovdqu %ymm0, c+{{.*}}(%rip)
+; AVX2-NEXT: vmovdqu %ymm0, .Lc$local+{{.*}}(%rip)
; AVX2-NEXT: vzeroupper
; AVX2-NEXT: retq
;
; AVX512-LABEL: PR42833:
; AVX512: # %bb.0:
; AVX512-NEXT: movl {{.*}}(%rip), %eax
-; AVX512-NEXT: vmovdqu c+{{.*}}(%rip), %ymm0
-; AVX512-NEXT: vmovdqu64 c+{{.*}}(%rip), %zmm1
-; AVX512-NEXT: addl c+{{.*}}(%rip), %eax
+; AVX512-NEXT: vmovdqu .Lc$local+{{.*}}(%rip), %ymm0
+; AVX512-NEXT: vmovdqu64 .Lc$local+{{.*}}(%rip), %zmm1
+; AVX512-NEXT: addl .Lc$local+{{.*}}(%rip), %eax
; AVX512-NEXT: vmovd %eax, %xmm2
; AVX512-NEXT: vpblendd {{.*#+}} ymm2 = ymm2[0],mem[1,2,3,4,5,6,7]
; AVX512-NEXT: vpaddd %ymm2, %ymm0, %ymm3
; AVX512-NEXT: vpsllvd %ymm2, %ymm0, %ymm0
; AVX512-NEXT: vpblendd {{.*#+}} ymm0 = ymm3[0],ymm0[1,2,3,4,5,6,7]
-; AVX512-NEXT: vmovdqa c+{{.*}}(%rip), %xmm2
-; AVX512-NEXT: vmovdqu %ymm0, c+{{.*}}(%rip)
-; AVX512-NEXT: vmovdqu c+{{.*}}(%rip), %ymm0
-; AVX512-NEXT: vmovdqu64 d+{{.*}}(%rip), %zmm3
+; AVX512-NEXT: vmovdqa .Lc$local+{{.*}}(%rip), %xmm2
+; AVX512-NEXT: vmovdqu %ymm0, .Lc$local+{{.*}}(%rip)
+; AVX512-NEXT: vmovdqu .Lc$local+{{.*}}(%rip), %ymm0
+; AVX512-NEXT: vmovdqu64 .Ld$local+{{.*}}(%rip), %zmm3
; AVX512-NEXT: vpinsrd $0, %eax, %xmm2, %xmm2
; AVX512-NEXT: vinserti32x4 $0, %xmm2, %zmm1, %zmm1
; AVX512-NEXT: vinserti64x4 $1, %ymm0, %zmm1, %zmm1
; AVX512-NEXT: vpsubd %zmm1, %zmm3, %zmm1
-; AVX512-NEXT: vmovdqu64 %zmm1, d+{{.*}}(%rip)
+; AVX512-NEXT: vmovdqu64 %zmm1, .Ld$local+{{.*}}(%rip)
; AVX512-NEXT: vpaddd %ymm0, %ymm0, %ymm0
-; AVX512-NEXT: vmovdqu %ymm0, c+{{.*}}(%rip)
+; AVX512-NEXT: vmovdqu %ymm0, .Lc$local+{{.*}}(%rip)
; AVX512-NEXT: vzeroupper
; AVX512-NEXT: retq
;
; XOP-LABEL: PR42833:
; XOP: # %bb.0:
-; XOP-NEXT: vmovdqa c+{{.*}}(%rip), %xmm0
+; XOP-NEXT: vmovdqa .Lc$local+{{.*}}(%rip), %xmm0
; XOP-NEXT: vmovd %xmm0, %eax
; XOP-NEXT: addl {{.*}}(%rip), %eax
; XOP-NEXT: vmovdqa {{.*#+}} xmm1 = <u,1,1,1>
; XOP-NEXT: vpinsrd $0, %eax, %xmm1, %xmm1
; XOP-NEXT: vpaddd %xmm1, %xmm0, %xmm2
-; XOP-NEXT: vmovdqa c+{{.*}}(%rip), %xmm3
+; XOP-NEXT: vmovdqa .Lc$local+{{.*}}(%rip), %xmm3
; XOP-NEXT: vpshld %xmm1, %xmm0, %xmm1
; XOP-NEXT: vpslld $1, %xmm3, %xmm3
; XOP-NEXT: vinsertf128 $1, %xmm3, %ymm1, %ymm1
; XOP-NEXT: vblendps {{.*#+}} ymm1 = ymm2[0],ymm1[1,2,3,4,5,6,7]
-; XOP-NEXT: vmovdqa d+{{.*}}(%rip), %xmm2
-; XOP-NEXT: vpsubd c+{{.*}}(%rip), %xmm2, %xmm2
-; XOP-NEXT: vmovups %ymm1, c+{{.*}}(%rip)
+; XOP-NEXT: vmovdqa .Ld$local+{{.*}}(%rip), %xmm2
+; XOP-NEXT: vpsubd .Lc$local+{{.*}}(%rip), %xmm2, %xmm2
+; XOP-NEXT: vmovups %ymm1, .Lc$local+{{.*}}(%rip)
; XOP-NEXT: vpinsrd $0, %eax, %xmm0, %xmm0
-; XOP-NEXT: vmovdqa d+{{.*}}(%rip), %xmm1
+; XOP-NEXT: vmovdqa .Ld$local+{{.*}}(%rip), %xmm1
; XOP-NEXT: vpsubd %xmm0, %xmm1, %xmm0
-; XOP-NEXT: vmovdqa d+{{.*}}(%rip), %xmm1
-; XOP-NEXT: vmovdqa c+{{.*}}(%rip), %xmm3
+; XOP-NEXT: vmovdqa .Ld$local+{{.*}}(%rip), %xmm1
+; XOP-NEXT: vmovdqa .Lc$local+{{.*}}(%rip), %xmm3
; XOP-NEXT: vpsubd %xmm3, %xmm1, %xmm1
-; XOP-NEXT: vmovdqa d+{{.*}}(%rip), %xmm4
-; XOP-NEXT: vmovdqa c+{{.*}}(%rip), %xmm5
+; XOP-NEXT: vmovdqa .Ld$local+{{.*}}(%rip), %xmm4
+; XOP-NEXT: vmovdqa .Lc$local+{{.*}}(%rip), %xmm5
; XOP-NEXT: vpsubd %xmm5, %xmm4, %xmm4
-; XOP-NEXT: vmovdqa %xmm2, d+{{.*}}(%rip)
-; XOP-NEXT: vmovdqa %xmm4, d+{{.*}}(%rip)
-; XOP-NEXT: vmovdqa %xmm1, d+{{.*}}(%rip)
-; XOP-NEXT: vmovdqa %xmm0, d+{{.*}}(%rip)
+; XOP-NEXT: vmovdqa %xmm2, .Ld$local+{{.*}}(%rip)
+; XOP-NEXT: vmovdqa %xmm4, .Ld$local+{{.*}}(%rip)
+; XOP-NEXT: vmovdqa %xmm1, .Ld$local+{{.*}}(%rip)
+; XOP-NEXT: vmovdqa %xmm0, .Ld$local+{{.*}}(%rip)
; XOP-NEXT: vpaddd %xmm3, %xmm3, %xmm0
; XOP-NEXT: vpaddd %xmm5, %xmm5, %xmm1
-; XOP-NEXT: vmovdqa %xmm1, c+{{.*}}(%rip)
-; XOP-NEXT: vmovdqa %xmm0, c+{{.*}}(%rip)
+; XOP-NEXT: vmovdqa %xmm1, .Lc$local+{{.*}}(%rip)
+; XOP-NEXT: vmovdqa %xmm0, .Lc$local+{{.*}}(%rip)
; XOP-NEXT: vzeroupper
; XOP-NEXT: retq
%1 = load i32, i32* @b, align 4