%struct.NSRect = type { [4 x float] }
-define void @foo(ptr %context) nounwind {
+define void @foo(i8* %context) nounwind {
; CHECK-LABEL: @foo(
; CHECK-NEXT: entry:
-; CHECK-NEXT: call void (i32, ...) @bar(i32 3, ptr byval([[STRUCT_NSRECT:%.*]]) align 4 [[CONTEXT:%.*]]) #[[ATTR0:[0-9]+]]
+; CHECK-NEXT: [[TMP1:%.*]] = bitcast i8* [[CONTEXT:%.*]] to %struct.NSRect*
+; CHECK-NEXT: call void (i32, ...) @bar(i32 3, %struct.NSRect* byval([[STRUCT_NSRECT:%.*]]) align 4 [[TMP1]]) #[[ATTR0:[0-9]+]]
; CHECK-NEXT: ret void
;
entry:
- call void (i32, ...) @bar( i32 3, ptr byval(%struct.NSRect) align 4 %context ) nounwind
+ %tmp1 = bitcast i8* %context to %struct.NSRect* ; <%struct.NSRect*> [#uses=1]
+ call void (i32, ...) @bar( i32 3, %struct.NSRect* byval(%struct.NSRect) align 4 %tmp1 ) nounwind
ret void
}
declare void @varargs(i32, ...)
-define void @test11(ptr %P) {
+define void @test11(i32* %P) {
; ALL-LABEL: @test11(
-; ALL-NEXT: call void (i32, ...) @varargs(i32 5, ptr [[P:%.*]])
+; ALL-NEXT: call void (i32, ...) @varargs(i32 5, i32* [[P:%.*]])
; ALL-NEXT: ret void
;
- call void (i32, ...) @varargs( i32 5, ptr %P )
+ %c = bitcast i32* %P to i16*
+ call void (i32, ...) @varargs( i32 5, i16* %c )
ret void
}
declare i32 @__gxx_personality_v0(...)
-define void @test_invoke_vararg_cast(ptr %a, ptr %b) personality ptr @__gxx_personality_v0 {
+define void @test_invoke_vararg_cast(i32* %a, i32* %b) personality i8* bitcast (i32 (...)* @__gxx_personality_v0 to i8*) {
; ALL-LABEL: @test_invoke_vararg_cast(
; ALL-NEXT: entry:
-; ALL-NEXT: invoke void (i32, ...) @varargs(i32 1, ptr [[B:%.*]], ptr [[A:%.*]])
+; ALL-NEXT: invoke void (i32, ...) @varargs(i32 1, i32* [[B:%.*]], i32* [[A:%.*]])
; ALL-NEXT: to label [[INVOKE_CONT:%.*]] unwind label [[LPAD:%.*]]
; ALL: invoke.cont:
; ALL-NEXT: ret void
; ALL: lpad:
-; ALL-NEXT: [[TMP0:%.*]] = landingpad { ptr, i32 }
+; ALL-NEXT: [[TMP0:%.*]] = landingpad { i8*, i32 }
; ALL-NEXT: cleanup
; ALL-NEXT: ret void
;
entry:
- invoke void (i32, ...) @varargs(i32 1, ptr %b, ptr %a)
+ %0 = bitcast i32* %b to i8*
+ %1 = bitcast i32* %a to i64*
+ invoke void (i32, ...) @varargs(i32 1, i8* %0, i64* %1)
to label %invoke.cont unwind label %lpad
invoke.cont:
ret void
lpad:
- %0 = landingpad { ptr, i32 }
+ %2 = landingpad { i8*, i32 }
cleanup
ret void
}
-define ptr @test13(i64 %A) {
+define i8* @test13(i64 %A) {
; ALL-LABEL: @test13(
-; ALL-NEXT: [[C:%.*]] = getelementptr [0 x i8], ptr @inbuf, i64 0, i64 [[A:%.*]]
-; ALL-NEXT: ret ptr [[C]]
+; ALL-NEXT: [[C:%.*]] = getelementptr [32832 x i8], [32832 x i8]* @inbuf, i64 0, i64 [[A:%.*]]
+; ALL-NEXT: ret i8* [[C]]
;
- %c = getelementptr [0 x i8], ptr @inbuf, i64 0, i64 %A
- ret ptr %c
+ %c = getelementptr [0 x i8], [0 x i8]* bitcast ([32832 x i8]* @inbuf to [0 x i8]*), i64 0, i64 %A
+ ret i8* %c
}
define i1 @test14(i8 %A) {
; ret bool %X
;}
-define i1 @test16(ptr %P) {
+define i1 @test16(i32* %P) {
; ALL-LABEL: @test16(
-; ALL-NEXT: [[C:%.*]] = icmp ne ptr [[P:%.*]], null
+; ALL-NEXT: [[C:%.*]] = icmp ne i32* [[P:%.*]], null
; ALL-NEXT: ret i1 [[C]]
;
- %c = icmp ne ptr %P, null
+ %c = icmp ne i32* %P, null
ret i1 %c
}
ret i32 %D
}
-define ptr @test27(ptr %A) {
+define [4 x float]* @test27([9 x [4 x float]]* %A) {
; ALL-LABEL: @test27(
-; ALL-NEXT: ret ptr [[A:%.*]]
+; ALL-NEXT: [[C:%.*]] = getelementptr [9 x [4 x float]], [9 x [4 x float]]* [[A:%.*]], i64 0, i64 0
+; ALL-NEXT: ret [4 x float]* [[C]]
;
- ret ptr %A
+ %c = bitcast [9 x [4 x float]]* %A to [4 x float]*
+ ret [4 x float]* %c
}
-define ptr @test28(ptr %A) {
+define float* @test28([4 x float]* %A) {
; ALL-LABEL: @test28(
-; ALL-NEXT: ret ptr [[A:%.*]]
+; ALL-NEXT: [[C:%.*]] = getelementptr [4 x float], [4 x float]* [[A:%.*]], i64 0, i64 0
+; ALL-NEXT: ret float* [[C]]
;
- ret ptr %A
+ %c = bitcast [4 x float]* %A to float*
+ ret float* %c
}
define i32 @test29(i32 %c1, i32 %c2) {
}
; PR1263
-define ptr @test41(ptr %t1) {
+define i32* @test41(i32* %t1) {
; ALL-LABEL: @test41(
-; ALL-NEXT: ret ptr [[T1:%.*]]
+; ALL-NEXT: ret i32* [[T1:%.*]]
;
- ret ptr %t1
+ %t64 = bitcast i32* %t1 to { i32 }*
+ %t65 = getelementptr { i32 }, { i32 }* %t64, i32 0, i32 0
+ ret i32* %t65
}
-define ptr addrspace(1) @test41_addrspacecast_smaller(ptr %t1) {
+define i32 addrspace(1)* @test41_addrspacecast_smaller(i32* %t1) {
; ALL-LABEL: @test41_addrspacecast_smaller(
-; ALL-NEXT: [[T64:%.*]] = addrspacecast ptr [[T1:%.*]] to ptr addrspace(1)
-; ALL-NEXT: ret ptr addrspace(1) [[T64]]
+; ALL-NEXT: [[T65:%.*]] = addrspacecast i32* [[T1:%.*]] to i32 addrspace(1)*
+; ALL-NEXT: ret i32 addrspace(1)* [[T65]]
;
- %t64 = addrspacecast ptr %t1 to ptr addrspace(1)
- ret ptr addrspace(1) %t64
+ %t64 = addrspacecast i32* %t1 to { i32 } addrspace(1)*
+ %t65 = getelementptr { i32 }, { i32 } addrspace(1)* %t64, i32 0, i32 0
+ ret i32 addrspace(1)* %t65
}
-define ptr @test41_addrspacecast_larger(ptr addrspace(1) %t1) {
+define i32* @test41_addrspacecast_larger(i32 addrspace(1)* %t1) {
; ALL-LABEL: @test41_addrspacecast_larger(
-; ALL-NEXT: [[T64:%.*]] = addrspacecast ptr addrspace(1) [[T1:%.*]] to ptr
-; ALL-NEXT: ret ptr [[T64]]
+; ALL-NEXT: [[T65:%.*]] = addrspacecast i32 addrspace(1)* [[T1:%.*]] to i32*
+; ALL-NEXT: ret i32* [[T65]]
;
- %t64 = addrspacecast ptr addrspace(1) %t1 to ptr
- ret ptr %t64
+ %t64 = addrspacecast i32 addrspace(1)* %t1 to { i32 }*
+ %t65 = getelementptr { i32 }, { i32 }* %t64, i32 0, i32 0
+ ret i32* %t65
}
define i32 @test42(i32 %X) {
define i64 @test50(i64 %x) {
; ALL-LABEL: @test50(
; ALL-NEXT: [[TMP1:%.*]] = shl i64 [[X:%.*]], 30
-; ALL-NEXT: [[SEXT:%.*]] = add i64 [[TMP1]], -4294967296
-; ALL-NEXT: [[E:%.*]] = ashr i64 [[SEXT]], 32
+; ALL-NEXT: [[TMP2:%.*]] = add i64 [[TMP1]], -4294967296
+; ALL-NEXT: [[E:%.*]] = ashr i64 [[TMP2]], 32
; ALL-NEXT: ret i64 [[E]]
;
%a = lshr i64 %x, 2
%s = type { i32, i32, i16 }
-define %s @test68(ptr %p, i64 %i) {
+define %s @test68(%s *%p, i64 %i) {
; ALL-LABEL: @test68(
-; ALL-NEXT: [[O:%.*]] = mul i64 [[I:%.*]], 12
-; ALL-NEXT: [[PP:%.*]] = getelementptr inbounds i8, ptr [[P:%.*]], i64 [[O]]
-; ALL-NEXT: [[L:%.*]] = load [[S:%.*]], ptr [[PP]], align 4
+; ALL-NEXT: [[PP1:%.*]] = getelementptr [[S:%.*]], %s* [[P:%.*]], i64 [[I:%.*]]
+; ALL-NEXT: [[L:%.*]] = load [[S]], %s* [[PP1]], align 4
; ALL-NEXT: ret [[S]] [[L]]
;
%o = mul i64 %i, 12
- %pp = getelementptr inbounds i8, ptr %p, i64 %o
- %l = load %s, ptr %pp
+ %q = bitcast %s* %p to i8*
+ %pp = getelementptr inbounds i8, i8* %q, i64 %o
+ %r = bitcast i8* %pp to %s*
+ %l = load %s, %s* %r
ret %s %l
}
; addrspacecasts should be eliminated.
-define %s @test68_addrspacecast(ptr %p, i64 %i) {
+define %s @test68_addrspacecast(%s* %p, i64 %i) {
; ALL-LABEL: @test68_addrspacecast(
-; ALL-NEXT: [[O:%.*]] = mul i64 [[I:%.*]], 12
-; ALL-NEXT: [[Q:%.*]] = addrspacecast ptr [[P:%.*]] to ptr addrspace(2)
-; ALL-NEXT: [[PP:%.*]] = getelementptr inbounds i8, ptr addrspace(2) [[Q]], i64 [[O]]
-; ALL-NEXT: [[R:%.*]] = addrspacecast ptr addrspace(2) [[PP]] to ptr
-; ALL-NEXT: [[L:%.*]] = load [[S:%.*]], ptr [[R]], align 4
+; ALL-NEXT: [[PP1:%.*]] = getelementptr [[S:%.*]], %s* [[P:%.*]], i64 [[I:%.*]]
+; ALL-NEXT: [[L:%.*]] = load [[S]], %s* [[PP1]], align 4
; ALL-NEXT: ret [[S]] [[L]]
;
%o = mul i64 %i, 12
- %q = addrspacecast ptr %p to ptr addrspace(2)
- %pp = getelementptr inbounds i8, ptr addrspace(2) %q, i64 %o
- %r = addrspacecast ptr addrspace(2) %pp to ptr
- %l = load %s, ptr %r
+ %q = addrspacecast %s* %p to i8 addrspace(2)*
+ %pp = getelementptr inbounds i8, i8 addrspace(2)* %q, i64 %o
+ %r = addrspacecast i8 addrspace(2)* %pp to %s*
+ %l = load %s, %s* %r
ret %s %l
}
-define %s @test68_addrspacecast_2(ptr %p, i64 %i) {
+define %s @test68_addrspacecast_2(%s* %p, i64 %i) {
; ALL-LABEL: @test68_addrspacecast_2(
-; ALL-NEXT: [[O:%.*]] = mul i64 [[I:%.*]], 12
-; ALL-NEXT: [[Q:%.*]] = addrspacecast ptr [[P:%.*]] to ptr addrspace(2)
-; ALL-NEXT: [[PP:%.*]] = getelementptr inbounds i8, ptr addrspace(2) [[Q]], i64 [[O]]
-; ALL-NEXT: [[R:%.*]] = addrspacecast ptr addrspace(2) [[PP]] to ptr addrspace(1)
-; ALL-NEXT: [[L:%.*]] = load [[S:%.*]], ptr addrspace(1) [[R]], align 4
+; ALL-NEXT: [[PP1:%.*]] = getelementptr [[S:%.*]], %s* [[P:%.*]], i64 [[I:%.*]]
+; ALL-NEXT: [[R:%.*]] = addrspacecast %s* [[PP1]] to [[S]] addrspace(1)*
+; ALL-NEXT: [[L:%.*]] = load [[S]], [[S]] addrspace(1)* [[R]], align 4
; ALL-NEXT: ret [[S]] [[L]]
;
%o = mul i64 %i, 12
- %q = addrspacecast ptr %p to ptr addrspace(2)
- %pp = getelementptr inbounds i8, ptr addrspace(2) %q, i64 %o
- %r = addrspacecast ptr addrspace(2) %pp to ptr addrspace(1)
- %l = load %s, ptr addrspace(1) %r
+ %q = addrspacecast %s* %p to i8 addrspace(2)*
+ %pp = getelementptr inbounds i8, i8 addrspace(2)* %q, i64 %o
+ %r = addrspacecast i8 addrspace(2)* %pp to %s addrspace(1)*
+ %l = load %s, %s addrspace(1)* %r
ret %s %l
}
-define %s @test68_as1(ptr addrspace(1) %p, i32 %i) {
+define %s @test68_as1(%s addrspace(1)* %p, i32 %i) {
; ALL-LABEL: @test68_as1(
-; ALL-NEXT: [[O:%.*]] = mul i32 [[I:%.*]], 12
-; ALL-NEXT: [[PP:%.*]] = getelementptr inbounds i8, ptr addrspace(1) [[P:%.*]], i32 [[O]]
-; ALL-NEXT: [[L:%.*]] = load [[S:%.*]], ptr addrspace(1) [[PP]], align 4
+; ALL-NEXT: [[PP1:%.*]] = getelementptr [[S:%.*]], [[S]] addrspace(1)* [[P:%.*]], i32 [[I:%.*]]
+; ALL-NEXT: [[L:%.*]] = load [[S]], [[S]] addrspace(1)* [[PP1]], align 4
; ALL-NEXT: ret [[S]] [[L]]
;
%o = mul i32 %i, 12
- %pp = getelementptr inbounds i8, ptr addrspace(1) %p, i32 %o
- %l = load %s, ptr addrspace(1) %pp
+ %q = bitcast %s addrspace(1)* %p to i8 addrspace(1)*
+ %pp = getelementptr inbounds i8, i8 addrspace(1)* %q, i32 %o
+ %r = bitcast i8 addrspace(1)* %pp to %s addrspace(1)*
+ %l = load %s, %s addrspace(1)* %r
ret %s %l
}
-define double @test69(ptr %p, i64 %i) {
+define double @test69(double *%p, i64 %i) {
; ALL-LABEL: @test69(
-; ALL-NEXT: [[O:%.*]] = shl nsw i64 [[I:%.*]], 3
-; ALL-NEXT: [[PP:%.*]] = getelementptr inbounds i8, ptr [[P:%.*]], i64 [[O]]
-; ALL-NEXT: [[L:%.*]] = load double, ptr [[PP]], align 8
+; ALL-NEXT: [[PP1:%.*]] = getelementptr inbounds double, double* [[P:%.*]], i64 [[I:%.*]]
+; ALL-NEXT: [[L:%.*]] = load double, double* [[PP1]], align 8
; ALL-NEXT: ret double [[L]]
;
%o = shl nsw i64 %i, 3
- %pp = getelementptr inbounds i8, ptr %p, i64 %o
- %l = load double, ptr %pp
+ %q = bitcast double* %p to i8*
+ %pp = getelementptr inbounds i8, i8* %q, i64 %o
+ %r = bitcast i8* %pp to double*
+ %l = load double, double* %r
ret double %l
}
-define %s @test70(ptr %p, i64 %i) {
+define %s @test70(%s *%p, i64 %i) {
; ALL-LABEL: @test70(
-; ALL-NEXT: [[O:%.*]] = mul nsw i64 [[I:%.*]], 36
-; ALL-NEXT: [[PP:%.*]] = getelementptr inbounds i8, ptr [[P:%.*]], i64 [[O]]
-; ALL-NEXT: [[L:%.*]] = load [[S:%.*]], ptr [[PP]], align 4
+; ALL-NEXT: [[O:%.*]] = mul nsw i64 [[I:%.*]], 3
+; ALL-NEXT: [[PP1:%.*]] = getelementptr inbounds [[S:%.*]], %s* [[P:%.*]], i64 [[O]]
+; ALL-NEXT: [[L:%.*]] = load [[S]], %s* [[PP1]], align 4
; ALL-NEXT: ret [[S]] [[L]]
;
%o = mul nsw i64 %i, 36
- %pp = getelementptr inbounds i8, ptr %p, i64 %o
- %l = load %s, ptr %pp
+ %q = bitcast %s* %p to i8*
+ %pp = getelementptr inbounds i8, i8* %q, i64 %o
+ %r = bitcast i8* %pp to %s*
+ %l = load %s, %s* %r
ret %s %l
}
-define double @test71(ptr %p, i64 %i) {
+define double @test71(double *%p, i64 %i) {
; ALL-LABEL: @test71(
-; ALL-NEXT: [[O:%.*]] = shl i64 [[I:%.*]], 5
-; ALL-NEXT: [[PP:%.*]] = getelementptr i8, ptr [[P:%.*]], i64 [[O]]
-; ALL-NEXT: [[L:%.*]] = load double, ptr [[PP]], align 8
+; ALL-NEXT: [[O:%.*]] = shl i64 [[I:%.*]], 2
+; ALL-NEXT: [[PP1:%.*]] = getelementptr double, double* [[P:%.*]], i64 [[O]]
+; ALL-NEXT: [[L:%.*]] = load double, double* [[PP1]], align 8
; ALL-NEXT: ret double [[L]]
;
%o = shl i64 %i, 5
- %pp = getelementptr i8, ptr %p, i64 %o
- %l = load double, ptr %pp
+ %q = bitcast double* %p to i8*
+ %pp = getelementptr i8, i8* %q, i64 %o
+ %r = bitcast i8* %pp to double*
+ %l = load double, double* %r
ret double %l
}
-define double @test72(ptr %p, i32 %i) {
+define double @test72(double *%p, i32 %i) {
; ALL-LABEL: @test72(
-; ALL-NEXT: [[SO:%.*]] = shl nsw i32 [[I:%.*]], 3
-; ALL-NEXT: [[O:%.*]] = sext i32 [[SO]] to i64
-; ALL-NEXT: [[PP:%.*]] = getelementptr inbounds i8, ptr [[P:%.*]], i64 [[O]]
-; ALL-NEXT: [[L:%.*]] = load double, ptr [[PP]], align 8
+; ALL-NEXT: [[O:%.*]] = sext i32 [[I:%.*]] to i64
+; ALL-NEXT: [[PP1:%.*]] = getelementptr inbounds double, double* [[P:%.*]], i64 [[O]]
+; ALL-NEXT: [[L:%.*]] = load double, double* [[PP1]], align 8
; ALL-NEXT: ret double [[L]]
;
%so = shl nsw i32 %i, 3
%o = sext i32 %so to i64
- %pp = getelementptr inbounds i8, ptr %p, i64 %o
- %l = load double, ptr %pp
+ %q = bitcast double* %p to i8*
+ %pp = getelementptr inbounds i8, i8* %q, i64 %o
+ %r = bitcast i8* %pp to double*
+ %l = load double, double* %r
ret double %l
}
-define double @test73(ptr %p, i128 %i) {
+define double @test73(double *%p, i128 %i) {
; ALL-LABEL: @test73(
; ALL-NEXT: [[I_TR:%.*]] = trunc i128 [[I:%.*]] to i64
-; ALL-NEXT: [[O:%.*]] = shl i64 [[I_TR]], 3
-; ALL-NEXT: [[PP:%.*]] = getelementptr inbounds i8, ptr [[P:%.*]], i64 [[O]]
-; ALL-NEXT: [[L:%.*]] = load double, ptr [[PP]], align 8
+; ALL-NEXT: [[PP1:%.*]] = getelementptr double, double* [[P:%.*]], i64 [[I_TR]]
+; ALL-NEXT: [[L:%.*]] = load double, double* [[PP1]], align 8
; ALL-NEXT: ret double [[L]]
;
%lo = shl nsw i128 %i, 3
%o = trunc i128 %lo to i64
- %pp = getelementptr inbounds i8, ptr %p, i64 %o
- %l = load double, ptr %pp
+ %q = bitcast double* %p to i8*
+ %pp = getelementptr inbounds i8, i8* %q, i64 %o
+ %r = bitcast i8* %pp to double*
+ %l = load double, double* %r
ret double %l
}
-define double @test74(ptr %p, i64 %i) {
+define double @test74(double *%p, i64 %i) {
; ALL-LABEL: @test74(
-; ALL-NEXT: [[PP:%.*]] = getelementptr inbounds i64, ptr [[P:%.*]], i64 [[I:%.*]]
-; ALL-NEXT: [[L:%.*]] = load double, ptr [[PP]], align 8
+; ALL-NEXT: [[PP1:%.*]] = getelementptr inbounds double, double* [[P:%.*]], i64 [[I:%.*]]
+; ALL-NEXT: [[L:%.*]] = load double, double* [[PP1]], align 8
; ALL-NEXT: ret double [[L]]
;
- %pp = getelementptr inbounds i64, ptr %p, i64 %i
- %l = load double, ptr %pp
+ %q = bitcast double* %p to i64*
+ %pp = getelementptr inbounds i64, i64* %q, i64 %i
+ %r = bitcast i64* %pp to double*
+ %l = load double, double* %r
ret double %l
}
-define ptr @test75(ptr %p, i32 %x) {
+define i32* @test75(i32* %p, i32 %x) {
; ALL-LABEL: @test75(
; ALL-NEXT: [[Y:%.*]] = shl i32 [[X:%.*]], 3
; ALL-NEXT: [[Z:%.*]] = sext i32 [[Y]] to i64
-; ALL-NEXT: [[R:%.*]] = getelementptr i8, ptr [[P:%.*]], i64 [[Z]]
-; ALL-NEXT: ret ptr [[R]]
+; ALL-NEXT: [[Q:%.*]] = bitcast i32* [[P:%.*]] to i8*
+; ALL-NEXT: [[R:%.*]] = getelementptr i8, i8* [[Q]], i64 [[Z]]
+; ALL-NEXT: [[S:%.*]] = bitcast i8* [[R]] to i32*
+; ALL-NEXT: ret i32* [[S]]
;
%y = shl i32 %x, 3
%z = sext i32 %y to i64
- %r = getelementptr i8, ptr %p, i64 %z
- ret ptr %r
+ %q = bitcast i32* %p to i8*
+ %r = getelementptr i8, i8* %q, i64 %z
+ %s = bitcast i8* %r to i32*
+ ret i32* %s
}
-define %s @test76(ptr %p, i64 %i, i64 %j) {
+define %s @test76(%s *%p, i64 %i, i64 %j) {
; ALL-LABEL: @test76(
-; ALL-NEXT: [[O:%.*]] = mul i64 [[I:%.*]], 12
-; ALL-NEXT: [[O2:%.*]] = mul nsw i64 [[O]], [[J:%.*]]
-; ALL-NEXT: [[PP:%.*]] = getelementptr inbounds i8, ptr [[P:%.*]], i64 [[O2]]
-; ALL-NEXT: [[L:%.*]] = load [[S:%.*]], ptr [[PP]], align 4
+; ALL-NEXT: [[O2:%.*]] = mul i64 [[I:%.*]], [[J:%.*]]
+; ALL-NEXT: [[PP1:%.*]] = getelementptr [[S:%.*]], %s* [[P:%.*]], i64 [[O2]]
+; ALL-NEXT: [[L:%.*]] = load [[S]], %s* [[PP1]], align 4
; ALL-NEXT: ret [[S]] [[L]]
;
%o = mul i64 %i, 12
%o2 = mul nsw i64 %o, %j
- %pp = getelementptr inbounds i8, ptr %p, i64 %o2
- %l = load %s, ptr %pp
+ %q = bitcast %s* %p to i8*
+ %pp = getelementptr inbounds i8, i8* %q, i64 %o2
+ %r = bitcast i8* %pp to %s*
+ %l = load %s, %s* %r
ret %s %l
}
-define %s @test77(ptr %p, i64 %i, i64 %j) {
+define %s @test77(%s *%p, i64 %i, i64 %j) {
; ALL-LABEL: @test77(
-; ALL-NEXT: [[O:%.*]] = mul nsw i64 [[I:%.*]], 36
+; ALL-NEXT: [[O:%.*]] = mul nsw i64 [[I:%.*]], 3
; ALL-NEXT: [[O2:%.*]] = mul nsw i64 [[O]], [[J:%.*]]
-; ALL-NEXT: [[PP:%.*]] = getelementptr inbounds i8, ptr [[P:%.*]], i64 [[O2]]
-; ALL-NEXT: [[L:%.*]] = load [[S:%.*]], ptr [[PP]], align 4
+; ALL-NEXT: [[PP1:%.*]] = getelementptr inbounds [[S:%.*]], %s* [[P:%.*]], i64 [[O2]]
+; ALL-NEXT: [[L:%.*]] = load [[S]], %s* [[PP1]], align 4
; ALL-NEXT: ret [[S]] [[L]]
;
%o = mul nsw i64 %i, 36
%o2 = mul nsw i64 %o, %j
- %pp = getelementptr inbounds i8, ptr %p, i64 %o2
- %l = load %s, ptr %pp
+ %q = bitcast %s* %p to i8*
+ %pp = getelementptr inbounds i8, i8* %q, i64 %o2
+ %r = bitcast i8* %pp to %s*
+ %l = load %s, %s* %r
ret %s %l
}
-define %s @test78(ptr %p, i64 %i, i64 %j, i32 %k, i32 %l, i128 %m, i128 %n) {
+define %s @test78(%s *%p, i64 %i, i64 %j, i32 %k, i32 %l, i128 %m, i128 %n) {
; ALL-LABEL: @test78(
-; ALL-NEXT: [[A:%.*]] = mul nsw i32 [[K:%.*]], 36
+; ALL-NEXT: [[A:%.*]] = mul nsw i32 [[K:%.*]], 3
; ALL-NEXT: [[B:%.*]] = mul nsw i32 [[A]], [[L:%.*]]
; ALL-NEXT: [[C:%.*]] = sext i32 [[B]] to i128
; ALL-NEXT: [[D:%.*]] = mul nsw i128 [[C]], [[M:%.*]]
; ALL-NEXT: [[E:%.*]] = mul i128 [[D]], [[N:%.*]]
; ALL-NEXT: [[F:%.*]] = trunc i128 [[E]] to i64
-; ALL-NEXT: [[G:%.*]] = mul nsw i64 [[F]], [[I:%.*]]
-; ALL-NEXT: [[H:%.*]] = mul nsw i64 [[G]], [[J:%.*]]
-; ALL-NEXT: [[PP:%.*]] = getelementptr inbounds i8, ptr [[P:%.*]], i64 [[H]]
-; ALL-NEXT: [[LOAD:%.*]] = load [[S:%.*]], ptr [[PP]], align 4
+; ALL-NEXT: [[G:%.*]] = mul i64 [[F]], [[I:%.*]]
+; ALL-NEXT: [[H:%.*]] = mul i64 [[G]], [[J:%.*]]
+; ALL-NEXT: [[PP1:%.*]] = getelementptr [[S:%.*]], %s* [[P:%.*]], i64 [[H]]
+; ALL-NEXT: [[LOAD:%.*]] = load [[S]], %s* [[PP1]], align 4
; ALL-NEXT: ret [[S]] [[LOAD]]
;
%a = mul nsw i32 %k, 36
%f = trunc i128 %e to i64
%g = mul nsw i64 %f, %i
%h = mul nsw i64 %g, %j
- %pp = getelementptr inbounds i8, ptr %p, i64 %h
- %load = load %s, ptr %pp
+ %q = bitcast %s* %p to i8*
+ %pp = getelementptr inbounds i8, i8* %q, i64 %h
+ %r = bitcast i8* %pp to %s*
+ %load = load %s, %s* %r
ret %s %load
}
-define %s @test79(ptr %p, i64 %i, i32 %j) {
+define %s @test79(%s *%p, i64 %i, i32 %j) {
; ALL-LABEL: @test79(
; ALL-NEXT: [[TMP1:%.*]] = trunc i64 [[I:%.*]] to i32
; ALL-NEXT: [[B:%.*]] = mul i32 [[TMP1]], 36
; ALL-NEXT: [[C:%.*]] = mul i32 [[B]], [[J:%.*]]
+; ALL-NEXT: [[Q:%.*]] = bitcast %s* [[P:%.*]] to i8*
; ALL-NEXT: [[TMP2:%.*]] = sext i32 [[C]] to i64
-; ALL-NEXT: [[PP:%.*]] = getelementptr inbounds i8, ptr [[P:%.*]], i64 [[TMP2]]
-; ALL-NEXT: [[L:%.*]] = load [[S:%.*]], ptr [[PP]], align 4
+; ALL-NEXT: [[PP:%.*]] = getelementptr inbounds i8, i8* [[Q]], i64 [[TMP2]]
+; ALL-NEXT: [[R:%.*]] = bitcast i8* [[PP]] to %s*
+; ALL-NEXT: [[L:%.*]] = load [[S:%.*]], %s* [[R]], align 4
; ALL-NEXT: ret [[S]] [[L]]
;
%a = mul nsw i64 %i, 36
%b = trunc i64 %a to i32
%c = mul i32 %b, %j
- %pp = getelementptr inbounds i8, ptr %p, i32 %c
- %l = load %s, ptr %pp
+ %q = bitcast %s* %p to i8*
+ %pp = getelementptr inbounds i8, i8* %q, i32 %c
+ %r = bitcast i8* %pp to %s*
+ %l = load %s, %s* %r
ret %s %l
}
-define double @test80(ptr %p, i32 %i) {
+define double @test80([100 x double]* %p, i32 %i) {
; ALL-LABEL: @test80(
-; ALL-NEXT: [[T:%.*]] = shl nsw i32 [[I:%.*]], 3
-; ALL-NEXT: [[TMP1:%.*]] = sext i32 [[T]] to i64
-; ALL-NEXT: [[PP:%.*]] = getelementptr i8, ptr [[P:%.*]], i64 [[TMP1]]
-; ALL-NEXT: [[L:%.*]] = load double, ptr [[PP]], align 8
+; ALL-NEXT: [[TMP1:%.*]] = sext i32 [[I:%.*]] to i64
+; ALL-NEXT: [[PP1:%.*]] = getelementptr [100 x double], [100 x double]* [[P:%.*]], i64 0, i64 [[TMP1]]
+; ALL-NEXT: [[L:%.*]] = load double, double* [[PP1]], align 8
; ALL-NEXT: ret double [[L]]
;
%t = shl nsw i32 %i, 3
- %pp = getelementptr i8, ptr %p, i32 %t
- %l = load double, ptr %pp
+ %q = bitcast [100 x double]* %p to i8*
+ %pp = getelementptr i8, i8* %q, i32 %t
+ %r = bitcast i8* %pp to double*
+ %l = load double, double* %r
ret double %l
}
-define double @test80_addrspacecast(ptr addrspace(1) %p, i32 %i) {
+define double @test80_addrspacecast([100 x double] addrspace(1)* %p, i32 %i) {
; ALL-LABEL: @test80_addrspacecast(
-; ALL-NEXT: [[T:%.*]] = shl nsw i32 [[I:%.*]], 3
-; ALL-NEXT: [[Q:%.*]] = addrspacecast ptr addrspace(1) [[P:%.*]] to ptr addrspace(2)
-; ALL-NEXT: [[TMP1:%.*]] = sext i32 [[T]] to i64
-; ALL-NEXT: [[PP:%.*]] = getelementptr i8, ptr addrspace(2) [[Q]], i64 [[TMP1]]
-; ALL-NEXT: [[R:%.*]] = addrspacecast ptr addrspace(2) [[PP]] to ptr addrspace(1)
-; ALL-NEXT: [[L:%.*]] = load double, ptr addrspace(1) [[R]], align 8
+; ALL-NEXT: [[PP1:%.*]] = getelementptr [100 x double], [100 x double] addrspace(1)* [[P:%.*]], i32 0, i32 [[I:%.*]]
+; ALL-NEXT: [[L:%.*]] = load double, double addrspace(1)* [[PP1]], align 8
; ALL-NEXT: ret double [[L]]
;
%t = shl nsw i32 %i, 3
- %q = addrspacecast ptr addrspace(1) %p to ptr addrspace(2)
- %pp = getelementptr i8, ptr addrspace(2) %q, i32 %t
- %r = addrspacecast ptr addrspace(2) %pp to ptr addrspace(1)
- %l = load double, ptr addrspace(1) %r
+ %q = addrspacecast [100 x double] addrspace(1)* %p to i8 addrspace(2)*
+ %pp = getelementptr i8, i8 addrspace(2)* %q, i32 %t
+ %r = addrspacecast i8 addrspace(2)* %pp to double addrspace(1)*
+ %l = load double, double addrspace(1)* %r
ret double %l
}
-define double @test80_addrspacecast_2(ptr addrspace(1) %p, i32 %i) {
+define double @test80_addrspacecast_2([100 x double] addrspace(1)* %p, i32 %i) {
; ALL-LABEL: @test80_addrspacecast_2(
-; ALL-NEXT: [[T:%.*]] = shl nsw i32 [[I:%.*]], 3
-; ALL-NEXT: [[Q:%.*]] = addrspacecast ptr addrspace(1) [[P:%.*]] to ptr addrspace(2)
-; ALL-NEXT: [[TMP1:%.*]] = sext i32 [[T]] to i64
-; ALL-NEXT: [[PP:%.*]] = getelementptr i8, ptr addrspace(2) [[Q]], i64 [[TMP1]]
-; ALL-NEXT: [[R:%.*]] = addrspacecast ptr addrspace(2) [[PP]] to ptr addrspace(3)
-; ALL-NEXT: [[L:%.*]] = load double, ptr addrspace(3) [[R]], align 8
+; ALL-NEXT: [[PP1:%.*]] = getelementptr [100 x double], [100 x double] addrspace(1)* [[P:%.*]], i32 0, i32 [[I:%.*]]
+; ALL-NEXT: [[R:%.*]] = addrspacecast double addrspace(1)* [[PP1]] to double addrspace(3)*
+; ALL-NEXT: [[L:%.*]] = load double, double addrspace(3)* [[R]], align 8
; ALL-NEXT: ret double [[L]]
;
%t = shl nsw i32 %i, 3
- %q = addrspacecast ptr addrspace(1) %p to ptr addrspace(2)
- %pp = getelementptr i8, ptr addrspace(2) %q, i32 %t
- %r = addrspacecast ptr addrspace(2) %pp to ptr addrspace(3)
- %l = load double, ptr addrspace(3) %r
+ %q = addrspacecast [100 x double] addrspace(1)* %p to i8 addrspace(2)*
+ %pp = getelementptr i8, i8 addrspace(2)* %q, i32 %t
+ %r = addrspacecast i8 addrspace(2)* %pp to double addrspace(3)*
+ %l = load double, double addrspace(3)* %r
ret double %l
}
-define double @test80_as1(ptr addrspace(1) %p, i16 %i) {
+define double @test80_as1([100 x double] addrspace(1)* %p, i16 %i) {
; ALL-LABEL: @test80_as1(
-; ALL-NEXT: [[T:%.*]] = shl nsw i16 [[I:%.*]], 3
-; ALL-NEXT: [[TMP1:%.*]] = sext i16 [[T]] to i32
-; ALL-NEXT: [[PP:%.*]] = getelementptr i8, ptr addrspace(1) [[P:%.*]], i32 [[TMP1]]
-; ALL-NEXT: [[L:%.*]] = load double, ptr addrspace(1) [[PP]], align 8
+; ALL-NEXT: [[TMP1:%.*]] = sext i16 [[I:%.*]] to i32
+; ALL-NEXT: [[PP1:%.*]] = getelementptr [100 x double], [100 x double] addrspace(1)* [[P:%.*]], i32 0, i32 [[TMP1]]
+; ALL-NEXT: [[L:%.*]] = load double, double addrspace(1)* [[PP1]], align 8
; ALL-NEXT: ret double [[L]]
;
%t = shl nsw i16 %i, 3
- %pp = getelementptr i8, ptr addrspace(1) %p, i16 %t
- %l = load double, ptr addrspace(1) %pp
+ %q = bitcast [100 x double] addrspace(1)* %p to i8 addrspace(1)*
+ %pp = getelementptr i8, i8 addrspace(1)* %q, i16 %t
+ %r = bitcast i8 addrspace(1)* %pp to double addrspace(1)*
+ %l = load double, double addrspace(1)* %r
ret double %l
}
-define double @test81(ptr %p, float %f) {
+define double @test81(double *%p, float %f) {
; ALL-LABEL: @test81(
; ALL-NEXT: [[I:%.*]] = fptosi float [[F:%.*]] to i64
-; ALL-NEXT: [[PP:%.*]] = getelementptr i8, ptr [[P:%.*]], i64 [[I]]
-; ALL-NEXT: [[L:%.*]] = load double, ptr [[PP]], align 8
+; ALL-NEXT: [[Q:%.*]] = bitcast double* [[P:%.*]] to i8*
+; ALL-NEXT: [[PP:%.*]] = getelementptr i8, i8* [[Q]], i64 [[I]]
+; ALL-NEXT: [[R:%.*]] = bitcast i8* [[PP]] to double*
+; ALL-NEXT: [[L:%.*]] = load double, double* [[R]], align 8
; ALL-NEXT: ret double [[L]]
;
%i = fptosi float %f to i64
- %pp = getelementptr i8, ptr %p, i64 %i
- %l = load double, ptr %pp
+ %q = bitcast double* %p to i8*
+ %pp = getelementptr i8, i8* %q, i64 %i
+ %r = bitcast i8* %pp to double*
+ %l = load double, double* %r
ret double %l
}
ret i16 %t
}
-define i32 @PR21388(ptr %v) {
+define i32 @PR21388(i32* %v) {
; ALL-LABEL: @PR21388(
-; ALL-NEXT: [[ICMP:%.*]] = icmp slt ptr [[V:%.*]], null
+; ALL-NEXT: [[ICMP:%.*]] = icmp slt i32* [[V:%.*]], null
; ALL-NEXT: [[SEXT:%.*]] = sext i1 [[ICMP]] to i32
; ALL-NEXT: ret i32 [[SEXT]]
;
- %icmp = icmp slt ptr %v, null
+ %icmp = icmp slt i32* %v, null
%sext = sext i1 %icmp to i32
ret i32 %sext
}
target datalayout = "p:32:32"
-define i1 @alloca_argument_compare(ptr %arg) {
+define i1 @alloca_argument_compare(i64* %arg) {
; CHECK-LABEL: @alloca_argument_compare(
; CHECK-NEXT: ret i1 false
;
%alloc = alloca i64
- %cmp = icmp eq ptr %arg, %alloc
+ %cmp = icmp eq i64* %arg, %alloc
ret i1 %cmp
}
-define i1 @alloca_argument_compare_swapped(ptr %arg) {
+define i1 @alloca_argument_compare_swapped(i64* %arg) {
; CHECK-LABEL: @alloca_argument_compare_swapped(
; CHECK-NEXT: ret i1 false
;
%alloc = alloca i64
- %cmp = icmp eq ptr %alloc, %arg
+ %cmp = icmp eq i64* %alloc, %arg
ret i1 %cmp
}
-define i1 @alloca_argument_compare_ne(ptr %arg) {
+define i1 @alloca_argument_compare_ne(i64* %arg) {
; CHECK-LABEL: @alloca_argument_compare_ne(
; CHECK-NEXT: ret i1 true
;
%alloc = alloca i64
- %cmp = icmp ne ptr %arg, %alloc
+ %cmp = icmp ne i64* %arg, %alloc
ret i1 %cmp
}
-define i1 @alloca_argument_compare_derived_ptrs(ptr %arg, i64 %x) {
+define i1 @alloca_argument_compare_derived_ptrs(i64* %arg, i64 %x) {
; CHECK-LABEL: @alloca_argument_compare_derived_ptrs(
; CHECK-NEXT: ret i1 false
;
%alloc = alloca i64, i64 8
- %p = getelementptr i64, ptr %arg, i64 %x
- %q = getelementptr i64, ptr %alloc, i64 3
- %cmp = icmp eq ptr %p, %q
+ %p = getelementptr i64, i64* %arg, i64 %x
+ %q = getelementptr i64, i64* %alloc, i64 3
+ %cmp = icmp eq i64* %p, %q
ret i1 %cmp
}
-declare void @escape(ptr)
-define i1 @alloca_argument_compare_escaped_alloca(ptr %arg) {
+declare void @escape(i64*)
+define i1 @alloca_argument_compare_escaped_alloca(i64* %arg) {
; CHECK-LABEL: @alloca_argument_compare_escaped_alloca(
; CHECK-NEXT: [[ALLOC:%.*]] = alloca i64, align 8
-; CHECK-NEXT: call void @escape(ptr nonnull [[ALLOC]])
-; CHECK-NEXT: [[CMP:%.*]] = icmp eq ptr [[ALLOC]], [[ARG:%.*]]
+; CHECK-NEXT: call void @escape(i64* nonnull [[ALLOC]])
+; CHECK-NEXT: [[CMP:%.*]] = icmp eq i64* [[ALLOC]], [[ARG:%.*]]
; CHECK-NEXT: ret i1 [[CMP]]
;
%alloc = alloca i64
- call void @escape(ptr %alloc)
- %cmp = icmp eq ptr %alloc, %arg
+ call void @escape(i64* %alloc)
+ %cmp = icmp eq i64* %alloc, %arg
ret i1 %cmp
}
declare void @check_compares(i1, i1)
-define void @alloca_argument_compare_two_compares(ptr %p) {
+define void @alloca_argument_compare_two_compares(i64* %p) {
; CHECK-LABEL: @alloca_argument_compare_two_compares(
; CHECK-NEXT: [[Q1:%.*]] = alloca [8 x i64], align 8
-; CHECK-NEXT: [[R:%.*]] = getelementptr i64, ptr [[P:%.*]], i32 1
-; CHECK-NEXT: [[S:%.*]] = getelementptr inbounds i64, ptr [[Q1]], i32 2
-; CHECK-NEXT: [[CMP1:%.*]] = icmp eq ptr [[Q1]], [[P]]
-; CHECK-NEXT: [[CMP2:%.*]] = icmp eq ptr [[R]], [[S]]
+; CHECK-NEXT: [[Q1_SUB:%.*]] = getelementptr inbounds [8 x i64], [8 x i64]* [[Q1]], i32 0, i32 0
+; CHECK-NEXT: [[R:%.*]] = getelementptr i64, i64* [[P:%.*]], i32 1
+; CHECK-NEXT: [[S:%.*]] = getelementptr inbounds [8 x i64], [8 x i64]* [[Q1]], i32 0, i32 2
+; CHECK-NEXT: [[CMP1:%.*]] = icmp eq i64* [[Q1_SUB]], [[P]]
+; CHECK-NEXT: [[CMP2:%.*]] = icmp eq i64* [[R]], [[S]]
; CHECK-NEXT: call void @check_compares(i1 [[CMP1]], i1 [[CMP2]])
; CHECK-NEXT: ret void
;
%q = alloca i64, i64 8
- %r = getelementptr i64, ptr %p, i64 1
- %s = getelementptr i64, ptr %q, i64 2
- %cmp1 = icmp eq ptr %p, %q
- %cmp2 = icmp eq ptr %r, %s
+ %r = getelementptr i64, i64* %p, i64 1
+ %s = getelementptr i64, i64* %q, i64 2
+ %cmp1 = icmp eq i64* %p, %q
+ %cmp2 = icmp eq i64* %r, %s
call void @check_compares(i1 %cmp1, i1 %cmp2)
ret void
; We will only fold if there is a single cmp.
}
-define i1 @alloca_argument_compare_escaped_through_store(ptr %arg, ptr %ptr) {
+define i1 @alloca_argument_compare_escaped_through_store(i64* %arg, i64** %ptr) {
; CHECK-LABEL: @alloca_argument_compare_escaped_through_store(
; CHECK-NEXT: [[ALLOC:%.*]] = alloca i64, align 8
-; CHECK-NEXT: [[CMP:%.*]] = icmp eq ptr [[ALLOC]], [[ARG:%.*]]
-; CHECK-NEXT: [[P:%.*]] = getelementptr inbounds i64, ptr [[ALLOC]], i32 1
-; CHECK-NEXT: store ptr [[P]], ptr [[PTR:%.*]], align 4
+; CHECK-NEXT: [[CMP:%.*]] = icmp eq i64* [[ALLOC]], [[ARG:%.*]]
+; CHECK-NEXT: [[P:%.*]] = getelementptr inbounds i64, i64* [[ALLOC]], i32 1
+; CHECK-NEXT: store i64* [[P]], i64** [[PTR:%.*]], align 4
; CHECK-NEXT: ret i1 [[CMP]]
;
%alloc = alloca i64
- %cmp = icmp eq ptr %alloc, %arg
- %p = getelementptr i64, ptr %alloc, i64 1
- store ptr %p, ptr %ptr
+ %cmp = icmp eq i64* %alloc, %arg
+ %p = getelementptr i64, i64* %alloc, i64 1
+ store i64* %p, i64** %ptr
ret i1 %cmp
}
-declare void @llvm.lifetime.start.p0(i64, ptr nocapture)
-declare void @llvm.lifetime.end.p0(i64, ptr nocapture)
-define i1 @alloca_argument_compare_benign_instrs(ptr %arg) {
+declare void @llvm.lifetime.start.p0i8(i64, i8* nocapture)
+declare void @llvm.lifetime.end.p0i8(i64, i8* nocapture)
+define i1 @alloca_argument_compare_benign_instrs(i8* %arg) {
; CHECK-LABEL: @alloca_argument_compare_benign_instrs(
; CHECK-NEXT: ret i1 false
;
%alloc = alloca i8
- call void @llvm.lifetime.start.p0(i64 1, ptr %alloc)
- %cmp = icmp eq ptr %arg, %alloc
- %x = load i8, ptr %arg
- store i8 %x, ptr %alloc
- call void @llvm.lifetime.end.p0(i64 1, ptr %alloc)
+ call void @llvm.lifetime.start.p0i8(i64 1, i8* %alloc)
+ %cmp = icmp eq i8* %arg, %alloc
+ %x = load i8, i8* %arg
+ store i8 %x, i8* %alloc
+ call void @llvm.lifetime.end.p0i8(i64 1, i8* %alloc)
ret i1 %cmp
}
-declare ptr @allocator()
+declare i64* @allocator()
define i1 @alloca_call_compare() {
; CHECK-LABEL: @alloca_call_compare(
-; CHECK-NEXT: [[Q:%.*]] = call ptr @allocator()
+; CHECK-NEXT: [[Q:%.*]] = call i64* @allocator()
; CHECK-NEXT: ret i1 false
;
%p = alloca i64
- %q = call ptr @allocator()
- %cmp = icmp eq ptr %p, %q
+ %q = call i64* @allocator()
+ %cmp = icmp eq i64* %p, %q
ret i1 %cmp
}
; These two functions represents either a) forging a pointer via inttoptr or
; b) indexing off an adjacent allocation. In either case, the operation is
; obscured by an uninlined helper and not visible to instcombine.
-declare ptr @hidden_inttoptr()
-declare ptr @hidden_offset(ptr %other)
+declare i8* @hidden_inttoptr()
+declare i8* @hidden_offset(i8* %other)
define i1 @ptrtoint_single_cmp() {
; CHECK-LABEL: @ptrtoint_single_cmp(
; CHECK-NEXT: ret i1 false
;
%m = alloca i8, i32 4
- %rhs = inttoptr i64 2048 to ptr
- %cmp = icmp eq ptr %m, %rhs
+ %rhs = inttoptr i64 2048 to i8*
+ %cmp = icmp eq i8* %m, %rhs
ret i1 %cmp
}
;
%m = alloca i8, i32 4
%n = alloca i8, i32 4
- %rhs = getelementptr i8, ptr %n, i32 4
- %cmp = icmp eq ptr %m, %rhs
+ %rhs = getelementptr i8, i8* %n, i32 4
+ %cmp = icmp eq i8* %m, %rhs
ret i1 %cmp
}
define void @neg_consistent_fold1() {
; CHECK-LABEL: @neg_consistent_fold1(
; CHECK-NEXT: [[M1:%.*]] = alloca [4 x i8], align 1
-; CHECK-NEXT: [[RHS2:%.*]] = call ptr @hidden_inttoptr()
-; CHECK-NEXT: [[CMP1:%.*]] = icmp eq ptr [[M1]], inttoptr (i64 2048 to ptr)
-; CHECK-NEXT: [[CMP2:%.*]] = icmp eq ptr [[M1]], [[RHS2]]
+; CHECK-NEXT: [[M1_SUB:%.*]] = getelementptr inbounds [4 x i8], [4 x i8]* [[M1]], i32 0, i32 0
+; CHECK-NEXT: [[RHS2:%.*]] = call i8* @hidden_inttoptr()
+; CHECK-NEXT: [[CMP1:%.*]] = icmp eq i8* [[M1_SUB]], inttoptr (i64 2048 to i8*)
+; CHECK-NEXT: [[CMP2:%.*]] = icmp eq i8* [[M1_SUB]], [[RHS2]]
; CHECK-NEXT: call void @witness(i1 [[CMP1]], i1 [[CMP2]])
; CHECK-NEXT: ret void
;
%m = alloca i8, i32 4
- %rhs = inttoptr i64 2048 to ptr
- %rhs2 = call ptr @hidden_inttoptr()
- %cmp1 = icmp eq ptr %m, %rhs
- %cmp2 = icmp eq ptr %m, %rhs2
+ %rhs = inttoptr i64 2048 to i8*
+ %rhs2 = call i8* @hidden_inttoptr()
+ %cmp1 = icmp eq i8* %m, %rhs
+ %cmp2 = icmp eq i8* %m, %rhs2
call void @witness(i1 %cmp1, i1 %cmp2)
ret void
}
; CHECK-LABEL: @neg_consistent_fold2(
; CHECK-NEXT: [[M1:%.*]] = alloca [4 x i8], align 1
; CHECK-NEXT: [[N2:%.*]] = alloca [4 x i8], align 1
-; CHECK-NEXT: [[RHS:%.*]] = getelementptr inbounds i8, ptr [[N2]], i32 4
-; CHECK-NEXT: [[RHS2:%.*]] = call ptr @hidden_offset(ptr nonnull [[N2]])
-; CHECK-NEXT: [[CMP1:%.*]] = icmp eq ptr [[M1]], [[RHS]]
-; CHECK-NEXT: [[CMP2:%.*]] = icmp eq ptr [[M1]], [[RHS2]]
+; CHECK-NEXT: [[N2_SUB:%.*]] = getelementptr inbounds [4 x i8], [4 x i8]* [[N2]], i32 0, i32 0
+; CHECK-NEXT: [[M1_SUB:%.*]] = getelementptr inbounds [4 x i8], [4 x i8]* [[M1]], i32 0, i32 0
+; CHECK-NEXT: [[RHS:%.*]] = getelementptr inbounds [4 x i8], [4 x i8]* [[N2]], i32 0, i32 4
+; CHECK-NEXT: [[RHS2:%.*]] = call i8* @hidden_offset(i8* nonnull [[N2_SUB]])
+; CHECK-NEXT: [[CMP1:%.*]] = icmp eq i8* [[M1_SUB]], [[RHS]]
+; CHECK-NEXT: [[CMP2:%.*]] = icmp eq i8* [[M1_SUB]], [[RHS2]]
; CHECK-NEXT: call void @witness(i1 [[CMP1]], i1 [[CMP2]])
; CHECK-NEXT: ret void
;
%m = alloca i8, i32 4
%n = alloca i8, i32 4
- %rhs = getelementptr i8, ptr %n, i32 4
- %rhs2 = call ptr @hidden_offset(ptr %n)
- %cmp1 = icmp eq ptr %m, %rhs
- %cmp2 = icmp eq ptr %m, %rhs2
+ %rhs = getelementptr i8, i8* %n, i32 4
+ %rhs2 = call i8* @hidden_offset(i8* %n)
+ %cmp1 = icmp eq i8* %m, %rhs
+ %cmp2 = icmp eq i8* %m, %rhs2
call void @witness(i1 %cmp1, i1 %cmp2)
ret void
}
define void @neg_consistent_fold3() {
; CHECK-LABEL: @neg_consistent_fold3(
-; CHECK-NEXT: [[M1:%.*]] = alloca [4 x i8], align 1
-; CHECK-NEXT: [[LGP:%.*]] = load ptr, ptr @gp, align 8
-; CHECK-NEXT: [[RHS2:%.*]] = call ptr @hidden_inttoptr()
-; CHECK-NEXT: [[CMP1:%.*]] = icmp eq ptr [[M1]], [[LGP]]
-; CHECK-NEXT: [[CMP2:%.*]] = icmp eq ptr [[M1]], [[RHS2]]
+; CHECK-NEXT: [[M1:%.*]] = alloca i32, align 1
+; CHECK-NEXT: [[M1_SUB:%.*]] = bitcast i32* [[M1]] to i8*
+; CHECK-NEXT: [[LGP:%.*]] = load i32*, i32** @gp, align 8
+; CHECK-NEXT: [[RHS2:%.*]] = call i8* @hidden_inttoptr()
+; CHECK-NEXT: [[CMP1:%.*]] = icmp eq i32* [[M1]], [[LGP]]
+; CHECK-NEXT: [[CMP2:%.*]] = icmp eq i8* [[RHS2]], [[M1_SUB]]
; CHECK-NEXT: call void @witness(i1 [[CMP1]], i1 [[CMP2]])
; CHECK-NEXT: ret void
;
%m = alloca i8, i32 4
- %lgp = load ptr, ptr @gp, align 8
- %rhs2 = call ptr @hidden_inttoptr()
- %cmp1 = icmp eq ptr %m, %lgp
- %cmp2 = icmp eq ptr %m, %rhs2
+ %bc = bitcast i8* %m to i32*
+ %lgp = load i32*, i32** @gp, align 8
+ %rhs2 = call i8* @hidden_inttoptr()
+ %cmp1 = icmp eq i32* %bc, %lgp
+ %cmp2 = icmp eq i8* %m, %rhs2
call void @witness(i1 %cmp1, i1 %cmp2)
ret void
}
; CHECK-NEXT: ret void
;
%m = alloca i8, i32 4
- %lgp = load ptr, ptr @gp, align 8
- %cmp1 = icmp eq ptr %m, %lgp
- %cmp2 = icmp eq ptr %m, %lgp
+ %bc = bitcast i8* %m to i32*
+ %lgp = load i32*, i32** @gp, align 8
+ %cmp1 = icmp eq i32* %bc, %lgp
+ %cmp2 = icmp eq i32* %bc, %lgp
call void @witness(i1 %cmp1, i1 %cmp2)
ret void
}
; assumption) not able to contain a comparison which might capture the
; address.
-declare void @unknown(ptr)
+declare void @unknown(i8*)
; TODO: Missing optimization
define i1 @consistent_nocapture_inttoptr() {
; CHECK-LABEL: @consistent_nocapture_inttoptr(
; CHECK-NEXT: [[M1:%.*]] = alloca [4 x i8], align 1
-; CHECK-NEXT: call void @unknown(ptr nocapture nonnull [[M1]])
-; CHECK-NEXT: [[CMP:%.*]] = icmp eq ptr [[M1]], inttoptr (i64 2048 to ptr)
+; CHECK-NEXT: [[M1_SUB:%.*]] = getelementptr inbounds [4 x i8], [4 x i8]* [[M1]], i32 0, i32 0
+; CHECK-NEXT: call void @unknown(i8* nocapture nonnull [[M1_SUB]])
+; CHECK-NEXT: [[CMP:%.*]] = icmp eq i8* [[M1_SUB]], inttoptr (i64 2048 to i8*)
; CHECK-NEXT: ret i1 [[CMP]]
;
%m = alloca i8, i32 4
- call void @unknown(ptr nocapture %m)
- %rhs = inttoptr i64 2048 to ptr
- %cmp = icmp eq ptr %m, %rhs
+ call void @unknown(i8* nocapture %m)
+ %rhs = inttoptr i64 2048 to i8*
+ %cmp = icmp eq i8* %m, %rhs
ret i1 %cmp
}
define i1 @consistent_nocapture_offset() {
; CHECK-LABEL: @consistent_nocapture_offset(
; CHECK-NEXT: [[M1:%.*]] = alloca [4 x i8], align 1
-; CHECK-NEXT: call void @unknown(ptr nocapture nonnull [[M1]])
+; CHECK-NEXT: [[M1_SUB:%.*]] = getelementptr inbounds [4 x i8], [4 x i8]* [[M1]], i32 0, i32 0
+; CHECK-NEXT: call void @unknown(i8* nocapture nonnull [[M1_SUB]])
; CHECK-NEXT: ret i1 false
;
%m = alloca i8, i32 4
- call void @unknown(ptr nocapture %m)
+ call void @unknown(i8* nocapture %m)
%n = alloca i8, i32 4
- %rhs = getelementptr i8, ptr %n, i32 4
- %cmp = icmp eq ptr %m, %rhs
+ %rhs = getelementptr i8, i8* %n, i32 4
+ %cmp = icmp eq i8* %m, %rhs
ret i1 %cmp
}
-@gp = global ptr null, align 8
+@gp = global i32* null, align 8
; TODO: Missing optimization
define i1 @consistent_nocapture_through_global() {
; CHECK-LABEL: @consistent_nocapture_through_global(
-; CHECK-NEXT: [[M1:%.*]] = alloca [4 x i8], align 1
-; CHECK-NEXT: call void @unknown(ptr nocapture nonnull [[M1]])
-; CHECK-NEXT: [[LGP:%.*]] = load ptr, ptr @gp, align 8, !nonnull !0
-; CHECK-NEXT: [[CMP:%.*]] = icmp eq ptr [[M1]], [[LGP]]
+; CHECK-NEXT: [[M1:%.*]] = alloca i32, align 1
+; CHECK-NEXT: [[M1_SUB:%.*]] = bitcast i32* [[M1]] to i8*
+; CHECK-NEXT: call void @unknown(i8* nocapture nonnull [[M1_SUB]])
+; CHECK-NEXT: [[LGP:%.*]] = load i32*, i32** @gp, align 8, !nonnull !0
+; CHECK-NEXT: [[CMP:%.*]] = icmp eq i32* [[M1]], [[LGP]]
; CHECK-NEXT: ret i1 [[CMP]]
;
%m = alloca i8, i32 4
- call void @unknown(ptr nocapture %m)
- %lgp = load ptr, ptr @gp, align 8, !nonnull !{}
- %cmp = icmp eq ptr %m, %lgp
+ call void @unknown(i8* nocapture %m)
+ %bc = bitcast i8* %m to i32*
+ %lgp = load i32*, i32** @gp, align 8, !nonnull !{}
+ %cmp = icmp eq i32* %bc, %lgp
ret i1 %cmp
}
; NOTE: Assertions have been autogenerated by utils/update_test_checks.py
; RUN: opt -passes=instcombine -S < %s | FileCheck %s
-@gp = global ptr null, align 8
+@gp = global i32* null, align 8
-declare noalias ptr @malloc(i64) allockind("alloc,uninitialized") allocsize(0)
+declare noalias i8* @malloc(i64) allockind("alloc,uninitialized") allocsize(0)
define i1 @compare_global_trivialeq() {
; CHECK-LABEL: @compare_global_trivialeq(
; CHECK-NEXT: ret i1 false
;
- %m = call ptr @malloc(i64 4)
- %lgp = load ptr, ptr @gp, align 8
- %cmp = icmp eq ptr %m, %lgp
+ %m = call i8* @malloc(i64 4)
+ %bc = bitcast i8* %m to i32*
+ %lgp = load i32*, i32** @gp, align 8
+ %cmp = icmp eq i32* %bc, %lgp
ret i1 %cmp
}
; CHECK-LABEL: @compare_global_trivialne(
; CHECK-NEXT: ret i1 true
;
- %m = call ptr @malloc(i64 4)
- %lgp = load ptr, ptr @gp, align 8
- %cmp = icmp ne ptr %m, %lgp
+ %m = call i8* @malloc(i64 4)
+ %bc = bitcast i8* %m to i32*
+ %lgp = load i32*, i32** @gp, align 8
+ %cmp = icmp ne i32* %bc, %lgp
ret i1 %cmp
}
declare void @f()
define i1 @compare_and_call_with_deopt() {
; CHECK-LABEL: @compare_and_call_with_deopt(
-; CHECK-NEXT: [[M:%.*]] = call dereferenceable_or_null(24) ptr @malloc(i64 24)
-; CHECK-NEXT: tail call void @f() [ "deopt"(ptr [[M]]) ]
+; CHECK-NEXT: [[M:%.*]] = call dereferenceable_or_null(24) i8* @malloc(i64 24)
+; CHECK-NEXT: tail call void @f() [ "deopt"(i8* [[M]]) ]
; CHECK-NEXT: ret i1 false
;
- %m = call ptr @malloc(i64 24)
- %lgp = load ptr, ptr @gp, align 8, !nonnull !0
- %cmp = icmp eq ptr %lgp, %m
- tail call void @f() [ "deopt"(ptr %m) ]
+ %m = call i8* @malloc(i64 24)
+ %bc = bitcast i8* %m to i32*
+ %lgp = load i32*, i32** @gp, align 8, !nonnull !0
+ %cmp = icmp eq i32* %lgp, %bc
+ tail call void @f() [ "deopt"(i8* %m) ]
ret i1 %cmp
}
; Same functon as above with deopt operand in function f, but comparison is NE
define i1 @compare_ne_and_call_with_deopt() {
; CHECK-LABEL: @compare_ne_and_call_with_deopt(
-; CHECK-NEXT: [[M:%.*]] = call dereferenceable_or_null(24) ptr @malloc(i64 24)
-; CHECK-NEXT: tail call void @f() [ "deopt"(ptr [[M]]) ]
+; CHECK-NEXT: [[M:%.*]] = call dereferenceable_or_null(24) i8* @malloc(i64 24)
+; CHECK-NEXT: tail call void @f() [ "deopt"(i8* [[M]]) ]
; CHECK-NEXT: ret i1 true
;
- %m = call ptr @malloc(i64 24)
- %lgp = load ptr, ptr @gp, align 8, !nonnull !0
- %cmp = icmp ne ptr %lgp, %m
- tail call void @f() [ "deopt"(ptr %m) ]
+ %m = call i8* @malloc(i64 24)
+ %bc = bitcast i8* %m to i32*
+ %lgp = load i32*, i32** @gp, align 8, !nonnull !0
+ %cmp = icmp ne i32* %lgp, %bc
+ tail call void @f() [ "deopt"(i8* %m) ]
ret i1 %cmp
}
; Same function as above, but global not marked nonnull, and we cannot fold the comparison
define i1 @compare_ne_global_maybe_null() {
; CHECK-LABEL: @compare_ne_global_maybe_null(
-; CHECK-NEXT: [[M:%.*]] = call dereferenceable_or_null(24) ptr @malloc(i64 24)
-; CHECK-NEXT: [[LGP:%.*]] = load ptr, ptr @gp, align 8
-; CHECK-NEXT: [[CMP:%.*]] = icmp ne ptr [[LGP]], [[M]]
-; CHECK-NEXT: tail call void @f() [ "deopt"(ptr [[M]]) ]
+; CHECK-NEXT: [[M:%.*]] = call dereferenceable_or_null(24) i8* @malloc(i64 24)
+; CHECK-NEXT: [[BC:%.*]] = bitcast i8* [[M]] to i32*
+; CHECK-NEXT: [[LGP:%.*]] = load i32*, i32** @gp, align 8
+; CHECK-NEXT: [[CMP:%.*]] = icmp ne i32* [[LGP]], [[BC]]
+; CHECK-NEXT: tail call void @f() [ "deopt"(i8* [[M]]) ]
; CHECK-NEXT: ret i1 [[CMP]]
;
- %m = call ptr @malloc(i64 24)
- %lgp = load ptr, ptr @gp
- %cmp = icmp ne ptr %lgp, %m
- tail call void @f() [ "deopt"(ptr %m) ]
+ %m = call i8* @malloc(i64 24)
+ %bc = bitcast i8* %m to i32*
+ %lgp = load i32*, i32** @gp
+ %cmp = icmp ne i32* %lgp, %bc
+ tail call void @f() [ "deopt"(i8* %m) ]
ret i1 %cmp
}
; FIXME: The comparison should fold to false since %m escapes (call to function escape)
; after the comparison.
-declare void @escape(ptr)
+declare void @escape(i8*)
define i1 @compare_and_call_after() {
; CHECK-LABEL: @compare_and_call_after(
-; CHECK-NEXT: [[M:%.*]] = call dereferenceable_or_null(24) ptr @malloc(i64 24)
-; CHECK-NEXT: [[LGP:%.*]] = load ptr, ptr @gp, align 8, !nonnull !0
-; CHECK-NEXT: [[CMP:%.*]] = icmp eq ptr [[M]], [[LGP]]
+; CHECK-NEXT: [[M:%.*]] = call dereferenceable_or_null(24) i8* @malloc(i64 24)
+; CHECK-NEXT: [[BC:%.*]] = bitcast i8* [[M]] to i32*
+; CHECK-NEXT: [[LGP:%.*]] = load i32*, i32** @gp, align 8, !nonnull !0
+; CHECK-NEXT: [[CMP:%.*]] = icmp eq i32* [[LGP]], [[BC]]
; CHECK-NEXT: br i1 [[CMP]], label [[ESCAPE_CALL:%.*]], label [[JUST_RETURN:%.*]]
; CHECK: escape_call:
-; CHECK-NEXT: call void @escape(ptr [[M]])
+; CHECK-NEXT: call void @escape(i8* [[M]])
; CHECK-NEXT: ret i1 true
; CHECK: just_return:
; CHECK-NEXT: ret i1 [[CMP]]
;
- %m = call ptr @malloc(i64 24)
- %lgp = load ptr, ptr @gp, align 8, !nonnull !0
- %cmp = icmp eq ptr %m, %lgp
+ %m = call i8* @malloc(i64 24)
+ %bc = bitcast i8* %m to i32*
+ %lgp = load i32*, i32** @gp, align 8, !nonnull !0
+ %cmp = icmp eq i32* %bc, %lgp
br i1 %cmp, label %escape_call, label %just_return
escape_call:
- call void @escape(ptr %m)
+ call void @escape(i8* %m)
ret i1 true
just_return:
; CHECK-LABEL: @compare_distinct_mallocs(
; CHECK-NEXT: ret i1 false
;
- %m = call ptr @malloc(i64 4)
- %n = call ptr @malloc(i64 4)
- %cmp = icmp eq ptr %m, %n
+ %m = call i8* @malloc(i64 4)
+ %n = call i8* @malloc(i64 4)
+ %cmp = icmp eq i8* %m, %n
ret i1 %cmp
}
; CHECK-LABEL: @compare_samepointer_under_bitcast(
; CHECK-NEXT: ret i1 true
;
- %m = call ptr @malloc(i64 4)
- %cmp = icmp eq ptr %m, %m
+ %m = call i8* @malloc(i64 4)
+ %bc = bitcast i8* %m to i32*
+ %bcback = bitcast i32* %bc to i8*
+ %cmp = icmp eq i8* %m, %bcback
ret i1 %cmp
}
; The malloc call for %m cannot be elided since it is used in the call to function f.
define i1 @compare_samepointer_escaped() {
; CHECK-LABEL: @compare_samepointer_escaped(
-; CHECK-NEXT: [[M:%.*]] = call dereferenceable_or_null(4) ptr @malloc(i64 4)
-; CHECK-NEXT: call void @f() [ "deopt"(ptr [[M]]) ]
+; CHECK-NEXT: [[M:%.*]] = call dereferenceable_or_null(4) i8* @malloc(i64 4)
+; CHECK-NEXT: call void @f() [ "deopt"(i8* [[M]]) ]
; CHECK-NEXT: ret i1 true
;
- %m = call ptr @malloc(i64 4)
- %cmp = icmp eq ptr %m, %m
- call void @f() [ "deopt"(ptr %m) ]
+ %m = call i8* @malloc(i64 4)
+ %bc = bitcast i8* %m to i32*
+ %bcback = bitcast i32* %bc to i8*
+ %cmp = icmp eq i8* %m, %bcback
+ call void @f() [ "deopt"(i8* %m) ]
ret i1 %cmp
}
; FIXME: Folding this %cmp2 when %m escapes through ret could be an issue with
; cross-threading data dependencies since we do not make the distinction between
; atomic and non-atomic loads in capture tracking.
-define ptr @compare_ret_escape(ptr %c) {
+define i8* @compare_ret_escape(i8* %c) {
; CHECK-LABEL: @compare_ret_escape(
-; CHECK-NEXT: [[M:%.*]] = call dereferenceable_or_null(4) ptr @malloc(i64 4)
-; CHECK-NEXT: [[N:%.*]] = call dereferenceable_or_null(4) ptr @malloc(i64 4)
-; CHECK-NEXT: [[CMP:%.*]] = icmp eq ptr [[N]], [[C:%.*]]
+; CHECK-NEXT: [[M:%.*]] = call dereferenceable_or_null(4) i8* @malloc(i64 4)
+; CHECK-NEXT: [[N:%.*]] = call dereferenceable_or_null(4) i8* @malloc(i64 4)
+; CHECK-NEXT: [[CMP:%.*]] = icmp eq i8* [[N]], [[C:%.*]]
; CHECK-NEXT: br i1 [[CMP]], label [[RETST:%.*]], label [[CHK:%.*]]
; CHECK: retst:
-; CHECK-NEXT: ret ptr [[M]]
+; CHECK-NEXT: ret i8* [[M]]
; CHECK: chk:
-; CHECK-NEXT: [[LGP:%.*]] = load ptr, ptr @gp, align 8, !nonnull !0
-; CHECK-NEXT: [[CMP2:%.*]] = icmp eq ptr [[M]], [[LGP]]
+; CHECK-NEXT: [[BC:%.*]] = bitcast i8* [[M]] to i32*
+; CHECK-NEXT: [[LGP:%.*]] = load i32*, i32** @gp, align 8, !nonnull !0
+; CHECK-NEXT: [[CMP2:%.*]] = icmp eq i32* [[LGP]], [[BC]]
; CHECK-NEXT: br i1 [[CMP2]], label [[RETST]], label [[CHK2:%.*]]
; CHECK: chk2:
-; CHECK-NEXT: ret ptr [[N]]
+; CHECK-NEXT: ret i8* [[N]]
;
- %m = call ptr @malloc(i64 4)
- %n = call ptr @malloc(i64 4)
- %cmp = icmp eq ptr %n, %c
+ %m = call i8* @malloc(i64 4)
+ %n = call i8* @malloc(i64 4)
+ %cmp = icmp eq i8* %n, %c
br i1 %cmp, label %retst, label %chk
retst:
- ret ptr %m
+ ret i8* %m
chk:
- %lgp = load ptr, ptr @gp, align 8, !nonnull !0
- %cmp2 = icmp eq ptr %m, %lgp
+ %bc = bitcast i8* %m to i32*
+ %lgp = load i32*, i32** @gp, align 8, !nonnull !0
+ %cmp2 = icmp eq i32* %bc, %lgp
br i1 %cmp2, label %retst, label %chk2
chk2:
- ret ptr %n
+ ret i8* %n
}
; The malloc call for %m cannot be elided since it is used in the call to function f.
; However, the cmp can be folded to true as %n doesnt escape and %m, %n are distinct allocations
define i1 @compare_distinct_pointer_escape() {
; CHECK-LABEL: @compare_distinct_pointer_escape(
-; CHECK-NEXT: [[M:%.*]] = call dereferenceable_or_null(4) ptr @malloc(i64 4)
-; CHECK-NEXT: tail call void @f() [ "deopt"(ptr [[M]]) ]
+; CHECK-NEXT: [[M:%.*]] = call dereferenceable_or_null(4) i8* @malloc(i64 4)
+; CHECK-NEXT: tail call void @f() [ "deopt"(i8* [[M]]) ]
; CHECK-NEXT: ret i1 true
;
- %m = call ptr @malloc(i64 4)
- %n = call ptr @malloc(i64 4)
- tail call void @f() [ "deopt"(ptr %m) ]
- %cmp = icmp ne ptr %m, %n
+ %m = call i8* @malloc(i64 4)
+ %n = call i8* @malloc(i64 4)
+ tail call void @f() [ "deopt"(i8* %m) ]
+ %cmp = icmp ne i8* %m, %n
ret i1 %cmp
}
; These two functions represents either a) forging a pointer via inttoptr or
; b) indexing off an adjacent allocation. In either case, the operation is
; obscured by an uninlined helper and not visible to instcombine.
-declare ptr @hidden_inttoptr()
-declare ptr @hidden_offset(ptr %other)
+declare i8* @hidden_inttoptr()
+declare i8* @hidden_offset(i8* %other)
; FIXME: Missed oppurtunity
define i1 @ptrtoint_single_cmp() {
; CHECK-LABEL: @ptrtoint_single_cmp(
-; CHECK-NEXT: [[M:%.*]] = call dereferenceable_or_null(4) ptr @malloc(i64 4)
-; CHECK-NEXT: [[CMP:%.*]] = icmp eq ptr [[M]], inttoptr (i64 2048 to ptr)
+; CHECK-NEXT: [[M:%.*]] = call dereferenceable_or_null(4) i8* @malloc(i64 4)
+; CHECK-NEXT: [[CMP:%.*]] = icmp eq i8* [[M]], inttoptr (i64 2048 to i8*)
; CHECK-NEXT: ret i1 [[CMP]]
;
- %m = call ptr @malloc(i64 4)
- %rhs = inttoptr i64 2048 to ptr
- %cmp = icmp eq ptr %m, %rhs
+ %m = call i8* @malloc(i64 4)
+ %rhs = inttoptr i64 2048 to i8*
+ %cmp = icmp eq i8* %m, %rhs
ret i1 %cmp
}
; CHECK-LABEL: @offset_single_cmp(
; CHECK-NEXT: ret i1 false
;
- %m = call ptr @malloc(i64 4)
- %n = call ptr @malloc(i64 4)
- %rhs = getelementptr i8, ptr %n, i32 4
- %cmp = icmp eq ptr %m, %rhs
+ %m = call i8* @malloc(i64 4)
+ %n = call i8* @malloc(i64 4)
+ %rhs = getelementptr i8, i8* %n, i32 4
+ %cmp = icmp eq i8* %m, %rhs
ret i1 %cmp
}
define void @neg_consistent_fold1() {
; CHECK-LABEL: @neg_consistent_fold1(
-; CHECK-NEXT: [[M:%.*]] = call dereferenceable_or_null(4) ptr @malloc(i64 4)
-; CHECK-NEXT: [[RHS2:%.*]] = call ptr @hidden_inttoptr()
-; CHECK-NEXT: [[CMP1:%.*]] = icmp eq ptr [[M]], inttoptr (i64 2048 to ptr)
-; CHECK-NEXT: [[CMP2:%.*]] = icmp eq ptr [[M]], [[RHS2]]
+; CHECK-NEXT: [[M:%.*]] = call dereferenceable_or_null(4) i8* @malloc(i64 4)
+; CHECK-NEXT: [[RHS2:%.*]] = call i8* @hidden_inttoptr()
+; CHECK-NEXT: [[CMP1:%.*]] = icmp eq i8* [[M]], inttoptr (i64 2048 to i8*)
+; CHECK-NEXT: [[CMP2:%.*]] = icmp eq i8* [[M]], [[RHS2]]
; CHECK-NEXT: call void @witness(i1 [[CMP1]], i1 [[CMP2]])
; CHECK-NEXT: ret void
;
- %m = call ptr @malloc(i64 4)
- %rhs = inttoptr i64 2048 to ptr
- %rhs2 = call ptr @hidden_inttoptr()
- %cmp1 = icmp eq ptr %m, %rhs
- %cmp2 = icmp eq ptr %m, %rhs2
+ %m = call i8* @malloc(i64 4)
+ %rhs = inttoptr i64 2048 to i8*
+ %rhs2 = call i8* @hidden_inttoptr()
+ %cmp1 = icmp eq i8* %m, %rhs
+ %cmp2 = icmp eq i8* %m, %rhs2
call void @witness(i1 %cmp1, i1 %cmp2)
ret void
}
define void @neg_consistent_fold2() {
; CHECK-LABEL: @neg_consistent_fold2(
-; CHECK-NEXT: [[M:%.*]] = call dereferenceable_or_null(4) ptr @malloc(i64 4)
-; CHECK-NEXT: [[N:%.*]] = call dereferenceable_or_null(4) ptr @malloc(i64 4)
-; CHECK-NEXT: [[RHS:%.*]] = getelementptr i8, ptr [[N]], i64 4
-; CHECK-NEXT: [[RHS2:%.*]] = call ptr @hidden_offset(ptr [[N]])
-; CHECK-NEXT: [[CMP1:%.*]] = icmp eq ptr [[M]], [[RHS]]
-; CHECK-NEXT: [[CMP2:%.*]] = icmp eq ptr [[M]], [[RHS2]]
+; CHECK-NEXT: [[M:%.*]] = call dereferenceable_or_null(4) i8* @malloc(i64 4)
+; CHECK-NEXT: [[N:%.*]] = call dereferenceable_or_null(4) i8* @malloc(i64 4)
+; CHECK-NEXT: [[RHS:%.*]] = getelementptr i8, i8* [[N]], i64 4
+; CHECK-NEXT: [[RHS2:%.*]] = call i8* @hidden_offset(i8* [[N]])
+; CHECK-NEXT: [[CMP1:%.*]] = icmp eq i8* [[M]], [[RHS]]
+; CHECK-NEXT: [[CMP2:%.*]] = icmp eq i8* [[M]], [[RHS2]]
; CHECK-NEXT: call void @witness(i1 [[CMP1]], i1 [[CMP2]])
; CHECK-NEXT: ret void
;
- %m = call ptr @malloc(i64 4)
- %n = call ptr @malloc(i64 4)
- %rhs = getelementptr i8, ptr %n, i32 4
- %rhs2 = call ptr @hidden_offset(ptr %n)
- %cmp1 = icmp eq ptr %m, %rhs
- %cmp2 = icmp eq ptr %m, %rhs2
+ %m = call i8* @malloc(i64 4)
+ %n = call i8* @malloc(i64 4)
+ %rhs = getelementptr i8, i8* %n, i32 4
+ %rhs2 = call i8* @hidden_offset(i8* %n)
+ %cmp1 = icmp eq i8* %m, %rhs
+ %cmp2 = icmp eq i8* %m, %rhs2
call void @witness(i1 %cmp1, i1 %cmp2)
ret void
}
define void @neg_consistent_fold3() {
; CHECK-LABEL: @neg_consistent_fold3(
-; CHECK-NEXT: [[M:%.*]] = call dereferenceable_or_null(4) ptr @malloc(i64 4)
-; CHECK-NEXT: [[LGP:%.*]] = load ptr, ptr @gp, align 8
-; CHECK-NEXT: [[RHS2:%.*]] = call ptr @hidden_inttoptr()
-; CHECK-NEXT: [[CMP1:%.*]] = icmp eq ptr [[M]], [[LGP]]
-; CHECK-NEXT: [[CMP2:%.*]] = icmp eq ptr [[M]], [[RHS2]]
+; CHECK-NEXT: [[M:%.*]] = call dereferenceable_or_null(4) i8* @malloc(i64 4)
+; CHECK-NEXT: [[BC:%.*]] = bitcast i8* [[M]] to i32*
+; CHECK-NEXT: [[LGP:%.*]] = load i32*, i32** @gp, align 8
+; CHECK-NEXT: [[RHS2:%.*]] = call i8* @hidden_inttoptr()
+; CHECK-NEXT: [[CMP1:%.*]] = icmp eq i32* [[LGP]], [[BC]]
+; CHECK-NEXT: [[CMP2:%.*]] = icmp eq i8* [[M]], [[RHS2]]
; CHECK-NEXT: call void @witness(i1 [[CMP1]], i1 [[CMP2]])
; CHECK-NEXT: ret void
;
- %m = call ptr @malloc(i64 4)
- %lgp = load ptr, ptr @gp, align 8
- %rhs2 = call ptr @hidden_inttoptr()
- %cmp1 = icmp eq ptr %m, %lgp
- %cmp2 = icmp eq ptr %m, %rhs2
+ %m = call i8* @malloc(i64 4)
+ %bc = bitcast i8* %m to i32*
+ %lgp = load i32*, i32** @gp, align 8
+ %rhs2 = call i8* @hidden_inttoptr()
+ %cmp1 = icmp eq i32* %bc, %lgp
+ %cmp2 = icmp eq i8* %m, %rhs2
call void @witness(i1 %cmp1, i1 %cmp2)
ret void
}
; CHECK-NEXT: call void @witness(i1 false, i1 false)
; CHECK-NEXT: ret void
;
- %m = call ptr @malloc(i64 4)
- %lgp = load ptr, ptr @gp, align 8
- %cmp1 = icmp eq ptr %m, %lgp
- %cmp2 = icmp eq ptr %m, %lgp
+ %m = call i8* @malloc(i64 4)
+ %bc = bitcast i8* %m to i32*
+ %lgp = load i32*, i32** @gp, align 8
+ %cmp1 = icmp eq i32* %bc, %lgp
+ %cmp2 = icmp eq i32* %bc, %lgp
call void @witness(i1 %cmp1, i1 %cmp2)
ret void
}
-declare void @unknown(ptr)
+declare void @unknown(i8*)
; Points out that a nocapture call can't cause a consistent result issue
; as it is (by assumption) not able to contain a comparison which might
define i1 @consistent_nocapture_inttoptr() {
; CHECK-LABEL: @consistent_nocapture_inttoptr(
-; CHECK-NEXT: [[M:%.*]] = call dereferenceable_or_null(4) ptr @malloc(i64 4)
-; CHECK-NEXT: call void @unknown(ptr nocapture [[M]])
-; CHECK-NEXT: [[CMP:%.*]] = icmp eq ptr [[M]], inttoptr (i64 2048 to ptr)
+; CHECK-NEXT: [[M:%.*]] = call dereferenceable_or_null(4) i8* @malloc(i64 4)
+; CHECK-NEXT: call void @unknown(i8* nocapture [[M]])
+; CHECK-NEXT: [[CMP:%.*]] = icmp eq i8* [[M]], inttoptr (i64 2048 to i8*)
; CHECK-NEXT: ret i1 [[CMP]]
;
- %m = call ptr @malloc(i64 4)
- call void @unknown(ptr nocapture %m)
- %rhs = inttoptr i64 2048 to ptr
- %cmp = icmp eq ptr %m, %rhs
+ %m = call i8* @malloc(i64 4)
+ call void @unknown(i8* nocapture %m)
+ %rhs = inttoptr i64 2048 to i8*
+ %cmp = icmp eq i8* %m, %rhs
ret i1 %cmp
}
define i1 @consistent_nocapture_offset() {
; CHECK-LABEL: @consistent_nocapture_offset(
-; CHECK-NEXT: [[M:%.*]] = call dereferenceable_or_null(4) ptr @malloc(i64 4)
-; CHECK-NEXT: call void @unknown(ptr nocapture [[M]])
+; CHECK-NEXT: [[M:%.*]] = call dereferenceable_or_null(4) i8* @malloc(i64 4)
+; CHECK-NEXT: call void @unknown(i8* nocapture [[M]])
; CHECK-NEXT: ret i1 false
;
- %m = call ptr @malloc(i64 4)
- call void @unknown(ptr nocapture %m)
- %n = call ptr @malloc(i64 4)
- %rhs = getelementptr i8, ptr %n, i32 4
- %cmp = icmp eq ptr %m, %rhs
+ %m = call i8* @malloc(i64 4)
+ call void @unknown(i8* nocapture %m)
+ %n = call i8* @malloc(i64 4)
+ %rhs = getelementptr i8, i8* %n, i32 4
+ %cmp = icmp eq i8* %m, %rhs
ret i1 %cmp
}
define i1 @consistent_nocapture_through_global() {
; CHECK-LABEL: @consistent_nocapture_through_global(
-; CHECK-NEXT: [[M:%.*]] = call dereferenceable_or_null(4) ptr @malloc(i64 4)
-; CHECK-NEXT: call void @unknown(ptr nocapture [[M]])
+; CHECK-NEXT: [[M:%.*]] = call dereferenceable_or_null(4) i8* @malloc(i64 4)
+; CHECK-NEXT: call void @unknown(i8* nocapture [[M]])
; CHECK-NEXT: ret i1 false
;
- %m = call ptr @malloc(i64 4)
- call void @unknown(ptr nocapture %m)
- %lgp = load ptr, ptr @gp, align 8, !nonnull !0
- %cmp = icmp eq ptr %m, %lgp
+ %m = call i8* @malloc(i64 4)
+ call void @unknown(i8* nocapture %m)
+ %bc = bitcast i8* %m to i32*
+ %lgp = load i32*, i32** @gp, align 8, !nonnull !0
+ %cmp = icmp eq i32* %bc, %lgp
ret i1 %cmp
}
; CHECK-LABEL: @two_nonnull_mallocs(
; CHECK-NEXT: ret i1 false
;
- %m = call nonnull ptr @malloc(i64 4)
- %n = call nonnull ptr @malloc(i64 4)
- %cmp = icmp eq ptr %m, %n
+ %m = call nonnull i8* @malloc(i64 4)
+ %n = call nonnull i8* @malloc(i64 4)
+ %cmp = icmp eq i8* %m, %n
ret i1 %cmp
}
; the comparison non-equal.
define i1 @two_nonnull_mallocs2() {
; CHECK-LABEL: @two_nonnull_mallocs2(
-; CHECK-NEXT: [[N:%.*]] = call nonnull dereferenceable(4) ptr @malloc(i64 4)
-; CHECK-NEXT: call void @unknown(ptr nonnull [[N]])
+; CHECK-NEXT: [[N:%.*]] = call nonnull dereferenceable(4) i8* @malloc(i64 4)
+; CHECK-NEXT: call void @unknown(i8* nonnull [[N]])
; CHECK-NEXT: ret i1 false
;
- %m = call nonnull ptr @malloc(i64 4)
- %n = call nonnull ptr @malloc(i64 4)
- call void @unknown(ptr %n)
- %cmp = icmp eq ptr %m, %n
+ %m = call nonnull i8* @malloc(i64 4)
+ %n = call nonnull i8* @malloc(i64 4)
+ call void @unknown(i8* %n)
+ %cmp = icmp eq i8* %m, %n
ret i1 %cmp
}
; TODO: We can fold this, but don't with the current scheme.
define i1 @two_nonnull_mallocs_hidden() {
; CHECK-LABEL: @two_nonnull_mallocs_hidden(
-; CHECK-NEXT: [[M:%.*]] = call nonnull dereferenceable(4) ptr @malloc(i64 4)
-; CHECK-NEXT: [[N:%.*]] = call nonnull dereferenceable(4) ptr @malloc(i64 4)
-; CHECK-NEXT: [[GEP1:%.*]] = getelementptr i8, ptr [[M]], i64 1
-; CHECK-NEXT: [[GEP2:%.*]] = getelementptr i8, ptr [[N]], i64 2
-; CHECK-NEXT: [[CMP:%.*]] = icmp eq ptr [[GEP1]], [[GEP2]]
+; CHECK-NEXT: [[M:%.*]] = call nonnull dereferenceable(4) i8* @malloc(i64 4)
+; CHECK-NEXT: [[N:%.*]] = call nonnull dereferenceable(4) i8* @malloc(i64 4)
+; CHECK-NEXT: [[GEP1:%.*]] = getelementptr i8, i8* [[M]], i64 1
+; CHECK-NEXT: [[GEP2:%.*]] = getelementptr i8, i8* [[N]], i64 2
+; CHECK-NEXT: [[CMP:%.*]] = icmp eq i8* [[GEP1]], [[GEP2]]
; CHECK-NEXT: ret i1 [[CMP]]
;
- %m = call nonnull ptr @malloc(i64 4)
- %n = call nonnull ptr @malloc(i64 4)
- %gep1 = getelementptr i8, ptr %m, i32 1
- %gep2 = getelementptr i8, ptr %n, i32 2
- %cmp = icmp eq ptr %gep1, %gep2
+ %m = call nonnull i8* @malloc(i64 4)
+ %n = call nonnull i8* @malloc(i64 4)
+ %gep1 = getelementptr i8, i8* %m, i32 1
+ %gep2 = getelementptr i8, i8* %n, i32 2
+ %cmp = icmp eq i8* %gep1, %gep2
ret i1 %cmp
}
%myStruct = type { float, [3 x float], [4 x float], i32 }
; make sure that we are not crashing when creating an illegal type
-define void @func(ptr addrspace(1) nocapture %p) nounwind {
+define void @func(%myStruct addrspace(1)* nocapture %p) nounwind {
; CHECK-LABEL: @func(
; CHECK-NEXT: ret void
;
- %B = addrspacecast ptr addrspace(1) %p to ptr
- %C = getelementptr inbounds %myStruct, ptr %B, i32 0, i32 1
- %D = getelementptr inbounds [3 x float], ptr %C, i32 0, i32 2
- %E = load float, ptr %D, align 4
+ %A = getelementptr inbounds %myStruct, %myStruct addrspace(1)* %p, i64 0
+ %B = addrspacecast %myStruct addrspace(1)* %A to %myStruct*
+ %C = getelementptr inbounds %myStruct, %myStruct* %B, i32 0, i32 1
+ %D = getelementptr inbounds [3 x float], [3 x float]* %C, i32 0, i32 2
+ %E = load float, float* %D, align 4
%F = fsub float %E, undef
ret void
}
@array = internal addrspace(3) global [256 x float] zeroinitializer, align 4
@scalar = internal addrspace(3) global float 0.000000e+00, align 4
-define void @keep_necessary_addrspacecast(i64 %i, ptr %out0, ptr %out1) {
+define void @keep_necessary_addrspacecast(i64 %i, float** %out0, float** %out1) {
; CHECK-LABEL: @keep_necessary_addrspacecast(
-; CHECK-NEXT: [[T0:%.*]] = getelementptr [256 x float], ptr addrspacecast (ptr addrspace(3) @array to ptr), i64 0, i64 [[I:%.*]]
-; CHECK-NEXT: [[T1:%.*]] = getelementptr [0 x float], ptr addrspacecast (ptr addrspace(3) @scalar to ptr), i64 0, i64 [[I]]
-; CHECK-NEXT: store ptr [[T0]], ptr [[OUT0:%.*]], align 4
-; CHECK-NEXT: store ptr [[T1]], ptr [[OUT1:%.*]], align 4
+; CHECK-NEXT: [[T01:%.*]] = getelementptr [256 x float], [256 x float] addrspace(3)* @array, i64 0, i64 [[I:%.*]]
+; CHECK-NEXT: [[T0:%.*]] = addrspacecast float addrspace(3)* [[T01]] to float*
+; CHECK-NEXT: [[TMP1:%.*]] = getelementptr float, float addrspace(3)* @scalar, i64 [[I]]
+; CHECK-NEXT: [[T1:%.*]] = addrspacecast float addrspace(3)* [[TMP1]] to float*
+; CHECK-NEXT: store float* [[T0]], float** [[OUT0:%.*]], align 4
+; CHECK-NEXT: store float* [[T1]], float** [[OUT1:%.*]], align 4
; CHECK-NEXT: ret void
;
- %t0 = getelementptr [256 x float], ptr addrspacecast (ptr addrspace(3) @array to ptr), i64 0, i64 %i
- %t1 = getelementptr [0 x float], ptr addrspacecast (ptr addrspace(3) @scalar to ptr), i64 0, i64 %i
- store ptr %t0, ptr %out0, align 4
- store ptr %t1, ptr %out1, align 4
+ %t0 = getelementptr [256 x float], [256 x float]* addrspacecast ([256 x float] addrspace(3)* @array to [256 x float]*), i64 0, i64 %i
+ %t1 = getelementptr [0 x float], [0 x float]* addrspacecast (float addrspace(3)* @scalar to [0 x float]*), i64 0, i64 %i
+ store float* %t0, float** %out0, align 4
+ store float* %t1, float** %out1, align 4
ret void
}
-declare void @escape_alloca(ptr)
+declare void @escape_alloca(i16*)
; check that addrspacecast is stripped when trying to mark a GEP as inbounds
define { i8, i8 } @inbounds_after_addrspacecast() {
; CHECK-LABEL: @inbounds_after_addrspacecast(
; CHECK-NEXT: [[T0:%.*]] = alloca i16, align 2
-; CHECK-NEXT: call void @escape_alloca(ptr nonnull [[T0]])
-; CHECK-NEXT: [[T1:%.*]] = addrspacecast ptr [[T0]] to ptr addrspace(11)
-; CHECK-NEXT: [[T2:%.*]] = getelementptr inbounds [2 x i8], ptr addrspace(11) [[T1]], i64 0, i64 1
-; CHECK-NEXT: [[T3:%.*]] = load i8, ptr addrspace(11) [[T2]], align 1
+; CHECK-NEXT: call void @escape_alloca(i16* nonnull [[T0]])
+; CHECK-NEXT: [[TMPCAST:%.*]] = bitcast i16* [[T0]] to [2 x i8]*
+; CHECK-NEXT: [[T1:%.*]] = addrspacecast [2 x i8]* [[TMPCAST]] to [2 x i8] addrspace(11)*
+; CHECK-NEXT: [[T2:%.*]] = getelementptr inbounds [2 x i8], [2 x i8] addrspace(11)* [[T1]], i64 0, i64 1
+; CHECK-NEXT: [[T3:%.*]] = load i8, i8 addrspace(11)* [[T2]], align 1
; CHECK-NEXT: [[INSERT:%.*]] = insertvalue { i8, i8 } zeroinitializer, i8 [[T3]], 1
; CHECK-NEXT: ret { i8, i8 } [[INSERT]]
;
%t0 = alloca i16, align 2
- call void @escape_alloca(ptr %t0)
- %t1 = addrspacecast ptr %t0 to ptr addrspace(11)
- %t2 = getelementptr [2 x i8], ptr addrspace(11) %t1, i64 0, i64 1
- %t3 = load i8, ptr addrspace(11) %t2, align 1
+ call void @escape_alloca(i16* %t0)
+ %tmpcast = bitcast i16* %t0 to [2 x i8]*
+ %t1 = addrspacecast [2 x i8]* %tmpcast to [2 x i8] addrspace(11)*
+ %t2 = getelementptr [2 x i8], [2 x i8] addrspace(11)* %t1, i64 0, i64 1
+ %t3 = load i8, i8 addrspace(11)* %t2, align 1
%insert = insertvalue { i8, i8 } zeroinitializer, i8 %t3, 1
ret { i8, i8 } %insert
}
declare spir_func <16 x i32> @my_extern_func()
; check that a bitcast is not generated when we need an addrspace cast
-define void @bitcast_after_gep(ptr %t0) {
+define void @bitcast_after_gep(<16 x i32>* %t0) {
; CHECK-LABEL: @bitcast_after_gep(
-; CHECK-NEXT: [[T2:%.*]] = addrspacecast ptr [[T0:%.*]] to ptr addrspace(3)
+; CHECK-NEXT: [[T4:%.*]] = addrspacecast <16 x i32>* [[T0:%.*]] to <16 x i32> addrspace(3)*
; CHECK-NEXT: [[CALL:%.*]] = call spir_func <16 x i32> @my_extern_func()
-; CHECK-NEXT: store <16 x i32> [[CALL]], ptr addrspace(3) [[T2]], align 64
+; CHECK-NEXT: store <16 x i32> [[CALL]], <16 x i32> addrspace(3)* [[T4]], align 64
; CHECK-NEXT: ret void
;
- %t2 = addrspacecast ptr %t0 to ptr addrspace(3)
+ %t1 = bitcast <16 x i32>* %t0 to [16 x i32]*
+ %t2 = addrspacecast [16 x i32]* %t1 to [16 x i32] addrspace(3)*
+ %t3 = getelementptr inbounds [16 x i32], [16 x i32] addrspace(3)* %t2, i64 0, i64 0
+ %t4 = bitcast i32 addrspace(3)* %t3 to <16 x i32> addrspace(3)*
%call = call spir_func <16 x i32> @my_extern_func()
- store <16 x i32> %call, ptr addrspace(3) %t2
+ store <16 x i32> %call, <16 x i32> addrspace(3)* %t4
ret void
}
@Global = external global [10 x i8]
; Test that two array indexing geps fold
-define ptr @test1(ptr %I) {
+define i32* @test1(i32* %I) {
; CHECK-LABEL: @test1(
-; CHECK-NEXT: [[B:%.*]] = getelementptr i32, ptr [[I:%.*]], i32 21
-; CHECK-NEXT: ret ptr [[B]]
+; CHECK-NEXT: [[B:%.*]] = getelementptr i32, i32* [[I:%.*]], i32 21
+; CHECK-NEXT: ret i32* [[B]]
;
- %A = getelementptr i32, ptr %I, i8 17
- %B = getelementptr i32, ptr %A, i16 4
- ret ptr %B
+ %A = getelementptr i32, i32* %I, i8 17
+ %B = getelementptr i32, i32* %A, i16 4
+ ret i32* %B
}
; Test that two getelementptr insts fold
-define ptr @test2(ptr %I) {
+define i32* @test2({ i32 }* %I) {
; CHECK-LABEL: @test2(
-; CHECK-NEXT: [[A:%.*]] = getelementptr { i32 }, ptr [[I:%.*]], i32 1
-; CHECK-NEXT: ret ptr [[A]]
+; CHECK-NEXT: [[B:%.*]] = getelementptr { i32 }, { i32 }* [[I:%.*]], i32 1, i32 0
+; CHECK-NEXT: ret i32* [[B]]
;
- %A = getelementptr { i32 }, ptr %I, i32 1
- ret ptr %A
+ %A = getelementptr { i32 }, { i32 }* %I, i32 1
+ %B = getelementptr { i32 }, { i32 }* %A, i32 0, i32 0
+ ret i32* %B
}
define void @test3(i8 %B) {
; This should be turned into a constexpr instead of being an instruction
; CHECK-LABEL: @test3(
-; CHECK-NEXT: store i8 [[B:%.*]], ptr getelementptr inbounds ([10 x i8], ptr @Global, i32 0, i32 4), align 1
+; CHECK-NEXT: store i8 [[B:%.*]], i8* getelementptr inbounds ([10 x i8], [10 x i8]* @Global, i32 0, i32 4), align 1
; CHECK-NEXT: ret void
;
- %A = getelementptr [10 x i8], ptr @Global, i32 0, i32 4
- store i8 %B, ptr %A
+ %A = getelementptr [10 x i8], [10 x i8]* @Global, i32 0, i32 4
+ store i8 %B, i8* %A
ret void
}
-%as1_ptr_struct = type { ptr addrspace(1) }
-%as2_ptr_struct = type { ptr addrspace(2) }
+%as1_ptr_struct = type { i32 addrspace(1)* }
+%as2_ptr_struct = type { i32 addrspace(2)* }
@global_as2 = addrspace(2) global i32 zeroinitializer
-@global_as1_as2_ptr = addrspace(1) global %as2_ptr_struct { ptr addrspace(2) @global_as2 }
+@global_as1_as2_ptr = addrspace(1) global %as2_ptr_struct { i32 addrspace(2)* @global_as2 }
; This should be turned into a constexpr instead of being an instruction
-define void @test_evaluate_gep_nested_as_ptrs(ptr addrspace(2) %B) {
+define void @test_evaluate_gep_nested_as_ptrs(i32 addrspace(2)* %B) {
; CHECK-LABEL: @test_evaluate_gep_nested_as_ptrs(
-; CHECK-NEXT: store ptr addrspace(2) [[B:%.*]], ptr addrspace(1) @global_as1_as2_ptr, align 8
+; CHECK-NEXT: store i32 addrspace(2)* [[B:%.*]], i32 addrspace(2)* addrspace(1)* getelementptr inbounds ([[AS2_PTR_STRUCT:%.*]], [[AS2_PTR_STRUCT]] addrspace(1)* @global_as1_as2_ptr, i32 0, i32 0), align 8
; CHECK-NEXT: ret void
;
- store ptr addrspace(2) %B, ptr addrspace(1) @global_as1_as2_ptr
+ %A = getelementptr %as2_ptr_struct, %as2_ptr_struct addrspace(1)* @global_as1_as2_ptr, i32 0, i32 0
+ store i32 addrspace(2)* %B, i32 addrspace(2)* addrspace(1)* %A
ret void
}
-@arst = addrspace(1) global [4 x ptr addrspace(2)] zeroinitializer
+@arst = addrspace(1) global [4 x i8 addrspace(2)*] zeroinitializer
-define void @test_evaluate_gep_as_ptrs_array(ptr addrspace(2) %B) {
+define void @test_evaluate_gep_as_ptrs_array(i8 addrspace(2)* %B) {
; CHECK-LABEL: @test_evaluate_gep_as_ptrs_array(
-; CHECK-NEXT: store ptr addrspace(2) [[B:%.*]], ptr addrspace(1) getelementptr inbounds ([4 x ptr addrspace(2)], ptr addrspace(1) @arst, i32 0, i32 2), align 16
+; CHECK-NEXT: store i8 addrspace(2)* [[B:%.*]], i8 addrspace(2)* addrspace(1)* getelementptr inbounds ([4 x i8 addrspace(2)*], [4 x i8 addrspace(2)*] addrspace(1)* @arst, i32 0, i32 2), align 16
; CHECK-NEXT: ret void
;
- %A = getelementptr [4 x ptr addrspace(2)], ptr addrspace(1) @arst, i16 0, i16 2
- store ptr addrspace(2) %B, ptr addrspace(1) %A
+ %A = getelementptr [4 x i8 addrspace(2)*], [4 x i8 addrspace(2)*] addrspace(1)* @arst, i16 0, i16 2
+ store i8 addrspace(2)* %B, i8 addrspace(2)* addrspace(1)* %A
ret void
}
-define ptr @test4(ptr %I, i32 %C, i32 %D) {
+define i32* @test4(i32* %I, i32 %C, i32 %D) {
; CHECK-LABEL: @test4(
-; CHECK-NEXT: [[A:%.*]] = getelementptr i32, ptr [[I:%.*]], i32 [[C:%.*]]
-; CHECK-NEXT: [[B:%.*]] = getelementptr i32, ptr [[A]], i32 [[D:%.*]]
-; CHECK-NEXT: ret ptr [[B]]
+; CHECK-NEXT: [[A:%.*]] = getelementptr i32, i32* [[I:%.*]], i32 [[C:%.*]]
+; CHECK-NEXT: [[B:%.*]] = getelementptr i32, i32* [[A]], i32 [[D:%.*]]
+; CHECK-NEXT: ret i32* [[B]]
;
- %A = getelementptr i32, ptr %I, i32 %C
- %B = getelementptr i32, ptr %A, i32 %D
- ret ptr %B
+ %A = getelementptr i32, i32* %I, i32 %C
+ %B = getelementptr i32, i32* %A, i32 %D
+ ret i32* %B
}
-define i1 @test5(ptr %x, ptr %y) {
+define i1 @test5({ i32, i32 }* %x, { i32, i32 }* %y) {
; CHECK-LABEL: @test5(
-; CHECK-NEXT: [[TMP_4:%.*]] = icmp eq ptr [[X:%.*]], [[Y:%.*]]
+; CHECK-NEXT: [[TMP_4:%.*]] = icmp eq { i32, i32 }* [[X:%.*]], [[Y:%.*]]
; CHECK-NEXT: ret i1 [[TMP_4]]
;
- %tmp.1 = getelementptr { i32, i32 }, ptr %x, i32 0, i32 1
- %tmp.3 = getelementptr { i32, i32 }, ptr %y, i32 0, i32 1
+ %tmp.1 = getelementptr { i32, i32 }, { i32, i32 }* %x, i32 0, i32 1
+ %tmp.3 = getelementptr { i32, i32 }, { i32, i32 }* %y, i32 0, i32 1
;; seteq x, y
- %tmp.4 = icmp eq ptr %tmp.1, %tmp.3
+ %tmp.4 = icmp eq i32* %tmp.1, %tmp.3
ret i1 %tmp.4
}
%S = type { i32, [ 100 x i32] }
-define <2 x i1> @test6(<2 x i32> %X, <2 x ptr> %P) nounwind {
+define <2 x i1> @test6(<2 x i32> %X, <2 x %S*> %P) nounwind {
; CHECK-LABEL: @test6(
-; CHECK-NEXT: [[A:%.*]] = getelementptr inbounds [[S:%.*]], <2 x ptr> [[P:%.*]], <2 x i32> zeroinitializer, <2 x i32> <i32 1, i32 1>, <2 x i32> [[X:%.*]]
-; CHECK-NEXT: [[C:%.*]] = icmp eq <2 x ptr> [[A]], [[P]]
+; CHECK-NEXT: [[C:%.*]] = icmp eq <2 x i32> [[X:%.*]], <i32 -1, i32 -1>
; CHECK-NEXT: ret <2 x i1> [[C]]
;
- %A = getelementptr inbounds %S, <2 x ptr> %P, <2 x i32> zeroinitializer, <2 x i32> <i32 1, i32 1>, <2 x i32> %X
- %B = getelementptr inbounds %S, <2 x ptr> %P, <2 x i32> <i32 0, i32 0>, <2 x i32> <i32 0, i32 0>
- %C = icmp eq <2 x ptr> %A, %B
+ %A = getelementptr inbounds %S, <2 x %S*> %P, <2 x i32> zeroinitializer, <2 x i32> <i32 1, i32 1>, <2 x i32> %X
+ %B = getelementptr inbounds %S, <2 x %S*> %P, <2 x i32> <i32 0, i32 0>, <2 x i32> <i32 0, i32 0>
+ %C = icmp eq <2 x i32*> %A, %B
ret <2 x i1> %C
}
; Same as above, but indices scalarized.
-define <2 x i1> @test6b(<2 x i32> %X, <2 x ptr> %P) nounwind {
+define <2 x i1> @test6b(<2 x i32> %X, <2 x %S*> %P) nounwind {
; CHECK-LABEL: @test6b(
-; CHECK-NEXT: [[A:%.*]] = getelementptr inbounds [[S:%.*]], <2 x ptr> [[P:%.*]], i32 0, i32 1, <2 x i32> [[X:%.*]]
-; CHECK-NEXT: [[C:%.*]] = icmp eq <2 x ptr> [[A]], [[P]]
+; CHECK-NEXT: [[C:%.*]] = icmp eq <2 x i32> [[X:%.*]], <i32 -1, i32 -1>
; CHECK-NEXT: ret <2 x i1> [[C]]
;
- %A = getelementptr inbounds %S, <2 x ptr> %P, i32 0, i32 1, <2 x i32> %X
- %B = getelementptr inbounds %S, <2 x ptr> %P, i32 0, i32 0
- %C = icmp eq <2 x ptr> %A, %B
+ %A = getelementptr inbounds %S, <2 x %S*> %P, i32 0, i32 1, <2 x i32> %X
+ %B = getelementptr inbounds %S, <2 x %S*> %P, i32 0, i32 0
+ %C = icmp eq <2 x i32*> %A, %B
ret <2 x i1> %C
}
@G = external global [3 x i8]
-define ptr @test7(i16 %Idx) {
+define i8* @test7(i16 %Idx) {
; CHECK-LABEL: @test7(
; CHECK-NEXT: [[ZE_IDX:%.*]] = zext i16 [[IDX:%.*]] to i32
-; CHECK-NEXT: [[TMP:%.*]] = getelementptr i8, ptr @G, i32 [[ZE_IDX]]
-; CHECK-NEXT: ret ptr [[TMP]]
+; CHECK-NEXT: [[TMP:%.*]] = getelementptr [3 x i8], [3 x i8]* @G, i32 0, i32 [[ZE_IDX]]
+; CHECK-NEXT: ret i8* [[TMP]]
;
%ZE_Idx = zext i16 %Idx to i32
- %tmp = getelementptr i8, ptr @G, i32 %ZE_Idx
- ret ptr %tmp
+ %tmp = getelementptr i8, i8* getelementptr ([3 x i8], [3 x i8]* @G, i32 0, i32 0), i32 %ZE_Idx
+ ret i8* %tmp
}
; Test folding of constantexpr geps into normal geps.
@Array = external global [40 x i32]
-define ptr @test8(i32 %X) {
+define i32 *@test8(i32 %X) {
; CHECK-LABEL: @test8(
-; CHECK-NEXT: [[A:%.*]] = getelementptr i32, ptr @Array, i32 [[X:%.*]]
-; CHECK-NEXT: ret ptr [[A]]
+; CHECK-NEXT: [[A:%.*]] = getelementptr [40 x i32], [40 x i32]* @Array, i32 0, i32 [[X:%.*]]
+; CHECK-NEXT: ret i32* [[A]]
;
- %A = getelementptr i32, ptr @Array, i32 %X
- ret ptr %A
+ %A = getelementptr i32, i32* getelementptr ([40 x i32], [40 x i32]* @Array, i32 0, i32 0), i32 %X
+ ret i32* %A
}
-define ptr @test9(ptr %base, i8 %ind) {
+define i32 *@test9(i32 *%base, i8 %ind) {
; CHECK-LABEL: @test9(
; CHECK-NEXT: [[TMP1:%.*]] = sext i8 [[IND:%.*]] to i32
-; CHECK-NEXT: [[RES:%.*]] = getelementptr i32, ptr [[BASE:%.*]], i32 [[TMP1]]
-; CHECK-NEXT: ret ptr [[RES]]
+; CHECK-NEXT: [[RES:%.*]] = getelementptr i32, i32* [[BASE:%.*]], i32 [[TMP1]]
+; CHECK-NEXT: ret i32* [[RES]]
;
- %res = getelementptr i32, ptr %base, i8 %ind
- ret ptr %res
+ %res = getelementptr i32, i32 *%base, i8 %ind
+ ret i32* %res
}
define i32 @test10() {
; CHECK-LABEL: @test10(
; CHECK-NEXT: ret i32 8
;
- %A = getelementptr { i32, double }, ptr null, i32 0, i32 1
- %B = ptrtoint ptr %A to i32
+ %A = getelementptr { i32, double }, { i32, double }* null, i32 0, i32 1
+ %B = ptrtoint double* %A to i32
ret i32 %B
}
define i16 @constant_fold_custom_dl() {
; CHECK-LABEL: @constant_fold_custom_dl(
; CHECK-NEXT: entry:
-; CHECK-NEXT: ret i16 ptrtoint (ptr addrspace(1) getelementptr (i8, ptr addrspace(1) getelementptr inbounds ([1000 x i8], ptr addrspace(1) @X_as1, i32 1, i32 0), i32 sext (i16 sub (i16 0, i16 ptrtoint (ptr addrspace(1) @X_as1 to i16)) to i32)) to i16)
+; CHECK-NEXT: ret i16 ptrtoint (i8 addrspace(1)* getelementptr (i8, i8 addrspace(1)* getelementptr inbounds ([1000 x i8], [1000 x i8] addrspace(1)* @X_as1, i32 1, i32 0), i32 sext (i16 sub (i16 0, i16 ptrtoint ([1000 x i8] addrspace(1)* @X_as1 to i16)) to i32)) to i16)
;
entry:
- %A = bitcast ptr addrspace(1) getelementptr inbounds ([1000 x i8], ptr addrspace(1) @X_as1, i64 1, i64 0) to ptr addrspace(1)
+ %A = bitcast i8 addrspace(1)* getelementptr inbounds ([1000 x i8], [1000 x i8] addrspace(1)* @X_as1, i64 1, i64 0) to i8 addrspace(1)*
+ %B = bitcast i8 addrspace(1)* getelementptr inbounds ([1000 x i8], [1000 x i8] addrspace(1)* @X_as1, i64 0, i64 0) to i8 addrspace(1)*
- %B2 = ptrtoint ptr addrspace(1) @X_as1 to i16
+ %B2 = ptrtoint i8 addrspace(1)* %B to i16
%C = sub i16 0, %B2
- %D = getelementptr i8, ptr addrspace(1) %A, i16 %C
- %E = ptrtoint ptr addrspace(1) %D to i16
+ %D = getelementptr i8, i8 addrspace(1)* %A, i16 %C
+ %E = ptrtoint i8 addrspace(1)* %D to i16
ret i16 %E
}
@block = global [64 x [8192 x i8]] zeroinitializer, align 1
-define <2 x ptr> @vectorindex1() {
+define <2 x i8*> @vectorindex1() {
; CHECK-LABEL: @vectorindex1(
-; CHECK-NEXT: ret <2 x ptr> getelementptr inbounds ([64 x [8192 x i8]], ptr @block, <2 x i64> zeroinitializer, <2 x i64> <i64 1, i64 2>, <2 x i64> zeroinitializer)
+; CHECK-NEXT: ret <2 x i8*> getelementptr inbounds ([64 x [8192 x i8]], [64 x [8192 x i8]]* @block, <2 x i64> zeroinitializer, <2 x i64> <i64 1, i64 2>, <2 x i64> zeroinitializer)
;
- %1 = getelementptr inbounds [64 x [8192 x i8]], ptr @block, i64 0, <2 x i64> <i64 0, i64 1>, i64 8192
- ret <2 x ptr> %1
+ %1 = getelementptr inbounds [64 x [8192 x i8]], [64 x [8192 x i8]]* @block, i64 0, <2 x i64> <i64 0, i64 1>, i64 8192
+ ret <2 x i8*> %1
}
-define <2 x ptr> @vectorindex2() {
+define <2 x i8*> @vectorindex2() {
; CHECK-LABEL: @vectorindex2(
-; CHECK-NEXT: ret <2 x ptr> getelementptr inbounds ([64 x [8192 x i8]], ptr @block, <2 x i64> zeroinitializer, <2 x i64> <i64 1, i64 2>, <2 x i64> <i64 8191, i64 1>)
+; CHECK-NEXT: ret <2 x i8*> getelementptr inbounds ([64 x [8192 x i8]], [64 x [8192 x i8]]* @block, <2 x i64> zeroinitializer, <2 x i64> <i64 1, i64 2>, <2 x i64> <i64 8191, i64 1>)
;
- %1 = getelementptr inbounds [64 x [8192 x i8]], ptr @block, i64 0, i64 1, <2 x i64> <i64 8191, i64 8193>
- ret <2 x ptr> %1
+ %1 = getelementptr inbounds [64 x [8192 x i8]], [64 x [8192 x i8]]* @block, i64 0, i64 1, <2 x i64> <i64 8191, i64 8193>
+ ret <2 x i8*> %1
}
-define <2 x ptr> @vectorindex3() {
+define <2 x i8*> @vectorindex3() {
; CHECK-LABEL: @vectorindex3(
-; CHECK-NEXT: ret <2 x ptr> getelementptr inbounds ([64 x [8192 x i8]], ptr @block, <2 x i64> zeroinitializer, <2 x i64> <i64 0, i64 2>, <2 x i64> <i64 8191, i64 1>)
+; CHECK-NEXT: ret <2 x i8*> getelementptr inbounds ([64 x [8192 x i8]], [64 x [8192 x i8]]* @block, <2 x i64> zeroinitializer, <2 x i64> <i64 0, i64 2>, <2 x i64> <i64 8191, i64 1>)
;
- %1 = getelementptr inbounds [64 x [8192 x i8]], ptr @block, i64 0, <2 x i64> <i64 0, i64 1>, <2 x i64> <i64 8191, i64 8193>
- ret <2 x ptr> %1
+ %1 = getelementptr inbounds [64 x [8192 x i8]], [64 x [8192 x i8]]* @block, i64 0, <2 x i64> <i64 0, i64 1>, <2 x i64> <i64 8191, i64 8193>
+ ret <2 x i8*> %1
}
; Negative test - datalayout's alloc size for the 2 types must match.
-define ptr @bitcast_vec_to_array_gep(ptr %x, i64 %y, i64 %z) {
+define i32* @bitcast_vec_to_array_gep(<7 x i32>* %x, i64 %y, i64 %z) {
; CHECK-LABEL: @bitcast_vec_to_array_gep(
-; CHECK-NEXT: [[GEP:%.*]] = getelementptr [7 x i32], ptr [[X:%.*]], i64 [[Y:%.*]], i64 [[Z:%.*]]
-; CHECK-NEXT: ret ptr [[GEP]]
+; CHECK-NEXT: [[ARR_PTR:%.*]] = bitcast <7 x i32>* [[X:%.*]] to [7 x i32]*
+; CHECK-NEXT: [[GEP:%.*]] = getelementptr [7 x i32], [7 x i32]* [[ARR_PTR]], i64 [[Y:%.*]], i64 [[Z:%.*]]
+; CHECK-NEXT: ret i32* [[GEP]]
;
- %gep = getelementptr [7 x i32], ptr %x, i64 %y, i64 %z
- ret ptr %gep
+ %arr_ptr = bitcast <7 x i32>* %x to [7 x i32]*
+ %gep = getelementptr [7 x i32], [7 x i32]* %arr_ptr, i64 %y, i64 %z
+ ret i32* %gep
}
; Negative test - datalayout's alloc size for the 2 types must match.
-define ptr @bitcast_array_to_vec_gep(ptr %x, i64 %y, i64 %z) {
+define i32* @bitcast_array_to_vec_gep([3 x i32]* %x, i64 %y, i64 %z) {
; CHECK-LABEL: @bitcast_array_to_vec_gep(
-; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds <3 x i32>, ptr [[X:%.*]], i64 [[Y:%.*]], i64 [[Z:%.*]]
-; CHECK-NEXT: ret ptr [[GEP]]
+; CHECK-NEXT: [[VEC_PTR:%.*]] = bitcast [3 x i32]* [[X:%.*]] to <3 x i32>*
+; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds <3 x i32>, <3 x i32>* [[VEC_PTR]], i64 [[Y:%.*]], i64 [[Z:%.*]]
+; CHECK-NEXT: ret i32* [[GEP]]
;
- %gep = getelementptr inbounds <3 x i32>, ptr %x, i64 %y, i64 %z
- ret ptr %gep
+ %vec_ptr = bitcast [3 x i32]* %x to <3 x i32>*
+ %gep = getelementptr inbounds <3 x i32>, <3 x i32>* %vec_ptr, i64 %y, i64 %z
+ ret i32* %gep
}
; Sizes and types match - safe to remove bitcast.
-define ptr @bitcast_vec_to_array_gep_matching_alloc_size(ptr %x, i64 %y, i64 %z) {
+define i32* @bitcast_vec_to_array_gep_matching_alloc_size(<4 x i32>* %x, i64 %y, i64 %z) {
; CHECK-LABEL: @bitcast_vec_to_array_gep_matching_alloc_size(
-; CHECK-NEXT: [[GEP:%.*]] = getelementptr [4 x i32], ptr [[X:%.*]], i64 [[Y:%.*]], i64 [[Z:%.*]]
-; CHECK-NEXT: ret ptr [[GEP]]
+; CHECK-NEXT: [[GEP:%.*]] = getelementptr <4 x i32>, <4 x i32>* [[X:%.*]], i64 [[Y:%.*]], i64 [[Z:%.*]]
+; CHECK-NEXT: ret i32* [[GEP]]
;
- %gep = getelementptr [4 x i32], ptr %x, i64 %y, i64 %z
- ret ptr %gep
+ %arr_ptr = bitcast <4 x i32>* %x to [4 x i32]*
+ %gep = getelementptr [4 x i32], [4 x i32]* %arr_ptr, i64 %y, i64 %z
+ ret i32* %gep
}
; Sizes and types match - safe to remove bitcast.
-define ptr @bitcast_array_to_vec_gep_matching_alloc_size(ptr %x, i64 %y, i64 %z) {
+define i32* @bitcast_array_to_vec_gep_matching_alloc_size([4 x i32]* %x, i64 %y, i64 %z) {
; CHECK-LABEL: @bitcast_array_to_vec_gep_matching_alloc_size(
-; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds <4 x i32>, ptr [[X:%.*]], i64 [[Y:%.*]], i64 [[Z:%.*]]
-; CHECK-NEXT: ret ptr [[GEP]]
+; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds [4 x i32], [4 x i32]* [[X:%.*]], i64 [[Y:%.*]], i64 [[Z:%.*]]
+; CHECK-NEXT: ret i32* [[GEP]]
;
- %gep = getelementptr inbounds <4 x i32>, ptr %x, i64 %y, i64 %z
- ret ptr %gep
+ %vec_ptr = bitcast [4 x i32]* %x to <4 x i32>*
+ %gep = getelementptr inbounds <4 x i32>, <4 x i32>* %vec_ptr, i64 %y, i64 %z
+ ret i32* %gep
}
; Negative test - datalayout's alloc size for the 2 types must match.
-define ptr addrspace(3) @bitcast_vec_to_array_addrspace(ptr %x, i64 %y, i64 %z) {
+define i32 addrspace(3)* @bitcast_vec_to_array_addrspace(<7 x i32>* %x, i64 %y, i64 %z) {
; CHECK-LABEL: @bitcast_vec_to_array_addrspace(
-; CHECK-NEXT: [[ASC:%.*]] = addrspacecast ptr [[X:%.*]] to ptr addrspace(3)
-; CHECK-NEXT: [[GEP:%.*]] = getelementptr [7 x i32], ptr addrspace(3) [[ASC]], i64 [[Y:%.*]], i64 [[Z:%.*]]
-; CHECK-NEXT: ret ptr addrspace(3) [[GEP]]
+; CHECK-NEXT: [[ARR_PTR:%.*]] = bitcast <7 x i32>* [[X:%.*]] to [7 x i32]*
+; CHECK-NEXT: [[ASC:%.*]] = addrspacecast [7 x i32]* [[ARR_PTR]] to [7 x i32] addrspace(3)*
+; CHECK-NEXT: [[GEP:%.*]] = getelementptr [7 x i32], [7 x i32] addrspace(3)* [[ASC]], i64 [[Y:%.*]], i64 [[Z:%.*]]
+; CHECK-NEXT: ret i32 addrspace(3)* [[GEP]]
;
- %asc = addrspacecast ptr %x to ptr addrspace(3)
- %gep = getelementptr [7 x i32], ptr addrspace(3) %asc, i64 %y, i64 %z
- ret ptr addrspace(3) %gep
+ %arr_ptr = bitcast <7 x i32>* %x to [7 x i32]*
+ %asc = addrspacecast [7 x i32]* %arr_ptr to [7 x i32] addrspace(3)*
+ %gep = getelementptr [7 x i32], [7 x i32] addrspace(3)* %asc, i64 %y, i64 %z
+ ret i32 addrspace(3)* %gep
}
; Negative test - datalayout's alloc size for the 2 types must match.
-define ptr addrspace(3) @inbounds_bitcast_vec_to_array_addrspace(ptr %x, i64 %y, i64 %z) {
+define i32 addrspace(3)* @inbounds_bitcast_vec_to_array_addrspace(<7 x i32>* %x, i64 %y, i64 %z) {
; CHECK-LABEL: @inbounds_bitcast_vec_to_array_addrspace(
-; CHECK-NEXT: [[ASC:%.*]] = addrspacecast ptr [[X:%.*]] to ptr addrspace(3)
-; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds [7 x i32], ptr addrspace(3) [[ASC]], i64 [[Y:%.*]], i64 [[Z:%.*]]
-; CHECK-NEXT: ret ptr addrspace(3) [[GEP]]
+; CHECK-NEXT: [[ARR_PTR:%.*]] = bitcast <7 x i32>* [[X:%.*]] to [7 x i32]*
+; CHECK-NEXT: [[ASC:%.*]] = addrspacecast [7 x i32]* [[ARR_PTR]] to [7 x i32] addrspace(3)*
+; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds [7 x i32], [7 x i32] addrspace(3)* [[ASC]], i64 [[Y:%.*]], i64 [[Z:%.*]]
+; CHECK-NEXT: ret i32 addrspace(3)* [[GEP]]
;
- %asc = addrspacecast ptr %x to ptr addrspace(3)
- %gep = getelementptr inbounds [7 x i32], ptr addrspace(3) %asc, i64 %y, i64 %z
- ret ptr addrspace(3) %gep
+ %arr_ptr = bitcast <7 x i32>* %x to [7 x i32]*
+ %asc = addrspacecast [7 x i32]* %arr_ptr to [7 x i32] addrspace(3)*
+ %gep = getelementptr inbounds [7 x i32], [7 x i32] addrspace(3)* %asc, i64 %y, i64 %z
+ ret i32 addrspace(3)* %gep
}
; Sizes and types match - safe to remove bitcast.
-define ptr addrspace(3) @bitcast_vec_to_array_addrspace_matching_alloc_size(ptr %x, i64 %y, i64 %z) {
+define i32 addrspace(3)* @bitcast_vec_to_array_addrspace_matching_alloc_size(<4 x i32>* %x, i64 %y, i64 %z) {
; CHECK-LABEL: @bitcast_vec_to_array_addrspace_matching_alloc_size(
-; CHECK-NEXT: [[ASC:%.*]] = addrspacecast ptr [[X:%.*]] to ptr addrspace(3)
-; CHECK-NEXT: [[GEP:%.*]] = getelementptr [4 x i32], ptr addrspace(3) [[ASC]], i64 [[Y:%.*]], i64 [[Z:%.*]]
-; CHECK-NEXT: ret ptr addrspace(3) [[GEP]]
+; CHECK-NEXT: [[GEP:%.*]] = getelementptr <4 x i32>, <4 x i32>* [[X:%.*]], i64 [[Y:%.*]], i64 [[Z:%.*]]
+; CHECK-NEXT: [[TMP1:%.*]] = addrspacecast i32* [[GEP]] to i32 addrspace(3)*
+; CHECK-NEXT: ret i32 addrspace(3)* [[TMP1]]
;
- %asc = addrspacecast ptr %x to ptr addrspace(3)
- %gep = getelementptr [4 x i32], ptr addrspace(3) %asc, i64 %y, i64 %z
- ret ptr addrspace(3) %gep
+ %arr_ptr = bitcast <4 x i32>* %x to [4 x i32]*
+ %asc = addrspacecast [4 x i32]* %arr_ptr to [4 x i32] addrspace(3)*
+ %gep = getelementptr [4 x i32], [4 x i32] addrspace(3)* %asc, i64 %y, i64 %z
+ ret i32 addrspace(3)* %gep
}
; Sizes and types match - safe to remove bitcast.
-define ptr addrspace(3) @inbounds_bitcast_vec_to_array_addrspace_matching_alloc_size(ptr %x, i64 %y, i64 %z) {
+define i32 addrspace(3)* @inbounds_bitcast_vec_to_array_addrspace_matching_alloc_size(<4 x i32>* %x, i64 %y, i64 %z) {
; CHECK-LABEL: @inbounds_bitcast_vec_to_array_addrspace_matching_alloc_size(
-; CHECK-NEXT: [[ASC:%.*]] = addrspacecast ptr [[X:%.*]] to ptr addrspace(3)
-; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds [4 x i32], ptr addrspace(3) [[ASC]], i64 [[Y:%.*]], i64 [[Z:%.*]]
-; CHECK-NEXT: ret ptr addrspace(3) [[GEP]]
+; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds <4 x i32>, <4 x i32>* [[X:%.*]], i64 [[Y:%.*]], i64 [[Z:%.*]]
+; CHECK-NEXT: [[TMP1:%.*]] = addrspacecast i32* [[GEP]] to i32 addrspace(3)*
+; CHECK-NEXT: ret i32 addrspace(3)* [[TMP1]]
;
- %asc = addrspacecast ptr %x to ptr addrspace(3)
- %gep = getelementptr inbounds [4 x i32], ptr addrspace(3) %asc, i64 %y, i64 %z
- ret ptr addrspace(3) %gep
+ %arr_ptr = bitcast <4 x i32>* %x to [4 x i32]*
+ %asc = addrspacecast [4 x i32]* %arr_ptr to [4 x i32] addrspace(3)*
+ %gep = getelementptr inbounds [4 x i32], [4 x i32] addrspace(3)* %asc, i64 %y, i64 %z
+ ret i32 addrspace(3)* %gep
}
-; Negative test - avoid doing bitcast on ptr, because '16' should be scaled by 'vscale'.
+; Negative test - avoid doing bitcast on i8*, because '16' should be scaled by 'vscale'.
-define ptr @test_accumulate_constant_offset_vscale_nonzero(<vscale x 16 x i1> %pg, ptr %base) {
-; CHECK-LABEL: @test_accumulate_constant_offset_vscale_nonzero(
-; CHECK-NEXT: [[GEP:%.*]] = getelementptr <vscale x 16 x i8>, ptr [[BASE:%.*]], i64 1, i64 4
-; CHECK-NEXT: ret ptr [[GEP]]
-;
- %gep = getelementptr <vscale x 16 x i8>, ptr %base, i64 1, i64 4
- ret ptr %gep
+define i8* @test_accumulate_constant_offset_vscale_nonzero(<vscale x 16 x i1> %pg, i8* %base) {
+; CHECK-LABEL: @test_accumulate_constant_offset_vscale_nonzero
+; CHECK-NEXT: %bc = bitcast i8* %base to <vscale x 16 x i8>*
+; CHECK-NEXT: %gep = getelementptr <vscale x 16 x i8>, <vscale x 16 x i8>* %bc, i64 1, i64 4
+; CHECK-NEXT: ret i8* %gep
+ %bc = bitcast i8* %base to <vscale x 16 x i8>*
+ %gep = getelementptr <vscale x 16 x i8>, <vscale x 16 x i8>* %bc, i64 1, i64 4
+ ret i8* %gep
}
-define ptr @test_accumulate_constant_offset_vscale_zero(<vscale x 16 x i1> %pg, ptr %base) {
-; CHECK-LABEL: @test_accumulate_constant_offset_vscale_zero(
-; CHECK-NEXT: [[GEP:%.*]] = getelementptr <vscale x 16 x i8>, ptr [[BASE:%.*]], i64 0, i64 4
-; CHECK-NEXT: ret ptr [[GEP]]
-;
- %gep = getelementptr <vscale x 16 x i8>, ptr %base, i64 0, i64 4
- ret ptr %gep
+define i8* @test_accumulate_constant_offset_vscale_zero(<vscale x 16 x i1> %pg, i8* %base) {
+; CHECK-LABEL: @test_accumulate_constant_offset_vscale_zero
+; CHECK-NEXT: %[[RES:.*]] = getelementptr i8, i8* %base, i64 4
+; CHECK-NEXT: ret i8* %[[RES]]
+ %bc = bitcast i8* %base to <vscale x 16 x i8>*
+ %gep = getelementptr <vscale x 16 x i8>, <vscale x 16 x i8>* %bc, i64 0, i64 4
+ ret i8* %gep
}
; NOTE: Assertions have been autogenerated by utils/update_test_checks.py
; RUN: opt -passes=instcombine -instcombine-infinite-loop-threshold=3 -S < %s | FileCheck %s
-%struct1 = type { ptr, i32, i32, i32 }
+%struct1 = type { %struct2*, i32, i32, i32 }
%struct2 = type { i32, i32 }
%struct3 = type { i32, %struct4, %struct4 }
%struct4 = type { %struct2, %struct2 }
-define i32 @test1(ptr %dm, i1 %tmp4, i64 %tmp9, i64 %tmp19) {
+define i32 @test1(%struct1* %dm, i1 %tmp4, i64 %tmp9, i64 %tmp19) {
; CHECK-LABEL: @test1(
; CHECK-NEXT: bb:
-; CHECK-NEXT: [[TMP1:%.*]] = load ptr, ptr [[DM:%.*]], align 8
+; CHECK-NEXT: [[TMP:%.*]] = getelementptr inbounds [[STRUCT1:%.*]], %struct1* [[DM:%.*]], i64 0, i32 0
+; CHECK-NEXT: [[TMP1:%.*]] = load %struct2*, %struct2** [[TMP]], align 8
; CHECK-NEXT: br i1 [[TMP4:%.*]], label [[BB1:%.*]], label [[BB2:%.*]]
; CHECK: bb1:
-; CHECK-NEXT: [[TMP10:%.*]] = getelementptr inbounds [[STRUCT2:%.*]], ptr [[TMP1]], i64 [[TMP9:%.*]]
-; CHECK-NEXT: store i32 0, ptr [[TMP10]], align 4
+; CHECK-NEXT: [[TMP11:%.*]] = getelementptr inbounds [[STRUCT2:%.*]], %struct2* [[TMP1]], i64 [[TMP9:%.*]], i32 0
+; CHECK-NEXT: store i32 0, i32* [[TMP11]], align 4
; CHECK-NEXT: br label [[BB3:%.*]]
; CHECK: bb2:
-; CHECK-NEXT: [[TMP20:%.*]] = getelementptr inbounds [[STRUCT2]], ptr [[TMP1]], i64 [[TMP19:%.*]]
-; CHECK-NEXT: store i32 0, ptr [[TMP20]], align 4
+; CHECK-NEXT: [[TMP21:%.*]] = getelementptr inbounds [[STRUCT2]], %struct2* [[TMP1]], i64 [[TMP19:%.*]], i32 0
+; CHECK-NEXT: store i32 0, i32* [[TMP21]], align 4
; CHECK-NEXT: br label [[BB3]]
; CHECK: bb3:
; CHECK-NEXT: [[TMP0:%.*]] = phi i64 [ [[TMP9]], [[BB1]] ], [ [[TMP19]], [[BB2]] ]
-; CHECK-NEXT: [[TMP24:%.*]] = getelementptr inbounds [[STRUCT2]], ptr [[TMP1]], i64 [[TMP0]], i32 1
-; CHECK-NEXT: [[TMP25:%.*]] = load i32, ptr [[TMP24]], align 4
+; CHECK-NEXT: [[TMP24:%.*]] = getelementptr inbounds [[STRUCT2]], %struct2* [[TMP1]], i64 [[TMP0]], i32 1
+; CHECK-NEXT: [[TMP25:%.*]] = load i32, i32* [[TMP24]], align 4
; CHECK-NEXT: ret i32 [[TMP25]]
;
bb:
- %tmp1 = load ptr, ptr %dm, align 8
+ %tmp = getelementptr inbounds %struct1, %struct1* %dm, i64 0, i32 0
+ %tmp1 = load %struct2*, %struct2** %tmp, align 8
br i1 %tmp4, label %bb1, label %bb2
bb1:
- %tmp10 = getelementptr inbounds %struct2, ptr %tmp1, i64 %tmp9
- store i32 0, ptr %tmp10, align 4
+ %tmp10 = getelementptr inbounds %struct2, %struct2* %tmp1, i64 %tmp9
+ %tmp11 = getelementptr inbounds %struct2, %struct2* %tmp10, i64 0, i32 0
+ store i32 0, i32* %tmp11, align 4
br label %bb3
bb2:
- %tmp20 = getelementptr inbounds %struct2, ptr %tmp1, i64 %tmp19
- store i32 0, ptr %tmp20, align 4
+ %tmp20 = getelementptr inbounds %struct2, %struct2* %tmp1, i64 %tmp19
+ %tmp21 = getelementptr inbounds %struct2, %struct2* %tmp20, i64 0, i32 0
+ store i32 0, i32* %tmp21, align 4
br label %bb3
bb3:
- %phi = phi ptr [ %tmp10, %bb1 ], [ %tmp20, %bb2 ]
- %tmp24 = getelementptr inbounds %struct2, ptr %phi, i64 0, i32 1
- %tmp25 = load i32, ptr %tmp24, align 4
+ %phi = phi %struct2* [ %tmp10, %bb1 ], [ %tmp20, %bb2 ]
+ %tmp24 = getelementptr inbounds %struct2, %struct2* %phi, i64 0, i32 1
+ %tmp25 = load i32, i32* %tmp24, align 4
ret i32 %tmp25
}
-define i32 @test2(ptr %dm, i1 %tmp4, i64 %tmp9, i64 %tmp19) {
+define i32 @test2(%struct1* %dm, i1 %tmp4, i64 %tmp9, i64 %tmp19) {
; CHECK-LABEL: @test2(
; CHECK-NEXT: bb:
-; CHECK-NEXT: [[TMP1:%.*]] = load ptr, ptr [[DM:%.*]], align 8
-; CHECK-NEXT: [[TMP10:%.*]] = getelementptr inbounds [[STRUCT2:%.*]], ptr [[TMP1]], i64 [[TMP9:%.*]]
-; CHECK-NEXT: store i32 0, ptr [[TMP10]], align 4
-; CHECK-NEXT: [[TMP20:%.*]] = getelementptr inbounds [[STRUCT2]], ptr [[TMP1]], i64 [[TMP19:%.*]]
-; CHECK-NEXT: store i32 0, ptr [[TMP20]], align 4
-; CHECK-NEXT: [[TMP24:%.*]] = getelementptr inbounds [[STRUCT2]], ptr [[TMP1]], i64 [[TMP9]], i32 1
-; CHECK-NEXT: [[TMP25:%.*]] = load i32, ptr [[TMP24]], align 4
+; CHECK-NEXT: [[TMP:%.*]] = getelementptr inbounds [[STRUCT1:%.*]], %struct1* [[DM:%.*]], i64 0, i32 0
+; CHECK-NEXT: [[TMP1:%.*]] = load %struct2*, %struct2** [[TMP]], align 8
+; CHECK-NEXT: [[TMP11:%.*]] = getelementptr inbounds [[STRUCT2:%.*]], %struct2* [[TMP1]], i64 [[TMP9:%.*]], i32 0
+; CHECK-NEXT: store i32 0, i32* [[TMP11]], align 4
+; CHECK-NEXT: [[TMP21:%.*]] = getelementptr inbounds [[STRUCT2]], %struct2* [[TMP1]], i64 [[TMP19:%.*]], i32 0
+; CHECK-NEXT: store i32 0, i32* [[TMP21]], align 4
+; CHECK-NEXT: [[TMP24:%.*]] = getelementptr inbounds [[STRUCT2]], %struct2* [[TMP1]], i64 [[TMP9]], i32 1
+; CHECK-NEXT: [[TMP25:%.*]] = load i32, i32* [[TMP24]], align 4
; CHECK-NEXT: ret i32 [[TMP25]]
;
bb:
- %tmp1 = load ptr, ptr %dm, align 8
- %tmp10 = getelementptr inbounds %struct2, ptr %tmp1, i64 %tmp9
- store i32 0, ptr %tmp10, align 4
- %tmp20 = getelementptr inbounds %struct2, ptr %tmp1, i64 %tmp19
- store i32 0, ptr %tmp20, align 4
- %tmp24 = getelementptr inbounds %struct2, ptr %tmp10, i64 0, i32 1
- %tmp25 = load i32, ptr %tmp24, align 4
+ %tmp = getelementptr inbounds %struct1, %struct1* %dm, i64 0, i32 0
+ %tmp1 = load %struct2*, %struct2** %tmp, align 8
+ %tmp10 = getelementptr inbounds %struct2, %struct2* %tmp1, i64 %tmp9
+ %tmp11 = getelementptr inbounds %struct2, %struct2* %tmp10, i64 0, i32 0
+ store i32 0, i32* %tmp11, align 4
+ %tmp20 = getelementptr inbounds %struct2, %struct2* %tmp1, i64 %tmp19
+ %tmp21 = getelementptr inbounds %struct2, %struct2* %tmp20, i64 0, i32 0
+ store i32 0, i32* %tmp21, align 4
+ %tmp24 = getelementptr inbounds %struct2, %struct2* %tmp10, i64 0, i32 1
+ %tmp25 = load i32, i32* %tmp24, align 4
ret i32 %tmp25
}
; Check that instcombine doesn't insert GEPs before landingpad.
-define i32 @test3(ptr %dm, i1 %tmp4, i64 %tmp9, i64 %tmp19, i64 %tmp20, i64 %tmp21) personality ptr @__gxx_personality_v0 {
+define i32 @test3(%struct3* %dm, i1 %tmp4, i64 %tmp9, i64 %tmp19, i64 %tmp20, i64 %tmp21) personality i8* bitcast (i32 (...)* @__gxx_personality_v0 to i8*) {
; CHECK-LABEL: @test3(
; CHECK-NEXT: bb:
; CHECK-NEXT: br i1 [[TMP4:%.*]], label [[BB1:%.*]], label [[BB2:%.*]]
; CHECK: bb1:
-; CHECK-NEXT: [[TMP1:%.*]] = getelementptr inbounds [[STRUCT3:%.*]], ptr [[DM:%.*]], i64 [[TMP19:%.*]], i32 1
-; CHECK-NEXT: store i32 0, ptr [[TMP1]], align 4
+; CHECK-NEXT: [[TMP11:%.*]] = getelementptr inbounds [[STRUCT3:%.*]], %struct3* [[DM:%.*]], i64 [[TMP19:%.*]], i32 1, i32 0, i32 0
+; CHECK-NEXT: store i32 0, i32* [[TMP11]], align 4
; CHECK-NEXT: br label [[BB3:%.*]]
; CHECK: bb2:
-; CHECK-NEXT: [[TMP12:%.*]] = getelementptr inbounds [[STRUCT3]], ptr [[DM]], i64 [[TMP20:%.*]], i32 1, i32 0, i32 1
-; CHECK-NEXT: store i32 0, ptr [[TMP12]], align 4
+; CHECK-NEXT: [[TMP12:%.*]] = getelementptr inbounds [[STRUCT3]], %struct3* [[DM]], i64 [[TMP20:%.*]], i32 1, i32 0, i32 1
+; CHECK-NEXT: store i32 0, i32* [[TMP12]], align 4
; CHECK-NEXT: br label [[BB3]]
; CHECK: bb3:
; CHECK-NEXT: [[TMP0:%.*]] = phi i64 [ [[TMP19]], [[BB1]] ], [ [[TMP20]], [[BB2]] ]
; CHECK: bb4:
; CHECK-NEXT: ret i32 0
; CHECK: bb5:
-; CHECK-NEXT: [[TMP27:%.*]] = landingpad { ptr, i32 }
-; CHECK-NEXT: catch ptr @_ZTIi
-; CHECK-NEXT: [[TMP1:%.*]] = getelementptr inbounds [[STRUCT3]], ptr [[DM]], i64 [[TMP0]], i32 1
-; CHECK-NEXT: [[TMP35:%.*]] = getelementptr inbounds [[STRUCT4:%.*]], ptr [[TMP1]], i64 [[TMP21:%.*]], i32 1, i32 1
-; CHECK-NEXT: [[TMP25:%.*]] = load i32, ptr [[TMP35]], align 4
+; CHECK-NEXT: [[TMP27:%.*]] = landingpad { i8*, i32 }
+; CHECK-NEXT: catch i8* bitcast (i8** @_ZTIi to i8*)
+; CHECK-NEXT: [[TMP1:%.*]] = getelementptr inbounds [[STRUCT3]], %struct3* [[DM]], i64 [[TMP0]], i32 1
+; CHECK-NEXT: [[TMP35:%.*]] = getelementptr inbounds [[STRUCT4:%.*]], %struct4* [[TMP1]], i64 [[TMP21:%.*]], i32 1, i32 1
+; CHECK-NEXT: [[TMP25:%.*]] = load i32, i32* [[TMP35]], align 4
; CHECK-NEXT: ret i32 [[TMP25]]
;
bb:
+ %tmp = getelementptr inbounds %struct3, %struct3* %dm, i64 0
br i1 %tmp4, label %bb1, label %bb2
bb1:
- %tmp1 = getelementptr inbounds %struct3, ptr %dm, i64 %tmp19, i32 1
- store i32 0, ptr %tmp1, align 4
+ %tmp1 = getelementptr inbounds %struct3, %struct3* %tmp, i64 %tmp19, i32 1
+ %tmp11 = getelementptr inbounds %struct4, %struct4* %tmp1, i64 0, i32 0, i32 0
+ store i32 0, i32* %tmp11, align 4
br label %bb3
bb2:
- %tmp2 = getelementptr inbounds %struct3, ptr %dm, i64 %tmp20, i32 1
- %tmp12 = getelementptr inbounds %struct4, ptr %tmp2, i64 0, i32 0, i32 1
- store i32 0, ptr %tmp12, align 4
+ %tmp2 = getelementptr inbounds %struct3, %struct3* %tmp, i64 %tmp20, i32 1
+ %tmp12 = getelementptr inbounds %struct4, %struct4* %tmp2, i64 0, i32 0, i32 1
+ store i32 0, i32* %tmp12, align 4
br label %bb3
bb3:
- %phi = phi ptr [ %tmp1, %bb1 ], [ %tmp2, %bb2 ]
+ %phi = phi %struct4* [ %tmp1, %bb1 ], [ %tmp2, %bb2 ]
%tmp22 = invoke i32 @foo1(i32 11) to label %bb4 unwind label %bb5
bb4:
ret i32 0
bb5:
- %tmp27 = landingpad { ptr, i32 } catch ptr @_ZTIi
- %tmp34 = getelementptr inbounds %struct4, ptr %phi, i64 %tmp21, i32 1
- %tmp35 = getelementptr inbounds %struct2, ptr %tmp34, i64 0, i32 1
- %tmp25 = load i32, ptr %tmp35, align 4
+ %tmp27 = landingpad { i8*, i32 } catch i8* bitcast (i8** @_ZTIi to i8*)
+ %tmp34 = getelementptr inbounds %struct4, %struct4* %phi, i64 %tmp21, i32 1
+ %tmp35 = getelementptr inbounds %struct2, %struct2* %tmp34, i64 0, i32 1
+ %tmp25 = load i32, i32* %tmp35, align 4
ret i32 %tmp25
}
-@_ZTIi = external constant ptr
+@_ZTIi = external constant i8*
declare i32 @__gxx_personality_v0(...)
declare i32 @foo1(i32)
; Check that instcombine doesn't fold GEPs into themselves through a loop
; back-edge.
-define ptr @test4(i32 %value, ptr %buffer) {
+define i8* @test4(i32 %value, i8* %buffer) {
; CHECK-LABEL: @test4(
; CHECK-NEXT: entry:
; CHECK-NEXT: [[CMP:%.*]] = icmp ugt i32 [[VALUE:%.*]], 127
; CHECK: loop.header:
; CHECK-NEXT: br label [[LOOP_BODY:%.*]]
; CHECK: loop.body:
-; CHECK-NEXT: [[BUFFER_PN:%.*]] = phi ptr [ [[BUFFER:%.*]], [[LOOP_HEADER]] ], [ [[LOOPPTR:%.*]], [[LOOP_BODY]] ]
+; CHECK-NEXT: [[BUFFER_PN:%.*]] = phi i8* [ [[BUFFER:%.*]], [[LOOP_HEADER]] ], [ [[LOOPPTR:%.*]], [[LOOP_BODY]] ]
; CHECK-NEXT: [[NEWVAL:%.*]] = phi i32 [ [[VALUE]], [[LOOP_HEADER]] ], [ [[SHR:%.*]], [[LOOP_BODY]] ]
-; CHECK-NEXT: [[LOOPPTR]] = getelementptr inbounds i8, ptr [[BUFFER_PN]], i64 1
+; CHECK-NEXT: [[LOOPPTR]] = getelementptr inbounds i8, i8* [[BUFFER_PN]], i64 1
; CHECK-NEXT: [[SHR]] = lshr i32 [[NEWVAL]], 7
; CHECK-NEXT: [[CMP2:%.*]] = icmp ugt i32 [[NEWVAL]], 16383
; CHECK-NEXT: br i1 [[CMP2]], label [[LOOP_BODY]], label [[LOOP_EXIT:%.*]]
; CHECK: loop.exit:
; CHECK-NEXT: br label [[EXIT]]
; CHECK: exit:
-; CHECK-NEXT: [[TMP0:%.*]] = phi ptr [ [[LOOPPTR]], [[LOOP_EXIT]] ], [ [[BUFFER]], [[ENTRY:%.*]] ]
-; CHECK-NEXT: [[INCPTR3:%.*]] = getelementptr inbounds i8, ptr [[TMP0]], i64 2
-; CHECK-NEXT: ret ptr [[INCPTR3]]
+; CHECK-NEXT: [[TMP0:%.*]] = phi i8* [ [[LOOPPTR]], [[LOOP_EXIT]] ], [ [[BUFFER]], [[ENTRY:%.*]] ]
+; CHECK-NEXT: [[INCPTR3:%.*]] = getelementptr inbounds i8, i8* [[TMP0]], i64 2
+; CHECK-NEXT: ret i8* [[INCPTR3]]
;
entry:
- %incptr = getelementptr inbounds i8, ptr %buffer, i64 1
+ %incptr = getelementptr inbounds i8, i8* %buffer, i64 1
%cmp = icmp ugt i32 %value, 127
br i1 %cmp, label %loop.header, label %exit
br label %loop.body
loop.body:
- %loopptr = phi ptr [ %incptr, %loop.header ], [ %incptr2, %loop.body ]
+ %loopptr = phi i8* [ %incptr, %loop.header ], [ %incptr2, %loop.body ]
%newval = phi i32 [ %value, %loop.header ], [ %shr, %loop.body ]
%shr = lshr i32 %newval, 7
- %incptr2 = getelementptr inbounds i8, ptr %loopptr, i64 1
+ %incptr2 = getelementptr inbounds i8, i8* %loopptr, i64 1
%cmp2 = icmp ugt i32 %shr, 127
br i1 %cmp2, label %loop.body, label %loop.exit
loop.exit:
- %exitptr = phi ptr [ %incptr2, %loop.body ]
+ %exitptr = phi i8* [ %incptr2, %loop.body ]
br label %exit
exit:
- %ptr2 = phi ptr [ %exitptr, %loop.exit ], [ %incptr, %entry ]
- %incptr3 = getelementptr inbounds i8, ptr %ptr2, i64 1
- ret ptr %incptr3
+ %ptr2 = phi i8* [ %exitptr, %loop.exit ], [ %incptr, %entry ]
+ %incptr3 = getelementptr inbounds i8, i8* %ptr2, i64 1
+ ret i8* %incptr3
}
@.str.4 = external unnamed_addr constant [100 x i8], align 1
; Instcombine shouldn't add new PHI nodes while folding GEPs if that will leave
; old PHI nodes behind as this is not clearly beneficial.
-define void @test5(ptr %idx, ptr %in) #0 {
+define void @test5(i16 *%idx, i8 **%in) #0 {
; CHECK-LABEL: @test5(
; CHECK-NEXT: entry:
-; CHECK-NEXT: [[TMP0:%.*]] = load ptr, ptr [[IN:%.*]], align 8
-; CHECK-NEXT: [[INCDEC_PTR:%.*]] = getelementptr inbounds i8, ptr [[TMP0]], i64 1
-; CHECK-NEXT: [[TMP1:%.*]] = load i8, ptr [[INCDEC_PTR]], align 1
+; CHECK-NEXT: [[TMP0:%.*]] = load i8*, i8** [[IN:%.*]], align 8
+; CHECK-NEXT: [[INCDEC_PTR:%.*]] = getelementptr inbounds i8, i8* [[TMP0]], i64 1
+; CHECK-NEXT: [[TMP1:%.*]] = load i8, i8* [[INCDEC_PTR]], align 1
; CHECK-NEXT: [[CMP23:%.*]] = icmp eq i8 [[TMP1]], 54
; CHECK-NEXT: br i1 [[CMP23]], label [[WHILE_COND:%.*]], label [[IF_THEN_25:%.*]]
; CHECK: if.then.25:
-; CHECK-NEXT: call void @g(ptr nonnull @.str.4)
+; CHECK-NEXT: call void @g(i8* nonnull getelementptr inbounds ([100 x i8], [100 x i8]* @.str.4, i64 0, i64 0))
; CHECK-NEXT: br label [[WHILE_COND]]
; CHECK: while.cond:
-; CHECK-NEXT: [[PTR:%.*]] = phi ptr [ [[INCDEC_PTR]], [[ENTRY:%.*]] ], [ [[INCDEC_PTR32:%.*]], [[WHILE_BODY:%.*]] ], [ [[INCDEC_PTR]], [[IF_THEN_25]] ]
-; CHECK-NEXT: [[TMP2:%.*]] = load i8, ptr [[PTR]], align 1
+; CHECK-NEXT: [[PTR:%.*]] = phi i8* [ [[INCDEC_PTR]], [[ENTRY:%.*]] ], [ [[INCDEC_PTR32:%.*]], [[WHILE_BODY:%.*]] ], [ [[INCDEC_PTR]], [[IF_THEN_25]] ]
+; CHECK-NEXT: [[TMP2:%.*]] = load i8, i8* [[PTR]], align 1
; CHECK-NEXT: [[AND:%.*]] = and i8 [[TMP2]], 64
; CHECK-NEXT: [[LNOT:%.*]] = icmp eq i8 [[AND]], 0
; CHECK-NEXT: br i1 [[LNOT]], label [[WHILE_BODY]], label [[WHILE_COND_33:%.*]]
; CHECK: while.body:
-; CHECK-NEXT: [[INCDEC_PTR32]] = getelementptr inbounds i8, ptr [[PTR]], i64 1
+; CHECK-NEXT: [[INCDEC_PTR32]] = getelementptr inbounds i8, i8* [[PTR]], i64 1
; CHECK-NEXT: br label [[WHILE_COND]]
; CHECK: while.cond.33:
-; CHECK-NEXT: [[INCDEC_PTR34:%.*]] = getelementptr inbounds i8, ptr [[PTR]], i64 1
+; CHECK-NEXT: [[INCDEC_PTR34:%.*]] = getelementptr inbounds i8, i8* [[PTR]], i64 1
; CHECK-NEXT: br label [[WHILE_COND_57:%.*]]
; CHECK: while.cond.57:
-; CHECK-NEXT: [[TMP3:%.*]] = load i8, ptr [[INCDEC_PTR34]], align 1
+; CHECK-NEXT: [[TMP3:%.*]] = load i8, i8* [[INCDEC_PTR34]], align 1
; CHECK-NEXT: [[TMP4:%.*]] = zext i8 [[TMP3]] to i64
-; CHECK-NEXT: [[ARRAYIDX61:%.*]] = getelementptr inbounds i16, ptr [[IDX:%.*]], i64 [[TMP4]]
-; CHECK-NEXT: [[TMP5:%.*]] = load i16, ptr [[ARRAYIDX61]], align 2
+; CHECK-NEXT: [[ARRAYIDX61:%.*]] = getelementptr inbounds i16, i16* [[IDX:%.*]], i64 [[TMP4]]
+; CHECK-NEXT: [[TMP5:%.*]] = load i16, i16* [[ARRAYIDX61]], align 2
; CHECK-NEXT: [[AND63:%.*]] = and i16 [[TMP5]], 2048
; CHECK-NEXT: [[TOBOOL64:%.*]] = icmp eq i16 [[AND63]], 0
; CHECK-NEXT: br i1 [[TOBOOL64]], label [[WHILE_COND_73:%.*]], label [[WHILE_COND_57]]
; CHECK-NEXT: br label [[WHILE_COND_73]]
;
entry:
- %0 = load ptr, ptr %in
- %incdec.ptr = getelementptr inbounds i8, ptr %0, i32 1
- %1 = load i8, ptr %incdec.ptr, align 1
+ %0 = load i8*, i8** %in
+ %incdec.ptr = getelementptr inbounds i8, i8* %0, i32 1
+ %1 = load i8, i8* %incdec.ptr, align 1
%cmp23 = icmp eq i8 %1, 54
br i1 %cmp23, label %while.cond, label %if.then.25
if.then.25:
- call void @g(ptr @.str.4)
+ call void @g(i8* getelementptr inbounds ([100 x i8], [100 x i8]* @.str.4, i32 0, i32 0))
br label %while.cond
while.cond:
- %Ptr = phi ptr [ %incdec.ptr, %entry ], [ %incdec.ptr32, %while.body], [%incdec.ptr, %if.then.25 ]
- %2 = load i8, ptr %Ptr
+ %Ptr = phi i8* [ %incdec.ptr, %entry ], [ %incdec.ptr32, %while.body], [%incdec.ptr, %if.then.25 ]
+ %2 = load i8, i8* %Ptr
%and = and i8 %2, 64
%lnot = icmp eq i8 %and, 0
br i1 %lnot, label %while.body, label %while.cond.33
while.body:
- %incdec.ptr32 = getelementptr inbounds i8, ptr %Ptr, i32 1
+ %incdec.ptr32 = getelementptr inbounds i8, i8* %Ptr, i32 1
br label %while.cond
while.cond.33:
- %incdec.ptr34 = getelementptr inbounds i8, ptr %Ptr, i32 1
+ %incdec.ptr34 = getelementptr inbounds i8, i8* %Ptr, i32 1
br label %while.cond.57
while.cond.57:
- %3 = load i8, ptr %incdec.ptr34, align 1
+ %3 = load i8, i8* %incdec.ptr34, align 1
%conv59 = zext i8 %3 to i32
- %arrayidx61 = getelementptr inbounds i16, ptr %idx, i32 %conv59
- %4 = load i16, ptr %arrayidx61, align 2
+ %arrayidx61 = getelementptr inbounds i16, i16* %idx, i32 %conv59
+ %4 = load i16, i16* %arrayidx61, align 2
%and63 = and i16 %4, 2048
%tobool64 = icmp eq i16 %and63, 0
br i1 %tobool64, label %while.cond.73, label %while.cond.57
br label %while.cond.73
}
-declare void @g(ptr)
+declare void @g(i8*)
@Global_as1 = external addrspace(1) global [10 x i8]
; Test noop elimination
-define ptr @test1(ptr %I) {
+define i32* @test1(i32* %I) {
; CHECK-LABEL: @test1(
-; CHECK-NEXT: ret ptr [[I:%.*]]
+; CHECK-NEXT: ret i32* [[I:%.*]]
;
- ret ptr %I
+ %A = getelementptr i32, i32* %I, i64 0
+ ret i32* %A
}
-define ptr addrspace(1) @test1_as1(ptr addrspace(1) %I) {
+define i32 addrspace(1)* @test1_as1(i32 addrspace(1)* %I) {
; CHECK-LABEL: @test1_as1(
-; CHECK-NEXT: ret ptr addrspace(1) [[I:%.*]]
+; CHECK-NEXT: ret i32 addrspace(1)* [[I:%.*]]
;
- ret ptr addrspace(1) %I
+ %A = getelementptr i32, i32 addrspace(1)* %I, i64 0
+ ret i32 addrspace(1)* %A
}
; Test noop elimination
-define ptr @test2(ptr %I) {
+define i32* @test2(i32* %I) {
; CHECK-LABEL: @test2(
-; CHECK-NEXT: ret ptr [[I:%.*]]
+; CHECK-NEXT: ret i32* [[I:%.*]]
;
- %A = getelementptr i32, ptr %I
- ret ptr %A
+ %A = getelementptr i32, i32* %I
+ ret i32* %A
}
; Test that two array indexing geps fold
-define ptr @test3(ptr %I) {
+define i32* @test3(i32* %I) {
; CHECK-LABEL: @test3(
-; CHECK-NEXT: [[B:%.*]] = getelementptr i32, ptr [[I:%.*]], i64 21
-; CHECK-NEXT: ret ptr [[B]]
+; CHECK-NEXT: [[B:%.*]] = getelementptr i32, i32* [[I:%.*]], i64 21
+; CHECK-NEXT: ret i32* [[B]]
;
- %A = getelementptr i32, ptr %I, i64 17
- %B = getelementptr i32, ptr %A, i64 4
- ret ptr %B
+ %A = getelementptr i32, i32* %I, i64 17
+ %B = getelementptr i32, i32* %A, i64 4
+ ret i32* %B
}
; Test that two getelementptr insts fold
-define ptr @test4(ptr %I) {
+define i32* @test4({ i32 }* %I) {
; CHECK-LABEL: @test4(
-; CHECK-NEXT: [[A:%.*]] = getelementptr { i32 }, ptr [[I:%.*]], i64 1
-; CHECK-NEXT: ret ptr [[A]]
+; CHECK-NEXT: [[B:%.*]] = getelementptr { i32 }, { i32 }* [[I:%.*]], i64 1, i32 0
+; CHECK-NEXT: ret i32* [[B]]
;
- %A = getelementptr { i32 }, ptr %I, i64 1
- ret ptr %A
+ %A = getelementptr { i32 }, { i32 }* %I, i64 1
+ %B = getelementptr { i32 }, { i32 }* %A, i64 0, i32 0
+ ret i32* %B
}
define void @test5(i8 %B) {
; This should be turned into a constexpr instead of being an instruction
; CHECK-LABEL: @test5(
-; CHECK-NEXT: store i8 [[B:%.*]], ptr getelementptr inbounds ([10 x i8], ptr @Global, i64 0, i64 4), align 1
+; CHECK-NEXT: store i8 [[B:%.*]], i8* getelementptr inbounds ([10 x i8], [10 x i8]* @Global, i64 0, i64 4), align 1
; CHECK-NEXT: ret void
;
- %A = getelementptr [10 x i8], ptr @Global, i64 0, i64 4
- store i8 %B, ptr %A
+ %A = getelementptr [10 x i8], [10 x i8]* @Global, i64 0, i64 4
+ store i8 %B, i8* %A
ret void
}
define void @test5_as1(i8 %B) {
; This should be turned into a constexpr instead of being an instruction
; CHECK-LABEL: @test5_as1(
-; CHECK-NEXT: store i8 [[B:%.*]], ptr addrspace(1) getelementptr inbounds ([10 x i8], ptr addrspace(1) @Global_as1, i16 0, i16 4), align 1
+; CHECK-NEXT: store i8 [[B:%.*]], i8 addrspace(1)* getelementptr inbounds ([10 x i8], [10 x i8] addrspace(1)* @Global_as1, i16 0, i16 4), align 1
; CHECK-NEXT: ret void
;
- %A = getelementptr [10 x i8], ptr addrspace(1) @Global_as1, i16 0, i16 4
- store i8 %B, ptr addrspace(1) %A
+ %A = getelementptr [10 x i8], [10 x i8] addrspace(1)* @Global_as1, i16 0, i16 4
+ store i8 %B, i8 addrspace(1)* %A
ret void
}
-%as1_ptr_struct = type { ptr addrspace(1) }
-%as2_ptr_struct = type { ptr addrspace(2) }
+%as1_ptr_struct = type { i32 addrspace(1)* }
+%as2_ptr_struct = type { i32 addrspace(2)* }
@global_as2 = addrspace(2) global i32 zeroinitializer
-@global_as1_as2_ptr = addrspace(1) global %as2_ptr_struct { ptr addrspace(2) @global_as2 }
+@global_as1_as2_ptr = addrspace(1) global %as2_ptr_struct { i32 addrspace(2)* @global_as2 }
; This should be turned into a constexpr instead of being an instruction
-define void @test_evaluate_gep_nested_as_ptrs(ptr addrspace(2) %B) {
+define void @test_evaluate_gep_nested_as_ptrs(i32 addrspace(2)* %B) {
; CHECK-LABEL: @test_evaluate_gep_nested_as_ptrs(
-; CHECK-NEXT: store ptr addrspace(2) [[B:%.*]], ptr addrspace(1) @global_as1_as2_ptr, align 8
+; CHECK-NEXT: store i32 addrspace(2)* [[B:%.*]], i32 addrspace(2)* addrspace(1)* getelementptr inbounds ([[AS2_PTR_STRUCT:%.*]], [[AS2_PTR_STRUCT]] addrspace(1)* @global_as1_as2_ptr, i16 0, i32 0), align 8
; CHECK-NEXT: ret void
;
- store ptr addrspace(2) %B, ptr addrspace(1) @global_as1_as2_ptr
+ %A = getelementptr %as2_ptr_struct, %as2_ptr_struct addrspace(1)* @global_as1_as2_ptr, i16 0, i32 0
+ store i32 addrspace(2)* %B, i32 addrspace(2)* addrspace(1)* %A
ret void
}
-@arst = addrspace(1) global [4 x ptr addrspace(2)] zeroinitializer
+@arst = addrspace(1) global [4 x i8 addrspace(2)*] zeroinitializer
-define void @test_evaluate_gep_as_ptrs_array(ptr addrspace(2) %B) {
+define void @test_evaluate_gep_as_ptrs_array(i8 addrspace(2)* %B) {
; CHECK-LABEL: @test_evaluate_gep_as_ptrs_array(
-; CHECK-NEXT: store ptr addrspace(2) [[B:%.*]], ptr addrspace(1) getelementptr inbounds ([4 x ptr addrspace(2)], ptr addrspace(1) @arst, i16 0, i16 2), align 4
+; CHECK-NEXT: store i8 addrspace(2)* [[B:%.*]], i8 addrspace(2)* addrspace(1)* getelementptr inbounds ([4 x i8 addrspace(2)*], [4 x i8 addrspace(2)*] addrspace(1)* @arst, i16 0, i16 2), align 4
; CHECK-NEXT: ret void
;
- %A = getelementptr [4 x ptr addrspace(2)], ptr addrspace(1) @arst, i16 0, i16 2
- store ptr addrspace(2) %B, ptr addrspace(1) %A
+ %A = getelementptr [4 x i8 addrspace(2)*], [4 x i8 addrspace(2)*] addrspace(1)* @arst, i16 0, i16 2
+ store i8 addrspace(2)* %B, i8 addrspace(2)* addrspace(1)* %A
ret void
}
-define ptr @test7(ptr %I, i64 %C, i64 %D) {
+define i32* @test7(i32* %I, i64 %C, i64 %D) {
; CHECK-LABEL: @test7(
-; CHECK-NEXT: [[A:%.*]] = getelementptr i32, ptr [[I:%.*]], i64 [[C:%.*]]
-; CHECK-NEXT: [[B:%.*]] = getelementptr i32, ptr [[A]], i64 [[D:%.*]]
-; CHECK-NEXT: ret ptr [[B]]
+; CHECK-NEXT: [[A:%.*]] = getelementptr i32, i32* [[I:%.*]], i64 [[C:%.*]]
+; CHECK-NEXT: [[B:%.*]] = getelementptr i32, i32* [[A]], i64 [[D:%.*]]
+; CHECK-NEXT: ret i32* [[B]]
;
- %A = getelementptr i32, ptr %I, i64 %C
- %B = getelementptr i32, ptr %A, i64 %D
- ret ptr %B
+ %A = getelementptr i32, i32* %I, i64 %C
+ %B = getelementptr i32, i32* %A, i64 %D
+ ret i32* %B
}
-define ptr @test8(ptr %X) {
+define i8* @test8([10 x i32]* %X) {
;; Fold into the cast.
; CHECK-LABEL: @test8(
-; CHECK-NEXT: ret ptr [[X:%.*]]
+; CHECK-NEXT: [[B:%.*]] = bitcast [10 x i32]* [[X:%.*]] to i8*
+; CHECK-NEXT: ret i8* [[B]]
;
- ret ptr %X
+ %A = getelementptr [10 x i32], [10 x i32]* %X, i64 0, i64 0
+ %B = bitcast i32* %A to i8*
+ ret i8* %B
}
define i32 @test9() {
; CHECK-LABEL: @test9(
; CHECK-NEXT: ret i32 8
;
- %A = getelementptr { i32, double }, ptr null, i32 0, i32 1
- %B = ptrtoint ptr %A to i32
+ %A = getelementptr { i32, double }, { i32, double }* null, i32 0, i32 1
+ %B = ptrtoint double* %A to i32
ret i32 %B
}
-define i1 @test10(ptr %x, ptr %y) {
+define i1 @test10({ i32, i32 }* %x, { i32, i32 }* %y) {
; CHECK-LABEL: @test10(
-; CHECK-NEXT: [[T4:%.*]] = icmp eq ptr [[X:%.*]], [[Y:%.*]]
+; CHECK-NEXT: [[T4:%.*]] = icmp eq { i32, i32 }* [[X:%.*]], [[Y:%.*]]
; CHECK-NEXT: ret i1 [[T4]]
;
- %t1 = getelementptr { i32, i32 }, ptr %x, i32 0, i32 1
- %t3 = getelementptr { i32, i32 }, ptr %y, i32 0, i32 1
- %t4 = icmp eq ptr %t1, %t3
+ %t1 = getelementptr { i32, i32 }, { i32, i32 }* %x, i32 0, i32 1
+ %t3 = getelementptr { i32, i32 }, { i32, i32 }* %y, i32 0, i32 1
+ %t4 = icmp eq i32* %t1, %t3
ret i1 %t4
}
-define i1 @test10_addrspacecast(ptr %x, ptr addrspace(3) %y) {
+define i1 @test10_addrspacecast({ i32, i32 }* %x, { i32, i32 } addrspace(3)* %y) {
; CHECK-LABEL: @test10_addrspacecast(
-; CHECK-NEXT: [[T1:%.*]] = getelementptr { i32, i32 }, ptr [[X:%.*]], i64 0, i32 1
-; CHECK-NEXT: [[T3:%.*]] = getelementptr { i32, i32 }, ptr addrspace(3) [[Y:%.*]], i64 0, i32 1
-; CHECK-NEXT: [[T3_C:%.*]] = addrspacecast ptr addrspace(3) [[T3]] to ptr
-; CHECK-NEXT: [[T4:%.*]] = icmp eq ptr [[T1]], [[T3_C]]
+; CHECK-NEXT: [[T1:%.*]] = getelementptr { i32, i32 }, { i32, i32 }* [[X:%.*]], i64 0, i32 1
+; CHECK-NEXT: [[T3:%.*]] = getelementptr { i32, i32 }, { i32, i32 } addrspace(3)* [[Y:%.*]], i64 0, i32 1
+; CHECK-NEXT: [[T3_C:%.*]] = addrspacecast i32 addrspace(3)* [[T3]] to i32*
+; CHECK-NEXT: [[T4:%.*]] = icmp eq i32* [[T1]], [[T3_C]]
; CHECK-NEXT: ret i1 [[T4]]
;
- %t1 = getelementptr { i32, i32 }, ptr %x, i32 0, i32 1
- %t3 = getelementptr { i32, i32 }, ptr addrspace(3) %y, i32 0, i32 1
- %t3.c = addrspacecast ptr addrspace(3) %t3 to ptr
- %t4 = icmp eq ptr %t1, %t3.c
+ %t1 = getelementptr { i32, i32 }, { i32, i32 }* %x, i32 0, i32 1
+ %t3 = getelementptr { i32, i32 }, { i32, i32 } addrspace(3)* %y, i32 0, i32 1
+ %t3.c = addrspacecast i32 addrspace(3)* %t3 to i32*
+ %t4 = icmp eq i32* %t1, %t3.c
ret i1 %t4
}
-define i1 @test11(ptr %X) {
+define i1 @test11({ i32, i32 }* %X) {
; CHECK-LABEL: @test11(
-; CHECK-NEXT: [[Q:%.*]] = icmp eq ptr [[X:%.*]], null
+; CHECK-NEXT: [[Q:%.*]] = icmp eq { i32, i32 }* [[X:%.*]], null
; CHECK-NEXT: ret i1 [[Q]]
;
- %P = getelementptr { i32, i32 }, ptr %X, i32 0, i32 0
- %Q = icmp eq ptr %P, null
+ %P = getelementptr { i32, i32 }, { i32, i32 }* %X, i32 0, i32 0
+ %Q = icmp eq i32* %P, null
ret i1 %Q
}
; PR4748
-define i32 @test12(ptr %a) {
+define i32 @test12(%struct.A* %a) {
; CHECK-LABEL: @test12(
; CHECK-NEXT: entry:
-; CHECK-NEXT: [[G3:%.*]] = getelementptr [[STRUCT_A:%.*]], ptr [[A:%.*]], i64 0, i32 1
-; CHECK-NEXT: store i32 10, ptr [[G3]], align 4
+; CHECK-NEXT: [[G3:%.*]] = getelementptr [[STRUCT_A:%.*]], %struct.A* [[A:%.*]], i64 0, i32 1
+; CHECK-NEXT: store i32 10, i32* [[G3]], align 4
; CHECK-NEXT: ret i32 10
;
entry:
- %g3 = getelementptr %struct.A, ptr %a, i32 0, i32 1
- store i32 10, ptr %g3, align 4
+ %g3 = getelementptr %struct.A, %struct.A* %a, i32 0, i32 1
+ store i32 10, i32* %g3, align 4
+ %g4 = getelementptr %struct.A, %struct.A* %a, i32 0, i32 0
+ %new_a = bitcast %struct.B* %g4 to %struct.A*
- %g5 = getelementptr %struct.A, ptr %a, i32 0, i32 1
- %a_a = load i32, ptr %g5, align 4
+ %g5 = getelementptr %struct.A, %struct.A* %new_a, i32 0, i32 1
+ %a_a = load i32, i32* %g5, align 4
ret i32 %a_a
}
; PR2235
%S = type { i32, [ 100 x i32] }
-define i1 @test13(i64 %X, ptr %P) {
+define i1 @test13(i64 %X, %S* %P) {
; CHECK-LABEL: @test13(
; CHECK-NEXT: [[C:%.*]] = icmp eq i64 [[X:%.*]], -1
; CHECK-NEXT: ret i1 [[C]]
;
- %A = getelementptr inbounds %S, ptr %P, i32 0, i32 1, i64 %X
- %C = icmp eq ptr %A, %P
+ %A = getelementptr inbounds %S, %S* %P, i32 0, i32 1, i64 %X
+ %B = getelementptr inbounds %S, %S* %P, i32 0, i32 0
+ %C = icmp eq i32* %A, %B
ret i1 %C
}
-define <2 x i1> @test13_vector(<2 x i64> %X, <2 x ptr> %P) nounwind {
+define <2 x i1> @test13_vector(<2 x i64> %X, <2 x %S*> %P) nounwind {
; CHECK-LABEL: @test13_vector(
-; CHECK-NEXT: [[A:%.*]] = getelementptr inbounds [[S:%.*]], <2 x ptr> [[P:%.*]], <2 x i64> zeroinitializer, <2 x i32> <i32 1, i32 1>, <2 x i64> [[X:%.*]]
-; CHECK-NEXT: [[C:%.*]] = icmp eq <2 x ptr> [[A]], [[P]]
+; CHECK-NEXT: [[C:%.*]] = icmp eq <2 x i64> [[X:%.*]], <i64 -1, i64 -1>
; CHECK-NEXT: ret <2 x i1> [[C]]
;
- %A = getelementptr inbounds %S, <2 x ptr> %P, <2 x i64> zeroinitializer, <2 x i32> <i32 1, i32 1>, <2 x i64> %X
- %B = getelementptr inbounds %S, <2 x ptr> %P, <2 x i64> <i64 0, i64 0>, <2 x i32> <i32 0, i32 0>
- %C = icmp eq <2 x ptr> %A, %B
+ %A = getelementptr inbounds %S, <2 x %S*> %P, <2 x i64> zeroinitializer, <2 x i32> <i32 1, i32 1>, <2 x i64> %X
+ %B = getelementptr inbounds %S, <2 x %S*> %P, <2 x i64> <i64 0, i64 0>, <2 x i32> <i32 0, i32 0>
+ %C = icmp eq <2 x i32*> %A, %B
ret <2 x i1> %C
}
-define <2 x i1> @test13_vector2(i64 %X, <2 x ptr> %P) nounwind {
+define <2 x i1> @test13_vector2(i64 %X, <2 x %S*> %P) nounwind {
; CHECK-LABEL: @test13_vector2(
-; CHECK-NEXT: [[A:%.*]] = getelementptr inbounds [[S:%.*]], <2 x ptr> [[P:%.*]], <2 x i64> zeroinitializer, <2 x i32> <i32 1, i32 1>, i64 [[X:%.*]]
-; CHECK-NEXT: [[C:%.*]] = icmp eq <2 x ptr> [[A]], [[P]]
+; CHECK-NEXT: [[DOTSPLATINSERT:%.*]] = insertelement <2 x i64> poison, i64 [[X:%.*]], i64 0
+; CHECK-NEXT: [[TMP1:%.*]] = shl <2 x i64> [[DOTSPLATINSERT]], <i64 2, i64 0>
+; CHECK-NEXT: [[TMP2:%.*]] = icmp eq <2 x i64> [[TMP1]], <i64 -4, i64 poison>
+; CHECK-NEXT: [[C:%.*]] = shufflevector <2 x i1> [[TMP2]], <2 x i1> poison, <2 x i32> zeroinitializer
; CHECK-NEXT: ret <2 x i1> [[C]]
;
- %A = getelementptr inbounds %S, <2 x ptr> %P, <2 x i64> zeroinitializer, <2 x i32> <i32 1, i32 1>, i64 %X
- %B = getelementptr inbounds %S, <2 x ptr> %P, <2 x i64> <i64 0, i64 0>, <2 x i32> <i32 0, i32 0>
- %C = icmp eq <2 x ptr> %A, %B
+ %A = getelementptr inbounds %S, <2 x %S*> %P, <2 x i64> zeroinitializer, <2 x i32> <i32 1, i32 1>, i64 %X
+ %B = getelementptr inbounds %S, <2 x %S*> %P, <2 x i64> <i64 0, i64 0>, <2 x i32> <i32 0, i32 0>
+ %C = icmp eq <2 x i32*> %A, %B
ret <2 x i1> %C
}
; This is a test of icmp + shl nuw in disguise - 4611... is 0x3fff...
-define <2 x i1> @test13_vector3(i64 %X, <2 x ptr> %P) nounwind {
+define <2 x i1> @test13_vector3(i64 %X, <2 x %S*> %P) nounwind {
; CHECK-LABEL: @test13_vector3(
; CHECK-NEXT: [[DOTSPLATINSERT:%.*]] = insertelement <2 x i64> poison, i64 [[X:%.*]], i64 0
; CHECK-NEXT: [[TMP1:%.*]] = shl <2 x i64> [[DOTSPLATINSERT]], <i64 2, i64 0>
; CHECK-NEXT: [[C:%.*]] = shufflevector <2 x i1> [[TMP2]], <2 x i1> poison, <2 x i32> zeroinitializer
; CHECK-NEXT: ret <2 x i1> [[C]]
;
- %A = getelementptr inbounds %S, <2 x ptr> %P, <2 x i64> zeroinitializer, <2 x i32> <i32 1, i32 1>, i64 %X
- %B = getelementptr inbounds %S, <2 x ptr> %P, <2 x i64> <i64 0, i64 0>, <2 x i32> <i32 1, i32 1>, i64 1
- %C = icmp eq <2 x ptr> %A, %B
+ %A = getelementptr inbounds %S, <2 x %S*> %P, <2 x i64> zeroinitializer, <2 x i32> <i32 1, i32 1>, i64 %X
+ %B = getelementptr inbounds %S, <2 x %S*> %P, <2 x i64> <i64 0, i64 0>, <2 x i32> <i32 1, i32 1>, i64 1
+ %C = icmp eq <2 x i32*> %A, %B
ret <2 x i1> %C
}
-define i1 @test13_as1(i16 %X, ptr addrspace(1) %P) {
+define i1 @test13_as1(i16 %X, %S addrspace(1)* %P) {
; CHECK-LABEL: @test13_as1(
; CHECK-NEXT: [[C:%.*]] = icmp eq i16 [[X:%.*]], -1
; CHECK-NEXT: ret i1 [[C]]
;
- %A = getelementptr inbounds %S, ptr addrspace(1) %P, i16 0, i32 1, i16 %X
- %C = icmp eq ptr addrspace(1) %A, %P
+ %A = getelementptr inbounds %S, %S addrspace(1)* %P, i16 0, i32 1, i16 %X
+ %B = getelementptr inbounds %S, %S addrspace(1)* %P, i16 0, i32 0
+ %C = icmp eq i32 addrspace(1)* %A, %B
ret i1 %C
}
-define <2 x i1> @test13_vector_as1(<2 x i16> %X, <2 x ptr addrspace(1)> %P) {
+define <2 x i1> @test13_vector_as1(<2 x i16> %X, <2 x %S addrspace(1)*> %P) {
; CHECK-LABEL: @test13_vector_as1(
-; CHECK-NEXT: [[A:%.*]] = getelementptr inbounds [[S:%.*]], <2 x ptr addrspace(1)> [[P:%.*]], <2 x i16> zeroinitializer, <2 x i32> <i32 1, i32 1>, <2 x i16> [[X:%.*]]
-; CHECK-NEXT: [[C:%.*]] = icmp eq <2 x ptr addrspace(1)> [[A]], [[P]]
+; CHECK-NEXT: [[C:%.*]] = icmp eq <2 x i16> [[X:%.*]], <i16 -1, i16 -1>
; CHECK-NEXT: ret <2 x i1> [[C]]
;
- %A = getelementptr inbounds %S, <2 x ptr addrspace(1)> %P, <2 x i16> <i16 0, i16 0>, <2 x i32> <i32 1, i32 1>, <2 x i16> %X
- %B = getelementptr inbounds %S, <2 x ptr addrspace(1)> %P, <2 x i16> <i16 0, i16 0>, <2 x i32> <i32 0, i32 0>
- %C = icmp eq <2 x ptr addrspace(1)> %A, %B
+ %A = getelementptr inbounds %S, <2 x %S addrspace(1)*> %P, <2 x i16> <i16 0, i16 0>, <2 x i32> <i32 1, i32 1>, <2 x i16> %X
+ %B = getelementptr inbounds %S, <2 x %S addrspace(1)*> %P, <2 x i16> <i16 0, i16 0>, <2 x i32> <i32 0, i32 0>
+ %C = icmp eq <2 x i32 addrspace(1)*> %A, %B
ret <2 x i1> %C
}
-define i1 @test13_i32(i32 %X, ptr %P) {
+define i1 @test13_i32(i32 %X, %S* %P) {
; CHECK-LABEL: @test13_i32(
; CHECK-NEXT: [[C:%.*]] = icmp eq i32 [[X:%.*]], -1
; CHECK-NEXT: ret i1 [[C]]
;
- %A = getelementptr inbounds %S, ptr %P, i32 0, i32 1, i32 %X
- %C = icmp eq ptr %A, %P
+ %A = getelementptr inbounds %S, %S* %P, i32 0, i32 1, i32 %X
+ %B = getelementptr inbounds %S, %S* %P, i32 0, i32 0
+ %C = icmp eq i32* %A, %B
ret i1 %C
}
-define i1 @test13_i16(i16 %X, ptr %P) {
+define i1 @test13_i16(i16 %X, %S* %P) {
; CHECK-LABEL: @test13_i16(
; CHECK-NEXT: [[C:%.*]] = icmp eq i16 [[X:%.*]], -1
; CHECK-NEXT: ret i1 [[C]]
;
- %A = getelementptr inbounds %S, ptr %P, i16 0, i32 1, i16 %X
- %C = icmp eq ptr %A, %P
+ %A = getelementptr inbounds %S, %S* %P, i16 0, i32 1, i16 %X
+ %B = getelementptr inbounds %S, %S* %P, i16 0, i32 0
+ %C = icmp eq i32* %A, %B
ret i1 %C
}
-define i1 @test13_i128(i128 %X, ptr %P) {
+define i1 @test13_i128(i128 %X, %S* %P) {
; CHECK-LABEL: @test13_i128(
; CHECK-NEXT: [[TMP1:%.*]] = trunc i128 [[X:%.*]] to i64
; CHECK-NEXT: [[C:%.*]] = icmp eq i64 [[TMP1]], -1
; CHECK-NEXT: ret i1 [[C]]
;
- %A = getelementptr inbounds %S, ptr %P, i128 0, i32 1, i128 %X
- %C = icmp eq ptr %A, %P
+ %A = getelementptr inbounds %S, %S* %P, i128 0, i32 1, i128 %X
+ %B = getelementptr inbounds %S, %S* %P, i128 0, i32 0
+ %C = icmp eq i32* %A, %B
ret i1 %C
}
@G = external global [3 x i8]
-define ptr @test14(i32 %idx) {
+define i8* @test14(i32 %idx) {
; CHECK-LABEL: @test14(
; CHECK-NEXT: [[ZEXT:%.*]] = zext i32 [[IDX:%.*]] to i64
-; CHECK-NEXT: [[T:%.*]] = getelementptr i8, ptr @G, i64 [[ZEXT]]
-; CHECK-NEXT: ret ptr [[T]]
+; CHECK-NEXT: [[T:%.*]] = getelementptr [3 x i8], [3 x i8]* @G, i64 0, i64 [[ZEXT]]
+; CHECK-NEXT: ret i8* [[T]]
;
%zext = zext i32 %idx to i64
- %t = getelementptr i8, ptr @G, i64 %zext
- ret ptr %t
+ %t = getelementptr i8, i8* getelementptr ([3 x i8], [3 x i8]* @G, i32 0, i32 0), i64 %zext
+ ret i8* %t
}
; Test folding of constantexpr geps into normal geps.
@Array = external global [40 x i32]
-define ptr @test15(i64 %X) {
+define i32 *@test15(i64 %X) {
; CHECK-LABEL: @test15(
-; CHECK-NEXT: [[A:%.*]] = getelementptr i32, ptr @Array, i64 [[X:%.*]]
-; CHECK-NEXT: ret ptr [[A]]
+; CHECK-NEXT: [[A:%.*]] = getelementptr [40 x i32], [40 x i32]* @Array, i64 0, i64 [[X:%.*]]
+; CHECK-NEXT: ret i32* [[A]]
;
- %A = getelementptr i32, ptr @Array, i64 %X
- ret ptr %A
+ %A = getelementptr i32, i32* getelementptr ([40 x i32], [40 x i32]* @Array, i64 0, i64 0), i64 %X
+ ret i32* %A
}
-define ptr @test16(ptr %X, i32 %Idx) {
+define i32* @test16(i32* %X, i32 %Idx) {
; CHECK-LABEL: @test16(
; CHECK-NEXT: [[TMP1:%.*]] = sext i32 [[IDX:%.*]] to i64
-; CHECK-NEXT: [[R:%.*]] = getelementptr i32, ptr [[X:%.*]], i64 [[TMP1]]
-; CHECK-NEXT: ret ptr [[R]]
+; CHECK-NEXT: [[R:%.*]] = getelementptr i32, i32* [[X:%.*]], i64 [[TMP1]]
+; CHECK-NEXT: ret i32* [[R]]
;
- %R = getelementptr i32, ptr %X, i32 %Idx
- ret ptr %R
+ %R = getelementptr i32, i32* %X, i32 %Idx
+ ret i32* %R
}
-define i1 @test17(ptr %P, i32 %I, i32 %J) {
+define i1 @test17(i16* %P, i32 %I, i32 %J) {
; CHECK-LABEL: @test17(
; CHECK-NEXT: [[C:%.*]] = icmp slt i32 [[I:%.*]], [[J:%.*]]
; CHECK-NEXT: ret i1 [[C]]
;
- %X = getelementptr inbounds i16, ptr %P, i32 %I
- %Y = getelementptr inbounds i16, ptr %P, i32 %J
- %C = icmp ult ptr %X, %Y
+ %X = getelementptr inbounds i16, i16* %P, i32 %I
+ %Y = getelementptr inbounds i16, i16* %P, i32 %J
+ %C = icmp ult i16* %X, %Y
ret i1 %C
}
-define i1 @test18(ptr %P, i32 %I) {
+define i1 @test18(i16* %P, i32 %I) {
; CHECK-LABEL: @test18(
; CHECK-NEXT: [[C:%.*]] = icmp slt i32 [[I:%.*]], 0
; CHECK-NEXT: ret i1 [[C]]
;
- %X = getelementptr inbounds i16, ptr %P, i32 %I
- %C = icmp ult ptr %X, %P
+ %X = getelementptr inbounds i16, i16* %P, i32 %I
+ %C = icmp ult i16* %X, %P
ret i1 %C
}
; Larger than the pointer size for a non-zero address space
-define i1 @test18_as1(ptr addrspace(1) %P, i32 %I) {
+define i1 @test18_as1(i16 addrspace(1)* %P, i32 %I) {
; CHECK-LABEL: @test18_as1(
; CHECK-NEXT: [[TMP1:%.*]] = and i32 [[I:%.*]], 32768
; CHECK-NEXT: [[C:%.*]] = icmp ne i32 [[TMP1]], 0
; CHECK-NEXT: ret i1 [[C]]
;
- %X = getelementptr inbounds i16, ptr addrspace(1) %P, i32 %I
- %C = icmp ult ptr addrspace(1) %X, %P
+ %X = getelementptr inbounds i16, i16 addrspace(1)* %P, i32 %I
+ %C = icmp ult i16 addrspace(1)* %X, %P
ret i1 %C
}
; Smaller than the pointer size for a non-zero address space
-define i1 @test18_as1_i32(ptr addrspace(1) %P, i32 %I) {
+define i1 @test18_as1_i32(i16 addrspace(1)* %P, i32 %I) {
; CHECK-LABEL: @test18_as1_i32(
; CHECK-NEXT: [[TMP1:%.*]] = and i32 [[I:%.*]], 32768
; CHECK-NEXT: [[C:%.*]] = icmp ne i32 [[TMP1]], 0
; CHECK-NEXT: ret i1 [[C]]
;
- %X = getelementptr inbounds i16, ptr addrspace(1) %P, i32 %I
- %C = icmp ult ptr addrspace(1) %X, %P
+ %X = getelementptr inbounds i16, i16 addrspace(1)* %P, i32 %I
+ %C = icmp ult i16 addrspace(1)* %X, %P
ret i1 %C
}
; Smaller than pointer size
-define i1 @test18_i16(ptr %P, i16 %I) {
+define i1 @test18_i16(i16* %P, i16 %I) {
; CHECK-LABEL: @test18_i16(
; CHECK-NEXT: [[C:%.*]] = icmp slt i16 [[I:%.*]], 0
; CHECK-NEXT: ret i1 [[C]]
;
- %X = getelementptr inbounds i16, ptr %P, i16 %I
- %C = icmp ult ptr %X, %P
+ %X = getelementptr inbounds i16, i16* %P, i16 %I
+ %C = icmp ult i16* %X, %P
ret i1 %C
}
; Same as pointer size
-define i1 @test18_i64(ptr %P, i64 %I) {
+define i1 @test18_i64(i16* %P, i64 %I) {
; CHECK-LABEL: @test18_i64(
; CHECK-NEXT: [[C:%.*]] = icmp slt i64 [[I:%.*]], 0
; CHECK-NEXT: ret i1 [[C]]
;
- %X = getelementptr inbounds i16, ptr %P, i64 %I
- %C = icmp ult ptr %X, %P
+ %X = getelementptr inbounds i16, i16* %P, i64 %I
+ %C = icmp ult i16* %X, %P
ret i1 %C
}
; Larger than the pointer size
-define i1 @test18_i128(ptr %P, i128 %I) {
+define i1 @test18_i128(i16* %P, i128 %I) {
; CHECK-LABEL: @test18_i128(
; CHECK-NEXT: [[TMP1:%.*]] = and i128 [[I:%.*]], 9223372036854775808
; CHECK-NEXT: [[C:%.*]] = icmp ne i128 [[TMP1]], 0
; CHECK-NEXT: ret i1 [[C]]
;
- %X = getelementptr inbounds i16, ptr %P, i128 %I
- %C = icmp ult ptr %X, %P
+ %X = getelementptr inbounds i16, i16* %P, i128 %I
+ %C = icmp ult i16* %X, %P
ret i1 %C
}
-define i32 @test19(ptr %P, i32 %A, i32 %B) {
+define i32 @test19(i32* %P, i32 %A, i32 %B) {
; CHECK-LABEL: @test19(
; CHECK-NEXT: [[T10:%.*]] = icmp eq i32 [[A:%.*]], [[B:%.*]]
; CHECK-NEXT: [[T11:%.*]] = zext i1 [[T10]] to i32
; CHECK-NEXT: ret i32 [[T11]]
;
- %t4 = getelementptr inbounds i32, ptr %P, i32 %A
- %t9 = getelementptr inbounds i32, ptr %P, i32 %B
- %t10 = icmp eq ptr %t4, %t9
+ %t4 = getelementptr inbounds i32, i32* %P, i32 %A
+ %t9 = getelementptr inbounds i32, i32* %P, i32 %B
+ %t10 = icmp eq i32* %t4, %t9
%t11 = zext i1 %t10 to i32
ret i32 %t11
}
-define i32 @test20(ptr %P, i32 %A, i32 %B) {
+define i32 @test20(i32* %P, i32 %A, i32 %B) {
; CHECK-LABEL: @test20(
; CHECK-NEXT: [[T6:%.*]] = icmp eq i32 [[A:%.*]], 0
; CHECK-NEXT: [[T7:%.*]] = zext i1 [[T6]] to i32
; CHECK-NEXT: ret i32 [[T7]]
;
- %t4 = getelementptr inbounds i32, ptr %P, i32 %A
- %t6 = icmp eq ptr %t4, %P
+ %t4 = getelementptr inbounds i32, i32* %P, i32 %A
+ %t6 = icmp eq i32* %t4, %P
%t7 = zext i1 %t6 to i32
ret i32 %t7
}
-define i32 @test20_as1(ptr addrspace(1) %P, i32 %A, i32 %B) {
+define i32 @test20_as1(i32 addrspace(1)* %P, i32 %A, i32 %B) {
; CHECK-LABEL: @test20_as1(
; CHECK-NEXT: [[TMP1:%.*]] = trunc i32 [[A:%.*]] to i16
; CHECK-NEXT: [[T6:%.*]] = icmp eq i16 [[TMP1]], 0
; CHECK-NEXT: [[T7:%.*]] = zext i1 [[T6]] to i32
; CHECK-NEXT: ret i32 [[T7]]
;
- %t4 = getelementptr inbounds i32, ptr addrspace(1) %P, i32 %A
- %t6 = icmp eq ptr addrspace(1) %t4, %P
+ %t4 = getelementptr inbounds i32, i32 addrspace(1)* %P, i32 %A
+ %t6 = icmp eq i32 addrspace(1)* %t4, %P
%t7 = zext i1 %t6 to i32
ret i32 %t7
}
define i32 @test21() {
; CHECK-LABEL: @test21(
; CHECK-NEXT: [[PBOB1:%.*]] = alloca [[INTSTRUCT:%.*]], align 8
-; CHECK-NEXT: [[RVAL:%.*]] = load i32, ptr [[PBOB1]], align 8
+; CHECK-NEXT: [[PBOBEL:%.*]] = getelementptr inbounds [[INTSTRUCT]], %intstruct* [[PBOB1]], i64 0, i32 0
+; CHECK-NEXT: [[RVAL:%.*]] = load i32, i32* [[PBOBEL]], align 8
; CHECK-NEXT: ret i32 [[RVAL]]
;
%pbob1 = alloca %intstruct
- %pbob2 = getelementptr %intstruct, ptr %pbob1
- %rval = load i32, ptr %pbob2
+ %pbob2 = getelementptr %intstruct, %intstruct* %pbob1
+ %pbobel = getelementptr %intstruct, %intstruct* %pbob2, i64 0, i32 0
+ %rval = load i32, i32* %pbobel
ret i32 %rval
}
-@A = global i32 1 ; <ptr> [#uses=1]
-@B = global i32 2 ; <ptr> [#uses=1]
+@A = global i32 1 ; <i32*> [#uses=1]
+@B = global i32 2 ; <i32*> [#uses=1]
define i1 @test22() {
; CHECK-LABEL: @test22(
-; CHECK-NEXT: ret i1 icmp ult (ptr getelementptr inbounds (i32, ptr @A, i64 1), ptr getelementptr (i32, ptr @B, i64 2))
+; CHECK-NEXT: ret i1 icmp ult (i32* getelementptr inbounds (i32, i32* @A, i64 1), i32* getelementptr (i32, i32* @B, i64 2))
;
- %C = icmp ult ptr getelementptr (i32, ptr @A, i64 1),
- getelementptr (i32, ptr @B, i64 2)
+ %C = icmp ult i32* getelementptr (i32, i32* @A, i64 1),
+ getelementptr (i32, i32* @B, i64 2)
ret i1 %C
}
; CHECK-LABEL: @test23(
; CHECK-NEXT: ret i1 false
;
- %A = getelementptr %X, ptr null, i64 0, i32 0, i64 0 ; <ptr> [#uses=1]
- %B = icmp ne ptr %A, null ; <i1> [#uses=1]
+ %A = getelementptr %X, %X* null, i64 0, i32 0, i64 0 ; <i32*> [#uses=1]
+ %B = icmp ne i32* %A, null ; <i1> [#uses=1]
ret i1 %B
}
define void @test25() {
; CHECK-LABEL: @test25(
; CHECK-NEXT: entry:
-; CHECK-NEXT: store i64 poison, ptr null, align 4294967296
+; CHECK-NEXT: store i64 poison, i64* null, align 4294967296
; CHECK-NEXT: tail call void @foo25(i32 0, i64 poison)
; CHECK-NEXT: unreachable
;
entry:
- %t = getelementptr { i64, i64, i64, i64 }, ptr null, i32 0, i32 3
- %t.upgrd.1 = load i64, ptr %t
- %t8.ui = load i64, ptr null
+ %t = getelementptr { i64, i64, i64, i64 }, { i64, i64, i64, i64 }* null, i32 0, i32 3
+ %t.upgrd.1 = load i64, i64* %t
+ %t8.ui = load i64, i64* null
%t8 = bitcast i64 %t8.ui to i64
%t9 = and i64 %t8, %t.upgrd.1
%sext = trunc i64 %t9 to i32
; PR1637
-define i1 @test26(ptr %arr) {
+define i1 @test26(i8* %arr) {
; CHECK-LABEL: @test26(
; CHECK-NEXT: ret i1 true
;
- %X = getelementptr i8, ptr %arr, i32 1
- %Y = getelementptr i8, ptr %arr, i32 1
- %test = icmp uge ptr %X, %Y
+ %X = getelementptr i8, i8* %arr, i32 1
+ %Y = getelementptr i8, i8* %arr, i32 1
+ %test = icmp uge i8* %X, %Y
ret i1 %test
}
%struct.__large_struct = type { [100 x i64] }
%struct.compat_siginfo = type { i32, i32, i32, { [29 x i32] } }
%struct.siginfo_t = type { i32, i32, i32, { { i32, i32, [0 x i8], %struct.sigval_t, i32 }, [88 x i8] } }
- %struct.sigval_t = type { ptr }
+ %struct.sigval_t = type { i8* }
-define i32 @test27(ptr %to, ptr %from) {
+define i32 @test27(%struct.compat_siginfo* %to, %struct.siginfo_t* %from) {
; CHECK-LABEL: @test27(
; CHECK-NEXT: entry:
-; CHECK-NEXT: [[FROM_ADDR:%.*]] = alloca ptr, align 8
-; CHECK-NEXT: [[T344:%.*]] = load ptr, ptr [[FROM_ADDR]], align 8
-; CHECK-NEXT: [[T348:%.*]] = getelementptr [[STRUCT_SIGINFO_T:%.*]], ptr [[T344]], i64 0, i32 3, i32 0, i32 3
-; CHECK-NEXT: [[T351:%.*]] = load i32, ptr [[T348]], align 8
-; CHECK-NEXT: [[T360:%.*]] = call i32 asm sideeffect "...", "=r,ir,*m,i,0,~{dirflag},~{fpsr},~{flags}"(i32 [[T351]], ptr elementtype([[STRUCT___LARGE_STRUCT:%.*]]) null, i32 -14, i32 0) #[[ATTR0:[0-9]+]]
+; CHECK-NEXT: [[FROM_ADDR:%.*]] = alloca %struct.siginfo_t*, align 8
+; CHECK-NEXT: [[T344:%.*]] = load %struct.siginfo_t*, %struct.siginfo_t** [[FROM_ADDR]], align 8
+; CHECK-NEXT: [[T349:%.*]] = getelementptr [[STRUCT_SIGINFO_T:%.*]], %struct.siginfo_t* [[T344]], i64 0, i32 3, i32 0, i32 3, i32 0
+; CHECK-NEXT: [[T349350:%.*]] = bitcast i8** [[T349]] to i32*
+; CHECK-NEXT: [[T351:%.*]] = load i32, i32* [[T349350]], align 8
+; CHECK-NEXT: [[T360:%.*]] = call i32 asm sideeffect "...", "=r,ir,*m,i,0,~{dirflag},~{fpsr},~{flags}"(i32 [[T351]], %struct.__large_struct* elementtype([[STRUCT___LARGE_STRUCT:%.*]]) null, i32 -14, i32 0) #[[ATTR0:[0-9]+]]
; CHECK-NEXT: unreachable
;
entry:
- %from_addr = alloca ptr
- %t344 = load ptr, ptr %from_addr, align 8
- %t345 = getelementptr %struct.siginfo_t, ptr %t344, i32 0, i32 3
- %t346 = getelementptr { { i32, i32, [0 x i8], %struct.sigval_t, i32 }, [88 x i8] }, ptr %t345, i32 0, i32 0
- %t348 = getelementptr { i32, i32, %struct.sigval_t }, ptr %t346, i32 0, i32 2
- %t351 = load i32, ptr %t348, align 8
+ %from_addr = alloca %struct.siginfo_t*
+ %t344 = load %struct.siginfo_t*, %struct.siginfo_t** %from_addr, align 8
+ %t345 = getelementptr %struct.siginfo_t, %struct.siginfo_t* %t344, i32 0, i32 3
+ %t346 = getelementptr { { i32, i32, [0 x i8], %struct.sigval_t, i32 }, [88 x i8] }, { { i32, i32, [0 x i8], %struct.sigval_t, i32 }, [88 x i8] }* %t345, i32 0, i32 0
+ %t346347 = bitcast { i32, i32, [0 x i8], %struct.sigval_t, i32 }* %t346 to { i32, i32, %struct.sigval_t }*
+ %t348 = getelementptr { i32, i32, %struct.sigval_t }, { i32, i32, %struct.sigval_t }* %t346347, i32 0, i32 2
+ %t349 = getelementptr %struct.sigval_t, %struct.sigval_t* %t348, i32 0, i32 0
+ %t349350 = bitcast i8** %t349 to i32*
+ %t351 = load i32, i32* %t349350, align 8
%t360 = call i32 asm sideeffect "...",
- "=r,ir,*m,i,0,~{dirflag},~{fpsr},~{flags}"( i32 %t351, ptr elementtype(%struct.__large_struct) null, i32 -14, i32 0 )
+ "=r,ir,*m,i,0,~{dirflag},~{fpsr},~{flags}"( i32 %t351, %struct.__large_struct* elementtype(%struct.__large_struct) null, i32 -14, i32 0 )
unreachable
}
; CHECK-LABEL: @test28(
; CHECK-NEXT: entry:
; CHECK-NEXT: [[ORIENTATIONS:%.*]] = alloca [1 x [1 x %struct.x]], align 8
-; CHECK-NEXT: [[T3:%.*]] = call i32 @puts(ptr noundef nonnull dereferenceable(1) @.str) #[[ATTR0]]
+; CHECK-NEXT: [[T3:%.*]] = call i32 @puts(i8* noundef nonnull dereferenceable(1) getelementptr inbounds ([6 x i8], [6 x i8]* @.str, i64 0, i64 0)) #[[ATTR0]]
; CHECK-NEXT: br label [[BB10:%.*]]
; CHECK: bb10:
; CHECK-NEXT: [[INDVAR:%.*]] = phi i32 [ 0, [[ENTRY:%.*]] ], [ [[INDVAR_NEXT:%.*]], [[BB10]] ]
; CHECK-NEXT: [[T12_REC:%.*]] = xor i32 [[INDVAR]], -1
; CHECK-NEXT: [[TMP0:%.*]] = sext i32 [[T12_REC]] to i64
-; CHECK-NEXT: [[T12:%.*]] = getelementptr inbounds [1 x [1 x %struct.x]], ptr [[ORIENTATIONS]], i64 1, i64 0, i64 [[TMP0]]
-; CHECK-NEXT: [[T16:%.*]] = call i32 (ptr, ...) @printf(ptr noundef nonnull dereferenceable(1) @.str1, ptr nonnull [[T12]]) #[[ATTR0]]
+; CHECK-NEXT: [[T12:%.*]] = getelementptr inbounds [1 x [1 x %struct.x]], [1 x [1 x %struct.x]]* [[ORIENTATIONS]], i64 1, i64 0, i64 [[TMP0]]
+; CHECK-NEXT: [[T16:%.*]] = call i32 (i8*, ...) @printf(i8* noundef nonnull dereferenceable(1) getelementptr inbounds ([12 x i8], [12 x i8]* @.str1, i64 0, i64 0), %struct.x* nonnull [[T12]]) #[[ATTR0]]
; CHECK-NEXT: [[T84:%.*]] = icmp eq i32 [[INDVAR]], 0
; CHECK-NEXT: [[INDVAR_NEXT]] = add i32 [[INDVAR]], 1
; CHECK-NEXT: br i1 [[T84]], label [[BB17:%.*]], label [[BB10]]
;
entry:
%orientations = alloca [1 x [1 x %struct.x]]
- %t3 = call i32 @puts( ptr @.str ) nounwind
- %t45 = getelementptr inbounds [1 x [1 x %struct.x]], ptr %orientations, i32 1, i32 0, i32 0
+ %t3 = call i32 @puts( i8* getelementptr ([6 x i8], [6 x i8]* @.str, i32 0, i32 0) ) nounwind
+ %t45 = getelementptr inbounds [1 x [1 x %struct.x]], [1 x [1 x %struct.x]]* %orientations, i32 1, i32 0, i32 0
+ %orientations62 = getelementptr [1 x [1 x %struct.x]], [1 x [1 x %struct.x]]* %orientations, i32 0, i32 0, i32 0
br label %bb10
bb10:
%indvar = phi i32 [ 0, %entry ], [ %indvar.next, %bb10 ]
%t.0.reg2mem.0.rec = mul i32 %indvar, -1
%t12.rec = add i32 %t.0.reg2mem.0.rec, -1
- %t12 = getelementptr inbounds %struct.x, ptr %t45, i32 %t12.rec
- %t16 = call i32 (ptr, ...) @printf( ptr nonnull dereferenceable(1) @.str1, ptr %t12 ) nounwind
- %t84 = icmp eq ptr %t12, %orientations
+ %t12 = getelementptr inbounds %struct.x, %struct.x* %t45, i32 %t12.rec
+ %t16 = call i32 (i8*, ...) @printf( i8* nonnull dereferenceable(1) getelementptr ([12 x i8], [12 x i8]* @.str1, i32 0, i32 0), %struct.x* %t12 ) nounwind
+ %t84 = icmp eq %struct.x* %t12, %orientations62
%indvar.next = add i32 %indvar, 1
br i1 %t84, label %bb17, label %bb10
ret i32 0
}
-declare i32 @puts(ptr)
+declare i32 @puts(i8*)
-declare i32 @printf(ptr, ...)
+declare i32 @printf(i8*, ...)
; rdar://6762290
%T = type <{ i64, i64, i64 }>
-define i32 @test29(ptr %start, i32 %X) nounwind {
+define i32 @test29(i8* %start, i32 %X) nounwind {
; CHECK-LABEL: @test29(
; CHECK-NEXT: entry:
-; CHECK-NEXT: store i64 poison, ptr null, align 4294967296
+; CHECK-NEXT: store i64 poison, i64* null, align 4294967296
; CHECK-NEXT: br i1 poison, label [[IF_THEN216:%.*]], label [[IF_END363:%.*]]
; CHECK: if.then216:
; CHECK-NEXT: ret i32 1
; CHECK-NEXT: ret i32 0
;
entry:
- %t3 = load i64, ptr null
- %add.ptr = getelementptr i8, ptr %start, i64 %t3
- %t158 = load i32, ptr null
- %add.ptr159 = getelementptr %T, ptr null, i32 %t158
- %add.ptr212 = getelementptr i8, ptr %start, i32 %X
- %cmp214 = icmp ugt ptr %add.ptr212, %add.ptr
+ %t3 = load i64, i64* null
+ %add.ptr = getelementptr i8, i8* %start, i64 %t3
+ %t158 = load i32, i32* null
+ %add.ptr159 = getelementptr %T, %T* null, i32 %t158
+ %add.ptr209 = getelementptr i8, i8* %start, i64 0
+ %add.ptr212 = getelementptr i8, i8* %add.ptr209, i32 %X
+ %cmp214 = icmp ugt i8* %add.ptr212, %add.ptr
br i1 %cmp214, label %if.then216, label %if.end363
if.then216:
; CHECK-NEXT: entry:
; CHECK-NEXT: [[TMP0:%.*]] = zext i32 [[N:%.*]] to i64
; CHECK-NEXT: [[TMP1:%.*]] = alloca i32, i64 [[TMP0]], align 4
-; CHECK-NEXT: call void @test30f(ptr nonnull [[TMP1]]) #[[ATTR0]]
+; CHECK-NEXT: call void @test30f(i32* nonnull [[TMP1]]) #[[ATTR0]]
; CHECK-NEXT: [[TMP2:%.*]] = sext i32 [[M:%.*]] to i64
-; CHECK-NEXT: [[TMP3:%.*]] = getelementptr [0 x i32], ptr [[TMP1]], i64 0, i64 [[TMP2]]
-; CHECK-NEXT: [[TMP4:%.*]] = load i32, ptr [[TMP3]], align 4
+; CHECK-NEXT: [[TMP3:%.*]] = getelementptr i32, i32* [[TMP1]], i64 [[TMP2]]
+; CHECK-NEXT: [[TMP4:%.*]] = load i32, i32* [[TMP3]], align 4
; CHECK-NEXT: ret i32 [[TMP4]]
;
entry:
%0 = alloca i32, i32 %n, align 4
- call void @test30f(ptr %0) nounwind
- %1 = getelementptr [0 x i32], ptr %0, i32 0, i32 %m
- %2 = load i32, ptr %1, align 4
- ret i32 %2
+ %1 = bitcast i32* %0 to [0 x i32]*
+ call void @test30f(i32* %0) nounwind
+ %2 = getelementptr [0 x i32], [0 x i32]* %1, i32 0, i32 %m
+ %3 = load i32, i32* %2, align 4
+ ret i32 %3
}
-declare void @test30f(ptr)
+declare void @test30f(i32*)
-define i1 @test31(ptr %A) {
+define i1 @test31(i32* %A) {
; CHECK-LABEL: @test31(
; CHECK-NEXT: ret i1 true
;
- %B = getelementptr i32, ptr %A, i32 1
- %C = getelementptr i32, ptr %A, i64 1
- %V = icmp eq ptr %B, %C
+ %B = getelementptr i32, i32* %A, i32 1
+ %C = getelementptr i32, i32* %A, i64 1
+ %V = icmp eq i32* %B, %C
ret i1 %V
}
; PR1345
-define ptr @test32(ptr %v) {
+define i8* @test32(i8* %v) {
; CHECK-LABEL: @test32(
-; CHECK-NEXT: [[A:%.*]] = alloca [4 x ptr], align 16
-; CHECK-NEXT: store ptr null, ptr [[A]], align 16
-; CHECK-NEXT: [[D:%.*]] = getelementptr inbounds { [16 x i8] }, ptr [[A]], i64 0, i32 0, i64 8
-; CHECK-NEXT: store ptr [[V:%.*]], ptr [[D]], align 8
-; CHECK-NEXT: [[F:%.*]] = getelementptr inbounds [4 x ptr], ptr [[A]], i64 0, i64 2
-; CHECK-NEXT: [[G:%.*]] = load ptr, ptr [[F]], align 16
-; CHECK-NEXT: ret ptr [[G]]
-;
- %A = alloca [4 x ptr], align 16
- store ptr null, ptr %A
- %D = getelementptr { [16 x i8] }, ptr %A, i32 0, i32 0, i32 8
- store ptr %v, ptr %D
- %F = getelementptr [4 x ptr], ptr %A, i32 0, i32 2
- %G = load ptr, ptr %F
- ret ptr %G
+; CHECK-NEXT: [[A:%.*]] = alloca [4 x i8*], align 16
+; CHECK-NEXT: [[B:%.*]] = getelementptr inbounds [4 x i8*], [4 x i8*]* [[A]], i64 0, i64 0
+; CHECK-NEXT: store i8* null, i8** [[B]], align 16
+; CHECK-NEXT: [[D:%.*]] = getelementptr inbounds [4 x i8*], [4 x i8*]* [[A]], i64 0, i64 1
+; CHECK-NEXT: store i8* [[V:%.*]], i8** [[D]], align 8
+; CHECK-NEXT: [[F:%.*]] = getelementptr inbounds [4 x i8*], [4 x i8*]* [[A]], i64 0, i64 2
+; CHECK-NEXT: [[G:%.*]] = load i8*, i8** [[F]], align 16
+; CHECK-NEXT: ret i8* [[G]]
+;
+ %A = alloca [4 x i8*], align 16
+ %B = getelementptr [4 x i8*], [4 x i8*]* %A, i32 0, i32 0
+ store i8* null, i8** %B
+ %C = bitcast [4 x i8*]* %A to { [16 x i8] }*
+ %D = getelementptr { [16 x i8] }, { [16 x i8] }* %C, i32 0, i32 0, i32 8
+ %E = bitcast i8* %D to i8**
+ store i8* %v, i8** %E
+ %F = getelementptr [4 x i8*], [4 x i8*]* %A, i32 0, i32 2
+ %G = load i8*, i8** %F
+ ret i8* %G
}
; PR3290
%struct.Key = type { { i32, i32 } }
%struct.anon = type <{ i8, [3 x i8], i32 }>
-define ptr @test33(ptr %A) {
+define i32* @test33(%struct.Key* %A) {
; CHECK-LABEL: @test33(
-; CHECK-NEXT: [[C:%.*]] = getelementptr [[STRUCT_ANON:%.*]], ptr [[A:%.*]], i64 0, i32 2
-; CHECK-NEXT: ret ptr [[C]]
+; CHECK-NEXT: [[TMP1:%.*]] = getelementptr [[STRUCT_KEY:%.*]], %struct.Key* [[A:%.*]], i64 0, i32 0, i32 1
+; CHECK-NEXT: ret i32* [[TMP1]]
;
- %C = getelementptr %struct.anon, ptr %A, i32 0, i32 2
- ret ptr %C
+ %B = bitcast %struct.Key* %A to %struct.anon*
+ %C = getelementptr %struct.anon, %struct.anon* %B, i32 0, i32 2
+ ret i32* %C
}
-define ptr addrspace(1) @test33_as1(ptr addrspace(1) %A) {
+define i32 addrspace(1)* @test33_as1(%struct.Key addrspace(1)* %A) {
; CHECK-LABEL: @test33_as1(
-; CHECK-NEXT: [[C:%.*]] = getelementptr [[STRUCT_ANON:%.*]], ptr addrspace(1) [[A:%.*]], i16 0, i32 2
-; CHECK-NEXT: ret ptr addrspace(1) [[C]]
+; CHECK-NEXT: [[TMP1:%.*]] = getelementptr [[STRUCT_KEY:%.*]], [[STRUCT_KEY]] addrspace(1)* [[A:%.*]], i16 0, i32 0, i32 1
+; CHECK-NEXT: ret i32 addrspace(1)* [[TMP1]]
;
- %C = getelementptr %struct.anon, ptr addrspace(1) %A, i32 0, i32 2
- ret ptr addrspace(1) %C
+ %B = bitcast %struct.Key addrspace(1)* %A to %struct.anon addrspace(1)*
+ %C = getelementptr %struct.anon, %struct.anon addrspace(1)* %B, i32 0, i32 2
+ ret i32 addrspace(1)* %C
}
-define ptr addrspace(1) @test33_array_as1(ptr addrspace(1) %A) {
+define i32 addrspace(1)* @test33_array_as1([10 x i32] addrspace(1)* %A) {
; CHECK-LABEL: @test33_array_as1(
-; CHECK-NEXT: [[C:%.*]] = getelementptr [5 x i32], ptr addrspace(1) [[A:%.*]], i16 0, i16 2
-; CHECK-NEXT: ret ptr addrspace(1) [[C]]
+; CHECK-NEXT: [[C:%.*]] = getelementptr [10 x i32], [10 x i32] addrspace(1)* [[A:%.*]], i16 0, i16 2
+; CHECK-NEXT: ret i32 addrspace(1)* [[C]]
;
- %C = getelementptr [5 x i32], ptr addrspace(1) %A, i32 0, i32 2
- ret ptr addrspace(1) %C
+ %B = bitcast [10 x i32] addrspace(1)* %A to [5 x i32] addrspace(1)*
+ %C = getelementptr [5 x i32], [5 x i32] addrspace(1)* %B, i32 0, i32 2
+ ret i32 addrspace(1)* %C
}
; Make sure the GEP indices use the right pointer sized integer
-define ptr addrspace(1) @test33_array_struct_as1(ptr addrspace(1) %A) {
+define i32 addrspace(1)* @test33_array_struct_as1([10 x %struct.Key] addrspace(1)* %A) {
; CHECK-LABEL: @test33_array_struct_as1(
-; CHECK-NEXT: [[C:%.*]] = getelementptr [20 x i32], ptr addrspace(1) [[A:%.*]], i16 0, i16 2
-; CHECK-NEXT: ret ptr addrspace(1) [[C]]
+; CHECK-NEXT: [[TMP1:%.*]] = getelementptr [10 x %struct.Key], [10 x %struct.Key] addrspace(1)* [[A:%.*]], i16 0, i16 1, i32 0, i32 0
+; CHECK-NEXT: ret i32 addrspace(1)* [[TMP1]]
;
- %C = getelementptr [20 x i32], ptr addrspace(1) %A, i32 0, i32 2
- ret ptr addrspace(1) %C
+ %B = bitcast [10 x %struct.Key] addrspace(1)* %A to [20 x i32] addrspace(1)*
+ %C = getelementptr [20 x i32], [20 x i32] addrspace(1)* %B, i32 0, i32 2
+ ret i32 addrspace(1)* %C
}
-define ptr addrspace(1) @test33_addrspacecast(ptr %A) {
+define i32 addrspace(1)* @test33_addrspacecast(%struct.Key* %A) {
; CHECK-LABEL: @test33_addrspacecast(
-; CHECK-NEXT: [[B:%.*]] = addrspacecast ptr [[A:%.*]] to ptr addrspace(1)
-; CHECK-NEXT: [[C:%.*]] = getelementptr [[STRUCT_ANON:%.*]], ptr addrspace(1) [[B]], i16 0, i32 2
-; CHECK-NEXT: ret ptr addrspace(1) [[C]]
+; CHECK-NEXT: [[C:%.*]] = getelementptr [[STRUCT_KEY:%.*]], %struct.Key* [[A:%.*]], i64 0, i32 0, i32 1
+; CHECK-NEXT: [[TMP1:%.*]] = addrspacecast i32* [[C]] to i32 addrspace(1)*
+; CHECK-NEXT: ret i32 addrspace(1)* [[TMP1]]
;
- %B = addrspacecast ptr %A to ptr addrspace(1)
- %C = getelementptr %struct.anon, ptr addrspace(1) %B, i32 0, i32 2
- ret ptr addrspace(1) %C
+ %B = addrspacecast %struct.Key* %A to %struct.anon addrspace(1)*
+ %C = getelementptr %struct.anon, %struct.anon addrspace(1)* %B, i32 0, i32 2
+ ret i32 addrspace(1)* %C
}
- %T2 = type { ptr, i8 }
-define ptr @test34(ptr %Val, i64 %V) nounwind {
+ %T2 = type { i8*, i8 }
+define i8* @test34(i8* %Val, i64 %V) nounwind {
; CHECK-LABEL: @test34(
; CHECK-NEXT: entry:
-; CHECK-NEXT: [[C_CAST:%.*]] = inttoptr i64 [[V:%.*]] to ptr
-; CHECK-NEXT: ret ptr [[C_CAST]]
+; CHECK-NEXT: [[C_CAST:%.*]] = inttoptr i64 [[V:%.*]] to i8*
+; CHECK-NEXT: ret i8* [[C_CAST]]
;
entry:
%A = alloca %T2, align 8
+ %mrv_gep = bitcast %T2* %A to i64*
+ %B = getelementptr %T2, %T2* %A, i64 0, i32 0
- store i64 %V, ptr %A
- %C = load ptr, ptr %A, align 8
- ret ptr %C
+ store i64 %V, i64* %mrv_gep
+ %C = load i8*, i8** %B, align 8
+ ret i8* %C
}
-%t0 = type { ptr, [19 x i8] }
-%t1 = type { ptr, [0 x i8] }
+%t0 = type { i8*, [19 x i8] }
+%t1 = type { i8*, [0 x i8] }
@array = external global [11 x i8]
define i32 @test35() nounwind {
; CHECK-LABEL: @test35(
-; CHECK-NEXT: [[TMP1:%.*]] = call i32 (ptr, ...) @printf(ptr noundef nonnull dereferenceable(1) @"\01LC8", ptr nonnull getelementptr inbounds ([[T0:%.*]], ptr @s, i64 0, i32 1, i64 0)) #[[ATTR0]]
+; CHECK-NEXT: [[TMP1:%.*]] = call i32 (i8*, ...) @printf(i8* noundef nonnull dereferenceable(1) getelementptr inbounds ([17 x i8], [17 x i8]* @"\01LC8", i64 0, i64 0), i8* nonnull getelementptr inbounds ([[T0:%.*]], %t0* @s, i64 0, i32 1, i64 0)) #[[ATTR0]]
; CHECK-NEXT: ret i32 0
;
- call i32 (ptr, ...) @printf(ptr @"\01LC8",
- ptr getelementptr (%t1, ptr @s, i32 0, i32 1, i32 0)) nounwind
+ call i32 (i8*, ...) @printf(i8* getelementptr ([17 x i8], [17 x i8]* @"\01LC8", i32 0, i32 0),
+ i8* getelementptr (%t1, %t1* bitcast (%t0* @s to %t1*), i32 0, i32 1, i32 0)) nounwind
ret i32 0
}
; Don't treat signed offsets as unsigned.
-define ptr @test36() nounwind {
+define i8* @test36() nounwind {
; CHECK-LABEL: @test36(
-; CHECK-NEXT: ret ptr getelementptr ([11 x i8], ptr @array, i64 -1, i64 10)
+; CHECK-NEXT: ret i8* getelementptr ([11 x i8], [11 x i8]* @array, i64 -1, i64 10)
;
- ret ptr getelementptr ([11 x i8], ptr @array, i32 0, i64 -1)
+ ret i8* getelementptr ([11 x i8], [11 x i8]* @array, i32 0, i64 -1)
}
; Instcombine shouldn't assume that gep(A,0,1) != gep(A,1,0).
; CHECK-LABEL: @test37(
; CHECK-NEXT: ret i1 true
;
- %t = icmp eq ptr getelementptr ([1 x i8], ptr @A37, i64 0, i64 1),
- getelementptr ([1 x i8], ptr @A37, i64 1, i64 0)
+ %t = icmp eq i8* getelementptr ([1 x i8], [1 x i8]* @A37, i64 0, i64 1),
+ getelementptr ([1 x i8], [1 x i8]* @A37, i64 1, i64 0)
ret i1 %t
}
; Test index promotion
-define ptr @test38(ptr %I, i32 %n) {
+define i32* @test38(i32* %I, i32 %n) {
; CHECK-LABEL: @test38(
; CHECK-NEXT: [[TMP1:%.*]] = sext i32 [[N:%.*]] to i64
-; CHECK-NEXT: [[A:%.*]] = getelementptr i32, ptr [[I:%.*]], i64 [[TMP1]]
-; CHECK-NEXT: ret ptr [[A]]
+; CHECK-NEXT: [[A:%.*]] = getelementptr i32, i32* [[I:%.*]], i64 [[TMP1]]
+; CHECK-NEXT: ret i32* [[A]]
;
- %A = getelementptr i32, ptr %I, i32 %n
- ret ptr %A
+ %A = getelementptr i32, i32* %I, i32 %n
+ ret i32* %A
}
; Test that we don't duplicate work when the second gep is a "bitcast".
-%pr10322_t = type { ptr }
-declare void @pr10322_f2(ptr)
-declare void @pr10322_f3(ptr)
-define void @pr10322_f1(ptr %foo) {
+%pr10322_t = type { i8* }
+declare void @pr10322_f2(%pr10322_t*)
+declare void @pr10322_f3(i8**)
+define void @pr10322_f1(%pr10322_t* %foo) {
; CHECK-LABEL: @pr10322_f1(
; CHECK-NEXT: entry:
-; CHECK-NEXT: [[ARRAYIDX8:%.*]] = getelementptr inbounds [[PR10322_T:%.*]], ptr [[FOO:%.*]], i64 2
-; CHECK-NEXT: call void @pr10322_f2(ptr nonnull [[ARRAYIDX8]]) #[[ATTR0]]
-; CHECK-NEXT: call void @pr10322_f3(ptr nonnull [[ARRAYIDX8]]) #[[ATTR0]]
+; CHECK-NEXT: [[ARRAYIDX8:%.*]] = getelementptr inbounds [[PR10322_T:%.*]], %pr10322_t* [[FOO:%.*]], i64 2
+; CHECK-NEXT: call void @pr10322_f2(%pr10322_t* nonnull [[ARRAYIDX8]]) #[[ATTR0]]
+; CHECK-NEXT: [[T2:%.*]] = getelementptr inbounds [[PR10322_T]], %pr10322_t* [[ARRAYIDX8]], i64 0, i32 0
+; CHECK-NEXT: call void @pr10322_f3(i8** nonnull [[T2]]) #[[ATTR0]]
; CHECK-NEXT: ret void
;
entry:
- %arrayidx8 = getelementptr inbounds %pr10322_t, ptr %foo, i64 2
- call void @pr10322_f2(ptr %arrayidx8) nounwind
- call void @pr10322_f3(ptr %arrayidx8) nounwind
+ %arrayidx8 = getelementptr inbounds %pr10322_t, %pr10322_t* %foo, i64 2
+ call void @pr10322_f2(%pr10322_t* %arrayidx8) nounwind
+ %t2 = getelementptr inbounds %pr10322_t, %pr10322_t* %arrayidx8, i64 0, i32 0
+ call void @pr10322_f3(i8** %t2) nounwind
ret void
}
%three_gep_t = type {i32}
%three_gep_t2 = type {%three_gep_t}
-define void @three_gep_f(ptr %x) {
+define void @three_gep_f(%three_gep_t2* %x) {
; CHECK-LABEL: @three_gep_f(
-; CHECK-NEXT: [[GEP1:%.*]] = getelementptr [[THREE_GEP_T2:%.*]], ptr [[X:%.*]], i64 2
-; CHECK-NEXT: call void @three_gep_h(ptr [[GEP1]])
-; CHECK-NEXT: call void @three_gep_g(ptr [[GEP1]])
+; CHECK-NEXT: [[GEP1:%.*]] = getelementptr [[THREE_GEP_T2:%.*]], %three_gep_t2* [[X:%.*]], i64 2
+; CHECK-NEXT: call void @three_gep_h(%three_gep_t2* [[GEP1]])
+; CHECK-NEXT: [[GEP3:%.*]] = getelementptr [[THREE_GEP_T2]], %three_gep_t2* [[GEP1]], i64 0, i32 0, i32 0
+; CHECK-NEXT: call void @three_gep_g(i32* [[GEP3]])
; CHECK-NEXT: ret void
;
- %gep1 = getelementptr %three_gep_t2, ptr %x, i64 2
- call void @three_gep_h(ptr %gep1)
- call void @three_gep_g(ptr %gep1)
+ %gep1 = getelementptr %three_gep_t2, %three_gep_t2* %x, i64 2
+ call void @three_gep_h(%three_gep_t2* %gep1)
+ %gep2 = getelementptr %three_gep_t2, %three_gep_t2* %gep1, i64 0, i32 0
+ %gep3 = getelementptr %three_gep_t, %three_gep_t* %gep2, i64 0, i32 0
+ call void @three_gep_g(i32* %gep3)
ret void
}
-declare void @three_gep_g(ptr)
-declare void @three_gep_h(ptr)
+declare void @three_gep_g(i32*)
+declare void @three_gep_h(%three_gep_t2*)
-%struct.ham = type { i32, ptr, ptr, ptr }
+%struct.ham = type { i32, %struct.zot*, %struct.zot*, %struct.zot* }
%struct.zot = type { i64, i8 }
-define void @test39(ptr %arg, i8 %arg1) nounwind {
+define void @test39(%struct.ham* %arg, i8 %arg1) nounwind {
; CHECK-LABEL: @test39(
-; CHECK-NEXT: [[T:%.*]] = getelementptr inbounds [[STRUCT_HAM:%.*]], ptr [[ARG:%.*]], i64 0, i32 2
-; CHECK-NEXT: [[T2:%.*]] = load ptr, ptr [[T]], align 8
-; CHECK-NEXT: [[T4:%.*]] = getelementptr inbounds i8, ptr [[T2]], i64 -8
-; CHECK-NEXT: store i8 [[ARG1:%.*]], ptr [[T4]], align 8
+; CHECK-NEXT: [[T:%.*]] = getelementptr inbounds [[STRUCT_HAM:%.*]], %struct.ham* [[ARG:%.*]], i64 0, i32 2
+; CHECK-NEXT: [[TMP1:%.*]] = bitcast %struct.zot** [[T]] to i8**
+; CHECK-NEXT: [[T21:%.*]] = load i8*, i8** [[TMP1]], align 8
+; CHECK-NEXT: [[T4:%.*]] = getelementptr inbounds i8, i8* [[T21]], i64 -8
+; CHECK-NEXT: store i8 [[ARG1:%.*]], i8* [[T4]], align 8
; CHECK-NEXT: ret void
;
- %t = getelementptr inbounds %struct.ham, ptr %arg, i64 0, i32 2
- %t2 = load ptr, ptr %t, align 8
- %t4 = getelementptr inbounds i8, ptr %t2, i64 -8
- store i8 %arg1, ptr %t4, align 8
+ %t = getelementptr inbounds %struct.ham, %struct.ham* %arg, i64 0, i32 2
+ %t2 = load %struct.zot*, %struct.zot** %t, align 8
+ %t3 = bitcast %struct.zot* %t2 to i8*
+ %t4 = getelementptr inbounds i8, i8* %t3, i64 -8
+ store i8 %arg1, i8* %t4, align 8
ret void
}
-define i1 @pr16483(ptr %a, ptr %b) {
+define i1 @pr16483([1 x i8]* %a, [1 x i8]* %b) {
; CHECK-LABEL: @pr16483(
-; CHECK-NEXT: [[CMP:%.*]] = icmp ult ptr [[A:%.*]], [[B:%.*]]
+; CHECK-NEXT: [[CMP:%.*]] = icmp ult [1 x i8]* [[A:%.*]], [[B:%.*]]
; CHECK-NEXT: ret i1 [[CMP]]
;
- %cmp = icmp ult ptr %a, %b
+ %c = getelementptr [1 x i8], [1 x i8]* %a, i32 0, i32 0
+ %d = getelementptr [1 x i8], [1 x i8]* %b, i32 0, i32 0
+ %cmp = icmp ult i8* %c, %d
ret i1 %cmp
}
-define i8 @test_gep_bitcast_as1(ptr addrspace(1) %arr, i16 %N) {
+define i8 @test_gep_bitcast_as1(i32 addrspace(1)* %arr, i16 %N) {
; CHECK-LABEL: @test_gep_bitcast_as1(
-; CHECK-NEXT: [[V:%.*]] = shl i16 [[N:%.*]], 2
-; CHECK-NEXT: [[T:%.*]] = getelementptr i8, ptr addrspace(1) [[ARR:%.*]], i16 [[V]]
-; CHECK-NEXT: [[X:%.*]] = load i8, ptr addrspace(1) [[T]], align 1
+; CHECK-NEXT: [[T1:%.*]] = getelementptr i32, i32 addrspace(1)* [[ARR:%.*]], i16 [[N:%.*]]
+; CHECK-NEXT: [[T:%.*]] = bitcast i32 addrspace(1)* [[T1]] to i8 addrspace(1)*
+; CHECK-NEXT: [[X:%.*]] = load i8, i8 addrspace(1)* [[T]], align 1
; CHECK-NEXT: ret i8 [[X]]
;
+ %cast = bitcast i32 addrspace(1)* %arr to i8 addrspace(1)*
%V = mul i16 %N, 4
- %t = getelementptr i8, ptr addrspace(1) %arr, i16 %V
- %x = load i8, ptr addrspace(1) %t
+ %t = getelementptr i8, i8 addrspace(1)* %cast, i16 %V
+ %x = load i8, i8 addrspace(1)* %t
ret i8 %x
}
; The element size of the array matches the element size of the pointer
-define i64 @test_gep_bitcast_array_same_size_element(ptr %arr, i64 %N) {
+define i64 @test_gep_bitcast_array_same_size_element([100 x double]* %arr, i64 %N) {
; CHECK-LABEL: @test_gep_bitcast_array_same_size_element(
; CHECK-NEXT: [[V:%.*]] = shl i64 [[N:%.*]], 3
-; CHECK-NEXT: [[T:%.*]] = getelementptr i64, ptr [[ARR:%.*]], i64 [[V]]
-; CHECK-NEXT: [[X:%.*]] = load i64, ptr [[T]], align 4
+; CHECK-NEXT: [[T1:%.*]] = getelementptr [100 x double], [100 x double]* [[ARR:%.*]], i64 0, i64 [[V]]
+; CHECK-NEXT: [[T:%.*]] = bitcast double* [[T1]] to i64*
+; CHECK-NEXT: [[X:%.*]] = load i64, i64* [[T]], align 4
; CHECK-NEXT: ret i64 [[X]]
;
+ %cast = bitcast [100 x double]* %arr to i64*
%V = mul i64 %N, 8
- %t = getelementptr i64, ptr %arr, i64 %V
- %x = load i64, ptr %t
+ %t = getelementptr i64, i64* %cast, i64 %V
+ %x = load i64, i64* %t
ret i64 %x
}
; gep should be done in the original address space.
-define i64 @test_gep_bitcast_array_same_size_element_addrspacecast(ptr %arr, i64 %N) {
+define i64 @test_gep_bitcast_array_same_size_element_addrspacecast([100 x double]* %arr, i64 %N) {
; CHECK-LABEL: @test_gep_bitcast_array_same_size_element_addrspacecast(
-; CHECK-NEXT: [[CAST:%.*]] = addrspacecast ptr [[ARR:%.*]] to ptr addrspace(3)
; CHECK-NEXT: [[V:%.*]] = shl i64 [[N:%.*]], 3
-; CHECK-NEXT: [[T:%.*]] = getelementptr i64, ptr addrspace(3) [[CAST]], i64 [[V]]
-; CHECK-NEXT: [[X:%.*]] = load i64, ptr addrspace(3) [[T]], align 4
+; CHECK-NEXT: [[T1:%.*]] = getelementptr [100 x double], [100 x double]* [[ARR:%.*]], i64 0, i64 [[V]]
+; CHECK-NEXT: [[TMP1:%.*]] = bitcast double* [[T1]] to i64*
+; CHECK-NEXT: [[T:%.*]] = addrspacecast i64* [[TMP1]] to i64 addrspace(3)*
+; CHECK-NEXT: [[X:%.*]] = load i64, i64 addrspace(3)* [[T]], align 4
; CHECK-NEXT: ret i64 [[X]]
;
- %cast = addrspacecast ptr %arr to ptr addrspace(3)
+ %cast = addrspacecast [100 x double]* %arr to i64 addrspace(3)*
%V = mul i64 %N, 8
- %t = getelementptr i64, ptr addrspace(3) %cast, i64 %V
- %x = load i64, ptr addrspace(3) %t
+ %t = getelementptr i64, i64 addrspace(3)* %cast, i64 %V
+ %x = load i64, i64 addrspace(3)* %t
ret i64 %x
}
; The element size of the array is different the element size of the pointer
-define i8 @test_gep_bitcast_array_different_size_element(ptr %arr, i64 %N) {
+define i8 @test_gep_bitcast_array_different_size_element([100 x double]* %arr, i64 %N) {
; CHECK-LABEL: @test_gep_bitcast_array_different_size_element(
-; CHECK-NEXT: [[V:%.*]] = shl i64 [[N:%.*]], 3
-; CHECK-NEXT: [[T:%.*]] = getelementptr i8, ptr [[ARR:%.*]], i64 [[V]]
-; CHECK-NEXT: [[X:%.*]] = load i8, ptr [[T]], align 1
+; CHECK-NEXT: [[T1:%.*]] = getelementptr [100 x double], [100 x double]* [[ARR:%.*]], i64 0, i64 [[N:%.*]]
+; CHECK-NEXT: [[T:%.*]] = bitcast double* [[T1]] to i8*
+; CHECK-NEXT: [[X:%.*]] = load i8, i8* [[T]], align 1
; CHECK-NEXT: ret i8 [[X]]
;
+ %cast = bitcast [100 x double]* %arr to i8*
%V = mul i64 %N, 8
- %t = getelementptr i8, ptr %arr, i64 %V
- %x = load i8, ptr %t
+ %t = getelementptr i8, i8* %cast, i64 %V
+ %x = load i8, i8* %t
ret i8 %x
}
-define i64 @test_gep_bitcast_array_same_size_element_as1(ptr addrspace(1) %arr, i16 %N) {
+define i64 @test_gep_bitcast_array_same_size_element_as1([100 x double] addrspace(1)* %arr, i16 %N) {
; CHECK-LABEL: @test_gep_bitcast_array_same_size_element_as1(
; CHECK-NEXT: [[V:%.*]] = shl i16 [[N:%.*]], 3
-; CHECK-NEXT: [[T:%.*]] = getelementptr i64, ptr addrspace(1) [[ARR:%.*]], i16 [[V]]
-; CHECK-NEXT: [[X:%.*]] = load i64, ptr addrspace(1) [[T]], align 4
+; CHECK-NEXT: [[T1:%.*]] = getelementptr [100 x double], [100 x double] addrspace(1)* [[ARR:%.*]], i16 0, i16 [[V]]
+; CHECK-NEXT: [[T:%.*]] = bitcast double addrspace(1)* [[T1]] to i64 addrspace(1)*
+; CHECK-NEXT: [[X:%.*]] = load i64, i64 addrspace(1)* [[T]], align 4
; CHECK-NEXT: ret i64 [[X]]
;
+ %cast = bitcast [100 x double] addrspace(1)* %arr to i64 addrspace(1)*
%V = mul i16 %N, 8
- %t = getelementptr i64, ptr addrspace(1) %arr, i16 %V
- %x = load i64, ptr addrspace(1) %t
+ %t = getelementptr i64, i64 addrspace(1)* %cast, i16 %V
+ %x = load i64, i64 addrspace(1)* %t
ret i64 %x
}
-define i8 @test_gep_bitcast_array_different_size_element_as1(ptr addrspace(1) %arr, i16 %N) {
+define i8 @test_gep_bitcast_array_different_size_element_as1([100 x double] addrspace(1)* %arr, i16 %N) {
; CHECK-LABEL: @test_gep_bitcast_array_different_size_element_as1(
-; CHECK-NEXT: [[V:%.*]] = shl i16 [[N:%.*]], 3
-; CHECK-NEXT: [[T:%.*]] = getelementptr i8, ptr addrspace(1) [[ARR:%.*]], i16 [[V]]
-; CHECK-NEXT: [[X:%.*]] = load i8, ptr addrspace(1) [[T]], align 1
+; CHECK-NEXT: [[T1:%.*]] = getelementptr [100 x double], [100 x double] addrspace(1)* [[ARR:%.*]], i16 0, i16 [[N:%.*]]
+; CHECK-NEXT: [[T:%.*]] = bitcast double addrspace(1)* [[T1]] to i8 addrspace(1)*
+; CHECK-NEXT: [[X:%.*]] = load i8, i8 addrspace(1)* [[T]], align 1
; CHECK-NEXT: ret i8 [[X]]
;
+ %cast = bitcast [100 x double] addrspace(1)* %arr to i8 addrspace(1)*
%V = mul i16 %N, 8
- %t = getelementptr i8, ptr addrspace(1) %arr, i16 %V
- %x = load i8, ptr addrspace(1) %t
+ %t = getelementptr i8, i8 addrspace(1)* %cast, i16 %V
+ %x = load i8, i8 addrspace(1)* %t
ret i8 %x
}
; CHECK-NEXT: ret i64 8
;
%array = alloca [3 x i32], align 4
- %gep = getelementptr inbounds [3 x i32], ptr %array, i64 0, i64 2
- %p = ptrtoint ptr %array to i64
+ %gep = getelementptr inbounds [3 x i32], [3 x i32]* %array, i64 0, i64 2
+ %gepi8 = bitcast i32* %gep to i8*
+ %p = ptrtoint [3 x i32]* %array to i64
%np = sub i64 0, %p
- %gep2 = getelementptr i8, ptr %gep, i64 %np
- %ret = ptrtoint ptr %gep2 to i64
+ %gep2 = getelementptr i8, i8* %gepi8, i64 %np
+ %ret = ptrtoint i8* %gep2 to i64
ret i64 %ret
}
-define i16 @test41(ptr addrspace(1) %array) {
+define i16 @test41([3 x i32] addrspace(1)* %array) {
; CHECK-LABEL: @test41(
; CHECK-NEXT: ret i16 8
;
- %gep = getelementptr inbounds [3 x i32], ptr addrspace(1) %array, i16 0, i16 2
- %p = ptrtoint ptr addrspace(1) %array to i16
+ %gep = getelementptr inbounds [3 x i32], [3 x i32] addrspace(1)* %array, i16 0, i16 2
+ %gepi8 = bitcast i32 addrspace(1)* %gep to i8 addrspace(1)*
+ %p = ptrtoint [3 x i32] addrspace(1)* %array to i16
%np = sub i16 0, %p
- %gep2 = getelementptr i8, ptr addrspace(1) %gep, i16 %np
- %ret = ptrtoint ptr addrspace(1) %gep2 to i16
+ %gep2 = getelementptr i8, i8 addrspace(1)* %gepi8, i16 %np
+ %ret = ptrtoint i8 addrspace(1)* %gep2 to i16
ret i16 %ret
}
-define ptr @test42i(ptr %c1, ptr %c2) {
+define i8* @test42i(i8* %c1, i8* %c2) {
; CHECK-LABEL: @test42i(
-; CHECK-NEXT: [[PTRTOINT:%.*]] = ptrtoint ptr [[C1:%.*]] to i64
+; CHECK-NEXT: [[PTRTOINT:%.*]] = ptrtoint i8* [[C1:%.*]] to i64
; CHECK-NEXT: [[SUB:%.*]] = sub i64 0, [[PTRTOINT]]
-; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds i8, ptr [[C2:%.*]], i64 [[SUB]]
-; CHECK-NEXT: ret ptr [[GEP]]
+; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds i8, i8* [[C2:%.*]], i64 [[SUB]]
+; CHECK-NEXT: ret i8* [[GEP]]
;
- %ptrtoint = ptrtoint ptr %c1 to i64
+ %ptrtoint = ptrtoint i8* %c1 to i64
%sub = sub i64 0, %ptrtoint
- %gep = getelementptr inbounds i8, ptr %c2, i64 %sub
- ret ptr %gep
+ %gep = getelementptr inbounds i8, i8* %c2, i64 %sub
+ ret i8* %gep
}
-define ptr @test42(ptr %c1, ptr %c2) {
+define i8* @test42(i8* %c1, i8* %c2) {
; CHECK-LABEL: @test42(
-; CHECK-NEXT: [[PTRTOINT:%.*]] = ptrtoint ptr [[C1:%.*]] to i64
+; CHECK-NEXT: [[PTRTOINT:%.*]] = ptrtoint i8* [[C1:%.*]] to i64
; CHECK-NEXT: [[SUB:%.*]] = sub i64 0, [[PTRTOINT]]
-; CHECK-NEXT: [[GEP:%.*]] = getelementptr i8, ptr [[C2:%.*]], i64 [[SUB]]
-; CHECK-NEXT: ret ptr [[GEP]]
+; CHECK-NEXT: [[GEP:%.*]] = getelementptr i8, i8* [[C2:%.*]], i64 [[SUB]]
+; CHECK-NEXT: ret i8* [[GEP]]
;
- %ptrtoint = ptrtoint ptr %c1 to i64
+ %ptrtoint = ptrtoint i8* %c1 to i64
%sub = sub i64 0, %ptrtoint
- %gep = getelementptr i8, ptr %c2, i64 %sub
- ret ptr %gep
+ %gep = getelementptr i8, i8* %c2, i64 %sub
+ ret i8* %gep
}
-define ptr @test43i(ptr %c1, ptr %c2) {
+define i16* @test43i(i16* %c1, i16* %c2) {
; CHECK-LABEL: @test43i(
-; CHECK-NEXT: [[PTRTOINT:%.*]] = ptrtoint ptr [[C1:%.*]] to i64
+; CHECK-NEXT: [[PTRTOINT:%.*]] = ptrtoint i16* [[C1:%.*]] to i64
; CHECK-NEXT: [[SUB:%.*]] = sub i64 0, [[PTRTOINT]]
; CHECK-NEXT: [[SHR:%.*]] = ashr i64 [[SUB]], 1
-; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds i16, ptr [[C2:%.*]], i64 [[SHR]]
-; CHECK-NEXT: ret ptr [[GEP]]
+; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds i16, i16* [[C2:%.*]], i64 [[SHR]]
+; CHECK-NEXT: ret i16* [[GEP]]
;
- %ptrtoint = ptrtoint ptr %c1 to i64
+ %ptrtoint = ptrtoint i16* %c1 to i64
%sub = sub i64 0, %ptrtoint
%shr = ashr i64 %sub, 1
- %gep = getelementptr inbounds i16, ptr %c2, i64 %shr
- ret ptr %gep
+ %gep = getelementptr inbounds i16, i16* %c2, i64 %shr
+ ret i16* %gep
}
-define ptr @test44i(ptr %c1, ptr %c2) {
+define %struct.C* @test44i(%struct.C* %c1, %struct.C* %c2) {
; CHECK-LABEL: @test44i(
-; CHECK-NEXT: [[PTRTOINT:%.*]] = ptrtoint ptr [[C1:%.*]] to i64
+; CHECK-NEXT: [[PTRTOINT:%.*]] = ptrtoint %struct.C* [[C1:%.*]] to i64
; CHECK-NEXT: [[SUB:%.*]] = sub i64 0, [[PTRTOINT]]
; CHECK-NEXT: [[SHR:%.*]] = sdiv i64 [[SUB]], 7
-; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds [[STRUCT_C:%.*]], ptr [[C2:%.*]], i64 [[SHR]]
-; CHECK-NEXT: ret ptr [[GEP]]
+; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds [[STRUCT_C:%.*]], %struct.C* [[C2:%.*]], i64 [[SHR]]
+; CHECK-NEXT: ret %struct.C* [[GEP]]
;
- %ptrtoint = ptrtoint ptr %c1 to i64
+ %ptrtoint = ptrtoint %struct.C* %c1 to i64
%sub = sub i64 0, %ptrtoint
%shr = sdiv i64 %sub, 7
- %gep = getelementptr inbounds %struct.C, ptr %c2, i64 %shr
- ret ptr %gep
+ %gep = getelementptr inbounds %struct.C, %struct.C* %c2, i64 %shr
+ ret %struct.C* %gep
}
-define ptr @test45(ptr %c1, ptr %c2) {
+define %struct.C* @test45(%struct.C* %c1, %struct.C** %c2) {
; CHECK-LABEL: @test45(
-; CHECK-NEXT: [[PTRTOINT1:%.*]] = ptrtoint ptr [[C1:%.*]] to i64
-; CHECK-NEXT: [[PTRTOINT2:%.*]] = ptrtoint ptr [[C2:%.*]] to i64
+; CHECK-NEXT: [[PTRTOINT1:%.*]] = ptrtoint %struct.C* [[C1:%.*]] to i64
+; CHECK-NEXT: [[PTRTOINT2:%.*]] = ptrtoint %struct.C** [[C2:%.*]] to i64
; CHECK-NEXT: [[SUB:%.*]] = sub i64 [[PTRTOINT2]], [[PTRTOINT1]]
; CHECK-NEXT: [[SHR:%.*]] = sdiv i64 [[SUB]], 7
-; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds [[STRUCT_C:%.*]], ptr [[C1]], i64 [[SHR]]
-; CHECK-NEXT: ret ptr [[GEP]]
+; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds [[STRUCT_C:%.*]], %struct.C* [[C1]], i64 [[SHR]]
+; CHECK-NEXT: ret %struct.C* [[GEP]]
;
- %ptrtoint1 = ptrtoint ptr %c1 to i64
- %ptrtoint2 = ptrtoint ptr %c2 to i64
+ %ptrtoint1 = ptrtoint %struct.C* %c1 to i64
+ %ptrtoint2 = ptrtoint %struct.C** %c2 to i64
%sub = sub i64 %ptrtoint2, %ptrtoint1 ; C2 - C1
%shr = sdiv i64 %sub, 7
- %gep = getelementptr inbounds %struct.C, ptr %c1, i64 %shr ; C1 + (C2 - C1)
- ret ptr %gep
+ %gep = getelementptr inbounds %struct.C, %struct.C* %c1, i64 %shr ; C1 + (C2 - C1)
+ ret %struct.C* %gep
}
-define ptr @test46(ptr %c1, ptr %c2, i64 %N) {
+define %struct.C* @test46(%struct.C* %c1, %struct.C* %c2, i64 %N) {
; CHECK-LABEL: @test46(
-; CHECK-NEXT: [[PTRTOINT:%.*]] = ptrtoint ptr [[C1:%.*]] to i64
+; CHECK-NEXT: [[PTRTOINT:%.*]] = ptrtoint %struct.C* [[C1:%.*]] to i64
; CHECK-NEXT: [[SUB:%.*]] = sub i64 0, [[PTRTOINT]]
; CHECK-NEXT: [[SDIV:%.*]] = sdiv i64 [[SUB]], [[N:%.*]]
-; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds [[STRUCT_C:%.*]], ptr [[C2:%.*]], i64 [[SDIV]]
-; CHECK-NEXT: ret ptr [[GEP]]
+; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds [[STRUCT_C:%.*]], %struct.C* [[C2:%.*]], i64 [[SDIV]]
+; CHECK-NEXT: ret %struct.C* [[GEP]]
;
- %ptrtoint = ptrtoint ptr %c1 to i64
+ %ptrtoint = ptrtoint %struct.C* %c1 to i64
%sub = sub i64 0, %ptrtoint
%sdiv = sdiv i64 %sub, %N
- %gep = getelementptr inbounds %struct.C, ptr %c2, i64 %sdiv
- ret ptr %gep
+ %gep = getelementptr inbounds %struct.C, %struct.C* %c2, i64 %sdiv
+ ret %struct.C* %gep
}
-define ptr @test47(ptr %I, i64 %C, i64 %D) {
+define i32* @test47(i32* %I, i64 %C, i64 %D) {
; CHECK-LABEL: @test47(
-; CHECK-NEXT: [[B:%.*]] = getelementptr i32, ptr [[I:%.*]], i64 [[D:%.*]]
-; CHECK-NEXT: ret ptr [[B]]
+; CHECK-NEXT: [[B:%.*]] = getelementptr i32, i32* [[I:%.*]], i64 [[D:%.*]]
+; CHECK-NEXT: ret i32* [[B]]
;
%sub = sub i64 %D, %C
- %A = getelementptr i32, ptr %I, i64 %C
- %B = getelementptr i32, ptr %A, i64 %sub
- ret ptr %B
+ %A = getelementptr i32, i32* %I, i64 %C
+ %B = getelementptr i32, i32* %A, i64 %sub
+ ret i32* %B
}
-define ptr @test48(ptr %I, i64 %C, i64 %D) {
+define i32* @test48(i32* %I, i64 %C, i64 %D) {
; CHECK-LABEL: @test48(
-; CHECK-NEXT: [[B:%.*]] = getelementptr i32, ptr [[I:%.*]], i64 [[D:%.*]]
-; CHECK-NEXT: ret ptr [[B]]
+; CHECK-NEXT: [[B:%.*]] = getelementptr i32, i32* [[I:%.*]], i64 [[D:%.*]]
+; CHECK-NEXT: ret i32* [[B]]
;
%sub = sub i64 %D, %C
- %A = getelementptr i32, ptr %I, i64 %sub
- %B = getelementptr i32, ptr %A, i64 %C
- ret ptr %B
+ %A = getelementptr i32, i32* %I, i64 %sub
+ %B = getelementptr i32, i32* %A, i64 %C
+ ret i32* %B
}
-define ptr @test49(ptr %I, i64 %C) {
+define i32* @test49(i32* %I, i64 %C) {
; CHECK-LABEL: @test49(
-; CHECK-NEXT: [[B:%.*]] = getelementptr i32, ptr [[I:%.*]], i64 -1
-; CHECK-NEXT: ret ptr [[B]]
+; CHECK-NEXT: [[B:%.*]] = getelementptr i32, i32* [[I:%.*]], i64 -1
+; CHECK-NEXT: ret i32* [[B]]
;
%notC = xor i64 -1, %C
- %A = getelementptr i32, ptr %I, i64 %C
- %B = getelementptr i32, ptr %A, i64 %notC
- ret ptr %B
+ %A = getelementptr i32, i32* %I, i64 %C
+ %B = getelementptr i32, i32* %A, i64 %notC
+ ret i32* %B
}
-define ptr addrspace(1) @ascast_0_gep(ptr %p) nounwind {
+define i32 addrspace(1)* @ascast_0_gep(i32* %p) nounwind {
; CHECK-LABEL: @ascast_0_gep(
-; CHECK-NEXT: [[X:%.*]] = addrspacecast ptr [[P:%.*]] to ptr addrspace(1)
-; CHECK-NEXT: ret ptr addrspace(1) [[X]]
+; CHECK-NEXT: [[X:%.*]] = addrspacecast i32* [[P:%.*]] to i32 addrspace(1)*
+; CHECK-NEXT: ret i32 addrspace(1)* [[X]]
;
- %x = addrspacecast ptr %p to ptr addrspace(1)
- ret ptr addrspace(1) %x
+ %gep = getelementptr i32, i32* %p, i32 0
+ %x = addrspacecast i32* %gep to i32 addrspace(1)*
+ ret i32 addrspace(1)* %x
}
; Do not merge the GEP and the addrspacecast, because it would undo the
; addrspacecast canonicalization.
-define ptr addrspace(1) @ascast_0_0_gep(ptr %p) nounwind {
+define i32 addrspace(1)* @ascast_0_0_gep([128 x i32]* %p) nounwind {
; CHECK-LABEL: @ascast_0_0_gep(
-; CHECK-NEXT: [[X:%.*]] = addrspacecast ptr [[P:%.*]] to ptr addrspace(1)
-; CHECK-NEXT: ret ptr addrspace(1) [[X]]
+; CHECK-NEXT: [[GEP:%.*]] = getelementptr [128 x i32], [128 x i32]* [[P:%.*]], i64 0, i64 0
+; CHECK-NEXT: [[X:%.*]] = addrspacecast i32* [[GEP]] to i32 addrspace(1)*
+; CHECK-NEXT: ret i32 addrspace(1)* [[X]]
;
- %x = addrspacecast ptr %p to ptr addrspace(1)
- ret ptr addrspace(1) %x
+ %gep = getelementptr [128 x i32], [128 x i32]* %p, i32 0, i32 0
+ %x = addrspacecast i32* %gep to i32 addrspace(1)*
+ ret i32 addrspace(1)* %x
}
-define <2 x ptr> @PR32414(ptr %ptr) {
+define <2 x i32*> @PR32414(i32** %ptr) {
; CHECK-LABEL: @PR32414(
-; CHECK-NEXT: [[T1:%.*]] = getelementptr inbounds i32, ptr [[PTR:%.*]], <2 x i64> <i64 0, i64 1>
-; CHECK-NEXT: ret <2 x ptr> [[T1]]
+; CHECK-NEXT: [[T0:%.*]] = bitcast i32** [[PTR:%.*]] to i32*
+; CHECK-NEXT: [[T1:%.*]] = getelementptr inbounds i32, i32* [[T0]], <2 x i64> <i64 0, i64 1>
+; CHECK-NEXT: ret <2 x i32*> [[T1]]
;
- %t1 = getelementptr inbounds i32, ptr %ptr, <2 x i64> <i64 0, i64 1>
- ret <2 x ptr> %t1
+ %t0 = bitcast i32** %ptr to i32*
+ %t1 = getelementptr inbounds i32, i32* %t0, <2 x i64> <i64 0, i64 1>
+ ret <2 x i32*> %t1
}
-define ptr @test_bitcast_nzgep(ptr %base, i64 %idx) {
+define i32* @test_bitcast_nzgep([1 x i32]* %base, i64 %idx) {
; CHECK-LABEL: @test_bitcast_nzgep(
-; CHECK-NEXT: [[PTR:%.*]] = getelementptr inbounds i32, ptr [[BASE:%.*]], i64 [[IDX:%.*]]
-; CHECK-NEXT: ret ptr [[PTR]]
+; CHECK-NEXT: [[PTR:%.*]] = getelementptr inbounds [1 x i32], [1 x i32]* [[BASE:%.*]], i64 0, i64 [[IDX:%.*]]
+; CHECK-NEXT: ret i32* [[PTR]]
;
- %ptr = getelementptr inbounds i32, ptr %base, i64 %idx
- ret ptr %ptr
+ %base2 = bitcast [1 x i32]* %base to i32*
+ %ptr = getelementptr inbounds i32, i32* %base2, i64 %idx
+ ret i32* %ptr
}
-define ptr @test_zgep_nzgep(ptr %base, i64 %idx) {
+define i32* @test_zgep_nzgep([1 x i32]* %base, i64 %idx) {
; CHECK-LABEL: @test_zgep_nzgep(
-; CHECK-NEXT: [[PTR:%.*]] = getelementptr inbounds i32, ptr [[BASE:%.*]], i64 [[IDX:%.*]]
-; CHECK-NEXT: ret ptr [[PTR]]
+; CHECK-NEXT: [[PTR:%.*]] = getelementptr inbounds [1 x i32], [1 x i32]* [[BASE:%.*]], i64 0, i64 [[IDX:%.*]]
+; CHECK-NEXT: ret i32* [[PTR]]
;
- %ptr = getelementptr inbounds i32, ptr %base, i64 %idx
- ret ptr %ptr
+ %base2 = getelementptr [1 x i32], [1 x i32]* %base, i64 0, i64 0
+ %ptr = getelementptr inbounds i32, i32* %base2, i64 %idx
+ ret i32* %ptr
}
-define ptr @test_nzgep_zgep(ptr %base, i64 %idx) {
+define i32* @test_nzgep_zgep([1 x i32]* %base, i64 %idx) {
; CHECK-LABEL: @test_nzgep_zgep(
-; CHECK-NEXT: [[BASE2:%.*]] = getelementptr inbounds [1 x i32], ptr [[BASE:%.*]], i64 [[IDX:%.*]]
-; CHECK-NEXT: ret ptr [[BASE2]]
+; CHECK-NEXT: [[PTR:%.*]] = getelementptr inbounds [1 x i32], [1 x i32]* [[BASE:%.*]], i64 [[IDX:%.*]], i64 0
+; CHECK-NEXT: ret i32* [[PTR]]
;
- %base2 = getelementptr inbounds [1 x i32], ptr %base, i64 %idx
- ret ptr %base2
+ %base2 = getelementptr inbounds [1 x i32], [1 x i32]* %base, i64 %idx
+ %ptr = getelementptr [1 x i32], [1 x i32]* %base2, i64 0, i64 0
+ ret i32* %ptr
}
-define ptr @test_gep_inbounds_of_gep(ptr %base) {
+define i32* @test_gep_inbounds_of_gep(i32* %base) {
; CHECK-LABEL: @test_gep_inbounds_of_gep(
-; CHECK-NEXT: [[PTR2:%.*]] = getelementptr i32, ptr [[BASE:%.*]], i64 8
-; CHECK-NEXT: ret ptr [[PTR2]]
+; CHECK-NEXT: [[PTR2:%.*]] = getelementptr i32, i32* [[BASE:%.*]], i64 8
+; CHECK-NEXT: ret i32* [[PTR2]]
;
- %ptr1 = getelementptr i32, ptr %base, i64 4
- %ptr2 = getelementptr inbounds i32, ptr %ptr1, i64 4
- ret ptr %ptr2
+ %ptr1 = getelementptr i32, i32* %base, i64 4
+ %ptr2 = getelementptr inbounds i32, i32* %ptr1, i64 4
+ ret i32* %ptr2
}
%struct.f = type { i32 }
@g0 = internal unnamed_addr constant %struct.f zeroinitializer, align 4
@g1 = internal unnamed_addr constant %struct.f { i32 -1 }, align 4
-define ptr @PR45084(i1 %cond) {
+define i32* @PR45084(i1 %cond) {
; CHECK-LABEL: @PR45084(
-; CHECK-NEXT: [[SEL:%.*]] = select i1 [[COND:%.*]], ptr @g0, ptr @g1, !prof [[PROF0:![0-9]+]]
-; CHECK-NEXT: ret ptr [[SEL]]
+; CHECK-NEXT: [[GEP:%.*]] = select i1 [[COND:%.*]], i32* getelementptr inbounds ([[STRUCT_F:%.*]], %struct.f* @g0, i64 0, i32 0), i32* getelementptr inbounds ([[STRUCT_F]], %struct.f* @g1, i64 0, i32 0), !prof [[PROF0:![0-9]+]]
+; CHECK-NEXT: ret i32* [[GEP]]
;
- %sel = select i1 %cond, ptr @g0, ptr @g1, !prof !0
- ret ptr %sel
+ %sel = select i1 %cond, %struct.f* @g0, %struct.f* @g1, !prof !0
+ %gep = getelementptr inbounds %struct.f, %struct.f* %sel, i64 0, i32 0
+ ret i32* %gep
}
-define ptr @PR45084_extra_use(i1 %cond, ptr %p) {
+define i32* @PR45084_extra_use(i1 %cond, %struct.f** %p) {
; CHECK-LABEL: @PR45084_extra_use(
-; CHECK-NEXT: [[SEL:%.*]] = select i1 [[COND:%.*]], ptr @g0, ptr @g1
-; CHECK-NEXT: store ptr [[SEL]], ptr [[P:%.*]], align 8
-; CHECK-NEXT: ret ptr [[SEL]]
+; CHECK-NEXT: [[SEL:%.*]] = select i1 [[COND:%.*]], %struct.f* @g0, %struct.f* @g1
+; CHECK-NEXT: store %struct.f* [[SEL]], %struct.f** [[P:%.*]], align 8
+; CHECK-NEXT: [[GEP:%.*]] = select i1 [[COND]], i32* getelementptr inbounds ([[STRUCT_F:%.*]], %struct.f* @g0, i64 0, i32 0), i32* getelementptr inbounds ([[STRUCT_F]], %struct.f* @g1, i64 0, i32 0)
+; CHECK-NEXT: ret i32* [[GEP]]
;
- %sel = select i1 %cond, ptr @g0, ptr @g1
- store ptr %sel, ptr %p
- ret ptr %sel
+ %sel = select i1 %cond, %struct.f* @g0, %struct.f* @g1
+ store %struct.f* %sel, %struct.f** %p
+ %gep = getelementptr %struct.f, %struct.f* %sel, i64 0, i32 0
+ ret i32* %gep
}
-define ptr @gep_null_inbounds(i64 %idx) {
+define i8* @gep_null_inbounds(i64 %idx) {
; CHECK-LABEL: @gep_null_inbounds(
-; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds i8, ptr null, i64 [[IDX:%.*]]
-; CHECK-NEXT: ret ptr [[GEP]]
+; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds i8, i8* null, i64 [[IDX:%.*]]
+; CHECK-NEXT: ret i8* [[GEP]]
;
- %gep = getelementptr inbounds i8, ptr null, i64 %idx
- ret ptr %gep
+ %gep = getelementptr inbounds i8, i8* null, i64 %idx
+ ret i8* %gep
}
-define ptr @gep_null_not_inbounds(i64 %idx) {
+define i8* @gep_null_not_inbounds(i64 %idx) {
; CHECK-LABEL: @gep_null_not_inbounds(
-; CHECK-NEXT: [[GEP:%.*]] = getelementptr i8, ptr null, i64 [[IDX:%.*]]
-; CHECK-NEXT: ret ptr [[GEP]]
+; CHECK-NEXT: [[GEP:%.*]] = getelementptr i8, i8* null, i64 [[IDX:%.*]]
+; CHECK-NEXT: ret i8* [[GEP]]
;
- %gep = getelementptr i8, ptr null, i64 %idx
- ret ptr %gep
+ %gep = getelementptr i8, i8* null, i64 %idx
+ ret i8* %gep
}
-define ptr @gep_null_defined(i64 %idx) null_pointer_is_valid {
+define i8* @gep_null_defined(i64 %idx) null_pointer_is_valid {
; CHECK-LABEL: @gep_null_defined(
-; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds i8, ptr null, i64 [[IDX:%.*]]
-; CHECK-NEXT: ret ptr [[GEP]]
+; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds i8, i8* null, i64 [[IDX:%.*]]
+; CHECK-NEXT: ret i8* [[GEP]]
;
- %gep = getelementptr inbounds i8, ptr null, i64 %idx
- ret ptr %gep
+ %gep = getelementptr inbounds i8, i8* null, i64 %idx
+ ret i8* %gep
}
-define ptr @gep_null_inbounds_different_type(i64 %idx1, i64 %idx2) {
+define i8* @gep_null_inbounds_different_type(i64 %idx1, i64 %idx2) {
; CHECK-LABEL: @gep_null_inbounds_different_type(
-; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds [0 x i8], ptr null, i64 0, i64 [[IDX2:%.*]]
-; CHECK-NEXT: ret ptr [[GEP]]
+; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds [0 x i8], [0 x i8]* null, i64 0, i64 [[IDX2:%.*]]
+; CHECK-NEXT: ret i8* [[GEP]]
;
- %gep = getelementptr inbounds [0 x i8], ptr null, i64 %idx1, i64 %idx2
- ret ptr %gep
+ %gep = getelementptr inbounds [0 x i8], [0 x i8]* null, i64 %idx1, i64 %idx2
+ ret i8* %gep
}
-define ptr @D98588(ptr %c1, i64 %offset) {
+define i8* @D98588(i8* %c1, i64 %offset) {
; CHECK-LABEL: @D98588(
; CHECK-NEXT: [[C2_NEXT_IDX:%.*]] = shl nsw i64 [[OFFSET:%.*]], 3
-; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds i8, ptr [[C1:%.*]], i64 [[C2_NEXT_IDX]]
-; CHECK-NEXT: ret ptr [[GEP]]
+; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds i8, i8* [[C1:%.*]], i64 [[C2_NEXT_IDX]]
+; CHECK-NEXT: ret i8* [[GEP]]
;
- %c2_next = getelementptr inbounds i64, ptr %c1, i64 %offset
- %ptrtoint1 = ptrtoint ptr %c1 to i64
- %ptrtoint2 = ptrtoint ptr %c2_next to i64
+ %c2 = bitcast i8* %c1 to i64*
+ %c2_next = getelementptr inbounds i64, i64* %c2, i64 %offset
+ %ptrtoint1 = ptrtoint i8* %c1 to i64
+ %ptrtoint2 = ptrtoint i64* %c2_next to i64
%sub = sub i64 %ptrtoint2, %ptrtoint1 ; C2 - C1
- %gep = getelementptr inbounds i8, ptr %c1, i64 %sub ; C1 + (C2 - C1)
- ret ptr %gep
+ %gep = getelementptr inbounds i8, i8* %c1, i64 %sub ; C1 + (C2 - C1)
+ ret i8* %gep
}
-declare noalias ptr @malloc(i64) nounwind allockind("alloc,uninitialized") allocsize(0)
+declare noalias i8* @malloc(i64) nounwind allockind("alloc,uninitialized") allocsize(0)
-define i32 @test_gep_bitcast_malloc(ptr %a) {
+define i32 @test_gep_bitcast_malloc(%struct.A* %a) {
; CHECK-LABEL: @test_gep_bitcast_malloc(
; CHECK-NEXT: entry:
-; CHECK-NEXT: [[CALL:%.*]] = call noalias dereferenceable_or_null(16) ptr @malloc(i64 16)
-; CHECK-NEXT: [[G3:%.*]] = getelementptr [[STRUCT_A:%.*]], ptr [[CALL]], i64 0, i32 2
-; CHECK-NEXT: [[A_C:%.*]] = load i32, ptr [[G3]], align 4
+; CHECK-NEXT: [[CALL:%.*]] = call noalias dereferenceable_or_null(16) i8* @malloc(i64 16)
+; CHECK-NEXT: [[B:%.*]] = bitcast i8* [[CALL]] to %struct.A*
+; CHECK-NEXT: [[G3:%.*]] = getelementptr [[STRUCT_A:%.*]], %struct.A* [[B]], i64 0, i32 2
+; CHECK-NEXT: [[A_C:%.*]] = load i32, i32* [[G3]], align 4
; CHECK-NEXT: ret i32 [[A_C]]
;
entry:
- %call = call noalias ptr @malloc(i64 16) #2
- %g3 = getelementptr %struct.A, ptr %call, i32 0, i32 2
- %a_c = load i32, ptr %g3, align 4
+ %call = call noalias i8* @malloc(i64 16) #2
+ %B = bitcast i8* %call to %struct.A*
+ %g3 = getelementptr %struct.A, %struct.A* %B, i32 0, i32 2
+ %a_c = load i32, i32* %g3, align 4
ret i32 %a_c
}
; Verify that memrchr calls with a string consisting of all the same
; characters are folded and those with mixed strings are not.
-declare ptr @memrchr(ptr, i32, i64)
+declare i8* @memrchr(i8*, i32, i64)
@a11111 = constant [5 x i8] c"\01\01\01\01\01"
@a1110111 = constant [7 x i8] c"\01\01\01\00\01\01\01"
; Fold memrchr(a11111, C, 5) to *a11111 == C ? a11111 + 5 - 1 : null.
-define ptr @fold_memrchr_a11111_c_5(i32 %C) {
+define i8* @fold_memrchr_a11111_c_5(i32 %C) {
; CHECK-LABEL: @fold_memrchr_a11111_c_5(
; CHECK-NEXT: [[TMP1:%.*]] = trunc i32 [[C:%.*]] to i8
; CHECK-NEXT: [[TMP2:%.*]] = icmp eq i8 [[TMP1]], 1
-; CHECK-NEXT: [[MEMRCHR_SEL:%.*]] = select i1 [[TMP2]], ptr getelementptr inbounds ([5 x i8], ptr @a11111, i64 0, i64 4), ptr null
-; CHECK-NEXT: ret ptr [[MEMRCHR_SEL]]
+; CHECK-NEXT: [[MEMRCHR_SEL:%.*]] = select i1 [[TMP2]], i8* getelementptr inbounds ([5 x i8], [5 x i8]* @a11111, i64 0, i64 4), i8* null
+; CHECK-NEXT: ret i8* [[MEMRCHR_SEL]]
;
- %ret = call ptr @memrchr(ptr @a11111, i32 %C, i64 5)
- ret ptr %ret
+ %ptr = getelementptr [5 x i8], [5 x i8]* @a11111, i64 0, i64 0
+ %ret = call i8* @memrchr(i8* %ptr, i32 %C, i64 5)
+ ret i8* %ret
}
; Fold memrchr(a11111, C, N) to N && *a11111 == C ? a11111 + N - 1 : null,
; on the assumption that N is in bounds.
-define ptr @fold_memrchr_a11111_c_n(i32 %C, i64 %N) {
+define i8* @fold_memrchr_a11111_c_n(i32 %C, i64 %N) {
; CHECK-LABEL: @fold_memrchr_a11111_c_n(
; CHECK-NEXT: [[TMP1:%.*]] = icmp ne i64 [[N:%.*]], 0
; CHECK-NEXT: [[TMP2:%.*]] = trunc i32 [[C:%.*]] to i8
; CHECK-NEXT: [[TMP3:%.*]] = icmp eq i8 [[TMP2]], 1
; CHECK-NEXT: [[TMP4:%.*]] = select i1 [[TMP1]], i1 [[TMP3]], i1 false
; CHECK-NEXT: [[TMP5:%.*]] = add i64 [[N]], -1
-; CHECK-NEXT: [[MEMRCHR_PTR_PLUS:%.*]] = getelementptr inbounds i8, ptr @a11111, i64 [[TMP5]]
-; CHECK-NEXT: [[MEMRCHR_SEL:%.*]] = select i1 [[TMP4]], ptr [[MEMRCHR_PTR_PLUS]], ptr null
-; CHECK-NEXT: ret ptr [[MEMRCHR_SEL]]
+; CHECK-NEXT: [[MEMRCHR_PTR_PLUS:%.*]] = getelementptr inbounds [5 x i8], [5 x i8]* @a11111, i64 0, i64 [[TMP5]]
+; CHECK-NEXT: [[MEMRCHR_SEL:%.*]] = select i1 [[TMP4]], i8* [[MEMRCHR_PTR_PLUS]], i8* null
+; CHECK-NEXT: ret i8* [[MEMRCHR_SEL]]
;
- %ret = call ptr @memrchr(ptr @a11111, i32 %C, i64 %N)
- ret ptr %ret
+ %ptr = getelementptr [5 x i8], [5 x i8]* @a11111, i64 0, i64 0
+ %ret = call i8* @memrchr(i8* %ptr, i32 %C, i64 %N)
+ ret i8* %ret
}
; Fold memrchr(a1110111, C, 3) to a1110111[2] == C ? a1110111 + 2 : null.
-define ptr @fold_memrchr_a1110111_c_3(i32 %C) {
+define i8* @fold_memrchr_a1110111_c_3(i32 %C) {
; CHECK-LABEL: @fold_memrchr_a1110111_c_3(
; CHECK-NEXT: [[TMP1:%.*]] = trunc i32 [[C:%.*]] to i8
; CHECK-NEXT: [[TMP2:%.*]] = icmp eq i8 [[TMP1]], 1
-; CHECK-NEXT: [[MEMRCHR_SEL:%.*]] = select i1 [[TMP2]], ptr getelementptr inbounds ([7 x i8], ptr @a1110111, i64 0, i64 2), ptr null
-; CHECK-NEXT: ret ptr [[MEMRCHR_SEL]]
+; CHECK-NEXT: [[MEMRCHR_SEL:%.*]] = select i1 [[TMP2]], i8* getelementptr inbounds ([7 x i8], [7 x i8]* @a1110111, i64 0, i64 2), i8* null
+; CHECK-NEXT: ret i8* [[MEMRCHR_SEL]]
;
- %ret = call ptr @memrchr(ptr @a1110111, i32 %C, i64 3)
- ret ptr %ret
+ %ptr = getelementptr [7 x i8], [7 x i8]* @a1110111, i64 0, i64 0
+ %ret = call i8* @memrchr(i8* %ptr, i32 %C, i64 3)
+ ret i8* %ret
}
; Don't fold memrchr(a1110111, C, 4).
-define ptr @call_memrchr_a1110111_c_4(i32 %C) {
+define i8* @call_memrchr_a1110111_c_4(i32 %C) {
; CHECK-LABEL: @call_memrchr_a1110111_c_4(
-; CHECK-NEXT: [[RET:%.*]] = call ptr @memrchr(ptr noundef nonnull dereferenceable(4) @a1110111, i32 [[C:%.*]], i64 4)
-; CHECK-NEXT: ret ptr [[RET]]
+; CHECK-NEXT: [[RET:%.*]] = call i8* @memrchr(i8* noundef nonnull dereferenceable(4) getelementptr inbounds ([7 x i8], [7 x i8]* @a1110111, i64 0, i64 0), i32 [[C:%.*]], i64 4)
+; CHECK-NEXT: ret i8* [[RET]]
;
- %ret = call ptr @memrchr(ptr @a1110111, i32 %C, i64 4)
- ret ptr %ret
+ %ptr = getelementptr [7 x i8], [7 x i8]* @a1110111, i64 0, i64 0
+ %ret = call i8* @memrchr(i8* %ptr, i32 %C, i64 4)
+ ret i8* %ret
}
; Don't fold memrchr(a1110111, C, 7).
-define ptr @call_memrchr_a1110111_c_7(i32 %C) {
+define i8* @call_memrchr_a1110111_c_7(i32 %C) {
; CHECK-LABEL: @call_memrchr_a1110111_c_7(
-; CHECK-NEXT: [[RET:%.*]] = call ptr @memrchr(ptr noundef nonnull dereferenceable(7) @a1110111, i32 [[C:%.*]], i64 7)
-; CHECK-NEXT: ret ptr [[RET]]
+; CHECK-NEXT: [[RET:%.*]] = call i8* @memrchr(i8* noundef nonnull dereferenceable(7) getelementptr inbounds ([7 x i8], [7 x i8]* @a1110111, i64 0, i64 0), i32 [[C:%.*]], i64 7)
+; CHECK-NEXT: ret i8* [[RET]]
;
- %ret = call ptr @memrchr(ptr @a1110111, i32 %C, i64 7)
- ret ptr %ret
+ %ptr = getelementptr [7 x i8], [7 x i8]* @a1110111, i64 0, i64 0
+ %ret = call i8* @memrchr(i8* %ptr, i32 %C, i64 7)
+ ret i8* %ret
}
; Don't fold memrchr(a1110111, C, N).
-define ptr @call_memrchr_a1110111_c_n(i32 %C, i64 %N) {
+define i8* @call_memrchr_a1110111_c_n(i32 %C, i64 %N) {
; CHECK-LABEL: @call_memrchr_a1110111_c_n(
-; CHECK-NEXT: [[RET:%.*]] = call ptr @memrchr(ptr nonnull @a1110111, i32 [[C:%.*]], i64 [[N:%.*]])
-; CHECK-NEXT: ret ptr [[RET]]
+; CHECK-NEXT: [[RET:%.*]] = call i8* @memrchr(i8* nonnull getelementptr inbounds ([7 x i8], [7 x i8]* @a1110111, i64 0, i64 0), i32 [[C:%.*]], i64 [[N:%.*]])
+; CHECK-NEXT: ret i8* [[RET]]
;
- %ret = call ptr @memrchr(ptr @a1110111, i32 %C, i64 %N)
- ret ptr %ret
+ %ptr = getelementptr [7 x i8], [7 x i8]* @a1110111, i64 0, i64 0
+ %ret = call i8* @memrchr(i8* %ptr, i32 %C, i64 %N)
+ ret i8* %ret
}
ret i1 %R
}
-define i32 @test16(i1 %C, ptr %P) {
+define i32 @test16(i1 %C, i32* %P) {
; CHECK-LABEL: @test16(
-; CHECK-NEXT: [[V:%.*]] = load i32, ptr [[P:%.*]], align 4
+; CHECK-NEXT: [[V:%.*]] = load i32, i32* [[P:%.*]], align 4
; CHECK-NEXT: ret i32 [[V]]
;
- %P2 = select i1 %C, ptr %P, ptr null
- %V = load i32, ptr %P2
+ %P2 = select i1 %C, i32* %P, i32* null
+ %V = load i32, i32* %P2
ret i32 %V
}
;; It may be legal to load from a null address in a non-zero address space
-define i32 @test16_neg(i1 %C, ptr addrspace(1) %P) {
+define i32 @test16_neg(i1 %C, i32 addrspace(1)* %P) {
; CHECK-LABEL: @test16_neg(
-; CHECK-NEXT: [[P2:%.*]] = select i1 [[C:%.*]], ptr addrspace(1) [[P:%.*]], ptr addrspace(1) null
-; CHECK-NEXT: [[V:%.*]] = load i32, ptr addrspace(1) [[P2]], align 4
+; CHECK-NEXT: [[P2:%.*]] = select i1 [[C:%.*]], i32 addrspace(1)* [[P:%.*]], i32 addrspace(1)* null
+; CHECK-NEXT: [[V:%.*]] = load i32, i32 addrspace(1)* [[P2]], align 4
; CHECK-NEXT: ret i32 [[V]]
;
- %P2 = select i1 %C, ptr addrspace(1) %P, ptr addrspace(1) null
- %V = load i32, ptr addrspace(1) %P2
+ %P2 = select i1 %C, i32 addrspace(1)* %P, i32 addrspace(1)* null
+ %V = load i32, i32 addrspace(1)* %P2
ret i32 %V
}
-define i32 @test16_neg2(i1 %C, ptr addrspace(1) %P) {
+define i32 @test16_neg2(i1 %C, i32 addrspace(1)* %P) {
; CHECK-LABEL: @test16_neg2(
-; CHECK-NEXT: [[P2:%.*]] = select i1 [[C:%.*]], ptr addrspace(1) null, ptr addrspace(1) [[P:%.*]]
-; CHECK-NEXT: [[V:%.*]] = load i32, ptr addrspace(1) [[P2]], align 4
+; CHECK-NEXT: [[P2:%.*]] = select i1 [[C:%.*]], i32 addrspace(1)* null, i32 addrspace(1)* [[P:%.*]]
+; CHECK-NEXT: [[V:%.*]] = load i32, i32 addrspace(1)* [[P2]], align 4
; CHECK-NEXT: ret i32 [[V]]
;
- %P2 = select i1 %C, ptr addrspace(1) null, ptr addrspace(1) %P
- %V = load i32, ptr addrspace(1) %P2
+ %P2 = select i1 %C, i32 addrspace(1)* null, i32 addrspace(1)* %P
+ %V = load i32, i32 addrspace(1)* %P2
ret i32 %V
}
;; It may be legal to load from a null address with null pointer valid attribute.
-define i32 @test16_no_null_opt(i1 %C, ptr %P) #0 {
+define i32 @test16_no_null_opt(i1 %C, i32* %P) #0 {
; CHECK-LABEL: @test16_no_null_opt(
-; CHECK-NEXT: [[P2:%.*]] = select i1 [[C:%.*]], ptr [[P:%.*]], ptr null
-; CHECK-NEXT: [[V:%.*]] = load i32, ptr [[P2]], align 4
+; CHECK-NEXT: [[P2:%.*]] = select i1 [[C:%.*]], i32* [[P:%.*]], i32* null
+; CHECK-NEXT: [[V:%.*]] = load i32, i32* [[P2]], align 4
; CHECK-NEXT: ret i32 [[V]]
;
- %P2 = select i1 %C, ptr %P, ptr null
- %V = load i32, ptr %P2
+ %P2 = select i1 %C, i32* %P, i32* null
+ %V = load i32, i32* %P2
ret i32 %V
}
-define i32 @test16_no_null_opt_2(i1 %C, ptr %P) #0 {
+define i32 @test16_no_null_opt_2(i1 %C, i32* %P) #0 {
; CHECK-LABEL: @test16_no_null_opt_2(
-; CHECK-NEXT: [[P2:%.*]] = select i1 [[C:%.*]], ptr null, ptr [[P:%.*]]
-; CHECK-NEXT: [[V:%.*]] = load i32, ptr [[P2]], align 4
+; CHECK-NEXT: [[P2:%.*]] = select i1 [[C:%.*]], i32* null, i32* [[P:%.*]]
+; CHECK-NEXT: [[V:%.*]] = load i32, i32* [[P2]], align 4
; CHECK-NEXT: ret i32 [[V]]
;
- %P2 = select i1 %C, ptr null, ptr %P
- %V = load i32, ptr %P2
+ %P2 = select i1 %C, i32* null, i32* %P
+ %V = load i32, i32* %P2
ret i32 %V
}
attributes #0 = { null_pointer_is_valid }
-define i1 @test17(ptr %X, i1 %C) {
+define i1 @test17(i32* %X, i1 %C) {
; CHECK-LABEL: @test17(
-; CHECK-NEXT: [[RV1:%.*]] = icmp eq ptr [[X:%.*]], null
+; CHECK-NEXT: [[RV1:%.*]] = icmp eq i32* [[X:%.*]], null
; CHECK-NEXT: [[NOT_C:%.*]] = xor i1 [[C:%.*]], true
; CHECK-NEXT: [[RV:%.*]] = select i1 [[NOT_C]], i1 true, i1 [[RV1]]
; CHECK-NEXT: ret i1 [[RV]]
;
- %R = select i1 %C, ptr %X, ptr null
- %RV = icmp eq ptr %R, null
+ %R = select i1 %C, i32* %X, i32* null
+ %RV = icmp eq i32* %R, null
ret i1 %RV
}
define i64 @test21(i32 %x) {
; CHECK-LABEL: @test21(
; CHECK-NEXT: [[X_LOBIT:%.*]] = ashr i32 [[X:%.*]], 31
-; CHECK-NEXT: [[RETVAL:%.*]] = sext i32 [[X_LOBIT]] to i64
-; CHECK-NEXT: ret i64 [[RETVAL]]
+; CHECK-NEXT: [[TMP1:%.*]] = sext i32 [[X_LOBIT]] to i64
+; CHECK-NEXT: ret i64 [[TMP1]]
;
%t = icmp slt i32 %x, 0
%retval = select i1 %t, i64 -1, i64 0
define i16 @test22(i32 %x) {
; CHECK-LABEL: @test22(
; CHECK-NEXT: [[X_LOBIT:%.*]] = ashr i32 [[X:%.*]], 31
-; CHECK-NEXT: [[RETVAL:%.*]] = trunc i32 [[X_LOBIT]] to i16
-; CHECK-NEXT: ret i16 [[RETVAL]]
+; CHECK-NEXT: [[TMP1:%.*]] = trunc i32 [[X_LOBIT]] to i16
+; CHECK-NEXT: ret i16 [[TMP1]]
;
%t = icmp slt i32 %x, 0
%retval = select i1 %t, i16 -1, i16 0
; SMAX(SMAX(x, y), x) -> SMAX(x, y)
define i32 @test30(i32 %x, i32 %y) {
; CHECK-LABEL: @test30(
-; CHECK-NEXT: [[COND:%.*]] = call i32 @llvm.smax.i32(i32 [[X:%.*]], i32 [[Y:%.*]])
-; CHECK-NEXT: ret i32 [[COND]]
+; CHECK-NEXT: [[TMP1:%.*]] = call i32 @llvm.smax.i32(i32 [[X:%.*]], i32 [[Y:%.*]])
+; CHECK-NEXT: ret i32 [[TMP1]]
;
%cmp = icmp sgt i32 %x, %y
%cond = select i1 %cmp, i32 %x, i32 %y
; UMAX(UMAX(x, y), x) -> UMAX(x, y)
define i32 @test31(i32 %x, i32 %y) {
; CHECK-LABEL: @test31(
-; CHECK-NEXT: [[COND:%.*]] = call i32 @llvm.umax.i32(i32 [[X:%.*]], i32 [[Y:%.*]])
-; CHECK-NEXT: ret i32 [[COND]]
+; CHECK-NEXT: [[TMP1:%.*]] = call i32 @llvm.umax.i32(i32 [[X:%.*]], i32 [[Y:%.*]])
+; CHECK-NEXT: ret i32 [[TMP1]]
;
%cmp = icmp ugt i32 %x, %y
%cond = select i1 %cmp, i32 %x, i32 %y
; SMIN(SMIN(x, y), x) -> SMIN(x, y)
define i32 @test32(i32 %x, i32 %y) {
; CHECK-LABEL: @test32(
-; CHECK-NEXT: [[COND:%.*]] = call i32 @llvm.smin.i32(i32 [[X:%.*]], i32 [[Y:%.*]])
-; CHECK-NEXT: ret i32 [[COND]]
+; CHECK-NEXT: [[TMP1:%.*]] = call i32 @llvm.smin.i32(i32 [[X:%.*]], i32 [[Y:%.*]])
+; CHECK-NEXT: ret i32 [[TMP1]]
;
%cmp = icmp sgt i32 %x, %y
%cond = select i1 %cmp, i32 %y, i32 %x
;
%zero = alloca i32
%one = alloca i32
- %ptr = select i1 %cond, ptr %zero, ptr %one
- %isnull = icmp eq ptr %ptr, null
+ %ptr = select i1 %cond, i32* %zero, i32* %one
+ %isnull = icmp eq i32* %ptr, null
ret i1 %isnull
}
%a = alloca i32
%b = alloca i32
%c = alloca i32
- %s = select i1 %cond, ptr %a, ptr %b
- %r = icmp eq ptr %s, %c
+ %s = select i1 %cond, i32* %a, i32* %b
+ %r = icmp eq i32* %s, %c
ret i1 %r
}
ret <vscale x 4 x float> %bc2
}
-define void @select_oneuse_bitcast(<vscale x 4 x float> %a, <vscale x 4 x float> %b, <vscale x 4 x i32> %c, <vscale x 4 x i32> %d, ptr %ptr1) {
+define void @select_oneuse_bitcast(<vscale x 4 x float> %a, <vscale x 4 x float> %b, <vscale x 4 x i32> %c, <vscale x 4 x i32> %d, <vscale x 4 x i32>* %ptr1) {
; CHECK-LABEL: @select_oneuse_bitcast(
; CHECK-NEXT: [[CMP:%.*]] = icmp ult <vscale x 4 x i32> [[C:%.*]], [[D:%.*]]
; CHECK-NEXT: [[SEL1_V:%.*]] = select <vscale x 4 x i1> [[CMP]], <vscale x 4 x float> [[A:%.*]], <vscale x 4 x float> [[B:%.*]]
-; CHECK-NEXT: store <vscale x 4 x float> [[SEL1_V]], ptr [[PTR1:%.*]], align 16
+; CHECK-NEXT: [[TMP1:%.*]] = bitcast <vscale x 4 x i32>* [[PTR1:%.*]] to <vscale x 4 x float>*
+; CHECK-NEXT: store <vscale x 4 x float> [[SEL1_V]], <vscale x 4 x float>* [[TMP1]], align 16
; CHECK-NEXT: ret void
;
%cmp = icmp ult <vscale x 4 x i32> %c, %d
%bc1 = bitcast <vscale x 4 x float> %a to <vscale x 4 x i32>
%bc2 = bitcast <vscale x 4 x float> %b to <vscale x 4 x i32>
%sel1 = select <vscale x 4 x i1> %cmp, <vscale x 4 x i32> %bc1, <vscale x 4 x i32> %bc2
- store <vscale x 4 x i32> %sel1, ptr %ptr1
+ store <vscale x 4 x i32> %sel1, <vscale x 4 x i32>* %ptr1
ret void
}
; Allow select promotion even if there are multiple uses of bitcasted ops.
; Hoisting the selects allows later pattern matching to see that these are min/max ops.
-define void @min_max_bitcast(<4 x float> %a, <4 x float> %b, ptr %ptr1, ptr %ptr2) {
+define void @min_max_bitcast(<4 x float> %a, <4 x float> %b, <4 x i32>* %ptr1, <4 x i32>* %ptr2) {
; CHECK-LABEL: @min_max_bitcast(
; CHECK-NEXT: [[CMP:%.*]] = fcmp olt <4 x float> [[A:%.*]], [[B:%.*]]
; CHECK-NEXT: [[SEL1_V:%.*]] = select <4 x i1> [[CMP]], <4 x float> [[A]], <4 x float> [[B]]
; CHECK-NEXT: [[SEL2_V:%.*]] = select <4 x i1> [[CMP]], <4 x float> [[B]], <4 x float> [[A]]
-; CHECK-NEXT: store <4 x float> [[SEL1_V]], ptr [[PTR1:%.*]], align 16
-; CHECK-NEXT: store <4 x float> [[SEL2_V]], ptr [[PTR2:%.*]], align 16
+; CHECK-NEXT: [[TMP1:%.*]] = bitcast <4 x i32>* [[PTR1:%.*]] to <4 x float>*
+; CHECK-NEXT: store <4 x float> [[SEL1_V]], <4 x float>* [[TMP1]], align 16
+; CHECK-NEXT: [[TMP2:%.*]] = bitcast <4 x i32>* [[PTR2:%.*]] to <4 x float>*
+; CHECK-NEXT: store <4 x float> [[SEL2_V]], <4 x float>* [[TMP2]], align 16
; CHECK-NEXT: ret void
;
%cmp = fcmp olt <4 x float> %a, %b
%bc2 = bitcast <4 x float> %b to <4 x i32>
%sel1 = select <4 x i1> %cmp, <4 x i32> %bc1, <4 x i32> %bc2
%sel2 = select <4 x i1> %cmp, <4 x i32> %bc2, <4 x i32> %bc1
- store <4 x i32> %sel1, ptr %ptr1
- store <4 x i32> %sel2, ptr %ptr2
+ store <4 x i32> %sel1, <4 x i32>* %ptr1
+ store <4 x i32> %sel2, <4 x i32>* %ptr2
ret void
}
-define void @min_max_bitcast1(<vscale x 4 x float> %a, <vscale x 4 x float> %b, ptr %ptr1, ptr %ptr2) {
+define void @min_max_bitcast1(<vscale x 4 x float> %a, <vscale x 4 x float> %b, <vscale x 4 x i32>* %ptr1, <vscale x 4 x i32>* %ptr2) {
; CHECK-LABEL: @min_max_bitcast1(
; CHECK-NEXT: [[CMP:%.*]] = fcmp olt <vscale x 4 x float> [[A:%.*]], [[B:%.*]]
; CHECK-NEXT: [[SEL1_V:%.*]] = select <vscale x 4 x i1> [[CMP]], <vscale x 4 x float> [[A]], <vscale x 4 x float> [[B]]
; CHECK-NEXT: [[SEL2_V:%.*]] = select <vscale x 4 x i1> [[CMP]], <vscale x 4 x float> [[B]], <vscale x 4 x float> [[A]]
-; CHECK-NEXT: store <vscale x 4 x float> [[SEL1_V]], ptr [[PTR1:%.*]], align 16
-; CHECK-NEXT: store <vscale x 4 x float> [[SEL2_V]], ptr [[PTR2:%.*]], align 16
+; CHECK-NEXT: [[TMP1:%.*]] = bitcast <vscale x 4 x i32>* [[PTR1:%.*]] to <vscale x 4 x float>*
+; CHECK-NEXT: store <vscale x 4 x float> [[SEL1_V]], <vscale x 4 x float>* [[TMP1]], align 16
+; CHECK-NEXT: [[TMP2:%.*]] = bitcast <vscale x 4 x i32>* [[PTR2:%.*]] to <vscale x 4 x float>*
+; CHECK-NEXT: store <vscale x 4 x float> [[SEL2_V]], <vscale x 4 x float>* [[TMP2]], align 16
; CHECK-NEXT: ret void
;
%cmp = fcmp olt <vscale x 4 x float> %a, %b
%bc2 = bitcast <vscale x 4 x float> %b to <vscale x 4 x i32>
%sel1 = select <vscale x 4 x i1> %cmp, <vscale x 4 x i32> %bc1, <vscale x 4 x i32> %bc2
%sel2 = select <vscale x 4 x i1> %cmp, <vscale x 4 x i32> %bc2, <vscale x 4 x i32> %bc1
- store <vscale x 4 x i32> %sel1, ptr %ptr1
- store <vscale x 4 x i32> %sel2, ptr %ptr2
+ store <vscale x 4 x i32> %sel1, <vscale x 4 x i32>* %ptr1
+ store <vscale x 4 x i32> %sel2, <vscale x 4 x i32>* %ptr2
ret void
}
; To avoid potential backend problems, we don't do the same transform for other casts.
-define void @truncs_before_selects(<4 x float> %f1, <4 x float> %f2, <4 x i64> %a, <4 x i64> %b, ptr %ptr1, ptr %ptr2) {
+define void @truncs_before_selects(<4 x float> %f1, <4 x float> %f2, <4 x i64> %a, <4 x i64> %b, <4 x i32>* %ptr1, <4 x i32>* %ptr2) {
; CHECK-LABEL: @truncs_before_selects(
; CHECK-NEXT: [[CMP:%.*]] = fcmp olt <4 x float> [[F1:%.*]], [[F2:%.*]]
; CHECK-NEXT: [[BC1:%.*]] = trunc <4 x i64> [[A:%.*]] to <4 x i32>
; CHECK-NEXT: [[BC2:%.*]] = trunc <4 x i64> [[B:%.*]] to <4 x i32>
; CHECK-NEXT: [[SEL1:%.*]] = select <4 x i1> [[CMP]], <4 x i32> [[BC1]], <4 x i32> [[BC2]]
; CHECK-NEXT: [[SEL2:%.*]] = select <4 x i1> [[CMP]], <4 x i32> [[BC2]], <4 x i32> [[BC1]]
-; CHECK-NEXT: store <4 x i32> [[SEL1]], ptr [[PTR1:%.*]], align 16
-; CHECK-NEXT: store <4 x i32> [[SEL2]], ptr [[PTR2:%.*]], align 16
+; CHECK-NEXT: store <4 x i32> [[SEL1]], <4 x i32>* [[PTR1:%.*]], align 16
+; CHECK-NEXT: store <4 x i32> [[SEL2]], <4 x i32>* [[PTR2:%.*]], align 16
; CHECK-NEXT: ret void
;
%cmp = fcmp olt <4 x float> %f1, %f2
%bc2 = trunc <4 x i64> %b to <4 x i32>
%sel1 = select <4 x i1> %cmp, <4 x i32> %bc1, <4 x i32> %bc2
%sel2 = select <4 x i1> %cmp, <4 x i32> %bc2, <4 x i32> %bc1
- store <4 x i32> %sel1, ptr %ptr1, align 16
- store <4 x i32> %sel2, ptr %ptr2, align 16
+ store <4 x i32> %sel1, <4 x i32>* %ptr1, align 16
+ store <4 x i32> %sel2, <4 x i32>* %ptr2, align 16
ret void
}
ret i32 %.and
}
-define i1 @test60(i32 %x, ptr %y) {
+define i1 @test60(i32 %x, i1* %y) {
; CHECK-LABEL: @test60(
; CHECK-NEXT: [[CMP:%.*]] = icmp eq i32 [[X:%.*]], 0
-; CHECK-NEXT: [[LOAD:%.*]] = load i1, ptr [[Y:%.*]], align 1
+; CHECK-NEXT: [[LOAD:%.*]] = load i1, i1* [[Y:%.*]], align 1
; CHECK-NEXT: [[CMP1:%.*]] = icmp slt i32 [[X]], 1
; CHECK-NEXT: [[SEL:%.*]] = select i1 [[CMP]], i1 [[LOAD]], i1 [[CMP1]]
; CHECK-NEXT: ret i1 [[SEL]]
;
%cmp = icmp eq i32 %x, 0
- %load = load i1, ptr %y, align 1
+ %load = load i1, i1* %y, align 1
%cmp1 = icmp slt i32 %x, 1
%sel = select i1 %cmp, i1 %load, i1 %cmp1
ret i1 %sel
}
@glbl = constant i32 10
-define i32 @test61(ptr %ptr) {
+define i32 @test61(i32* %ptr) {
; CHECK-LABEL: @test61(
; CHECK-NEXT: ret i32 10
;
- %A = load i32, ptr %ptr
- %B = icmp eq ptr %ptr, @glbl
+ %A = load i32, i32* %ptr
+ %B = icmp eq i32* %ptr, @glbl
%C = select i1 %B, i32 %A, i32 10
ret i32 %C
}
; The load here must not be speculated around the select. One side of the
; select is trivially dereferenceable but may have a lower alignment than the
; load does.
-define i32 @test76(i1 %flag, ptr %x) {
+define i32 @test76(i1 %flag, i32* %x) {
; CHECK-LABEL: @test76(
-; CHECK-NEXT: store i32 0, ptr [[X:%.*]], align 4
-; CHECK-NEXT: [[P:%.*]] = select i1 [[FLAG:%.*]], ptr @under_aligned, ptr [[X]]
-; CHECK-NEXT: [[V:%.*]] = load i32, ptr [[P]], align 4
+; CHECK-NEXT: store i32 0, i32* [[X:%.*]], align 4
+; CHECK-NEXT: [[P:%.*]] = select i1 [[FLAG:%.*]], i32* @under_aligned, i32* [[X]]
+; CHECK-NEXT: [[V:%.*]] = load i32, i32* [[P]], align 4
; CHECK-NEXT: ret i32 [[V]]
;
- store i32 0, ptr %x
- %p = select i1 %flag, ptr @under_aligned, ptr %x
- %v = load i32, ptr %p
+ store i32 0, i32* %x
+ %p = select i1 %flag, i32* @under_aligned, i32* %x
+ %v = load i32, i32* %p
ret i32 %v
}
-declare void @scribble_on_i32(ptr)
+declare void @scribble_on_i32(i32*)
; The load here must not be speculated around the select. One side of the
; select is trivially dereferenceable but may have a lower alignment than the
; load does.
-define i32 @test77(i1 %flag, ptr %x) {
+define i32 @test77(i1 %flag, i32* %x) {
; CHECK-LABEL: @test77(
; CHECK-NEXT: [[UNDER_ALIGNED:%.*]] = alloca i32, align 1
-; CHECK-NEXT: call void @scribble_on_i32(ptr nonnull [[UNDER_ALIGNED]])
-; CHECK-NEXT: store i32 0, ptr [[X:%.*]], align 4
-; CHECK-NEXT: [[P:%.*]] = select i1 [[FLAG:%.*]], ptr [[UNDER_ALIGNED]], ptr [[X]]
-; CHECK-NEXT: [[V:%.*]] = load i32, ptr [[P]], align 4
+; CHECK-NEXT: call void @scribble_on_i32(i32* nonnull [[UNDER_ALIGNED]])
+; CHECK-NEXT: store i32 0, i32* [[X:%.*]], align 4
+; CHECK-NEXT: [[P:%.*]] = select i1 [[FLAG:%.*]], i32* [[UNDER_ALIGNED]], i32* [[X]]
+; CHECK-NEXT: [[V:%.*]] = load i32, i32* [[P]], align 4
; CHECK-NEXT: ret i32 [[V]]
;
%under_aligned = alloca i32, align 1
- call void @scribble_on_i32(ptr %under_aligned)
- store i32 0, ptr %x
- %p = select i1 %flag, ptr %under_aligned, ptr %x
- %v = load i32, ptr %p
+ call void @scribble_on_i32(i32* %under_aligned)
+ store i32 0, i32* %x
+ %p = select i1 %flag, i32* %under_aligned, i32* %x
+ %v = load i32, i32* %p
ret i32 %v
}
-define i32 @test78(i1 %flag, ptr %x, ptr %y, ptr %z) {
+define i32 @test78(i1 %flag, i32* %x, i32* %y, i32* %z) {
; Test that we can speculate the loads around the select even when we can't
; fold the load completely away.
; CHECK-LABEL: @test78(
; CHECK-NEXT: entry:
-; CHECK-NEXT: store i32 0, ptr [[X:%.*]], align 4
-; CHECK-NEXT: store i32 0, ptr [[Y:%.*]], align 4
-; CHECK-NEXT: store i32 42, ptr [[Z:%.*]], align 4
-; CHECK-NEXT: [[X_VAL:%.*]] = load i32, ptr [[X]], align 4
-; CHECK-NEXT: [[Y_VAL:%.*]] = load i32, ptr [[Y]], align 4
+; CHECK-NEXT: store i32 0, i32* [[X:%.*]], align 4
+; CHECK-NEXT: store i32 0, i32* [[Y:%.*]], align 4
+; CHECK-NEXT: store i32 42, i32* [[Z:%.*]], align 4
+; CHECK-NEXT: [[X_VAL:%.*]] = load i32, i32* [[X]], align 4
+; CHECK-NEXT: [[Y_VAL:%.*]] = load i32, i32* [[Y]], align 4
; CHECK-NEXT: [[V:%.*]] = select i1 [[FLAG:%.*]], i32 [[X_VAL]], i32 [[Y_VAL]]
; CHECK-NEXT: ret i32 [[V]]
;
entry:
- store i32 0, ptr %x
- store i32 0, ptr %y
+ store i32 0, i32* %x
+ store i32 0, i32* %y
; Block forwarding by storing to %z which could alias either %x or %y.
- store i32 42, ptr %z
- %p = select i1 %flag, ptr %x, ptr %y
- %v = load i32, ptr %p
+ store i32 42, i32* %z
+ %p = select i1 %flag, i32* %x, i32* %y
+ %v = load i32, i32* %p
ret i32 %v
}
; Test that we can speculate the loads around the select even when we can't
; fold the load completely away.
-define i32 @test78_deref(i1 %flag, ptr dereferenceable(4) align 4 %x, ptr dereferenceable(4) align 4 %y, ptr %z) nofree nosync {
+define i32 @test78_deref(i1 %flag, i32* dereferenceable(4) align 4 %x, i32* dereferenceable(4) align 4 %y, i32* %z) nofree nosync {
; CHECK-LABEL: @test78_deref(
-; CHECK-NEXT: [[X_VAL:%.*]] = load i32, ptr [[X:%.*]], align 4
-; CHECK-NEXT: [[Y_VAL:%.*]] = load i32, ptr [[Y:%.*]], align 4
+; CHECK-NEXT: [[X_VAL:%.*]] = load i32, i32* [[X:%.*]], align 4
+; CHECK-NEXT: [[Y_VAL:%.*]] = load i32, i32* [[Y:%.*]], align 4
; CHECK-NEXT: [[V:%.*]] = select i1 [[FLAG:%.*]], i32 [[X_VAL]], i32 [[Y_VAL]]
; CHECK-NEXT: ret i32 [[V]]
;
- %p = select i1 %flag, ptr %x, ptr %y
- %v = load i32, ptr %p
+ %p = select i1 %flag, i32* %x, i32* %y
+ %v = load i32, i32* %p
ret i32 %v
}
; The same as @test78 but we can't speculate the load because it can trap
; if under-aligned.
-define i32 @test78_neg(i1 %flag, ptr %x, ptr %y, ptr %z) {
+define i32 @test78_neg(i1 %flag, i32* %x, i32* %y, i32* %z) {
; CHECK-LABEL: @test78_neg(
-; CHECK-NEXT: store i32 0, ptr [[X:%.*]], align 4
-; CHECK-NEXT: store i32 0, ptr [[Y:%.*]], align 4
-; CHECK-NEXT: store i32 42, ptr [[Z:%.*]], align 4
-; CHECK-NEXT: [[P:%.*]] = select i1 [[FLAG:%.*]], ptr [[X]], ptr [[Y]]
-; CHECK-NEXT: [[V:%.*]] = load i32, ptr [[P]], align 16
+; CHECK-NEXT: store i32 0, i32* [[X:%.*]], align 4
+; CHECK-NEXT: store i32 0, i32* [[Y:%.*]], align 4
+; CHECK-NEXT: store i32 42, i32* [[Z:%.*]], align 4
+; CHECK-NEXT: [[P:%.*]] = select i1 [[FLAG:%.*]], i32* [[X]], i32* [[Y]]
+; CHECK-NEXT: [[V:%.*]] = load i32, i32* [[P]], align 16
; CHECK-NEXT: ret i32 [[V]]
;
- store i32 0, ptr %x
- store i32 0, ptr %y
+ store i32 0, i32* %x
+ store i32 0, i32* %y
; Block forwarding by storing to %z which could alias either %x or %y.
- store i32 42, ptr %z
- %p = select i1 %flag, ptr %x, ptr %y
- %v = load i32, ptr %p, align 16
+ store i32 42, i32* %z
+ %p = select i1 %flag, i32* %x, i32* %y
+ %v = load i32, i32* %p, align 16
ret i32 %v
}
; The same as @test78_deref but we can't speculate the load because
; one of the arguments is not sufficiently dereferenceable.
-define i32 @test78_deref_neg(i1 %flag, ptr dereferenceable(2) %x, ptr dereferenceable(4) %y, ptr %z) nofree nosync {
+define i32 @test78_deref_neg(i1 %flag, i32* dereferenceable(2) %x, i32* dereferenceable(4) %y, i32* %z) nofree nosync {
; CHECK-LABEL: @test78_deref_neg(
-; CHECK-NEXT: [[P:%.*]] = select i1 [[FLAG:%.*]], ptr [[X:%.*]], ptr [[Y:%.*]]
-; CHECK-NEXT: [[V:%.*]] = load i32, ptr [[P]], align 4
+; CHECK-NEXT: [[P:%.*]] = select i1 [[FLAG:%.*]], i32* [[X:%.*]], i32* [[Y:%.*]]
+; CHECK-NEXT: [[V:%.*]] = load i32, i32* [[P]], align 4
; CHECK-NEXT: ret i32 [[V]]
;
- %p = select i1 %flag, ptr %x, ptr %y
- %v = load i32, ptr %p
+ %p = select i1 %flag, i32* %x, i32* %y
+ %v = load i32, i32* %p
ret i32 %v
}
; Test that we can speculate the loads around the select even when we can't
; fold the load completely away.
-define float @test79(i1 %flag, ptr %x, ptr %y, ptr %z) {
+define float @test79(i1 %flag, float* %x, i32* %y, i32* %z) {
; CHECK-LABEL: @test79(
-; CHECK-NEXT: store i32 0, ptr [[X:%.*]], align 4
-; CHECK-NEXT: store i32 0, ptr [[Y:%.*]], align 4
-; CHECK-NEXT: store i32 42, ptr [[Z:%.*]], align 4
-; CHECK-NEXT: [[X_VAL:%.*]] = load float, ptr [[X]], align 4
-; CHECK-NEXT: [[Y_VAL:%.*]] = load float, ptr [[Y]], align 4
-; CHECK-NEXT: [[V:%.*]] = select i1 [[FLAG:%.*]], float [[X_VAL]], float [[Y_VAL]]
+; CHECK-NEXT: [[X1:%.*]] = bitcast float* [[X:%.*]] to i32*
+; CHECK-NEXT: [[Y1:%.*]] = bitcast i32* [[Y:%.*]] to float*
+; CHECK-NEXT: store i32 0, i32* [[X1]], align 4
+; CHECK-NEXT: store i32 0, i32* [[Y]], align 4
+; CHECK-NEXT: store i32 42, i32* [[Z:%.*]], align 4
+; CHECK-NEXT: [[X_VAL:%.*]] = load float, float* [[X]], align 4
+; CHECK-NEXT: [[Y1_VAL:%.*]] = load float, float* [[Y1]], align 4
+; CHECK-NEXT: [[V:%.*]] = select i1 [[FLAG:%.*]], float [[X_VAL]], float [[Y1_VAL]]
; CHECK-NEXT: ret float [[V]]
;
- store i32 0, ptr %x
- store i32 0, ptr %y
+ %x1 = bitcast float* %x to i32*
+ %y1 = bitcast i32* %y to float*
+ store i32 0, i32* %x1
+ store i32 0, i32* %y
; Block forwarding by storing to %z which could alias either %x or %y.
- store i32 42, ptr %z
- %p = select i1 %flag, ptr %x, ptr %y
- %v = load float, ptr %p
+ store i32 42, i32* %z
+ %p = select i1 %flag, float* %x, float* %y1
+ %v = load float, float* %p
ret float %v
}
; CHECK-LABEL: @test80(
; CHECK-NEXT: [[X:%.*]] = alloca i32, align 4
; CHECK-NEXT: [[Y:%.*]] = alloca i32, align 4
-; CHECK-NEXT: call void @scribble_on_i32(ptr nonnull [[X]])
-; CHECK-NEXT: call void @scribble_on_i32(ptr nonnull [[Y]])
-; CHECK-NEXT: [[T:%.*]] = load i32, ptr [[X]], align 4
-; CHECK-NEXT: store i32 [[T]], ptr [[Y]], align 4
+; CHECK-NEXT: call void @scribble_on_i32(i32* nonnull [[X]])
+; CHECK-NEXT: call void @scribble_on_i32(i32* nonnull [[Y]])
+; CHECK-NEXT: [[T:%.*]] = load i32, i32* [[X]], align 4
+; CHECK-NEXT: store i32 [[T]], i32* [[Y]], align 4
; CHECK-NEXT: ret i32 [[T]]
;
%x = alloca i32
%y = alloca i32
- call void @scribble_on_i32(ptr %x)
- call void @scribble_on_i32(ptr %y)
- %t = load i32, ptr %x
- store i32 %t, ptr %y
- %p = select i1 %flag, ptr %x, ptr %y
- %v = load i32, ptr %p
+ call void @scribble_on_i32(i32* %x)
+ call void @scribble_on_i32(i32* %y)
+ %t = load i32, i32* %x
+ store i32 %t, i32* %y
+ %p = select i1 %flag, i32* %x, i32* %y
+ %v = load i32, i32* %p
ret i32 %v
}
; differently typed pointers.
define float @test81(i1 %flag) {
; CHECK-LABEL: @test81(
-; CHECK-NEXT: [[X:%.*]] = alloca float, align 4
+; CHECK-NEXT: [[X:%.*]] = alloca i32, align 4
; CHECK-NEXT: [[Y:%.*]] = alloca i32, align 4
-; CHECK-NEXT: call void @scribble_on_i32(ptr nonnull [[X]])
-; CHECK-NEXT: call void @scribble_on_i32(ptr nonnull [[Y]])
-; CHECK-NEXT: [[T:%.*]] = load i32, ptr [[X]], align 4
-; CHECK-NEXT: store i32 [[T]], ptr [[Y]], align 4
+; CHECK-NEXT: call void @scribble_on_i32(i32* nonnull [[X]])
+; CHECK-NEXT: call void @scribble_on_i32(i32* nonnull [[Y]])
+; CHECK-NEXT: [[T:%.*]] = load i32, i32* [[X]], align 4
+; CHECK-NEXT: store i32 [[T]], i32* [[Y]], align 4
; CHECK-NEXT: [[V:%.*]] = bitcast i32 [[T]] to float
; CHECK-NEXT: ret float [[V]]
;
%x = alloca float
%y = alloca i32
- call void @scribble_on_i32(ptr %x)
- call void @scribble_on_i32(ptr %y)
- %t = load i32, ptr %x
- store i32 %t, ptr %y
- %p = select i1 %flag, ptr %x, ptr %y
- %v = load float, ptr %p
+ %x1 = bitcast float* %x to i32*
+ %y1 = bitcast i32* %y to float*
+ call void @scribble_on_i32(i32* %x1)
+ call void @scribble_on_i32(i32* %y)
+ %t = load i32, i32* %x1
+ store i32 %t, i32* %y
+ %p = select i1 %flag, float* %x, float* %y1
+ %v = load float, float* %p
ret float %v
}
; CHECK-LABEL: @test82(
; CHECK-NEXT: [[X:%.*]] = alloca float, align 4
; CHECK-NEXT: [[Y:%.*]] = alloca i32, align 4
-; CHECK-NEXT: call void @scribble_on_i32(ptr nonnull [[X]])
-; CHECK-NEXT: call void @scribble_on_i32(ptr nonnull [[Y]])
-; CHECK-NEXT: [[T:%.*]] = load float, ptr [[X]], align 4
-; CHECK-NEXT: store float [[T]], ptr [[Y]], align 4
+; CHECK-NEXT: [[X1:%.*]] = bitcast float* [[X]] to i32*
+; CHECK-NEXT: [[Y1:%.*]] = bitcast i32* [[Y]] to float*
+; CHECK-NEXT: call void @scribble_on_i32(i32* nonnull [[X1]])
+; CHECK-NEXT: call void @scribble_on_i32(i32* nonnull [[Y]])
+; CHECK-NEXT: [[T:%.*]] = load float, float* [[X]], align 4
+; CHECK-NEXT: store float [[T]], float* [[Y1]], align 4
; CHECK-NEXT: [[V:%.*]] = bitcast float [[T]] to i32
; CHECK-NEXT: ret i32 [[V]]
;
%x = alloca float
%y = alloca i32
- call void @scribble_on_i32(ptr %x)
- call void @scribble_on_i32(ptr %y)
- %t = load float, ptr %x
- store float %t, ptr %y
- %p = select i1 %flag, ptr %x, ptr %y
- %v = load i32, ptr %p
+ %x1 = bitcast float* %x to i32*
+ %y1 = bitcast i32* %y to float*
+ call void @scribble_on_i32(i32* %x1)
+ call void @scribble_on_i32(i32* %y)
+ %t = load float, float* %x
+ store float %t, float* %y1
+ %p = select i1 %flag, i32* %x1, i32* %y
+ %v = load i32, i32* %p
ret i32 %v
}
-declare void @scribble_on_i64(ptr)
-declare void @scribble_on_i128(ptr)
+declare void @scribble_on_i64(i64*)
+declare void @scribble_on_i128(i128*)
; Test that we can speculate the load around the select even though they use
; differently typed pointers and requires inttoptr casts.
-define ptr @test83(i1 %flag) {
+define i8* @test83(i1 %flag) {
; CHECK-LABEL: @test83(
-; CHECK-NEXT: [[X:%.*]] = alloca ptr, align 8
-; CHECK-NEXT: [[Y:%.*]] = alloca i64, align 8
-; CHECK-NEXT: call void @scribble_on_i64(ptr nonnull [[X]])
-; CHECK-NEXT: call void @scribble_on_i64(ptr nonnull [[Y]])
-; CHECK-NEXT: [[T:%.*]] = load i64, ptr [[X]], align 8
-; CHECK-NEXT: store i64 [[T]], ptr [[Y]], align 8
-; CHECK-NEXT: [[V:%.*]] = inttoptr i64 [[T]] to ptr
-; CHECK-NEXT: ret ptr [[V]]
-;
- %x = alloca ptr
+; CHECK-NEXT: [[X:%.*]] = alloca i8*, align 8
+; CHECK-NEXT: [[Y:%.*]] = alloca i8*, align 8
+; CHECK-NEXT: [[TMPCAST:%.*]] = bitcast i8** [[Y]] to i64*
+; CHECK-NEXT: [[X1:%.*]] = bitcast i8** [[X]] to i64*
+; CHECK-NEXT: call void @scribble_on_i64(i64* nonnull [[X1]])
+; CHECK-NEXT: call void @scribble_on_i64(i64* nonnull [[TMPCAST]])
+; CHECK-NEXT: [[T:%.*]] = load i64, i64* [[X1]], align 8
+; CHECK-NEXT: store i64 [[T]], i64* [[TMPCAST]], align 8
+; CHECK-NEXT: [[V:%.*]] = inttoptr i64 [[T]] to i8*
+; CHECK-NEXT: ret i8* [[V]]
+;
+ %x = alloca i8*
%y = alloca i64
- call void @scribble_on_i64(ptr %x)
- call void @scribble_on_i64(ptr %y)
- %t = load i64, ptr %x
- store i64 %t, ptr %y
- %p = select i1 %flag, ptr %x, ptr %y
- %v = load ptr, ptr %p
- ret ptr %v
+ %x1 = bitcast i8** %x to i64*
+ %y1 = bitcast i64* %y to i8**
+ call void @scribble_on_i64(i64* %x1)
+ call void @scribble_on_i64(i64* %y)
+ %t = load i64, i64* %x1
+ store i64 %t, i64* %y
+ %p = select i1 %flag, i8** %x, i8** %y1
+ %v = load i8*, i8** %p
+ ret i8* %v
}
; Test that we can speculate the load around the select even though they use
; differently typed pointers and requires a ptrtoint cast.
define i64 @test84(i1 %flag) {
; CHECK-LABEL: @test84(
-; CHECK-NEXT: [[X:%.*]] = alloca ptr, align 8
-; CHECK-NEXT: [[Y:%.*]] = alloca i64, align 8
-; CHECK-NEXT: call void @scribble_on_i64(ptr nonnull [[X]])
-; CHECK-NEXT: call void @scribble_on_i64(ptr nonnull [[Y]])
-; CHECK-NEXT: [[T:%.*]] = load ptr, ptr [[X]], align 8
-; CHECK-NEXT: store ptr [[T]], ptr [[Y]], align 8
-; CHECK-NEXT: [[V:%.*]] = ptrtoint ptr [[T]] to i64
+; CHECK-NEXT: [[X:%.*]] = alloca i8*, align 8
+; CHECK-NEXT: [[Y:%.*]] = alloca i8*, align 8
+; CHECK-NEXT: [[TMPCAST:%.*]] = bitcast i8** [[Y]] to i64*
+; CHECK-NEXT: [[X1:%.*]] = bitcast i8** [[X]] to i64*
+; CHECK-NEXT: call void @scribble_on_i64(i64* nonnull [[X1]])
+; CHECK-NEXT: call void @scribble_on_i64(i64* nonnull [[TMPCAST]])
+; CHECK-NEXT: [[T:%.*]] = load i8*, i8** [[X]], align 8
+; CHECK-NEXT: store i8* [[T]], i8** [[Y]], align 8
+; CHECK-NEXT: [[V:%.*]] = ptrtoint i8* [[T]] to i64
; CHECK-NEXT: ret i64 [[V]]
;
- %x = alloca ptr
+ %x = alloca i8*
%y = alloca i64
- call void @scribble_on_i64(ptr %x)
- call void @scribble_on_i64(ptr %y)
- %t = load ptr, ptr %x
- store ptr %t, ptr %y
- %p = select i1 %flag, ptr %x, ptr %y
- %v = load i64, ptr %p
+ %x1 = bitcast i8** %x to i64*
+ %y1 = bitcast i64* %y to i8**
+ call void @scribble_on_i64(i64* %x1)
+ call void @scribble_on_i64(i64* %y)
+ %t = load i8*, i8** %x
+ store i8* %t, i8** %y1
+ %p = select i1 %flag, i64* %x1, i64* %y
+ %v = load i64, i64* %p
ret i64 %v
}
; Test that we can't speculate the load around the select. The load of the
; pointer doesn't load all of the stored integer bits. We could fix this, but it
; would require endianness checks and other nastiness.
-define ptr @test85(i1 %flag) {
+define i8* @test85(i1 %flag) {
; CHECK-LABEL: @test85(
-; CHECK-NEXT: [[X:%.*]] = alloca [2 x ptr], align 8
+; CHECK-NEXT: [[X1:%.*]] = alloca [2 x i8*], align 8
; CHECK-NEXT: [[Y:%.*]] = alloca i128, align 8
-; CHECK-NEXT: call void @scribble_on_i128(ptr nonnull [[X]])
-; CHECK-NEXT: call void @scribble_on_i128(ptr nonnull [[Y]])
-; CHECK-NEXT: [[T:%.*]] = load i128, ptr [[X]], align 8
-; CHECK-NEXT: store i128 [[T]], ptr [[Y]], align 8
-; CHECK-NEXT: [[X_VAL:%.*]] = load ptr, ptr [[X]], align 8
-; CHECK-NEXT: [[Y_VAL:%.*]] = load ptr, ptr [[Y]], align 8
-; CHECK-NEXT: [[V:%.*]] = select i1 [[FLAG:%.*]], ptr [[X_VAL]], ptr [[Y_VAL]]
-; CHECK-NEXT: ret ptr [[V]]
-;
- %x = alloca [2 x ptr]
+; CHECK-NEXT: [[X1_SUB:%.*]] = getelementptr inbounds [2 x i8*], [2 x i8*]* [[X1]], i64 0, i64 0
+; CHECK-NEXT: [[X2:%.*]] = bitcast [2 x i8*]* [[X1]] to i128*
+; CHECK-NEXT: [[Y1:%.*]] = bitcast i128* [[Y]] to i8**
+; CHECK-NEXT: call void @scribble_on_i128(i128* nonnull [[X2]])
+; CHECK-NEXT: call void @scribble_on_i128(i128* nonnull [[Y]])
+; CHECK-NEXT: [[T:%.*]] = load i128, i128* [[X2]], align 8
+; CHECK-NEXT: store i128 [[T]], i128* [[Y]], align 8
+; CHECK-NEXT: [[X1_SUB_VAL:%.*]] = load i8*, i8** [[X1_SUB]], align 8
+; CHECK-NEXT: [[Y1_VAL:%.*]] = load i8*, i8** [[Y1]], align 8
+; CHECK-NEXT: [[V:%.*]] = select i1 [[FLAG:%.*]], i8* [[X1_SUB_VAL]], i8* [[Y1_VAL]]
+; CHECK-NEXT: ret i8* [[V]]
+;
+ %x = alloca [2 x i8*]
%y = alloca i128
- call void @scribble_on_i128(ptr %x)
- call void @scribble_on_i128(ptr %y)
- %t = load i128, ptr %x
- store i128 %t, ptr %y
- %p = select i1 %flag, ptr %x, ptr %y
- %v = load ptr, ptr %p
- ret ptr %v
+ %x1 = bitcast [2 x i8*]* %x to i8**
+ %x2 = bitcast i8** %x1 to i128*
+ %y1 = bitcast i128* %y to i8**
+ call void @scribble_on_i128(i128* %x2)
+ call void @scribble_on_i128(i128* %y)
+ %t = load i128, i128* %x2
+ store i128 %t, i128* %y
+ %p = select i1 %flag, i8** %x1, i8** %y1
+ %v = load i8*, i8** %p
+ ret i8* %v
}
; Test that we can't speculate the load around the select when the integer size
; the bits of the integer.
define i128 @test86(i1 %flag) {
; CHECK-LABEL: @test86(
-; CHECK-NEXT: [[X:%.*]] = alloca [2 x ptr], align 8
+; CHECK-NEXT: [[X1:%.*]] = alloca [2 x i8*], align 8
; CHECK-NEXT: [[Y:%.*]] = alloca i128, align 8
-; CHECK-NEXT: call void @scribble_on_i128(ptr nonnull [[X]])
-; CHECK-NEXT: call void @scribble_on_i128(ptr nonnull [[Y]])
-; CHECK-NEXT: [[T:%.*]] = load ptr, ptr [[X]], align 8
-; CHECK-NEXT: store ptr [[T]], ptr [[Y]], align 8
-; CHECK-NEXT: [[X_VAL:%.*]] = load i128, ptr [[X]], align 8
-; CHECK-NEXT: [[Y_VAL:%.*]] = load i128, ptr [[Y]], align 8
-; CHECK-NEXT: [[V:%.*]] = select i1 [[FLAG:%.*]], i128 [[X_VAL]], i128 [[Y_VAL]]
+; CHECK-NEXT: [[X1_SUB:%.*]] = getelementptr inbounds [2 x i8*], [2 x i8*]* [[X1]], i64 0, i64 0
+; CHECK-NEXT: [[X2:%.*]] = bitcast [2 x i8*]* [[X1]] to i128*
+; CHECK-NEXT: [[Y1:%.*]] = bitcast i128* [[Y]] to i8**
+; CHECK-NEXT: call void @scribble_on_i128(i128* nonnull [[X2]])
+; CHECK-NEXT: call void @scribble_on_i128(i128* nonnull [[Y]])
+; CHECK-NEXT: [[T:%.*]] = load i8*, i8** [[X1_SUB]], align 8
+; CHECK-NEXT: store i8* [[T]], i8** [[Y1]], align 8
+; CHECK-NEXT: [[X2_VAL:%.*]] = load i128, i128* [[X2]], align 8
+; CHECK-NEXT: [[Y_VAL:%.*]] = load i128, i128* [[Y]], align 8
+; CHECK-NEXT: [[V:%.*]] = select i1 [[FLAG:%.*]], i128 [[X2_VAL]], i128 [[Y_VAL]]
; CHECK-NEXT: ret i128 [[V]]
;
- %x = alloca [2 x ptr]
+ %x = alloca [2 x i8*]
%y = alloca i128
- call void @scribble_on_i128(ptr %x)
- call void @scribble_on_i128(ptr %y)
- %t = load ptr, ptr %x
- store ptr %t, ptr %y
- %p = select i1 %flag, ptr %x, ptr %y
- %v = load i128, ptr %p
+ %x1 = bitcast [2 x i8*]* %x to i8**
+ %x2 = bitcast i8** %x1 to i128*
+ %y1 = bitcast i128* %y to i8**
+ call void @scribble_on_i128(i128* %x2)
+ call void @scribble_on_i128(i128* %y)
+ %t = load i8*, i8** %x1
+ store i8* %t, i8** %y1
+ %p = select i1 %flag, i128* %x2, i128* %y
+ %v = load i128, i128* %p
ret i128 %v
}
ret i32 %sel
}
-define i32 @PR23757_ne(i32 %x, ptr %p) {
+define i32 @PR23757_ne(i32 %x, i1* %p) {
; CHECK-LABEL: @PR23757_ne(
; CHECK-NEXT: [[CMP:%.*]] = icmp ne i32 [[X:%.*]], 2147483647
-; CHECK-NEXT: store i1 [[CMP]], ptr [[P:%.*]], align 1
+; CHECK-NEXT: store i1 [[CMP]], i1* [[P:%.*]], align 1
; CHECK-NEXT: ret i32 -2147483648
;
%cmp = icmp ne i32 %x, 2147483647
- store i1 %cmp, ptr %p ; thwart predicate canonicalization
+ store i1 %cmp, i1* %p ; thwart predicate canonicalization
%add = add nsw i32 %x, 1
%sel = select i1 %cmp, i32 -2147483648, i32 %add
ret i32 %sel
}
-define i32 @PR23757_ne_swapped(i32 %x, ptr %p) {
+define i32 @PR23757_ne_swapped(i32 %x, i1* %p) {
; CHECK-LABEL: @PR23757_ne_swapped(
; CHECK-NEXT: [[CMP:%.*]] = icmp ne i32 [[X:%.*]], 2147483647
-; CHECK-NEXT: store i1 [[CMP]], ptr [[P:%.*]], align 1
+; CHECK-NEXT: store i1 [[CMP]], i1* [[P:%.*]], align 1
; CHECK-NEXT: [[ADD:%.*]] = add i32 [[X]], 1
; CHECK-NEXT: ret i32 [[ADD]]
;
%cmp = icmp ne i32 %x, 2147483647
- store i1 %cmp, ptr %p ; thwart predicate canonicalization
+ store i1 %cmp, i1* %p ; thwart predicate canonicalization
%add = add nsw i32 %x, 1
%sel = select i1 %cmp, i32 %add, i32 -2147483648
ret i32 %sel
define i32 @PR27137(i32 %a) {
; CHECK-LABEL: @PR27137(
; CHECK-NEXT: [[TMP1:%.*]] = call i32 @llvm.smin.i32(i32 [[A:%.*]], i32 0)
-; CHECK-NEXT: [[S1:%.*]] = xor i32 [[TMP1]], -1
-; CHECK-NEXT: ret i32 [[S1]]
+; CHECK-NEXT: [[TMP2:%.*]] = xor i32 [[TMP1]], -1
+; CHECK-NEXT: ret i32 [[TMP2]]
;
%not_a = xor i32 %a, -1
%c0 = icmp slt i32 %a, 0
define i32 @select_icmp_slt0_xor(i32 %x) {
; CHECK-LABEL: @select_icmp_slt0_xor(
-; CHECK-NEXT: [[X_XOR:%.*]] = or i32 [[X:%.*]], -2147483648
-; CHECK-NEXT: ret i32 [[X_XOR]]
+; CHECK-NEXT: [[TMP1:%.*]] = or i32 [[X:%.*]], -2147483648
+; CHECK-NEXT: ret i32 [[TMP1]]
;
%cmp = icmp slt i32 %x, zeroinitializer
%xor = xor i32 %x, 2147483648
define <2 x i32> @select_icmp_slt0_xor_vec(<2 x i32> %x) {
; CHECK-LABEL: @select_icmp_slt0_xor_vec(
-; CHECK-NEXT: [[X_XOR:%.*]] = or <2 x i32> [[X:%.*]], <i32 -2147483648, i32 -2147483648>
-; CHECK-NEXT: ret <2 x i32> [[X_XOR]]
+; CHECK-NEXT: [[TMP1:%.*]] = or <2 x i32> [[X:%.*]], <i32 -2147483648, i32 -2147483648>
+; CHECK-NEXT: ret <2 x i32> [[TMP1]]
;
%cmp = icmp slt <2 x i32> %x, zeroinitializer
%xor = xor <2 x i32> %x, <i32 2147483648, i32 2147483648>
define <4 x i32> @cannot_canonicalize_to_shuffle1(<4 x i32> %a, <4 x i32> %b) {
; CHECK-LABEL: @cannot_canonicalize_to_shuffle1(
-; CHECK-NEXT: [[SEL:%.*]] = select <4 x i1> bitcast (i4 ptrtoint (ptr @g to i4) to <4 x i1>), <4 x i32> [[A:%.*]], <4 x i32> [[B:%.*]]
+; CHECK-NEXT: [[SEL:%.*]] = select <4 x i1> bitcast (i4 ptrtoint (i32* @g to i4) to <4 x i1>), <4 x i32> [[A:%.*]], <4 x i32> [[B:%.*]]
; CHECK-NEXT: ret <4 x i32> [[SEL]]
;
- %sel = select <4 x i1> bitcast (i4 ptrtoint (ptr @g to i4) to <4 x i1>), <4 x i32> %a, <4 x i32> %b
+ %sel = select <4 x i1> bitcast (i4 ptrtoint (i32* @g to i4) to <4 x i1>), <4 x i32> %a, <4 x i32> %b
ret <4 x i32> %sel
}
define <4 x i32> @cannot_canonicalize_to_shuffle2(<4 x i32> %a, <4 x i32> %b) {
; CHECK-LABEL: @cannot_canonicalize_to_shuffle2(
-; CHECK-NEXT: [[SEL:%.*]] = select <4 x i1> <i1 true, i1 undef, i1 false, i1 icmp sle (i16 ptrtoint (ptr @g to i16), i16 4)>, <4 x i32> [[A:%.*]], <4 x i32> [[B:%.*]]
+; CHECK-NEXT: [[SEL:%.*]] = select <4 x i1> <i1 true, i1 undef, i1 false, i1 icmp sle (i16 ptrtoint (i32* @g to i16), i16 4)>, <4 x i32> [[A:%.*]], <4 x i32> [[B:%.*]]
; CHECK-NEXT: ret <4 x i32> [[SEL]]
;
- %sel = select <4 x i1> <i1 true, i1 undef, i1 false, i1 icmp sle (i16 ptrtoint (ptr @g to i16), i16 4)>, <4 x i32> %a, <4 x i32> %b
+ %sel = select <4 x i1> <i1 true, i1 undef, i1 false, i1 icmp sle (i16 ptrtoint (i32* @g to i16), i16 4)>, <4 x i32> %a, <4 x i32> %b
ret <4 x i32> %sel
}
declare i32 @__gxx_personality_v0(...)
declare i1 @foo()
-define i32 @test_invoke_neg(i32 %x, i32 %y) nounwind uwtable ssp personality ptr @__gxx_personality_v0 {
+define i32 @test_invoke_neg(i32 %x, i32 %y) nounwind uwtable ssp personality i8* bitcast (i32 (...)* @__gxx_personality_v0 to i8*) {
; CHECK-LABEL: @test_invoke_neg(
; CHECK-NEXT: entry:
; CHECK-NEXT: [[COND:%.*]] = invoke i1 @foo()
declare i32 @bar()
-define i32 @test_invoke_2_neg(i1 %cond, i32 %x, i32 %y) nounwind uwtable ssp personality ptr @__gxx_personality_v0 {
+define i32 @test_invoke_2_neg(i1 %cond, i32 %x, i32 %y) nounwind uwtable ssp personality i8* bitcast (i32 (...)* @__gxx_personality_v0 to i8*) {
; CHECK-LABEL: @test_invoke_2_neg(
; CHECK-NEXT: entry:
; CHECK-NEXT: br i1 [[COND:%.*]], label [[IF_TRUE:%.*]], label [[IF_FALSE:%.*]]
ret i32 %sel
}
-define i32 @test_select_into_phi_not_idom_no_dom_input_1(i1 %cond, i32 %A, i32 %B, ptr %p) {
+define i32 @test_select_into_phi_not_idom_no_dom_input_1(i1 %cond, i32 %A, i32 %B, i32 *%p) {
; CHECK-LABEL: @test_select_into_phi_not_idom_no_dom_input_1(
; CHECK-NEXT: entry:
; CHECK-NEXT: br i1 [[COND:%.*]], label [[IF_TRUE:%.*]], label [[IF_FALSE:%.*]]
; CHECK: if.true:
-; CHECK-NEXT: [[C:%.*]] = load i32, ptr [[P:%.*]], align 4
+; CHECK-NEXT: [[C:%.*]] = load i32, i32* [[P:%.*]], align 4
; CHECK-NEXT: br label [[MERGE:%.*]]
; CHECK: if.false:
; CHECK-NEXT: br label [[MERGE]]
br i1 %cond, label %if.true, label %if.false
if.true:
- %C = load i32, ptr %p
+ %C = load i32, i32* %p
br label %merge
if.false:
ret i32 %sel
}
-define i32 @test_select_into_phi_not_idom_no_dom_input_2(i1 %cond, i32 %A, i32 %B, ptr %p) {
+define i32 @test_select_into_phi_not_idom_no_dom_input_2(i1 %cond, i32 %A, i32 %B, i32 *%p) {
; CHECK-LABEL: @test_select_into_phi_not_idom_no_dom_input_2(
; CHECK-NEXT: entry:
; CHECK-NEXT: br i1 [[COND:%.*]], label [[IF_TRUE:%.*]], label [[IF_FALSE:%.*]]
; CHECK: if.true:
; CHECK-NEXT: br label [[MERGE:%.*]]
; CHECK: if.false:
-; CHECK-NEXT: [[C:%.*]] = load i32, ptr [[P:%.*]], align 4
+; CHECK-NEXT: [[C:%.*]] = load i32, i32* [[P:%.*]], align 4
; CHECK-NEXT: br label [[MERGE]]
; CHECK: merge:
; CHECK-NEXT: [[SEL:%.*]] = phi i32 [ [[C]], [[IF_FALSE]] ], [ [[B:%.*]], [[IF_TRUE]] ]
br label %merge
if.false:
- %C = load i32, ptr %p
+ %C = load i32, i32* %p
br label %merge
merge:
}
; TODO: Dropping the inbounds flag should not be necessary for this fold.
-define ptr @select_replacement_gep_inbounds(ptr %base, i64 %offset) {
+define i8* @select_replacement_gep_inbounds(i8* %base, i64 %offset) {
; CHECK-LABEL: @select_replacement_gep_inbounds(
-; CHECK-NEXT: [[GEP:%.*]] = getelementptr i8, ptr [[BASE:%.*]], i64 [[OFFSET:%.*]]
-; CHECK-NEXT: ret ptr [[GEP]]
+; CHECK-NEXT: [[GEP:%.*]] = getelementptr i8, i8* [[BASE:%.*]], i64 [[OFFSET:%.*]]
+; CHECK-NEXT: ret i8* [[GEP]]
;
%cmp = icmp eq i64 %offset, 0
- %gep = getelementptr inbounds i8, ptr %base, i64 %offset
- %sel = select i1 %cmp, ptr %base, ptr %gep
- ret ptr %sel
+ %gep = getelementptr inbounds i8, i8* %base, i64 %offset
+ %sel = select i1 %cmp, i8* %base, i8* %gep
+ ret i8* %sel
}
define <2 x i1> @partial_true_undef_condval(<2 x i1> %x) {
define i64 @udiv_of_select_constexpr(i1 %c, i64 %x) {
; CHECK-LABEL: @udiv_of_select_constexpr(
-; CHECK-NEXT: [[SEL:%.*]] = select i1 [[C:%.*]], i64 [[X:%.*]], i64 ptrtoint (ptr @glbl to i64)
+; CHECK-NEXT: [[SEL:%.*]] = select i1 [[C:%.*]], i64 [[X:%.*]], i64 ptrtoint (i32* @glbl to i64)
; CHECK-NEXT: [[OP:%.*]] = udiv i64 [[SEL]], 3
; CHECK-NEXT: ret i64 [[OP]]
;
- %sel = select i1 %c, i64 %x, i64 ptrtoint (ptr @glbl to i64)
+ %sel = select i1 %c, i64 %x, i64 ptrtoint (i32* @glbl to i64)
%op = udiv i64 %sel, 3
ret i64 %op
}
define i64 @udiv_of_select_constexpr_commuted(i1 %c, i64 %x) {
; CHECK-LABEL: @udiv_of_select_constexpr_commuted(
-; CHECK-NEXT: [[SEL:%.*]] = select i1 [[C:%.*]], i64 ptrtoint (ptr @glbl to i64), i64 [[X:%.*]]
+; CHECK-NEXT: [[SEL:%.*]] = select i1 [[C:%.*]], i64 ptrtoint (i32* @glbl to i64), i64 [[X:%.*]]
; CHECK-NEXT: [[OP:%.*]] = udiv i64 [[SEL]], 3
; CHECK-NEXT: ret i64 [[OP]]
;
- %sel = select i1 %c, i64 ptrtoint (ptr @glbl to i64), i64 %x
+ %sel = select i1 %c, i64 ptrtoint (i32* @glbl to i64), i64 %x
%op = udiv i64 %sel, 3
ret i64 %op
}
; RUN: opt -passes=instcombine -S < %s | FileCheck %s
; Function Attrs: noinline uwtable
-define i32 @foo(ptr nocapture writeonly %arg) {
+define i32 @foo(i32* nocapture writeonly %arg) {
; CHECK-LABEL: @foo(
; CHECK-NEXT: bb:
; CHECK-NEXT: [[VAR:%.*]] = call i32 @baz()
-; CHECK-NEXT: store i32 [[VAR]], ptr [[ARG:%.*]], align 4
+; CHECK-NEXT: store i32 [[VAR]], i32* [[ARG:%.*]], align 4
; CHECK-NEXT: [[VAR1:%.*]] = call i32 @baz()
; CHECK-NEXT: ret i32 [[VAR1]]
;
bb:
%var = call i32 @baz()
- store i32 %var, ptr %arg, align 4
+ store i32 %var, i32* %arg, align 4
%var1 = call i32 @baz()
ret i32 %var1
}
; CHECK-NEXT: bb:
; CHECK-NEXT: [[VAR:%.*]] = alloca i32, align 4
; CHECK-NEXT: [[VAR1:%.*]] = alloca i32, align 4
-; CHECK-NEXT: call void @llvm.lifetime.start.p0(i64 4, ptr nonnull [[VAR]])
-; CHECK-NEXT: [[VAR3:%.*]] = call i32 @foo(ptr nonnull writeonly [[VAR]])
+; CHECK-NEXT: [[VAR2:%.*]] = bitcast i32* [[VAR]] to i8*
+; CHECK-NEXT: call void @llvm.lifetime.start.p0i8(i64 4, i8* nonnull [[VAR2]])
+; CHECK-NEXT: [[VAR3:%.*]] = call i32 @foo(i32* nonnull writeonly [[VAR]])
; CHECK-NEXT: [[VAR4:%.*]] = icmp eq i32 [[VAR3]], 0
; CHECK-NEXT: br i1 [[VAR4]], label [[BB5:%.*]], label [[BB14:%.*]]
; CHECK: bb5:
-; CHECK-NEXT: call void @llvm.lifetime.start.p0(i64 4, ptr nonnull [[VAR1]])
-; CHECK-NEXT: [[VAR8:%.*]] = load i32, ptr [[VAR]], align 4
+; CHECK-NEXT: [[VAR6:%.*]] = bitcast i32* [[VAR1]] to i8*
+; CHECK-NEXT: call void @llvm.lifetime.start.p0i8(i64 4, i8* nonnull [[VAR6]])
+; CHECK-NEXT: [[VAR8:%.*]] = load i32, i32* [[VAR]], align 4
; CHECK-NEXT: [[VAR9:%.*]] = icmp eq i32 [[VAR8]], 0
-; CHECK-NEXT: [[VAR7:%.*]] = call i32 @foo(ptr nonnull writeonly [[VAR1]])
+; CHECK-NEXT: [[VAR7:%.*]] = call i32 @foo(i32* nonnull writeonly [[VAR1]])
; CHECK-NEXT: br i1 [[VAR9]], label [[BB10:%.*]], label [[BB_CRIT_EDGE:%.*]]
; CHECK: bb10:
; CHECK-NEXT: [[VAR11:%.*]] = call i32 @bar()
; CHECK-NEXT: br label [[BB12]]
; CHECK: bb12:
; CHECK-NEXT: [[VAR13:%.*]] = phi i32 [ [[VAR11]], [[BB10]] ], [ [[VAR7]], [[BB_CRIT_EDGE]] ]
-; CHECK-NEXT: call void @llvm.lifetime.end.p0(i64 4, ptr nonnull [[VAR1]])
+; CHECK-NEXT: call void @llvm.lifetime.end.p0i8(i64 4, i8* nonnull [[VAR6]])
; CHECK-NEXT: br label [[BB14]]
; CHECK: bb14:
; CHECK-NEXT: [[VAR15:%.*]] = phi i32 [ [[VAR13]], [[BB12]] ], [ 0, [[BB:%.*]] ]
-; CHECK-NEXT: call void @llvm.lifetime.end.p0(i64 4, ptr nonnull [[VAR]])
+; CHECK-NEXT: call void @llvm.lifetime.end.p0i8(i64 4, i8* nonnull [[VAR2]])
; CHECK-NEXT: ret i32 [[VAR15]]
;
bb:
%var = alloca i32, align 4
%var1 = alloca i32, align 4
- call void @llvm.lifetime.start.p0(i64 4, ptr nonnull %var) #4
- %var3 = call i32 @foo(ptr nonnull writeonly %var)
+ %var2 = bitcast i32* %var to i8*
+ call void @llvm.lifetime.start.p0i8(i64 4, i8* nonnull %var2) #4
+ %var3 = call i32 @foo(i32* nonnull writeonly %var)
%var4 = icmp eq i32 %var3, 0
br i1 %var4, label %bb5, label %bb14
bb5: ; preds = %bb
- call void @llvm.lifetime.start.p0(i64 4, ptr nonnull %var1) #4
- %var8 = load i32, ptr %var, align 4
+ %var6 = bitcast i32* %var1 to i8*
+ call void @llvm.lifetime.start.p0i8(i64 4, i8* nonnull %var6) #4
+ %var8 = load i32, i32* %var, align 4
%var9 = icmp eq i32 %var8, 0
- %var7 = call i32 @foo(ptr nonnull writeonly %var1)
+ %var7 = call i32 @foo(i32* nonnull writeonly %var1)
br i1 %var9, label %bb10, label %bb_crit_edge
bb10: ; preds = %bb5
bb12: ; preds = %bb10, %bb5
%var13 = phi i32 [ %var11, %bb10 ], [ %var7, %bb_crit_edge ]
- call void @llvm.lifetime.end.p0(i64 4, ptr nonnull %var1) #4
+ call void @llvm.lifetime.end.p0i8(i64 4, i8* nonnull %var6) #4
br label %bb14
bb14: ; preds = %bb12, %bb
%var15 = phi i32 [ %var13, %bb12 ], [ 0, %bb ]
- call void @llvm.lifetime.end.p0(i64 4, ptr nonnull %var)
+ call void @llvm.lifetime.end.p0i8(i64 4, i8* nonnull %var2)
ret i32 %var15
}
-declare i32 @unknown(ptr %dest)
-declare i32 @unknown.as2(ptr addrspace(2) %dest)
+declare i32 @unknown(i32* %dest)
+declare i32 @unknown.as2(i32 addrspace(2)* %dest)
define i32 @sink_write_to_use(i1 %c) {
; CHECK-LABEL: @sink_write_to_use(
; CHECK: early_return:
; CHECK-NEXT: ret i32 0
; CHECK: use_block:
-; CHECK-NEXT: [[VAR3:%.*]] = call i32 @unknown(ptr nonnull writeonly [[VAR]]) #[[ATTR1:[0-9]+]]
+; CHECK-NEXT: [[VAR3:%.*]] = call i32 @unknown(i32* nonnull writeonly [[VAR]]) #[[ATTR1:[0-9]+]]
; CHECK-NEXT: ret i32 [[VAR3]]
;
entry:
%var = alloca i32, align 4
- %var3 = call i32 @unknown(ptr writeonly %var) argmemonly nounwind willreturn
+ %var3 = call i32 @unknown(i32* writeonly %var) argmemonly nounwind willreturn
br i1 %c, label %early_return, label %use_block
early_return:
; CHECK: early_return:
; CHECK-NEXT: ret i32 0
; CHECK: use_block:
-; CHECK-NEXT: [[VAR3:%.*]] = call i32 @unknown(ptr nonnull [[VAR]]) #[[ATTR1]]
+; CHECK-NEXT: [[VAR3:%.*]] = call i32 @unknown(i32* nonnull [[VAR]]) #[[ATTR1]]
; CHECK-NEXT: ret i32 [[VAR3]]
;
entry:
%var = alloca i32, align 4
- %var3 = call i32 @unknown(ptr %var) argmemonly nounwind willreturn
+ %var3 = call i32 @unknown(i32* %var) argmemonly nounwind willreturn
br i1 %c, label %early_return, label %use_block
early_return:
; CHECK: early_return:
; CHECK-NEXT: ret i32 0
; CHECK: use_block:
-; CHECK-NEXT: [[VAR3:%.*]] = call i32 @unknown(ptr nonnull [[VAR]]) #[[ATTR1]]
+; CHECK-NEXT: [[BITCAST:%.*]] = bitcast i8* [[VAR]] to i32*
+; CHECK-NEXT: [[VAR3:%.*]] = call i32 @unknown(i32* nonnull [[BITCAST]]) #[[ATTR1]]
; CHECK-NEXT: ret i32 [[VAR3]]
;
entry:
%var = alloca i8, align 8
- %var3 = call i32 @unknown(ptr %var) argmemonly nounwind willreturn
+ %bitcast = bitcast i8* %var to i32*
+ %var3 = call i32 @unknown(i32* %bitcast) argmemonly nounwind willreturn
br i1 %c, label %early_return, label %use_block
early_return:
define i32 @sink_gep1(i1 %c) {
; CHECK-LABEL: @sink_gep1(
; CHECK-NEXT: entry:
-; CHECK-NEXT: [[VAR:%.*]] = alloca i64, align 8
+; CHECK-NEXT: [[VAR1:%.*]] = alloca [2 x i32], align 8
; CHECK-NEXT: br i1 [[C:%.*]], label [[EARLY_RETURN:%.*]], label [[USE_BLOCK:%.*]]
; CHECK: early_return:
; CHECK-NEXT: ret i32 0
; CHECK: use_block:
-; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds i32, ptr [[VAR]], i64 1
-; CHECK-NEXT: [[VAR3:%.*]] = call i32 @unknown(ptr nonnull [[GEP]]) #[[ATTR1]]
+; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds [2 x i32], [2 x i32]* [[VAR1]], i64 0, i64 1
+; CHECK-NEXT: [[VAR3:%.*]] = call i32 @unknown(i32* nonnull [[GEP]]) #[[ATTR1]]
; CHECK-NEXT: ret i32 [[VAR3]]
;
entry:
%var = alloca i64, align 8
- %gep = getelementptr i32, ptr %var, i32 1
- %var3 = call i32 @unknown(ptr %gep) argmemonly nounwind willreturn
+ %bitcast = bitcast i64* %var to i32*
+ %gep = getelementptr i32, i32* %bitcast, i32 1
+ %var3 = call i32 @unknown(i32* %gep) argmemonly nounwind willreturn
br i1 %c, label %early_return, label %use_block
early_return:
define i32 @sink_gep2(i1 %c) {
; CHECK-LABEL: @sink_gep2(
; CHECK-NEXT: entry:
-; CHECK-NEXT: [[VAR:%.*]] = alloca i64, align 8
+; CHECK-NEXT: [[VAR1:%.*]] = alloca [2 x i32], align 8
; CHECK-NEXT: br i1 [[C:%.*]], label [[EARLY_RETURN:%.*]], label [[USE_BLOCK:%.*]]
; CHECK: early_return:
; CHECK-NEXT: ret i32 0
; CHECK: use_block:
-; CHECK-NEXT: [[VAR3:%.*]] = call i32 @unknown(ptr nonnull [[VAR]]) #[[ATTR1]]
+; CHECK-NEXT: [[VAR1_SUB:%.*]] = getelementptr inbounds [2 x i32], [2 x i32]* [[VAR1]], i64 0, i64 0
+; CHECK-NEXT: [[VAR3:%.*]] = call i32 @unknown(i32* nonnull [[VAR1_SUB]]) #[[ATTR1]]
; CHECK-NEXT: ret i32 [[VAR3]]
;
entry:
%var = alloca i64, align 8
- %var3 = call i32 @unknown(ptr %var) argmemonly nounwind willreturn
+ %bitcast = bitcast i64* %var to i32*
+ %var3 = call i32 @unknown(i32* %bitcast) argmemonly nounwind willreturn
br i1 %c, label %early_return, label %use_block
early_return:
; CHECK: early_return:
; CHECK-NEXT: ret i32 0
; CHECK: use_block:
-; CHECK-NEXT: [[CAST:%.*]] = addrspacecast ptr [[VAR]] to ptr addrspace(2)
-; CHECK-NEXT: [[VAR3:%.*]] = call i32 @unknown.as2(ptr addrspace(2) [[CAST]]) #[[ATTR1]]
+; CHECK-NEXT: [[CAST:%.*]] = addrspacecast i32* [[VAR]] to i32 addrspace(2)*
+; CHECK-NEXT: [[VAR3:%.*]] = call i32 @unknown.as2(i32 addrspace(2)* [[CAST]]) #[[ATTR1]]
; CHECK-NEXT: ret i32 [[VAR3]]
;
entry:
%var = alloca i32, align 8
- %cast = addrspacecast ptr %var to ptr addrspace(2)
- %var3 = call i32 @unknown.as2(ptr addrspace(2) %cast) argmemonly nounwind willreturn
+ %cast = addrspacecast i32* %var to i32 addrspace(2)*
+ %var3 = call i32 @unknown.as2(i32 addrspace(2)* %cast) argmemonly nounwind willreturn
br i1 %c, label %early_return, label %use_block
early_return:
; CHECK-LABEL: @neg_infinite_loop(
; CHECK-NEXT: entry:
; CHECK-NEXT: [[VAR:%.*]] = alloca i32, align 4
-; CHECK-NEXT: [[VAR3:%.*]] = call i32 @unknown(ptr nonnull [[VAR]]) #[[ATTR2:[0-9]+]]
+; CHECK-NEXT: [[VAR3:%.*]] = call i32 @unknown(i32* nonnull [[VAR]]) #[[ATTR2:[0-9]+]]
; CHECK-NEXT: br i1 [[C:%.*]], label [[EARLY_RETURN:%.*]], label [[USE_BLOCK:%.*]]
; CHECK: early_return:
; CHECK-NEXT: ret i32 0
;
entry:
%var = alloca i32, align 4
- %var3 = call i32 @unknown(ptr %var) argmemonly nounwind
+ %var3 = call i32 @unknown(i32* %var) argmemonly nounwind
br i1 %c, label %early_return, label %use_block
early_return:
; CHECK-LABEL: @neg_throw(
; CHECK-NEXT: entry:
; CHECK-NEXT: [[VAR:%.*]] = alloca i32, align 4
-; CHECK-NEXT: [[VAR3:%.*]] = call i32 @unknown(ptr nonnull [[VAR]]) #[[ATTR3:[0-9]+]]
+; CHECK-NEXT: [[VAR3:%.*]] = call i32 @unknown(i32* nonnull [[VAR]]) #[[ATTR3:[0-9]+]]
; CHECK-NEXT: br i1 [[C:%.*]], label [[EARLY_RETURN:%.*]], label [[USE_BLOCK:%.*]]
; CHECK: early_return:
; CHECK-NEXT: ret i32 0
;
entry:
%var = alloca i32, align 4
- %var3 = call i32 @unknown(ptr %var) argmemonly willreturn
+ %var3 = call i32 @unknown(i32* %var) argmemonly willreturn
br i1 %c, label %early_return, label %use_block
early_return:
; CHECK-LABEL: @neg_unknown_write(
; CHECK-NEXT: entry:
; CHECK-NEXT: [[VAR:%.*]] = alloca i32, align 4
-; CHECK-NEXT: [[VAR3:%.*]] = call i32 @unknown(ptr nonnull [[VAR]]) #[[ATTR4:[0-9]+]]
+; CHECK-NEXT: [[VAR3:%.*]] = call i32 @unknown(i32* nonnull [[VAR]]) #[[ATTR4:[0-9]+]]
; CHECK-NEXT: br i1 [[C:%.*]], label [[EARLY_RETURN:%.*]], label [[USE_BLOCK:%.*]]
; CHECK: early_return:
; CHECK-NEXT: ret i32 0
;
entry:
%var = alloca i32, align 4
- %var3 = call i32 @unknown(ptr %var) nounwind willreturn
+ %var3 = call i32 @unknown(i32* %var) nounwind willreturn
br i1 %c, label %early_return, label %use_block
early_return:
; CHECK-LABEL: @sink_lifetime1(
; CHECK-NEXT: entry:
; CHECK-NEXT: [[VAR:%.*]] = alloca i32, align 4
-; CHECK-NEXT: call void @llvm.lifetime.start.p0(i64 4, ptr nonnull [[VAR]])
-; CHECK-NEXT: [[VAR3:%.*]] = call i32 @unknown(ptr nonnull [[VAR]]) #[[ATTR1]]
+; CHECK-NEXT: [[BITCAST:%.*]] = bitcast i32* [[VAR]] to i8*
+; CHECK-NEXT: call void @llvm.lifetime.start.p0i8(i64 4, i8* nonnull [[BITCAST]])
+; CHECK-NEXT: [[VAR3:%.*]] = call i32 @unknown(i32* nonnull [[VAR]]) #[[ATTR1]]
; CHECK-NEXT: br i1 [[C:%.*]], label [[EARLY_RETURN:%.*]], label [[USE_BLOCK:%.*]]
; CHECK: early_return:
; CHECK-NEXT: ret i32 0
; CHECK: use_block:
-; CHECK-NEXT: call void @llvm.lifetime.end.p0(i64 4, ptr nonnull [[VAR]])
+; CHECK-NEXT: call void @llvm.lifetime.end.p0i8(i64 4, i8* nonnull [[BITCAST]])
; CHECK-NEXT: ret i32 [[VAR3]]
;
entry:
%var = alloca i32, align 4
- call void @llvm.lifetime.start.p0(i64 4, ptr %var)
- %var3 = call i32 @unknown(ptr %var) argmemonly nounwind willreturn
+ %bitcast = bitcast i32* %var to i8*
+ call void @llvm.lifetime.start.p0i8(i64 4, i8* %bitcast)
+ %var3 = call i32 @unknown(i32* %var) argmemonly nounwind willreturn
br i1 %c, label %early_return, label %use_block
early_return:
ret i32 0
use_block:
- call void @llvm.lifetime.end.p0(i64 4, ptr %var)
+ call void @llvm.lifetime.end.p0i8(i64 4, i8* %bitcast)
ret i32 %var3
}
; CHECK-LABEL: @sink_lifetime2(
; CHECK-NEXT: entry:
; CHECK-NEXT: [[VAR:%.*]] = alloca i32, align 4
-; CHECK-NEXT: call void @llvm.lifetime.start.p0(i64 4, ptr nonnull [[VAR]])
-; CHECK-NEXT: [[VAR3:%.*]] = call i32 @unknown(ptr nonnull [[VAR]]) #[[ATTR1]]
+; CHECK-NEXT: [[BITCAST:%.*]] = bitcast i32* [[VAR]] to i8*
+; CHECK-NEXT: call void @llvm.lifetime.start.p0i8(i64 4, i8* nonnull [[BITCAST]])
+; CHECK-NEXT: [[VAR3:%.*]] = call i32 @unknown(i32* nonnull [[VAR]]) #[[ATTR1]]
; CHECK-NEXT: br i1 [[C:%.*]], label [[MERGE:%.*]], label [[USE_BLOCK:%.*]]
; CHECK: merge:
; CHECK-NEXT: [[RET:%.*]] = phi i32 [ 0, [[ENTRY:%.*]] ], [ [[VAR3]], [[USE_BLOCK]] ]
-; CHECK-NEXT: call void @llvm.lifetime.end.p0(i64 4, ptr nonnull [[VAR]])
+; CHECK-NEXT: call void @llvm.lifetime.end.p0i8(i64 4, i8* nonnull [[BITCAST]])
; CHECK-NEXT: ret i32 [[RET]]
; CHECK: use_block:
; CHECK-NEXT: br label [[MERGE]]
;
entry:
%var = alloca i32, align 4
- call void @llvm.lifetime.start.p0(i64 4, ptr %var)
- %var3 = call i32 @unknown(ptr %var) argmemonly nounwind willreturn
+ %bitcast = bitcast i32* %var to i8*
+ call void @llvm.lifetime.start.p0i8(i64 4, i8* %bitcast)
+ %var3 = call i32 @unknown(i32* %var) argmemonly nounwind willreturn
br i1 %c, label %merge, label %use_block
merge:
%ret = phi i32 [0, %entry], [%var3, %use_block]
- call void @llvm.lifetime.end.p0(i64 4, ptr %var)
+ call void @llvm.lifetime.end.p0i8(i64 4, i8* %bitcast)
ret i32 %ret
use_block:
; CHECK: early_return:
; CHECK-NEXT: ret i32 0
; CHECK: use_block:
-; CHECK-NEXT: [[VAR3:%.*]] = call i32 @unknown(ptr nonnull [[VAR]]) #[[ATTR1]]
+; CHECK-NEXT: [[VAR3:%.*]] = call i32 @unknown(i32* nonnull [[VAR]]) #[[ATTR1]]
; CHECK-NEXT: ret i32 [[VAR3]]
;
entry:
%var = alloca i32, align 4
- call void @llvm.lifetime.start.p0(i64 4, ptr %var)
- call void @llvm.lifetime.end.p0(i64 4, ptr %var)
+ %bitcast = bitcast i32* %var to i8*
+ call void @llvm.lifetime.start.p0i8(i64 4, i8* %bitcast)
+ call void @llvm.lifetime.end.p0i8(i64 4, i8* %bitcast)
; If unknown accesses %var, that's UB
- %var3 = call i32 @unknown(ptr %var) argmemonly nounwind willreturn
+ %var3 = call i32 @unknown(i32* %var) argmemonly nounwind willreturn
br i1 %c, label %early_return, label %use_block
early_return:
; CHECK-LABEL: @sink_lifetime4a(
; CHECK-NEXT: entry:
; CHECK-NEXT: [[VAR:%.*]] = alloca i32, align 4
-; CHECK-NEXT: call void @llvm.lifetime.start.p0(i64 4, ptr nonnull [[VAR]])
-; CHECK-NEXT: [[VAR3:%.*]] = call i32 @unknown(ptr nonnull [[VAR]]) #[[ATTR1]]
-; CHECK-NEXT: call void @llvm.lifetime.end.p0(i64 4, ptr nonnull [[VAR]])
+; CHECK-NEXT: [[BITCAST:%.*]] = bitcast i32* [[VAR]] to i8*
+; CHECK-NEXT: call void @llvm.lifetime.start.p0i8(i64 4, i8* nonnull [[BITCAST]])
+; CHECK-NEXT: [[VAR3:%.*]] = call i32 @unknown(i32* nonnull [[VAR]]) #[[ATTR1]]
+; CHECK-NEXT: call void @llvm.lifetime.end.p0i8(i64 4, i8* nonnull [[BITCAST]])
; CHECK-NEXT: br i1 [[C:%.*]], label [[EARLY_RETURN:%.*]], label [[USE_BLOCK:%.*]]
; CHECK: early_return:
; CHECK-NEXT: ret i32 0
;
entry:
%var = alloca i32, align 4
- call void @llvm.lifetime.start.p0(i64 4, ptr %var)
- %var3 = call i32 @unknown(ptr %var) argmemonly nounwind willreturn
- call void @llvm.lifetime.end.p0(i64 4, ptr %var)
+ %bitcast = bitcast i32* %var to i8*
+ call void @llvm.lifetime.start.p0i8(i64 4, i8* %bitcast)
+ %var3 = call i32 @unknown(i32* %var) argmemonly nounwind willreturn
+ call void @llvm.lifetime.end.p0i8(i64 4, i8* %bitcast)
br i1 %c, label %early_return, label %use_block
early_return:
; CHECK-LABEL: @sink_lifetime4b(
; CHECK-NEXT: entry:
; CHECK-NEXT: [[VAR:%.*]] = alloca i32, align 4
-; CHECK-NEXT: call void @llvm.lifetime.start.p0(i64 4, ptr nonnull [[VAR]])
-; CHECK-NEXT: [[VAR3:%.*]] = call i32 @unknown(ptr nonnull writeonly [[VAR]]) #[[ATTR1]]
-; CHECK-NEXT: call void @llvm.lifetime.end.p0(i64 4, ptr nonnull [[VAR]])
+; CHECK-NEXT: [[BITCAST:%.*]] = bitcast i32* [[VAR]] to i8*
+; CHECK-NEXT: call void @llvm.lifetime.start.p0i8(i64 4, i8* nonnull [[BITCAST]])
+; CHECK-NEXT: [[VAR3:%.*]] = call i32 @unknown(i32* nonnull writeonly [[VAR]]) #[[ATTR1]]
+; CHECK-NEXT: call void @llvm.lifetime.end.p0i8(i64 4, i8* nonnull [[BITCAST]])
; CHECK-NEXT: br i1 [[C:%.*]], label [[EARLY_RETURN:%.*]], label [[USE_BLOCK:%.*]]
; CHECK: early_return:
; CHECK-NEXT: ret i32 0
;
entry:
%var = alloca i32, align 4
- call void @llvm.lifetime.start.p0(i64 4, ptr %var)
- %var3 = call i32 @unknown(ptr writeonly %var) argmemonly nounwind willreturn
- call void @llvm.lifetime.end.p0(i64 4, ptr %var)
+ %bitcast = bitcast i32* %var to i8*
+ call void @llvm.lifetime.start.p0i8(i64 4, i8* %bitcast)
+ %var3 = call i32 @unknown(i32* writeonly %var) argmemonly nounwind willreturn
+ call void @llvm.lifetime.end.p0i8(i64 4, i8* %bitcast)
br i1 %c, label %early_return, label %use_block
early_return:
; CHECK-LABEL: @sink_atomicrmw_to_use(
; CHECK-NEXT: entry:
; CHECK-NEXT: [[VAR:%.*]] = alloca i32, align 4
-; CHECK-NEXT: store i32 0, ptr [[VAR]], align 4
-; CHECK-NEXT: [[VAR3:%.*]] = atomicrmw add ptr [[VAR]], i32 1 seq_cst, align 4
+; CHECK-NEXT: store i32 0, i32* [[VAR]], align 4
+; CHECK-NEXT: [[VAR3:%.*]] = atomicrmw add i32* [[VAR]], i32 1 seq_cst, align 4
; CHECK-NEXT: br i1 [[C:%.*]], label [[EARLY_RETURN:%.*]], label [[USE_BLOCK:%.*]]
; CHECK: early_return:
; CHECK-NEXT: ret i32 0
;
entry:
%var = alloca i32, align 4
- store i32 0, ptr %var
- %var3 = atomicrmw add ptr %var, i32 1 seq_cst, align 4
+ store i32 0, i32* %var
+ %var3 = atomicrmw add i32* %var, i32 1 seq_cst, align 4
br i1 %c, label %early_return, label %use_block
early_return:
declare i32 @bar()
-declare void @llvm.lifetime.start.p0(i64 immarg, ptr nocapture)
-declare void @llvm.lifetime.end.p0(i64 immarg, ptr nocapture)
+declare void @llvm.lifetime.start.p0i8(i64 immarg, i8* nocapture)
+declare void @llvm.lifetime.end.p0i8(i64 immarg, i8* nocapture)
@b = common global [32 x i8] zeroinitializer, align 1
@percent_s = constant [3 x i8] c"%s\00"
-declare ptr @stpcpy(ptr, ptr)
+declare i8* @stpcpy(i8*, i8*)
-define ptr @test_simplify1() {
+define i8* @test_simplify1() {
; CHECK-LABEL: @test_simplify1(
-; CHECK-NEXT: call void @llvm.memcpy.p0.p0.i32(ptr noundef nonnull align 1 dereferenceable(6) @a, ptr noundef nonnull align 1 dereferenceable(6) @hello, i32 6, i1 false)
-; CHECK-NEXT: ret ptr getelementptr inbounds ([32 x i8], ptr @a, i32 0, i32 5)
+; CHECK-NEXT: call void @llvm.memcpy.p0i8.p0i8.i32(i8* noundef nonnull align 1 dereferenceable(6) getelementptr inbounds ([32 x i8], [32 x i8]* @a, i32 0, i32 0), i8* noundef nonnull align 1 dereferenceable(6) getelementptr inbounds ([6 x i8], [6 x i8]* @hello, i32 0, i32 0), i32 6, i1 false)
+; CHECK-NEXT: ret i8* getelementptr inbounds ([32 x i8], [32 x i8]* @a, i32 0, i32 5)
;
- %ret = call ptr @stpcpy(ptr @a, ptr @hello)
- ret ptr %ret
+ %dst = getelementptr [32 x i8], [32 x i8]* @a, i32 0, i32 0
+ %src = getelementptr [6 x i8], [6 x i8]* @hello, i32 0, i32 0
+ %ret = call i8* @stpcpy(i8* %dst, i8* %src)
+ ret i8* %ret
}
-define ptr @test_simplify2() {
+define i8* @test_simplify2() {
; CHECK-LABEL: @test_simplify2(
-; CHECK-NEXT: [[STRLEN:%.*]] = call i32 @strlen(ptr noundef nonnull dereferenceable(1) @a)
-; CHECK-NEXT: [[RET:%.*]] = getelementptr inbounds i8, ptr @a, i32 [[STRLEN]]
-; CHECK-NEXT: ret ptr [[RET]]
+; CHECK-NEXT: [[STRLEN:%.*]] = call i32 @strlen(i8* noundef nonnull dereferenceable(1) getelementptr inbounds ([32 x i8], [32 x i8]* @a, i32 0, i32 0))
+; CHECK-NEXT: [[TMP1:%.*]] = getelementptr inbounds [32 x i8], [32 x i8]* @a, i32 0, i32 [[STRLEN]]
+; CHECK-NEXT: ret i8* [[TMP1]]
;
- %ret = call ptr @stpcpy(ptr @a, ptr @a)
- ret ptr %ret
+ %dst = getelementptr [32 x i8], [32 x i8]* @a, i32 0, i32 0
+ %ret = call i8* @stpcpy(i8* %dst, i8* %dst)
+ ret i8* %ret
}
-define void @test_simplify3(ptr %dst) {
+define void @test_simplify3(i8* %dst) {
; CHECK-LABEL: @test_simplify3(
-; CHECK-NEXT: call void @llvm.memcpy.p0.p0.i32(ptr noundef nonnull align 1 dereferenceable(6) [[DST:%.*]], ptr noundef nonnull align 1 dereferenceable(6) @hello, i32 6, i1 false)
+; CHECK-NEXT: call void @llvm.memcpy.p0i8.p0i8.i32(i8* noundef nonnull align 1 dereferenceable(6) [[DST:%.*]], i8* noundef nonnull align 1 dereferenceable(6) getelementptr inbounds ([6 x i8], [6 x i8]* @hello, i32 0, i32 0), i32 6, i1 false)
; CHECK-NEXT: ret void
;
- call ptr @stpcpy(ptr dereferenceable(80) %dst, ptr @hello)
+ %src = getelementptr [6 x i8], [6 x i8]* @hello, i32 0, i32 0
+ call i8* @stpcpy(i8* dereferenceable(80) %dst, i8* %src)
ret void
}
-define ptr @test_no_simplify1() {
+define i8* @test_no_simplify1() {
; CHECK-LABEL: @test_no_simplify1(
-; CHECK-NEXT: [[RET:%.*]] = call ptr @stpcpy(ptr nonnull @a, ptr nonnull @b)
-; CHECK-NEXT: ret ptr [[RET]]
+; CHECK-NEXT: [[RET:%.*]] = call i8* @stpcpy(i8* nonnull getelementptr inbounds ([32 x i8], [32 x i8]* @a, i32 0, i32 0), i8* nonnull getelementptr inbounds ([32 x i8], [32 x i8]* @b, i32 0, i32 0))
+; CHECK-NEXT: ret i8* [[RET]]
;
- %ret = call ptr @stpcpy(ptr @a, ptr @b)
- ret ptr %ret
+ %dst = getelementptr [32 x i8], [32 x i8]* @a, i32 0, i32 0
+ %src = getelementptr [32 x i8], [32 x i8]* @b, i32 0, i32 0
+ %ret = call i8* @stpcpy(i8* %dst, i8* %src)
+ ret i8* %ret
}
-define ptr @test_no_simplify2(ptr %dst, ptr %src) {
+define i8* @test_no_simplify2(i8* %dst, i8* %src) {
; CHECK-LABEL: @test_no_simplify2(
-; CHECK-NEXT: [[RET:%.*]] = musttail call ptr @stpcpy(ptr [[DST:%.*]], ptr [[SRC:%.*]])
-; CHECK-NEXT: ret ptr [[RET]]
+; CHECK-NEXT: [[RET:%.*]] = musttail call i8* @stpcpy(i8* [[DST:%.*]], i8* [[SRC:%.*]])
+; CHECK-NEXT: ret i8* [[RET]]
;
- %ret = musttail call ptr @stpcpy(ptr %dst, ptr %src)
- ret ptr %ret
+ %ret = musttail call i8* @stpcpy(i8* %dst, i8* %src)
+ ret i8* %ret
}
-define ptr @test_no_incompatible_attr() {
+define i8* @test_no_incompatible_attr() {
; CHECK-LABEL: @test_no_incompatible_attr(
-; CHECK-NEXT: call void @llvm.memcpy.p0.p0.i32(ptr noundef nonnull align 1 dereferenceable(6) @a, ptr noundef nonnull align 1 dereferenceable(6) @hello, i32 6, i1 false)
-; CHECK-NEXT: ret ptr getelementptr inbounds ([32 x i8], ptr @a, i32 0, i32 5)
+; CHECK-NEXT: call void @llvm.memcpy.p0i8.p0i8.i32(i8* noundef nonnull align 1 dereferenceable(6) getelementptr inbounds ([32 x i8], [32 x i8]* @a, i32 0, i32 0), i8* noundef nonnull align 1 dereferenceable(6) getelementptr inbounds ([6 x i8], [6 x i8]* @hello, i32 0, i32 0), i32 6, i1 false)
+; CHECK-NEXT: ret i8* getelementptr inbounds ([32 x i8], [32 x i8]* @a, i32 0, i32 5)
;
- %ret = call dereferenceable(1) ptr @stpcpy(ptr @a, ptr @hello)
- ret ptr %ret
+ %dst = getelementptr [32 x i8], [32 x i8]* @a, i32 0, i32 0
+ %src = getelementptr [6 x i8], [6 x i8]* @hello, i32 0, i32 0
+ %ret = call dereferenceable(1) i8* @stpcpy(i8* %dst, i8* %src)
+ ret i8* %ret
}
; Check cases where slen >= strlen (src).
-define ptr @test_simplify1() {
+define i8* @test_simplify1() {
; CHECK-LABEL: @test_simplify1(
-; CHECK-NEXT: call void @llvm.memcpy.p0.p0.i32(ptr noundef nonnull align 1 dereferenceable(12) @a, ptr noundef nonnull align 1 dereferenceable(12) @.str, i32 12, i1 false)
-; CHECK-NEXT: ret ptr getelementptr inbounds ([60 x i8], ptr @a, i32 0, i32 11)
+; CHECK-NEXT: call void @llvm.memcpy.p0i8.p0i8.i32(i8* noundef nonnull align 1 dereferenceable(12) getelementptr inbounds ([60 x i8], [60 x i8]* @a, i32 0, i32 0), i8* noundef nonnull align 1 dereferenceable(12) getelementptr inbounds ([12 x i8], [12 x i8]* @.str, i32 0, i32 0), i32 12, i1 false)
+; CHECK-NEXT: ret i8* getelementptr inbounds ([60 x i8], [60 x i8]* @a, i32 0, i32 11)
;
+ %dst = getelementptr inbounds [60 x i8], [60 x i8]* @a, i32 0, i32 0
+ %src = getelementptr inbounds [12 x i8], [12 x i8]* @.str, i32 0, i32 0
- %ret = call ptr @__stpcpy_chk(ptr @a, ptr @.str, i32 60)
- ret ptr %ret
+ %ret = call i8* @__stpcpy_chk(i8* %dst, i8* %src, i32 60)
+ ret i8* %ret
}
-define ptr @test_simplify2() {
+define i8* @test_simplify2() {
; CHECK-LABEL: @test_simplify2(
-; CHECK-NEXT: call void @llvm.memcpy.p0.p0.i32(ptr noundef nonnull align 1 dereferenceable(12) @a, ptr noundef nonnull align 1 dereferenceable(12) @.str, i32 12, i1 false)
-; CHECK-NEXT: ret ptr getelementptr inbounds ([60 x i8], ptr @a, i32 0, i32 11)
+; CHECK-NEXT: call void @llvm.memcpy.p0i8.p0i8.i32(i8* noundef nonnull align 1 dereferenceable(12) getelementptr inbounds ([60 x i8], [60 x i8]* @a, i32 0, i32 0), i8* noundef nonnull align 1 dereferenceable(12) getelementptr inbounds ([12 x i8], [12 x i8]* @.str, i32 0, i32 0), i32 12, i1 false)
+; CHECK-NEXT: ret i8* getelementptr inbounds ([60 x i8], [60 x i8]* @a, i32 0, i32 11)
;
+ %dst = getelementptr inbounds [60 x i8], [60 x i8]* @a, i32 0, i32 0
+ %src = getelementptr inbounds [12 x i8], [12 x i8]* @.str, i32 0, i32 0
- %ret = call ptr @__stpcpy_chk(ptr @a, ptr @.str, i32 12)
- ret ptr %ret
+ %ret = call i8* @__stpcpy_chk(i8* %dst, i8* %src, i32 12)
+ ret i8* %ret
}
-define ptr @test_simplify3() {
+define i8* @test_simplify3() {
; CHECK-LABEL: @test_simplify3(
-; CHECK-NEXT: call void @llvm.memcpy.p0.p0.i32(ptr noundef nonnull align 1 dereferenceable(12) @a, ptr noundef nonnull align 1 dereferenceable(12) @.str, i32 12, i1 false)
-; CHECK-NEXT: ret ptr getelementptr inbounds ([60 x i8], ptr @a, i32 0, i32 11)
+; CHECK-NEXT: call void @llvm.memcpy.p0i8.p0i8.i32(i8* noundef nonnull align 1 dereferenceable(12) getelementptr inbounds ([60 x i8], [60 x i8]* @a, i32 0, i32 0), i8* noundef nonnull align 1 dereferenceable(12) getelementptr inbounds ([12 x i8], [12 x i8]* @.str, i32 0, i32 0), i32 12, i1 false)
+; CHECK-NEXT: ret i8* getelementptr inbounds ([60 x i8], [60 x i8]* @a, i32 0, i32 11)
;
+ %dst = getelementptr inbounds [60 x i8], [60 x i8]* @a, i32 0, i32 0
+ %src = getelementptr inbounds [12 x i8], [12 x i8]* @.str, i32 0, i32 0
- %ret = call ptr @__stpcpy_chk(ptr @a, ptr @.str, i32 -1)
- ret ptr %ret
+ %ret = call i8* @__stpcpy_chk(i8* %dst, i8* %src, i32 -1)
+ ret i8* %ret
}
-define ptr @test_simplify1_tail() {
+define i8* @test_simplify1_tail() {
; CHECK-LABEL: @test_simplify1_tail(
-; CHECK-NEXT: tail call void @llvm.memcpy.p0.p0.i32(ptr noundef nonnull align 1 dereferenceable(12) @a, ptr noundef nonnull align 1 dereferenceable(12) @.str, i32 12, i1 false)
-; CHECK-NEXT: ret ptr getelementptr inbounds ([60 x i8], ptr @a, i32 0, i32 11)
+; CHECK-NEXT: tail call void @llvm.memcpy.p0i8.p0i8.i32(i8* noundef nonnull align 1 dereferenceable(12) getelementptr inbounds ([60 x i8], [60 x i8]* @a, i32 0, i32 0), i8* noundef nonnull align 1 dereferenceable(12) getelementptr inbounds ([12 x i8], [12 x i8]* @.str, i32 0, i32 0), i32 12, i1 false)
+; CHECK-NEXT: ret i8* getelementptr inbounds ([60 x i8], [60 x i8]* @a, i32 0, i32 11)
;
+ %dst = getelementptr inbounds [60 x i8], [60 x i8]* @a, i32 0, i32 0
+ %src = getelementptr inbounds [12 x i8], [12 x i8]* @.str, i32 0, i32 0
- %ret = tail call ptr @__stpcpy_chk(ptr @a, ptr @.str, i32 60)
- ret ptr %ret
+ %ret = tail call i8* @__stpcpy_chk(i8* %dst, i8* %src, i32 60)
+ ret i8* %ret
}
; Check cases where there are no string constants.
-define ptr @test_simplify4() {
+define i8* @test_simplify4() {
; CHECK-LABEL: @test_simplify4(
-; CHECK-NEXT: [[STPCPY:%.*]] = call ptr @stpcpy(ptr nonnull @a, ptr nonnull @b)
-; CHECK-NEXT: ret ptr [[STPCPY]]
+; CHECK-NEXT: [[STPCPY:%.*]] = call i8* @stpcpy(i8* nonnull getelementptr inbounds ([60 x i8], [60 x i8]* @a, i32 0, i32 0), i8* nonnull getelementptr inbounds ([60 x i8], [60 x i8]* @b, i32 0, i32 0))
+; CHECK-NEXT: ret i8* [[STPCPY]]
;
+ %dst = getelementptr inbounds [60 x i8], [60 x i8]* @a, i32 0, i32 0
+ %src = getelementptr inbounds [60 x i8], [60 x i8]* @b, i32 0, i32 0
- %ret = call ptr @__stpcpy_chk(ptr @a, ptr @b, i32 -1)
- ret ptr %ret
+ %ret = call i8* @__stpcpy_chk(i8* %dst, i8* %src, i32 -1)
+ ret i8* %ret
}
-define ptr @test_simplify4_tail() {
+define i8* @test_simplify4_tail() {
; CHECK-LABEL: @test_simplify4_tail(
-; CHECK-NEXT: [[STPCPY:%.*]] = tail call ptr @stpcpy(ptr nonnull @a, ptr nonnull @b)
-; CHECK-NEXT: ret ptr [[STPCPY]]
+; CHECK-NEXT: [[STPCPY:%.*]] = tail call i8* @stpcpy(i8* nonnull getelementptr inbounds ([60 x i8], [60 x i8]* @a, i32 0, i32 0), i8* nonnull getelementptr inbounds ([60 x i8], [60 x i8]* @b, i32 0, i32 0))
+; CHECK-NEXT: ret i8* [[STPCPY]]
;
+ %dst = getelementptr inbounds [60 x i8], [60 x i8]* @a, i32 0, i32 0
+ %src = getelementptr inbounds [60 x i8], [60 x i8]* @b, i32 0, i32 0
- %ret = tail call ptr @__stpcpy_chk(ptr @a, ptr @b, i32 -1)
- ret ptr %ret
+ %ret = tail call i8* @__stpcpy_chk(i8* %dst, i8* %src, i32 -1)
+ ret i8* %ret
}
; Check case where the string length is not constant.
-define ptr @test_simplify5() {
+define i8* @test_simplify5() {
; CHECK-LABEL: @test_simplify5(
-; CHECK-NEXT: [[LEN:%.*]] = call i32 @llvm.objectsize.i32.p0(ptr @a, i1 false, i1 false, i1 false)
-; CHECK-NEXT: [[TMP1:%.*]] = call ptr @__memcpy_chk(ptr nonnull @a, ptr nonnull @.str, i32 12, i32 [[LEN]])
-; CHECK-NEXT: ret ptr getelementptr inbounds ([60 x i8], ptr @a, i32 0, i32 11)
+; CHECK-NEXT: [[LEN:%.*]] = call i32 @llvm.objectsize.i32.p0i8(i8* getelementptr inbounds ([60 x i8], [60 x i8]* @a, i32 0, i32 0), i1 false, i1 false, i1 false)
+; CHECK-NEXT: [[TMP1:%.*]] = call i8* @__memcpy_chk(i8* nonnull getelementptr inbounds ([60 x i8], [60 x i8]* @a, i32 0, i32 0), i8* nonnull getelementptr inbounds ([12 x i8], [12 x i8]* @.str, i32 0, i32 0), i32 12, i32 [[LEN]])
+; CHECK-NEXT: ret i8* getelementptr inbounds ([60 x i8], [60 x i8]* @a, i32 0, i32 11)
;
+ %dst = getelementptr inbounds [60 x i8], [60 x i8]* @a, i32 0, i32 0
+ %src = getelementptr inbounds [12 x i8], [12 x i8]* @.str, i32 0, i32 0
- %len = call i32 @llvm.objectsize.i32.p0(ptr @a, i1 false, i1 false, i1 false)
- %ret = call ptr @__stpcpy_chk(ptr @a, ptr @.str, i32 %len)
- ret ptr %ret
+ %len = call i32 @llvm.objectsize.i32.p0i8(i8* %dst, i1 false, i1 false, i1 false)
+ %ret = call i8* @__stpcpy_chk(i8* %dst, i8* %src, i32 %len)
+ ret i8* %ret
}
; Check case where the source and destination are the same.
-define ptr @test_simplify6() {
+define i8* @test_simplify6() {
; CHECK-LABEL: @test_simplify6(
-; CHECK-NEXT: [[STRLEN:%.*]] = call i32 @strlen(ptr noundef nonnull dereferenceable(1) @a)
-; CHECK-NEXT: [[RET:%.*]] = getelementptr inbounds i8, ptr @a, i32 [[STRLEN]]
-; CHECK-NEXT: ret ptr [[RET]]
+; CHECK-NEXT: [[STRLEN:%.*]] = call i32 @strlen(i8* noundef nonnull dereferenceable(1) getelementptr inbounds ([60 x i8], [60 x i8]* @a, i32 0, i32 0))
+; CHECK-NEXT: [[TMP1:%.*]] = getelementptr inbounds [60 x i8], [60 x i8]* @a, i32 0, i32 [[STRLEN]]
+; CHECK-NEXT: ret i8* [[TMP1]]
;
+ %dst = getelementptr inbounds [60 x i8], [60 x i8]* @a, i32 0, i32 0
- %len = call i32 @llvm.objectsize.i32.p0(ptr @a, i1 false, i1 false, i1 false)
- %ret = call ptr @__stpcpy_chk(ptr @a, ptr @a, i32 %len)
- ret ptr %ret
+ %len = call i32 @llvm.objectsize.i32.p0i8(i8* %dst, i1 false, i1 false, i1 false)
+ %ret = call i8* @__stpcpy_chk(i8* %dst, i8* %dst, i32 %len)
+ ret i8* %ret
}
; Check cases where there are no string constants, and is a tail call.
-define ptr @test_simplify7() {
+define i8* @test_simplify7() {
; CHECK-LABEL: @test_simplify7(
-; CHECK-NEXT: [[STPCPY:%.*]] = tail call ptr @stpcpy(ptr nonnull @a, ptr nonnull @b)
-; CHECK-NEXT: ret ptr [[STPCPY]]
+; CHECK-NEXT: [[STPCPY:%.*]] = tail call i8* @stpcpy(i8* nonnull getelementptr inbounds ([60 x i8], [60 x i8]* @a, i32 0, i32 0), i8* nonnull getelementptr inbounds ([60 x i8], [60 x i8]* @b, i32 0, i32 0))
+; CHECK-NEXT: ret i8* [[STPCPY]]
;
+ %dst = getelementptr inbounds [60 x i8], [60 x i8]* @a, i32 0, i32 0
+ %src = getelementptr inbounds [60 x i8], [60 x i8]* @b, i32 0, i32 0
- %ret = tail call ptr @__stpcpy_chk(ptr @a, ptr @b, i32 -1)
- ret ptr %ret
+ %ret = tail call i8* @__stpcpy_chk(i8* %dst, i8* %src, i32 -1)
+ ret i8* %ret
}
; Check case where slen < strlen (src).
-define ptr @test_no_simplify1() {
+define i8* @test_no_simplify1() {
; CHECK-LABEL: @test_no_simplify1(
-; CHECK-NEXT: [[RET:%.*]] = call ptr @__stpcpy_chk(ptr nonnull @a, ptr nonnull @b, i32 8)
-; CHECK-NEXT: ret ptr [[RET]]
+; CHECK-NEXT: [[RET:%.*]] = call i8* @__stpcpy_chk(i8* nonnull getelementptr inbounds ([60 x i8], [60 x i8]* @a, i32 0, i32 0), i8* nonnull getelementptr inbounds ([60 x i8], [60 x i8]* @b, i32 0, i32 0), i32 8)
+; CHECK-NEXT: ret i8* [[RET]]
;
+ %dst = getelementptr inbounds [60 x i8], [60 x i8]* @a, i32 0, i32 0
+ %src = getelementptr inbounds [60 x i8], [60 x i8]* @b, i32 0, i32 0
- %ret = call ptr @__stpcpy_chk(ptr @a, ptr @b, i32 8)
- ret ptr %ret
+ %ret = call i8* @__stpcpy_chk(i8* %dst, i8* %src, i32 8)
+ ret i8* %ret
}
-declare ptr @__stpcpy_chk(ptr, ptr, i32) nounwind
-declare i32 @llvm.objectsize.i32.p0(ptr, i1, i1, i1) nounwind readonly
+declare i8* @__stpcpy_chk(i8*, i8*, i32) nounwind
+declare i32 @llvm.objectsize.i32.p0i8(i8*, i1, i1, i1) nounwind readonly