From: Yaxun Liu Date: Thu, 16 Nov 2017 16:32:16 +0000 (+0000) Subject: Let llvm.invariant.group.barrier accepts pointer to any address space X-Git-Tag: llvmorg-6.0.0-rc1~3232 X-Git-Url: http://review.tizen.org/git/?a=commitdiff_plain;h=407ca36b2775322d352562a993bedb3a07d56ed6;p=platform%2Fupstream%2Fllvm.git Let llvm.invariant.group.barrier accepts pointer to any address space llvm.invariant.group.barrier may accept pointers to arbitrary address space. This patch let it accept pointers to i8 in any address space and returns pointer to i8 in the same address space. Differential Revision: https://reviews.llvm.org/D39973 llvm-svn: 318413 --- diff --git a/llvm/docs/LangRef.rst b/llvm/docs/LangRef.rst index b097230..a091cc1 100644 --- a/llvm/docs/LangRef.rst +++ b/llvm/docs/LangRef.rst @@ -12817,10 +12817,13 @@ This intrinsic indicates that the memory is mutable again. Syntax: """"""" +This is an overloaded intrinsic. The memory object can belong to any address +space. The returned pointer must belong to the same address space as the +argument. :: - declare i8* @llvm.invariant.group.barrier(i8* ) + declare i8* @llvm.invariant.group.barrier.p0i8(i8* ) Overview: """"""""" diff --git a/llvm/include/llvm/IR/IRBuilder.h b/llvm/include/llvm/IR/IRBuilder.h index fed2cbf..e687ca6 100644 --- a/llvm/include/llvm/IR/IRBuilder.h +++ b/llvm/include/llvm/IR/IRBuilder.h @@ -1810,26 +1810,28 @@ public: /// \brief Create an invariant.group.barrier intrinsic call, that stops /// optimizer to propagate equality using invariant.group metadata. - /// If Ptr type is different from i8*, it's casted to i8* before call - /// and casted back to Ptr type after call. + /// If Ptr type is different from pointer to i8, it's casted to pointer to i8 + /// in the same address space before call and casted back to Ptr type after + /// call. Value *CreateInvariantGroupBarrier(Value *Ptr) { + assert(isa(Ptr->getType()) && + "invariant.group.barrier only applies to pointers."); + auto *PtrType = Ptr->getType(); + auto *Int8PtrTy = getInt8PtrTy(PtrType->getPointerAddressSpace()); + if (PtrType != Int8PtrTy) + Ptr = CreateBitCast(Ptr, Int8PtrTy); Module *M = BB->getParent()->getParent(); - Function *FnInvariantGroupBarrier = Intrinsic::getDeclaration(M, - Intrinsic::invariant_group_barrier); + Function *FnInvariantGroupBarrier = Intrinsic::getDeclaration( + M, Intrinsic::invariant_group_barrier, {Int8PtrTy}); - Type *ArgumentAndReturnType = FnInvariantGroupBarrier->getReturnType(); - assert(ArgumentAndReturnType == - FnInvariantGroupBarrier->getFunctionType()->getParamType(0) && - "InvariantGroupBarrier should take and return the same type"); - Type *PtrType = Ptr->getType(); - - bool PtrTypeConversionNeeded = PtrType != ArgumentAndReturnType; - if (PtrTypeConversionNeeded) - Ptr = CreateBitCast(Ptr, ArgumentAndReturnType); + assert(FnInvariantGroupBarrier->getReturnType() == Int8PtrTy && + FnInvariantGroupBarrier->getFunctionType()->getParamType(0) == + Int8PtrTy && + "InvariantGroupBarrier should take and return the same type"); CallInst *Fn = CreateCall(FnInvariantGroupBarrier, {Ptr}); - if (PtrTypeConversionNeeded) + if (PtrType != Int8PtrTy) return CreateBitCast(Fn, PtrType); return Fn; } diff --git a/llvm/include/llvm/IR/Intrinsics.td b/llvm/include/llvm/IR/Intrinsics.td index 9560ae6..07de056 100644 --- a/llvm/include/llvm/IR/Intrinsics.td +++ b/llvm/include/llvm/IR/Intrinsics.td @@ -711,8 +711,8 @@ def int_invariant_end : Intrinsic<[], // which is valid. // The argument also can't be marked with 'returned' attribute, because // it would remove barrier. -def int_invariant_group_barrier : Intrinsic<[llvm_ptr_ty], - [llvm_ptr_ty], +def int_invariant_group_barrier : Intrinsic<[llvm_anyptr_ty], + [LLVMMatchType<0>], [IntrReadMem, IntrArgMemOnly]>; //===------------------------ Stackmap Intrinsics -------------------------===// diff --git a/llvm/test/Analysis/MemorySSA/invariant-groups.ll b/llvm/test/Analysis/MemorySSA/invariant-groups.ll index 26a0a32..eb8ad96 100644 --- a/llvm/test/Analysis/MemorySSA/invariant-groups.ll +++ b/llvm/test/Analysis/MemorySSA/invariant-groups.ll @@ -17,8 +17,8 @@ define i32 @foo(i32* %a) { %1 = bitcast i32* %a to i8* ; CHECK: MemoryUse(2) -; CHECK-NEXT: %a8 = call i8* @llvm.invariant.group.barrier(i8* %1) - %a8 = call i8* @llvm.invariant.group.barrier(i8* %1) +; CHECK-NEXT: %a8 = call i8* @llvm.invariant.group.barrier.p0i8(i8* %1) + %a8 = call i8* @llvm.invariant.group.barrier.p0i8(i8* %1) %a32 = bitcast i8* %a8 to i32* ; This have to be MemoryUse(2), because we can't skip the barrier based on @@ -36,8 +36,8 @@ define i32 @skipBarrier(i32* %a) { %1 = bitcast i32* %a to i8* ; CHECK: MemoryUse(1) -; CHECK-NEXT: %a8 = call i8* @llvm.invariant.group.barrier(i8* %1) - %a8 = call i8* @llvm.invariant.group.barrier(i8* %1) +; CHECK-NEXT: %a8 = call i8* @llvm.invariant.group.barrier.p0i8(i8* %1) + %a8 = call i8* @llvm.invariant.group.barrier.p0i8(i8* %1) %a32 = bitcast i8* %a8 to i32* ; We can skip the barrier only if the "skip" is not based on !invariant.group. @@ -55,8 +55,8 @@ define i32 @skipBarrier2(i32* %a) { %1 = bitcast i32* %a to i8* ; CHECK: MemoryUse(liveOnEntry) -; CHECK-NEXT: %a8 = call i8* @llvm.invariant.group.barrier(i8* %1) - %a8 = call i8* @llvm.invariant.group.barrier(i8* %1) +; CHECK-NEXT: %a8 = call i8* @llvm.invariant.group.barrier.p0i8(i8* %1) + %a8 = call i8* @llvm.invariant.group.barrier.p0i8(i8* %1) %a32 = bitcast i8* %a8 to i32* ; We can skip the barrier only if the "skip" is not based on !invariant.group. @@ -86,8 +86,8 @@ define i32 @handleInvariantGroups(i32* %a) { store i32 1, i32* @g, align 4 %1 = bitcast i32* %a to i8* ; CHECK: MemoryUse(2) -; CHECK-NEXT: %a8 = call i8* @llvm.invariant.group.barrier(i8* %1) - %a8 = call i8* @llvm.invariant.group.barrier(i8* %1) +; CHECK-NEXT: %a8 = call i8* @llvm.invariant.group.barrier.p0i8(i8* %1) + %a8 = call i8* @llvm.invariant.group.barrier.p0i8(i8* %1) %a32 = bitcast i8* %a8 to i32* ; CHECK: MemoryUse(2) @@ -145,8 +145,8 @@ entry: call void @clobber8(i8* %p) ; CHECK: MemoryUse(2) -; CHECK-NEXT: %after = call i8* @llvm.invariant.group.barrier(i8* %p) - %after = call i8* @llvm.invariant.group.barrier(i8* %p) +; CHECK-NEXT: %after = call i8* @llvm.invariant.group.barrier.p0i8(i8* %p) + %after = call i8* @llvm.invariant.group.barrier.p0i8(i8* %p) br i1 undef, label %Loop.Body, label %Loop.End Loop.Body: @@ -192,8 +192,8 @@ entry: call void @clobber8(i8* %p) ; CHECK: MemoryUse(2) -; CHECK-NEXT: %after = call i8* @llvm.invariant.group.barrier(i8* %p) - %after = call i8* @llvm.invariant.group.barrier(i8* %p) +; CHECK-NEXT: %after = call i8* @llvm.invariant.group.barrier.p0i8(i8* %p) + %after = call i8* @llvm.invariant.group.barrier.p0i8(i8* %p) br i1 undef, label %Loop.Body, label %Loop.End Loop.Body: @@ -253,8 +253,8 @@ entry: ; CHECK-NEXT: call void @clobber call void @clobber8(i8* %p) ; CHECK: MemoryUse(2) -; CHECK-NEXT: %after = call i8* @llvm.invariant.group.barrier(i8* %p) - %after = call i8* @llvm.invariant.group.barrier(i8* %p) +; CHECK-NEXT: %after = call i8* @llvm.invariant.group.barrier.p0i8(i8* %p) + %after = call i8* @llvm.invariant.group.barrier.p0i8(i8* %p) br i1 undef, label %Loop.Pre, label %Loop.End Loop.Pre: @@ -293,7 +293,7 @@ Ret: ret i8 %3 } -declare i8* @llvm.invariant.group.barrier(i8*) +declare i8* @llvm.invariant.group.barrier.p0i8(i8*) declare void @clobber(i32*) declare void @clobber8(i8*) diff --git a/llvm/test/CodeGen/AMDGPU/promote-alloca-invariant-markers.ll b/llvm/test/CodeGen/AMDGPU/promote-alloca-invariant-markers.ll index 776b151..768d520 100644 --- a/llvm/test/CodeGen/AMDGPU/promote-alloca-invariant-markers.ll +++ b/llvm/test/CodeGen/AMDGPU/promote-alloca-invariant-markers.ll @@ -1,22 +1,23 @@ -; RUN: llc -amdgpu-scalarize-global-loads=false -march=amdgcn -mattr=+promote-alloca -verify-machineinstrs < %s | FileCheck -check-prefix=GCN %s +; RUN: llc -amdgpu-scalarize-global-loads=false -march=amdgcn -mtriple=amdgcn---amdgiz -mattr=+promote-alloca -verify-machineinstrs < %s | FileCheck -check-prefix=GCN %s +target datalayout = "A5" -declare {}* @llvm.invariant.start.p0i8(i64, i8* nocapture) #0 -declare void @llvm.invariant.end.p0i8({}*, i64, i8* nocapture) #0 -declare i8* @llvm.invariant.group.barrier(i8*) #1 +declare {}* @llvm.invariant.start.p5i8(i64, i8 addrspace(5)* nocapture) #0 +declare void @llvm.invariant.end.p5i8({}*, i64, i8 addrspace(5)* nocapture) #0 +declare i8 addrspace(5)* @llvm.invariant.group.barrier.p5i8(i8 addrspace(5)*) #1 ; GCN-LABEL: {{^}}use_invariant_promotable_lds: ; GCN: buffer_load_dword ; GCN: ds_write_b32 define amdgpu_kernel void @use_invariant_promotable_lds(i32 addrspace(1)* %arg) #2 { bb: - %tmp = alloca i32, align 4 - %tmp1 = bitcast i32* %tmp to i8* + %tmp = alloca i32, align 4, addrspace(5) + %tmp1 = bitcast i32 addrspace(5)* %tmp to i8 addrspace(5)* %tmp2 = getelementptr inbounds i32, i32 addrspace(1)* %arg, i64 1 %tmp3 = load i32, i32 addrspace(1)* %tmp2 - store i32 %tmp3, i32* %tmp - %tmp4 = call {}* @llvm.invariant.start.p0i8(i64 4, i8* %tmp1) #0 - call void @llvm.invariant.end.p0i8({}* %tmp4, i64 4, i8* %tmp1) #0 - %tmp5 = call i8* @llvm.invariant.group.barrier(i8* %tmp1) #1 + store i32 %tmp3, i32 addrspace(5)* %tmp + %tmp4 = call {}* @llvm.invariant.start.p5i8(i64 4, i8 addrspace(5)* %tmp1) #0 + call void @llvm.invariant.end.p5i8({}* %tmp4, i64 4, i8 addrspace(5)* %tmp1) #0 + %tmp5 = call i8 addrspace(5)* @llvm.invariant.group.barrier.p5i8(i8 addrspace(5)* %tmp1) #1 ret void } diff --git a/llvm/test/Other/invariant.group.barrier.ll b/llvm/test/Other/invariant.group.barrier.ll index 5b9b54f..f10e4a1 100644 --- a/llvm/test/Other/invariant.group.barrier.ll +++ b/llvm/test/Other/invariant.group.barrier.ll @@ -12,10 +12,10 @@ define i8 @optimizable() { entry: %ptr = alloca i8 store i8 42, i8* %ptr, !invariant.group !0 -; CHECK: call i8* @llvm.invariant.group.barrier - %ptr2 = call i8* @llvm.invariant.group.barrier(i8* %ptr) -; CHECK-NOT: call i8* @llvm.invariant.group.barrier - %ptr3 = call i8* @llvm.invariant.group.barrier(i8* %ptr) +; CHECK: call i8* @llvm.invariant.group.barrier.p0i8 + %ptr2 = call i8* @llvm.invariant.group.barrier.p0i8(i8* %ptr) +; CHECK-NOT: call i8* @llvm.invariant.group.barrier.p0i8 + %ptr3 = call i8* @llvm.invariant.group.barrier.p0i8(i8* %ptr) ; CHECK: call void @clobber(i8* {{.*}}%ptr) call void @clobber(i8* %ptr) @@ -34,11 +34,11 @@ define i8 @unoptimizable() { entry: %ptr = alloca i8 store i8 42, i8* %ptr, !invariant.group !0 -; CHECK: call i8* @llvm.invariant.group.barrier - %ptr2 = call i8* @llvm.invariant.group.barrier(i8* %ptr) +; CHECK: call i8* @llvm.invariant.group.barrier.p0i8 + %ptr2 = call i8* @llvm.invariant.group.barrier.p0i8(i8* %ptr) call void @clobber(i8* %ptr) -; CHECK: call i8* @llvm.invariant.group.barrier - %ptr3 = call i8* @llvm.invariant.group.barrier(i8* %ptr) +; CHECK: call i8* @llvm.invariant.group.barrier.p0i8 + %ptr3 = call i8* @llvm.invariant.group.barrier.p0i8(i8* %ptr) ; CHECK: call void @clobber(i8* {{.*}}%ptr) call void @clobber(i8* %ptr) ; CHECK: call void @use(i8* {{.*}}%ptr2) @@ -55,8 +55,8 @@ declare void @use(i8* readonly) declare void @clobber(i8*) ; CHECK: Function Attrs: argmemonly nounwind readonly -; CHECK-NEXT: declare i8* @llvm.invariant.group.barrier(i8*) -declare i8* @llvm.invariant.group.barrier(i8*) +; CHECK-NEXT: declare i8* @llvm.invariant.group.barrier.p0i8(i8*) +declare i8* @llvm.invariant.group.barrier.p0i8(i8*) !0 = !{} diff --git a/llvm/test/Transforms/CodeGenPrepare/invariant.group.ll b/llvm/test/Transforms/CodeGenPrepare/invariant.group.ll index bde04a5..042e58b 100644 --- a/llvm/test/Transforms/CodeGenPrepare/invariant.group.ll +++ b/llvm/test/Transforms/CodeGenPrepare/invariant.group.ll @@ -6,10 +6,10 @@ define void @foo() { enter: ; CHECK-NOT: !invariant.group - ; CHECK-NOT: @llvm.invariant.group.barrier( + ; CHECK-NOT: @llvm.invariant.group.barrier.p0i8( ; CHECK: %val = load i8, i8* @tmp, !tbaa %val = load i8, i8* @tmp, !invariant.group !0, !tbaa !{!1, !1, i64 0} - %ptr = call i8* @llvm.invariant.group.barrier(i8* @tmp) + %ptr = call i8* @llvm.invariant.group.barrier.p0i8(i8* @tmp) ; CHECK: store i8 42, i8* @tmp store i8 42, i8* %ptr, !invariant.group !0 @@ -18,7 +18,7 @@ enter: } ; CHECK-LABEL: } -declare i8* @llvm.invariant.group.barrier(i8*) +declare i8* @llvm.invariant.group.barrier.p0i8(i8*) !0 = !{!"something"} !1 = !{!"x", !0} diff --git a/llvm/test/Transforms/GVN/invariant.group.ll b/llvm/test/Transforms/GVN/invariant.group.ll index 787a103..1bc1f49 100644 --- a/llvm/test/Transforms/GVN/invariant.group.ll +++ b/llvm/test/Transforms/GVN/invariant.group.ll @@ -25,7 +25,7 @@ define i8 @optimizable1() { entry: %ptr = alloca i8 store i8 42, i8* %ptr, !invariant.group !0 - %ptr2 = call i8* @llvm.invariant.group.barrier(i8* %ptr) + %ptr2 = call i8* @llvm.invariant.group.barrier.p0i8(i8* %ptr) %a = load i8, i8* %ptr, !invariant.group !0 call void @foo(i8* %ptr2); call to use %ptr2 @@ -242,7 +242,7 @@ define i8 @optimizable4() { entry: %ptr = alloca i8 store i8 42, i8* %ptr, !invariant.group !0 - %ptr2 = call i8* @llvm.invariant.group.barrier(i8* %ptr) + %ptr2 = call i8* @llvm.invariant.group.barrier.p0i8(i8* %ptr) ; CHECK-NOT: load %a = load i8, i8* %ptr2, !invariant.group !0 @@ -314,7 +314,7 @@ entry: ; CHECK: store i8 %unknownValue, i8* %ptr, !invariant.group !0 store i8 %unknownValue, i8* %ptr, !invariant.group !0 - %newPtr2 = call i8* @llvm.invariant.group.barrier(i8* %ptr) + %newPtr2 = call i8* @llvm.invariant.group.barrier.p0i8(i8* %ptr) ; CHECK-NOT: load %d = load i8, i8* %newPtr2, !invariant.group !0 ; CHECK: ret i8 %unknownValue @@ -441,7 +441,7 @@ declare void @_ZN1A3fooEv(%struct.A*) declare void @_ZN1AC1Ev(%struct.A*) declare void @fooBit(i1*, i1) -declare i8* @llvm.invariant.group.barrier(i8*) +declare i8* @llvm.invariant.group.barrier.p0i8(i8*) ; Function Attrs: nounwind declare void @llvm.assume(i1 %cmp.vtables) #0