Index: llvm/include/llvm/Frontend/OpenMP/OMPKinds.def =================================================================== --- llvm/include/llvm/Frontend/OpenMP/OMPKinds.def +++ llvm/include/llvm/Frontend/OpenMP/OMPKinds.def @@ -483,9 +483,9 @@ VoidPtrPtr, Int64Ptr, Int64Ptr, VoidPtrPtr) __OMP_RTL(__tgt_target_data_begin_nowait_mapper, false, Void, Int64, Int32, VoidPtrPtr, VoidPtrPtr, Int64Ptr, Int64Ptr, VoidPtrPtr) -__OMP_RTL(__tgt_target_data_begin_mapper_issue, false, AsyncInfo, Int64, Int32, - VoidPtrPtr, VoidPtrPtr, Int64Ptr, Int64Ptr, VoidPtrPtr) -__OMP_RTL(__tgt_target_data_begin_mapper_wait, false, Void, Int64, AsyncInfo) +__OMP_RTL(__tgt_target_data_begin_mapper_issue, false, Void, Int64, Int32, + VoidPtrPtr, VoidPtrPtr, Int64Ptr, Int64Ptr, VoidPtrPtr, AsyncInfoPtr) +__OMP_RTL(__tgt_target_data_begin_mapper_wait, false, Void, Int64, AsyncInfoPtr) __OMP_RTL(__tgt_target_data_end_mapper, false, Void, Int64, Int32, VoidPtrPtr, VoidPtrPtr, Int64Ptr, Int64Ptr, VoidPtrPtr) __OMP_RTL(__tgt_target_data_end_nowait_mapper, false, Void, Int64, Int32, Index: llvm/lib/Transforms/IPO/OpenMPOpt.cpp =================================================================== --- llvm/lib/Transforms/IPO/OpenMPOpt.cpp +++ llvm/lib/Transforms/IPO/OpenMPOpt.cpp @@ -812,7 +812,16 @@ /// Splits \p RuntimeCall into its "issue" and "wait" counterparts. bool splitTargetDataBeginRTC(CallInst &RuntimeCall, Instruction &WaitMovementPoint) { + // Create stack allocated handle (__tgt_async_info) at the beginning of the + // function. Used for storing information of the async transfer, allowing to + // wait on it later. auto &IRBuilder = OMPInfoCache.OMPBuilder; + auto *F = RuntimeCall.getCaller(); + Instruction *FirstInst = &(F->getEntryBlock().front()); + AllocaInst *Handle = new AllocaInst(IRBuilder.AsyncInfo, + F->getAddressSpace(), + "handle", FirstInst); + // Add "issue" runtime call declaration: // declare %struct.tgt_async_info @__tgt_target_data_begin_issue(i64, i32, // i8**, i8**, i64*, i64*) @@ -823,9 +832,10 @@ SmallVector Args; for (auto &Arg : RuntimeCall.args()) Args.push_back(Arg.get()); + Args.push_back(Handle); CallInst *IssueCallsite = - CallInst::Create(IssueDecl, Args, "handle", &RuntimeCall); + CallInst::Create(IssueDecl, Args, /*NameStr=*/"", &RuntimeCall); RuntimeCall.eraseFromParent(); // Add "wait" runtime call declaration: @@ -836,7 +846,7 @@ // Add call site to WaitDecl. Value *WaitParams[2] = { IssueCallsite->getArgOperand(0), // device_id. - IssueCallsite // returned handle. + Handle // handle to wait on. }; CallInst::Create(WaitDecl, WaitParams, /*NameStr=*/"", &WaitMovementPoint); Index: llvm/test/Transforms/OpenMP/hide_mem_transfer_latency.ll =================================================================== --- llvm/test/Transforms/OpenMP/hide_mem_transfer_latency.ll +++ llvm/test/Transforms/OpenMP/hide_mem_transfer_latency.ll @@ -38,8 +38,11 @@ ; return random + a; ;} define dso_local double @heavyComputation1() { -; CHECK-LABEL: define {{[^@]+}}@heavyComputation1() +; CHECK-LABEL: define {{[^@]+}}@heavyComputation1() { ; CHECK-NEXT: entry: + +; CHECK-NEXT: %handle = alloca %struct.__tgt_async_info, align 8 + ; CHECK-NEXT: %a = alloca double, align 8 ; CHECK-NEXT: %.offload_baseptrs = alloca [1 x i8*], align 8 ; CHECK-NEXT: %.offload_ptrs = alloca [1 x i8*], align 8 @@ -58,11 +61,11 @@ ; CHECK-NEXT: %4 = bitcast [1 x i8*]* %.offload_ptrs to double** ; CHECK-NEXT: store double* %a, double** %4, align 8 -; CHECK-NEXT: %handle = call %struct.__tgt_async_info @__tgt_target_data_begin_mapper_issue(i64 -1, i32 1, i8** %1, i8** %3, i64* getelementptr inbounds ([1 x i64], [1 x i64]* @.offload_sizes.1, i64 0, i64 0), i64* getelementptr inbounds ([1 x i64], [1 x i64]* @.offload_maptypes, i64 0, i64 0), i8** null) +; CHECK-NEXT: call void @__tgt_target_data_begin_mapper_issue(i64 -1, i32 1, i8** %1, i8** %3, i64* getelementptr inbounds ([1 x i64], [1 x i64]* @.offload_sizes.1, i64 0, i64 0), i64* getelementptr inbounds ([1 x i64], [1 x i64]* @.offload_maptypes, i64 0, i64 0), i8** null, %struct.__tgt_async_info* %handle) ; CHECK-NEXT: %5 = bitcast double* %a to i64* -; CHECK-NEXT: call void @__tgt_target_data_begin_mapper_wait(i64 -1, %struct.__tgt_async_info %handle) +; CHECK-NEXT: call void @__tgt_target_data_begin_mapper_wait(i64 -1, %struct.__tgt_async_info* %handle) ; CHECK-NEXT: %6 = load i64, i64* %5, align 8 ; CHECK-NEXT: %7 = getelementptr inbounds [1 x i8*], [1 x i8*]* %.offload_baseptrs4, i64 0, i64 0 @@ -157,7 +160,7 @@ ; return random; ;} define dso_local i32 @heavyComputation2(double* %a, i32 %size) { -; CHECK-LABEL: define {{[^@]+}}@heavyComputation2(double* %a, i32 %size) +; CHECK-LABEL: define {{[^@]+}}@heavyComputation2(double* %a, i32 %size) { ; CHECK-NEXT: entry: ; CHECK-NEXT: %size.addr = alloca i32, align 4 ; CHECK-NEXT: %.offload_baseptrs = alloca [2 x i8*], align 8 @@ -297,7 +300,7 @@ ; return random; ;} define dso_local i32 @heavyComputation3(double* noalias %a, i32 %size) { -; CHECK-LABEL: define {{[^@]+}}@heavyComputation3(double* noalias %a, i32 %size) +; CHECK-LABEL: define {{[^@]+}}@heavyComputation3(double* noalias %a, i32 %size) { ; CHECK-NEXT: entry: ; CHECK-NEXT: %size.addr = alloca i32, align 4 ; CHECK-NEXT: %.offload_baseptrs = alloca [2 x i8*], align 8 @@ -435,8 +438,11 @@ ; return random; ;} define dso_local i32 @dataTransferOnly1(double* noalias %a, i32 %size) { -; CHECK-LABEL: define {{[^@]+}}@dataTransferOnly1(double* noalias %a, i32 %size) +; CHECK-LABEL: define {{[^@]+}}@dataTransferOnly1(double* noalias %a, i32 %size) { ; CHECK-NEXT: entry: + +; CHECK-NEXT: %handle = alloca %struct.__tgt_async_info, align 8 + ; CHECK-NEXT: %.offload_baseptrs = alloca [1 x i8*], align 8 ; CHECK-NEXT: %.offload_ptrs = alloca [1 x i8*], align 8 ; CHECK-NEXT: %.offload_sizes = alloca [1 x i64], align 8 @@ -452,11 +458,11 @@ ; CHECK-NEXT: %5 = getelementptr inbounds [1 x i64], [1 x i64]* %.offload_sizes, i64 0, i64 0 ; CHECK-NEXT: store i64 %0, i64* %5, align 8 -; CHECK-NEXT: %handle = call %struct.__tgt_async_info @__tgt_target_data_begin_mapper_issue(i64 -1, i32 1, i8** %1, i8** %3, i64* %5, i64* getelementptr inbounds ([1 x i64], [1 x i64]* @.offload_maptypes.5, i64 0, i64 0), i8** null) +; CHECK-NEXT: call void @__tgt_target_data_begin_mapper_issue(i64 -1, i32 1, i8** %1, i8** %3, i64* %5, i64* getelementptr inbounds ([1 x i64], [1 x i64]* @.offload_maptypes.5, i64 0, i64 0), i8** null, %struct.__tgt_async_info* %handle) ; CHECK-NEXT: %rem = urem i32 %call, %size -; CHECK-NEXT: call void @__tgt_target_data_begin_mapper_wait(i64 -1, %struct.__tgt_async_info %handle) +; CHECK-NEXT: call void @__tgt_target_data_begin_mapper_wait(i64 -1, %struct.__tgt_async_info* %handle) ; CHECK-NEXT: call void @__tgt_target_data_end_mapper(i64 -1, i32 1, i8** nonnull %1, i8** nonnull %3, i64* nonnull %5, i64* getelementptr inbounds ([1 x i64], [1 x i64]* @.offload_maptypes.5, i64 0, i64 0), i8** null) ; CHECK-NEXT: ret i32 %rem @@ -493,5 +499,5 @@ declare dso_local i32 @rand(...) -; CHECK: declare %struct.__tgt_async_info @__tgt_target_data_begin_mapper_issue(i64, i32, i8**, i8**, i64*, i64*, i8**) -; CHECK: declare void @__tgt_target_data_begin_mapper_wait(i64, %struct.__tgt_async_info) +; CHECK: declare void @__tgt_target_data_begin_mapper_issue(i64, i32, i8**, i8**, i64*, i64*, i8**, %struct.__tgt_async_info*) +; CHECK: declare void @__tgt_target_data_begin_mapper_wait(i64, %struct.__tgt_async_info*)