| //===- MapInfoFinalization.cpp -----------------------------------------===// |
| // |
| // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. |
| // See https://llvm.org/LICENSE.txt for license information. |
| // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception |
| // |
| //===----------------------------------------------------------------------===// |
| |
| //===----------------------------------------------------------------------===// |
| /// \file |
| /// An OpenMP dialect related pass for FIR/HLFIR which performs some |
| /// pre-processing of MapInfoOp's after the module has been lowered to |
| /// finalize them. |
| /// |
| /// For example, it expands MapInfoOp's containing descriptor related |
| /// types (fir::BoxType's) into multiple MapInfoOp's containing the parent |
| /// descriptor and pointer member components for individual mapping, |
| /// treating the descriptor type as a record type for later lowering in the |
| /// OpenMP dialect. |
| /// |
| /// The pass also adds MapInfoOp's that are members of a parent object but are |
| /// not directly used in the body of a target region to its BlockArgument list |
| /// to maintain consistency across all MapInfoOp's tied to a region directly or |
| /// indirectly via a parent object. |
| //===----------------------------------------------------------------------===// |
| |
| #include "flang/Optimizer/Builder/DirectivesCommon.h" |
| #include "flang/Optimizer/Builder/FIRBuilder.h" |
| #include "flang/Optimizer/Builder/HLFIRTools.h" |
| #include "flang/Optimizer/Dialect/FIRType.h" |
| #include "flang/Optimizer/Dialect/Support/KindMapping.h" |
| #include "flang/Optimizer/HLFIR/HLFIROps.h" |
| #include "flang/Optimizer/OpenMP/Passes.h" |
| #include "mlir/Analysis/SliceAnalysis.h" |
| #include "mlir/Dialect/Func/IR/FuncOps.h" |
| #include "mlir/Dialect/OpenMP/OpenMPDialect.h" |
| #include "mlir/IR/BuiltinDialect.h" |
| #include "mlir/IR/BuiltinOps.h" |
| #include "mlir/IR/Operation.h" |
| #include "mlir/IR/SymbolTable.h" |
| #include "mlir/Pass/Pass.h" |
| #include "mlir/Support/LLVM.h" |
| #include "llvm/ADT/BitmaskEnum.h" |
| #include "llvm/ADT/SmallPtrSet.h" |
| #include "llvm/ADT/StringSet.h" |
| #include "llvm/Support/raw_ostream.h" |
| #include <algorithm> |
| #include <cstddef> |
| #include <iterator> |
| #include <numeric> |
| |
| #define DEBUG_TYPE "omp-map-info-finalization" |
| |
| namespace flangomp { |
| #define GEN_PASS_DEF_MAPINFOFINALIZATIONPASS |
| #include "flang/Optimizer/OpenMP/Passes.h.inc" |
| } // namespace flangomp |
| |
| namespace { |
| class MapInfoFinalizationPass |
| : public flangomp::impl::MapInfoFinalizationPassBase< |
| MapInfoFinalizationPass> { |
| /// Helper class tracking a members parent and its |
| /// placement in the parents member list |
| struct ParentAndPlacement { |
| mlir::omp::MapInfoOp parent; |
| size_t index; |
| }; |
| |
| /// Tracks any intermediate function/subroutine local allocations we |
| /// generate for the descriptors of box type dummy arguments, so that |
| /// we can retrieve it for subsequent reuses within the functions |
| /// scope. |
| /// |
| /// descriptor defining op |
| /// | corresponding local alloca |
| /// | | |
| std::map<mlir::Operation *, mlir::Value> localBoxAllocas; |
| |
| // List of deferrable descriptors to process at the end of |
| // the pass. |
| llvm::SmallVector<mlir::Operation *> deferrableDesc; |
| |
| /// Return true if the given path exists in a list of paths. |
| static bool |
| containsPath(const llvm::SmallVectorImpl<llvm::SmallVector<int64_t>> &paths, |
| llvm::ArrayRef<int64_t> path) { |
| return llvm::any_of(paths, [&](const llvm::SmallVector<int64_t> &p) { |
| return p.size() == path.size() && |
| std::equal(p.begin(), p.end(), path.begin()); |
| }); |
| } |
| |
| /// Return true if the given path is already present in |
| /// op.getMembersIndexAttr(). |
| static bool mappedIndexPathExists(mlir::omp::MapInfoOp op, |
| llvm::ArrayRef<int64_t> indexPath) { |
| if (mlir::ArrayAttr attr = op.getMembersIndexAttr()) { |
| for (mlir::Attribute list : attr) { |
| auto listAttr = mlir::cast<mlir::ArrayAttr>(list); |
| if (listAttr.size() != indexPath.size()) |
| continue; |
| bool allEq = true; |
| for (auto [i, val] : llvm::enumerate(listAttr)) { |
| if (mlir::cast<mlir::IntegerAttr>(val).getInt() != indexPath[i]) { |
| allEq = false; |
| break; |
| } |
| } |
| if (allEq) |
| return true; |
| } |
| } |
| return false; |
| } |
| |
| /// Build a compact string key for an index path for set-based |
| /// deduplication. Format: "N:v0,v1,..." where N is the length. |
| static void buildPathKey(llvm::ArrayRef<int64_t> path, |
| llvm::SmallString<64> &outKey) { |
| outKey.clear(); |
| llvm::raw_svector_ostream os(outKey); |
| os << path.size() << ':'; |
| for (size_t i = 0; i < path.size(); ++i) { |
| if (i) |
| os << ','; |
| os << path[i]; |
| } |
| } |
| |
| /// Return true if the module has an OpenMP requires clause that includes |
| /// unified_shared_memory. |
| static bool moduleRequiresUSM(mlir::ModuleOp module) { |
| assert(module && "invalid module"); |
| if (auto req = module->getAttrOfType<mlir::omp::ClauseRequiresAttr>( |
| "omp.requires")) |
| return mlir::omp::bitEnumContainsAll( |
| req.getValue(), mlir::omp::ClauseRequires::unified_shared_memory); |
| return false; |
| } |
| |
| /// Create the member map for coordRef and append it (and its index |
| /// path) to the provided new* vectors, if it is not already present. |
| void appendMemberMapIfNew( |
| mlir::omp::MapInfoOp op, fir::FirOpBuilder &builder, mlir::Location loc, |
| mlir::Value coordRef, llvm::ArrayRef<int64_t> indexPath, |
| llvm::StringRef memberName, |
| llvm::SmallVectorImpl<mlir::Value> &newMapOpsForFields, |
| llvm::SmallVectorImpl<llvm::SmallVector<int64_t>> &newMemberIndexPaths) { |
| // Local de-dup within this op invocation. |
| if (containsPath(newMemberIndexPaths, indexPath)) |
| return; |
| // Global de-dup against already present member indices. |
| if (mappedIndexPathExists(op, indexPath)) |
| return; |
| |
| if (op.getMapperId()) { |
| mlir::omp::DeclareMapperOp symbol = |
| mlir::SymbolTable::lookupNearestSymbolFrom< |
| mlir::omp::DeclareMapperOp>(op, op.getMapperIdAttr()); |
| assert(symbol && "missing symbol for declare mapper identifier"); |
| mlir::omp::DeclareMapperInfoOp mapperInfo = symbol.getDeclareMapperInfo(); |
| // TODO: Probably a way to cache these keys in someway so we don't |
| // constantly go through the process of rebuilding them on every check, to |
| // save some cycles, but it can wait for a subsequent patch. |
| for (auto v : mapperInfo.getMapVars()) { |
| mlir::omp::MapInfoOp map = |
| mlir::cast<mlir::omp::MapInfoOp>(v.getDefiningOp()); |
| if (!map.getMembers().empty() && mappedIndexPathExists(map, indexPath)) |
| return; |
| } |
| } |
| |
| builder.setInsertionPoint(op); |
| fir::factory::AddrAndBoundsInfo info = fir::factory::getDataOperandBaseAddr( |
| builder, coordRef, /*isOptional=*/false, loc); |
| llvm::SmallVector<mlir::Value> bounds = fir::factory::genImplicitBoundsOps< |
| mlir::omp::MapBoundsOp, mlir::omp::MapBoundsType>( |
| builder, info, |
| hlfir::translateToExtendedValue(loc, builder, hlfir::Entity{coordRef}) |
| .first, |
| /*dataExvIsAssumedSize=*/false, loc); |
| |
| mlir::omp::MapInfoOp fieldMapOp = mlir::omp::MapInfoOp::create( |
| builder, loc, coordRef.getType(), coordRef, |
| mlir::TypeAttr::get(fir::unwrapRefType(coordRef.getType())), |
| op.getMapTypeAttr(), |
| builder.getAttr<mlir::omp::VariableCaptureKindAttr>( |
| mlir::omp::VariableCaptureKind::ByRef), |
| /*varPtrPtr=*/mlir::Value{}, /*members=*/mlir::ValueRange{}, |
| /*members_index=*/mlir::ArrayAttr{}, bounds, |
| /*mapperId=*/mlir::FlatSymbolRefAttr(), |
| builder.getStringAttr(op.getNameAttr().strref() + "." + memberName + |
| ".implicit_map"), |
| /*partial_map=*/builder.getBoolAttr(false)); |
| |
| newMapOpsForFields.emplace_back(fieldMapOp); |
| newMemberIndexPaths.emplace_back(indexPath.begin(), indexPath.end()); |
| } |
| |
| // Check if the declaration operation we have refers to a dummy |
| // function argument. |
| bool isDummyArgument(mlir::Value mappedValue) { |
| if (auto declareOp = mlir::dyn_cast_if_present<hlfir::DeclareOp>( |
| mappedValue.getDefiningOp())) |
| if (auto dummyScope = declareOp.getDummyScope()) |
| return true; |
| return false; |
| } |
| |
| // Relevant for OpenMP < 5.2, where attach semantics and rules don't exist. |
| // As descriptors were an unspoken implementation detail in these versions |
| // there's certain cases where the user (and the compiler implementation) |
| // can create data mapping errors by having temporary descriptors stuck |
| // in memory. The main example is calling an 'target enter data map' |
| // without a corresponding exit on an assumed shape or size dummy |
| // argument, a local stack descriptor is generated, gets mapped and |
| // is then left on device. A user doesn't realize what they've done as |
| // the OpenMP specification isn't explicit on descriptor handling in |
| // earlier versions and as far as Fortran is concerned this si something |
| // hidden from a user. To avoid this we can defer the descriptor mapping |
| // in these cases until target or target data regions, when we can be |
| // sure they have a clear limited scope on device. |
| bool canDeferDescriptorMapping(mlir::Value descriptor) { |
| if (fir::isAllocatableType(descriptor.getType()) || |
| fir::isPointerType(descriptor.getType())) |
| return false; |
| if (isDummyArgument(descriptor) && |
| (fir::isAssumedType(descriptor.getType()) || |
| fir::isAssumedShape(descriptor.getType()))) |
| return true; |
| return false; |
| } |
| |
| /// getMemberUserList gathers all users of a particular MapInfoOp that are |
| /// other MapInfoOp's and places them into the mapMemberUsers list, which |
| /// records the map that the current argument MapInfoOp "op" is part of |
| /// alongside the placement of "op" in the recorded users members list. The |
| /// intent of the generated list is to find all MapInfoOp's that may be |
| /// considered parents of the passed in "op" and in which it shows up in the |
| /// member list, alongside collecting the placement information of "op" in its |
| /// parents member list. |
| void |
| getMemberUserList(mlir::omp::MapInfoOp op, |
| llvm::SmallVectorImpl<ParentAndPlacement> &mapMemberUsers) { |
| for (auto *user : op->getUsers()) |
| if (auto map = mlir::dyn_cast_if_present<mlir::omp::MapInfoOp>(user)) |
| for (auto [i, mapMember] : llvm::enumerate(map.getMembers())) |
| if (mapMember.getDefiningOp() == op) |
| mapMemberUsers.push_back({map, i}); |
| } |
| |
| void getAsIntegers(llvm::ArrayRef<mlir::Attribute> values, |
| llvm::SmallVectorImpl<int64_t> &ints) { |
| ints.reserve(values.size()); |
| llvm::transform(values, std::back_inserter(ints), |
| [](mlir::Attribute value) { |
| return mlir::cast<mlir::IntegerAttr>(value).getInt(); |
| }); |
| } |
| |
| /// This function will expand a MapInfoOp's member indices back into a vector |
| /// so that they can be trivially modified as unfortunately the attribute type |
| /// that's used does not have modifiable fields at the moment (generally |
| /// awkward to work with) |
| void getMemberIndicesAsVectors( |
| mlir::omp::MapInfoOp mapInfo, |
| llvm::SmallVectorImpl<llvm::SmallVector<int64_t>> &indices) { |
| indices.reserve(mapInfo.getMembersIndexAttr().getValue().size()); |
| llvm::transform(mapInfo.getMembersIndexAttr().getValue(), |
| std::back_inserter(indices), [this](mlir::Attribute value) { |
| auto memberIndex = mlir::cast<mlir::ArrayAttr>(value); |
| llvm::SmallVector<int64_t> indexes; |
| getAsIntegers(memberIndex.getValue(), indexes); |
| return indexes; |
| }); |
| } |
| |
| /// When provided a MapInfoOp containing a descriptor type that |
| /// we must expand into multiple maps this function will extract |
| /// the value from it and return it, in certain cases we must |
| /// generate a new allocation to store into so that the |
| /// fir::BoxOffsetOp we utilise to access the descriptor datas |
| /// base address can be utilised. |
| mlir::Value getDescriptorFromBoxMap(mlir::omp::MapInfoOp boxMap, |
| fir::FirOpBuilder &builder, |
| bool &canDescBeDeferred) { |
| mlir::Value descriptor = boxMap.getVarPtr(); |
| if (!fir::isTypeWithDescriptor(boxMap.getVarType())) |
| if (auto addrOp = mlir::dyn_cast_if_present<fir::BoxAddrOp>( |
| boxMap.getVarPtr().getDefiningOp())) |
| descriptor = addrOp.getVal(); |
| |
| canDescBeDeferred = canDeferDescriptorMapping(descriptor); |
| |
| if (!mlir::isa<fir::BaseBoxType>(descriptor.getType()) && |
| !fir::factory::isOptionalArgument(descriptor.getDefiningOp())) |
| return descriptor; |
| |
| mlir::Value &alloca = localBoxAllocas[descriptor.getDefiningOp()]; |
| mlir::Location loc = boxMap->getLoc(); |
| |
| if (!alloca) { |
| // The fir::BoxOffsetOp only works with !fir.ref<!fir.box<...>> types, as |
| // allowing it to access non-reference box operations can cause some |
| // problematic SSA IR. However, in the case of assumed shape's the type |
| // is not a !fir.ref, in these cases to retrieve the appropriate |
| // !fir.ref<!fir.box<...>> to access the data we need to map we must |
| // perform an alloca and then store to it and retrieve the data from the |
| // new alloca. |
| mlir::OpBuilder::InsertPoint insPt = builder.saveInsertionPoint(); |
| mlir::Block *allocaBlock = builder.getAllocaBlock(); |
| assert(allocaBlock && "No alloca block found for this top level op"); |
| builder.setInsertionPointToStart(allocaBlock); |
| |
| mlir::Type allocaType = descriptor.getType(); |
| if (fir::isBoxAddress(allocaType)) |
| allocaType = fir::unwrapRefType(allocaType); |
| alloca = fir::AllocaOp::create(builder, loc, allocaType); |
| builder.restoreInsertionPoint(insPt); |
| } |
| |
| // We should only emit a store if the passed in data is present, it is |
| // possible a user passes in no argument to an optional parameter, in which |
| // case we cannot store or we'll segfault on the emitted memcpy. |
| // TODO: We currently emit a present -> load/store every time we use a |
| // mapped value that requires a local allocation, this isn't the most |
| // efficient, although, it is more correct in a lot of situations. One |
| // such situation is emitting a this series of instructions in separate |
| // segments of a branch (e.g. two target regions in separate else/if branch |
| // mapping the same function argument), however, it would be nice to be able |
| // to optimize these situations e.g. raising the load/store out of the |
| // branch if possible. But perhaps this is best left to lower level |
| // optimisation passes. |
| auto isPresent = |
| fir::IsPresentOp::create(builder, loc, builder.getI1Type(), descriptor); |
| builder.genIfOp(loc, {}, isPresent, false) |
| .genThen([&]() { |
| descriptor = builder.loadIfRef(loc, descriptor); |
| fir::StoreOp::create(builder, loc, descriptor, alloca); |
| }) |
| .end(); |
| return alloca; |
| } |
| |
| /// Function that generates a FIR operation accessing the descriptor's |
| /// base address (BoxOffsetOp) and a MapInfoOp for it. The most |
| /// important thing to note is that we normally move the bounds from |
| /// the descriptor map onto the base address map. |
| mlir::omp::MapInfoOp |
| genBaseAddrMap(mlir::Value descriptor, mlir::OperandRange bounds, |
| mlir::omp::ClauseMapFlags mapType, fir::FirOpBuilder &builder, |
| mlir::FlatSymbolRefAttr mapperId = mlir::FlatSymbolRefAttr()) { |
| mlir::Location loc = descriptor.getLoc(); |
| mlir::Value baseAddrAddr = fir::BoxOffsetOp::create( |
| builder, loc, descriptor, fir::BoxFieldAttr::base_addr); |
| |
| mlir::Type underlyingVarType = |
| llvm::cast<mlir::omp::PointerLikeType>( |
| fir::unwrapRefType(baseAddrAddr.getType())) |
| .getElementType(); |
| if (auto seqType = llvm::dyn_cast<fir::SequenceType>(underlyingVarType)) |
| if (seqType.hasDynamicExtents()) |
| underlyingVarType = seqType.getEleTy(); |
| |
| // Member of the descriptor pointing at the allocated data |
| return mlir::omp::MapInfoOp::create( |
| builder, loc, baseAddrAddr.getType(), descriptor, |
| mlir::TypeAttr::get(underlyingVarType), |
| builder.getAttr<mlir::omp::ClauseMapFlagsAttr>(mapType), |
| builder.getAttr<mlir::omp::VariableCaptureKindAttr>( |
| mlir::omp::VariableCaptureKind::ByRef), |
| baseAddrAddr, /*members=*/mlir::SmallVector<mlir::Value>{}, |
| /*membersIndex=*/mlir::ArrayAttr{}, bounds, |
| /*mapperId=*/mapperId, |
| /*name=*/builder.getStringAttr(""), |
| /*partial_map=*/builder.getBoolAttr(false)); |
| } |
| |
| /// This function adjusts the member indices vector to include a new |
| /// base address member. We take the position of the descriptor in |
| /// the member indices list, which is the index data that the base |
| /// addresses index will be based off of, as the base address is |
| /// a member of the descriptor. We must also alter other members |
| /// that are members of this descriptor to account for the addition |
| /// of the base address index. |
| void adjustMemberIndices( |
| llvm::SmallVectorImpl<llvm::SmallVector<int64_t>> &memberIndices, |
| size_t memberIndex) { |
| llvm::SmallVector<int64_t> baseAddrIndex = memberIndices[memberIndex]; |
| |
| // If we find another member that is "derived/a member of" the descriptor |
| // that is not the descriptor itself, we must insert a 0 for the new base |
| // address we have just added for the descriptor into the list at the |
| // appropriate position to maintain correctness of the positional/index data |
| // for that member. |
| for (llvm::SmallVector<int64_t> &member : memberIndices) |
| if (member.size() > baseAddrIndex.size() && |
| std::equal(baseAddrIndex.begin(), baseAddrIndex.end(), |
| member.begin())) |
| member.insert(std::next(member.begin(), baseAddrIndex.size()), 0); |
| |
| // Add the base address index to the main base address member data |
| baseAddrIndex.push_back(0); |
| |
| // Insert our newly created baseAddrIndex into the larger list of indices at |
| // the correct location. |
| memberIndices.insert(std::next(memberIndices.begin(), memberIndex + 1), |
| baseAddrIndex); |
| } |
| |
| /// Adjusts the descriptor's map type. The main alteration that is done |
| /// currently is transforming the map type to `OMP_MAP_TO` where possible. |
| /// This is because we will always need to map the descriptor to device |
| /// (or at the very least it seems to be the case currently with the |
| /// current lowered kernel IR), as without the appropriate descriptor |
| /// information on the device there is a risk of the kernel IR |
| /// requesting for various data that will not have been copied to |
| /// perform things like indexing. This can cause segfaults and |
| /// memory access errors. However, we do not need this data mapped |
| /// back to the host from the device, as per the OpenMP spec we cannot alter |
| /// the data via resizing or deletion on the device. Discarding any |
| /// descriptor alterations via no map back is reasonable (and required |
| /// for certain segments of descriptor data like the type descriptor that are |
| /// global constants). This alteration is only inapplicable to `target exit` |
| /// and `target update` currently, and that's due to `target exit` not |
| /// allowing `to` mappings, and `target update` not allowing both `to` and |
| /// `from` simultaneously. We currently try to maintain the `implicit` flag |
| /// where necessary, although it does not seem strictly required. |
| mlir::omp::ClauseMapFlags |
| getDescriptorMapType(mlir::omp::ClauseMapFlags mapTypeFlag, |
| mlir::Operation *target) { |
| using mapFlags = mlir::omp::ClauseMapFlags; |
| if (llvm::isa_and_nonnull<mlir::omp::TargetExitDataOp, |
| mlir::omp::TargetUpdateOp>(target)) |
| return mapTypeFlag; |
| |
| mapFlags flags = |
| mapFlags::to | (mapTypeFlag & (mapFlags::implicit | mapFlags::always)); |
| |
| // Descriptors for objects will always be copied. This is because the |
| // descriptor can be rematerialized by the compiler, and so the address |
| // of the descriptor for a given object at one place in the code may |
| // differ from that address in another place. The contents of the |
| // descriptor (the base address in particular) will remain unchanged |
| // though. |
| // TODO/FIXME: We currently cannot have MAP_CLOSE and MAP_ALWAYS on |
| // the descriptor at once, these are mutually exclusive and when |
| // both are applied the runtime will fail to map. |
| flags |= ((mapFlags(mapTypeFlag) & mapFlags::close) == mapFlags::close) |
| ? mapFlags::close |
| : mapFlags::always; |
| |
| // For unified_shared_memory, we additionally add `CLOSE` on the descriptor |
| // to ensure device-local placement where required by tests relying on USM + |
| // close semantics. |
| if (moduleRequiresUSM(target->getParentOfType<mlir::ModuleOp>())) |
| flags |= mapFlags::close; |
| return flags; |
| } |
| |
| /// Check if the mapOp is present in the HasDeviceAddr clause on |
| /// the userOp. Only applies to TargetOp. |
| bool isHasDeviceAddr(mlir::omp::MapInfoOp mapOp, mlir::Operation &userOp) { |
| if (auto targetOp = llvm::dyn_cast<mlir::omp::TargetOp>(userOp)) { |
| for (mlir::Value hda : targetOp.getHasDeviceAddrVars()) { |
| if (hda.getDefiningOp() == mapOp) |
| return true; |
| } |
| } |
| return false; |
| } |
| |
| bool isUseDeviceAddr(mlir::omp::MapInfoOp mapOp, mlir::Operation &userOp) { |
| if (auto targetDataOp = llvm::dyn_cast<mlir::omp::TargetDataOp>(userOp)) { |
| for (mlir::Value uda : targetDataOp.getUseDeviceAddrVars()) { |
| if (uda.getDefiningOp() == mapOp) |
| return true; |
| } |
| } |
| return false; |
| } |
| |
| bool isUseDevicePtr(mlir::omp::MapInfoOp mapOp, mlir::Operation &userOp) { |
| if (auto targetDataOp = llvm::dyn_cast<mlir::omp::TargetDataOp>(userOp)) { |
| for (mlir::Value udp : targetDataOp.getUseDevicePtrVars()) { |
| if (udp.getDefiningOp() == mapOp) |
| return true; |
| } |
| } |
| return false; |
| } |
| |
| // Expand mappings of type(C_PTR) to map their `__address` field explicitly |
| // as a single pointer-sized member (USM-gated at callsite). This helps in |
| // USM scenarios to ensure the pointer-sized mapping is used. |
| mlir::omp::MapInfoOp genCptrMemberMap(mlir::omp::MapInfoOp op, |
| fir::FirOpBuilder &builder) { |
| if (!op.getMembers().empty()) |
| return op; |
| |
| mlir::Type varTy = fir::unwrapRefType(op.getVarPtr().getType()); |
| if (!mlir::isa<fir::RecordType>(varTy)) |
| return op; |
| auto recTy = mlir::cast<fir::RecordType>(varTy); |
| // If not a builtin C_PTR record, skip. |
| if (!recTy.getName().ends_with("__builtin_c_ptr")) |
| return op; |
| |
| // Find the index of the c_ptr address component named "__address". |
| int32_t fieldIdx = recTy.getFieldIndex("__address"); |
| if (fieldIdx < 0) |
| return op; |
| |
| mlir::Location loc = op.getVarPtr().getLoc(); |
| mlir::Type memTy = recTy.getType(fieldIdx); |
| fir::IntOrValue idxConst = |
| mlir::IntegerAttr::get(builder.getI32Type(), fieldIdx); |
| mlir::Value coord = fir::CoordinateOp::create( |
| builder, loc, builder.getRefType(memTy), op.getVarPtr(), |
| llvm::SmallVector<fir::IntOrValue, 1>{idxConst}); |
| |
| // Child for the `__address` member. |
| llvm::SmallVector<llvm::SmallVector<int64_t>> memberIdx = {{0}}; |
| mlir::ArrayAttr newMembersAttr = builder.create2DI64ArrayAttr(memberIdx); |
| // Force CLOSE in USM paths so the pointer gets device-local placement |
| // when required by tests relying on USM + close semantics. |
| mlir::omp::ClauseMapFlagsAttr mapTypeAttr = |
| builder.getAttr<mlir::omp::ClauseMapFlagsAttr>( |
| op.getMapType() | mlir::omp::ClauseMapFlags::close); |
| |
| mlir::omp::MapInfoOp memberMap = mlir::omp::MapInfoOp::create( |
| builder, loc, coord.getType(), coord, |
| mlir::TypeAttr::get(fir::unwrapRefType(coord.getType())), mapTypeAttr, |
| builder.getAttr<mlir::omp::VariableCaptureKindAttr>( |
| mlir::omp::VariableCaptureKind::ByRef), |
| /*varPtrPtr=*/mlir::Value{}, |
| /*members=*/llvm::SmallVector<mlir::Value>{}, |
| /*member_index=*/mlir::ArrayAttr{}, |
| /*bounds=*/op.getBounds(), |
| /*mapperId=*/mlir::FlatSymbolRefAttr(), |
| /*name=*/op.getNameAttr(), |
| /*partial_map=*/builder.getBoolAttr(false)); |
| |
| // Rebuild the parent as a container with the `__address` member. |
| mlir::omp::MapInfoOp newParent = mlir::omp::MapInfoOp::create( |
| builder, op.getLoc(), op.getResult().getType(), op.getVarPtr(), |
| op.getVarTypeAttr(), mapTypeAttr, op.getMapCaptureTypeAttr(), |
| /*varPtrPtr=*/mlir::Value{}, |
| /*members=*/llvm::SmallVector<mlir::Value>{memberMap}, |
| /*member_index=*/newMembersAttr, |
| /*bounds=*/llvm::SmallVector<mlir::Value>{}, |
| /*mapperId=*/mlir::FlatSymbolRefAttr(), op.getNameAttr(), |
| /*partial_map=*/builder.getBoolAttr(false)); |
| op.replaceAllUsesWith(newParent.getResult()); |
| op->erase(); |
| return newParent; |
| } |
| |
| mlir::omp::MapInfoOp genDescriptorMemberMaps(mlir::omp::MapInfoOp op, |
| fir::FirOpBuilder &builder, |
| mlir::Operation *target) { |
| llvm::SmallVector<ParentAndPlacement> mapMemberUsers; |
| getMemberUserList(op, mapMemberUsers); |
| |
| // TODO: map the addendum segment of the descriptor, similarly to the |
| // base address/data pointer member. |
| bool descCanBeDeferred = false; |
| mlir::Value descriptor = |
| getDescriptorFromBoxMap(op, builder, descCanBeDeferred); |
| |
| mlir::ArrayAttr newMembersAttr; |
| mlir::SmallVector<mlir::Value> newMembers; |
| llvm::SmallVector<llvm::SmallVector<int64_t>> memberIndices; |
| bool isHasDeviceAddrFlag = isHasDeviceAddr(op, *target); |
| |
| if (!mapMemberUsers.empty() || !op.getMembers().empty()) |
| getMemberIndicesAsVectors( |
| !mapMemberUsers.empty() ? mapMemberUsers[0].parent : op, |
| memberIndices); |
| |
| // If the operation that we are expanding with a descriptor has a user |
| // (parent), then we have to expand the parent's member indices to reflect |
| // the adjusted member indices for the base address insertion. However, if |
| // it does not then we are expanding a MapInfoOp without any pre-existing |
| // member information to now have one new member for the base address, or |
| // we are expanding a parent that is a descriptor and we have to adjust |
| // all of its members to reflect the insertion of the base address. |
| // |
| // If we're expanding a top-level descriptor for a map operation that |
| // resulted from "has_device_addr" clause, then we want the base pointer |
| // from the descriptor to be used verbatim, i.e. without additional |
| // remapping. To avoid this remapping, simply don't generate any map |
| // information for the descriptor members. |
| mlir::FlatSymbolRefAttr mapperId = op.getMapperIdAttr(); |
| if (!mapMemberUsers.empty()) { |
| // Currently, there should only be one user per map when this pass |
| // is executed. Either a parent map, holding the current map in its |
| // member list, or a target operation that holds a map clause. This |
| // may change in the future if we aim to refactor the MLIR for map |
| // clauses to allow sharing of duplicate maps across target |
| // operations. |
| assert(mapMemberUsers.size() == 1 && |
| "OMPMapInfoFinalization currently only supports single users of a " |
| "MapInfoOp"); |
| auto baseAddr = genBaseAddrMap(descriptor, op.getBounds(), |
| op.getMapType(), builder, mapperId); |
| ParentAndPlacement mapUser = mapMemberUsers[0]; |
| adjustMemberIndices(memberIndices, mapUser.index); |
| llvm::SmallVector<mlir::Value> newMemberOps; |
| for (auto v : mapUser.parent.getMembers()) { |
| newMemberOps.push_back(v); |
| if (v == op) |
| newMemberOps.push_back(baseAddr); |
| } |
| mapUser.parent.getMembersMutable().assign(newMemberOps); |
| mapUser.parent.setMembersIndexAttr( |
| builder.create2DI64ArrayAttr(memberIndices)); |
| } else if (!isHasDeviceAddrFlag) { |
| auto baseAddr = genBaseAddrMap(descriptor, op.getBounds(), |
| op.getMapType(), builder, mapperId); |
| newMembers.push_back(baseAddr); |
| if (!op.getMembers().empty()) { |
| for (auto &indices : memberIndices) |
| indices.insert(indices.begin(), 0); |
| memberIndices.insert(memberIndices.begin(), {0}); |
| newMembersAttr = builder.create2DI64ArrayAttr(memberIndices); |
| newMembers.append(op.getMembers().begin(), op.getMembers().end()); |
| } else { |
| llvm::SmallVector<llvm::SmallVector<int64_t>> memberIdx = {{0}}; |
| newMembersAttr = builder.create2DI64ArrayAttr(memberIdx); |
| } |
| } |
| |
| // Descriptors for objects listed on the `has_device_addr` will always |
| // be copied. This is because the descriptor can be rematerialized by the |
| // compiler, and so the address of the descriptor for a given object at |
| // one place in the code may differ from that address in another place. |
| // The contents of the descriptor (the base address in particular) will |
| // remain unchanged though. |
| mlir::omp::ClauseMapFlags mapType = op.getMapType(); |
| if (isHasDeviceAddrFlag) { |
| mapType |= mlir::omp::ClauseMapFlags::always; |
| } |
| |
| mlir::omp::MapInfoOp newDescParentMapOp = mlir::omp::MapInfoOp::create( |
| builder, op->getLoc(), op.getResult().getType(), descriptor, |
| mlir::TypeAttr::get(fir::unwrapRefType(descriptor.getType())), |
| builder.getAttr<mlir::omp::ClauseMapFlagsAttr>( |
| getDescriptorMapType(mapType, target)), |
| op.getMapCaptureTypeAttr(), /*varPtrPtr=*/mlir::Value{}, newMembers, |
| newMembersAttr, /*bounds=*/mlir::SmallVector<mlir::Value>{}, |
| /*mapperId=*/mlir::FlatSymbolRefAttr(), op.getNameAttr(), |
| /*partial_map=*/builder.getBoolAttr(false)); |
| op.replaceAllUsesWith(newDescParentMapOp.getResult()); |
| op->erase(); |
| |
| if (descCanBeDeferred) |
| deferrableDesc.push_back(newDescParentMapOp); |
| |
| return newDescParentMapOp; |
| } |
| |
| // We add all mapped record members not directly used in the target region |
| // to the block arguments in front of their parent and we place them into |
| // the map operands list for consistency. |
| // |
| // These indirect uses (via accesses to their parent) will still be |
| // mapped individually in most cases, and a parent mapping doesn't |
| // guarantee the parent will be mapped in its totality, partial |
| // mapping is common. |
| // |
| // For example: |
| // map(tofrom: x%y) |
| // |
| // Will generate a mapping for "x" (the parent) and "y" (the member). |
| // The parent "x" will not be mapped, but the member "y" will. |
| // However, we must have the parent as a BlockArg and MapOperand |
| // in these cases, to maintain the correct uses within the region and |
| // to help tracking that the member is part of a larger object. |
| // |
| // In the case of: |
| // map(tofrom: x%y, x%z) |
| // |
| // The parent member becomes more critical, as we perform a partial |
| // structure mapping where we link the mapping of the members y |
| // and z together via the parent x. We do this at a kernel argument |
| // level in LLVM IR and not just MLIR, which is important to maintain |
| // similarity to Clang and for the runtime to do the correct thing. |
| // However, we still do not map the structure in its totality but |
| // rather we generate an un-sized "binding" map entry for it. |
| // |
| // In the case of: |
| // map(tofrom: x, x%y, x%z) |
| // |
| // We do actually map the entirety of "x", so the explicit mapping of |
| // x%y, x%z becomes unnecessary. It is redundant to write this from a |
| // Fortran OpenMP perspective (although it is legal), as even if the |
| // members were allocatables or pointers, we are mandated by the |
| // specification to map these (and any recursive components) in their |
| // entirety, which is different to the C++ equivalent, which requires |
| // explicit mapping of these segments. |
| void addImplicitMembersToTarget(mlir::omp::MapInfoOp op, |
| fir::FirOpBuilder &builder, |
| mlir::Operation *target) { |
| auto mapClauseOwner = |
| llvm::dyn_cast_if_present<mlir::omp::MapClauseOwningOpInterface>( |
| target); |
| // TargetDataOp is technically a MapClauseOwningOpInterface, so we |
| // do not need to explicitly check for the extra cases here for use_device |
| // addr/ptr |
| if (!mapClauseOwner) |
| return; |
| |
| auto addOperands = [&](mlir::MutableOperandRange &mutableOpRange, |
| mlir::Operation *directiveOp, |
| unsigned blockArgInsertIndex = 0) { |
| if (!llvm::is_contained(mutableOpRange.getAsOperandRange(), |
| op.getResult())) |
| return; |
| |
| // There doesn't appear to be a simple way to convert MutableOperandRange |
| // to a vector currently, so we instead use a for_each to populate our |
| // vector. |
| llvm::SmallVector<mlir::Value> newMapOps; |
| newMapOps.reserve(mutableOpRange.size()); |
| llvm::for_each( |
| mutableOpRange.getAsOperandRange(), |
| [&newMapOps](mlir::Value oper) { newMapOps.push_back(oper); }); |
| |
| for (auto mapMember : op.getMembers()) { |
| if (llvm::is_contained(mutableOpRange.getAsOperandRange(), mapMember)) |
| continue; |
| newMapOps.push_back(mapMember); |
| if (directiveOp) { |
| directiveOp->getRegion(0).insertArgument( |
| blockArgInsertIndex, mapMember.getType(), mapMember.getLoc()); |
| blockArgInsertIndex++; |
| } |
| } |
| |
| mutableOpRange.assign(newMapOps); |
| }; |
| |
| auto argIface = |
| llvm::dyn_cast<mlir::omp::BlockArgOpenMPOpInterface>(target); |
| |
| if (auto mapClauseOwner = |
| llvm::dyn_cast<mlir::omp::MapClauseOwningOpInterface>(target)) { |
| mlir::MutableOperandRange mapMutableOpRange = |
| mapClauseOwner.getMapVarsMutable(); |
| unsigned blockArgInsertIndex = |
| argIface |
| ? argIface.getMapBlockArgsStart() + argIface.numMapBlockArgs() |
| : 0; |
| addOperands(mapMutableOpRange, |
| llvm::dyn_cast_if_present<mlir::omp::TargetOp>( |
| argIface.getOperation()), |
| blockArgInsertIndex); |
| } |
| |
| if (auto targetDataOp = llvm::dyn_cast<mlir::omp::TargetDataOp>(target)) { |
| mlir::MutableOperandRange useDevAddrMutableOpRange = |
| targetDataOp.getUseDeviceAddrVarsMutable(); |
| addOperands(useDevAddrMutableOpRange, target, |
| argIface.getUseDeviceAddrBlockArgsStart() + |
| argIface.numUseDeviceAddrBlockArgs()); |
| |
| mlir::MutableOperandRange useDevPtrMutableOpRange = |
| targetDataOp.getUseDevicePtrVarsMutable(); |
| addOperands(useDevPtrMutableOpRange, target, |
| argIface.getUseDevicePtrBlockArgsStart() + |
| argIface.numUseDevicePtrBlockArgs()); |
| } else if (auto targetOp = llvm::dyn_cast<mlir::omp::TargetOp>(target)) { |
| mlir::MutableOperandRange hasDevAddrMutableOpRange = |
| targetOp.getHasDeviceAddrVarsMutable(); |
| addOperands(hasDevAddrMutableOpRange, target, |
| argIface.getHasDeviceAddrBlockArgsStart() + |
| argIface.numHasDeviceAddrBlockArgs()); |
| } |
| } |
| |
| // We retrieve the first user that is a Target operation, of which |
| // there should only be one currently. Every MapInfoOp can be tied to |
| // at most one Target operation and at the minimum no operations. |
| // This may change in the future with IR cleanups/modifications, |
| // in which case this pass will need updating to support cases |
| // where a map can have more than one user and more than one of |
| // those users can be a Target operation. For now, we simply |
| // return the first target operation encountered, which may |
| // be on the parent MapInfoOp in the case of a member mapping. |
| // In that case, we traverse the MapInfoOp chain until we |
| // find the first TargetOp user. |
| mlir::Operation *getFirstTargetUser(mlir::omp::MapInfoOp mapOp) { |
| for (auto *user : mapOp->getUsers()) { |
| if (llvm::isa<mlir::omp::TargetOp, mlir::omp::TargetDataOp, |
| mlir::omp::TargetUpdateOp, mlir::omp::TargetExitDataOp, |
| mlir::omp::TargetEnterDataOp, |
| mlir::omp::DeclareMapperInfoOp>(user)) |
| return user; |
| |
| if (auto mapUser = llvm::dyn_cast<mlir::omp::MapInfoOp>(user)) |
| return getFirstTargetUser(mapUser); |
| } |
| |
| return nullptr; |
| } |
| |
| void addImplicitDescriptorMapToTargetDataOp(mlir::omp::MapInfoOp op, |
| fir::FirOpBuilder &builder, |
| mlir::Operation &target) { |
| // Checks if the map is present as an explicit map already on the target |
| // data directive, and not just present on a use_device_addr/ptr, as if |
| // that's the case, we should not need to add an implicit map for the |
| // descriptor. |
| auto explicitMappingPresent = [](mlir::omp::MapInfoOp op, |
| mlir::omp::TargetDataOp tarData) { |
| // Verify top-level descriptor mapping is at least equal with same |
| // varPtr, the map type should always be To for a descriptor, which is |
| // all we really care about for this mapping as we aim to make sure the |
| // descriptor is always present on device if we're expecting to access |
| // the underlying data. |
| if (tarData.getMapVars().empty()) |
| return false; |
| |
| for (mlir::Value mapVar : tarData.getMapVars()) { |
| auto mapOp = llvm::cast<mlir::omp::MapInfoOp>(mapVar.getDefiningOp()); |
| if (mapOp.getVarPtr() == op.getVarPtr() && |
| mapOp.getVarPtrPtr() == op.getVarPtrPtr()) { |
| return true; |
| } |
| } |
| |
| return false; |
| }; |
| |
| // if we're not a top level descriptor with members (e.g. member of a |
| // derived type), we do not want to perform this step. |
| if (!llvm::isa<mlir::omp::TargetDataOp>(target) || op.getMembers().empty()) |
| return; |
| |
| if (!isUseDeviceAddr(op, target) && !isUseDevicePtr(op, target)) |
| return; |
| |
| auto targetDataOp = llvm::cast<mlir::omp::TargetDataOp>(target); |
| if (explicitMappingPresent(op, targetDataOp)) |
| return; |
| |
| mlir::omp::MapInfoOp newDescParentMapOp = mlir::omp::MapInfoOp::create( |
| builder, op->getLoc(), op.getResult().getType(), op.getVarPtr(), |
| op.getVarTypeAttr(), |
| builder.getAttr<mlir::omp::ClauseMapFlagsAttr>( |
| mlir::omp::ClauseMapFlags::to | mlir::omp::ClauseMapFlags::always), |
| op.getMapCaptureTypeAttr(), /*varPtrPtr=*/mlir::Value{}, |
| mlir::SmallVector<mlir::Value>{}, mlir::ArrayAttr{}, |
| /*bounds=*/mlir::SmallVector<mlir::Value>{}, |
| /*mapperId*/ mlir::FlatSymbolRefAttr(), op.getNameAttr(), |
| /*partial_map=*/builder.getBoolAttr(false)); |
| |
| targetDataOp.getMapVarsMutable().append({newDescParentMapOp}); |
| } |
| |
| void removeTopLevelDescriptor(mlir::omp::MapInfoOp op, |
| fir::FirOpBuilder &builder, |
| mlir::Operation *target) { |
| if (llvm::isa<mlir::omp::TargetOp, mlir::omp::TargetDataOp, |
| mlir::omp::DeclareMapperInfoOp>(target)) |
| return; |
| |
| // if we're not a top level descriptor with members (e.g. member of a |
| // derived type), we do not want to perform this step. |
| if (op.getMembers().empty()) |
| return; |
| |
| mlir::SmallVector<mlir::Value> members = op.getMembers(); |
| mlir::omp::MapInfoOp baseAddr = |
| mlir::dyn_cast_or_null<mlir::omp::MapInfoOp>( |
| members.front().getDefiningOp()); |
| assert(baseAddr && "Expected member to be MapInfoOp"); |
| members.erase(members.begin()); |
| |
| llvm::SmallVector<llvm::SmallVector<int64_t>> memberIndices; |
| getMemberIndicesAsVectors(op, memberIndices); |
| |
| // Can skip the extra processing if there's only 1 member as it'd |
| // be the base addresses, which we're promoting to the parent. |
| mlir::ArrayAttr membersAttr; |
| if (memberIndices.size() > 1) { |
| memberIndices.erase(memberIndices.begin()); |
| membersAttr = builder.create2DI64ArrayAttr(memberIndices); |
| } |
| |
| // VarPtrPtr is tied to detecting if something is a pointer in the later |
| // lowering currently, this at the moment comes tied with |
| // OMP_MAP_PTR_AND_OBJ being applied which breaks the problem this tries to |
| // solve by emitting a 8-byte mapping tied to the descriptor address (even |
| // if we only emit a single map). So we circumvent this by removing the |
| // varPtrPtr mapping, however, a side affect of this is we lose the |
| // additional load from the backend tied to this which is required for |
| // correctness and getting the correct address of the data to perform our |
| // mapping. So we do our load at this stage. |
| // TODO/FIXME: Tidy up the OMP_MAP_PTR_AND_OBJ and varPtrPtr being tied to |
| // if something is a pointer to try and tidy up the implementation a bit. |
| // This is an unfortunate complexity from push-back from upstream. We |
| // could also emit a load at this level for all base addresses as well, |
| // which in turn will simplify the later lowering a bit as well. But first |
| // need to see how well this alteration works. |
| auto loadBaseAddr = |
| builder.loadIfRef(op->getLoc(), baseAddr.getVarPtrPtr()); |
| mlir::omp::MapInfoOp newBaseAddrMapOp = mlir::omp::MapInfoOp::create( |
| builder, op->getLoc(), loadBaseAddr.getType(), loadBaseAddr, |
| baseAddr.getVarTypeAttr(), baseAddr.getMapTypeAttr(), |
| baseAddr.getMapCaptureTypeAttr(), mlir::Value{}, members, membersAttr, |
| baseAddr.getBounds(), |
| /*mapperId*/ mlir::FlatSymbolRefAttr(), op.getNameAttr(), |
| /*partial_map=*/builder.getBoolAttr(false)); |
| op.replaceAllUsesWith(newBaseAddrMapOp.getResult()); |
| op->erase(); |
| baseAddr.erase(); |
| } |
| |
| static bool hasADescriptor(mlir::Operation *varOp, mlir::Type varType) { |
| if (fir::isTypeWithDescriptor(varType) || |
| mlir::isa<fir::BoxCharType>(varType) || |
| mlir::isa_and_present<fir::BoxAddrOp>(varOp)) |
| return true; |
| return false; |
| } |
| |
| // This pass executes on omp::MapInfoOp's containing descriptor based types |
| // (allocatables, pointers, assumed shape etc.) and expanding them into |
| // multiple omp::MapInfoOp's for each pointer member contained within the |
| // descriptor. |
| // |
| // From the perspective of the MLIR pass manager this runs on the top level |
| // operation (usually function) containing the MapInfoOp because this pass |
| // will mutate siblings of MapInfoOp. |
| void runOnOperation() override { |
| mlir::ModuleOp module = getOperation(); |
| if (!module) |
| module = getOperation()->getParentOfType<mlir::ModuleOp>(); |
| fir::KindMapping kindMap = fir::getKindMapping(module); |
| fir::FirOpBuilder builder{module, std::move(kindMap)}; |
| |
| // We wish to maintain some function level scope (currently |
| // just local function scope variables used to load and store box |
| // variables into so we can access their base address, an |
| // quirk of box_offset requires us to have an in memory box, but Fortran |
| // in certain cases does not provide this) whilst not subjecting |
| // ourselves to the possibility of race conditions while this pass |
| // undergoes frequent re-iteration for the near future. So we loop |
| // over function in the module and then map.info inside of those. |
| getOperation()->walk([&](mlir::Operation *func) { |
| if (!mlir::isa<mlir::func::FuncOp, mlir::omp::DeclareMapperOp>(func)) |
| return; |
| // clear all local allocations we made for any boxes in any prior |
| // iterations from previous function scopes. |
| localBoxAllocas.clear(); |
| deferrableDesc.clear(); |
| |
| // Next, walk `omp.map.info` ops to see if any record members should be |
| // implicitly mapped. |
| func->walk([&](mlir::omp::MapInfoOp op) { |
| mlir::Type underlyingType = |
| fir::unwrapRefType(op.getVarPtr().getType()); |
| |
| // TODO Test with and support more complicated cases; like arrays for |
| // records, for example. |
| if (!fir::isRecordWithAllocatableMember(underlyingType)) |
| return mlir::WalkResult::advance(); |
| |
| // TODO For now, only consider `omp.target` ops. Other ops that support |
| // `map` clauses will follow later. |
| mlir::omp::TargetOp target = |
| mlir::dyn_cast_if_present<mlir::omp::TargetOp>( |
| getFirstTargetUser(op)); |
| |
| if (!target) |
| return mlir::WalkResult::advance(); |
| |
| auto mapClauseOwner = |
| llvm::dyn_cast<mlir::omp::MapClauseOwningOpInterface>(*target); |
| |
| int64_t mapVarIdx = mapClauseOwner.getOperandIndexForMap(op); |
| assert(mapVarIdx >= 0 && |
| mapVarIdx < |
| static_cast<int64_t>(mapClauseOwner.getMapVars().size())); |
| |
| auto argIface = |
| llvm::dyn_cast<mlir::omp::BlockArgOpenMPOpInterface>(*target); |
| // TODO How should `map` block argument that correspond to: `private`, |
| // `use_device_addr`, `use_device_ptr`, be handled? |
| mlir::BlockArgument opBlockArg = argIface.getMapBlockArgs()[mapVarIdx]; |
| llvm::SetVector<mlir::Operation *> mapVarForwardSlice; |
| mlir::getForwardSlice(opBlockArg, &mapVarForwardSlice); |
| |
| mapVarForwardSlice.remove_if([&](mlir::Operation *sliceOp) { |
| // TODO Support coordinate_of ops. |
| // |
| // TODO Support call ops by recursively examining the forward slice of |
| // the corresponding parameter to the field in the called function. |
| return !mlir::isa<hlfir::DesignateOp>(sliceOp); |
| }); |
| |
| auto recordType = mlir::cast<fir::RecordType>(underlyingType); |
| llvm::SmallVector<mlir::Value> newMapOpsForFields; |
| llvm::SmallVector<llvm::SmallVector<int64_t>> newMemberIndexPaths; |
| |
| // 1) Handle direct top-level allocatable fields. |
| for (auto fieldMemTyPair : recordType.getTypeList()) { |
| auto &field = fieldMemTyPair.first; |
| auto memTy = fieldMemTyPair.second; |
| |
| if (!fir::isAllocatableType(memTy)) |
| continue; |
| |
| bool referenced = llvm::any_of(mapVarForwardSlice, [&](auto *opv) { |
| auto designateOp = mlir::dyn_cast<hlfir::DesignateOp>(opv); |
| return designateOp && designateOp.getComponent() && |
| designateOp.getComponent()->strref() == field; |
| }); |
| if (!referenced) |
| continue; |
| |
| int32_t fieldIdx = recordType.getFieldIndex(field); |
| builder.setInsertionPoint(op); |
| fir::IntOrValue idxConst = |
| mlir::IntegerAttr::get(builder.getI32Type(), fieldIdx); |
| auto fieldCoord = fir::CoordinateOp::create( |
| builder, op.getLoc(), builder.getRefType(memTy), op.getVarPtr(), |
| llvm::SmallVector<fir::IntOrValue, 1>{idxConst}); |
| int64_t fieldIdx64 = static_cast<int64_t>(fieldIdx); |
| llvm::SmallVector<int64_t, 1> idxPath{fieldIdx64}; |
| appendMemberMapIfNew(op, builder, op.getLoc(), fieldCoord, idxPath, |
| field, newMapOpsForFields, newMemberIndexPaths); |
| } |
| |
| // Handle nested allocatable fields along any component chain |
| // referenced in the region via HLFIR designates. |
| llvm::SmallVector<llvm::SmallVector<int64_t>> seenIndexPaths; |
| for (mlir::Operation *sliceOp : mapVarForwardSlice) { |
| auto designateOp = mlir::dyn_cast<hlfir::DesignateOp>(sliceOp); |
| if (!designateOp || !designateOp.getComponent()) |
| continue; |
| llvm::SmallVector<llvm::StringRef> compPathReversed; |
| compPathReversed.push_back(designateOp.getComponent()->strref()); |
| mlir::Value curBase = designateOp.getMemref(); |
| bool rootedAtMapArg = false; |
| while (true) { |
| if (auto parentDes = curBase.getDefiningOp<hlfir::DesignateOp>()) { |
| if (!parentDes.getComponent()) |
| break; |
| compPathReversed.push_back(parentDes.getComponent()->strref()); |
| curBase = parentDes.getMemref(); |
| continue; |
| } |
| if (auto decl = curBase.getDefiningOp<hlfir::DeclareOp>()) { |
| if (auto barg = |
| mlir::dyn_cast<mlir::BlockArgument>(decl.getMemref())) |
| rootedAtMapArg = (barg == opBlockArg); |
| } else if (auto blockArg = |
| mlir::dyn_cast_or_null<mlir::BlockArgument>( |
| curBase)) { |
| rootedAtMapArg = (blockArg == opBlockArg); |
| } |
| break; |
| } |
| // Only process nested paths (2+ components). Single-component paths |
| // for direct fields are handled above. |
| if (!rootedAtMapArg || compPathReversed.size() < 2) |
| continue; |
| builder.setInsertionPoint(op); |
| llvm::SmallVector<int64_t> indexPath; |
| mlir::Type curTy = underlyingType; |
| mlir::Value coordRef = op.getVarPtr(); |
| bool validPath = true; |
| for (llvm::StringRef compName : llvm::reverse(compPathReversed)) { |
| auto recTy = mlir::dyn_cast<fir::RecordType>(curTy); |
| if (!recTy) { |
| validPath = false; |
| break; |
| } |
| int32_t idx = recTy.getFieldIndex(compName); |
| if (idx < 0) { |
| validPath = false; |
| break; |
| } |
| indexPath.push_back(idx); |
| mlir::Type memTy = recTy.getType(idx); |
| fir::IntOrValue idxConst = |
| mlir::IntegerAttr::get(builder.getI32Type(), idx); |
| coordRef = fir::CoordinateOp::create( |
| builder, op.getLoc(), builder.getRefType(memTy), coordRef, |
| llvm::SmallVector<fir::IntOrValue, 1>{idxConst}); |
| curTy = memTy; |
| } |
| if (!validPath) |
| continue; |
| if (auto finalRefTy = |
| mlir::dyn_cast<fir::ReferenceType>(coordRef.getType())) { |
| mlir::Type eleTy = finalRefTy.getElementType(); |
| if (fir::isAllocatableType(eleTy)) { |
| if (!containsPath(seenIndexPaths, indexPath)) { |
| seenIndexPaths.emplace_back(indexPath.begin(), indexPath.end()); |
| appendMemberMapIfNew(op, builder, op.getLoc(), coordRef, |
| indexPath, compPathReversed.front(), |
| newMapOpsForFields, newMemberIndexPaths); |
| } |
| } |
| } |
| } |
| |
| if (newMapOpsForFields.empty()) |
| return mlir::WalkResult::advance(); |
| |
| // Deduplicate by index path to avoid emitting duplicate members for |
| // the same component. Use a set-based key to keep this near O(n). |
| llvm::SmallVector<mlir::Value> dedupMapOps; |
| llvm::SmallVector<llvm::SmallVector<int64_t>> dedupIndexPaths; |
| llvm::StringSet<> seenKeys; |
| for (auto [i, mapOp] : llvm::enumerate(newMapOpsForFields)) { |
| const auto &path = newMemberIndexPaths[i]; |
| llvm::SmallString<64> key; |
| buildPathKey(path, key); |
| if (seenKeys.contains(key)) |
| continue; |
| seenKeys.insert(key); |
| dedupMapOps.push_back(mapOp); |
| dedupIndexPaths.emplace_back(path.begin(), path.end()); |
| } |
| op.getMembersMutable().append(dedupMapOps); |
| llvm::SmallVector<llvm::SmallVector<int64_t>> newMemberIndices; |
| if (mlir::ArrayAttr oldAttr = op.getMembersIndexAttr()) |
| for (mlir::Attribute indexList : oldAttr) { |
| llvm::SmallVector<int64_t> listVec; |
| |
| for (mlir::Attribute index : mlir::cast<mlir::ArrayAttr>(indexList)) |
| listVec.push_back(mlir::cast<mlir::IntegerAttr>(index).getInt()); |
| |
| newMemberIndices.emplace_back(std::move(listVec)); |
| } |
| for (auto &path : dedupIndexPaths) |
| newMemberIndices.emplace_back(path); |
| |
| op.setMembersIndexAttr(builder.create2DI64ArrayAttr(newMemberIndices)); |
| // Set to partial map only if there is no user-defined mapper. |
| op.setPartialMap(op.getMapperIdAttr() == nullptr); |
| |
| return mlir::WalkResult::advance(); |
| }); |
| |
| // Expand type(C_PTR) only when unified_shared_memory is required, |
| // to ensure device-visible pointer size/behavior in USM scenarios |
| // without changing default expectations elsewhere. |
| func->walk([&](mlir::omp::MapInfoOp op) { |
| // Only expand C_PTR members when unified_shared_memory is required. |
| if (!moduleRequiresUSM(func->getParentOfType<mlir::ModuleOp>())) |
| return; |
| builder.setInsertionPoint(op); |
| genCptrMemberMap(op, builder); |
| }); |
| |
| func->walk([&](mlir::omp::MapInfoOp op) { |
| // TODO: Currently only supports a single user for the MapInfoOp. This |
| // is fine for the moment, as the Fortran frontend will generate a |
| // new MapInfoOp with at most one user currently. In the case of |
| // members of other objects, like derived types, the user would be the |
| // parent. In cases where it's a regular non-member map, the user would |
| // be the target operation it is being mapped by. |
| // |
| // However, when/if we optimise/cleanup the IR we will have to extend |
| // this pass to support multiple users, as we may wish to have a map |
| // be re-used by multiple users (e.g. across multiple targets that map |
| // the variable and have identical map properties). |
| assert(llvm::hasSingleElement(op->getUsers()) && |
| "OMPMapInfoFinalization currently only supports single users " |
| "of a MapInfoOp"); |
| |
| if (hasADescriptor(op.getVarPtr().getDefiningOp(), |
| fir::unwrapRefType(op.getVarType()))) { |
| builder.setInsertionPoint(op); |
| mlir::Operation *targetUser = getFirstTargetUser(op); |
| assert(targetUser && "expected user of map operation was not found"); |
| genDescriptorMemberMaps(op, builder, targetUser); |
| } |
| }); |
| |
| // Now that we've expanded all of our boxes into a descriptor and base |
| // address map where necessary, we check if the map owner is an |
| // enter/exit/target data directive, and if they are we drop the initial |
| // descriptor (top-level parent) and replace it with the |
| // base_address/data. |
| // |
| // This circumvents issues with stack allocated descriptors bound to |
| // device colliding which in Flang is rather trivial for a user to do by |
| // accident due to the rather pervasive local intermediate descriptor |
| // generation that occurs whenever you pass boxes around different scopes. |
| // In OpenMP 6+ mapping these would be a user error as the tools required |
| // to circumvent these issues are provided by the spec (ref_ptr/ptee map |
| // types), but in prior specifications these tools are not available and |
| // it becomes an implementation issue for us to solve. |
| // |
| // We do this by dropping the top-level descriptor which will be the stack |
| // descriptor when we perform enter/exit maps, as we don't want these to |
| // be bound until necessary which is when we utilise the descriptor type |
| // within a target region. At which point we map the relevant descriptor |
| // data and the runtime should correctly associate the data with the |
| // descriptor and bind together and allow clean mapping and execution. |
| for (auto *op : deferrableDesc) { |
| auto mapOp = llvm::dyn_cast<mlir::omp::MapInfoOp>(op); |
| mlir::Operation *targetUser = getFirstTargetUser(mapOp); |
| assert(targetUser && "expected user of map operation was not found"); |
| builder.setInsertionPoint(mapOp); |
| removeTopLevelDescriptor(mapOp, builder, targetUser); |
| addImplicitDescriptorMapToTargetDataOp(mapOp, builder, *targetUser); |
| } |
| |
| // Wait until after we have generated all of our maps to add them onto |
| // the target's block arguments, simplifying the process as there would be |
| // no need to avoid accidental duplicate additions. |
| func->walk([&](mlir::omp::MapInfoOp op) { |
| mlir::Operation *targetUser = getFirstTargetUser(op); |
| assert(targetUser && "expected user of map operation was not found"); |
| addImplicitMembersToTarget(op, builder, targetUser); |
| }); |
| }); |
| } |
| }; |
| |
| } // namespace |