| //===-- HexagonISelLowering.h - Hexagon DAG Lowering Interface --*- C++ -*-===// |
| // |
| // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. |
| // See https://llvm.org/LICENSE.txt for license information. |
| // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception |
| // |
| //===----------------------------------------------------------------------===// |
| // |
| // This file defines the interfaces that Hexagon uses to lower LLVM code into a |
| // selection DAG. |
| // |
| //===----------------------------------------------------------------------===// |
| |
| #ifndef LLVM_LIB_TARGET_HEXAGON_HEXAGONISELLOWERING_H |
| #define LLVM_LIB_TARGET_HEXAGON_HEXAGONISELLOWERING_H |
| |
| #include "Hexagon.h" |
| #include "MCTargetDesc/HexagonMCTargetDesc.h" |
| #include "llvm/ADT/StringRef.h" |
| #include "llvm/CodeGen/ISDOpcodes.h" |
| #include "llvm/CodeGen/SelectionDAGNodes.h" |
| #include "llvm/CodeGen/TargetLowering.h" |
| #include "llvm/CodeGen/ValueTypes.h" |
| #include "llvm/IR/CallingConv.h" |
| #include "llvm/IR/InlineAsm.h" |
| #include "llvm/Support/MachineValueType.h" |
| #include <cstdint> |
| #include <utility> |
| |
| namespace llvm { |
| |
| namespace HexagonISD { |
| |
| enum NodeType : unsigned { |
| OP_BEGIN = ISD::BUILTIN_OP_END, |
| |
| CONST32 = OP_BEGIN, |
| CONST32_GP, // For marking data present in GP. |
| ADDC, // Add with carry: (X, Y, Cin) -> (X+Y, Cout). |
| SUBC, // Sub with carry: (X, Y, Cin) -> (X+~Y+Cin, Cout). |
| ALLOCA, |
| |
| AT_GOT, // Index in GOT. |
| AT_PCREL, // Offset relative to PC. |
| |
| CALL, // Function call. |
| CALLnr, // Function call that does not return. |
| CALLR, |
| |
| RET_FLAG, // Return with a flag operand. |
| BARRIER, // Memory barrier. |
| JT, // Jump table. |
| CP, // Constant pool. |
| |
| COMBINE, |
| VASL, |
| VASR, |
| VLSR, |
| |
| TSTBIT, |
| INSERT, |
| EXTRACTU, |
| VEXTRACTW, |
| VINSERTW0, |
| VROR, |
| TC_RETURN, |
| EH_RETURN, |
| DCFETCH, |
| READCYCLE, |
| PTRUE, |
| PFALSE, |
| D2P, // Convert 8-byte value to 8-bit predicate register. [*] |
| P2D, // Convert 8-bit predicate register to 8-byte value. [*] |
| V2Q, // Convert HVX vector to a vector predicate reg. [*] |
| Q2V, // Convert vector predicate to an HVX vector. [*] |
| // [*] The equivalence is defined as "Q <=> (V != 0)", |
| // where the != operation compares bytes. |
| // Note: V != 0 is implemented as V >u 0. |
| QCAT, |
| QTRUE, |
| QFALSE, |
| TYPECAST, // No-op that's used to convert between different legal |
| // types in a register. |
| VALIGN, // Align two vectors (in Op0, Op1) to one that would have |
| // been loaded from address in Op2. |
| VALIGNADDR, // Align vector address: Op0 & -Op1, except when it is |
| // an address in a vector load, then it's a no-op. |
| VPACKL, // Pack low parts of the input vector to the front of the |
| // output. For example v64i16 VPACKL(v32i32) will pick |
| // the low halfwords and pack them into the first 32 |
| // halfwords of the output. The rest of the output is |
| // unspecified. |
| VUNPACK, // Unpacking into low elements with sign extension. |
| VUNPACKU, // Unpacking into low elements with zero extension. |
| ISEL, // Marker for nodes that were created during ISel, and |
| // which need explicit selection (would have been left |
| // unselected otherwise). |
| OP_END |
| }; |
| |
| } // end namespace HexagonISD |
| |
| class HexagonSubtarget; |
| |
| class HexagonTargetLowering : public TargetLowering { |
| int VarArgsFrameOffset; // Frame offset to start of varargs area. |
| const HexagonTargetMachine &HTM; |
| const HexagonSubtarget &Subtarget; |
| |
| bool CanReturnSmallStruct(const Function* CalleeFn, unsigned& RetSize) |
| const; |
| |
| public: |
| explicit HexagonTargetLowering(const TargetMachine &TM, |
| const HexagonSubtarget &ST); |
| |
| bool isHVXVectorType(MVT Ty) const; |
| |
| /// IsEligibleForTailCallOptimization - Check whether the call is eligible |
| /// for tail call optimization. Targets which want to do tail call |
| /// optimization should implement this function. |
| bool IsEligibleForTailCallOptimization(SDValue Callee, |
| CallingConv::ID CalleeCC, bool isVarArg, bool isCalleeStructRet, |
| bool isCallerStructRet, const SmallVectorImpl<ISD::OutputArg> &Outs, |
| const SmallVectorImpl<SDValue> &OutVals, |
| const SmallVectorImpl<ISD::InputArg> &Ins, SelectionDAG& DAG) const; |
| |
| bool getTgtMemIntrinsic(IntrinsicInfo &Info, const CallInst &I, |
| MachineFunction &MF, |
| unsigned Intrinsic) const override; |
| |
| bool isTruncateFree(Type *Ty1, Type *Ty2) const override; |
| bool isTruncateFree(EVT VT1, EVT VT2) const override; |
| |
| bool isCheapToSpeculateCttz() const override { return true; } |
| bool isCheapToSpeculateCtlz() const override { return true; } |
| bool isCtlzFast() const override { return true; } |
| |
| bool hasBitTest(SDValue X, SDValue Y) const override; |
| |
| bool allowTruncateForTailCall(Type *Ty1, Type *Ty2) const override; |
| |
| /// Return true if an FMA operation is faster than a pair of mul and add |
| /// instructions. fmuladd intrinsics will be expanded to FMAs when this |
| /// method returns true (and FMAs are legal), otherwise fmuladd is |
| /// expanded to mul + add. |
| bool isFMAFasterThanFMulAndFAdd(const MachineFunction &, |
| EVT) const override; |
| |
| // Should we expand the build vector with shuffles? |
| bool shouldExpandBuildVectorWithShuffles(EVT VT, |
| unsigned DefinedValues) const override; |
| |
| bool isShuffleMaskLegal(ArrayRef<int> Mask, EVT VT) const override; |
| TargetLoweringBase::LegalizeTypeAction getPreferredVectorAction(MVT VT) |
| const override; |
| |
| SDValue LowerOperation(SDValue Op, SelectionDAG &DAG) const override; |
| void LowerOperationWrapper(SDNode *N, SmallVectorImpl<SDValue> &Results, |
| SelectionDAG &DAG) const override; |
| void ReplaceNodeResults(SDNode *N, SmallVectorImpl<SDValue> &Results, |
| SelectionDAG &DAG) const override; |
| |
| const char *getTargetNodeName(unsigned Opcode) const override; |
| |
| SDValue LowerBUILD_VECTOR(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerCONCAT_VECTORS(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerEXTRACT_VECTOR_ELT(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerEXTRACT_SUBVECTOR(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerINSERT_VECTOR_ELT(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerINSERT_SUBVECTOR(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerVECTOR_SHUFFLE(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerVECTOR_SHIFT(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerROTL(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerBITCAST(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerANY_EXTEND(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerSIGN_EXTEND(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerZERO_EXTEND(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerLoad(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerStore(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerUnalignedLoad(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerUAddSubO(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerAddSubCarry(SDValue Op, SelectionDAG &DAG) const; |
| |
| SDValue LowerDYNAMIC_STACKALLOC(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerINLINEASM(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerPREFETCH(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerREADCYCLECOUNTER(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerEH_LABEL(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerEH_RETURN(SDValue Op, SelectionDAG &DAG) const; |
| SDValue |
| LowerFormalArguments(SDValue Chain, CallingConv::ID CallConv, bool isVarArg, |
| const SmallVectorImpl<ISD::InputArg> &Ins, |
| const SDLoc &dl, SelectionDAG &DAG, |
| SmallVectorImpl<SDValue> &InVals) const override; |
| SDValue LowerGLOBALADDRESS(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerBlockAddress(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerGlobalTLSAddress(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerToTLSGeneralDynamicModel(GlobalAddressSDNode *GA, |
| SelectionDAG &DAG) const; |
| SDValue LowerToTLSInitialExecModel(GlobalAddressSDNode *GA, |
| SelectionDAG &DAG) const; |
| SDValue LowerToTLSLocalExecModel(GlobalAddressSDNode *GA, |
| SelectionDAG &DAG) const; |
| SDValue GetDynamicTLSAddr(SelectionDAG &DAG, SDValue Chain, |
| GlobalAddressSDNode *GA, SDValue InFlag, EVT PtrVT, |
| unsigned ReturnReg, unsigned char OperandFlags) const; |
| SDValue LowerGLOBAL_OFFSET_TABLE(SDValue Op, SelectionDAG &DAG) const; |
| |
| SDValue LowerCall(TargetLowering::CallLoweringInfo &CLI, |
| SmallVectorImpl<SDValue> &InVals) const override; |
| SDValue LowerCallResult(SDValue Chain, SDValue InFlag, |
| CallingConv::ID CallConv, bool isVarArg, |
| const SmallVectorImpl<ISD::InputArg> &Ins, |
| const SDLoc &dl, SelectionDAG &DAG, |
| SmallVectorImpl<SDValue> &InVals, |
| const SmallVectorImpl<SDValue> &OutVals, |
| SDValue Callee) const; |
| |
| SDValue LowerSETCC(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerVSELECT(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerFRAMEADDR(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerATOMIC_FENCE(SDValue Op, SelectionDAG& DAG) const; |
| SDValue LowerRETURNADDR(SDValue Op, SelectionDAG &DAG) const; |
| |
| bool CanLowerReturn(CallingConv::ID CallConv, |
| MachineFunction &MF, bool isVarArg, |
| const SmallVectorImpl<ISD::OutputArg> &Outs, |
| LLVMContext &Context) const override; |
| |
| SDValue LowerReturn(SDValue Chain, CallingConv::ID CallConv, bool isVarArg, |
| const SmallVectorImpl<ISD::OutputArg> &Outs, |
| const SmallVectorImpl<SDValue> &OutVals, |
| const SDLoc &dl, SelectionDAG &DAG) const override; |
| |
| SDValue PerformDAGCombine(SDNode *N, DAGCombinerInfo &DCI) const override; |
| |
| bool mayBeEmittedAsTailCall(const CallInst *CI) const override; |
| |
| Register getRegisterByName(const char* RegName, LLT VT, |
| const MachineFunction &MF) const override; |
| |
| /// If a physical register, this returns the register that receives the |
| /// exception address on entry to an EH pad. |
| Register |
| getExceptionPointerRegister(const Constant *PersonalityFn) const override { |
| return Hexagon::R0; |
| } |
| |
| /// If a physical register, this returns the register that receives the |
| /// exception typeid on entry to a landing pad. |
| Register |
| getExceptionSelectorRegister(const Constant *PersonalityFn) const override { |
| return Hexagon::R1; |
| } |
| |
| SDValue LowerVASTART(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerVACOPY(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerConstantPool(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerJumpTable(SDValue Op, SelectionDAG &DAG) const; |
| |
| EVT getSetCCResultType(const DataLayout &, LLVMContext &C, |
| EVT VT) const override { |
| if (!VT.isVector()) |
| return MVT::i1; |
| else |
| return EVT::getVectorVT(C, MVT::i1, VT.getVectorNumElements()); |
| } |
| |
| bool getPostIndexedAddressParts(SDNode *N, SDNode *Op, |
| SDValue &Base, SDValue &Offset, |
| ISD::MemIndexedMode &AM, |
| SelectionDAG &DAG) const override; |
| |
| ConstraintType getConstraintType(StringRef Constraint) const override; |
| |
| std::pair<unsigned, const TargetRegisterClass *> |
| getRegForInlineAsmConstraint(const TargetRegisterInfo *TRI, |
| StringRef Constraint, MVT VT) const override; |
| |
| // Intrinsics |
| SDValue LowerINTRINSIC_WO_CHAIN(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerINTRINSIC_VOID(SDValue Op, SelectionDAG &DAG) const; |
| /// isLegalAddressingMode - Return true if the addressing mode represented |
| /// by AM is legal for this target, for a load/store of the specified type. |
| /// The type may be VoidTy, in which case only return true if the addressing |
| /// mode is legal for a load/store of any legal type. |
| /// TODO: Handle pre/postinc as well. |
| bool isLegalAddressingMode(const DataLayout &DL, const AddrMode &AM, |
| Type *Ty, unsigned AS, |
| Instruction *I = nullptr) const override; |
| /// Return true if folding a constant offset with the given GlobalAddress |
| /// is legal. It is frequently not legal in PIC relocation models. |
| bool isOffsetFoldingLegal(const GlobalAddressSDNode *GA) const override; |
| |
| bool isFPImmLegal(const APFloat &Imm, EVT VT, |
| bool ForCodeSize) const override; |
| |
| /// isLegalICmpImmediate - Return true if the specified immediate is legal |
| /// icmp immediate, that is the target has icmp instructions which can |
| /// compare a register against the immediate without having to materialize |
| /// the immediate into a register. |
| bool isLegalICmpImmediate(int64_t Imm) const override; |
| |
| EVT getOptimalMemOpType(const MemOp &Op, |
| const AttributeList &FuncAttributes) const override; |
| |
| bool allowsMemoryAccess(LLVMContext &Context, const DataLayout &DL, EVT VT, |
| unsigned AddrSpace, Align Alignment, |
| MachineMemOperand::Flags Flags, |
| bool *Fast) const override; |
| |
| bool allowsMisalignedMemoryAccesses(EVT VT, unsigned AddrSpace, |
| Align Alignment, |
| MachineMemOperand::Flags Flags, |
| bool *Fast) const override; |
| |
| /// Returns relocation base for the given PIC jumptable. |
| SDValue getPICJumpTableRelocBase(SDValue Table, SelectionDAG &DAG) |
| const override; |
| |
| bool shouldReduceLoadWidth(SDNode *Load, ISD::LoadExtType ExtTy, |
| EVT NewVT) const override; |
| |
| // Handling of atomic RMW instructions. |
| Value *emitLoadLinked(IRBuilderBase &Builder, Type *ValueTy, Value *Addr, |
| AtomicOrdering Ord) const override; |
| Value *emitStoreConditional(IRBuilderBase &Builder, Value *Val, Value *Addr, |
| AtomicOrdering Ord) const override; |
| AtomicExpansionKind shouldExpandAtomicLoadInIR(LoadInst *LI) const override; |
| bool shouldExpandAtomicStoreInIR(StoreInst *SI) const override; |
| AtomicExpansionKind |
| shouldExpandAtomicCmpXchgInIR(AtomicCmpXchgInst *AI) const override; |
| |
| AtomicExpansionKind |
| shouldExpandAtomicRMWInIR(AtomicRMWInst *AI) const override { |
| return AtomicExpansionKind::LLSC; |
| } |
| |
| private: |
| void initializeHVXLowering(); |
| unsigned getPreferredHvxVectorAction(MVT VecTy) const; |
| |
| bool validateConstPtrAlignment(SDValue Ptr, Align NeedAlign, const SDLoc &dl, |
| SelectionDAG &DAG) const; |
| SDValue replaceMemWithUndef(SDValue Op, SelectionDAG &DAG) const; |
| |
| std::pair<SDValue,int> getBaseAndOffset(SDValue Addr) const; |
| |
| bool getBuildVectorConstInts(ArrayRef<SDValue> Values, MVT VecTy, |
| SelectionDAG &DAG, |
| MutableArrayRef<ConstantInt*> Consts) const; |
| SDValue buildVector32(ArrayRef<SDValue> Elem, const SDLoc &dl, MVT VecTy, |
| SelectionDAG &DAG) const; |
| SDValue buildVector64(ArrayRef<SDValue> Elem, const SDLoc &dl, MVT VecTy, |
| SelectionDAG &DAG) const; |
| SDValue extractVector(SDValue VecV, SDValue IdxV, const SDLoc &dl, |
| MVT ValTy, MVT ResTy, SelectionDAG &DAG) const; |
| SDValue insertVector(SDValue VecV, SDValue ValV, SDValue IdxV, |
| const SDLoc &dl, MVT ValTy, SelectionDAG &DAG) const; |
| SDValue expandPredicate(SDValue Vec32, const SDLoc &dl, |
| SelectionDAG &DAG) const; |
| SDValue contractPredicate(SDValue Vec64, const SDLoc &dl, |
| SelectionDAG &DAG) const; |
| SDValue getVectorShiftByInt(SDValue Op, SelectionDAG &DAG) const; |
| SDValue appendUndef(SDValue Val, MVT ResTy, SelectionDAG &DAG) const; |
| |
| bool isUndef(SDValue Op) const { |
| if (Op.isMachineOpcode()) |
| return Op.getMachineOpcode() == TargetOpcode::IMPLICIT_DEF; |
| return Op.getOpcode() == ISD::UNDEF; |
| } |
| SDValue getInstr(unsigned MachineOpc, const SDLoc &dl, MVT Ty, |
| ArrayRef<SDValue> Ops, SelectionDAG &DAG) const { |
| SDNode *N = DAG.getMachineNode(MachineOpc, dl, Ty, Ops); |
| return SDValue(N, 0); |
| } |
| SDValue getZero(const SDLoc &dl, MVT Ty, SelectionDAG &DAG) const; |
| |
| using VectorPair = std::pair<SDValue, SDValue>; |
| using TypePair = std::pair<MVT, MVT>; |
| |
| SDValue getInt(unsigned IntId, MVT ResTy, ArrayRef<SDValue> Ops, |
| const SDLoc &dl, SelectionDAG &DAG) const; |
| |
| MVT ty(SDValue Op) const { |
| return Op.getValueType().getSimpleVT(); |
| } |
| TypePair ty(const VectorPair &Ops) const { |
| return { Ops.first.getValueType().getSimpleVT(), |
| Ops.second.getValueType().getSimpleVT() }; |
| } |
| MVT tyScalar(MVT Ty) const { |
| if (!Ty.isVector()) |
| return Ty; |
| return MVT::getIntegerVT(Ty.getSizeInBits()); |
| } |
| MVT tyVector(MVT Ty, MVT ElemTy) const { |
| if (Ty.isVector() && Ty.getVectorElementType() == ElemTy) |
| return Ty; |
| unsigned TyWidth = Ty.getSizeInBits(); |
| unsigned ElemWidth = ElemTy.getSizeInBits(); |
| assert((TyWidth % ElemWidth) == 0); |
| return MVT::getVectorVT(ElemTy, TyWidth/ElemWidth); |
| } |
| |
| MVT typeJoin(const TypePair &Tys) const; |
| TypePair typeSplit(MVT Ty) const; |
| MVT typeExtElem(MVT VecTy, unsigned Factor) const; |
| MVT typeTruncElem(MVT VecTy, unsigned Factor) const; |
| |
| SDValue opJoin(const VectorPair &Ops, const SDLoc &dl, |
| SelectionDAG &DAG) const; |
| VectorPair opSplit(SDValue Vec, const SDLoc &dl, SelectionDAG &DAG) const; |
| SDValue opCastElem(SDValue Vec, MVT ElemTy, SelectionDAG &DAG) const; |
| |
| bool allowsHvxMemoryAccess(MVT VecTy, MachineMemOperand::Flags Flags, |
| bool *Fast) const; |
| bool allowsHvxMisalignedMemoryAccesses(MVT VecTy, |
| MachineMemOperand::Flags Flags, |
| bool *Fast) const; |
| |
| bool isHvxSingleTy(MVT Ty) const; |
| bool isHvxPairTy(MVT Ty) const; |
| bool isHvxBoolTy(MVT Ty) const; |
| SDValue convertToByteIndex(SDValue ElemIdx, MVT ElemTy, |
| SelectionDAG &DAG) const; |
| SDValue getIndexInWord32(SDValue Idx, MVT ElemTy, SelectionDAG &DAG) const; |
| SDValue getByteShuffle(const SDLoc &dl, SDValue Op0, SDValue Op1, |
| ArrayRef<int> Mask, SelectionDAG &DAG) const; |
| |
| SDValue buildHvxVectorReg(ArrayRef<SDValue> Values, const SDLoc &dl, |
| MVT VecTy, SelectionDAG &DAG) const; |
| SDValue buildHvxVectorPred(ArrayRef<SDValue> Values, const SDLoc &dl, |
| MVT VecTy, SelectionDAG &DAG) const; |
| SDValue createHvxPrefixPred(SDValue PredV, const SDLoc &dl, |
| unsigned BitBytes, bool ZeroFill, |
| SelectionDAG &DAG) const; |
| SDValue extractHvxElementReg(SDValue VecV, SDValue IdxV, const SDLoc &dl, |
| MVT ResTy, SelectionDAG &DAG) const; |
| SDValue extractHvxElementPred(SDValue VecV, SDValue IdxV, const SDLoc &dl, |
| MVT ResTy, SelectionDAG &DAG) const; |
| SDValue insertHvxElementReg(SDValue VecV, SDValue IdxV, SDValue ValV, |
| const SDLoc &dl, SelectionDAG &DAG) const; |
| SDValue insertHvxElementPred(SDValue VecV, SDValue IdxV, SDValue ValV, |
| const SDLoc &dl, SelectionDAG &DAG) const; |
| SDValue extractHvxSubvectorReg(SDValue VecV, SDValue IdxV, const SDLoc &dl, |
| MVT ResTy, SelectionDAG &DAG) const; |
| SDValue extractHvxSubvectorPred(SDValue VecV, SDValue IdxV, const SDLoc &dl, |
| MVT ResTy, SelectionDAG &DAG) const; |
| SDValue insertHvxSubvectorReg(SDValue VecV, SDValue SubV, SDValue IdxV, |
| const SDLoc &dl, SelectionDAG &DAG) const; |
| SDValue insertHvxSubvectorPred(SDValue VecV, SDValue SubV, SDValue IdxV, |
| const SDLoc &dl, SelectionDAG &DAG) const; |
| SDValue extendHvxVectorPred(SDValue VecV, const SDLoc &dl, MVT ResTy, |
| bool ZeroExt, SelectionDAG &DAG) const; |
| SDValue compressHvxPred(SDValue VecQ, const SDLoc &dl, MVT ResTy, |
| SelectionDAG &DAG) const; |
| |
| SDValue LowerHvxBuildVector(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerHvxConcatVectors(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerHvxExtractElement(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerHvxInsertElement(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerHvxExtractSubvector(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerHvxInsertSubvector(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerHvxBitcast(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerHvxAnyExt(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerHvxSignExt(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerHvxZeroExt(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerHvxCttz(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerHvxMul(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerHvxMulh(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerHvxSetCC(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerHvxExtend(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerHvxSelect(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerHvxShift(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerHvxIntrinsic(SDValue Op, SelectionDAG &DAG) const; |
| SDValue LowerHvxMaskedOp(SDValue Op, SelectionDAG &DAG) const; |
| |
| SDValue SplitHvxPairOp(SDValue Op, SelectionDAG &DAG) const; |
| SDValue SplitHvxMemOp(SDValue Op, SelectionDAG &DAG) const; |
| SDValue WidenHvxLoad(SDValue Op, SelectionDAG &DAG) const; |
| SDValue WidenHvxStore(SDValue Op, SelectionDAG &DAG) const; |
| SDValue WidenHvxSetCC(SDValue Op, SelectionDAG &DAG) const; |
| SDValue WidenHvxExtend(SDValue Op, SelectionDAG &DAG) const; |
| SDValue WidenHvxTruncate(SDValue Op, SelectionDAG &DAG) const; |
| |
| std::pair<const TargetRegisterClass*, uint8_t> |
| findRepresentativeClass(const TargetRegisterInfo *TRI, MVT VT) |
| const override; |
| |
| bool shouldWidenToHvx(MVT Ty, SelectionDAG &DAG) const; |
| bool isHvxOperation(SDNode *N, SelectionDAG &DAG) const; |
| SDValue LowerHvxOperation(SDValue Op, SelectionDAG &DAG) const; |
| void LowerHvxOperationWrapper(SDNode *N, SmallVectorImpl<SDValue> &Results, |
| SelectionDAG &DAG) const; |
| void ReplaceHvxNodeResults(SDNode *N, SmallVectorImpl<SDValue> &Results, |
| SelectionDAG &DAG) const; |
| SDValue PerformHvxDAGCombine(SDNode *N, DAGCombinerInfo &DCI) const; |
| }; |
| |
| } // end namespace llvm |
| |
| #endif // LLVM_LIB_TARGET_HEXAGON_HEXAGONISELLOWERING_H |