| //===-- SIInstrInfo.td -----------------------------------------------------===// | 
 | // | 
 | // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. | 
 | // See https://llvm.org/LICENSE.txt for license information. | 
 | // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception | 
 | // | 
 | //===----------------------------------------------------------------------===// | 
 |  | 
 | def isWave32 : Predicate<"Subtarget->isWave32()">, | 
 |   AssemblerPredicate <(all_of FeatureWavefrontSize32)>; | 
 | def isWave64 : Predicate<"Subtarget->isWave64()">, | 
 |   AssemblerPredicate <(all_of FeatureWavefrontSize64)>; | 
 |  | 
 | class AMDGPUMnemonicAlias<string From, string To, string VariantName = ""> | 
 |     : MnemonicAlias<From, To, VariantName>, PredicateControl; | 
 |  | 
 | // Except for the NONE field, this must be kept in sync with the | 
 | // SIEncodingFamily enum in SIInstrInfo.cpp and the columns of the | 
 | // getMCOpcodeGen table. | 
 | def SIEncodingFamily { | 
 |   int NONE = -1; | 
 |   int SI = 0; | 
 |   int VI = 1; | 
 |   int SDWA = 2; | 
 |   int SDWA9 = 3; | 
 |   int GFX80 = 4; | 
 |   int GFX9 = 5; | 
 |   int GFX10 = 6; | 
 |   int SDWA10 = 7; | 
 |   int GFX90A = 8; | 
 |   int GFX940 = 9; | 
 |   int GFX11 = 10; | 
 |   int GFX12 = 11; | 
 |   int GFX1250 = 12; | 
 | } | 
 |  | 
 | //===----------------------------------------------------------------------===// | 
 | // Subtarget info | 
 | //===----------------------------------------------------------------------===// | 
 |  | 
 | class GFXGen<Predicate pred, string dn, string suffix, int sub> { | 
 |   Predicate AssemblerPredicate = pred; | 
 |   string DecoderNamespace = dn; | 
 |   string Suffix = suffix; | 
 |   int Subtarget = sub; | 
 | } | 
 |  | 
 | def GFX1250Gen       : GFXGen<isGFX125xOnly, "GFX1250", "_gfx1250", SIEncodingFamily.GFX1250>; | 
 | def GFX12Not12_50Gen : GFXGen<isGFX12Not12_50, "GFX12", "_gfx12", SIEncodingFamily.GFX12>; | 
 | def GFX12Gen : GFXGen<isGFX12Only, "GFX12", "_gfx12", SIEncodingFamily.GFX12>; | 
 | def GFX11Gen : GFXGen<isGFX11Only, "GFX11", "_gfx11", SIEncodingFamily.GFX11>; | 
 | def GFX10Gen : GFXGen<isGFX10Only, "GFX10", "_gfx10", SIEncodingFamily.GFX10>; | 
 |  | 
 | //===----------------------------------------------------------------------===// | 
 | // SI DAG Nodes | 
 | //===----------------------------------------------------------------------===// | 
 |  | 
 | def AMDGPUclamp : SDNode<"AMDGPUISD::CLAMP", SDTFPUnaryOp>; | 
 |  | 
 | def SDTSBufferLoad : SDTypeProfile<1, 3, | 
 |     [                    // vdata | 
 |      SDTCisVT<1, v4i32>, // rsrc | 
 |      SDTCisVT<2, i32>,   // offset(imm) | 
 |      SDTCisVT<3, i32>]>; // cachepolicy | 
 |  | 
 | def SIsbuffer_load : SDNode<"AMDGPUISD::SBUFFER_LOAD", SDTSBufferLoad, | 
 |                             [SDNPMayLoad, SDNPMemOperand]>; | 
 |  | 
 | def SIsbuffer_load_byte : SDNode<"AMDGPUISD::SBUFFER_LOAD_BYTE", SDTSBufferLoad, | 
 |                                  [SDNPMayLoad, SDNPMemOperand]>; | 
 |  | 
 | def SIsbuffer_load_ubyte | 
 |     : SDNode<"AMDGPUISD::SBUFFER_LOAD_UBYTE", SDTSBufferLoad, | 
 |              [SDNPMayLoad, SDNPMemOperand]>; | 
 |  | 
 | def SIsbuffer_load_short | 
 |     : SDNode<"AMDGPUISD::SBUFFER_LOAD_SHORT", SDTSBufferLoad, | 
 |              [SDNPMayLoad, SDNPMemOperand]>; | 
 |  | 
 | def SIsbuffer_load_ushort | 
 |     : SDNode<"AMDGPUISD::SBUFFER_LOAD_USHORT", SDTSBufferLoad, | 
 |              [SDNPMayLoad, SDNPMemOperand]>; | 
 |  | 
 | def SIds_ordered_count : SDNode<"AMDGPUISD::DS_ORDERED_COUNT", | 
 |   SDTypeProfile<1, 2, [SDTCisVT<0, i32>, SDTCisVT<1, i32>, SDTCisVT<2, i16>]>, | 
 |   [SDNPMayLoad, SDNPMayStore, SDNPMemOperand, SDNPHasChain, SDNPInGlue] | 
 | >; | 
 |  | 
 | def SDTAtomic2_f32 : SDTypeProfile<1, 2, [ | 
 |   SDTCisSameAs<0,2>, SDTCisFP<0>, SDTCisPtrTy<1> | 
 | ]>; | 
 |  | 
 | // load_d16_{lo|hi} ptr, tied_input | 
 | def SIload_d16 : SDTypeProfile<1, 2, [ | 
 |   SDTCisPtrTy<1>, | 
 |   SDTCisSameAs<0, 2> | 
 | ]>; | 
 |  | 
 |  | 
 | def SDTtbuffer_load : SDTypeProfile<1, 8, | 
 |   [                     // vdata | 
 |    SDTCisVT<1, v4i32>,  // rsrc | 
 |    SDTCisVT<2, i32>,    // vindex(VGPR) | 
 |    SDTCisVT<3, i32>,    // voffset(VGPR) | 
 |    SDTCisVT<4, i32>,    // soffset(SGPR) | 
 |    SDTCisVT<5, i32>,    // offset(imm) | 
 |    SDTCisVT<6, i32>,    // format(imm) | 
 |    SDTCisVT<7, i32>,    // cachepolicy, swizzled buffer(imm) | 
 |    SDTCisVT<8, i1>      // idxen(imm) | 
 |   ]>; | 
 |  | 
 | def SItbuffer_load :   SDNode<"AMDGPUISD::TBUFFER_LOAD_FORMAT", SDTtbuffer_load, | 
 |                               [SDNPMayLoad, SDNPMemOperand, SDNPHasChain]>; | 
 | def SItbuffer_load_d16 : SDNode<"AMDGPUISD::TBUFFER_LOAD_FORMAT_D16", | 
 |                                 SDTtbuffer_load, | 
 |                                 [SDNPMayLoad, SDNPMemOperand, SDNPHasChain]>; | 
 |  | 
 | def SDTtbuffer_store : SDTypeProfile<0, 9, | 
 |     [                     // vdata | 
 |      SDTCisVT<1, v4i32>,  // rsrc | 
 |      SDTCisVT<2, i32>,    // vindex(VGPR) | 
 |      SDTCisVT<3, i32>,    // voffset(VGPR) | 
 |      SDTCisVT<4, i32>,    // soffset(SGPR) | 
 |      SDTCisVT<5, i32>,    // offset(imm) | 
 |      SDTCisVT<6, i32>,    // format(imm) | 
 |      SDTCisVT<7, i32>,    // cachepolicy, swizzled buffer(imm) | 
 |      SDTCisVT<8, i1>      // idxen(imm) | 
 |     ]>; | 
 |  | 
 | def SItbuffer_store : SDNode<"AMDGPUISD::TBUFFER_STORE_FORMAT", SDTtbuffer_store, | 
 |                              [SDNPMayStore, SDNPMemOperand, SDNPHasChain]>; | 
 | def SItbuffer_store_d16 : SDNode<"AMDGPUISD::TBUFFER_STORE_FORMAT_D16", | 
 |                                 SDTtbuffer_store, | 
 |                                 [SDNPMayStore, SDNPMemOperand, SDNPHasChain]>; | 
 |  | 
 | def SDTBufferLoad : SDTypeProfile<1, 7, | 
 |     [                    // vdata | 
 |      SDTCisVT<1, v4i32>, // rsrc | 
 |      SDTCisVT<2, i32>,   // vindex(VGPR) | 
 |      SDTCisVT<3, i32>,   // voffset(VGPR) | 
 |      SDTCisVT<4, i32>,   // soffset(SGPR) | 
 |      SDTCisVT<5, i32>,   // offset(imm) | 
 |      SDTCisVT<6, i32>,   // cachepolicy, swizzled buffer(imm) | 
 |      SDTCisVT<7, i1>]>;  // idxen(imm) | 
 |  | 
 | def SIbuffer_load : SDNode <"AMDGPUISD::BUFFER_LOAD", SDTBufferLoad, | 
 |                             [SDNPMemOperand, SDNPHasChain, SDNPMayLoad]>; | 
 | def SIbuffer_load_ubyte : SDNode <"AMDGPUISD::BUFFER_LOAD_UBYTE", SDTBufferLoad, | 
 |                             [SDNPMemOperand, SDNPHasChain, SDNPMayLoad]>; | 
 | def SIbuffer_load_ushort : SDNode <"AMDGPUISD::BUFFER_LOAD_USHORT", SDTBufferLoad, | 
 |                             [SDNPMemOperand, SDNPHasChain, SDNPMayLoad]>; | 
 | def SIbuffer_load_byte : SDNode <"AMDGPUISD::BUFFER_LOAD_BYTE", SDTBufferLoad, | 
 |                             [SDNPMemOperand, SDNPHasChain, SDNPMayLoad]>; | 
 | def SIbuffer_load_short: SDNode <"AMDGPUISD::BUFFER_LOAD_SHORT", SDTBufferLoad, | 
 |                             [SDNPMemOperand, SDNPHasChain, SDNPMayLoad]>; | 
 | def SIbuffer_load_tfe : SDNode <"AMDGPUISD::BUFFER_LOAD_TFE", SDTBufferLoad, | 
 |                             [SDNPMemOperand, SDNPHasChain, SDNPMayLoad]>; | 
 | def SIbuffer_load_ubyte_tfe : SDNode <"AMDGPUISD::BUFFER_LOAD_UBYTE_TFE", SDTBufferLoad, | 
 |                             [SDNPMemOperand, SDNPHasChain, SDNPMayLoad]>; | 
 | def SIbuffer_load_ushort_tfe : SDNode <"AMDGPUISD::BUFFER_LOAD_USHORT_TFE", SDTBufferLoad, | 
 |                             [SDNPMemOperand, SDNPHasChain, SDNPMayLoad]>; | 
 | def SIbuffer_load_byte_tfe : SDNode <"AMDGPUISD::BUFFER_LOAD_BYTE_TFE", SDTBufferLoad, | 
 |                             [SDNPMemOperand, SDNPHasChain, SDNPMayLoad]>; | 
 | def SIbuffer_load_short_tfe: SDNode <"AMDGPUISD::BUFFER_LOAD_SHORT_TFE", SDTBufferLoad, | 
 |                             [SDNPMemOperand, SDNPHasChain, SDNPMayLoad]>; | 
 | def SIbuffer_load_format : SDNode <"AMDGPUISD::BUFFER_LOAD_FORMAT", SDTBufferLoad, | 
 |                             [SDNPMemOperand, SDNPHasChain, SDNPMayLoad]>; | 
 | def SIbuffer_load_format_tfe : SDNode <"AMDGPUISD::BUFFER_LOAD_FORMAT_TFE", SDTBufferLoad, | 
 |                                [SDNPMemOperand, SDNPHasChain, SDNPMayLoad]>; | 
 | def SIbuffer_load_format_d16 : SDNode <"AMDGPUISD::BUFFER_LOAD_FORMAT_D16", | 
 |                                 SDTBufferLoad, | 
 |                                 [SDNPMemOperand, SDNPHasChain, SDNPMayLoad]>; | 
 |  | 
 | def SDTBufferStore : SDTypeProfile<0, 8, | 
 |     [                    // vdata | 
 |      SDTCisVT<1, v4i32>, // rsrc | 
 |      SDTCisVT<2, i32>,   // vindex(VGPR) | 
 |      SDTCisVT<3, i32>,   // voffset(VGPR) | 
 |      SDTCisVT<4, i32>,   // soffset(SGPR) | 
 |      SDTCisVT<5, i32>,   // offset(imm) | 
 |      SDTCisVT<6, i32>,   // cachepolicy, swizzled buffer(imm) | 
 |      SDTCisVT<7, i1>]>;  // idxen(imm) | 
 |  | 
 | def SIbuffer_store : SDNode <"AMDGPUISD::BUFFER_STORE", SDTBufferStore, | 
 |                              [SDNPMayStore, SDNPMemOperand, SDNPHasChain]>; | 
 | def SIbuffer_store_byte: SDNode <"AMDGPUISD::BUFFER_STORE_BYTE", | 
 |                          SDTBufferStore, | 
 |                          [SDNPMayStore, SDNPMemOperand, SDNPHasChain]>; | 
 | def SIbuffer_store_short : SDNode <"AMDGPUISD::BUFFER_STORE_SHORT", | 
 |                            SDTBufferStore, | 
 |                            [SDNPMayStore, SDNPMemOperand, SDNPHasChain]>; | 
 | def SIbuffer_store_format : SDNode <"AMDGPUISD::BUFFER_STORE_FORMAT", | 
 |                             SDTBufferStore, | 
 |                             [SDNPMayStore, SDNPMemOperand, SDNPHasChain]>; | 
 | def SIbuffer_store_format_d16 : SDNode <"AMDGPUISD::BUFFER_STORE_FORMAT_D16", | 
 |                             SDTBufferStore, | 
 |                             [SDNPMayStore, SDNPMemOperand, SDNPHasChain]>; | 
 |  | 
 | multiclass SDBufferAtomic<string opcode> { | 
 |   def "" : SDNode <opcode, | 
 |     SDTypeProfile<1, 8, | 
 |          [SDTCisVT<2, v4i32>, // rsrc | 
 |          SDTCisVT<3, i32>,   // vindex(VGPR) | 
 |          SDTCisVT<4, i32>,   // voffset(VGPR) | 
 |          SDTCisVT<5, i32>,   // soffset(SGPR) | 
 |          SDTCisVT<6, i32>,   // offset(imm) | 
 |          SDTCisVT<7, i32>,   // cachepolicy(imm) | 
 |          SDTCisVT<8, i1>]>,  // idxen(imm) | 
 |     [SDNPMemOperand, SDNPHasChain, SDNPMayLoad, SDNPMayStore] | 
 |   >; | 
 |   def "_noret" : PatFrag< | 
 |     (ops node:$vdata_in, node:$rsrc, node:$vindex, node:$voffset, node:$soffset, | 
 |       node:$offset, node:$cachepolicy, node:$idxen), | 
 |     (!cast<SDNode>(NAME) node:$vdata_in, node:$rsrc, node:$vindex, | 
 |       node:$voffset, node:$soffset, node:$offset, node:$cachepolicy, | 
 |       node:$idxen)> { | 
 |     let HasNoUse = true; | 
 |   } | 
 | } | 
 |  | 
 | defm SIbuffer_atomic_swap : SDBufferAtomic <"AMDGPUISD::BUFFER_ATOMIC_SWAP">; | 
 | defm SIbuffer_atomic_add : SDBufferAtomic <"AMDGPUISD::BUFFER_ATOMIC_ADD">; | 
 | defm SIbuffer_atomic_sub : SDBufferAtomic <"AMDGPUISD::BUFFER_ATOMIC_SUB">; | 
 | defm SIbuffer_atomic_smin : SDBufferAtomic <"AMDGPUISD::BUFFER_ATOMIC_SMIN">; | 
 | defm SIbuffer_atomic_umin : SDBufferAtomic <"AMDGPUISD::BUFFER_ATOMIC_UMIN">; | 
 | defm SIbuffer_atomic_smax : SDBufferAtomic <"AMDGPUISD::BUFFER_ATOMIC_SMAX">; | 
 | defm SIbuffer_atomic_umax : SDBufferAtomic <"AMDGPUISD::BUFFER_ATOMIC_UMAX">; | 
 | defm SIbuffer_atomic_and : SDBufferAtomic <"AMDGPUISD::BUFFER_ATOMIC_AND">; | 
 | defm SIbuffer_atomic_or : SDBufferAtomic <"AMDGPUISD::BUFFER_ATOMIC_OR">; | 
 | defm SIbuffer_atomic_xor : SDBufferAtomic <"AMDGPUISD::BUFFER_ATOMIC_XOR">; | 
 | defm SIbuffer_atomic_inc : SDBufferAtomic <"AMDGPUISD::BUFFER_ATOMIC_INC">; | 
 | defm SIbuffer_atomic_dec : SDBufferAtomic <"AMDGPUISD::BUFFER_ATOMIC_DEC">; | 
 | defm SIbuffer_atomic_csub : SDBufferAtomic <"AMDGPUISD::BUFFER_ATOMIC_CSUB">; | 
 | defm SIbuffer_atomic_fadd : SDBufferAtomic <"AMDGPUISD::BUFFER_ATOMIC_FADD">; | 
 | defm SIbuffer_atomic_fmin : SDBufferAtomic <"AMDGPUISD::BUFFER_ATOMIC_FMIN">; | 
 | defm SIbuffer_atomic_fmax : SDBufferAtomic <"AMDGPUISD::BUFFER_ATOMIC_FMAX">; | 
 | defm SIbuffer_atomic_cond_sub_u32 : SDBufferAtomic <"AMDGPUISD::BUFFER_ATOMIC_COND_SUB_U32">; | 
 |  | 
 | def SIbuffer_atomic_cmpswap : SDNode <"AMDGPUISD::BUFFER_ATOMIC_CMPSWAP", | 
 |   SDTypeProfile<1, 9, | 
 |     [SDTCisVT<3, v4i32>, // rsrc | 
 |      SDTCisVT<4, i32>,   // vindex(VGPR) | 
 |      SDTCisVT<5, i32>,   // voffset(VGPR) | 
 |      SDTCisVT<6, i32>,   // soffset(SGPR) | 
 |      SDTCisVT<7, i32>,   // offset(imm) | 
 |      SDTCisVT<8, i32>,   // cachepolicy(imm) | 
 |      SDTCisVT<9, i1>]>,  // idxen(imm) | 
 |   [SDNPMemOperand, SDNPHasChain, SDNPMayLoad, SDNPMayStore] | 
 | >; | 
 |  | 
 | def SIbuffer_atomic_cmpswap_noret : PatFrag< | 
 |   (ops node:$src, node:$cmp, node:$rsrc, node:$vindex, node:$voffset, | 
 |     node:$soffset, node:$offset, node:$cachepolicy, node:$idxen), | 
 |   (SIbuffer_atomic_cmpswap node:$src, node:$cmp, node:$rsrc, node:$vindex, | 
 |     node:$voffset, node:$soffset, node:$offset, node:$cachepolicy, | 
 |     node:$idxen)> { | 
 |   let HasNoUse = true; | 
 | } | 
 |  | 
 | class SDGlobalAtomicNoRtn<string opcode, ValueType ty> : SDNode <opcode, | 
 |   SDTypeProfile<0, 2, | 
 |       [SDTCisPtrTy<0>,     // vaddr | 
 |        SDTCisVT<1, ty>]>,  // vdata | 
 |   [SDNPMemOperand, SDNPHasChain, SDNPMayLoad, SDNPMayStore] | 
 | >; | 
 |  | 
 | def SIpc_add_rel_offset : SDNode<"AMDGPUISD::PC_ADD_REL_OFFSET", | 
 |   SDTypeProfile<1, 2, [SDTCisVT<0, iPTR>, SDTCisSameAs<0,1>, SDTCisSameAs<0,2>]> | 
 | >; | 
 |  | 
 | def SIlds : SDNode<"AMDGPUISD::LDS", | 
 |   SDTypeProfile<1, 1, [SDTCisVT<0, iPTR>, SDTCisSameAs<0,1>]> | 
 | >; | 
 |  | 
 | def SIload_d16_lo : SDNode<"AMDGPUISD::LOAD_D16_LO", | 
 |   SIload_d16, | 
 |   [SDNPMayLoad, SDNPMemOperand, SDNPHasChain] | 
 | >; | 
 |  | 
 | def SIload_d16_lo_u8 : SDNode<"AMDGPUISD::LOAD_D16_LO_U8", | 
 |   SIload_d16, | 
 |   [SDNPMayLoad, SDNPMemOperand, SDNPHasChain] | 
 | >; | 
 |  | 
 | def SIload_d16_lo_i8 : SDNode<"AMDGPUISD::LOAD_D16_LO_I8", | 
 |   SIload_d16, | 
 |   [SDNPMayLoad, SDNPMemOperand, SDNPHasChain] | 
 | >; | 
 |  | 
 | def SIload_d16_hi : SDNode<"AMDGPUISD::LOAD_D16_HI", | 
 |   SIload_d16, | 
 |   [SDNPMayLoad, SDNPMemOperand, SDNPHasChain] | 
 | >; | 
 |  | 
 | def SIload_d16_hi_u8 : SDNode<"AMDGPUISD::LOAD_D16_HI_U8", | 
 |   SIload_d16, | 
 |   [SDNPMayLoad, SDNPMemOperand, SDNPHasChain] | 
 | >; | 
 |  | 
 | def SIload_d16_hi_i8 : SDNode<"AMDGPUISD::LOAD_D16_HI_I8", | 
 |   SIload_d16, | 
 |   [SDNPMayLoad, SDNPMemOperand, SDNPHasChain] | 
 | >; | 
 |  | 
 | def SIdenorm_mode : SDNode<"AMDGPUISD::DENORM_MODE", | 
 |   SDTypeProfile<0 ,1, [SDTCisInt<0>]>, | 
 |   [SDNPHasChain, SDNPOptInGlue, SDNPOutGlue] | 
 | >; | 
 |  | 
 |  | 
 | // Optimize v_mfma_scale* instructions to avoid the scale if the | 
 | // scales are known 0. | 
 | class UnscaledMFMAOptimizationPat<SDPatternOperator intrin> : PatFrag< | 
 |   (ops node:$srca, node:$srcb, node:$srcc, | 
 |        node:$cbsz, node:$blgp), | 
 |   (intrin $srca, $srcb, $srcc, $cbsz, $blgp, | 
 |           srcvalue, 0, srcvalue, 0) | 
 | >; | 
 |  | 
 | def mfma_f32_16x16x128_f8f6f4 : UnscaledMFMAOptimizationPat<int_amdgcn_mfma_scale_f32_16x16x128_f8f6f4>; | 
 | def mfma_f32_32x32x64_f8f6f4 : UnscaledMFMAOptimizationPat<int_amdgcn_mfma_scale_f32_32x32x64_f8f6f4>; | 
 |  | 
 | //===----------------------------------------------------------------------===// | 
 | // ValueType helpers | 
 | //===----------------------------------------------------------------------===// | 
 |  | 
 | class isIntType<ValueType SrcVT> { | 
 |   bit ret = !and(SrcVT.isInteger, !ne(SrcVT.Value, i1.Value)); | 
 | } | 
 |  | 
 | def SDTSBufferPrefetch : SDTypeProfile<0, 3, | 
 |     [SDTCisVT<0, v4i32>, // rsrc | 
 |      SDTCisVT<1, i32>,   // offset(imm) | 
 |      SDTCisVT<2, i32>]>; // length | 
 |  | 
 | def SIsbuffer_prefetch : SDNode<"AMDGPUISD::SBUFFER_PREFETCH_DATA", SDTSBufferPrefetch, | 
 |                                 [SDNPMayLoad, SDNPMayStore, SDNPMemOperand, SDNPHasChain]>; | 
 |  | 
 | //===----------------------------------------------------------------------===// | 
 | // SDNodes PatFrags for loads/stores with a glue input. | 
 | // This is for SDNodes and PatFrag for local loads and stores to | 
 | // enable s_mov_b32 m0, -1 to be glued to the memory instructions. | 
 | // | 
 | // These mirror the regular load/store PatFrags and rely on special | 
 | // processing during Select() to add the glued copy. | 
 | // | 
 | //===----------------------------------------------------------------------===// | 
 |  | 
 | def AMDGPUld_glue : SDNode <"ISD::LOAD", SDTLoad, | 
 |   [SDNPHasChain, SDNPMayLoad, SDNPMemOperand, SDNPInGlue] | 
 | >; | 
 |  | 
 | def AMDGPUatomic_ld_glue : SDNode <"ISD::ATOMIC_LOAD", SDTAtomicLoad, | 
 |   [SDNPHasChain, SDNPMayLoad, SDNPMemOperand, SDNPInGlue] | 
 | >; | 
 |  | 
 | def unindexedload_glue : PatFrag <(ops node:$ptr), (AMDGPUld_glue node:$ptr)> { | 
 |   let IsLoad = 1; | 
 |   let IsUnindexed = 1; | 
 | } | 
 |  | 
 | def load_glue : PatFrag <(ops node:$ptr), (unindexedload_glue node:$ptr)> { | 
 |   let IsLoad = 1; | 
 |   let IsNonExtLoad = 1; | 
 | } | 
 |  | 
 | def atomic_load_nonext_glue : | 
 |   PatFrag<(ops node:$ptr), (AMDGPUatomic_ld_glue node:$ptr)> { | 
 |   let IsAtomic = true; // FIXME: Should be IsLoad and/or IsAtomic? | 
 |   let IsNonExtLoad = true; | 
 | } | 
 |  | 
 | def atomic_load_zext_glue : | 
 |   PatFrag<(ops node:$ptr), (AMDGPUatomic_ld_glue node:$ptr)> { | 
 |   let IsAtomic = true; // FIXME: Should be IsLoad and/or IsAtomic? | 
 |   let IsZeroExtLoad = true; | 
 | } | 
 |  | 
 | def atomic_load_sext_glue : | 
 |   PatFrag<(ops node:$ptr), (AMDGPUatomic_ld_glue node:$ptr)> { | 
 |   let IsAtomic = true; // FIXME: Should be IsLoad and/or IsAtomic? | 
 |   let IsSignExtLoad = true; | 
 | } | 
 |  | 
 | def atomic_load_aext_glue : | 
 |   PatFrag<(ops node:$ptr), (AMDGPUatomic_ld_glue node:$ptr)> { | 
 |   let IsAtomic = true; // FIXME: Should be IsLoad and/or IsAtomic? | 
 |   let IsAnyExtLoad = true; | 
 | } | 
 |  | 
 | def atomic_load_nonext_16_glue : PatFrag<(ops node:$ptr), | 
 |   (atomic_load_nonext_glue node:$ptr)> { | 
 |   let IsAtomic = 1; | 
 |   let MemoryVT = i16; | 
 | } | 
 |  | 
 | def atomic_load_nonext_32_glue : PatFrag<(ops node:$ptr), | 
 |   (atomic_load_nonext_glue node:$ptr)> { | 
 |   let IsAtomic = 1; | 
 |   let MemoryVT = i32; | 
 | } | 
 |  | 
 | def atomic_load_nonext_64_glue : PatFrag<(ops node:$ptr), | 
 |   (atomic_load_nonext_glue node:$ptr)> { | 
 |   let IsAtomic = 1; | 
 |   let MemoryVT = i64; | 
 | } | 
 |  | 
 | def atomic_load_zext_8_glue : PatFrag<(ops node:$ptr), | 
 |   (atomic_load_zext_glue node:$ptr)> { | 
 |   let IsAtomic = 1; | 
 |   let MemoryVT = i8; | 
 | } | 
 |  | 
 | def atomic_load_sext_8_glue : PatFrag<(ops node:$ptr), | 
 |   (atomic_load_sext_glue node:$ptr)> { | 
 |   let IsAtomic = 1; | 
 |   let MemoryVT = i8; | 
 | } | 
 |  | 
 | def atomic_load_aext_8_glue : PatFrag<(ops node:$ptr), | 
 |   (atomic_load_aext_glue node:$ptr)> { | 
 |   let IsAtomic = 1; | 
 |   let MemoryVT = i8; | 
 | } | 
 |  | 
 | def atomic_load_zext_16_glue : PatFrag<(ops node:$ptr), | 
 |   (atomic_load_zext_glue node:$ptr)> { | 
 |   let IsAtomic = 1; | 
 |   let MemoryVT = i16; | 
 | } | 
 |  | 
 | def atomic_load_sext_16_glue : PatFrag<(ops node:$ptr), | 
 |   (atomic_load_sext_glue node:$ptr)> { | 
 |   let IsAtomic = 1; | 
 |   let MemoryVT = i16; | 
 | } | 
 |  | 
 | def atomic_load_aext_16_glue : PatFrag<(ops node:$ptr), | 
 |   (atomic_load_aext_glue node:$ptr)> { | 
 |   let IsAtomic = 1; | 
 |   let MemoryVT = i16; | 
 | } | 
 |  | 
 | def extload_glue : PatFrag<(ops node:$ptr), (unindexedload_glue node:$ptr)> { | 
 |   let IsLoad = 1; | 
 |   let IsAnyExtLoad = 1; | 
 | } | 
 |  | 
 | def sextload_glue : PatFrag<(ops node:$ptr), (unindexedload_glue node:$ptr)> { | 
 |   let IsLoad = 1; | 
 |   let IsSignExtLoad = 1; | 
 | } | 
 |  | 
 | def zextload_glue : PatFrag<(ops node:$ptr), (unindexedload_glue node:$ptr)> { | 
 |   let IsLoad = 1; | 
 |   let IsZeroExtLoad = 1; | 
 | } | 
 |  | 
 | def extloadi8_glue : PatFrag<(ops node:$ptr), (extload_glue node:$ptr)> { | 
 |   let IsLoad = 1; | 
 |   let MemoryVT = i8; | 
 | } | 
 |  | 
 | def zextloadi8_glue : PatFrag<(ops node:$ptr), (zextload_glue node:$ptr)> { | 
 |   let IsLoad = 1; | 
 |   let MemoryVT = i8; | 
 | } | 
 |  | 
 | def extloadi16_glue : PatFrag<(ops node:$ptr), (extload_glue node:$ptr)> { | 
 |   let IsLoad = 1; | 
 |   let MemoryVT = i16; | 
 | } | 
 |  | 
 | def zextloadi16_glue : PatFrag<(ops node:$ptr), (zextload_glue node:$ptr)> { | 
 |   let IsLoad = 1; | 
 |   let MemoryVT = i16; | 
 | } | 
 |  | 
 | def sextloadi8_glue : PatFrag<(ops node:$ptr), (sextload_glue node:$ptr)> { | 
 |   let IsLoad = 1; | 
 |   let MemoryVT = i8; | 
 | } | 
 |  | 
 | def sextloadi16_glue : PatFrag<(ops node:$ptr), (sextload_glue node:$ptr)> { | 
 |   let IsLoad = 1; | 
 |   let MemoryVT = i16; | 
 | } | 
 |  | 
 |  | 
 | let IsLoad = 1, AddressSpaces = LoadAddress_local.AddrSpaces in { | 
 | def load_local_m0 : PatFrag<(ops node:$ptr), (load_glue node:$ptr)> { | 
 |   let IsNonExtLoad = 1; | 
 | } | 
 |  | 
 | def extloadi8_local_m0 : PatFrag<(ops node:$ptr), (extloadi8_glue node:$ptr)>; | 
 | def sextloadi8_local_m0 : PatFrag<(ops node:$ptr), (sextloadi8_glue node:$ptr)>; | 
 | def zextloadi8_local_m0 : PatFrag<(ops node:$ptr), (zextloadi8_glue node:$ptr)>; | 
 |  | 
 | def extloadi16_local_m0 : PatFrag<(ops node:$ptr), (extloadi16_glue node:$ptr)>; | 
 | def sextloadi16_local_m0 : PatFrag<(ops node:$ptr), (sextloadi16_glue node:$ptr)>; | 
 | def zextloadi16_local_m0 : PatFrag<(ops node:$ptr), (zextloadi16_glue node:$ptr)>; | 
 | } // End IsLoad = 1, , AddressSpaces = LoadAddress_local.AddrSpaces | 
 |  | 
 | def load_align8_local_m0 : PatFrag<(ops node:$ptr), | 
 |                                    (load_local_m0 node:$ptr)> { | 
 |   let IsLoad = 1; | 
 |   int MinAlignment = 8; | 
 | } | 
 |  | 
 | def load_align16_local_m0 : PatFrag<(ops node:$ptr), | 
 |                                    (load_local_m0 node:$ptr)> { | 
 |   let IsLoad = 1; | 
 |   int MinAlignment = 16; | 
 | } | 
 |  | 
 | let IsAtomic = 1, AddressSpaces = LoadAddress_local.AddrSpaces in { | 
 | def atomic_load_nonext_16_local_m0 : PatFrag<(ops node:$ptr), | 
 |                                       (atomic_load_nonext_16_glue node:$ptr)>; | 
 | def atomic_load_nonext_32_local_m0 : PatFrag<(ops node:$ptr), | 
 |                                       (atomic_load_nonext_32_glue node:$ptr)>; | 
 | def atomic_load_nonext_64_local_m0 : PatFrag<(ops node:$ptr), | 
 |                                        (atomic_load_nonext_64_glue node:$ptr)>; | 
 |  | 
 | def atomic_load_zext_8_local_m0 : PatFrag<(ops node:$ptr), | 
 |                                       (atomic_load_zext_8_glue node:$ptr)>; | 
 | def atomic_load_sext_8_local_m0 : PatFrag<(ops node:$ptr), | 
 |                                       (atomic_load_sext_8_glue node:$ptr)>; | 
 | def atomic_load_aext_8_local_m0 : PatFrag<(ops node:$ptr), | 
 |                                       (atomic_load_aext_8_glue node:$ptr)>; | 
 | def atomic_load_zext_16_local_m0 : PatFrag<(ops node:$ptr), | 
 |                                       (atomic_load_zext_16_glue node:$ptr)>; | 
 | def atomic_load_sext_16_local_m0 : PatFrag<(ops node:$ptr), | 
 |                                       (atomic_load_sext_16_glue node:$ptr)>; | 
 | def atomic_load_aext_16_local_m0 : PatFrag<(ops node:$ptr), | 
 |                                       (atomic_load_aext_16_glue node:$ptr)>; | 
 | } // End let AddressSpaces = LoadAddress_local.AddrSpaces | 
 |  | 
 |  | 
 | def AMDGPUst_glue : SDNode <"ISD::STORE", SDTStore, | 
 |   [SDNPHasChain, SDNPMayStore, SDNPMemOperand, SDNPInGlue] | 
 | >; | 
 |  | 
 | def AMDGPUatomic_st_glue : SDNode <"ISD::ATOMIC_STORE", SDTAtomicStore, | 
 |   [SDNPHasChain, SDNPMayStore, SDNPMemOperand, SDNPInGlue] | 
 | >; | 
 |  | 
 | def unindexedstore_glue : PatFrag<(ops node:$val, node:$ptr), | 
 |                                    (AMDGPUst_glue node:$val, node:$ptr)> { | 
 |   let IsStore = 1; | 
 |   let IsUnindexed = 1; | 
 | } | 
 |  | 
 | def store_glue : PatFrag<(ops node:$val, node:$ptr), | 
 |                          (unindexedstore_glue node:$val, node:$ptr)> { | 
 |   let IsStore = 1; | 
 |   let IsTruncStore = 0; | 
 | } | 
 |  | 
 | def truncstore_glue : PatFrag<(ops node:$val, node:$ptr), | 
 |   (unindexedstore_glue node:$val, node:$ptr)> { | 
 |   let IsStore = 1; | 
 |   let IsTruncStore = 1; | 
 | } | 
 |  | 
 | def truncstorei8_glue : PatFrag<(ops node:$val, node:$ptr), | 
 |                            (truncstore_glue node:$val, node:$ptr)> { | 
 |   let IsStore = 1; | 
 |   let MemoryVT = i8; | 
 | } | 
 |  | 
 | def truncstorei16_glue : PatFrag<(ops node:$val, node:$ptr), | 
 |                            (truncstore_glue node:$val, node:$ptr)> { | 
 |   let IsStore = 1; | 
 |   let MemoryVT = i16; | 
 | } | 
 |  | 
 | let IsStore = 1, AddressSpaces = StoreAddress_local.AddrSpaces in { | 
 | def store_local_m0 : PatFrag<(ops node:$val, node:$ptr), | 
 |                              (store_glue node:$val, node:$ptr)>; | 
 | def truncstorei8_local_m0 : PatFrag<(ops node:$val, node:$ptr), | 
 |                                     (truncstorei8_glue node:$val, node:$ptr)>; | 
 | def truncstorei16_local_m0 : PatFrag<(ops node:$val, node:$ptr), | 
 |                                     (truncstorei16_glue node:$val, node:$ptr)>; | 
 | } | 
 |  | 
 | def store_align8_local_m0 : PatFrag <(ops node:$value, node:$ptr), | 
 |                                      (store_local_m0 node:$value, node:$ptr)>, | 
 |                             Aligned<8> { | 
 |   let IsStore = 1; | 
 | } | 
 |  | 
 | def store_align16_local_m0 : PatFrag <(ops node:$value, node:$ptr), | 
 |                                      (store_local_m0 node:$value, node:$ptr)>, | 
 |                             Aligned<16> { | 
 |   let IsStore = 1; | 
 | } | 
 |  | 
 | let PredicateCode = [{return cast<MemSDNode>(N)->getAlign() < 4;}], | 
 |     GISelPredicateCode = [{return (*MI.memoperands_begin())->getAlign() < 4;}], | 
 |     AddressSpaces = [ AddrSpaces.Local ] in { | 
 | def load_align_less_than_4_local : PatFrag<(ops node:$ptr), | 
 |                                            (load_local node:$ptr)> { | 
 |   let IsLoad = 1; | 
 |   let IsNonExtLoad = 1; | 
 | } | 
 |  | 
 | def load_align_less_than_4_local_m0 : PatFrag<(ops node:$ptr), | 
 |                                               (load_local_m0 node:$ptr)> { | 
 |   let IsLoad = 1; | 
 |   let IsNonExtLoad = 1; | 
 | } | 
 |  | 
 | def store_align_less_than_4_local : PatFrag <(ops node:$value, node:$ptr), | 
 |                                              (store_local node:$value, node:$ptr)> { | 
 |   let IsStore = 1; | 
 |   let IsTruncStore = 0; | 
 | } | 
 |  | 
 | def store_align_less_than_4_local_m0 : PatFrag <(ops node:$value, node:$ptr), | 
 |                                                 (store_local_m0 node:$value, node:$ptr)> { | 
 |   let IsStore = 1; | 
 |   let IsTruncStore = 0; | 
 | } | 
 | } | 
 |  | 
 | def atomic_store_8_glue : PatFrag < | 
 |   (ops node:$ptr, node:$value), | 
 |   (AMDGPUatomic_st_glue node:$ptr, node:$value)> { | 
 |   let IsAtomic = 1; | 
 |   let MemoryVT = i8; | 
 | } | 
 |  | 
 | def atomic_store_16_glue : PatFrag < | 
 |   (ops node:$ptr, node:$value), | 
 |   (AMDGPUatomic_st_glue node:$ptr, node:$value)> { | 
 |   let IsAtomic = 1; | 
 |   let MemoryVT = i16; | 
 | } | 
 |  | 
 | def atomic_store_32_glue : PatFrag < | 
 |   (ops node:$ptr, node:$value), | 
 |   (AMDGPUatomic_st_glue node:$ptr, node:$value)> { | 
 |   let IsAtomic = 1; | 
 |   let MemoryVT = i32; | 
 | } | 
 |  | 
 | def atomic_store_64_glue : PatFrag < | 
 |   (ops node:$ptr, node:$value), | 
 |   (AMDGPUatomic_st_glue node:$ptr, node:$value)> { | 
 |   let IsAtomic = 1; | 
 |   let MemoryVT = i64; | 
 | } | 
 |  | 
 | let IsAtomic = 1, AddressSpaces = StoreAddress_local.AddrSpaces in { | 
 | def atomic_store_8_local_m0 : PatFrag<(ops node:$val, node:$ptr), | 
 |                                        (atomic_store_8_glue node:$val, node:$ptr)>; | 
 | def atomic_store_16_local_m0 : PatFrag<(ops node:$val, node:$ptr), | 
 |                                        (atomic_store_16_glue node:$val, node:$ptr)>; | 
 | def atomic_store_32_local_m0 : PatFrag<(ops node:$val, node:$ptr), | 
 |                                        (atomic_store_32_glue node:$val, node:$ptr)>; | 
 | def atomic_store_64_local_m0 : PatFrag<(ops node:$val, node:$ptr), | 
 |                                        (atomic_store_64_glue node:$val, node:$ptr)>; | 
 | } // End let IsAtomic = 1, AddressSpaces = StoreAddress_local.AddrSpaces | 
 |  | 
 |  | 
 | //===----------------------------------------------------------------------===// | 
 | // SDNodes PatFrags for a16 loads and stores with 3 components. | 
 | // v3f16/v3i16 is widened to v4f16/v4i16, so we need to match on the memory | 
 | // load/store size. | 
 | //===----------------------------------------------------------------------===// | 
 |  | 
 | class mubuf_intrinsic_load<SDPatternOperator name, ValueType vt> : PatFrag < | 
 |   (ops node:$rsrc, node:$vindex, node:$voffset, node:$soffset, node:$offset, | 
 |             node:$auxiliary, node:$idxen), | 
 |   (name node:$rsrc, node:$vindex, node:$voffset, node:$soffset, node:$offset, | 
 |             node:$auxiliary, node:$idxen)> { | 
 |   let IsLoad = 1; | 
 |   let MemoryVT = vt; | 
 | } | 
 |  | 
 | class mubuf_intrinsic_store<SDPatternOperator name, ValueType vt> : PatFrag < | 
 |   (ops node:$vdata, node:$rsrc, node:$vindex, node:$voffset, node:$soffset, node:$offset, | 
 |             node:$auxiliary, node:$idxen), | 
 |   (name node:$vdata, node:$rsrc, node:$vindex, node:$voffset, node:$soffset, node:$offset, | 
 |             node:$auxiliary, node:$idxen)> { | 
 |   let IsStore = 1; | 
 |   let MemoryVT = vt; | 
 | } | 
 |  | 
 | class mtbuf_intrinsic_load<SDPatternOperator name, ValueType vt> : PatFrag < | 
 |   (ops node:$rsrc, node:$vindex, node:$voffset, node:$soffset, node:$offset, | 
 |             node:$format, node:$auxiliary, node:$idxen), | 
 |   (name node:$rsrc, node:$vindex, node:$voffset, node:$soffset, node:$offset, | 
 |             node:$format, node:$auxiliary, node:$idxen)> { | 
 |   let IsLoad = 1; | 
 |   let MemoryVT = vt; | 
 | } | 
 |  | 
 | class mtbuf_intrinsic_store<SDPatternOperator name, ValueType vt> : PatFrag < | 
 |   (ops node:$vdata, node:$rsrc, node:$vindex, node:$voffset, node:$soffset, node:$offset, | 
 |             node:$format, node:$auxiliary, node:$idxen), | 
 |   (name node:$vdata, node:$rsrc, node:$vindex, node:$voffset, node:$soffset, node:$offset, | 
 |             node:$format, node:$auxiliary, node:$idxen)> { | 
 |   let IsStore = 1; | 
 |   let MemoryVT = vt; | 
 | } | 
 |  | 
 | //===----------------------------------------------------------------------===// | 
 | // SDNodes PatFrags for d16 loads | 
 | //===----------------------------------------------------------------------===// | 
 |  | 
 | class LoadD16Frag <SDPatternOperator op> : PatFrag< | 
 |   (ops node:$ptr, node:$tied_in), | 
 |   (op node:$ptr, node:$tied_in)> { | 
 |   let IsLoad = 1; | 
 | } | 
 |  | 
 | foreach as = [ "global", "flat", "constant", "local", "private", "region" ] in { | 
 | let AddressSpaces = !cast<AddressSpaceList>("LoadAddress_"#as).AddrSpaces in { | 
 |  | 
 | def load_d16_hi_#as : LoadD16Frag <SIload_d16_hi>; | 
 |  | 
 | def az_extloadi8_d16_hi_#as : LoadD16Frag <SIload_d16_hi_u8> { | 
 |   let MemoryVT = i8; | 
 | } | 
 |  | 
 | def sextloadi8_d16_hi_#as : LoadD16Frag <SIload_d16_hi_i8> { | 
 |   let MemoryVT = i8; | 
 | } | 
 |  | 
 | def load_d16_lo_#as : LoadD16Frag <SIload_d16_lo>; | 
 |  | 
 | def az_extloadi8_d16_lo_#as : LoadD16Frag <SIload_d16_lo_u8> { | 
 |   let MemoryVT = i8; | 
 | } | 
 |  | 
 | def sextloadi8_d16_lo_#as : LoadD16Frag <SIload_d16_lo_i8> { | 
 |   let MemoryVT = i8; | 
 | } | 
 |  | 
 | } // End let AddressSpaces = ... | 
 | } // End foreach AddrSpace | 
 |  | 
 | def lshr_rev : PatFrag < | 
 |   (ops node:$src1, node:$src0), | 
 |   (srl $src0, $src1) | 
 | >; | 
 |  | 
 | def ashr_rev : PatFrag < | 
 |   (ops node:$src1, node:$src0), | 
 |   (sra $src0, $src1) | 
 | >; | 
 |  | 
 | def lshl_rev : PatFrag < | 
 |   (ops node:$src1, node:$src0), | 
 |   (shl $src0, $src1) | 
 | >; | 
 |  | 
 | def add_ctpop : PatFrag < | 
 |   (ops node:$src0, node:$src1), | 
 |   (add (ctpop $src0), $src1) | 
 | >; | 
 |  | 
 | def xnor : PatFrag < | 
 |   (ops node:$src0, node:$src1), | 
 |   (not (xor $src0, $src1)) | 
 | >; | 
 |  | 
 | foreach I = 1-4 in { | 
 | def shl#I#_add : PatFrag < | 
 |   (ops node:$src0, node:$src1), | 
 |   (add (shl_oneuse $src0, (i32 I)), $src1)> { | 
 |   // FIXME: Poor substitute for disabling pattern in SelectionDAG | 
 |   let PredicateCode = [{return false;}]; | 
 |   let GISelPredicateCode = [{return true;}]; | 
 | } | 
 | } | 
 |  | 
 | multiclass SIAtomicM0Glue2 <string op_name, bit is_amdgpu = 0, | 
 |                             SDTypeProfile tc = SDTAtomic2, | 
 |                             bit IsInt = 1> { | 
 |  | 
 |   def _glue : SDNode < | 
 |     !if(is_amdgpu, "AMDGPUISD", "ISD")#"::ATOMIC_"#op_name, tc, | 
 |     [SDNPHasChain, SDNPMayStore, SDNPMayLoad, SDNPMemOperand, SDNPInGlue] | 
 |   >; | 
 |  | 
 |   let AddressSpaces = StoreAddress_local.AddrSpaces in { | 
 |  | 
 |     if IsInt then { | 
 |       defm _local_m0 : binary_atomic_op <!cast<SDNode>(NAME#"_glue")>; | 
 |       defm _local_m0 : noret_binary_atomic_op <!cast<SDNode>(NAME#"_glue")>; | 
 |     } else { | 
 |       defm _local_m0 : binary_atomic_op_fp <!cast<SDNode>(NAME#"_glue")>; | 
 |       defm _local_m0 : noret_binary_atomic_op_fp <!cast<SDNode>(NAME#"_glue")>; | 
 |      } | 
 |   } | 
 |  | 
 |   let AddressSpaces = StoreAddress_region.AddrSpaces in { | 
 |     if IsInt then { | 
 |       defm _region_m0 : binary_atomic_op <!cast<SDNode>(NAME#"_glue")>; | 
 |       defm _region_m0 : noret_binary_atomic_op <!cast<SDNode>(NAME#"_glue")>; | 
 |     } else { | 
 |       defm _region_m0 : binary_atomic_op_fp <!cast<SDNode>(NAME#"_glue")>; | 
 |       defm _region_m0 : noret_binary_atomic_op_fp <!cast<SDNode>(NAME#"_glue")>; | 
 |     } | 
 |   } | 
 | } | 
 |  | 
 | defm atomic_load_add : SIAtomicM0Glue2 <"LOAD_ADD">; | 
 | defm atomic_load_sub : SIAtomicM0Glue2 <"LOAD_SUB">; | 
 | defm atomic_load_uinc_wrap : SIAtomicM0Glue2 <"LOAD_UINC_WRAP">; | 
 | defm atomic_load_udec_wrap : SIAtomicM0Glue2 <"LOAD_UDEC_WRAP">; | 
 | defm atomic_load_and : SIAtomicM0Glue2 <"LOAD_AND">; | 
 | defm atomic_load_min : SIAtomicM0Glue2 <"LOAD_MIN">; | 
 | defm atomic_load_max : SIAtomicM0Glue2 <"LOAD_MAX">; | 
 | defm atomic_load_or : SIAtomicM0Glue2 <"LOAD_OR">; | 
 | defm atomic_load_xor : SIAtomicM0Glue2 <"LOAD_XOR">; | 
 | defm atomic_load_umin : SIAtomicM0Glue2 <"LOAD_UMIN">; | 
 | defm atomic_load_umax : SIAtomicM0Glue2 <"LOAD_UMAX">; | 
 | defm atomic_swap : SIAtomicM0Glue2 <"SWAP">; | 
 | defm atomic_load_fadd : SIAtomicM0Glue2 <"LOAD_FADD", 0, SDTAtomic2_f32, 0>; | 
 | defm atomic_load_fmin : SIAtomicM0Glue2 <"LOAD_FMIN", 0, SDTAtomic2_f32, 0>; | 
 | defm atomic_load_fmax : SIAtomicM0Glue2 <"LOAD_FMAX", 0, SDTAtomic2_f32, 0>; | 
 |  | 
 | def as_i1timm : SDNodeXForm<timm, [{ | 
 |   return CurDAG->getTargetConstant(N->getZExtValue(), SDLoc(N), MVT::i1); | 
 | }]>; | 
 |  | 
 | def as_i1timm_zext : SDNodeXForm<timm, [{ | 
 |   return CurDAG->getTargetConstant(N->getZExtValue(), SDLoc(N), MVT::i32); | 
 | }]>; | 
 |  | 
 | def as_i8imm : SDNodeXForm<imm, [{ | 
 |   return CurDAG->getTargetConstant(N->getZExtValue(), SDLoc(N), MVT::i8); | 
 | }]>; | 
 |  | 
 | def as_i8timm : SDNodeXForm<timm, [{ | 
 |   return CurDAG->getSignedTargetConstant(N->getSExtValue(), SDLoc(N), MVT::i16); | 
 | }]>; | 
 |  | 
 | def as_i16imm : SDNodeXForm<imm, [{ | 
 |   return CurDAG->getSignedTargetConstant(N->getSExtValue(), SDLoc(N), MVT::i16); | 
 | }]>; | 
 |  | 
 | def as_i16timm : SDNodeXForm<timm, [{ | 
 |   // Explicit cast, as this is used with both signed and unsigned immediates. | 
 |   return CurDAG->getSignedTargetConstant(int16_t(N->getSExtValue()), SDLoc(N), | 
 |                                          MVT::i16); | 
 | }]>; | 
 |  | 
 | def as_i32imm: SDNodeXForm<imm, [{ | 
 |   return CurDAG->getSignedTargetConstant(N->getSExtValue(), SDLoc(N), MVT::i32); | 
 | }]>; | 
 |  | 
 | def as_i32timm: SDNodeXForm<timm, [{ | 
 |   return CurDAG->getSignedTargetConstant(N->getSExtValue(), SDLoc(N), MVT::i32); | 
 | }]>; | 
 |  | 
 | def as_i64imm: SDNodeXForm<imm, [{ | 
 |   return CurDAG->getSignedTargetConstant(N->getSExtValue(), SDLoc(N), MVT::i64); | 
 | }]>; | 
 |  | 
 | def cond_as_i32imm: SDNodeXForm<cond, [{ | 
 |   return CurDAG->getTargetConstant(N->get(), SDLoc(N), MVT::i32); | 
 | }]>; | 
 |  | 
 | // Copied from the AArch64 backend: | 
 | def bitcast_fpimm_to_i16 : SDNodeXForm<fpimm, [{ | 
 | return CurDAG->getTargetConstant( | 
 |   N->getValueAPF().bitcastToAPInt().getZExtValue(), SDLoc(N), MVT::i16); | 
 | }]>; | 
 |  | 
 | def bitcast_fpimm_to_i32 : SDNodeXForm<fpimm, [{ | 
 | return CurDAG->getTargetConstant( | 
 |   N->getValueAPF().bitcastToAPInt().getZExtValue(), SDLoc(N), MVT::i32); | 
 | }]>; | 
 |  | 
 | def frameindex_to_targetframeindex : SDNodeXForm<frameindex, [{ | 
 |   auto FI = cast<FrameIndexSDNode>(N); | 
 |   return CurDAG->getTargetFrameIndex(FI->getIndex(), MVT::i32); | 
 | }]>; | 
 |  | 
 | // Copied from the AArch64 backend: | 
 | def bitcast_fpimm_to_i64 : SDNodeXForm<fpimm, [{ | 
 | return CurDAG->getTargetConstant( | 
 |   N->getValueAPF().bitcastToAPInt().getZExtValue(), SDLoc(N), MVT::i64); | 
 | }]>; | 
 |  | 
 | def as_hw_round_mode : SDNodeXForm<timm, [{ | 
 |   // "round.towardzero" -> TowardZero 0        -> FP_ROUND_ROUND_TO_ZERO 3 | 
 |   // "round.tonearest"  -> NearestTiesToEven 1 -> FP_ROUND_ROUND_TO_NEAREST 0 | 
 |   // "round.upward"     -> TowardPositive 2    -> FP_ROUND_ROUND_TO_INF 1 | 
 |   // "round.downward    -> TowardNegative 3    -> FP_ROUND_ROUND_TO_NEGINF 2 | 
 |   return CurDAG->getTargetConstant((N->getSExtValue() + 3) % 4, SDLoc(N), | 
 |                                     MVT::i32); | 
 | }]>; | 
 |  | 
 | def SupportedRoundMode : TImmLeaf<i32, [{ | 
 |   return Imm == (int)RoundingMode::TowardZero || | 
 |          Imm == (int)RoundingMode::NearestTiesToEven || | 
 |          Imm == (int)RoundingMode::TowardPositive || | 
 |          Imm == (int)RoundingMode::TowardNegative; | 
 | }]>; | 
 |  | 
 | class bitextract_imm<int bitnum> : SDNodeXForm<imm, [{ | 
 |   uint64_t Imm = N->getZExtValue(); | 
 |   unsigned Bit = (Imm >> }] # bitnum # [{ ) & 1; | 
 |   return CurDAG->getTargetConstant(Bit, SDLoc(N), MVT::i1); | 
 | }]>; | 
 |  | 
 | def SIMM16bit : TImmLeaf <i32, | 
 |   [{return isInt<16>(Imm) || isUInt<16>(Imm);}], | 
 |   as_i16timm | 
 | >; | 
 |  | 
 | def i64imm_32bit : ImmLeaf<i64, [{ | 
 |   return (Imm & 0xffffffffULL) == static_cast<uint64_t>(Imm); | 
 | }]>; | 
 |  | 
 | def InlineImm64 : IntImmLeaf<i64, [{ | 
 |   return isInlineImmediate(Imm); | 
 | }]>; | 
 |  | 
 | def InlineImmFP32 : FPImmLeaf<f32, [{ | 
 |   return isInlineImmediate(Imm); | 
 | }]>; | 
 |  | 
 | def InlineImmFP64 : FPImmLeaf<f64, [{ | 
 |   return isInlineImmediate(Imm); | 
 | }]>; | 
 |  | 
 |  | 
 | class VGPRImm <dag frag> : PatLeaf<frag, [{ | 
 |   return isVGPRImm(N); | 
 | }]> { | 
 |   let GISelPredicateCode = [{return true;}]; | 
 | } | 
 |  | 
 | def NegateImm : SDNodeXForm<imm, [{ | 
 |   return CurDAG->getSignedConstant(-N->getSExtValue(), SDLoc(N), MVT::i32); | 
 | }]>; | 
 |  | 
 | // TODO: When FP inline imm values work? | 
 | def NegSubInlineConst32 : ImmLeaf<i32, [{ | 
 |   return Imm < -16 && Imm >= -64; | 
 | }], NegateImm>; | 
 |  | 
 | def NegSubInlineIntConst16 : ImmLeaf<i16, [{ | 
 |   return Imm < -16 && Imm >= -64; | 
 | }], NegateImm>; | 
 |  | 
 | def ShiftAmt32Imm : ImmLeaf <i32, [{ | 
 |   return Imm < 32; | 
 | }]>; | 
 |  | 
 | def fp16_zeros_high_16bits : PatLeaf<(f16 VGPR_32:$src), [{ | 
 |   return fp16SrcZerosHighBits(N->getOpcode()); | 
 | }]>; | 
 |  | 
 | def MFMALdScaleXForm : SDNodeXForm<timm, [{ | 
 |   unsigned Val = N->getZExtValue(); | 
 |   unsigned New = 0; | 
 |   if (Val & 0x1) | 
 |     New |= SISrcMods::OP_SEL_0; | 
 |   if (Val & 0x2) | 
 |     New |= SISrcMods::OP_SEL_1; | 
 |   return CurDAG->getTargetConstant(New, SDLoc(N), MVT::i32); | 
 | }]>; | 
 |  | 
 | def is_canonicalized : PatLeaf<(fAny srcvalue:$src), [{ | 
 |   const SITargetLowering &Lowering = | 
 |       *static_cast<const SITargetLowering *>(getTargetLowering()); | 
 |   return Lowering.isCanonicalized(*CurDAG, Op); | 
 | }]> { | 
 |   let GISelPredicateCode = [{ | 
 |     const SITargetLowering *TLI = static_cast<const SITargetLowering *>( | 
 |         MF.getSubtarget().getTargetLowering()); | 
 |     const MachineOperand &Dst = MI.getOperand(0); | 
 |     assert(Dst.isDef()); | 
 |     return TLI->isCanonicalized(Dst.getReg(), MF); | 
 |    }]; | 
 | } | 
 |  | 
 | //===----------------------------------------------------------------------===// | 
 | // MUBUF/SMEM Patterns | 
 | //===----------------------------------------------------------------------===// | 
 |  | 
 | def extract_cpol : SDNodeXForm<timm, [{ | 
 |   return CurDAG->getTargetConstant( | 
 |       N->getZExtValue() & (Subtarget->getGeneration() >= AMDGPUSubtarget::GFX12 | 
 |                                ? AMDGPU::CPol::ALL | 
 |                                : AMDGPU::CPol::ALL_pregfx12), | 
 |       SDLoc(N), MVT::i8); | 
 | }]>; | 
 |  | 
 | def extract_swz : SDNodeXForm<timm, [{ | 
 |   const bool Swizzle = | 
 |       N->getZExtValue() & (Subtarget->getGeneration() >= AMDGPUSubtarget::GFX12 | 
 |                                ? AMDGPU::CPol::SWZ | 
 |                                : AMDGPU::CPol::SWZ_pregfx12); | 
 |   return CurDAG->getTargetConstant(Swizzle, SDLoc(N), MVT::i8); | 
 | }]>; | 
 |  | 
 | def extract_cpol_set_glc : SDNodeXForm<timm, [{ | 
 |   const uint32_t cpol = N->getZExtValue() & (Subtarget->getGeneration() >= AMDGPUSubtarget::GFX12 | 
 |                                ? AMDGPU::CPol::ALL | 
 |                                : AMDGPU::CPol::ALL_pregfx12); | 
 |   return CurDAG->getTargetConstant(cpol | AMDGPU::CPol::GLC, SDLoc(N), MVT::i8); | 
 | }]>; | 
 |  | 
 | //===----------------------------------------------------------------------===// | 
 | // Custom Operands | 
 | //===----------------------------------------------------------------------===// | 
 |  | 
 | def SOPPBrTarget : CustomOperand<OtherVT> { | 
 |   let PrintMethod = "printOperand"; | 
 |   let EncoderMethod = "getSOPPBrEncoding"; | 
 |   let DecoderMethod = "decodeSOPPBrTarget"; | 
 |   let OperandType = "OPERAND_PCREL"; | 
 | } | 
 |  | 
 | def si_ga : Operand<iPTR>; | 
 |  | 
 | def InterpSlot : CustomOperand<i32>; | 
 |  | 
 | // It appears to be necessary to create a separate operand for this to | 
 | // be able to parse attr<num> with no space. | 
 | def InterpAttr : CustomOperand<i32>; | 
 |  | 
 | def InterpAttrChan : ImmOperand<i32>; | 
 |  | 
 | def SplitBarrier : ImmOperand<i32> { | 
 |   let OperandNamespace = "AMDGPU"; | 
 |   let OperandType = "OPERAND_INLINE_SPLIT_BARRIER_INT32"; | 
 |   let DecoderMethod = "decodeSplitBarrier"; | 
 |   let PrintMethod = "printOperand"; | 
 | } | 
 |  | 
 | def VReg32OrOffClass : AsmOperandClass { | 
 |   let Name = "VReg32OrOff"; | 
 |   let ParserMethod = "parseVReg32OrOff"; | 
 | } | 
 |  | 
 | def SendMsg : CustomOperand<i32>; | 
 |  | 
 | def Swizzle : CustomOperand<i16, 1>; | 
 |  | 
 | def Endpgm : CustomOperand<i16, 1>; | 
 |  | 
 | def SWaitCnt : CustomOperand<i32>; | 
 |  | 
 | def DepCtr : CustomOperand<i32>; | 
 |  | 
 | def SDelayALU : CustomOperand<i32>; | 
 |  | 
 | include "SIInstrFormats.td" | 
 | include "VIInstrFormats.td" | 
 |  | 
 | def BoolReg : AsmOperandClass { | 
 |   let Name = "BoolReg"; | 
 |   let ParserMethod = "parseBoolReg"; | 
 |   let RenderMethod = "addRegOperands"; | 
 | } | 
 |  | 
 | class BoolRC : RegisterOperand<SReg_1> { | 
 |   let ParserMatchClass = BoolReg; | 
 |   let DecoderMethod = "decodeBoolReg"; | 
 | } | 
 |  | 
 | def SSrc_i1 : RegisterOperand<SReg_1_XEXEC> { | 
 |   let ParserMatchClass = BoolReg; | 
 |   let DecoderMethod = "decodeBoolReg"; | 
 | } | 
 |  | 
 | def VOPDstS64orS32 : BoolRC { | 
 |   let PrintMethod = "printVOPDst"; | 
 | } | 
 |  | 
 | // SCSrc_i1 is the operand for pseudo instructions only. | 
 | // Boolean immediates shall not be exposed to codegen instructions. | 
 | def SCSrc_i1 : RegisterOperand<SReg_1_XEXEC> { | 
 |   let OperandNamespace = "AMDGPU"; | 
 |   let OperandType = "OPERAND_REG_IMM_INT32"; | 
 |   let ParserMatchClass = BoolReg; | 
 |   let DecoderMethod = "decodeBoolReg"; | 
 | } | 
 |  | 
 | // ===----------------------------------------------------------------------===// | 
 | // ExpSrc* Special cases for exp src operands which are printed as | 
 | // "off" depending on en operand. | 
 | // ===----------------------------------------------------------------------===// | 
 |  | 
 | def ExpSrc0 : RegisterOperand<VGPR_32> { | 
 |   let PrintMethod = "printExpSrc0"; | 
 |   let ParserMatchClass = VReg32OrOffClass; | 
 | } | 
 |  | 
 | def ExpSrc1 : RegisterOperand<VGPR_32> { | 
 |   let PrintMethod = "printExpSrc1"; | 
 |   let ParserMatchClass = VReg32OrOffClass; | 
 | } | 
 |  | 
 | def ExpSrc2 : RegisterOperand<VGPR_32> { | 
 |   let PrintMethod = "printExpSrc2"; | 
 |   let ParserMatchClass = VReg32OrOffClass; | 
 | } | 
 |  | 
 | def ExpSrc3 : RegisterOperand<VGPR_32> { | 
 |   let PrintMethod = "printExpSrc3"; | 
 |   let ParserMatchClass = VReg32OrOffClass; | 
 | } | 
 |  | 
 | class SDWASrc<ValueType vt> : RegisterOperand<VS_32> { | 
 |   let OperandNamespace = "AMDGPU"; | 
 |   string Type = !if(vt.isFP, "FP", "INT"); | 
 |   let OperandType = "OPERAND_REG_INLINE_C_"#Type#vt.Size; | 
 |   let DecoderMethod = "decodeSDWASrc"#vt.Size; | 
 |   let EncoderMethod = "getSDWASrcEncoding"; | 
 | } | 
 |  | 
 | def SDWASrc_i32 : SDWASrc<i32>; | 
 | def SDWASrc_i16 : SDWASrc<i16>; | 
 | def SDWASrc_f32 : SDWASrc<f32>; | 
 | def SDWASrc_f16 : SDWASrc<f16>; | 
 |  | 
 | def SDWAVopcDst : BoolRC { | 
 |   let OperandNamespace = "AMDGPU"; | 
 |   let OperandType = "OPERAND_SDWA_VOPC_DST"; | 
 |   let EncoderMethod = "getSDWAVopcDstEncoding"; | 
 |   let DecoderMethod = "decodeSDWAVopcDst"; | 
 |   let PrintMethod = "printVOPDst"; | 
 | } | 
 |  | 
 | class NamedIntOperand<string prefix, bit Optional = 1, string name = NAME> | 
 |     : CustomOperand<i32, Optional, name> { | 
 |   string Prefix = prefix; | 
 |  | 
 |   let PredicateMethod = | 
 |     "getPredicate([](const AMDGPUOperand &Op) -> bool { "# | 
 |     "return Op.isImmTy(AMDGPUOperand::"#ImmTy#"); })"; | 
 |  | 
 |   string Validator = "[](int64_t V) { return true; }"; | 
 |   string ConvertMethod = "[](int64_t &V) { return "#Validator#"(V); }"; | 
 |   let ParserMethod = | 
 |     "[this](OperandVector &Operands) -> ParseStatus { "# | 
 |     "return parseIntWithPrefix(\""#Prefix#"\", Operands, "# | 
 |     "AMDGPUOperand::"#ImmTy#", "#ConvertMethod#"); }"; | 
 |  | 
 |   bit PrintInHex = 0; | 
 |   bit AlwaysPrint = 0; | 
 |   let PrintMethod = "[this](const MCInst *MI, unsigned OpNo, " | 
 |                     "const MCSubtargetInfo &STI, raw_ostream &O) { " | 
 |                     "printNamedInt(MI, OpNo, STI, O, \""#Prefix#"\", "# | 
 |                     !if(PrintInHex, "true", "false")#", "# | 
 |                     !if(AlwaysPrint, "true", "false")#"); }"; | 
 | } | 
 |  | 
 | class NamedBitOperand<string Id, string Name = NAME> | 
 |     : CustomOperand<i1, 1, Name> { | 
 |   let PredicateMethod = "isImmTy<AMDGPUOperand::"#ImmTy#">"; | 
 |   let ParserMethod = | 
 |     "[this](OperandVector &Operands) -> ParseStatus { "# | 
 |     "return parseNamedBit(\""#Id#"\", Operands, AMDGPUOperand::"#ImmTy#"); }"; | 
 |   let PrintMethod = "[this](const MCInst *MI, unsigned OpNo, "# | 
 |     "const MCSubtargetInfo &STI, raw_ostream &O) { "# | 
 |     "printNamedBit(MI, OpNo, O, \""#Id#"\"); }"; | 
 | } | 
 |  | 
 | class DefaultOperand<CustomOperand Op, int Value> | 
 |   : OperandWithDefaultOps<Op.Type, (ops (Op.Type Value))>, | 
 |     CustomOperandProps<1> { | 
 |   let ParserMatchClass = Op.ParserMatchClass; | 
 |   let PrintMethod = Op.PrintMethod; | 
 | } | 
 |  | 
 | class SDWAOperand<string Id, string Name = NAME> | 
 |     : CustomOperand<i32, 1, Name> { | 
 |   let ParserMethod = | 
 |     "[this](OperandVector &Operands) -> ParseStatus { "# | 
 |     "return parseSDWASel(Operands, \""#Id#"\", AMDGPUOperand::"#ImmTy#"); }"; | 
 | } | 
 |  | 
 | class ArrayOperand0<string Id, string Name = NAME> | 
 |   : OperandWithDefaultOps<i32, (ops (i32 0))>, | 
 |     CustomOperandProps<1, Name> { | 
 |   let ParserMethod = | 
 |     "[this](OperandVector &Operands) -> ParseStatus { "# | 
 |     "return parseOperandArrayWithPrefix(\""#Id#"\", Operands, "# | 
 |     "AMDGPUOperand::"#ImmTy#"); }"; | 
 | } | 
 |  | 
 | let ImmTy = "ImmTyOffset" in | 
 | def flat_offset : CustomOperand<i32, 1, "FlatOffset">; | 
 | let PrintMethod = "printOffset" in | 
 | def Offset : NamedIntOperand<"offset">; | 
 | let Validator = "isUInt<8>" in { | 
 | def Offset0 : NamedIntOperand<"offset0">; | 
 | def Offset1 : NamedIntOperand<"offset1">; | 
 | } | 
 |  | 
 | def gds : NamedBitOperand<"gds", "GDS">; | 
 |  | 
 | def omod : CustomOperand<i32, 1, "OModSI">; | 
 | def omod0 : DefaultOperand<omod, 0>; | 
 |  | 
 | // We need to make the cases with a default of 0 distinct from no | 
 | // default to help deal with some cases where the operand appears | 
 | // before a mandatory operand. | 
 | def Clamp : NamedBitOperand<"clamp">; | 
 | def Clamp0 : DefaultOperand<Clamp, 0>; | 
 | def highmod : NamedBitOperand<"high", "High">; | 
 |  | 
 | def CPol : CustomOperand<i32, 1>; | 
 | def CPol_0 : DefaultOperand<CPol, 0>; | 
 | def CPol_GLC1 : DefaultOperand<CPol, 1>; | 
 | def CPol_GLC : ValuePredicatedOperand<CPol, "Op.getImm() & CPol::GLC">; | 
 | def CPol_NonGLC : ValuePredicatedOperand<CPol, "!(Op.getImm() & CPol::GLC)", 1>; | 
 | def CPol_GLC_WithDefault : DefaultOperand<CPol_GLC, !shl(1, CPolBit.GLC)>; | 
 | def CPol_NonGLC_WithDefault : DefaultOperand<CPol_NonGLC, 0>; | 
 |  | 
 | def TFE : NamedBitOperand<"tfe">; | 
 | def UNorm : NamedBitOperand<"unorm">; | 
 | def DA : NamedBitOperand<"da">; | 
 | def R128A16 : CustomOperand<i1, 1>; | 
 | def A16 : NamedBitOperand<"a16">; | 
 | def D16 : NamedBitOperand<"d16">; | 
 | def LWE : NamedBitOperand<"lwe">; | 
 | def exp_compr : NamedBitOperand<"compr", "ExpCompr">; | 
 | def exp_vm : NamedBitOperand<"vm", "ExpVM">; | 
 |  | 
 | def FORMAT : CustomOperand<i8>; | 
 |  | 
 | let PrintInHex = 1 in | 
 | def DMask : NamedIntOperand<"dmask">; | 
 |  | 
 | def Dim : CustomOperand<i8, /*optional=*/1>; | 
 |  | 
 | def dst_sel : SDWAOperand<"dst_sel", "SDWADstSel">; | 
 | def src0_sel : SDWAOperand<"src0_sel", "SDWASrc0Sel">; | 
 | def src1_sel : SDWAOperand<"src1_sel", "SDWASrc1Sel">; | 
 | def dst_unused : CustomOperand<i32, 1, "SDWADstUnused">; | 
 |  | 
 | def op_sel0 : ArrayOperand0<"op_sel", "OpSel">; | 
 | def op_sel_hi0 : ArrayOperand0<"op_sel_hi", "OpSelHi">; | 
 | def neg_lo0 : ArrayOperand0<"neg_lo", "NegLo">; | 
 | def neg_hi0 : ArrayOperand0<"neg_hi", "NegHi">; | 
 |  | 
 | def IndexKey16bit : CustomOperand<i32, 1>; | 
 | def IndexKey8bit : CustomOperand<i32, 1>; | 
 |  | 
 | def dpp8 : CustomOperand<i32, 0, "DPP8">; | 
 | def dpp_ctrl : CustomOperand<i32, 0, "DPPCtrl">; | 
 |  | 
 | let DefaultValue = "0xf", PrintInHex = 1, AlwaysPrint = 1 in { | 
 | def DppRowMask : NamedIntOperand<"row_mask">; | 
 | def DppBankMask : NamedIntOperand<"bank_mask">; | 
 | } | 
 | def DppBoundCtrl : NamedIntOperand<"bound_ctrl"> { | 
 |   let ConvertMethod = "[this] (int64_t &BC) -> bool { return convertDppBoundCtrl(BC); }"; | 
 |   let PrintMethod = "printDppBoundCtrl"; | 
 | } | 
 |  | 
 | let DecoderMethod = "decodeDpp8FI", PrintMethod = "printDppFI" in | 
 | def Dpp8FI : NamedIntOperand<"fi", 1, "DppFI">; | 
 | let PrintMethod = "printDppFI" in | 
 | def Dpp16FI : NamedIntOperand<"fi", 1, "DppFI">; | 
 |  | 
 | def blgp : CustomOperand<i32, 1, "BLGP">; | 
 | def CBSZ : NamedIntOperand<"cbsz"> { | 
 |   let Validator = "isUInt<3>"; | 
 | } | 
 | def ABID : NamedIntOperand<"abid"> { | 
 |   let Validator = "isUInt<4>"; | 
 | } | 
 | def hwreg : CustomOperand<i32, 0, "Hwreg">; | 
 |  | 
 | def exp_tgt : CustomOperand<i32, 0, "ExpTgt">; | 
 |  | 
 | let AlwaysPrint = 1 in { | 
 | def WaitVDST : NamedIntOperand<"wait_vdst"> { | 
 |   let Validator = "isUInt<4>"; | 
 | } | 
 | def WaitEXP : NamedIntOperand<"wait_exp"> { | 
 |   let Validator = "isUInt<3>"; | 
 | } | 
 | def WaitVAVDst : NamedIntOperand<"wait_va_vdst"> { | 
 |   let Validator = "isUInt<4>"; | 
 | } | 
 | def WaitVMVSrc : NamedIntOperand<"wait_vm_vsrc"> { | 
 |   let Validator = "isUInt<1>"; | 
 | } | 
 | } // End AlwaysPrint = 1 | 
 |  | 
 | def ByteSel : NamedIntOperand<"byte_sel"> { | 
 |   let Validator = "isUInt<2>"; | 
 | } | 
 | def ByteSel0 : DefaultOperand<ByteSel, 0>; | 
 |  | 
 | let PrintMethod = "printBitOp3" in | 
 | def BitOp3 : NamedIntOperand<"bitop3">; | 
 | def bitop3_0 : DefaultOperand<BitOp3, 0>; | 
 |  | 
 | class KImmFPOperand<ValueType vt> : ImmOperand<vt> { | 
 |   let OperandNamespace = "AMDGPU"; | 
 |   let OperandType = "OPERAND_KIMM"#vt.Size; | 
 |   let PrintMethod = "printU"#vt.Size#"ImmOperand"; | 
 |   let DecoderMethod = "decodeOperand_KImmFP"; | 
 | } | 
 |  | 
 | // 32-bit VALU immediate operand that uses the constant bus. | 
 | def KImmFP32 : KImmFPOperand<i32>; | 
 |  | 
 | // 32-bit VALU immediate operand with a 16-bit value that uses the | 
 | // constant bus. | 
 | def KImmFP16 : KImmFPOperand<i16>; | 
 |  | 
 | class FPInputModsMatchClass <int opSize> : AsmOperandClass { | 
 |   let Name = "RegOrImmWithFP"#opSize#"InputMods"; | 
 |   let ParserMethod = "parseRegOrImmWithFPInputMods"; | 
 |   let PredicateMethod = "isRegOrImmWithFP"#opSize#"InputMods"; | 
 | } | 
 |  | 
 | class FPVCSrcInputModsMatchClass <int opSize> : FPInputModsMatchClass <opSize> { | 
 |   let Name = "RegOrInlineImmWithFP"#opSize#"InputMods"; | 
 |   let PredicateMethod = "isRegOrInlineImmWithFP"#opSize#"InputMods"; | 
 | } | 
 |  | 
 | def FP16InputModsMatchClass : FPInputModsMatchClass<16>; | 
 | class FPT16InputModsMatchClass<bit IsFake16> : FPInputModsMatchClass<16> { | 
 |   let Name = !if(IsFake16, "RegOrImmWithFPFake16InputMods", | 
 |                  "RegOrImmWithFPT16InputMods"); | 
 |   let PredicateMethod = "isRegOrImmWithFPT16InputMods<" # | 
 |                         !if(IsFake16, "true", "false") # ">"; | 
 | } | 
 | def FP32InputModsMatchClass : FPInputModsMatchClass<32>; | 
 | def FP64InputModsMatchClass : FPInputModsMatchClass<64>; | 
 |  | 
 | class FP16VCSrcInputModsMatchClass<bit IsFake16> | 
 |     : FPVCSrcInputModsMatchClass<16> { | 
 |   let Name = !if(IsFake16, "RegOrInlineImmWithFPFake16InputMods", | 
 |                  "RegOrInlineImmWithFPT16InputMods"); | 
 |   let PredicateMethod = "isRegOrInlineImmWithFP16InputMods<" # | 
 |                         !if(IsFake16, "true", "false") # ">"; | 
 | } | 
 | def FP32VCSrcInputModsMatchClass : FPVCSrcInputModsMatchClass<32>; | 
 |  | 
 | class InputMods <AsmOperandClass matchClass> : Operand <i32> { | 
 |   let OperandNamespace = "AMDGPU"; | 
 |   let OperandType = "OPERAND_INPUT_MODS"; | 
 |   let ParserMatchClass = matchClass; | 
 | } | 
 |  | 
 | class FPInputMods <FPInputModsMatchClass matchClass> : InputMods <matchClass> { | 
 |   let PrintMethod = "printOperandAndFPInputMods"; | 
 | } | 
 |  | 
 | def FP16InputMods : FPInputMods<FP16InputModsMatchClass>; | 
 | class FPT16InputMods<bit IsFake16> : FPInputMods<FPT16InputModsMatchClass<IsFake16>> { | 
 |   let EncoderMethod = "getMachineOpValueT16"; | 
 | } | 
 | def FP32InputMods : FPInputMods<FP32InputModsMatchClass>; | 
 | def FP32T16DstInputMods : FPInputMods<FP32InputModsMatchClass> { | 
 |   let EncoderMethod = "getMachineOpValueT16"; | 
 | } | 
 | def FP64InputMods : FPInputMods<FP64InputModsMatchClass>; | 
 |  | 
 | class FPT16VCSrcInputMods<bit IsFake16 = 1> | 
 |   : FPInputMods<FP16VCSrcInputModsMatchClass<IsFake16>> { | 
 |   let EncoderMethod = "getMachineOpValueT16"; | 
 | } | 
 | def FP32VCSrcInputMods : FPInputMods<FP32VCSrcInputModsMatchClass>; | 
 |  | 
 | class IntInputModsMatchClass <int opSize> : AsmOperandClass { | 
 |   let Name = "RegOrImmWithInt"#opSize#"InputMods"; | 
 |   let ParserMethod = "parseRegOrImmWithIntInputMods"; | 
 |   let PredicateMethod = "isRegOrImmWithInt"#opSize#"InputMods"; | 
 | } | 
 | class IntVCSrcInputModsMatchClass <int opSize> : IntInputModsMatchClass <opSize> { | 
 |   let Name = "RegOrInlineImmWithInt"#opSize#"InputMods"; | 
 |   let PredicateMethod = "isRegOrInlineImmWithInt"#opSize#"InputMods"; | 
 | } | 
 | class IntT16InputModsMatchClass<bit IsFake16> : IntInputModsMatchClass<16> { | 
 |   let Name = !if(IsFake16, "RegOrImmWithIntFake16InputMods", | 
 |                  "RegOrImmWithIntT16InputMods"); | 
 |   let PredicateMethod = "isRegOrImmWithIntT16InputMods<" # | 
 |                         !if(IsFake16, "true", "false") # ">"; | 
 | } | 
 | def Int32InputModsMatchClass : IntInputModsMatchClass<32>; | 
 | def Int64InputModsMatchClass : IntInputModsMatchClass<64>; | 
 | def Int32VCSrcInputModsMatchClass : IntVCSrcInputModsMatchClass<32>; | 
 | class IntT16VCSrcInputModsMatchClass<bit IsFake16> : IntInputModsMatchClass<16> { | 
 |   let Name = !if(IsFake16, "RegOrInlineImmWithIntFake16InputMods", | 
 |                  "RegOrInlineImmWithIntT16InputMods"); | 
 |   let PredicateMethod = "isRegOrInlineImmWithIntT16InputMods<" # | 
 |                         !if(IsFake16, "true", "false") # ">"; | 
 | } | 
 |  | 
 | class IntInputMods <IntInputModsMatchClass matchClass> : InputMods <matchClass> { | 
 |   let PrintMethod = "printOperandAndIntInputMods"; | 
 | } | 
 | class IntT16InputMods<bit IsFake16> : IntInputMods<IntT16InputModsMatchClass<IsFake16>> { | 
 |   let EncoderMethod = "getMachineOpValueT16"; | 
 | } | 
 | def Int32InputMods : IntInputMods<Int32InputModsMatchClass>; | 
 | def Int32T16DstInputMods : IntInputMods<Int32InputModsMatchClass> { | 
 |   let EncoderMethod = "getMachineOpValueT16"; | 
 | } | 
 | def Int64InputMods : IntInputMods<Int64InputModsMatchClass>; | 
 | def Int32VCSrcInputMods : IntInputMods<Int32VCSrcInputModsMatchClass>; | 
 | class IntT16VCSrcInputMods<bit IsFake16 = 1> | 
 |     : IntInputMods<IntT16VCSrcInputModsMatchClass<IsFake16>> { | 
 |   let EncoderMethod = "getMachineOpValueT16"; | 
 | } | 
 |  | 
 | class OpSelModsMatchClass : AsmOperandClass { | 
 |   let Name = "OpSelMods"; | 
 |   let ParserMethod = "parseRegOrImm"; | 
 |   let PredicateMethod = "isRegOrImm"; | 
 | } | 
 |  | 
 | def IntOpSelModsMatchClass : OpSelModsMatchClass; | 
 | def IntOpSelMods : InputMods<IntOpSelModsMatchClass>; | 
 |  | 
 | class FPSDWAInputModsMatchClass <int opSize> : AsmOperandClass { | 
 |   let Name = "SDWAWithFP"#opSize#"InputMods"; | 
 |   let ParserMethod = "parseRegOrImmWithFPInputMods"; | 
 |   let PredicateMethod = "isSDWAFP"#opSize#"Operand"; | 
 | } | 
 |  | 
 | def FP16SDWAInputModsMatchClass : FPSDWAInputModsMatchClass<16>; | 
 | def FP32SDWAInputModsMatchClass : FPSDWAInputModsMatchClass<32>; | 
 |  | 
 | class FPSDWAInputMods <FPSDWAInputModsMatchClass matchClass> : | 
 |   InputMods <matchClass> { | 
 |   let PrintMethod = "printOperandAndFPInputMods"; | 
 | } | 
 |  | 
 | def FP16SDWAInputMods : FPSDWAInputMods<FP16SDWAInputModsMatchClass>; | 
 | def FP32SDWAInputMods : FPSDWAInputMods<FP32SDWAInputModsMatchClass>; | 
 |  | 
 | def FPVRegInputModsMatchClass : AsmOperandClass { | 
 |   let Name = "VRegWithFPInputMods"; | 
 |   let ParserMethod = "parseRegWithFPInputMods"; | 
 |   let PredicateMethod = "isVRegWithInputMods"; | 
 | } | 
 |  | 
 | def FPVRegInputMods : InputMods <FPVRegInputModsMatchClass> { | 
 |   let PrintMethod = "printOperandAndFPInputMods"; | 
 | } | 
 |  | 
 | def FPVRegT16DstInputMods : InputMods <FPVRegInputModsMatchClass> { | 
 |   let PrintMethod = "printOperandAndFPInputMods"; | 
 |   let EncoderMethod = "getMachineOpValueT16"; | 
 | } | 
 |  | 
 | class FPT16_Lo128VRegInputModsMatchClass<bit IsFake16> : AsmOperandClass { | 
 |   let Name = !if(IsFake16, "Fake16_Lo128VRegWithFPInputMods", | 
 |                  "T16_Lo128VRegWithFPInputMods"); | 
 |   let ParserMethod = "parseRegWithFPInputMods"; | 
 |   let PredicateMethod = "isT16_Lo128VRegWithInputMods<" # | 
 |                         !if(IsFake16, "true", "false") # ">"; | 
 | } | 
 |  | 
 | class FPT16VRegInputModsMatchClass<bit IsFake16> : AsmOperandClass { | 
 |   let Name = !if(IsFake16, "Fake16VRegWithFPInputMods", | 
 |                  "T16VRegWithFPInputMods"); | 
 |   let ParserMethod = "parseRegWithFPInputMods"; | 
 |   let PredicateMethod = "isT16VRegWithInputMods<" # | 
 |                         !if(IsFake16, "true", "false") # ">"; | 
 | } | 
 |  | 
 | class FPT16_Lo128VRegInputMods<bit IsFake16 = 1> | 
 |     : InputMods <FPT16_Lo128VRegInputModsMatchClass<IsFake16>> { | 
 |   let PrintMethod = "printOperandAndFPInputMods"; | 
 |   let EncoderMethod = "getMachineOpValueT16Lo128"; | 
 | } | 
 |  | 
 | class FPT16VRegInputMods<bit IsFake16 = 1> | 
 |     : InputMods <FPT16VRegInputModsMatchClass<IsFake16>> { | 
 |   let PrintMethod = "printOperandAndFPInputMods"; | 
 |   let EncoderMethod = "getMachineOpValueT16"; | 
 | } | 
 |  | 
 | class IntSDWAInputModsMatchClass <int opSize> : AsmOperandClass { | 
 |   let Name = "SDWAWithInt"#opSize#"InputMods"; | 
 |   let ParserMethod = "parseRegOrImmWithIntInputMods"; | 
 |   let PredicateMethod = "isSDWAInt"#opSize#"Operand"; | 
 | } | 
 |  | 
 | def Int16SDWAInputModsMatchClass : IntSDWAInputModsMatchClass<16>; | 
 | def Int32SDWAInputModsMatchClass : IntSDWAInputModsMatchClass<32>; | 
 | def Bin32SDWAInputModsMatchClass : IntSDWAInputModsMatchClass<32> { | 
 |   let Name = "SDWAWithBin32InputMods"; | 
 |   let ParserMethod = "parseRegOrImm"; | 
 | } | 
 |  | 
 | class IntSDWAInputMods <IntSDWAInputModsMatchClass matchClass> : | 
 |   InputMods <matchClass> { | 
 |   let PrintMethod = "printOperandAndIntInputMods"; | 
 | } | 
 |  | 
 | def Int16SDWAInputMods : IntSDWAInputMods<Int16SDWAInputModsMatchClass>; | 
 | def Int32SDWAInputMods : IntSDWAInputMods<Int32SDWAInputModsMatchClass>; | 
 | def Bin32SDWAInputMods : IntSDWAInputMods<Bin32SDWAInputModsMatchClass>; | 
 |  | 
 | def IntVRegInputModsMatchClass : AsmOperandClass { | 
 |   let Name = "VRegWithIntInputMods"; | 
 |   let ParserMethod = "parseRegWithIntInputMods"; | 
 |   let PredicateMethod = "isVRegWithInputMods"; | 
 | } | 
 |  | 
 | class IntT16_Lo128VRegInputModsMatchClass<bit IsFake16 = 1> : AsmOperandClass { | 
 |   let Name = !if(IsFake16, "Fake16_Lo128VRegWithIntInputMods", | 
 |                  "T16_Lo128VRegWithIntInputMods"); | 
 |   let ParserMethod = "parseRegWithIntInputMods"; | 
 |   let PredicateMethod = "isT16_Lo128VRegWithInputMods<" # | 
 |                         !if(IsFake16, "true", "false") # ">"; | 
 | } | 
 |  | 
 | class IntT16VRegInputModsMatchClass<bit IsFake16 = 1> : AsmOperandClass { | 
 |   let Name = !if(IsFake16, "Fake16VRegWithIntInputMods", | 
 |                  "T16VRegWithIntInputMods"); | 
 |   let ParserMethod = "parseRegWithIntInputMods"; | 
 |   let PredicateMethod = "isT16VRegWithInputMods<" # | 
 |                         !if(IsFake16, "true", "false") # ">"; | 
 | } | 
 |  | 
 | class IntT16_Lo128VRegInputMods<bit IsFake16 = 1> | 
 |     : InputMods <IntT16_Lo128VRegInputModsMatchClass<IsFake16>> { | 
 |   let PrintMethod = "printOperandAndIntInputMods"; | 
 |   let EncoderMethod = "getMachineOpValueT16Lo128"; | 
 | } | 
 |  | 
 | class IntT16VRegInputMods<bit IsFake16 = 1> | 
 |     : InputMods <IntT16VRegInputModsMatchClass<IsFake16>> { | 
 |   let PrintMethod = "printOperandAndIntInputMods"; | 
 |   let EncoderMethod = "getMachineOpValueT16"; | 
 | } | 
 |  | 
 | def IntVRegInputMods : InputMods <IntVRegInputModsMatchClass> { | 
 |   let PrintMethod = "printOperandAndIntInputMods"; | 
 | } | 
 |  | 
 | def IntVRegT16DstInputMods : InputMods <IntVRegInputModsMatchClass> { | 
 |   let PrintMethod = "printOperandAndIntInputMods"; | 
 |   let EncoderMethod = "getMachineOpValueT16"; | 
 | } | 
 |  | 
 | class PackedFPInputModsMatchClass <int opSize> : AsmOperandClass { | 
 |   let Name = "PackedFP"#opSize#"InputMods"; | 
 |   let ParserMethod = "parseRegOrImmWithFPInputMods"; | 
 |   let PredicateMethod = "isPackedFP"#opSize#"InputMods"; | 
 | } | 
 |  | 
 | class PackedVGPRFPInputModsMatchClass <int opSize> : PackedFPInputModsMatchClass<opSize> { | 
 |   let PredicateMethod = "isPackedVGPRFP"#opSize#"InputMods"; | 
 | } | 
 |  | 
 | class PackedIntInputModsMatchClass <int opSize> : AsmOperandClass { | 
 |   let Name = "PackedInt"#opSize#"InputMods"; | 
 |   let ParserMethod = "parseRegOrImm"; | 
 |   let PredicateMethod = "isRegOrImm"; | 
 | //  let PredicateMethod = "isPackedInt"#opSize#"InputMods"; | 
 | } | 
 |  | 
 | def PackedF16InputModsMatchClass : PackedFPInputModsMatchClass<16>; | 
 | def PackedI16InputModsMatchClass : PackedIntInputModsMatchClass<16>; | 
 | def PackedVGPRF32InputModsMatchClass : PackedVGPRFPInputModsMatchClass<32>; | 
 |  | 
 | class PackedFPInputMods <PackedFPInputModsMatchClass matchClass> : InputMods <matchClass> { | 
 |   let PrintMethod = "printOperandAndFPInputMods"; | 
 | } | 
 |  | 
 | class PackedIntInputMods <PackedIntInputModsMatchClass matchClass> : InputMods <matchClass> { | 
 |   //let PrintMethod = "printPackedIntInputMods"; | 
 | } | 
 |  | 
 | def PackedF16InputMods : PackedFPInputMods<PackedF16InputModsMatchClass>; | 
 | def PackedI16InputMods : PackedIntInputMods<PackedI16InputModsMatchClass>; | 
 | def PackedVGPRF32InputMods : PackedFPInputMods<PackedVGPRF32InputModsMatchClass>; | 
 |  | 
 | def MFMALdScaleModifierOp : TImmLeaf<i32, [{ | 
 |   return isUInt<2>(Imm); | 
 | }], MFMALdScaleXForm>; | 
 |  | 
 | //===----------------------------------------------------------------------===// | 
 | // Complex patterns | 
 | //===----------------------------------------------------------------------===// | 
 |  | 
 | def DS1Addr1Offset : ComplexPattern<iPTR, 2, "SelectDS1Addr1Offset">; | 
 | def DS64Bit4ByteAligned : ComplexPattern<iPTR, 3, "SelectDS64Bit4ByteAligned">; | 
 | def DS128Bit8ByteAligned : ComplexPattern<iPTR, 3, "SelectDS128Bit8ByteAligned">; | 
 |  | 
 | def MOVRELOffset : ComplexPattern<iPTR, 2, "SelectMOVRELOffset">; | 
 |  | 
 | def VOP3Mods0 : ComplexPattern<untyped, 4, "SelectVOP3Mods0">; | 
 |  | 
 | // Modifiers for floating point instructions. | 
 | def VOP3Mods  : ComplexPattern<untyped, 2, "SelectVOP3Mods">; | 
 |  | 
 | // VOP3 modifiers used for instructions that do not read canonicalized | 
 | // floating point values (i.e. integer operations with FP source | 
 | // modifiers) | 
 | def VOP3ModsNonCanonicalizing : ComplexPattern<untyped, 2, | 
 |   "SelectVOP3ModsNonCanonicalizing">; | 
 |  | 
 | def VOP3NoMods : ComplexPattern<untyped, 1, "SelectVOP3NoMods">; | 
 |  | 
 | def VOP3OMods : ComplexPattern<untyped, 3, "SelectVOP3OMods">; | 
 |  | 
 | def VOP3PMods  : ComplexPattern<untyped, 2, "SelectVOP3PMods">; | 
 |  | 
 | def VOP3PModsDOT  : ComplexPattern<untyped, 2, "SelectVOP3PModsDOT">; | 
 | def VOP3PModsNeg  : ComplexPattern<untyped, 1, "SelectVOP3PModsNeg">; | 
 | def WMMAOpSelVOP3PMods  : ComplexPattern<untyped, 1, "SelectWMMAOpSelVOP3PMods">; | 
 |  | 
 | def WMMAModsF32NegAbs  : ComplexPattern<untyped, 2, "SelectWMMAModsF32NegAbs">; | 
 | def WMMAModsF16Neg  : ComplexPattern<untyped, 2, "SelectWMMAModsF16Neg">; | 
 | def WMMAModsF16NegAbs  : ComplexPattern<untyped, 2, "SelectWMMAModsF16NegAbs">; | 
 | def WMMAVISrc  : ComplexPattern<untyped, 1, "SelectWMMAVISrc">; | 
 | def SWMMACIndex8  : ComplexPattern<untyped, 2, "SelectSWMMACIndex8">; | 
 | def SWMMACIndex16  : ComplexPattern<untyped, 2, "SelectSWMMACIndex16">; | 
 |  | 
 | def VOP3OpSel  : ComplexPattern<untyped, 2, "SelectVOP3OpSel">; | 
 |  | 
 | def VOP3OpSelMods  : ComplexPattern<untyped, 2, "SelectVOP3OpSelMods">; | 
 |  | 
 | def VOP3PMadMixModsExt : ComplexPattern<untyped, 2, "SelectVOP3PMadMixModsExt">; | 
 | def VOP3PMadMixMods : ComplexPattern<untyped, 2, "SelectVOP3PMadMixMods">; | 
 |  | 
 | def VINTERPMods  : ComplexPattern<untyped, 2, "SelectVINTERPMods">; | 
 | def VINTERPModsHi  : ComplexPattern<untyped, 2, "SelectVINTERPModsHi">; | 
 |  | 
 | //===----------------------------------------------------------------------===// | 
 | // SI assembler operands | 
 | //===----------------------------------------------------------------------===// | 
 |  | 
 | def SIOperand { | 
 |   int ZERO = 0x80; | 
 |   int VCC = 0x6A; | 
 |   int FLAT_SCR = 0x68; | 
 | } | 
 |  | 
 | // This should be kept in sync with SISrcMods enum | 
 | def SRCMODS { | 
 |   int NONE = 0; | 
 |   int NEG = 1; | 
 |   int ABS = 2; | 
 |   int NEG_ABS = 3; | 
 |  | 
 |   int NEG_HI = ABS; | 
 |   int OP_SEL_0 = 4; | 
 |   int OP_SEL_1 = 8; | 
 |   int DST_OP_SEL = 8; | 
 | } | 
 |  | 
 | def DSTCLAMP { | 
 |   int NONE = 0; | 
 |   int ENABLE = 1; | 
 | } | 
 |  | 
 | def DSTOMOD { | 
 |   int NONE = 0; | 
 | } | 
 |  | 
 | def HWREG { | 
 |   int MODE = 1; | 
 |   int STATUS = 2; | 
 |   int TRAPSTS = 3; | 
 |   int HW_ID = 4; | 
 |   int GPR_ALLOC = 5; | 
 |   int LDS_ALLOC = 6; | 
 |   int IB_STS = 7; | 
 |   int MEM_BASES = 15; | 
 |   int TBA_LO = 16; | 
 |   int TBA_HI = 17; | 
 |   int TMA_LO = 18; | 
 |   int TMA_HI = 19; | 
 |   int FLAT_SCR_LO = 20; | 
 |   int FLAT_SCR_HI = 21; | 
 |   int XNACK_MASK = 22; | 
 |   int POPS_PACKER = 25; | 
 |   int SHADER_CYCLES = 29; | 
 | } | 
 |  | 
 | class getHwRegImm<int Reg, int Offset = 0, int Size = 32> { | 
 |   int ret = !and(!or(Reg, | 
 |                      !shl(Offset, 6), | 
 |                      !shl(!add(Size, -1), 11)), 65535); | 
 | } | 
 |  | 
 | //===----------------------------------------------------------------------===// | 
 | // | 
 | // SI Instruction multiclass helpers. | 
 | // | 
 | // Instructions with _32 take 32-bit operands. | 
 | // Instructions with _64 take 64-bit operands. | 
 | // | 
 | // VOP_* instructions can use either a 32-bit or 64-bit encoding.  The 32-bit | 
 | // encoding is the standard encoding, but instruction that make use of | 
 | // any of the instruction modifiers must use the 64-bit encoding. | 
 | // | 
 | // Instructions with _e32 use the 32-bit encoding. | 
 | // Instructions with _e64 use the 64-bit encoding. | 
 | // | 
 | //===----------------------------------------------------------------------===// | 
 |  | 
 | class SIMCInstr <string pseudo, int subtarget> { | 
 |   string PseudoInstr = pseudo; | 
 |   int Subtarget = subtarget; | 
 | } | 
 |  | 
 | //===----------------------------------------------------------------------===// | 
 | // Vector ALU classes | 
 | //===----------------------------------------------------------------------===// | 
 |  | 
 | class getNumSrcArgs<ValueType Src0, ValueType Src1, ValueType Src2> { | 
 |   int ret = | 
 |     !if (!eq(Src0.Value, untyped.Value),      0, | 
 |       !if (!eq(Src1.Value, untyped.Value),    1,   // VOP1 | 
 |          !if (!eq(Src2.Value, untyped.Value), 2,   // VOP2 | 
 |                                               3))); // VOP3 | 
 | } | 
 |  | 
 | // Returns the register class to use for the destination of VOP[123C] | 
 | // instructions for the given VT. | 
 | class getVALUDstForVT<ValueType VT, bit IsTrue16 = 0, bit IsVOP3Encoding = 0> { | 
 |   defvar op16 = !if(IsTrue16, !if (IsVOP3Encoding, VOPDstOperand_t16, | 
 |                                    VOPDstOperand_t16Lo128), | 
 |                     VOPDstOperand<VGPR_32>); | 
 |   RegisterOperand ret = !cond(!eq(VT.Size, 1024) : VOPDstOperand<VReg_1024>, | 
 |                               !eq(VT.Size, 512) : VOPDstOperand<VReg_512>, | 
 |                               !eq(VT.Size, 256) : VOPDstOperand<VReg_256>, | 
 |                               !eq(VT.Size, 192) : VOPDstOperand<VReg_192>, | 
 |                               !eq(VT.Size, 128) : VOPDstOperand<VReg_128>, | 
 |                               !eq(VT.Size, 64)  : VOPDstOperand<VReg_64>, | 
 |                               !eq(VT.Size, 32)  : VOPDstOperand<VGPR_32>, | 
 |                               !eq(VT.Size, 16)  : op16, | 
 |                               1                 : VOPDstS64orS32); // else VT == i1 | 
 | } | 
 |  | 
 | class getVALUDstForVT_fake16<ValueType VT> { | 
 |   RegisterOperand ret = !if(!eq(VT.Size, 32), VOPDstOperand<VGPR_32>, | 
 |                           !if(!eq(VT.Size, 128), VOPDstOperand<VReg_128>, | 
 |                             !if(!eq(VT.Size, 64), VOPDstOperand<VReg_64>, | 
 |                               !if(!eq(VT.Size, 16), VOPDstOperand<VGPR_32_Lo128>, | 
 |                               VOPDstS64orS32)))); // else VT == i1 | 
 | } | 
 |  | 
 | // Returns the register class to use for the destination of VOP[12C] | 
 | // instructions with SDWA extension | 
 | class getSDWADstForVT<ValueType VT> { | 
 |   RegisterOperand ret = !if(!eq(VT.Size, 1), | 
 |                             SDWAVopcDst, // VOPC | 
 |                             VOPDstOperand<VGPR_32>); // VOP1/2 32-bit dst | 
 | } | 
 |  | 
 | // Returns the register class to use for source 0 of VOP[12C] | 
 | // instructions for the given VT. | 
 | class getVOPSrc0ForVT<ValueType VT, bit IsTrue16, bit IsFake16 = 1> { | 
 |   RegisterOperand ret = | 
 |   !cond(!eq(VT, i64)    : VSrc_b64, | 
 |         !eq(VT, f64)    : VSrc_f64, | 
 |         !eq(VT, i32)    : VSrc_b32, | 
 |         !eq(VT, f32)    : VSrc_f32, | 
 |         !eq(VT, i16)    : !if(IsTrue16, | 
 |                               !if(IsFake16, VSrcFake16_b16_Lo128, VSrcT_b16_Lo128), | 
 |                               VSrc_b16), | 
 |         !eq(VT, f16)    : !if(IsTrue16, | 
 |                               !if(IsFake16, VSrcFake16_f16_Lo128, VSrcT_f16_Lo128), | 
 |                               VSrc_f16), | 
 |         !eq(VT, bf16)   : !if(IsTrue16, | 
 |                               !if(IsFake16, VSrcFake16_bf16_Lo128, VSrcT_bf16_Lo128), | 
 |                               VSrc_bf16), | 
 |         !eq(VT, v2i16)  : VSrc_v2b16, | 
 |         !eq(VT, v2f16)  : VSrc_v2f16, | 
 |         !eq(VT, v2bf16) : VSrc_v2bf16, | 
 |         !eq(VT, v4f16)  : AVSrc_64, | 
 |         !eq(VT, v4bf16) : AVSrc_64, | 
 |         1               : VSrc_b32); | 
 | } | 
 |  | 
 | class getSOPSrcForVT<ValueType VT> { | 
 |   RegisterOperand ret = !if(!eq(VT.Size, 64), SSrc_b64, SSrc_b32); | 
 | } | 
 |  | 
 | // Returns the vreg register class to use for source operand given VT | 
 | class getVregSrcForVT<ValueType VT, bit IsTrue16 = 0, bit IsFake16 = 1> { | 
 |   RegisterOperand ret = | 
 |   !cond(!eq(VT.Size, 512) : RegisterOperand<VReg_512>, | 
 |         !eq(VT.Size, 192) : RegisterOperand<VReg_192>, | 
 |         !eq(VT.Size, 128) : RegisterOperand<VReg_128>, | 
 |         !eq(VT.Size, 96)  : RegisterOperand<VReg_96>, | 
 |         !eq(VT.Size, 64)  : RegisterOperand<VReg_64>, | 
 |         !eq(VT.Size, 48)  : RegisterOperand<VReg_64>, | 
 |         !eq(VT.Size, 16)  : !if(IsTrue16, | 
 |                                 !if(IsFake16, VGPRSrc_32_Lo128, VGPRSrc_16_Lo128), | 
 |                                 RegisterOperand<VGPR_32>), | 
 |         1                 : RegisterOperand<VGPR_32>); | 
 | } | 
 |  | 
 | class getSDWASrcForVT <ValueType VT> { | 
 |   RegisterOperand retFlt = !if(!eq(VT.Size, 16), SDWASrc_f16, SDWASrc_f32); | 
 |   RegisterOperand retInt = !if(!eq(VT.Size, 16), SDWASrc_i16, SDWASrc_i32); | 
 |   RegisterOperand ret = !if(VT.isFP, retFlt, retInt); | 
 | } | 
 |  | 
 | // Returns the register class to use for sources of VOP3 instructions for the | 
 | // given VT. | 
 | class getVOP3SrcForVT<ValueType VT, bit IsTrue16 = 0> { | 
 |   RegisterOperand ret = | 
 |   !cond(!eq(VT, f64)       : VSrc_f64, | 
 |         !eq(VT, f32)       : VSrc_f32, | 
 |         !eq(VT, f16)       : !if(IsTrue16, VSrcT_f16, VSrc_f16), | 
 |         !eq(VT, bf16)      : !if(IsTrue16, VSrcT_bf16, VSrc_bf16), | 
 |         !eq(VT, i16)       : !if(IsTrue16, VSrcT_b16, VSrc_b16), | 
 |         !eq(VT, i1)        : SSrc_i1, | 
 |         !eq(VT, v2f32)     : VSrc_v2f32, | 
 |         !eq(VT, v2i32)     : VSrc_v2b32, | 
 |         !eq(VT, v2f16)     : VSrc_v2f16, | 
 |         !eq(VT, v2bf16)    : VSrc_v2bf16, | 
 |         !eq(VT, v2i16)     : VSrc_v2b16, | 
 |         !eq(VT, v4f16)     : AVSrc_64, | 
 |         !eq(VT, v4bf16)    : AVSrc_64, | 
 |         !eq(VT.Size, 1024) : VRegSrc_1024, | 
 |         !eq(VT.Size, 512)  : VRegSrc_512, | 
 |         !eq(VT.Size, 256)  : VRegSrc_256, | 
 |         !eq(VT.Size, 192)  : VRegSrc_192, | 
 |         !eq(VT.Size, 128)  : VRegSrc_128, | 
 |         !eq(VT.Size, 96)   : VRegSrc_96, | 
 |         !eq(VT.Size, 64)   : VSrc_b64, | 
 |         1                  : VSrc_b32); | 
 | } | 
 |  | 
 | // VGPR only VOP3 src with 9 bit encoding | 
 | class getVOP3VRegSrcForVT<ValueType VT> { | 
 |   RegisterOperand ret = !cond(!eq(VT.Size, 1024) : VRegSrc_1024, | 
 |                               !eq(VT.Size, 512)  : VRegSrc_512, | 
 |                               !eq(VT.Size, 256)  : VRegSrc_256, | 
 |                               !eq(VT.Size, 192)  : VRegSrc_192, | 
 |                               !eq(VT.Size, 128)  : VRegSrc_128, | 
 |                               !eq(VT.Size, 96)   : VRegSrc_96, | 
 |                               !eq(VT.Size, 64)   : VRegSrc_64, | 
 |                               1 : VRegSrc_32); | 
 | } | 
 |  | 
 | // Src2 of VOP3 DPP instructions cannot be a literal | 
 | class getVOP3DPPSrcForVT<ValueType VT, bit IsFake16 = 1> { | 
 |   RegisterOperand ret = | 
 |   !cond(!eq(VT, i1)     : SSrc_i1, | 
 |         !eq(VT, i16)    : !if (IsFake16, VCSrc_b16, VCSrcT_b16), | 
 |         !eq(VT, f16)    : !if (IsFake16, VCSrc_f16, VCSrcT_f16), | 
 |         !eq(VT, bf16)   : !if (IsFake16, VCSrc_bf16, VCSrcT_bf16), | 
 |         !eq(VT, v2i16)  : VCSrc_v2b16, | 
 |         !eq(VT, v2f16)  : VCSrc_v2f16, | 
 |         !eq(VT, v2bf16) : VCSrc_v2bf16, | 
 |         !eq(VT, f32)    : VCSrc_f32, | 
 |         1               : VCSrc_b32); | 
 | } | 
 |  | 
 | // Float or packed int | 
 | class isModifierType<ValueType SrcVT> { | 
 |   bit ret = !or(!eq(SrcVT.Value, f16.Value), | 
 |                 !eq(SrcVT.Value, bf16.Value), | 
 |                 !eq(SrcVT.Value, f32.Value), | 
 |                 !eq(SrcVT.Value, f64.Value), | 
 |                 !eq(SrcVT.Value, v2f16.Value), | 
 |                 !eq(SrcVT.Value, v2i16.Value), | 
 |                 !eq(SrcVT.Value, v2bf16.Value), | 
 |                 !eq(SrcVT.Value, v2f32.Value), | 
 |                 !eq(SrcVT.Value, v2i32.Value), | 
 |                 !eq(SrcVT.Value, v4f16.Value), | 
 |                 !eq(SrcVT.Value, v4i16.Value), | 
 |                 !eq(SrcVT.Value, v4bf16.Value), | 
 |                 !eq(SrcVT.Value, v4f32.Value), | 
 |                 !eq(SrcVT.Value, v4i32.Value), | 
 |                 !eq(SrcVT.Value, v8f16.Value), | 
 |                 !eq(SrcVT.Value, v8i16.Value), | 
 |                 !eq(SrcVT.Value, v8bf16.Value), | 
 |                 !eq(SrcVT.Value, v8f32.Value), | 
 |                 !eq(SrcVT.Value, v8i32.Value), | 
 |                 !eq(SrcVT.Value, v16f16.Value), | 
 |                 !eq(SrcVT.Value, v16i16.Value), | 
 |                 !eq(SrcVT.Value, v16bf16.Value)); | 
 | } | 
 |  | 
 | // Return type of input modifiers operand for specified input operand. | 
 | // True16: If the destination is a 16-bit value, the src0 modifier must hold | 
 | // dst's opsel bit. Use a dummy value for DstVT if getting the mod for a src operand besides 0. | 
 | // 64-bit src types are not implemented for True16 dst. | 
 | class getSrc0Mod <ValueType VT, ValueType DstVT, bit IsTrue16 = 0, bit IsFake16 = 1> { | 
 |   defvar T16Dst =  !if(!eq(VT.Size, 64), | 
 |                      !if(VT.isFP, FP64InputMods, Int64InputMods), | 
 |                      !if(!eq(VT.Size, 16), | 
 |                          !if(VT.isFP, !if(IsTrue16, FPT16InputMods<IsFake16>, FP16InputMods), | 
 |                                       !if(IsTrue16, IntT16InputMods<IsFake16>, IntOpSelMods)), | 
 |                          !if(VT.isFP, FP32T16DstInputMods, Int32T16DstInputMods))); | 
 |   defvar Normal =  !if(!eq(VT.Size, 64), | 
 |                      !if(VT.isFP, FP64InputMods, Int64InputMods), | 
 |                      !if(!eq(VT.Size, 16), | 
 |                          !if(VT.isFP, !if(IsTrue16, FPT16InputMods<IsFake16>, FP16InputMods), | 
 |                                       !if(IsTrue16, IntT16InputMods<IsFake16>, IntOpSelMods)), | 
 |                          !if(VT.isFP, FP32InputMods, Int32InputMods))); | 
 |   Operand ret = !if(!and(IsTrue16, !eq(DstVT.Size, 16)), T16Dst, Normal); | 
 | } | 
 |  | 
 | class getSrcMod<ValueType VT, bit IsTrue16 = 0, bit IsFake16 = 1> : getSrc0Mod<VT, f128/*Dummy Arg*/, IsTrue16, IsFake16>; | 
 |  | 
 | // Return type of input modifiers operand specified input operand for DPP | 
 | class getSrcModDPP <ValueType VT> { | 
 |   Operand ret = !if(VT.isFP, FPVRegInputMods, IntVRegInputMods); | 
 | } | 
 |  | 
 | class getSrcModDPP_t16 <ValueType VT, bit IsFake16 = 1> { | 
 |   Operand ret = | 
 |       !if (VT.isFP, | 
 |            !if (!or(!eq(VT.Value, f16.Value), !eq(VT.Value, bf16.Value)), | 
 |                 FPT16_Lo128VRegInputMods<IsFake16>, FPVRegInputMods), | 
 |            !if (!eq(VT.Value, i16.Value), | 
 |                 IntT16_Lo128VRegInputMods<IsFake16>, IntVRegInputMods)); | 
 | } | 
 |  | 
 | // Return type of input modifiers operand for specified input operand for DPP | 
 | // True16: If the destination is a 16-bit value, the src0 modifier must hold | 
 | // dst's opsel bit. Use a dummy value for DstVT if getting the mod for a src operand besides 0. | 
 | // 64-bit src types are not implemented for True16 dst. | 
 | class getSrc0ModVOP3DPP <ValueType VT, ValueType DstVT, bit IsFake16 = 1> { | 
 |   defvar T16Dst = | 
 |       !if (VT.isFP, | 
 |            !if (!or(!eq(VT.Value, f16.Value), !eq(VT.Value, bf16.Value)), | 
 |                 FPT16VRegInputMods<IsFake16>, FPVRegT16DstInputMods), | 
 |            !if (!eq(VT.Value, i16.Value), IntT16VRegInputMods<IsFake16>, | 
 |                 IntVRegT16DstInputMods)); | 
 |   defvar Normal = | 
 |       !if (VT.isFP, | 
 |            !if (!or(!eq(VT.Value, f16.Value), !eq(VT.Value, bf16.Value)), | 
 |                 FPT16VRegInputMods<IsFake16>, FPVRegInputMods), | 
 |            !if (!eq(VT.Value, i16.Value), | 
 |                 IntT16VRegInputMods<IsFake16>, | 
 |                 IntVRegInputMods)); | 
 |   Operand ret = !if(!and(!not(IsFake16), !eq(DstVT.Size, 16)), T16Dst, Normal); | 
 | } | 
 |  | 
 | // GFX11 only supports VGPR src1, but the restriction is done in AsmParser | 
 | // and GCNDPPCombine. | 
 | class getSrcModVOP3DPP<ValueType VT, bit IsFake16 = 1> { | 
 |   Operand ret = | 
 |       !if (VT.isFP, | 
 |            !if (!or(!eq(VT.Value, f16.Value), !eq(VT.Value, bf16.Value)), | 
 |                 FPT16VCSrcInputMods<IsFake16>, FP32VCSrcInputMods), | 
 |            !if (!eq(VT.Value, i16.Value), | 
 |                 IntT16VCSrcInputMods<IsFake16>, | 
 |                 Int32VCSrcInputMods)); | 
 | } | 
 |  | 
 | // Return type of input modifiers operand specified input operand for SDWA | 
 | class getSrcModSDWA <ValueType VT> { | 
 |   Operand ret = !if(!eq(VT.Value, f16.Value), FP16SDWAInputMods, | 
 |                 !if(!eq(VT.Value, f32.Value), FP32SDWAInputMods, | 
 |                 !if(!eq(VT.Value, i16.Value), Int16SDWAInputMods, | 
 |                 !if(!eq(VT.Value, bf16.Value), FP16SDWAInputMods, | 
 |                 Int32SDWAInputMods)))); | 
 | } | 
 |  | 
 | // Returns the input arguments for VOP[12C] instructions for the given SrcVT. | 
 | class getIns32 <RegisterOperand Src0RC, RegisterOperand Src1RC, int NumSrcArgs> { | 
 |   dag ret = !if(!eq(NumSrcArgs, 1), (ins Src0RC:$src0),               // VOP1 | 
 |             !if(!eq(NumSrcArgs, 2), (ins Src0RC:$src0, Src1RC:$src1), // VOP2 | 
 |                                     (ins))); | 
 | } | 
 |  | 
 | // Returns the input arguments for VOP3 instructions for the given SrcVT. | 
 | class getIns64 <RegisterOperand Src0RC, RegisterOperand Src1RC, | 
 |                 RegisterOperand Src2RC, int NumSrcArgs, | 
 |                 bit HasClamp, bit HasModifiers, bit HasSrc2Mods, bit HasOMod, | 
 |                 Operand Src0Mod, Operand Src1Mod, Operand Src2Mod, | 
 |                 bit HasFP8ByteSel = 0, bit HasFP8DstByteSel = 0, | 
 |                 bit HasBitOp3 = 0> { | 
 |   dag src0 = !if(!ge(NumSrcArgs, 1), | 
 |                  !if (HasModifiers, | 
 |                       (ins Src0Mod:$src0_modifiers, Src0RC:$src0), | 
 |                       (ins Src0RC:$src0)), | 
 |                  (ins)); | 
 |   dag src1 = !if(!ge(NumSrcArgs, 2), | 
 |                  !if (HasModifiers, | 
 |                       (ins Src1Mod:$src1_modifiers, Src1RC:$src1), | 
 |                       (ins Src1RC:$src1)), | 
 |                  (ins)); | 
 |   dag src2 = !if(!ge(NumSrcArgs, 3), | 
 |                  !if (HasSrc2Mods, | 
 |                       (ins Src2Mod:$src2_modifiers, Src2RC:$src2), | 
 |                       (ins Src2RC:$src2)), | 
 |                  (ins)); | 
 |   // If there is vdst_in after clamp with HasFP8DstByteSel we cannot use | 
 |   // Clamp0 with default value, all default operands must be at the end. | 
 |   dag clamp = !if(HasClamp, !if(HasFP8DstByteSel, (ins Clamp:$clamp), | 
 |                                                   (ins Clamp0:$clamp)), | 
 |                             (ins)); | 
 |   dag omod = !if(HasOMod, (ins omod0:$omod), (ins)); | 
 |   dag bytesel = !if(HasFP8ByteSel, | 
 |                     !con(!if(HasFP8DstByteSel, (ins VGPR_32:$vdst_in), (ins)), | 
 |                          (ins ByteSel0:$byte_sel)), | 
 |                     (ins)); | 
 |   dag bitop3 = !if(HasBitOp3, (ins bitop3_0:$bitop3), (ins)); | 
 |  | 
 |   dag ret = !con(src0, src1, src2, clamp, omod, bytesel, bitop3); | 
 | } | 
 |  | 
 | class getInsVOP3Base<RegisterOperand Src0RC, RegisterOperand Src1RC, | 
 |                 RegisterOperand Src2RC, int NumSrcArgs, | 
 |                 bit HasClamp, bit HasModifiers, bit HasSrc2Mods, bit HasOMod, | 
 |                 Operand Src0Mod, Operand Src1Mod, Operand Src2Mod, bit HasOpSel, | 
 |                 bit HasFP8ByteSel = 0, bit HasFP8DstByteSel = 0, bit HasBitOp3 = 0> { | 
 |   // getInst64 handles clamp and omod. implicit mutex between vop3p and omod | 
 |   dag base = getIns64 <Src0RC, Src1RC, Src2RC, NumSrcArgs, | 
 |                 HasClamp, HasModifiers, HasSrc2Mods, HasOMod, | 
 |                 Src0Mod, Src1Mod, Src2Mod, HasFP8ByteSel, HasFP8DstByteSel>.ret; | 
 |   dag opsel = (ins op_sel0:$op_sel); | 
 |   dag bitop3 = (ins bitop3_0:$bitop3); | 
 |   dag ret = !con(base, !if(HasBitOp3, bitop3, (ins)), !if(HasOpSel, opsel, (ins))); | 
 | } | 
 |  | 
 | class getInsVOP3P <RegisterOperand Src0RC, RegisterOperand Src1RC, | 
 |                    RegisterOperand Src2RC, int NumSrcArgs, bit HasClamp, bit HasOpSel, | 
 |                    bit HasNeg, | 
 |                    Operand Src0Mod, Operand Src1Mod, Operand Src2Mod> { | 
 |   dag base = getInsVOP3Base<Src0RC, Src1RC, Src2RC, NumSrcArgs, | 
 |                     HasClamp, 1/*HasModifiers*/, 1/*HasSrc2Mods*/, | 
 |                     0/*HasOMod*/, Src0Mod, Src1Mod, Src2Mod, HasOpSel>.ret; | 
 |  | 
 |   dag vop3pOpsel = (ins op_sel_hi0:$op_sel_hi); | 
 |   dag vop3p_neg = !if(HasNeg, (ins neg_lo0:$neg_lo, neg_hi0:$neg_hi), (ins)); | 
 |  | 
 |   dag vop3pFields = !con(!if(HasOpSel, vop3pOpsel, (ins)), vop3p_neg); | 
 |   dag ret = !con(base, vop3pFields); | 
 | } | 
 |  | 
 | class getInsVOP3OpSel <RegisterOperand Src0RC, RegisterOperand Src1RC, | 
 |                        RegisterOperand Src2RC, int NumSrcArgs, | 
 |                        bit HasClamp, bit HasOMod, | 
 |                        Operand Src0Mod, Operand Src1Mod, Operand Src2Mod, | 
 |                        bit HasFP8ByteSel = 0, bit HasFP8DstByteSel = 0, bit HasBitOp3 = 0> { | 
 |   dag ret = getInsVOP3Base<Src0RC, Src1RC, | 
 |                     Src2RC, NumSrcArgs, | 
 |                     HasClamp, 1/*HasModifiers*/, 1/*HasSrc2Mods*/, HasOMod, | 
 |                     Src0Mod, Src1Mod, Src2Mod, /*HasOpSel=*/1, | 
 |                     HasFP8ByteSel, HasFP8DstByteSel, HasBitOp3>.ret; | 
 | } | 
 |  | 
 | class getInsDPPBase <RegisterOperand OldRC, RegisterOperand Src0RC, RegisterOperand Src1RC, | 
 |                      RegisterOperand Src2RC, int NumSrcArgs, bit HasModifiers, | 
 |                      Operand Src0Mod, Operand Src1Mod, Operand Src2Mod, bit HasOld> { | 
 |   dag ret = !if(!eq(NumSrcArgs, 0), | 
 |                 // VOP1 without input operands (V_NOP) | 
 |                 (ins ), | 
 |                 !con( | 
 |                   !if(HasOld ,(ins OldRC:$old), (ins)), | 
 |                   !if (!eq(NumSrcArgs, 1), | 
 |                     !if (HasModifiers, | 
 |                       // VOP1_DPP with modifiers | 
 |                       (ins Src0Mod:$src0_modifiers, Src0RC:$src0) | 
 |                     /* else */, | 
 |                       // VOP1_DPP without modifiers | 
 |                       (ins Src0RC:$src0) | 
 |                     /* endif */), | 
 |                   !if (!eq(NumSrcArgs, 2), | 
 |                     !if (HasModifiers, | 
 |                       // VOP2_DPP with modifiers | 
 |                       (ins Src0Mod:$src0_modifiers, Src0RC:$src0, | 
 |                        Src1Mod:$src1_modifiers, Src1RC:$src1) | 
 |                     /* else */, | 
 |                       // VOP2_DPP without modifiers | 
 |                       (ins Src0RC:$src0, Src1RC:$src1) | 
 |                     ) | 
 |                     /* NumSrcArgs == 3, VOP3 */, | 
 |                     !if (HasModifiers, | 
 |                       // VOP3_DPP with modifiers | 
 |                       (ins Src0Mod:$src0_modifiers, Src0RC:$src0, | 
 |                        Src1Mod:$src1_modifiers, Src1RC:$src1, | 
 |                        Src2Mod:$src2_modifiers, Src2RC:$src2) | 
 |                     /* else */, | 
 |                       // VOP3_DPP without modifiers | 
 |                       (ins Src0RC:$src0, Src1RC:$src1, | 
 |                        Src2RC:$src2) | 
 |                       ) | 
 |                     ) | 
 |                   ) | 
 |                 ) | 
 |             ); | 
 | } | 
 |  | 
 | class getInsDPP <RegisterOperand OldRC, RegisterOperand Src0RC, RegisterOperand Src1RC, | 
 |                  RegisterOperand Src2RC, int NumSrcArgs, bit HasModifiers, | 
 |                  Operand Src0Mod, Operand Src1Mod, Operand Src2Mod, bit HasOld = 1> { | 
 |   dag ret = !con(getInsDPPBase<OldRC, Src0RC, Src1RC, Src2RC, NumSrcArgs, | 
 |                            HasModifiers, Src0Mod, Src1Mod, Src2Mod, HasOld>.ret, | 
 |                  (ins dpp_ctrl:$dpp_ctrl, DppRowMask:$row_mask, | 
 |                       DppBankMask:$bank_mask, DppBoundCtrl:$bound_ctrl)); | 
 | } | 
 |  | 
 | class getInsDPP16 <RegisterOperand OldRC, RegisterOperand Src0RC, RegisterOperand Src1RC, | 
 |                    RegisterOperand Src2RC, int NumSrcArgs, bit HasModifiers, | 
 |                    Operand Src0Mod, Operand Src1Mod, Operand Src2Mod, bit HasOld = 1> { | 
 |   dag ret = !con(getInsDPP<OldRC, Src0RC, Src1RC, Src2RC, NumSrcArgs, | 
 |                            HasModifiers, Src0Mod, Src1Mod, Src2Mod, HasOld>.ret, | 
 |                  (ins Dpp16FI:$fi)); | 
 | } | 
 |  | 
 | class getInsDPP8 <RegisterOperand OldRC, RegisterOperand Src0RC, RegisterOperand Src1RC, | 
 |                   RegisterOperand Src2RC, int NumSrcArgs, bit HasModifiers, | 
 |                   Operand Src0Mod, Operand Src1Mod, Operand Src2Mod, bit HasOld = 1> { | 
 |   dag ret = !con(getInsDPPBase<OldRC, Src0RC, Src1RC, Src2RC, NumSrcArgs, | 
 |                            HasModifiers, Src0Mod, Src1Mod, Src2Mod, HasOld>.ret, | 
 |                  (ins dpp8:$dpp8, Dpp8FI:$fi)); | 
 | } | 
 |  | 
 | class getInsVOP3DPPBase<dag VOP3Base, RegisterOperand OldRC, int NumSrcArgs, bit HasOld> { | 
 |   dag old = ( ins OldRC:$old ); | 
 |   dag base = VOP3Base; | 
 |   dag ret =  !con( | 
 |                 !if(!and(HasOld,!ne(NumSrcArgs, 0)), old, (ins)), | 
 |                 base | 
 |               ); | 
 | } | 
 |  | 
 | class getInsVOP3DPP<dag VOP3Base, RegisterOperand OldRC, int NumSrcArgs, bit HasOld = 1> { | 
 |   dag ret = !con(getInsVOP3DPPBase<VOP3Base,OldRC,NumSrcArgs,HasOld>.ret, | 
 |                  (ins dpp_ctrl:$dpp_ctrl, DppRowMask:$row_mask, | 
 |                       DppBankMask:$bank_mask, DppBoundCtrl:$bound_ctrl)); | 
 | } | 
 |  | 
 | class getInsVOP3DPP16<dag VOP3Base, RegisterOperand OldRC, int NumSrcArgs, bit HasOld = 1> { | 
 |   dag ret = !con(getInsVOP3DPP<VOP3Base,OldRC,NumSrcArgs,HasOld>.ret, | 
 |                  (ins Dpp16FI:$fi)); | 
 | } | 
 |  | 
 | class getInsVOP3DPP8<dag VOP3Base, RegisterOperand OldRC, int NumSrcArgs, bit HasOld = 1> { | 
 |   dag ret = !con(getInsVOP3DPPBase<VOP3Base,OldRC,NumSrcArgs,HasOld>.ret, | 
 |                  (ins dpp8:$dpp8, Dpp8FI:$fi)); | 
 | } | 
 |  | 
 | // Ins for SDWA | 
 | class getInsSDWA <RegisterOperand Src0RC, RegisterOperand Src1RC, int NumSrcArgs, | 
 |                   bit HasSDWAOMod, Operand Src0Mod, Operand Src1Mod, | 
 |                   ValueType DstVT> { | 
 |  | 
 |   dag ret = !if(!eq(NumSrcArgs, 0), | 
 |                // VOP1 without input operands (V_NOP) | 
 |                (ins), | 
 |             !if(!eq(NumSrcArgs, 1), | 
 |                // VOP1 | 
 |                !if(!not(HasSDWAOMod), | 
 |                   // VOP1_SDWA without omod | 
 |                   (ins Src0Mod:$src0_modifiers, Src0RC:$src0, | 
 |                        Clamp:$clamp, | 
 |                        dst_sel:$dst_sel, dst_unused:$dst_unused, | 
 |                        src0_sel:$src0_sel), | 
 |                   // VOP1_SDWA with omod | 
 |                   (ins Src0Mod:$src0_modifiers, Src0RC:$src0, | 
 |                        Clamp:$clamp, omod:$omod, | 
 |                        dst_sel:$dst_sel, dst_unused:$dst_unused, | 
 |                        src0_sel:$src0_sel)), | 
 |             !if(!eq(NumSrcArgs, 2), | 
 |                !if(!eq(DstVT.Size, 1), | 
 |                   // VOPC_SDWA | 
 |                   (ins Src0Mod:$src0_modifiers, Src0RC:$src0, | 
 |                        Src1Mod:$src1_modifiers, Src1RC:$src1, | 
 |                        Clamp:$clamp, src0_sel:$src0_sel, src1_sel:$src1_sel), | 
 |                   // VOP2_SDWA | 
 |                   !if(!not(HasSDWAOMod), | 
 |                      // VOP2_SDWA without omod | 
 |                      (ins Src0Mod:$src0_modifiers, Src0RC:$src0, | 
 |                           Src1Mod:$src1_modifiers, Src1RC:$src1, | 
 |                           Clamp:$clamp, | 
 |                           dst_sel:$dst_sel, dst_unused:$dst_unused, | 
 |                           src0_sel:$src0_sel, src1_sel:$src1_sel), | 
 |                      // VOP2_SDWA with omod | 
 |                      (ins Src0Mod:$src0_modifiers, Src0RC:$src0, | 
 |                           Src1Mod:$src1_modifiers, Src1RC:$src1, | 
 |                           Clamp:$clamp, omod:$omod, | 
 |                           dst_sel:$dst_sel, dst_unused:$dst_unused, | 
 |                           src0_sel:$src0_sel, src1_sel:$src1_sel))), | 
 |             (ins)/* endif */))); | 
 | } | 
 |  | 
 | // Outs for DPP | 
 | class getOutsDPP <bit HasDst, ValueType DstVT, RegisterOperand DstRCDPP> { | 
 |   dag ret = !if(HasDst, | 
 |                 !if(!eq(DstVT.Size, 1), | 
 |                     (outs), // no dst for VOPC, we use "vcc"-token as dst in SDWA VOPC instructions | 
 |                     (outs DstRCDPP:$vdst)), | 
 |                 (outs)); // V_NOP | 
 | } | 
 |  | 
 | // Outs for SDWA | 
 | class getOutsSDWA <bit HasDst, ValueType DstVT, RegisterOperand DstRCSDWA> { | 
 |   dag ret = !if(HasDst, | 
 |                 !if(!eq(DstVT.Size, 1), | 
 |                     (outs DstRCSDWA:$sdst), | 
 |                     (outs DstRCSDWA:$vdst)), | 
 |                 (outs)); // V_NOP | 
 | } | 
 |  | 
 | // Returns the assembly string for the inputs and outputs of a VOP[12C] | 
 | // instruction. | 
 | class getAsm32 <bit HasDst, int NumSrcArgs, ValueType DstVT = i32> { | 
 |   string dst = !if(!eq(DstVT.Size, 1), "$sdst", "$vdst"); // use $sdst for VOPC | 
 |   string src0 = ", $src0"; | 
 |   string src1 = ", $src1"; | 
 |   string src2 = ", $src2"; | 
 |   string ret = !if(HasDst, dst, "") # | 
 |                !if(!eq(NumSrcArgs, 1), src0, "") # | 
 |                !if(!eq(NumSrcArgs, 2), src0#src1, "") # | 
 |                !if(!eq(NumSrcArgs, 3), src0#src1#src2, ""); | 
 | } | 
 |  | 
 | class getAsmVOPDPart <int NumSrcArgs, string XorY> { | 
 |   string dst = "$vdst" # XorY; | 
 |   string src0 = ", $src0" # XorY; | 
 |   string src1 = ", $vsrc1" # XorY; | 
 |   string ret = dst # | 
 |                !if(!ge(NumSrcArgs, 1), src0, "") # | 
 |                !if(!ge(NumSrcArgs, 2), src1, ""); | 
 | } | 
 |  | 
 | // Returns the assembly string for the inputs and outputs of a VOP3P | 
 | // instruction. | 
 | class getAsmVOP3P <bit HasDst, int NumSrcArgs, bit HasNeg, | 
 |                    bit HasClamp, bit HasOpSel> { | 
 |   string dst = !if(HasDst, "$vdst"# !if(!gt(NumSrcArgs, 0), ",", ""), ""); | 
 |   string src0 = !if(!eq(NumSrcArgs, 1), " $src0", " $src0,"); | 
 |   string src1 = !if(!eq(NumSrcArgs, 1), "", | 
 |                    !if(!eq(NumSrcArgs, 2), " $src1", | 
 |                                            " $src1,")); | 
 |   string src2 = !if(!eq(NumSrcArgs, 3), " $src2", ""); | 
 |  | 
 |   string mods = !if(HasNeg, "$neg_lo$neg_hi", ""); | 
 |   string clamp = !if(HasClamp, "$clamp", ""); | 
 |   string opsel = !if(HasOpSel, "$op_sel$op_sel_hi", ""); | 
 |  | 
 |   // Each modifier is printed as an array of bits for each operand, so | 
 |   // all operands are printed as part of src0_modifiers. | 
 |   string ret = dst#src0#src1#src2#opsel#mods#clamp; | 
 | } | 
 |  | 
 | class getAsmDPP <bit HasDst, int NumSrcArgs, bit HasModifiers, ValueType DstVT = i32> { | 
 |   string dst = !if(HasDst, | 
 |                    !if(!eq(DstVT.Size, 1), | 
 |                        "$sdst", | 
 |                        "$vdst"), | 
 |                     ""); // use $sdst for VOPC | 
 |   string src0 = !if(!eq(NumSrcArgs, 1), "$src0_modifiers", "$src0_modifiers,"); | 
 |   string src1 = !if(!eq(NumSrcArgs, 1), "", | 
 |                    !if(!eq(NumSrcArgs, 2), " $src1_modifiers", | 
 |                                            " $src1_modifiers,")); | 
 |   string args = !if(!not(HasModifiers), | 
 |                      getAsm32<0, NumSrcArgs, DstVT>.ret, | 
 |                      ", "#src0#src1); | 
 |   string ret = dst#args#" $dpp_ctrl$row_mask$bank_mask$bound_ctrl"; | 
 | } | 
 |  | 
 | class getAsmDPP16 <bit HasDst, int NumSrcArgs, bit HasModifiers, ValueType DstVT = i32> { | 
 |   string ret = getAsmDPP<HasDst, NumSrcArgs, HasModifiers, DstVT>.ret#"$fi"; | 
 | } | 
 |  | 
 | class getAsmDPP8 <bit HasDst, int NumSrcArgs, bit HasModifiers, ValueType DstVT = i32> | 
 |   : getAsmDPP<HasDst, NumSrcArgs, HasModifiers, DstVT>{ | 
 |   let ret = dst#args#" $dpp8$fi"; | 
 | } | 
 |  | 
 | class getAsmVOP3Base <int NumSrcArgs, bit HasDst, bit HasClamp, | 
 |                        bit HasOpSel, bit HasOMod, bit IsVOP3P, | 
 |                        bit HasNeg, bit Src0HasMods, | 
 |                        bit Src1HasMods, bit Src2HasMods, ValueType DstVT = i32, | 
 |                        bit HasByteSel = 0, bit HasBitOp3 = 0> { | 
 |   string dst = !if(HasDst, | 
 |                    !if(!eq(DstVT.Size, 1), | 
 |                        "$sdst", | 
 |                        "$vdst"), | 
 |                     ""); // use $sdst for VOPC | 
 |   string src0nomods = !if(!eq(NumSrcArgs, 1), "$src0", "$src0,"); | 
 |   string src1nomods = !if(!eq(NumSrcArgs, 1), "", | 
 |                     !if(!eq(NumSrcArgs, 2), " $src1", | 
 |                                             " $src1,")); | 
 |   string src2nomods = !if(!eq(NumSrcArgs, 3), " $src2", ""); | 
 |  | 
 |   string src0mods = !if(!eq(NumSrcArgs, 1), "$src0_modifiers", "$src0_modifiers,"); | 
 |   string src1mods = !if(!eq(NumSrcArgs, 1), "", | 
 |                     !if(!eq(NumSrcArgs, 2), " $src1_modifiers", | 
 |                                             " $src1_modifiers,")); | 
 |   string src2mods = !if(!eq(NumSrcArgs, 3), " $src2_modifiers", ""); | 
 |  | 
 |   string src0 = !if(Src0HasMods, src0mods, src0nomods); | 
 |   string src1 = !if(Src1HasMods, src1mods, src1nomods); | 
 |   string src2 = !if(Src2HasMods, src2mods, src2nomods); | 
 |   string opsel = !if(HasOpSel, "$op_sel", ""); | 
 |   string bytesel = !if(HasByteSel, "$byte_sel", ""); | 
 |   string bitop3 = !if(HasBitOp3, "$bitop3", ""); | 
 |   string 3PMods = !if(IsVOP3P, | 
 |                       !if(HasOpSel, "$op_sel_hi", "") | 
 |                         #!if(HasNeg, "$neg_lo$neg_hi", ""), | 
 |                       ""); | 
 |   string clamp = !if(HasClamp, "$clamp", ""); | 
 |   string omod = !if(HasOMod, "$omod", ""); | 
 |  | 
 |   string ret = dst#!if(!eq(NumSrcArgs,0), | 
 |                        "", | 
 |                        !if(HasDst,", ", "")#src0#src1#src2#bitop3#opsel#bytesel#3PMods#clamp#omod); | 
 | } | 
 |  | 
 | class getAsmVOP3DPP<string base> { | 
 |   string ret = base # " $dpp_ctrl$row_mask$bank_mask$bound_ctrl"; | 
 | } | 
 |  | 
 | class getAsmVOP3DPP16<string base> { | 
 |   string ret = getAsmVOP3DPP<base>.ret # "$fi"; | 
 | } | 
 |  | 
 | class getAsmVOP3DPP8<string base> { | 
 |   string ret = base # " $dpp8$fi"; | 
 | } | 
 |  | 
 |  | 
 | class getAsmSDWA <bit HasDst, int NumSrcArgs, ValueType DstVT = i32> { | 
 |   string dst = !if(HasDst, | 
 |                    !if(!eq(DstVT.Size, 1), | 
 |                        " vcc", // use vcc token as dst for VOPC instructions | 
 |                        "$vdst"), | 
 |                     ""); | 
 |   string src0 = "$src0_modifiers"; | 
 |   string src1 = "$src1_modifiers"; | 
 |   string args = !if(!eq(NumSrcArgs, 0), | 
 |                     "", | 
 |                     !if(!eq(NumSrcArgs, 1), | 
 |                         ", "#src0#"$clamp", | 
 |                         ", "#src0#", "#src1#"$clamp" | 
 |                      ) | 
 |                 ); | 
 |   string sdwa = !if(!eq(NumSrcArgs, 0), | 
 |                     "", | 
 |                     !if(!eq(NumSrcArgs, 1), | 
 |                         " $dst_sel $dst_unused $src0_sel", | 
 |                         !if(!eq(DstVT.Size, 1), | 
 |                             " $src0_sel $src1_sel", // No dst_sel and dst_unused for VOPC | 
 |                             " $dst_sel $dst_unused $src0_sel $src1_sel" | 
 |                         ) | 
 |                     ) | 
 |                 ); | 
 |   string ret = dst#args#sdwa; | 
 | } | 
 |  | 
 | class getAsmSDWA9 <bit HasDst, bit HasOMod, int NumSrcArgs, | 
 |                    ValueType DstVT = i32> { | 
 |   string dst = !if(HasDst, | 
 |                    !if(!eq(DstVT.Size, 1), | 
 |                        "$sdst", // VOPC | 
 |                        "$vdst"), // VOP1/2 | 
 |                     ""); | 
 |   string src0 = "$src0_modifiers"; | 
 |   string src1 = "$src1_modifiers"; | 
 |   string out_mods = !if(!not(HasOMod), "$clamp", "$clamp$omod"); | 
 |   string args = !if(!eq(NumSrcArgs, 0), "", | 
 |                     !if(!eq(NumSrcArgs, 1), | 
 |                         ", "#src0, | 
 |                         ", "#src0#", "#src1 | 
 |                      ) | 
 |                 ); | 
 |   string sdwa = !if(!eq(NumSrcArgs, 0), "", | 
 |                     !if(!eq(NumSrcArgs, 1), | 
 |                         out_mods#" $dst_sel $dst_unused $src0_sel", | 
 |                         !if(!eq(DstVT.Size, 1), | 
 |                             " $src0_sel $src1_sel", // No dst_sel, dst_unused and output modifiers for VOPC | 
 |                             out_mods#" $dst_sel $dst_unused $src0_sel $src1_sel" | 
 |                         ) | 
 |                     ) | 
 |                 ); | 
 |   string ret = dst#args#sdwa; | 
 | } | 
 |  | 
 | class getHas64BitOps <int NumSrcArgs, ValueType DstVT, ValueType Src0VT, | 
 |                       ValueType Src1VT> { | 
 |   bit ret = !if(!eq(NumSrcArgs, 3), | 
 |                 0, | 
 |                 !if(!eq(DstVT.Size, 64), | 
 |                     1, | 
 |                     !if(!eq(Src0VT.Size, 64), | 
 |                         1, | 
 |                         !if(!eq(Src1VT.Size, 64), | 
 |                             1, | 
 |                             0 | 
 |                         ) | 
 |                     ) | 
 |                 ) | 
 |             ); | 
 | } | 
 |  | 
 | class getHasSDWA <int NumSrcArgs, ValueType DstVT = i32, ValueType Src0VT = i32, | 
 |                   ValueType Src1VT = i32> { | 
 |   bit ret = !if(!eq(NumSrcArgs, 3), | 
 |                 0, // NumSrcArgs == 3 - No SDWA for VOP3 | 
 |                 !if(!eq(DstVT.Size, 64), | 
 |                     0, // 64-bit dst - No SDWA for 64-bit operands | 
 |                     !if(!eq(Src0VT.Size, 64), | 
 |                         0, // 64-bit src0 | 
 |                         !if(!eq(Src1VT.Size, 64), | 
 |                             0, // 64-bit src2 | 
 |                             1 | 
 |                         ) | 
 |                     ) | 
 |                 ) | 
 |             ); | 
 | } | 
 |  | 
 | class getHasDPP <int NumSrcArgs> { | 
 |   bit ret = !if(!eq(NumSrcArgs, 3), | 
 |                 0, // NumSrcArgs == 3 - No DPP for VOP3 | 
 |                 1); | 
 | } | 
 |  | 
 | class getHasExt32BitDPP <int NumSrcArgs, ValueType DstVT = i32, ValueType Src0VT = i32, | 
 |                  ValueType Src1VT = i32> { | 
 |   bit ret = !and(getHasDPP<NumSrcArgs>.ret, | 
 |                  !not(getHas64BitOps<NumSrcArgs, DstVT, Src0VT, Src1VT>.ret)); | 
 | } | 
 |  | 
 | class getHasExt64BitDPP <int NumSrcArgs, ValueType DstVT = i32, ValueType Src0VT = i32, | 
 |                  ValueType Src1VT = i32> { | 
 |   bit ret = !and(getHasDPP<NumSrcArgs>.ret, | 
 |                  getHas64BitOps<NumSrcArgs, DstVT, Src0VT, Src1VT>.ret); | 
 | } | 
 |  | 
 | // Function that checks if instruction supports DPP and SDWA | 
 | class getHasExt <int NumSrcArgs, ValueType DstVT = i32, ValueType Src0VT = i32, | 
 |                  ValueType Src1VT = i32> { | 
 |   bit ret = !or(getHasDPP<NumSrcArgs>.ret, | 
 |                 getHasSDWA<NumSrcArgs, DstVT, Src0VT, Src1VT>.ret); | 
 | } | 
 |  | 
 | // Return an AGPR+VGPR operand class for the given VGPR register class. | 
 | class getLdStRegisterOperand<RegisterClass RC> { | 
 |   // This type of operands is only used in pseudo instructions helping | 
 |   // code generation and thus doesn't need encoding and decoding methods. | 
 |   // It also doesn't need to support AGPRs, because GFX908/A/40 do not | 
 |   // support True16. | 
 |   defvar VLdSt_16 = RegisterOperand<VGPR_16>; | 
 |  | 
 |   RegisterOperand ret = | 
 |     !cond(!eq(RC.Size, 16)   : VLdSt_16, | 
 |           !eq(RC.Size, 32)   : AVLdSt_32, | 
 |           !eq(RC.Size, 64)   : AVLdSt_64, | 
 |           !eq(RC.Size, 96)   : AVLdSt_96, | 
 |           !eq(RC.Size, 128)  : AVLdSt_128, | 
 |           !eq(RC.Size, 160)  : AVLdSt_160, | 
 |           !eq(RC.Size, 1024) : AVLdSt_1024); | 
 | } | 
 |  | 
 | class getHasVOP3DPP <ValueType DstVT = i32, ValueType Src0VT = i32, | 
 |                  ValueType Src1VT = i32, ValueType Src2VT = i32> { | 
 |   bit ret =    !if(!eq(DstVT.Size, 64), | 
 |                     0, // 64-bit dst No DPP for 64-bit operands | 
 |                     !if(!eq(Src0VT.Size, 64), | 
 |                         0, // 64-bit src0 | 
 |                         !if(!eq(Src1VT.Size, 64), | 
 |                             0, // 64-bit src1 | 
 |                             !if(!eq(Src2VT.Size, 64), | 
 |                                 0, // 64-bit src2 | 
 |                                 1 | 
 |                             ) | 
 |                         ) | 
 |                     ) | 
 |                 ); | 
 | } | 
 |  | 
 |  | 
 | def PatGenMode { | 
 |   int NoPattern = 0; | 
 |   int Pattern   = 1; | 
 | } | 
 |  | 
 | class VOPProfile <list<ValueType> _ArgVT, bit _EnableClamp = 0> { | 
 |  | 
 |   field list<ValueType> ArgVT = _ArgVT; | 
 |   field bit EnableClamp = _EnableClamp; | 
 |   field bit IsTrue16 = 0; | 
 |   field bit IsRealTrue16 = 0; | 
 |  | 
 |   field ValueType DstVT = ArgVT[0]; | 
 |   field ValueType Src0VT = ArgVT[1]; | 
 |   field ValueType Src1VT = ArgVT[2]; | 
 |   field ValueType Src2VT = ArgVT[3]; | 
 |   field RegisterOperand DstRC = getVALUDstForVT<DstVT>.ret; | 
 |   field RegisterOperand DstRCDPP = DstRC; | 
 |   field RegisterOperand DstRC64 = DstRC; | 
 |   field RegisterOperand DstRCVOP3DPP = DstRC64; | 
 |   field RegisterOperand DstRCSDWA = getSDWADstForVT<DstVT>.ret; | 
 |   field RegisterOperand Src0RC32 = getVOPSrc0ForVT<Src0VT, IsTrue16>.ret; | 
 |   field RegisterOperand Src1RC32 = getVregSrcForVT<Src1VT>.ret; | 
 |   field RegisterOperand Src0RC64 = getVOP3SrcForVT<Src0VT>.ret; | 
 |   field RegisterOperand Src1RC64 = getVOP3SrcForVT<Src1VT>.ret; | 
 |   field RegisterOperand Src2RC64 = getVOP3SrcForVT<Src2VT>.ret; | 
 |   field RegisterOperand Src0DPP = getVregSrcForVT<Src0VT>.ret; | 
 |   field RegisterOperand Src1DPP = getVregSrcForVT<Src1VT>.ret; | 
 |   field RegisterOperand Src2DPP = getVregSrcForVT<Src2VT>.ret; | 
 |   field RegisterOperand Src0VOP3DPP = VGPRSrc_32; | 
 |   field RegisterOperand Src1VOP3DPP = getVOP3DPPSrcForVT<Src1VT>.ret; | 
 |   field RegisterOperand Src2VOP3DPP = getVOP3DPPSrcForVT<Src2VT>.ret; | 
 |   field RegisterOperand Src0SDWA = getSDWASrcForVT<Src0VT>.ret; | 
 |   field RegisterOperand Src1SDWA = getSDWASrcForVT<Src0VT>.ret; | 
 |   field Operand Src0Mod = getSrc0Mod<Src0VT, DstVT>.ret; | 
 |   field Operand Src1Mod = getSrcMod<Src1VT>.ret; | 
 |   field Operand Src2Mod = getSrcMod<Src2VT>.ret; | 
 |   field Operand Src0ModDPP = getSrcModDPP<Src0VT>.ret; | 
 |   field Operand Src1ModDPP = getSrcModDPP<Src1VT>.ret; | 
 |   field Operand Src2ModDPP = getSrcModDPP<Src2VT>.ret; | 
 |   field Operand Src0ModVOP3DPP = getSrc0ModVOP3DPP<Src0VT, DstVT>.ret; | 
 |   field Operand Src1ModVOP3DPP = getSrcModVOP3DPP<Src1VT>.ret; | 
 |   field Operand Src2ModVOP3DPP = getSrcModVOP3DPP<Src2VT>.ret; | 
 |   field Operand Src0ModSDWA = getSrcModSDWA<Src0VT>.ret; | 
 |   field Operand Src1ModSDWA = getSrcModSDWA<Src1VT>.ret; | 
 |  | 
 |  | 
 |   field bit IsMAI = 0; | 
 |   field bit IsVOP3P = 0; | 
 |   field bit IsDOT = 0; | 
 |   field bit IsSingle = 0; | 
 |   field bit IsWMMA = 0; | 
 |   field bit IsSWMMAC = 0; | 
 |  | 
 |   field bit HasFP8SrcByteSel = 0; | 
 |   field bit HasFP8DstByteSel = 0; | 
 |   field bit HasFP4DstByteSel = 0; | 
 |   field bit HasFP8ByteSel = !or(HasFP8SrcByteSel, HasFP8DstByteSel); | 
 |   field bit HasBitOp3 = 0; | 
 |  | 
 |   field bit HasDst = !ne(DstVT.Value, untyped.Value); | 
 |   field bit HasDst32 = HasDst; | 
 |   field bit EmitDst = HasDst; // force dst encoding, see v_movreld_b32 special case | 
 |   field bit EmitDstSel = EmitDst; | 
 |   field int NumSrcArgs = getNumSrcArgs<Src0VT, Src1VT, Src2VT>.ret; | 
 |   field bit HasSrc0 = !ne(Src0VT.Value, untyped.Value); | 
 |   field bit HasSrc1 = !ne(Src1VT.Value, untyped.Value); | 
 |   field bit HasSrc2 = !ne(Src2VT.Value, untyped.Value); | 
 |  | 
 |   field bit HasSrc0FloatMods = Src0VT.isFP; | 
 |   field bit HasSrc1FloatMods = Src1VT.isFP; | 
 |   field bit HasSrc2FloatMods = Src2VT.isFP; | 
 |  | 
 |   field bit HasSrc0IntMods = isIntType<Src0VT>.ret; | 
 |   field bit HasSrc1IntMods = isIntType<Src1VT>.ret; | 
 |   field bit HasSrc2IntMods = isIntType<Src2VT>.ret; | 
 |  | 
 |   field bit HasClamp = !or(isModifierType<Src0VT>.ret, EnableClamp); | 
 |   field bit HasSDWAClamp = EmitDst; | 
 |   field bit HasFPClamp = !and(DstVT.isFP, HasClamp); | 
 |   field bit HasIntClamp = !if(DstVT.isFP, 0, HasClamp); | 
 |   field bit HasClampLo = HasClamp; | 
 |   field bit HasClampHi = !and(DstVT.isVector, HasClamp); | 
 |   field bit HasHigh = 0; | 
 |  | 
 |   field bit IsPacked = Src0VT.isVector; | 
 |   field bit HasOpSel = IsPacked; | 
 |   field bit HasOMod = !if(IsVOP3P, 0, DstVT.isFP); | 
 |   field bit HasSDWAOMod = DstVT.isFP; | 
 |  | 
 |   field bit HasModifiers = !or(isModifierType<Src0VT>.ret, | 
 |                                isModifierType<Src1VT>.ret, | 
 |                                isModifierType<Src2VT>.ret, | 
 |                                HasOMod); | 
 |   field bit HasNeg = HasModifiers; | 
 |  | 
 |   field bit HasSrc0Mods = HasModifiers; | 
 |   field bit HasSrc1Mods = !if(HasModifiers, !or(HasSrc1FloatMods, HasSrc1IntMods), 0); | 
 |   field bit HasSrc2Mods = !if(HasModifiers, !or(HasSrc2FloatMods, HasSrc2IntMods), 0); | 
 |  | 
 |   field bit HasExt = getHasExt<NumSrcArgs, DstVT, Src0VT, Src1VT>.ret; | 
 |   field bit HasExtVOP3DPP = getHasVOP3DPP<DstVT, Src0VT, Src1VT, Src2VT>.ret; | 
 |   field bit HasExtDPP = !or(getHasDPP<NumSrcArgs>.ret, HasExtVOP3DPP); | 
 |   field bit HasExt32BitDPP = getHasExt32BitDPP<NumSrcArgs, DstVT, Src0VT, Src1VT>.ret; | 
 |   field bit HasExt64BitDPP = getHasExt64BitDPP<NumSrcArgs, DstVT, Src0VT, Src1VT>.ret; | 
 |   field bit HasExtSDWA = getHasSDWA<NumSrcArgs, DstVT, Src0VT, Src1VT>.ret; | 
 |   field bit HasExtSDWA9 = HasExtSDWA; | 
 |   field int NeedPatGen = PatGenMode.NoPattern; | 
 |  | 
 |   field Operand Src0PackedMod = !if(HasSrc0FloatMods, PackedF16InputMods, PackedI16InputMods); | 
 |   field Operand Src1PackedMod = !if(HasSrc1FloatMods, PackedF16InputMods, PackedI16InputMods); | 
 |   field Operand Src2PackedMod = !if(HasSrc2FloatMods, PackedF16InputMods, PackedI16InputMods); | 
 |  | 
 |   field dag Outs = !if(HasDst,(outs DstRC:$vdst),(outs)); | 
 |  | 
 |   // VOP3b instructions are a special case with a second explicit | 
 |   // output. This is manually overridden for them. | 
 |   field dag Outs32 = Outs; | 
 |   field dag Outs64 = !if(HasDst,(outs DstRC64:$vdst),(outs)); | 
 |   field dag OutsDPP = getOutsDPP<HasDst, DstVT, DstRCDPP>.ret; | 
 |   field dag OutsDPP8 = OutsDPP; | 
 |   field dag OutsVOP3DPP = getOutsDPP<HasDst, DstVT, DstRCVOP3DPP>.ret; | 
 |   field dag OutsVOP3DPP8 = OutsVOP3DPP; | 
 |   field dag OutsSDWA = getOutsSDWA<HasDst, DstVT, DstRCSDWA>.ret; | 
 |  | 
 |   field dag Ins32 = getIns32<Src0RC32, Src1RC32, NumSrcArgs>.ret; | 
 |   field dag Ins64 = getIns64<Src0RC64, Src1RC64, Src2RC64, NumSrcArgs, | 
 |                              HasClamp, HasModifiers, HasSrc2Mods, | 
 |                              HasOMod, Src0Mod, Src1Mod, Src2Mod, | 
 |                              HasFP8ByteSel, HasFP8DstByteSel, HasBitOp3>.ret; | 
 |   field dag InsVOP3P = getInsVOP3P<Src0RC64, Src1RC64, Src2RC64, | 
 |                                    NumSrcArgs, HasClamp, HasOpSel, HasNeg, | 
 |                                    Src0PackedMod, Src1PackedMod, Src2PackedMod>.ret; | 
 |   field dag InsVOP3OpSel = getInsVOP3OpSel<Src0RC64, Src1RC64, Src2RC64, | 
 |                                 NumSrcArgs, HasClamp, HasOMod, | 
 |                                 Src0Mod, Src1Mod, Src2Mod, | 
 |                                 HasFP8ByteSel, HasFP8DstByteSel, HasBitOp3>.ret; | 
 |   field dag InsDPP = !if(HasExtDPP, | 
 |                          getInsDPP<DstRCDPP, Src0DPP, Src1DPP, Src2DPP, NumSrcArgs, | 
 |                                    HasModifiers, Src0ModDPP, Src1ModDPP, Src2ModDPP>.ret, | 
 |                          (ins)); | 
 |   field dag InsDPP16 = getInsDPP16<DstRCDPP, Src0DPP, Src1DPP, Src2DPP, NumSrcArgs, | 
 |                                    HasModifiers, Src0ModDPP, Src1ModDPP, Src2ModDPP>.ret; | 
 |   field dag InsDPP8 = getInsDPP8<DstRCDPP, Src0DPP, Src1DPP, Src2DPP, | 
 |                                  NumSrcArgs, HasModifiers, | 
 |                                  Src0ModDPP, Src1ModDPP, Src2ModDPP>.ret; | 
 |   defvar InsVOP3DPPBase = getInsVOP3Base<Src0VOP3DPP, Src1VOP3DPP, | 
 |                   Src2VOP3DPP, NumSrcArgs, HasClamp, HasModifiers, HasSrc2Mods, HasOMod, | 
 |                   Src0ModVOP3DPP, Src1ModVOP3DPP, Src2ModVOP3DPP, HasOpSel, | 
 |                   HasFP8ByteSel, HasFP8DstByteSel, HasBitOp3>.ret; | 
 |   defvar InsVOP3PDPPBase = getInsVOP3P<Src0VOP3DPP, Src1VOP3DPP, | 
 |                   Src2VOP3DPP, NumSrcArgs, HasClamp, HasOpSel, HasNeg, | 
 |                   Src0ModVOP3DPP, Src1ModVOP3DPP, Src2ModVOP3DPP>.ret; | 
 |  | 
 |   field dag InsVOP3Base = !if(IsVOP3P, InsVOP3PDPPBase, InsVOP3DPPBase); | 
 |  | 
 |   field dag InsVOP3DPP = getInsVOP3DPP<InsVOP3Base, DstRCVOP3DPP, NumSrcArgs>.ret; | 
 |   field dag InsVOP3DPP16 = getInsVOP3DPP16<InsVOP3Base, DstRCVOP3DPP, NumSrcArgs>.ret; | 
 |   field dag InsVOP3DPP8 = getInsVOP3DPP8<InsVOP3Base, DstRCVOP3DPP, NumSrcArgs>.ret; | 
 |   field dag InsSDWA = getInsSDWA<Src0SDWA, Src1SDWA, NumSrcArgs, | 
 |                                  HasSDWAOMod, Src0ModSDWA, Src1ModSDWA, | 
 |                                  DstVT>.ret; | 
 |   field dag InsVOPDX = (ins Src0RC32:$src0X, Src1RC32:$vsrc1X); | 
 |   // It is a slight misnomer to use the f32 operand type for non-float | 
 |   // operands, but this operand type will only be used if the other dual | 
 |   // component is FMAAK or FMAMK | 
 |   field dag InsVOPDX_immX = (ins !if(!eq(Src0VT.Size, 32), VSrc_f32, VSrc_f16):$src0X, VGPR_32:$vsrc1X); | 
 |   field dag InsVOPDY = (ins Src0RC32:$src0Y, Src1RC32:$vsrc1Y); | 
 |  | 
 |   field string Asm32 = getAsm32<HasDst, NumSrcArgs, DstVT>.ret; | 
 |   field string AsmDPP = !if(HasExtDPP, | 
 |                             getAsmDPP<HasDst, NumSrcArgs, HasModifiers, DstVT>.ret, ""); | 
 |   field string AsmDPP16 = getAsmDPP16<HasDst, NumSrcArgs, HasModifiers, DstVT>.ret; | 
 |   // DPP8 encoding has no fields for modifiers, and it is enforced by setting | 
 |   // the asm operand name via this HasModifiers flag | 
 |   field string AsmDPP8 = getAsmDPP8<HasDst, NumSrcArgs, 0 /*HasModifiers*/, DstVT>.ret; | 
 |   field string AsmVOP3Base = getAsmVOP3Base<NumSrcArgs, HasDst, HasClamp, | 
 |    HasOpSel, HasOMod, IsVOP3P, HasNeg, HasSrc0FloatMods, HasSrc1FloatMods, | 
 |    HasSrc2FloatMods, DstVT, HasFP8ByteSel, HasBitOp3>.ret; | 
 |   field string Asm64 = AsmVOP3Base; | 
 |   field string AsmVOP3P = getAsmVOP3P<HasDst, NumSrcArgs, HasNeg, HasClamp, HasOpSel>.ret; | 
 |   field string AsmVOP3DPP = getAsmVOP3DPP<AsmVOP3Base>.ret; | 
 |   field string AsmVOP3DPP16 = getAsmVOP3DPP16<AsmVOP3Base>.ret; | 
 |   field string AsmVOP3DPP8 = getAsmVOP3DPP8<AsmVOP3Base>.ret; | 
 |   field string AsmSDWA = getAsmSDWA<HasDst, NumSrcArgs, DstVT>.ret; | 
 |   field string AsmSDWA9 = getAsmSDWA9<HasDst, HasSDWAOMod, NumSrcArgs, DstVT>.ret; | 
 |   field string AsmVOPDX = getAsmVOPDPart<NumSrcArgs, "X">.ret; | 
 |   field string AsmVOPDY = getAsmVOPDPart<NumSrcArgs, "Y">.ret; | 
 |   field string TieRegDPP = "$old"; | 
 |   field bit IsSMFMAC = false; | 
 |   field bit HasAbid = !and(IsMAI, HasSrc1); | 
 | } | 
 |  | 
 |   class VOP_NO_EXT <VOPProfile p> : VOPProfile <p.ArgVT> { | 
 |   let HasExt = 0; | 
 |   let HasExtDPP = 0; | 
 |   let HasExtVOP3DPP = 0; | 
 |   let HasExt32BitDPP = 0; | 
 |   let HasExt64BitDPP = 0; | 
 |   let HasExtSDWA = 0; | 
 |   let HasExtSDWA9 = 0; | 
 | } | 
 |  | 
 | class VOP_PAT_GEN <VOPProfile p, int mode=PatGenMode.NoPattern> : VOPProfile <p.ArgVT> { | 
 |   let NeedPatGen = mode; | 
 | } | 
 |  | 
 | // VOPC_Profile_t16, VOPC_NoSdst_Profile_t16, VOPC_Class_Profile_t16, | 
 | // VOPC_Class_NoSdst_Profile_t16, and  VOP_MAC_F16_t16 do not inherit from this | 
 | // class, so copy changes to this class in those profiles | 
 | class VOPProfile_True16<VOPProfile P> : VOPProfile<P.ArgVT> { | 
 |   let IsTrue16 = 1; | 
 |   let IsRealTrue16 = 1; | 
 |  | 
 |   let HasOpSel = 1; | 
 |   let HasModifiers = 1; // All instructions at least have OpSel. | 
 |  | 
 |   // Most DstVT are 16-bit, but not all. | 
 |   let DstRC = getVALUDstForVT<DstVT, 1 /*IsTrue16*/, 0 /*IsVOP3Encoding*/>.ret; | 
 |   let Src0RC32 = getVOPSrc0ForVT<Src0VT, 1 /*IsTrue16*/, 0 /*IsFake16*/>.ret; | 
 |   let Src1RC32 = getVregSrcForVT<Src1VT, 1 /*IsTrue16*/, 0 /*IsFake16*/>.ret; | 
 |   let Src0DPP = getVregSrcForVT<Src0VT, 1 /*IsTrue16*/, 0 /*IsFake16*/>.ret; | 
 |   let Src1DPP = getVregSrcForVT<Src1VT, 1 /*IsTrue16*/, 0 /*IsFake16*/>.ret; | 
 |   let Src2DPP = getVregSrcForVT<Src2VT, 1 /*IsTrue16*/, 0 /*IsFake16*/>.ret; | 
 |   let Src0ModDPP = getSrcModDPP_t16<Src0VT, 0 /*IsFake16*/>.ret; | 
 |   let Src1ModDPP = getSrcModDPP_t16<Src1VT, 0 /*IsFake16*/>.ret; | 
 |   let Src2ModDPP = getSrcModDPP_t16<Src2VT, 0 /*IsFake16*/>.ret; | 
 |   let Src0VOP3DPP = !if (!eq(Src0VT.Size, 16), VGPRSrc_16, VGPRSrc_32); | 
 |   let Src1VOP3DPP = getVOP3DPPSrcForVT<Src1VT, 0 /*IsFake16*/>.ret; | 
 |   let Src2VOP3DPP = getVOP3DPPSrcForVT<Src2VT, 0 /*IsFake16*/>.ret; | 
 |   let Src0ModVOP3DPP = getSrc0ModVOP3DPP<Src0VT, DstVT, 0/*IsFake16*/>.ret; | 
 |   let Src1ModVOP3DPP = getSrcModVOP3DPP<Src1VT, 0 /*IsFake16*/>.ret; | 
 |   let Src2ModVOP3DPP = getSrcModVOP3DPP<Src2VT, 0 /*IsFake16*/>.ret; | 
 |  | 
 |   let DstRC64 = getVALUDstForVT<DstVT, 1 /*IsTrue16*/, 1 /*IsVOP3Encoding*/>.ret; | 
 |   let Src0RC64 = getVOP3SrcForVT<Src0VT, 1 /*IsTrue16*/>.ret; | 
 |   let Src1RC64 = getVOP3SrcForVT<Src1VT, 1 /*IsTrue16*/>.ret; | 
 |   let Src2RC64 = getVOP3SrcForVT<Src2VT, 1 /*IsTrue16*/>.ret; | 
 |   let Src0Mod = getSrc0Mod<Src0VT, DstVT, 1/*IsTrue16*/, 0/*IsFake16*/>.ret; | 
 |   let Src1Mod = getSrcMod<Src1VT, 1 /*IsTrue16*/, 0/*IsFake16*/>.ret; | 
 |   let Src2Mod = getSrcMod<Src2VT, 1 /*IsTrue16*/, 0/*IsFake16*/>.ret; | 
 | } | 
 |  | 
 | class VOPProfile_Fake16<VOPProfile P> : VOPProfile<P.ArgVT> { | 
 |   let IsTrue16 = 1; | 
 |   // Most DstVT are 16-bit, but not all | 
 |   let DstRC = getVALUDstForVT_fake16<DstVT>.ret; | 
 |   let DstRC64 = getVALUDstForVT<DstVT>.ret; | 
 |   let Src1RC32 = getVregSrcForVT<Src1VT, 1/*IsTrue16*/, 1/*IsFake16*/>.ret; | 
 |   let Src0DPP = getVregSrcForVT<Src0VT, 1/*IsTrue16*/, 1/*IsFake16*/>.ret; | 
 |   let Src1DPP = getVregSrcForVT<Src1VT, 1/*IsTrue16*/, 1/*IsFake16*/>.ret; | 
 |   let Src2DPP = getVregSrcForVT<Src2VT, 1/*IsTrue16*/, 1/*IsFake16*/>.ret; | 
 |   let Src0ModDPP = getSrcModDPP_t16<Src0VT, 1/*IsFake16*/>.ret; | 
 |   let Src1ModDPP = getSrcModDPP_t16<Src1VT, 1/*IsFake16*/>.ret; | 
 |   let Src2ModDPP = getSrcModDPP_t16<Src2VT, 1/*IsFake16*/>.ret; | 
 |   let Src0Mod = getSrc0Mod<Src0VT, DstVT, 1/*IsTrue16*/, 1/*IsFake16*/>.ret; | 
 |   let Src1Mod = getSrcMod<Src1VT, 1 /*IsTrue16*/, 1/*IsFake16*/>.ret; | 
 |   let Src2Mod = getSrcMod<Src2VT, 1 /*IsTrue16*/, 1/*IsFake16*/>.ret; | 
 |   let Src1VOP3DPP = getVOP3DPPSrcForVT<Src1VT, 1 /*IsFake16*/>.ret; | 
 |   let Src2VOP3DPP = getVOP3DPPSrcForVT<Src2VT, 1 /*IsFake16*/>.ret; | 
 |   let Src0ModVOP3DPP = getSrc0ModVOP3DPP<Src0VT, DstVT, 1/*IsFake16*/>.ret; | 
 |   let Src1ModVOP3DPP = getSrcModVOP3DPP<Src1VT, 1/*IsFake16*/>.ret; | 
 |   let Src2ModVOP3DPP = getSrcModVOP3DPP<Src2VT, 1/*IsFake16*/>.ret; | 
 | } | 
 |  | 
 | def VOP_F16_F16 : VOPProfile<[f16, f16, untyped, untyped]>; | 
 | def VOP_F16_I16 : VOPProfile <[f16, i16, untyped, untyped]>; | 
 | def VOP_I16_F16 : VOPProfile <[i16, f16, untyped, untyped]>; | 
 | def VOP_I16_I16 : VOPProfile <[i16, i16, untyped, untyped]>; | 
 |  | 
 | def VOP_F16_F16_F16 : VOPProfile <[f16, f16, f16, untyped]>; | 
 | def VOP_F16_F16_I16 : VOPProfile <[f16, f16, i16, untyped]>; | 
 | def VOP_F16_F16_I32 : VOPProfile <[f16, f16, i32, untyped]>; | 
 | def VOP_I16_I16_I16 : VOPProfile <[i16, i16, i16, untyped]>; | 
 | def VOP_I16_I16_I16_ARITH : VOPProfile <[i16, i16, i16, untyped], /*EnableClamp=*/1>; | 
 |  | 
 | def VOP_I16_I16_I16_I16 : VOPProfile <[i16, i16, i16, i16, untyped]>; | 
 | def VOP_F16_F16_F16_F16 : VOPProfile <[f16, f16, f16, f16, untyped]>; | 
 |  | 
 | def VOP_I32_I16_I16_I32 : VOPProfile <[i32, i16, i16, i32, untyped]>; | 
 | def VOP_I32_I16 : VOPProfile <[i32, i16, untyped, untyped]>; | 
 | def VOP_I16_I32 : VOPProfile <[i16, i32, untyped, untyped]>; | 
 |  | 
 | def VOP_V2F16_V2F16_V2F16 : VOPProfile <[v2f16, v2f16, v2f16, untyped]>; | 
 | def VOP_V2I16_V2I16_V2I16 : VOPProfile <[v2i16, v2i16, v2i16, untyped]>; | 
 | def VOP_B32_F16_F16 : VOPProfile <[i32, f16, f16, untyped]>; | 
 |  | 
 | def VOP_V2F16_V2F16_V2F16_V2F16 : VOPProfile <[v2f16, v2f16, v2f16, v2f16]>; | 
 | def VOP_V2I16_V2I16_V2I16_V2I16 : VOPProfile <[v2i16, v2i16, v2i16, v2i16]>; | 
 | def VOP_V2I16_F32_F32 : VOPProfile <[v2i16, f32, f32, untyped]>; | 
 | def VOP_V2I16_I32_I32 : VOPProfile <[v2i16, i32, i32, untyped]>; | 
 |  | 
 | def VOP_F16_V2F16_V2F16_F16 : VOPProfile <[f16, v2f16, v2f16, f16]>; | 
 | def VOP_BF16_V2BF16_V2BF16_BF16: VOPProfile <[bf16, v2bf16, v2bf16, bf16]>; | 
 | def VOP_F32_V2BF16_V2BF16_F32 : VOPProfile <[f32, v2bf16, v2bf16, f32]>; | 
 |  | 
 | def VOP_F32_V2F16_V2F16_V2F16 : VOPProfile <[f32, v2f16, v2f16, v2f16]>; | 
 |  | 
 | def VOP_NONE : VOPProfile <[untyped, untyped, untyped, untyped]>; | 
 |  | 
 | def VOP_F32_F32 : VOPProfile <[f32, f32, untyped, untyped]>; | 
 | def VOP_F32_F64 : VOPProfile <[f32, f64, untyped, untyped]>; | 
 | def VOP_F32_I32 : VOPProfile <[f32, i32, untyped, untyped]>; | 
 | def VOP_F64_F32 : VOPProfile <[f64, f32, untyped, untyped]>; | 
 | def VOP_F64_F64 : VOPProfile <[f64, f64, untyped, untyped]>; | 
 | def VOP_F64_I32 : VOPProfile <[f64, i32, untyped, untyped]>; | 
 | def VOP_I32_F32 : VOPProfile <[i32, f32, untyped, untyped]>; | 
 | def VOP_I32_F64 : VOPProfile <[i32, f64, untyped, untyped]>; | 
 | def VOP_I32_I32 : VOPProfile <[i32, i32, untyped, untyped]>; | 
 | def VOP_F16_F32 : VOPProfile <[f16, f32, untyped, untyped]>; | 
 | def VOP_F32_F16 : VOPProfile <[f32, f16, untyped, untyped]>; | 
 | def VOP_I64_I64 : VOPProfile <[i64, i64, untyped, untyped]>; | 
 | def VOP_F32_BF16 : VOPProfile <[f32, bf16, untyped, untyped]>; | 
 |  | 
 | def VOP_F32_F32_F16 : VOPProfile <[f32, f32, f16, untyped]>; | 
 | def VOP_F32_F32_F32 : VOPProfile <[f32, f32, f32, untyped]>; | 
 | def VOP_F32_F32_I32 : VOPProfile <[f32, f32, i32, untyped]>; | 
 | def VOP_F64_F64_F64 : VOPProfile <[f64, f64, f64, untyped]>; | 
 | def VOP_F64_F64_I32 : VOPProfile <[f64, f64, i32, untyped]>; | 
 | def VOP_I32_F32_F32 : VOPProfile <[i32, f32, f32, untyped]>; | 
 | def VOP_I32_F32_I32 : VOPProfile <[i32, f32, i32, untyped]>; | 
 | def VOP_I32_I32_I32 : VOPProfile <[i32, i32, i32, untyped]>; | 
 | def VOP_I16_F32_F32 : VOPProfile <[i16, f32, f32, untyped]>; | 
 | def VOP_I32_I32_I32_ARITH : VOPProfile <[i32, i32, i32, untyped], /*EnableClamp=*/1>; | 
 | def VOP_V2F16_F32_F32 : VOPProfile <[v2f16, f32, f32, untyped]>; | 
 | def VOP_F32_F16_F16_F16 : VOPProfile <[f32, f16, f16, f16]>; | 
 | def VOP_V2BF16_F32_F32 : VOPProfile <[v2bf16, f32, f32, untyped]>; | 
 | def VOP_V32F32_V6I32_F32 : VOPProfile <[v32f32, v6i32, f32, untyped]>; | 
 | def VOP_V32F16_V6I32_F32 : VOPProfile <[v32f16, v6i32, f32, untyped]>; | 
 | def VOP_V32BF16_V6I32_F32 : VOPProfile <[v32bf16, v6i32, f32, untyped]>; | 
 | def VOP_V6I32_V32F16_F32 : VOPProfile<[v6i32, v32f16, f32, untyped]>; | 
 | def VOP_V6I32_V32BF16_F32 : VOPProfile<[v6i32, v32bf16, f32, untyped]>; | 
 | def VOP_V6I32_V16F32_V16F32_F32 : VOPProfile<[v6i32, v16f32, v16f32, f32]>; | 
 | def VOP_V2F16_I32_F32 : VOPProfile<[v2f16, i32, f32, untyped]>; | 
 | def VOP_V2I16_F32_F32_F32 : VOPProfile<[v2i16, f32, f32, f32]>; | 
 | def VOP_V2I16_V2F16_F32 : VOPProfile<[v2i16, v2f16, f32, untyped]>; | 
 | def VOP_V2I16_V2BF16_F32 : VOPProfile<[v2i16, v2bf16, f32, untyped]>; | 
 | def VOP_I32_F32_F32_F32 : VOPProfile<[i32, f32, f32, f32]>; | 
 | def VOP_I32_V2F32_I32_F32 : VOPProfile<[i32, v2f32, i32, f32]>; | 
 | def VOP_I32_V2F16_F32_F32 : VOPProfile<[i32, v2f16, f32, f32]>; | 
 | def VOP_I32_V2BF16_F32_F32: VOPProfile<[i32, v2bf16, f32, f32]>; | 
 | def VOP_BF16_F32_I32 : VOPProfile<[bf16, f32, i32, untyped]>; | 
 | def VOP_F16_F32_I32 : VOPProfile<[f16, f32, i32, untyped]>; | 
 | def VOP_I32_BF16_I32_F32 : VOPProfile<[i32, bf16, i32, f32]>; | 
 | def VOP_I32_F16_I32_F32 : VOPProfile<[i32, f16, i32, f32]>; | 
 | def VOP_I32_F32_I32_F32 : VOPProfile<[i32, f32, i32, f32]>; | 
 |  | 
 | def VOP_V6I32_V32BF16_I32_F32 : VOPProfile<[v6i32, v32bf16, i32, f32]>; | 
 | def VOP_V6I32_V32F16_I32_F32 : VOPProfile<[v6i32, v32f16, i32, f32]>; | 
 | def VOP_V6I32_V32F32_I32_F32 : VOPProfile<[v6i32, v32f32, i32, f32]>; | 
 |  | 
 | def VOP_I64_I64_I32 : VOPProfile <[i64, i64, i32, untyped]>; | 
 | def VOP_I64_I32_I64 : VOPProfile <[i64, i32, i64, untyped]>; | 
 | def VOP_I64_I64_I64 : VOPProfile <[i64, i64, i64, untyped]>; | 
 |  | 
 | def VOP_F16_F32_F16_F32 : VOPProfile <[f16, f32, f16, f32]>; | 
 | def VOP_F32_F32_F16_F16 : VOPProfile <[f32, f32, f16, f16]>; | 
 | def VOP_F32_F32_F32_F32 : VOPProfile <[f32, f32, f32, f32]>; | 
 | def VOP_F64_F64_F64_F64 : VOPProfile <[f64, f64, f64, f64]>; | 
 | def VOP_I32_I32_I32_I32 : VOPProfile <[i32, i32, i32, i32]>; | 
 | def VOP_I32_I32_I32_I16 : VOPProfile <[i32, i32, i32, i16]>; | 
 | def VOP_I64_I32_I32_I64 : VOPProfile <[i64, i32, i32, i64]>; | 
 | def VOP_I32_F32_I32_I32 : VOPProfile <[i32, f32, i32, i32]>; | 
 | def VOP_I64_I64_I32_I64 : VOPProfile <[i64, i64, i32, i64]>; | 
 | def VOP_V4I32_I64_I32_V4I32 : VOPProfile <[v4i32, i64, i32, v4i32]>; | 
 | def VOP_I16_I32_I32_I32 : VOPProfile <[i16, i32, i32, i32]>; | 
 |  | 
 | def VOP_F32_V2F16_V2F16_F32 : VOPProfile <[f32, v2f16, v2f16, f32]>; | 
 | def VOP_I32_V2I16_V2I16_I32 : VOPProfile <[i32, v2i16, v2i16, i32]>; | 
 |  | 
 | def VOP_V4F32_F32_F32_V4F32       : VOPProfile <[v4f32,  f32,   f32,   v4f32]>; | 
 | def VOP_V16F32_F32_F32_V16F32     : VOPProfile <[v16f32, f32,   f32,   v16f32]>; | 
 | def VOP_V32F32_F32_F32_V32F32     : VOPProfile <[v32f32, f32,   f32,   v32f32]>; | 
 | def VOP_V4F32_V4F16_V4F16_V4F32   : VOPProfile <[v4f32,  v4f16, v4f16, v4f32]>; | 
 | def VOP_V16F32_V4F16_V4F16_V16F32 : VOPProfile <[v16f32, v4f16, v4f16, v16f32]>; | 
 | def VOP_V32F32_V4F16_V4F16_V32F32 : VOPProfile <[v32f32, v4f16, v4f16, v32f32]>; | 
 | def VOP_V4F32_V2I16_V2I16_V4F32   : VOPProfile <[v4f32,  v2i16, v2i16, v4f32]>; | 
 | def VOP_V16F32_V2I16_V2I16_V16F32 : VOPProfile <[v16f32, v2i16, v2i16, v16f32]>; | 
 | def VOP_V32F32_V2I16_V2I16_V32F32 : VOPProfile <[v32f32, v2i16, v2i16, v32f32]>; | 
 | def VOP_V4I32_I32_I32_V4I32       : VOPProfile <[v4i32,  i32,   i32,   v4i32]>; | 
 | def VOP_V16I32_I32_I32_V16I32     : VOPProfile <[v16i32, i32,   i32,   v16i32]>; | 
 | def VOP_V32I32_I32_I32_V32I32     : VOPProfile <[v32i32, i32,   i32,   v32i32]>; | 
 |  | 
 | def VOP_V4F64_F64_F64_V4F64 : VOPProfile <[v4f64, f64, f64, v4f64]>; | 
 | def VOP_V1F64_F64_F64_V1F64 : VOPProfile <[v1f64, f64, f64, v1f64]>; | 
 |  | 
 | def VOP_V2F32_V2F32_V2F32_V2F32   : VOPProfile <[v2f32,  v2f32, v2f32, v2f32]>; | 
 | def VOP_V2F32_V2F32_V2F32         : VOPProfile <[v2f32,  v2f32, v2f32, untyped]>; | 
 | def VOP_V2I32_V2I32_V2I32         : VOPProfile <[v2i32,  v2i32, v2i32, untyped]>; | 
 | def VOP_V4F32_V4I16_V4I16_V4F32   : VOPProfile <[v4f32,  v4i16, v4i16, v4f32]>; | 
 | def VOP_V16F32_V4I16_V4I16_V16F32 : VOPProfile <[v16f32, v4i16, v4i16, v16f32]>; | 
 | def VOP_V32F32_V4I16_V4I16_V32F32 : VOPProfile <[v32f32, v4i16, v4i16, v32f32]>; | 
 |  | 
 | def VOP_V4I32_I64_I64_V4I32       : VOPProfile <[v4i32,  i64,   i64,   v4i32]>; | 
 | def VOP_V16I32_I64_I64_V16I32     : VOPProfile <[v16i32, i64,   i64,   v16i32]>; | 
 | def VOP_V4F32_V2F32_V2F32_V4F32   : VOPProfile <[v4f32,  v2f32, v2f32, v4f32]>; | 
 | def VOP_V16F32_V2F32_V2F32_V16F32 : VOPProfile <[v16f32, v2f32, v2f32, v16f32]>; | 
 | def VOP_V4F32_I64_I64_V4F32       : VOPProfile <[v4f32,  i64,   i64,   v4f32]>; | 
 | def VOP_V16F32_I64_I64_V16F32     : VOPProfile <[v16f32, i64,   i64,   v16f32]>; | 
 |  | 
 | def VOP_V4F32_V4F16_V8F16_I32     : VOPProfile <[v4f32,  v4f16, v8f16, i32]>; | 
 | def VOP_V4F32_V8F16_V16F16_I32    : VOPProfile <[v4f32,  v8f16, v16f16, i32]>; | 
 | def VOP_V4F32_V8BF16_V16BF16_I32  : VOPProfile <[v4f32,  v8bf16, v16bf16, i32]>; | 
 | def VOP_V16F32_V4F16_V8F16_I32    : VOPProfile <[v16f32, v4f16, v8f16, i32]>; | 
 | def VOP_V16F32_V8F16_V16F16_I32   : VOPProfile <[v16f32, v8f16, v16f16, i32]>; | 
 | def VOP_V16F32_V8BF16_V16BF16_I32 : VOPProfile <[v16f32, v8bf16, v16bf16, i32]>; | 
 | def VOP_V4F32_V4I16_V8I16_I32     : VOPProfile <[v4f32,  v4i16, v8i16, i32]>; | 
 | def VOP_V16F32_V4I16_V8I16_I32    : VOPProfile <[v16f32, v4i16, v8i16, i32]>; | 
 | def VOP_V4I32_V2I32_V4I32_I32     : VOPProfile <[v4i32,  v2i32, v4i32, i32]>; | 
 | def VOP_V16I32_V2I32_V4I32_I32    : VOPProfile <[v16i32, v2i32, v4i32, i32]>; | 
 | def VOP_V4F32_V2I32_V4I32_I32     : VOPProfile <[v4f32,  v2i32, v4i32, i32]>; | 
 | def VOP_V16F32_V2I32_V4I32_I32    : VOPProfile <[v16f32, v2i32, v4i32, i32]>; | 
 | def VOP_V4I32_V4I32_V8I32_I32     : VOPProfile <[v4i32,  v4i32, v8i32, i32]>; | 
 | def VOP_V16I32_V4I32_V8I32_I32    : VOPProfile <[v16i32, v4i32, v8i32, i32]>; | 
 | def VOP_V4F32_V4I32_V8I32_I32     : VOPProfile <[v4f32, v4i32, v8i32, i32]>; | 
 | def VOP_V16F32_V4I32_V8I32_I32    : VOPProfile <[v16f32, v4i32, v8i32, i32]>; | 
 |  | 
 | def VOP_V4F32_V8F16_V8F16_V4F32   : VOPProfile <[v4f32,  v8f16, v8f16, v4f32]>; | 
 | def VOP_V16F32_V8F16_V8F16_V16F32 : VOPProfile <[v16f32, v8f16, v8f16, v16f32]>; | 
 | def VOP_V16F32_V8BF16_V8BF16_V16F32 : VOPProfile <[v16f32, v8bf16, v8bf16, v16f32]>; | 
 | def VOP_V4F32_V8BF16_V8BF16_V4F32 : VOPProfile <[v4f32, v8bf16, v8bf16, v4f32]>; | 
 | def VOP_V4F32_V8I32_V8I32_V4F32   : VOPProfile <[v4f32,  v8i32, v8i32, v4f32]>; | 
 |  | 
 | def VOP_V4F32_V8I32_V6I32_V4F32   : VOPProfile <[v4f32,  v8i32, v6i32, v4f32]>; | 
 | def VOP_V4F32_V6I32_V8I32_V4F32   : VOPProfile <[v4f32,  v6i32, v8i32, v4f32]>; | 
 | def VOP_V4F32_V6I32_V6I32_V4F32   : VOPProfile <[v4f32,  v6i32, v6i32, v4f32]>; | 
 |  | 
 | def VOP_V4F32_V8I32_V4I32_V4F32   : VOPProfile <[v4f32,  v8i32, v4i32, v4f32]>; | 
 | def VOP_V4F32_V4I32_V8I32_V4F32   : VOPProfile <[v4f32,  v4i32, v8i32, v4f32]>; | 
 | def VOP_V4F32_V6I32_V4I32_V4F32   : VOPProfile <[v4f32,  v6i32, v4i32, v4f32]>; | 
 | def VOP_V4F32_V4I32_V6I32_V4F32   : VOPProfile <[v4f32,  v4i32, v6i32, v4f32]>; | 
 | def VOP_V4F32_V4I32_V4I32_V4F32   : VOPProfile <[v4f32,  v4i32, v4i32, v4f32]>; | 
 |  | 
 | def VOP_V16F32_V8I32_V8I32_V16F32   : VOPProfile <[v16f32, v8i32, v8i32, v16f32]>; | 
 | def VOP_V16F32_V8I32_V6I32_V16F32   : VOPProfile <[v16f32, v8i32, v6i32, v16f32]>; | 
 | def VOP_V16F32_V6I32_V8I32_V16F32   : VOPProfile <[v16f32, v6i32, v8i32, v16f32]>; | 
 | def VOP_V16F32_V6I32_V6I32_V16F32   : VOPProfile <[v16f32, v6i32, v6i32, v16f32]>; | 
 |  | 
 | def VOP_V16F32_V8I32_V4I32_V16F32   : VOPProfile <[v16f32, v8i32, v4i32, v16f32]>; | 
 | def VOP_V16F32_V4I32_V8I32_V16F32   : VOPProfile <[v16f32, v4i32, v8i32, v16f32]>; | 
 | def VOP_V16F32_V6I32_V4I32_V16F32   : VOPProfile <[v16f32, v6i32, v4i32, v16f32]>; | 
 | def VOP_V16F32_V4I32_V6I32_V16F32   : VOPProfile <[v16f32, v4i32, v6i32, v16f32]>; | 
 | def VOP_V16F32_V4I32_V4I32_V16F32   : VOPProfile <[v16f32, v4i32, v4i32, v16f32]>; | 
 |  | 
 | def VOP_V4I32_V4I32_V4I32_V4I32     : VOPProfile <[v4i32,  v4i32, v4i32, v4i32]>; | 
 | def VOP_V16I32_V4I32_V4I32_V16I32   : VOPProfile <[v16i32,  v4i32, v4i32, v16i32]>; | 
 |  | 
 |  | 
 | class Commutable_REV <string revOp, bit isOrig> { | 
 |   string RevOp = revOp; | 
 |   bit IsOrig = isOrig; | 
 | } | 
 |  | 
 | //===----------------------------------------------------------------------===// | 
 | // Interpolation opcodes | 
 | //===----------------------------------------------------------------------===// | 
 |  | 
 | class VINTRPDstOperand <RegisterClass rc> : RegisterOperand <rc, "printVINTRPDst">; | 
 |  | 
 | class VINTRP_Pseudo <string opName, dag outs, dag ins, list<dag> pattern> : | 
 |   VINTRPCommon <outs, ins, "", pattern>, | 
 |   SIMCInstr<opName, SIEncodingFamily.NONE> { | 
 |   let isPseudo = 1; | 
 |   let isCodeGenOnly = 1; | 
 | } | 
 |  | 
 | // FIXME-GFX10: WIP. | 
 | class VINTRP_Real_si <bits <2> op, string opName, dag outs, dag ins, | 
 |                       string asm, int encodingFamily> : | 
 |   VINTRPCommon <outs, ins, asm, []>, | 
 |   VINTRPe <op>, | 
 |   SIMCInstr<opName, encodingFamily> { | 
 | } | 
 |  | 
 | class VINTRP_Real_vi <bits <2> op, string opName, dag outs, dag ins, | 
 |                       string asm> : | 
 |   VINTRPCommon <outs, ins, asm, []>, | 
 |   VINTRPe_vi <op>, | 
 |   SIMCInstr<opName, SIEncodingFamily.VI> { | 
 |   let AssemblerPredicate = isGFX8GFX9; | 
 |   let DecoderNamespace = "GFX8"; | 
 | } | 
 |  | 
 | // FIXME-GFX10: WIP. | 
 | multiclass VINTRP_m <bits <2> op, dag outs, dag ins, string asm, | 
 |                      list<dag> pattern = []> { | 
 |   def "" : VINTRP_Pseudo <NAME, outs, ins, pattern>; | 
 |  | 
 |   let AssemblerPredicate = isGFX6GFX7, DecoderNamespace = "GFX6GFX7" in { | 
 |     def _si : VINTRP_Real_si <op, NAME, outs, ins, asm, SIEncodingFamily.SI>; | 
 |   } // End AssemblerPredicate = isGFX6GFX7, DecoderNamespace = "GFX6GFX7" | 
 |  | 
 |   def _vi : VINTRP_Real_vi <op, NAME, outs, ins, asm>; | 
 |  | 
 |   let AssemblerPredicate = isGFX10Only, DecoderNamespace = "GFX10" in { | 
 |     def _gfx10 : VINTRP_Real_si<op, NAME, outs, ins, asm, SIEncodingFamily.GFX10>; | 
 |   } // End AssemblerPredicate = isGFX10Only, DecoderNamespace = "GFX10" | 
 | } | 
 |  | 
 | //===----------------------------------------------------------------------===// | 
 | // Vector instruction mappings | 
 | //===----------------------------------------------------------------------===// | 
 |  | 
 | // Maps an opcode in e32 form to its e64 equivalent | 
 | def getVOPe64 : InstrMapping { | 
 |   let FilterClass = "VOP"; | 
 |   let RowFields = ["OpName"]; | 
 |   let ColFields = ["Size", "VOP3"]; | 
 |   let KeyCol = ["4", "0"]; | 
 |   let ValueCols = [["8", "1"]]; | 
 | } | 
 |  | 
 | // Maps an opcode in e64 form to its e32 equivalent | 
 | def getVOPe32 : InstrMapping { | 
 |   let FilterClass = "VOP"; | 
 |   let RowFields = ["OpName"]; | 
 |   let ColFields = ["Size", "VOP3"]; | 
 |   let KeyCol = ["8", "1"]; | 
 |   let ValueCols = [["4", "0"]]; | 
 | } | 
 |  | 
 | // Maps ordinary instructions to their SDWA counterparts | 
 | def getSDWAOp : InstrMapping { | 
 |   let FilterClass = "VOP"; | 
 |   let RowFields = ["OpName"]; | 
 |   let ColFields = ["AsmVariantName"]; | 
 |   let KeyCol = ["Default"]; | 
 |   let ValueCols = [["SDWA"]]; | 
 | } | 
 |  | 
 | // Maps SDWA instructions to their ordinary counterparts | 
 | def getBasicFromSDWAOp : InstrMapping { | 
 |   let FilterClass = "VOP"; | 
 |   let RowFields = ["OpName"]; | 
 |   let ColFields = ["AsmVariantName"]; | 
 |   let KeyCol = ["SDWA"]; | 
 |   let ValueCols = [["Default"]]; | 
 | } | 
 |  | 
 | // Maps ordinary instructions to their DPP counterparts | 
 | def getDPPOp32 : InstrMapping { | 
 |   let FilterClass = "VOP"; | 
 |   let RowFields = ["OpName"]; | 
 |   let ColFields = ["AsmVariantName"]; | 
 |   let KeyCol = ["Default"]; | 
 |   let ValueCols = [["DPP"]]; | 
 | } | 
 |  | 
 | def getDPPOp64 : InstrMapping { | 
 |   let FilterClass = "VOP"; | 
 |   let RowFields = ["OpName"]; | 
 |   let ColFields = ["AsmVariantName"]; | 
 |   let KeyCol = ["VOP3"]; | 
 |   let ValueCols = [["VOP3_DPP"]]; | 
 | } | 
 |  | 
 | // Maps an commuted opcode to its original version | 
 | def getCommuteOrig : InstrMapping { | 
 |   let FilterClass = "Commutable_REV"; | 
 |   let RowFields = ["RevOp"]; | 
 |   let ColFields = ["IsOrig"]; | 
 |   let KeyCol = ["0"]; | 
 |   let ValueCols = [["1"]]; | 
 | } | 
 |  | 
 | // Maps an original opcode to its commuted version | 
 | def getCommuteRev : InstrMapping { | 
 |   let FilterClass = "Commutable_REV"; | 
 |   let RowFields = ["RevOp"]; | 
 |   let ColFields = ["IsOrig"]; | 
 |   let KeyCol = ["1"]; | 
 |   let ValueCols = [["0"]]; | 
 | } | 
 |  | 
 | def getMCOpcodeGen : InstrMapping { | 
 |   let FilterClass = "SIMCInstr"; | 
 |   let RowFields = ["PseudoInstr"]; | 
 |   let ColFields = ["Subtarget"]; | 
 |   let KeyCol = [!cast<string>(SIEncodingFamily.NONE)]; | 
 |   // These columns must be kept in sync with the SIEncodingFamily enumeration. | 
 |   let ValueCols = [[!cast<string>(SIEncodingFamily.SI)], | 
 |                    [!cast<string>(SIEncodingFamily.VI)], | 
 |                    [!cast<string>(SIEncodingFamily.SDWA)], | 
 |                    [!cast<string>(SIEncodingFamily.SDWA9)], | 
 |                    // GFX80 encoding is added to work around a multiple matching | 
 |                    // issue for buffer instructions with unpacked d16 data. This | 
 |                    // does not actually change the encoding, and thus may be | 
 |                    // removed later. | 
 |                    [!cast<string>(SIEncodingFamily.GFX80)], | 
 |                    [!cast<string>(SIEncodingFamily.GFX9)], | 
 |                    [!cast<string>(SIEncodingFamily.GFX10)], | 
 |                    [!cast<string>(SIEncodingFamily.SDWA10)], | 
 |                    [!cast<string>(SIEncodingFamily.GFX90A)], | 
 |                    [!cast<string>(SIEncodingFamily.GFX940)], | 
 |                    [!cast<string>(SIEncodingFamily.GFX11)], | 
 |                    [!cast<string>(SIEncodingFamily.GFX12)], | 
 |                    [!cast<string>(SIEncodingFamily.GFX1250)]]; | 
 | } | 
 |  | 
 | // Get equivalent SOPK instruction. | 
 | def getSOPKOp : InstrMapping { | 
 |   let FilterClass = "SOPKInstTable"; | 
 |   let RowFields = ["BaseCmpOp"]; | 
 |   let ColFields = ["IsSOPK"]; | 
 |   let KeyCol = ["0"]; | 
 |   let ValueCols = [["1"]]; | 
 | } | 
 |  | 
 | def getAddr64Inst : InstrMapping { | 
 |   let FilterClass = "MUBUFAddr64Table"; | 
 |   let RowFields = ["OpName"]; | 
 |   let ColFields = ["IsAddr64"]; | 
 |   let KeyCol = ["0"]; | 
 |   let ValueCols = [["1"]]; | 
 | } | 
 |  | 
 | def getIfAddr64Inst : InstrMapping { | 
 |   let FilterClass = "MUBUFAddr64Table"; | 
 |   let RowFields = ["OpName"]; | 
 |   let ColFields = ["IsAddr64"]; | 
 |   let KeyCol = ["1"]; | 
 |   let ValueCols = [["1"]]; | 
 | } | 
 |  | 
 | // Maps a GLOBAL to its SADDR form. | 
 | def getGlobalSaddrOp : InstrMapping { | 
 |   let FilterClass = "GlobalSaddrTable"; | 
 |   let RowFields = ["SaddrOp"]; | 
 |   let ColFields = ["IsSaddr"]; | 
 |   let KeyCol = ["0"]; | 
 |   let ValueCols = [["1"]]; | 
 | } | 
 |  | 
 | // Maps a GLOBAL SADDR to its VADDR form. | 
 | def getGlobalVaddrOp : InstrMapping { | 
 |   let FilterClass = "GlobalSaddrTable"; | 
 |   let RowFields = ["SaddrOp"]; | 
 |   let ColFields = ["IsSaddr"]; | 
 |   let KeyCol = ["1"]; | 
 |   let ValueCols = [["0"]]; | 
 | } | 
 |  | 
 | // Maps a v_cmpx opcode with sdst to opcode without sdst. | 
 | def getVCMPXNoSDstOp : InstrMapping { | 
 |   let FilterClass = "VCMPXNoSDstTable"; | 
 |   let RowFields = ["NoSDstOp"]; | 
 |   let ColFields = ["HasSDst"]; | 
 |   let KeyCol = ["1"]; | 
 |   let ValueCols = [["0"]]; | 
 | } | 
 |  | 
 | // Maps a SOPP to a SOPP with S_NOP | 
 | def getSOPPWithRelaxation : InstrMapping { | 
 |   let FilterClass = "SOPPRelaxTable"; | 
 |   let RowFields = ["KeyName"]; | 
 |   let ColFields = ["IsRelaxed"]; | 
 |   let KeyCol = ["0"]; | 
 |   let ValueCols = [["1"]]; | 
 | } | 
 |  | 
 | // Maps flat scratch opcodes by addressing modes | 
 | def getFlatScratchInstSTfromSS : InstrMapping { | 
 |   let FilterClass = "FlatScratchInst"; | 
 |   let RowFields = ["SVOp"]; | 
 |   let ColFields = ["Mode"]; | 
 |   let KeyCol = ["SS"]; | 
 |   let ValueCols = [["ST"]]; | 
 | } | 
 |  | 
 | def getFlatScratchInstSSfromSV : InstrMapping { | 
 |   let FilterClass = "FlatScratchInst"; | 
 |   let RowFields = ["SVOp"]; | 
 |   let ColFields = ["Mode"]; | 
 |   let KeyCol = ["SV"]; | 
 |   let ValueCols = [["SS"]]; | 
 | } | 
 |  | 
 | def getFlatScratchInstSVfromSVS : InstrMapping { | 
 |   let FilterClass = "FlatScratchInst"; | 
 |   let RowFields = ["SVOp"]; | 
 |   let ColFields = ["Mode"]; | 
 |   let KeyCol = ["SVS"]; | 
 |   let ValueCols = [["SV"]]; | 
 | } | 
 |  | 
 | def getFlatScratchInstSVfromSS : InstrMapping { | 
 |   let FilterClass = "FlatScratchInst"; | 
 |   let RowFields = ["SVOp"]; | 
 |   let ColFields = ["Mode"]; | 
 |   let KeyCol = ["SS"]; | 
 |   let ValueCols = [["SV"]]; | 
 | } | 
 |  | 
 | def getMFMAEarlyClobberOp : InstrMapping { | 
 |   let FilterClass = "MFMATable"; | 
 |   let RowFields = ["FMAOp"]; | 
 |   let ColFields = ["IsMac"]; | 
 |   let KeyCol = ["1"]; | 
 |   let ValueCols = [["0"]]; | 
 | } | 
 |  | 
 | // Map from an mfma using VGPRs to one using AGPRs. | 
 | def getMFMASrcCVDstAGPROp : InstrMapping { | 
 |   let FilterClass = "MFMATable"; | 
 |   let RowFields = ["AGPROp"]; | 
 |   let ColFields = ["MFMAKind"]; | 
 |   let KeyCol = ["VGPR"]; | 
 |   let ValueCols = [["AGPR"]]; | 
 | } | 
 |  | 
 | // Maps an v_cmp instruction to its v_cmpx equivalent. | 
 | def getVCMPXOpFromVCMP : InstrMapping { | 
 |   let FilterClass = "VCMPVCMPXTable"; | 
 |   let RowFields = ["VCMPOp"]; | 
 |   let ColFields = ["IsVCMPX"]; | 
 |   let KeyCol = ["0"]; | 
 |   let ValueCols = [["1"]]; | 
 | } | 
 |  | 
 | // Map encoded mfma(_scale)?_f8f6f4 instructions depending on the | 
 | // number of registers required for the used format. | 
 | def getMFMA_F8F6F4_WithSize : GenericTable { | 
 |   let FilterClass = "MFMA_F8F6F4_WithSizeTable"; | 
 |   let CppTypeName = "MFMA_F8F6F4_Info"; | 
 |   let Fields = [ "Opcode", "F8F8Opcode", "NumRegsSrcA", "NumRegsSrcB" ]; | 
 |   let PrimaryKey = [ "NumRegsSrcA", "NumRegsSrcB", "F8F8Opcode" ]; | 
 |   let PrimaryKeyName = "getMFMA_F8F6F4_InstWithNumRegs" ; | 
 | } | 
 |  | 
 | def isMFMA_F8F6F4Table : GenericTable { | 
 |   let FilterClass = "MFMA_F8F6F4_WithSizeTable"; | 
 |   let CppTypeName = "MFMA_F8F6F4_Info"; | 
 | //  let Fields = [ "Opcode" ]; | 
 |   let Fields = [ "Opcode", "F8F8Opcode", "NumRegsSrcA", "NumRegsSrcB" ]; | 
 |   let PrimaryKey = [ "Opcode" ]; | 
 |   let PrimaryKeyName = "isMFMA_F8F6F4" ; | 
 | } | 
 |  | 
 | def FP4FP8DstByteSelTable : GenericTable { | 
 |   let FilterClass = "VOP3_Pseudo"; | 
 |   let CppTypeName = "FP4FP8DstByteSelInfo"; | 
 |   let Fields = ["Opcode", "HasFP8DstByteSel", "HasFP4DstByteSel"]; | 
 |  | 
 |   let PrimaryKey = ["Opcode"]; | 
 |   let PrimaryKeyName = "getFP4FP8DstByteSelHelper"; | 
 | } | 
 |  | 
 | def VOPDComponentTable : GenericTable { | 
 |   let FilterClass = "VOPD_Component"; | 
 |   let CppTypeName = "VOPDComponentInfo"; | 
 |   let Fields = ["BaseVOP", "VOPDOp", "CanBeVOPDX"]; | 
 |   let PrimaryKey = ["BaseVOP"]; | 
 |   let PrimaryKeyName = "getVOPDComponentHelper"; | 
 | } | 
 |  | 
 | def getVOPDBaseFromComponent : SearchIndex { | 
 |   let Table = VOPDComponentTable; | 
 |   let Key = ["VOPDOp"]; | 
 | } | 
 |  | 
 | def VOPDPairs : GenericTable { | 
 |   let FilterClass = "VOPD_Base"; | 
 |   let CppTypeName = "VOPDInfo"; | 
 |   let Fields = ["Opcode", "OpX", "OpY", "SubTgt"]; | 
 |   let PrimaryKey = ["Opcode"]; | 
 |   let PrimaryKeyName = "getVOPDOpcodeHelper"; | 
 | } | 
 |  | 
 | def getVOPDInfoFromComponentOpcodes : SearchIndex { | 
 |   let Table = VOPDPairs; | 
 |   let Key = ["OpX", "OpY", "SubTgt"]; | 
 | } | 
 |  | 
 | include "SIInstructions.td" | 
 |  | 
 | include "DSInstructions.td" | 
 | include "MIMGInstructions.td" |