| //=- AArch64RegisterInfo.td - Describe the AArch64 Registers -*- tablegen -*-=// |
| // |
| // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. |
| // See https://llvm.org/LICENSE.txt for license information. |
| // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception |
| // |
| //===----------------------------------------------------------------------===// |
| // |
| // |
| //===----------------------------------------------------------------------===// |
| |
| |
| class AArch64Reg<bits<16> enc, string n, list<Register> subregs = [], |
| list<string> altNames = []> |
| : Register<n, altNames> { |
| let HWEncoding = enc; |
| let Namespace = "AArch64"; |
| let SubRegs = subregs; |
| } |
| |
| let Namespace = "AArch64" in { |
| def sub_32 : SubRegIndex<32>; |
| |
| def bsub : SubRegIndex<8>; |
| def hsub : SubRegIndex<16>; |
| def ssub : SubRegIndex<32>; |
| def dsub : SubRegIndex<64>; |
| def sube32 : SubRegIndex<32>; |
| def subo32 : SubRegIndex<32>; |
| def sube64 : SubRegIndex<64>; |
| def subo64 : SubRegIndex<64>; |
| // SVE |
| def zsub : SubRegIndex<128>; |
| // Note: zsub_hi should never be used directly because it represents |
| // the scalable part of the SVE vector and cannot be manipulated as a |
| // subvector in the same way the lower 128bits can. |
| def zsub_hi : SubRegIndex<128>; |
| // Note: Code depends on these having consecutive numbers |
| def dsub0 : SubRegIndex<64>; |
| def dsub1 : SubRegIndex<64>; |
| def dsub2 : SubRegIndex<64>; |
| def dsub3 : SubRegIndex<64>; |
| // Note: Code depends on these having consecutive numbers |
| def qsub0 : SubRegIndex<128>; |
| def qsub1 : SubRegIndex<128>; |
| def qsub2 : SubRegIndex<128>; |
| def qsub3 : SubRegIndex<128>; |
| // Note: Code depends on these having consecutive numbers |
| def zasubb : SubRegIndex<2048>; // (16 x 16)/1 bytes = 2048 bits |
| def zasubh0 : SubRegIndex<1024>; // (16 x 16)/2 bytes = 1024 bits |
| def zasubh1 : SubRegIndex<1024>; // (16 x 16)/2 bytes = 1024 bits |
| def zasubs0 : SubRegIndex<512>; // (16 x 16)/4 bytes = 512 bits |
| def zasubs1 : SubRegIndex<512>; // (16 x 16)/4 bytes = 512 bits |
| def zasubd0 : SubRegIndex<256>; // (16 x 16)/8 bytes = 256 bits |
| def zasubd1 : SubRegIndex<256>; // (16 x 16)/8 bytes = 256 bits |
| def zasubq0 : SubRegIndex<128>; // (16 x 16)/16 bytes = 128 bits |
| def zasubq1 : SubRegIndex<128>; // (16 x 16)/16 bytes = 128 bits |
| } |
| |
| let Namespace = "AArch64" in { |
| def vreg : RegAltNameIndex; |
| def vlist1 : RegAltNameIndex; |
| } |
| |
| //===----------------------------------------------------------------------===// |
| // Registers |
| //===----------------------------------------------------------------------===// |
| def W0 : AArch64Reg<0, "w0" >, DwarfRegNum<[0]>; |
| def W1 : AArch64Reg<1, "w1" >, DwarfRegNum<[1]>; |
| def W2 : AArch64Reg<2, "w2" >, DwarfRegNum<[2]>; |
| def W3 : AArch64Reg<3, "w3" >, DwarfRegNum<[3]>; |
| def W4 : AArch64Reg<4, "w4" >, DwarfRegNum<[4]>; |
| def W5 : AArch64Reg<5, "w5" >, DwarfRegNum<[5]>; |
| def W6 : AArch64Reg<6, "w6" >, DwarfRegNum<[6]>; |
| def W7 : AArch64Reg<7, "w7" >, DwarfRegNum<[7]>; |
| def W8 : AArch64Reg<8, "w8" >, DwarfRegNum<[8]>; |
| def W9 : AArch64Reg<9, "w9" >, DwarfRegNum<[9]>; |
| def W10 : AArch64Reg<10, "w10">, DwarfRegNum<[10]>; |
| def W11 : AArch64Reg<11, "w11">, DwarfRegNum<[11]>; |
| def W12 : AArch64Reg<12, "w12">, DwarfRegNum<[12]>; |
| def W13 : AArch64Reg<13, "w13">, DwarfRegNum<[13]>; |
| def W14 : AArch64Reg<14, "w14">, DwarfRegNum<[14]>; |
| def W15 : AArch64Reg<15, "w15">, DwarfRegNum<[15]>; |
| def W16 : AArch64Reg<16, "w16">, DwarfRegNum<[16]>; |
| def W17 : AArch64Reg<17, "w17">, DwarfRegNum<[17]>; |
| def W18 : AArch64Reg<18, "w18">, DwarfRegNum<[18]>; |
| def W19 : AArch64Reg<19, "w19">, DwarfRegNum<[19]>; |
| def W20 : AArch64Reg<20, "w20">, DwarfRegNum<[20]>; |
| def W21 : AArch64Reg<21, "w21">, DwarfRegNum<[21]>; |
| def W22 : AArch64Reg<22, "w22">, DwarfRegNum<[22]>; |
| def W23 : AArch64Reg<23, "w23">, DwarfRegNum<[23]>; |
| def W24 : AArch64Reg<24, "w24">, DwarfRegNum<[24]>; |
| def W25 : AArch64Reg<25, "w25">, DwarfRegNum<[25]>; |
| def W26 : AArch64Reg<26, "w26">, DwarfRegNum<[26]>; |
| def W27 : AArch64Reg<27, "w27">, DwarfRegNum<[27]>; |
| def W28 : AArch64Reg<28, "w28">, DwarfRegNum<[28]>; |
| def W29 : AArch64Reg<29, "w29">, DwarfRegNum<[29]>; |
| def W30 : AArch64Reg<30, "w30">, DwarfRegNum<[30]>; |
| def WSP : AArch64Reg<31, "wsp">, DwarfRegNum<[31]>; |
| def WZR : AArch64Reg<31, "wzr">, DwarfRegAlias<WSP>; |
| |
| let SubRegIndices = [sub_32] in { |
| def X0 : AArch64Reg<0, "x0", [W0]>, DwarfRegAlias<W0>; |
| def X1 : AArch64Reg<1, "x1", [W1]>, DwarfRegAlias<W1>; |
| def X2 : AArch64Reg<2, "x2", [W2]>, DwarfRegAlias<W2>; |
| def X3 : AArch64Reg<3, "x3", [W3]>, DwarfRegAlias<W3>; |
| def X4 : AArch64Reg<4, "x4", [W4]>, DwarfRegAlias<W4>; |
| def X5 : AArch64Reg<5, "x5", [W5]>, DwarfRegAlias<W5>; |
| def X6 : AArch64Reg<6, "x6", [W6]>, DwarfRegAlias<W6>; |
| def X7 : AArch64Reg<7, "x7", [W7]>, DwarfRegAlias<W7>; |
| def X8 : AArch64Reg<8, "x8", [W8]>, DwarfRegAlias<W8>; |
| def X9 : AArch64Reg<9, "x9", [W9]>, DwarfRegAlias<W9>; |
| def X10 : AArch64Reg<10, "x10", [W10]>, DwarfRegAlias<W10>; |
| def X11 : AArch64Reg<11, "x11", [W11]>, DwarfRegAlias<W11>; |
| def X12 : AArch64Reg<12, "x12", [W12]>, DwarfRegAlias<W12>; |
| def X13 : AArch64Reg<13, "x13", [W13]>, DwarfRegAlias<W13>; |
| def X14 : AArch64Reg<14, "x14", [W14]>, DwarfRegAlias<W14>; |
| def X15 : AArch64Reg<15, "x15", [W15]>, DwarfRegAlias<W15>; |
| def X16 : AArch64Reg<16, "x16", [W16]>, DwarfRegAlias<W16>; |
| def X17 : AArch64Reg<17, "x17", [W17]>, DwarfRegAlias<W17>; |
| def X18 : AArch64Reg<18, "x18", [W18]>, DwarfRegAlias<W18>; |
| def X19 : AArch64Reg<19, "x19", [W19]>, DwarfRegAlias<W19>; |
| def X20 : AArch64Reg<20, "x20", [W20]>, DwarfRegAlias<W20>; |
| def X21 : AArch64Reg<21, "x21", [W21]>, DwarfRegAlias<W21>; |
| def X22 : AArch64Reg<22, "x22", [W22]>, DwarfRegAlias<W22>; |
| def X23 : AArch64Reg<23, "x23", [W23]>, DwarfRegAlias<W23>; |
| def X24 : AArch64Reg<24, "x24", [W24]>, DwarfRegAlias<W24>; |
| def X25 : AArch64Reg<25, "x25", [W25]>, DwarfRegAlias<W25>; |
| def X26 : AArch64Reg<26, "x26", [W26]>, DwarfRegAlias<W26>; |
| def X27 : AArch64Reg<27, "x27", [W27]>, DwarfRegAlias<W27>; |
| def X28 : AArch64Reg<28, "x28", [W28]>, DwarfRegAlias<W28>; |
| def FP : AArch64Reg<29, "x29", [W29]>, DwarfRegAlias<W29>; |
| def LR : AArch64Reg<30, "x30", [W30]>, DwarfRegAlias<W30>; |
| def SP : AArch64Reg<31, "sp", [WSP]>, DwarfRegAlias<WSP>; |
| def XZR : AArch64Reg<31, "xzr", [WZR]>, DwarfRegAlias<WSP>; |
| } |
| |
| // Condition code register. |
| def NZCV : AArch64Reg<0, "nzcv">; |
| |
| // First fault status register |
| def FFR : AArch64Reg<0, "ffr">, DwarfRegNum<[47]>; |
| |
| // Purely virtual Vector Granule (VG) Dwarf register |
| def VG : AArch64Reg<0, "vg">, DwarfRegNum<[46]>; |
| |
| // GPR register classes with the intersections of GPR32/GPR32sp and |
| // GPR64/GPR64sp for use by the coalescer. |
| def GPR32common : RegisterClass<"AArch64", [i32], 32, (sequence "W%u", 0, 30)> { |
| let AltOrders = [(rotl GPR32common, 8)]; |
| let AltOrderSelect = [{ return 1; }]; |
| } |
| def GPR64common : RegisterClass<"AArch64", [i64], 64, |
| (add (sequence "X%u", 0, 28), FP, LR)> { |
| let AltOrders = [(rotl GPR64common, 8)]; |
| let AltOrderSelect = [{ return 1; }]; |
| } |
| // GPR register classes which exclude SP/WSP. |
| def GPR32 : RegisterClass<"AArch64", [i32], 32, (add GPR32common, WZR)> { |
| let AltOrders = [(rotl GPR32, 8)]; |
| let AltOrderSelect = [{ return 1; }]; |
| } |
| def GPR64 : RegisterClass<"AArch64", [i64], 64, (add GPR64common, XZR)> { |
| let AltOrders = [(rotl GPR64, 8)]; |
| let AltOrderSelect = [{ return 1; }]; |
| } |
| |
| // GPR register classes which include SP/WSP. |
| def GPR32sp : RegisterClass<"AArch64", [i32], 32, (add GPR32common, WSP)> { |
| let AltOrders = [(rotl GPR32sp, 8)]; |
| let AltOrderSelect = [{ return 1; }]; |
| } |
| def GPR64sp : RegisterClass<"AArch64", [i64], 64, (add GPR64common, SP)> { |
| let AltOrders = [(rotl GPR64sp, 8)]; |
| let AltOrderSelect = [{ return 1; }]; |
| } |
| |
| def GPR32sponly : RegisterClass<"AArch64", [i32], 32, (add WSP)>; |
| def GPR64sponly : RegisterClass<"AArch64", [i64], 64, (add SP)>; |
| |
| def GPR64spPlus0Operand : AsmOperandClass { |
| let Name = "GPR64sp0"; |
| let RenderMethod = "addRegOperands"; |
| let PredicateMethod = "isGPR64<AArch64::GPR64spRegClassID>"; |
| let ParserMethod = "tryParseGPR64sp0Operand"; |
| } |
| |
| def GPR64sp0 : RegisterOperand<GPR64sp> { |
| let ParserMatchClass = GPR64spPlus0Operand; |
| } |
| |
| // GPR32/GPR64 but with zero-register substitution enabled. |
| // TODO: Roll this out to GPR32/GPR64/GPR32all/GPR64all. |
| def GPR32z : RegisterOperand<GPR32> { |
| let GIZeroRegister = WZR; |
| } |
| def GPR64z : RegisterOperand<GPR64> { |
| let GIZeroRegister = XZR; |
| } |
| |
| // GPR argument registers. |
| def GPR32arg : RegisterClass<"AArch64", [i32], 32, (sequence "W%u", 0, 7)>; |
| def GPR64arg : RegisterClass<"AArch64", [i64], 64, (sequence "X%u", 0, 7)>; |
| |
| // GPR register classes which include WZR/XZR AND SP/WSP. This is not a |
| // constraint used by any instructions, it is used as a common super-class. |
| def GPR32all : RegisterClass<"AArch64", [i32], 32, (add GPR32common, WZR, WSP)>; |
| def GPR64all : RegisterClass<"AArch64", [i64], 64, (add GPR64common, XZR, SP)>; |
| |
| // For tail calls, we can't use callee-saved registers, as they are restored |
| // to the saved value before the tail call, which would clobber a call address. |
| // This is for indirect tail calls to store the address of the destination. |
| def tcGPR64 : RegisterClass<"AArch64", [i64], 64, (sub GPR64common, X19, X20, X21, |
| X22, X23, X24, X25, X26, |
| X27, X28, FP, LR)>; |
| |
| // Restricted set of tail call registers, for use when branch target |
| // enforcement is enabled. These are the only registers which can be used to |
| // indirectly branch (not call) to the "BTI c" instruction at the start of a |
| // BTI-protected function. |
| def rtcGPR64 : RegisterClass<"AArch64", [i64], 64, (add X16, X17)>; |
| |
| // Register set that excludes registers that are reserved for procedure calls. |
| // This is used for pseudo-instructions that are actually implemented using a |
| // procedure call. |
| def GPR64noip : RegisterClass<"AArch64", [i64], 64, (sub GPR64, X16, X17, LR)>; |
| |
| // GPR register classes for post increment amount of vector load/store that |
| // has alternate printing when Rm=31 and prints a constant immediate value |
| // equal to the total number of bytes transferred. |
| |
| // FIXME: TableGen *should* be able to do these itself now. There appears to be |
| // a bug in counting how many operands a Post-indexed MCInst should have which |
| // means the aliases don't trigger. |
| def GPR64pi1 : RegisterOperand<GPR64, "printPostIncOperand<1>">; |
| def GPR64pi2 : RegisterOperand<GPR64, "printPostIncOperand<2>">; |
| def GPR64pi3 : RegisterOperand<GPR64, "printPostIncOperand<3>">; |
| def GPR64pi4 : RegisterOperand<GPR64, "printPostIncOperand<4>">; |
| def GPR64pi6 : RegisterOperand<GPR64, "printPostIncOperand<6>">; |
| def GPR64pi8 : RegisterOperand<GPR64, "printPostIncOperand<8>">; |
| def GPR64pi12 : RegisterOperand<GPR64, "printPostIncOperand<12>">; |
| def GPR64pi16 : RegisterOperand<GPR64, "printPostIncOperand<16>">; |
| def GPR64pi24 : RegisterOperand<GPR64, "printPostIncOperand<24>">; |
| def GPR64pi32 : RegisterOperand<GPR64, "printPostIncOperand<32>">; |
| def GPR64pi48 : RegisterOperand<GPR64, "printPostIncOperand<48>">; |
| def GPR64pi64 : RegisterOperand<GPR64, "printPostIncOperand<64>">; |
| |
| // Condition code regclass. |
| def CCR : RegisterClass<"AArch64", [i32], 32, (add NZCV)> { |
| let CopyCost = -1; // Don't allow copying of status registers. |
| |
| // CCR is not allocatable. |
| let isAllocatable = 0; |
| } |
| |
| //===----------------------------------------------------------------------===// |
| // Floating Point Scalar Registers |
| //===----------------------------------------------------------------------===// |
| |
| def B0 : AArch64Reg<0, "b0">, DwarfRegNum<[64]>; |
| def B1 : AArch64Reg<1, "b1">, DwarfRegNum<[65]>; |
| def B2 : AArch64Reg<2, "b2">, DwarfRegNum<[66]>; |
| def B3 : AArch64Reg<3, "b3">, DwarfRegNum<[67]>; |
| def B4 : AArch64Reg<4, "b4">, DwarfRegNum<[68]>; |
| def B5 : AArch64Reg<5, "b5">, DwarfRegNum<[69]>; |
| def B6 : AArch64Reg<6, "b6">, DwarfRegNum<[70]>; |
| def B7 : AArch64Reg<7, "b7">, DwarfRegNum<[71]>; |
| def B8 : AArch64Reg<8, "b8">, DwarfRegNum<[72]>; |
| def B9 : AArch64Reg<9, "b9">, DwarfRegNum<[73]>; |
| def B10 : AArch64Reg<10, "b10">, DwarfRegNum<[74]>; |
| def B11 : AArch64Reg<11, "b11">, DwarfRegNum<[75]>; |
| def B12 : AArch64Reg<12, "b12">, DwarfRegNum<[76]>; |
| def B13 : AArch64Reg<13, "b13">, DwarfRegNum<[77]>; |
| def B14 : AArch64Reg<14, "b14">, DwarfRegNum<[78]>; |
| def B15 : AArch64Reg<15, "b15">, DwarfRegNum<[79]>; |
| def B16 : AArch64Reg<16, "b16">, DwarfRegNum<[80]>; |
| def B17 : AArch64Reg<17, "b17">, DwarfRegNum<[81]>; |
| def B18 : AArch64Reg<18, "b18">, DwarfRegNum<[82]>; |
| def B19 : AArch64Reg<19, "b19">, DwarfRegNum<[83]>; |
| def B20 : AArch64Reg<20, "b20">, DwarfRegNum<[84]>; |
| def B21 : AArch64Reg<21, "b21">, DwarfRegNum<[85]>; |
| def B22 : AArch64Reg<22, "b22">, DwarfRegNum<[86]>; |
| def B23 : AArch64Reg<23, "b23">, DwarfRegNum<[87]>; |
| def B24 : AArch64Reg<24, "b24">, DwarfRegNum<[88]>; |
| def B25 : AArch64Reg<25, "b25">, DwarfRegNum<[89]>; |
| def B26 : AArch64Reg<26, "b26">, DwarfRegNum<[90]>; |
| def B27 : AArch64Reg<27, "b27">, DwarfRegNum<[91]>; |
| def B28 : AArch64Reg<28, "b28">, DwarfRegNum<[92]>; |
| def B29 : AArch64Reg<29, "b29">, DwarfRegNum<[93]>; |
| def B30 : AArch64Reg<30, "b30">, DwarfRegNum<[94]>; |
| def B31 : AArch64Reg<31, "b31">, DwarfRegNum<[95]>; |
| |
| let SubRegIndices = [bsub] in { |
| def H0 : AArch64Reg<0, "h0", [B0]>, DwarfRegAlias<B0>; |
| def H1 : AArch64Reg<1, "h1", [B1]>, DwarfRegAlias<B1>; |
| def H2 : AArch64Reg<2, "h2", [B2]>, DwarfRegAlias<B2>; |
| def H3 : AArch64Reg<3, "h3", [B3]>, DwarfRegAlias<B3>; |
| def H4 : AArch64Reg<4, "h4", [B4]>, DwarfRegAlias<B4>; |
| def H5 : AArch64Reg<5, "h5", [B5]>, DwarfRegAlias<B5>; |
| def H6 : AArch64Reg<6, "h6", [B6]>, DwarfRegAlias<B6>; |
| def H7 : AArch64Reg<7, "h7", [B7]>, DwarfRegAlias<B7>; |
| def H8 : AArch64Reg<8, "h8", [B8]>, DwarfRegAlias<B8>; |
| def H9 : AArch64Reg<9, "h9", [B9]>, DwarfRegAlias<B9>; |
| def H10 : AArch64Reg<10, "h10", [B10]>, DwarfRegAlias<B10>; |
| def H11 : AArch64Reg<11, "h11", [B11]>, DwarfRegAlias<B11>; |
| def H12 : AArch64Reg<12, "h12", [B12]>, DwarfRegAlias<B12>; |
| def H13 : AArch64Reg<13, "h13", [B13]>, DwarfRegAlias<B13>; |
| def H14 : AArch64Reg<14, "h14", [B14]>, DwarfRegAlias<B14>; |
| def H15 : AArch64Reg<15, "h15", [B15]>, DwarfRegAlias<B15>; |
| def H16 : AArch64Reg<16, "h16", [B16]>, DwarfRegAlias<B16>; |
| def H17 : AArch64Reg<17, "h17", [B17]>, DwarfRegAlias<B17>; |
| def H18 : AArch64Reg<18, "h18", [B18]>, DwarfRegAlias<B18>; |
| def H19 : AArch64Reg<19, "h19", [B19]>, DwarfRegAlias<B19>; |
| def H20 : AArch64Reg<20, "h20", [B20]>, DwarfRegAlias<B20>; |
| def H21 : AArch64Reg<21, "h21", [B21]>, DwarfRegAlias<B21>; |
| def H22 : AArch64Reg<22, "h22", [B22]>, DwarfRegAlias<B22>; |
| def H23 : AArch64Reg<23, "h23", [B23]>, DwarfRegAlias<B23>; |
| def H24 : AArch64Reg<24, "h24", [B24]>, DwarfRegAlias<B24>; |
| def H25 : AArch64Reg<25, "h25", [B25]>, DwarfRegAlias<B25>; |
| def H26 : AArch64Reg<26, "h26", [B26]>, DwarfRegAlias<B26>; |
| def H27 : AArch64Reg<27, "h27", [B27]>, DwarfRegAlias<B27>; |
| def H28 : AArch64Reg<28, "h28", [B28]>, DwarfRegAlias<B28>; |
| def H29 : AArch64Reg<29, "h29", [B29]>, DwarfRegAlias<B29>; |
| def H30 : AArch64Reg<30, "h30", [B30]>, DwarfRegAlias<B30>; |
| def H31 : AArch64Reg<31, "h31", [B31]>, DwarfRegAlias<B31>; |
| } |
| |
| let SubRegIndices = [hsub] in { |
| def S0 : AArch64Reg<0, "s0", [H0]>, DwarfRegAlias<B0>; |
| def S1 : AArch64Reg<1, "s1", [H1]>, DwarfRegAlias<B1>; |
| def S2 : AArch64Reg<2, "s2", [H2]>, DwarfRegAlias<B2>; |
| def S3 : AArch64Reg<3, "s3", [H3]>, DwarfRegAlias<B3>; |
| def S4 : AArch64Reg<4, "s4", [H4]>, DwarfRegAlias<B4>; |
| def S5 : AArch64Reg<5, "s5", [H5]>, DwarfRegAlias<B5>; |
| def S6 : AArch64Reg<6, "s6", [H6]>, DwarfRegAlias<B6>; |
| def S7 : AArch64Reg<7, "s7", [H7]>, DwarfRegAlias<B7>; |
| def S8 : AArch64Reg<8, "s8", [H8]>, DwarfRegAlias<B8>; |
| def S9 : AArch64Reg<9, "s9", [H9]>, DwarfRegAlias<B9>; |
| def S10 : AArch64Reg<10, "s10", [H10]>, DwarfRegAlias<B10>; |
| def S11 : AArch64Reg<11, "s11", [H11]>, DwarfRegAlias<B11>; |
| def S12 : AArch64Reg<12, "s12", [H12]>, DwarfRegAlias<B12>; |
| def S13 : AArch64Reg<13, "s13", [H13]>, DwarfRegAlias<B13>; |
| def S14 : AArch64Reg<14, "s14", [H14]>, DwarfRegAlias<B14>; |
| def S15 : AArch64Reg<15, "s15", [H15]>, DwarfRegAlias<B15>; |
| def S16 : AArch64Reg<16, "s16", [H16]>, DwarfRegAlias<B16>; |
| def S17 : AArch64Reg<17, "s17", [H17]>, DwarfRegAlias<B17>; |
| def S18 : AArch64Reg<18, "s18", [H18]>, DwarfRegAlias<B18>; |
| def S19 : AArch64Reg<19, "s19", [H19]>, DwarfRegAlias<B19>; |
| def S20 : AArch64Reg<20, "s20", [H20]>, DwarfRegAlias<B20>; |
| def S21 : AArch64Reg<21, "s21", [H21]>, DwarfRegAlias<B21>; |
| def S22 : AArch64Reg<22, "s22", [H22]>, DwarfRegAlias<B22>; |
| def S23 : AArch64Reg<23, "s23", [H23]>, DwarfRegAlias<B23>; |
| def S24 : AArch64Reg<24, "s24", [H24]>, DwarfRegAlias<B24>; |
| def S25 : AArch64Reg<25, "s25", [H25]>, DwarfRegAlias<B25>; |
| def S26 : AArch64Reg<26, "s26", [H26]>, DwarfRegAlias<B26>; |
| def S27 : AArch64Reg<27, "s27", [H27]>, DwarfRegAlias<B27>; |
| def S28 : AArch64Reg<28, "s28", [H28]>, DwarfRegAlias<B28>; |
| def S29 : AArch64Reg<29, "s29", [H29]>, DwarfRegAlias<B29>; |
| def S30 : AArch64Reg<30, "s30", [H30]>, DwarfRegAlias<B30>; |
| def S31 : AArch64Reg<31, "s31", [H31]>, DwarfRegAlias<B31>; |
| } |
| |
| let SubRegIndices = [ssub], RegAltNameIndices = [vreg, vlist1] in { |
| def D0 : AArch64Reg<0, "d0", [S0], ["v0", ""]>, DwarfRegAlias<B0>; |
| def D1 : AArch64Reg<1, "d1", [S1], ["v1", ""]>, DwarfRegAlias<B1>; |
| def D2 : AArch64Reg<2, "d2", [S2], ["v2", ""]>, DwarfRegAlias<B2>; |
| def D3 : AArch64Reg<3, "d3", [S3], ["v3", ""]>, DwarfRegAlias<B3>; |
| def D4 : AArch64Reg<4, "d4", [S4], ["v4", ""]>, DwarfRegAlias<B4>; |
| def D5 : AArch64Reg<5, "d5", [S5], ["v5", ""]>, DwarfRegAlias<B5>; |
| def D6 : AArch64Reg<6, "d6", [S6], ["v6", ""]>, DwarfRegAlias<B6>; |
| def D7 : AArch64Reg<7, "d7", [S7], ["v7", ""]>, DwarfRegAlias<B7>; |
| def D8 : AArch64Reg<8, "d8", [S8], ["v8", ""]>, DwarfRegAlias<B8>; |
| def D9 : AArch64Reg<9, "d9", [S9], ["v9", ""]>, DwarfRegAlias<B9>; |
| def D10 : AArch64Reg<10, "d10", [S10], ["v10", ""]>, DwarfRegAlias<B10>; |
| def D11 : AArch64Reg<11, "d11", [S11], ["v11", ""]>, DwarfRegAlias<B11>; |
| def D12 : AArch64Reg<12, "d12", [S12], ["v12", ""]>, DwarfRegAlias<B12>; |
| def D13 : AArch64Reg<13, "d13", [S13], ["v13", ""]>, DwarfRegAlias<B13>; |
| def D14 : AArch64Reg<14, "d14", [S14], ["v14", ""]>, DwarfRegAlias<B14>; |
| def D15 : AArch64Reg<15, "d15", [S15], ["v15", ""]>, DwarfRegAlias<B15>; |
| def D16 : AArch64Reg<16, "d16", [S16], ["v16", ""]>, DwarfRegAlias<B16>; |
| def D17 : AArch64Reg<17, "d17", [S17], ["v17", ""]>, DwarfRegAlias<B17>; |
| def D18 : AArch64Reg<18, "d18", [S18], ["v18", ""]>, DwarfRegAlias<B18>; |
| def D19 : AArch64Reg<19, "d19", [S19], ["v19", ""]>, DwarfRegAlias<B19>; |
| def D20 : AArch64Reg<20, "d20", [S20], ["v20", ""]>, DwarfRegAlias<B20>; |
| def D21 : AArch64Reg<21, "d21", [S21], ["v21", ""]>, DwarfRegAlias<B21>; |
| def D22 : AArch64Reg<22, "d22", [S22], ["v22", ""]>, DwarfRegAlias<B22>; |
| def D23 : AArch64Reg<23, "d23", [S23], ["v23", ""]>, DwarfRegAlias<B23>; |
| def D24 : AArch64Reg<24, "d24", [S24], ["v24", ""]>, DwarfRegAlias<B24>; |
| def D25 : AArch64Reg<25, "d25", [S25], ["v25", ""]>, DwarfRegAlias<B25>; |
| def D26 : AArch64Reg<26, "d26", [S26], ["v26", ""]>, DwarfRegAlias<B26>; |
| def D27 : AArch64Reg<27, "d27", [S27], ["v27", ""]>, DwarfRegAlias<B27>; |
| def D28 : AArch64Reg<28, "d28", [S28], ["v28", ""]>, DwarfRegAlias<B28>; |
| def D29 : AArch64Reg<29, "d29", [S29], ["v29", ""]>, DwarfRegAlias<B29>; |
| def D30 : AArch64Reg<30, "d30", [S30], ["v30", ""]>, DwarfRegAlias<B30>; |
| def D31 : AArch64Reg<31, "d31", [S31], ["v31", ""]>, DwarfRegAlias<B31>; |
| } |
| |
| let SubRegIndices = [dsub], RegAltNameIndices = [vreg, vlist1] in { |
| def Q0 : AArch64Reg<0, "q0", [D0], ["v0", ""]>, DwarfRegAlias<B0>; |
| def Q1 : AArch64Reg<1, "q1", [D1], ["v1", ""]>, DwarfRegAlias<B1>; |
| def Q2 : AArch64Reg<2, "q2", [D2], ["v2", ""]>, DwarfRegAlias<B2>; |
| def Q3 : AArch64Reg<3, "q3", [D3], ["v3", ""]>, DwarfRegAlias<B3>; |
| def Q4 : AArch64Reg<4, "q4", [D4], ["v4", ""]>, DwarfRegAlias<B4>; |
| def Q5 : AArch64Reg<5, "q5", [D5], ["v5", ""]>, DwarfRegAlias<B5>; |
| def Q6 : AArch64Reg<6, "q6", [D6], ["v6", ""]>, DwarfRegAlias<B6>; |
| def Q7 : AArch64Reg<7, "q7", [D7], ["v7", ""]>, DwarfRegAlias<B7>; |
| def Q8 : AArch64Reg<8, "q8", [D8], ["v8", ""]>, DwarfRegAlias<B8>; |
| def Q9 : AArch64Reg<9, "q9", [D9], ["v9", ""]>, DwarfRegAlias<B9>; |
| def Q10 : AArch64Reg<10, "q10", [D10], ["v10", ""]>, DwarfRegAlias<B10>; |
| def Q11 : AArch64Reg<11, "q11", [D11], ["v11", ""]>, DwarfRegAlias<B11>; |
| def Q12 : AArch64Reg<12, "q12", [D12], ["v12", ""]>, DwarfRegAlias<B12>; |
| def Q13 : AArch64Reg<13, "q13", [D13], ["v13", ""]>, DwarfRegAlias<B13>; |
| def Q14 : AArch64Reg<14, "q14", [D14], ["v14", ""]>, DwarfRegAlias<B14>; |
| def Q15 : AArch64Reg<15, "q15", [D15], ["v15", ""]>, DwarfRegAlias<B15>; |
| def Q16 : AArch64Reg<16, "q16", [D16], ["v16", ""]>, DwarfRegAlias<B16>; |
| def Q17 : AArch64Reg<17, "q17", [D17], ["v17", ""]>, DwarfRegAlias<B17>; |
| def Q18 : AArch64Reg<18, "q18", [D18], ["v18", ""]>, DwarfRegAlias<B18>; |
| def Q19 : AArch64Reg<19, "q19", [D19], ["v19", ""]>, DwarfRegAlias<B19>; |
| def Q20 : AArch64Reg<20, "q20", [D20], ["v20", ""]>, DwarfRegAlias<B20>; |
| def Q21 : AArch64Reg<21, "q21", [D21], ["v21", ""]>, DwarfRegAlias<B21>; |
| def Q22 : AArch64Reg<22, "q22", [D22], ["v22", ""]>, DwarfRegAlias<B22>; |
| def Q23 : AArch64Reg<23, "q23", [D23], ["v23", ""]>, DwarfRegAlias<B23>; |
| def Q24 : AArch64Reg<24, "q24", [D24], ["v24", ""]>, DwarfRegAlias<B24>; |
| def Q25 : AArch64Reg<25, "q25", [D25], ["v25", ""]>, DwarfRegAlias<B25>; |
| def Q26 : AArch64Reg<26, "q26", [D26], ["v26", ""]>, DwarfRegAlias<B26>; |
| def Q27 : AArch64Reg<27, "q27", [D27], ["v27", ""]>, DwarfRegAlias<B27>; |
| def Q28 : AArch64Reg<28, "q28", [D28], ["v28", ""]>, DwarfRegAlias<B28>; |
| def Q29 : AArch64Reg<29, "q29", [D29], ["v29", ""]>, DwarfRegAlias<B29>; |
| def Q30 : AArch64Reg<30, "q30", [D30], ["v30", ""]>, DwarfRegAlias<B30>; |
| def Q31 : AArch64Reg<31, "q31", [D31], ["v31", ""]>, DwarfRegAlias<B31>; |
| } |
| |
| def FPR8 : RegisterClass<"AArch64", [untyped], 8, (sequence "B%u", 0, 31)> { |
| let Size = 8; |
| } |
| def FPR16 : RegisterClass<"AArch64", [f16, bf16], 16, (sequence "H%u", 0, 31)> { |
| let Size = 16; |
| } |
| |
| def FPR16_lo : RegisterClass<"AArch64", [f16], 16, (trunc FPR16, 16)> { |
| let Size = 16; |
| } |
| def FPR32 : RegisterClass<"AArch64", [f32, i32], 32,(sequence "S%u", 0, 31)>; |
| def FPR64 : RegisterClass<"AArch64", [f64, i64, v2f32, v1f64, v8i8, v4i16, v2i32, |
| v1i64, v4f16, v4bf16], |
| 64, (sequence "D%u", 0, 31)>; |
| def FPR64_lo : RegisterClass<"AArch64", |
| [v8i8, v4i16, v2i32, v1i64, v4f16, v4bf16, v2f32, |
| v1f64], |
| 64, (trunc FPR64, 16)>; |
| |
| // We don't (yet) have an f128 legal type, so don't use that here. We |
| // normalize 128-bit vectors to v2f64 for arg passing and such, so use |
| // that here. |
| def FPR128 : RegisterClass<"AArch64", |
| [v16i8, v8i16, v4i32, v2i64, v4f32, v2f64, f128, |
| v8f16, v8bf16], |
| 128, (sequence "Q%u", 0, 31)>; |
| |
| // The lower 16 vector registers. Some instructions can only take registers |
| // in this range. |
| def FPR128_lo : RegisterClass<"AArch64", |
| [v16i8, v8i16, v4i32, v2i64, v4f32, v2f64, v8f16, |
| v8bf16], |
| 128, (trunc FPR128, 16)>; |
| |
| // Pairs, triples, and quads of 64-bit vector registers. |
| def DSeqPairs : RegisterTuples<[dsub0, dsub1], [(rotl FPR64, 0), (rotl FPR64, 1)]>; |
| def DSeqTriples : RegisterTuples<[dsub0, dsub1, dsub2], |
| [(rotl FPR64, 0), (rotl FPR64, 1), |
| (rotl FPR64, 2)]>; |
| def DSeqQuads : RegisterTuples<[dsub0, dsub1, dsub2, dsub3], |
| [(rotl FPR64, 0), (rotl FPR64, 1), |
| (rotl FPR64, 2), (rotl FPR64, 3)]>; |
| def DD : RegisterClass<"AArch64", [untyped], 64, (add DSeqPairs)> { |
| let Size = 128; |
| } |
| def DDD : RegisterClass<"AArch64", [untyped], 64, (add DSeqTriples)> { |
| let Size = 192; |
| } |
| def DDDD : RegisterClass<"AArch64", [untyped], 64, (add DSeqQuads)> { |
| let Size = 256; |
| } |
| |
| // Pairs, triples, and quads of 128-bit vector registers. |
| def QSeqPairs : RegisterTuples<[qsub0, qsub1], [(rotl FPR128, 0), (rotl FPR128, 1)]>; |
| def QSeqTriples : RegisterTuples<[qsub0, qsub1, qsub2], |
| [(rotl FPR128, 0), (rotl FPR128, 1), |
| (rotl FPR128, 2)]>; |
| def QSeqQuads : RegisterTuples<[qsub0, qsub1, qsub2, qsub3], |
| [(rotl FPR128, 0), (rotl FPR128, 1), |
| (rotl FPR128, 2), (rotl FPR128, 3)]>; |
| def QQ : RegisterClass<"AArch64", [untyped], 128, (add QSeqPairs)> { |
| let Size = 256; |
| } |
| def QQQ : RegisterClass<"AArch64", [untyped], 128, (add QSeqTriples)> { |
| let Size = 384; |
| } |
| def QQQQ : RegisterClass<"AArch64", [untyped], 128, (add QSeqQuads)> { |
| let Size = 512; |
| } |
| |
| |
| // Vector operand versions of the FP registers. Alternate name printing and |
| // assembler matching. |
| def VectorReg64AsmOperand : AsmOperandClass { |
| let Name = "VectorReg64"; |
| let PredicateMethod = "isNeonVectorReg"; |
| } |
| def VectorReg128AsmOperand : AsmOperandClass { |
| let Name = "VectorReg128"; |
| let PredicateMethod = "isNeonVectorReg"; |
| } |
| |
| def V64 : RegisterOperand<FPR64, "printVRegOperand"> { |
| let ParserMatchClass = VectorReg64AsmOperand; |
| } |
| |
| def V128 : RegisterOperand<FPR128, "printVRegOperand"> { |
| let ParserMatchClass = VectorReg128AsmOperand; |
| } |
| |
| def VectorRegLoAsmOperand : AsmOperandClass { |
| let Name = "VectorRegLo"; |
| let PredicateMethod = "isNeonVectorRegLo"; |
| } |
| def V64_lo : RegisterOperand<FPR64_lo, "printVRegOperand"> { |
| let ParserMatchClass = VectorRegLoAsmOperand; |
| } |
| def V128_lo : RegisterOperand<FPR128_lo, "printVRegOperand"> { |
| let ParserMatchClass = VectorRegLoAsmOperand; |
| } |
| |
| class TypedVecListAsmOperand<int count, string vecty, int lanes, int eltsize> |
| : AsmOperandClass { |
| let Name = "TypedVectorList" # count # "_" # lanes # eltsize; |
| |
| let PredicateMethod |
| = "isTypedVectorList<RegKind::NeonVector, " # count # ", " # lanes # ", " # eltsize # ">"; |
| let RenderMethod = "addVectorListOperands<" # vecty # ", " # count # ">"; |
| } |
| |
| class TypedVecListRegOperand<RegisterClass Reg, int lanes, string eltsize> |
| : RegisterOperand<Reg, "printTypedVectorList<" # lanes # ", '" |
| # eltsize # "'>">; |
| |
| multiclass VectorList<int count, RegisterClass Reg64, RegisterClass Reg128> { |
| // With implicit types (probably on instruction instead). E.g. { v0, v1 } |
| def _64AsmOperand : AsmOperandClass { |
| let Name = NAME # "64"; |
| let PredicateMethod = "isImplicitlyTypedVectorList<RegKind::NeonVector, " # count # ">"; |
| let RenderMethod = "addVectorListOperands<AArch64Operand::VecListIdx_DReg, " # count # ">"; |
| } |
| |
| def "64" : RegisterOperand<Reg64, "printImplicitlyTypedVectorList"> { |
| let ParserMatchClass = !cast<AsmOperandClass>(NAME # "_64AsmOperand"); |
| } |
| |
| def _128AsmOperand : AsmOperandClass { |
| let Name = NAME # "128"; |
| let PredicateMethod = "isImplicitlyTypedVectorList<RegKind::NeonVector, " # count # ">"; |
| let RenderMethod = "addVectorListOperands<AArch64Operand::VecListIdx_QReg, " # count # ">"; |
| } |
| |
| def "128" : RegisterOperand<Reg128, "printImplicitlyTypedVectorList"> { |
| let ParserMatchClass = !cast<AsmOperandClass>(NAME # "_128AsmOperand"); |
| } |
| |
| // 64-bit register lists with explicit type. |
| |
| // { v0.8b, v1.8b } |
| def _8bAsmOperand : TypedVecListAsmOperand<count, "AArch64Operand::VecListIdx_DReg", 8, 8>; |
| def "8b" : TypedVecListRegOperand<Reg64, 8, "b"> { |
| let ParserMatchClass = !cast<AsmOperandClass>(NAME # "_8bAsmOperand"); |
| } |
| |
| // { v0.4h, v1.4h } |
| def _4hAsmOperand : TypedVecListAsmOperand<count, "AArch64Operand::VecListIdx_DReg", 4, 16>; |
| def "4h" : TypedVecListRegOperand<Reg64, 4, "h"> { |
| let ParserMatchClass = !cast<AsmOperandClass>(NAME # "_4hAsmOperand"); |
| } |
| |
| // { v0.2s, v1.2s } |
| def _2sAsmOperand : TypedVecListAsmOperand<count, "AArch64Operand::VecListIdx_DReg", 2, 32>; |
| def "2s" : TypedVecListRegOperand<Reg64, 2, "s"> { |
| let ParserMatchClass = !cast<AsmOperandClass>(NAME # "_2sAsmOperand"); |
| } |
| |
| // { v0.1d, v1.1d } |
| def _1dAsmOperand : TypedVecListAsmOperand<count, "AArch64Operand::VecListIdx_DReg", 1, 64>; |
| def "1d" : TypedVecListRegOperand<Reg64, 1, "d"> { |
| let ParserMatchClass = !cast<AsmOperandClass>(NAME # "_1dAsmOperand"); |
| } |
| |
| // 128-bit register lists with explicit type |
| |
| // { v0.16b, v1.16b } |
| def _16bAsmOperand : TypedVecListAsmOperand<count, "AArch64Operand::VecListIdx_QReg", 16, 8>; |
| def "16b" : TypedVecListRegOperand<Reg128, 16, "b"> { |
| let ParserMatchClass = !cast<AsmOperandClass>(NAME # "_16bAsmOperand"); |
| } |
| |
| // { v0.8h, v1.8h } |
| def _8hAsmOperand : TypedVecListAsmOperand<count, "AArch64Operand::VecListIdx_QReg", 8, 16>; |
| def "8h" : TypedVecListRegOperand<Reg128, 8, "h"> { |
| let ParserMatchClass = !cast<AsmOperandClass>(NAME # "_8hAsmOperand"); |
| } |
| |
| // { v0.4s, v1.4s } |
| def _4sAsmOperand : TypedVecListAsmOperand<count, "AArch64Operand::VecListIdx_QReg", 4, 32>; |
| def "4s" : TypedVecListRegOperand<Reg128, 4, "s"> { |
| let ParserMatchClass = !cast<AsmOperandClass>(NAME # "_4sAsmOperand"); |
| } |
| |
| // { v0.2d, v1.2d } |
| def _2dAsmOperand : TypedVecListAsmOperand<count, "AArch64Operand::VecListIdx_QReg", 2, 64>; |
| def "2d" : TypedVecListRegOperand<Reg128, 2, "d"> { |
| let ParserMatchClass = !cast<AsmOperandClass>(NAME # "_2dAsmOperand"); |
| } |
| |
| // { v0.b, v1.b } |
| def _bAsmOperand : TypedVecListAsmOperand<count, "AArch64Operand::VecListIdx_QReg", 0, 8>; |
| def "b" : TypedVecListRegOperand<Reg128, 0, "b"> { |
| let ParserMatchClass = !cast<AsmOperandClass>(NAME # "_bAsmOperand"); |
| } |
| |
| // { v0.h, v1.h } |
| def _hAsmOperand : TypedVecListAsmOperand<count, "AArch64Operand::VecListIdx_QReg", 0, 16>; |
| def "h" : TypedVecListRegOperand<Reg128, 0, "h"> { |
| let ParserMatchClass = !cast<AsmOperandClass>(NAME # "_hAsmOperand"); |
| } |
| |
| // { v0.s, v1.s } |
| def _sAsmOperand : TypedVecListAsmOperand<count, "AArch64Operand::VecListIdx_QReg", 0, 32>; |
| def "s" : TypedVecListRegOperand<Reg128, 0, "s"> { |
| let ParserMatchClass = !cast<AsmOperandClass>(NAME # "_sAsmOperand"); |
| } |
| |
| // { v0.d, v1.d } |
| def _dAsmOperand : TypedVecListAsmOperand<count, "AArch64Operand::VecListIdx_QReg", 0, 64>; |
| def "d" : TypedVecListRegOperand<Reg128, 0, "d"> { |
| let ParserMatchClass = !cast<AsmOperandClass>(NAME # "_dAsmOperand"); |
| } |
| |
| |
| } |
| |
| defm VecListOne : VectorList<1, FPR64, FPR128>; |
| defm VecListTwo : VectorList<2, DD, QQ>; |
| defm VecListThree : VectorList<3, DDD, QQQ>; |
| defm VecListFour : VectorList<4, DDDD, QQQQ>; |
| |
| class FPRAsmOperand<string RC> : AsmOperandClass { |
| let Name = "FPRAsmOperand" # RC; |
| let PredicateMethod = "isGPR64<AArch64::" # RC # "RegClassID>"; |
| let RenderMethod = "addRegOperands"; |
| } |
| |
| // Register operand versions of the scalar FP registers. |
| def FPR8Op : RegisterOperand<FPR8, "printOperand"> { |
| let ParserMatchClass = FPRAsmOperand<"FPR8">; |
| } |
| |
| def FPR16Op : RegisterOperand<FPR16, "printOperand"> { |
| let ParserMatchClass = FPRAsmOperand<"FPR16">; |
| } |
| |
| def FPR16Op_lo : RegisterOperand<FPR16_lo, "printOperand"> { |
| let ParserMatchClass = FPRAsmOperand<"FPR16_lo">; |
| } |
| |
| def FPR32Op : RegisterOperand<FPR32, "printOperand"> { |
| let ParserMatchClass = FPRAsmOperand<"FPR32">; |
| } |
| |
| def FPR64Op : RegisterOperand<FPR64, "printOperand"> { |
| let ParserMatchClass = FPRAsmOperand<"FPR64">; |
| } |
| |
| def FPR128Op : RegisterOperand<FPR128, "printOperand"> { |
| let ParserMatchClass = FPRAsmOperand<"FPR128">; |
| } |
| |
| //===----------------------------------------------------------------------===// |
| // ARMv8.1a atomic CASP register operands |
| |
| |
| def WSeqPairs : RegisterTuples<[sube32, subo32], |
| [(decimate (rotl GPR32, 0), 2), |
| (decimate (rotl GPR32, 1), 2)]>; |
| def XSeqPairs : RegisterTuples<[sube64, subo64], |
| [(decimate (rotl GPR64, 0), 2), |
| (decimate (rotl GPR64, 1), 2)]>; |
| |
| def WSeqPairsClass : RegisterClass<"AArch64", [untyped], 32, |
| (add WSeqPairs)>{ |
| let Size = 64; |
| } |
| def XSeqPairsClass : RegisterClass<"AArch64", [untyped], 64, |
| (add XSeqPairs)>{ |
| let Size = 128; |
| } |
| |
| |
| let RenderMethod = "addRegOperands", ParserMethod="tryParseGPRSeqPair" in { |
| def WSeqPairsAsmOperandClass : AsmOperandClass { let Name = "WSeqPair"; } |
| def XSeqPairsAsmOperandClass : AsmOperandClass { let Name = "XSeqPair"; } |
| } |
| |
| def WSeqPairClassOperand : |
| RegisterOperand<WSeqPairsClass, "printGPRSeqPairsClassOperand<32>"> { |
| let ParserMatchClass = WSeqPairsAsmOperandClass; |
| } |
| def XSeqPairClassOperand : |
| RegisterOperand<XSeqPairsClass, "printGPRSeqPairsClassOperand<64>"> { |
| let ParserMatchClass = XSeqPairsAsmOperandClass; |
| } |
| |
| |
| //===----- END: v8.1a atomic CASP register operands -----------------------===// |
| |
| //===----------------------------------------------------------------------===// |
| // Armv8.7a accelerator extension register operands: 8 consecutive GPRs |
| // starting with an even one |
| |
| let Namespace = "AArch64" in { |
| foreach i = 0-7 in |
| def "x8sub_"#i : SubRegIndex<64, !mul(64, i)>; |
| } |
| |
| def Tuples8X : RegisterTuples< |
| !foreach(i, [0,1,2,3,4,5,6,7], !cast<SubRegIndex>("x8sub_"#i)), |
| !foreach(i, [0,1,2,3,4,5,6,7], (trunc (decimate (rotl GPR64, i), 2), 12))>; |
| |
| def GPR64x8Class : RegisterClass<"AArch64", [i64x8], 512, (trunc Tuples8X, 12)> { |
| let Size = 512; |
| } |
| def GPR64x8AsmOp : AsmOperandClass { |
| let Name = "GPR64x8"; |
| let ParserMethod = "tryParseGPR64x8"; |
| let RenderMethod = "addRegOperands"; |
| } |
| def GPR64x8 : RegisterOperand<GPR64x8Class, "printGPR64x8"> { |
| let ParserMatchClass = GPR64x8AsmOp; |
| let PrintMethod = "printGPR64x8"; |
| } |
| |
| //===----- END: v8.7a accelerator extension register operands -------------===// |
| |
| // SVE predicate registers |
| def P0 : AArch64Reg<0, "p0">, DwarfRegNum<[48]>; |
| def P1 : AArch64Reg<1, "p1">, DwarfRegNum<[49]>; |
| def P2 : AArch64Reg<2, "p2">, DwarfRegNum<[50]>; |
| def P3 : AArch64Reg<3, "p3">, DwarfRegNum<[51]>; |
| def P4 : AArch64Reg<4, "p4">, DwarfRegNum<[52]>; |
| def P5 : AArch64Reg<5, "p5">, DwarfRegNum<[53]>; |
| def P6 : AArch64Reg<6, "p6">, DwarfRegNum<[54]>; |
| def P7 : AArch64Reg<7, "p7">, DwarfRegNum<[55]>; |
| def P8 : AArch64Reg<8, "p8">, DwarfRegNum<[56]>; |
| def P9 : AArch64Reg<9, "p9">, DwarfRegNum<[57]>; |
| def P10 : AArch64Reg<10, "p10">, DwarfRegNum<[58]>; |
| def P11 : AArch64Reg<11, "p11">, DwarfRegNum<[59]>; |
| def P12 : AArch64Reg<12, "p12">, DwarfRegNum<[60]>; |
| def P13 : AArch64Reg<13, "p13">, DwarfRegNum<[61]>; |
| def P14 : AArch64Reg<14, "p14">, DwarfRegNum<[62]>; |
| def P15 : AArch64Reg<15, "p15">, DwarfRegNum<[63]>; |
| |
| // The part of SVE registers that don't overlap Neon registers. |
| // These are only used as part of clobber lists. |
| def Z0_HI : AArch64Reg<0, "z0_hi">; |
| def Z1_HI : AArch64Reg<1, "z1_hi">; |
| def Z2_HI : AArch64Reg<2, "z2_hi">; |
| def Z3_HI : AArch64Reg<3, "z3_hi">; |
| def Z4_HI : AArch64Reg<4, "z4_hi">; |
| def Z5_HI : AArch64Reg<5, "z5_hi">; |
| def Z6_HI : AArch64Reg<6, "z6_hi">; |
| def Z7_HI : AArch64Reg<7, "z7_hi">; |
| def Z8_HI : AArch64Reg<8, "z8_hi">; |
| def Z9_HI : AArch64Reg<9, "z9_hi">; |
| def Z10_HI : AArch64Reg<10, "z10_hi">; |
| def Z11_HI : AArch64Reg<11, "z11_hi">; |
| def Z12_HI : AArch64Reg<12, "z12_hi">; |
| def Z13_HI : AArch64Reg<13, "z13_hi">; |
| def Z14_HI : AArch64Reg<14, "z14_hi">; |
| def Z15_HI : AArch64Reg<15, "z15_hi">; |
| def Z16_HI : AArch64Reg<16, "z16_hi">; |
| def Z17_HI : AArch64Reg<17, "z17_hi">; |
| def Z18_HI : AArch64Reg<18, "z18_hi">; |
| def Z19_HI : AArch64Reg<19, "z19_hi">; |
| def Z20_HI : AArch64Reg<20, "z20_hi">; |
| def Z21_HI : AArch64Reg<21, "z21_hi">; |
| def Z22_HI : AArch64Reg<22, "z22_hi">; |
| def Z23_HI : AArch64Reg<23, "z23_hi">; |
| def Z24_HI : AArch64Reg<24, "z24_hi">; |
| def Z25_HI : AArch64Reg<25, "z25_hi">; |
| def Z26_HI : AArch64Reg<26, "z26_hi">; |
| def Z27_HI : AArch64Reg<27, "z27_hi">; |
| def Z28_HI : AArch64Reg<28, "z28_hi">; |
| def Z29_HI : AArch64Reg<29, "z29_hi">; |
| def Z30_HI : AArch64Reg<30, "z30_hi">; |
| def Z31_HI : AArch64Reg<31, "z31_hi">; |
| |
| // SVE variable-size vector registers |
| let SubRegIndices = [zsub,zsub_hi] in { |
| def Z0 : AArch64Reg<0, "z0", [Q0, Z0_HI]>, DwarfRegNum<[96]>; |
| def Z1 : AArch64Reg<1, "z1", [Q1, Z1_HI]>, DwarfRegNum<[97]>; |
| def Z2 : AArch64Reg<2, "z2", [Q2, Z2_HI]>, DwarfRegNum<[98]>; |
| def Z3 : AArch64Reg<3, "z3", [Q3, Z3_HI]>, DwarfRegNum<[99]>; |
| def Z4 : AArch64Reg<4, "z4", [Q4, Z4_HI]>, DwarfRegNum<[100]>; |
| def Z5 : AArch64Reg<5, "z5", [Q5, Z5_HI]>, DwarfRegNum<[101]>; |
| def Z6 : AArch64Reg<6, "z6", [Q6, Z6_HI]>, DwarfRegNum<[102]>; |
| def Z7 : AArch64Reg<7, "z7", [Q7, Z7_HI]>, DwarfRegNum<[103]>; |
| def Z8 : AArch64Reg<8, "z8", [Q8, Z8_HI]>, DwarfRegNum<[104]>; |
| def Z9 : AArch64Reg<9, "z9", [Q9, Z9_HI]>, DwarfRegNum<[105]>; |
| def Z10 : AArch64Reg<10, "z10", [Q10, Z10_HI]>, DwarfRegNum<[106]>; |
| def Z11 : AArch64Reg<11, "z11", [Q11, Z11_HI]>, DwarfRegNum<[107]>; |
| def Z12 : AArch64Reg<12, "z12", [Q12, Z12_HI]>, DwarfRegNum<[108]>; |
| def Z13 : AArch64Reg<13, "z13", [Q13, Z13_HI]>, DwarfRegNum<[109]>; |
| def Z14 : AArch64Reg<14, "z14", [Q14, Z14_HI]>, DwarfRegNum<[110]>; |
| def Z15 : AArch64Reg<15, "z15", [Q15, Z15_HI]>, DwarfRegNum<[111]>; |
| def Z16 : AArch64Reg<16, "z16", [Q16, Z16_HI]>, DwarfRegNum<[112]>; |
| def Z17 : AArch64Reg<17, "z17", [Q17, Z17_HI]>, DwarfRegNum<[113]>; |
| def Z18 : AArch64Reg<18, "z18", [Q18, Z18_HI]>, DwarfRegNum<[114]>; |
| def Z19 : AArch64Reg<19, "z19", [Q19, Z19_HI]>, DwarfRegNum<[115]>; |
| def Z20 : AArch64Reg<20, "z20", [Q20, Z20_HI]>, DwarfRegNum<[116]>; |
| def Z21 : AArch64Reg<21, "z21", [Q21, Z21_HI]>, DwarfRegNum<[117]>; |
| def Z22 : AArch64Reg<22, "z22", [Q22, Z22_HI]>, DwarfRegNum<[118]>; |
| def Z23 : AArch64Reg<23, "z23", [Q23, Z23_HI]>, DwarfRegNum<[119]>; |
| def Z24 : AArch64Reg<24, "z24", [Q24, Z24_HI]>, DwarfRegNum<[120]>; |
| def Z25 : AArch64Reg<25, "z25", [Q25, Z25_HI]>, DwarfRegNum<[121]>; |
| def Z26 : AArch64Reg<26, "z26", [Q26, Z26_HI]>, DwarfRegNum<[122]>; |
| def Z27 : AArch64Reg<27, "z27", [Q27, Z27_HI]>, DwarfRegNum<[123]>; |
| def Z28 : AArch64Reg<28, "z28", [Q28, Z28_HI]>, DwarfRegNum<[124]>; |
| def Z29 : AArch64Reg<29, "z29", [Q29, Z29_HI]>, DwarfRegNum<[125]>; |
| def Z30 : AArch64Reg<30, "z30", [Q30, Z30_HI]>, DwarfRegNum<[126]>; |
| def Z31 : AArch64Reg<31, "z31", [Q31, Z31_HI]>, DwarfRegNum<[127]>; |
| } |
| |
| // Enum describing the element size for destructive |
| // operations. |
| class ElementSizeEnum<bits<3> val> { |
| bits<3> Value = val; |
| } |
| |
| def ElementSizeNone : ElementSizeEnum<0>; |
| def ElementSizeB : ElementSizeEnum<1>; |
| def ElementSizeH : ElementSizeEnum<2>; |
| def ElementSizeS : ElementSizeEnum<3>; |
| def ElementSizeD : ElementSizeEnum<4>; |
| def ElementSizeQ : ElementSizeEnum<5>; // Unused |
| |
| class SVERegOp <string Suffix, AsmOperandClass C, |
| ElementSizeEnum Size, |
| RegisterClass RC> : RegisterOperand<RC> { |
| ElementSizeEnum ElementSize; |
| |
| let ElementSize = Size; |
| let PrintMethod = !if(!eq(Suffix, ""), |
| "printSVERegOp<>", |
| "printSVERegOp<'" # Suffix # "'>"); |
| let ParserMatchClass = C; |
| } |
| |
| class PPRRegOp <string Suffix, AsmOperandClass C, ElementSizeEnum Size, |
| RegisterClass RC> : SVERegOp<Suffix, C, Size, RC> {} |
| class ZPRRegOp <string Suffix, AsmOperandClass C, ElementSizeEnum Size, |
| RegisterClass RC> : SVERegOp<Suffix, C, Size, RC> {} |
| |
| //****************************************************************************** |
| |
| // SVE predicate register classes. |
| class PPRClass<int lastreg> : RegisterClass< |
| "AArch64", |
| [ nxv16i1, nxv8i1, nxv4i1, nxv2i1 ], 16, |
| (sequence "P%u", 0, lastreg)> { |
| let Size = 16; |
| } |
| |
| def PPR : PPRClass<15>; |
| def PPR_3b : PPRClass<7>; // Restricted 3 bit SVE predicate register class. |
| |
| class PPRAsmOperand <string name, string RegClass, int Width>: AsmOperandClass { |
| let Name = "SVE" # name # "Reg"; |
| let PredicateMethod = "isSVEPredicateVectorRegOfWidth<" |
| # Width # ", " # "AArch64::" # RegClass # "RegClassID>"; |
| let DiagnosticType = "InvalidSVE" # name # "Reg"; |
| let RenderMethod = "addRegOperands"; |
| let ParserMethod = "tryParseSVEPredicateVector"; |
| } |
| |
| def PPRAsmOpAny : PPRAsmOperand<"PredicateAny", "PPR", 0>; |
| def PPRAsmOp8 : PPRAsmOperand<"PredicateB", "PPR", 8>; |
| def PPRAsmOp16 : PPRAsmOperand<"PredicateH", "PPR", 16>; |
| def PPRAsmOp32 : PPRAsmOperand<"PredicateS", "PPR", 32>; |
| def PPRAsmOp64 : PPRAsmOperand<"PredicateD", "PPR", 64>; |
| |
| def PPRAny : PPRRegOp<"", PPRAsmOpAny, ElementSizeNone, PPR>; |
| def PPR8 : PPRRegOp<"b", PPRAsmOp8, ElementSizeB, PPR>; |
| def PPR16 : PPRRegOp<"h", PPRAsmOp16, ElementSizeH, PPR>; |
| def PPR32 : PPRRegOp<"s", PPRAsmOp32, ElementSizeS, PPR>; |
| def PPR64 : PPRRegOp<"d", PPRAsmOp64, ElementSizeD, PPR>; |
| |
| def PPRAsmOp3bAny : PPRAsmOperand<"Predicate3bAny", "PPR_3b", 0>; |
| |
| def PPR3bAny : PPRRegOp<"", PPRAsmOp3bAny, ElementSizeNone, PPR_3b>; |
| |
| //****************************************************************************** |
| |
| // SVE vector register classes |
| class ZPRClass<int lastreg> : RegisterClass<"AArch64", |
| [nxv16i8, nxv8i16, nxv4i32, nxv2i64, |
| nxv2f16, nxv4f16, nxv8f16, |
| nxv2bf16, nxv4bf16, nxv8bf16, |
| nxv2f32, nxv4f32, |
| nxv2f64], |
| 128, (sequence "Z%u", 0, lastreg)> { |
| let Size = 128; |
| } |
| |
| def ZPR : ZPRClass<31>; |
| def ZPR_4b : ZPRClass<15>; // Restricted 4 bit SVE vector register class. |
| def ZPR_3b : ZPRClass<7>; // Restricted 3 bit SVE vector register class. |
| |
| class ZPRAsmOperand<string name, int Width, string RegClassSuffix = ""> |
| : AsmOperandClass { |
| let Name = "SVE" # name # "Reg"; |
| let PredicateMethod = "isSVEDataVectorRegOfWidth<" |
| # Width # ", AArch64::ZPR" |
| # RegClassSuffix # "RegClassID>"; |
| let RenderMethod = "addRegOperands"; |
| let DiagnosticType = "InvalidZPR" # RegClassSuffix # Width; |
| let ParserMethod = "tryParseSVEDataVector<false, " |
| # !if(!eq(Width, 0), "false", "true") # ">"; |
| } |
| |
| def ZPRAsmOpAny : ZPRAsmOperand<"VectorAny", 0>; |
| def ZPRAsmOp8 : ZPRAsmOperand<"VectorB", 8>; |
| def ZPRAsmOp16 : ZPRAsmOperand<"VectorH", 16>; |
| def ZPRAsmOp32 : ZPRAsmOperand<"VectorS", 32>; |
| def ZPRAsmOp64 : ZPRAsmOperand<"VectorD", 64>; |
| def ZPRAsmOp128 : ZPRAsmOperand<"VectorQ", 128>; |
| |
| def ZPRAny : ZPRRegOp<"", ZPRAsmOpAny, ElementSizeNone, ZPR>; |
| def ZPR8 : ZPRRegOp<"b", ZPRAsmOp8, ElementSizeB, ZPR>; |
| def ZPR16 : ZPRRegOp<"h", ZPRAsmOp16, ElementSizeH, ZPR>; |
| def ZPR32 : ZPRRegOp<"s", ZPRAsmOp32, ElementSizeS, ZPR>; |
| def ZPR64 : ZPRRegOp<"d", ZPRAsmOp64, ElementSizeD, ZPR>; |
| def ZPR128 : ZPRRegOp<"q", ZPRAsmOp128, ElementSizeQ, ZPR>; |
| |
| def ZPRAsmOp3b8 : ZPRAsmOperand<"Vector3bB", 8, "_3b">; |
| def ZPRAsmOp3b16 : ZPRAsmOperand<"Vector3bH", 16, "_3b">; |
| def ZPRAsmOp3b32 : ZPRAsmOperand<"Vector3bS", 32, "_3b">; |
| |
| def ZPR3b8 : ZPRRegOp<"b", ZPRAsmOp3b8, ElementSizeB, ZPR_3b>; |
| def ZPR3b16 : ZPRRegOp<"h", ZPRAsmOp3b16, ElementSizeH, ZPR_3b>; |
| def ZPR3b32 : ZPRRegOp<"s", ZPRAsmOp3b32, ElementSizeS, ZPR_3b>; |
| |
| def ZPRAsmOp4b16 : ZPRAsmOperand<"Vector4bH", 16, "_4b">; |
| def ZPRAsmOp4b32 : ZPRAsmOperand<"Vector4bS", 32, "_4b">; |
| def ZPRAsmOp4b64 : ZPRAsmOperand<"Vector4bD", 64, "_4b">; |
| |
| def ZPR4b16 : ZPRRegOp<"h", ZPRAsmOp4b16, ElementSizeH, ZPR_4b>; |
| def ZPR4b32 : ZPRRegOp<"s", ZPRAsmOp4b32, ElementSizeS, ZPR_4b>; |
| def ZPR4b64 : ZPRRegOp<"d", ZPRAsmOp4b64, ElementSizeD, ZPR_4b>; |
| |
| class FPRasZPR<int Width> : AsmOperandClass{ |
| let Name = "FPR" # Width # "asZPR"; |
| let PredicateMethod = "isFPRasZPR<AArch64::FPR" # Width # "RegClassID>"; |
| let RenderMethod = "addFPRasZPRRegOperands<" # Width # ">"; |
| } |
| |
| class FPRasZPROperand<int Width> : RegisterOperand<ZPR> { |
| let ParserMatchClass = FPRasZPR<Width>; |
| let PrintMethod = "printZPRasFPR<" # Width # ">"; |
| } |
| |
| def FPR8asZPR : FPRasZPROperand<8>; |
| def FPR16asZPR : FPRasZPROperand<16>; |
| def FPR32asZPR : FPRasZPROperand<32>; |
| def FPR64asZPR : FPRasZPROperand<64>; |
| def FPR128asZPR : FPRasZPROperand<128>; |
| |
| let Namespace = "AArch64" in { |
| def zsub0 : SubRegIndex<128, -1>; |
| def zsub1 : SubRegIndex<128, -1>; |
| def zsub2 : SubRegIndex<128, -1>; |
| def zsub3 : SubRegIndex<128, -1>; |
| } |
| |
| // Pairs, triples, and quads of SVE vector registers. |
| def ZSeqPairs : RegisterTuples<[zsub0, zsub1], [(rotl ZPR, 0), (rotl ZPR, 1)]>; |
| def ZSeqTriples : RegisterTuples<[zsub0, zsub1, zsub2], [(rotl ZPR, 0), (rotl ZPR, 1), (rotl ZPR, 2)]>; |
| def ZSeqQuads : RegisterTuples<[zsub0, zsub1, zsub2, zsub3], [(rotl ZPR, 0), (rotl ZPR, 1), (rotl ZPR, 2), (rotl ZPR, 3)]>; |
| |
| def ZPR2 : RegisterClass<"AArch64", [untyped], 128, (add ZSeqPairs)> { |
| let Size = 256; |
| } |
| def ZPR3 : RegisterClass<"AArch64", [untyped], 128, (add ZSeqTriples)> { |
| let Size = 384; |
| } |
| def ZPR4 : RegisterClass<"AArch64", [untyped], 128, (add ZSeqQuads)> { |
| let Size = 512; |
| } |
| |
| class ZPRVectorList<int ElementWidth, int NumRegs> : AsmOperandClass { |
| let Name = "SVEVectorList" # NumRegs # ElementWidth; |
| let ParserMethod = "tryParseVectorList<RegKind::SVEDataVector>"; |
| let PredicateMethod = |
| "isTypedVectorList<RegKind::SVEDataVector, " #NumRegs #", 0, " #ElementWidth #">"; |
| let RenderMethod = "addVectorListOperands<AArch64Operand::VecListIdx_ZReg, " # NumRegs # ">"; |
| } |
| |
| def Z_b : RegisterOperand<ZPR, "printTypedVectorList<0,'b'>"> { |
| let ParserMatchClass = ZPRVectorList<8, 1>; |
| } |
| |
| def Z_h : RegisterOperand<ZPR, "printTypedVectorList<0,'h'>"> { |
| let ParserMatchClass = ZPRVectorList<16, 1>; |
| } |
| |
| def Z_s : RegisterOperand<ZPR, "printTypedVectorList<0,'s'>"> { |
| let ParserMatchClass = ZPRVectorList<32, 1>; |
| } |
| |
| def Z_d : RegisterOperand<ZPR, "printTypedVectorList<0,'d'>"> { |
| let ParserMatchClass = ZPRVectorList<64, 1>; |
| } |
| |
| def ZZ_b : RegisterOperand<ZPR2, "printTypedVectorList<0,'b'>"> { |
| let ParserMatchClass = ZPRVectorList<8, 2>; |
| } |
| |
| def ZZ_h : RegisterOperand<ZPR2, "printTypedVectorList<0,'h'>"> { |
| let ParserMatchClass = ZPRVectorList<16, 2>; |
| } |
| |
| def ZZ_s : RegisterOperand<ZPR2, "printTypedVectorList<0,'s'>"> { |
| let ParserMatchClass = ZPRVectorList<32, 2>; |
| } |
| |
| def ZZ_d : RegisterOperand<ZPR2, "printTypedVectorList<0,'d'>"> { |
| let ParserMatchClass = ZPRVectorList<64, 2>; |
| } |
| |
| def ZZZ_b : RegisterOperand<ZPR3, "printTypedVectorList<0,'b'>"> { |
| let ParserMatchClass = ZPRVectorList<8, 3>; |
| } |
| |
| def ZZZ_h : RegisterOperand<ZPR3, "printTypedVectorList<0,'h'>"> { |
| let ParserMatchClass = ZPRVectorList<16, 3>; |
| } |
| |
| def ZZZ_s : RegisterOperand<ZPR3, "printTypedVectorList<0,'s'>"> { |
| let ParserMatchClass = ZPRVectorList<32, 3>; |
| } |
| |
| def ZZZ_d : RegisterOperand<ZPR3, "printTypedVectorList<0,'d'>"> { |
| let ParserMatchClass = ZPRVectorList<64, 3>; |
| } |
| |
| def ZZZZ_b : RegisterOperand<ZPR4, "printTypedVectorList<0,'b'>"> { |
| let ParserMatchClass = ZPRVectorList<8, 4>; |
| } |
| |
| def ZZZZ_h : RegisterOperand<ZPR4, "printTypedVectorList<0,'h'>"> { |
| let ParserMatchClass = ZPRVectorList<16, 4>; |
| } |
| |
| def ZZZZ_s : RegisterOperand<ZPR4, "printTypedVectorList<0,'s'>"> { |
| let ParserMatchClass = ZPRVectorList<32, 4>; |
| } |
| |
| def ZZZZ_d : RegisterOperand<ZPR4, "printTypedVectorList<0,'d'>"> { |
| let ParserMatchClass = ZPRVectorList<64, 4>; |
| } |
| |
| class ZPRExtendAsmOperand<string ShiftExtend, int RegWidth, int Scale, |
| bit ScaleAlwaysSame = 0b0> : AsmOperandClass { |
| let Name = "ZPRExtend" # ShiftExtend # RegWidth # Scale |
| # !if(ScaleAlwaysSame, "Only", ""); |
| |
| let PredicateMethod = "isSVEDataVectorRegWithShiftExtend<" |
| # RegWidth # ", AArch64::ZPRRegClassID, " |
| # "AArch64_AM::" # ShiftExtend # ", " |
| # Scale # ", " |
| # !if(ScaleAlwaysSame, "true", "false") |
| # ">"; |
| let DiagnosticType = "InvalidZPR" # RegWidth # ShiftExtend # Scale; |
| let RenderMethod = "addRegOperands"; |
| let ParserMethod = "tryParseSVEDataVector<true, true>"; |
| } |
| |
| class ZPRExtendRegisterOperand<bit SignExtend, bit IsLSL, string Repr, |
| int RegWidth, int Scale, string Suffix = ""> |
| : RegisterOperand<ZPR> { |
| let ParserMatchClass = |
| !cast<AsmOperandClass>("ZPR" # RegWidth # "AsmOpndExt" # Repr # Scale # Suffix); |
| let PrintMethod = "printRegWithShiftExtend<" |
| # !if(SignExtend, "true", "false") # ", " |
| # Scale # ", " |
| # !if(IsLSL, "'x'", "'w'") # ", " |
| # !if(!eq(RegWidth, 32), "'s'", "'d'") # ">"; |
| } |
| |
| foreach RegWidth = [32, 64] in { |
| // UXTW(8|16|32|64) |
| def ZPR#RegWidth#AsmOpndExtUXTW8Only : ZPRExtendAsmOperand<"UXTW", RegWidth, 8, 0b1>; |
| def ZPR#RegWidth#AsmOpndExtUXTW8 : ZPRExtendAsmOperand<"UXTW", RegWidth, 8>; |
| def ZPR#RegWidth#AsmOpndExtUXTW16 : ZPRExtendAsmOperand<"UXTW", RegWidth, 16>; |
| def ZPR#RegWidth#AsmOpndExtUXTW32 : ZPRExtendAsmOperand<"UXTW", RegWidth, 32>; |
| def ZPR#RegWidth#AsmOpndExtUXTW64 : ZPRExtendAsmOperand<"UXTW", RegWidth, 64>; |
| |
| def ZPR#RegWidth#ExtUXTW8Only : ZPRExtendRegisterOperand<0b0, 0b0, "UXTW", RegWidth, 8, "Only">; |
| def ZPR#RegWidth#ExtUXTW8 : ZPRExtendRegisterOperand<0b0, 0b0, "UXTW", RegWidth, 8>; |
| def ZPR#RegWidth#ExtUXTW16 : ZPRExtendRegisterOperand<0b0, 0b0, "UXTW", RegWidth, 16>; |
| def ZPR#RegWidth#ExtUXTW32 : ZPRExtendRegisterOperand<0b0, 0b0, "UXTW", RegWidth, 32>; |
| def ZPR#RegWidth#ExtUXTW64 : ZPRExtendRegisterOperand<0b0, 0b0, "UXTW", RegWidth, 64>; |
| |
| // SXTW(8|16|32|64) |
| def ZPR#RegWidth#AsmOpndExtSXTW8Only : ZPRExtendAsmOperand<"SXTW", RegWidth, 8, 0b1>; |
| def ZPR#RegWidth#AsmOpndExtSXTW8 : ZPRExtendAsmOperand<"SXTW", RegWidth, 8>; |
| def ZPR#RegWidth#AsmOpndExtSXTW16 : ZPRExtendAsmOperand<"SXTW", RegWidth, 16>; |
| def ZPR#RegWidth#AsmOpndExtSXTW32 : ZPRExtendAsmOperand<"SXTW", RegWidth, 32>; |
| def ZPR#RegWidth#AsmOpndExtSXTW64 : ZPRExtendAsmOperand<"SXTW", RegWidth, 64>; |
| |
| def ZPR#RegWidth#ExtSXTW8Only : ZPRExtendRegisterOperand<0b1, 0b0, "SXTW", RegWidth, 8, "Only">; |
| def ZPR#RegWidth#ExtSXTW8 : ZPRExtendRegisterOperand<0b1, 0b0, "SXTW", RegWidth, 8>; |
| def ZPR#RegWidth#ExtSXTW16 : ZPRExtendRegisterOperand<0b1, 0b0, "SXTW", RegWidth, 16>; |
| def ZPR#RegWidth#ExtSXTW32 : ZPRExtendRegisterOperand<0b1, 0b0, "SXTW", RegWidth, 32>; |
| def ZPR#RegWidth#ExtSXTW64 : ZPRExtendRegisterOperand<0b1, 0b0, "SXTW", RegWidth, 64>; |
| |
| // LSL(8|16|32|64) |
| def ZPR#RegWidth#AsmOpndExtLSL8 : ZPRExtendAsmOperand<"LSL", RegWidth, 8>; |
| def ZPR#RegWidth#AsmOpndExtLSL16 : ZPRExtendAsmOperand<"LSL", RegWidth, 16>; |
| def ZPR#RegWidth#AsmOpndExtLSL32 : ZPRExtendAsmOperand<"LSL", RegWidth, 32>; |
| def ZPR#RegWidth#AsmOpndExtLSL64 : ZPRExtendAsmOperand<"LSL", RegWidth, 64>; |
| def ZPR#RegWidth#ExtLSL8 : ZPRExtendRegisterOperand<0b0, 0b1, "LSL", RegWidth, 8>; |
| def ZPR#RegWidth#ExtLSL16 : ZPRExtendRegisterOperand<0b0, 0b1, "LSL", RegWidth, 16>; |
| def ZPR#RegWidth#ExtLSL32 : ZPRExtendRegisterOperand<0b0, 0b1, "LSL", RegWidth, 32>; |
| def ZPR#RegWidth#ExtLSL64 : ZPRExtendRegisterOperand<0b0, 0b1, "LSL", RegWidth, 64>; |
| } |
| |
| class GPR64ShiftExtendAsmOperand <string AsmOperandName, int Scale, string RegClass> : AsmOperandClass { |
| let Name = AsmOperandName # Scale; |
| let PredicateMethod = "isGPR64WithShiftExtend<AArch64::"#RegClass#"RegClassID, " # Scale # ">"; |
| let DiagnosticType = "Invalid" # AsmOperandName # Scale; |
| let RenderMethod = "addRegOperands"; |
| let ParserMethod = "tryParseGPROperand<true>"; |
| } |
| |
| class GPR64ExtendRegisterOperand<string Name, int Scale, RegisterClass RegClass> : RegisterOperand<RegClass>{ |
| let ParserMatchClass = !cast<AsmOperandClass>(Name); |
| let PrintMethod = "printRegWithShiftExtend<false, " # Scale # ", 'x', 0>"; |
| } |
| |
| foreach Scale = [8, 16, 32, 64, 128] in { |
| def GPR64shiftedAsmOpnd # Scale : GPR64ShiftExtendAsmOperand<"GPR64shifted", Scale, "GPR64">; |
| def GPR64shifted # Scale : GPR64ExtendRegisterOperand<"GPR64shiftedAsmOpnd" # Scale, Scale, GPR64>; |
| |
| def GPR64NoXZRshiftedAsmOpnd # Scale : GPR64ShiftExtendAsmOperand<"GPR64NoXZRshifted", Scale, "GPR64common">; |
| def GPR64NoXZRshifted # Scale : GPR64ExtendRegisterOperand<"GPR64NoXZRshiftedAsmOpnd" # Scale, Scale, GPR64common>; |
| } |
| |
| // Accumulator array tiles. |
| def ZAQ0 : AArch64Reg<0, "za0.q">; |
| def ZAQ1 : AArch64Reg<1, "za1.q">; |
| def ZAQ2 : AArch64Reg<2, "za2.q">; |
| def ZAQ3 : AArch64Reg<3, "za3.q">; |
| def ZAQ4 : AArch64Reg<4, "za4.q">; |
| def ZAQ5 : AArch64Reg<5, "za5.q">; |
| def ZAQ6 : AArch64Reg<6, "za6.q">; |
| def ZAQ7 : AArch64Reg<7, "za7.q">; |
| def ZAQ8 : AArch64Reg<8, "za8.q">; |
| def ZAQ9 : AArch64Reg<9, "za9.q">; |
| def ZAQ10 : AArch64Reg<10, "za10.q">; |
| def ZAQ11 : AArch64Reg<11, "za11.q">; |
| def ZAQ12 : AArch64Reg<12, "za12.q">; |
| def ZAQ13 : AArch64Reg<13, "za13.q">; |
| def ZAQ14 : AArch64Reg<14, "za14.q">; |
| def ZAQ15 : AArch64Reg<15, "za15.q">; |
| |
| let SubRegIndices = [zasubq0, zasubq1] in { |
| def ZAD0 : AArch64Reg<0, "za0.d", [ZAQ0, ZAQ8]>; |
| def ZAD1 : AArch64Reg<1, "za1.d", [ZAQ1, ZAQ9]>; |
| def ZAD2 : AArch64Reg<2, "za2.d", [ZAQ2, ZAQ10]>; |
| def ZAD3 : AArch64Reg<3, "za3.d", [ZAQ3, ZAQ11]>; |
| def ZAD4 : AArch64Reg<4, "za4.d", [ZAQ4, ZAQ12]>; |
| def ZAD5 : AArch64Reg<5, "za5.d", [ZAQ5, ZAQ13]>; |
| def ZAD6 : AArch64Reg<6, "za6.d", [ZAQ6, ZAQ14]>; |
| def ZAD7 : AArch64Reg<7, "za7.d", [ZAQ7, ZAQ15]>; |
| } |
| |
| let SubRegIndices = [zasubd0, zasubd1] in { |
| def ZAS0 : AArch64Reg<0, "za0.s", [ZAD0, ZAD4]>; |
| def ZAS1 : AArch64Reg<1, "za1.s", [ZAD1, ZAD5]>; |
| def ZAS2 : AArch64Reg<2, "za2.s", [ZAD2, ZAD6]>; |
| def ZAS3 : AArch64Reg<3, "za3.s", [ZAD3, ZAD7]>; |
| } |
| |
| let SubRegIndices = [zasubs0, zasubs1] in { |
| def ZAH0 : AArch64Reg<0, "za0.h", [ZAS0, ZAS2]>; |
| def ZAH1 : AArch64Reg<1, "za1.h", [ZAS1, ZAS3]>; |
| } |
| |
| let SubRegIndices = [zasubh0, zasubh1] in { |
| def ZAB0 : AArch64Reg<0, "za0.b", [ZAH0, ZAH1]>; |
| } |
| |
| let SubRegIndices = [zasubb] in { |
| def ZA : AArch64Reg<0, "za", [ZAB0]>; |
| } |
| |
| // SME Register Classes |
| |
| // Accumulator array |
| def MPR : RegisterClass<"AArch64", [untyped], 2048, (add ZA)> { |
| let Size = 2048; |
| } |
| |
| // Accumulator array as single tiles |
| def MPR8 : RegisterClass<"AArch64", [untyped], 2048, (add (sequence "ZAB%u", 0, 0))> { |
| let Size = 2048; |
| } |
| def MPR16 : RegisterClass<"AArch64", [untyped], 1024, (add (sequence "ZAH%u", 0, 1))> { |
| let Size = 1024; |
| } |
| def MPR32 : RegisterClass<"AArch64", [untyped], 512, (add (sequence "ZAS%u", 0, 3))> { |
| let Size = 512; |
| } |
| def MPR64 : RegisterClass<"AArch64", [untyped], 256, (add (sequence "ZAD%u", 0, 7))> { |
| let Size = 256; |
| } |
| def MPR128 : RegisterClass<"AArch64", [untyped], 128, (add (sequence "ZAQ%u", 0, 15))> { |
| let Size = 128; |
| } |
| |
| // SME Register Operands |
| // There are three types of SME matrix register operands: |
| // * Tiles: |
| // |
| // These tiles make up the larger accumulator matrix. The tile representation |
| // has an element type suffix, e.g. za0.b or za15.q and can be any of the |
| // registers: |
| // ZAQ0..ZAQ15 |
| // ZAD0..ZAD7 |
| // ZAS0..ZAS3 |
| // ZAH0..ZAH1 |
| // or ZAB0 |
| // |
| // * Tile vectors: |
| // |
| // Their representation is similar to regular tiles, but they have an extra |
| // 'h' or 'v' to tell how the vector at [reg+offset] is layed out in the tile, |
| // horizontally or vertically. |
| // |
| // e.g. za1h.h or za15v.q, which corresponds to vectors in registers ZAH1 and |
| // ZAQ15, respectively. The horizontal/vertical is more a property of the |
| // instruction, than a property of the asm-operand itself, or its register. |
| // The distinction is required for the parsing/printing of the operand, |
| // as from a compiler's perspective, the whole tile is read/written. |
| // |
| // * Accumulator matrix: |
| // |
| // This is the entire matrix accumulator register ZA (<=> ZAB0), printed as |
| // 'za'. |
| |
| // |
| // Tiles |
| // |
| |
| class MatrixTileAsmOperand<string RC, int EltSize> : AsmOperandClass { |
| let Name = "MatrixTile" # EltSize; |
| let DiagnosticType = "Invalid" # Name; |
| let ParserMethod = "tryParseMatrixRegister"; |
| let RenderMethod = "addMatrixOperands"; |
| let PredicateMethod = "isMatrixRegOperand<" |
| # "MatrixKind::Tile" # ", " |
| # EltSize # ", AArch64::" # RC # "RegClassID>"; |
| } |
| |
| class MatrixTileOperand<int EltSize, int NumBitsForTile, RegisterClass RC> |
| : RegisterOperand<RC> { |
| let ParserMatchClass = MatrixTileAsmOperand<!cast<string>(RC), EltSize>; |
| let DecoderMethod = "DecodeMatrixTile<" # NumBitsForTile # ">"; |
| let PrintMethod = "printMatrixTile"; |
| } |
| |
| def TileOp32 : MatrixTileOperand<32, 2, MPR32>; |
| def TileOp64 : MatrixTileOperand<64, 3, MPR64>; |
| |
| // |
| // Tile vectors (horizontal and vertical) |
| // |
| |
| class MatrixTileVectorAsmOperand<string RC, int EltSize, int IsVertical> |
| : AsmOperandClass { |
| let Name = "MatrixTileVector" # !if(IsVertical, "V", "H") # EltSize; |
| let DiagnosticType = "Invalid" # Name; |
| let ParserMethod = "tryParseMatrixRegister"; |
| let RenderMethod = "addMatrixOperands"; |
| let PredicateMethod = "isMatrixRegOperand<" |
| # "MatrixKind::" |
| # !if(IsVertical, "Col", "Row") # ", " |
| # EltSize # ", AArch64::" # RC # "RegClassID>"; |
| } |
| |
| class MatrixTileVectorOperand<int EltSize, int NumBitsForTile, |
| RegisterClass RC, int IsVertical> |
| : RegisterOperand<RC> { |
| let ParserMatchClass = MatrixTileVectorAsmOperand<!cast<string>(RC), EltSize, |
| IsVertical>; |
| let DecoderMethod = "DecodeMatrixTile<" # NumBitsForTile # ">"; |
| let PrintMethod = "printMatrixTileVector<" # IsVertical # ">"; |
| } |
| |
| def TileVectorOpH8 : MatrixTileVectorOperand< 8, 0, MPR8, 0>; |
| def TileVectorOpH16 : MatrixTileVectorOperand< 16, 1, MPR16, 0>; |
| def TileVectorOpH32 : MatrixTileVectorOperand< 32, 2, MPR32, 0>; |
| def TileVectorOpH64 : MatrixTileVectorOperand< 64, 3, MPR64, 0>; |
| def TileVectorOpH128 : MatrixTileVectorOperand<128, 4, MPR128, 0>; |
| |
| def TileVectorOpV8 : MatrixTileVectorOperand< 8, 0, MPR8, 1>; |
| def TileVectorOpV16 : MatrixTileVectorOperand< 16, 1, MPR16, 1>; |
| def TileVectorOpV32 : MatrixTileVectorOperand< 32, 2, MPR32, 1>; |
| def TileVectorOpV64 : MatrixTileVectorOperand< 64, 3, MPR64, 1>; |
| def TileVectorOpV128 : MatrixTileVectorOperand<128, 4, MPR128, 1>; |
| |
| // |
| // Accumulator matrix |
| // |
| |
| class MatrixAsmOperand<string RC, int EltSize> : AsmOperandClass { |
| let Name = "Matrix"; |
| let DiagnosticType = "Invalid" # Name; |
| let ParserMethod = "tryParseMatrixRegister"; |
| let RenderMethod = "addMatrixOperands"; |
| let PredicateMethod = "isMatrixRegOperand<" |
| # "MatrixKind::Array" # ", " |
| # EltSize # ", AArch64::" # RC # "RegClassID>"; |
| } |
| |
| class MatrixOperand<RegisterClass RC, int EltSize> : RegisterOperand<RC> { |
| let ParserMatchClass = MatrixAsmOperand<!cast<string>(RC), EltSize>; |
| let PrintMethod = "printMatrix<" # EltSize # ">"; |
| } |
| |
| def MatrixOp : MatrixOperand<MPR, 0>; |
| |
| class MatrixTileListAsmOperand : AsmOperandClass { |
| let Name = "MatrixTileList"; |
| let ParserMethod = "tryParseMatrixTileList"; |
| let RenderMethod = "addMatrixTileListOperands"; |
| let PredicateMethod = "isMatrixTileList"; |
| } |
| |
| class MatrixTileListOperand : Operand<i8> { |
| let ParserMatchClass = MatrixTileListAsmOperand<>; |
| let DecoderMethod = "DecodeMatrixTileListRegisterClass"; |
| let EncoderMethod = "EncodeMatrixTileListRegisterClass"; |
| let PrintMethod = "printMatrixTileList"; |
| } |
| |
| def MatrixTileList : MatrixTileListOperand<>; |
| |
| def MatrixIndexGPR32_12_15 : RegisterClass<"AArch64", [i32], 32, (sequence "W%u", 12, 15)> { |
| let DiagnosticType = "InvalidMatrixIndexGPR32_12_15"; |
| } |
| def MatrixIndexGPR32Op12_15 : RegisterOperand<MatrixIndexGPR32_12_15> { |
| let EncoderMethod = "encodeMatrixIndexGPR32"; |
| } |
| |
| def SVCROperand : AsmOperandClass { |
| let Name = "SVCR"; |
| let ParserMethod = "tryParseSVCR"; |
| let DiagnosticType = "Invalid" # Name; |
| } |
| |
| def svcr_op : Operand<i32> { |
| let ParserMatchClass = SVCROperand; |
| let PrintMethod = "printSVCROp"; |
| let DecoderMethod = "DecodeSVCROp"; |
| let MCOperandPredicate = [{ |
| if (!MCOp.isImm()) |
| return false; |
| return AArch64SVCR::lookupSVCRByEncoding(MCOp.getImm()) != nullptr; |
| }]; |
| } |