blob: 235538ff3ddb38910f3afb61393c92b944fc6419 [file] [log] [blame]
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
; RUN: llc < %s | FileCheck %s
target triple = "aarch64-unknown-linux-gnu"
; Ensure we don't crash by trying to fold fixed length frame indexes into
; loads/stores that don't support an appropriate addressing mode, hence creating
; too many extra vregs during frame lowering, when we don't have an emergency
; spill slot.
define dso_local void @func1(i64* %v1, i64* %v2, i64* %v3, i64* %v4, i64* %v5, i64* %v6, i64* %v7, i64* %v8,
; CHECK-LABEL: func1:
; CHECK: // %bb.0:
; CHECK-NEXT: str x25, [sp, #-64]! // 8-byte Folded Spill
; CHECK-NEXT: .cfi_def_cfa_offset 64
; CHECK-NEXT: stp x24, x23, [sp, #16] // 16-byte Folded Spill
; CHECK-NEXT: stp x22, x21, [sp, #32] // 16-byte Folded Spill
; CHECK-NEXT: stp x20, x19, [sp, #48] // 16-byte Folded Spill
; CHECK-NEXT: .cfi_offset w19, -8
; CHECK-NEXT: .cfi_offset w20, -16
; CHECK-NEXT: .cfi_offset w21, -24
; CHECK-NEXT: .cfi_offset w22, -32
; CHECK-NEXT: .cfi_offset w23, -40
; CHECK-NEXT: .cfi_offset w24, -48
; CHECK-NEXT: .cfi_offset w25, -64
; CHECK-NEXT: add x8, sp, #64
; CHECK-NEXT: add x9, sp, #128
; CHECK-NEXT: add x10, sp, #160
; CHECK-NEXT: add x11, sp, #192
; CHECK-NEXT: ptrue p0.d
; CHECK-NEXT: add x20, sp, #192
; CHECK-NEXT: ld1d { z0.d }, p0/z, [x8]
; CHECK-NEXT: ld1d { z1.d }, p0/z, [x9]
; CHECK-NEXT: ld1d { z2.d }, p0/z, [x10]
; CHECK-NEXT: ld1d { z3.d }, p0/z, [x11]
; CHECK-NEXT: ldp x18, x19, [sp, #368]
; CHECK-NEXT: add x21, sp, #160
; CHECK-NEXT: add x22, sp, #128
; CHECK-NEXT: ldp x24, x14, [sp, #296]
; CHECK-NEXT: add x23, sp, #64
; CHECK-NEXT: ldr x25, [sp, #288]
; CHECK-NEXT: ldp x9, x8, [sp, #344]
; CHECK-NEXT: ldp x11, x10, [sp, #328]
; CHECK-NEXT: ldp x13, x12, [sp, #312]
; CHECK-NEXT: ldr x15, [sp, #120]
; CHECK-NEXT: ldur q4, [sp, #104]
; CHECK-NEXT: ldp x16, x17, [sp, #224]
; CHECK-NEXT: st1d { z3.d }, p0, [x20]
; CHECK-NEXT: st1d { z2.d }, p0, [x21]
; CHECK-NEXT: st1d { z1.d }, p0, [x22]
; CHECK-NEXT: st1d { z0.d }, p0, [x23]
; CHECK-NEXT: stp x18, x19, [sp, #368]
; CHECK-NEXT: stp x25, x24, [sp, #288]
; CHECK-NEXT: ldp x20, x19, [sp, #48] // 16-byte Folded Reload
; CHECK-NEXT: stp x16, x17, [sp, #224]
; CHECK-NEXT: ldp x22, x21, [sp, #32] // 16-byte Folded Reload
; CHECK-NEXT: stur q4, [sp, #104]
; CHECK-NEXT: ldp x24, x23, [sp, #16] // 16-byte Folded Reload
; CHECK-NEXT: str x15, [sp, #120]
; CHECK-NEXT: stp x14, x13, [sp, #304]
; CHECK-NEXT: stp x12, x11, [sp, #320]
; CHECK-NEXT: stp x10, x9, [sp, #336]
; CHECK-NEXT: str x8, [sp, #352]
; CHECK-NEXT: ldr x25, [sp], #64 // 8-byte Folded Reload
; CHECK-NEXT: b func2
i64* %v9, i64* %v10, i64* %v11, i64* %v12, i64* %v13, i64* %v14, i64* %v15, i64* %v16,
i64* %v17, i64* %v18, i64* %v19, i64* %v20, i64* %v21, i64* %v22, i64* %v23, i64* %v24,
i64* %v25, i64* %v26, i64* %v27, i64* %v28, i64* %v29, i64* %v30, i64* %v31, i64* %v32,
i64* %v33, i64* %v34, i64* %v35, i64* %v36, i64* %v37, i64* %v38, i64* %v39, i64* %v40,
i64* %v41, i64* %v42, i64* %v43, i64* %v44, i64* %v45, i64* %v46, i64* %v47, i64* %v48,
i64 %v49) #0 {
tail call void @func2(i64* %v1, i64* %v2, i64* %v3, i64* %v4, i64* %v5, i64* %v6, i64* %v7, i64* %v8,
i64* %v9, i64* %v10, i64* %v11, i64* %v12, i64* undef, i64* %v14, i64* %v15, i64* %v16,
i64* %v17, i64* %v18, i64* %v19, i64* %v20, i64* %v21, i64* %v22, i64* %v23, i64* %v24,
i64* %v25, i64* %v26, i64* %v27, i64* %v28, i64* %v29, i64* %v30, i64* undef, i64* undef,
i64* undef, i64* undef, i64* undef, i64* undef, i64* %v37, i64* %v38, i64* %v39, i64* %v40,
i64* %v41, i64* %v42, i64* %v43, i64* %v44, i64* %v45, i64* undef, i64* %v47, i64* %v48,
i64 undef)
ret void
}
declare dso_local void @func2(i64*, i64*, i64*, i64*, i64*, i64*, i64*, i64*,
i64*, i64*, i64*, i64*, i64*, i64*, i64*, i64*,
i64*, i64*, i64*, i64*, i64*, i64*, i64*, i64*,
i64*, i64*, i64*, i64*, i64*, i64*, i64*, i64*,
i64*, i64*, i64*, i64*, i64*, i64*, i64*, i64*,
i64*, i64*, i64*, i64*, i64*, i64*, i64*, i64*,
i64)
attributes #0 = { "target-features"="+sve" vscale_range(2,2) }