| ; RUN: llc < %s -mtriple=aarch64 -mattr=+mte | FileCheck %s |
| |
| define i8* @small_alloca() { |
| entry: |
| ; CHECK-LABEL: small_alloca: |
| ; CHECK: irg [[R:x[0-9]+]], sp{{$}} |
| ; CHECK-NEXT: addg x0, [[R]], #0, #1 |
| ; CHECK: ret |
| %a = alloca i8, align 16 |
| %q = call i8* @llvm.aarch64.irg.sp(i64 0) |
| %q1 = call i8* @llvm.aarch64.tagp.p0i8(i8* %a, i8* %q, i64 1) |
| ret i8* %q1 |
| } |
| |
| ; Two large allocas. One's offset overflows addg immediate. |
| define void @huge_allocas() { |
| entry: |
| ; CHECK-LABEL: huge_allocas: |
| ; CHECK: irg [[R:x[0-9]+]], sp{{$}} |
| ; CHECK: add [[TMP:x[0-9]+]], [[R]], #3088 |
| ; CHECK: addg x0, [[TMP]], #1008, #1 |
| ; CHECK: addg x1, [[R]], #0, #2 |
| ; CHECK: bl use2 |
| %a = alloca i8, i64 4096, align 16 |
| %b = alloca i8, i64 4096, align 16 |
| %base = call i8* @llvm.aarch64.irg.sp(i64 0) |
| %a_t = call i8* @llvm.aarch64.tagp.p0i8(i8* %a, i8* %base, i64 1) |
| %b_t = call i8* @llvm.aarch64.tagp.p0i8(i8* %b, i8* %base, i64 2) |
| call void @use2(i8* %a_t, i8* %b_t) |
| ret void |
| } |
| |
| ; Realigned stack frame. IRG uses value of SP after realignment, |
| ; ADDG for the first stack allocation has offset 0. |
| define void @realign() { |
| entry: |
| ; CHECK-LABEL: realign: |
| ; CHECK: mov x29, sp |
| ; CHECK: and sp, x{{[0-9]*}}, #0xffffffffffffffc0 |
| ; CHECK: irg [[R:x[0-9]+]], sp{{$}} |
| ; CHECK: addg x0, [[R]], #0, #1 |
| ; CHECK: bl use |
| %a = alloca i8, i64 4096, align 64 |
| %base = call i8* @llvm.aarch64.irg.sp(i64 0) |
| %a_t = call i8* @llvm.aarch64.tagp.p0i8(i8* %a, i8* %base, i64 1) |
| call void @use(i8* %a_t) |
| ret void |
| } |
| |
| ; With a dynamic alloca, IRG has to use FP with non-zero offset. |
| ; ADDG offset for the single static alloca is still zero. |
| define void @dynamic_alloca(i64 %size) { |
| entry: |
| ; CHECK-LABEL: dynamic_alloca: |
| ; CHECK: sub [[R:x[0-9]+]], x29, #[[OFS:[0-9]+]] |
| ; CHECK: irg [[R]], [[R]] |
| ; CHECK: addg x1, [[R]], #0, #1 |
| ; CHECK: sub x0, x29, #[[OFS]] |
| ; CHECK: bl use2 |
| %base = call i8* @llvm.aarch64.irg.sp(i64 0) |
| %a = alloca i128, i64 %size, align 16 |
| %b = alloca i8, i64 16, align 16 |
| %b_t = call i8* @llvm.aarch64.tagp.p0i8(i8* %b, i8* %base, i64 1) |
| call void @use2(i8* %b, i8* %b_t) |
| ret void |
| } |
| |
| ; Both dynamic alloca and realigned frame. |
| ; After initial realignment, generate the base pointer. |
| ; IRG uses the base pointer w/o offset. |
| ; Offsets for tagged and untagged pointers to the same alloca match. |
| define void @dynamic_alloca_and_realign(i64 %size) { |
| entryz: |
| ; CHECK-LABEL: dynamic_alloca_and_realign: |
| ; CHECK: and sp, x{{.*}}, #0xffffffffffffffc0 |
| ; CHECK: mov x19, sp |
| ; CHECK: irg [[R:x[0-9]+]], x19 |
| ; CHECK: addg x1, [[R]], #[[OFS:[0-9]+]], #1 |
| ; CHECK: add x0, x19, #[[OFS]] |
| ; CHECK: bl use2 |
| %base = call i8* @llvm.aarch64.irg.sp(i64 0) |
| %a = alloca i128, i64 %size, align 64 |
| %b = alloca i8, i64 16, align 16 |
| %b_t = call i8* @llvm.aarch64.tagp.p0i8(i8* %b, i8* %base, i64 1) |
| call void @use2(i8* %b, i8* %b_t) |
| ret void |
| } |
| |
| declare void @use(i8*) |
| declare void @use2(i8*, i8*) |
| |
| declare i8* @llvm.aarch64.irg.sp(i64 %exclude) |
| declare i8* @llvm.aarch64.tagp.p0i8(i8* %p, i8* %tag, i64 %ofs) |