blob: 6bca8343402914642106408bc11d5e95759fbc54 [file] [edit]
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py UTC_ARGS: --version 6
; RUN: llc -mattr=+sve < %s | FileCheck %s
target triple = "aarch64-unknown-linux-gnu"
define <vscale x 16 x i8> @reverse_masked_load() nounwind {
; CHECK-LABEL: reverse_masked_load:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: sub sp, sp, #16
; CHECK-NEXT: ptrue p0.b, vl6
; CHECK-NEXT: mov w8, #1799 // =0x707
; CHECK-NEXT: add x9, sp, #8
; CHECK-NEXT: strh w8, [sp, #12]
; CHECK-NEXT: mov w8, #117901063 // =0x7070707
; CHECK-NEXT: rev p0.b, p0.b
; CHECK-NEXT: str w8, [sp, #8]
; CHECK-NEXT: add x8, x9, #6
; CHECK-NEXT: ld1b { z0.b }, p0/z, [x8, #-1, mul vl]
; CHECK-NEXT: add sp, sp, #16
; CHECK-NEXT: ret
entry:
%l_2 = alloca [6 x i8], align 1
call void @llvm.memset.p0.i64(ptr noundef nonnull align 1 dereferenceable(6) %l_2, i8 7, i64 6, i1 false)
%1 = getelementptr inbounds nuw i8, ptr %l_2, i64 6
%2 = tail call i64 @llvm.vscale.i64()
%3 = mul nsw i64 %2, -16
%5 = getelementptr i8, ptr %1, i64 %3
%active.lane.mask = tail call <vscale x 16 x i1> @llvm.get.active.lane.mask(i64 0, i64 6)
%reverse = tail call <vscale x 16 x i1> @llvm.vector.reverse(<vscale x 16 x i1> %active.lane.mask)
%masked.load = call <vscale x 16 x i8> @llvm.masked.load(ptr align 1 %5, <vscale x 16 x i1> %reverse, <vscale x 16 x i8> zeroinitializer)
ret <vscale x 16 x i8> %masked.load
}