blob: e580dfe176493bcdc22aab440d341e2aec69ab69 [file] [edit]
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
; RUN: llc --mtriple=loongarch64 -mattr=+d,-lsx < %s | FileCheck %s --check-prefix=LA64
; RUN: llc --mtriple=loongarch32 -mattr=+d < %s | FileCheck %s --check-prefix=LA32
; RUN: llc --mtriple=loongarch64 -mattr=+d,+lsx < %s | FileCheck %s --check-prefix=LSX
; RUN: llc --mtriple=loongarch64 -mattr=+d,+lasx < %s | FileCheck %s --check-prefix=LASX
; ISSUE-186645 - Merge 64-bit operations as 128-bit operations while noimplicitfloat
; attribute is set are not legal. However, we can generate a paired 64-bit loads and
; stores, without using floating point registers.
define void @memcpy16_noimplicitfloat(ptr %p0, ptr %p1) noimplicitfloat {
; LA64-LABEL: memcpy16_noimplicitfloat:
; LA64: # %bb.0:
; LA64-NEXT: ld.d $a2, $a1, 8
; LA64-NEXT: st.d $a2, $a0, 8
; LA64-NEXT: ld.d $a1, $a1, 0
; LA64-NEXT: st.d $a1, $a0, 0
; LA64-NEXT: ret
;
; LA32-LABEL: memcpy16_noimplicitfloat:
; LA32: # %bb.0:
; LA32-NEXT: ld.w $a2, $a1, 12
; LA32-NEXT: st.w $a2, $a0, 12
; LA32-NEXT: ld.w $a2, $a1, 8
; LA32-NEXT: st.w $a2, $a0, 8
; LA32-NEXT: ld.w $a2, $a1, 4
; LA32-NEXT: st.w $a2, $a0, 4
; LA32-NEXT: ld.w $a1, $a1, 0
; LA32-NEXT: st.w $a1, $a0, 0
; LA32-NEXT: ret
;
; LSX-LABEL: memcpy16_noimplicitfloat:
; LSX: # %bb.0:
; LSX-NEXT: ld.d $a2, $a1, 8
; LSX-NEXT: st.d $a2, $a0, 8
; LSX-NEXT: ld.d $a1, $a1, 0
; LSX-NEXT: st.d $a1, $a0, 0
; LSX-NEXT: ret
;
; LASX-LABEL: memcpy16_noimplicitfloat:
; LASX: # %bb.0:
; LASX-NEXT: ld.d $a2, $a1, 8
; LASX-NEXT: st.d $a2, $a0, 8
; LASX-NEXT: ld.d $a1, $a1, 0
; LASX-NEXT: st.d $a1, $a0, 0
; LASX-NEXT: ret
call void @llvm.memcpy.p0.p0.i64(ptr align 4 %p0, ptr align 4 %p1, i64 16, i1 false)
ret void
}
define void @memcpy16_aligned(ptr %p0, ptr %p1) {
; LA64-LABEL: memcpy16_aligned:
; LA64: # %bb.0:
; LA64-NEXT: ld.d $a2, $a1, 8
; LA64-NEXT: st.d $a2, $a0, 8
; LA64-NEXT: ld.d $a1, $a1, 0
; LA64-NEXT: st.d $a1, $a0, 0
; LA64-NEXT: ret
;
; LA32-LABEL: memcpy16_aligned:
; LA32: # %bb.0:
; LA32-NEXT: ld.w $a2, $a1, 12
; LA32-NEXT: st.w $a2, $a0, 12
; LA32-NEXT: ld.w $a2, $a1, 8
; LA32-NEXT: st.w $a2, $a0, 8
; LA32-NEXT: ld.w $a2, $a1, 4
; LA32-NEXT: st.w $a2, $a0, 4
; LA32-NEXT: ld.w $a1, $a1, 0
; LA32-NEXT: st.w $a1, $a0, 0
; LA32-NEXT: ret
;
; LSX-LABEL: memcpy16_aligned:
; LSX: # %bb.0:
; LSX-NEXT: vld $vr0, $a1, 0
; LSX-NEXT: vst $vr0, $a0, 0
; LSX-NEXT: ret
;
; LASX-LABEL: memcpy16_aligned:
; LASX: # %bb.0:
; LASX-NEXT: vld $vr0, $a1, 0
; LASX-NEXT: vst $vr0, $a0, 0
; LASX-NEXT: ret
call void @llvm.memcpy.p0.p0.i64(ptr align 4 %p0, ptr align 4 %p1, i64 16, i1 false)
ret void
}
define void @memcpy32_aligned(ptr %p0, ptr %p1) {
; LA64-LABEL: memcpy32_aligned:
; LA64: # %bb.0:
; LA64-NEXT: ld.d $a2, $a1, 24
; LA64-NEXT: st.d $a2, $a0, 24
; LA64-NEXT: ld.d $a2, $a1, 16
; LA64-NEXT: st.d $a2, $a0, 16
; LA64-NEXT: ld.d $a2, $a1, 8
; LA64-NEXT: st.d $a2, $a0, 8
; LA64-NEXT: ld.d $a1, $a1, 0
; LA64-NEXT: st.d $a1, $a0, 0
; LA64-NEXT: ret
;
; LA32-LABEL: memcpy32_aligned:
; LA32: # %bb.0:
; LA32-NEXT: ld.w $a2, $a1, 28
; LA32-NEXT: st.w $a2, $a0, 28
; LA32-NEXT: ld.w $a2, $a1, 24
; LA32-NEXT: st.w $a2, $a0, 24
; LA32-NEXT: ld.w $a2, $a1, 20
; LA32-NEXT: st.w $a2, $a0, 20
; LA32-NEXT: ld.w $a2, $a1, 16
; LA32-NEXT: st.w $a2, $a0, 16
; LA32-NEXT: ld.w $a2, $a1, 12
; LA32-NEXT: st.w $a2, $a0, 12
; LA32-NEXT: ld.w $a2, $a1, 8
; LA32-NEXT: st.w $a2, $a0, 8
; LA32-NEXT: ld.w $a2, $a1, 4
; LA32-NEXT: st.w $a2, $a0, 4
; LA32-NEXT: ld.w $a1, $a1, 0
; LA32-NEXT: st.w $a1, $a0, 0
; LA32-NEXT: ret
;
; LSX-LABEL: memcpy32_aligned:
; LSX: # %bb.0:
; LSX-NEXT: vld $vr0, $a1, 16
; LSX-NEXT: vst $vr0, $a0, 16
; LSX-NEXT: vld $vr0, $a1, 0
; LSX-NEXT: vst $vr0, $a0, 0
; LSX-NEXT: ret
;
; LASX-LABEL: memcpy32_aligned:
; LASX: # %bb.0:
; LASX-NEXT: xvld $xr0, $a1, 0
; LASX-NEXT: xvst $xr0, $a0, 0
; LASX-NEXT: ret
call void @llvm.memcpy.p0.p0.i64(ptr align 4 %p0, ptr align 4 %p1, i64 32, i1 false)
ret void
}
declare void @llvm.memcpy.p0.p0.i64(ptr, ptr, i64, i1)