| ; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py |
| ; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mattr=+avx512bf16 | FileCheck %s --check-prefix=X64 |
| ; RUN: llc < %s -mtriple=i686-unknown-unknown -mattr=+avx512bf16 | FileCheck %s --check-prefix=X86 |
| |
| define dso_local void @funbf16(ptr readonly %src, ptr writeonly %dst) { |
| ; X64-LABEL: funbf16: |
| ; X64: # %bb.0: # %entry |
| ; X64-NEXT: vmovups (%rdi), %xmm0 |
| ; X64-NEXT: vmovups %xmm0, (%rsi) |
| ; X64-NEXT: vmovaps (%rdi), %xmm0 |
| ; X64-NEXT: vmovaps %xmm0, (%rsi) |
| ; X64-NEXT: vmovups (%rdi), %ymm0 |
| ; X64-NEXT: vmovups %ymm0, (%rsi) |
| ; X64-NEXT: vmovaps (%rdi), %ymm0 |
| ; X64-NEXT: vmovaps %ymm0, (%rsi) |
| ; X64-NEXT: vzeroupper |
| ; X64-NEXT: retq |
| ; |
| ; X86-LABEL: funbf16: |
| ; X86: # %bb.0: # %entry |
| ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax |
| ; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx |
| ; X86-NEXT: vmovups (%ecx), %xmm0 |
| ; X86-NEXT: vmovups %xmm0, (%eax) |
| ; X86-NEXT: vmovaps (%ecx), %xmm0 |
| ; X86-NEXT: vmovaps %xmm0, (%eax) |
| ; X86-NEXT: vmovups (%ecx), %ymm0 |
| ; X86-NEXT: vmovups %ymm0, (%eax) |
| ; X86-NEXT: vmovaps (%ecx), %ymm0 |
| ; X86-NEXT: vmovaps %ymm0, (%eax) |
| ; X86-NEXT: vzeroupper |
| ; X86-NEXT: retl |
| entry: |
| %0 = load <8 x bfloat>, ptr %src, align 1 |
| store <8 x bfloat> %0, ptr %dst, align 1 |
| %1 = load <8 x bfloat>, ptr %src, align 32 |
| store <8 x bfloat> %1, ptr %dst, align 32 |
| %2 = load <16 x bfloat>, ptr %src, align 1 |
| store <16 x bfloat> %2, ptr %dst, align 1 |
| %3 = load <16 x bfloat>, ptr %src, align 32 |
| store <16 x bfloat> %3, ptr %dst, align 32 |
| ret void |
| } |