; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py ; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mattr=+avx512bf16 | FileCheck %s --check-prefix=X64 ; RUN: llc < %s -mtriple=i686-unknown-unknown -mattr=+avx512bf16 | FileCheck %s --check-prefix=X86 define dso_local void @funbf16(ptr readonly %src, ptr writeonly %dst) { ; X64-LABEL: funbf16: ; X64: # %bb.0: # %entry ; X64-NEXT: vmovups (%rdi), %xmm0 ; X64-NEXT: vmovups %xmm0, (%rsi) ; X64-NEXT: vmovaps (%rdi), %xmm0 ; X64-NEXT: vmovaps %xmm0, (%rsi) ; X64-NEXT: vmovups (%rdi), %ymm0 ; X64-NEXT: vmovups %ymm0, (%rsi) ; X64-NEXT: vmovaps (%rdi), %ymm0 ; X64-NEXT: vmovaps %ymm0, (%rsi) ; X64-NEXT: vzeroupper ; X64-NEXT: retq ; ; X86-LABEL: funbf16: ; X86: # %bb.0: # %entry ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx ; X86-NEXT: vmovups (%ecx), %xmm0 ; X86-NEXT: vmovups %xmm0, (%eax) ; X86-NEXT: vmovaps (%ecx), %xmm0 ; X86-NEXT: vmovaps %xmm0, (%eax) ; X86-NEXT: vmovups (%ecx), %ymm0 ; X86-NEXT: vmovups %ymm0, (%eax) ; X86-NEXT: vmovaps (%ecx), %ymm0 ; X86-NEXT: vmovaps %ymm0, (%eax) ; X86-NEXT: vzeroupper ; X86-NEXT: retl entry: %0 = load <8 x bfloat>, ptr %src, align 1 store <8 x bfloat> %0, ptr %dst, align 1 %1 = load <8 x bfloat>, ptr %src, align 32 store <8 x bfloat> %1, ptr %dst, align 32 %2 = load <16 x bfloat>, ptr %src, align 1 store <16 x bfloat> %2, ptr %dst, align 1 %3 = load <16 x bfloat>, ptr %src, align 32 store <16 x bfloat> %3, ptr %dst, align 32 ret void }