# NOTE: Assertions have been autogenerated by utils/update_mir_test_checks.py # RUN: llc -mtriple=aarch64-unknown-unknown -run-pass=legalizer -mattr=+fullfp16 -verify-machineinstrs %s -o - | FileCheck %s --check-prefixes=CHECK,CHECK-FP16 # RUN: llc -mtriple=aarch64-unknown-unknown -run-pass=legalizer -verify-machineinstrs %s -o - | FileCheck %s --check-prefixes=CHECK,CHECK-NOFP16 ... --- name: s16_legal_with_full_fp16 alignment: 4 body: | bb.0: liveins: $h0, $h1 ; CHECK-FP16-LABEL: name: s16_legal_with_full_fp16 ; CHECK-FP16: liveins: $h0, $h1 ; CHECK-FP16-NEXT: {{ $}} ; CHECK-FP16-NEXT: %a:_(s16) = COPY $h0 ; CHECK-FP16-NEXT: %b:_(s16) = COPY $h1 ; CHECK-FP16-NEXT: %minnum:_(s16) = G_FMINNUM %a, %b ; CHECK-FP16-NEXT: $h0 = COPY %minnum(s16) ; CHECK-FP16-NEXT: RET_ReallyLR implicit $h0 ; ; CHECK-NOFP16-LABEL: name: s16_legal_with_full_fp16 ; CHECK-NOFP16: liveins: $h0, $h1 ; CHECK-NOFP16-NEXT: {{ $}} ; CHECK-NOFP16-NEXT: %a:_(s16) = COPY $h0 ; CHECK-NOFP16-NEXT: %b:_(s16) = COPY $h1 ; CHECK-NOFP16-NEXT: [[FPEXT:%[0-9]+]]:_(s32) = G_FPEXT %a(s16) ; CHECK-NOFP16-NEXT: [[FPEXT1:%[0-9]+]]:_(s32) = G_FPEXT %b(s16) ; CHECK-NOFP16-NEXT: [[FMINNUM:%[0-9]+]]:_(s32) = G_FMINNUM [[FPEXT]], [[FPEXT1]] ; CHECK-NOFP16-NEXT: %minnum:_(s16) = G_FPTRUNC [[FMINNUM]](s32) ; CHECK-NOFP16-NEXT: $h0 = COPY %minnum(s16) ; CHECK-NOFP16-NEXT: RET_ReallyLR implicit $h0 %a:_(s16) = COPY $h0 %b:_(s16) = COPY $h1 %minnum:_(s16) = G_FMINNUM %a, %b $h0 = COPY %minnum(s16) RET_ReallyLR implicit $h0 ... --- name: s32_legal alignment: 4 body: | bb.0: liveins: $s0, $s1 ; CHECK-LABEL: name: s32_legal ; CHECK: liveins: $s0, $s1 ; CHECK-NEXT: {{ $}} ; CHECK-NEXT: %a:_(s32) = COPY $s0 ; CHECK-NEXT: %b:_(s32) = COPY $s1 ; CHECK-NEXT: %minnum:_(s32) = G_FMINNUM %a, %b ; CHECK-NEXT: $s0 = COPY %minnum(s32) ; CHECK-NEXT: RET_ReallyLR implicit $s0 %a:_(s32) = COPY $s0 %b:_(s32) = COPY $s1 %minnum:_(s32) = G_FMINNUM %a, %b $s0 = COPY %minnum(s32) RET_ReallyLR implicit $s0 ... --- name: s64_legal alignment: 4 body: | bb.0: liveins: $d0, $d1 ; CHECK-LABEL: name: s64_legal ; CHECK: liveins: $d0, $d1 ; CHECK-NEXT: {{ $}} ; CHECK-NEXT: %a:_(s64) = COPY $d0 ; CHECK-NEXT: %b:_(s64) = COPY $d1 ; CHECK-NEXT: %minnum:_(s64) = G_FMINNUM %a, %b ; CHECK-NEXT: $d0 = COPY %minnum(s64) ; CHECK-NEXT: RET_ReallyLR implicit $d0 %a:_(s64) = COPY $d0 %b:_(s64) = COPY $d1 %minnum:_(s64) = G_FMINNUM %a, %b $d0 = COPY %minnum(s64) RET_ReallyLR implicit $d0 ... --- name: s128_libcall alignment: 4 body: | bb.0: liveins: $q0, $q1 ; CHECK-LABEL: name: s128_libcall ; CHECK: liveins: $q0, $q1 ; CHECK-NEXT: {{ $}} ; CHECK-NEXT: %a:_(s128) = COPY $q0 ; CHECK-NEXT: %b:_(s128) = COPY $q1 ; CHECK-NEXT: ADJCALLSTACKDOWN 0, 0, implicit-def $sp, implicit $sp ; CHECK-NEXT: $q0 = COPY %a(s128) ; CHECK-NEXT: $q1 = COPY %b(s128) ; CHECK-NEXT: BL &fminl, csr_aarch64_aapcs, implicit-def $lr, implicit $sp, implicit $q0, implicit $q1, implicit-def $q0 ; CHECK-NEXT: ADJCALLSTACKUP 0, 0, implicit-def $sp, implicit $sp ; CHECK-NEXT: %minnum:_(s128) = COPY $q0 ; CHECK-NEXT: $q0 = COPY %minnum(s128) ; CHECK-NEXT: RET_ReallyLR implicit $q0 %a:_(s128) = COPY $q0 %b:_(s128) = COPY $q1 %minnum:_(s128) = G_FMINNUM %a, %b $q0 = COPY %minnum(s128) RET_ReallyLR implicit $q0 ... --- name: v4s32_legal alignment: 4 body: | bb.0: liveins: $q0, $q1 ; CHECK-LABEL: name: v4s32_legal ; CHECK: liveins: $q0, $q1 ; CHECK-NEXT: {{ $}} ; CHECK-NEXT: %a:_(<4 x s32>) = COPY $q0 ; CHECK-NEXT: %b:_(<4 x s32>) = COPY $q1 ; CHECK-NEXT: %minnum:_(<4 x s32>) = G_FMINNUM %a, %b ; CHECK-NEXT: $q0 = COPY %minnum(<4 x s32>) ; CHECK-NEXT: RET_ReallyLR implicit $q0 %a:_(<4 x s32>) = COPY $q0 %b:_(<4 x s32>) = COPY $q1 %minnum:_(<4 x s32>) = G_FMINNUM %a, %b $q0 = COPY %minnum(<4 x s32>) RET_ReallyLR implicit $q0 ... --- name: v3s32_widen alignment: 4 body: | bb.1.entry: liveins: $q0, $q1 ; CHECK-LABEL: name: v3s32_widen ; CHECK: liveins: $q0, $q1 ; CHECK-NEXT: {{ $}} ; CHECK-NEXT: [[COPY:%[0-9]+]]:_(<2 x s64>) = COPY $q0 ; CHECK-NEXT: [[BITCAST:%[0-9]+]]:_(<4 x s32>) = G_BITCAST [[COPY]](<2 x s64>) ; CHECK-NEXT: [[COPY1:%[0-9]+]]:_(<2 x s64>) = COPY $q1 ; CHECK-NEXT: [[BITCAST1:%[0-9]+]]:_(<4 x s32>) = G_BITCAST [[COPY1]](<2 x s64>) ; CHECK-NEXT: [[FMINNUM:%[0-9]+]]:_(<4 x s32>) = G_FMINNUM [[BITCAST]], [[BITCAST1]] ; CHECK-NEXT: $q0 = COPY [[FMINNUM]](<4 x s32>) ; CHECK-NEXT: RET_ReallyLR implicit $q0 %2:_(<2 x s64>) = COPY $q0 %3:_(<4 x s32>) = G_BITCAST %2:_(<2 x s64>) %4:_(s32), %5:_(s32), %6:_(s32), %7:_(s32) = G_UNMERGE_VALUES %3:_(<4 x s32>) %0:_(<3 x s32>) = G_BUILD_VECTOR %4:_(s32), %5:_(s32), %6:_(s32) %8:_(<2 x s64>) = COPY $q1 %9:_(<4 x s32>) = G_BITCAST %8:_(<2 x s64>) %10:_(s32), %11:_(s32), %12:_(s32), %13:_(s32) = G_UNMERGE_VALUES %9:_(<4 x s32>) %1:_(<3 x s32>) = G_BUILD_VECTOR %10:_(s32), %11:_(s32), %12:_(s32) %14:_(<3 x s32>) = G_FMINNUM %0:_, %1:_ %15:_(s32), %16:_(s32), %17:_(s32) = G_UNMERGE_VALUES %14:_(<3 x s32>) %18:_(s32) = G_IMPLICIT_DEF %19:_(<4 x s32>) = G_BUILD_VECTOR %15:_(s32), %16:_(s32), %17:_(s32), %18:_(s32) $q0 = COPY %19:_(<4 x s32>) RET_ReallyLR implicit $q0 ...