; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py ; RUN: llc < %s | FileCheck %s target triple = "aarch64-unknown-linux-gnu" ; INCP define i64 @cntp_add_all_active_nxv16i1(i64 %x, %pg) #0 { ; CHECK-LABEL: cntp_add_all_active_nxv16i1: ; CHECK: // %bb.0: ; CHECK-NEXT: incp x0, p0.b ; CHECK-NEXT: ret %1 = tail call @llvm.aarch64.sve.ptrue.nxv16i1(i32 31) %2 = tail call i64 @llvm.aarch64.sve.cntp.nxv16i1( %1, %pg) %add = add i64 %2, %x ret i64 %add } define i64 @cntp_add_all_active_nxv8i1(i64 %x, %pg) #0 { ; CHECK-LABEL: cntp_add_all_active_nxv8i1: ; CHECK: // %bb.0: ; CHECK-NEXT: incp x0, p0.h ; CHECK-NEXT: ret %1 = tail call @llvm.aarch64.sve.ptrue.nxv8i1(i32 31) %2 = tail call i64 @llvm.aarch64.sve.cntp.nxv8i1( %1, %pg) %add = add i64 %2, %x ret i64 %add } define i64 @cntp_add_all_active_nxv4i1(i64 %x, %pg) #0 { ; CHECK-LABEL: cntp_add_all_active_nxv4i1: ; CHECK: // %bb.0: ; CHECK-NEXT: incp x0, p0.s ; CHECK-NEXT: ret %1 = tail call @llvm.aarch64.sve.ptrue.nxv4i1(i32 31) %2 = tail call i64 @llvm.aarch64.sve.cntp.nxv4i1( %1, %pg) %add = add i64 %2, %x ret i64 %add } define i64 @cntp_add_all_active_nxv2i1(i64 %x, %pg) #0 { ; CHECK-LABEL: cntp_add_all_active_nxv2i1: ; CHECK: // %bb.0: ; CHECK-NEXT: incp x0, p0.d ; CHECK-NEXT: ret %1 = tail call @llvm.aarch64.sve.ptrue.nxv2i1(i32 31) %2 = tail call i64 @llvm.aarch64.sve.cntp.nxv2i1( %1, %pg) %add = add i64 %2, %x ret i64 %add } define i64 @cntp_add_all_active_nxv8i1_via_cast(i64 %x, %pg) #0 { ; CHECK-LABEL: cntp_add_all_active_nxv8i1_via_cast: ; CHECK: // %bb.0: ; CHECK-NEXT: incp x0, p0.h ; CHECK-NEXT: ret %1 = tail call @llvm.aarch64.sve.ptrue.nxv16i1(i32 31) %2 = tail call @llvm.aarch64.sve.convert.from.svbool.nxv8i1( %1) %3 = tail call i64 @llvm.aarch64.sve.cntp.nxv8i1( %2, %pg) %add = add i64 %3, %x ret i64 %add } define i64 @cntp_add_all_active_nxv2i1_multiuse(i64 %x, %pg) #0 { ; CHECK-LABEL: cntp_add_all_active_nxv2i1_multiuse: ; CHECK: // %bb.0: ; CHECK-NEXT: ptrue p1.d ; CHECK-NEXT: cntp x8, p1, p0.d ; CHECK-NEXT: add x9, x8, x0 ; CHECK-NEXT: madd x0, x8, x0, x9 ; CHECK-NEXT: ret %1 = tail call @llvm.aarch64.sve.ptrue.nxv2i1(i32 31) %2 = tail call i64 @llvm.aarch64.sve.cntp.nxv2i1( %1, %pg) %add = add i64 %2, %x %mul = mul i64 %2, %x %res = add i64 %add, %mul ret i64 %res } define i64 @cntp_add_same_active_nxv16i1(i64 %x, %pg) #0 { ; CHECK-LABEL: cntp_add_same_active_nxv16i1: ; CHECK: // %bb.0: ; CHECK-NEXT: incp x0, p0.b ; CHECK-NEXT: ret %1 = tail call i64 @llvm.aarch64.sve.cntp.nxv16i1( %pg, %pg) %add = add i64 %1, %x ret i64 %add } define i64 @cntp_add_same_active_nxv8i1(i64 %x, %pg) #0 { ; CHECK-LABEL: cntp_add_same_active_nxv8i1: ; CHECK: // %bb.0: ; CHECK-NEXT: incp x0, p0.h ; CHECK-NEXT: ret %1 = tail call i64 @llvm.aarch64.sve.cntp.nxv8i1( %pg, %pg) %add = add i64 %1, %x ret i64 %add } define i64 @cntp_add_same_active_nxv4i1(i64 %x, %pg) #0 { ; CHECK-LABEL: cntp_add_same_active_nxv4i1: ; CHECK: // %bb.0: ; CHECK-NEXT: incp x0, p0.s ; CHECK-NEXT: ret %1 = tail call i64 @llvm.aarch64.sve.cntp.nxv4i1( %pg, %pg) %add = add i64 %1, %x ret i64 %add } define i64 @cntp_add_same_active_nxv2i1(i64 %x, %pg) #0 { ; CHECK-LABEL: cntp_add_same_active_nxv2i1: ; CHECK: // %bb.0: ; CHECK-NEXT: incp x0, p0.d ; CHECK-NEXT: ret %1 = tail call i64 @llvm.aarch64.sve.cntp.nxv2i1( %pg, %pg) %add = add i64 %1, %x ret i64 %add } define i64 @cntp_add_same_active_nxv2i1_multiuse(i64 %x, %pg) #0 { ; CHECK-LABEL: cntp_add_same_active_nxv2i1_multiuse: ; CHECK: // %bb.0: ; CHECK-NEXT: cntp x8, p0, p0.d ; CHECK-NEXT: add x9, x8, x0 ; CHECK-NEXT: madd x0, x8, x0, x9 ; CHECK-NEXT: ret %1 = tail call i64 @llvm.aarch64.sve.cntp.nxv2i1( %pg, %pg) %add = add i64 %1, %x %mul = mul i64 %1, %x %res = add i64 %add, %mul ret i64 %res } ; DECP define i64 @cntp_sub_all_active_nxv16i1(i64 %x, %pg) #0 { ; CHECK-LABEL: cntp_sub_all_active_nxv16i1: ; CHECK: // %bb.0: ; CHECK-NEXT: decp x0, p0.b ; CHECK-NEXT: ret %1 = tail call @llvm.aarch64.sve.ptrue.nxv16i1(i32 31) %2 = tail call i64 @llvm.aarch64.sve.cntp.nxv16i1( %1, %pg) %sub = sub i64 %x, %2 ret i64 %sub } define i64 @cntp_sub_all_active_nxv8i1(i64 %x, %pg) #0 { ; CHECK-LABEL: cntp_sub_all_active_nxv8i1: ; CHECK: // %bb.0: ; CHECK-NEXT: decp x0, p0.h ; CHECK-NEXT: ret %1 = tail call @llvm.aarch64.sve.ptrue.nxv8i1(i32 31) %2 = tail call i64 @llvm.aarch64.sve.cntp.nxv8i1( %1, %pg) %sub = sub i64 %x, %2 ret i64 %sub } define i64 @cntp_sub_all_active_nxv4i1(i64 %x, %pg) #0 { ; CHECK-LABEL: cntp_sub_all_active_nxv4i1: ; CHECK: // %bb.0: ; CHECK-NEXT: decp x0, p0.s ; CHECK-NEXT: ret %1 = tail call @llvm.aarch64.sve.ptrue.nxv4i1(i32 31) %2 = tail call i64 @llvm.aarch64.sve.cntp.nxv4i1( %1, %pg) %sub = sub i64 %x, %2 ret i64 %sub } define i64 @cntp_sub_all_active_nxv2i1(i64 %x, %pg) #0 { ; CHECK-LABEL: cntp_sub_all_active_nxv2i1: ; CHECK: // %bb.0: ; CHECK-NEXT: decp x0, p0.d ; CHECK-NEXT: ret %1 = tail call @llvm.aarch64.sve.ptrue.nxv2i1(i32 31) %2 = tail call i64 @llvm.aarch64.sve.cntp.nxv2i1( %1, %pg) %sub = sub i64 %x, %2 ret i64 %sub } define i64 @cntp_sub_all_active_nxv8i1_via_cast(i64 %x, %pg) #0 { ; CHECK-LABEL: cntp_sub_all_active_nxv8i1_via_cast: ; CHECK: // %bb.0: ; CHECK-NEXT: decp x0, p0.h ; CHECK-NEXT: ret %1 = tail call @llvm.aarch64.sve.ptrue.nxv16i1(i32 31) %2 = tail call @llvm.aarch64.sve.convert.from.svbool.nxv8i1( %1) %3 = tail call i64 @llvm.aarch64.sve.cntp.nxv8i1( %2, %pg) %sub = sub i64 %x, %3 ret i64 %sub } define i64 @cntp_sub_all_active_nxv2i1_multiuse(i64 %x, %pg) #0 { ; CHECK-LABEL: cntp_sub_all_active_nxv2i1_multiuse: ; CHECK: // %bb.0: ; CHECK-NEXT: ptrue p1.d ; CHECK-NEXT: cntp x8, p1, p0.d ; CHECK-NEXT: sub x9, x8, x0 ; CHECK-NEXT: madd x0, x8, x0, x9 ; CHECK-NEXT: ret %1 = tail call @llvm.aarch64.sve.ptrue.nxv2i1(i32 31) %2 = tail call i64 @llvm.aarch64.sve.cntp.nxv2i1( %1, %pg) %sub = sub i64 %2, %x %mul = mul i64 %2, %x %res = add i64 %sub, %mul ret i64 %res } define i64 @cntp_sub_same_active_nxv16i1(i64 %x, %pg) #0 { ; CHECK-LABEL: cntp_sub_same_active_nxv16i1: ; CHECK: // %bb.0: ; CHECK-NEXT: decp x0, p0.b ; CHECK-NEXT: ret %1 = tail call i64 @llvm.aarch64.sve.cntp.nxv16i1( %pg, %pg) %sub = sub i64 %x, %1 ret i64 %sub } define i64 @cntp_sub_same_active_nxv8i1(i64 %x, %pg) #0 { ; CHECK-LABEL: cntp_sub_same_active_nxv8i1: ; CHECK: // %bb.0: ; CHECK-NEXT: decp x0, p0.h ; CHECK-NEXT: ret %1 = tail call i64 @llvm.aarch64.sve.cntp.nxv8i1( %pg, %pg) %sub = sub i64 %x, %1 ret i64 %sub } define i64 @cntp_sub_same_active_nxv4i1(i64 %x, %pg) #0 { ; CHECK-LABEL: cntp_sub_same_active_nxv4i1: ; CHECK: // %bb.0: ; CHECK-NEXT: decp x0, p0.s ; CHECK-NEXT: ret %1 = tail call i64 @llvm.aarch64.sve.cntp.nxv4i1( %pg, %pg) %sub = sub i64 %x, %1 ret i64 %sub } define i64 @cntp_sub_same_active_nxv2i1(i64 %x, %pg) #0 { ; CHECK-LABEL: cntp_sub_same_active_nxv2i1: ; CHECK: // %bb.0: ; CHECK-NEXT: decp x0, p0.d ; CHECK-NEXT: ret %1 = tail call i64 @llvm.aarch64.sve.cntp.nxv2i1( %pg, %pg) %sub = sub i64 %x, %1 ret i64 %sub } define i64 @cntp_sub_same_active_nxv2i1_multiuse(i64 %x, %pg) #0 { ; CHECK-LABEL: cntp_sub_same_active_nxv2i1_multiuse: ; CHECK: // %bb.0: ; CHECK-NEXT: cntp x8, p0, p0.d ; CHECK-NEXT: sub x9, x8, x0 ; CHECK-NEXT: madd x0, x8, x0, x9 ; CHECK-NEXT: ret %1 = tail call i64 @llvm.aarch64.sve.cntp.nxv2i1( %pg, %pg) %sub = sub i64 %1, %x %mul = mul i64 %1, %x %res = add i64 %sub, %mul ret i64 %res } declare @llvm.aarch64.sve.convert.from.svbool.nxv8i1() declare @llvm.aarch64.sve.convert.from.svbool.nxv4i1() declare @llvm.aarch64.sve.convert.from.svbool.nxv2i1() declare @llvm.aarch64.sve.ptrue.nxv16i1(i32) declare @llvm.aarch64.sve.ptrue.nxv8i1(i32) declare @llvm.aarch64.sve.ptrue.nxv4i1(i32) declare @llvm.aarch64.sve.ptrue.nxv2i1(i32) declare i64 @llvm.aarch64.sve.cntp.nxv16i1(, ) declare i64 @llvm.aarch64.sve.cntp.nxv8i1(, ) declare i64 @llvm.aarch64.sve.cntp.nxv4i1(, ) declare i64 @llvm.aarch64.sve.cntp.nxv2i1(, ) attributes #0 = { "target-features"="+sve" }