; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py ; RUN: llc < %s | FileCheck %s target triple = "aarch64-unknown-linux-gnu" ; INCP define i64 @cntp_add_all_active_nxv16i1(i64 %x, <vscale x 16 x i1> %pg) #0 { ; CHECK-LABEL: cntp_add_all_active_nxv16i1: ; CHECK: // %bb.0: ; CHECK-NEXT: incp x0, p0.b ; CHECK-NEXT: ret %1 = tail call <vscale x 16 x i1> @llvm.aarch64.sve.ptrue.nxv16i1(i32 31) %2 = tail call i64 @llvm.aarch64.sve.cntp.nxv16i1(<vscale x 16 x i1> %1, <vscale x 16 x i1> %pg) %add = add i64 %2, %x ret i64 %add } define i64 @cntp_add_all_active_nxv8i1(i64 %x, <vscale x 8 x i1> %pg) #0 { ; CHECK-LABEL: cntp_add_all_active_nxv8i1: ; CHECK: // %bb.0: ; CHECK-NEXT: incp x0, p0.h ; CHECK-NEXT: ret %1 = tail call <vscale x 8 x i1> @llvm.aarch64.sve.ptrue.nxv8i1(i32 31) %2 = tail call i64 @llvm.aarch64.sve.cntp.nxv8i1(<vscale x 8 x i1> %1, <vscale x 8 x i1> %pg) %add = add i64 %2, %x ret i64 %add } define i64 @cntp_add_all_active_nxv4i1(i64 %x, <vscale x 4 x i1> %pg) #0 { ; CHECK-LABEL: cntp_add_all_active_nxv4i1: ; CHECK: // %bb.0: ; CHECK-NEXT: incp x0, p0.s ; CHECK-NEXT: ret %1 = tail call <vscale x 4 x i1> @llvm.aarch64.sve.ptrue.nxv4i1(i32 31) %2 = tail call i64 @llvm.aarch64.sve.cntp.nxv4i1(<vscale x 4 x i1> %1, <vscale x 4 x i1> %pg) %add = add i64 %2, %x ret i64 %add } define i64 @cntp_add_all_active_nxv2i1(i64 %x, <vscale x 2 x i1> %pg) #0 { ; CHECK-LABEL: cntp_add_all_active_nxv2i1: ; CHECK: // %bb.0: ; CHECK-NEXT: incp x0, p0.d ; CHECK-NEXT: ret %1 = tail call <vscale x 2 x i1> @llvm.aarch64.sve.ptrue.nxv2i1(i32 31) %2 = tail call i64 @llvm.aarch64.sve.cntp.nxv2i1(<vscale x 2 x i1> %1, <vscale x 2 x i1> %pg) %add = add i64 %2, %x ret i64 %add } define i64 @cntp_add_all_active_nxv8i1_via_cast(i64 %x, <vscale x 8 x i1> %pg) #0 { ; CHECK-LABEL: cntp_add_all_active_nxv8i1_via_cast: ; CHECK: // %bb.0: ; CHECK-NEXT: incp x0, p0.h ; CHECK-NEXT: ret %1 = tail call <vscale x 16 x i1> @llvm.aarch64.sve.ptrue.nxv16i1(i32 31) %2 = tail call <vscale x 8 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv8i1(<vscale x 16 x i1> %1) %3 = tail call i64 @llvm.aarch64.sve.cntp.nxv8i1(<vscale x 8 x i1> %2, <vscale x 8 x i1> %pg) %add = add i64 %3, %x ret i64 %add } define i64 @cntp_add_all_active_nxv2i1_oneuse(i64 %x, <vscale x 2 x i1> %pg) #0 { ; CHECK-LABEL: cntp_add_all_active_nxv2i1_oneuse: ; CHECK: // %bb.0: ; CHECK-NEXT: ptrue p1.d ; CHECK-NEXT: cntp x8, p1, p0.d ; CHECK-NEXT: add x9, x8, x0 ; CHECK-NEXT: madd x0, x8, x0, x9 ; CHECK-NEXT: ret %1 = tail call <vscale x 2 x i1> @llvm.aarch64.sve.ptrue.nxv2i1(i32 31) %2 = tail call i64 @llvm.aarch64.sve.cntp.nxv2i1(<vscale x 2 x i1> %1, <vscale x 2 x i1> %pg) %add = add i64 %2, %x %mul = mul i64 %2, %x %res = add i64 %add, %mul ret i64 %res } define i64 @cntp_add_same_active_nxv16i1(i64 %x, <vscale x 16 x i1> %pg) #0 { ; CHECK-LABEL: cntp_add_same_active_nxv16i1: ; CHECK: // %bb.0: ; CHECK-NEXT: incp x0, p0.b ; CHECK-NEXT: ret %1 = tail call i64 @llvm.aarch64.sve.cntp.nxv16i1(<vscale x 16 x i1> %pg, <vscale x 16 x i1> %pg) %add = add i64 %1, %x ret i64 %add } define i64 @cntp_add_same_active_nxv8i1(i64 %x, <vscale x 8 x i1> %pg) #0 { ; CHECK-LABEL: cntp_add_same_active_nxv8i1: ; CHECK: // %bb.0: ; CHECK-NEXT: incp x0, p0.h ; CHECK-NEXT: ret %1 = tail call i64 @llvm.aarch64.sve.cntp.nxv8i1(<vscale x 8 x i1> %pg, <vscale x 8 x i1> %pg) %add = add i64 %1, %x ret i64 %add } define i64 @cntp_add_same_active_nxv4i1(i64 %x, <vscale x 4 x i1> %pg) #0 { ; CHECK-LABEL: cntp_add_same_active_nxv4i1: ; CHECK: // %bb.0: ; CHECK-NEXT: incp x0, p0.s ; CHECK-NEXT: ret %1 = tail call i64 @llvm.aarch64.sve.cntp.nxv4i1(<vscale x 4 x i1> %pg, <vscale x 4 x i1> %pg) %add = add i64 %1, %x ret i64 %add } define i64 @cntp_add_same_active_nxv2i1(i64 %x, <vscale x 2 x i1> %pg) #0 { ; CHECK-LABEL: cntp_add_same_active_nxv2i1: ; CHECK: // %bb.0: ; CHECK-NEXT: incp x0, p0.d ; CHECK-NEXT: ret %1 = tail call i64 @llvm.aarch64.sve.cntp.nxv2i1(<vscale x 2 x i1> %pg, <vscale x 2 x i1> %pg) %add = add i64 %1, %x ret i64 %add } define i64 @cntp_add_same_active_nxv2i1_oneuse(i64 %x, <vscale x 2 x i1> %pg) #0 { ; CHECK-LABEL: cntp_add_same_active_nxv2i1_oneuse: ; CHECK: // %bb.0: ; CHECK-NEXT: cntp x8, p0, p0.d ; CHECK-NEXT: add x9, x8, x0 ; CHECK-NEXT: madd x0, x8, x0, x9 ; CHECK-NEXT: ret %1 = tail call i64 @llvm.aarch64.sve.cntp.nxv2i1(<vscale x 2 x i1> %pg, <vscale x 2 x i1> %pg) %add = add i64 %1, %x %mul = mul i64 %1, %x %res = add i64 %add, %mul ret i64 %res } ; DECP define i64 @cntp_sub_all_active_nxv16i1(i64 %x, <vscale x 16 x i1> %pg) #0 { ; CHECK-LABEL: cntp_sub_all_active_nxv16i1: ; CHECK: // %bb.0: ; CHECK-NEXT: decp x0, p0.b ; CHECK-NEXT: ret %1 = tail call <vscale x 16 x i1> @llvm.aarch64.sve.ptrue.nxv16i1(i32 31) %2 = tail call i64 @llvm.aarch64.sve.cntp.nxv16i1(<vscale x 16 x i1> %1, <vscale x 16 x i1> %pg) %sub = sub i64 %x, %2 ret i64 %sub } define i64 @cntp_sub_all_active_nxv8i1(i64 %x, <vscale x 8 x i1> %pg) #0 { ; CHECK-LABEL: cntp_sub_all_active_nxv8i1: ; CHECK: // %bb.0: ; CHECK-NEXT: decp x0, p0.h ; CHECK-NEXT: ret %1 = tail call <vscale x 8 x i1> @llvm.aarch64.sve.ptrue.nxv8i1(i32 31) %2 = tail call i64 @llvm.aarch64.sve.cntp.nxv8i1(<vscale x 8 x i1> %1, <vscale x 8 x i1> %pg) %sub = sub i64 %x, %2 ret i64 %sub } define i64 @cntp_sub_all_active_nxv4i1(i64 %x, <vscale x 4 x i1> %pg) #0 { ; CHECK-LABEL: cntp_sub_all_active_nxv4i1: ; CHECK: // %bb.0: ; CHECK-NEXT: decp x0, p0.s ; CHECK-NEXT: ret %1 = tail call <vscale x 4 x i1> @llvm.aarch64.sve.ptrue.nxv4i1(i32 31) %2 = tail call i64 @llvm.aarch64.sve.cntp.nxv4i1(<vscale x 4 x i1> %1, <vscale x 4 x i1> %pg) %sub = sub i64 %x, %2 ret i64 %sub } define i64 @cntp_sub_all_active_nxv2i1(i64 %x, <vscale x 2 x i1> %pg) #0 { ; CHECK-LABEL: cntp_sub_all_active_nxv2i1: ; CHECK: // %bb.0: ; CHECK-NEXT: decp x0, p0.d ; CHECK-NEXT: ret %1 = tail call <vscale x 2 x i1> @llvm.aarch64.sve.ptrue.nxv2i1(i32 31) %2 = tail call i64 @llvm.aarch64.sve.cntp.nxv2i1(<vscale x 2 x i1> %1, <vscale x 2 x i1> %pg) %sub = sub i64 %x, %2 ret i64 %sub } define i64 @cntp_sub_all_active_nxv8i1_via_cast(i64 %x, <vscale x 8 x i1> %pg) #0 { ; CHECK-LABEL: cntp_sub_all_active_nxv8i1_via_cast: ; CHECK: // %bb.0: ; CHECK-NEXT: decp x0, p0.h ; CHECK-NEXT: ret %1 = tail call <vscale x 16 x i1> @llvm.aarch64.sve.ptrue.nxv16i1(i32 31) %2 = tail call <vscale x 8 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv8i1(<vscale x 16 x i1> %1) %3 = tail call i64 @llvm.aarch64.sve.cntp.nxv8i1(<vscale x 8 x i1> %2, <vscale x 8 x i1> %pg) %sub = sub i64 %x, %3 ret i64 %sub } define i64 @cntp_sub_all_active_nxv2i1_multiuse(i64 %x, <vscale x 2 x i1> %pg) #0 { ; CHECK-LABEL: cntp_sub_all_active_nxv2i1_multiuse: ; CHECK: // %bb.0: ; CHECK-NEXT: ptrue p1.d ; CHECK-NEXT: cntp x8, p1, p0.d ; CHECK-NEXT: sub x9, x8, x0 ; CHECK-NEXT: madd x0, x8, x0, x9 ; CHECK-NEXT: ret %1 = tail call <vscale x 2 x i1> @llvm.aarch64.sve.ptrue.nxv2i1(i32 31) %2 = tail call i64 @llvm.aarch64.sve.cntp.nxv2i1(<vscale x 2 x i1> %1, <vscale x 2 x i1> %pg) %add = sub i64 %2, %x %mul = mul i64 %2, %x %res = add i64 %add, %mul ret i64 %res } define i64 @cntp_sub_same_active_nxv16i1(i64 %x, <vscale x 16 x i1> %pg) #0 { ; CHECK-LABEL: cntp_sub_same_active_nxv16i1: ; CHECK: // %bb.0: ; CHECK-NEXT: decp x0, p0.b ; CHECK-NEXT: ret %1 = tail call i64 @llvm.aarch64.sve.cntp.nxv16i1(<vscale x 16 x i1> %pg, <vscale x 16 x i1> %pg) %sub = sub i64 %x, %1 ret i64 %sub } define i64 @cntp_sub_same_active_nxv8i1(i64 %x, <vscale x 8 x i1> %pg) #0 { ; CHECK-LABEL: cntp_sub_same_active_nxv8i1: ; CHECK: // %bb.0: ; CHECK-NEXT: decp x0, p0.h ; CHECK-NEXT: ret %1 = tail call i64 @llvm.aarch64.sve.cntp.nxv8i1(<vscale x 8 x i1> %pg, <vscale x 8 x i1> %pg) %sub = sub i64 %x, %1 ret i64 %sub } define i64 @cntp_sub_same_active_nxv4i1(i64 %x, <vscale x 4 x i1> %pg) #0 { ; CHECK-LABEL: cntp_sub_same_active_nxv4i1: ; CHECK: // %bb.0: ; CHECK-NEXT: decp x0, p0.s ; CHECK-NEXT: ret %1 = tail call i64 @llvm.aarch64.sve.cntp.nxv4i1(<vscale x 4 x i1> %pg, <vscale x 4 x i1> %pg) %sub = sub i64 %x, %1 ret i64 %sub } define i64 @cntp_sub_same_active_nxv2i1(i64 %x, <vscale x 2 x i1> %pg) #0 { ; CHECK-LABEL: cntp_sub_same_active_nxv2i1: ; CHECK: // %bb.0: ; CHECK-NEXT: decp x0, p0.d ; CHECK-NEXT: ret %1 = tail call i64 @llvm.aarch64.sve.cntp.nxv2i1(<vscale x 2 x i1> %pg, <vscale x 2 x i1> %pg) %sub = sub i64 %x, %1 ret i64 %sub } define i64 @cntp_sub_same_active_nxv2i1_multiuse(i64 %x, <vscale x 2 x i1> %pg) #0 { ; CHECK-LABEL: cntp_sub_same_active_nxv2i1_multiuse: ; CHECK: // %bb.0: ; CHECK-NEXT: cntp x8, p0, p0.d ; CHECK-NEXT: sub x9, x8, x0 ; CHECK-NEXT: madd x0, x8, x0, x9 ; CHECK-NEXT: ret %1 = tail call i64 @llvm.aarch64.sve.cntp.nxv2i1(<vscale x 2 x i1> %pg, <vscale x 2 x i1> %pg) %add = sub i64 %1, %x %mul = mul i64 %1, %x %res = add i64 %add, %mul ret i64 %res } declare <vscale x 8 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv8i1(<vscale x 16 x i1>) declare <vscale x 4 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv4i1(<vscale x 16 x i1>) declare <vscale x 2 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv2i1(<vscale x 16 x i1>) declare <vscale x 16 x i1> @llvm.aarch64.sve.ptrue.nxv16i1(i32) declare <vscale x 8 x i1> @llvm.aarch64.sve.ptrue.nxv8i1(i32) declare <vscale x 4 x i1> @llvm.aarch64.sve.ptrue.nxv4i1(i32) declare <vscale x 2 x i1> @llvm.aarch64.sve.ptrue.nxv2i1(i32) declare i64 @llvm.aarch64.sve.cntp.nxv16i1(<vscale x 16 x i1>, <vscale x 16 x i1>) declare i64 @llvm.aarch64.sve.cntp.nxv8i1(<vscale x 8 x i1>, <vscale x 8 x i1>) declare i64 @llvm.aarch64.sve.cntp.nxv4i1(<vscale x 4 x i1>, <vscale x 4 x i1>) declare i64 @llvm.aarch64.sve.cntp.nxv2i1(<vscale x 2 x i1>, <vscale x 2 x i1>) attributes #0 = { "target-features"="+sve" }