; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py ; RUN: llc -mtriple=aarch64-linux-gnu -mattr=+sve < %s | FileCheck %s ; ; CLASTA (Vectors) ; define <vscale x 16 x i8> @clasta_i8(<vscale x 16 x i1> %pg, <vscale x 16 x i8> %a, <vscale x 16 x i8> %b) { ; CHECK-LABEL: clasta_i8: ; CHECK: // %bb.0: ; CHECK-NEXT: clasta z0.b, p0, z0.b, z1.b ; CHECK-NEXT: ret %out = call <vscale x 16 x i8> @llvm.aarch64.sve.clasta.nxv16i8(<vscale x 16 x i1> %pg, <vscale x 16 x i8> %a, <vscale x 16 x i8> %b) ret <vscale x 16 x i8> %out } define <vscale x 8 x i16> @clasta_i16(<vscale x 8 x i1> %pg, <vscale x 8 x i16> %a, <vscale x 8 x i16> %b) { ; CHECK-LABEL: clasta_i16: ; CHECK: // %bb.0: ; CHECK-NEXT: clasta z0.h, p0, z0.h, z1.h ; CHECK-NEXT: ret %out = call <vscale x 8 x i16> @llvm.aarch64.sve.clasta.nxv8i16(<vscale x 8 x i1> %pg, <vscale x 8 x i16> %a, <vscale x 8 x i16> %b) ret <vscale x 8 x i16> %out } define <vscale x 4 x i32> @clasta_i32(<vscale x 4 x i1> %pg, <vscale x 4 x i32> %a, <vscale x 4 x i32> %b) { ; CHECK-LABEL: clasta_i32: ; CHECK: // %bb.0: ; CHECK-NEXT: clasta z0.s, p0, z0.s, z1.s ; CHECK-NEXT: ret %out = call <vscale x 4 x i32> @llvm.aarch64.sve.clasta.nxv4i32(<vscale x 4 x i1> %pg, <vscale x 4 x i32> %a, <vscale x 4 x i32> %b) ret <vscale x 4 x i32> %out } define <vscale x 2 x i64> @clasta_i64(<vscale x 2 x i1> %pg, <vscale x 2 x i64> %a, <vscale x 2 x i64> %b) { ; CHECK-LABEL: clasta_i64: ; CHECK: // %bb.0: ; CHECK-NEXT: clasta z0.d, p0, z0.d, z1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x i64> @llvm.aarch64.sve.clasta.nxv2i64(<vscale x 2 x i1> %pg, <vscale x 2 x i64> %a, <vscale x 2 x i64> %b) ret <vscale x 2 x i64> %out } define <vscale x 8 x half> @clasta_f16(<vscale x 8 x i1> %pg, <vscale x 8 x half> %a, <vscale x 8 x half> %b) { ; CHECK-LABEL: clasta_f16: ; CHECK: // %bb.0: ; CHECK-NEXT: clasta z0.h, p0, z0.h, z1.h ; CHECK-NEXT: ret %out = call <vscale x 8 x half> @llvm.aarch64.sve.clasta.nxv8f16(<vscale x 8 x i1> %pg, <vscale x 8 x half> %a, <vscale x 8 x half> %b) ret <vscale x 8 x half> %out } define <vscale x 8 x bfloat> @clasta_bf16(<vscale x 8 x i1> %pg, <vscale x 8 x bfloat> %a, <vscale x 8 x bfloat> %b) #0 { ; CHECK-LABEL: clasta_bf16: ; CHECK: // %bb.0: ; CHECK-NEXT: clasta z0.h, p0, z0.h, z1.h ; CHECK-NEXT: ret %out = call <vscale x 8 x bfloat> @llvm.aarch64.sve.clasta.nxv8bf16(<vscale x 8 x i1> %pg, <vscale x 8 x bfloat> %a, <vscale x 8 x bfloat> %b) ret <vscale x 8 x bfloat> %out } define <vscale x 4 x float> @clasta_f32(<vscale x 4 x i1> %pg, <vscale x 4 x float> %a, <vscale x 4 x float> %b) { ; CHECK-LABEL: clasta_f32: ; CHECK: // %bb.0: ; CHECK-NEXT: clasta z0.s, p0, z0.s, z1.s ; CHECK-NEXT: ret %out = call <vscale x 4 x float> @llvm.aarch64.sve.clasta.nxv4f32(<vscale x 4 x i1> %pg, <vscale x 4 x float> %a, <vscale x 4 x float> %b) ret <vscale x 4 x float> %out } define <vscale x 2 x double> @clasta_f64(<vscale x 2 x i1> %pg, <vscale x 2 x double> %a, <vscale x 2 x double> %b) { ; CHECK-LABEL: clasta_f64: ; CHECK: // %bb.0: ; CHECK-NEXT: clasta z0.d, p0, z0.d, z1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x double> @llvm.aarch64.sve.clasta.nxv2f64(<vscale x 2 x i1> %pg, <vscale x 2 x double> %a, <vscale x 2 x double> %b) ret <vscale x 2 x double> %out } ; ; CLASTA (Scalar) ; define i8 @clasta_n_i8(<vscale x 16 x i1> %pg, i8 %a, <vscale x 16 x i8> %b) { ; CHECK-LABEL: clasta_n_i8: ; CHECK: // %bb.0: ; CHECK-NEXT: clasta w0, p0, w0, z0.b ; CHECK-NEXT: ret %out = call i8 @llvm.aarch64.sve.clasta.n.nxv16i8(<vscale x 16 x i1> %pg, i8 %a, <vscale x 16 x i8> %b) ret i8 %out } define i16 @clasta_n_i16(<vscale x 8 x i1> %pg, i16 %a, <vscale x 8 x i16> %b) { ; CHECK-LABEL: clasta_n_i16: ; CHECK: // %bb.0: ; CHECK-NEXT: clasta w0, p0, w0, z0.h ; CHECK-NEXT: ret %out = call i16 @llvm.aarch64.sve.clasta.n.nxv8i16(<vscale x 8 x i1> %pg, i16 %a, <vscale x 8 x i16> %b) ret i16 %out } define i32 @clasta_n_i32(<vscale x 4 x i1> %pg, i32 %a, <vscale x 4 x i32> %b) { ; CHECK-LABEL: clasta_n_i32: ; CHECK: // %bb.0: ; CHECK-NEXT: clasta w0, p0, w0, z0.s ; CHECK-NEXT: ret %out = call i32 @llvm.aarch64.sve.clasta.n.nxv4i32(<vscale x 4 x i1> %pg, i32 %a, <vscale x 4 x i32> %b) ret i32 %out } define i64 @clasta_n_i64(<vscale x 2 x i1> %pg, i64 %a, <vscale x 2 x i64> %b) { ; CHECK-LABEL: clasta_n_i64: ; CHECK: // %bb.0: ; CHECK-NEXT: clasta x0, p0, x0, z0.d ; CHECK-NEXT: ret %out = call i64 @llvm.aarch64.sve.clasta.n.nxv2i64(<vscale x 2 x i1> %pg, i64 %a, <vscale x 2 x i64> %b) ret i64 %out } define half @clasta_n_f16(<vscale x 8 x i1> %pg, half %a, <vscale x 8 x half> %b) { ; CHECK-LABEL: clasta_n_f16: ; CHECK: // %bb.0: ; CHECK-NEXT: clasta h0, p0, h0, z1.h ; CHECK-NEXT: ret %out = call half @llvm.aarch64.sve.clasta.n.nxv8f16(<vscale x 8 x i1> %pg, half %a, <vscale x 8 x half> %b) ret half %out } define bfloat @clasta_n_bf16(<vscale x 8 x i1> %pg, bfloat %a, <vscale x 8 x bfloat> %b) #0 { ; CHECK-LABEL: clasta_n_bf16: ; CHECK: // %bb.0: ; CHECK-NEXT: clasta h0, p0, h0, z1.h ; CHECK-NEXT: ret %out = call bfloat @llvm.aarch64.sve.clasta.n.nxv8bf16(<vscale x 8 x i1> %pg, bfloat %a, <vscale x 8 x bfloat> %b) ret bfloat %out } define float @clasta_n_f32(<vscale x 4 x i1> %pg, float %a, <vscale x 4 x float> %b) { ; CHECK-LABEL: clasta_n_f32: ; CHECK: // %bb.0: ; CHECK-NEXT: clasta s0, p0, s0, z1.s ; CHECK-NEXT: ret %out = call float @llvm.aarch64.sve.clasta.n.nxv4f32(<vscale x 4 x i1> %pg, float %a, <vscale x 4 x float> %b) ret float %out } define double @clasta_n_f64(<vscale x 2 x i1> %pg, double %a, <vscale x 2 x double> %b) { ; CHECK-LABEL: clasta_n_f64: ; CHECK: // %bb.0: ; CHECK-NEXT: clasta d0, p0, d0, z1.d ; CHECK-NEXT: ret %out = call double @llvm.aarch64.sve.clasta.n.nxv2f64(<vscale x 2 x i1> %pg, double %a, <vscale x 2 x double> %b) ret double %out } ; ; CLASTB (Vectors) ; define <vscale x 16 x i8> @clastb_i8(<vscale x 16 x i1> %pg, <vscale x 16 x i8> %a, <vscale x 16 x i8> %b) { ; CHECK-LABEL: clastb_i8: ; CHECK: // %bb.0: ; CHECK-NEXT: clastb z0.b, p0, z0.b, z1.b ; CHECK-NEXT: ret %out = call <vscale x 16 x i8> @llvm.aarch64.sve.clastb.nxv16i8(<vscale x 16 x i1> %pg, <vscale x 16 x i8> %a, <vscale x 16 x i8> %b) ret <vscale x 16 x i8> %out } define <vscale x 8 x i16> @clastb_i16(<vscale x 8 x i1> %pg, <vscale x 8 x i16> %a, <vscale x 8 x i16> %b) { ; CHECK-LABEL: clastb_i16: ; CHECK: // %bb.0: ; CHECK-NEXT: clastb z0.h, p0, z0.h, z1.h ; CHECK-NEXT: ret %out = call <vscale x 8 x i16> @llvm.aarch64.sve.clastb.nxv8i16(<vscale x 8 x i1> %pg, <vscale x 8 x i16> %a, <vscale x 8 x i16> %b) ret <vscale x 8 x i16> %out } define <vscale x 4 x i32> @clastb_i32(<vscale x 4 x i1> %pg, <vscale x 4 x i32> %a, <vscale x 4 x i32> %b) { ; CHECK-LABEL: clastb_i32: ; CHECK: // %bb.0: ; CHECK-NEXT: clastb z0.s, p0, z0.s, z1.s ; CHECK-NEXT: ret %out = call <vscale x 4 x i32> @llvm.aarch64.sve.clastb.nxv4i32(<vscale x 4 x i1> %pg, <vscale x 4 x i32> %a, <vscale x 4 x i32> %b) ret <vscale x 4 x i32> %out } define <vscale x 2 x i64> @clastb_i64(<vscale x 2 x i1> %pg, <vscale x 2 x i64> %a, <vscale x 2 x i64> %b) { ; CHECK-LABEL: clastb_i64: ; CHECK: // %bb.0: ; CHECK-NEXT: clastb z0.d, p0, z0.d, z1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x i64> @llvm.aarch64.sve.clastb.nxv2i64(<vscale x 2 x i1> %pg, <vscale x 2 x i64> %a, <vscale x 2 x i64> %b) ret <vscale x 2 x i64> %out } define <vscale x 8 x half> @clastb_f16(<vscale x 8 x i1> %pg, <vscale x 8 x half> %a, <vscale x 8 x half> %b) { ; CHECK-LABEL: clastb_f16: ; CHECK: // %bb.0: ; CHECK-NEXT: clastb z0.h, p0, z0.h, z1.h ; CHECK-NEXT: ret %out = call <vscale x 8 x half> @llvm.aarch64.sve.clastb.nxv8f16(<vscale x 8 x i1> %pg, <vscale x 8 x half> %a, <vscale x 8 x half> %b) ret <vscale x 8 x half> %out } define <vscale x 8 x bfloat> @clastb_bf16(<vscale x 8 x i1> %pg, <vscale x 8 x bfloat> %a, <vscale x 8 x bfloat> %b) #0 { ; CHECK-LABEL: clastb_bf16: ; CHECK: // %bb.0: ; CHECK-NEXT: clastb z0.h, p0, z0.h, z1.h ; CHECK-NEXT: ret %out = call <vscale x 8 x bfloat> @llvm.aarch64.sve.clastb.nxv8bf16(<vscale x 8 x i1> %pg, <vscale x 8 x bfloat> %a, <vscale x 8 x bfloat> %b) ret <vscale x 8 x bfloat> %out } define <vscale x 4 x float> @clastb_f32(<vscale x 4 x i1> %pg, <vscale x 4 x float> %a, <vscale x 4 x float> %b) { ; CHECK-LABEL: clastb_f32: ; CHECK: // %bb.0: ; CHECK-NEXT: clastb z0.s, p0, z0.s, z1.s ; CHECK-NEXT: ret %out = call <vscale x 4 x float> @llvm.aarch64.sve.clastb.nxv4f32(<vscale x 4 x i1> %pg, <vscale x 4 x float> %a, <vscale x 4 x float> %b) ret <vscale x 4 x float> %out } define <vscale x 2 x double> @clastb_f64(<vscale x 2 x i1> %pg, <vscale x 2 x double> %a, <vscale x 2 x double> %b) { ; CHECK-LABEL: clastb_f64: ; CHECK: // %bb.0: ; CHECK-NEXT: clastb z0.d, p0, z0.d, z1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x double> @llvm.aarch64.sve.clastb.nxv2f64(<vscale x 2 x i1> %pg, <vscale x 2 x double> %a, <vscale x 2 x double> %b) ret <vscale x 2 x double> %out } ; ; CLASTB (Scalar) ; define i8 @clastb_n_i8(<vscale x 16 x i1> %pg, i8 %a, <vscale x 16 x i8> %b) { ; CHECK-LABEL: clastb_n_i8: ; CHECK: // %bb.0: ; CHECK-NEXT: clastb w0, p0, w0, z0.b ; CHECK-NEXT: ret %out = call i8 @llvm.aarch64.sve.clastb.n.nxv16i8(<vscale x 16 x i1> %pg, i8 %a, <vscale x 16 x i8> %b) ret i8 %out } define i16 @clastb_n_i16(<vscale x 8 x i1> %pg, i16 %a, <vscale x 8 x i16> %b) { ; CHECK-LABEL: clastb_n_i16: ; CHECK: // %bb.0: ; CHECK-NEXT: clastb w0, p0, w0, z0.h ; CHECK-NEXT: ret %out = call i16 @llvm.aarch64.sve.clastb.n.nxv8i16(<vscale x 8 x i1> %pg, i16 %a, <vscale x 8 x i16> %b) ret i16 %out } define i32 @clastb_n_i32(<vscale x 4 x i1> %pg, i32 %a, <vscale x 4 x i32> %b) { ; CHECK-LABEL: clastb_n_i32: ; CHECK: // %bb.0: ; CHECK-NEXT: clastb w0, p0, w0, z0.s ; CHECK-NEXT: ret %out = call i32 @llvm.aarch64.sve.clastb.n.nxv4i32(<vscale x 4 x i1> %pg, i32 %a, <vscale x 4 x i32> %b) ret i32 %out } define i64 @clastb_n_i64(<vscale x 2 x i1> %pg, i64 %a, <vscale x 2 x i64> %b) { ; CHECK-LABEL: clastb_n_i64: ; CHECK: // %bb.0: ; CHECK-NEXT: clastb x0, p0, x0, z0.d ; CHECK-NEXT: ret %out = call i64 @llvm.aarch64.sve.clastb.n.nxv2i64(<vscale x 2 x i1> %pg, i64 %a, <vscale x 2 x i64> %b) ret i64 %out } define half @clastb_n_f16(<vscale x 8 x i1> %pg, half %a, <vscale x 8 x half> %b) { ; CHECK-LABEL: clastb_n_f16: ; CHECK: // %bb.0: ; CHECK-NEXT: clastb h0, p0, h0, z1.h ; CHECK-NEXT: ret %out = call half @llvm.aarch64.sve.clastb.n.nxv8f16(<vscale x 8 x i1> %pg, half %a, <vscale x 8 x half> %b) ret half %out } define bfloat @clastb_n_bf16(<vscale x 8 x i1> %pg, bfloat %a, <vscale x 8 x bfloat> %b) #0 { ; CHECK-LABEL: clastb_n_bf16: ; CHECK: // %bb.0: ; CHECK-NEXT: clastb h0, p0, h0, z1.h ; CHECK-NEXT: ret %out = call bfloat @llvm.aarch64.sve.clastb.n.nxv8bf16(<vscale x 8 x i1> %pg, bfloat %a, <vscale x 8 x bfloat> %b) ret bfloat %out } define float @clastb_n_f32(<vscale x 4 x i1> %pg, float %a, <vscale x 4 x float> %b) { ; CHECK-LABEL: clastb_n_f32: ; CHECK: // %bb.0: ; CHECK-NEXT: clastb s0, p0, s0, z1.s ; CHECK-NEXT: ret %out = call float @llvm.aarch64.sve.clastb.n.nxv4f32(<vscale x 4 x i1> %pg, float %a, <vscale x 4 x float> %b) ret float %out } define double @clastb_n_f64(<vscale x 2 x i1> %pg, double %a, <vscale x 2 x double> %b) { ; CHECK-LABEL: clastb_n_f64: ; CHECK: // %bb.0: ; CHECK-NEXT: clastb d0, p0, d0, z1.d ; CHECK-NEXT: ret %out = call double @llvm.aarch64.sve.clastb.n.nxv2f64(<vscale x 2 x i1> %pg, double %a, <vscale x 2 x double> %b) ret double %out } ; ; DUPQ ; define <vscale x 16 x i8> @dupq_i8(<vscale x 16 x i8> %a) { ; CHECK-LABEL: dupq_i8: ; CHECK: // %bb.0: ; CHECK-NEXT: mov z0.q, q0 ; CHECK-NEXT: ret %out = call <vscale x 16 x i8> @llvm.aarch64.sve.dupq.lane.nxv16i8(<vscale x 16 x i8> %a, i64 0) ret <vscale x 16 x i8> %out } define <vscale x 8 x i16> @dupq_i16(<vscale x 8 x i16> %a) { ; CHECK-LABEL: dupq_i16: ; CHECK: // %bb.0: ; CHECK-NEXT: mov z0.q, z0.q[1] ; CHECK-NEXT: ret %out = call <vscale x 8 x i16> @llvm.aarch64.sve.dupq.lane.nxv8i16(<vscale x 8 x i16> %a, i64 1) ret <vscale x 8 x i16> %out } define <vscale x 4 x i32> @dupq_i32(<vscale x 4 x i32> %a) { ; CHECK-LABEL: dupq_i32: ; CHECK: // %bb.0: ; CHECK-NEXT: mov z0.q, z0.q[2] ; CHECK-NEXT: ret %out = call <vscale x 4 x i32> @llvm.aarch64.sve.dupq.lane.nxv4i32(<vscale x 4 x i32> %a, i64 2) ret <vscale x 4 x i32> %out } define <vscale x 2 x i64> @dupq_i64(<vscale x 2 x i64> %a) { ; CHECK-LABEL: dupq_i64: ; CHECK: // %bb.0: ; CHECK-NEXT: mov z0.q, z0.q[3] ; CHECK-NEXT: ret %out = call <vscale x 2 x i64> @llvm.aarch64.sve.dupq.lane.nxv2i64(<vscale x 2 x i64> %a, i64 3) ret <vscale x 2 x i64> %out } define <vscale x 8 x half> @dupq_f16(<vscale x 8 x half> %a) { ; CHECK-LABEL: dupq_f16: ; CHECK: // %bb.0: ; CHECK-NEXT: mov z0.q, q0 ; CHECK-NEXT: ret %out = call <vscale x 8 x half> @llvm.aarch64.sve.dupq.lane.nxv8f16(<vscale x 8 x half> %a, i64 0) ret <vscale x 8 x half> %out } define <vscale x 8 x bfloat> @dupq_bf16(<vscale x 8 x bfloat> %a) #0 { ; CHECK-LABEL: dupq_bf16: ; CHECK: // %bb.0: ; CHECK-NEXT: mov z0.q, q0 ; CHECK-NEXT: ret %out = call <vscale x 8 x bfloat> @llvm.aarch64.sve.dupq.lane.nxv8bf16(<vscale x 8 x bfloat> %a, i64 0) ret <vscale x 8 x bfloat> %out } define <vscale x 4 x float> @dupq_f32(<vscale x 4 x float> %a) { ; CHECK-LABEL: dupq_f32: ; CHECK: // %bb.0: ; CHECK-NEXT: mov z0.q, z0.q[1] ; CHECK-NEXT: ret %out = call <vscale x 4 x float> @llvm.aarch64.sve.dupq.lane.nxv4f32(<vscale x 4 x float> %a, i64 1) ret <vscale x 4 x float> %out } define <vscale x 2 x double> @dupq_f64(<vscale x 2 x double> %a) { ; CHECK-LABEL: dupq_f64: ; CHECK: // %bb.0: ; CHECK-NEXT: mov z0.q, z0.q[2] ; CHECK-NEXT: ret %out = call <vscale x 2 x double> @llvm.aarch64.sve.dupq.lane.nxv2f64(<vscale x 2 x double> %a, i64 2) ret <vscale x 2 x double> %out } ; ; DUPQ_LANE ; define <vscale x 16 x i8> @dupq_lane_i8(<vscale x 16 x i8> %a, i64 %idx) { ; CHECK-LABEL: dupq_lane_i8: ; CHECK: // %bb.0: ; CHECK-NEXT: add x8, x0, x0 ; CHECK-NEXT: index z1.d, #0, #1 ; CHECK-NEXT: and z1.d, z1.d, #0x1 ; CHECK-NEXT: mov z2.d, x8 ; CHECK-NEXT: add z1.d, z1.d, z2.d ; CHECK-NEXT: tbl z0.d, { z0.d }, z1.d ; CHECK-NEXT: ret %out = call <vscale x 16 x i8> @llvm.aarch64.sve.dupq.lane.nxv16i8(<vscale x 16 x i8> %a, i64 %idx) ret <vscale x 16 x i8> %out } ; NOTE: Identical operation to dupq_lane_i8 (i.e. element type is irrelevant). define <vscale x 8 x i16> @dupq_lane_i16(<vscale x 8 x i16> %a, i64 %idx) { ; CHECK-LABEL: dupq_lane_i16: ; CHECK: // %bb.0: ; CHECK-NEXT: add x8, x0, x0 ; CHECK-NEXT: index z1.d, #0, #1 ; CHECK-NEXT: and z1.d, z1.d, #0x1 ; CHECK-NEXT: mov z2.d, x8 ; CHECK-NEXT: add z1.d, z1.d, z2.d ; CHECK-NEXT: tbl z0.d, { z0.d }, z1.d ; CHECK-NEXT: ret %out = call <vscale x 8 x i16> @llvm.aarch64.sve.dupq.lane.nxv8i16(<vscale x 8 x i16> %a, i64 %idx) ret <vscale x 8 x i16> %out } ; NOTE: Identical operation to dupq_lane_i8 (i.e. element type is irrelevant). define <vscale x 4 x i32> @dupq_lane_i32(<vscale x 4 x i32> %a, i64 %idx) { ; CHECK-LABEL: dupq_lane_i32: ; CHECK: // %bb.0: ; CHECK-NEXT: add x8, x0, x0 ; CHECK-NEXT: index z1.d, #0, #1 ; CHECK-NEXT: and z1.d, z1.d, #0x1 ; CHECK-NEXT: mov z2.d, x8 ; CHECK-NEXT: add z1.d, z1.d, z2.d ; CHECK-NEXT: tbl z0.d, { z0.d }, z1.d ; CHECK-NEXT: ret %out = call <vscale x 4 x i32> @llvm.aarch64.sve.dupq.lane.nxv4i32(<vscale x 4 x i32> %a, i64 %idx) ret <vscale x 4 x i32> %out } ; NOTE: Identical operation to dupq_lane_i8 (i.e. element type is irrelevant). define <vscale x 2 x i64> @dupq_lane_i64(<vscale x 2 x i64> %a, i64 %idx) { ; CHECK-LABEL: dupq_lane_i64: ; CHECK: // %bb.0: ; CHECK-NEXT: add x8, x0, x0 ; CHECK-NEXT: index z1.d, #0, #1 ; CHECK-NEXT: and z1.d, z1.d, #0x1 ; CHECK-NEXT: mov z2.d, x8 ; CHECK-NEXT: add z1.d, z1.d, z2.d ; CHECK-NEXT: tbl z0.d, { z0.d }, z1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x i64> @llvm.aarch64.sve.dupq.lane.nxv2i64(<vscale x 2 x i64> %a, i64 %idx) ret <vscale x 2 x i64> %out } ; NOTE: Identical operation to dupq_lane_i8 (i.e. element type is irrelevant). define <vscale x 8 x half> @dupq_lane_f16(<vscale x 8 x half> %a, i64 %idx) { ; CHECK-LABEL: dupq_lane_f16: ; CHECK: // %bb.0: ; CHECK-NEXT: add x8, x0, x0 ; CHECK-NEXT: index z1.d, #0, #1 ; CHECK-NEXT: and z1.d, z1.d, #0x1 ; CHECK-NEXT: mov z2.d, x8 ; CHECK-NEXT: add z1.d, z1.d, z2.d ; CHECK-NEXT: tbl z0.d, { z0.d }, z1.d ; CHECK-NEXT: ret %out = call <vscale x 8 x half> @llvm.aarch64.sve.dupq.lane.nxv8f16(<vscale x 8 x half> %a, i64 %idx) ret <vscale x 8 x half> %out } ; NOTE: Identical operation to dupq_lane_i8 (i.e. element type is irrelevant). define <vscale x 8 x bfloat> @dupq_lane_bf16(<vscale x 8 x bfloat> %a, i64 %idx) #0 { ; CHECK-LABEL: dupq_lane_bf16: ; CHECK: // %bb.0: ; CHECK-NEXT: add x8, x0, x0 ; CHECK-NEXT: index z1.d, #0, #1 ; CHECK-NEXT: and z1.d, z1.d, #0x1 ; CHECK-NEXT: mov z2.d, x8 ; CHECK-NEXT: add z1.d, z1.d, z2.d ; CHECK-NEXT: tbl z0.d, { z0.d }, z1.d ; CHECK-NEXT: ret %out = call <vscale x 8 x bfloat> @llvm.aarch64.sve.dupq.lane.nxv8bf16(<vscale x 8 x bfloat> %a, i64 %idx) ret <vscale x 8 x bfloat> %out } ; NOTE: Identical operation to dupq_lane_i8 (i.e. element type is irrelevant). define <vscale x 4 x float> @dupq_lane_f32(<vscale x 4 x float> %a, i64 %idx) { ; CHECK-LABEL: dupq_lane_f32: ; CHECK: // %bb.0: ; CHECK-NEXT: add x8, x0, x0 ; CHECK-NEXT: index z1.d, #0, #1 ; CHECK-NEXT: and z1.d, z1.d, #0x1 ; CHECK-NEXT: mov z2.d, x8 ; CHECK-NEXT: add z1.d, z1.d, z2.d ; CHECK-NEXT: tbl z0.d, { z0.d }, z1.d ; CHECK-NEXT: ret %out = call <vscale x 4 x float> @llvm.aarch64.sve.dupq.lane.nxv4f32(<vscale x 4 x float> %a, i64 %idx) ret <vscale x 4 x float> %out } ; NOTE: Identical operation to dupq_lane_i8 (i.e. element type is irrelevant). define <vscale x 2 x double> @dupq_lane_f64(<vscale x 2 x double> %a, i64 %idx) { ; CHECK-LABEL: dupq_lane_f64: ; CHECK: // %bb.0: ; CHECK-NEXT: add x8, x0, x0 ; CHECK-NEXT: index z1.d, #0, #1 ; CHECK-NEXT: and z1.d, z1.d, #0x1 ; CHECK-NEXT: mov z2.d, x8 ; CHECK-NEXT: add z1.d, z1.d, z2.d ; CHECK-NEXT: tbl z0.d, { z0.d }, z1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x double> @llvm.aarch64.sve.dupq.lane.nxv2f64(<vscale x 2 x double> %a, i64 %idx) ret <vscale x 2 x double> %out } ; NOTE: Index out of range (0-3) define <vscale x 2 x i64> @dupq_i64_range(<vscale x 2 x i64> %a) { ; CHECK-LABEL: dupq_i64_range: ; CHECK: // %bb.0: ; CHECK-NEXT: index z1.d, #0, #1 ; CHECK-NEXT: and z1.d, z1.d, #0x1 ; CHECK-NEXT: add z1.d, z1.d, #8 // =0x8 ; CHECK-NEXT: tbl z0.d, { z0.d }, z1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x i64> @llvm.aarch64.sve.dupq.lane.nxv2i64(<vscale x 2 x i64> %a, i64 4) ret <vscale x 2 x i64> %out } ; ; EXT ; define <vscale x 16 x i8> @ext_i8(<vscale x 16 x i8> %a, <vscale x 16 x i8> %b) { ; CHECK-LABEL: ext_i8: ; CHECK: // %bb.0: ; CHECK-NEXT: ext z0.b, z0.b, z1.b, #255 ; CHECK-NEXT: ret %out = call <vscale x 16 x i8> @llvm.aarch64.sve.ext.nxv16i8(<vscale x 16 x i8> %a, <vscale x 16 x i8> %b, i32 255) ret <vscale x 16 x i8> %out } define <vscale x 8 x i16> @ext_i16(<vscale x 8 x i16> %a, <vscale x 8 x i16> %b) { ; CHECK-LABEL: ext_i16: ; CHECK: // %bb.0: ; CHECK-NEXT: ext z0.b, z0.b, z1.b, #0 ; CHECK-NEXT: ret %out = call <vscale x 8 x i16> @llvm.aarch64.sve.ext.nxv8i16(<vscale x 8 x i16> %a, <vscale x 8 x i16> %b, i32 0) ret <vscale x 8 x i16> %out } define <vscale x 4 x i32> @ext_i32(<vscale x 4 x i32> %a, <vscale x 4 x i32> %b) { ; CHECK-LABEL: ext_i32: ; CHECK: // %bb.0: ; CHECK-NEXT: ext z0.b, z0.b, z1.b, #4 ; CHECK-NEXT: ret %out = call <vscale x 4 x i32> @llvm.aarch64.sve.ext.nxv4i32(<vscale x 4 x i32> %a, <vscale x 4 x i32> %b, i32 1) ret <vscale x 4 x i32> %out } define <vscale x 2 x i64> @ext_i64(<vscale x 2 x i64> %a, <vscale x 2 x i64> %b) { ; CHECK-LABEL: ext_i64: ; CHECK: // %bb.0: ; CHECK-NEXT: ext z0.b, z0.b, z1.b, #16 ; CHECK-NEXT: ret %out = call <vscale x 2 x i64> @llvm.aarch64.sve.ext.nxv2i64(<vscale x 2 x i64> %a, <vscale x 2 x i64> %b, i32 2) ret <vscale x 2 x i64> %out } define <vscale x 8 x bfloat> @ext_bf16(<vscale x 8 x bfloat> %a, <vscale x 8 x bfloat> %b) #0 { ; CHECK-LABEL: ext_bf16: ; CHECK: // %bb.0: ; CHECK-NEXT: ext z0.b, z0.b, z1.b, #6 ; CHECK-NEXT: ret %out = call <vscale x 8 x bfloat> @llvm.aarch64.sve.ext.nxv8bf16(<vscale x 8 x bfloat> %a, <vscale x 8 x bfloat> %b, i32 3) ret <vscale x 8 x bfloat> %out } define <vscale x 8 x half> @ext_f16(<vscale x 8 x half> %a, <vscale x 8 x half> %b) { ; CHECK-LABEL: ext_f16: ; CHECK: // %bb.0: ; CHECK-NEXT: ext z0.b, z0.b, z1.b, #6 ; CHECK-NEXT: ret %out = call <vscale x 8 x half> @llvm.aarch64.sve.ext.nxv8f16(<vscale x 8 x half> %a, <vscale x 8 x half> %b, i32 3) ret <vscale x 8 x half> %out } define <vscale x 4 x float> @ext_f32(<vscale x 4 x float> %a, <vscale x 4 x float> %b) { ; CHECK-LABEL: ext_f32: ; CHECK: // %bb.0: ; CHECK-NEXT: ext z0.b, z0.b, z1.b, #16 ; CHECK-NEXT: ret %out = call <vscale x 4 x float> @llvm.aarch64.sve.ext.nxv4f32(<vscale x 4 x float> %a, <vscale x 4 x float> %b, i32 4) ret <vscale x 4 x float> %out } define <vscale x 2 x double> @ext_f64(<vscale x 2 x double> %a, <vscale x 2 x double> %b) { ; CHECK-LABEL: ext_f64: ; CHECK: // %bb.0: ; CHECK-NEXT: ext z0.b, z0.b, z1.b, #40 ; CHECK-NEXT: ret %out = call <vscale x 2 x double> @llvm.aarch64.sve.ext.nxv2f64(<vscale x 2 x double> %a, <vscale x 2 x double> %b, i32 5) ret <vscale x 2 x double> %out } ; ; LASTA ; define i8 @lasta_i8(<vscale x 16 x i1> %pg, <vscale x 16 x i8> %a) { ; CHECK-LABEL: lasta_i8: ; CHECK: // %bb.0: ; CHECK-NEXT: lasta w0, p0, z0.b ; CHECK-NEXT: ret %res = call i8 @llvm.aarch64.sve.lasta.nxv16i8(<vscale x 16 x i1> %pg, <vscale x 16 x i8> %a) ret i8 %res } define i16 @lasta_i16(<vscale x 8 x i1> %pg, <vscale x 8 x i16> %a) { ; CHECK-LABEL: lasta_i16: ; CHECK: // %bb.0: ; CHECK-NEXT: lasta w0, p0, z0.h ; CHECK-NEXT: ret %res = call i16 @llvm.aarch64.sve.lasta.nxv8i16(<vscale x 8 x i1> %pg, <vscale x 8 x i16> %a) ret i16 %res } define i32 @lasta_i32(<vscale x 4 x i1> %pg, <vscale x 4 x i32> %a) { ; CHECK-LABEL: lasta_i32: ; CHECK: // %bb.0: ; CHECK-NEXT: lasta w0, p0, z0.s ; CHECK-NEXT: ret %res = call i32 @llvm.aarch64.sve.lasta.nxv4i32(<vscale x 4 x i1> %pg, <vscale x 4 x i32> %a) ret i32 %res } define i64 @lasta_i64(<vscale x 2 x i1> %pg, <vscale x 2 x i64> %a) { ; CHECK-LABEL: lasta_i64: ; CHECK: // %bb.0: ; CHECK-NEXT: lasta x0, p0, z0.d ; CHECK-NEXT: ret %res = call i64 @llvm.aarch64.sve.lasta.nxv2i64(<vscale x 2 x i1> %pg, <vscale x 2 x i64> %a) ret i64 %res } define half @lasta_f16(<vscale x 8 x i1> %pg, <vscale x 8 x half> %a) { ; CHECK-LABEL: lasta_f16: ; CHECK: // %bb.0: ; CHECK-NEXT: lasta h0, p0, z0.h ; CHECK-NEXT: ret %res = call half @llvm.aarch64.sve.lasta.nxv8f16(<vscale x 8 x i1> %pg, <vscale x 8 x half> %a) ret half %res } define bfloat @lasta_bf16(<vscale x 8 x i1> %pg, <vscale x 8 x bfloat> %a) #0 { ; CHECK-LABEL: lasta_bf16: ; CHECK: // %bb.0: ; CHECK-NEXT: lasta h0, p0, z0.h ; CHECK-NEXT: ret %res = call bfloat @llvm.aarch64.sve.lasta.nxv8bf16(<vscale x 8 x i1> %pg, <vscale x 8 x bfloat> %a) ret bfloat %res } define float @lasta_f32(<vscale x 4 x i1> %pg, <vscale x 4 x float> %a) { ; CHECK-LABEL: lasta_f32: ; CHECK: // %bb.0: ; CHECK-NEXT: lasta s0, p0, z0.s ; CHECK-NEXT: ret %res = call float @llvm.aarch64.sve.lasta.nxv4f32(<vscale x 4 x i1> %pg, <vscale x 4 x float> %a) ret float %res } define float @lasta_f32_v2(<vscale x 2 x i1> %pg, <vscale x 2 x float> %a) { ; CHECK-LABEL: lasta_f32_v2: ; CHECK: // %bb.0: ; CHECK-NEXT: lasta s0, p0, z0.s ; CHECK-NEXT: ret %res = call float @llvm.aarch64.sve.lasta.nxv2f32(<vscale x 2 x i1> %pg, <vscale x 2 x float> %a) ret float %res } define double @lasta_f64(<vscale x 2 x i1> %pg, <vscale x 2 x double> %a) { ; CHECK-LABEL: lasta_f64: ; CHECK: // %bb.0: ; CHECK-NEXT: lasta d0, p0, z0.d ; CHECK-NEXT: ret %res = call double @llvm.aarch64.sve.lasta.nxv2f64(<vscale x 2 x i1> %pg, <vscale x 2 x double> %a) ret double %res } ; ; LASTB ; define i8 @lastb_i8(<vscale x 16 x i1> %pg, <vscale x 16 x i8> %a) { ; CHECK-LABEL: lastb_i8: ; CHECK: // %bb.0: ; CHECK-NEXT: lastb w0, p0, z0.b ; CHECK-NEXT: ret %res = call i8 @llvm.aarch64.sve.lastb.nxv16i8(<vscale x 16 x i1> %pg, <vscale x 16 x i8> %a) ret i8 %res } define i16 @lastb_i16(<vscale x 8 x i1> %pg, <vscale x 8 x i16> %a) { ; CHECK-LABEL: lastb_i16: ; CHECK: // %bb.0: ; CHECK-NEXT: lastb w0, p0, z0.h ; CHECK-NEXT: ret %res = call i16 @llvm.aarch64.sve.lastb.nxv8i16(<vscale x 8 x i1> %pg, <vscale x 8 x i16> %a) ret i16 %res } define i32 @lastb_i32(<vscale x 4 x i1> %pg, <vscale x 4 x i32> %a) { ; CHECK-LABEL: lastb_i32: ; CHECK: // %bb.0: ; CHECK-NEXT: lastb w0, p0, z0.s ; CHECK-NEXT: ret %res = call i32 @llvm.aarch64.sve.lastb.nxv4i32(<vscale x 4 x i1> %pg, <vscale x 4 x i32> %a) ret i32 %res } define i64 @lastb_i64(<vscale x 2 x i1> %pg, <vscale x 2 x i64> %a) { ; CHECK-LABEL: lastb_i64: ; CHECK: // %bb.0: ; CHECK-NEXT: lastb x0, p0, z0.d ; CHECK-NEXT: ret %res = call i64 @llvm.aarch64.sve.lastb.nxv2i64(<vscale x 2 x i1> %pg, <vscale x 2 x i64> %a) ret i64 %res } define half @lastb_f16(<vscale x 8 x i1> %pg, <vscale x 8 x half> %a) { ; CHECK-LABEL: lastb_f16: ; CHECK: // %bb.0: ; CHECK-NEXT: lastb h0, p0, z0.h ; CHECK-NEXT: ret %res = call half @llvm.aarch64.sve.lastb.nxv8f16(<vscale x 8 x i1> %pg, <vscale x 8 x half> %a) ret half %res } define bfloat @lastb_bf16(<vscale x 8 x i1> %pg, <vscale x 8 x bfloat> %a) #0 { ; CHECK-LABEL: lastb_bf16: ; CHECK: // %bb.0: ; CHECK-NEXT: lastb h0, p0, z0.h ; CHECK-NEXT: ret %res = call bfloat @llvm.aarch64.sve.lastb.nxv8bf16(<vscale x 8 x i1> %pg, <vscale x 8 x bfloat> %a) ret bfloat %res } define float @lastb_f32(<vscale x 4 x i1> %pg, <vscale x 4 x float> %a) { ; CHECK-LABEL: lastb_f32: ; CHECK: // %bb.0: ; CHECK-NEXT: lastb s0, p0, z0.s ; CHECK-NEXT: ret %res = call float @llvm.aarch64.sve.lastb.nxv4f32(<vscale x 4 x i1> %pg, <vscale x 4 x float> %a) ret float %res } define float @lastb_f32_v2(<vscale x 2 x i1> %pg, <vscale x 2 x float> %a) { ; CHECK-LABEL: lastb_f32_v2: ; CHECK: // %bb.0: ; CHECK-NEXT: lastb s0, p0, z0.s ; CHECK-NEXT: ret %res = call float @llvm.aarch64.sve.lastb.nxv2f32(<vscale x 2 x i1> %pg, <vscale x 2 x float> %a) ret float %res } define double @lastb_f64(<vscale x 2 x i1> %pg, <vscale x 2 x double> %a) { ; CHECK-LABEL: lastb_f64: ; CHECK: // %bb.0: ; CHECK-NEXT: lastb d0, p0, z0.d ; CHECK-NEXT: ret %res = call double @llvm.aarch64.sve.lastb.nxv2f64(<vscale x 2 x i1> %pg, <vscale x 2 x double> %a) ret double %res } ; ; COMPACT ; define <vscale x 4 x i32> @compact_i32(<vscale x 4 x i1> %pg, <vscale x 4 x i32> %a) { ; CHECK-LABEL: compact_i32: ; CHECK: // %bb.0: ; CHECK-NEXT: compact z0.s, p0, z0.s ; CHECK-NEXT: ret %out = call <vscale x 4 x i32> @llvm.aarch64.sve.compact.nxv4i32(<vscale x 4 x i1> %pg, <vscale x 4 x i32> %a) ret <vscale x 4 x i32> %out } define <vscale x 2 x i64> @compact_i64(<vscale x 2 x i1> %pg, <vscale x 2 x i64> %a) { ; CHECK-LABEL: compact_i64: ; CHECK: // %bb.0: ; CHECK-NEXT: compact z0.d, p0, z0.d ; CHECK-NEXT: ret %out = call <vscale x 2 x i64> @llvm.aarch64.sve.compact.nxv2i64(<vscale x 2 x i1> %pg, <vscale x 2 x i64> %a) ret <vscale x 2 x i64> %out } define <vscale x 4 x float> @compact_f32(<vscale x 4 x i1> %pg, <vscale x 4 x float> %a) { ; CHECK-LABEL: compact_f32: ; CHECK: // %bb.0: ; CHECK-NEXT: compact z0.s, p0, z0.s ; CHECK-NEXT: ret %out = call <vscale x 4 x float> @llvm.aarch64.sve.compact.nxv4f32(<vscale x 4 x i1> %pg, <vscale x 4 x float> %a) ret <vscale x 4 x float> %out } define <vscale x 2 x double> @compact_f64(<vscale x 2 x i1> %pg, <vscale x 2 x double> %a) { ; CHECK-LABEL: compact_f64: ; CHECK: // %bb.0: ; CHECK-NEXT: compact z0.d, p0, z0.d ; CHECK-NEXT: ret %out = call <vscale x 2 x double> @llvm.aarch64.sve.compact.nxv2f64(<vscale x 2 x i1> %pg, <vscale x 2 x double> %a) ret <vscale x 2 x double> %out } ; ; REV ; define <vscale x 16 x i1> @rev_b8( <vscale x 16 x i1> %a) { ; CHECK-LABEL: rev_b8: ; CHECK: // %bb.0: ; CHECK-NEXT: rev p0.b, p0.b ; CHECK-NEXT: ret %res = call <vscale x 16 x i1> @llvm.aarch64.sve.rev.nxv16i1(<vscale x 16 x i1> %a) ret <vscale x 16 x i1> %res } define <vscale x 8 x i1> @rev_b16(<vscale x 8 x i1> %a) { ; CHECK-LABEL: rev_b16: ; CHECK: // %bb.0: ; CHECK-NEXT: rev p0.h, p0.h ; CHECK-NEXT: ret %res = call <vscale x 8 x i1> @llvm.aarch64.sve.rev.nxv8i1(<vscale x 8 x i1> %a) ret <vscale x 8 x i1> %res } define <vscale x 4 x i1> @rev_b32(<vscale x 4 x i1> %a) { ; CHECK-LABEL: rev_b32: ; CHECK: // %bb.0: ; CHECK-NEXT: rev p0.s, p0.s ; CHECK-NEXT: ret %res = call <vscale x 4 x i1> @llvm.aarch64.sve.rev.nxv4i1(<vscale x 4 x i1> %a) ret <vscale x 4 x i1> %res } define <vscale x 2 x i1> @rev_b64(<vscale x 2 x i1> %a) { ; CHECK-LABEL: rev_b64: ; CHECK: // %bb.0: ; CHECK-NEXT: rev p0.d, p0.d ; CHECK-NEXT: ret %res = call <vscale x 2 x i1> @llvm.aarch64.sve.rev.nxv2i1(<vscale x 2 x i1> %a) ret <vscale x 2 x i1> %res } define <vscale x 16 x i8> @rev_i8( <vscale x 16 x i8> %a) { ; CHECK-LABEL: rev_i8: ; CHECK: // %bb.0: ; CHECK-NEXT: rev z0.b, z0.b ; CHECK-NEXT: ret %res = call <vscale x 16 x i8> @llvm.aarch64.sve.rev.nxv16i8(<vscale x 16 x i8> %a) ret <vscale x 16 x i8> %res } define <vscale x 8 x i16> @rev_i16(<vscale x 8 x i16> %a) { ; CHECK-LABEL: rev_i16: ; CHECK: // %bb.0: ; CHECK-NEXT: rev z0.h, z0.h ; CHECK-NEXT: ret %res = call <vscale x 8 x i16> @llvm.aarch64.sve.rev.nxv8i16(<vscale x 8 x i16> %a) ret <vscale x 8 x i16> %res } define <vscale x 4 x i32> @rev_i32(<vscale x 4 x i32> %a) { ; CHECK-LABEL: rev_i32: ; CHECK: // %bb.0: ; CHECK-NEXT: rev z0.s, z0.s ; CHECK-NEXT: ret %res = call <vscale x 4 x i32> @llvm.aarch64.sve.rev.nxv4i32(<vscale x 4 x i32> %a) ret <vscale x 4 x i32> %res } define <vscale x 2 x i64> @rev_i64(<vscale x 2 x i64> %a) { ; CHECK-LABEL: rev_i64: ; CHECK: // %bb.0: ; CHECK-NEXT: rev z0.d, z0.d ; CHECK-NEXT: ret %res = call <vscale x 2 x i64> @llvm.aarch64.sve.rev.nxv2i64(<vscale x 2 x i64> %a) ret <vscale x 2 x i64> %res } define <vscale x 8 x bfloat> @rev_bf16(<vscale x 8 x bfloat> %a) #0 { ; CHECK-LABEL: rev_bf16: ; CHECK: // %bb.0: ; CHECK-NEXT: rev z0.h, z0.h ; CHECK-NEXT: ret %res = call <vscale x 8 x bfloat> @llvm.aarch64.sve.rev.nxv8bf16(<vscale x 8 x bfloat> %a) ret <vscale x 8 x bfloat> %res } define <vscale x 8 x half> @rev_f16(<vscale x 8 x half> %a) { ; CHECK-LABEL: rev_f16: ; CHECK: // %bb.0: ; CHECK-NEXT: rev z0.h, z0.h ; CHECK-NEXT: ret %res = call <vscale x 8 x half> @llvm.aarch64.sve.rev.nxv8f16(<vscale x 8 x half> %a) ret <vscale x 8 x half> %res } define <vscale x 4 x float> @rev_f32(<vscale x 4 x float> %a) { ; CHECK-LABEL: rev_f32: ; CHECK: // %bb.0: ; CHECK-NEXT: rev z0.s, z0.s ; CHECK-NEXT: ret %res = call <vscale x 4 x float> @llvm.aarch64.sve.rev.nxv4f32(<vscale x 4 x float> %a) ret <vscale x 4 x float> %res } define <vscale x 2 x double> @rev_f64(<vscale x 2 x double> %a) { ; CHECK-LABEL: rev_f64: ; CHECK: // %bb.0: ; CHECK-NEXT: rev z0.d, z0.d ; CHECK-NEXT: ret %res = call <vscale x 2 x double> @llvm.aarch64.sve.rev.nxv2f64(<vscale x 2 x double> %a) ret <vscale x 2 x double> %res } ; ; SPLICE ; define <vscale x 16 x i8> @splice_i8(<vscale x 16 x i1> %pg, <vscale x 16 x i8> %a, <vscale x 16 x i8> %b) { ; CHECK-LABEL: splice_i8: ; CHECK: // %bb.0: ; CHECK-NEXT: splice z0.b, p0, z0.b, z1.b ; CHECK-NEXT: ret %out = call <vscale x 16 x i8> @llvm.aarch64.sve.splice.nxv16i8(<vscale x 16 x i1> %pg, <vscale x 16 x i8> %a, <vscale x 16 x i8> %b) ret <vscale x 16 x i8> %out } define <vscale x 8 x i16> @splice_i16(<vscale x 8 x i1> %pg, <vscale x 8 x i16> %a, <vscale x 8 x i16> %b) { ; CHECK-LABEL: splice_i16: ; CHECK: // %bb.0: ; CHECK-NEXT: splice z0.h, p0, z0.h, z1.h ; CHECK-NEXT: ret %out = call <vscale x 8 x i16> @llvm.aarch64.sve.splice.nxv8i16(<vscale x 8 x i1> %pg, <vscale x 8 x i16> %a, <vscale x 8 x i16> %b) ret <vscale x 8 x i16> %out } define <vscale x 4 x i32> @splice_i32(<vscale x 4 x i1> %pg, <vscale x 4 x i32> %a, <vscale x 4 x i32> %b) { ; CHECK-LABEL: splice_i32: ; CHECK: // %bb.0: ; CHECK-NEXT: splice z0.s, p0, z0.s, z1.s ; CHECK-NEXT: ret %out = call <vscale x 4 x i32> @llvm.aarch64.sve.splice.nxv4i32(<vscale x 4 x i1> %pg, <vscale x 4 x i32> %a, <vscale x 4 x i32> %b) ret <vscale x 4 x i32> %out } define <vscale x 2 x i64> @splice_i64(<vscale x 2 x i1> %pg, <vscale x 2 x i64> %a, <vscale x 2 x i64> %b) { ; CHECK-LABEL: splice_i64: ; CHECK: // %bb.0: ; CHECK-NEXT: splice z0.d, p0, z0.d, z1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x i64> @llvm.aarch64.sve.splice.nxv2i64(<vscale x 2 x i1> %pg, <vscale x 2 x i64> %a, <vscale x 2 x i64> %b) ret <vscale x 2 x i64> %out } define <vscale x 8 x bfloat> @splice_bf16(<vscale x 8 x i1> %pg, <vscale x 8 x bfloat> %a, <vscale x 8 x bfloat> %b) #0 { ; CHECK-LABEL: splice_bf16: ; CHECK: // %bb.0: ; CHECK-NEXT: splice z0.h, p0, z0.h, z1.h ; CHECK-NEXT: ret %out = call <vscale x 8 x bfloat> @llvm.aarch64.sve.splice.nxv8bf16(<vscale x 8 x i1> %pg, <vscale x 8 x bfloat> %a, <vscale x 8 x bfloat> %b) ret <vscale x 8 x bfloat> %out } define <vscale x 8 x half> @splice_f16(<vscale x 8 x i1> %pg, <vscale x 8 x half> %a, <vscale x 8 x half> %b) { ; CHECK-LABEL: splice_f16: ; CHECK: // %bb.0: ; CHECK-NEXT: splice z0.h, p0, z0.h, z1.h ; CHECK-NEXT: ret %out = call <vscale x 8 x half> @llvm.aarch64.sve.splice.nxv8f16(<vscale x 8 x i1> %pg, <vscale x 8 x half> %a, <vscale x 8 x half> %b) ret <vscale x 8 x half> %out } define <vscale x 4 x float> @splice_f32(<vscale x 4 x i1> %pg, <vscale x 4 x float> %a, <vscale x 4 x float> %b) { ; CHECK-LABEL: splice_f32: ; CHECK: // %bb.0: ; CHECK-NEXT: splice z0.s, p0, z0.s, z1.s ; CHECK-NEXT: ret %out = call <vscale x 4 x float> @llvm.aarch64.sve.splice.nxv4f32(<vscale x 4 x i1> %pg, <vscale x 4 x float> %a, <vscale x 4 x float> %b) ret <vscale x 4 x float> %out } define <vscale x 2 x double> @splice_f64(<vscale x 2 x i1> %pg, <vscale x 2 x double> %a, <vscale x 2 x double> %b) { ; CHECK-LABEL: splice_f64: ; CHECK: // %bb.0: ; CHECK-NEXT: splice z0.d, p0, z0.d, z1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x double> @llvm.aarch64.sve.splice.nxv2f64(<vscale x 2 x i1> %pg, <vscale x 2 x double> %a, <vscale x 2 x double> %b) ret <vscale x 2 x double> %out } ; ; SUNPKHI ; define <vscale x 8 x i16> @sunpkhi_i16(<vscale x 16 x i8> %a) { ; CHECK-LABEL: sunpkhi_i16: ; CHECK: // %bb.0: ; CHECK-NEXT: sunpkhi z0.h, z0.b ; CHECK-NEXT: ret %res = call <vscale x 8 x i16> @llvm.aarch64.sve.sunpkhi.nxv8i16(<vscale x 16 x i8> %a) ret <vscale x 8 x i16> %res } define <vscale x 4 x i32> @sunpkhi_i32(<vscale x 8 x i16> %a) { ; CHECK-LABEL: sunpkhi_i32: ; CHECK: // %bb.0: ; CHECK-NEXT: sunpkhi z0.s, z0.h ; CHECK-NEXT: ret %res = call <vscale x 4 x i32> @llvm.aarch64.sve.sunpkhi.nxv4i32(<vscale x 8 x i16> %a) ret <vscale x 4 x i32> %res } define <vscale x 2 x i64> @sunpkhi_i64(<vscale x 4 x i32> %a) { ; CHECK-LABEL: sunpkhi_i64: ; CHECK: // %bb.0: ; CHECK-NEXT: sunpkhi z0.d, z0.s ; CHECK-NEXT: ret %res = call <vscale x 2 x i64> @llvm.aarch64.sve.sunpkhi.nxv2i64(<vscale x 4 x i32> %a) ret <vscale x 2 x i64> %res } ; ; SUNPKLO ; define <vscale x 8 x i16> @sunpklo_i16(<vscale x 16 x i8> %a) { ; CHECK-LABEL: sunpklo_i16: ; CHECK: // %bb.0: ; CHECK-NEXT: sunpklo z0.h, z0.b ; CHECK-NEXT: ret %res = call <vscale x 8 x i16> @llvm.aarch64.sve.sunpklo.nxv8i16(<vscale x 16 x i8> %a) ret <vscale x 8 x i16> %res } define <vscale x 4 x i32> @sunpklo_i32(<vscale x 8 x i16> %a) { ; CHECK-LABEL: sunpklo_i32: ; CHECK: // %bb.0: ; CHECK-NEXT: sunpklo z0.s, z0.h ; CHECK-NEXT: ret %res = call <vscale x 4 x i32> @llvm.aarch64.sve.sunpklo.nxv4i32(<vscale x 8 x i16> %a) ret <vscale x 4 x i32> %res } define <vscale x 2 x i64> @sunpklo_i64(<vscale x 4 x i32> %a) { ; CHECK-LABEL: sunpklo_i64: ; CHECK: // %bb.0: ; CHECK-NEXT: sunpklo z0.d, z0.s ; CHECK-NEXT: ret %res = call <vscale x 2 x i64> @llvm.aarch64.sve.sunpklo.nxv2i64(<vscale x 4 x i32> %a) ret <vscale x 2 x i64> %res } ; ; TBL ; define <vscale x 16 x i8> @tbl_i8(<vscale x 16 x i8> %a, <vscale x 16 x i8> %b) { ; CHECK-LABEL: tbl_i8: ; CHECK: // %bb.0: ; CHECK-NEXT: tbl z0.b, { z0.b }, z1.b ; CHECK-NEXT: ret %out = call <vscale x 16 x i8> @llvm.aarch64.sve.tbl.nxv16i8(<vscale x 16 x i8> %a, <vscale x 16 x i8> %b) ret <vscale x 16 x i8> %out } define <vscale x 8 x i16> @tbl_i16(<vscale x 8 x i16> %a, <vscale x 8 x i16> %b) { ; CHECK-LABEL: tbl_i16: ; CHECK: // %bb.0: ; CHECK-NEXT: tbl z0.h, { z0.h }, z1.h ; CHECK-NEXT: ret %out = call <vscale x 8 x i16> @llvm.aarch64.sve.tbl.nxv8i16(<vscale x 8 x i16> %a, <vscale x 8 x i16> %b) ret <vscale x 8 x i16> %out } define <vscale x 4 x i32> @tbl_i32(<vscale x 4 x i32> %a, <vscale x 4 x i32> %b) { ; CHECK-LABEL: tbl_i32: ; CHECK: // %bb.0: ; CHECK-NEXT: tbl z0.s, { z0.s }, z1.s ; CHECK-NEXT: ret %out = call <vscale x 4 x i32> @llvm.aarch64.sve.tbl.nxv4i32(<vscale x 4 x i32> %a, <vscale x 4 x i32> %b) ret <vscale x 4 x i32> %out } define <vscale x 2 x i64> @tbl_i64(<vscale x 2 x i64> %a, <vscale x 2 x i64> %b) { ; CHECK-LABEL: tbl_i64: ; CHECK: // %bb.0: ; CHECK-NEXT: tbl z0.d, { z0.d }, z1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x i64> @llvm.aarch64.sve.tbl.nxv2i64(<vscale x 2 x i64> %a, <vscale x 2 x i64> %b) ret <vscale x 2 x i64> %out } define <vscale x 8 x half> @tbl_f16(<vscale x 8 x half> %a, <vscale x 8 x i16> %b) { ; CHECK-LABEL: tbl_f16: ; CHECK: // %bb.0: ; CHECK-NEXT: tbl z0.h, { z0.h }, z1.h ; CHECK-NEXT: ret %out = call <vscale x 8 x half> @llvm.aarch64.sve.tbl.nxv8f16(<vscale x 8 x half> %a, <vscale x 8 x i16> %b) ret <vscale x 8 x half> %out } define <vscale x 8 x bfloat> @tbl_bf16(<vscale x 8 x bfloat> %a, <vscale x 8 x i16> %b) #0 { ; CHECK-LABEL: tbl_bf16: ; CHECK: // %bb.0: ; CHECK-NEXT: tbl z0.h, { z0.h }, z1.h ; CHECK-NEXT: ret %out = call <vscale x 8 x bfloat> @llvm.aarch64.sve.tbl.nxv8bf16(<vscale x 8 x bfloat> %a, <vscale x 8 x i16> %b) ret <vscale x 8 x bfloat> %out } define <vscale x 4 x float> @tbl_f32(<vscale x 4 x float> %a, <vscale x 4 x i32> %b) { ; CHECK-LABEL: tbl_f32: ; CHECK: // %bb.0: ; CHECK-NEXT: tbl z0.s, { z0.s }, z1.s ; CHECK-NEXT: ret %out = call <vscale x 4 x float> @llvm.aarch64.sve.tbl.nxv4f32(<vscale x 4 x float> %a, <vscale x 4 x i32> %b) ret <vscale x 4 x float> %out } define <vscale x 2 x double> @tbl_f64(<vscale x 2 x double> %a, <vscale x 2 x i64> %b) { ; CHECK-LABEL: tbl_f64: ; CHECK: // %bb.0: ; CHECK-NEXT: tbl z0.d, { z0.d }, z1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x double> @llvm.aarch64.sve.tbl.nxv2f64(<vscale x 2 x double> %a, <vscale x 2 x i64> %b) ret <vscale x 2 x double> %out } ; ; UUNPKHI ; define <vscale x 8 x i16> @uunpkhi_i16(<vscale x 16 x i8> %a) { ; CHECK-LABEL: uunpkhi_i16: ; CHECK: // %bb.0: ; CHECK-NEXT: uunpkhi z0.h, z0.b ; CHECK-NEXT: ret %res = call <vscale x 8 x i16> @llvm.aarch64.sve.uunpkhi.nxv8i16(<vscale x 16 x i8> %a) ret <vscale x 8 x i16> %res } define <vscale x 4 x i32> @uunpkhi_i32(<vscale x 8 x i16> %a) { ; CHECK-LABEL: uunpkhi_i32: ; CHECK: // %bb.0: ; CHECK-NEXT: uunpkhi z0.s, z0.h ; CHECK-NEXT: ret %res = call <vscale x 4 x i32> @llvm.aarch64.sve.uunpkhi.nxv4i32(<vscale x 8 x i16> %a) ret <vscale x 4 x i32> %res } define <vscale x 2 x i64> @uunpkhi_i64(<vscale x 4 x i32> %a) { ; CHECK-LABEL: uunpkhi_i64: ; CHECK: // %bb.0: ; CHECK-NEXT: uunpkhi z0.d, z0.s ; CHECK-NEXT: ret %res = call <vscale x 2 x i64> @llvm.aarch64.sve.uunpkhi.nxv2i64(<vscale x 4 x i32> %a) ret <vscale x 2 x i64> %res } ; ; UUNPKLO ; define <vscale x 8 x i16> @uunpklo_i16(<vscale x 16 x i8> %a) { ; CHECK-LABEL: uunpklo_i16: ; CHECK: // %bb.0: ; CHECK-NEXT: uunpklo z0.h, z0.b ; CHECK-NEXT: ret %res = call <vscale x 8 x i16> @llvm.aarch64.sve.uunpklo.nxv8i16(<vscale x 16 x i8> %a) ret <vscale x 8 x i16> %res } define <vscale x 4 x i32> @uunpklo_i32(<vscale x 8 x i16> %a) { ; CHECK-LABEL: uunpklo_i32: ; CHECK: // %bb.0: ; CHECK-NEXT: uunpklo z0.s, z0.h ; CHECK-NEXT: ret %res = call <vscale x 4 x i32> @llvm.aarch64.sve.uunpklo.nxv4i32(<vscale x 8 x i16> %a) ret <vscale x 4 x i32> %res } define <vscale x 2 x i64> @uunpklo_i64(<vscale x 4 x i32> %a) { ; CHECK-LABEL: uunpklo_i64: ; CHECK: // %bb.0: ; CHECK-NEXT: uunpklo z0.d, z0.s ; CHECK-NEXT: ret %res = call <vscale x 2 x i64> @llvm.aarch64.sve.uunpklo.nxv2i64(<vscale x 4 x i32> %a) ret <vscale x 2 x i64> %res } ; ; TRN1 ; define <vscale x 16 x i1> @trn1_b8(<vscale x 16 x i1> %a, <vscale x 16 x i1> %b) { ; CHECK-LABEL: trn1_b8: ; CHECK: // %bb.0: ; CHECK-NEXT: trn1 p0.b, p0.b, p1.b ; CHECK-NEXT: ret %out = call <vscale x 16 x i1> @llvm.aarch64.sve.trn1.nxv16i1(<vscale x 16 x i1> %a, <vscale x 16 x i1> %b) ret <vscale x 16 x i1> %out } define <vscale x 8 x i1> @trn1_b16(<vscale x 8 x i1> %a, <vscale x 8 x i1> %b) { ; CHECK-LABEL: trn1_b16: ; CHECK: // %bb.0: ; CHECK-NEXT: trn1 p0.h, p0.h, p1.h ; CHECK-NEXT: ret %out = call <vscale x 8 x i1> @llvm.aarch64.sve.trn1.nxv8i1(<vscale x 8 x i1> %a, <vscale x 8 x i1> %b) ret <vscale x 8 x i1> %out } define <vscale x 4 x i1> @trn1_b32(<vscale x 4 x i1> %a, <vscale x 4 x i1> %b) { ; CHECK-LABEL: trn1_b32: ; CHECK: // %bb.0: ; CHECK-NEXT: trn1 p0.s, p0.s, p1.s ; CHECK-NEXT: ret %out = call <vscale x 4 x i1> @llvm.aarch64.sve.trn1.nxv4i1(<vscale x 4 x i1> %a, <vscale x 4 x i1> %b) ret <vscale x 4 x i1> %out } define <vscale x 2 x i1> @trn1_b64(<vscale x 2 x i1> %a, <vscale x 2 x i1> %b) { ; CHECK-LABEL: trn1_b64: ; CHECK: // %bb.0: ; CHECK-NEXT: trn1 p0.d, p0.d, p1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x i1> @llvm.aarch64.sve.trn1.nxv2i1(<vscale x 2 x i1> %a, <vscale x 2 x i1> %b) ret <vscale x 2 x i1> %out } define <vscale x 16 x i8> @trn1_i8(<vscale x 16 x i8> %a, <vscale x 16 x i8> %b) { ; CHECK-LABEL: trn1_i8: ; CHECK: // %bb.0: ; CHECK-NEXT: trn1 z0.b, z0.b, z1.b ; CHECK-NEXT: ret %out = call <vscale x 16 x i8> @llvm.aarch64.sve.trn1.nxv16i8(<vscale x 16 x i8> %a, <vscale x 16 x i8> %b) ret <vscale x 16 x i8> %out } define <vscale x 8 x i16> @trn1_i16(<vscale x 8 x i16> %a, <vscale x 8 x i16> %b) { ; CHECK-LABEL: trn1_i16: ; CHECK: // %bb.0: ; CHECK-NEXT: trn1 z0.h, z0.h, z1.h ; CHECK-NEXT: ret %out = call <vscale x 8 x i16> @llvm.aarch64.sve.trn1.nxv8i16(<vscale x 8 x i16> %a, <vscale x 8 x i16> %b) ret <vscale x 8 x i16> %out } define <vscale x 4 x i32> @trn1_i32(<vscale x 4 x i32> %a, <vscale x 4 x i32> %b) { ; CHECK-LABEL: trn1_i32: ; CHECK: // %bb.0: ; CHECK-NEXT: trn1 z0.s, z0.s, z1.s ; CHECK-NEXT: ret %out = call <vscale x 4 x i32> @llvm.aarch64.sve.trn1.nxv4i32(<vscale x 4 x i32> %a, <vscale x 4 x i32> %b) ret <vscale x 4 x i32> %out } define <vscale x 2 x i64> @trn1_i64(<vscale x 2 x i64> %a, <vscale x 2 x i64> %b) { ; CHECK-LABEL: trn1_i64: ; CHECK: // %bb.0: ; CHECK-NEXT: trn1 z0.d, z0.d, z1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x i64> @llvm.aarch64.sve.trn1.nxv2i64(<vscale x 2 x i64> %a, <vscale x 2 x i64> %b) ret <vscale x 2 x i64> %out } define <vscale x 2 x half> @trn1_f16_v2(<vscale x 2 x half> %a, <vscale x 2 x half> %b) { ; CHECK-LABEL: trn1_f16_v2: ; CHECK: // %bb.0: ; CHECK-NEXT: trn1 z0.d, z0.d, z1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x half> @llvm.aarch64.sve.trn1.nxv2f16(<vscale x 2 x half> %a, <vscale x 2 x half> %b) ret <vscale x 2 x half> %out } define <vscale x 4 x half> @trn1_f16_v4(<vscale x 4 x half> %a, <vscale x 4 x half> %b) { ; CHECK-LABEL: trn1_f16_v4: ; CHECK: // %bb.0: ; CHECK-NEXT: trn1 z0.s, z0.s, z1.s ; CHECK-NEXT: ret %out = call <vscale x 4 x half> @llvm.aarch64.sve.trn1.nxv4f16(<vscale x 4 x half> %a, <vscale x 4 x half> %b) ret <vscale x 4 x half> %out } define <vscale x 8 x bfloat> @trn1_bf16(<vscale x 8 x bfloat> %a, <vscale x 8 x bfloat> %b) #0 { ; CHECK-LABEL: trn1_bf16: ; CHECK: // %bb.0: ; CHECK-NEXT: trn1 z0.h, z0.h, z1.h ; CHECK-NEXT: ret %out = call <vscale x 8 x bfloat> @llvm.aarch64.sve.trn1.nxv8bf16(<vscale x 8 x bfloat> %a, <vscale x 8 x bfloat> %b) ret <vscale x 8 x bfloat> %out } define <vscale x 8 x half> @trn1_f16(<vscale x 8 x half> %a, <vscale x 8 x half> %b) { ; CHECK-LABEL: trn1_f16: ; CHECK: // %bb.0: ; CHECK-NEXT: trn1 z0.h, z0.h, z1.h ; CHECK-NEXT: ret %out = call <vscale x 8 x half> @llvm.aarch64.sve.trn1.nxv8f16(<vscale x 8 x half> %a, <vscale x 8 x half> %b) ret <vscale x 8 x half> %out } define <vscale x 2 x float> @trn1_f32_v2(<vscale x 2 x float> %a, <vscale x 2 x float> %b) { ; CHECK-LABEL: trn1_f32_v2: ; CHECK: // %bb.0: ; CHECK-NEXT: trn1 z0.d, z0.d, z1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x float> @llvm.aarch64.sve.trn1.nxv2f32(<vscale x 2 x float> %a, <vscale x 2 x float> %b) ret <vscale x 2 x float> %out } define <vscale x 4 x float> @trn1_f32(<vscale x 4 x float> %a, <vscale x 4 x float> %b) { ; CHECK-LABEL: trn1_f32: ; CHECK: // %bb.0: ; CHECK-NEXT: trn1 z0.s, z0.s, z1.s ; CHECK-NEXT: ret %out = call <vscale x 4 x float> @llvm.aarch64.sve.trn1.nxv4f32(<vscale x 4 x float> %a, <vscale x 4 x float> %b) ret <vscale x 4 x float> %out } define <vscale x 2 x double> @trn1_f64(<vscale x 2 x double> %a, <vscale x 2 x double> %b) { ; CHECK-LABEL: trn1_f64: ; CHECK: // %bb.0: ; CHECK-NEXT: trn1 z0.d, z0.d, z1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x double> @llvm.aarch64.sve.trn1.nxv2f64(<vscale x 2 x double> %a, <vscale x 2 x double> %b) ret <vscale x 2 x double> %out } ; ; TRN2 ; define <vscale x 16 x i1> @trn2_b8(<vscale x 16 x i1> %a, <vscale x 16 x i1> %b) { ; CHECK-LABEL: trn2_b8: ; CHECK: // %bb.0: ; CHECK-NEXT: trn2 p0.b, p0.b, p1.b ; CHECK-NEXT: ret %out = call <vscale x 16 x i1> @llvm.aarch64.sve.trn2.nxv16i1(<vscale x 16 x i1> %a, <vscale x 16 x i1> %b) ret <vscale x 16 x i1> %out } define <vscale x 8 x i1> @trn2_b16(<vscale x 8 x i1> %a, <vscale x 8 x i1> %b) { ; CHECK-LABEL: trn2_b16: ; CHECK: // %bb.0: ; CHECK-NEXT: trn2 p0.h, p0.h, p1.h ; CHECK-NEXT: ret %out = call <vscale x 8 x i1> @llvm.aarch64.sve.trn2.nxv8i1(<vscale x 8 x i1> %a, <vscale x 8 x i1> %b) ret <vscale x 8 x i1> %out } define <vscale x 4 x i1> @trn2_b32(<vscale x 4 x i1> %a, <vscale x 4 x i1> %b) { ; CHECK-LABEL: trn2_b32: ; CHECK: // %bb.0: ; CHECK-NEXT: trn2 p0.s, p0.s, p1.s ; CHECK-NEXT: ret %out = call <vscale x 4 x i1> @llvm.aarch64.sve.trn2.nxv4i1(<vscale x 4 x i1> %a, <vscale x 4 x i1> %b) ret <vscale x 4 x i1> %out } define <vscale x 2 x i1> @trn2_b64(<vscale x 2 x i1> %a, <vscale x 2 x i1> %b) { ; CHECK-LABEL: trn2_b64: ; CHECK: // %bb.0: ; CHECK-NEXT: trn2 p0.d, p0.d, p1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x i1> @llvm.aarch64.sve.trn2.nxv2i1(<vscale x 2 x i1> %a, <vscale x 2 x i1> %b) ret <vscale x 2 x i1> %out } define <vscale x 16 x i8> @trn2_i8(<vscale x 16 x i8> %a, <vscale x 16 x i8> %b) { ; CHECK-LABEL: trn2_i8: ; CHECK: // %bb.0: ; CHECK-NEXT: trn2 z0.b, z0.b, z1.b ; CHECK-NEXT: ret %out = call <vscale x 16 x i8> @llvm.aarch64.sve.trn2.nxv16i8(<vscale x 16 x i8> %a, <vscale x 16 x i8> %b) ret <vscale x 16 x i8> %out } define <vscale x 8 x i16> @trn2_i16(<vscale x 8 x i16> %a, <vscale x 8 x i16> %b) { ; CHECK-LABEL: trn2_i16: ; CHECK: // %bb.0: ; CHECK-NEXT: trn2 z0.h, z0.h, z1.h ; CHECK-NEXT: ret %out = call <vscale x 8 x i16> @llvm.aarch64.sve.trn2.nxv8i16(<vscale x 8 x i16> %a, <vscale x 8 x i16> %b) ret <vscale x 8 x i16> %out } define <vscale x 4 x i32> @trn2_i32(<vscale x 4 x i32> %a, <vscale x 4 x i32> %b) { ; CHECK-LABEL: trn2_i32: ; CHECK: // %bb.0: ; CHECK-NEXT: trn2 z0.s, z0.s, z1.s ; CHECK-NEXT: ret %out = call <vscale x 4 x i32> @llvm.aarch64.sve.trn2.nxv4i32(<vscale x 4 x i32> %a, <vscale x 4 x i32> %b) ret <vscale x 4 x i32> %out } define <vscale x 2 x i64> @trn2_i64(<vscale x 2 x i64> %a, <vscale x 2 x i64> %b) { ; CHECK-LABEL: trn2_i64: ; CHECK: // %bb.0: ; CHECK-NEXT: trn2 z0.d, z0.d, z1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x i64> @llvm.aarch64.sve.trn2.nxv2i64(<vscale x 2 x i64> %a, <vscale x 2 x i64> %b) ret <vscale x 2 x i64> %out } define <vscale x 2 x half> @trn2_f16_v2(<vscale x 2 x half> %a, <vscale x 2 x half> %b) { ; CHECK-LABEL: trn2_f16_v2: ; CHECK: // %bb.0: ; CHECK-NEXT: trn2 z0.d, z0.d, z1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x half> @llvm.aarch64.sve.trn2.nxv2f16(<vscale x 2 x half> %a, <vscale x 2 x half> %b) ret <vscale x 2 x half> %out } define <vscale x 4 x half> @trn2_f16_v4(<vscale x 4 x half> %a, <vscale x 4 x half> %b) { ; CHECK-LABEL: trn2_f16_v4: ; CHECK: // %bb.0: ; CHECK-NEXT: trn2 z0.s, z0.s, z1.s ; CHECK-NEXT: ret %out = call <vscale x 4 x half> @llvm.aarch64.sve.trn2.nxv4f16(<vscale x 4 x half> %a, <vscale x 4 x half> %b) ret <vscale x 4 x half> %out } define <vscale x 8 x bfloat> @trn2_bf16(<vscale x 8 x bfloat> %a, <vscale x 8 x bfloat> %b) #0 { ; CHECK-LABEL: trn2_bf16: ; CHECK: // %bb.0: ; CHECK-NEXT: trn2 z0.h, z0.h, z1.h ; CHECK-NEXT: ret %out = call <vscale x 8 x bfloat> @llvm.aarch64.sve.trn2.nxv8bf16(<vscale x 8 x bfloat> %a, <vscale x 8 x bfloat> %b) ret <vscale x 8 x bfloat> %out } define <vscale x 8 x half> @trn2_f16(<vscale x 8 x half> %a, <vscale x 8 x half> %b) { ; CHECK-LABEL: trn2_f16: ; CHECK: // %bb.0: ; CHECK-NEXT: trn2 z0.h, z0.h, z1.h ; CHECK-NEXT: ret %out = call <vscale x 8 x half> @llvm.aarch64.sve.trn2.nxv8f16(<vscale x 8 x half> %a, <vscale x 8 x half> %b) ret <vscale x 8 x half> %out } define <vscale x 2 x float> @trn2_f32_v2(<vscale x 2 x float> %a, <vscale x 2 x float> %b) { ; CHECK-LABEL: trn2_f32_v2: ; CHECK: // %bb.0: ; CHECK-NEXT: trn2 z0.d, z0.d, z1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x float> @llvm.aarch64.sve.trn2.nxv2f32(<vscale x 2 x float> %a, <vscale x 2 x float> %b) ret <vscale x 2 x float> %out } define <vscale x 4 x float> @trn2_f32(<vscale x 4 x float> %a, <vscale x 4 x float> %b) { ; CHECK-LABEL: trn2_f32: ; CHECK: // %bb.0: ; CHECK-NEXT: trn2 z0.s, z0.s, z1.s ; CHECK-NEXT: ret %out = call <vscale x 4 x float> @llvm.aarch64.sve.trn2.nxv4f32(<vscale x 4 x float> %a, <vscale x 4 x float> %b) ret <vscale x 4 x float> %out } define <vscale x 2 x double> @trn2_f64(<vscale x 2 x double> %a, <vscale x 2 x double> %b) { ; CHECK-LABEL: trn2_f64: ; CHECK: // %bb.0: ; CHECK-NEXT: trn2 z0.d, z0.d, z1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x double> @llvm.aarch64.sve.trn2.nxv2f64(<vscale x 2 x double> %a, <vscale x 2 x double> %b) ret <vscale x 2 x double> %out } ; ; UZP1 ; define <vscale x 16 x i1> @uzp1_b8(<vscale x 16 x i1> %a, <vscale x 16 x i1> %b) { ; CHECK-LABEL: uzp1_b8: ; CHECK: // %bb.0: ; CHECK-NEXT: uzp1 p0.b, p0.b, p1.b ; CHECK-NEXT: ret %out = call <vscale x 16 x i1> @llvm.aarch64.sve.uzp1.nxv16i1(<vscale x 16 x i1> %a, <vscale x 16 x i1> %b) ret <vscale x 16 x i1> %out } define <vscale x 8 x i1> @uzp1_b16(<vscale x 8 x i1> %a, <vscale x 8 x i1> %b) { ; CHECK-LABEL: uzp1_b16: ; CHECK: // %bb.0: ; CHECK-NEXT: uzp1 p0.h, p0.h, p1.h ; CHECK-NEXT: ret %out = call <vscale x 8 x i1> @llvm.aarch64.sve.uzp1.nxv8i1(<vscale x 8 x i1> %a, <vscale x 8 x i1> %b) ret <vscale x 8 x i1> %out } define <vscale x 4 x i1> @uzp1_b32(<vscale x 4 x i1> %a, <vscale x 4 x i1> %b) { ; CHECK-LABEL: uzp1_b32: ; CHECK: // %bb.0: ; CHECK-NEXT: uzp1 p0.s, p0.s, p1.s ; CHECK-NEXT: ret %out = call <vscale x 4 x i1> @llvm.aarch64.sve.uzp1.nxv4i1(<vscale x 4 x i1> %a, <vscale x 4 x i1> %b) ret <vscale x 4 x i1> %out } define <vscale x 2 x i1> @uzp1_b64(<vscale x 2 x i1> %a, <vscale x 2 x i1> %b) { ; CHECK-LABEL: uzp1_b64: ; CHECK: // %bb.0: ; CHECK-NEXT: uzp1 p0.d, p0.d, p1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x i1> @llvm.aarch64.sve.uzp1.nxv2i1(<vscale x 2 x i1> %a, <vscale x 2 x i1> %b) ret <vscale x 2 x i1> %out } define <vscale x 16 x i8> @uzp1_i8(<vscale x 16 x i8> %a, <vscale x 16 x i8> %b) { ; CHECK-LABEL: uzp1_i8: ; CHECK: // %bb.0: ; CHECK-NEXT: uzp1 z0.b, z0.b, z1.b ; CHECK-NEXT: ret %out = call <vscale x 16 x i8> @llvm.aarch64.sve.uzp1.nxv16i8(<vscale x 16 x i8> %a, <vscale x 16 x i8> %b) ret <vscale x 16 x i8> %out } define <vscale x 8 x i16> @uzp1_i16(<vscale x 8 x i16> %a, <vscale x 8 x i16> %b) { ; CHECK-LABEL: uzp1_i16: ; CHECK: // %bb.0: ; CHECK-NEXT: uzp1 z0.h, z0.h, z1.h ; CHECK-NEXT: ret %out = call <vscale x 8 x i16> @llvm.aarch64.sve.uzp1.nxv8i16(<vscale x 8 x i16> %a, <vscale x 8 x i16> %b) ret <vscale x 8 x i16> %out } define <vscale x 4 x i32> @uzp1_i32(<vscale x 4 x i32> %a, <vscale x 4 x i32> %b) { ; CHECK-LABEL: uzp1_i32: ; CHECK: // %bb.0: ; CHECK-NEXT: uzp1 z0.s, z0.s, z1.s ; CHECK-NEXT: ret %out = call <vscale x 4 x i32> @llvm.aarch64.sve.uzp1.nxv4i32(<vscale x 4 x i32> %a, <vscale x 4 x i32> %b) ret <vscale x 4 x i32> %out } define <vscale x 2 x i64> @uzp1_i64(<vscale x 2 x i64> %a, <vscale x 2 x i64> %b) { ; CHECK-LABEL: uzp1_i64: ; CHECK: // %bb.0: ; CHECK-NEXT: uzp1 z0.d, z0.d, z1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x i64> @llvm.aarch64.sve.uzp1.nxv2i64(<vscale x 2 x i64> %a, <vscale x 2 x i64> %b) ret <vscale x 2 x i64> %out } define <vscale x 2 x half> @uzp1_f16_v2(<vscale x 2 x half> %a, <vscale x 2 x half> %b) { ; CHECK-LABEL: uzp1_f16_v2: ; CHECK: // %bb.0: ; CHECK-NEXT: uzp1 z0.d, z0.d, z1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x half> @llvm.aarch64.sve.uzp1.nxv2f16(<vscale x 2 x half> %a, <vscale x 2 x half> %b) ret <vscale x 2 x half> %out } define <vscale x 4 x half> @uzp1_f16_v4(<vscale x 4 x half> %a, <vscale x 4 x half> %b) { ; CHECK-LABEL: uzp1_f16_v4: ; CHECK: // %bb.0: ; CHECK-NEXT: uzp1 z0.s, z0.s, z1.s ; CHECK-NEXT: ret %out = call <vscale x 4 x half> @llvm.aarch64.sve.uzp1.nxv4f16(<vscale x 4 x half> %a, <vscale x 4 x half> %b) ret <vscale x 4 x half> %out } define <vscale x 8 x bfloat> @uzp1_bf16(<vscale x 8 x bfloat> %a, <vscale x 8 x bfloat> %b) #0 { ; CHECK-LABEL: uzp1_bf16: ; CHECK: // %bb.0: ; CHECK-NEXT: uzp1 z0.h, z0.h, z1.h ; CHECK-NEXT: ret %out = call <vscale x 8 x bfloat> @llvm.aarch64.sve.uzp1.nxv8bf16(<vscale x 8 x bfloat> %a, <vscale x 8 x bfloat> %b) ret <vscale x 8 x bfloat> %out } define <vscale x 8 x half> @uzp1_f16(<vscale x 8 x half> %a, <vscale x 8 x half> %b) { ; CHECK-LABEL: uzp1_f16: ; CHECK: // %bb.0: ; CHECK-NEXT: uzp1 z0.h, z0.h, z1.h ; CHECK-NEXT: ret %out = call <vscale x 8 x half> @llvm.aarch64.sve.uzp1.nxv8f16(<vscale x 8 x half> %a, <vscale x 8 x half> %b) ret <vscale x 8 x half> %out } define <vscale x 2 x float> @uzp1_f32_v2(<vscale x 2 x float> %a, <vscale x 2 x float> %b) { ; CHECK-LABEL: uzp1_f32_v2: ; CHECK: // %bb.0: ; CHECK-NEXT: uzp1 z0.d, z0.d, z1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x float> @llvm.aarch64.sve.uzp1.nxv2f32(<vscale x 2 x float> %a, <vscale x 2 x float> %b) ret <vscale x 2 x float> %out } define <vscale x 4 x float> @uzp1_f32(<vscale x 4 x float> %a, <vscale x 4 x float> %b) { ; CHECK-LABEL: uzp1_f32: ; CHECK: // %bb.0: ; CHECK-NEXT: uzp1 z0.s, z0.s, z1.s ; CHECK-NEXT: ret %out = call <vscale x 4 x float> @llvm.aarch64.sve.uzp1.nxv4f32(<vscale x 4 x float> %a, <vscale x 4 x float> %b) ret <vscale x 4 x float> %out } define <vscale x 2 x double> @uzp1_f64(<vscale x 2 x double> %a, <vscale x 2 x double> %b) { ; CHECK-LABEL: uzp1_f64: ; CHECK: // %bb.0: ; CHECK-NEXT: uzp1 z0.d, z0.d, z1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x double> @llvm.aarch64.sve.uzp1.nxv2f64(<vscale x 2 x double> %a, <vscale x 2 x double> %b) ret <vscale x 2 x double> %out } ; ; UZP2 ; define <vscale x 16 x i1> @uzp2_b8(<vscale x 16 x i1> %a, <vscale x 16 x i1> %b) { ; CHECK-LABEL: uzp2_b8: ; CHECK: // %bb.0: ; CHECK-NEXT: uzp2 p0.b, p0.b, p1.b ; CHECK-NEXT: ret %out = call <vscale x 16 x i1> @llvm.aarch64.sve.uzp2.nxv16i1(<vscale x 16 x i1> %a, <vscale x 16 x i1> %b) ret <vscale x 16 x i1> %out } define <vscale x 8 x i1> @uzp2_b16(<vscale x 8 x i1> %a, <vscale x 8 x i1> %b) { ; CHECK-LABEL: uzp2_b16: ; CHECK: // %bb.0: ; CHECK-NEXT: uzp2 p0.h, p0.h, p1.h ; CHECK-NEXT: ret %out = call <vscale x 8 x i1> @llvm.aarch64.sve.uzp2.nxv8i1(<vscale x 8 x i1> %a, <vscale x 8 x i1> %b) ret <vscale x 8 x i1> %out } define <vscale x 4 x i1> @uzp2_b32(<vscale x 4 x i1> %a, <vscale x 4 x i1> %b) { ; CHECK-LABEL: uzp2_b32: ; CHECK: // %bb.0: ; CHECK-NEXT: uzp2 p0.s, p0.s, p1.s ; CHECK-NEXT: ret %out = call <vscale x 4 x i1> @llvm.aarch64.sve.uzp2.nxv4i1(<vscale x 4 x i1> %a, <vscale x 4 x i1> %b) ret <vscale x 4 x i1> %out } define <vscale x 2 x i1> @uzp2_b64(<vscale x 2 x i1> %a, <vscale x 2 x i1> %b) { ; CHECK-LABEL: uzp2_b64: ; CHECK: // %bb.0: ; CHECK-NEXT: uzp2 p0.d, p0.d, p1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x i1> @llvm.aarch64.sve.uzp2.nxv2i1(<vscale x 2 x i1> %a, <vscale x 2 x i1> %b) ret <vscale x 2 x i1> %out } define <vscale x 16 x i8> @uzp2_i8(<vscale x 16 x i8> %a, <vscale x 16 x i8> %b) { ; CHECK-LABEL: uzp2_i8: ; CHECK: // %bb.0: ; CHECK-NEXT: uzp2 z0.b, z0.b, z1.b ; CHECK-NEXT: ret %out = call <vscale x 16 x i8> @llvm.aarch64.sve.uzp2.nxv16i8(<vscale x 16 x i8> %a, <vscale x 16 x i8> %b) ret <vscale x 16 x i8> %out } define <vscale x 8 x i16> @uzp2_i16(<vscale x 8 x i16> %a, <vscale x 8 x i16> %b) { ; CHECK-LABEL: uzp2_i16: ; CHECK: // %bb.0: ; CHECK-NEXT: uzp2 z0.h, z0.h, z1.h ; CHECK-NEXT: ret %out = call <vscale x 8 x i16> @llvm.aarch64.sve.uzp2.nxv8i16(<vscale x 8 x i16> %a, <vscale x 8 x i16> %b) ret <vscale x 8 x i16> %out } define <vscale x 4 x i32> @uzp2_i32(<vscale x 4 x i32> %a, <vscale x 4 x i32> %b) { ; CHECK-LABEL: uzp2_i32: ; CHECK: // %bb.0: ; CHECK-NEXT: uzp2 z0.s, z0.s, z1.s ; CHECK-NEXT: ret %out = call <vscale x 4 x i32> @llvm.aarch64.sve.uzp2.nxv4i32(<vscale x 4 x i32> %a, <vscale x 4 x i32> %b) ret <vscale x 4 x i32> %out } define <vscale x 2 x i64> @uzp2_i64(<vscale x 2 x i64> %a, <vscale x 2 x i64> %b) { ; CHECK-LABEL: uzp2_i64: ; CHECK: // %bb.0: ; CHECK-NEXT: uzp2 z0.d, z0.d, z1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x i64> @llvm.aarch64.sve.uzp2.nxv2i64(<vscale x 2 x i64> %a, <vscale x 2 x i64> %b) ret <vscale x 2 x i64> %out } define <vscale x 2 x half> @uzp2_f16_v2(<vscale x 2 x half> %a, <vscale x 2 x half> %b) { ; CHECK-LABEL: uzp2_f16_v2: ; CHECK: // %bb.0: ; CHECK-NEXT: uzp2 z0.d, z0.d, z1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x half> @llvm.aarch64.sve.uzp2.nxv2f16(<vscale x 2 x half> %a, <vscale x 2 x half> %b) ret <vscale x 2 x half> %out } define <vscale x 4 x half> @uzp2_f16_v4(<vscale x 4 x half> %a, <vscale x 4 x half> %b) { ; CHECK-LABEL: uzp2_f16_v4: ; CHECK: // %bb.0: ; CHECK-NEXT: uzp2 z0.s, z0.s, z1.s ; CHECK-NEXT: ret %out = call <vscale x 4 x half> @llvm.aarch64.sve.uzp2.nxv4f16(<vscale x 4 x half> %a, <vscale x 4 x half> %b) ret <vscale x 4 x half> %out } define <vscale x 8 x bfloat> @uzp2_bf16(<vscale x 8 x bfloat> %a, <vscale x 8 x bfloat> %b) #0 { ; CHECK-LABEL: uzp2_bf16: ; CHECK: // %bb.0: ; CHECK-NEXT: uzp2 z0.h, z0.h, z1.h ; CHECK-NEXT: ret %out = call <vscale x 8 x bfloat> @llvm.aarch64.sve.uzp2.nxv8bf16(<vscale x 8 x bfloat> %a, <vscale x 8 x bfloat> %b) ret <vscale x 8 x bfloat> %out } define <vscale x 8 x half> @uzp2_f16(<vscale x 8 x half> %a, <vscale x 8 x half> %b) { ; CHECK-LABEL: uzp2_f16: ; CHECK: // %bb.0: ; CHECK-NEXT: uzp2 z0.h, z0.h, z1.h ; CHECK-NEXT: ret %out = call <vscale x 8 x half> @llvm.aarch64.sve.uzp2.nxv8f16(<vscale x 8 x half> %a, <vscale x 8 x half> %b) ret <vscale x 8 x half> %out } define <vscale x 2 x float> @uzp2_f32_v2(<vscale x 2 x float> %a, <vscale x 2 x float> %b) { ; CHECK-LABEL: uzp2_f32_v2: ; CHECK: // %bb.0: ; CHECK-NEXT: uzp2 z0.d, z0.d, z1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x float> @llvm.aarch64.sve.uzp2.nxv2f32(<vscale x 2 x float> %a, <vscale x 2 x float> %b) ret <vscale x 2 x float> %out } define <vscale x 4 x float> @uzp2_f32(<vscale x 4 x float> %a, <vscale x 4 x float> %b) { ; CHECK-LABEL: uzp2_f32: ; CHECK: // %bb.0: ; CHECK-NEXT: uzp2 z0.s, z0.s, z1.s ; CHECK-NEXT: ret %out = call <vscale x 4 x float> @llvm.aarch64.sve.uzp2.nxv4f32(<vscale x 4 x float> %a, <vscale x 4 x float> %b) ret <vscale x 4 x float> %out } define <vscale x 2 x double> @uzp2_f64(<vscale x 2 x double> %a, <vscale x 2 x double> %b) { ; CHECK-LABEL: uzp2_f64: ; CHECK: // %bb.0: ; CHECK-NEXT: uzp2 z0.d, z0.d, z1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x double> @llvm.aarch64.sve.uzp2.nxv2f64(<vscale x 2 x double> %a, <vscale x 2 x double> %b) ret <vscale x 2 x double> %out } ; ; ZIP1 ; define <vscale x 16 x i1> @zip1_b8(<vscale x 16 x i1> %a, <vscale x 16 x i1> %b) { ; CHECK-LABEL: zip1_b8: ; CHECK: // %bb.0: ; CHECK-NEXT: zip1 p0.b, p0.b, p1.b ; CHECK-NEXT: ret %out = call <vscale x 16 x i1> @llvm.aarch64.sve.zip1.nxv16i1(<vscale x 16 x i1> %a, <vscale x 16 x i1> %b) ret <vscale x 16 x i1> %out } define <vscale x 8 x i1> @zip1_b16(<vscale x 8 x i1> %a, <vscale x 8 x i1> %b) { ; CHECK-LABEL: zip1_b16: ; CHECK: // %bb.0: ; CHECK-NEXT: zip1 p0.h, p0.h, p1.h ; CHECK-NEXT: ret %out = call <vscale x 8 x i1> @llvm.aarch64.sve.zip1.nxv8i1(<vscale x 8 x i1> %a, <vscale x 8 x i1> %b) ret <vscale x 8 x i1> %out } define <vscale x 4 x i1> @zip1_b32(<vscale x 4 x i1> %a, <vscale x 4 x i1> %b) { ; CHECK-LABEL: zip1_b32: ; CHECK: // %bb.0: ; CHECK-NEXT: zip1 p0.s, p0.s, p1.s ; CHECK-NEXT: ret %out = call <vscale x 4 x i1> @llvm.aarch64.sve.zip1.nxv4i1(<vscale x 4 x i1> %a, <vscale x 4 x i1> %b) ret <vscale x 4 x i1> %out } define <vscale x 2 x i1> @zip1_b64(<vscale x 2 x i1> %a, <vscale x 2 x i1> %b) { ; CHECK-LABEL: zip1_b64: ; CHECK: // %bb.0: ; CHECK-NEXT: zip1 p0.d, p0.d, p1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x i1> @llvm.aarch64.sve.zip1.nxv2i1(<vscale x 2 x i1> %a, <vscale x 2 x i1> %b) ret <vscale x 2 x i1> %out } define <vscale x 16 x i8> @zip1_i8(<vscale x 16 x i8> %a, <vscale x 16 x i8> %b) { ; CHECK-LABEL: zip1_i8: ; CHECK: // %bb.0: ; CHECK-NEXT: zip1 z0.b, z0.b, z1.b ; CHECK-NEXT: ret %out = call <vscale x 16 x i8> @llvm.aarch64.sve.zip1.nxv16i8(<vscale x 16 x i8> %a, <vscale x 16 x i8> %b) ret <vscale x 16 x i8> %out } define <vscale x 8 x i16> @zip1_i16(<vscale x 8 x i16> %a, <vscale x 8 x i16> %b) { ; CHECK-LABEL: zip1_i16: ; CHECK: // %bb.0: ; CHECK-NEXT: zip1 z0.h, z0.h, z1.h ; CHECK-NEXT: ret %out = call <vscale x 8 x i16> @llvm.aarch64.sve.zip1.nxv8i16(<vscale x 8 x i16> %a, <vscale x 8 x i16> %b) ret <vscale x 8 x i16> %out } define <vscale x 4 x i32> @zip1_i32(<vscale x 4 x i32> %a, <vscale x 4 x i32> %b) { ; CHECK-LABEL: zip1_i32: ; CHECK: // %bb.0: ; CHECK-NEXT: zip1 z0.s, z0.s, z1.s ; CHECK-NEXT: ret %out = call <vscale x 4 x i32> @llvm.aarch64.sve.zip1.nxv4i32(<vscale x 4 x i32> %a, <vscale x 4 x i32> %b) ret <vscale x 4 x i32> %out } define <vscale x 2 x i64> @zip1_i64(<vscale x 2 x i64> %a, <vscale x 2 x i64> %b) { ; CHECK-LABEL: zip1_i64: ; CHECK: // %bb.0: ; CHECK-NEXT: zip1 z0.d, z0.d, z1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x i64> @llvm.aarch64.sve.zip1.nxv2i64(<vscale x 2 x i64> %a, <vscale x 2 x i64> %b) ret <vscale x 2 x i64> %out } define <vscale x 2 x half> @zip1_f16_v2(<vscale x 2 x half> %a, <vscale x 2 x half> %b) { ; CHECK-LABEL: zip1_f16_v2: ; CHECK: // %bb.0: ; CHECK-NEXT: zip1 z0.d, z0.d, z1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x half> @llvm.aarch64.sve.zip1.nxv2f16(<vscale x 2 x half> %a, <vscale x 2 x half> %b) ret <vscale x 2 x half> %out } define <vscale x 4 x half> @zip1_f16_v4(<vscale x 4 x half> %a, <vscale x 4 x half> %b) { ; CHECK-LABEL: zip1_f16_v4: ; CHECK: // %bb.0: ; CHECK-NEXT: zip1 z0.s, z0.s, z1.s ; CHECK-NEXT: ret %out = call <vscale x 4 x half> @llvm.aarch64.sve.zip1.nxv4f16(<vscale x 4 x half> %a, <vscale x 4 x half> %b) ret <vscale x 4 x half> %out } define <vscale x 8 x bfloat> @zip1_bf16(<vscale x 8 x bfloat> %a, <vscale x 8 x bfloat> %b) #0 { ; CHECK-LABEL: zip1_bf16: ; CHECK: // %bb.0: ; CHECK-NEXT: zip1 z0.h, z0.h, z1.h ; CHECK-NEXT: ret %out = call <vscale x 8 x bfloat> @llvm.aarch64.sve.zip1.nxv8bf16(<vscale x 8 x bfloat> %a, <vscale x 8 x bfloat> %b) ret <vscale x 8 x bfloat> %out } define <vscale x 8 x half> @zip1_f16(<vscale x 8 x half> %a, <vscale x 8 x half> %b) { ; CHECK-LABEL: zip1_f16: ; CHECK: // %bb.0: ; CHECK-NEXT: zip1 z0.h, z0.h, z1.h ; CHECK-NEXT: ret %out = call <vscale x 8 x half> @llvm.aarch64.sve.zip1.nxv8f16(<vscale x 8 x half> %a, <vscale x 8 x half> %b) ret <vscale x 8 x half> %out } define <vscale x 2 x float> @zip1_f32_v2(<vscale x 2 x float> %a, <vscale x 2 x float> %b) { ; CHECK-LABEL: zip1_f32_v2: ; CHECK: // %bb.0: ; CHECK-NEXT: zip1 z0.d, z0.d, z1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x float> @llvm.aarch64.sve.zip1.nxv2f32(<vscale x 2 x float> %a, <vscale x 2 x float> %b) ret <vscale x 2 x float> %out } define <vscale x 4 x float> @zip1_f32(<vscale x 4 x float> %a, <vscale x 4 x float> %b) { ; CHECK-LABEL: zip1_f32: ; CHECK: // %bb.0: ; CHECK-NEXT: zip1 z0.s, z0.s, z1.s ; CHECK-NEXT: ret %out = call <vscale x 4 x float> @llvm.aarch64.sve.zip1.nxv4f32(<vscale x 4 x float> %a, <vscale x 4 x float> %b) ret <vscale x 4 x float> %out } define <vscale x 2 x double> @zip1_f64(<vscale x 2 x double> %a, <vscale x 2 x double> %b) { ; CHECK-LABEL: zip1_f64: ; CHECK: // %bb.0: ; CHECK-NEXT: zip1 z0.d, z0.d, z1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x double> @llvm.aarch64.sve.zip1.nxv2f64(<vscale x 2 x double> %a, <vscale x 2 x double> %b) ret <vscale x 2 x double> %out } ; ; ZIP2 ; define <vscale x 16 x i1> @zip2_b8(<vscale x 16 x i1> %a, <vscale x 16 x i1> %b) { ; CHECK-LABEL: zip2_b8: ; CHECK: // %bb.0: ; CHECK-NEXT: zip2 p0.b, p0.b, p1.b ; CHECK-NEXT: ret %out = call <vscale x 16 x i1> @llvm.aarch64.sve.zip2.nxv16i1(<vscale x 16 x i1> %a, <vscale x 16 x i1> %b) ret <vscale x 16 x i1> %out } define <vscale x 8 x i1> @zip2_b16(<vscale x 8 x i1> %a, <vscale x 8 x i1> %b) { ; CHECK-LABEL: zip2_b16: ; CHECK: // %bb.0: ; CHECK-NEXT: zip2 p0.h, p0.h, p1.h ; CHECK-NEXT: ret %out = call <vscale x 8 x i1> @llvm.aarch64.sve.zip2.nxv8i1(<vscale x 8 x i1> %a, <vscale x 8 x i1> %b) ret <vscale x 8 x i1> %out } define <vscale x 4 x i1> @zip2_b32(<vscale x 4 x i1> %a, <vscale x 4 x i1> %b) { ; CHECK-LABEL: zip2_b32: ; CHECK: // %bb.0: ; CHECK-NEXT: zip2 p0.s, p0.s, p1.s ; CHECK-NEXT: ret %out = call <vscale x 4 x i1> @llvm.aarch64.sve.zip2.nxv4i1(<vscale x 4 x i1> %a, <vscale x 4 x i1> %b) ret <vscale x 4 x i1> %out } define <vscale x 2 x i1> @zip2_b64(<vscale x 2 x i1> %a, <vscale x 2 x i1> %b) { ; CHECK-LABEL: zip2_b64: ; CHECK: // %bb.0: ; CHECK-NEXT: zip2 p0.d, p0.d, p1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x i1> @llvm.aarch64.sve.zip2.nxv2i1(<vscale x 2 x i1> %a, <vscale x 2 x i1> %b) ret <vscale x 2 x i1> %out } define <vscale x 16 x i8> @zip2_i8(<vscale x 16 x i8> %a, <vscale x 16 x i8> %b) { ; CHECK-LABEL: zip2_i8: ; CHECK: // %bb.0: ; CHECK-NEXT: zip2 z0.b, z0.b, z1.b ; CHECK-NEXT: ret %out = call <vscale x 16 x i8> @llvm.aarch64.sve.zip2.nxv16i8(<vscale x 16 x i8> %a, <vscale x 16 x i8> %b) ret <vscale x 16 x i8> %out } define <vscale x 8 x i16> @zip2_i16(<vscale x 8 x i16> %a, <vscale x 8 x i16> %b) { ; CHECK-LABEL: zip2_i16: ; CHECK: // %bb.0: ; CHECK-NEXT: zip2 z0.h, z0.h, z1.h ; CHECK-NEXT: ret %out = call <vscale x 8 x i16> @llvm.aarch64.sve.zip2.nxv8i16(<vscale x 8 x i16> %a, <vscale x 8 x i16> %b) ret <vscale x 8 x i16> %out } define <vscale x 4 x i32> @zip2_i32(<vscale x 4 x i32> %a, <vscale x 4 x i32> %b) { ; CHECK-LABEL: zip2_i32: ; CHECK: // %bb.0: ; CHECK-NEXT: zip2 z0.s, z0.s, z1.s ; CHECK-NEXT: ret %out = call <vscale x 4 x i32> @llvm.aarch64.sve.zip2.nxv4i32(<vscale x 4 x i32> %a, <vscale x 4 x i32> %b) ret <vscale x 4 x i32> %out } define <vscale x 2 x i64> @zip2_i64(<vscale x 2 x i64> %a, <vscale x 2 x i64> %b) { ; CHECK-LABEL: zip2_i64: ; CHECK: // %bb.0: ; CHECK-NEXT: zip2 z0.d, z0.d, z1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x i64> @llvm.aarch64.sve.zip2.nxv2i64(<vscale x 2 x i64> %a, <vscale x 2 x i64> %b) ret <vscale x 2 x i64> %out } define <vscale x 2 x half> @zip2_f16_v2(<vscale x 2 x half> %a, <vscale x 2 x half> %b) { ; CHECK-LABEL: zip2_f16_v2: ; CHECK: // %bb.0: ; CHECK-NEXT: zip2 z0.d, z0.d, z1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x half> @llvm.aarch64.sve.zip2.nxv2f16(<vscale x 2 x half> %a, <vscale x 2 x half> %b) ret <vscale x 2 x half> %out } define <vscale x 4 x half> @zip2_f16_v4(<vscale x 4 x half> %a, <vscale x 4 x half> %b) { ; CHECK-LABEL: zip2_f16_v4: ; CHECK: // %bb.0: ; CHECK-NEXT: zip2 z0.s, z0.s, z1.s ; CHECK-NEXT: ret %out = call <vscale x 4 x half> @llvm.aarch64.sve.zip2.nxv4f16(<vscale x 4 x half> %a, <vscale x 4 x half> %b) ret <vscale x 4 x half> %out } define <vscale x 8 x bfloat> @zip2_bf16(<vscale x 8 x bfloat> %a, <vscale x 8 x bfloat> %b) #0 { ; CHECK-LABEL: zip2_bf16: ; CHECK: // %bb.0: ; CHECK-NEXT: zip2 z0.h, z0.h, z1.h ; CHECK-NEXT: ret %out = call <vscale x 8 x bfloat> @llvm.aarch64.sve.zip2.nxv8bf16(<vscale x 8 x bfloat> %a, <vscale x 8 x bfloat> %b) ret <vscale x 8 x bfloat> %out } define <vscale x 8 x half> @zip2_f16(<vscale x 8 x half> %a, <vscale x 8 x half> %b) { ; CHECK-LABEL: zip2_f16: ; CHECK: // %bb.0: ; CHECK-NEXT: zip2 z0.h, z0.h, z1.h ; CHECK-NEXT: ret %out = call <vscale x 8 x half> @llvm.aarch64.sve.zip2.nxv8f16(<vscale x 8 x half> %a, <vscale x 8 x half> %b) ret <vscale x 8 x half> %out } define <vscale x 2 x float> @zip2_f32_v2(<vscale x 2 x float> %a, <vscale x 2 x float> %b) { ; CHECK-LABEL: zip2_f32_v2: ; CHECK: // %bb.0: ; CHECK-NEXT: zip2 z0.d, z0.d, z1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x float> @llvm.aarch64.sve.zip2.nxv2f32(<vscale x 2 x float> %a, <vscale x 2 x float> %b) ret <vscale x 2 x float> %out } define <vscale x 4 x float> @zip2_f32(<vscale x 4 x float> %a, <vscale x 4 x float> %b) { ; CHECK-LABEL: zip2_f32: ; CHECK: // %bb.0: ; CHECK-NEXT: zip2 z0.s, z0.s, z1.s ; CHECK-NEXT: ret %out = call <vscale x 4 x float> @llvm.aarch64.sve.zip2.nxv4f32(<vscale x 4 x float> %a, <vscale x 4 x float> %b) ret <vscale x 4 x float> %out } define <vscale x 2 x double> @zip2_f64(<vscale x 2 x double> %a, <vscale x 2 x double> %b) { ; CHECK-LABEL: zip2_f64: ; CHECK: // %bb.0: ; CHECK-NEXT: zip2 z0.d, z0.d, z1.d ; CHECK-NEXT: ret %out = call <vscale x 2 x double> @llvm.aarch64.sve.zip2.nxv2f64(<vscale x 2 x double> %a, <vscale x 2 x double> %b) ret <vscale x 2 x double> %out } declare <vscale x 16 x i8> @llvm.aarch64.sve.clasta.nxv16i8(<vscale x 16 x i1>, <vscale x 16 x i8>, <vscale x 16 x i8>) declare <vscale x 8 x i16> @llvm.aarch64.sve.clasta.nxv8i16(<vscale x 8 x i1>, <vscale x 8 x i16>, <vscale x 8 x i16>) declare <vscale x 4 x i32> @llvm.aarch64.sve.clasta.nxv4i32(<vscale x 4 x i1>, <vscale x 4 x i32>, <vscale x 4 x i32>) declare <vscale x 2 x i64> @llvm.aarch64.sve.clasta.nxv2i64(<vscale x 2 x i1>, <vscale x 2 x i64>, <vscale x 2 x i64>) declare <vscale x 8 x half> @llvm.aarch64.sve.clasta.nxv8f16(<vscale x 8 x i1>, <vscale x 8 x half>, <vscale x 8 x half>) declare <vscale x 8 x bfloat> @llvm.aarch64.sve.clasta.nxv8bf16(<vscale x 8 x i1>, <vscale x 8 x bfloat>, <vscale x 8 x bfloat>) declare <vscale x 4 x float> @llvm.aarch64.sve.clasta.nxv4f32(<vscale x 4 x i1>, <vscale x 4 x float>, <vscale x 4 x float>) declare <vscale x 2 x double> @llvm.aarch64.sve.clasta.nxv2f64(<vscale x 2 x i1>, <vscale x 2 x double>, <vscale x 2 x double>) declare i8 @llvm.aarch64.sve.clasta.n.nxv16i8(<vscale x 16 x i1>, i8, <vscale x 16 x i8>) declare i16 @llvm.aarch64.sve.clasta.n.nxv8i16(<vscale x 8 x i1>, i16, <vscale x 8 x i16>) declare i32 @llvm.aarch64.sve.clasta.n.nxv4i32(<vscale x 4 x i1>, i32, <vscale x 4 x i32>) declare i64 @llvm.aarch64.sve.clasta.n.nxv2i64(<vscale x 2 x i1>, i64, <vscale x 2 x i64>) declare half @llvm.aarch64.sve.clasta.n.nxv8f16(<vscale x 8 x i1>, half, <vscale x 8 x half>) declare bfloat @llvm.aarch64.sve.clasta.n.nxv8bf16(<vscale x 8 x i1>, bfloat, <vscale x 8 x bfloat>) declare float @llvm.aarch64.sve.clasta.n.nxv4f32(<vscale x 4 x i1>, float, <vscale x 4 x float>) declare double @llvm.aarch64.sve.clasta.n.nxv2f64(<vscale x 2 x i1>, double, <vscale x 2 x double>) declare <vscale x 16 x i8> @llvm.aarch64.sve.clastb.nxv16i8(<vscale x 16 x i1>, <vscale x 16 x i8>, <vscale x 16 x i8>) declare <vscale x 8 x i16> @llvm.aarch64.sve.clastb.nxv8i16(<vscale x 8 x i1>, <vscale x 8 x i16>, <vscale x 8 x i16>) declare <vscale x 4 x i32> @llvm.aarch64.sve.clastb.nxv4i32(<vscale x 4 x i1>, <vscale x 4 x i32>, <vscale x 4 x i32>) declare <vscale x 2 x i64> @llvm.aarch64.sve.clastb.nxv2i64(<vscale x 2 x i1>, <vscale x 2 x i64>, <vscale x 2 x i64>) declare <vscale x 8 x half> @llvm.aarch64.sve.clastb.nxv8f16(<vscale x 8 x i1>, <vscale x 8 x half>, <vscale x 8 x half>) declare <vscale x 8 x bfloat> @llvm.aarch64.sve.clastb.nxv8bf16(<vscale x 8 x i1>, <vscale x 8 x bfloat>, <vscale x 8 x bfloat>) declare <vscale x 4 x float> @llvm.aarch64.sve.clastb.nxv4f32(<vscale x 4 x i1>, <vscale x 4 x float>, <vscale x 4 x float>) declare <vscale x 2 x double> @llvm.aarch64.sve.clastb.nxv2f64(<vscale x 2 x i1>, <vscale x 2 x double>, <vscale x 2 x double>) declare i8 @llvm.aarch64.sve.clastb.n.nxv16i8(<vscale x 16 x i1>, i8, <vscale x 16 x i8>) declare i16 @llvm.aarch64.sve.clastb.n.nxv8i16(<vscale x 8 x i1>, i16, <vscale x 8 x i16>) declare i32 @llvm.aarch64.sve.clastb.n.nxv4i32(<vscale x 4 x i1>, i32, <vscale x 4 x i32>) declare i64 @llvm.aarch64.sve.clastb.n.nxv2i64(<vscale x 2 x i1>, i64, <vscale x 2 x i64>) declare half @llvm.aarch64.sve.clastb.n.nxv8f16(<vscale x 8 x i1>, half, <vscale x 8 x half>) declare bfloat @llvm.aarch64.sve.clastb.n.nxv8bf16(<vscale x 8 x i1>, bfloat, <vscale x 8 x bfloat>) declare float @llvm.aarch64.sve.clastb.n.nxv4f32(<vscale x 4 x i1>, float, <vscale x 4 x float>) declare double @llvm.aarch64.sve.clastb.n.nxv2f64(<vscale x 2 x i1>, double, <vscale x 2 x double>) declare <vscale x 4 x i32> @llvm.aarch64.sve.compact.nxv4i32(<vscale x 4 x i1>, <vscale x 4 x i32>) declare <vscale x 2 x i64> @llvm.aarch64.sve.compact.nxv2i64(<vscale x 2 x i1>, <vscale x 2 x i64>) declare <vscale x 4 x float> @llvm.aarch64.sve.compact.nxv4f32(<vscale x 4 x i1>, <vscale x 4 x float>) declare <vscale x 2 x double> @llvm.aarch64.sve.compact.nxv2f64(<vscale x 2 x i1>, <vscale x 2 x double>) declare <vscale x 16 x i8> @llvm.aarch64.sve.dupq.lane.nxv16i8(<vscale x 16 x i8>, i64) declare <vscale x 8 x i16> @llvm.aarch64.sve.dupq.lane.nxv8i16(<vscale x 8 x i16>, i64) declare <vscale x 4 x i32> @llvm.aarch64.sve.dupq.lane.nxv4i32(<vscale x 4 x i32>, i64) declare <vscale x 2 x i64> @llvm.aarch64.sve.dupq.lane.nxv2i64(<vscale x 2 x i64>, i64) declare <vscale x 8 x half> @llvm.aarch64.sve.dupq.lane.nxv8f16(<vscale x 8 x half>, i64) declare <vscale x 8 x bfloat> @llvm.aarch64.sve.dupq.lane.nxv8bf16(<vscale x 8 x bfloat>, i64) declare <vscale x 4 x float> @llvm.aarch64.sve.dupq.lane.nxv4f32(<vscale x 4 x float>, i64) declare <vscale x 2 x double> @llvm.aarch64.sve.dupq.lane.nxv2f64(<vscale x 2 x double>, i64) declare <vscale x 16 x i8> @llvm.aarch64.sve.ext.nxv16i8(<vscale x 16 x i8>, <vscale x 16 x i8>, i32) declare <vscale x 8 x i16> @llvm.aarch64.sve.ext.nxv8i16(<vscale x 8 x i16>, <vscale x 8 x i16>, i32) declare <vscale x 4 x i32> @llvm.aarch64.sve.ext.nxv4i32(<vscale x 4 x i32>, <vscale x 4 x i32>, i32) declare <vscale x 2 x i64> @llvm.aarch64.sve.ext.nxv2i64(<vscale x 2 x i64>, <vscale x 2 x i64>, i32) declare <vscale x 8 x bfloat> @llvm.aarch64.sve.ext.nxv8bf16(<vscale x 8 x bfloat>, <vscale x 8 x bfloat>, i32) declare <vscale x 8 x half> @llvm.aarch64.sve.ext.nxv8f16(<vscale x 8 x half>, <vscale x 8 x half>, i32) declare <vscale x 4 x float> @llvm.aarch64.sve.ext.nxv4f32(<vscale x 4 x float>, <vscale x 4 x float>, i32) declare <vscale x 2 x double> @llvm.aarch64.sve.ext.nxv2f64(<vscale x 2 x double>, <vscale x 2 x double>, i32) declare i8 @llvm.aarch64.sve.lasta.nxv16i8(<vscale x 16 x i1>, <vscale x 16 x i8>) declare i16 @llvm.aarch64.sve.lasta.nxv8i16(<vscale x 8 x i1>, <vscale x 8 x i16>) declare i32 @llvm.aarch64.sve.lasta.nxv4i32(<vscale x 4 x i1>, <vscale x 4 x i32>) declare i64 @llvm.aarch64.sve.lasta.nxv2i64(<vscale x 2 x i1>, <vscale x 2 x i64>) declare half @llvm.aarch64.sve.lasta.nxv8f16(<vscale x 8 x i1>, <vscale x 8 x half>) declare bfloat @llvm.aarch64.sve.lasta.nxv8bf16(<vscale x 8 x i1>, <vscale x 8 x bfloat>) declare float @llvm.aarch64.sve.lasta.nxv2f32(<vscale x 2 x i1>, <vscale x 2 x float>) declare float @llvm.aarch64.sve.lasta.nxv4f32(<vscale x 4 x i1>, <vscale x 4 x float>) declare double @llvm.aarch64.sve.lasta.nxv2f64(<vscale x 2 x i1>, <vscale x 2 x double>) declare i8 @llvm.aarch64.sve.lastb.nxv16i8(<vscale x 16 x i1>, <vscale x 16 x i8>) declare i16 @llvm.aarch64.sve.lastb.nxv8i16(<vscale x 8 x i1>, <vscale x 8 x i16>) declare i32 @llvm.aarch64.sve.lastb.nxv4i32(<vscale x 4 x i1>, <vscale x 4 x i32>) declare i64 @llvm.aarch64.sve.lastb.nxv2i64(<vscale x 2 x i1>, <vscale x 2 x i64>) declare half @llvm.aarch64.sve.lastb.nxv8f16(<vscale x 8 x i1>, <vscale x 8 x half>) declare bfloat @llvm.aarch64.sve.lastb.nxv8bf16(<vscale x 8 x i1>, <vscale x 8 x bfloat>) declare float @llvm.aarch64.sve.lastb.nxv2f32(<vscale x 2 x i1>, <vscale x 2 x float>) declare float @llvm.aarch64.sve.lastb.nxv4f32(<vscale x 4 x i1>, <vscale x 4 x float>) declare double @llvm.aarch64.sve.lastb.nxv2f64(<vscale x 2 x i1>, <vscale x 2 x double>) declare <vscale x 16 x i1> @llvm.aarch64.sve.rev.nxv16i1(<vscale x 16 x i1>) declare <vscale x 8 x i1> @llvm.aarch64.sve.rev.nxv8i1(<vscale x 8 x i1>) declare <vscale x 4 x i1> @llvm.aarch64.sve.rev.nxv4i1(<vscale x 4 x i1>) declare <vscale x 2 x i1> @llvm.aarch64.sve.rev.nxv2i1(<vscale x 2 x i1>) declare <vscale x 16 x i8> @llvm.aarch64.sve.rev.nxv16i8(<vscale x 16 x i8>) declare <vscale x 8 x i16> @llvm.aarch64.sve.rev.nxv8i16(<vscale x 8 x i16>) declare <vscale x 4 x i32> @llvm.aarch64.sve.rev.nxv4i32(<vscale x 4 x i32>) declare <vscale x 2 x i64> @llvm.aarch64.sve.rev.nxv2i64(<vscale x 2 x i64>) declare <vscale x 8 x bfloat> @llvm.aarch64.sve.rev.nxv8bf16(<vscale x 8 x bfloat>) declare <vscale x 8 x half> @llvm.aarch64.sve.rev.nxv8f16(<vscale x 8 x half>) declare <vscale x 4 x float> @llvm.aarch64.sve.rev.nxv4f32(<vscale x 4 x float>) declare <vscale x 2 x double> @llvm.aarch64.sve.rev.nxv2f64(<vscale x 2 x double>) declare <vscale x 16 x i8> @llvm.aarch64.sve.splice.nxv16i8(<vscale x 16 x i1>, <vscale x 16 x i8>, <vscale x 16 x i8>) declare <vscale x 8 x i16> @llvm.aarch64.sve.splice.nxv8i16(<vscale x 8 x i1>, <vscale x 8 x i16>, <vscale x 8 x i16>) declare <vscale x 4 x i32> @llvm.aarch64.sve.splice.nxv4i32(<vscale x 4 x i1>, <vscale x 4 x i32>, <vscale x 4 x i32>) declare <vscale x 2 x i64> @llvm.aarch64.sve.splice.nxv2i64(<vscale x 2 x i1>, <vscale x 2 x i64>, <vscale x 2 x i64>) declare <vscale x 8 x bfloat> @llvm.aarch64.sve.splice.nxv8bf16(<vscale x 8 x i1>, <vscale x 8 x bfloat>, <vscale x 8 x bfloat>) declare <vscale x 8 x half> @llvm.aarch64.sve.splice.nxv8f16(<vscale x 8 x i1>, <vscale x 8 x half>, <vscale x 8 x half>) declare <vscale x 4 x float> @llvm.aarch64.sve.splice.nxv4f32(<vscale x 4 x i1>, <vscale x 4 x float>, <vscale x 4 x float>) declare <vscale x 2 x double> @llvm.aarch64.sve.splice.nxv2f64(<vscale x 2 x i1>, <vscale x 2 x double>, <vscale x 2 x double>) declare <vscale x 8 x i16> @llvm.aarch64.sve.sunpkhi.nxv8i16(<vscale x 16 x i8>) declare <vscale x 4 x i32> @llvm.aarch64.sve.sunpkhi.nxv4i32(<vscale x 8 x i16>) declare <vscale x 2 x i64> @llvm.aarch64.sve.sunpkhi.nxv2i64(<vscale x 4 x i32>) declare <vscale x 8 x i16> @llvm.aarch64.sve.sunpklo.nxv8i16(<vscale x 16 x i8>) declare <vscale x 4 x i32> @llvm.aarch64.sve.sunpklo.nxv4i32(<vscale x 8 x i16>) declare <vscale x 2 x i64> @llvm.aarch64.sve.sunpklo.nxv2i64(<vscale x 4 x i32>) declare <vscale x 16 x i8> @llvm.aarch64.sve.tbl.nxv16i8(<vscale x 16 x i8>, <vscale x 16 x i8>) declare <vscale x 8 x i16> @llvm.aarch64.sve.tbl.nxv8i16(<vscale x 8 x i16>, <vscale x 8 x i16>) declare <vscale x 4 x i32> @llvm.aarch64.sve.tbl.nxv4i32(<vscale x 4 x i32>, <vscale x 4 x i32>) declare <vscale x 2 x i64> @llvm.aarch64.sve.tbl.nxv2i64(<vscale x 2 x i64>, <vscale x 2 x i64>) declare <vscale x 8 x half> @llvm.aarch64.sve.tbl.nxv8f16(<vscale x 8 x half>, <vscale x 8 x i16>) declare <vscale x 8 x bfloat> @llvm.aarch64.sve.tbl.nxv8bf16(<vscale x 8 x bfloat>, <vscale x 8 x i16>) declare <vscale x 4 x float> @llvm.aarch64.sve.tbl.nxv4f32(<vscale x 4 x float>, <vscale x 4 x i32>) declare <vscale x 2 x double> @llvm.aarch64.sve.tbl.nxv2f64(<vscale x 2 x double>, <vscale x 2 x i64>) declare <vscale x 8 x i16> @llvm.aarch64.sve.uunpkhi.nxv8i16(<vscale x 16 x i8>) declare <vscale x 4 x i32> @llvm.aarch64.sve.uunpkhi.nxv4i32(<vscale x 8 x i16>) declare <vscale x 2 x i64> @llvm.aarch64.sve.uunpkhi.nxv2i64(<vscale x 4 x i32>) declare <vscale x 8 x i16> @llvm.aarch64.sve.uunpklo.nxv8i16(<vscale x 16 x i8>) declare <vscale x 4 x i32> @llvm.aarch64.sve.uunpklo.nxv4i32(<vscale x 8 x i16>) declare <vscale x 2 x i64> @llvm.aarch64.sve.uunpklo.nxv2i64(<vscale x 4 x i32>) declare <vscale x 16 x i1> @llvm.aarch64.sve.trn1.nxv16i1(<vscale x 16 x i1>, <vscale x 16 x i1>) declare <vscale x 8 x i1> @llvm.aarch64.sve.trn1.nxv8i1(<vscale x 8 x i1>, <vscale x 8 x i1>) declare <vscale x 4 x i1> @llvm.aarch64.sve.trn1.nxv4i1(<vscale x 4 x i1>, <vscale x 4 x i1>) declare <vscale x 2 x i1> @llvm.aarch64.sve.trn1.nxv2i1(<vscale x 2 x i1>, <vscale x 2 x i1>) declare <vscale x 16 x i8> @llvm.aarch64.sve.trn1.nxv16i8(<vscale x 16 x i8>, <vscale x 16 x i8>) declare <vscale x 8 x i16> @llvm.aarch64.sve.trn1.nxv8i16(<vscale x 8 x i16>, <vscale x 8 x i16>) declare <vscale x 4 x i32> @llvm.aarch64.sve.trn1.nxv4i32(<vscale x 4 x i32>, <vscale x 4 x i32>) declare <vscale x 2 x i64> @llvm.aarch64.sve.trn1.nxv2i64(<vscale x 2 x i64>, <vscale x 2 x i64>) declare <vscale x 2 x half> @llvm.aarch64.sve.trn1.nxv2f16(<vscale x 2 x half>, <vscale x 2 x half>) declare <vscale x 4 x half> @llvm.aarch64.sve.trn1.nxv4f16(<vscale x 4 x half>, <vscale x 4 x half>) declare <vscale x 8 x bfloat> @llvm.aarch64.sve.trn1.nxv8bf16(<vscale x 8 x bfloat>, <vscale x 8 x bfloat>) declare <vscale x 8 x half> @llvm.aarch64.sve.trn1.nxv8f16(<vscale x 8 x half>, <vscale x 8 x half>) declare <vscale x 2 x float> @llvm.aarch64.sve.trn1.nxv2f32(<vscale x 2 x float>, <vscale x 2 x float>) declare <vscale x 4 x float> @llvm.aarch64.sve.trn1.nxv4f32(<vscale x 4 x float>, <vscale x 4 x float>) declare <vscale x 2 x double> @llvm.aarch64.sve.trn1.nxv2f64(<vscale x 2 x double>, <vscale x 2 x double>) declare <vscale x 16 x i1> @llvm.aarch64.sve.trn2.nxv16i1(<vscale x 16 x i1>, <vscale x 16 x i1>) declare <vscale x 8 x i1> @llvm.aarch64.sve.trn2.nxv8i1(<vscale x 8 x i1>, <vscale x 8 x i1>) declare <vscale x 4 x i1> @llvm.aarch64.sve.trn2.nxv4i1(<vscale x 4 x i1>, <vscale x 4 x i1>) declare <vscale x 2 x i1> @llvm.aarch64.sve.trn2.nxv2i1(<vscale x 2 x i1>, <vscale x 2 x i1>) declare <vscale x 16 x i8> @llvm.aarch64.sve.trn2.nxv16i8(<vscale x 16 x i8>, <vscale x 16 x i8>) declare <vscale x 8 x i16> @llvm.aarch64.sve.trn2.nxv8i16(<vscale x 8 x i16>, <vscale x 8 x i16>) declare <vscale x 4 x i32> @llvm.aarch64.sve.trn2.nxv4i32(<vscale x 4 x i32>, <vscale x 4 x i32>) declare <vscale x 2 x i64> @llvm.aarch64.sve.trn2.nxv2i64(<vscale x 2 x i64>, <vscale x 2 x i64>) declare <vscale x 2 x half> @llvm.aarch64.sve.trn2.nxv2f16(<vscale x 2 x half>, <vscale x 2 x half>) declare <vscale x 4 x half> @llvm.aarch64.sve.trn2.nxv4f16(<vscale x 4 x half>, <vscale x 4 x half>) declare <vscale x 8 x bfloat> @llvm.aarch64.sve.trn2.nxv8bf16(<vscale x 8 x bfloat>, <vscale x 8 x bfloat>) declare <vscale x 8 x half> @llvm.aarch64.sve.trn2.nxv8f16(<vscale x 8 x half>, <vscale x 8 x half>) declare <vscale x 2 x float> @llvm.aarch64.sve.trn2.nxv2f32(<vscale x 2 x float>, <vscale x 2 x float>) declare <vscale x 4 x float> @llvm.aarch64.sve.trn2.nxv4f32(<vscale x 4 x float>, <vscale x 4 x float>) declare <vscale x 2 x double> @llvm.aarch64.sve.trn2.nxv2f64(<vscale x 2 x double>, <vscale x 2 x double>) declare <vscale x 16 x i1> @llvm.aarch64.sve.uzp1.nxv16i1(<vscale x 16 x i1>, <vscale x 16 x i1>) declare <vscale x 8 x i1> @llvm.aarch64.sve.uzp1.nxv8i1(<vscale x 8 x i1>, <vscale x 8 x i1>) declare <vscale x 4 x i1> @llvm.aarch64.sve.uzp1.nxv4i1(<vscale x 4 x i1>, <vscale x 4 x i1>) declare <vscale x 2 x i1> @llvm.aarch64.sve.uzp1.nxv2i1(<vscale x 2 x i1>, <vscale x 2 x i1>) declare <vscale x 16 x i8> @llvm.aarch64.sve.uzp1.nxv16i8(<vscale x 16 x i8>, <vscale x 16 x i8>) declare <vscale x 8 x i16> @llvm.aarch64.sve.uzp1.nxv8i16(<vscale x 8 x i16>, <vscale x 8 x i16>) declare <vscale x 4 x i32> @llvm.aarch64.sve.uzp1.nxv4i32(<vscale x 4 x i32>, <vscale x 4 x i32>) declare <vscale x 2 x i64> @llvm.aarch64.sve.uzp1.nxv2i64(<vscale x 2 x i64>, <vscale x 2 x i64>) declare <vscale x 2 x half> @llvm.aarch64.sve.uzp1.nxv2f16(<vscale x 2 x half>, <vscale x 2 x half>) declare <vscale x 4 x half> @llvm.aarch64.sve.uzp1.nxv4f16(<vscale x 4 x half>, <vscale x 4 x half>) declare <vscale x 8 x bfloat> @llvm.aarch64.sve.uzp1.nxv8bf16(<vscale x 8 x bfloat>, <vscale x 8 x bfloat>) declare <vscale x 8 x half> @llvm.aarch64.sve.uzp1.nxv8f16(<vscale x 8 x half>, <vscale x 8 x half>) declare <vscale x 2 x float> @llvm.aarch64.sve.uzp1.nxv2f32(<vscale x 2 x float>, <vscale x 2 x float>) declare <vscale x 4 x float> @llvm.aarch64.sve.uzp1.nxv4f32(<vscale x 4 x float>, <vscale x 4 x float>) declare <vscale x 2 x double> @llvm.aarch64.sve.uzp1.nxv2f64(<vscale x 2 x double>, <vscale x 2 x double>) declare <vscale x 16 x i1> @llvm.aarch64.sve.uzp2.nxv16i1(<vscale x 16 x i1>, <vscale x 16 x i1>) declare <vscale x 8 x i1> @llvm.aarch64.sve.uzp2.nxv8i1(<vscale x 8 x i1>, <vscale x 8 x i1>) declare <vscale x 4 x i1> @llvm.aarch64.sve.uzp2.nxv4i1(<vscale x 4 x i1>, <vscale x 4 x i1>) declare <vscale x 2 x i1> @llvm.aarch64.sve.uzp2.nxv2i1(<vscale x 2 x i1>, <vscale x 2 x i1>) declare <vscale x 16 x i8> @llvm.aarch64.sve.uzp2.nxv16i8(<vscale x 16 x i8>, <vscale x 16 x i8>) declare <vscale x 8 x i16> @llvm.aarch64.sve.uzp2.nxv8i16(<vscale x 8 x i16>, <vscale x 8 x i16>) declare <vscale x 4 x i32> @llvm.aarch64.sve.uzp2.nxv4i32(<vscale x 4 x i32>, <vscale x 4 x i32>) declare <vscale x 2 x i64> @llvm.aarch64.sve.uzp2.nxv2i64(<vscale x 2 x i64>, <vscale x 2 x i64>) declare <vscale x 2 x half> @llvm.aarch64.sve.uzp2.nxv2f16(<vscale x 2 x half>, <vscale x 2 x half>) declare <vscale x 4 x half> @llvm.aarch64.sve.uzp2.nxv4f16(<vscale x 4 x half>, <vscale x 4 x half>) declare <vscale x 8 x bfloat> @llvm.aarch64.sve.uzp2.nxv8bf16(<vscale x 8 x bfloat>, <vscale x 8 x bfloat>) declare <vscale x 8 x half> @llvm.aarch64.sve.uzp2.nxv8f16(<vscale x 8 x half>, <vscale x 8 x half>) declare <vscale x 2 x float> @llvm.aarch64.sve.uzp2.nxv2f32(<vscale x 2 x float>, <vscale x 2 x float>) declare <vscale x 4 x float> @llvm.aarch64.sve.uzp2.nxv4f32(<vscale x 4 x float>, <vscale x 4 x float>) declare <vscale x 2 x double> @llvm.aarch64.sve.uzp2.nxv2f64(<vscale x 2 x double>, <vscale x 2 x double>) declare <vscale x 16 x i1> @llvm.aarch64.sve.zip1.nxv16i1(<vscale x 16 x i1>, <vscale x 16 x i1>) declare <vscale x 8 x i1> @llvm.aarch64.sve.zip1.nxv8i1(<vscale x 8 x i1>, <vscale x 8 x i1>) declare <vscale x 4 x i1> @llvm.aarch64.sve.zip1.nxv4i1(<vscale x 4 x i1>, <vscale x 4 x i1>) declare <vscale x 2 x i1> @llvm.aarch64.sve.zip1.nxv2i1(<vscale x 2 x i1>, <vscale x 2 x i1>) declare <vscale x 16 x i8> @llvm.aarch64.sve.zip1.nxv16i8(<vscale x 16 x i8>, <vscale x 16 x i8>) declare <vscale x 8 x i16> @llvm.aarch64.sve.zip1.nxv8i16(<vscale x 8 x i16>, <vscale x 8 x i16>) declare <vscale x 4 x i32> @llvm.aarch64.sve.zip1.nxv4i32(<vscale x 4 x i32>, <vscale x 4 x i32>) declare <vscale x 2 x i64> @llvm.aarch64.sve.zip1.nxv2i64(<vscale x 2 x i64>, <vscale x 2 x i64>) declare <vscale x 2 x half> @llvm.aarch64.sve.zip1.nxv2f16(<vscale x 2 x half>, <vscale x 2 x half>) declare <vscale x 4 x half> @llvm.aarch64.sve.zip1.nxv4f16(<vscale x 4 x half>, <vscale x 4 x half>) declare <vscale x 8 x bfloat> @llvm.aarch64.sve.zip1.nxv8bf16(<vscale x 8 x bfloat>, <vscale x 8 x bfloat>) declare <vscale x 8 x half> @llvm.aarch64.sve.zip1.nxv8f16(<vscale x 8 x half>, <vscale x 8 x half>) declare <vscale x 2 x float> @llvm.aarch64.sve.zip1.nxv2f32(<vscale x 2 x float>, <vscale x 2 x float>) declare <vscale x 4 x float> @llvm.aarch64.sve.zip1.nxv4f32(<vscale x 4 x float>, <vscale x 4 x float>) declare <vscale x 2 x double> @llvm.aarch64.sve.zip1.nxv2f64(<vscale x 2 x double>, <vscale x 2 x double>) declare <vscale x 16 x i1> @llvm.aarch64.sve.zip2.nxv16i1(<vscale x 16 x i1>, <vscale x 16 x i1>) declare <vscale x 8 x i1> @llvm.aarch64.sve.zip2.nxv8i1(<vscale x 8 x i1>, <vscale x 8 x i1>) declare <vscale x 4 x i1> @llvm.aarch64.sve.zip2.nxv4i1(<vscale x 4 x i1>, <vscale x 4 x i1>) declare <vscale x 2 x i1> @llvm.aarch64.sve.zip2.nxv2i1(<vscale x 2 x i1>, <vscale x 2 x i1>) declare <vscale x 16 x i8> @llvm.aarch64.sve.zip2.nxv16i8(<vscale x 16 x i8>, <vscale x 16 x i8>) declare <vscale x 8 x i16> @llvm.aarch64.sve.zip2.nxv8i16(<vscale x 8 x i16>, <vscale x 8 x i16>) declare <vscale x 4 x i32> @llvm.aarch64.sve.zip2.nxv4i32(<vscale x 4 x i32>, <vscale x 4 x i32>) declare <vscale x 2 x i64> @llvm.aarch64.sve.zip2.nxv2i64(<vscale x 2 x i64>, <vscale x 2 x i64>) declare <vscale x 2 x half> @llvm.aarch64.sve.zip2.nxv2f16(<vscale x 2 x half>, <vscale x 2 x half>) declare <vscale x 4 x half> @llvm.aarch64.sve.zip2.nxv4f16(<vscale x 4 x half>, <vscale x 4 x half>) declare <vscale x 8 x bfloat> @llvm.aarch64.sve.zip2.nxv8bf16(<vscale x 8 x bfloat>, <vscale x 8 x bfloat>) declare <vscale x 8 x half> @llvm.aarch64.sve.zip2.nxv8f16(<vscale x 8 x half>, <vscale x 8 x half>) declare <vscale x 2 x float> @llvm.aarch64.sve.zip2.nxv2f32(<vscale x 2 x float>, <vscale x 2 x float>) declare <vscale x 4 x float> @llvm.aarch64.sve.zip2.nxv4f32(<vscale x 4 x float>, <vscale x 4 x float>) declare <vscale x 2 x double> @llvm.aarch64.sve.zip2.nxv2f64(<vscale x 2 x double>, <vscale x 2 x double>) declare <vscale x 2 x double> @llvm.vector.insert.nxv2f64.v2f64(<vscale x 2 x double>, <2 x double>, i64) declare <vscale x 4 x float> @llvm.vector.insert.nxv4f32.v4f32(<vscale x 4 x float>, <4 x float>, i64) declare <vscale x 8 x half> @llvm.vector.insert.nxv8f16.v8f16(<vscale x 8 x half>, <8 x half>, i64) declare <vscale x 2 x i64> @llvm.vector.insert.nxv2i64.v2i64(<vscale x 2 x i64>, <2 x i64>, i64) declare <vscale x 4 x i32> @llvm.vector.insert.nxv4i32.v4i32(<vscale x 4 x i32>, <4 x i32>, i64) declare <vscale x 8 x i16> @llvm.vector.insert.nxv8i16.v8i16(<vscale x 8 x i16>, <8 x i16>, i64) declare <vscale x 16 x i8> @llvm.vector.insert.nxv16i8.v16i8(<vscale x 16 x i8>, <16 x i8>, i64) declare <vscale x 8 x bfloat> @llvm.vector.insert.nxv8bf16.v8bf16(<vscale x 8 x bfloat>, <8 x bfloat>, i64) ; +bf16 is required for the bfloat version. attributes #0 = { "target-features"="+sve,+bf16" }