; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py ; RUN: llc -mtriple=x86_64-linux -mcpu=skylake-avx512 < %s | FileCheck %s define void @scatter_scale_512(ptr %result, <4 x i64> %idx, <4 x i1> %mask) { ; CHECK-LABEL: scatter_scale_512: ; CHECK: # %bb.0: ; CHECK-NEXT: vpslld $31, %xmm1, %xmm1 ; CHECK-NEXT: vpmovd2m %xmm1, %k1 ; CHECK-NEXT: vpsllq $9, %ymm0, %ymm0 ; CHECK-NEXT: vpxor %xmm1, %xmm1, %xmm1 ; CHECK-NEXT: vscatterqpd %ymm1, (%rdi,%ymm0) {%k1} ; CHECK-NEXT: vzeroupper ; CHECK-NEXT: retq %gep = getelementptr inbounds [512 x i8], ptr %result, <4 x i64> %idx call void @llvm.masked.scatter.v4f64.v4p0(<4 x double> zeroinitializer, <4 x ptr> %gep, i32 0, <4 x i1> %mask) ret void } define void @scatter_scale_16(ptr %result, <4 x i64> %idx, <4 x i1> %mask) { ; CHECK-LABEL: scatter_scale_16: ; CHECK: # %bb.0: ; CHECK-NEXT: vpslld $31, %xmm1, %xmm1 ; CHECK-NEXT: vpmovd2m %xmm1, %k1 ; CHECK-NEXT: vpsllq $4, %ymm0, %ymm0 ; CHECK-NEXT: vpxor %xmm1, %xmm1, %xmm1 ; CHECK-NEXT: vscatterqpd %ymm1, (%rdi,%ymm0) {%k1} ; CHECK-NEXT: vzeroupper ; CHECK-NEXT: retq %gep = getelementptr inbounds [16 x i8], ptr %result, <4 x i64> %idx call void @llvm.masked.scatter.v4f64.v4p0(<4 x double> zeroinitializer, <4 x ptr> %gep, i32 0, <4 x i1> %mask) ret void } define void @scatter_scale_8(ptr %result, <4 x i64> %idx, <4 x i1> %mask) { ; CHECK-LABEL: scatter_scale_8: ; CHECK: # %bb.0: ; CHECK-NEXT: vpslld $31, %xmm1, %xmm1 ; CHECK-NEXT: vpmovd2m %xmm1, %k1 ; CHECK-NEXT: vpxor %xmm1, %xmm1, %xmm1 ; CHECK-NEXT: vscatterqpd %ymm1, (%rdi,%ymm0,8) {%k1} ; CHECK-NEXT: vzeroupper ; CHECK-NEXT: retq %gep = getelementptr inbounds [8 x i8], ptr %result, <4 x i64> %idx call void @llvm.masked.scatter.v4f64.v4p0(<4 x double> zeroinitializer, <4 x ptr> %gep, i32 0, <4 x i1> %mask) ret void } define void @scatter_scale_4(ptr %result, <4 x i64> %idx, <4 x i1> %mask) { ; CHECK-LABEL: scatter_scale_4: ; CHECK: # %bb.0: ; CHECK-NEXT: vpslld $31, %xmm1, %xmm1 ; CHECK-NEXT: vpmovd2m %xmm1, %k1 ; CHECK-NEXT: vpxor %xmm1, %xmm1, %xmm1 ; CHECK-NEXT: vscatterqpd %ymm1, (%rdi,%ymm0,4) {%k1} ; CHECK-NEXT: vzeroupper ; CHECK-NEXT: retq %gep = getelementptr inbounds [4 x i8], ptr %result, <4 x i64> %idx call void @llvm.masked.scatter.v4f64.v4p0(<4 x double> zeroinitializer, <4 x ptr> %gep, i32 0, <4 x i1> %mask) ret void } define void @scatter_scale_3(ptr %result, <4 x i64> %idx, <4 x i1> %mask) { ; CHECK-LABEL: scatter_scale_3: ; CHECK: # %bb.0: ; CHECK-NEXT: vpslld $31, %xmm1, %xmm1 ; CHECK-NEXT: vpmovd2m %xmm1, %k1 ; CHECK-NEXT: vpaddq %ymm0, %ymm0, %ymm1 ; CHECK-NEXT: vpaddq %ymm0, %ymm1, %ymm0 ; CHECK-NEXT: vpxor %xmm1, %xmm1, %xmm1 ; CHECK-NEXT: vscatterqpd %ymm1, (%rdi,%ymm0) {%k1} ; CHECK-NEXT: vzeroupper ; CHECK-NEXT: retq %gep = getelementptr inbounds [3 x i8], ptr %result, <4 x i64> %idx call void @llvm.masked.scatter.v4f64.v4p0(<4 x double> zeroinitializer, <4 x ptr> %gep, i32 0, <4 x i1> %mask) ret void } define void @scatter_scale_1(ptr %result, <4 x i64> %idx, <4 x i1> %mask) { ; CHECK-LABEL: scatter_scale_1: ; CHECK: # %bb.0: ; CHECK-NEXT: vpslld $31, %xmm1, %xmm1 ; CHECK-NEXT: vpmovd2m %xmm1, %k1 ; CHECK-NEXT: vpxor %xmm1, %xmm1, %xmm1 ; CHECK-NEXT: vscatterqpd %ymm1, (%rdi,%ymm0) {%k1} ; CHECK-NEXT: vzeroupper ; CHECK-NEXT: retq %gep = getelementptr inbounds [1 x i8], ptr %result, <4 x i64> %idx call void @llvm.masked.scatter.v4f64.v4p0(<4 x double> zeroinitializer, <4 x ptr> %gep, i32 0, <4 x i1> %mask) ret void } define <4 x double> @gather_scale_512(ptr %result, <4 x i64> %idx, <4 x i1> %mask) { ; CHECK-LABEL: gather_scale_512: ; CHECK: # %bb.0: ; CHECK-NEXT: vpslld $31, %xmm1, %xmm1 ; CHECK-NEXT: vpmovd2m %xmm1, %k1 ; CHECK-NEXT: vpsllq $9, %ymm0, %ymm1 ; CHECK-NEXT: vpxor %xmm0, %xmm0, %xmm0 ; CHECK-NEXT: vgatherqpd (%rdi,%ymm1), %ymm0 {%k1} ; CHECK-NEXT: retq %gep = getelementptr inbounds [512 x i8], ptr %result, <4 x i64> %idx %res = call <4 x double> @llvm.masked.gather.v4f64.v4p0(<4 x ptr> %gep, i32 0, <4 x i1> %mask, <4 x double> zeroinitializer) ret <4 x double> %res } define <4 x double> @gather_scale_16(ptr %result, <4 x i64> %idx, <4 x i1> %mask) { ; CHECK-LABEL: gather_scale_16: ; CHECK: # %bb.0: ; CHECK-NEXT: vpslld $31, %xmm1, %xmm1 ; CHECK-NEXT: vpmovd2m %xmm1, %k1 ; CHECK-NEXT: vpsllq $4, %ymm0, %ymm1 ; CHECK-NEXT: vpxor %xmm0, %xmm0, %xmm0 ; CHECK-NEXT: vgatherqpd (%rdi,%ymm1), %ymm0 {%k1} ; CHECK-NEXT: retq %gep = getelementptr inbounds [16 x i8], ptr %result, <4 x i64> %idx %res = call <4 x double> @llvm.masked.gather.v4f64.v4p0(<4 x ptr> %gep, i32 0, <4 x i1> %mask, <4 x double> zeroinitializer) ret <4 x double> %res } define <4 x double> @gather_scale_8(ptr %result, <4 x i64> %idx, <4 x i1> %mask) { ; CHECK-LABEL: gather_scale_8: ; CHECK: # %bb.0: ; CHECK-NEXT: vpslld $31, %xmm1, %xmm1 ; CHECK-NEXT: vpmovd2m %xmm1, %k1 ; CHECK-NEXT: vpxor %xmm1, %xmm1, %xmm1 ; CHECK-NEXT: vgatherqpd (%rdi,%ymm0,8), %ymm1 {%k1} ; CHECK-NEXT: vmovapd %ymm1, %ymm0 ; CHECK-NEXT: retq %gep = getelementptr inbounds [8 x i8], ptr %result, <4 x i64> %idx %res = call <4 x double> @llvm.masked.gather.v4f64.v4p0(<4 x ptr> %gep, i32 0, <4 x i1> %mask, <4 x double> zeroinitializer) ret <4 x double> %res } define <4 x double> @gather_scale_4(ptr %result, <4 x i64> %idx, <4 x i1> %mask) { ; CHECK-LABEL: gather_scale_4: ; CHECK: # %bb.0: ; CHECK-NEXT: vpslld $31, %xmm1, %xmm1 ; CHECK-NEXT: vpmovd2m %xmm1, %k1 ; CHECK-NEXT: vpxor %xmm1, %xmm1, %xmm1 ; CHECK-NEXT: vgatherqpd (%rdi,%ymm0,4), %ymm1 {%k1} ; CHECK-NEXT: vmovapd %ymm1, %ymm0 ; CHECK-NEXT: retq %gep = getelementptr inbounds [4 x i8], ptr %result, <4 x i64> %idx %res = call <4 x double> @llvm.masked.gather.v4f64.v4p0(<4 x ptr> %gep, i32 0, <4 x i1> %mask, <4 x double> zeroinitializer) ret <4 x double> %res } define <4 x double> @gather_scale_3(ptr %result, <4 x i64> %idx, <4 x i1> %mask) { ; CHECK-LABEL: gather_scale_3: ; CHECK: # %bb.0: ; CHECK-NEXT: vpslld $31, %xmm1, %xmm1 ; CHECK-NEXT: vpmovd2m %xmm1, %k1 ; CHECK-NEXT: vpaddq %ymm0, %ymm0, %ymm1 ; CHECK-NEXT: vpaddq %ymm0, %ymm1, %ymm1 ; CHECK-NEXT: vpxor %xmm0, %xmm0, %xmm0 ; CHECK-NEXT: vgatherqpd (%rdi,%ymm1), %ymm0 {%k1} ; CHECK-NEXT: retq %gep = getelementptr inbounds [3 x i8], ptr %result, <4 x i64> %idx %res = call <4 x double> @llvm.masked.gather.v4f64.v4p0(<4 x ptr> %gep, i32 0, <4 x i1> %mask, <4 x double> zeroinitializer) ret <4 x double> %res } define <4 x double> @gather_scale_1(ptr %result, <4 x i64> %idx, <4 x i1> %mask) { ; CHECK-LABEL: gather_scale_1: ; CHECK: # %bb.0: ; CHECK-NEXT: vpslld $31, %xmm1, %xmm1 ; CHECK-NEXT: vpmovd2m %xmm1, %k1 ; CHECK-NEXT: vpxor %xmm1, %xmm1, %xmm1 ; CHECK-NEXT: vgatherqpd (%rdi,%ymm0), %ymm1 {%k1} ; CHECK-NEXT: vmovapd %ymm1, %ymm0 ; CHECK-NEXT: retq %gep = getelementptr inbounds [1 x i8], ptr %result, <4 x i64> %idx %res = call <4 x double> @llvm.masked.gather.v4f64.v4p0(<4 x ptr> %gep, i32 0, <4 x i1> %mask, <4 x double> zeroinitializer) ret <4 x double> %res } declare void @llvm.masked.scatter.v4f64.v4p0(<4 x double>, <4 x ptr>, i32 immarg, <4 x i1>) declare <4 x double> @llvm.masked.gather.v4f64.v4p0(<4 x ptr>, i32 immarg, <4 x i1>, <4 x double>)