; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py ; RUN: llc < %s -mtriple=x86_64-unknown-linux-gnu -mattr=+sse2 | FileCheck %s --check-prefix=SSE2 ; RUN: llc < %s -mtriple=x86_64-unknown-linux-gnu -mattr=+avx2 | FileCheck %s --check-prefix=AVX2 define void @trunc_shl_7_v4i32_v4i64(ptr addrspace(1) %out, ptr addrspace(1) %in) { ; SSE2-LABEL: trunc_shl_7_v4i32_v4i64: ; SSE2: # %bb.0: ; SSE2-NEXT: movaps (%rsi), %xmm0 ; SSE2-NEXT: shufps {{.*#+}} xmm0 = xmm0[0,2],mem[0,2] ; SSE2-NEXT: pslld $7, %xmm0 ; SSE2-NEXT: movdqa %xmm0, (%rdi) ; SSE2-NEXT: retq ; ; AVX2-LABEL: trunc_shl_7_v4i32_v4i64: ; AVX2: # %bb.0: ; AVX2-NEXT: vmovaps (%rsi), %xmm0 ; AVX2-NEXT: vshufps {{.*#+}} xmm0 = xmm0[0,2],mem[0,2] ; AVX2-NEXT: vpslld $7, %xmm0, %xmm0 ; AVX2-NEXT: vmovdqa %xmm0, (%rdi) ; AVX2-NEXT: retq %val = load <4 x i64>, ptr addrspace(1) %in %shl = shl <4 x i64> %val, <i64 7, i64 7, i64 7, i64 7> %trunc = trunc <4 x i64> %shl to <4 x i32> store <4 x i32> %trunc, ptr addrspace(1) %out ret void } define <8 x i16> @trunc_shl_15_v8i16_v8i32(<8 x i32> %a) { ; SSE2-LABEL: trunc_shl_15_v8i16_v8i32: ; SSE2: # %bb.0: ; SSE2-NEXT: pslld $16, %xmm1 ; SSE2-NEXT: psrad $16, %xmm1 ; SSE2-NEXT: pslld $16, %xmm0 ; SSE2-NEXT: psrad $16, %xmm0 ; SSE2-NEXT: packssdw %xmm1, %xmm0 ; SSE2-NEXT: psllw $15, %xmm0 ; SSE2-NEXT: retq ; ; AVX2-LABEL: trunc_shl_15_v8i16_v8i32: ; AVX2: # %bb.0: ; AVX2-NEXT: vpshufb {{.*#+}} ymm0 = ymm0[0,1,4,5,8,9,12,13,u,u,u,u,u,u,u,u,16,17,20,21,24,25,28,29,u,u,u,u,u,u,u,u] ; AVX2-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,2,2,3] ; AVX2-NEXT: vpsllw $15, %xmm0, %xmm0 ; AVX2-NEXT: vzeroupper ; AVX2-NEXT: retq %shl = shl <8 x i32> %a, <i32 15, i32 15, i32 15, i32 15, i32 15, i32 15, i32 15, i32 15> %conv = trunc <8 x i32> %shl to <8 x i16> ret <8 x i16> %conv } define <8 x i16> @trunc_shl_16_v8i16_v8i32(<8 x i32> %a) { ; SSE2-LABEL: trunc_shl_16_v8i16_v8i32: ; SSE2: # %bb.0: ; SSE2-NEXT: xorps %xmm0, %xmm0 ; SSE2-NEXT: retq ; ; AVX2-LABEL: trunc_shl_16_v8i16_v8i32: ; AVX2: # %bb.0: ; AVX2-NEXT: vxorps %xmm0, %xmm0, %xmm0 ; AVX2-NEXT: retq %shl = shl <8 x i32> %a, <i32 16, i32 16, i32 16, i32 16, i32 16, i32 16, i32 16, i32 16> %conv = trunc <8 x i32> %shl to <8 x i16> ret <8 x i16> %conv } define <8 x i16> @trunc_shl_17_v8i16_v8i32(<8 x i32> %a) { ; SSE2-LABEL: trunc_shl_17_v8i16_v8i32: ; SSE2: # %bb.0: ; SSE2-NEXT: xorps %xmm0, %xmm0 ; SSE2-NEXT: retq ; ; AVX2-LABEL: trunc_shl_17_v8i16_v8i32: ; AVX2: # %bb.0: ; AVX2-NEXT: vxorps %xmm0, %xmm0, %xmm0 ; AVX2-NEXT: retq %shl = shl <8 x i32> %a, <i32 17, i32 17, i32 17, i32 17, i32 17, i32 17, i32 17, i32 17> %conv = trunc <8 x i32> %shl to <8 x i16> ret <8 x i16> %conv } define void @trunc_shl_31_i32_i64(ptr %out, ptr %in) { ; SSE2-LABEL: trunc_shl_31_i32_i64: ; SSE2: # %bb.0: ; SSE2-NEXT: movl (%rsi), %eax ; SSE2-NEXT: shll $31, %eax ; SSE2-NEXT: movl %eax, (%rdi) ; SSE2-NEXT: retq ; ; AVX2-LABEL: trunc_shl_31_i32_i64: ; AVX2: # %bb.0: ; AVX2-NEXT: movl (%rsi), %eax ; AVX2-NEXT: shll $31, %eax ; AVX2-NEXT: movl %eax, (%rdi) ; AVX2-NEXT: retq %val = load i64, ptr %in %shl = shl i64 %val, 31 %trunc = trunc i64 %shl to i32 store i32 %trunc, ptr %out ret void } define void @trunc_shl_32_i32_i64(ptr %out, ptr %in) { ; SSE2-LABEL: trunc_shl_32_i32_i64: ; SSE2: # %bb.0: ; SSE2-NEXT: movl $0, (%rdi) ; SSE2-NEXT: retq ; ; AVX2-LABEL: trunc_shl_32_i32_i64: ; AVX2: # %bb.0: ; AVX2-NEXT: movl $0, (%rdi) ; AVX2-NEXT: retq %val = load i64, ptr %in %shl = shl i64 %val, 32 %trunc = trunc i64 %shl to i32 store i32 %trunc, ptr %out ret void } define void @trunc_shl_15_i16_i64(ptr %out, ptr %in) { ; SSE2-LABEL: trunc_shl_15_i16_i64: ; SSE2: # %bb.0: ; SSE2-NEXT: movl (%rsi), %eax ; SSE2-NEXT: shll $15, %eax ; SSE2-NEXT: movw %ax, (%rdi) ; SSE2-NEXT: retq ; ; AVX2-LABEL: trunc_shl_15_i16_i64: ; AVX2: # %bb.0: ; AVX2-NEXT: movl (%rsi), %eax ; AVX2-NEXT: shll $15, %eax ; AVX2-NEXT: movw %ax, (%rdi) ; AVX2-NEXT: retq %val = load i64, ptr %in %shl = shl i64 %val, 15 %trunc = trunc i64 %shl to i16 store i16 %trunc, ptr %out ret void } define void @trunc_shl_16_i16_i64(ptr %out, ptr %in) { ; SSE2-LABEL: trunc_shl_16_i16_i64: ; SSE2: # %bb.0: ; SSE2-NEXT: movw $0, (%rdi) ; SSE2-NEXT: retq ; ; AVX2-LABEL: trunc_shl_16_i16_i64: ; AVX2: # %bb.0: ; AVX2-NEXT: movw $0, (%rdi) ; AVX2-NEXT: retq %val = load i64, ptr %in %shl = shl i64 %val, 16 %trunc = trunc i64 %shl to i16 store i16 %trunc, ptr %out ret void } define void @trunc_shl_7_i8_i64(ptr %out, ptr %in) { ; SSE2-LABEL: trunc_shl_7_i8_i64: ; SSE2: # %bb.0: ; SSE2-NEXT: movzbl (%rsi), %eax ; SSE2-NEXT: shlb $7, %al ; SSE2-NEXT: movb %al, (%rdi) ; SSE2-NEXT: retq ; ; AVX2-LABEL: trunc_shl_7_i8_i64: ; AVX2: # %bb.0: ; AVX2-NEXT: movzbl (%rsi), %eax ; AVX2-NEXT: shlb $7, %al ; AVX2-NEXT: movb %al, (%rdi) ; AVX2-NEXT: retq %val = load i64, ptr %in %shl = shl i64 %val, 7 %trunc = trunc i64 %shl to i8 store i8 %trunc, ptr %out ret void } define void @trunc_shl_8_i8_i64(ptr %out, ptr %in) { ; SSE2-LABEL: trunc_shl_8_i8_i64: ; SSE2: # %bb.0: ; SSE2-NEXT: movb $0, (%rdi) ; SSE2-NEXT: retq ; ; AVX2-LABEL: trunc_shl_8_i8_i64: ; AVX2: # %bb.0: ; AVX2-NEXT: movb $0, (%rdi) ; AVX2-NEXT: retq %val = load i64, ptr %in %shl = shl i64 %val, 8 %trunc = trunc i64 %shl to i8 store i8 %trunc, ptr %out ret void }