; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py ; RUN: llc < %s -mtriple=i686-unknown -mattr=+avx512bw,+avx512vl | FileCheck %s --check-prefixes=CHECK,X86 ; RUN: llc < %s -mtriple=x86_64-unknown -mattr=+avx512bw,+avx512vl | FileCheck %s --check-prefixes=CHECK,X64 declare <16 x i16> @llvm.x86.avx512.maskz.vpermt2var.hi.256(<16 x i16>, <16 x i16>, <16 x i16>, i16) declare <16 x i16> @llvm.x86.avx512.mask.vpermi2var.hi.256(<16 x i16>, <16 x i16>, <16 x i16>, i16) define <16 x i16> @combine_vpermt2var_16i16_identity(<16 x i16> %x0, <16 x i16> %x1) { ; CHECK-LABEL: combine_vpermt2var_16i16_identity: ; CHECK: # %bb.0: ; CHECK-NEXT: ret{{[l|q]}} %res0 = call <16 x i16> @llvm.x86.avx512.maskz.vpermt2var.hi.256(<16 x i16> , <16 x i16> %x0, <16 x i16> %x1, i16 -1) %res1 = call <16 x i16> @llvm.x86.avx512.maskz.vpermt2var.hi.256(<16 x i16> , <16 x i16> %res0, <16 x i16> %res0, i16 -1) ret <16 x i16> %res1 } define <16 x i16> @combine_vpermt2var_16i16_identity_mask(<16 x i16> %x0, <16 x i16> %x1, i16 %m) { ; X86-LABEL: combine_vpermt2var_16i16_identity_mask: ; X86: # %bb.0: ; X86-NEXT: vmovdqa {{.*#+}} ymm1 = [15,14,13,12,11,10,9,8,7,6,5,4,3,2,1,0] ; X86-NEXT: kmovw {{[0-9]+}}(%esp), %k1 ; X86-NEXT: vpermi2w %ymm0, %ymm0, %ymm1 {%k1} {z} ; X86-NEXT: vmovdqa {{.*#+}} ymm0 = [15,30,13,28,11,26,9,24,7,22,5,20,3,18,1,16] ; X86-NEXT: vpermi2w %ymm1, %ymm1, %ymm0 {%k1} {z} ; X86-NEXT: retl ; ; X64-LABEL: combine_vpermt2var_16i16_identity_mask: ; X64: # %bb.0: ; X64-NEXT: vmovdqa {{.*#+}} ymm1 = [15,14,13,12,11,10,9,8,7,6,5,4,3,2,1,0] ; X64-NEXT: kmovd %edi, %k1 ; X64-NEXT: vpermi2w %ymm0, %ymm0, %ymm1 {%k1} {z} ; X64-NEXT: vmovdqa {{.*#+}} ymm0 = [15,30,13,28,11,26,9,24,7,22,5,20,3,18,1,16] ; X64-NEXT: vpermi2w %ymm1, %ymm1, %ymm0 {%k1} {z} ; X64-NEXT: retq %res0 = call <16 x i16> @llvm.x86.avx512.maskz.vpermt2var.hi.256(<16 x i16> , <16 x i16> %x0, <16 x i16> %x1, i16 %m) %res1 = call <16 x i16> @llvm.x86.avx512.maskz.vpermt2var.hi.256(<16 x i16> , <16 x i16> %res0, <16 x i16> %res0, i16 %m) ret <16 x i16> %res1 } define <16 x i16> @combine_vpermi2var_16i16_as_permw(<16 x i16> %x0, <16 x i16> %x1) { ; CHECK-LABEL: combine_vpermi2var_16i16_as_permw: ; CHECK: # %bb.0: ; CHECK-NEXT: vmovdqa {{.*#+}} ymm1 = [15,0,14,1,13,2,12,3,11,4,10,5,9,6,8,7] ; CHECK-NEXT: vpermw %ymm0, %ymm1, %ymm0 ; CHECK-NEXT: ret{{[l|q]}} %res0 = call <16 x i16> @llvm.x86.avx512.mask.vpermi2var.hi.256(<16 x i16> %x0, <16 x i16> , <16 x i16> %x1, i16 -1) %res1 = call <16 x i16> @llvm.x86.avx512.mask.vpermi2var.hi.256(<16 x i16> %res0, <16 x i16> , <16 x i16> %res0, i16 -1) ret <16 x i16> %res1 } define <16 x i16> @combine_vpermt2var_vpermi2var_16i16_as_vperm2(<16 x i16> %x0, <16 x i16> %x1) { ; CHECK-LABEL: combine_vpermt2var_vpermi2var_16i16_as_vperm2: ; CHECK: # %bb.0: ; CHECK-NEXT: vmovdqa {{.*#+}} ymm2 = [0,31,2,2,4,29,6,27,8,25,10,23,12,21,14,19] ; CHECK-NEXT: vpermt2w %ymm1, %ymm2, %ymm0 ; CHECK-NEXT: ret{{[l|q]}} %res0 = call <16 x i16> @llvm.x86.avx512.mask.vpermi2var.hi.256(<16 x i16> %x0, <16 x i16> , <16 x i16> %x1, i16 -1) %res1 = call <16 x i16> @llvm.x86.avx512.maskz.vpermt2var.hi.256(<16 x i16> , <16 x i16> %res0, <16 x i16> %res0, i16 -1) ret <16 x i16> %res1 } define <16 x i16> @combine_vpermt2var_vpermi2var_16i16_as_unpckhwd(<16 x i16> %a0, <16 x i16> %a1) { ; CHECK-LABEL: combine_vpermt2var_vpermi2var_16i16_as_unpckhwd: ; CHECK: # %bb.0: ; CHECK-NEXT: vpunpckhwd {{.*#+}} ymm0 = ymm1[4],ymm0[4],ymm1[5],ymm0[5],ymm1[6],ymm0[6],ymm1[7],ymm0[7],ymm1[12],ymm0[12],ymm1[13],ymm0[13],ymm1[14],ymm0[14],ymm1[15],ymm0[15] ; CHECK-NEXT: ret{{[l|q]}} %res0 = call <16 x i16> @llvm.x86.avx512.mask.vpermi2var.hi.256(<16 x i16> %a0, <16 x i16> , <16 x i16> %a1, i16 -1) ret <16 x i16> %res0 } define <16 x i16> @combine_vpermt2var_vpermi2var_16i16_as_unpcklwd(<16 x i16> %a0, <16 x i16> %a1) { ; CHECK-LABEL: combine_vpermt2var_vpermi2var_16i16_as_unpcklwd: ; CHECK: # %bb.0: ; CHECK-NEXT: vpunpcklwd {{.*#+}} ymm0 = ymm0[0],ymm1[0],ymm0[1],ymm1[1],ymm0[2],ymm1[2],ymm0[3],ymm1[3],ymm0[8],ymm1[8],ymm0[9],ymm1[9],ymm0[10],ymm1[10],ymm0[11],ymm1[11] ; CHECK-NEXT: ret{{[l|q]}} %res0 = call <16 x i16> @llvm.x86.avx512.maskz.vpermt2var.hi.256(<16 x i16> , <16 x i16> %a0, <16 x i16> %a1, i16 -1) ret <16 x i16> %res0 } define <16 x i8> @combine_shuffle_vrotri_v2i64(<2 x i64> %a0) { ; CHECK-LABEL: combine_shuffle_vrotri_v2i64: ; CHECK: # %bb.0: ; CHECK-NEXT: vpshufb {{.*#+}} xmm0 = xmm0[13,12,11,10,9,8,15,14,5,4,3,2,1,0,7,6] ; CHECK-NEXT: ret{{[l|q]}} %1 = call <2 x i64> @llvm.fshr.v2i64(<2 x i64> %a0, <2 x i64> %a0, <2 x i64> ) %2 = bitcast <2 x i64> %1 to <16 x i8> %3 = shufflevector <16 x i8> %2, <16 x i8> undef, <16 x i32> ret <16 x i8> %3 } declare <2 x i64> @llvm.fshr.v2i64(<2 x i64>, <2 x i64>, <2 x i64>) define <16 x i8> @combine_shuffle_vrotli_v4i32(<4 x i32> %a0) { ; CHECK-LABEL: combine_shuffle_vrotli_v4i32: ; CHECK: # %bb.0: ; CHECK-NEXT: vpshufb {{.*#+}} xmm0 = xmm0[14,13,12,15,10,9,8,11,6,5,4,7,2,1,0,3] ; CHECK-NEXT: ret{{[l|q]}} %1 = call <4 x i32> @llvm.fshl.v4i32(<4 x i32> %a0, <4 x i32> %a0, <4 x i32> ) %2 = bitcast <4 x i32> %1 to <16 x i8> %3 = shufflevector <16 x i8> %2, <16 x i8> undef, <16 x i32> ret <16 x i8> %3 } declare <4 x i32> @llvm.fshl.v4i32(<4 x i32>, <4 x i32>, <4 x i32>) define void @PR46178(i16* %0) { ; X86-LABEL: PR46178: ; X86: # %bb.0: ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: vmovdqu 0, %ymm0 ; X86-NEXT: vmovdqu (%eax), %ymm1 ; X86-NEXT: vpmovqw %ymm0, %xmm0 ; X86-NEXT: vpmovqw %ymm1, %xmm1 ; X86-NEXT: vpsllw $8, %xmm1, %xmm1 ; X86-NEXT: vpsraw $8, %xmm1, %xmm1 ; X86-NEXT: vpsllw $8, %xmm0, %xmm0 ; X86-NEXT: vpsraw $8, %xmm0, %xmm0 ; X86-NEXT: vshufpd {{.*#+}} ymm0 = ymm0[0],ymm1[0],ymm0[2],ymm1[3] ; X86-NEXT: vmovupd %ymm0, (%eax) ; X86-NEXT: vzeroupper ; X86-NEXT: retl ; ; X64-LABEL: PR46178: ; X64: # %bb.0: ; X64-NEXT: vmovdqu 0, %ymm0 ; X64-NEXT: vmovdqu (%rax), %ymm1 ; X64-NEXT: vpmovqw %ymm0, %xmm0 ; X64-NEXT: vpmovqw %ymm1, %xmm1 ; X64-NEXT: vinserti128 $1, %xmm1, %ymm0, %ymm0 ; X64-NEXT: vpsllw $8, %ymm0, %ymm0 ; X64-NEXT: vpsraw $8, %ymm0, %ymm0 ; X64-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,2,2,3] ; X64-NEXT: vmovdqa %xmm0, %xmm0 ; X64-NEXT: vmovdqu %ymm0, (%rdi) ; X64-NEXT: vzeroupper ; X64-NEXT: retq %2 = load <4 x i64>, <4 x i64>* null, align 8 %3 = load <4 x i64>, <4 x i64>* undef, align 8 %4 = trunc <4 x i64> %2 to <4 x i16> %5 = trunc <4 x i64> %3 to <4 x i16> %6 = shl <4 x i16> %4, %7 = shl <4 x i16> %5, %8 = ashr exact <4 x i16> %6, %9 = ashr exact <4 x i16> %7, %10 = bitcast i16* %0 to <4 x i16>* %11 = getelementptr inbounds i16, i16* %0, i64 4 %12 = bitcast i16* %11 to <4 x i16>* %13 = getelementptr inbounds i16, i16* %0, i64 8 %14 = bitcast i16* %13 to <4 x i16>* %15 = getelementptr inbounds i16, i16* %0, i64 12 %16 = bitcast i16* %15 to <4 x i16>* store <4 x i16> %8, <4 x i16>* %10, align 2 store <4 x i16> %9, <4 x i16>* %12, align 2 store <4 x i16> zeroinitializer, <4 x i16>* %14, align 2 store <4 x i16> zeroinitializer, <4 x i16>* %16, align 2 ret void } define <8 x i32> @PR46393(<8 x i16> %a0, i8 %a1) { ; X86-LABEL: PR46393: ; X86: # %bb.0: ; X86-NEXT: vpmovzxwd {{.*#+}} ymm0 = xmm0[0],zero,xmm0[1],zero,xmm0[2],zero,xmm0[3],zero,xmm0[4],zero,xmm0[5],zero,xmm0[6],zero,xmm0[7],zero ; X86-NEXT: movzbl {{[0-9]+}}(%esp), %eax ; X86-NEXT: kmovd %eax, %k1 ; X86-NEXT: vpslld $16, %ymm0, %ymm0 {%k1} {z} ; X86-NEXT: retl ; ; X64-LABEL: PR46393: ; X64: # %bb.0: ; X64-NEXT: vpmovzxwd {{.*#+}} ymm0 = xmm0[0],zero,xmm0[1],zero,xmm0[2],zero,xmm0[3],zero,xmm0[4],zero,xmm0[5],zero,xmm0[6],zero,xmm0[7],zero ; X64-NEXT: kmovd %edi, %k1 ; X64-NEXT: vpslld $16, %ymm0, %ymm0 {%k1} {z} ; X64-NEXT: retq %zext = sext <8 x i16> %a0 to <8 x i32> %mask = bitcast i8 %a1 to <8 x i1> %shl = shl nuw <8 x i32> %zext, %sel = select <8 x i1> %mask, <8 x i32> %shl, <8 x i32> zeroinitializer ret <8 x i32> %sel }