162 lines
5.6 KiB
LLVM
162 lines
5.6 KiB
LLVM
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
|
|
; RUN: llc < %s -mtriple=i686-unknown-unknown -mattr=+avx512dq | FileCheck %s --check-prefixes=CHECK,X86
|
|
; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mattr=+avx512dq | FileCheck %s --check-prefixes=CHECK,X64
|
|
|
|
; NOTE: This should use IR equivalent to what is generated by clang/test/CodeGen/avx512dq-builtins.c
|
|
|
|
define zeroext i8 @test_mm512_mask_fpclass_pd_mask(i8 zeroext %__U, <8 x double> %__A) {
|
|
; X86-LABEL: test_mm512_mask_fpclass_pd_mask:
|
|
; X86: # %bb.0: # %entry
|
|
; X86-NEXT: vfpclasspd $4, %zmm0, %k0
|
|
; X86-NEXT: kmovw %k0, %eax
|
|
; X86-NEXT: andb {{[0-9]+}}(%esp), %al
|
|
; X86-NEXT: # kill: def $al killed $al killed $eax
|
|
; X86-NEXT: vzeroupper
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-LABEL: test_mm512_mask_fpclass_pd_mask:
|
|
; X64: # %bb.0: # %entry
|
|
; X64-NEXT: vfpclasspd $4, %zmm0, %k0
|
|
; X64-NEXT: kmovw %k0, %eax
|
|
; X64-NEXT: andb %dil, %al
|
|
; X64-NEXT: # kill: def $al killed $al killed $eax
|
|
; X64-NEXT: vzeroupper
|
|
; X64-NEXT: retq
|
|
entry:
|
|
%0 = tail call <8 x i1> @llvm.x86.avx512.fpclass.pd.512(<8 x double> %__A, i32 4)
|
|
%1 = bitcast i8 %__U to <8 x i1>
|
|
%2 = and <8 x i1> %0, %1
|
|
%3 = bitcast <8 x i1> %2 to i8
|
|
ret i8 %3
|
|
}
|
|
|
|
declare <8 x i1> @llvm.x86.avx512.fpclass.pd.512(<8 x double>, i32)
|
|
|
|
define zeroext i8 @test_mm512_fpclass_pd_mask(<8 x double> %__A) {
|
|
; CHECK-LABEL: test_mm512_fpclass_pd_mask:
|
|
; CHECK: # %bb.0: # %entry
|
|
; CHECK-NEXT: vfpclasspd $4, %zmm0, %k0
|
|
; CHECK-NEXT: kmovw %k0, %eax
|
|
; CHECK-NEXT: # kill: def $al killed $al killed $eax
|
|
; CHECK-NEXT: vzeroupper
|
|
; CHECK-NEXT: ret{{[l|q]}}
|
|
entry:
|
|
%0 = tail call <8 x i1> @llvm.x86.avx512.fpclass.pd.512(<8 x double> %__A, i32 4)
|
|
%1 = bitcast <8 x i1> %0 to i8
|
|
ret i8 %1
|
|
}
|
|
|
|
define zeroext i16 @test_mm512_mask_fpclass_ps_mask(i16 zeroext %__U, <16 x float> %__A) {
|
|
; X86-LABEL: test_mm512_mask_fpclass_ps_mask:
|
|
; X86: # %bb.0: # %entry
|
|
; X86-NEXT: vfpclassps $4, %zmm0, %k0
|
|
; X86-NEXT: kmovw %k0, %eax
|
|
; X86-NEXT: andw {{[0-9]+}}(%esp), %ax
|
|
; X86-NEXT: # kill: def $ax killed $ax killed $eax
|
|
; X86-NEXT: vzeroupper
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-LABEL: test_mm512_mask_fpclass_ps_mask:
|
|
; X64: # %bb.0: # %entry
|
|
; X64-NEXT: vfpclassps $4, %zmm0, %k0
|
|
; X64-NEXT: kmovw %k0, %eax
|
|
; X64-NEXT: andl %edi, %eax
|
|
; X64-NEXT: # kill: def $ax killed $ax killed $eax
|
|
; X64-NEXT: vzeroupper
|
|
; X64-NEXT: retq
|
|
entry:
|
|
%0 = tail call <16 x i1> @llvm.x86.avx512.fpclass.ps.512(<16 x float> %__A, i32 4)
|
|
%1 = bitcast i16 %__U to <16 x i1>
|
|
%2 = and <16 x i1> %0, %1
|
|
%3 = bitcast <16 x i1> %2 to i16
|
|
ret i16 %3
|
|
}
|
|
|
|
declare <16 x i1> @llvm.x86.avx512.fpclass.ps.512(<16 x float>, i32)
|
|
|
|
define zeroext i16 @test_mm512_fpclass_ps_mask(<16 x float> %__A) {
|
|
; CHECK-LABEL: test_mm512_fpclass_ps_mask:
|
|
; CHECK: # %bb.0: # %entry
|
|
; CHECK-NEXT: vfpclassps $4, %zmm0, %k0
|
|
; CHECK-NEXT: kmovw %k0, %eax
|
|
; CHECK-NEXT: # kill: def $ax killed $ax killed $eax
|
|
; CHECK-NEXT: vzeroupper
|
|
; CHECK-NEXT: ret{{[l|q]}}
|
|
entry:
|
|
%0 = tail call <16 x i1> @llvm.x86.avx512.fpclass.ps.512(<16 x float> %__A, i32 4)
|
|
%1 = bitcast <16 x i1> %0 to i16
|
|
ret i16 %1
|
|
}
|
|
|
|
define zeroext i8 @test_mm_fpclass_sd_mask(<4 x float> %__A) {
|
|
; CHECK-LABEL: test_mm_fpclass_sd_mask:
|
|
; CHECK: # %bb.0: # %entry
|
|
; CHECK-NEXT: vfpclasssd $2, %xmm0, %k0
|
|
; CHECK-NEXT: kmovw %k0, %eax
|
|
; CHECK-NEXT: # kill: def $al killed $al killed $eax
|
|
; CHECK-NEXT: ret{{[l|q]}}
|
|
entry:
|
|
%0 = bitcast <4 x float> %__A to <2 x double>
|
|
%1 = tail call i8 @llvm.x86.avx512.mask.fpclass.sd(<2 x double> %0, i32 2, i8 -1)
|
|
ret i8 %1
|
|
}
|
|
|
|
declare i8 @llvm.x86.avx512.mask.fpclass.sd(<2 x double>, i32, i8)
|
|
|
|
define zeroext i8 @test_mm_mask_fpclass_sd_mask(i8 zeroext %__U, <4 x float> %__A) {
|
|
; X86-LABEL: test_mm_mask_fpclass_sd_mask:
|
|
; X86: # %bb.0: # %entry
|
|
; X86-NEXT: kmovb {{[0-9]+}}(%esp), %k1
|
|
; X86-NEXT: vfpclasssd $2, %xmm0, %k0 {%k1}
|
|
; X86-NEXT: kmovw %k0, %eax
|
|
; X86-NEXT: # kill: def $al killed $al killed $eax
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-LABEL: test_mm_mask_fpclass_sd_mask:
|
|
; X64: # %bb.0: # %entry
|
|
; X64-NEXT: kmovw %edi, %k1
|
|
; X64-NEXT: vfpclasssd $2, %xmm0, %k0 {%k1}
|
|
; X64-NEXT: kmovw %k0, %eax
|
|
; X64-NEXT: # kill: def $al killed $al killed $eax
|
|
; X64-NEXT: retq
|
|
entry:
|
|
%0 = bitcast <4 x float> %__A to <2 x double>
|
|
%1 = tail call i8 @llvm.x86.avx512.mask.fpclass.sd(<2 x double> %0, i32 2, i8 %__U)
|
|
ret i8 %1
|
|
}
|
|
|
|
define zeroext i8 @test_mm_fpclass_ss_mask(<4 x float> %__A) {
|
|
; CHECK-LABEL: test_mm_fpclass_ss_mask:
|
|
; CHECK: # %bb.0: # %entry
|
|
; CHECK-NEXT: vfpclassss $2, %xmm0, %k0
|
|
; CHECK-NEXT: kmovw %k0, %eax
|
|
; CHECK-NEXT: # kill: def $al killed $al killed $eax
|
|
; CHECK-NEXT: ret{{[l|q]}}
|
|
entry:
|
|
%0 = tail call i8 @llvm.x86.avx512.mask.fpclass.ss(<4 x float> %__A, i32 2, i8 -1)
|
|
ret i8 %0
|
|
}
|
|
|
|
declare i8 @llvm.x86.avx512.mask.fpclass.ss(<4 x float>, i32, i8)
|
|
|
|
define zeroext i8 @test_mm_mask_fpclass_ss_mask(i8 zeroext %__U, <4 x float> %__A) {
|
|
; X86-LABEL: test_mm_mask_fpclass_ss_mask:
|
|
; X86: # %bb.0: # %entry
|
|
; X86-NEXT: kmovb {{[0-9]+}}(%esp), %k1
|
|
; X86-NEXT: vfpclassss $2, %xmm0, %k0 {%k1}
|
|
; X86-NEXT: kmovw %k0, %eax
|
|
; X86-NEXT: # kill: def $al killed $al killed $eax
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-LABEL: test_mm_mask_fpclass_ss_mask:
|
|
; X64: # %bb.0: # %entry
|
|
; X64-NEXT: kmovw %edi, %k1
|
|
; X64-NEXT: vfpclassss $2, %xmm0, %k0 {%k1}
|
|
; X64-NEXT: kmovw %k0, %eax
|
|
; X64-NEXT: # kill: def $al killed $al killed $eax
|
|
; X64-NEXT: retq
|
|
entry:
|
|
%0 = tail call i8 @llvm.x86.avx512.mask.fpclass.ss(<4 x float> %__A, i32 2, i8 %__U)
|
|
ret i8 %0
|
|
}
|