; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py ; RUN: llc -mtriple=riscv32 -verify-machineinstrs < %s \ ; RUN: | FileCheck %s -check-prefix=RV32I ; RUN: llc -mtriple=riscv64 -verify-machineinstrs < %s \ ; RUN: | FileCheck %s -check-prefix=RV64I ; RUN: llc -mtriple=riscv32 -verify-machineinstrs -mattr=+f \ ; RUN: -target-abi ilp32f < %s | FileCheck %s -check-prefix=RV32IF ; RUN: llc -mtriple=riscv32 -verify-machineinstrs -mattr=+f -mattr=+d \ ; RUN: -target-abi ilp32d < %s | FileCheck %s -check-prefix=RV32IFD ; RUN: llc -mtriple=riscv64 -verify-machineinstrs -mattr=+f -mattr=+d \ ; RUN: -target-abi lp64d < %s | FileCheck %s -check-prefix=RV64IFD ; RUN: llc -mtriple=riscv32 -verify-machineinstrs -mattr=+f \ ; RUN: -mattr=+experimental-zfh -target-abi ilp32f < %s \ ; RUN: | FileCheck %s -check-prefix=RV32IFZFH ; RUN: llc -mtriple=riscv32 -verify-machineinstrs -mattr=+f -mattr=+d \ ; RUN: -mattr=+experimental-zfh -target-abi ilp32d < %s \ ; RUN: | FileCheck %s -check-prefix=RV32IFDZFH ; RUN: llc -mtriple=riscv64 -verify-machineinstrs -mattr=+f -mattr=+d \ ; RUN: -mattr=+experimental-zfh -target-abi lp64d < %s \ ; RUN: | FileCheck %s -check-prefix=RV64IFDZFH ; Test fcopysign scenarios where the sign argument is casted to the type of the ; magnitude argument. Those casts can be folded away by the DAGCombiner. declare double @llvm.copysign.f64(double, double) declare float @llvm.copysign.f32(float, float) declare half @llvm.copysign.f16(half, half) define double @fold_promote_d_s(double %a, float %b) nounwind { ; RV32I-LABEL: fold_promote_d_s: ; RV32I: # %bb.0: ; RV32I-NEXT: lui a3, 524288 ; RV32I-NEXT: and a2, a2, a3 ; RV32I-NEXT: addi a3, a3, -1 ; RV32I-NEXT: and a1, a1, a3 ; RV32I-NEXT: or a1, a1, a2 ; RV32I-NEXT: ret ; ; RV64I-LABEL: fold_promote_d_s: ; RV64I: # %bb.0: ; RV64I-NEXT: addi a2, zero, -1 ; RV64I-NEXT: slli a2, a2, 63 ; RV64I-NEXT: addi a2, a2, -1 ; RV64I-NEXT: and a0, a0, a2 ; RV64I-NEXT: lui a2, 524288 ; RV64I-NEXT: and a1, a1, a2 ; RV64I-NEXT: slli a1, a1, 32 ; RV64I-NEXT: or a0, a0, a1 ; RV64I-NEXT: ret ; ; RV32IF-LABEL: fold_promote_d_s: ; RV32IF: # %bb.0: ; RV32IF-NEXT: fmv.x.w a2, fa0 ; RV32IF-NEXT: lui a3, 524288 ; RV32IF-NEXT: and a2, a2, a3 ; RV32IF-NEXT: addi a3, a3, -1 ; RV32IF-NEXT: and a1, a1, a3 ; RV32IF-NEXT: or a1, a1, a2 ; RV32IF-NEXT: ret ; ; RV32IFD-LABEL: fold_promote_d_s: ; RV32IFD: # %bb.0: ; RV32IFD-NEXT: fcvt.d.s ft0, fa1 ; RV32IFD-NEXT: fsgnj.d fa0, fa0, ft0 ; RV32IFD-NEXT: ret ; ; RV64IFD-LABEL: fold_promote_d_s: ; RV64IFD: # %bb.0: ; RV64IFD-NEXT: fcvt.d.s ft0, fa1 ; RV64IFD-NEXT: fsgnj.d fa0, fa0, ft0 ; RV64IFD-NEXT: ret ; ; RV32IFZFH-LABEL: fold_promote_d_s: ; RV32IFZFH: # %bb.0: ; RV32IFZFH-NEXT: fmv.x.w a2, fa0 ; RV32IFZFH-NEXT: lui a3, 524288 ; RV32IFZFH-NEXT: and a2, a2, a3 ; RV32IFZFH-NEXT: addi a3, a3, -1 ; RV32IFZFH-NEXT: and a1, a1, a3 ; RV32IFZFH-NEXT: or a1, a1, a2 ; RV32IFZFH-NEXT: ret ; ; RV32IFDZFH-LABEL: fold_promote_d_s: ; RV32IFDZFH: # %bb.0: ; RV32IFDZFH-NEXT: fcvt.d.s ft0, fa1 ; RV32IFDZFH-NEXT: fsgnj.d fa0, fa0, ft0 ; RV32IFDZFH-NEXT: ret ; ; RV64IFDZFH-LABEL: fold_promote_d_s: ; RV64IFDZFH: # %bb.0: ; RV64IFDZFH-NEXT: fcvt.d.s ft0, fa1 ; RV64IFDZFH-NEXT: fsgnj.d fa0, fa0, ft0 ; RV64IFDZFH-NEXT: ret %c = fpext float %b to double %t = call double @llvm.copysign.f64(double %a, double %c) ret double %t } define double @fold_promote_d_h(double %a, half %b) nounwind { ; RV32I-LABEL: fold_promote_d_h: ; RV32I: # %bb.0: ; RV32I-NEXT: lui a3, 524288 ; RV32I-NEXT: addi a3, a3, -1 ; RV32I-NEXT: and a1, a1, a3 ; RV32I-NEXT: lui a3, 8 ; RV32I-NEXT: and a2, a2, a3 ; RV32I-NEXT: slli a2, a2, 16 ; RV32I-NEXT: or a1, a1, a2 ; RV32I-NEXT: ret ; ; RV64I-LABEL: fold_promote_d_h: ; RV64I: # %bb.0: ; RV64I-NEXT: addi a2, zero, -1 ; RV64I-NEXT: slli a2, a2, 63 ; RV64I-NEXT: addi a2, a2, -1 ; RV64I-NEXT: and a0, a0, a2 ; RV64I-NEXT: lui a2, 8 ; RV64I-NEXT: and a1, a1, a2 ; RV64I-NEXT: slli a1, a1, 48 ; RV64I-NEXT: or a0, a0, a1 ; RV64I-NEXT: ret ; ; RV32IF-LABEL: fold_promote_d_h: ; RV32IF: # %bb.0: ; RV32IF-NEXT: fmv.x.w a2, fa0 ; RV32IF-NEXT: lui a3, 524288 ; RV32IF-NEXT: and a2, a2, a3 ; RV32IF-NEXT: addi a3, a3, -1 ; RV32IF-NEXT: and a1, a1, a3 ; RV32IF-NEXT: or a1, a1, a2 ; RV32IF-NEXT: ret ; ; RV32IFD-LABEL: fold_promote_d_h: ; RV32IFD: # %bb.0: ; RV32IFD-NEXT: fcvt.d.s ft0, fa1 ; RV32IFD-NEXT: fsgnj.d fa0, fa0, ft0 ; RV32IFD-NEXT: ret ; ; RV64IFD-LABEL: fold_promote_d_h: ; RV64IFD: # %bb.0: ; RV64IFD-NEXT: fcvt.d.s ft0, fa1 ; RV64IFD-NEXT: fsgnj.d fa0, fa0, ft0 ; RV64IFD-NEXT: ret ; ; RV32IFZFH-LABEL: fold_promote_d_h: ; RV32IFZFH: # %bb.0: ; RV32IFZFH-NEXT: fmv.x.h a2, fa0 ; RV32IFZFH-NEXT: lui a3, 524288 ; RV32IFZFH-NEXT: addi a3, a3, -1 ; RV32IFZFH-NEXT: and a1, a1, a3 ; RV32IFZFH-NEXT: lui a3, 8 ; RV32IFZFH-NEXT: and a2, a2, a3 ; RV32IFZFH-NEXT: slli a2, a2, 16 ; RV32IFZFH-NEXT: or a1, a1, a2 ; RV32IFZFH-NEXT: ret ; ; RV32IFDZFH-LABEL: fold_promote_d_h: ; RV32IFDZFH: # %bb.0: ; RV32IFDZFH-NEXT: fcvt.d.h ft0, fa1 ; RV32IFDZFH-NEXT: fsgnj.d fa0, fa0, ft0 ; RV32IFDZFH-NEXT: ret ; ; RV64IFDZFH-LABEL: fold_promote_d_h: ; RV64IFDZFH: # %bb.0: ; RV64IFDZFH-NEXT: fcvt.d.h ft0, fa1 ; RV64IFDZFH-NEXT: fsgnj.d fa0, fa0, ft0 ; RV64IFDZFH-NEXT: ret %c = fpext half %b to double %t = call double @llvm.copysign.f64(double %a, double %c) ret double %t } define float @fold_promote_f_h(float %a, half %b) nounwind { ; RV32I-LABEL: fold_promote_f_h: ; RV32I: # %bb.0: ; RV32I-NEXT: lui a2, 524288 ; RV32I-NEXT: addi a2, a2, -1 ; RV32I-NEXT: and a0, a0, a2 ; RV32I-NEXT: lui a2, 8 ; RV32I-NEXT: and a1, a1, a2 ; RV32I-NEXT: slli a1, a1, 16 ; RV32I-NEXT: or a0, a0, a1 ; RV32I-NEXT: ret ; ; RV64I-LABEL: fold_promote_f_h: ; RV64I: # %bb.0: ; RV64I-NEXT: lui a2, 524288 ; RV64I-NEXT: addiw a2, a2, -1 ; RV64I-NEXT: and a0, a0, a2 ; RV64I-NEXT: lui a2, 1048568 ; RV64I-NEXT: and a1, a1, a2 ; RV64I-NEXT: slli a1, a1, 16 ; RV64I-NEXT: or a0, a0, a1 ; RV64I-NEXT: ret ; ; RV32IF-LABEL: fold_promote_f_h: ; RV32IF: # %bb.0: ; RV32IF-NEXT: fsgnj.s fa0, fa0, fa1 ; RV32IF-NEXT: ret ; ; RV32IFD-LABEL: fold_promote_f_h: ; RV32IFD: # %bb.0: ; RV32IFD-NEXT: fsgnj.s fa0, fa0, fa1 ; RV32IFD-NEXT: ret ; ; RV64IFD-LABEL: fold_promote_f_h: ; RV64IFD: # %bb.0: ; RV64IFD-NEXT: fsgnj.s fa0, fa0, fa1 ; RV64IFD-NEXT: ret ; ; RV32IFZFH-LABEL: fold_promote_f_h: ; RV32IFZFH: # %bb.0: ; RV32IFZFH-NEXT: fcvt.s.h ft0, fa1 ; RV32IFZFH-NEXT: fsgnj.s fa0, fa0, ft0 ; RV32IFZFH-NEXT: ret ; ; RV32IFDZFH-LABEL: fold_promote_f_h: ; RV32IFDZFH: # %bb.0: ; RV32IFDZFH-NEXT: fcvt.s.h ft0, fa1 ; RV32IFDZFH-NEXT: fsgnj.s fa0, fa0, ft0 ; RV32IFDZFH-NEXT: ret ; ; RV64IFDZFH-LABEL: fold_promote_f_h: ; RV64IFDZFH: # %bb.0: ; RV64IFDZFH-NEXT: fcvt.s.h ft0, fa1 ; RV64IFDZFH-NEXT: fsgnj.s fa0, fa0, ft0 ; RV64IFDZFH-NEXT: ret %c = fpext half %b to float %t = call float @llvm.copysign.f32(float %a, float %c) ret float %t } define float @fold_demote_s_d(float %a, double %b) nounwind { ; RV32I-LABEL: fold_demote_s_d: ; RV32I: # %bb.0: ; RV32I-NEXT: lui a1, 524288 ; RV32I-NEXT: and a2, a2, a1 ; RV32I-NEXT: addi a1, a1, -1 ; RV32I-NEXT: and a0, a0, a1 ; RV32I-NEXT: or a0, a0, a2 ; RV32I-NEXT: ret ; ; RV64I-LABEL: fold_demote_s_d: ; RV64I: # %bb.0: ; RV64I-NEXT: lui a2, 524288 ; RV64I-NEXT: addiw a2, a2, -1 ; RV64I-NEXT: and a0, a0, a2 ; RV64I-NEXT: addi a2, zero, -1 ; RV64I-NEXT: slli a2, a2, 63 ; RV64I-NEXT: and a1, a1, a2 ; RV64I-NEXT: srli a1, a1, 32 ; RV64I-NEXT: or a0, a0, a1 ; RV64I-NEXT: ret ; ; RV32IF-LABEL: fold_demote_s_d: ; RV32IF: # %bb.0: ; RV32IF-NEXT: fmv.w.x ft0, a1 ; RV32IF-NEXT: fsgnj.s fa0, fa0, ft0 ; RV32IF-NEXT: ret ; ; RV32IFD-LABEL: fold_demote_s_d: ; RV32IFD: # %bb.0: ; RV32IFD-NEXT: fcvt.s.d ft0, fa1 ; RV32IFD-NEXT: fsgnj.s fa0, fa0, ft0 ; RV32IFD-NEXT: ret ; ; RV64IFD-LABEL: fold_demote_s_d: ; RV64IFD: # %bb.0: ; RV64IFD-NEXT: fcvt.s.d ft0, fa1 ; RV64IFD-NEXT: fsgnj.s fa0, fa0, ft0 ; RV64IFD-NEXT: ret ; ; RV32IFZFH-LABEL: fold_demote_s_d: ; RV32IFZFH: # %bb.0: ; RV32IFZFH-NEXT: fmv.w.x ft0, a1 ; RV32IFZFH-NEXT: fsgnj.s fa0, fa0, ft0 ; RV32IFZFH-NEXT: ret ; ; RV32IFDZFH-LABEL: fold_demote_s_d: ; RV32IFDZFH: # %bb.0: ; RV32IFDZFH-NEXT: fcvt.s.d ft0, fa1 ; RV32IFDZFH-NEXT: fsgnj.s fa0, fa0, ft0 ; RV32IFDZFH-NEXT: ret ; ; RV64IFDZFH-LABEL: fold_demote_s_d: ; RV64IFDZFH: # %bb.0: ; RV64IFDZFH-NEXT: fcvt.s.d ft0, fa1 ; RV64IFDZFH-NEXT: fsgnj.s fa0, fa0, ft0 ; RV64IFDZFH-NEXT: ret %c = fptrunc double %b to float %t = call float @llvm.copysign.f32(float %a, float %c) ret float %t } define half @fold_demote_h_s(half %a, float %b) nounwind { ; RV32I-LABEL: fold_demote_h_s: ; RV32I: # %bb.0: ; RV32I-NEXT: addi sp, sp, -16 ; RV32I-NEXT: sw ra, 12(sp) # 4-byte Folded Spill ; RV32I-NEXT: sw s0, 8(sp) # 4-byte Folded Spill ; RV32I-NEXT: mv s0, a1 ; RV32I-NEXT: lui a1, 16 ; RV32I-NEXT: addi a1, a1, -1 ; RV32I-NEXT: and a0, a0, a1 ; RV32I-NEXT: call __gnu_h2f_ieee@plt ; RV32I-NEXT: lui a1, 524288 ; RV32I-NEXT: and a2, s0, a1 ; RV32I-NEXT: addi a1, a1, -1 ; RV32I-NEXT: and a0, a0, a1 ; RV32I-NEXT: or a0, a0, a2 ; RV32I-NEXT: call __gnu_f2h_ieee@plt ; RV32I-NEXT: lw s0, 8(sp) # 4-byte Folded Reload ; RV32I-NEXT: lw ra, 12(sp) # 4-byte Folded Reload ; RV32I-NEXT: addi sp, sp, 16 ; RV32I-NEXT: ret ; ; RV64I-LABEL: fold_demote_h_s: ; RV64I: # %bb.0: ; RV64I-NEXT: addi sp, sp, -16 ; RV64I-NEXT: sd ra, 8(sp) # 8-byte Folded Spill ; RV64I-NEXT: sd s0, 0(sp) # 8-byte Folded Spill ; RV64I-NEXT: mv s0, a1 ; RV64I-NEXT: lui a1, 16 ; RV64I-NEXT: addiw a1, a1, -1 ; RV64I-NEXT: and a0, a0, a1 ; RV64I-NEXT: call __gnu_h2f_ieee@plt ; RV64I-NEXT: lui a1, 524288 ; RV64I-NEXT: and a2, s0, a1 ; RV64I-NEXT: addiw a1, a1, -1 ; RV64I-NEXT: and a0, a0, a1 ; RV64I-NEXT: or a0, a0, a2 ; RV64I-NEXT: call __gnu_f2h_ieee@plt ; RV64I-NEXT: ld s0, 0(sp) # 8-byte Folded Reload ; RV64I-NEXT: ld ra, 8(sp) # 8-byte Folded Reload ; RV64I-NEXT: addi sp, sp, 16 ; RV64I-NEXT: ret ; ; RV32IF-LABEL: fold_demote_h_s: ; RV32IF: # %bb.0: ; RV32IF-NEXT: addi sp, sp, -16 ; RV32IF-NEXT: sw ra, 12(sp) # 4-byte Folded Spill ; RV32IF-NEXT: fsw fs0, 8(sp) # 4-byte Folded Spill ; RV32IF-NEXT: fmv.s fs0, fa1 ; RV32IF-NEXT: call __gnu_f2h_ieee@plt ; RV32IF-NEXT: call __gnu_h2f_ieee@plt ; RV32IF-NEXT: fsgnj.s fa0, fa0, fs0 ; RV32IF-NEXT: flw fs0, 8(sp) # 4-byte Folded Reload ; RV32IF-NEXT: lw ra, 12(sp) # 4-byte Folded Reload ; RV32IF-NEXT: addi sp, sp, 16 ; RV32IF-NEXT: ret ; ; RV32IFD-LABEL: fold_demote_h_s: ; RV32IFD: # %bb.0: ; RV32IFD-NEXT: addi sp, sp, -16 ; RV32IFD-NEXT: sw ra, 12(sp) # 4-byte Folded Spill ; RV32IFD-NEXT: fsd fs0, 0(sp) # 8-byte Folded Spill ; RV32IFD-NEXT: fmv.s fs0, fa1 ; RV32IFD-NEXT: call __gnu_f2h_ieee@plt ; RV32IFD-NEXT: call __gnu_h2f_ieee@plt ; RV32IFD-NEXT: fsgnj.s fa0, fa0, fs0 ; RV32IFD-NEXT: fld fs0, 0(sp) # 8-byte Folded Reload ; RV32IFD-NEXT: lw ra, 12(sp) # 4-byte Folded Reload ; RV32IFD-NEXT: addi sp, sp, 16 ; RV32IFD-NEXT: ret ; ; RV64IFD-LABEL: fold_demote_h_s: ; RV64IFD: # %bb.0: ; RV64IFD-NEXT: addi sp, sp, -16 ; RV64IFD-NEXT: sd ra, 8(sp) # 8-byte Folded Spill ; RV64IFD-NEXT: fsd fs0, 0(sp) # 8-byte Folded Spill ; RV64IFD-NEXT: fmv.s fs0, fa1 ; RV64IFD-NEXT: call __gnu_f2h_ieee@plt ; RV64IFD-NEXT: call __gnu_h2f_ieee@plt ; RV64IFD-NEXT: fsgnj.s fa0, fa0, fs0 ; RV64IFD-NEXT: fld fs0, 0(sp) # 8-byte Folded Reload ; RV64IFD-NEXT: ld ra, 8(sp) # 8-byte Folded Reload ; RV64IFD-NEXT: addi sp, sp, 16 ; RV64IFD-NEXT: ret ; ; RV32IFZFH-LABEL: fold_demote_h_s: ; RV32IFZFH: # %bb.0: ; RV32IFZFH-NEXT: fcvt.h.s ft0, fa1 ; RV32IFZFH-NEXT: fsgnj.h fa0, fa0, ft0 ; RV32IFZFH-NEXT: ret ; ; RV32IFDZFH-LABEL: fold_demote_h_s: ; RV32IFDZFH: # %bb.0: ; RV32IFDZFH-NEXT: fcvt.h.s ft0, fa1 ; RV32IFDZFH-NEXT: fsgnj.h fa0, fa0, ft0 ; RV32IFDZFH-NEXT: ret ; ; RV64IFDZFH-LABEL: fold_demote_h_s: ; RV64IFDZFH: # %bb.0: ; RV64IFDZFH-NEXT: fcvt.h.s ft0, fa1 ; RV64IFDZFH-NEXT: fsgnj.h fa0, fa0, ft0 ; RV64IFDZFH-NEXT: ret %c = fptrunc float %b to half %t = call half @llvm.copysign.f16(half %a, half %c) ret half %t } define half @fold_demote_h_d(half %a, double %b) nounwind { ; RV32I-LABEL: fold_demote_h_d: ; RV32I: # %bb.0: ; RV32I-NEXT: addi sp, sp, -16 ; RV32I-NEXT: sw ra, 12(sp) # 4-byte Folded Spill ; RV32I-NEXT: sw s0, 8(sp) # 4-byte Folded Spill ; RV32I-NEXT: mv s0, a2 ; RV32I-NEXT: lui a1, 16 ; RV32I-NEXT: addi a1, a1, -1 ; RV32I-NEXT: and a0, a0, a1 ; RV32I-NEXT: call __gnu_h2f_ieee@plt ; RV32I-NEXT: lui a1, 524288 ; RV32I-NEXT: and a2, s0, a1 ; RV32I-NEXT: addi a1, a1, -1 ; RV32I-NEXT: and a0, a0, a1 ; RV32I-NEXT: or a0, a0, a2 ; RV32I-NEXT: call __gnu_f2h_ieee@plt ; RV32I-NEXT: lw s0, 8(sp) # 4-byte Folded Reload ; RV32I-NEXT: lw ra, 12(sp) # 4-byte Folded Reload ; RV32I-NEXT: addi sp, sp, 16 ; RV32I-NEXT: ret ; ; RV64I-LABEL: fold_demote_h_d: ; RV64I: # %bb.0: ; RV64I-NEXT: addi sp, sp, -16 ; RV64I-NEXT: sd ra, 8(sp) # 8-byte Folded Spill ; RV64I-NEXT: sd s0, 0(sp) # 8-byte Folded Spill ; RV64I-NEXT: mv s0, a1 ; RV64I-NEXT: lui a1, 16 ; RV64I-NEXT: addiw a1, a1, -1 ; RV64I-NEXT: and a0, a0, a1 ; RV64I-NEXT: call __gnu_h2f_ieee@plt ; RV64I-NEXT: lui a1, 524288 ; RV64I-NEXT: addiw a1, a1, -1 ; RV64I-NEXT: and a0, a0, a1 ; RV64I-NEXT: addi a1, zero, -1 ; RV64I-NEXT: slli a1, a1, 63 ; RV64I-NEXT: and a1, s0, a1 ; RV64I-NEXT: srli a1, a1, 32 ; RV64I-NEXT: or a0, a0, a1 ; RV64I-NEXT: call __gnu_f2h_ieee@plt ; RV64I-NEXT: ld s0, 0(sp) # 8-byte Folded Reload ; RV64I-NEXT: ld ra, 8(sp) # 8-byte Folded Reload ; RV64I-NEXT: addi sp, sp, 16 ; RV64I-NEXT: ret ; ; RV32IF-LABEL: fold_demote_h_d: ; RV32IF: # %bb.0: ; RV32IF-NEXT: addi sp, sp, -16 ; RV32IF-NEXT: sw ra, 12(sp) # 4-byte Folded Spill ; RV32IF-NEXT: sw s0, 8(sp) # 4-byte Folded Spill ; RV32IF-NEXT: mv s0, a1 ; RV32IF-NEXT: call __gnu_f2h_ieee@plt ; RV32IF-NEXT: call __gnu_h2f_ieee@plt ; RV32IF-NEXT: fmv.w.x ft0, s0 ; RV32IF-NEXT: fsgnj.s fa0, fa0, ft0 ; RV32IF-NEXT: lw s0, 8(sp) # 4-byte Folded Reload ; RV32IF-NEXT: lw ra, 12(sp) # 4-byte Folded Reload ; RV32IF-NEXT: addi sp, sp, 16 ; RV32IF-NEXT: ret ; ; RV32IFD-LABEL: fold_demote_h_d: ; RV32IFD: # %bb.0: ; RV32IFD-NEXT: addi sp, sp, -16 ; RV32IFD-NEXT: sw ra, 12(sp) # 4-byte Folded Spill ; RV32IFD-NEXT: fsd fs0, 0(sp) # 8-byte Folded Spill ; RV32IFD-NEXT: fmv.d fs0, fa1 ; RV32IFD-NEXT: call __gnu_f2h_ieee@plt ; RV32IFD-NEXT: call __gnu_h2f_ieee@plt ; RV32IFD-NEXT: fcvt.s.d ft0, fs0 ; RV32IFD-NEXT: fsgnj.s fa0, fa0, ft0 ; RV32IFD-NEXT: fld fs0, 0(sp) # 8-byte Folded Reload ; RV32IFD-NEXT: lw ra, 12(sp) # 4-byte Folded Reload ; RV32IFD-NEXT: addi sp, sp, 16 ; RV32IFD-NEXT: ret ; ; RV64IFD-LABEL: fold_demote_h_d: ; RV64IFD: # %bb.0: ; RV64IFD-NEXT: addi sp, sp, -16 ; RV64IFD-NEXT: sd ra, 8(sp) # 8-byte Folded Spill ; RV64IFD-NEXT: fsd fs0, 0(sp) # 8-byte Folded Spill ; RV64IFD-NEXT: fmv.d fs0, fa1 ; RV64IFD-NEXT: call __gnu_f2h_ieee@plt ; RV64IFD-NEXT: call __gnu_h2f_ieee@plt ; RV64IFD-NEXT: fcvt.s.d ft0, fs0 ; RV64IFD-NEXT: fsgnj.s fa0, fa0, ft0 ; RV64IFD-NEXT: fld fs0, 0(sp) # 8-byte Folded Reload ; RV64IFD-NEXT: ld ra, 8(sp) # 8-byte Folded Reload ; RV64IFD-NEXT: addi sp, sp, 16 ; RV64IFD-NEXT: ret ; ; RV32IFZFH-LABEL: fold_demote_h_d: ; RV32IFZFH: # %bb.0: ; RV32IFZFH-NEXT: srli a0, a1, 16 ; RV32IFZFH-NEXT: fmv.h.x ft0, a0 ; RV32IFZFH-NEXT: fsgnj.h fa0, fa0, ft0 ; RV32IFZFH-NEXT: ret ; ; RV32IFDZFH-LABEL: fold_demote_h_d: ; RV32IFDZFH: # %bb.0: ; RV32IFDZFH-NEXT: fcvt.h.d ft0, fa1 ; RV32IFDZFH-NEXT: fsgnj.h fa0, fa0, ft0 ; RV32IFDZFH-NEXT: ret ; ; RV64IFDZFH-LABEL: fold_demote_h_d: ; RV64IFDZFH: # %bb.0: ; RV64IFDZFH-NEXT: fcvt.h.d ft0, fa1 ; RV64IFDZFH-NEXT: fsgnj.h fa0, fa0, ft0 ; RV64IFDZFH-NEXT: ret %c = fptrunc double %b to half %t = call half @llvm.copysign.f16(half %a, half %c) ret half %t }