
(FPCore (J l K U) :precision binary64 (+ (* (* J (- (exp l) (exp (- l)))) (cos (/ K 2.0))) U))
double code(double J, double l, double K, double U) {
return ((J * (exp(l) - exp(-l))) * cos((K / 2.0))) + U;
}
real(8) function code(j, l, k, u)
real(8), intent (in) :: j
real(8), intent (in) :: l
real(8), intent (in) :: k
real(8), intent (in) :: u
code = ((j * (exp(l) - exp(-l))) * cos((k / 2.0d0))) + u
end function
public static double code(double J, double l, double K, double U) {
return ((J * (Math.exp(l) - Math.exp(-l))) * Math.cos((K / 2.0))) + U;
}
def code(J, l, K, U): return ((J * (math.exp(l) - math.exp(-l))) * math.cos((K / 2.0))) + U
function code(J, l, K, U) return Float64(Float64(Float64(J * Float64(exp(l) - exp(Float64(-l)))) * cos(Float64(K / 2.0))) + U) end
function tmp = code(J, l, K, U) tmp = ((J * (exp(l) - exp(-l))) * cos((K / 2.0))) + U; end
code[J_, l_, K_, U_] := N[(N[(N[(J * N[(N[Exp[l], $MachinePrecision] - N[Exp[(-l)], $MachinePrecision]), $MachinePrecision]), $MachinePrecision] * N[Cos[N[(K / 2.0), $MachinePrecision]], $MachinePrecision]), $MachinePrecision] + U), $MachinePrecision]
\begin{array}{l}
\\
\left(J \cdot \left(e^{\ell} - e^{-\ell}\right)\right) \cdot \cos \left(\frac{K}{2}\right) + U
\end{array}
Sampling outcomes in binary64 precision:
Herbie found 14 alternatives:
| Alternative | Accuracy | Speedup |
|---|
(FPCore (J l K U) :precision binary64 (+ (* (* J (- (exp l) (exp (- l)))) (cos (/ K 2.0))) U))
double code(double J, double l, double K, double U) {
return ((J * (exp(l) - exp(-l))) * cos((K / 2.0))) + U;
}
real(8) function code(j, l, k, u)
real(8), intent (in) :: j
real(8), intent (in) :: l
real(8), intent (in) :: k
real(8), intent (in) :: u
code = ((j * (exp(l) - exp(-l))) * cos((k / 2.0d0))) + u
end function
public static double code(double J, double l, double K, double U) {
return ((J * (Math.exp(l) - Math.exp(-l))) * Math.cos((K / 2.0))) + U;
}
def code(J, l, K, U): return ((J * (math.exp(l) - math.exp(-l))) * math.cos((K / 2.0))) + U
function code(J, l, K, U) return Float64(Float64(Float64(J * Float64(exp(l) - exp(Float64(-l)))) * cos(Float64(K / 2.0))) + U) end
function tmp = code(J, l, K, U) tmp = ((J * (exp(l) - exp(-l))) * cos((K / 2.0))) + U; end
code[J_, l_, K_, U_] := N[(N[(N[(J * N[(N[Exp[l], $MachinePrecision] - N[Exp[(-l)], $MachinePrecision]), $MachinePrecision]), $MachinePrecision] * N[Cos[N[(K / 2.0), $MachinePrecision]], $MachinePrecision]), $MachinePrecision] + U), $MachinePrecision]
\begin{array}{l}
\\
\left(J \cdot \left(e^{\ell} - e^{-\ell}\right)\right) \cdot \cos \left(\frac{K}{2}\right) + U
\end{array}
(FPCore (J l K U) :precision binary64 (fma (* (* (sinh l) (cos (/ K -2.0))) 2.0) J U))
double code(double J, double l, double K, double U) {
return fma(((sinh(l) * cos((K / -2.0))) * 2.0), J, U);
}
function code(J, l, K, U) return fma(Float64(Float64(sinh(l) * cos(Float64(K / -2.0))) * 2.0), J, U) end
code[J_, l_, K_, U_] := N[(N[(N[(N[Sinh[l], $MachinePrecision] * N[Cos[N[(K / -2.0), $MachinePrecision]], $MachinePrecision]), $MachinePrecision] * 2.0), $MachinePrecision] * J + U), $MachinePrecision]
\begin{array}{l}
\\
\mathsf{fma}\left(\left(\sinh \ell \cdot \cos \left(\frac{K}{-2}\right)\right) \cdot 2, J, U\right)
\end{array}
Initial program 83.3%
lift-+.f64N/A
lift-*.f64N/A
lift-*.f64N/A
associate-*l*N/A
*-commutativeN/A
lower-fma.f64N/A
Applied rewrites100.0%
(FPCore (J l K U)
:precision binary64
(let* ((t_0 (cos (/ K 2.0))))
(if (<= t_0 -0.45)
(fma (* (* (sinh l) (fma (* K K) -0.125 1.0)) 2.0) J U)
(if (<= t_0 -0.02)
(fma (* (+ J J) l) (cos (* -0.5 K)) U)
(fma (* (* (sinh l) 1.0) 2.0) J U)))))
double code(double J, double l, double K, double U) {
double t_0 = cos((K / 2.0));
double tmp;
if (t_0 <= -0.45) {
tmp = fma(((sinh(l) * fma((K * K), -0.125, 1.0)) * 2.0), J, U);
} else if (t_0 <= -0.02) {
tmp = fma(((J + J) * l), cos((-0.5 * K)), U);
} else {
tmp = fma(((sinh(l) * 1.0) * 2.0), J, U);
}
return tmp;
}
function code(J, l, K, U) t_0 = cos(Float64(K / 2.0)) tmp = 0.0 if (t_0 <= -0.45) tmp = fma(Float64(Float64(sinh(l) * fma(Float64(K * K), -0.125, 1.0)) * 2.0), J, U); elseif (t_0 <= -0.02) tmp = fma(Float64(Float64(J + J) * l), cos(Float64(-0.5 * K)), U); else tmp = fma(Float64(Float64(sinh(l) * 1.0) * 2.0), J, U); end return tmp end
code[J_, l_, K_, U_] := Block[{t$95$0 = N[Cos[N[(K / 2.0), $MachinePrecision]], $MachinePrecision]}, If[LessEqual[t$95$0, -0.45], N[(N[(N[(N[Sinh[l], $MachinePrecision] * N[(N[(K * K), $MachinePrecision] * -0.125 + 1.0), $MachinePrecision]), $MachinePrecision] * 2.0), $MachinePrecision] * J + U), $MachinePrecision], If[LessEqual[t$95$0, -0.02], N[(N[(N[(J + J), $MachinePrecision] * l), $MachinePrecision] * N[Cos[N[(-0.5 * K), $MachinePrecision]], $MachinePrecision] + U), $MachinePrecision], N[(N[(N[(N[Sinh[l], $MachinePrecision] * 1.0), $MachinePrecision] * 2.0), $MachinePrecision] * J + U), $MachinePrecision]]]]
\begin{array}{l}
\\
\begin{array}{l}
t_0 := \cos \left(\frac{K}{2}\right)\\
\mathbf{if}\;t\_0 \leq -0.45:\\
\;\;\;\;\mathsf{fma}\left(\left(\sinh \ell \cdot \mathsf{fma}\left(K \cdot K, -0.125, 1\right)\right) \cdot 2, J, U\right)\\
\mathbf{elif}\;t\_0 \leq -0.02:\\
\;\;\;\;\mathsf{fma}\left(\left(J + J\right) \cdot \ell, \cos \left(-0.5 \cdot K\right), U\right)\\
\mathbf{else}:\\
\;\;\;\;\mathsf{fma}\left(\left(\sinh \ell \cdot 1\right) \cdot 2, J, U\right)\\
\end{array}
\end{array}
if (cos.f64 (/.f64 K #s(literal 2 binary64))) < -0.450000000000000011Initial program 88.8%
lift-+.f64N/A
lift-*.f64N/A
lift-*.f64N/A
associate-*l*N/A
*-commutativeN/A
lower-fma.f64N/A
Applied rewrites100.0%
Taylor expanded in K around 0
+-commutativeN/A
*-commutativeN/A
lower-fma.f64N/A
unpow2N/A
lower-*.f6472.0
Applied rewrites72.0%
if -0.450000000000000011 < (cos.f64 (/.f64 K #s(literal 2 binary64))) < -0.0200000000000000004Initial program 78.9%
Taylor expanded in l around 0
+-commutativeN/A
*-commutativeN/A
associate-*r*N/A
associate-*l*N/A
*-commutativeN/A
associate-*r*N/A
associate-*r*N/A
lower-fma.f64N/A
*-commutativeN/A
lower-*.f64N/A
lower-*.f64N/A
cos-neg-revN/A
lower-cos.f64N/A
distribute-lft-neg-inN/A
lower-*.f64N/A
metadata-eval84.6
Applied rewrites84.6%
Applied rewrites84.6%
if -0.0200000000000000004 < (cos.f64 (/.f64 K #s(literal 2 binary64))) Initial program 82.2%
lift-+.f64N/A
lift-*.f64N/A
lift-*.f64N/A
associate-*l*N/A
*-commutativeN/A
lower-fma.f64N/A
Applied rewrites100.0%
Taylor expanded in K around 0
Applied rewrites95.5%
(FPCore (J l K U) :precision binary64 (if (<= (cos (/ K 2.0)) 0.172) (fma (* (cos (* -0.5 K)) (* J (fma (* l l) 0.3333333333333333 2.0))) l U) (fma (* (* (sinh l) 1.0) 2.0) J U)))
double code(double J, double l, double K, double U) {
double tmp;
if (cos((K / 2.0)) <= 0.172) {
tmp = fma((cos((-0.5 * K)) * (J * fma((l * l), 0.3333333333333333, 2.0))), l, U);
} else {
tmp = fma(((sinh(l) * 1.0) * 2.0), J, U);
}
return tmp;
}
function code(J, l, K, U) tmp = 0.0 if (cos(Float64(K / 2.0)) <= 0.172) tmp = fma(Float64(cos(Float64(-0.5 * K)) * Float64(J * fma(Float64(l * l), 0.3333333333333333, 2.0))), l, U); else tmp = fma(Float64(Float64(sinh(l) * 1.0) * 2.0), J, U); end return tmp end
code[J_, l_, K_, U_] := If[LessEqual[N[Cos[N[(K / 2.0), $MachinePrecision]], $MachinePrecision], 0.172], N[(N[(N[Cos[N[(-0.5 * K), $MachinePrecision]], $MachinePrecision] * N[(J * N[(N[(l * l), $MachinePrecision] * 0.3333333333333333 + 2.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] * l + U), $MachinePrecision], N[(N[(N[(N[Sinh[l], $MachinePrecision] * 1.0), $MachinePrecision] * 2.0), $MachinePrecision] * J + U), $MachinePrecision]]
\begin{array}{l}
\\
\begin{array}{l}
\mathbf{if}\;\cos \left(\frac{K}{2}\right) \leq 0.172:\\
\;\;\;\;\mathsf{fma}\left(\cos \left(-0.5 \cdot K\right) \cdot \left(J \cdot \mathsf{fma}\left(\ell \cdot \ell, 0.3333333333333333, 2\right)\right), \ell, U\right)\\
\mathbf{else}:\\
\;\;\;\;\mathsf{fma}\left(\left(\sinh \ell \cdot 1\right) \cdot 2, J, U\right)\\
\end{array}
\end{array}
if (cos.f64 (/.f64 K #s(literal 2 binary64))) < 0.17199999999999999Initial program 86.3%
Taylor expanded in l around 0
+-commutativeN/A
*-commutativeN/A
lower-fma.f64N/A
Applied rewrites87.8%
if 0.17199999999999999 < (cos.f64 (/.f64 K #s(literal 2 binary64))) Initial program 82.0%
lift-+.f64N/A
lift-*.f64N/A
lift-*.f64N/A
associate-*l*N/A
*-commutativeN/A
lower-fma.f64N/A
Applied rewrites100.0%
Taylor expanded in K around 0
Applied rewrites95.8%
(FPCore (J l K U) :precision binary64 (if (<= (cos (/ K 2.0)) -0.02) (fma (* (+ J J) l) (cos (* -0.5 K)) U) (fma (* (* (sinh l) 1.0) 2.0) J U)))
double code(double J, double l, double K, double U) {
double tmp;
if (cos((K / 2.0)) <= -0.02) {
tmp = fma(((J + J) * l), cos((-0.5 * K)), U);
} else {
tmp = fma(((sinh(l) * 1.0) * 2.0), J, U);
}
return tmp;
}
function code(J, l, K, U) tmp = 0.0 if (cos(Float64(K / 2.0)) <= -0.02) tmp = fma(Float64(Float64(J + J) * l), cos(Float64(-0.5 * K)), U); else tmp = fma(Float64(Float64(sinh(l) * 1.0) * 2.0), J, U); end return tmp end
code[J_, l_, K_, U_] := If[LessEqual[N[Cos[N[(K / 2.0), $MachinePrecision]], $MachinePrecision], -0.02], N[(N[(N[(J + J), $MachinePrecision] * l), $MachinePrecision] * N[Cos[N[(-0.5 * K), $MachinePrecision]], $MachinePrecision] + U), $MachinePrecision], N[(N[(N[(N[Sinh[l], $MachinePrecision] * 1.0), $MachinePrecision] * 2.0), $MachinePrecision] * J + U), $MachinePrecision]]
\begin{array}{l}
\\
\begin{array}{l}
\mathbf{if}\;\cos \left(\frac{K}{2}\right) \leq -0.02:\\
\;\;\;\;\mathsf{fma}\left(\left(J + J\right) \cdot \ell, \cos \left(-0.5 \cdot K\right), U\right)\\
\mathbf{else}:\\
\;\;\;\;\mathsf{fma}\left(\left(\sinh \ell \cdot 1\right) \cdot 2, J, U\right)\\
\end{array}
\end{array}
if (cos.f64 (/.f64 K #s(literal 2 binary64))) < -0.0200000000000000004Initial program 86.2%
Taylor expanded in l around 0
+-commutativeN/A
*-commutativeN/A
associate-*r*N/A
associate-*l*N/A
*-commutativeN/A
associate-*r*N/A
associate-*r*N/A
lower-fma.f64N/A
*-commutativeN/A
lower-*.f64N/A
lower-*.f64N/A
cos-neg-revN/A
lower-cos.f64N/A
distribute-lft-neg-inN/A
lower-*.f64N/A
metadata-eval67.0
Applied rewrites67.0%
Applied rewrites67.0%
if -0.0200000000000000004 < (cos.f64 (/.f64 K #s(literal 2 binary64))) Initial program 82.2%
lift-+.f64N/A
lift-*.f64N/A
lift-*.f64N/A
associate-*l*N/A
*-commutativeN/A
lower-fma.f64N/A
Applied rewrites100.0%
Taylor expanded in K around 0
Applied rewrites95.5%
(FPCore (J l K U) :precision binary64 (if (<= (cos (/ K 2.0)) -0.45) (fma (* (* (fma (* K K) -0.125 1.0) l) 2.0) J U) (fma (* (* (sinh l) 1.0) 2.0) J U)))
double code(double J, double l, double K, double U) {
double tmp;
if (cos((K / 2.0)) <= -0.45) {
tmp = fma(((fma((K * K), -0.125, 1.0) * l) * 2.0), J, U);
} else {
tmp = fma(((sinh(l) * 1.0) * 2.0), J, U);
}
return tmp;
}
function code(J, l, K, U) tmp = 0.0 if (cos(Float64(K / 2.0)) <= -0.45) tmp = fma(Float64(Float64(fma(Float64(K * K), -0.125, 1.0) * l) * 2.0), J, U); else tmp = fma(Float64(Float64(sinh(l) * 1.0) * 2.0), J, U); end return tmp end
code[J_, l_, K_, U_] := If[LessEqual[N[Cos[N[(K / 2.0), $MachinePrecision]], $MachinePrecision], -0.45], N[(N[(N[(N[(N[(K * K), $MachinePrecision] * -0.125 + 1.0), $MachinePrecision] * l), $MachinePrecision] * 2.0), $MachinePrecision] * J + U), $MachinePrecision], N[(N[(N[(N[Sinh[l], $MachinePrecision] * 1.0), $MachinePrecision] * 2.0), $MachinePrecision] * J + U), $MachinePrecision]]
\begin{array}{l}
\\
\begin{array}{l}
\mathbf{if}\;\cos \left(\frac{K}{2}\right) \leq -0.45:\\
\;\;\;\;\mathsf{fma}\left(\left(\mathsf{fma}\left(K \cdot K, -0.125, 1\right) \cdot \ell\right) \cdot 2, J, U\right)\\
\mathbf{else}:\\
\;\;\;\;\mathsf{fma}\left(\left(\sinh \ell \cdot 1\right) \cdot 2, J, U\right)\\
\end{array}
\end{array}
if (cos.f64 (/.f64 K #s(literal 2 binary64))) < -0.450000000000000011Initial program 88.8%
lift-+.f64N/A
lift-*.f64N/A
lift-*.f64N/A
associate-*l*N/A
*-commutativeN/A
lower-fma.f64N/A
Applied rewrites100.0%
Taylor expanded in l around 0
*-commutativeN/A
lower-*.f64N/A
cos-neg-revN/A
lower-cos.f64N/A
distribute-lft-neg-inN/A
metadata-evalN/A
lower-*.f6460.9
Applied rewrites60.9%
Taylor expanded in K around 0
Applied rewrites66.4%
if -0.450000000000000011 < (cos.f64 (/.f64 K #s(literal 2 binary64))) Initial program 81.9%
lift-+.f64N/A
lift-*.f64N/A
lift-*.f64N/A
associate-*l*N/A
*-commutativeN/A
lower-fma.f64N/A
Applied rewrites100.0%
Taylor expanded in K around 0
Applied rewrites92.5%
(FPCore (J l K U)
:precision binary64
(+
(*
(*
J
(*
(fma
(fma
(fma 0.0003968253968253968 (* l l) 0.016666666666666666)
(* l l)
0.3333333333333333)
(* l l)
2.0)
l))
(cos (/ K 2.0)))
U))
double code(double J, double l, double K, double U) {
return ((J * (fma(fma(fma(0.0003968253968253968, (l * l), 0.016666666666666666), (l * l), 0.3333333333333333), (l * l), 2.0) * l)) * cos((K / 2.0))) + U;
}
function code(J, l, K, U) return Float64(Float64(Float64(J * Float64(fma(fma(fma(0.0003968253968253968, Float64(l * l), 0.016666666666666666), Float64(l * l), 0.3333333333333333), Float64(l * l), 2.0) * l)) * cos(Float64(K / 2.0))) + U) end
code[J_, l_, K_, U_] := N[(N[(N[(J * N[(N[(N[(N[(0.0003968253968253968 * N[(l * l), $MachinePrecision] + 0.016666666666666666), $MachinePrecision] * N[(l * l), $MachinePrecision] + 0.3333333333333333), $MachinePrecision] * N[(l * l), $MachinePrecision] + 2.0), $MachinePrecision] * l), $MachinePrecision]), $MachinePrecision] * N[Cos[N[(K / 2.0), $MachinePrecision]], $MachinePrecision]), $MachinePrecision] + U), $MachinePrecision]
\begin{array}{l}
\\
\left(J \cdot \left(\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(0.0003968253968253968, \ell \cdot \ell, 0.016666666666666666\right), \ell \cdot \ell, 0.3333333333333333\right), \ell \cdot \ell, 2\right) \cdot \ell\right)\right) \cdot \cos \left(\frac{K}{2}\right) + U
\end{array}
Initial program 83.3%
Taylor expanded in l around 0
*-commutativeN/A
lower-*.f64N/A
+-commutativeN/A
*-commutativeN/A
lower-fma.f64N/A
+-commutativeN/A
*-commutativeN/A
lower-fma.f64N/A
+-commutativeN/A
lower-fma.f64N/A
unpow2N/A
lower-*.f64N/A
unpow2N/A
lower-*.f64N/A
unpow2N/A
lower-*.f6495.8
Applied rewrites95.8%
(FPCore (J l K U)
:precision binary64
(if (<= K 1e-6)
(fma (* (* (sinh l) (fma (* K K) -0.125 1.0)) 2.0) J U)
(+
(*
(*
J
(*
(fma (fma 0.016666666666666666 (* l l) 0.3333333333333333) (* l l) 2.0)
l))
(cos (/ K 2.0)))
U)))
double code(double J, double l, double K, double U) {
double tmp;
if (K <= 1e-6) {
tmp = fma(((sinh(l) * fma((K * K), -0.125, 1.0)) * 2.0), J, U);
} else {
tmp = ((J * (fma(fma(0.016666666666666666, (l * l), 0.3333333333333333), (l * l), 2.0) * l)) * cos((K / 2.0))) + U;
}
return tmp;
}
function code(J, l, K, U) tmp = 0.0 if (K <= 1e-6) tmp = fma(Float64(Float64(sinh(l) * fma(Float64(K * K), -0.125, 1.0)) * 2.0), J, U); else tmp = Float64(Float64(Float64(J * Float64(fma(fma(0.016666666666666666, Float64(l * l), 0.3333333333333333), Float64(l * l), 2.0) * l)) * cos(Float64(K / 2.0))) + U); end return tmp end
code[J_, l_, K_, U_] := If[LessEqual[K, 1e-6], N[(N[(N[(N[Sinh[l], $MachinePrecision] * N[(N[(K * K), $MachinePrecision] * -0.125 + 1.0), $MachinePrecision]), $MachinePrecision] * 2.0), $MachinePrecision] * J + U), $MachinePrecision], N[(N[(N[(J * N[(N[(N[(0.016666666666666666 * N[(l * l), $MachinePrecision] + 0.3333333333333333), $MachinePrecision] * N[(l * l), $MachinePrecision] + 2.0), $MachinePrecision] * l), $MachinePrecision]), $MachinePrecision] * N[Cos[N[(K / 2.0), $MachinePrecision]], $MachinePrecision]), $MachinePrecision] + U), $MachinePrecision]]
\begin{array}{l}
\\
\begin{array}{l}
\mathbf{if}\;K \leq 10^{-6}:\\
\;\;\;\;\mathsf{fma}\left(\left(\sinh \ell \cdot \mathsf{fma}\left(K \cdot K, -0.125, 1\right)\right) \cdot 2, J, U\right)\\
\mathbf{else}:\\
\;\;\;\;\left(J \cdot \left(\mathsf{fma}\left(\mathsf{fma}\left(0.016666666666666666, \ell \cdot \ell, 0.3333333333333333\right), \ell \cdot \ell, 2\right) \cdot \ell\right)\right) \cdot \cos \left(\frac{K}{2}\right) + U\\
\end{array}
\end{array}
if K < 9.99999999999999955e-7Initial program 81.0%
lift-+.f64N/A
lift-*.f64N/A
lift-*.f64N/A
associate-*l*N/A
*-commutativeN/A
lower-fma.f64N/A
Applied rewrites100.0%
Taylor expanded in K around 0
+-commutativeN/A
*-commutativeN/A
lower-fma.f64N/A
unpow2N/A
lower-*.f6477.6
Applied rewrites77.6%
if 9.99999999999999955e-7 < K Initial program 88.2%
Taylor expanded in l around 0
*-commutativeN/A
lower-*.f64N/A
+-commutativeN/A
*-commutativeN/A
lower-fma.f64N/A
+-commutativeN/A
lower-fma.f64N/A
unpow2N/A
lower-*.f64N/A
unpow2N/A
lower-*.f6495.4
Applied rewrites95.4%
(FPCore (J l K U) :precision binary64 (if (<= K 0.115) (fma (* (* (sinh l) (fma (* K K) -0.125 1.0)) 2.0) J U) (+ (* (* J (* (fma (* l l) 0.3333333333333333 2.0) l)) (cos (/ K 2.0))) U)))
double code(double J, double l, double K, double U) {
double tmp;
if (K <= 0.115) {
tmp = fma(((sinh(l) * fma((K * K), -0.125, 1.0)) * 2.0), J, U);
} else {
tmp = ((J * (fma((l * l), 0.3333333333333333, 2.0) * l)) * cos((K / 2.0))) + U;
}
return tmp;
}
function code(J, l, K, U) tmp = 0.0 if (K <= 0.115) tmp = fma(Float64(Float64(sinh(l) * fma(Float64(K * K), -0.125, 1.0)) * 2.0), J, U); else tmp = Float64(Float64(Float64(J * Float64(fma(Float64(l * l), 0.3333333333333333, 2.0) * l)) * cos(Float64(K / 2.0))) + U); end return tmp end
code[J_, l_, K_, U_] := If[LessEqual[K, 0.115], N[(N[(N[(N[Sinh[l], $MachinePrecision] * N[(N[(K * K), $MachinePrecision] * -0.125 + 1.0), $MachinePrecision]), $MachinePrecision] * 2.0), $MachinePrecision] * J + U), $MachinePrecision], N[(N[(N[(J * N[(N[(N[(l * l), $MachinePrecision] * 0.3333333333333333 + 2.0), $MachinePrecision] * l), $MachinePrecision]), $MachinePrecision] * N[Cos[N[(K / 2.0), $MachinePrecision]], $MachinePrecision]), $MachinePrecision] + U), $MachinePrecision]]
\begin{array}{l}
\\
\begin{array}{l}
\mathbf{if}\;K \leq 0.115:\\
\;\;\;\;\mathsf{fma}\left(\left(\sinh \ell \cdot \mathsf{fma}\left(K \cdot K, -0.125, 1\right)\right) \cdot 2, J, U\right)\\
\mathbf{else}:\\
\;\;\;\;\left(J \cdot \left(\mathsf{fma}\left(\ell \cdot \ell, 0.3333333333333333, 2\right) \cdot \ell\right)\right) \cdot \cos \left(\frac{K}{2}\right) + U\\
\end{array}
\end{array}
if K < 0.115000000000000005Initial program 81.0%
lift-+.f64N/A
lift-*.f64N/A
lift-*.f64N/A
associate-*l*N/A
*-commutativeN/A
lower-fma.f64N/A
Applied rewrites100.0%
Taylor expanded in K around 0
+-commutativeN/A
*-commutativeN/A
lower-fma.f64N/A
unpow2N/A
lower-*.f6477.6
Applied rewrites77.6%
if 0.115000000000000005 < K Initial program 88.2%
Taylor expanded in l around 0
*-commutativeN/A
lower-*.f64N/A
+-commutativeN/A
*-commutativeN/A
lower-fma.f64N/A
unpow2N/A
lower-*.f6493.0
Applied rewrites93.0%
(FPCore (J l K U) :precision binary64 (if (<= (cos (/ K 2.0)) 5.5e-5) (fma (* (* 2.0 l) (* 2.0 l)) J U) (* (fma (/ (* J l) U) 2.0 1.0) U)))
double code(double J, double l, double K, double U) {
double tmp;
if (cos((K / 2.0)) <= 5.5e-5) {
tmp = fma(((2.0 * l) * (2.0 * l)), J, U);
} else {
tmp = fma(((J * l) / U), 2.0, 1.0) * U;
}
return tmp;
}
function code(J, l, K, U) tmp = 0.0 if (cos(Float64(K / 2.0)) <= 5.5e-5) tmp = fma(Float64(Float64(2.0 * l) * Float64(2.0 * l)), J, U); else tmp = Float64(fma(Float64(Float64(J * l) / U), 2.0, 1.0) * U); end return tmp end
code[J_, l_, K_, U_] := If[LessEqual[N[Cos[N[(K / 2.0), $MachinePrecision]], $MachinePrecision], 5.5e-5], N[(N[(N[(2.0 * l), $MachinePrecision] * N[(2.0 * l), $MachinePrecision]), $MachinePrecision] * J + U), $MachinePrecision], N[(N[(N[(N[(J * l), $MachinePrecision] / U), $MachinePrecision] * 2.0 + 1.0), $MachinePrecision] * U), $MachinePrecision]]
\begin{array}{l}
\\
\begin{array}{l}
\mathbf{if}\;\cos \left(\frac{K}{2}\right) \leq 5.5 \cdot 10^{-5}:\\
\;\;\;\;\mathsf{fma}\left(\left(2 \cdot \ell\right) \cdot \left(2 \cdot \ell\right), J, U\right)\\
\mathbf{else}:\\
\;\;\;\;\mathsf{fma}\left(\frac{J \cdot \ell}{U}, 2, 1\right) \cdot U\\
\end{array}
\end{array}
if (cos.f64 (/.f64 K #s(literal 2 binary64))) < 5.5000000000000002e-5Initial program 86.4%
Taylor expanded in l around 0
+-commutativeN/A
*-commutativeN/A
associate-*r*N/A
associate-*l*N/A
*-commutativeN/A
associate-*r*N/A
associate-*r*N/A
lower-fma.f64N/A
*-commutativeN/A
lower-*.f64N/A
lower-*.f64N/A
cos-neg-revN/A
lower-cos.f64N/A
distribute-lft-neg-inN/A
lower-*.f64N/A
metadata-eval66.1
Applied rewrites66.1%
Taylor expanded in K around 0
Applied rewrites37.3%
Applied rewrites37.3%
Applied rewrites57.8%
if 5.5000000000000002e-5 < (cos.f64 (/.f64 K #s(literal 2 binary64))) Initial program 82.1%
Taylor expanded in l around 0
+-commutativeN/A
*-commutativeN/A
associate-*r*N/A
associate-*l*N/A
*-commutativeN/A
associate-*r*N/A
associate-*r*N/A
lower-fma.f64N/A
*-commutativeN/A
lower-*.f64N/A
lower-*.f64N/A
cos-neg-revN/A
lower-cos.f64N/A
distribute-lft-neg-inN/A
lower-*.f64N/A
metadata-eval71.0
Applied rewrites71.0%
Taylor expanded in K around 0
Applied rewrites66.5%
Taylor expanded in U around inf
Applied rewrites71.1%
(FPCore (J l K U)
:precision binary64
(if (<= l -7.2e+254)
(fma (* (* (fma (* K K) -0.125 1.0) l) 2.0) J U)
(if (or (<= l -1.05e+17) (not (<= l 0.0033)))
(fma (pow (* 2.0 l) 3.0) J U)
(fma (+ l l) J U))))
double code(double J, double l, double K, double U) {
double tmp;
if (l <= -7.2e+254) {
tmp = fma(((fma((K * K), -0.125, 1.0) * l) * 2.0), J, U);
} else if ((l <= -1.05e+17) || !(l <= 0.0033)) {
tmp = fma(pow((2.0 * l), 3.0), J, U);
} else {
tmp = fma((l + l), J, U);
}
return tmp;
}
function code(J, l, K, U) tmp = 0.0 if (l <= -7.2e+254) tmp = fma(Float64(Float64(fma(Float64(K * K), -0.125, 1.0) * l) * 2.0), J, U); elseif ((l <= -1.05e+17) || !(l <= 0.0033)) tmp = fma((Float64(2.0 * l) ^ 3.0), J, U); else tmp = fma(Float64(l + l), J, U); end return tmp end
code[J_, l_, K_, U_] := If[LessEqual[l, -7.2e+254], N[(N[(N[(N[(N[(K * K), $MachinePrecision] * -0.125 + 1.0), $MachinePrecision] * l), $MachinePrecision] * 2.0), $MachinePrecision] * J + U), $MachinePrecision], If[Or[LessEqual[l, -1.05e+17], N[Not[LessEqual[l, 0.0033]], $MachinePrecision]], N[(N[Power[N[(2.0 * l), $MachinePrecision], 3.0], $MachinePrecision] * J + U), $MachinePrecision], N[(N[(l + l), $MachinePrecision] * J + U), $MachinePrecision]]]
\begin{array}{l}
\\
\begin{array}{l}
\mathbf{if}\;\ell \leq -7.2 \cdot 10^{+254}:\\
\;\;\;\;\mathsf{fma}\left(\left(\mathsf{fma}\left(K \cdot K, -0.125, 1\right) \cdot \ell\right) \cdot 2, J, U\right)\\
\mathbf{elif}\;\ell \leq -1.05 \cdot 10^{+17} \lor \neg \left(\ell \leq 0.0033\right):\\
\;\;\;\;\mathsf{fma}\left({\left(2 \cdot \ell\right)}^{3}, J, U\right)\\
\mathbf{else}:\\
\;\;\;\;\mathsf{fma}\left(\ell + \ell, J, U\right)\\
\end{array}
\end{array}
if l < -7.19999999999999954e254Initial program 100.0%
lift-+.f64N/A
lift-*.f64N/A
lift-*.f64N/A
associate-*l*N/A
*-commutativeN/A
lower-fma.f64N/A
Applied rewrites100.0%
Taylor expanded in l around 0
*-commutativeN/A
lower-*.f64N/A
cos-neg-revN/A
lower-cos.f64N/A
distribute-lft-neg-inN/A
metadata-evalN/A
lower-*.f6451.7
Applied rewrites51.7%
Taylor expanded in K around 0
Applied rewrites75.7%
if -7.19999999999999954e254 < l < -1.05e17 or 0.0033 < l Initial program 100.0%
Taylor expanded in l around 0
+-commutativeN/A
*-commutativeN/A
associate-*r*N/A
associate-*l*N/A
*-commutativeN/A
associate-*r*N/A
associate-*r*N/A
lower-fma.f64N/A
*-commutativeN/A
lower-*.f64N/A
lower-*.f64N/A
cos-neg-revN/A
lower-cos.f64N/A
distribute-lft-neg-inN/A
lower-*.f64N/A
metadata-eval31.3
Applied rewrites31.3%
Taylor expanded in K around 0
Applied rewrites24.0%
Applied rewrites24.0%
Applied rewrites61.9%
if -1.05e17 < l < 0.0033Initial program 69.7%
Taylor expanded in l around 0
+-commutativeN/A
*-commutativeN/A
associate-*r*N/A
associate-*l*N/A
*-commutativeN/A
associate-*r*N/A
associate-*r*N/A
lower-fma.f64N/A
*-commutativeN/A
lower-*.f64N/A
lower-*.f64N/A
cos-neg-revN/A
lower-cos.f64N/A
distribute-lft-neg-inN/A
lower-*.f64N/A
metadata-eval99.2
Applied rewrites99.2%
Taylor expanded in K around 0
Applied rewrites86.3%
Applied rewrites86.3%
Final simplification76.0%
(FPCore (J l K U) :precision binary64 (if (or (<= l -5e+185) (not (<= l 0.0033))) (fma (* (* 2.0 l) (* 2.0 l)) J U) (fma (+ l l) J U)))
double code(double J, double l, double K, double U) {
double tmp;
if ((l <= -5e+185) || !(l <= 0.0033)) {
tmp = fma(((2.0 * l) * (2.0 * l)), J, U);
} else {
tmp = fma((l + l), J, U);
}
return tmp;
}
function code(J, l, K, U) tmp = 0.0 if ((l <= -5e+185) || !(l <= 0.0033)) tmp = fma(Float64(Float64(2.0 * l) * Float64(2.0 * l)), J, U); else tmp = fma(Float64(l + l), J, U); end return tmp end
code[J_, l_, K_, U_] := If[Or[LessEqual[l, -5e+185], N[Not[LessEqual[l, 0.0033]], $MachinePrecision]], N[(N[(N[(2.0 * l), $MachinePrecision] * N[(2.0 * l), $MachinePrecision]), $MachinePrecision] * J + U), $MachinePrecision], N[(N[(l + l), $MachinePrecision] * J + U), $MachinePrecision]]
\begin{array}{l}
\\
\begin{array}{l}
\mathbf{if}\;\ell \leq -5 \cdot 10^{+185} \lor \neg \left(\ell \leq 0.0033\right):\\
\;\;\;\;\mathsf{fma}\left(\left(2 \cdot \ell\right) \cdot \left(2 \cdot \ell\right), J, U\right)\\
\mathbf{else}:\\
\;\;\;\;\mathsf{fma}\left(\ell + \ell, J, U\right)\\
\end{array}
\end{array}
if l < -4.9999999999999999e185 or 0.0033 < l Initial program 100.0%
Taylor expanded in l around 0
+-commutativeN/A
*-commutativeN/A
associate-*r*N/A
associate-*l*N/A
*-commutativeN/A
associate-*r*N/A
associate-*r*N/A
lower-fma.f64N/A
*-commutativeN/A
lower-*.f64N/A
lower-*.f64N/A
cos-neg-revN/A
lower-cos.f64N/A
distribute-lft-neg-inN/A
lower-*.f64N/A
metadata-eval38.0
Applied rewrites38.0%
Taylor expanded in K around 0
Applied rewrites27.3%
Applied rewrites27.3%
Applied rewrites54.3%
if -4.9999999999999999e185 < l < 0.0033Initial program 74.5%
Taylor expanded in l around 0
+-commutativeN/A
*-commutativeN/A
associate-*r*N/A
associate-*l*N/A
*-commutativeN/A
associate-*r*N/A
associate-*r*N/A
lower-fma.f64N/A
*-commutativeN/A
lower-*.f64N/A
lower-*.f64N/A
cos-neg-revN/A
lower-cos.f64N/A
distribute-lft-neg-inN/A
lower-*.f64N/A
metadata-eval86.2
Applied rewrites86.2%
Taylor expanded in K around 0
Applied rewrites74.7%
Applied rewrites74.7%
Final simplification67.7%
(FPCore (J l K U) :precision binary64 (if (<= l -4.4e+20) (fma (* (* (fma (* K K) -0.125 1.0) l) 2.0) J U) (if (<= l 0.0033) (fma (+ l l) J U) (fma (* (* 2.0 l) (* 2.0 l)) J U))))
double code(double J, double l, double K, double U) {
double tmp;
if (l <= -4.4e+20) {
tmp = fma(((fma((K * K), -0.125, 1.0) * l) * 2.0), J, U);
} else if (l <= 0.0033) {
tmp = fma((l + l), J, U);
} else {
tmp = fma(((2.0 * l) * (2.0 * l)), J, U);
}
return tmp;
}
function code(J, l, K, U) tmp = 0.0 if (l <= -4.4e+20) tmp = fma(Float64(Float64(fma(Float64(K * K), -0.125, 1.0) * l) * 2.0), J, U); elseif (l <= 0.0033) tmp = fma(Float64(l + l), J, U); else tmp = fma(Float64(Float64(2.0 * l) * Float64(2.0 * l)), J, U); end return tmp end
code[J_, l_, K_, U_] := If[LessEqual[l, -4.4e+20], N[(N[(N[(N[(N[(K * K), $MachinePrecision] * -0.125 + 1.0), $MachinePrecision] * l), $MachinePrecision] * 2.0), $MachinePrecision] * J + U), $MachinePrecision], If[LessEqual[l, 0.0033], N[(N[(l + l), $MachinePrecision] * J + U), $MachinePrecision], N[(N[(N[(2.0 * l), $MachinePrecision] * N[(2.0 * l), $MachinePrecision]), $MachinePrecision] * J + U), $MachinePrecision]]]
\begin{array}{l}
\\
\begin{array}{l}
\mathbf{if}\;\ell \leq -4.4 \cdot 10^{+20}:\\
\;\;\;\;\mathsf{fma}\left(\left(\mathsf{fma}\left(K \cdot K, -0.125, 1\right) \cdot \ell\right) \cdot 2, J, U\right)\\
\mathbf{elif}\;\ell \leq 0.0033:\\
\;\;\;\;\mathsf{fma}\left(\ell + \ell, J, U\right)\\
\mathbf{else}:\\
\;\;\;\;\mathsf{fma}\left(\left(2 \cdot \ell\right) \cdot \left(2 \cdot \ell\right), J, U\right)\\
\end{array}
\end{array}
if l < -4.4e20Initial program 100.0%
lift-+.f64N/A
lift-*.f64N/A
lift-*.f64N/A
associate-*l*N/A
*-commutativeN/A
lower-fma.f64N/A
Applied rewrites100.0%
Taylor expanded in l around 0
*-commutativeN/A
lower-*.f64N/A
cos-neg-revN/A
lower-cos.f64N/A
distribute-lft-neg-inN/A
metadata-evalN/A
lower-*.f6431.2
Applied rewrites31.2%
Taylor expanded in K around 0
Applied rewrites48.7%
if -4.4e20 < l < 0.0033Initial program 69.9%
Taylor expanded in l around 0
+-commutativeN/A
*-commutativeN/A
associate-*r*N/A
associate-*l*N/A
*-commutativeN/A
associate-*r*N/A
associate-*r*N/A
lower-fma.f64N/A
*-commutativeN/A
lower-*.f64N/A
lower-*.f64N/A
cos-neg-revN/A
lower-cos.f64N/A
distribute-lft-neg-inN/A
lower-*.f64N/A
metadata-eval98.5
Applied rewrites98.5%
Taylor expanded in K around 0
Applied rewrites85.7%
Applied rewrites85.7%
if 0.0033 < l Initial program 100.0%
Taylor expanded in l around 0
+-commutativeN/A
*-commutativeN/A
associate-*r*N/A
associate-*l*N/A
*-commutativeN/A
associate-*r*N/A
associate-*r*N/A
lower-fma.f64N/A
*-commutativeN/A
lower-*.f64N/A
lower-*.f64N/A
cos-neg-revN/A
lower-cos.f64N/A
distribute-lft-neg-inN/A
lower-*.f64N/A
metadata-eval35.9
Applied rewrites35.9%
Taylor expanded in K around 0
Applied rewrites30.4%
Applied rewrites30.4%
Applied rewrites57.1%
(FPCore (J l K U) :precision binary64 (if (<= l -4.4e+20) (fma (* (fma -0.125 (* K K) 1.0) (* J 2.0)) l U) (if (<= l 0.0033) (fma (+ l l) J U) (fma (* (* 2.0 l) (* 2.0 l)) J U))))
double code(double J, double l, double K, double U) {
double tmp;
if (l <= -4.4e+20) {
tmp = fma((fma(-0.125, (K * K), 1.0) * (J * 2.0)), l, U);
} else if (l <= 0.0033) {
tmp = fma((l + l), J, U);
} else {
tmp = fma(((2.0 * l) * (2.0 * l)), J, U);
}
return tmp;
}
function code(J, l, K, U) tmp = 0.0 if (l <= -4.4e+20) tmp = fma(Float64(fma(-0.125, Float64(K * K), 1.0) * Float64(J * 2.0)), l, U); elseif (l <= 0.0033) tmp = fma(Float64(l + l), J, U); else tmp = fma(Float64(Float64(2.0 * l) * Float64(2.0 * l)), J, U); end return tmp end
code[J_, l_, K_, U_] := If[LessEqual[l, -4.4e+20], N[(N[(N[(-0.125 * N[(K * K), $MachinePrecision] + 1.0), $MachinePrecision] * N[(J * 2.0), $MachinePrecision]), $MachinePrecision] * l + U), $MachinePrecision], If[LessEqual[l, 0.0033], N[(N[(l + l), $MachinePrecision] * J + U), $MachinePrecision], N[(N[(N[(2.0 * l), $MachinePrecision] * N[(2.0 * l), $MachinePrecision]), $MachinePrecision] * J + U), $MachinePrecision]]]
\begin{array}{l}
\\
\begin{array}{l}
\mathbf{if}\;\ell \leq -4.4 \cdot 10^{+20}:\\
\;\;\;\;\mathsf{fma}\left(\mathsf{fma}\left(-0.125, K \cdot K, 1\right) \cdot \left(J \cdot 2\right), \ell, U\right)\\
\mathbf{elif}\;\ell \leq 0.0033:\\
\;\;\;\;\mathsf{fma}\left(\ell + \ell, J, U\right)\\
\mathbf{else}:\\
\;\;\;\;\mathsf{fma}\left(\left(2 \cdot \ell\right) \cdot \left(2 \cdot \ell\right), J, U\right)\\
\end{array}
\end{array}
if l < -4.4e20Initial program 100.0%
Taylor expanded in l around 0
+-commutativeN/A
*-commutativeN/A
associate-*r*N/A
associate-*l*N/A
*-commutativeN/A
associate-*r*N/A
associate-*r*N/A
lower-fma.f64N/A
*-commutativeN/A
lower-*.f64N/A
lower-*.f64N/A
cos-neg-revN/A
lower-cos.f64N/A
distribute-lft-neg-inN/A
lower-*.f64N/A
metadata-eval31.2
Applied rewrites31.2%
Applied rewrites31.2%
Taylor expanded in K around 0
Applied rewrites43.4%
if -4.4e20 < l < 0.0033Initial program 69.9%
Taylor expanded in l around 0
+-commutativeN/A
*-commutativeN/A
associate-*r*N/A
associate-*l*N/A
*-commutativeN/A
associate-*r*N/A
associate-*r*N/A
lower-fma.f64N/A
*-commutativeN/A
lower-*.f64N/A
lower-*.f64N/A
cos-neg-revN/A
lower-cos.f64N/A
distribute-lft-neg-inN/A
lower-*.f64N/A
metadata-eval98.5
Applied rewrites98.5%
Taylor expanded in K around 0
Applied rewrites85.7%
Applied rewrites85.7%
if 0.0033 < l Initial program 100.0%
Taylor expanded in l around 0
+-commutativeN/A
*-commutativeN/A
associate-*r*N/A
associate-*l*N/A
*-commutativeN/A
associate-*r*N/A
associate-*r*N/A
lower-fma.f64N/A
*-commutativeN/A
lower-*.f64N/A
lower-*.f64N/A
cos-neg-revN/A
lower-cos.f64N/A
distribute-lft-neg-inN/A
lower-*.f64N/A
metadata-eval35.9
Applied rewrites35.9%
Taylor expanded in K around 0
Applied rewrites30.4%
Applied rewrites30.4%
Applied rewrites57.1%
(FPCore (J l K U) :precision binary64 (fma (+ l l) J U))
double code(double J, double l, double K, double U) {
return fma((l + l), J, U);
}
function code(J, l, K, U) return fma(Float64(l + l), J, U) end
code[J_, l_, K_, U_] := N[(N[(l + l), $MachinePrecision] * J + U), $MachinePrecision]
\begin{array}{l}
\\
\mathsf{fma}\left(\ell + \ell, J, U\right)
\end{array}
Initial program 83.3%
Taylor expanded in l around 0
+-commutativeN/A
*-commutativeN/A
associate-*r*N/A
associate-*l*N/A
*-commutativeN/A
associate-*r*N/A
associate-*r*N/A
lower-fma.f64N/A
*-commutativeN/A
lower-*.f64N/A
lower-*.f64N/A
cos-neg-revN/A
lower-cos.f64N/A
distribute-lft-neg-inN/A
lower-*.f64N/A
metadata-eval69.6
Applied rewrites69.6%
Taylor expanded in K around 0
Applied rewrites58.4%
Applied rewrites58.4%
herbie shell --seed 2024344
(FPCore (J l K U)
:name "Maksimov and Kolovsky, Equation (4)"
:precision binary64
(+ (* (* J (- (exp l) (exp (- l)))) (cos (/ K 2.0))) U))