
(FPCore (f) :precision binary64 (let* ((t_0 (* (/ PI 4.0) f)) (t_1 (exp t_0)) (t_2 (exp (- t_0)))) (- (* (/ 1.0 (/ PI 4.0)) (log (/ (+ t_1 t_2) (- t_1 t_2)))))))
double code(double f) {
double t_0 = (((double) M_PI) / 4.0) * f;
double t_1 = exp(t_0);
double t_2 = exp(-t_0);
return -((1.0 / (((double) M_PI) / 4.0)) * log(((t_1 + t_2) / (t_1 - t_2))));
}
public static double code(double f) {
double t_0 = (Math.PI / 4.0) * f;
double t_1 = Math.exp(t_0);
double t_2 = Math.exp(-t_0);
return -((1.0 / (Math.PI / 4.0)) * Math.log(((t_1 + t_2) / (t_1 - t_2))));
}
def code(f): t_0 = (math.pi / 4.0) * f t_1 = math.exp(t_0) t_2 = math.exp(-t_0) return -((1.0 / (math.pi / 4.0)) * math.log(((t_1 + t_2) / (t_1 - t_2))))
function code(f) t_0 = Float64(Float64(pi / 4.0) * f) t_1 = exp(t_0) t_2 = exp(Float64(-t_0)) return Float64(-Float64(Float64(1.0 / Float64(pi / 4.0)) * log(Float64(Float64(t_1 + t_2) / Float64(t_1 - t_2))))) end
function tmp = code(f) t_0 = (pi / 4.0) * f; t_1 = exp(t_0); t_2 = exp(-t_0); tmp = -((1.0 / (pi / 4.0)) * log(((t_1 + t_2) / (t_1 - t_2)))); end
code[f_] := Block[{t$95$0 = N[(N[(Pi / 4.0), $MachinePrecision] * f), $MachinePrecision]}, Block[{t$95$1 = N[Exp[t$95$0], $MachinePrecision]}, Block[{t$95$2 = N[Exp[(-t$95$0)], $MachinePrecision]}, (-N[(N[(1.0 / N[(Pi / 4.0), $MachinePrecision]), $MachinePrecision] * N[Log[N[(N[(t$95$1 + t$95$2), $MachinePrecision] / N[(t$95$1 - t$95$2), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]), $MachinePrecision])]]]
\begin{array}{l}
\\
\begin{array}{l}
t_0 := \frac{\pi}{4} \cdot f\\
t_1 := e^{t\_0}\\
t_2 := e^{-t\_0}\\
-\frac{1}{\frac{\pi}{4}} \cdot \log \left(\frac{t\_1 + t\_2}{t\_1 - t\_2}\right)
\end{array}
\end{array}
Herbie found 6 alternatives:
| Alternative | Accuracy | Speedup |
|---|
(FPCore (f) :precision binary64 (let* ((t_0 (* (/ PI 4.0) f)) (t_1 (exp t_0)) (t_2 (exp (- t_0)))) (- (* (/ 1.0 (/ PI 4.0)) (log (/ (+ t_1 t_2) (- t_1 t_2)))))))
double code(double f) {
double t_0 = (((double) M_PI) / 4.0) * f;
double t_1 = exp(t_0);
double t_2 = exp(-t_0);
return -((1.0 / (((double) M_PI) / 4.0)) * log(((t_1 + t_2) / (t_1 - t_2))));
}
public static double code(double f) {
double t_0 = (Math.PI / 4.0) * f;
double t_1 = Math.exp(t_0);
double t_2 = Math.exp(-t_0);
return -((1.0 / (Math.PI / 4.0)) * Math.log(((t_1 + t_2) / (t_1 - t_2))));
}
def code(f): t_0 = (math.pi / 4.0) * f t_1 = math.exp(t_0) t_2 = math.exp(-t_0) return -((1.0 / (math.pi / 4.0)) * math.log(((t_1 + t_2) / (t_1 - t_2))))
function code(f) t_0 = Float64(Float64(pi / 4.0) * f) t_1 = exp(t_0) t_2 = exp(Float64(-t_0)) return Float64(-Float64(Float64(1.0 / Float64(pi / 4.0)) * log(Float64(Float64(t_1 + t_2) / Float64(t_1 - t_2))))) end
function tmp = code(f) t_0 = (pi / 4.0) * f; t_1 = exp(t_0); t_2 = exp(-t_0); tmp = -((1.0 / (pi / 4.0)) * log(((t_1 + t_2) / (t_1 - t_2)))); end
code[f_] := Block[{t$95$0 = N[(N[(Pi / 4.0), $MachinePrecision] * f), $MachinePrecision]}, Block[{t$95$1 = N[Exp[t$95$0], $MachinePrecision]}, Block[{t$95$2 = N[Exp[(-t$95$0)], $MachinePrecision]}, (-N[(N[(1.0 / N[(Pi / 4.0), $MachinePrecision]), $MachinePrecision] * N[Log[N[(N[(t$95$1 + t$95$2), $MachinePrecision] / N[(t$95$1 - t$95$2), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]), $MachinePrecision])]]]
\begin{array}{l}
\\
\begin{array}{l}
t_0 := \frac{\pi}{4} \cdot f\\
t_1 := e^{t\_0}\\
t_2 := e^{-t\_0}\\
-\frac{1}{\frac{\pi}{4}} \cdot \log \left(\frac{t\_1 + t\_2}{t\_1 - t\_2}\right)
\end{array}
\end{array}
(FPCore (f)
:precision binary64
(let* ((t_0 (* (* PI f) -0.25)) (t_1 (exp t_0)))
(if (<= f 5.0)
(/ (* (log (/ (cosh t_0) (sinh (* (* PI f) 0.25)))) -4.0) PI)
(* (/ 4.0 PI) (- (log (/ (+ 1.0 t_1) (- 1.0 t_1))))))))
double code(double f) {
double t_0 = (((double) M_PI) * f) * -0.25;
double t_1 = exp(t_0);
double tmp;
if (f <= 5.0) {
tmp = (log((cosh(t_0) / sinh(((((double) M_PI) * f) * 0.25)))) * -4.0) / ((double) M_PI);
} else {
tmp = (4.0 / ((double) M_PI)) * -log(((1.0 + t_1) / (1.0 - t_1)));
}
return tmp;
}
public static double code(double f) {
double t_0 = (Math.PI * f) * -0.25;
double t_1 = Math.exp(t_0);
double tmp;
if (f <= 5.0) {
tmp = (Math.log((Math.cosh(t_0) / Math.sinh(((Math.PI * f) * 0.25)))) * -4.0) / Math.PI;
} else {
tmp = (4.0 / Math.PI) * -Math.log(((1.0 + t_1) / (1.0 - t_1)));
}
return tmp;
}
def code(f): t_0 = (math.pi * f) * -0.25 t_1 = math.exp(t_0) tmp = 0 if f <= 5.0: tmp = (math.log((math.cosh(t_0) / math.sinh(((math.pi * f) * 0.25)))) * -4.0) / math.pi else: tmp = (4.0 / math.pi) * -math.log(((1.0 + t_1) / (1.0 - t_1))) return tmp
function code(f) t_0 = Float64(Float64(pi * f) * -0.25) t_1 = exp(t_0) tmp = 0.0 if (f <= 5.0) tmp = Float64(Float64(log(Float64(cosh(t_0) / sinh(Float64(Float64(pi * f) * 0.25)))) * -4.0) / pi); else tmp = Float64(Float64(4.0 / pi) * Float64(-log(Float64(Float64(1.0 + t_1) / Float64(1.0 - t_1))))); end return tmp end
function tmp_2 = code(f) t_0 = (pi * f) * -0.25; t_1 = exp(t_0); tmp = 0.0; if (f <= 5.0) tmp = (log((cosh(t_0) / sinh(((pi * f) * 0.25)))) * -4.0) / pi; else tmp = (4.0 / pi) * -log(((1.0 + t_1) / (1.0 - t_1))); end tmp_2 = tmp; end
code[f_] := Block[{t$95$0 = N[(N[(Pi * f), $MachinePrecision] * -0.25), $MachinePrecision]}, Block[{t$95$1 = N[Exp[t$95$0], $MachinePrecision]}, If[LessEqual[f, 5.0], N[(N[(N[Log[N[(N[Cosh[t$95$0], $MachinePrecision] / N[Sinh[N[(N[(Pi * f), $MachinePrecision] * 0.25), $MachinePrecision]], $MachinePrecision]), $MachinePrecision]], $MachinePrecision] * -4.0), $MachinePrecision] / Pi), $MachinePrecision], N[(N[(4.0 / Pi), $MachinePrecision] * (-N[Log[N[(N[(1.0 + t$95$1), $MachinePrecision] / N[(1.0 - t$95$1), $MachinePrecision]), $MachinePrecision]], $MachinePrecision])), $MachinePrecision]]]]
\begin{array}{l}
\\
\begin{array}{l}
t_0 := \left(\pi \cdot f\right) \cdot -0.25\\
t_1 := e^{t\_0}\\
\mathbf{if}\;f \leq 5:\\
\;\;\;\;\frac{\log \left(\frac{\cosh t\_0}{\sinh \left(\left(\pi \cdot f\right) \cdot 0.25\right)}\right) \cdot -4}{\pi}\\
\mathbf{else}:\\
\;\;\;\;\frac{4}{\pi} \cdot \left(-\log \left(\frac{1 + t\_1}{1 - t\_1}\right)\right)\\
\end{array}
\end{array}
if f < 5Initial program 6.9%
Taylor expanded in f around inf
*-commutativeN/A
lower-*.f64N/A
Applied rewrites97.1%
lift-*.f64N/A
Applied rewrites97.1%
if 5 < f Initial program 6.9%
Taylor expanded in f around 0
Applied rewrites5.3%
Taylor expanded in f around 0
Applied rewrites6.2%
lift-neg.f64N/A
lift-*.f64N/A
distribute-rgt-neg-inN/A
lower-*.f64N/A
Applied rewrites6.2%
Taylor expanded in f around 0
*-commutativeN/A
lower-*.f64N/A
*-commutativeN/A
lift-*.f64N/A
lift-PI.f646.2
Applied rewrites6.2%
Taylor expanded in f around 0
*-commutativeN/A
lower-*.f64N/A
*-commutativeN/A
lift-*.f64N/A
lift-PI.f646.2
Applied rewrites6.2%
(FPCore (f) :precision binary64 (/ (* (log (/ (cosh (* (* PI f) -0.25)) (sinh (* (* PI f) 0.25)))) -4.0) PI))
double code(double f) {
return (log((cosh(((((double) M_PI) * f) * -0.25)) / sinh(((((double) M_PI) * f) * 0.25)))) * -4.0) / ((double) M_PI);
}
public static double code(double f) {
return (Math.log((Math.cosh(((Math.PI * f) * -0.25)) / Math.sinh(((Math.PI * f) * 0.25)))) * -4.0) / Math.PI;
}
def code(f): return (math.log((math.cosh(((math.pi * f) * -0.25)) / math.sinh(((math.pi * f) * 0.25)))) * -4.0) / math.pi
function code(f) return Float64(Float64(log(Float64(cosh(Float64(Float64(pi * f) * -0.25)) / sinh(Float64(Float64(pi * f) * 0.25)))) * -4.0) / pi) end
function tmp = code(f) tmp = (log((cosh(((pi * f) * -0.25)) / sinh(((pi * f) * 0.25)))) * -4.0) / pi; end
code[f_] := N[(N[(N[Log[N[(N[Cosh[N[(N[(Pi * f), $MachinePrecision] * -0.25), $MachinePrecision]], $MachinePrecision] / N[Sinh[N[(N[(Pi * f), $MachinePrecision] * 0.25), $MachinePrecision]], $MachinePrecision]), $MachinePrecision]], $MachinePrecision] * -4.0), $MachinePrecision] / Pi), $MachinePrecision]
\begin{array}{l}
\\
\frac{\log \left(\frac{\cosh \left(\left(\pi \cdot f\right) \cdot -0.25\right)}{\sinh \left(\left(\pi \cdot f\right) \cdot 0.25\right)}\right) \cdot -4}{\pi}
\end{array}
Initial program 6.9%
Taylor expanded in f around inf
*-commutativeN/A
lower-*.f64N/A
Applied rewrites97.1%
lift-*.f64N/A
Applied rewrites97.1%
(FPCore (f) :precision binary64 (/ (* (log (/ (fma (* 0.03125 (* f f)) (* PI PI) 1.0) (sinh (* (* PI f) 0.25)))) -4.0) PI))
double code(double f) {
return (log((fma((0.03125 * (f * f)), (((double) M_PI) * ((double) M_PI)), 1.0) / sinh(((((double) M_PI) * f) * 0.25)))) * -4.0) / ((double) M_PI);
}
function code(f) return Float64(Float64(log(Float64(fma(Float64(0.03125 * Float64(f * f)), Float64(pi * pi), 1.0) / sinh(Float64(Float64(pi * f) * 0.25)))) * -4.0) / pi) end
code[f_] := N[(N[(N[Log[N[(N[(N[(0.03125 * N[(f * f), $MachinePrecision]), $MachinePrecision] * N[(Pi * Pi), $MachinePrecision] + 1.0), $MachinePrecision] / N[Sinh[N[(N[(Pi * f), $MachinePrecision] * 0.25), $MachinePrecision]], $MachinePrecision]), $MachinePrecision]], $MachinePrecision] * -4.0), $MachinePrecision] / Pi), $MachinePrecision]
\begin{array}{l}
\\
\frac{\log \left(\frac{\mathsf{fma}\left(0.03125 \cdot \left(f \cdot f\right), \pi \cdot \pi, 1\right)}{\sinh \left(\left(\pi \cdot f\right) \cdot 0.25\right)}\right) \cdot -4}{\pi}
\end{array}
Initial program 6.9%
Taylor expanded in f around inf
*-commutativeN/A
lower-*.f64N/A
Applied rewrites97.1%
lift-*.f64N/A
Applied rewrites97.1%
Taylor expanded in f around 0
+-commutativeN/A
associate-*r*N/A
lower-fma.f64N/A
lower-*.f64N/A
unpow2N/A
lower-*.f64N/A
unpow2N/A
lower-*.f64N/A
lift-PI.f64N/A
lift-PI.f6496.2
Applied rewrites96.2%
(FPCore (f) :precision binary64 (* (log (/ (fma (* (* f f) 0.03125) (* PI PI) 1.0) (sinh (* (* PI f) 0.25)))) (/ -4.0 PI)))
double code(double f) {
return log((fma(((f * f) * 0.03125), (((double) M_PI) * ((double) M_PI)), 1.0) / sinh(((((double) M_PI) * f) * 0.25)))) * (-4.0 / ((double) M_PI));
}
function code(f) return Float64(log(Float64(fma(Float64(Float64(f * f) * 0.03125), Float64(pi * pi), 1.0) / sinh(Float64(Float64(pi * f) * 0.25)))) * Float64(-4.0 / pi)) end
code[f_] := N[(N[Log[N[(N[(N[(N[(f * f), $MachinePrecision] * 0.03125), $MachinePrecision] * N[(Pi * Pi), $MachinePrecision] + 1.0), $MachinePrecision] / N[Sinh[N[(N[(Pi * f), $MachinePrecision] * 0.25), $MachinePrecision]], $MachinePrecision]), $MachinePrecision]], $MachinePrecision] * N[(-4.0 / Pi), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\log \left(\frac{\mathsf{fma}\left(\left(f \cdot f\right) \cdot 0.03125, \pi \cdot \pi, 1\right)}{\sinh \left(\left(\pi \cdot f\right) \cdot 0.25\right)}\right) \cdot \frac{-4}{\pi}
\end{array}
Initial program 6.9%
Taylor expanded in f around inf
*-commutativeN/A
lower-*.f64N/A
Applied rewrites97.1%
lift-*.f64N/A
Applied rewrites97.1%
Taylor expanded in f around 0
+-commutativeN/A
associate-*r*N/A
lower-fma.f64N/A
lower-*.f64N/A
unpow2N/A
lower-*.f64N/A
unpow2N/A
lower-*.f64N/A
lift-PI.f64N/A
lift-PI.f6496.2
Applied rewrites96.2%
lift-PI.f64N/A
lift-/.f64N/A
lift-*.f64N/A
associate-/l*N/A
lower-*.f64N/A
Applied rewrites96.1%
(FPCore (f) :precision binary64 (* (/ (log (/ (fma (* 0.0625 (* f f)) (* PI PI) 2.0) (* f (* PI 0.5)))) PI) -4.0))
double code(double f) {
return (log((fma((0.0625 * (f * f)), (((double) M_PI) * ((double) M_PI)), 2.0) / (f * (((double) M_PI) * 0.5)))) / ((double) M_PI)) * -4.0;
}
function code(f) return Float64(Float64(log(Float64(fma(Float64(0.0625 * Float64(f * f)), Float64(pi * pi), 2.0) / Float64(f * Float64(pi * 0.5)))) / pi) * -4.0) end
code[f_] := N[(N[(N[Log[N[(N[(N[(0.0625 * N[(f * f), $MachinePrecision]), $MachinePrecision] * N[(Pi * Pi), $MachinePrecision] + 2.0), $MachinePrecision] / N[(f * N[(Pi * 0.5), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]], $MachinePrecision] / Pi), $MachinePrecision] * -4.0), $MachinePrecision]
\begin{array}{l}
\\
\frac{\log \left(\frac{\mathsf{fma}\left(0.0625 \cdot \left(f \cdot f\right), \pi \cdot \pi, 2\right)}{f \cdot \left(\pi \cdot 0.5\right)}\right)}{\pi} \cdot -4
\end{array}
Initial program 6.9%
Taylor expanded in f around inf
*-commutativeN/A
lower-*.f64N/A
Applied rewrites97.1%
Taylor expanded in f around 0
lower-*.f64N/A
distribute-rgt-out--N/A
metadata-evalN/A
lift-*.f64N/A
lift-PI.f6495.9
Applied rewrites95.9%
Taylor expanded in f around 0
+-commutativeN/A
associate-*r*N/A
lower-fma.f64N/A
lower-*.f64N/A
unpow2N/A
lower-*.f64N/A
unpow2N/A
lower-*.f64N/A
lift-PI.f64N/A
lift-PI.f6495.9
Applied rewrites95.9%
(FPCore (f) :precision binary64 (* (/ (log (/ 4.0 (* f PI))) PI) -4.0))
double code(double f) {
return (log((4.0 / (f * ((double) M_PI)))) / ((double) M_PI)) * -4.0;
}
public static double code(double f) {
return (Math.log((4.0 / (f * Math.PI))) / Math.PI) * -4.0;
}
def code(f): return (math.log((4.0 / (f * math.pi))) / math.pi) * -4.0
function code(f) return Float64(Float64(log(Float64(4.0 / Float64(f * pi))) / pi) * -4.0) end
function tmp = code(f) tmp = (log((4.0 / (f * pi))) / pi) * -4.0; end
code[f_] := N[(N[(N[Log[N[(4.0 / N[(f * Pi), $MachinePrecision]), $MachinePrecision]], $MachinePrecision] / Pi), $MachinePrecision] * -4.0), $MachinePrecision]
\begin{array}{l}
\\
\frac{\log \left(\frac{4}{f \cdot \pi}\right)}{\pi} \cdot -4
\end{array}
Initial program 6.9%
Taylor expanded in f around 0
*-commutativeN/A
lower-*.f64N/A
Applied rewrites95.8%
Taylor expanded in f around 0
lower-/.f64N/A
lower-*.f64N/A
lift-PI.f6495.8
Applied rewrites95.8%
herbie shell --seed 2025142
(FPCore (f)
:name "VandenBroeck and Keller, Equation (20)"
:precision binary64
(- (* (/ 1.0 (/ PI 4.0)) (log (/ (+ (exp (* (/ PI 4.0) f)) (exp (- (* (/ PI 4.0) f)))) (- (exp (* (/ PI 4.0) f)) (exp (- (* (/ PI 4.0) f)))))))))