
(FPCore (f) :precision binary64 (let* ((t_0 (* (/ PI 4.0) f)) (t_1 (exp t_0)) (t_2 (exp (- t_0)))) (- (* (/ 1.0 (/ PI 4.0)) (log (/ (+ t_1 t_2) (- t_1 t_2)))))))
double code(double f) {
double t_0 = (((double) M_PI) / 4.0) * f;
double t_1 = exp(t_0);
double t_2 = exp(-t_0);
return -((1.0 / (((double) M_PI) / 4.0)) * log(((t_1 + t_2) / (t_1 - t_2))));
}
public static double code(double f) {
double t_0 = (Math.PI / 4.0) * f;
double t_1 = Math.exp(t_0);
double t_2 = Math.exp(-t_0);
return -((1.0 / (Math.PI / 4.0)) * Math.log(((t_1 + t_2) / (t_1 - t_2))));
}
def code(f): t_0 = (math.pi / 4.0) * f t_1 = math.exp(t_0) t_2 = math.exp(-t_0) return -((1.0 / (math.pi / 4.0)) * math.log(((t_1 + t_2) / (t_1 - t_2))))
function code(f) t_0 = Float64(Float64(pi / 4.0) * f) t_1 = exp(t_0) t_2 = exp(Float64(-t_0)) return Float64(-Float64(Float64(1.0 / Float64(pi / 4.0)) * log(Float64(Float64(t_1 + t_2) / Float64(t_1 - t_2))))) end
function tmp = code(f) t_0 = (pi / 4.0) * f; t_1 = exp(t_0); t_2 = exp(-t_0); tmp = -((1.0 / (pi / 4.0)) * log(((t_1 + t_2) / (t_1 - t_2)))); end
code[f_] := Block[{t$95$0 = N[(N[(Pi / 4.0), $MachinePrecision] * f), $MachinePrecision]}, Block[{t$95$1 = N[Exp[t$95$0], $MachinePrecision]}, Block[{t$95$2 = N[Exp[(-t$95$0)], $MachinePrecision]}, (-N[(N[(1.0 / N[(Pi / 4.0), $MachinePrecision]), $MachinePrecision] * N[Log[N[(N[(t$95$1 + t$95$2), $MachinePrecision] / N[(t$95$1 - t$95$2), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]), $MachinePrecision])]]]
\begin{array}{l}
\\
\begin{array}{l}
t_0 := \frac{\pi}{4} \cdot f\\
t_1 := e^{t\_0}\\
t_2 := e^{-t\_0}\\
-\frac{1}{\frac{\pi}{4}} \cdot \log \left(\frac{t\_1 + t\_2}{t\_1 - t\_2}\right)
\end{array}
\end{array}
Sampling outcomes in binary64 precision:
Herbie found 7 alternatives:
| Alternative | Accuracy | Speedup |
|---|
(FPCore (f) :precision binary64 (let* ((t_0 (* (/ PI 4.0) f)) (t_1 (exp t_0)) (t_2 (exp (- t_0)))) (- (* (/ 1.0 (/ PI 4.0)) (log (/ (+ t_1 t_2) (- t_1 t_2)))))))
double code(double f) {
double t_0 = (((double) M_PI) / 4.0) * f;
double t_1 = exp(t_0);
double t_2 = exp(-t_0);
return -((1.0 / (((double) M_PI) / 4.0)) * log(((t_1 + t_2) / (t_1 - t_2))));
}
public static double code(double f) {
double t_0 = (Math.PI / 4.0) * f;
double t_1 = Math.exp(t_0);
double t_2 = Math.exp(-t_0);
return -((1.0 / (Math.PI / 4.0)) * Math.log(((t_1 + t_2) / (t_1 - t_2))));
}
def code(f): t_0 = (math.pi / 4.0) * f t_1 = math.exp(t_0) t_2 = math.exp(-t_0) return -((1.0 / (math.pi / 4.0)) * math.log(((t_1 + t_2) / (t_1 - t_2))))
function code(f) t_0 = Float64(Float64(pi / 4.0) * f) t_1 = exp(t_0) t_2 = exp(Float64(-t_0)) return Float64(-Float64(Float64(1.0 / Float64(pi / 4.0)) * log(Float64(Float64(t_1 + t_2) / Float64(t_1 - t_2))))) end
function tmp = code(f) t_0 = (pi / 4.0) * f; t_1 = exp(t_0); t_2 = exp(-t_0); tmp = -((1.0 / (pi / 4.0)) * log(((t_1 + t_2) / (t_1 - t_2)))); end
code[f_] := Block[{t$95$0 = N[(N[(Pi / 4.0), $MachinePrecision] * f), $MachinePrecision]}, Block[{t$95$1 = N[Exp[t$95$0], $MachinePrecision]}, Block[{t$95$2 = N[Exp[(-t$95$0)], $MachinePrecision]}, (-N[(N[(1.0 / N[(Pi / 4.0), $MachinePrecision]), $MachinePrecision] * N[Log[N[(N[(t$95$1 + t$95$2), $MachinePrecision] / N[(t$95$1 - t$95$2), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]), $MachinePrecision])]]]
\begin{array}{l}
\\
\begin{array}{l}
t_0 := \frac{\pi}{4} \cdot f\\
t_1 := e^{t\_0}\\
t_2 := e^{-t\_0}\\
-\frac{1}{\frac{\pi}{4}} \cdot \log \left(\frac{t\_1 + t\_2}{t\_1 - t\_2}\right)
\end{array}
\end{array}
(FPCore (f)
:precision binary64
(/
(*
(-
(- (log (* 2.0 (cosh (* -0.25 (* f PI))))) (log 2.0))
(log (sinh (* (* f PI) 0.25))))
-4.0)
PI))
double code(double f) {
return (((log((2.0 * cosh((-0.25 * (f * ((double) M_PI)))))) - log(2.0)) - log(sinh(((f * ((double) M_PI)) * 0.25)))) * -4.0) / ((double) M_PI);
}
public static double code(double f) {
return (((Math.log((2.0 * Math.cosh((-0.25 * (f * Math.PI))))) - Math.log(2.0)) - Math.log(Math.sinh(((f * Math.PI) * 0.25)))) * -4.0) / Math.PI;
}
def code(f): return (((math.log((2.0 * math.cosh((-0.25 * (f * math.pi))))) - math.log(2.0)) - math.log(math.sinh(((f * math.pi) * 0.25)))) * -4.0) / math.pi
function code(f) return Float64(Float64(Float64(Float64(log(Float64(2.0 * cosh(Float64(-0.25 * Float64(f * pi))))) - log(2.0)) - log(sinh(Float64(Float64(f * pi) * 0.25)))) * -4.0) / pi) end
function tmp = code(f) tmp = (((log((2.0 * cosh((-0.25 * (f * pi))))) - log(2.0)) - log(sinh(((f * pi) * 0.25)))) * -4.0) / pi; end
code[f_] := N[(N[(N[(N[(N[Log[N[(2.0 * N[Cosh[N[(-0.25 * N[(f * Pi), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]), $MachinePrecision]], $MachinePrecision] - N[Log[2.0], $MachinePrecision]), $MachinePrecision] - N[Log[N[Sinh[N[(N[(f * Pi), $MachinePrecision] * 0.25), $MachinePrecision]], $MachinePrecision]], $MachinePrecision]), $MachinePrecision] * -4.0), $MachinePrecision] / Pi), $MachinePrecision]
\begin{array}{l}
\\
\frac{\left(\left(\log \left(2 \cdot \cosh \left(-0.25 \cdot \left(f \cdot \pi\right)\right)\right) - \log 2\right) - \log \sinh \left(\left(f \cdot \pi\right) \cdot 0.25\right)\right) \cdot -4}{\pi}
\end{array}
Initial program 9.1%
Taylor expanded in f around inf
*-commutativeN/A
lower-*.f64N/A
Applied rewrites97.8%
Applied rewrites97.8%
lift-log.f64N/A
lift-/.f64N/A
lift-cosh.f64N/A
lift-*.f64N/A
lift-PI.f64N/A
lift-*.f64N/A
lift-sinh.f64N/A
lift-*.f64N/A
lift-PI.f64N/A
lift-*.f64N/A
log-divN/A
lower--.f64N/A
Applied rewrites97.8%
lift-log.f64N/A
lift-cosh.f64N/A
cosh-defN/A
lift-*.f64N/A
lift-PI.f64N/A
lift-*.f64N/A
*-commutativeN/A
lift-*.f64N/A
lift-PI.f64N/A
lift-*.f64N/A
*-commutativeN/A
rec-expN/A
Applied rewrites97.9%
(FPCore (f) :precision binary64 (let* ((t_0 (* (* f PI) 0.25))) (/ (* (log (/ (cosh t_0) (sinh t_0))) -4.0) PI)))
double code(double f) {
double t_0 = (f * ((double) M_PI)) * 0.25;
return (log((cosh(t_0) / sinh(t_0))) * -4.0) / ((double) M_PI);
}
public static double code(double f) {
double t_0 = (f * Math.PI) * 0.25;
return (Math.log((Math.cosh(t_0) / Math.sinh(t_0))) * -4.0) / Math.PI;
}
def code(f): t_0 = (f * math.pi) * 0.25 return (math.log((math.cosh(t_0) / math.sinh(t_0))) * -4.0) / math.pi
function code(f) t_0 = Float64(Float64(f * pi) * 0.25) return Float64(Float64(log(Float64(cosh(t_0) / sinh(t_0))) * -4.0) / pi) end
function tmp = code(f) t_0 = (f * pi) * 0.25; tmp = (log((cosh(t_0) / sinh(t_0))) * -4.0) / pi; end
code[f_] := Block[{t$95$0 = N[(N[(f * Pi), $MachinePrecision] * 0.25), $MachinePrecision]}, N[(N[(N[Log[N[(N[Cosh[t$95$0], $MachinePrecision] / N[Sinh[t$95$0], $MachinePrecision]), $MachinePrecision]], $MachinePrecision] * -4.0), $MachinePrecision] / Pi), $MachinePrecision]]
\begin{array}{l}
\\
\begin{array}{l}
t_0 := \left(f \cdot \pi\right) \cdot 0.25\\
\frac{\log \left(\frac{\cosh t\_0}{\sinh t\_0}\right) \cdot -4}{\pi}
\end{array}
\end{array}
Initial program 9.1%
Taylor expanded in f around inf
*-commutativeN/A
lower-*.f64N/A
Applied rewrites97.8%
Applied rewrites97.8%
(FPCore (f) :precision binary64 (let* ((t_0 (* (* f PI) 0.25))) (* (log (/ (cosh t_0) (sinh t_0))) (/ -4.0 PI))))
double code(double f) {
double t_0 = (f * ((double) M_PI)) * 0.25;
return log((cosh(t_0) / sinh(t_0))) * (-4.0 / ((double) M_PI));
}
public static double code(double f) {
double t_0 = (f * Math.PI) * 0.25;
return Math.log((Math.cosh(t_0) / Math.sinh(t_0))) * (-4.0 / Math.PI);
}
def code(f): t_0 = (f * math.pi) * 0.25 return math.log((math.cosh(t_0) / math.sinh(t_0))) * (-4.0 / math.pi)
function code(f) t_0 = Float64(Float64(f * pi) * 0.25) return Float64(log(Float64(cosh(t_0) / sinh(t_0))) * Float64(-4.0 / pi)) end
function tmp = code(f) t_0 = (f * pi) * 0.25; tmp = log((cosh(t_0) / sinh(t_0))) * (-4.0 / pi); end
code[f_] := Block[{t$95$0 = N[(N[(f * Pi), $MachinePrecision] * 0.25), $MachinePrecision]}, N[(N[Log[N[(N[Cosh[t$95$0], $MachinePrecision] / N[Sinh[t$95$0], $MachinePrecision]), $MachinePrecision]], $MachinePrecision] * N[(-4.0 / Pi), $MachinePrecision]), $MachinePrecision]]
\begin{array}{l}
\\
\begin{array}{l}
t_0 := \left(f \cdot \pi\right) \cdot 0.25\\
\log \left(\frac{\cosh t\_0}{\sinh t\_0}\right) \cdot \frac{-4}{\pi}
\end{array}
\end{array}
Initial program 9.1%
Taylor expanded in f around inf
*-commutativeN/A
lower-*.f64N/A
Applied rewrites97.8%
Applied rewrites97.8%
lift-PI.f64N/A
lift-/.f64N/A
Applied rewrites97.7%
(FPCore (f)
:precision binary64
(/
(log
(/
(fma
(fma (fma (* PI 0.020833333333333332) -2.0 (* 0.125 PI)) f 0.0)
f
(/ 4.0 PI))
f))
(/ (- PI) 4.0)))
double code(double f) {
return log((fma(fma(fma((((double) M_PI) * 0.020833333333333332), -2.0, (0.125 * ((double) M_PI))), f, 0.0), f, (4.0 / ((double) M_PI))) / f)) / (-((double) M_PI) / 4.0);
}
function code(f) return Float64(log(Float64(fma(fma(fma(Float64(pi * 0.020833333333333332), -2.0, Float64(0.125 * pi)), f, 0.0), f, Float64(4.0 / pi)) / f)) / Float64(Float64(-pi) / 4.0)) end
code[f_] := N[(N[Log[N[(N[(N[(N[(N[(Pi * 0.020833333333333332), $MachinePrecision] * -2.0 + N[(0.125 * Pi), $MachinePrecision]), $MachinePrecision] * f + 0.0), $MachinePrecision] * f + N[(4.0 / Pi), $MachinePrecision]), $MachinePrecision] / f), $MachinePrecision]], $MachinePrecision] / N[((-Pi) / 4.0), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\frac{\log \left(\frac{\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(\pi \cdot 0.020833333333333332, -2, 0.125 \cdot \pi\right), f, 0\right), f, \frac{4}{\pi}\right)}{f}\right)}{\frac{-\pi}{4}}
\end{array}
Initial program 9.1%
Taylor expanded in f around 0
Applied rewrites95.6%
Applied rewrites95.7%
Final simplification95.7%
(FPCore (f) :precision binary64 (* (/ -1.0 (/ PI 4.0)) (log (/ (fma (/ 2.0 PI) 2.0 (* (* PI 0.08333333333333333) (* f f))) f))))
double code(double f) {
return (-1.0 / (((double) M_PI) / 4.0)) * log((fma((2.0 / ((double) M_PI)), 2.0, ((((double) M_PI) * 0.08333333333333333) * (f * f))) / f));
}
function code(f) return Float64(Float64(-1.0 / Float64(pi / 4.0)) * log(Float64(fma(Float64(2.0 / pi), 2.0, Float64(Float64(pi * 0.08333333333333333) * Float64(f * f))) / f))) end
code[f_] := N[(N[(-1.0 / N[(Pi / 4.0), $MachinePrecision]), $MachinePrecision] * N[Log[N[(N[(N[(2.0 / Pi), $MachinePrecision] * 2.0 + N[(N[(Pi * 0.08333333333333333), $MachinePrecision] * N[(f * f), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] / f), $MachinePrecision]], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\frac{-1}{\frac{\pi}{4}} \cdot \log \left(\frac{\mathsf{fma}\left(\frac{2}{\pi}, 2, \left(\pi \cdot 0.08333333333333333\right) \cdot \left(f \cdot f\right)\right)}{f}\right)
\end{array}
Initial program 9.1%
Taylor expanded in f around 0
Applied rewrites95.6%
Taylor expanded in f around 0
*-commutativeN/A
lower-*.f64N/A
distribute-rgt-outN/A
lower-*.f64N/A
lift-PI.f64N/A
metadata-evalN/A
unpow2N/A
lower-*.f6495.6
Applied rewrites95.6%
Final simplification95.6%
(FPCore (f) :precision binary64 (* (/ (- (log (/ 2.0 (* 0.5 PI))) (log f)) PI) -4.0))
double code(double f) {
return ((log((2.0 / (0.5 * ((double) M_PI)))) - log(f)) / ((double) M_PI)) * -4.0;
}
public static double code(double f) {
return ((Math.log((2.0 / (0.5 * Math.PI))) - Math.log(f)) / Math.PI) * -4.0;
}
def code(f): return ((math.log((2.0 / (0.5 * math.pi))) - math.log(f)) / math.pi) * -4.0
function code(f) return Float64(Float64(Float64(log(Float64(2.0 / Float64(0.5 * pi))) - log(f)) / pi) * -4.0) end
function tmp = code(f) tmp = ((log((2.0 / (0.5 * pi))) - log(f)) / pi) * -4.0; end
code[f_] := N[(N[(N[(N[Log[N[(2.0 / N[(0.5 * Pi), $MachinePrecision]), $MachinePrecision]], $MachinePrecision] - N[Log[f], $MachinePrecision]), $MachinePrecision] / Pi), $MachinePrecision] * -4.0), $MachinePrecision]
\begin{array}{l}
\\
\frac{\log \left(\frac{2}{0.5 \cdot \pi}\right) - \log f}{\pi} \cdot -4
\end{array}
Initial program 9.1%
Taylor expanded in f around 0
*-commutativeN/A
lower-*.f64N/A
Applied rewrites95.2%
lift-log.f64N/A
lift-/.f64N/A
lift-*.f64N/A
lift-PI.f64N/A
lift-*.f64N/A
metadata-evalN/A
distribute-rgt-out--N/A
associate-/r*N/A
metadata-evalN/A
associate-*r/N/A
log-divN/A
associate-*r/N/A
metadata-evalN/A
lower--.f64N/A
Applied rewrites95.2%
(FPCore (f) :precision binary64 (* (/ (log (/ 4.0 (* f PI))) PI) -4.0))
double code(double f) {
return (log((4.0 / (f * ((double) M_PI)))) / ((double) M_PI)) * -4.0;
}
public static double code(double f) {
return (Math.log((4.0 / (f * Math.PI))) / Math.PI) * -4.0;
}
def code(f): return (math.log((4.0 / (f * math.pi))) / math.pi) * -4.0
function code(f) return Float64(Float64(log(Float64(4.0 / Float64(f * pi))) / pi) * -4.0) end
function tmp = code(f) tmp = (log((4.0 / (f * pi))) / pi) * -4.0; end
code[f_] := N[(N[(N[Log[N[(4.0 / N[(f * Pi), $MachinePrecision]), $MachinePrecision]], $MachinePrecision] / Pi), $MachinePrecision] * -4.0), $MachinePrecision]
\begin{array}{l}
\\
\frac{\log \left(\frac{4}{f \cdot \pi}\right)}{\pi} \cdot -4
\end{array}
Initial program 9.1%
Taylor expanded in f around 0
*-commutativeN/A
lower-*.f64N/A
Applied rewrites95.2%
Taylor expanded in f around 0
lower-/.f64N/A
lower-*.f64N/A
lift-PI.f6495.2
Applied rewrites95.2%
herbie shell --seed 2025064
(FPCore (f)
:name "VandenBroeck and Keller, Equation (20)"
:precision binary64
(- (* (/ 1.0 (/ PI 4.0)) (log (/ (+ (exp (* (/ PI 4.0) f)) (exp (- (* (/ PI 4.0) f)))) (- (exp (* (/ PI 4.0) f)) (exp (- (* (/ PI 4.0) f)))))))))