
(FPCore (f) :precision binary64 (let* ((t_0 (* (/ PI 4.0) f)) (t_1 (exp t_0)) (t_2 (exp (- t_0)))) (- (* (/ 1.0 (/ PI 4.0)) (log (/ (+ t_1 t_2) (- t_1 t_2)))))))
double code(double f) {
double t_0 = (((double) M_PI) / 4.0) * f;
double t_1 = exp(t_0);
double t_2 = exp(-t_0);
return -((1.0 / (((double) M_PI) / 4.0)) * log(((t_1 + t_2) / (t_1 - t_2))));
}
public static double code(double f) {
double t_0 = (Math.PI / 4.0) * f;
double t_1 = Math.exp(t_0);
double t_2 = Math.exp(-t_0);
return -((1.0 / (Math.PI / 4.0)) * Math.log(((t_1 + t_2) / (t_1 - t_2))));
}
def code(f): t_0 = (math.pi / 4.0) * f t_1 = math.exp(t_0) t_2 = math.exp(-t_0) return -((1.0 / (math.pi / 4.0)) * math.log(((t_1 + t_2) / (t_1 - t_2))))
function code(f) t_0 = Float64(Float64(pi / 4.0) * f) t_1 = exp(t_0) t_2 = exp(Float64(-t_0)) return Float64(-Float64(Float64(1.0 / Float64(pi / 4.0)) * log(Float64(Float64(t_1 + t_2) / Float64(t_1 - t_2))))) end
function tmp = code(f) t_0 = (pi / 4.0) * f; t_1 = exp(t_0); t_2 = exp(-t_0); tmp = -((1.0 / (pi / 4.0)) * log(((t_1 + t_2) / (t_1 - t_2)))); end
code[f_] := Block[{t$95$0 = N[(N[(Pi / 4.0), $MachinePrecision] * f), $MachinePrecision]}, Block[{t$95$1 = N[Exp[t$95$0], $MachinePrecision]}, Block[{t$95$2 = N[Exp[(-t$95$0)], $MachinePrecision]}, (-N[(N[(1.0 / N[(Pi / 4.0), $MachinePrecision]), $MachinePrecision] * N[Log[N[(N[(t$95$1 + t$95$2), $MachinePrecision] / N[(t$95$1 - t$95$2), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]), $MachinePrecision])]]]
\begin{array}{l}
\\
\begin{array}{l}
t_0 := \frac{\pi}{4} \cdot f\\
t_1 := e^{t_0}\\
t_2 := e^{-t_0}\\
-\frac{1}{\frac{\pi}{4}} \cdot \log \left(\frac{t_1 + t_2}{t_1 - t_2}\right)
\end{array}
\end{array}
Sampling outcomes in binary64 precision:
Herbie found 11 alternatives:
| Alternative | Accuracy | Speedup |
|---|
(FPCore (f) :precision binary64 (let* ((t_0 (* (/ PI 4.0) f)) (t_1 (exp t_0)) (t_2 (exp (- t_0)))) (- (* (/ 1.0 (/ PI 4.0)) (log (/ (+ t_1 t_2) (- t_1 t_2)))))))
double code(double f) {
double t_0 = (((double) M_PI) / 4.0) * f;
double t_1 = exp(t_0);
double t_2 = exp(-t_0);
return -((1.0 / (((double) M_PI) / 4.0)) * log(((t_1 + t_2) / (t_1 - t_2))));
}
public static double code(double f) {
double t_0 = (Math.PI / 4.0) * f;
double t_1 = Math.exp(t_0);
double t_2 = Math.exp(-t_0);
return -((1.0 / (Math.PI / 4.0)) * Math.log(((t_1 + t_2) / (t_1 - t_2))));
}
def code(f): t_0 = (math.pi / 4.0) * f t_1 = math.exp(t_0) t_2 = math.exp(-t_0) return -((1.0 / (math.pi / 4.0)) * math.log(((t_1 + t_2) / (t_1 - t_2))))
function code(f) t_0 = Float64(Float64(pi / 4.0) * f) t_1 = exp(t_0) t_2 = exp(Float64(-t_0)) return Float64(-Float64(Float64(1.0 / Float64(pi / 4.0)) * log(Float64(Float64(t_1 + t_2) / Float64(t_1 - t_2))))) end
function tmp = code(f) t_0 = (pi / 4.0) * f; t_1 = exp(t_0); t_2 = exp(-t_0); tmp = -((1.0 / (pi / 4.0)) * log(((t_1 + t_2) / (t_1 - t_2)))); end
code[f_] := Block[{t$95$0 = N[(N[(Pi / 4.0), $MachinePrecision] * f), $MachinePrecision]}, Block[{t$95$1 = N[Exp[t$95$0], $MachinePrecision]}, Block[{t$95$2 = N[Exp[(-t$95$0)], $MachinePrecision]}, (-N[(N[(1.0 / N[(Pi / 4.0), $MachinePrecision]), $MachinePrecision] * N[Log[N[(N[(t$95$1 + t$95$2), $MachinePrecision] / N[(t$95$1 - t$95$2), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]), $MachinePrecision])]]]
\begin{array}{l}
\\
\begin{array}{l}
t_0 := \frac{\pi}{4} \cdot f\\
t_1 := e^{t_0}\\
t_2 := e^{-t_0}\\
-\frac{1}{\frac{\pi}{4}} \cdot \log \left(\frac{t_1 + t_2}{t_1 - t_2}\right)
\end{array}
\end{array}
(FPCore (f)
:precision binary64
(*
(log
(/
(+ (exp (* (/ PI 4.0) f)) (exp (* f (- (/ PI 4.0)))))
(fma
f
(* PI 0.5)
(fma
(pow f 5.0)
(* (pow PI 5.0) 1.6276041666666666e-5)
(fma
(pow f 3.0)
(* (pow PI 3.0) 0.005208333333333333)
(* (pow f 7.0) (* (pow PI 7.0) 2.422030009920635e-8)))))))
(/ -1.0 (/ PI 4.0))))
double code(double f) {
return log(((exp(((((double) M_PI) / 4.0) * f)) + exp((f * -(((double) M_PI) / 4.0)))) / fma(f, (((double) M_PI) * 0.5), fma(pow(f, 5.0), (pow(((double) M_PI), 5.0) * 1.6276041666666666e-5), fma(pow(f, 3.0), (pow(((double) M_PI), 3.0) * 0.005208333333333333), (pow(f, 7.0) * (pow(((double) M_PI), 7.0) * 2.422030009920635e-8))))))) * (-1.0 / (((double) M_PI) / 4.0));
}
function code(f) return Float64(log(Float64(Float64(exp(Float64(Float64(pi / 4.0) * f)) + exp(Float64(f * Float64(-Float64(pi / 4.0))))) / fma(f, Float64(pi * 0.5), fma((f ^ 5.0), Float64((pi ^ 5.0) * 1.6276041666666666e-5), fma((f ^ 3.0), Float64((pi ^ 3.0) * 0.005208333333333333), Float64((f ^ 7.0) * Float64((pi ^ 7.0) * 2.422030009920635e-8))))))) * Float64(-1.0 / Float64(pi / 4.0))) end
code[f_] := N[(N[Log[N[(N[(N[Exp[N[(N[(Pi / 4.0), $MachinePrecision] * f), $MachinePrecision]], $MachinePrecision] + N[Exp[N[(f * (-N[(Pi / 4.0), $MachinePrecision])), $MachinePrecision]], $MachinePrecision]), $MachinePrecision] / N[(f * N[(Pi * 0.5), $MachinePrecision] + N[(N[Power[f, 5.0], $MachinePrecision] * N[(N[Power[Pi, 5.0], $MachinePrecision] * 1.6276041666666666e-5), $MachinePrecision] + N[(N[Power[f, 3.0], $MachinePrecision] * N[(N[Power[Pi, 3.0], $MachinePrecision] * 0.005208333333333333), $MachinePrecision] + N[(N[Power[f, 7.0], $MachinePrecision] * N[(N[Power[Pi, 7.0], $MachinePrecision] * 2.422030009920635e-8), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]], $MachinePrecision] * N[(-1.0 / N[(Pi / 4.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\log \left(\frac{e^{\frac{\pi}{4} \cdot f} + e^{f \cdot \left(-\frac{\pi}{4}\right)}}{\mathsf{fma}\left(f, \pi \cdot 0.5, \mathsf{fma}\left({f}^{5}, {\pi}^{5} \cdot 1.6276041666666666 \cdot 10^{-5}, \mathsf{fma}\left({f}^{3}, {\pi}^{3} \cdot 0.005208333333333333, {f}^{7} \cdot \left({\pi}^{7} \cdot 2.422030009920635 \cdot 10^{-8}\right)\right)\right)\right)}\right) \cdot \frac{-1}{\frac{\pi}{4}}
\end{array}
Initial program 6.9%
Taylor expanded in f around 0 97.5%
fma-def97.5%
distribute-rgt-out--97.5%
metadata-eval97.5%
associate-+r+97.5%
+-commutative97.5%
Simplified97.5%
Final simplification97.5%
(FPCore (f)
:precision binary64
(*
(log
(/
(+ (exp (* (/ PI 4.0) f)) (exp (* f (- (/ PI 4.0)))))
(fma
(pow f 3.0)
(* (pow PI 3.0) 0.005208333333333333)
(fma
f
(* PI 0.5)
(* (pow f 5.0) (* (pow PI 5.0) 1.6276041666666666e-5))))))
(/ -1.0 (/ PI 4.0))))
double code(double f) {
return log(((exp(((((double) M_PI) / 4.0) * f)) + exp((f * -(((double) M_PI) / 4.0)))) / fma(pow(f, 3.0), (pow(((double) M_PI), 3.0) * 0.005208333333333333), fma(f, (((double) M_PI) * 0.5), (pow(f, 5.0) * (pow(((double) M_PI), 5.0) * 1.6276041666666666e-5)))))) * (-1.0 / (((double) M_PI) / 4.0));
}
function code(f) return Float64(log(Float64(Float64(exp(Float64(Float64(pi / 4.0) * f)) + exp(Float64(f * Float64(-Float64(pi / 4.0))))) / fma((f ^ 3.0), Float64((pi ^ 3.0) * 0.005208333333333333), fma(f, Float64(pi * 0.5), Float64((f ^ 5.0) * Float64((pi ^ 5.0) * 1.6276041666666666e-5)))))) * Float64(-1.0 / Float64(pi / 4.0))) end
code[f_] := N[(N[Log[N[(N[(N[Exp[N[(N[(Pi / 4.0), $MachinePrecision] * f), $MachinePrecision]], $MachinePrecision] + N[Exp[N[(f * (-N[(Pi / 4.0), $MachinePrecision])), $MachinePrecision]], $MachinePrecision]), $MachinePrecision] / N[(N[Power[f, 3.0], $MachinePrecision] * N[(N[Power[Pi, 3.0], $MachinePrecision] * 0.005208333333333333), $MachinePrecision] + N[(f * N[(Pi * 0.5), $MachinePrecision] + N[(N[Power[f, 5.0], $MachinePrecision] * N[(N[Power[Pi, 5.0], $MachinePrecision] * 1.6276041666666666e-5), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]], $MachinePrecision] * N[(-1.0 / N[(Pi / 4.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\log \left(\frac{e^{\frac{\pi}{4} \cdot f} + e^{f \cdot \left(-\frac{\pi}{4}\right)}}{\mathsf{fma}\left({f}^{3}, {\pi}^{3} \cdot 0.005208333333333333, \mathsf{fma}\left(f, \pi \cdot 0.5, {f}^{5} \cdot \left({\pi}^{5} \cdot 1.6276041666666666 \cdot 10^{-5}\right)\right)\right)}\right) \cdot \frac{-1}{\frac{\pi}{4}}
\end{array}
Initial program 6.9%
Taylor expanded in f around 0 97.4%
associate-+r+97.4%
+-commutative97.4%
associate-+l+97.4%
fma-def97.4%
distribute-rgt-out--97.4%
metadata-eval97.4%
Simplified97.4%
Final simplification97.4%
(FPCore (f)
:precision binary64
(fma
(/ (- (log (/ 4.0 PI)) (log f)) PI)
-4.0
(fma
(*
(/ (pow f 2.0) PI)
(fma
PI
(*
0.5
(fma
(/ 0.005208333333333333 (* 0.5 (/ 0.5 PI)))
-2.0
(* 0.0625 (* PI 2.0))))
0.0))
-2.0
0.0)))
double code(double f) {
return fma(((log((4.0 / ((double) M_PI))) - log(f)) / ((double) M_PI)), -4.0, fma(((pow(f, 2.0) / ((double) M_PI)) * fma(((double) M_PI), (0.5 * fma((0.005208333333333333 / (0.5 * (0.5 / ((double) M_PI)))), -2.0, (0.0625 * (((double) M_PI) * 2.0)))), 0.0)), -2.0, 0.0));
}
function code(f) return fma(Float64(Float64(log(Float64(4.0 / pi)) - log(f)) / pi), -4.0, fma(Float64(Float64((f ^ 2.0) / pi) * fma(pi, Float64(0.5 * fma(Float64(0.005208333333333333 / Float64(0.5 * Float64(0.5 / pi))), -2.0, Float64(0.0625 * Float64(pi * 2.0)))), 0.0)), -2.0, 0.0)) end
code[f_] := N[(N[(N[(N[Log[N[(4.0 / Pi), $MachinePrecision]], $MachinePrecision] - N[Log[f], $MachinePrecision]), $MachinePrecision] / Pi), $MachinePrecision] * -4.0 + N[(N[(N[(N[Power[f, 2.0], $MachinePrecision] / Pi), $MachinePrecision] * N[(Pi * N[(0.5 * N[(N[(0.005208333333333333 / N[(0.5 * N[(0.5 / Pi), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] * -2.0 + N[(0.0625 * N[(Pi * 2.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] + 0.0), $MachinePrecision]), $MachinePrecision] * -2.0 + 0.0), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\mathsf{fma}\left(\frac{\log \left(\frac{4}{\pi}\right) - \log f}{\pi}, -4, \mathsf{fma}\left(\frac{{f}^{2}}{\pi} \cdot \mathsf{fma}\left(\pi, 0.5 \cdot \mathsf{fma}\left(\frac{0.005208333333333333}{0.5 \cdot \frac{0.5}{\pi}}, -2, 0.0625 \cdot \left(\pi \cdot 2\right)\right), 0\right), -2, 0\right)\right)
\end{array}
Initial program 6.9%
distribute-lft-neg-in6.9%
*-commutative6.9%
Simplified6.9%
Taylor expanded in f around 0 97.4%
Simplified97.4%
Final simplification97.4%
(FPCore (f)
:precision binary64
(*
(log
(fma
f
(fma
(/ 0.005208333333333333 (* 0.5 (/ 0.5 PI)))
-2.0
(* 0.0625 (* PI 2.0)))
(/ (/ 4.0 PI) f)))
(/ -1.0 (/ PI 4.0))))
double code(double f) {
return log(fma(f, fma((0.005208333333333333 / (0.5 * (0.5 / ((double) M_PI)))), -2.0, (0.0625 * (((double) M_PI) * 2.0))), ((4.0 / ((double) M_PI)) / f))) * (-1.0 / (((double) M_PI) / 4.0));
}
function code(f) return Float64(log(fma(f, fma(Float64(0.005208333333333333 / Float64(0.5 * Float64(0.5 / pi))), -2.0, Float64(0.0625 * Float64(pi * 2.0))), Float64(Float64(4.0 / pi) / f))) * Float64(-1.0 / Float64(pi / 4.0))) end
code[f_] := N[(N[Log[N[(f * N[(N[(0.005208333333333333 / N[(0.5 * N[(0.5 / Pi), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] * -2.0 + N[(0.0625 * N[(Pi * 2.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] + N[(N[(4.0 / Pi), $MachinePrecision] / f), $MachinePrecision]), $MachinePrecision]], $MachinePrecision] * N[(-1.0 / N[(Pi / 4.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\log \left(\mathsf{fma}\left(f, \mathsf{fma}\left(\frac{0.005208333333333333}{0.5 \cdot \frac{0.5}{\pi}}, -2, 0.0625 \cdot \left(\pi \cdot 2\right)\right), \frac{\frac{4}{\pi}}{f}\right)\right) \cdot \frac{-1}{\frac{\pi}{4}}
\end{array}
Initial program 6.9%
Taylor expanded in f around 0 97.3%
Simplified97.3%
Final simplification97.3%
(FPCore (f) :precision binary64 (- (* 4.0 (/ (- (log f) (log (/ 4.0 PI))) PI)) (* 0.125 (* PI (pow f 2.0)))))
double code(double f) {
return (4.0 * ((log(f) - log((4.0 / ((double) M_PI)))) / ((double) M_PI))) - (0.125 * (((double) M_PI) * pow(f, 2.0)));
}
public static double code(double f) {
return (4.0 * ((Math.log(f) - Math.log((4.0 / Math.PI))) / Math.PI)) - (0.125 * (Math.PI * Math.pow(f, 2.0)));
}
def code(f): return (4.0 * ((math.log(f) - math.log((4.0 / math.pi))) / math.pi)) - (0.125 * (math.pi * math.pow(f, 2.0)))
function code(f) return Float64(Float64(4.0 * Float64(Float64(log(f) - log(Float64(4.0 / pi))) / pi)) - Float64(0.125 * Float64(pi * (f ^ 2.0)))) end
function tmp = code(f) tmp = (4.0 * ((log(f) - log((4.0 / pi))) / pi)) - (0.125 * (pi * (f ^ 2.0))); end
code[f_] := N[(N[(4.0 * N[(N[(N[Log[f], $MachinePrecision] - N[Log[N[(4.0 / Pi), $MachinePrecision]], $MachinePrecision]), $MachinePrecision] / Pi), $MachinePrecision]), $MachinePrecision] - N[(0.125 * N[(Pi * N[Power[f, 2.0], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
4 \cdot \frac{\log f - \log \left(\frac{4}{\pi}\right)}{\pi} - 0.125 \cdot \left(\pi \cdot {f}^{2}\right)
\end{array}
Initial program 6.9%
Taylor expanded in f around 0 96.6%
distribute-rgt-out--96.6%
metadata-eval96.6%
Simplified96.6%
Taylor expanded in f around 0 96.8%
Final simplification96.8%
(FPCore (f) :precision binary64 (/ (- (log (fma 0.125 (* PI f) (/ (/ 4.0 PI) f)))) (* PI 0.25)))
double code(double f) {
return -log(fma(0.125, (((double) M_PI) * f), ((4.0 / ((double) M_PI)) / f))) / (((double) M_PI) * 0.25);
}
function code(f) return Float64(Float64(-log(fma(0.125, Float64(pi * f), Float64(Float64(4.0 / pi) / f)))) / Float64(pi * 0.25)) end
code[f_] := N[((-N[Log[N[(0.125 * N[(Pi * f), $MachinePrecision] + N[(N[(4.0 / Pi), $MachinePrecision] / f), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]) / N[(Pi * 0.25), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\frac{-\log \left(\mathsf{fma}\left(0.125, \pi \cdot f, \frac{\frac{4}{\pi}}{f}\right)\right)}{\pi \cdot 0.25}
\end{array}
Initial program 6.9%
Taylor expanded in f around 0 96.6%
distribute-rgt-out--96.6%
metadata-eval96.6%
Simplified96.6%
Taylor expanded in f around 0 96.6%
rem-cbrt-cube96.3%
Applied egg-rr96.3%
associate-*l/96.4%
*-un-lft-identity96.4%
fma-def96.4%
*-commutative96.4%
*-commutative96.4%
div-inv96.4%
associate-/r*96.4%
rem-cbrt-cube96.7%
div-inv96.7%
metadata-eval96.7%
Applied egg-rr96.7%
Final simplification96.7%
(FPCore (f) :precision binary64 (/ (* -4.0 (- (log (/ 2.0 (* PI 0.5))) (log f))) PI))
double code(double f) {
return (-4.0 * (log((2.0 / (((double) M_PI) * 0.5))) - log(f))) / ((double) M_PI);
}
public static double code(double f) {
return (-4.0 * (Math.log((2.0 / (Math.PI * 0.5))) - Math.log(f))) / Math.PI;
}
def code(f): return (-4.0 * (math.log((2.0 / (math.pi * 0.5))) - math.log(f))) / math.pi
function code(f) return Float64(Float64(-4.0 * Float64(log(Float64(2.0 / Float64(pi * 0.5))) - log(f))) / pi) end
function tmp = code(f) tmp = (-4.0 * (log((2.0 / (pi * 0.5))) - log(f))) / pi; end
code[f_] := N[(N[(-4.0 * N[(N[Log[N[(2.0 / N[(Pi * 0.5), $MachinePrecision]), $MachinePrecision]], $MachinePrecision] - N[Log[f], $MachinePrecision]), $MachinePrecision]), $MachinePrecision] / Pi), $MachinePrecision]
\begin{array}{l}
\\
\frac{-4 \cdot \left(\log \left(\frac{2}{\pi \cdot 0.5}\right) - \log f\right)}{\pi}
\end{array}
Initial program 6.9%
distribute-lft-neg-in6.9%
*-commutative6.9%
Simplified6.9%
Taylor expanded in f around 0 96.6%
*-commutative96.6%
associate-*l/96.6%
mul-1-neg96.6%
unsub-neg96.6%
distribute-rgt-out--96.6%
metadata-eval96.6%
Simplified96.6%
Final simplification96.6%
(FPCore (f) :precision binary64 (* (log (+ (* 0.125 (* PI f)) (* 4.0 (/ 1.0 (* PI f))))) (/ -1.0 (/ PI 4.0))))
double code(double f) {
return log(((0.125 * (((double) M_PI) * f)) + (4.0 * (1.0 / (((double) M_PI) * f))))) * (-1.0 / (((double) M_PI) / 4.0));
}
public static double code(double f) {
return Math.log(((0.125 * (Math.PI * f)) + (4.0 * (1.0 / (Math.PI * f))))) * (-1.0 / (Math.PI / 4.0));
}
def code(f): return math.log(((0.125 * (math.pi * f)) + (4.0 * (1.0 / (math.pi * f))))) * (-1.0 / (math.pi / 4.0))
function code(f) return Float64(log(Float64(Float64(0.125 * Float64(pi * f)) + Float64(4.0 * Float64(1.0 / Float64(pi * f))))) * Float64(-1.0 / Float64(pi / 4.0))) end
function tmp = code(f) tmp = log(((0.125 * (pi * f)) + (4.0 * (1.0 / (pi * f))))) * (-1.0 / (pi / 4.0)); end
code[f_] := N[(N[Log[N[(N[(0.125 * N[(Pi * f), $MachinePrecision]), $MachinePrecision] + N[(4.0 * N[(1.0 / N[(Pi * f), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]], $MachinePrecision] * N[(-1.0 / N[(Pi / 4.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\log \left(0.125 \cdot \left(\pi \cdot f\right) + 4 \cdot \frac{1}{\pi \cdot f}\right) \cdot \frac{-1}{\frac{\pi}{4}}
\end{array}
Initial program 6.9%
Taylor expanded in f around 0 96.6%
distribute-rgt-out--96.6%
metadata-eval96.6%
Simplified96.6%
Taylor expanded in f around 0 96.6%
Final simplification96.6%
(FPCore (f) :precision binary64 (/ (* -4.0 (log1p (+ -1.0 (/ (/ 2.0 (* PI 0.5)) f)))) PI))
double code(double f) {
return (-4.0 * log1p((-1.0 + ((2.0 / (((double) M_PI) * 0.5)) / f)))) / ((double) M_PI);
}
public static double code(double f) {
return (-4.0 * Math.log1p((-1.0 + ((2.0 / (Math.PI * 0.5)) / f)))) / Math.PI;
}
def code(f): return (-4.0 * math.log1p((-1.0 + ((2.0 / (math.pi * 0.5)) / f)))) / math.pi
function code(f) return Float64(Float64(-4.0 * log1p(Float64(-1.0 + Float64(Float64(2.0 / Float64(pi * 0.5)) / f)))) / pi) end
code[f_] := N[(N[(-4.0 * N[Log[1 + N[(-1.0 + N[(N[(2.0 / N[(Pi * 0.5), $MachinePrecision]), $MachinePrecision] / f), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]), $MachinePrecision] / Pi), $MachinePrecision]
\begin{array}{l}
\\
\frac{-4 \cdot \mathsf{log1p}\left(-1 + \frac{\frac{2}{\pi \cdot 0.5}}{f}\right)}{\pi}
\end{array}
Initial program 6.9%
distribute-lft-neg-in6.9%
*-commutative6.9%
Simplified6.9%
Taylor expanded in f around 0 96.6%
*-commutative96.6%
associate-*l/96.6%
mul-1-neg96.6%
unsub-neg96.6%
distribute-rgt-out--96.6%
metadata-eval96.6%
Simplified96.6%
log1p-expm1-u96.6%
expm1-udef96.6%
diff-log96.6%
add-exp-log96.6%
Applied egg-rr96.6%
Final simplification96.6%
(FPCore (f) :precision binary64 (* -4.0 (/ (log (/ 4.0 (* PI f))) PI)))
double code(double f) {
return -4.0 * (log((4.0 / (((double) M_PI) * f))) / ((double) M_PI));
}
public static double code(double f) {
return -4.0 * (Math.log((4.0 / (Math.PI * f))) / Math.PI);
}
def code(f): return -4.0 * (math.log((4.0 / (math.pi * f))) / math.pi)
function code(f) return Float64(-4.0 * Float64(log(Float64(4.0 / Float64(pi * f))) / pi)) end
function tmp = code(f) tmp = -4.0 * (log((4.0 / (pi * f))) / pi); end
code[f_] := N[(-4.0 * N[(N[Log[N[(4.0 / N[(Pi * f), $MachinePrecision]), $MachinePrecision]], $MachinePrecision] / Pi), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
-4 \cdot \frac{\log \left(\frac{4}{\pi \cdot f}\right)}{\pi}
\end{array}
Initial program 6.9%
distribute-lft-neg-in6.9%
*-commutative6.9%
Simplified6.9%
Taylor expanded in f around 0 96.6%
*-commutative96.6%
associate-*l/96.6%
mul-1-neg96.6%
unsub-neg96.6%
distribute-rgt-out--96.6%
metadata-eval96.6%
Simplified96.6%
Taylor expanded in f around 0 96.6%
log-div96.6%
associate-*r/96.6%
log-div96.6%
remove-double-neg96.6%
mul-1-neg96.6%
log-rec96.6%
associate-*r/96.6%
div-sub96.6%
log-rec96.6%
mul-1-neg96.6%
remove-double-neg96.6%
div-sub96.6%
Simplified96.6%
Final simplification96.6%
(FPCore (f) :precision binary64 (* 4.0 (/ (- (log 0.0)) PI)))
double code(double f) {
return 4.0 * (-log(0.0) / ((double) M_PI));
}
public static double code(double f) {
return 4.0 * (-Math.log(0.0) / Math.PI);
}
def code(f): return 4.0 * (-math.log(0.0) / math.pi)
function code(f) return Float64(4.0 * Float64(Float64(-log(0.0)) / pi)) end
function tmp = code(f) tmp = 4.0 * (-log(0.0) / pi); end
code[f_] := N[(4.0 * N[((-N[Log[0.0], $MachinePrecision]) / Pi), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
4 \cdot \frac{-\log 0}{\pi}
\end{array}
Initial program 6.9%
Taylor expanded in f around 0 96.5%
Taylor expanded in f around inf 0.7%
distribute-rgt-out0.7%
distribute-rgt-out--0.7%
metadata-eval0.7%
metadata-eval0.7%
mul0-rgt0.7%
Simplified0.7%
Final simplification0.7%
herbie shell --seed 2024013
(FPCore (f)
:name "VandenBroeck and Keller, Equation (20)"
:precision binary64
(- (* (/ 1.0 (/ PI 4.0)) (log (/ (+ (exp (* (/ PI 4.0) f)) (exp (- (* (/ PI 4.0) f)))) (- (exp (* (/ PI 4.0) f)) (exp (- (* (/ PI 4.0) f)))))))))