
(FPCore (f) :precision binary64 (let* ((t_0 (* (/ PI 4.0) f)) (t_1 (exp t_0)) (t_2 (exp (- t_0)))) (- (* (/ 1.0 (/ PI 4.0)) (log (/ (+ t_1 t_2) (- t_1 t_2)))))))
double code(double f) {
double t_0 = (((double) M_PI) / 4.0) * f;
double t_1 = exp(t_0);
double t_2 = exp(-t_0);
return -((1.0 / (((double) M_PI) / 4.0)) * log(((t_1 + t_2) / (t_1 - t_2))));
}
public static double code(double f) {
double t_0 = (Math.PI / 4.0) * f;
double t_1 = Math.exp(t_0);
double t_2 = Math.exp(-t_0);
return -((1.0 / (Math.PI / 4.0)) * Math.log(((t_1 + t_2) / (t_1 - t_2))));
}
def code(f): t_0 = (math.pi / 4.0) * f t_1 = math.exp(t_0) t_2 = math.exp(-t_0) return -((1.0 / (math.pi / 4.0)) * math.log(((t_1 + t_2) / (t_1 - t_2))))
function code(f) t_0 = Float64(Float64(pi / 4.0) * f) t_1 = exp(t_0) t_2 = exp(Float64(-t_0)) return Float64(-Float64(Float64(1.0 / Float64(pi / 4.0)) * log(Float64(Float64(t_1 + t_2) / Float64(t_1 - t_2))))) end
function tmp = code(f) t_0 = (pi / 4.0) * f; t_1 = exp(t_0); t_2 = exp(-t_0); tmp = -((1.0 / (pi / 4.0)) * log(((t_1 + t_2) / (t_1 - t_2)))); end
code[f_] := Block[{t$95$0 = N[(N[(Pi / 4.0), $MachinePrecision] * f), $MachinePrecision]}, Block[{t$95$1 = N[Exp[t$95$0], $MachinePrecision]}, Block[{t$95$2 = N[Exp[(-t$95$0)], $MachinePrecision]}, (-N[(N[(1.0 / N[(Pi / 4.0), $MachinePrecision]), $MachinePrecision] * N[Log[N[(N[(t$95$1 + t$95$2), $MachinePrecision] / N[(t$95$1 - t$95$2), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]), $MachinePrecision])]]]
\begin{array}{l}
\\
\begin{array}{l}
t_0 := \frac{\pi}{4} \cdot f\\
t_1 := e^{t_0}\\
t_2 := e^{-t_0}\\
-\frac{1}{\frac{\pi}{4}} \cdot \log \left(\frac{t_1 + t_2}{t_1 - t_2}\right)
\end{array}
\end{array}
Sampling outcomes in binary64 precision:
Herbie found 10 alternatives:
| Alternative | Accuracy | Speedup |
|---|
(FPCore (f) :precision binary64 (let* ((t_0 (* (/ PI 4.0) f)) (t_1 (exp t_0)) (t_2 (exp (- t_0)))) (- (* (/ 1.0 (/ PI 4.0)) (log (/ (+ t_1 t_2) (- t_1 t_2)))))))
double code(double f) {
double t_0 = (((double) M_PI) / 4.0) * f;
double t_1 = exp(t_0);
double t_2 = exp(-t_0);
return -((1.0 / (((double) M_PI) / 4.0)) * log(((t_1 + t_2) / (t_1 - t_2))));
}
public static double code(double f) {
double t_0 = (Math.PI / 4.0) * f;
double t_1 = Math.exp(t_0);
double t_2 = Math.exp(-t_0);
return -((1.0 / (Math.PI / 4.0)) * Math.log(((t_1 + t_2) / (t_1 - t_2))));
}
def code(f): t_0 = (math.pi / 4.0) * f t_1 = math.exp(t_0) t_2 = math.exp(-t_0) return -((1.0 / (math.pi / 4.0)) * math.log(((t_1 + t_2) / (t_1 - t_2))))
function code(f) t_0 = Float64(Float64(pi / 4.0) * f) t_1 = exp(t_0) t_2 = exp(Float64(-t_0)) return Float64(-Float64(Float64(1.0 / Float64(pi / 4.0)) * log(Float64(Float64(t_1 + t_2) / Float64(t_1 - t_2))))) end
function tmp = code(f) t_0 = (pi / 4.0) * f; t_1 = exp(t_0); t_2 = exp(-t_0); tmp = -((1.0 / (pi / 4.0)) * log(((t_1 + t_2) / (t_1 - t_2)))); end
code[f_] := Block[{t$95$0 = N[(N[(Pi / 4.0), $MachinePrecision] * f), $MachinePrecision]}, Block[{t$95$1 = N[Exp[t$95$0], $MachinePrecision]}, Block[{t$95$2 = N[Exp[(-t$95$0)], $MachinePrecision]}, (-N[(N[(1.0 / N[(Pi / 4.0), $MachinePrecision]), $MachinePrecision] * N[Log[N[(N[(t$95$1 + t$95$2), $MachinePrecision] / N[(t$95$1 - t$95$2), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]), $MachinePrecision])]]]
\begin{array}{l}
\\
\begin{array}{l}
t_0 := \frac{\pi}{4} \cdot f\\
t_1 := e^{t_0}\\
t_2 := e^{-t_0}\\
-\frac{1}{\frac{\pi}{4}} \cdot \log \left(\frac{t_1 + t_2}{t_1 - t_2}\right)
\end{array}
\end{array}
(FPCore (f)
:precision binary64
(fma
-4.0
(/ (- (log (/ (/ 2.0 PI) 0.5)) (log f)) PI)
(*
-2.0
(+
(* (/ f PI) (* PI 0.0))
(*
(/ (pow f 2.0) PI)
(fma
(* PI 0.5)
(+
(* 0.0625 (/ PI 0.5))
(*
-2.0
(* (/ (pow PI 3.0) (* 0.25 (pow PI 2.0))) 0.005208333333333333)))
(* 0.0 (pow (* PI 0.5) 2.0))))))))
double code(double f) {
return fma(-4.0, ((log(((2.0 / ((double) M_PI)) / 0.5)) - log(f)) / ((double) M_PI)), (-2.0 * (((f / ((double) M_PI)) * (((double) M_PI) * 0.0)) + ((pow(f, 2.0) / ((double) M_PI)) * fma((((double) M_PI) * 0.5), ((0.0625 * (((double) M_PI) / 0.5)) + (-2.0 * ((pow(((double) M_PI), 3.0) / (0.25 * pow(((double) M_PI), 2.0))) * 0.005208333333333333))), (0.0 * pow((((double) M_PI) * 0.5), 2.0)))))));
}
function code(f) return fma(-4.0, Float64(Float64(log(Float64(Float64(2.0 / pi) / 0.5)) - log(f)) / pi), Float64(-2.0 * Float64(Float64(Float64(f / pi) * Float64(pi * 0.0)) + Float64(Float64((f ^ 2.0) / pi) * fma(Float64(pi * 0.5), Float64(Float64(0.0625 * Float64(pi / 0.5)) + Float64(-2.0 * Float64(Float64((pi ^ 3.0) / Float64(0.25 * (pi ^ 2.0))) * 0.005208333333333333))), Float64(0.0 * (Float64(pi * 0.5) ^ 2.0))))))) end
code[f_] := N[(-4.0 * N[(N[(N[Log[N[(N[(2.0 / Pi), $MachinePrecision] / 0.5), $MachinePrecision]], $MachinePrecision] - N[Log[f], $MachinePrecision]), $MachinePrecision] / Pi), $MachinePrecision] + N[(-2.0 * N[(N[(N[(f / Pi), $MachinePrecision] * N[(Pi * 0.0), $MachinePrecision]), $MachinePrecision] + N[(N[(N[Power[f, 2.0], $MachinePrecision] / Pi), $MachinePrecision] * N[(N[(Pi * 0.5), $MachinePrecision] * N[(N[(0.0625 * N[(Pi / 0.5), $MachinePrecision]), $MachinePrecision] + N[(-2.0 * N[(N[(N[Power[Pi, 3.0], $MachinePrecision] / N[(0.25 * N[Power[Pi, 2.0], $MachinePrecision]), $MachinePrecision]), $MachinePrecision] * 0.005208333333333333), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] + N[(0.0 * N[Power[N[(Pi * 0.5), $MachinePrecision], 2.0], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\mathsf{fma}\left(-4, \frac{\log \left(\frac{\frac{2}{\pi}}{0.5}\right) - \log f}{\pi}, -2 \cdot \left(\frac{f}{\pi} \cdot \left(\pi \cdot 0\right) + \frac{{f}^{2}}{\pi} \cdot \mathsf{fma}\left(\pi \cdot 0.5, 0.0625 \cdot \frac{\pi}{0.5} + -2 \cdot \left(\frac{{\pi}^{3}}{0.25 \cdot {\pi}^{2}} \cdot 0.005208333333333333\right), 0 \cdot {\left(\pi \cdot 0.5\right)}^{2}\right)\right)\right)
\end{array}
Initial program 8.3%
distribute-lft-neg-in8.3%
*-commutative8.3%
Simplified8.3%
Taylor expanded in f around 0 95.9%
Simplified95.9%
fma-udef95.9%
associate-/r*95.9%
pow195.9%
pow-div95.9%
metadata-eval95.9%
pow195.9%
associate-/r/95.9%
*-commutative95.9%
unpow-prod-down95.9%
metadata-eval95.9%
Applied egg-rr95.9%
Final simplification95.9%
(FPCore (f)
:precision binary64
(*
(log
(fma
f
(fma
0.0625
(/ (pow PI 2.0) (* PI 0.5))
(* -2.0 (/ (pow PI 3.0) (/ (pow (* PI 0.5) 2.0) 0.005208333333333333))))
(pow (sqrt (/ (/ 2.0 PI) (* 0.5 f))) 2.0)))
(/ -4.0 PI)))
double code(double f) {
return log(fma(f, fma(0.0625, (pow(((double) M_PI), 2.0) / (((double) M_PI) * 0.5)), (-2.0 * (pow(((double) M_PI), 3.0) / (pow((((double) M_PI) * 0.5), 2.0) / 0.005208333333333333)))), pow(sqrt(((2.0 / ((double) M_PI)) / (0.5 * f))), 2.0))) * (-4.0 / ((double) M_PI));
}
function code(f) return Float64(log(fma(f, fma(0.0625, Float64((pi ^ 2.0) / Float64(pi * 0.5)), Float64(-2.0 * Float64((pi ^ 3.0) / Float64((Float64(pi * 0.5) ^ 2.0) / 0.005208333333333333)))), (sqrt(Float64(Float64(2.0 / pi) / Float64(0.5 * f))) ^ 2.0))) * Float64(-4.0 / pi)) end
code[f_] := N[(N[Log[N[(f * N[(0.0625 * N[(N[Power[Pi, 2.0], $MachinePrecision] / N[(Pi * 0.5), $MachinePrecision]), $MachinePrecision] + N[(-2.0 * N[(N[Power[Pi, 3.0], $MachinePrecision] / N[(N[Power[N[(Pi * 0.5), $MachinePrecision], 2.0], $MachinePrecision] / 0.005208333333333333), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] + N[Power[N[Sqrt[N[(N[(2.0 / Pi), $MachinePrecision] / N[(0.5 * f), $MachinePrecision]), $MachinePrecision]], $MachinePrecision], 2.0], $MachinePrecision]), $MachinePrecision]], $MachinePrecision] * N[(-4.0 / Pi), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\log \left(\mathsf{fma}\left(f, \mathsf{fma}\left(0.0625, \frac{{\pi}^{2}}{\pi \cdot 0.5}, -2 \cdot \frac{{\pi}^{3}}{\frac{{\left(\pi \cdot 0.5\right)}^{2}}{0.005208333333333333}}\right), {\left(\sqrt{\frac{\frac{2}{\pi}}{0.5 \cdot f}}\right)}^{2}\right)\right) \cdot \frac{-4}{\pi}
\end{array}
Initial program 8.3%
distribute-lft-neg-in8.3%
*-commutative8.3%
Simplified8.3%
Taylor expanded in f around 0 95.7%
Simplified95.7%
add-sqr-sqrt95.7%
pow295.7%
associate-/l/95.7%
Applied egg-rr95.7%
Final simplification95.7%
(FPCore (f)
:precision binary64
(*
(/ -4.0 PI)
(log
(fma
f
(+ (* 0.0625 (/ PI 0.5)) (* -2.0 (* PI 0.020833333333333332)))
(/ (/ (/ 2.0 PI) 0.5) f)))))
double code(double f) {
return (-4.0 / ((double) M_PI)) * log(fma(f, ((0.0625 * (((double) M_PI) / 0.5)) + (-2.0 * (((double) M_PI) * 0.020833333333333332))), (((2.0 / ((double) M_PI)) / 0.5) / f)));
}
function code(f) return Float64(Float64(-4.0 / pi) * log(fma(f, Float64(Float64(0.0625 * Float64(pi / 0.5)) + Float64(-2.0 * Float64(pi * 0.020833333333333332))), Float64(Float64(Float64(2.0 / pi) / 0.5) / f)))) end
code[f_] := N[(N[(-4.0 / Pi), $MachinePrecision] * N[Log[N[(f * N[(N[(0.0625 * N[(Pi / 0.5), $MachinePrecision]), $MachinePrecision] + N[(-2.0 * N[(Pi * 0.020833333333333332), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] + N[(N[(N[(2.0 / Pi), $MachinePrecision] / 0.5), $MachinePrecision] / f), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\frac{-4}{\pi} \cdot \log \left(\mathsf{fma}\left(f, 0.0625 \cdot \frac{\pi}{0.5} + -2 \cdot \left(\pi \cdot 0.020833333333333332\right), \frac{\frac{\frac{2}{\pi}}{0.5}}{f}\right)\right)
\end{array}
Initial program 8.3%
distribute-lft-neg-in8.3%
*-commutative8.3%
Simplified8.3%
Taylor expanded in f around 0 95.7%
Simplified95.7%
fma-udef95.9%
associate-/r*95.9%
pow195.9%
pow-div95.9%
metadata-eval95.9%
pow195.9%
associate-/r/95.9%
*-commutative95.9%
unpow-prod-down95.9%
metadata-eval95.9%
Applied egg-rr95.7%
expm1-log1p-u95.7%
expm1-udef95.7%
associate-*l/95.7%
*-commutative95.7%
times-frac95.7%
metadata-eval95.7%
Applied egg-rr95.7%
expm1-def95.7%
expm1-log1p95.7%
cube-mult95.7%
unpow295.7%
associate-/l*95.7%
*-inverses95.7%
/-rgt-identity95.7%
Simplified95.7%
Final simplification95.7%
(FPCore (f) :precision binary64 (+ (* -4.0 (/ (- (log (/ 4.0 PI)) (log f)) PI)) (* -0.125 (* PI (pow f 2.0)))))
double code(double f) {
return (-4.0 * ((log((4.0 / ((double) M_PI))) - log(f)) / ((double) M_PI))) + (-0.125 * (((double) M_PI) * pow(f, 2.0)));
}
public static double code(double f) {
return (-4.0 * ((Math.log((4.0 / Math.PI)) - Math.log(f)) / Math.PI)) + (-0.125 * (Math.PI * Math.pow(f, 2.0)));
}
def code(f): return (-4.0 * ((math.log((4.0 / math.pi)) - math.log(f)) / math.pi)) + (-0.125 * (math.pi * math.pow(f, 2.0)))
function code(f) return Float64(Float64(-4.0 * Float64(Float64(log(Float64(4.0 / pi)) - log(f)) / pi)) + Float64(-0.125 * Float64(pi * (f ^ 2.0)))) end
function tmp = code(f) tmp = (-4.0 * ((log((4.0 / pi)) - log(f)) / pi)) + (-0.125 * (pi * (f ^ 2.0))); end
code[f_] := N[(N[(-4.0 * N[(N[(N[Log[N[(4.0 / Pi), $MachinePrecision]], $MachinePrecision] - N[Log[f], $MachinePrecision]), $MachinePrecision] / Pi), $MachinePrecision]), $MachinePrecision] + N[(-0.125 * N[(Pi * N[Power[f, 2.0], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
-4 \cdot \frac{\log \left(\frac{4}{\pi}\right) - \log f}{\pi} + -0.125 \cdot \left(\pi \cdot {f}^{2}\right)
\end{array}
Initial program 8.3%
distribute-lft-neg-in8.3%
*-commutative8.3%
Simplified8.3%
Taylor expanded in f around 0 95.2%
distribute-rgt-out--95.2%
metadata-eval95.2%
Simplified95.2%
Taylor expanded in f around 0 95.4%
Final simplification95.4%
(FPCore (f) :precision binary64 (* (/ -4.0 PI) (log (+ (* 0.125 (* PI f)) (* 4.0 (/ 1.0 (* PI f)))))))
double code(double f) {
return (-4.0 / ((double) M_PI)) * log(((0.125 * (((double) M_PI) * f)) + (4.0 * (1.0 / (((double) M_PI) * f)))));
}
public static double code(double f) {
return (-4.0 / Math.PI) * Math.log(((0.125 * (Math.PI * f)) + (4.0 * (1.0 / (Math.PI * f)))));
}
def code(f): return (-4.0 / math.pi) * math.log(((0.125 * (math.pi * f)) + (4.0 * (1.0 / (math.pi * f)))))
function code(f) return Float64(Float64(-4.0 / pi) * log(Float64(Float64(0.125 * Float64(pi * f)) + Float64(4.0 * Float64(1.0 / Float64(pi * f)))))) end
function tmp = code(f) tmp = (-4.0 / pi) * log(((0.125 * (pi * f)) + (4.0 * (1.0 / (pi * f))))); end
code[f_] := N[(N[(-4.0 / Pi), $MachinePrecision] * N[Log[N[(N[(0.125 * N[(Pi * f), $MachinePrecision]), $MachinePrecision] + N[(4.0 * N[(1.0 / N[(Pi * f), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\frac{-4}{\pi} \cdot \log \left(0.125 \cdot \left(\pi \cdot f\right) + 4 \cdot \frac{1}{\pi \cdot f}\right)
\end{array}
Initial program 8.3%
distribute-lft-neg-in8.3%
*-commutative8.3%
Simplified8.3%
Taylor expanded in f around 0 95.2%
distribute-rgt-out--95.2%
metadata-eval95.2%
Simplified95.2%
Taylor expanded in f around 0 95.2%
Final simplification95.2%
(FPCore (f) :precision binary64 (* -4.0 (/ (- (log (/ (/ 2.0 PI) 0.5)) (log f)) PI)))
double code(double f) {
return -4.0 * ((log(((2.0 / ((double) M_PI)) / 0.5)) - log(f)) / ((double) M_PI));
}
public static double code(double f) {
return -4.0 * ((Math.log(((2.0 / Math.PI) / 0.5)) - Math.log(f)) / Math.PI);
}
def code(f): return -4.0 * ((math.log(((2.0 / math.pi) / 0.5)) - math.log(f)) / math.pi)
function code(f) return Float64(-4.0 * Float64(Float64(log(Float64(Float64(2.0 / pi) / 0.5)) - log(f)) / pi)) end
function tmp = code(f) tmp = -4.0 * ((log(((2.0 / pi) / 0.5)) - log(f)) / pi); end
code[f_] := N[(-4.0 * N[(N[(N[Log[N[(N[(2.0 / Pi), $MachinePrecision] / 0.5), $MachinePrecision]], $MachinePrecision] - N[Log[f], $MachinePrecision]), $MachinePrecision] / Pi), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
-4 \cdot \frac{\log \left(\frac{\frac{2}{\pi}}{0.5}\right) - \log f}{\pi}
\end{array}
Initial program 8.3%
distribute-lft-neg-in8.3%
*-commutative8.3%
Simplified8.3%
Taylor expanded in f around 0 95.2%
mul-1-neg95.2%
unsub-neg95.2%
distribute-rgt-out--95.2%
metadata-eval95.2%
metadata-eval95.2%
associate-/r*95.2%
metadata-eval95.2%
Simplified95.2%
Final simplification95.2%
(FPCore (f) :precision binary64 (* (/ -4.0 PI) (fabs (log (/ 4.0 (* PI f))))))
double code(double f) {
return (-4.0 / ((double) M_PI)) * fabs(log((4.0 / (((double) M_PI) * f))));
}
public static double code(double f) {
return (-4.0 / Math.PI) * Math.abs(Math.log((4.0 / (Math.PI * f))));
}
def code(f): return (-4.0 / math.pi) * math.fabs(math.log((4.0 / (math.pi * f))))
function code(f) return Float64(Float64(-4.0 / pi) * abs(log(Float64(4.0 / Float64(pi * f))))) end
function tmp = code(f) tmp = (-4.0 / pi) * abs(log((4.0 / (pi * f)))); end
code[f_] := N[(N[(-4.0 / Pi), $MachinePrecision] * N[Abs[N[Log[N[(4.0 / N[(Pi * f), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\frac{-4}{\pi} \cdot \left|\log \left(\frac{4}{\pi \cdot f}\right)\right|
\end{array}
Initial program 8.3%
distribute-lft-neg-in8.3%
*-commutative8.3%
Simplified8.3%
Taylor expanded in f around 0 95.0%
*-commutative95.0%
associate-/r*95.0%
distribute-rgt-out--95.0%
metadata-eval95.0%
metadata-eval95.0%
associate-/r*95.0%
metadata-eval95.0%
Simplified95.0%
Taylor expanded in f around 0 95.2%
Simplified95.0%
add-sqr-sqrt94.5%
sqrt-unprod95.2%
pow295.2%
*-commutative95.2%
Applied egg-rr95.2%
unpow295.2%
rem-sqrt-square95.2%
Simplified95.2%
Final simplification95.2%
(FPCore (f) :precision binary64 (* (/ -4.0 PI) (log (/ 4.0 (* PI f)))))
double code(double f) {
return (-4.0 / ((double) M_PI)) * log((4.0 / (((double) M_PI) * f)));
}
public static double code(double f) {
return (-4.0 / Math.PI) * Math.log((4.0 / (Math.PI * f)));
}
def code(f): return (-4.0 / math.pi) * math.log((4.0 / (math.pi * f)))
function code(f) return Float64(Float64(-4.0 / pi) * log(Float64(4.0 / Float64(pi * f)))) end
function tmp = code(f) tmp = (-4.0 / pi) * log((4.0 / (pi * f))); end
code[f_] := N[(N[(-4.0 / Pi), $MachinePrecision] * N[Log[N[(4.0 / N[(Pi * f), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\frac{-4}{\pi} \cdot \log \left(\frac{4}{\pi \cdot f}\right)
\end{array}
Initial program 8.3%
distribute-lft-neg-in8.3%
*-commutative8.3%
Simplified8.3%
Taylor expanded in f around 0 95.0%
*-commutative95.0%
associate-/r*95.0%
distribute-rgt-out--95.0%
metadata-eval95.0%
metadata-eval95.0%
associate-/r*95.0%
metadata-eval95.0%
Simplified95.0%
Taylor expanded in f around 0 95.2%
Simplified95.0%
Final simplification95.0%
(FPCore (f) :precision binary64 (/ (* -4.0 (log (/ 4.0 (* PI f)))) PI))
double code(double f) {
return (-4.0 * log((4.0 / (((double) M_PI) * f)))) / ((double) M_PI);
}
public static double code(double f) {
return (-4.0 * Math.log((4.0 / (Math.PI * f)))) / Math.PI;
}
def code(f): return (-4.0 * math.log((4.0 / (math.pi * f)))) / math.pi
function code(f) return Float64(Float64(-4.0 * log(Float64(4.0 / Float64(pi * f)))) / pi) end
function tmp = code(f) tmp = (-4.0 * log((4.0 / (pi * f)))) / pi; end
code[f_] := N[(N[(-4.0 * N[Log[N[(4.0 / N[(Pi * f), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]), $MachinePrecision] / Pi), $MachinePrecision]
\begin{array}{l}
\\
\frac{-4 \cdot \log \left(\frac{4}{\pi \cdot f}\right)}{\pi}
\end{array}
Initial program 8.3%
distribute-lft-neg-in8.3%
*-commutative8.3%
Simplified8.3%
Taylor expanded in f around 0 95.0%
*-commutative95.0%
associate-/r*95.0%
distribute-rgt-out--95.0%
metadata-eval95.0%
metadata-eval95.0%
associate-/r*95.0%
metadata-eval95.0%
Simplified95.0%
Taylor expanded in f around 0 95.2%
Simplified95.0%
*-commutative95.0%
associate-*r/95.2%
*-commutative95.2%
Applied egg-rr95.2%
Final simplification95.2%
(FPCore (f) :precision binary64 (/ (* -4.0 (log 0.0)) PI))
double code(double f) {
return (-4.0 * log(0.0)) / ((double) M_PI);
}
public static double code(double f) {
return (-4.0 * Math.log(0.0)) / Math.PI;
}
def code(f): return (-4.0 * math.log(0.0)) / math.pi
function code(f) return Float64(Float64(-4.0 * log(0.0)) / pi) end
function tmp = code(f) tmp = (-4.0 * log(0.0)) / pi; end
code[f_] := N[(N[(-4.0 * N[Log[0.0], $MachinePrecision]), $MachinePrecision] / Pi), $MachinePrecision]
\begin{array}{l}
\\
\frac{-4 \cdot \log 0}{\pi}
\end{array}
Initial program 8.3%
distribute-lft-neg-in8.3%
*-commutative8.3%
Simplified8.3%
Taylor expanded in f around 0 95.0%
Taylor expanded in f around inf 0.7%
associate-*r/0.7%
distribute-rgt-out0.7%
distribute-rgt-out--0.7%
metadata-eval0.7%
metadata-eval0.7%
mul0-rgt0.7%
Simplified0.7%
Final simplification0.7%
herbie shell --seed 2023332
(FPCore (f)
:name "VandenBroeck and Keller, Equation (20)"
:precision binary64
(- (* (/ 1.0 (/ PI 4.0)) (log (/ (+ (exp (* (/ PI 4.0) f)) (exp (- (* (/ PI 4.0) f)))) (- (exp (* (/ PI 4.0) f)) (exp (- (* (/ PI 4.0) f)))))))))