
(FPCore (x) :precision binary64 (* (/ 1.0 2.0) (log (/ (+ 1.0 x) (- 1.0 x)))))
double code(double x) {
return (1.0 / 2.0) * log(((1.0 + x) / (1.0 - x)));
}
real(8) function code(x)
real(8), intent (in) :: x
code = (1.0d0 / 2.0d0) * log(((1.0d0 + x) / (1.0d0 - x)))
end function
public static double code(double x) {
return (1.0 / 2.0) * Math.log(((1.0 + x) / (1.0 - x)));
}
def code(x): return (1.0 / 2.0) * math.log(((1.0 + x) / (1.0 - x)))
function code(x) return Float64(Float64(1.0 / 2.0) * log(Float64(Float64(1.0 + x) / Float64(1.0 - x)))) end
function tmp = code(x) tmp = (1.0 / 2.0) * log(((1.0 + x) / (1.0 - x))); end
code[x_] := N[(N[(1.0 / 2.0), $MachinePrecision] * N[Log[N[(N[(1.0 + x), $MachinePrecision] / N[(1.0 - x), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\frac{1}{2} \cdot \log \left(\frac{1 + x}{1 - x}\right)
\end{array}
Sampling outcomes in binary64 precision:
Herbie found 5 alternatives:
| Alternative | Accuracy | Speedup |
|---|
(FPCore (x) :precision binary64 (* (/ 1.0 2.0) (log (/ (+ 1.0 x) (- 1.0 x)))))
double code(double x) {
return (1.0 / 2.0) * log(((1.0 + x) / (1.0 - x)));
}
real(8) function code(x)
real(8), intent (in) :: x
code = (1.0d0 / 2.0d0) * log(((1.0d0 + x) / (1.0d0 - x)))
end function
public static double code(double x) {
return (1.0 / 2.0) * Math.log(((1.0 + x) / (1.0 - x)));
}
def code(x): return (1.0 / 2.0) * math.log(((1.0 + x) / (1.0 - x)))
function code(x) return Float64(Float64(1.0 / 2.0) * log(Float64(Float64(1.0 + x) / Float64(1.0 - x)))) end
function tmp = code(x) tmp = (1.0 / 2.0) * log(((1.0 + x) / (1.0 - x))); end
code[x_] := N[(N[(1.0 / 2.0), $MachinePrecision] * N[Log[N[(N[(1.0 + x), $MachinePrecision] / N[(1.0 - x), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\frac{1}{2} \cdot \log \left(\frac{1 + x}{1 - x}\right)
\end{array}
(FPCore (x) :precision binary64 (* (- (log1p (- x)) (log1p x)) -0.5))
double code(double x) {
return (log1p(-x) - log1p(x)) * -0.5;
}
public static double code(double x) {
return (Math.log1p(-x) - Math.log1p(x)) * -0.5;
}
def code(x): return (math.log1p(-x) - math.log1p(x)) * -0.5
function code(x) return Float64(Float64(log1p(Float64(-x)) - log1p(x)) * -0.5) end
code[x_] := N[(N[(N[Log[1 + (-x)], $MachinePrecision] - N[Log[1 + x], $MachinePrecision]), $MachinePrecision] * -0.5), $MachinePrecision]
\begin{array}{l}
\\
\left(\mathsf{log1p}\left(-x\right) - \mathsf{log1p}\left(x\right)\right) \cdot -0.5
\end{array}
Initial program 9.8%
*-commutative9.8%
log-div9.8%
sub-neg9.8%
remove-double-neg9.8%
sub-neg9.8%
+-commutative9.8%
neg-sub09.8%
associate--r-9.8%
sub-neg9.8%
log-div9.8%
neg-sub09.8%
distribute-lft-neg-in9.8%
distribute-rgt-neg-in9.8%
Simplified100.0%
Final simplification100.0%
(FPCore (x) :precision binary64 (let* ((t_0 (/ (+ x 1.0) (- 1.0 x)))) (if (<= t_0 1.0005) (* -0.5 (* x -2.0)) (* 0.5 (log t_0)))))
double code(double x) {
double t_0 = (x + 1.0) / (1.0 - x);
double tmp;
if (t_0 <= 1.0005) {
tmp = -0.5 * (x * -2.0);
} else {
tmp = 0.5 * log(t_0);
}
return tmp;
}
real(8) function code(x)
real(8), intent (in) :: x
real(8) :: t_0
real(8) :: tmp
t_0 = (x + 1.0d0) / (1.0d0 - x)
if (t_0 <= 1.0005d0) then
tmp = (-0.5d0) * (x * (-2.0d0))
else
tmp = 0.5d0 * log(t_0)
end if
code = tmp
end function
public static double code(double x) {
double t_0 = (x + 1.0) / (1.0 - x);
double tmp;
if (t_0 <= 1.0005) {
tmp = -0.5 * (x * -2.0);
} else {
tmp = 0.5 * Math.log(t_0);
}
return tmp;
}
def code(x): t_0 = (x + 1.0) / (1.0 - x) tmp = 0 if t_0 <= 1.0005: tmp = -0.5 * (x * -2.0) else: tmp = 0.5 * math.log(t_0) return tmp
function code(x) t_0 = Float64(Float64(x + 1.0) / Float64(1.0 - x)) tmp = 0.0 if (t_0 <= 1.0005) tmp = Float64(-0.5 * Float64(x * -2.0)); else tmp = Float64(0.5 * log(t_0)); end return tmp end
function tmp_2 = code(x) t_0 = (x + 1.0) / (1.0 - x); tmp = 0.0; if (t_0 <= 1.0005) tmp = -0.5 * (x * -2.0); else tmp = 0.5 * log(t_0); end tmp_2 = tmp; end
code[x_] := Block[{t$95$0 = N[(N[(x + 1.0), $MachinePrecision] / N[(1.0 - x), $MachinePrecision]), $MachinePrecision]}, If[LessEqual[t$95$0, 1.0005], N[(-0.5 * N[(x * -2.0), $MachinePrecision]), $MachinePrecision], N[(0.5 * N[Log[t$95$0], $MachinePrecision]), $MachinePrecision]]]
\begin{array}{l}
\\
\begin{array}{l}
t_0 := \frac{x + 1}{1 - x}\\
\mathbf{if}\;t_0 \leq 1.0005:\\
\;\;\;\;-0.5 \cdot \left(x \cdot -2\right)\\
\mathbf{else}:\\
\;\;\;\;0.5 \cdot \log t_0\\
\end{array}
\end{array}
if (/.f64 (+.f64 1 x) (-.f64 1 x)) < 1.00049999999999994Initial program 8.5%
*-commutative8.5%
log-div8.5%
sub-neg8.5%
remove-double-neg8.5%
sub-neg8.5%
+-commutative8.5%
neg-sub08.5%
associate--r-8.5%
sub-neg8.5%
log-div8.4%
neg-sub08.4%
distribute-lft-neg-in8.4%
distribute-rgt-neg-in8.4%
Simplified100.0%
Taylor expanded in x around 0 98.8%
*-commutative98.8%
Simplified98.8%
if 1.00049999999999994 < (/.f64 (+.f64 1 x) (-.f64 1 x)) Initial program 95.2%
metadata-eval95.2%
Simplified95.2%
Final simplification98.8%
(FPCore (x) :precision binary64 (* -0.5 (+ (* x -2.0) (* -0.6666666666666666 (pow x 3.0)))))
double code(double x) {
return -0.5 * ((x * -2.0) + (-0.6666666666666666 * pow(x, 3.0)));
}
real(8) function code(x)
real(8), intent (in) :: x
code = (-0.5d0) * ((x * (-2.0d0)) + ((-0.6666666666666666d0) * (x ** 3.0d0)))
end function
public static double code(double x) {
return -0.5 * ((x * -2.0) + (-0.6666666666666666 * Math.pow(x, 3.0)));
}
def code(x): return -0.5 * ((x * -2.0) + (-0.6666666666666666 * math.pow(x, 3.0)))
function code(x) return Float64(-0.5 * Float64(Float64(x * -2.0) + Float64(-0.6666666666666666 * (x ^ 3.0)))) end
function tmp = code(x) tmp = -0.5 * ((x * -2.0) + (-0.6666666666666666 * (x ^ 3.0))); end
code[x_] := N[(-0.5 * N[(N[(x * -2.0), $MachinePrecision] + N[(-0.6666666666666666 * N[Power[x, 3.0], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
-0.5 \cdot \left(x \cdot -2 + -0.6666666666666666 \cdot {x}^{3}\right)
\end{array}
Initial program 9.8%
*-commutative9.8%
log-div9.8%
sub-neg9.8%
remove-double-neg9.8%
sub-neg9.8%
+-commutative9.8%
neg-sub09.8%
associate--r-9.8%
sub-neg9.8%
log-div9.8%
neg-sub09.8%
distribute-lft-neg-in9.8%
distribute-rgt-neg-in9.8%
Simplified100.0%
Taylor expanded in x around 0 98.4%
Final simplification98.4%
(FPCore (x) :precision binary64 (* -0.5 (* x -2.0)))
double code(double x) {
return -0.5 * (x * -2.0);
}
real(8) function code(x)
real(8), intent (in) :: x
code = (-0.5d0) * (x * (-2.0d0))
end function
public static double code(double x) {
return -0.5 * (x * -2.0);
}
def code(x): return -0.5 * (x * -2.0)
function code(x) return Float64(-0.5 * Float64(x * -2.0)) end
function tmp = code(x) tmp = -0.5 * (x * -2.0); end
code[x_] := N[(-0.5 * N[(x * -2.0), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
-0.5 \cdot \left(x \cdot -2\right)
\end{array}
Initial program 9.8%
*-commutative9.8%
log-div9.8%
sub-neg9.8%
remove-double-neg9.8%
sub-neg9.8%
+-commutative9.8%
neg-sub09.8%
associate--r-9.8%
sub-neg9.8%
log-div9.8%
neg-sub09.8%
distribute-lft-neg-in9.8%
distribute-rgt-neg-in9.8%
Simplified100.0%
Taylor expanded in x around 0 97.8%
*-commutative97.8%
Simplified97.8%
Final simplification97.8%
(FPCore (x) :precision binary64 0.0)
double code(double x) {
return 0.0;
}
real(8) function code(x)
real(8), intent (in) :: x
code = 0.0d0
end function
public static double code(double x) {
return 0.0;
}
def code(x): return 0.0
function code(x) return 0.0 end
function tmp = code(x) tmp = 0.0; end
code[x_] := 0.0
\begin{array}{l}
\\
0
\end{array}
Initial program 9.8%
metadata-eval9.8%
Simplified9.8%
log-div9.8%
log1p-udef22.4%
sub-neg22.4%
log1p-udef100.0%
*-un-lft-identity100.0%
cancel-sign-sub-inv100.0%
metadata-eval100.0%
add-sqr-sqrt47.5%
sqrt-unprod32.4%
sqr-neg32.4%
sqrt-unprod3.0%
add-sqr-sqrt5.1%
neg-mul-15.1%
Applied egg-rr5.1%
sub-neg5.1%
+-inverses5.1%
Simplified5.1%
Final simplification5.1%
herbie shell --seed 2023338
(FPCore (x)
:name "Hyperbolic arc-(co)tangent"
:precision binary64
(* (/ 1.0 2.0) (log (/ (+ 1.0 x) (- 1.0 x)))))