
(FPCore (x) :precision binary64 (+ (- (exp x) 2.0) (exp (- x))))
double code(double x) {
return (exp(x) - 2.0) + exp(-x);
}
real(8) function code(x)
real(8), intent (in) :: x
code = (exp(x) - 2.0d0) + exp(-x)
end function
public static double code(double x) {
return (Math.exp(x) - 2.0) + Math.exp(-x);
}
def code(x): return (math.exp(x) - 2.0) + math.exp(-x)
function code(x) return Float64(Float64(exp(x) - 2.0) + exp(Float64(-x))) end
function tmp = code(x) tmp = (exp(x) - 2.0) + exp(-x); end
code[x_] := N[(N[(N[Exp[x], $MachinePrecision] - 2.0), $MachinePrecision] + N[Exp[(-x)], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\left(e^{x} - 2\right) + e^{-x}
\end{array}
Sampling outcomes in binary64 precision:
Herbie found 6 alternatives:
| Alternative | Accuracy | Speedup |
|---|
(FPCore (x) :precision binary64 (+ (- (exp x) 2.0) (exp (- x))))
double code(double x) {
return (exp(x) - 2.0) + exp(-x);
}
real(8) function code(x)
real(8), intent (in) :: x
code = (exp(x) - 2.0d0) + exp(-x)
end function
public static double code(double x) {
return (Math.exp(x) - 2.0) + Math.exp(-x);
}
def code(x): return (math.exp(x) - 2.0) + math.exp(-x)
function code(x) return Float64(Float64(exp(x) - 2.0) + exp(Float64(-x))) end
function tmp = code(x) tmp = (exp(x) - 2.0) + exp(-x); end
code[x_] := N[(N[(N[Exp[x], $MachinePrecision] - 2.0), $MachinePrecision] + N[Exp[(-x)], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\left(e^{x} - 2\right) + e^{-x}
\end{array}
x_m = (fabs.f64 x)
(FPCore (x_m)
:precision binary64
(let* ((t_0 (+ (- (exp x_m) 2.0) (exp (- x_m)))))
(if (<= t_0 1e-5)
(*
(* x_m x_m)
(+
1.0
(*
(* x_m x_m)
(+ 0.08333333333333333 (* (* x_m x_m) 0.002777777777777778)))))
t_0)))x_m = fabs(x);
double code(double x_m) {
double t_0 = (exp(x_m) - 2.0) + exp(-x_m);
double tmp;
if (t_0 <= 1e-5) {
tmp = (x_m * x_m) * (1.0 + ((x_m * x_m) * (0.08333333333333333 + ((x_m * x_m) * 0.002777777777777778))));
} else {
tmp = t_0;
}
return tmp;
}
x_m = abs(x)
real(8) function code(x_m)
real(8), intent (in) :: x_m
real(8) :: t_0
real(8) :: tmp
t_0 = (exp(x_m) - 2.0d0) + exp(-x_m)
if (t_0 <= 1d-5) then
tmp = (x_m * x_m) * (1.0d0 + ((x_m * x_m) * (0.08333333333333333d0 + ((x_m * x_m) * 0.002777777777777778d0))))
else
tmp = t_0
end if
code = tmp
end function
x_m = Math.abs(x);
public static double code(double x_m) {
double t_0 = (Math.exp(x_m) - 2.0) + Math.exp(-x_m);
double tmp;
if (t_0 <= 1e-5) {
tmp = (x_m * x_m) * (1.0 + ((x_m * x_m) * (0.08333333333333333 + ((x_m * x_m) * 0.002777777777777778))));
} else {
tmp = t_0;
}
return tmp;
}
x_m = math.fabs(x) def code(x_m): t_0 = (math.exp(x_m) - 2.0) + math.exp(-x_m) tmp = 0 if t_0 <= 1e-5: tmp = (x_m * x_m) * (1.0 + ((x_m * x_m) * (0.08333333333333333 + ((x_m * x_m) * 0.002777777777777778)))) else: tmp = t_0 return tmp
x_m = abs(x) function code(x_m) t_0 = Float64(Float64(exp(x_m) - 2.0) + exp(Float64(-x_m))) tmp = 0.0 if (t_0 <= 1e-5) tmp = Float64(Float64(x_m * x_m) * Float64(1.0 + Float64(Float64(x_m * x_m) * Float64(0.08333333333333333 + Float64(Float64(x_m * x_m) * 0.002777777777777778))))); else tmp = t_0; end return tmp end
x_m = abs(x); function tmp_2 = code(x_m) t_0 = (exp(x_m) - 2.0) + exp(-x_m); tmp = 0.0; if (t_0 <= 1e-5) tmp = (x_m * x_m) * (1.0 + ((x_m * x_m) * (0.08333333333333333 + ((x_m * x_m) * 0.002777777777777778)))); else tmp = t_0; end tmp_2 = tmp; end
x_m = N[Abs[x], $MachinePrecision]
code[x$95$m_] := Block[{t$95$0 = N[(N[(N[Exp[x$95$m], $MachinePrecision] - 2.0), $MachinePrecision] + N[Exp[(-x$95$m)], $MachinePrecision]), $MachinePrecision]}, If[LessEqual[t$95$0, 1e-5], N[(N[(x$95$m * x$95$m), $MachinePrecision] * N[(1.0 + N[(N[(x$95$m * x$95$m), $MachinePrecision] * N[(0.08333333333333333 + N[(N[(x$95$m * x$95$m), $MachinePrecision] * 0.002777777777777778), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision], t$95$0]]
\begin{array}{l}
x_m = \left|x\right|
\\
\begin{array}{l}
t_0 := \left(e^{x\_m} - 2\right) + e^{-x\_m}\\
\mathbf{if}\;t\_0 \leq 10^{-5}:\\
\;\;\;\;\left(x\_m \cdot x\_m\right) \cdot \left(1 + \left(x\_m \cdot x\_m\right) \cdot \left(0.08333333333333333 + \left(x\_m \cdot x\_m\right) \cdot 0.002777777777777778\right)\right)\\
\mathbf{else}:\\
\;\;\;\;t\_0\\
\end{array}
\end{array}
if (+.f64 (-.f64 (exp.f64 x) #s(literal 2 binary64)) (exp.f64 (neg.f64 x))) < 1.00000000000000008e-5Initial program 54.1%
Taylor expanded in x around 0 100.0%
*-commutative100.0%
Simplified100.0%
unpow2100.0%
Applied egg-rr100.0%
unpow2100.0%
Applied egg-rr100.0%
unpow2100.0%
Applied egg-rr100.0%
if 1.00000000000000008e-5 < (+.f64 (-.f64 (exp.f64 x) #s(literal 2 binary64)) (exp.f64 (neg.f64 x))) Initial program 98.3%
x_m = (fabs.f64 x)
(FPCore (x_m)
:precision binary64
(if (<= (+ (- (exp x_m) 2.0) (exp (- x_m))) 1e-5)
(*
(* x_m x_m)
(+
1.0
(*
(* x_m x_m)
(+ 0.08333333333333333 (* (* x_m x_m) 0.002777777777777778)))))
(pow (cbrt (+ (* 2.0 (cosh x_m)) -2.0)) 3.0)))x_m = fabs(x);
double code(double x_m) {
double tmp;
if (((exp(x_m) - 2.0) + exp(-x_m)) <= 1e-5) {
tmp = (x_m * x_m) * (1.0 + ((x_m * x_m) * (0.08333333333333333 + ((x_m * x_m) * 0.002777777777777778))));
} else {
tmp = pow(cbrt(((2.0 * cosh(x_m)) + -2.0)), 3.0);
}
return tmp;
}
x_m = Math.abs(x);
public static double code(double x_m) {
double tmp;
if (((Math.exp(x_m) - 2.0) + Math.exp(-x_m)) <= 1e-5) {
tmp = (x_m * x_m) * (1.0 + ((x_m * x_m) * (0.08333333333333333 + ((x_m * x_m) * 0.002777777777777778))));
} else {
tmp = Math.pow(Math.cbrt(((2.0 * Math.cosh(x_m)) + -2.0)), 3.0);
}
return tmp;
}
x_m = abs(x) function code(x_m) tmp = 0.0 if (Float64(Float64(exp(x_m) - 2.0) + exp(Float64(-x_m))) <= 1e-5) tmp = Float64(Float64(x_m * x_m) * Float64(1.0 + Float64(Float64(x_m * x_m) * Float64(0.08333333333333333 + Float64(Float64(x_m * x_m) * 0.002777777777777778))))); else tmp = cbrt(Float64(Float64(2.0 * cosh(x_m)) + -2.0)) ^ 3.0; end return tmp end
x_m = N[Abs[x], $MachinePrecision] code[x$95$m_] := If[LessEqual[N[(N[(N[Exp[x$95$m], $MachinePrecision] - 2.0), $MachinePrecision] + N[Exp[(-x$95$m)], $MachinePrecision]), $MachinePrecision], 1e-5], N[(N[(x$95$m * x$95$m), $MachinePrecision] * N[(1.0 + N[(N[(x$95$m * x$95$m), $MachinePrecision] * N[(0.08333333333333333 + N[(N[(x$95$m * x$95$m), $MachinePrecision] * 0.002777777777777778), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision], N[Power[N[Power[N[(N[(2.0 * N[Cosh[x$95$m], $MachinePrecision]), $MachinePrecision] + -2.0), $MachinePrecision], 1/3], $MachinePrecision], 3.0], $MachinePrecision]]
\begin{array}{l}
x_m = \left|x\right|
\\
\begin{array}{l}
\mathbf{if}\;\left(e^{x\_m} - 2\right) + e^{-x\_m} \leq 10^{-5}:\\
\;\;\;\;\left(x\_m \cdot x\_m\right) \cdot \left(1 + \left(x\_m \cdot x\_m\right) \cdot \left(0.08333333333333333 + \left(x\_m \cdot x\_m\right) \cdot 0.002777777777777778\right)\right)\\
\mathbf{else}:\\
\;\;\;\;{\left(\sqrt[3]{2 \cdot \cosh x\_m + -2}\right)}^{3}\\
\end{array}
\end{array}
if (+.f64 (-.f64 (exp.f64 x) #s(literal 2 binary64)) (exp.f64 (neg.f64 x))) < 1.00000000000000008e-5Initial program 54.1%
Taylor expanded in x around 0 100.0%
*-commutative100.0%
Simplified100.0%
unpow2100.0%
Applied egg-rr100.0%
unpow2100.0%
Applied egg-rr100.0%
unpow2100.0%
Applied egg-rr100.0%
if 1.00000000000000008e-5 < (+.f64 (-.f64 (exp.f64 x) #s(literal 2 binary64)) (exp.f64 (neg.f64 x))) Initial program 98.3%
add-cube-cbrt97.4%
pow397.9%
Applied egg-rr98.5%
x_m = (fabs.f64 x)
(FPCore (x_m)
:precision binary64
(if (<= x_m 0.032)
(*
(* x_m x_m)
(+
1.0
(*
(* x_m x_m)
(+ 0.08333333333333333 (* (* x_m x_m) 0.002777777777777778)))))
(- (* 2.0 (cosh x_m)) 2.0)))x_m = fabs(x);
double code(double x_m) {
double tmp;
if (x_m <= 0.032) {
tmp = (x_m * x_m) * (1.0 + ((x_m * x_m) * (0.08333333333333333 + ((x_m * x_m) * 0.002777777777777778))));
} else {
tmp = (2.0 * cosh(x_m)) - 2.0;
}
return tmp;
}
x_m = abs(x)
real(8) function code(x_m)
real(8), intent (in) :: x_m
real(8) :: tmp
if (x_m <= 0.032d0) then
tmp = (x_m * x_m) * (1.0d0 + ((x_m * x_m) * (0.08333333333333333d0 + ((x_m * x_m) * 0.002777777777777778d0))))
else
tmp = (2.0d0 * cosh(x_m)) - 2.0d0
end if
code = tmp
end function
x_m = Math.abs(x);
public static double code(double x_m) {
double tmp;
if (x_m <= 0.032) {
tmp = (x_m * x_m) * (1.0 + ((x_m * x_m) * (0.08333333333333333 + ((x_m * x_m) * 0.002777777777777778))));
} else {
tmp = (2.0 * Math.cosh(x_m)) - 2.0;
}
return tmp;
}
x_m = math.fabs(x) def code(x_m): tmp = 0 if x_m <= 0.032: tmp = (x_m * x_m) * (1.0 + ((x_m * x_m) * (0.08333333333333333 + ((x_m * x_m) * 0.002777777777777778)))) else: tmp = (2.0 * math.cosh(x_m)) - 2.0 return tmp
x_m = abs(x) function code(x_m) tmp = 0.0 if (x_m <= 0.032) tmp = Float64(Float64(x_m * x_m) * Float64(1.0 + Float64(Float64(x_m * x_m) * Float64(0.08333333333333333 + Float64(Float64(x_m * x_m) * 0.002777777777777778))))); else tmp = Float64(Float64(2.0 * cosh(x_m)) - 2.0); end return tmp end
x_m = abs(x); function tmp_2 = code(x_m) tmp = 0.0; if (x_m <= 0.032) tmp = (x_m * x_m) * (1.0 + ((x_m * x_m) * (0.08333333333333333 + ((x_m * x_m) * 0.002777777777777778)))); else tmp = (2.0 * cosh(x_m)) - 2.0; end tmp_2 = tmp; end
x_m = N[Abs[x], $MachinePrecision] code[x$95$m_] := If[LessEqual[x$95$m, 0.032], N[(N[(x$95$m * x$95$m), $MachinePrecision] * N[(1.0 + N[(N[(x$95$m * x$95$m), $MachinePrecision] * N[(0.08333333333333333 + N[(N[(x$95$m * x$95$m), $MachinePrecision] * 0.002777777777777778), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision], N[(N[(2.0 * N[Cosh[x$95$m], $MachinePrecision]), $MachinePrecision] - 2.0), $MachinePrecision]]
\begin{array}{l}
x_m = \left|x\right|
\\
\begin{array}{l}
\mathbf{if}\;x\_m \leq 0.032:\\
\;\;\;\;\left(x\_m \cdot x\_m\right) \cdot \left(1 + \left(x\_m \cdot x\_m\right) \cdot \left(0.08333333333333333 + \left(x\_m \cdot x\_m\right) \cdot 0.002777777777777778\right)\right)\\
\mathbf{else}:\\
\;\;\;\;2 \cdot \cosh x\_m - 2\\
\end{array}
\end{array}
if x < 0.032000000000000001Initial program 54.9%
Taylor expanded in x around 0 98.7%
*-commutative98.7%
Simplified98.7%
unpow298.7%
Applied egg-rr98.7%
unpow298.7%
Applied egg-rr98.7%
unpow298.7%
Applied egg-rr98.7%
if 0.032000000000000001 < x Initial program 100.0%
+-commutative100.0%
associate-+r-100.0%
add-sqr-sqrt0.0%
sqrt-unprod18.7%
sqr-neg18.7%
sqrt-unprod18.7%
add-sqr-sqrt18.7%
add-sqr-sqrt18.7%
sqrt-unprod18.7%
sqr-neg18.7%
sqrt-unprod0.0%
add-sqr-sqrt100.0%
cosh-undef100.0%
Applied egg-rr100.0%
x_m = (fabs.f64 x)
(FPCore (x_m)
:precision binary64
(*
(* x_m x_m)
(+
1.0
(*
(* x_m x_m)
(+ 0.08333333333333333 (* (* x_m x_m) 0.002777777777777778))))))x_m = fabs(x);
double code(double x_m) {
return (x_m * x_m) * (1.0 + ((x_m * x_m) * (0.08333333333333333 + ((x_m * x_m) * 0.002777777777777778))));
}
x_m = abs(x)
real(8) function code(x_m)
real(8), intent (in) :: x_m
code = (x_m * x_m) * (1.0d0 + ((x_m * x_m) * (0.08333333333333333d0 + ((x_m * x_m) * 0.002777777777777778d0))))
end function
x_m = Math.abs(x);
public static double code(double x_m) {
return (x_m * x_m) * (1.0 + ((x_m * x_m) * (0.08333333333333333 + ((x_m * x_m) * 0.002777777777777778))));
}
x_m = math.fabs(x) def code(x_m): return (x_m * x_m) * (1.0 + ((x_m * x_m) * (0.08333333333333333 + ((x_m * x_m) * 0.002777777777777778))))
x_m = abs(x) function code(x_m) return Float64(Float64(x_m * x_m) * Float64(1.0 + Float64(Float64(x_m * x_m) * Float64(0.08333333333333333 + Float64(Float64(x_m * x_m) * 0.002777777777777778))))) end
x_m = abs(x); function tmp = code(x_m) tmp = (x_m * x_m) * (1.0 + ((x_m * x_m) * (0.08333333333333333 + ((x_m * x_m) * 0.002777777777777778)))); end
x_m = N[Abs[x], $MachinePrecision] code[x$95$m_] := N[(N[(x$95$m * x$95$m), $MachinePrecision] * N[(1.0 + N[(N[(x$95$m * x$95$m), $MachinePrecision] * N[(0.08333333333333333 + N[(N[(x$95$m * x$95$m), $MachinePrecision] * 0.002777777777777778), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
x_m = \left|x\right|
\\
\left(x\_m \cdot x\_m\right) \cdot \left(1 + \left(x\_m \cdot x\_m\right) \cdot \left(0.08333333333333333 + \left(x\_m \cdot x\_m\right) \cdot 0.002777777777777778\right)\right)
\end{array}
Initial program 55.1%
Taylor expanded in x around 0 98.3%
*-commutative98.3%
Simplified98.3%
unpow298.3%
Applied egg-rr98.3%
unpow298.3%
Applied egg-rr98.3%
unpow298.3%
Applied egg-rr98.3%
x_m = (fabs.f64 x) (FPCore (x_m) :precision binary64 (* (* x_m x_m) (+ 1.0 (* (* x_m x_m) 0.08333333333333333))))
x_m = fabs(x);
double code(double x_m) {
return (x_m * x_m) * (1.0 + ((x_m * x_m) * 0.08333333333333333));
}
x_m = abs(x)
real(8) function code(x_m)
real(8), intent (in) :: x_m
code = (x_m * x_m) * (1.0d0 + ((x_m * x_m) * 0.08333333333333333d0))
end function
x_m = Math.abs(x);
public static double code(double x_m) {
return (x_m * x_m) * (1.0 + ((x_m * x_m) * 0.08333333333333333));
}
x_m = math.fabs(x) def code(x_m): return (x_m * x_m) * (1.0 + ((x_m * x_m) * 0.08333333333333333))
x_m = abs(x) function code(x_m) return Float64(Float64(x_m * x_m) * Float64(1.0 + Float64(Float64(x_m * x_m) * 0.08333333333333333))) end
x_m = abs(x); function tmp = code(x_m) tmp = (x_m * x_m) * (1.0 + ((x_m * x_m) * 0.08333333333333333)); end
x_m = N[Abs[x], $MachinePrecision] code[x$95$m_] := N[(N[(x$95$m * x$95$m), $MachinePrecision] * N[(1.0 + N[(N[(x$95$m * x$95$m), $MachinePrecision] * 0.08333333333333333), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
x_m = \left|x\right|
\\
\left(x\_m \cdot x\_m\right) \cdot \left(1 + \left(x\_m \cdot x\_m\right) \cdot 0.08333333333333333\right)
\end{array}
Initial program 55.1%
Taylor expanded in x around 0 98.3%
*-commutative98.3%
Simplified98.3%
Taylor expanded in x around 0 98.1%
+-commutative98.1%
Simplified98.1%
unpow298.3%
Applied egg-rr98.1%
unpow298.3%
Applied egg-rr98.1%
Final simplification98.1%
x_m = (fabs.f64 x) (FPCore (x_m) :precision binary64 (* x_m x_m))
x_m = fabs(x);
double code(double x_m) {
return x_m * x_m;
}
x_m = abs(x)
real(8) function code(x_m)
real(8), intent (in) :: x_m
code = x_m * x_m
end function
x_m = Math.abs(x);
public static double code(double x_m) {
return x_m * x_m;
}
x_m = math.fabs(x) def code(x_m): return x_m * x_m
x_m = abs(x) function code(x_m) return Float64(x_m * x_m) end
x_m = abs(x); function tmp = code(x_m) tmp = x_m * x_m; end
x_m = N[Abs[x], $MachinePrecision] code[x$95$m_] := N[(x$95$m * x$95$m), $MachinePrecision]
\begin{array}{l}
x_m = \left|x\right|
\\
x\_m \cdot x\_m
\end{array}
Initial program 55.1%
Taylor expanded in x around 0 97.7%
unpow298.3%
Applied egg-rr97.7%
(FPCore (x) :precision binary64 (let* ((t_0 (sinh (/ x 2.0)))) (* 4.0 (* t_0 t_0))))
double code(double x) {
double t_0 = sinh((x / 2.0));
return 4.0 * (t_0 * t_0);
}
real(8) function code(x)
real(8), intent (in) :: x
real(8) :: t_0
t_0 = sinh((x / 2.0d0))
code = 4.0d0 * (t_0 * t_0)
end function
public static double code(double x) {
double t_0 = Math.sinh((x / 2.0));
return 4.0 * (t_0 * t_0);
}
def code(x): t_0 = math.sinh((x / 2.0)) return 4.0 * (t_0 * t_0)
function code(x) t_0 = sinh(Float64(x / 2.0)) return Float64(4.0 * Float64(t_0 * t_0)) end
function tmp = code(x) t_0 = sinh((x / 2.0)); tmp = 4.0 * (t_0 * t_0); end
code[x_] := Block[{t$95$0 = N[Sinh[N[(x / 2.0), $MachinePrecision]], $MachinePrecision]}, N[(4.0 * N[(t$95$0 * t$95$0), $MachinePrecision]), $MachinePrecision]]
\begin{array}{l}
\\
\begin{array}{l}
t_0 := \sinh \left(\frac{x}{2}\right)\\
4 \cdot \left(t\_0 \cdot t\_0\right)
\end{array}
\end{array}
herbie shell --seed 2024118
(FPCore (x)
:name "exp2 (problem 3.3.7)"
:precision binary64
:pre (<= (fabs x) 710.0)
:alt
(! :herbie-platform default (* 4 (* (sinh (/ x 2)) (sinh (/ x 2)))))
(+ (- (exp x) 2.0) (exp (- x))))