?

Average Accuracy: 53.3% → 99.2%
Time: 13.1s
Precision: binary64
Cost: 13632

?

\[\frac{\left(1 + \frac{1}{\varepsilon}\right) \cdot e^{-\left(1 - \varepsilon\right) \cdot x} - \left(\frac{1}{\varepsilon} - 1\right) \cdot e^{-\left(1 + \varepsilon\right) \cdot x}}{2} \]
\[\begin{array}{l} t_0 := \frac{x + 1}{e^{x}}\\ \frac{t_0 + t_0}{2} \end{array} \]
(FPCore (x eps)
 :precision binary64
 (/
  (-
   (* (+ 1.0 (/ 1.0 eps)) (exp (- (* (- 1.0 eps) x))))
   (* (- (/ 1.0 eps) 1.0) (exp (- (* (+ 1.0 eps) x)))))
  2.0))
(FPCore (x eps)
 :precision binary64
 (let* ((t_0 (/ (+ x 1.0) (exp x)))) (/ (+ t_0 t_0) 2.0)))
double code(double x, double eps) {
	return (((1.0 + (1.0 / eps)) * exp(-((1.0 - eps) * x))) - (((1.0 / eps) - 1.0) * exp(-((1.0 + eps) * x)))) / 2.0;
}
double code(double x, double eps) {
	double t_0 = (x + 1.0) / exp(x);
	return (t_0 + t_0) / 2.0;
}
real(8) function code(x, eps)
    real(8), intent (in) :: x
    real(8), intent (in) :: eps
    code = (((1.0d0 + (1.0d0 / eps)) * exp(-((1.0d0 - eps) * x))) - (((1.0d0 / eps) - 1.0d0) * exp(-((1.0d0 + eps) * x)))) / 2.0d0
end function
real(8) function code(x, eps)
    real(8), intent (in) :: x
    real(8), intent (in) :: eps
    real(8) :: t_0
    t_0 = (x + 1.0d0) / exp(x)
    code = (t_0 + t_0) / 2.0d0
end function
public static double code(double x, double eps) {
	return (((1.0 + (1.0 / eps)) * Math.exp(-((1.0 - eps) * x))) - (((1.0 / eps) - 1.0) * Math.exp(-((1.0 + eps) * x)))) / 2.0;
}
public static double code(double x, double eps) {
	double t_0 = (x + 1.0) / Math.exp(x);
	return (t_0 + t_0) / 2.0;
}
def code(x, eps):
	return (((1.0 + (1.0 / eps)) * math.exp(-((1.0 - eps) * x))) - (((1.0 / eps) - 1.0) * math.exp(-((1.0 + eps) * x)))) / 2.0
def code(x, eps):
	t_0 = (x + 1.0) / math.exp(x)
	return (t_0 + t_0) / 2.0
function code(x, eps)
	return Float64(Float64(Float64(Float64(1.0 + Float64(1.0 / eps)) * exp(Float64(-Float64(Float64(1.0 - eps) * x)))) - Float64(Float64(Float64(1.0 / eps) - 1.0) * exp(Float64(-Float64(Float64(1.0 + eps) * x))))) / 2.0)
end
function code(x, eps)
	t_0 = Float64(Float64(x + 1.0) / exp(x))
	return Float64(Float64(t_0 + t_0) / 2.0)
end
function tmp = code(x, eps)
	tmp = (((1.0 + (1.0 / eps)) * exp(-((1.0 - eps) * x))) - (((1.0 / eps) - 1.0) * exp(-((1.0 + eps) * x)))) / 2.0;
end
function tmp = code(x, eps)
	t_0 = (x + 1.0) / exp(x);
	tmp = (t_0 + t_0) / 2.0;
end
code[x_, eps_] := N[(N[(N[(N[(1.0 + N[(1.0 / eps), $MachinePrecision]), $MachinePrecision] * N[Exp[(-N[(N[(1.0 - eps), $MachinePrecision] * x), $MachinePrecision])], $MachinePrecision]), $MachinePrecision] - N[(N[(N[(1.0 / eps), $MachinePrecision] - 1.0), $MachinePrecision] * N[Exp[(-N[(N[(1.0 + eps), $MachinePrecision] * x), $MachinePrecision])], $MachinePrecision]), $MachinePrecision]), $MachinePrecision] / 2.0), $MachinePrecision]
code[x_, eps_] := Block[{t$95$0 = N[(N[(x + 1.0), $MachinePrecision] / N[Exp[x], $MachinePrecision]), $MachinePrecision]}, N[(N[(t$95$0 + t$95$0), $MachinePrecision] / 2.0), $MachinePrecision]]
\frac{\left(1 + \frac{1}{\varepsilon}\right) \cdot e^{-\left(1 - \varepsilon\right) \cdot x} - \left(\frac{1}{\varepsilon} - 1\right) \cdot e^{-\left(1 + \varepsilon\right) \cdot x}}{2}
\begin{array}{l}
t_0 := \frac{x + 1}{e^{x}}\\
\frac{t_0 + t_0}{2}
\end{array}

Error?

Try it out?

Your Program's Arguments

Results

Enter valid numbers for all inputs

Derivation?

  1. Initial program 53.3%

    \[\frac{\left(1 + \frac{1}{\varepsilon}\right) \cdot e^{-\left(1 - \varepsilon\right) \cdot x} - \left(\frac{1}{\varepsilon} - 1\right) \cdot e^{-\left(1 + \varepsilon\right) \cdot x}}{2} \]
  2. Simplified53.3%

    \[\leadsto \color{blue}{\frac{\left(1 + \frac{1}{\varepsilon}\right) \cdot e^{\left(1 - \varepsilon\right) \cdot \left(-x\right)} - \left(\frac{1}{\varepsilon} + -1\right) \cdot e^{\left(1 + \varepsilon\right) \cdot \left(-x\right)}}{2}} \]
    Proof

    [Start]53.3

    \[ \frac{\left(1 + \frac{1}{\varepsilon}\right) \cdot e^{-\left(1 - \varepsilon\right) \cdot x} - \left(\frac{1}{\varepsilon} - 1\right) \cdot e^{-\left(1 + \varepsilon\right) \cdot x}}{2} \]

    distribute-rgt-neg-in [=>]53.3

    \[ \frac{\left(1 + \frac{1}{\varepsilon}\right) \cdot e^{\color{blue}{\left(1 - \varepsilon\right) \cdot \left(-x\right)}} - \left(\frac{1}{\varepsilon} - 1\right) \cdot e^{-\left(1 + \varepsilon\right) \cdot x}}{2} \]

    sub-neg [=>]53.3

    \[ \frac{\left(1 + \frac{1}{\varepsilon}\right) \cdot e^{\left(1 - \varepsilon\right) \cdot \left(-x\right)} - \color{blue}{\left(\frac{1}{\varepsilon} + \left(-1\right)\right)} \cdot e^{-\left(1 + \varepsilon\right) \cdot x}}{2} \]

    metadata-eval [=>]53.3

    \[ \frac{\left(1 + \frac{1}{\varepsilon}\right) \cdot e^{\left(1 - \varepsilon\right) \cdot \left(-x\right)} - \left(\frac{1}{\varepsilon} + \color{blue}{-1}\right) \cdot e^{-\left(1 + \varepsilon\right) \cdot x}}{2} \]

    distribute-rgt-neg-in [=>]53.3

    \[ \frac{\left(1 + \frac{1}{\varepsilon}\right) \cdot e^{\left(1 - \varepsilon\right) \cdot \left(-x\right)} - \left(\frac{1}{\varepsilon} + -1\right) \cdot e^{\color{blue}{\left(1 + \varepsilon\right) \cdot \left(-x\right)}}}{2} \]
  3. Taylor expanded in eps around 0 99.2%

    \[\leadsto \frac{\color{blue}{\left(e^{-1 \cdot x} \cdot x + e^{-1 \cdot x}\right) - \left(-1 \cdot \left(e^{-1 \cdot x} \cdot x\right) + -1 \cdot e^{-1 \cdot x}\right)}}{2} \]
  4. Simplified99.2%

    \[\leadsto \frac{\color{blue}{\left(x + 1\right) \cdot e^{-x} + 1 \cdot \left(\left(x + 1\right) \cdot e^{-x}\right)}}{2} \]
    Proof

    [Start]99.2

    \[ \frac{\left(e^{-1 \cdot x} \cdot x + e^{-1 \cdot x}\right) - \left(-1 \cdot \left(e^{-1 \cdot x} \cdot x\right) + -1 \cdot e^{-1 \cdot x}\right)}{2} \]

    sub-neg [=>]99.2

    \[ \frac{\color{blue}{\left(e^{-1 \cdot x} \cdot x + e^{-1 \cdot x}\right) + \left(-\left(-1 \cdot \left(e^{-1 \cdot x} \cdot x\right) + -1 \cdot e^{-1 \cdot x}\right)\right)}}{2} \]

    *-commutative [=>]99.2

    \[ \frac{\left(\color{blue}{x \cdot e^{-1 \cdot x}} + e^{-1 \cdot x}\right) + \left(-\left(-1 \cdot \left(e^{-1 \cdot x} \cdot x\right) + -1 \cdot e^{-1 \cdot x}\right)\right)}{2} \]

    distribute-lft1-in [=>]99.2

    \[ \frac{\color{blue}{\left(x + 1\right) \cdot e^{-1 \cdot x}} + \left(-\left(-1 \cdot \left(e^{-1 \cdot x} \cdot x\right) + -1 \cdot e^{-1 \cdot x}\right)\right)}{2} \]

    mul-1-neg [=>]99.2

    \[ \frac{\left(x + 1\right) \cdot e^{\color{blue}{-x}} + \left(-\left(-1 \cdot \left(e^{-1 \cdot x} \cdot x\right) + -1 \cdot e^{-1 \cdot x}\right)\right)}{2} \]

    distribute-lft-out [=>]99.2

    \[ \frac{\left(x + 1\right) \cdot e^{-x} + \left(-\color{blue}{-1 \cdot \left(e^{-1 \cdot x} \cdot x + e^{-1 \cdot x}\right)}\right)}{2} \]

    distribute-lft-neg-in [=>]99.2

    \[ \frac{\left(x + 1\right) \cdot e^{-x} + \color{blue}{\left(--1\right) \cdot \left(e^{-1 \cdot x} \cdot x + e^{-1 \cdot x}\right)}}{2} \]

    metadata-eval [=>]99.2

    \[ \frac{\left(x + 1\right) \cdot e^{-x} + \color{blue}{1} \cdot \left(e^{-1 \cdot x} \cdot x + e^{-1 \cdot x}\right)}{2} \]

    *-commutative [=>]99.2

    \[ \frac{\left(x + 1\right) \cdot e^{-x} + 1 \cdot \left(\color{blue}{x \cdot e^{-1 \cdot x}} + e^{-1 \cdot x}\right)}{2} \]

    distribute-lft1-in [=>]99.2

    \[ \frac{\left(x + 1\right) \cdot e^{-x} + 1 \cdot \color{blue}{\left(\left(x + 1\right) \cdot e^{-1 \cdot x}\right)}}{2} \]

    mul-1-neg [=>]99.2

    \[ \frac{\left(x + 1\right) \cdot e^{-x} + 1 \cdot \left(\left(x + 1\right) \cdot e^{\color{blue}{-x}}\right)}{2} \]
  5. Applied egg-rr99.2%

    \[\leadsto \frac{\color{blue}{\frac{x + 1}{e^{x}}} + 1 \cdot \left(\left(x + 1\right) \cdot e^{-x}\right)}{2} \]
    Proof
  6. No proof available- proof too large to flatten.
  7. Applied egg-rr99.2%

    \[\leadsto \frac{\frac{x + 1}{e^{x}} + 1 \cdot \color{blue}{\frac{x + 1}{e^{x}}}}{2} \]
    Proof
  8. No proof available- proof too large to flatten.
  9. Final simplification99.2%

    \[\leadsto \frac{\frac{x + 1}{e^{x}} + \frac{x + 1}{e^{x}}}{2} \]

Alternatives

Alternative 1
Accuracy98.5%
Cost13632
\[\frac{e^{x \cdot \left(-1 + \varepsilon\right)} + e^{x \cdot \left(-1 - \varepsilon\right)}}{2} \]
Alternative 2
Accuracy98.5%
Cost6852
\[\begin{array}{l} \mathbf{if}\;x \leq 1.5:\\ \;\;\;\;\frac{\left(1 + \left(x \cdot x\right) \cdot \left(x \cdot \left(0.3333333333333333 + x \cdot -0.125\right) + -0.5\right)\right) + \left(1 + \left(x \cdot x\right) \cdot -0.5\right)}{2}\\ \mathbf{else}:\\ \;\;\;\;\frac{\frac{x}{e^{x}}}{2}\\ \end{array} \]
Alternative 3
Accuracy97.7%
Cost6784
\[\frac{2 \cdot e^{-x}}{2} \]
Alternative 4
Accuracy98.5%
Cost1732
\[\begin{array}{l} \mathbf{if}\;x \leq 1.6:\\ \;\;\;\;\frac{\left(1 + \left(x \cdot x\right) \cdot \left(x \cdot \left(0.3333333333333333 + x \cdot -0.125\right) + -0.5\right)\right) + \left(1 + \left(x \cdot x\right) \cdot -0.5\right)}{2}\\ \mathbf{else}:\\ \;\;\;\;0\\ \end{array} \]
Alternative 5
Accuracy98.5%
Cost580
\[\begin{array}{l} \mathbf{if}\;x \leq 1.4:\\ \;\;\;\;\frac{2 - x \cdot x}{2}\\ \mathbf{else}:\\ \;\;\;\;0\\ \end{array} \]
Alternative 6
Accuracy98.3%
Cost196
\[\begin{array}{l} \mathbf{if}\;x \leq 360:\\ \;\;\;\;1\\ \mathbf{else}:\\ \;\;\;\;0\\ \end{array} \]
Alternative 7
Accuracy26.6%
Cost64
\[0 \]

Error

Reproduce?

herbie shell --seed 2023133 
(FPCore (x eps)
  :name "NMSE Section 6.1 mentioned, A"
  :precision binary64
  (/ (- (* (+ 1.0 (/ 1.0 eps)) (exp (- (* (- 1.0 eps) x)))) (* (- (/ 1.0 eps) 1.0) (exp (- (* (+ 1.0 eps) x))))) 2.0))