ENA, Section 1.4, Exercise 4b, n=2

Percentage Accurate: 76.0% → 100.0%
Time: 6.1s
Alternatives: 5
Speedup: 29.6×

Specification

?
\[\left(-1000000000 \leq x \land x \leq 1000000000\right) \land \left(-1 \leq \varepsilon \land \varepsilon \leq 1\right)\]
\[\begin{array}{l} \\ {\left(x + \varepsilon\right)}^{2} - {x}^{2} \end{array} \]
(FPCore (x eps) :precision binary64 (- (pow (+ x eps) 2.0) (pow x 2.0)))
double code(double x, double eps) {
	return pow((x + eps), 2.0) - pow(x, 2.0);
}
real(8) function code(x, eps)
    real(8), intent (in) :: x
    real(8), intent (in) :: eps
    code = ((x + eps) ** 2.0d0) - (x ** 2.0d0)
end function
public static double code(double x, double eps) {
	return Math.pow((x + eps), 2.0) - Math.pow(x, 2.0);
}
def code(x, eps):
	return math.pow((x + eps), 2.0) - math.pow(x, 2.0)
function code(x, eps)
	return Float64((Float64(x + eps) ^ 2.0) - (x ^ 2.0))
end
function tmp = code(x, eps)
	tmp = ((x + eps) ^ 2.0) - (x ^ 2.0);
end
code[x_, eps_] := N[(N[Power[N[(x + eps), $MachinePrecision], 2.0], $MachinePrecision] - N[Power[x, 2.0], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}

\\
{\left(x + \varepsilon\right)}^{2} - {x}^{2}
\end{array}

Sampling outcomes in binary64 precision:

Local Percentage Accuracy vs ?

The average percentage accuracy by input value. Horizontal axis shows value of an input variable; the variable is choosen in the title. Vertical axis is accuracy; higher is better. Red represent the original program, while blue represents Herbie's suggestion. These can be toggled with buttons below the plot. The line is an average while dots represent individual samples.

Accuracy vs Speed?

Herbie found 5 alternatives:

AlternativeAccuracySpeedup
The accuracy (vertical axis) and speed (horizontal axis) of each alternatives. Up and to the right is better. The red square shows the initial program, and each blue circle shows an alternative.The line shows the best available speed-accuracy tradeoffs.

Initial Program: 76.0% accurate, 1.0× speedup?

\[\begin{array}{l} \\ {\left(x + \varepsilon\right)}^{2} - {x}^{2} \end{array} \]
(FPCore (x eps) :precision binary64 (- (pow (+ x eps) 2.0) (pow x 2.0)))
double code(double x, double eps) {
	return pow((x + eps), 2.0) - pow(x, 2.0);
}
real(8) function code(x, eps)
    real(8), intent (in) :: x
    real(8), intent (in) :: eps
    code = ((x + eps) ** 2.0d0) - (x ** 2.0d0)
end function
public static double code(double x, double eps) {
	return Math.pow((x + eps), 2.0) - Math.pow(x, 2.0);
}
def code(x, eps):
	return math.pow((x + eps), 2.0) - math.pow(x, 2.0)
function code(x, eps)
	return Float64((Float64(x + eps) ^ 2.0) - (x ^ 2.0))
end
function tmp = code(x, eps)
	tmp = ((x + eps) ^ 2.0) - (x ^ 2.0);
end
code[x_, eps_] := N[(N[Power[N[(x + eps), $MachinePrecision], 2.0], $MachinePrecision] - N[Power[x, 2.0], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}

\\
{\left(x + \varepsilon\right)}^{2} - {x}^{2}
\end{array}

Alternative 1: 100.0% accurate, 1.9× speedup?

\[\begin{array}{l} \\ \mathsf{fma}\left(\varepsilon, \varepsilon, x \cdot \left(\varepsilon \cdot 2\right)\right) \end{array} \]
(FPCore (x eps) :precision binary64 (fma eps eps (* x (* eps 2.0))))
double code(double x, double eps) {
	return fma(eps, eps, (x * (eps * 2.0)));
}
function code(x, eps)
	return fma(eps, eps, Float64(x * Float64(eps * 2.0)))
end
code[x_, eps_] := N[(eps * eps + N[(x * N[(eps * 2.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}

\\
\mathsf{fma}\left(\varepsilon, \varepsilon, x \cdot \left(\varepsilon \cdot 2\right)\right)
\end{array}
Derivation
  1. Initial program 74.4%

    \[{\left(x + \varepsilon\right)}^{2} - {x}^{2} \]
  2. Step-by-step derivation
    1. +-commutative74.4%

      \[\leadsto {\color{blue}{\left(\varepsilon + x\right)}}^{2} - {x}^{2} \]
    2. unpow274.4%

      \[\leadsto \color{blue}{\left(\varepsilon + x\right) \cdot \left(\varepsilon + x\right)} - {x}^{2} \]
    3. unpow274.4%

      \[\leadsto \left(\varepsilon + x\right) \cdot \left(\varepsilon + x\right) - \color{blue}{x \cdot x} \]
    4. difference-of-squares74.4%

      \[\leadsto \color{blue}{\left(\left(\varepsilon + x\right) + x\right) \cdot \left(\left(\varepsilon + x\right) - x\right)} \]
    5. sub-neg74.4%

      \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \color{blue}{\left(\left(\varepsilon + x\right) + \left(-x\right)\right)} \]
    6. distribute-lft-in74.3%

      \[\leadsto \color{blue}{\left(\left(\varepsilon + x\right) + x\right) \cdot \left(\varepsilon + x\right) + \left(\left(\varepsilon + x\right) + x\right) \cdot \left(-x\right)} \]
    7. +-commutative74.3%

      \[\leadsto \color{blue}{\left(\left(\varepsilon + x\right) + x\right) \cdot \left(-x\right) + \left(\left(\varepsilon + x\right) + x\right) \cdot \left(\varepsilon + x\right)} \]
    8. distribute-lft-in74.4%

      \[\leadsto \color{blue}{\left(\left(\varepsilon + x\right) + x\right) \cdot \left(\left(-x\right) + \left(\varepsilon + x\right)\right)} \]
    9. associate-+l+74.4%

      \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \color{blue}{\left(\left(\left(-x\right) + \varepsilon\right) + x\right)} \]
    10. remove-double-neg74.4%

      \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \left(\left(\left(-x\right) + \varepsilon\right) + \color{blue}{\left(-\left(-x\right)\right)}\right) \]
    11. sub-neg74.4%

      \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \color{blue}{\left(\left(\left(-x\right) + \varepsilon\right) - \left(-x\right)\right)} \]
    12. +-commutative74.4%

      \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \left(\color{blue}{\left(\varepsilon + \left(-x\right)\right)} - \left(-x\right)\right) \]
    13. associate--l+99.9%

      \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \color{blue}{\left(\varepsilon + \left(\left(-x\right) - \left(-x\right)\right)\right)} \]
    14. +-inverses99.9%

      \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \left(\varepsilon + \color{blue}{0}\right) \]
    15. +-rgt-identity99.9%

      \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \color{blue}{\varepsilon} \]
    16. *-commutative99.9%

      \[\leadsto \color{blue}{\varepsilon \cdot \left(\left(\varepsilon + x\right) + x\right)} \]
    17. associate-+l+100.0%

      \[\leadsto \varepsilon \cdot \color{blue}{\left(\varepsilon + \left(x + x\right)\right)} \]
    18. count-2100.0%

      \[\leadsto \varepsilon \cdot \left(\varepsilon + \color{blue}{2 \cdot x}\right) \]
    19. *-commutative100.0%

      \[\leadsto \varepsilon \cdot \left(\varepsilon + \color{blue}{x \cdot 2}\right) \]
  3. Simplified100.0%

    \[\leadsto \color{blue}{\varepsilon \cdot \left(\varepsilon + x \cdot 2\right)} \]
  4. Add Preprocessing
  5. Step-by-step derivation
    1. distribute-rgt-in100.0%

      \[\leadsto \color{blue}{\varepsilon \cdot \varepsilon + \left(x \cdot 2\right) \cdot \varepsilon} \]
    2. fma-define100.0%

      \[\leadsto \color{blue}{\mathsf{fma}\left(\varepsilon, \varepsilon, \left(x \cdot 2\right) \cdot \varepsilon\right)} \]
    3. associate-*l*100.0%

      \[\leadsto \mathsf{fma}\left(\varepsilon, \varepsilon, \color{blue}{x \cdot \left(2 \cdot \varepsilon\right)}\right) \]
  6. Applied egg-rr100.0%

    \[\leadsto \color{blue}{\mathsf{fma}\left(\varepsilon, \varepsilon, x \cdot \left(2 \cdot \varepsilon\right)\right)} \]
  7. Final simplification100.0%

    \[\leadsto \mathsf{fma}\left(\varepsilon, \varepsilon, x \cdot \left(\varepsilon \cdot 2\right)\right) \]
  8. Add Preprocessing

Alternative 2: 90.2% accurate, 13.8× speedup?

\[\begin{array}{l} \\ \begin{array}{l} \mathbf{if}\;x \leq -4.3 \cdot 10^{-116} \lor \neg \left(x \leq 1.56 \cdot 10^{-74}\right):\\ \;\;\;\;x \cdot \left(\varepsilon \cdot 2\right)\\ \mathbf{else}:\\ \;\;\;\;\varepsilon \cdot \varepsilon\\ \end{array} \end{array} \]
(FPCore (x eps)
 :precision binary64
 (if (or (<= x -4.3e-116) (not (<= x 1.56e-74)))
   (* x (* eps 2.0))
   (* eps eps)))
double code(double x, double eps) {
	double tmp;
	if ((x <= -4.3e-116) || !(x <= 1.56e-74)) {
		tmp = x * (eps * 2.0);
	} else {
		tmp = eps * eps;
	}
	return tmp;
}
real(8) function code(x, eps)
    real(8), intent (in) :: x
    real(8), intent (in) :: eps
    real(8) :: tmp
    if ((x <= (-4.3d-116)) .or. (.not. (x <= 1.56d-74))) then
        tmp = x * (eps * 2.0d0)
    else
        tmp = eps * eps
    end if
    code = tmp
end function
public static double code(double x, double eps) {
	double tmp;
	if ((x <= -4.3e-116) || !(x <= 1.56e-74)) {
		tmp = x * (eps * 2.0);
	} else {
		tmp = eps * eps;
	}
	return tmp;
}
def code(x, eps):
	tmp = 0
	if (x <= -4.3e-116) or not (x <= 1.56e-74):
		tmp = x * (eps * 2.0)
	else:
		tmp = eps * eps
	return tmp
function code(x, eps)
	tmp = 0.0
	if ((x <= -4.3e-116) || !(x <= 1.56e-74))
		tmp = Float64(x * Float64(eps * 2.0));
	else
		tmp = Float64(eps * eps);
	end
	return tmp
end
function tmp_2 = code(x, eps)
	tmp = 0.0;
	if ((x <= -4.3e-116) || ~((x <= 1.56e-74)))
		tmp = x * (eps * 2.0);
	else
		tmp = eps * eps;
	end
	tmp_2 = tmp;
end
code[x_, eps_] := If[Or[LessEqual[x, -4.3e-116], N[Not[LessEqual[x, 1.56e-74]], $MachinePrecision]], N[(x * N[(eps * 2.0), $MachinePrecision]), $MachinePrecision], N[(eps * eps), $MachinePrecision]]
\begin{array}{l}

\\
\begin{array}{l}
\mathbf{if}\;x \leq -4.3 \cdot 10^{-116} \lor \neg \left(x \leq 1.56 \cdot 10^{-74}\right):\\
\;\;\;\;x \cdot \left(\varepsilon \cdot 2\right)\\

\mathbf{else}:\\
\;\;\;\;\varepsilon \cdot \varepsilon\\


\end{array}
\end{array}
Derivation
  1. Split input into 2 regimes
  2. if x < -4.2999999999999997e-116 or 1.5600000000000001e-74 < x

    1. Initial program 23.9%

      \[{\left(x + \varepsilon\right)}^{2} - {x}^{2} \]
    2. Step-by-step derivation
      1. +-commutative23.9%

        \[\leadsto {\color{blue}{\left(\varepsilon + x\right)}}^{2} - {x}^{2} \]
      2. unpow223.9%

        \[\leadsto \color{blue}{\left(\varepsilon + x\right) \cdot \left(\varepsilon + x\right)} - {x}^{2} \]
      3. unpow223.9%

        \[\leadsto \left(\varepsilon + x\right) \cdot \left(\varepsilon + x\right) - \color{blue}{x \cdot x} \]
      4. difference-of-squares23.9%

        \[\leadsto \color{blue}{\left(\left(\varepsilon + x\right) + x\right) \cdot \left(\left(\varepsilon + x\right) - x\right)} \]
      5. sub-neg23.9%

        \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \color{blue}{\left(\left(\varepsilon + x\right) + \left(-x\right)\right)} \]
      6. distribute-lft-in23.8%

        \[\leadsto \color{blue}{\left(\left(\varepsilon + x\right) + x\right) \cdot \left(\varepsilon + x\right) + \left(\left(\varepsilon + x\right) + x\right) \cdot \left(-x\right)} \]
      7. +-commutative23.8%

        \[\leadsto \color{blue}{\left(\left(\varepsilon + x\right) + x\right) \cdot \left(-x\right) + \left(\left(\varepsilon + x\right) + x\right) \cdot \left(\varepsilon + x\right)} \]
      8. distribute-lft-in23.9%

        \[\leadsto \color{blue}{\left(\left(\varepsilon + x\right) + x\right) \cdot \left(\left(-x\right) + \left(\varepsilon + x\right)\right)} \]
      9. associate-+l+23.9%

        \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \color{blue}{\left(\left(\left(-x\right) + \varepsilon\right) + x\right)} \]
      10. remove-double-neg23.9%

        \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \left(\left(\left(-x\right) + \varepsilon\right) + \color{blue}{\left(-\left(-x\right)\right)}\right) \]
      11. sub-neg23.9%

        \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \color{blue}{\left(\left(\left(-x\right) + \varepsilon\right) - \left(-x\right)\right)} \]
      12. +-commutative23.9%

        \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \left(\color{blue}{\left(\varepsilon + \left(-x\right)\right)} - \left(-x\right)\right) \]
      13. associate--l+99.9%

        \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \color{blue}{\left(\varepsilon + \left(\left(-x\right) - \left(-x\right)\right)\right)} \]
      14. +-inverses99.9%

        \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \left(\varepsilon + \color{blue}{0}\right) \]
      15. +-rgt-identity99.9%

        \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \color{blue}{\varepsilon} \]
      16. *-commutative99.9%

        \[\leadsto \color{blue}{\varepsilon \cdot \left(\left(\varepsilon + x\right) + x\right)} \]
      17. associate-+l+100.0%

        \[\leadsto \varepsilon \cdot \color{blue}{\left(\varepsilon + \left(x + x\right)\right)} \]
      18. count-2100.0%

        \[\leadsto \varepsilon \cdot \left(\varepsilon + \color{blue}{2 \cdot x}\right) \]
      19. *-commutative100.0%

        \[\leadsto \varepsilon \cdot \left(\varepsilon + \color{blue}{x \cdot 2}\right) \]
    3. Simplified100.0%

      \[\leadsto \color{blue}{\varepsilon \cdot \left(\varepsilon + x \cdot 2\right)} \]
    4. Add Preprocessing
    5. Taylor expanded in x around inf 100.0%

      \[\leadsto \color{blue}{x \cdot \left(2 \cdot \varepsilon + \frac{{\varepsilon}^{2}}{x}\right)} \]
    6. Step-by-step derivation
      1. *-commutative100.0%

        \[\leadsto x \cdot \left(\color{blue}{\varepsilon \cdot 2} + \frac{{\varepsilon}^{2}}{x}\right) \]
      2. unpow2100.0%

        \[\leadsto x \cdot \left(\varepsilon \cdot 2 + \frac{\color{blue}{\varepsilon \cdot \varepsilon}}{x}\right) \]
      3. associate-/l*100.0%

        \[\leadsto x \cdot \left(\varepsilon \cdot 2 + \color{blue}{\varepsilon \cdot \frac{\varepsilon}{x}}\right) \]
      4. distribute-lft-out99.9%

        \[\leadsto x \cdot \color{blue}{\left(\varepsilon \cdot \left(2 + \frac{\varepsilon}{x}\right)\right)} \]
    7. Simplified99.9%

      \[\leadsto \color{blue}{x \cdot \left(\varepsilon \cdot \left(2 + \frac{\varepsilon}{x}\right)\right)} \]
    8. Taylor expanded in eps around 0 89.2%

      \[\leadsto x \cdot \left(\varepsilon \cdot \color{blue}{2}\right) \]

    if -4.2999999999999997e-116 < x < 1.5600000000000001e-74

    1. Initial program 95.6%

      \[{\left(x + \varepsilon\right)}^{2} - {x}^{2} \]
    2. Step-by-step derivation
      1. +-commutative95.6%

        \[\leadsto {\color{blue}{\left(\varepsilon + x\right)}}^{2} - {x}^{2} \]
      2. unpow295.6%

        \[\leadsto \color{blue}{\left(\varepsilon + x\right) \cdot \left(\varepsilon + x\right)} - {x}^{2} \]
      3. unpow295.6%

        \[\leadsto \left(\varepsilon + x\right) \cdot \left(\varepsilon + x\right) - \color{blue}{x \cdot x} \]
      4. difference-of-squares95.7%

        \[\leadsto \color{blue}{\left(\left(\varepsilon + x\right) + x\right) \cdot \left(\left(\varepsilon + x\right) - x\right)} \]
      5. sub-neg95.7%

        \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \color{blue}{\left(\left(\varepsilon + x\right) + \left(-x\right)\right)} \]
      6. distribute-lft-in95.7%

        \[\leadsto \color{blue}{\left(\left(\varepsilon + x\right) + x\right) \cdot \left(\varepsilon + x\right) + \left(\left(\varepsilon + x\right) + x\right) \cdot \left(-x\right)} \]
      7. +-commutative95.7%

        \[\leadsto \color{blue}{\left(\left(\varepsilon + x\right) + x\right) \cdot \left(-x\right) + \left(\left(\varepsilon + x\right) + x\right) \cdot \left(\varepsilon + x\right)} \]
      8. distribute-lft-in95.7%

        \[\leadsto \color{blue}{\left(\left(\varepsilon + x\right) + x\right) \cdot \left(\left(-x\right) + \left(\varepsilon + x\right)\right)} \]
      9. associate-+l+95.7%

        \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \color{blue}{\left(\left(\left(-x\right) + \varepsilon\right) + x\right)} \]
      10. remove-double-neg95.7%

        \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \left(\left(\left(-x\right) + \varepsilon\right) + \color{blue}{\left(-\left(-x\right)\right)}\right) \]
      11. sub-neg95.7%

        \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \color{blue}{\left(\left(\left(-x\right) + \varepsilon\right) - \left(-x\right)\right)} \]
      12. +-commutative95.7%

        \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \left(\color{blue}{\left(\varepsilon + \left(-x\right)\right)} - \left(-x\right)\right) \]
      13. associate--l+100.0%

        \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \color{blue}{\left(\varepsilon + \left(\left(-x\right) - \left(-x\right)\right)\right)} \]
      14. +-inverses100.0%

        \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \left(\varepsilon + \color{blue}{0}\right) \]
      15. +-rgt-identity100.0%

        \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \color{blue}{\varepsilon} \]
      16. *-commutative100.0%

        \[\leadsto \color{blue}{\varepsilon \cdot \left(\left(\varepsilon + x\right) + x\right)} \]
      17. associate-+l+100.0%

        \[\leadsto \varepsilon \cdot \color{blue}{\left(\varepsilon + \left(x + x\right)\right)} \]
      18. count-2100.0%

        \[\leadsto \varepsilon \cdot \left(\varepsilon + \color{blue}{2 \cdot x}\right) \]
      19. *-commutative100.0%

        \[\leadsto \varepsilon \cdot \left(\varepsilon + \color{blue}{x \cdot 2}\right) \]
    3. Simplified100.0%

      \[\leadsto \color{blue}{\varepsilon \cdot \left(\varepsilon + x \cdot 2\right)} \]
    4. Add Preprocessing
    5. Taylor expanded in eps around inf 94.6%

      \[\leadsto \varepsilon \cdot \color{blue}{\varepsilon} \]
  3. Recombined 2 regimes into one program.
  4. Final simplification93.0%

    \[\leadsto \begin{array}{l} \mathbf{if}\;x \leq -4.3 \cdot 10^{-116} \lor \neg \left(x \leq 1.56 \cdot 10^{-74}\right):\\ \;\;\;\;x \cdot \left(\varepsilon \cdot 2\right)\\ \mathbf{else}:\\ \;\;\;\;\varepsilon \cdot \varepsilon\\ \end{array} \]
  5. Add Preprocessing

Alternative 3: 90.1% accurate, 13.8× speedup?

\[\begin{array}{l} \\ \begin{array}{l} \mathbf{if}\;x \leq -6.6 \cdot 10^{-115} \lor \neg \left(x \leq 2.35 \cdot 10^{-74}\right):\\ \;\;\;\;2 \cdot \left(\varepsilon \cdot x\right)\\ \mathbf{else}:\\ \;\;\;\;\varepsilon \cdot \varepsilon\\ \end{array} \end{array} \]
(FPCore (x eps)
 :precision binary64
 (if (or (<= x -6.6e-115) (not (<= x 2.35e-74)))
   (* 2.0 (* eps x))
   (* eps eps)))
double code(double x, double eps) {
	double tmp;
	if ((x <= -6.6e-115) || !(x <= 2.35e-74)) {
		tmp = 2.0 * (eps * x);
	} else {
		tmp = eps * eps;
	}
	return tmp;
}
real(8) function code(x, eps)
    real(8), intent (in) :: x
    real(8), intent (in) :: eps
    real(8) :: tmp
    if ((x <= (-6.6d-115)) .or. (.not. (x <= 2.35d-74))) then
        tmp = 2.0d0 * (eps * x)
    else
        tmp = eps * eps
    end if
    code = tmp
end function
public static double code(double x, double eps) {
	double tmp;
	if ((x <= -6.6e-115) || !(x <= 2.35e-74)) {
		tmp = 2.0 * (eps * x);
	} else {
		tmp = eps * eps;
	}
	return tmp;
}
def code(x, eps):
	tmp = 0
	if (x <= -6.6e-115) or not (x <= 2.35e-74):
		tmp = 2.0 * (eps * x)
	else:
		tmp = eps * eps
	return tmp
function code(x, eps)
	tmp = 0.0
	if ((x <= -6.6e-115) || !(x <= 2.35e-74))
		tmp = Float64(2.0 * Float64(eps * x));
	else
		tmp = Float64(eps * eps);
	end
	return tmp
end
function tmp_2 = code(x, eps)
	tmp = 0.0;
	if ((x <= -6.6e-115) || ~((x <= 2.35e-74)))
		tmp = 2.0 * (eps * x);
	else
		tmp = eps * eps;
	end
	tmp_2 = tmp;
end
code[x_, eps_] := If[Or[LessEqual[x, -6.6e-115], N[Not[LessEqual[x, 2.35e-74]], $MachinePrecision]], N[(2.0 * N[(eps * x), $MachinePrecision]), $MachinePrecision], N[(eps * eps), $MachinePrecision]]
\begin{array}{l}

\\
\begin{array}{l}
\mathbf{if}\;x \leq -6.6 \cdot 10^{-115} \lor \neg \left(x \leq 2.35 \cdot 10^{-74}\right):\\
\;\;\;\;2 \cdot \left(\varepsilon \cdot x\right)\\

\mathbf{else}:\\
\;\;\;\;\varepsilon \cdot \varepsilon\\


\end{array}
\end{array}
Derivation
  1. Split input into 2 regimes
  2. if x < -6.59999999999999979e-115 or 2.35000000000000005e-74 < x

    1. Initial program 23.9%

      \[{\left(x + \varepsilon\right)}^{2} - {x}^{2} \]
    2. Step-by-step derivation
      1. +-commutative23.9%

        \[\leadsto {\color{blue}{\left(\varepsilon + x\right)}}^{2} - {x}^{2} \]
      2. unpow223.9%

        \[\leadsto \color{blue}{\left(\varepsilon + x\right) \cdot \left(\varepsilon + x\right)} - {x}^{2} \]
      3. unpow223.9%

        \[\leadsto \left(\varepsilon + x\right) \cdot \left(\varepsilon + x\right) - \color{blue}{x \cdot x} \]
      4. difference-of-squares23.9%

        \[\leadsto \color{blue}{\left(\left(\varepsilon + x\right) + x\right) \cdot \left(\left(\varepsilon + x\right) - x\right)} \]
      5. sub-neg23.9%

        \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \color{blue}{\left(\left(\varepsilon + x\right) + \left(-x\right)\right)} \]
      6. distribute-lft-in23.8%

        \[\leadsto \color{blue}{\left(\left(\varepsilon + x\right) + x\right) \cdot \left(\varepsilon + x\right) + \left(\left(\varepsilon + x\right) + x\right) \cdot \left(-x\right)} \]
      7. +-commutative23.8%

        \[\leadsto \color{blue}{\left(\left(\varepsilon + x\right) + x\right) \cdot \left(-x\right) + \left(\left(\varepsilon + x\right) + x\right) \cdot \left(\varepsilon + x\right)} \]
      8. distribute-lft-in23.9%

        \[\leadsto \color{blue}{\left(\left(\varepsilon + x\right) + x\right) \cdot \left(\left(-x\right) + \left(\varepsilon + x\right)\right)} \]
      9. associate-+l+23.9%

        \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \color{blue}{\left(\left(\left(-x\right) + \varepsilon\right) + x\right)} \]
      10. remove-double-neg23.9%

        \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \left(\left(\left(-x\right) + \varepsilon\right) + \color{blue}{\left(-\left(-x\right)\right)}\right) \]
      11. sub-neg23.9%

        \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \color{blue}{\left(\left(\left(-x\right) + \varepsilon\right) - \left(-x\right)\right)} \]
      12. +-commutative23.9%

        \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \left(\color{blue}{\left(\varepsilon + \left(-x\right)\right)} - \left(-x\right)\right) \]
      13. associate--l+99.9%

        \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \color{blue}{\left(\varepsilon + \left(\left(-x\right) - \left(-x\right)\right)\right)} \]
      14. +-inverses99.9%

        \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \left(\varepsilon + \color{blue}{0}\right) \]
      15. +-rgt-identity99.9%

        \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \color{blue}{\varepsilon} \]
      16. *-commutative99.9%

        \[\leadsto \color{blue}{\varepsilon \cdot \left(\left(\varepsilon + x\right) + x\right)} \]
      17. associate-+l+100.0%

        \[\leadsto \varepsilon \cdot \color{blue}{\left(\varepsilon + \left(x + x\right)\right)} \]
      18. count-2100.0%

        \[\leadsto \varepsilon \cdot \left(\varepsilon + \color{blue}{2 \cdot x}\right) \]
      19. *-commutative100.0%

        \[\leadsto \varepsilon \cdot \left(\varepsilon + \color{blue}{x \cdot 2}\right) \]
    3. Simplified100.0%

      \[\leadsto \color{blue}{\varepsilon \cdot \left(\varepsilon + x \cdot 2\right)} \]
    4. Add Preprocessing
    5. Taylor expanded in eps around 0 89.1%

      \[\leadsto \color{blue}{2 \cdot \left(\varepsilon \cdot x\right)} \]
    6. Step-by-step derivation
      1. *-commutative89.1%

        \[\leadsto 2 \cdot \color{blue}{\left(x \cdot \varepsilon\right)} \]
    7. Simplified89.1%

      \[\leadsto \color{blue}{2 \cdot \left(x \cdot \varepsilon\right)} \]

    if -6.59999999999999979e-115 < x < 2.35000000000000005e-74

    1. Initial program 95.6%

      \[{\left(x + \varepsilon\right)}^{2} - {x}^{2} \]
    2. Step-by-step derivation
      1. +-commutative95.6%

        \[\leadsto {\color{blue}{\left(\varepsilon + x\right)}}^{2} - {x}^{2} \]
      2. unpow295.6%

        \[\leadsto \color{blue}{\left(\varepsilon + x\right) \cdot \left(\varepsilon + x\right)} - {x}^{2} \]
      3. unpow295.6%

        \[\leadsto \left(\varepsilon + x\right) \cdot \left(\varepsilon + x\right) - \color{blue}{x \cdot x} \]
      4. difference-of-squares95.7%

        \[\leadsto \color{blue}{\left(\left(\varepsilon + x\right) + x\right) \cdot \left(\left(\varepsilon + x\right) - x\right)} \]
      5. sub-neg95.7%

        \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \color{blue}{\left(\left(\varepsilon + x\right) + \left(-x\right)\right)} \]
      6. distribute-lft-in95.7%

        \[\leadsto \color{blue}{\left(\left(\varepsilon + x\right) + x\right) \cdot \left(\varepsilon + x\right) + \left(\left(\varepsilon + x\right) + x\right) \cdot \left(-x\right)} \]
      7. +-commutative95.7%

        \[\leadsto \color{blue}{\left(\left(\varepsilon + x\right) + x\right) \cdot \left(-x\right) + \left(\left(\varepsilon + x\right) + x\right) \cdot \left(\varepsilon + x\right)} \]
      8. distribute-lft-in95.7%

        \[\leadsto \color{blue}{\left(\left(\varepsilon + x\right) + x\right) \cdot \left(\left(-x\right) + \left(\varepsilon + x\right)\right)} \]
      9. associate-+l+95.7%

        \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \color{blue}{\left(\left(\left(-x\right) + \varepsilon\right) + x\right)} \]
      10. remove-double-neg95.7%

        \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \left(\left(\left(-x\right) + \varepsilon\right) + \color{blue}{\left(-\left(-x\right)\right)}\right) \]
      11. sub-neg95.7%

        \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \color{blue}{\left(\left(\left(-x\right) + \varepsilon\right) - \left(-x\right)\right)} \]
      12. +-commutative95.7%

        \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \left(\color{blue}{\left(\varepsilon + \left(-x\right)\right)} - \left(-x\right)\right) \]
      13. associate--l+100.0%

        \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \color{blue}{\left(\varepsilon + \left(\left(-x\right) - \left(-x\right)\right)\right)} \]
      14. +-inverses100.0%

        \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \left(\varepsilon + \color{blue}{0}\right) \]
      15. +-rgt-identity100.0%

        \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \color{blue}{\varepsilon} \]
      16. *-commutative100.0%

        \[\leadsto \color{blue}{\varepsilon \cdot \left(\left(\varepsilon + x\right) + x\right)} \]
      17. associate-+l+100.0%

        \[\leadsto \varepsilon \cdot \color{blue}{\left(\varepsilon + \left(x + x\right)\right)} \]
      18. count-2100.0%

        \[\leadsto \varepsilon \cdot \left(\varepsilon + \color{blue}{2 \cdot x}\right) \]
      19. *-commutative100.0%

        \[\leadsto \varepsilon \cdot \left(\varepsilon + \color{blue}{x \cdot 2}\right) \]
    3. Simplified100.0%

      \[\leadsto \color{blue}{\varepsilon \cdot \left(\varepsilon + x \cdot 2\right)} \]
    4. Add Preprocessing
    5. Taylor expanded in eps around inf 94.6%

      \[\leadsto \varepsilon \cdot \color{blue}{\varepsilon} \]
  3. Recombined 2 regimes into one program.
  4. Final simplification93.0%

    \[\leadsto \begin{array}{l} \mathbf{if}\;x \leq -6.6 \cdot 10^{-115} \lor \neg \left(x \leq 2.35 \cdot 10^{-74}\right):\\ \;\;\;\;2 \cdot \left(\varepsilon \cdot x\right)\\ \mathbf{else}:\\ \;\;\;\;\varepsilon \cdot \varepsilon\\ \end{array} \]
  5. Add Preprocessing

Alternative 4: 100.0% accurate, 29.6× speedup?

\[\begin{array}{l} \\ \varepsilon \cdot \left(\varepsilon + x \cdot 2\right) \end{array} \]
(FPCore (x eps) :precision binary64 (* eps (+ eps (* x 2.0))))
double code(double x, double eps) {
	return eps * (eps + (x * 2.0));
}
real(8) function code(x, eps)
    real(8), intent (in) :: x
    real(8), intent (in) :: eps
    code = eps * (eps + (x * 2.0d0))
end function
public static double code(double x, double eps) {
	return eps * (eps + (x * 2.0));
}
def code(x, eps):
	return eps * (eps + (x * 2.0))
function code(x, eps)
	return Float64(eps * Float64(eps + Float64(x * 2.0)))
end
function tmp = code(x, eps)
	tmp = eps * (eps + (x * 2.0));
end
code[x_, eps_] := N[(eps * N[(eps + N[(x * 2.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}

\\
\varepsilon \cdot \left(\varepsilon + x \cdot 2\right)
\end{array}
Derivation
  1. Initial program 74.4%

    \[{\left(x + \varepsilon\right)}^{2} - {x}^{2} \]
  2. Step-by-step derivation
    1. +-commutative74.4%

      \[\leadsto {\color{blue}{\left(\varepsilon + x\right)}}^{2} - {x}^{2} \]
    2. unpow274.4%

      \[\leadsto \color{blue}{\left(\varepsilon + x\right) \cdot \left(\varepsilon + x\right)} - {x}^{2} \]
    3. unpow274.4%

      \[\leadsto \left(\varepsilon + x\right) \cdot \left(\varepsilon + x\right) - \color{blue}{x \cdot x} \]
    4. difference-of-squares74.4%

      \[\leadsto \color{blue}{\left(\left(\varepsilon + x\right) + x\right) \cdot \left(\left(\varepsilon + x\right) - x\right)} \]
    5. sub-neg74.4%

      \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \color{blue}{\left(\left(\varepsilon + x\right) + \left(-x\right)\right)} \]
    6. distribute-lft-in74.3%

      \[\leadsto \color{blue}{\left(\left(\varepsilon + x\right) + x\right) \cdot \left(\varepsilon + x\right) + \left(\left(\varepsilon + x\right) + x\right) \cdot \left(-x\right)} \]
    7. +-commutative74.3%

      \[\leadsto \color{blue}{\left(\left(\varepsilon + x\right) + x\right) \cdot \left(-x\right) + \left(\left(\varepsilon + x\right) + x\right) \cdot \left(\varepsilon + x\right)} \]
    8. distribute-lft-in74.4%

      \[\leadsto \color{blue}{\left(\left(\varepsilon + x\right) + x\right) \cdot \left(\left(-x\right) + \left(\varepsilon + x\right)\right)} \]
    9. associate-+l+74.4%

      \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \color{blue}{\left(\left(\left(-x\right) + \varepsilon\right) + x\right)} \]
    10. remove-double-neg74.4%

      \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \left(\left(\left(-x\right) + \varepsilon\right) + \color{blue}{\left(-\left(-x\right)\right)}\right) \]
    11. sub-neg74.4%

      \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \color{blue}{\left(\left(\left(-x\right) + \varepsilon\right) - \left(-x\right)\right)} \]
    12. +-commutative74.4%

      \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \left(\color{blue}{\left(\varepsilon + \left(-x\right)\right)} - \left(-x\right)\right) \]
    13. associate--l+99.9%

      \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \color{blue}{\left(\varepsilon + \left(\left(-x\right) - \left(-x\right)\right)\right)} \]
    14. +-inverses99.9%

      \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \left(\varepsilon + \color{blue}{0}\right) \]
    15. +-rgt-identity99.9%

      \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \color{blue}{\varepsilon} \]
    16. *-commutative99.9%

      \[\leadsto \color{blue}{\varepsilon \cdot \left(\left(\varepsilon + x\right) + x\right)} \]
    17. associate-+l+100.0%

      \[\leadsto \varepsilon \cdot \color{blue}{\left(\varepsilon + \left(x + x\right)\right)} \]
    18. count-2100.0%

      \[\leadsto \varepsilon \cdot \left(\varepsilon + \color{blue}{2 \cdot x}\right) \]
    19. *-commutative100.0%

      \[\leadsto \varepsilon \cdot \left(\varepsilon + \color{blue}{x \cdot 2}\right) \]
  3. Simplified100.0%

    \[\leadsto \color{blue}{\varepsilon \cdot \left(\varepsilon + x \cdot 2\right)} \]
  4. Add Preprocessing
  5. Add Preprocessing

Alternative 5: 73.5% accurate, 69.0× speedup?

\[\begin{array}{l} \\ \varepsilon \cdot \varepsilon \end{array} \]
(FPCore (x eps) :precision binary64 (* eps eps))
double code(double x, double eps) {
	return eps * eps;
}
real(8) function code(x, eps)
    real(8), intent (in) :: x
    real(8), intent (in) :: eps
    code = eps * eps
end function
public static double code(double x, double eps) {
	return eps * eps;
}
def code(x, eps):
	return eps * eps
function code(x, eps)
	return Float64(eps * eps)
end
function tmp = code(x, eps)
	tmp = eps * eps;
end
code[x_, eps_] := N[(eps * eps), $MachinePrecision]
\begin{array}{l}

\\
\varepsilon \cdot \varepsilon
\end{array}
Derivation
  1. Initial program 74.4%

    \[{\left(x + \varepsilon\right)}^{2} - {x}^{2} \]
  2. Step-by-step derivation
    1. +-commutative74.4%

      \[\leadsto {\color{blue}{\left(\varepsilon + x\right)}}^{2} - {x}^{2} \]
    2. unpow274.4%

      \[\leadsto \color{blue}{\left(\varepsilon + x\right) \cdot \left(\varepsilon + x\right)} - {x}^{2} \]
    3. unpow274.4%

      \[\leadsto \left(\varepsilon + x\right) \cdot \left(\varepsilon + x\right) - \color{blue}{x \cdot x} \]
    4. difference-of-squares74.4%

      \[\leadsto \color{blue}{\left(\left(\varepsilon + x\right) + x\right) \cdot \left(\left(\varepsilon + x\right) - x\right)} \]
    5. sub-neg74.4%

      \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \color{blue}{\left(\left(\varepsilon + x\right) + \left(-x\right)\right)} \]
    6. distribute-lft-in74.3%

      \[\leadsto \color{blue}{\left(\left(\varepsilon + x\right) + x\right) \cdot \left(\varepsilon + x\right) + \left(\left(\varepsilon + x\right) + x\right) \cdot \left(-x\right)} \]
    7. +-commutative74.3%

      \[\leadsto \color{blue}{\left(\left(\varepsilon + x\right) + x\right) \cdot \left(-x\right) + \left(\left(\varepsilon + x\right) + x\right) \cdot \left(\varepsilon + x\right)} \]
    8. distribute-lft-in74.4%

      \[\leadsto \color{blue}{\left(\left(\varepsilon + x\right) + x\right) \cdot \left(\left(-x\right) + \left(\varepsilon + x\right)\right)} \]
    9. associate-+l+74.4%

      \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \color{blue}{\left(\left(\left(-x\right) + \varepsilon\right) + x\right)} \]
    10. remove-double-neg74.4%

      \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \left(\left(\left(-x\right) + \varepsilon\right) + \color{blue}{\left(-\left(-x\right)\right)}\right) \]
    11. sub-neg74.4%

      \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \color{blue}{\left(\left(\left(-x\right) + \varepsilon\right) - \left(-x\right)\right)} \]
    12. +-commutative74.4%

      \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \left(\color{blue}{\left(\varepsilon + \left(-x\right)\right)} - \left(-x\right)\right) \]
    13. associate--l+99.9%

      \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \color{blue}{\left(\varepsilon + \left(\left(-x\right) - \left(-x\right)\right)\right)} \]
    14. +-inverses99.9%

      \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \left(\varepsilon + \color{blue}{0}\right) \]
    15. +-rgt-identity99.9%

      \[\leadsto \left(\left(\varepsilon + x\right) + x\right) \cdot \color{blue}{\varepsilon} \]
    16. *-commutative99.9%

      \[\leadsto \color{blue}{\varepsilon \cdot \left(\left(\varepsilon + x\right) + x\right)} \]
    17. associate-+l+100.0%

      \[\leadsto \varepsilon \cdot \color{blue}{\left(\varepsilon + \left(x + x\right)\right)} \]
    18. count-2100.0%

      \[\leadsto \varepsilon \cdot \left(\varepsilon + \color{blue}{2 \cdot x}\right) \]
    19. *-commutative100.0%

      \[\leadsto \varepsilon \cdot \left(\varepsilon + \color{blue}{x \cdot 2}\right) \]
  3. Simplified100.0%

    \[\leadsto \color{blue}{\varepsilon \cdot \left(\varepsilon + x \cdot 2\right)} \]
  4. Add Preprocessing
  5. Taylor expanded in eps around inf 72.4%

    \[\leadsto \varepsilon \cdot \color{blue}{\varepsilon} \]
  6. Add Preprocessing

Reproduce

?
herbie shell --seed 2024158 
(FPCore (x eps)
  :name "ENA, Section 1.4, Exercise 4b, n=2"
  :precision binary64
  :pre (and (and (<= -1000000000.0 x) (<= x 1000000000.0)) (and (<= -1.0 eps) (<= eps 1.0)))
  (- (pow (+ x eps) 2.0) (pow x 2.0)))