2tan (problem 3.3.2)

Percentage Accurate: 62.2% → 99.6%
Time: 16.0s
Alternatives: 11
Speedup: 17.3×

Specification

?
\[\left(\left(-10000 \leq x \land x \leq 10000\right) \land 10^{-16} \cdot \left|x\right| < \varepsilon\right) \land \varepsilon < \left|x\right|\]
\[\begin{array}{l} \\ \tan \left(x + \varepsilon\right) - \tan x \end{array} \]
(FPCore (x eps) :precision binary64 (- (tan (+ x eps)) (tan x)))
double code(double x, double eps) {
	return tan((x + eps)) - tan(x);
}
real(8) function code(x, eps)
    real(8), intent (in) :: x
    real(8), intent (in) :: eps
    code = tan((x + eps)) - tan(x)
end function
public static double code(double x, double eps) {
	return Math.tan((x + eps)) - Math.tan(x);
}
def code(x, eps):
	return math.tan((x + eps)) - math.tan(x)
function code(x, eps)
	return Float64(tan(Float64(x + eps)) - tan(x))
end
function tmp = code(x, eps)
	tmp = tan((x + eps)) - tan(x);
end
code[x_, eps_] := N[(N[Tan[N[(x + eps), $MachinePrecision]], $MachinePrecision] - N[Tan[x], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}

\\
\tan \left(x + \varepsilon\right) - \tan x
\end{array}

Sampling outcomes in binary64 precision:

Local Percentage Accuracy vs ?

The average percentage accuracy by input value. Horizontal axis shows value of an input variable; the variable is choosen in the title. Vertical axis is accuracy; higher is better. Red represent the original program, while blue represents Herbie's suggestion. These can be toggled with buttons below the plot. The line is an average while dots represent individual samples.

Accuracy vs Speed?

Herbie found 11 alternatives:

AlternativeAccuracySpeedup
The accuracy (vertical axis) and speed (horizontal axis) of each alternatives. Up and to the right is better. The red square shows the initial program, and each blue circle shows an alternative.The line shows the best available speed-accuracy tradeoffs.

Initial Program: 62.2% accurate, 1.0× speedup?

\[\begin{array}{l} \\ \tan \left(x + \varepsilon\right) - \tan x \end{array} \]
(FPCore (x eps) :precision binary64 (- (tan (+ x eps)) (tan x)))
double code(double x, double eps) {
	return tan((x + eps)) - tan(x);
}
real(8) function code(x, eps)
    real(8), intent (in) :: x
    real(8), intent (in) :: eps
    code = tan((x + eps)) - tan(x)
end function
public static double code(double x, double eps) {
	return Math.tan((x + eps)) - Math.tan(x);
}
def code(x, eps):
	return math.tan((x + eps)) - math.tan(x)
function code(x, eps)
	return Float64(tan(Float64(x + eps)) - tan(x))
end
function tmp = code(x, eps)
	tmp = tan((x + eps)) - tan(x);
end
code[x_, eps_] := N[(N[Tan[N[(x + eps), $MachinePrecision]], $MachinePrecision] - N[Tan[x], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}

\\
\tan \left(x + \varepsilon\right) - \tan x
\end{array}

Alternative 1: 99.6% accurate, 0.1× speedup?

\[\begin{array}{l} \\ \begin{array}{l} t_0 := {\tan x}^{2}\\ t_1 := 0.5 \cdot \cos \left(x + x\right)\\ \mathsf{fma}\left(\mathsf{fma}\left(\varepsilon, \mathsf{fma}\left(\varepsilon, \frac{\left(0.5 - t\_1\right) + {\left(\sin x \cdot \tan x\right)}^{2}}{0.5 + t\_1} - \mathsf{fma}\left(t\_0, -0.3333333333333333, -0.3333333333333333\right), \frac{\mathsf{fma}\left(\sin x, t\_0, \sin x\right)}{\cos x}\right), t\_0\right), \varepsilon, \varepsilon\right) \end{array} \end{array} \]
(FPCore (x eps)
 :precision binary64
 (let* ((t_0 (pow (tan x) 2.0)) (t_1 (* 0.5 (cos (+ x x)))))
   (fma
    (fma
     eps
     (fma
      eps
      (-
       (/ (+ (- 0.5 t_1) (pow (* (sin x) (tan x)) 2.0)) (+ 0.5 t_1))
       (fma t_0 -0.3333333333333333 -0.3333333333333333))
      (/ (fma (sin x) t_0 (sin x)) (cos x)))
     t_0)
    eps
    eps)))
double code(double x, double eps) {
	double t_0 = pow(tan(x), 2.0);
	double t_1 = 0.5 * cos((x + x));
	return fma(fma(eps, fma(eps, ((((0.5 - t_1) + pow((sin(x) * tan(x)), 2.0)) / (0.5 + t_1)) - fma(t_0, -0.3333333333333333, -0.3333333333333333)), (fma(sin(x), t_0, sin(x)) / cos(x))), t_0), eps, eps);
}
function code(x, eps)
	t_0 = tan(x) ^ 2.0
	t_1 = Float64(0.5 * cos(Float64(x + x)))
	return fma(fma(eps, fma(eps, Float64(Float64(Float64(Float64(0.5 - t_1) + (Float64(sin(x) * tan(x)) ^ 2.0)) / Float64(0.5 + t_1)) - fma(t_0, -0.3333333333333333, -0.3333333333333333)), Float64(fma(sin(x), t_0, sin(x)) / cos(x))), t_0), eps, eps)
end
code[x_, eps_] := Block[{t$95$0 = N[Power[N[Tan[x], $MachinePrecision], 2.0], $MachinePrecision]}, Block[{t$95$1 = N[(0.5 * N[Cos[N[(x + x), $MachinePrecision]], $MachinePrecision]), $MachinePrecision]}, N[(N[(eps * N[(eps * N[(N[(N[(N[(0.5 - t$95$1), $MachinePrecision] + N[Power[N[(N[Sin[x], $MachinePrecision] * N[Tan[x], $MachinePrecision]), $MachinePrecision], 2.0], $MachinePrecision]), $MachinePrecision] / N[(0.5 + t$95$1), $MachinePrecision]), $MachinePrecision] - N[(t$95$0 * -0.3333333333333333 + -0.3333333333333333), $MachinePrecision]), $MachinePrecision] + N[(N[(N[Sin[x], $MachinePrecision] * t$95$0 + N[Sin[x], $MachinePrecision]), $MachinePrecision] / N[Cos[x], $MachinePrecision]), $MachinePrecision]), $MachinePrecision] + t$95$0), $MachinePrecision] * eps + eps), $MachinePrecision]]]
\begin{array}{l}

\\
\begin{array}{l}
t_0 := {\tan x}^{2}\\
t_1 := 0.5 \cdot \cos \left(x + x\right)\\
\mathsf{fma}\left(\mathsf{fma}\left(\varepsilon, \mathsf{fma}\left(\varepsilon, \frac{\left(0.5 - t\_1\right) + {\left(\sin x \cdot \tan x\right)}^{2}}{0.5 + t\_1} - \mathsf{fma}\left(t\_0, -0.3333333333333333, -0.3333333333333333\right), \frac{\mathsf{fma}\left(\sin x, t\_0, \sin x\right)}{\cos x}\right), t\_0\right), \varepsilon, \varepsilon\right)
\end{array}
\end{array}
Derivation
  1. Initial program 59.9%

    \[\tan \left(x + \varepsilon\right) - \tan x \]
  2. Add Preprocessing
  3. Taylor expanded in eps around 0

    \[\leadsto \color{blue}{\varepsilon \cdot \left(\left(1 + \varepsilon \cdot \left(-1 \cdot \left(\varepsilon \cdot \left(\frac{1}{6} + \left(-1 \cdot \frac{{\sin x}^{2} \cdot \left(1 - -1 \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right)}{{\cos x}^{2}} + \left(\frac{-1}{2} \cdot \left(1 - -1 \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right) + \frac{1}{6} \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right)\right)\right)\right) - -1 \cdot \frac{\sin x \cdot \left(1 - -1 \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right)}{\cos x}\right)\right) - -1 \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right)} \]
  4. Applied rewrites98.6%

    \[\leadsto \color{blue}{\mathsf{fma}\left(\varepsilon, \mathsf{fma}\left(\varepsilon, \mathsf{fma}\left(\varepsilon, \left(\frac{{\sin x}^{2} + \frac{{\sin x}^{4}}{{\cos x}^{2}}}{{\cos x}^{2}} - \mathsf{fma}\left(0.16666666666666666, \frac{{\sin x}^{2}}{{\cos x}^{2}}, \mathsf{fma}\left(-0.5, \frac{{\sin x}^{2}}{{\cos x}^{2}}, -0.5\right)\right)\right) + -0.16666666666666666, \frac{\sin x + \frac{{\sin x}^{3}}{{\cos x}^{2}}}{\cos x}\right), \frac{{\sin x}^{2}}{{\cos x}^{2}}\right), \varepsilon\right)} \]
  5. Applied rewrites98.6%

    \[\leadsto \mathsf{fma}\left(\mathsf{fma}\left(\varepsilon, \mathsf{fma}\left(\varepsilon, \frac{\left(0.5 - 0.5 \cdot \cos \left(x + x\right)\right) + {\left(\sin x \cdot \tan x\right)}^{2}}{0.5 + 0.5 \cdot \cos \left(x + x\right)} - \mathsf{fma}\left({\tan x}^{2}, -0.3333333333333333, -0.3333333333333333\right), \frac{\mathsf{fma}\left(\sin x, {\tan x}^{2}, \sin x\right)}{\cos x}\right), {\tan x}^{2}\right), \color{blue}{\varepsilon}, \varepsilon\right) \]
  6. Add Preprocessing

Alternative 2: 99.5% accurate, 0.2× speedup?

\[\begin{array}{l} \\ \begin{array}{l} t_0 := {\cos x}^{2}\\ \mathsf{fma}\left(\varepsilon, \mathsf{fma}\left(\varepsilon, \mathsf{fma}\left(\varepsilon, 0.3333333333333333, \frac{\sin x + \frac{{\sin x}^{3}}{t\_0}}{\cos x}\right), \frac{{\sin x}^{2}}{t\_0}\right), \varepsilon\right) \end{array} \end{array} \]
(FPCore (x eps)
 :precision binary64
 (let* ((t_0 (pow (cos x) 2.0)))
   (fma
    eps
    (fma
     eps
     (fma
      eps
      0.3333333333333333
      (/ (+ (sin x) (/ (pow (sin x) 3.0) t_0)) (cos x)))
     (/ (pow (sin x) 2.0) t_0))
    eps)))
double code(double x, double eps) {
	double t_0 = pow(cos(x), 2.0);
	return fma(eps, fma(eps, fma(eps, 0.3333333333333333, ((sin(x) + (pow(sin(x), 3.0) / t_0)) / cos(x))), (pow(sin(x), 2.0) / t_0)), eps);
}
function code(x, eps)
	t_0 = cos(x) ^ 2.0
	return fma(eps, fma(eps, fma(eps, 0.3333333333333333, Float64(Float64(sin(x) + Float64((sin(x) ^ 3.0) / t_0)) / cos(x))), Float64((sin(x) ^ 2.0) / t_0)), eps)
end
code[x_, eps_] := Block[{t$95$0 = N[Power[N[Cos[x], $MachinePrecision], 2.0], $MachinePrecision]}, N[(eps * N[(eps * N[(eps * 0.3333333333333333 + N[(N[(N[Sin[x], $MachinePrecision] + N[(N[Power[N[Sin[x], $MachinePrecision], 3.0], $MachinePrecision] / t$95$0), $MachinePrecision]), $MachinePrecision] / N[Cos[x], $MachinePrecision]), $MachinePrecision]), $MachinePrecision] + N[(N[Power[N[Sin[x], $MachinePrecision], 2.0], $MachinePrecision] / t$95$0), $MachinePrecision]), $MachinePrecision] + eps), $MachinePrecision]]
\begin{array}{l}

\\
\begin{array}{l}
t_0 := {\cos x}^{2}\\
\mathsf{fma}\left(\varepsilon, \mathsf{fma}\left(\varepsilon, \mathsf{fma}\left(\varepsilon, 0.3333333333333333, \frac{\sin x + \frac{{\sin x}^{3}}{t\_0}}{\cos x}\right), \frac{{\sin x}^{2}}{t\_0}\right), \varepsilon\right)
\end{array}
\end{array}
Derivation
  1. Initial program 59.9%

    \[\tan \left(x + \varepsilon\right) - \tan x \]
  2. Add Preprocessing
  3. Taylor expanded in eps around 0

    \[\leadsto \color{blue}{\varepsilon \cdot \left(\left(1 + \varepsilon \cdot \left(-1 \cdot \left(\varepsilon \cdot \left(\frac{1}{6} + \left(-1 \cdot \frac{{\sin x}^{2} \cdot \left(1 - -1 \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right)}{{\cos x}^{2}} + \left(\frac{-1}{2} \cdot \left(1 - -1 \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right) + \frac{1}{6} \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right)\right)\right)\right) - -1 \cdot \frac{\sin x \cdot \left(1 - -1 \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right)}{\cos x}\right)\right) - -1 \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right)} \]
  4. Applied rewrites98.6%

    \[\leadsto \color{blue}{\mathsf{fma}\left(\varepsilon, \mathsf{fma}\left(\varepsilon, \mathsf{fma}\left(\varepsilon, \left(\frac{{\sin x}^{2} + \frac{{\sin x}^{4}}{{\cos x}^{2}}}{{\cos x}^{2}} - \mathsf{fma}\left(0.16666666666666666, \frac{{\sin x}^{2}}{{\cos x}^{2}}, \mathsf{fma}\left(-0.5, \frac{{\sin x}^{2}}{{\cos x}^{2}}, -0.5\right)\right)\right) + -0.16666666666666666, \frac{\sin x + \frac{{\sin x}^{3}}{{\cos x}^{2}}}{\cos x}\right), \frac{{\sin x}^{2}}{{\cos x}^{2}}\right), \varepsilon\right)} \]
  5. Taylor expanded in x around 0

    \[\leadsto \mathsf{fma}\left(\varepsilon, \mathsf{fma}\left(\varepsilon, \mathsf{fma}\left(\varepsilon, \frac{1}{3}, \frac{\sin x + \frac{{\sin x}^{3}}{{\cos x}^{2}}}{\cos x}\right), \frac{{\sin x}^{2}}{{\cos x}^{2}}\right), \varepsilon\right) \]
  6. Step-by-step derivation
    1. Applied rewrites98.6%

      \[\leadsto \mathsf{fma}\left(\varepsilon, \mathsf{fma}\left(\varepsilon, \mathsf{fma}\left(\varepsilon, 0.3333333333333333, \frac{\sin x + \frac{{\sin x}^{3}}{{\cos x}^{2}}}{\cos x}\right), \frac{{\sin x}^{2}}{{\cos x}^{2}}\right), \varepsilon\right) \]
    2. Add Preprocessing

    Alternative 3: 99.4% accurate, 0.4× speedup?

    \[\begin{array}{l} \\ \mathsf{fma}\left(\mathsf{fma}\left(\tan x + {\tan x}^{3}, \varepsilon, {\tan x}^{2}\right), \varepsilon, \varepsilon\right) \end{array} \]
    (FPCore (x eps)
     :precision binary64
     (fma (fma (+ (tan x) (pow (tan x) 3.0)) eps (pow (tan x) 2.0)) eps eps))
    double code(double x, double eps) {
    	return fma(fma((tan(x) + pow(tan(x), 3.0)), eps, pow(tan(x), 2.0)), eps, eps);
    }
    
    function code(x, eps)
    	return fma(fma(Float64(tan(x) + (tan(x) ^ 3.0)), eps, (tan(x) ^ 2.0)), eps, eps)
    end
    
    code[x_, eps_] := N[(N[(N[(N[Tan[x], $MachinePrecision] + N[Power[N[Tan[x], $MachinePrecision], 3.0], $MachinePrecision]), $MachinePrecision] * eps + N[Power[N[Tan[x], $MachinePrecision], 2.0], $MachinePrecision]), $MachinePrecision] * eps + eps), $MachinePrecision]
    
    \begin{array}{l}
    
    \\
    \mathsf{fma}\left(\mathsf{fma}\left(\tan x + {\tan x}^{3}, \varepsilon, {\tan x}^{2}\right), \varepsilon, \varepsilon\right)
    \end{array}
    
    Derivation
    1. Initial program 59.9%

      \[\tan \left(x + \varepsilon\right) - \tan x \]
    2. Add Preprocessing
    3. Taylor expanded in eps around 0

      \[\leadsto \color{blue}{\varepsilon \cdot \left(\left(1 + \varepsilon \cdot \left(-1 \cdot \left(\varepsilon \cdot \left(\frac{1}{6} + \left(-1 \cdot \frac{{\sin x}^{2} \cdot \left(1 - -1 \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right)}{{\cos x}^{2}} + \left(\frac{-1}{2} \cdot \left(1 - -1 \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right) + \frac{1}{6} \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right)\right)\right)\right) - -1 \cdot \frac{\sin x \cdot \left(1 - -1 \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right)}{\cos x}\right)\right) - -1 \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right)} \]
    4. Applied rewrites98.6%

      \[\leadsto \color{blue}{\mathsf{fma}\left(\varepsilon, \mathsf{fma}\left(\varepsilon, \mathsf{fma}\left(\varepsilon, \left(\frac{{\sin x}^{2} + \frac{{\sin x}^{4}}{{\cos x}^{2}}}{{\cos x}^{2}} - \mathsf{fma}\left(0.16666666666666666, \frac{{\sin x}^{2}}{{\cos x}^{2}}, \mathsf{fma}\left(-0.5, \frac{{\sin x}^{2}}{{\cos x}^{2}}, -0.5\right)\right)\right) + -0.16666666666666666, \frac{\sin x + \frac{{\sin x}^{3}}{{\cos x}^{2}}}{\cos x}\right), \frac{{\sin x}^{2}}{{\cos x}^{2}}\right), \varepsilon\right)} \]
    5. Applied rewrites98.6%

      \[\leadsto \mathsf{fma}\left(\mathsf{fma}\left(\varepsilon, \mathsf{fma}\left(\varepsilon, \frac{\left(0.5 - 0.5 \cdot \cos \left(x + x\right)\right) + {\left(\sin x \cdot \tan x\right)}^{2}}{0.5 + 0.5 \cdot \cos \left(x + x\right)} - \mathsf{fma}\left({\tan x}^{2}, -0.3333333333333333, -0.3333333333333333\right), \frac{\mathsf{fma}\left(\sin x, {\tan x}^{2}, \sin x\right)}{\cos x}\right), {\tan x}^{2}\right), \color{blue}{\varepsilon}, \varepsilon\right) \]
    6. Taylor expanded in eps around 0

      \[\leadsto \mathsf{fma}\left(\mathsf{fma}\left(\varepsilon, \frac{\sin x}{\cos x} + \frac{{\sin x}^{3}}{{\cos x}^{3}}, {\tan x}^{2}\right), \varepsilon, \varepsilon\right) \]
    7. Step-by-step derivation
      1. Applied rewrites98.6%

        \[\leadsto \mathsf{fma}\left(\mathsf{fma}\left(\varepsilon, \frac{\sin x}{\cos x} + \frac{{\sin x}^{3}}{{\cos x}^{3}}, {\tan x}^{2}\right), \varepsilon, \varepsilon\right) \]
      2. Step-by-step derivation
        1. Applied rewrites98.6%

          \[\leadsto \mathsf{fma}\left(\mathsf{fma}\left({\tan x}^{3} + \tan x, \varepsilon, {\tan x}^{2}\right), \varepsilon, \varepsilon\right) \]
        2. Final simplification98.6%

          \[\leadsto \mathsf{fma}\left(\mathsf{fma}\left(\tan x + {\tan x}^{3}, \varepsilon, {\tan x}^{2}\right), \varepsilon, \varepsilon\right) \]
        3. Add Preprocessing

        Alternative 4: 99.1% accurate, 0.9× speedup?

        \[\begin{array}{l} \\ \mathsf{fma}\left(\mathsf{fma}\left(\varepsilon, \mathsf{fma}\left(\varepsilon, 0.3333333333333333, x\right), {\tan x}^{2}\right), \varepsilon, \varepsilon\right) \end{array} \]
        (FPCore (x eps)
         :precision binary64
         (fma (fma eps (fma eps 0.3333333333333333 x) (pow (tan x) 2.0)) eps eps))
        double code(double x, double eps) {
        	return fma(fma(eps, fma(eps, 0.3333333333333333, x), pow(tan(x), 2.0)), eps, eps);
        }
        
        function code(x, eps)
        	return fma(fma(eps, fma(eps, 0.3333333333333333, x), (tan(x) ^ 2.0)), eps, eps)
        end
        
        code[x_, eps_] := N[(N[(eps * N[(eps * 0.3333333333333333 + x), $MachinePrecision] + N[Power[N[Tan[x], $MachinePrecision], 2.0], $MachinePrecision]), $MachinePrecision] * eps + eps), $MachinePrecision]
        
        \begin{array}{l}
        
        \\
        \mathsf{fma}\left(\mathsf{fma}\left(\varepsilon, \mathsf{fma}\left(\varepsilon, 0.3333333333333333, x\right), {\tan x}^{2}\right), \varepsilon, \varepsilon\right)
        \end{array}
        
        Derivation
        1. Initial program 59.9%

          \[\tan \left(x + \varepsilon\right) - \tan x \]
        2. Add Preprocessing
        3. Taylor expanded in eps around 0

          \[\leadsto \color{blue}{\varepsilon \cdot \left(\left(1 + \varepsilon \cdot \left(-1 \cdot \left(\varepsilon \cdot \left(\frac{1}{6} + \left(-1 \cdot \frac{{\sin x}^{2} \cdot \left(1 - -1 \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right)}{{\cos x}^{2}} + \left(\frac{-1}{2} \cdot \left(1 - -1 \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right) + \frac{1}{6} \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right)\right)\right)\right) - -1 \cdot \frac{\sin x \cdot \left(1 - -1 \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right)}{\cos x}\right)\right) - -1 \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right)} \]
        4. Applied rewrites98.6%

          \[\leadsto \color{blue}{\mathsf{fma}\left(\varepsilon, \mathsf{fma}\left(\varepsilon, \mathsf{fma}\left(\varepsilon, \left(\frac{{\sin x}^{2} + \frac{{\sin x}^{4}}{{\cos x}^{2}}}{{\cos x}^{2}} - \mathsf{fma}\left(0.16666666666666666, \frac{{\sin x}^{2}}{{\cos x}^{2}}, \mathsf{fma}\left(-0.5, \frac{{\sin x}^{2}}{{\cos x}^{2}}, -0.5\right)\right)\right) + -0.16666666666666666, \frac{\sin x + \frac{{\sin x}^{3}}{{\cos x}^{2}}}{\cos x}\right), \frac{{\sin x}^{2}}{{\cos x}^{2}}\right), \varepsilon\right)} \]
        5. Applied rewrites98.6%

          \[\leadsto \mathsf{fma}\left(\mathsf{fma}\left(\varepsilon, \mathsf{fma}\left(\varepsilon, \frac{\left(0.5 - 0.5 \cdot \cos \left(x + x\right)\right) + {\left(\sin x \cdot \tan x\right)}^{2}}{0.5 + 0.5 \cdot \cos \left(x + x\right)} - \mathsf{fma}\left({\tan x}^{2}, -0.3333333333333333, -0.3333333333333333\right), \frac{\mathsf{fma}\left(\sin x, {\tan x}^{2}, \sin x\right)}{\cos x}\right), {\tan x}^{2}\right), \color{blue}{\varepsilon}, \varepsilon\right) \]
        6. Taylor expanded in x around 0

          \[\leadsto \mathsf{fma}\left(\mathsf{fma}\left(\varepsilon, x + \frac{1}{3} \cdot \varepsilon, {\tan x}^{2}\right), \varepsilon, \varepsilon\right) \]
        7. Step-by-step derivation
          1. Applied rewrites98.4%

            \[\leadsto \mathsf{fma}\left(\mathsf{fma}\left(\varepsilon, \mathsf{fma}\left(\varepsilon, 0.3333333333333333, x\right), {\tan x}^{2}\right), \varepsilon, \varepsilon\right) \]
          2. Add Preprocessing

          Alternative 5: 99.1% accurate, 1.0× speedup?

          \[\begin{array}{l} \\ \varepsilon \cdot \left({\tan x}^{2} + 1\right) \end{array} \]
          (FPCore (x eps) :precision binary64 (* eps (+ (pow (tan x) 2.0) 1.0)))
          double code(double x, double eps) {
          	return eps * (pow(tan(x), 2.0) + 1.0);
          }
          
          real(8) function code(x, eps)
              real(8), intent (in) :: x
              real(8), intent (in) :: eps
              code = eps * ((tan(x) ** 2.0d0) + 1.0d0)
          end function
          
          public static double code(double x, double eps) {
          	return eps * (Math.pow(Math.tan(x), 2.0) + 1.0);
          }
          
          def code(x, eps):
          	return eps * (math.pow(math.tan(x), 2.0) + 1.0)
          
          function code(x, eps)
          	return Float64(eps * Float64((tan(x) ^ 2.0) + 1.0))
          end
          
          function tmp = code(x, eps)
          	tmp = eps * ((tan(x) ^ 2.0) + 1.0);
          end
          
          code[x_, eps_] := N[(eps * N[(N[Power[N[Tan[x], $MachinePrecision], 2.0], $MachinePrecision] + 1.0), $MachinePrecision]), $MachinePrecision]
          
          \begin{array}{l}
          
          \\
          \varepsilon \cdot \left({\tan x}^{2} + 1\right)
          \end{array}
          
          Derivation
          1. Initial program 59.9%

            \[\tan \left(x + \varepsilon\right) - \tan x \]
          2. Add Preprocessing
          3. Step-by-step derivation
            1. lift--.f64N/A

              \[\leadsto \color{blue}{\tan \left(x + \varepsilon\right) - \tan x} \]
            2. lift-tan.f64N/A

              \[\leadsto \color{blue}{\tan \left(x + \varepsilon\right)} - \tan x \]
            3. tan-quotN/A

              \[\leadsto \color{blue}{\frac{\sin \left(x + \varepsilon\right)}{\cos \left(x + \varepsilon\right)}} - \tan x \]
            4. clear-numN/A

              \[\leadsto \color{blue}{\frac{1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)}}} - \tan x \]
            5. lift-tan.f64N/A

              \[\leadsto \frac{1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)}} - \color{blue}{\tan x} \]
            6. tan-quotN/A

              \[\leadsto \frac{1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)}} - \color{blue}{\frac{\sin x}{\cos x}} \]
            7. clear-numN/A

              \[\leadsto \frac{1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)}} - \color{blue}{\frac{1}{\frac{\cos x}{\sin x}}} \]
            8. frac-subN/A

              \[\leadsto \color{blue}{\frac{1 \cdot \frac{\cos x}{\sin x} - \frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)} \cdot 1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)} \cdot \frac{\cos x}{\sin x}}} \]
            9. lower-/.f64N/A

              \[\leadsto \color{blue}{\frac{1 \cdot \frac{\cos x}{\sin x} - \frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)} \cdot 1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)} \cdot \frac{\cos x}{\sin x}}} \]
          4. Applied rewrites31.1%

            \[\leadsto \color{blue}{\frac{1 \cdot \frac{1}{\tan x} - \frac{1}{\tan \left(x + \varepsilon\right)} \cdot 1}{\frac{1}{\tan \left(x + \varepsilon\right)} \cdot \frac{1}{\tan x}}} \]
          5. Taylor expanded in eps around 0

            \[\leadsto \color{blue}{\varepsilon \cdot \left(1 - -1 \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right)} \]
          6. Step-by-step derivation
            1. cancel-sign-sub-invN/A

              \[\leadsto \varepsilon \cdot \color{blue}{\left(1 + \left(\mathsf{neg}\left(-1\right)\right) \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right)} \]
            2. metadata-evalN/A

              \[\leadsto \varepsilon \cdot \left(1 + \color{blue}{1} \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right) \]
            3. *-lft-identityN/A

              \[\leadsto \varepsilon \cdot \left(1 + \color{blue}{\frac{{\sin x}^{2}}{{\cos x}^{2}}}\right) \]
            4. +-commutativeN/A

              \[\leadsto \varepsilon \cdot \color{blue}{\left(\frac{{\sin x}^{2}}{{\cos x}^{2}} + 1\right)} \]
            5. distribute-rgt-inN/A

              \[\leadsto \color{blue}{\frac{{\sin x}^{2}}{{\cos x}^{2}} \cdot \varepsilon + 1 \cdot \varepsilon} \]
            6. *-lft-identityN/A

              \[\leadsto \frac{{\sin x}^{2}}{{\cos x}^{2}} \cdot \varepsilon + \color{blue}{\varepsilon} \]
            7. lower-fma.f64N/A

              \[\leadsto \color{blue}{\mathsf{fma}\left(\frac{{\sin x}^{2}}{{\cos x}^{2}}, \varepsilon, \varepsilon\right)} \]
            8. lower-/.f64N/A

              \[\leadsto \mathsf{fma}\left(\color{blue}{\frac{{\sin x}^{2}}{{\cos x}^{2}}}, \varepsilon, \varepsilon\right) \]
            9. lower-pow.f64N/A

              \[\leadsto \mathsf{fma}\left(\frac{\color{blue}{{\sin x}^{2}}}{{\cos x}^{2}}, \varepsilon, \varepsilon\right) \]
            10. lower-sin.f64N/A

              \[\leadsto \mathsf{fma}\left(\frac{{\color{blue}{\sin x}}^{2}}{{\cos x}^{2}}, \varepsilon, \varepsilon\right) \]
            11. lower-pow.f64N/A

              \[\leadsto \mathsf{fma}\left(\frac{{\sin x}^{2}}{\color{blue}{{\cos x}^{2}}}, \varepsilon, \varepsilon\right) \]
            12. lower-cos.f6498.0

              \[\leadsto \mathsf{fma}\left(\frac{{\sin x}^{2}}{{\color{blue}{\cos x}}^{2}}, \varepsilon, \varepsilon\right) \]
          7. Applied rewrites98.0%

            \[\leadsto \color{blue}{\mathsf{fma}\left(\frac{{\sin x}^{2}}{{\cos x}^{2}}, \varepsilon, \varepsilon\right)} \]
          8. Step-by-step derivation
            1. Applied rewrites98.0%

              \[\leadsto \left({\tan x}^{2} + 1\right) \cdot \color{blue}{\varepsilon} \]
            2. Final simplification98.0%

              \[\leadsto \varepsilon \cdot \left({\tan x}^{2} + 1\right) \]
            3. Add Preprocessing

            Alternative 6: 99.1% accurate, 1.0× speedup?

            \[\begin{array}{l} \\ \mathsf{fma}\left({\tan x}^{2}, \varepsilon, \varepsilon\right) \end{array} \]
            (FPCore (x eps) :precision binary64 (fma (pow (tan x) 2.0) eps eps))
            double code(double x, double eps) {
            	return fma(pow(tan(x), 2.0), eps, eps);
            }
            
            function code(x, eps)
            	return fma((tan(x) ^ 2.0), eps, eps)
            end
            
            code[x_, eps_] := N[(N[Power[N[Tan[x], $MachinePrecision], 2.0], $MachinePrecision] * eps + eps), $MachinePrecision]
            
            \begin{array}{l}
            
            \\
            \mathsf{fma}\left({\tan x}^{2}, \varepsilon, \varepsilon\right)
            \end{array}
            
            Derivation
            1. Initial program 59.9%

              \[\tan \left(x + \varepsilon\right) - \tan x \]
            2. Add Preprocessing
            3. Step-by-step derivation
              1. lift--.f64N/A

                \[\leadsto \color{blue}{\tan \left(x + \varepsilon\right) - \tan x} \]
              2. lift-tan.f64N/A

                \[\leadsto \color{blue}{\tan \left(x + \varepsilon\right)} - \tan x \]
              3. tan-quotN/A

                \[\leadsto \color{blue}{\frac{\sin \left(x + \varepsilon\right)}{\cos \left(x + \varepsilon\right)}} - \tan x \]
              4. clear-numN/A

                \[\leadsto \color{blue}{\frac{1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)}}} - \tan x \]
              5. lift-tan.f64N/A

                \[\leadsto \frac{1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)}} - \color{blue}{\tan x} \]
              6. tan-quotN/A

                \[\leadsto \frac{1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)}} - \color{blue}{\frac{\sin x}{\cos x}} \]
              7. clear-numN/A

                \[\leadsto \frac{1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)}} - \color{blue}{\frac{1}{\frac{\cos x}{\sin x}}} \]
              8. frac-subN/A

                \[\leadsto \color{blue}{\frac{1 \cdot \frac{\cos x}{\sin x} - \frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)} \cdot 1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)} \cdot \frac{\cos x}{\sin x}}} \]
              9. lower-/.f64N/A

                \[\leadsto \color{blue}{\frac{1 \cdot \frac{\cos x}{\sin x} - \frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)} \cdot 1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)} \cdot \frac{\cos x}{\sin x}}} \]
            4. Applied rewrites31.1%

              \[\leadsto \color{blue}{\frac{1 \cdot \frac{1}{\tan x} - \frac{1}{\tan \left(x + \varepsilon\right)} \cdot 1}{\frac{1}{\tan \left(x + \varepsilon\right)} \cdot \frac{1}{\tan x}}} \]
            5. Taylor expanded in eps around 0

              \[\leadsto \color{blue}{\varepsilon \cdot \left(1 - -1 \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right)} \]
            6. Step-by-step derivation
              1. cancel-sign-sub-invN/A

                \[\leadsto \varepsilon \cdot \color{blue}{\left(1 + \left(\mathsf{neg}\left(-1\right)\right) \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right)} \]
              2. metadata-evalN/A

                \[\leadsto \varepsilon \cdot \left(1 + \color{blue}{1} \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right) \]
              3. *-lft-identityN/A

                \[\leadsto \varepsilon \cdot \left(1 + \color{blue}{\frac{{\sin x}^{2}}{{\cos x}^{2}}}\right) \]
              4. +-commutativeN/A

                \[\leadsto \varepsilon \cdot \color{blue}{\left(\frac{{\sin x}^{2}}{{\cos x}^{2}} + 1\right)} \]
              5. distribute-rgt-inN/A

                \[\leadsto \color{blue}{\frac{{\sin x}^{2}}{{\cos x}^{2}} \cdot \varepsilon + 1 \cdot \varepsilon} \]
              6. *-lft-identityN/A

                \[\leadsto \frac{{\sin x}^{2}}{{\cos x}^{2}} \cdot \varepsilon + \color{blue}{\varepsilon} \]
              7. lower-fma.f64N/A

                \[\leadsto \color{blue}{\mathsf{fma}\left(\frac{{\sin x}^{2}}{{\cos x}^{2}}, \varepsilon, \varepsilon\right)} \]
              8. lower-/.f64N/A

                \[\leadsto \mathsf{fma}\left(\color{blue}{\frac{{\sin x}^{2}}{{\cos x}^{2}}}, \varepsilon, \varepsilon\right) \]
              9. lower-pow.f64N/A

                \[\leadsto \mathsf{fma}\left(\frac{\color{blue}{{\sin x}^{2}}}{{\cos x}^{2}}, \varepsilon, \varepsilon\right) \]
              10. lower-sin.f64N/A

                \[\leadsto \mathsf{fma}\left(\frac{{\color{blue}{\sin x}}^{2}}{{\cos x}^{2}}, \varepsilon, \varepsilon\right) \]
              11. lower-pow.f64N/A

                \[\leadsto \mathsf{fma}\left(\frac{{\sin x}^{2}}{\color{blue}{{\cos x}^{2}}}, \varepsilon, \varepsilon\right) \]
              12. lower-cos.f6498.0

                \[\leadsto \mathsf{fma}\left(\frac{{\sin x}^{2}}{{\color{blue}{\cos x}}^{2}}, \varepsilon, \varepsilon\right) \]
            7. Applied rewrites98.0%

              \[\leadsto \color{blue}{\mathsf{fma}\left(\frac{{\sin x}^{2}}{{\cos x}^{2}}, \varepsilon, \varepsilon\right)} \]
            8. Step-by-step derivation
              1. Applied rewrites98.0%

                \[\leadsto \mathsf{fma}\left({\tan x}^{2}, \color{blue}{\varepsilon}, \varepsilon\right) \]
              2. Add Preprocessing

              Alternative 7: 98.5% accurate, 4.1× speedup?

              \[\begin{array}{l} \\ \mathsf{fma}\left(\left(x \cdot x\right) \cdot \mathsf{fma}\left(x \cdot x, \mathsf{fma}\left(x \cdot x, \mathsf{fma}\left(x \cdot x, 0.19682539682539682, 0.37777777777777777\right), 0.6666666666666666\right), 1\right), \varepsilon, \varepsilon\right) \end{array} \]
              (FPCore (x eps)
               :precision binary64
               (fma
                (*
                 (* x x)
                 (fma
                  (* x x)
                  (fma
                   (* x x)
                   (fma (* x x) 0.19682539682539682 0.37777777777777777)
                   0.6666666666666666)
                  1.0))
                eps
                eps))
              double code(double x, double eps) {
              	return fma(((x * x) * fma((x * x), fma((x * x), fma((x * x), 0.19682539682539682, 0.37777777777777777), 0.6666666666666666), 1.0)), eps, eps);
              }
              
              function code(x, eps)
              	return fma(Float64(Float64(x * x) * fma(Float64(x * x), fma(Float64(x * x), fma(Float64(x * x), 0.19682539682539682, 0.37777777777777777), 0.6666666666666666), 1.0)), eps, eps)
              end
              
              code[x_, eps_] := N[(N[(N[(x * x), $MachinePrecision] * N[(N[(x * x), $MachinePrecision] * N[(N[(x * x), $MachinePrecision] * N[(N[(x * x), $MachinePrecision] * 0.19682539682539682 + 0.37777777777777777), $MachinePrecision] + 0.6666666666666666), $MachinePrecision] + 1.0), $MachinePrecision]), $MachinePrecision] * eps + eps), $MachinePrecision]
              
              \begin{array}{l}
              
              \\
              \mathsf{fma}\left(\left(x \cdot x\right) \cdot \mathsf{fma}\left(x \cdot x, \mathsf{fma}\left(x \cdot x, \mathsf{fma}\left(x \cdot x, 0.19682539682539682, 0.37777777777777777\right), 0.6666666666666666\right), 1\right), \varepsilon, \varepsilon\right)
              \end{array}
              
              Derivation
              1. Initial program 59.9%

                \[\tan \left(x + \varepsilon\right) - \tan x \]
              2. Add Preprocessing
              3. Step-by-step derivation
                1. lift--.f64N/A

                  \[\leadsto \color{blue}{\tan \left(x + \varepsilon\right) - \tan x} \]
                2. lift-tan.f64N/A

                  \[\leadsto \color{blue}{\tan \left(x + \varepsilon\right)} - \tan x \]
                3. tan-quotN/A

                  \[\leadsto \color{blue}{\frac{\sin \left(x + \varepsilon\right)}{\cos \left(x + \varepsilon\right)}} - \tan x \]
                4. clear-numN/A

                  \[\leadsto \color{blue}{\frac{1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)}}} - \tan x \]
                5. lift-tan.f64N/A

                  \[\leadsto \frac{1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)}} - \color{blue}{\tan x} \]
                6. tan-quotN/A

                  \[\leadsto \frac{1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)}} - \color{blue}{\frac{\sin x}{\cos x}} \]
                7. clear-numN/A

                  \[\leadsto \frac{1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)}} - \color{blue}{\frac{1}{\frac{\cos x}{\sin x}}} \]
                8. frac-subN/A

                  \[\leadsto \color{blue}{\frac{1 \cdot \frac{\cos x}{\sin x} - \frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)} \cdot 1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)} \cdot \frac{\cos x}{\sin x}}} \]
                9. lower-/.f64N/A

                  \[\leadsto \color{blue}{\frac{1 \cdot \frac{\cos x}{\sin x} - \frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)} \cdot 1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)} \cdot \frac{\cos x}{\sin x}}} \]
              4. Applied rewrites31.1%

                \[\leadsto \color{blue}{\frac{1 \cdot \frac{1}{\tan x} - \frac{1}{\tan \left(x + \varepsilon\right)} \cdot 1}{\frac{1}{\tan \left(x + \varepsilon\right)} \cdot \frac{1}{\tan x}}} \]
              5. Taylor expanded in eps around 0

                \[\leadsto \color{blue}{\varepsilon \cdot \left(1 - -1 \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right)} \]
              6. Step-by-step derivation
                1. cancel-sign-sub-invN/A

                  \[\leadsto \varepsilon \cdot \color{blue}{\left(1 + \left(\mathsf{neg}\left(-1\right)\right) \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right)} \]
                2. metadata-evalN/A

                  \[\leadsto \varepsilon \cdot \left(1 + \color{blue}{1} \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right) \]
                3. *-lft-identityN/A

                  \[\leadsto \varepsilon \cdot \left(1 + \color{blue}{\frac{{\sin x}^{2}}{{\cos x}^{2}}}\right) \]
                4. +-commutativeN/A

                  \[\leadsto \varepsilon \cdot \color{blue}{\left(\frac{{\sin x}^{2}}{{\cos x}^{2}} + 1\right)} \]
                5. distribute-rgt-inN/A

                  \[\leadsto \color{blue}{\frac{{\sin x}^{2}}{{\cos x}^{2}} \cdot \varepsilon + 1 \cdot \varepsilon} \]
                6. *-lft-identityN/A

                  \[\leadsto \frac{{\sin x}^{2}}{{\cos x}^{2}} \cdot \varepsilon + \color{blue}{\varepsilon} \]
                7. lower-fma.f64N/A

                  \[\leadsto \color{blue}{\mathsf{fma}\left(\frac{{\sin x}^{2}}{{\cos x}^{2}}, \varepsilon, \varepsilon\right)} \]
                8. lower-/.f64N/A

                  \[\leadsto \mathsf{fma}\left(\color{blue}{\frac{{\sin x}^{2}}{{\cos x}^{2}}}, \varepsilon, \varepsilon\right) \]
                9. lower-pow.f64N/A

                  \[\leadsto \mathsf{fma}\left(\frac{\color{blue}{{\sin x}^{2}}}{{\cos x}^{2}}, \varepsilon, \varepsilon\right) \]
                10. lower-sin.f64N/A

                  \[\leadsto \mathsf{fma}\left(\frac{{\color{blue}{\sin x}}^{2}}{{\cos x}^{2}}, \varepsilon, \varepsilon\right) \]
                11. lower-pow.f64N/A

                  \[\leadsto \mathsf{fma}\left(\frac{{\sin x}^{2}}{\color{blue}{{\cos x}^{2}}}, \varepsilon, \varepsilon\right) \]
                12. lower-cos.f6498.0

                  \[\leadsto \mathsf{fma}\left(\frac{{\sin x}^{2}}{{\color{blue}{\cos x}}^{2}}, \varepsilon, \varepsilon\right) \]
              7. Applied rewrites98.0%

                \[\leadsto \color{blue}{\mathsf{fma}\left(\frac{{\sin x}^{2}}{{\cos x}^{2}}, \varepsilon, \varepsilon\right)} \]
              8. Taylor expanded in x around 0

                \[\leadsto \mathsf{fma}\left({x}^{2} \cdot \left(1 + {x}^{2} \cdot \left(\frac{2}{3} + {x}^{2} \cdot \left(\frac{17}{45} + \frac{62}{315} \cdot {x}^{2}\right)\right)\right), \varepsilon, \varepsilon\right) \]
              9. Step-by-step derivation
                1. Applied rewrites97.9%

                  \[\leadsto \mathsf{fma}\left(\left(x \cdot x\right) \cdot \mathsf{fma}\left(x \cdot x, \mathsf{fma}\left(x \cdot x, \mathsf{fma}\left(x \cdot x, 0.19682539682539682, 0.37777777777777777\right), 0.6666666666666666\right), 1\right), \varepsilon, \varepsilon\right) \]
                2. Add Preprocessing

                Alternative 8: 98.5% accurate, 5.3× speedup?

                \[\begin{array}{l} \\ \mathsf{fma}\left(\left(x \cdot x\right) \cdot \mathsf{fma}\left(x \cdot x, \mathsf{fma}\left(x \cdot x, 0.37777777777777777, 0.6666666666666666\right), 1\right), \varepsilon, \varepsilon\right) \end{array} \]
                (FPCore (x eps)
                 :precision binary64
                 (fma
                  (*
                   (* x x)
                   (fma (* x x) (fma (* x x) 0.37777777777777777 0.6666666666666666) 1.0))
                  eps
                  eps))
                double code(double x, double eps) {
                	return fma(((x * x) * fma((x * x), fma((x * x), 0.37777777777777777, 0.6666666666666666), 1.0)), eps, eps);
                }
                
                function code(x, eps)
                	return fma(Float64(Float64(x * x) * fma(Float64(x * x), fma(Float64(x * x), 0.37777777777777777, 0.6666666666666666), 1.0)), eps, eps)
                end
                
                code[x_, eps_] := N[(N[(N[(x * x), $MachinePrecision] * N[(N[(x * x), $MachinePrecision] * N[(N[(x * x), $MachinePrecision] * 0.37777777777777777 + 0.6666666666666666), $MachinePrecision] + 1.0), $MachinePrecision]), $MachinePrecision] * eps + eps), $MachinePrecision]
                
                \begin{array}{l}
                
                \\
                \mathsf{fma}\left(\left(x \cdot x\right) \cdot \mathsf{fma}\left(x \cdot x, \mathsf{fma}\left(x \cdot x, 0.37777777777777777, 0.6666666666666666\right), 1\right), \varepsilon, \varepsilon\right)
                \end{array}
                
                Derivation
                1. Initial program 59.9%

                  \[\tan \left(x + \varepsilon\right) - \tan x \]
                2. Add Preprocessing
                3. Step-by-step derivation
                  1. lift--.f64N/A

                    \[\leadsto \color{blue}{\tan \left(x + \varepsilon\right) - \tan x} \]
                  2. lift-tan.f64N/A

                    \[\leadsto \color{blue}{\tan \left(x + \varepsilon\right)} - \tan x \]
                  3. tan-quotN/A

                    \[\leadsto \color{blue}{\frac{\sin \left(x + \varepsilon\right)}{\cos \left(x + \varepsilon\right)}} - \tan x \]
                  4. clear-numN/A

                    \[\leadsto \color{blue}{\frac{1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)}}} - \tan x \]
                  5. lift-tan.f64N/A

                    \[\leadsto \frac{1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)}} - \color{blue}{\tan x} \]
                  6. tan-quotN/A

                    \[\leadsto \frac{1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)}} - \color{blue}{\frac{\sin x}{\cos x}} \]
                  7. clear-numN/A

                    \[\leadsto \frac{1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)}} - \color{blue}{\frac{1}{\frac{\cos x}{\sin x}}} \]
                  8. frac-subN/A

                    \[\leadsto \color{blue}{\frac{1 \cdot \frac{\cos x}{\sin x} - \frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)} \cdot 1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)} \cdot \frac{\cos x}{\sin x}}} \]
                  9. lower-/.f64N/A

                    \[\leadsto \color{blue}{\frac{1 \cdot \frac{\cos x}{\sin x} - \frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)} \cdot 1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)} \cdot \frac{\cos x}{\sin x}}} \]
                4. Applied rewrites31.1%

                  \[\leadsto \color{blue}{\frac{1 \cdot \frac{1}{\tan x} - \frac{1}{\tan \left(x + \varepsilon\right)} \cdot 1}{\frac{1}{\tan \left(x + \varepsilon\right)} \cdot \frac{1}{\tan x}}} \]
                5. Taylor expanded in eps around 0

                  \[\leadsto \color{blue}{\varepsilon \cdot \left(1 - -1 \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right)} \]
                6. Step-by-step derivation
                  1. cancel-sign-sub-invN/A

                    \[\leadsto \varepsilon \cdot \color{blue}{\left(1 + \left(\mathsf{neg}\left(-1\right)\right) \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right)} \]
                  2. metadata-evalN/A

                    \[\leadsto \varepsilon \cdot \left(1 + \color{blue}{1} \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right) \]
                  3. *-lft-identityN/A

                    \[\leadsto \varepsilon \cdot \left(1 + \color{blue}{\frac{{\sin x}^{2}}{{\cos x}^{2}}}\right) \]
                  4. +-commutativeN/A

                    \[\leadsto \varepsilon \cdot \color{blue}{\left(\frac{{\sin x}^{2}}{{\cos x}^{2}} + 1\right)} \]
                  5. distribute-rgt-inN/A

                    \[\leadsto \color{blue}{\frac{{\sin x}^{2}}{{\cos x}^{2}} \cdot \varepsilon + 1 \cdot \varepsilon} \]
                  6. *-lft-identityN/A

                    \[\leadsto \frac{{\sin x}^{2}}{{\cos x}^{2}} \cdot \varepsilon + \color{blue}{\varepsilon} \]
                  7. lower-fma.f64N/A

                    \[\leadsto \color{blue}{\mathsf{fma}\left(\frac{{\sin x}^{2}}{{\cos x}^{2}}, \varepsilon, \varepsilon\right)} \]
                  8. lower-/.f64N/A

                    \[\leadsto \mathsf{fma}\left(\color{blue}{\frac{{\sin x}^{2}}{{\cos x}^{2}}}, \varepsilon, \varepsilon\right) \]
                  9. lower-pow.f64N/A

                    \[\leadsto \mathsf{fma}\left(\frac{\color{blue}{{\sin x}^{2}}}{{\cos x}^{2}}, \varepsilon, \varepsilon\right) \]
                  10. lower-sin.f64N/A

                    \[\leadsto \mathsf{fma}\left(\frac{{\color{blue}{\sin x}}^{2}}{{\cos x}^{2}}, \varepsilon, \varepsilon\right) \]
                  11. lower-pow.f64N/A

                    \[\leadsto \mathsf{fma}\left(\frac{{\sin x}^{2}}{\color{blue}{{\cos x}^{2}}}, \varepsilon, \varepsilon\right) \]
                  12. lower-cos.f6498.0

                    \[\leadsto \mathsf{fma}\left(\frac{{\sin x}^{2}}{{\color{blue}{\cos x}}^{2}}, \varepsilon, \varepsilon\right) \]
                7. Applied rewrites98.0%

                  \[\leadsto \color{blue}{\mathsf{fma}\left(\frac{{\sin x}^{2}}{{\cos x}^{2}}, \varepsilon, \varepsilon\right)} \]
                8. Taylor expanded in x around 0

                  \[\leadsto \mathsf{fma}\left({x}^{2} \cdot \left(1 + {x}^{2} \cdot \left(\frac{2}{3} + \frac{17}{45} \cdot {x}^{2}\right)\right), \varepsilon, \varepsilon\right) \]
                9. Step-by-step derivation
                  1. Applied rewrites97.8%

                    \[\leadsto \mathsf{fma}\left(\left(x \cdot x\right) \cdot \mathsf{fma}\left(x \cdot x, \mathsf{fma}\left(x \cdot x, 0.37777777777777777, 0.6666666666666666\right), 1\right), \varepsilon, \varepsilon\right) \]
                  2. Add Preprocessing

                  Alternative 9: 98.5% accurate, 7.4× speedup?

                  \[\begin{array}{l} \\ \mathsf{fma}\left(x \cdot \left(x \cdot \mathsf{fma}\left(x \cdot x, 0.6666666666666666, 1\right)\right), \varepsilon, \varepsilon\right) \end{array} \]
                  (FPCore (x eps)
                   :precision binary64
                   (fma (* x (* x (fma (* x x) 0.6666666666666666 1.0))) eps eps))
                  double code(double x, double eps) {
                  	return fma((x * (x * fma((x * x), 0.6666666666666666, 1.0))), eps, eps);
                  }
                  
                  function code(x, eps)
                  	return fma(Float64(x * Float64(x * fma(Float64(x * x), 0.6666666666666666, 1.0))), eps, eps)
                  end
                  
                  code[x_, eps_] := N[(N[(x * N[(x * N[(N[(x * x), $MachinePrecision] * 0.6666666666666666 + 1.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] * eps + eps), $MachinePrecision]
                  
                  \begin{array}{l}
                  
                  \\
                  \mathsf{fma}\left(x \cdot \left(x \cdot \mathsf{fma}\left(x \cdot x, 0.6666666666666666, 1\right)\right), \varepsilon, \varepsilon\right)
                  \end{array}
                  
                  Derivation
                  1. Initial program 59.9%

                    \[\tan \left(x + \varepsilon\right) - \tan x \]
                  2. Add Preprocessing
                  3. Step-by-step derivation
                    1. lift--.f64N/A

                      \[\leadsto \color{blue}{\tan \left(x + \varepsilon\right) - \tan x} \]
                    2. lift-tan.f64N/A

                      \[\leadsto \color{blue}{\tan \left(x + \varepsilon\right)} - \tan x \]
                    3. tan-quotN/A

                      \[\leadsto \color{blue}{\frac{\sin \left(x + \varepsilon\right)}{\cos \left(x + \varepsilon\right)}} - \tan x \]
                    4. clear-numN/A

                      \[\leadsto \color{blue}{\frac{1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)}}} - \tan x \]
                    5. lift-tan.f64N/A

                      \[\leadsto \frac{1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)}} - \color{blue}{\tan x} \]
                    6. tan-quotN/A

                      \[\leadsto \frac{1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)}} - \color{blue}{\frac{\sin x}{\cos x}} \]
                    7. clear-numN/A

                      \[\leadsto \frac{1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)}} - \color{blue}{\frac{1}{\frac{\cos x}{\sin x}}} \]
                    8. frac-subN/A

                      \[\leadsto \color{blue}{\frac{1 \cdot \frac{\cos x}{\sin x} - \frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)} \cdot 1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)} \cdot \frac{\cos x}{\sin x}}} \]
                    9. lower-/.f64N/A

                      \[\leadsto \color{blue}{\frac{1 \cdot \frac{\cos x}{\sin x} - \frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)} \cdot 1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)} \cdot \frac{\cos x}{\sin x}}} \]
                  4. Applied rewrites31.1%

                    \[\leadsto \color{blue}{\frac{1 \cdot \frac{1}{\tan x} - \frac{1}{\tan \left(x + \varepsilon\right)} \cdot 1}{\frac{1}{\tan \left(x + \varepsilon\right)} \cdot \frac{1}{\tan x}}} \]
                  5. Taylor expanded in eps around 0

                    \[\leadsto \color{blue}{\varepsilon \cdot \left(1 - -1 \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right)} \]
                  6. Step-by-step derivation
                    1. cancel-sign-sub-invN/A

                      \[\leadsto \varepsilon \cdot \color{blue}{\left(1 + \left(\mathsf{neg}\left(-1\right)\right) \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right)} \]
                    2. metadata-evalN/A

                      \[\leadsto \varepsilon \cdot \left(1 + \color{blue}{1} \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right) \]
                    3. *-lft-identityN/A

                      \[\leadsto \varepsilon \cdot \left(1 + \color{blue}{\frac{{\sin x}^{2}}{{\cos x}^{2}}}\right) \]
                    4. +-commutativeN/A

                      \[\leadsto \varepsilon \cdot \color{blue}{\left(\frac{{\sin x}^{2}}{{\cos x}^{2}} + 1\right)} \]
                    5. distribute-rgt-inN/A

                      \[\leadsto \color{blue}{\frac{{\sin x}^{2}}{{\cos x}^{2}} \cdot \varepsilon + 1 \cdot \varepsilon} \]
                    6. *-lft-identityN/A

                      \[\leadsto \frac{{\sin x}^{2}}{{\cos x}^{2}} \cdot \varepsilon + \color{blue}{\varepsilon} \]
                    7. lower-fma.f64N/A

                      \[\leadsto \color{blue}{\mathsf{fma}\left(\frac{{\sin x}^{2}}{{\cos x}^{2}}, \varepsilon, \varepsilon\right)} \]
                    8. lower-/.f64N/A

                      \[\leadsto \mathsf{fma}\left(\color{blue}{\frac{{\sin x}^{2}}{{\cos x}^{2}}}, \varepsilon, \varepsilon\right) \]
                    9. lower-pow.f64N/A

                      \[\leadsto \mathsf{fma}\left(\frac{\color{blue}{{\sin x}^{2}}}{{\cos x}^{2}}, \varepsilon, \varepsilon\right) \]
                    10. lower-sin.f64N/A

                      \[\leadsto \mathsf{fma}\left(\frac{{\color{blue}{\sin x}}^{2}}{{\cos x}^{2}}, \varepsilon, \varepsilon\right) \]
                    11. lower-pow.f64N/A

                      \[\leadsto \mathsf{fma}\left(\frac{{\sin x}^{2}}{\color{blue}{{\cos x}^{2}}}, \varepsilon, \varepsilon\right) \]
                    12. lower-cos.f6498.0

                      \[\leadsto \mathsf{fma}\left(\frac{{\sin x}^{2}}{{\color{blue}{\cos x}}^{2}}, \varepsilon, \varepsilon\right) \]
                  7. Applied rewrites98.0%

                    \[\leadsto \color{blue}{\mathsf{fma}\left(\frac{{\sin x}^{2}}{{\cos x}^{2}}, \varepsilon, \varepsilon\right)} \]
                  8. Taylor expanded in x around 0

                    \[\leadsto \mathsf{fma}\left({x}^{2} \cdot \left(1 + \frac{2}{3} \cdot {x}^{2}\right), \varepsilon, \varepsilon\right) \]
                  9. Step-by-step derivation
                    1. Applied rewrites97.7%

                      \[\leadsto \mathsf{fma}\left(\left(\mathsf{fma}\left(x \cdot x, 0.6666666666666666, 1\right) \cdot x\right) \cdot x, \varepsilon, \varepsilon\right) \]
                    2. Final simplification97.7%

                      \[\leadsto \mathsf{fma}\left(x \cdot \left(x \cdot \mathsf{fma}\left(x \cdot x, 0.6666666666666666, 1\right)\right), \varepsilon, \varepsilon\right) \]
                    3. Add Preprocessing

                    Alternative 10: 98.4% accurate, 14.8× speedup?

                    \[\begin{array}{l} \\ \varepsilon \cdot \left(1 + x \cdot x\right) \end{array} \]
                    (FPCore (x eps) :precision binary64 (* eps (+ 1.0 (* x x))))
                    double code(double x, double eps) {
                    	return eps * (1.0 + (x * x));
                    }
                    
                    real(8) function code(x, eps)
                        real(8), intent (in) :: x
                        real(8), intent (in) :: eps
                        code = eps * (1.0d0 + (x * x))
                    end function
                    
                    public static double code(double x, double eps) {
                    	return eps * (1.0 + (x * x));
                    }
                    
                    def code(x, eps):
                    	return eps * (1.0 + (x * x))
                    
                    function code(x, eps)
                    	return Float64(eps * Float64(1.0 + Float64(x * x)))
                    end
                    
                    function tmp = code(x, eps)
                    	tmp = eps * (1.0 + (x * x));
                    end
                    
                    code[x_, eps_] := N[(eps * N[(1.0 + N[(x * x), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
                    
                    \begin{array}{l}
                    
                    \\
                    \varepsilon \cdot \left(1 + x \cdot x\right)
                    \end{array}
                    
                    Derivation
                    1. Initial program 59.9%

                      \[\tan \left(x + \varepsilon\right) - \tan x \]
                    2. Add Preprocessing
                    3. Step-by-step derivation
                      1. lift--.f64N/A

                        \[\leadsto \color{blue}{\tan \left(x + \varepsilon\right) - \tan x} \]
                      2. lift-tan.f64N/A

                        \[\leadsto \color{blue}{\tan \left(x + \varepsilon\right)} - \tan x \]
                      3. tan-quotN/A

                        \[\leadsto \color{blue}{\frac{\sin \left(x + \varepsilon\right)}{\cos \left(x + \varepsilon\right)}} - \tan x \]
                      4. clear-numN/A

                        \[\leadsto \color{blue}{\frac{1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)}}} - \tan x \]
                      5. lift-tan.f64N/A

                        \[\leadsto \frac{1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)}} - \color{blue}{\tan x} \]
                      6. tan-quotN/A

                        \[\leadsto \frac{1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)}} - \color{blue}{\frac{\sin x}{\cos x}} \]
                      7. clear-numN/A

                        \[\leadsto \frac{1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)}} - \color{blue}{\frac{1}{\frac{\cos x}{\sin x}}} \]
                      8. frac-subN/A

                        \[\leadsto \color{blue}{\frac{1 \cdot \frac{\cos x}{\sin x} - \frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)} \cdot 1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)} \cdot \frac{\cos x}{\sin x}}} \]
                      9. lower-/.f64N/A

                        \[\leadsto \color{blue}{\frac{1 \cdot \frac{\cos x}{\sin x} - \frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)} \cdot 1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)} \cdot \frac{\cos x}{\sin x}}} \]
                    4. Applied rewrites31.1%

                      \[\leadsto \color{blue}{\frac{1 \cdot \frac{1}{\tan x} - \frac{1}{\tan \left(x + \varepsilon\right)} \cdot 1}{\frac{1}{\tan \left(x + \varepsilon\right)} \cdot \frac{1}{\tan x}}} \]
                    5. Taylor expanded in eps around 0

                      \[\leadsto \color{blue}{\varepsilon \cdot \left(1 - -1 \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right)} \]
                    6. Step-by-step derivation
                      1. cancel-sign-sub-invN/A

                        \[\leadsto \varepsilon \cdot \color{blue}{\left(1 + \left(\mathsf{neg}\left(-1\right)\right) \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right)} \]
                      2. metadata-evalN/A

                        \[\leadsto \varepsilon \cdot \left(1 + \color{blue}{1} \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right) \]
                      3. *-lft-identityN/A

                        \[\leadsto \varepsilon \cdot \left(1 + \color{blue}{\frac{{\sin x}^{2}}{{\cos x}^{2}}}\right) \]
                      4. +-commutativeN/A

                        \[\leadsto \varepsilon \cdot \color{blue}{\left(\frac{{\sin x}^{2}}{{\cos x}^{2}} + 1\right)} \]
                      5. distribute-rgt-inN/A

                        \[\leadsto \color{blue}{\frac{{\sin x}^{2}}{{\cos x}^{2}} \cdot \varepsilon + 1 \cdot \varepsilon} \]
                      6. *-lft-identityN/A

                        \[\leadsto \frac{{\sin x}^{2}}{{\cos x}^{2}} \cdot \varepsilon + \color{blue}{\varepsilon} \]
                      7. lower-fma.f64N/A

                        \[\leadsto \color{blue}{\mathsf{fma}\left(\frac{{\sin x}^{2}}{{\cos x}^{2}}, \varepsilon, \varepsilon\right)} \]
                      8. lower-/.f64N/A

                        \[\leadsto \mathsf{fma}\left(\color{blue}{\frac{{\sin x}^{2}}{{\cos x}^{2}}}, \varepsilon, \varepsilon\right) \]
                      9. lower-pow.f64N/A

                        \[\leadsto \mathsf{fma}\left(\frac{\color{blue}{{\sin x}^{2}}}{{\cos x}^{2}}, \varepsilon, \varepsilon\right) \]
                      10. lower-sin.f64N/A

                        \[\leadsto \mathsf{fma}\left(\frac{{\color{blue}{\sin x}}^{2}}{{\cos x}^{2}}, \varepsilon, \varepsilon\right) \]
                      11. lower-pow.f64N/A

                        \[\leadsto \mathsf{fma}\left(\frac{{\sin x}^{2}}{\color{blue}{{\cos x}^{2}}}, \varepsilon, \varepsilon\right) \]
                      12. lower-cos.f6498.0

                        \[\leadsto \mathsf{fma}\left(\frac{{\sin x}^{2}}{{\color{blue}{\cos x}}^{2}}, \varepsilon, \varepsilon\right) \]
                    7. Applied rewrites98.0%

                      \[\leadsto \color{blue}{\mathsf{fma}\left(\frac{{\sin x}^{2}}{{\cos x}^{2}}, \varepsilon, \varepsilon\right)} \]
                    8. Taylor expanded in x around 0

                      \[\leadsto \mathsf{fma}\left({x}^{2}, \varepsilon, \varepsilon\right) \]
                    9. Step-by-step derivation
                      1. Applied rewrites97.6%

                        \[\leadsto \mathsf{fma}\left(x \cdot x, \varepsilon, \varepsilon\right) \]
                      2. Step-by-step derivation
                        1. Applied rewrites97.6%

                          \[\leadsto \left(1 + x \cdot x\right) \cdot \color{blue}{\varepsilon} \]
                        2. Final simplification97.6%

                          \[\leadsto \varepsilon \cdot \left(1 + x \cdot x\right) \]
                        3. Add Preprocessing

                        Alternative 11: 98.4% accurate, 17.3× speedup?

                        \[\begin{array}{l} \\ \mathsf{fma}\left(x \cdot x, \varepsilon, \varepsilon\right) \end{array} \]
                        (FPCore (x eps) :precision binary64 (fma (* x x) eps eps))
                        double code(double x, double eps) {
                        	return fma((x * x), eps, eps);
                        }
                        
                        function code(x, eps)
                        	return fma(Float64(x * x), eps, eps)
                        end
                        
                        code[x_, eps_] := N[(N[(x * x), $MachinePrecision] * eps + eps), $MachinePrecision]
                        
                        \begin{array}{l}
                        
                        \\
                        \mathsf{fma}\left(x \cdot x, \varepsilon, \varepsilon\right)
                        \end{array}
                        
                        Derivation
                        1. Initial program 59.9%

                          \[\tan \left(x + \varepsilon\right) - \tan x \]
                        2. Add Preprocessing
                        3. Step-by-step derivation
                          1. lift--.f64N/A

                            \[\leadsto \color{blue}{\tan \left(x + \varepsilon\right) - \tan x} \]
                          2. lift-tan.f64N/A

                            \[\leadsto \color{blue}{\tan \left(x + \varepsilon\right)} - \tan x \]
                          3. tan-quotN/A

                            \[\leadsto \color{blue}{\frac{\sin \left(x + \varepsilon\right)}{\cos \left(x + \varepsilon\right)}} - \tan x \]
                          4. clear-numN/A

                            \[\leadsto \color{blue}{\frac{1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)}}} - \tan x \]
                          5. lift-tan.f64N/A

                            \[\leadsto \frac{1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)}} - \color{blue}{\tan x} \]
                          6. tan-quotN/A

                            \[\leadsto \frac{1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)}} - \color{blue}{\frac{\sin x}{\cos x}} \]
                          7. clear-numN/A

                            \[\leadsto \frac{1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)}} - \color{blue}{\frac{1}{\frac{\cos x}{\sin x}}} \]
                          8. frac-subN/A

                            \[\leadsto \color{blue}{\frac{1 \cdot \frac{\cos x}{\sin x} - \frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)} \cdot 1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)} \cdot \frac{\cos x}{\sin x}}} \]
                          9. lower-/.f64N/A

                            \[\leadsto \color{blue}{\frac{1 \cdot \frac{\cos x}{\sin x} - \frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)} \cdot 1}{\frac{\cos \left(x + \varepsilon\right)}{\sin \left(x + \varepsilon\right)} \cdot \frac{\cos x}{\sin x}}} \]
                        4. Applied rewrites31.1%

                          \[\leadsto \color{blue}{\frac{1 \cdot \frac{1}{\tan x} - \frac{1}{\tan \left(x + \varepsilon\right)} \cdot 1}{\frac{1}{\tan \left(x + \varepsilon\right)} \cdot \frac{1}{\tan x}}} \]
                        5. Taylor expanded in eps around 0

                          \[\leadsto \color{blue}{\varepsilon \cdot \left(1 - -1 \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right)} \]
                        6. Step-by-step derivation
                          1. cancel-sign-sub-invN/A

                            \[\leadsto \varepsilon \cdot \color{blue}{\left(1 + \left(\mathsf{neg}\left(-1\right)\right) \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right)} \]
                          2. metadata-evalN/A

                            \[\leadsto \varepsilon \cdot \left(1 + \color{blue}{1} \cdot \frac{{\sin x}^{2}}{{\cos x}^{2}}\right) \]
                          3. *-lft-identityN/A

                            \[\leadsto \varepsilon \cdot \left(1 + \color{blue}{\frac{{\sin x}^{2}}{{\cos x}^{2}}}\right) \]
                          4. +-commutativeN/A

                            \[\leadsto \varepsilon \cdot \color{blue}{\left(\frac{{\sin x}^{2}}{{\cos x}^{2}} + 1\right)} \]
                          5. distribute-rgt-inN/A

                            \[\leadsto \color{blue}{\frac{{\sin x}^{2}}{{\cos x}^{2}} \cdot \varepsilon + 1 \cdot \varepsilon} \]
                          6. *-lft-identityN/A

                            \[\leadsto \frac{{\sin x}^{2}}{{\cos x}^{2}} \cdot \varepsilon + \color{blue}{\varepsilon} \]
                          7. lower-fma.f64N/A

                            \[\leadsto \color{blue}{\mathsf{fma}\left(\frac{{\sin x}^{2}}{{\cos x}^{2}}, \varepsilon, \varepsilon\right)} \]
                          8. lower-/.f64N/A

                            \[\leadsto \mathsf{fma}\left(\color{blue}{\frac{{\sin x}^{2}}{{\cos x}^{2}}}, \varepsilon, \varepsilon\right) \]
                          9. lower-pow.f64N/A

                            \[\leadsto \mathsf{fma}\left(\frac{\color{blue}{{\sin x}^{2}}}{{\cos x}^{2}}, \varepsilon, \varepsilon\right) \]
                          10. lower-sin.f64N/A

                            \[\leadsto \mathsf{fma}\left(\frac{{\color{blue}{\sin x}}^{2}}{{\cos x}^{2}}, \varepsilon, \varepsilon\right) \]
                          11. lower-pow.f64N/A

                            \[\leadsto \mathsf{fma}\left(\frac{{\sin x}^{2}}{\color{blue}{{\cos x}^{2}}}, \varepsilon, \varepsilon\right) \]
                          12. lower-cos.f6498.0

                            \[\leadsto \mathsf{fma}\left(\frac{{\sin x}^{2}}{{\color{blue}{\cos x}}^{2}}, \varepsilon, \varepsilon\right) \]
                        7. Applied rewrites98.0%

                          \[\leadsto \color{blue}{\mathsf{fma}\left(\frac{{\sin x}^{2}}{{\cos x}^{2}}, \varepsilon, \varepsilon\right)} \]
                        8. Taylor expanded in x around 0

                          \[\leadsto \mathsf{fma}\left({x}^{2}, \varepsilon, \varepsilon\right) \]
                        9. Step-by-step derivation
                          1. Applied rewrites97.6%

                            \[\leadsto \mathsf{fma}\left(x \cdot x, \varepsilon, \varepsilon\right) \]
                          2. Add Preprocessing

                          Developer Target 1: 99.9% accurate, 0.6× speedup?

                          \[\begin{array}{l} \\ \frac{\sin \varepsilon}{\cos x \cdot \cos \left(x + \varepsilon\right)} \end{array} \]
                          (FPCore (x eps) :precision binary64 (/ (sin eps) (* (cos x) (cos (+ x eps)))))
                          double code(double x, double eps) {
                          	return sin(eps) / (cos(x) * cos((x + eps)));
                          }
                          
                          real(8) function code(x, eps)
                              real(8), intent (in) :: x
                              real(8), intent (in) :: eps
                              code = sin(eps) / (cos(x) * cos((x + eps)))
                          end function
                          
                          public static double code(double x, double eps) {
                          	return Math.sin(eps) / (Math.cos(x) * Math.cos((x + eps)));
                          }
                          
                          def code(x, eps):
                          	return math.sin(eps) / (math.cos(x) * math.cos((x + eps)))
                          
                          function code(x, eps)
                          	return Float64(sin(eps) / Float64(cos(x) * cos(Float64(x + eps))))
                          end
                          
                          function tmp = code(x, eps)
                          	tmp = sin(eps) / (cos(x) * cos((x + eps)));
                          end
                          
                          code[x_, eps_] := N[(N[Sin[eps], $MachinePrecision] / N[(N[Cos[x], $MachinePrecision] * N[Cos[N[(x + eps), $MachinePrecision]], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
                          
                          \begin{array}{l}
                          
                          \\
                          \frac{\sin \varepsilon}{\cos x \cdot \cos \left(x + \varepsilon\right)}
                          \end{array}
                          

                          Developer Target 2: 62.3% accurate, 0.4× speedup?

                          \[\begin{array}{l} \\ \frac{\tan x + \tan \varepsilon}{1 - \tan x \cdot \tan \varepsilon} - \tan x \end{array} \]
                          (FPCore (x eps)
                           :precision binary64
                           (- (/ (+ (tan x) (tan eps)) (- 1.0 (* (tan x) (tan eps)))) (tan x)))
                          double code(double x, double eps) {
                          	return ((tan(x) + tan(eps)) / (1.0 - (tan(x) * tan(eps)))) - tan(x);
                          }
                          
                          real(8) function code(x, eps)
                              real(8), intent (in) :: x
                              real(8), intent (in) :: eps
                              code = ((tan(x) + tan(eps)) / (1.0d0 - (tan(x) * tan(eps)))) - tan(x)
                          end function
                          
                          public static double code(double x, double eps) {
                          	return ((Math.tan(x) + Math.tan(eps)) / (1.0 - (Math.tan(x) * Math.tan(eps)))) - Math.tan(x);
                          }
                          
                          def code(x, eps):
                          	return ((math.tan(x) + math.tan(eps)) / (1.0 - (math.tan(x) * math.tan(eps)))) - math.tan(x)
                          
                          function code(x, eps)
                          	return Float64(Float64(Float64(tan(x) + tan(eps)) / Float64(1.0 - Float64(tan(x) * tan(eps)))) - tan(x))
                          end
                          
                          function tmp = code(x, eps)
                          	tmp = ((tan(x) + tan(eps)) / (1.0 - (tan(x) * tan(eps)))) - tan(x);
                          end
                          
                          code[x_, eps_] := N[(N[(N[(N[Tan[x], $MachinePrecision] + N[Tan[eps], $MachinePrecision]), $MachinePrecision] / N[(1.0 - N[(N[Tan[x], $MachinePrecision] * N[Tan[eps], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] - N[Tan[x], $MachinePrecision]), $MachinePrecision]
                          
                          \begin{array}{l}
                          
                          \\
                          \frac{\tan x + \tan \varepsilon}{1 - \tan x \cdot \tan \varepsilon} - \tan x
                          \end{array}
                          

                          Developer Target 3: 99.1% accurate, 1.0× speedup?

                          \[\begin{array}{l} \\ \varepsilon + \left(\varepsilon \cdot \tan x\right) \cdot \tan x \end{array} \]
                          (FPCore (x eps) :precision binary64 (+ eps (* (* eps (tan x)) (tan x))))
                          double code(double x, double eps) {
                          	return eps + ((eps * tan(x)) * tan(x));
                          }
                          
                          real(8) function code(x, eps)
                              real(8), intent (in) :: x
                              real(8), intent (in) :: eps
                              code = eps + ((eps * tan(x)) * tan(x))
                          end function
                          
                          public static double code(double x, double eps) {
                          	return eps + ((eps * Math.tan(x)) * Math.tan(x));
                          }
                          
                          def code(x, eps):
                          	return eps + ((eps * math.tan(x)) * math.tan(x))
                          
                          function code(x, eps)
                          	return Float64(eps + Float64(Float64(eps * tan(x)) * tan(x)))
                          end
                          
                          function tmp = code(x, eps)
                          	tmp = eps + ((eps * tan(x)) * tan(x));
                          end
                          
                          code[x_, eps_] := N[(eps + N[(N[(eps * N[Tan[x], $MachinePrecision]), $MachinePrecision] * N[Tan[x], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
                          
                          \begin{array}{l}
                          
                          \\
                          \varepsilon + \left(\varepsilon \cdot \tan x\right) \cdot \tan x
                          \end{array}
                          

                          Reproduce

                          ?
                          herbie shell --seed 2024227 
                          (FPCore (x eps)
                            :name "2tan (problem 3.3.2)"
                            :precision binary64
                            :pre (and (and (and (<= -10000.0 x) (<= x 10000.0)) (< (* 1e-16 (fabs x)) eps)) (< eps (fabs x)))
                          
                            :alt
                            (! :herbie-platform default (/ (sin eps) (* (cos x) (cos (+ x eps)))))
                          
                            :alt
                            (! :herbie-platform default (- (/ (+ (tan x) (tan eps)) (- 1 (* (tan x) (tan eps)))) (tan x)))
                          
                            :alt
                            (! :herbie-platform default (+ eps (* eps (tan x) (tan x))))
                          
                            (- (tan (+ x eps)) (tan x)))