Kahan p13 Example 3

Percentage Accurate: 100.0% → 100.0%
Time: 13.6s
Alternatives: 11
Speedup: 1.8×

Specification

?
\[\begin{array}{l} \\ \begin{array}{l} t_1 := 2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\\ 1 - \frac{1}{2 + t\_1 \cdot t\_1} \end{array} \end{array} \]
(FPCore (t)
 :precision binary64
 (let* ((t_1 (- 2.0 (/ (/ 2.0 t) (+ 1.0 (/ 1.0 t))))))
   (- 1.0 (/ 1.0 (+ 2.0 (* t_1 t_1))))))
double code(double t) {
	double t_1 = 2.0 - ((2.0 / t) / (1.0 + (1.0 / t)));
	return 1.0 - (1.0 / (2.0 + (t_1 * t_1)));
}
real(8) function code(t)
    real(8), intent (in) :: t
    real(8) :: t_1
    t_1 = 2.0d0 - ((2.0d0 / t) / (1.0d0 + (1.0d0 / t)))
    code = 1.0d0 - (1.0d0 / (2.0d0 + (t_1 * t_1)))
end function
public static double code(double t) {
	double t_1 = 2.0 - ((2.0 / t) / (1.0 + (1.0 / t)));
	return 1.0 - (1.0 / (2.0 + (t_1 * t_1)));
}
def code(t):
	t_1 = 2.0 - ((2.0 / t) / (1.0 + (1.0 / t)))
	return 1.0 - (1.0 / (2.0 + (t_1 * t_1)))
function code(t)
	t_1 = Float64(2.0 - Float64(Float64(2.0 / t) / Float64(1.0 + Float64(1.0 / t))))
	return Float64(1.0 - Float64(1.0 / Float64(2.0 + Float64(t_1 * t_1))))
end
function tmp = code(t)
	t_1 = 2.0 - ((2.0 / t) / (1.0 + (1.0 / t)));
	tmp = 1.0 - (1.0 / (2.0 + (t_1 * t_1)));
end
code[t_] := Block[{t$95$1 = N[(2.0 - N[(N[(2.0 / t), $MachinePrecision] / N[(1.0 + N[(1.0 / t), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]}, N[(1.0 - N[(1.0 / N[(2.0 + N[(t$95$1 * t$95$1), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]]
\begin{array}{l}

\\
\begin{array}{l}
t_1 := 2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\\
1 - \frac{1}{2 + t\_1 \cdot t\_1}
\end{array}
\end{array}

Sampling outcomes in binary64 precision:

Local Percentage Accuracy vs ?

The average percentage accuracy by input value. Horizontal axis shows value of an input variable; the variable is choosen in the title. Vertical axis is accuracy; higher is better. Red represent the original program, while blue represents Herbie's suggestion. These can be toggled with buttons below the plot. The line is an average while dots represent individual samples.

Accuracy vs Speed?

Herbie found 11 alternatives:

AlternativeAccuracySpeedup
The accuracy (vertical axis) and speed (horizontal axis) of each alternatives. Up and to the right is better. The red square shows the initial program, and each blue circle shows an alternative.The line shows the best available speed-accuracy tradeoffs.

Initial Program: 100.0% accurate, 1.0× speedup?

\[\begin{array}{l} \\ \begin{array}{l} t_1 := 2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\\ 1 - \frac{1}{2 + t\_1 \cdot t\_1} \end{array} \end{array} \]
(FPCore (t)
 :precision binary64
 (let* ((t_1 (- 2.0 (/ (/ 2.0 t) (+ 1.0 (/ 1.0 t))))))
   (- 1.0 (/ 1.0 (+ 2.0 (* t_1 t_1))))))
double code(double t) {
	double t_1 = 2.0 - ((2.0 / t) / (1.0 + (1.0 / t)));
	return 1.0 - (1.0 / (2.0 + (t_1 * t_1)));
}
real(8) function code(t)
    real(8), intent (in) :: t
    real(8) :: t_1
    t_1 = 2.0d0 - ((2.0d0 / t) / (1.0d0 + (1.0d0 / t)))
    code = 1.0d0 - (1.0d0 / (2.0d0 + (t_1 * t_1)))
end function
public static double code(double t) {
	double t_1 = 2.0 - ((2.0 / t) / (1.0 + (1.0 / t)));
	return 1.0 - (1.0 / (2.0 + (t_1 * t_1)));
}
def code(t):
	t_1 = 2.0 - ((2.0 / t) / (1.0 + (1.0 / t)))
	return 1.0 - (1.0 / (2.0 + (t_1 * t_1)))
function code(t)
	t_1 = Float64(2.0 - Float64(Float64(2.0 / t) / Float64(1.0 + Float64(1.0 / t))))
	return Float64(1.0 - Float64(1.0 / Float64(2.0 + Float64(t_1 * t_1))))
end
function tmp = code(t)
	t_1 = 2.0 - ((2.0 / t) / (1.0 + (1.0 / t)));
	tmp = 1.0 - (1.0 / (2.0 + (t_1 * t_1)));
end
code[t_] := Block[{t$95$1 = N[(2.0 - N[(N[(2.0 / t), $MachinePrecision] / N[(1.0 + N[(1.0 / t), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]}, N[(1.0 - N[(1.0 / N[(2.0 + N[(t$95$1 * t$95$1), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]]
\begin{array}{l}

\\
\begin{array}{l}
t_1 := 2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\\
1 - \frac{1}{2 + t\_1 \cdot t\_1}
\end{array}
\end{array}

Alternative 1: 100.0% accurate, 1.7× speedup?

\[\begin{array}{l} \\ 1 + \frac{1}{\frac{\frac{-2}{1 + t} - -2}{-0.5 + \frac{-0.5}{t}} - 2} \end{array} \]
(FPCore (t)
 :precision binary64
 (+ 1.0 (/ 1.0 (- (/ (- (/ -2.0 (+ 1.0 t)) -2.0) (+ -0.5 (/ -0.5 t))) 2.0))))
double code(double t) {
	return 1.0 + (1.0 / ((((-2.0 / (1.0 + t)) - -2.0) / (-0.5 + (-0.5 / t))) - 2.0));
}
real(8) function code(t)
    real(8), intent (in) :: t
    code = 1.0d0 + (1.0d0 / (((((-2.0d0) / (1.0d0 + t)) - (-2.0d0)) / ((-0.5d0) + ((-0.5d0) / t))) - 2.0d0))
end function
public static double code(double t) {
	return 1.0 + (1.0 / ((((-2.0 / (1.0 + t)) - -2.0) / (-0.5 + (-0.5 / t))) - 2.0));
}
def code(t):
	return 1.0 + (1.0 / ((((-2.0 / (1.0 + t)) - -2.0) / (-0.5 + (-0.5 / t))) - 2.0))
function code(t)
	return Float64(1.0 + Float64(1.0 / Float64(Float64(Float64(Float64(-2.0 / Float64(1.0 + t)) - -2.0) / Float64(-0.5 + Float64(-0.5 / t))) - 2.0)))
end
function tmp = code(t)
	tmp = 1.0 + (1.0 / ((((-2.0 / (1.0 + t)) - -2.0) / (-0.5 + (-0.5 / t))) - 2.0));
end
code[t_] := N[(1.0 + N[(1.0 / N[(N[(N[(N[(-2.0 / N[(1.0 + t), $MachinePrecision]), $MachinePrecision] - -2.0), $MachinePrecision] / N[(-0.5 + N[(-0.5 / t), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] - 2.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}

\\
1 + \frac{1}{\frac{\frac{-2}{1 + t} - -2}{-0.5 + \frac{-0.5}{t}} - 2}
\end{array}
Derivation
  1. Initial program 100.0%

    \[1 - \frac{1}{2 + \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right) \cdot \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right)} \]
  2. Add Preprocessing
  3. Step-by-step derivation
    1. flip--N/A

      \[\leadsto 1 - \frac{1}{2 + \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right) \cdot \color{blue}{\frac{2 \cdot 2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}} \cdot \frac{\frac{2}{t}}{1 + \frac{1}{t}}}{2 + \frac{\frac{2}{t}}{1 + \frac{1}{t}}}}} \]
    2. clear-numN/A

      \[\leadsto 1 - \frac{1}{2 + \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right) \cdot \color{blue}{\frac{1}{\frac{2 + \frac{\frac{2}{t}}{1 + \frac{1}{t}}}{2 \cdot 2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}} \cdot \frac{\frac{2}{t}}{1 + \frac{1}{t}}}}}} \]
    3. un-div-invN/A

      \[\leadsto 1 - \frac{1}{2 + \color{blue}{\frac{2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}}{\frac{2 + \frac{\frac{2}{t}}{1 + \frac{1}{t}}}{2 \cdot 2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}} \cdot \frac{\frac{2}{t}}{1 + \frac{1}{t}}}}}} \]
    4. /-lowering-/.f64N/A

      \[\leadsto 1 - \frac{1}{2 + \color{blue}{\frac{2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}}{\frac{2 + \frac{\frac{2}{t}}{1 + \frac{1}{t}}}{2 \cdot 2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}} \cdot \frac{\frac{2}{t}}{1 + \frac{1}{t}}}}}} \]
  4. Applied egg-rr100.0%

    \[\leadsto 1 - \frac{1}{2 + \color{blue}{\frac{2 - \frac{2}{\mathsf{fma}\left(t, \frac{1}{t}, t\right)}}{\frac{1}{2 - \frac{2}{\mathsf{fma}\left(t, \frac{1}{t}, t\right)}}}}} \]
  5. Step-by-step derivation
    1. div-subN/A

      \[\leadsto 1 - \frac{1}{2 + \color{blue}{\left(\frac{2}{\frac{1}{2 - \frac{2}{t \cdot \frac{1}{t} + t}}} - \frac{\frac{2}{t \cdot \frac{1}{t} + t}}{\frac{1}{2 - \frac{2}{t \cdot \frac{1}{t} + t}}}\right)}} \]
    2. frac-2negN/A

      \[\leadsto 1 - \frac{1}{2 + \left(\color{blue}{\frac{\mathsf{neg}\left(2\right)}{\mathsf{neg}\left(\frac{1}{2 - \frac{2}{t \cdot \frac{1}{t} + t}}\right)}} - \frac{\frac{2}{t \cdot \frac{1}{t} + t}}{\frac{1}{2 - \frac{2}{t \cdot \frac{1}{t} + t}}}\right)} \]
    3. metadata-evalN/A

      \[\leadsto 1 - \frac{1}{2 + \left(\frac{\color{blue}{-2}}{\mathsf{neg}\left(\frac{1}{2 - \frac{2}{t \cdot \frac{1}{t} + t}}\right)} - \frac{\frac{2}{t \cdot \frac{1}{t} + t}}{\frac{1}{2 - \frac{2}{t \cdot \frac{1}{t} + t}}}\right)} \]
    4. frac-2negN/A

      \[\leadsto 1 - \frac{1}{2 + \left(\frac{-2}{\mathsf{neg}\left(\frac{1}{2 - \frac{2}{t \cdot \frac{1}{t} + t}}\right)} - \color{blue}{\frac{\mathsf{neg}\left(\frac{2}{t \cdot \frac{1}{t} + t}\right)}{\mathsf{neg}\left(\frac{1}{2 - \frac{2}{t \cdot \frac{1}{t} + t}}\right)}}\right)} \]
    5. sub-divN/A

      \[\leadsto 1 - \frac{1}{2 + \color{blue}{\frac{-2 - \left(\mathsf{neg}\left(\frac{2}{t \cdot \frac{1}{t} + t}\right)\right)}{\mathsf{neg}\left(\frac{1}{2 - \frac{2}{t \cdot \frac{1}{t} + t}}\right)}}} \]
    6. /-lowering-/.f64N/A

      \[\leadsto 1 - \frac{1}{2 + \color{blue}{\frac{-2 - \left(\mathsf{neg}\left(\frac{2}{t \cdot \frac{1}{t} + t}\right)\right)}{\mathsf{neg}\left(\frac{1}{2 - \frac{2}{t \cdot \frac{1}{t} + t}}\right)}}} \]
  6. Applied egg-rr100.0%

    \[\leadsto 1 - \frac{1}{2 + \color{blue}{\frac{-2 - \frac{-2}{t + 1}}{\frac{1}{-2 + \frac{2}{t + 1}}}}} \]
  7. Taylor expanded in t around 0

    \[\leadsto 1 - \frac{1}{2 + \frac{-2 - \frac{-2}{t + 1}}{\color{blue}{\frac{\frac{-1}{2} \cdot t - \frac{1}{2}}{t}}}} \]
  8. Step-by-step derivation
    1. div-subN/A

      \[\leadsto 1 - \frac{1}{2 + \frac{-2 - \frac{-2}{t + 1}}{\color{blue}{\frac{\frac{-1}{2} \cdot t}{t} - \frac{\frac{1}{2}}{t}}}} \]
    2. metadata-evalN/A

      \[\leadsto 1 - \frac{1}{2 + \frac{-2 - \frac{-2}{t + 1}}{\frac{\frac{-1}{2} \cdot t}{t} - \frac{\color{blue}{\frac{1}{2} \cdot 1}}{t}}} \]
    3. associate-*r/N/A

      \[\leadsto 1 - \frac{1}{2 + \frac{-2 - \frac{-2}{t + 1}}{\frac{\frac{-1}{2} \cdot t}{t} - \color{blue}{\frac{1}{2} \cdot \frac{1}{t}}}} \]
    4. sub-negN/A

      \[\leadsto 1 - \frac{1}{2 + \frac{-2 - \frac{-2}{t + 1}}{\color{blue}{\frac{\frac{-1}{2} \cdot t}{t} + \left(\mathsf{neg}\left(\frac{1}{2} \cdot \frac{1}{t}\right)\right)}}} \]
    5. associate-/l*N/A

      \[\leadsto 1 - \frac{1}{2 + \frac{-2 - \frac{-2}{t + 1}}{\color{blue}{\frac{-1}{2} \cdot \frac{t}{t}} + \left(\mathsf{neg}\left(\frac{1}{2} \cdot \frac{1}{t}\right)\right)}} \]
    6. *-inversesN/A

      \[\leadsto 1 - \frac{1}{2 + \frac{-2 - \frac{-2}{t + 1}}{\frac{-1}{2} \cdot \color{blue}{1} + \left(\mathsf{neg}\left(\frac{1}{2} \cdot \frac{1}{t}\right)\right)}} \]
    7. metadata-evalN/A

      \[\leadsto 1 - \frac{1}{2 + \frac{-2 - \frac{-2}{t + 1}}{\color{blue}{\frac{-1}{2}} + \left(\mathsf{neg}\left(\frac{1}{2} \cdot \frac{1}{t}\right)\right)}} \]
    8. mul-1-negN/A

      \[\leadsto 1 - \frac{1}{2 + \frac{-2 - \frac{-2}{t + 1}}{\frac{-1}{2} + \color{blue}{-1 \cdot \left(\frac{1}{2} \cdot \frac{1}{t}\right)}}} \]
    9. +-lowering-+.f64N/A

      \[\leadsto 1 - \frac{1}{2 + \frac{-2 - \frac{-2}{t + 1}}{\color{blue}{\frac{-1}{2} + -1 \cdot \left(\frac{1}{2} \cdot \frac{1}{t}\right)}}} \]
    10. associate-*r/N/A

      \[\leadsto 1 - \frac{1}{2 + \frac{-2 - \frac{-2}{t + 1}}{\frac{-1}{2} + -1 \cdot \color{blue}{\frac{\frac{1}{2} \cdot 1}{t}}}} \]
    11. metadata-evalN/A

      \[\leadsto 1 - \frac{1}{2 + \frac{-2 - \frac{-2}{t + 1}}{\frac{-1}{2} + -1 \cdot \frac{\color{blue}{\frac{1}{2}}}{t}}} \]
    12. associate-*r/N/A

      \[\leadsto 1 - \frac{1}{2 + \frac{-2 - \frac{-2}{t + 1}}{\frac{-1}{2} + \color{blue}{\frac{-1 \cdot \frac{1}{2}}{t}}}} \]
    13. metadata-evalN/A

      \[\leadsto 1 - \frac{1}{2 + \frac{-2 - \frac{-2}{t + 1}}{\frac{-1}{2} + \frac{\color{blue}{\frac{-1}{2}}}{t}}} \]
    14. /-lowering-/.f64100.0

      \[\leadsto 1 - \frac{1}{2 + \frac{-2 - \frac{-2}{t + 1}}{-0.5 + \color{blue}{\frac{-0.5}{t}}}} \]
  9. Simplified100.0%

    \[\leadsto 1 - \frac{1}{2 + \frac{-2 - \frac{-2}{t + 1}}{\color{blue}{-0.5 + \frac{-0.5}{t}}}} \]
  10. Final simplification100.0%

    \[\leadsto 1 + \frac{1}{\frac{\frac{-2}{1 + t} - -2}{-0.5 + \frac{-0.5}{t}} - 2} \]
  11. Add Preprocessing

Alternative 2: 99.5% accurate, 0.8× speedup?

\[\begin{array}{l} \\ \begin{array}{l} t_1 := 2 + \frac{\frac{2}{t}}{-1 + \frac{-1}{t}}\\ \mathbf{if}\;t\_1 \cdot t\_1 \leq 0.2:\\ \;\;\;\;\mathsf{fma}\left(t, \mathsf{fma}\left(t \cdot t, -2 + t, t\right), 0.5\right)\\ \mathbf{else}:\\ \;\;\;\;0.8333333333333334 + \frac{\frac{0.037037037037037035 + \frac{0.04938271604938271}{t}}{t} + -0.2222222222222222}{t}\\ \end{array} \end{array} \]
(FPCore (t)
 :precision binary64
 (let* ((t_1 (+ 2.0 (/ (/ 2.0 t) (+ -1.0 (/ -1.0 t))))))
   (if (<= (* t_1 t_1) 0.2)
     (fma t (fma (* t t) (+ -2.0 t) t) 0.5)
     (+
      0.8333333333333334
      (/
       (+
        (/ (+ 0.037037037037037035 (/ 0.04938271604938271 t)) t)
        -0.2222222222222222)
       t)))))
double code(double t) {
	double t_1 = 2.0 + ((2.0 / t) / (-1.0 + (-1.0 / t)));
	double tmp;
	if ((t_1 * t_1) <= 0.2) {
		tmp = fma(t, fma((t * t), (-2.0 + t), t), 0.5);
	} else {
		tmp = 0.8333333333333334 + ((((0.037037037037037035 + (0.04938271604938271 / t)) / t) + -0.2222222222222222) / t);
	}
	return tmp;
}
function code(t)
	t_1 = Float64(2.0 + Float64(Float64(2.0 / t) / Float64(-1.0 + Float64(-1.0 / t))))
	tmp = 0.0
	if (Float64(t_1 * t_1) <= 0.2)
		tmp = fma(t, fma(Float64(t * t), Float64(-2.0 + t), t), 0.5);
	else
		tmp = Float64(0.8333333333333334 + Float64(Float64(Float64(Float64(0.037037037037037035 + Float64(0.04938271604938271 / t)) / t) + -0.2222222222222222) / t));
	end
	return tmp
end
code[t_] := Block[{t$95$1 = N[(2.0 + N[(N[(2.0 / t), $MachinePrecision] / N[(-1.0 + N[(-1.0 / t), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]}, If[LessEqual[N[(t$95$1 * t$95$1), $MachinePrecision], 0.2], N[(t * N[(N[(t * t), $MachinePrecision] * N[(-2.0 + t), $MachinePrecision] + t), $MachinePrecision] + 0.5), $MachinePrecision], N[(0.8333333333333334 + N[(N[(N[(N[(0.037037037037037035 + N[(0.04938271604938271 / t), $MachinePrecision]), $MachinePrecision] / t), $MachinePrecision] + -0.2222222222222222), $MachinePrecision] / t), $MachinePrecision]), $MachinePrecision]]]
\begin{array}{l}

\\
\begin{array}{l}
t_1 := 2 + \frac{\frac{2}{t}}{-1 + \frac{-1}{t}}\\
\mathbf{if}\;t\_1 \cdot t\_1 \leq 0.2:\\
\;\;\;\;\mathsf{fma}\left(t, \mathsf{fma}\left(t \cdot t, -2 + t, t\right), 0.5\right)\\

\mathbf{else}:\\
\;\;\;\;0.8333333333333334 + \frac{\frac{0.037037037037037035 + \frac{0.04938271604938271}{t}}{t} + -0.2222222222222222}{t}\\


\end{array}
\end{array}
Derivation
  1. Split input into 2 regimes
  2. if (*.f64 (-.f64 #s(literal 2 binary64) (/.f64 (/.f64 #s(literal 2 binary64) t) (+.f64 #s(literal 1 binary64) (/.f64 #s(literal 1 binary64) t)))) (-.f64 #s(literal 2 binary64) (/.f64 (/.f64 #s(literal 2 binary64) t) (+.f64 #s(literal 1 binary64) (/.f64 #s(literal 1 binary64) t))))) < 0.20000000000000001

    1. Initial program 100.0%

      \[1 - \frac{1}{2 + \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right) \cdot \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right)} \]
    2. Add Preprocessing
    3. Taylor expanded in t around 0

      \[\leadsto \color{blue}{\frac{1}{2} + {t}^{2} \cdot \left(1 + t \cdot \left(t - 2\right)\right)} \]
    4. Step-by-step derivation
      1. +-commutativeN/A

        \[\leadsto \color{blue}{{t}^{2} \cdot \left(1 + t \cdot \left(t - 2\right)\right) + \frac{1}{2}} \]
      2. unpow2N/A

        \[\leadsto \color{blue}{\left(t \cdot t\right)} \cdot \left(1 + t \cdot \left(t - 2\right)\right) + \frac{1}{2} \]
      3. associate-*l*N/A

        \[\leadsto \color{blue}{t \cdot \left(t \cdot \left(1 + t \cdot \left(t - 2\right)\right)\right)} + \frac{1}{2} \]
      4. accelerator-lowering-fma.f64N/A

        \[\leadsto \color{blue}{\mathsf{fma}\left(t, t \cdot \left(1 + t \cdot \left(t - 2\right)\right), \frac{1}{2}\right)} \]
      5. +-commutativeN/A

        \[\leadsto \mathsf{fma}\left(t, t \cdot \color{blue}{\left(t \cdot \left(t - 2\right) + 1\right)}, \frac{1}{2}\right) \]
      6. distribute-lft-inN/A

        \[\leadsto \mathsf{fma}\left(t, \color{blue}{t \cdot \left(t \cdot \left(t - 2\right)\right) + t \cdot 1}, \frac{1}{2}\right) \]
      7. associate-*r*N/A

        \[\leadsto \mathsf{fma}\left(t, \color{blue}{\left(t \cdot t\right) \cdot \left(t - 2\right)} + t \cdot 1, \frac{1}{2}\right) \]
      8. unpow2N/A

        \[\leadsto \mathsf{fma}\left(t, \color{blue}{{t}^{2}} \cdot \left(t - 2\right) + t \cdot 1, \frac{1}{2}\right) \]
      9. *-rgt-identityN/A

        \[\leadsto \mathsf{fma}\left(t, {t}^{2} \cdot \left(t - 2\right) + \color{blue}{t}, \frac{1}{2}\right) \]
      10. accelerator-lowering-fma.f64N/A

        \[\leadsto \mathsf{fma}\left(t, \color{blue}{\mathsf{fma}\left({t}^{2}, t - 2, t\right)}, \frac{1}{2}\right) \]
      11. unpow2N/A

        \[\leadsto \mathsf{fma}\left(t, \mathsf{fma}\left(\color{blue}{t \cdot t}, t - 2, t\right), \frac{1}{2}\right) \]
      12. *-lowering-*.f64N/A

        \[\leadsto \mathsf{fma}\left(t, \mathsf{fma}\left(\color{blue}{t \cdot t}, t - 2, t\right), \frac{1}{2}\right) \]
      13. sub-negN/A

        \[\leadsto \mathsf{fma}\left(t, \mathsf{fma}\left(t \cdot t, \color{blue}{t + \left(\mathsf{neg}\left(2\right)\right)}, t\right), \frac{1}{2}\right) \]
      14. metadata-evalN/A

        \[\leadsto \mathsf{fma}\left(t, \mathsf{fma}\left(t \cdot t, t + \color{blue}{-2}, t\right), \frac{1}{2}\right) \]
      15. +-commutativeN/A

        \[\leadsto \mathsf{fma}\left(t, \mathsf{fma}\left(t \cdot t, \color{blue}{-2 + t}, t\right), \frac{1}{2}\right) \]
      16. +-lowering-+.f6499.5

        \[\leadsto \mathsf{fma}\left(t, \mathsf{fma}\left(t \cdot t, \color{blue}{-2 + t}, t\right), 0.5\right) \]
    5. Simplified99.5%

      \[\leadsto \color{blue}{\mathsf{fma}\left(t, \mathsf{fma}\left(t \cdot t, -2 + t, t\right), 0.5\right)} \]

    if 0.20000000000000001 < (*.f64 (-.f64 #s(literal 2 binary64) (/.f64 (/.f64 #s(literal 2 binary64) t) (+.f64 #s(literal 1 binary64) (/.f64 #s(literal 1 binary64) t)))) (-.f64 #s(literal 2 binary64) (/.f64 (/.f64 #s(literal 2 binary64) t) (+.f64 #s(literal 1 binary64) (/.f64 #s(literal 1 binary64) t)))))

    1. Initial program 100.0%

      \[1 - \frac{1}{2 + \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right) \cdot \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right)} \]
    2. Add Preprocessing
    3. Taylor expanded in t around inf

      \[\leadsto \color{blue}{\left(\frac{5}{6} + \left(\frac{\frac{1}{27}}{{t}^{2}} + \frac{4}{81} \cdot \frac{1}{{t}^{3}}\right)\right) - \frac{2}{9} \cdot \frac{1}{t}} \]
    4. Simplified99.2%

      \[\leadsto \color{blue}{0.8333333333333334 + \frac{\frac{0.037037037037037035 + \frac{0.04938271604938271}{t}}{t} + -0.2222222222222222}{t}} \]
  3. Recombined 2 regimes into one program.
  4. Final simplification99.3%

    \[\leadsto \begin{array}{l} \mathbf{if}\;\left(2 + \frac{\frac{2}{t}}{-1 + \frac{-1}{t}}\right) \cdot \left(2 + \frac{\frac{2}{t}}{-1 + \frac{-1}{t}}\right) \leq 0.2:\\ \;\;\;\;\mathsf{fma}\left(t, \mathsf{fma}\left(t \cdot t, -2 + t, t\right), 0.5\right)\\ \mathbf{else}:\\ \;\;\;\;0.8333333333333334 + \frac{\frac{0.037037037037037035 + \frac{0.04938271604938271}{t}}{t} + -0.2222222222222222}{t}\\ \end{array} \]
  5. Add Preprocessing

Alternative 3: 99.4% accurate, 0.9× speedup?

\[\begin{array}{l} \\ \begin{array}{l} t_1 := 2 + \frac{\frac{2}{t}}{-1 + \frac{-1}{t}}\\ \mathbf{if}\;t\_1 \cdot t\_1 \leq 0.2:\\ \;\;\;\;\mathsf{fma}\left(t, \mathsf{fma}\left(t \cdot t, -2 + t, t\right), 0.5\right)\\ \mathbf{else}:\\ \;\;\;\;0.8333333333333334 + \frac{\mathsf{fma}\left(t, -0.2222222222222222, 0.037037037037037035\right)}{t \cdot t}\\ \end{array} \end{array} \]
(FPCore (t)
 :precision binary64
 (let* ((t_1 (+ 2.0 (/ (/ 2.0 t) (+ -1.0 (/ -1.0 t))))))
   (if (<= (* t_1 t_1) 0.2)
     (fma t (fma (* t t) (+ -2.0 t) t) 0.5)
     (+
      0.8333333333333334
      (/ (fma t -0.2222222222222222 0.037037037037037035) (* t t))))))
double code(double t) {
	double t_1 = 2.0 + ((2.0 / t) / (-1.0 + (-1.0 / t)));
	double tmp;
	if ((t_1 * t_1) <= 0.2) {
		tmp = fma(t, fma((t * t), (-2.0 + t), t), 0.5);
	} else {
		tmp = 0.8333333333333334 + (fma(t, -0.2222222222222222, 0.037037037037037035) / (t * t));
	}
	return tmp;
}
function code(t)
	t_1 = Float64(2.0 + Float64(Float64(2.0 / t) / Float64(-1.0 + Float64(-1.0 / t))))
	tmp = 0.0
	if (Float64(t_1 * t_1) <= 0.2)
		tmp = fma(t, fma(Float64(t * t), Float64(-2.0 + t), t), 0.5);
	else
		tmp = Float64(0.8333333333333334 + Float64(fma(t, -0.2222222222222222, 0.037037037037037035) / Float64(t * t)));
	end
	return tmp
end
code[t_] := Block[{t$95$1 = N[(2.0 + N[(N[(2.0 / t), $MachinePrecision] / N[(-1.0 + N[(-1.0 / t), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]}, If[LessEqual[N[(t$95$1 * t$95$1), $MachinePrecision], 0.2], N[(t * N[(N[(t * t), $MachinePrecision] * N[(-2.0 + t), $MachinePrecision] + t), $MachinePrecision] + 0.5), $MachinePrecision], N[(0.8333333333333334 + N[(N[(t * -0.2222222222222222 + 0.037037037037037035), $MachinePrecision] / N[(t * t), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]]]
\begin{array}{l}

\\
\begin{array}{l}
t_1 := 2 + \frac{\frac{2}{t}}{-1 + \frac{-1}{t}}\\
\mathbf{if}\;t\_1 \cdot t\_1 \leq 0.2:\\
\;\;\;\;\mathsf{fma}\left(t, \mathsf{fma}\left(t \cdot t, -2 + t, t\right), 0.5\right)\\

\mathbf{else}:\\
\;\;\;\;0.8333333333333334 + \frac{\mathsf{fma}\left(t, -0.2222222222222222, 0.037037037037037035\right)}{t \cdot t}\\


\end{array}
\end{array}
Derivation
  1. Split input into 2 regimes
  2. if (*.f64 (-.f64 #s(literal 2 binary64) (/.f64 (/.f64 #s(literal 2 binary64) t) (+.f64 #s(literal 1 binary64) (/.f64 #s(literal 1 binary64) t)))) (-.f64 #s(literal 2 binary64) (/.f64 (/.f64 #s(literal 2 binary64) t) (+.f64 #s(literal 1 binary64) (/.f64 #s(literal 1 binary64) t))))) < 0.20000000000000001

    1. Initial program 100.0%

      \[1 - \frac{1}{2 + \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right) \cdot \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right)} \]
    2. Add Preprocessing
    3. Taylor expanded in t around 0

      \[\leadsto \color{blue}{\frac{1}{2} + {t}^{2} \cdot \left(1 + t \cdot \left(t - 2\right)\right)} \]
    4. Step-by-step derivation
      1. +-commutativeN/A

        \[\leadsto \color{blue}{{t}^{2} \cdot \left(1 + t \cdot \left(t - 2\right)\right) + \frac{1}{2}} \]
      2. unpow2N/A

        \[\leadsto \color{blue}{\left(t \cdot t\right)} \cdot \left(1 + t \cdot \left(t - 2\right)\right) + \frac{1}{2} \]
      3. associate-*l*N/A

        \[\leadsto \color{blue}{t \cdot \left(t \cdot \left(1 + t \cdot \left(t - 2\right)\right)\right)} + \frac{1}{2} \]
      4. accelerator-lowering-fma.f64N/A

        \[\leadsto \color{blue}{\mathsf{fma}\left(t, t \cdot \left(1 + t \cdot \left(t - 2\right)\right), \frac{1}{2}\right)} \]
      5. +-commutativeN/A

        \[\leadsto \mathsf{fma}\left(t, t \cdot \color{blue}{\left(t \cdot \left(t - 2\right) + 1\right)}, \frac{1}{2}\right) \]
      6. distribute-lft-inN/A

        \[\leadsto \mathsf{fma}\left(t, \color{blue}{t \cdot \left(t \cdot \left(t - 2\right)\right) + t \cdot 1}, \frac{1}{2}\right) \]
      7. associate-*r*N/A

        \[\leadsto \mathsf{fma}\left(t, \color{blue}{\left(t \cdot t\right) \cdot \left(t - 2\right)} + t \cdot 1, \frac{1}{2}\right) \]
      8. unpow2N/A

        \[\leadsto \mathsf{fma}\left(t, \color{blue}{{t}^{2}} \cdot \left(t - 2\right) + t \cdot 1, \frac{1}{2}\right) \]
      9. *-rgt-identityN/A

        \[\leadsto \mathsf{fma}\left(t, {t}^{2} \cdot \left(t - 2\right) + \color{blue}{t}, \frac{1}{2}\right) \]
      10. accelerator-lowering-fma.f64N/A

        \[\leadsto \mathsf{fma}\left(t, \color{blue}{\mathsf{fma}\left({t}^{2}, t - 2, t\right)}, \frac{1}{2}\right) \]
      11. unpow2N/A

        \[\leadsto \mathsf{fma}\left(t, \mathsf{fma}\left(\color{blue}{t \cdot t}, t - 2, t\right), \frac{1}{2}\right) \]
      12. *-lowering-*.f64N/A

        \[\leadsto \mathsf{fma}\left(t, \mathsf{fma}\left(\color{blue}{t \cdot t}, t - 2, t\right), \frac{1}{2}\right) \]
      13. sub-negN/A

        \[\leadsto \mathsf{fma}\left(t, \mathsf{fma}\left(t \cdot t, \color{blue}{t + \left(\mathsf{neg}\left(2\right)\right)}, t\right), \frac{1}{2}\right) \]
      14. metadata-evalN/A

        \[\leadsto \mathsf{fma}\left(t, \mathsf{fma}\left(t \cdot t, t + \color{blue}{-2}, t\right), \frac{1}{2}\right) \]
      15. +-commutativeN/A

        \[\leadsto \mathsf{fma}\left(t, \mathsf{fma}\left(t \cdot t, \color{blue}{-2 + t}, t\right), \frac{1}{2}\right) \]
      16. +-lowering-+.f6499.5

        \[\leadsto \mathsf{fma}\left(t, \mathsf{fma}\left(t \cdot t, \color{blue}{-2 + t}, t\right), 0.5\right) \]
    5. Simplified99.5%

      \[\leadsto \color{blue}{\mathsf{fma}\left(t, \mathsf{fma}\left(t \cdot t, -2 + t, t\right), 0.5\right)} \]

    if 0.20000000000000001 < (*.f64 (-.f64 #s(literal 2 binary64) (/.f64 (/.f64 #s(literal 2 binary64) t) (+.f64 #s(literal 1 binary64) (/.f64 #s(literal 1 binary64) t)))) (-.f64 #s(literal 2 binary64) (/.f64 (/.f64 #s(literal 2 binary64) t) (+.f64 #s(literal 1 binary64) (/.f64 #s(literal 1 binary64) t)))))

    1. Initial program 100.0%

      \[1 - \frac{1}{2 + \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right) \cdot \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right)} \]
    2. Add Preprocessing
    3. Taylor expanded in t around inf

      \[\leadsto \color{blue}{\left(\frac{5}{6} + \frac{\frac{1}{27}}{{t}^{2}}\right) - \frac{2}{9} \cdot \frac{1}{t}} \]
    4. Step-by-step derivation
      1. associate--l+N/A

        \[\leadsto \color{blue}{\frac{5}{6} + \left(\frac{\frac{1}{27}}{{t}^{2}} - \frac{2}{9} \cdot \frac{1}{t}\right)} \]
      2. unpow2N/A

        \[\leadsto \frac{5}{6} + \left(\frac{\frac{1}{27}}{\color{blue}{t \cdot t}} - \frac{2}{9} \cdot \frac{1}{t}\right) \]
      3. associate-/r*N/A

        \[\leadsto \frac{5}{6} + \left(\color{blue}{\frac{\frac{\frac{1}{27}}{t}}{t}} - \frac{2}{9} \cdot \frac{1}{t}\right) \]
      4. metadata-evalN/A

        \[\leadsto \frac{5}{6} + \left(\frac{\frac{\color{blue}{\frac{1}{27} \cdot 1}}{t}}{t} - \frac{2}{9} \cdot \frac{1}{t}\right) \]
      5. associate-*r/N/A

        \[\leadsto \frac{5}{6} + \left(\frac{\color{blue}{\frac{1}{27} \cdot \frac{1}{t}}}{t} - \frac{2}{9} \cdot \frac{1}{t}\right) \]
      6. associate-*r/N/A

        \[\leadsto \frac{5}{6} + \left(\frac{\frac{1}{27} \cdot \frac{1}{t}}{t} - \color{blue}{\frac{\frac{2}{9} \cdot 1}{t}}\right) \]
      7. metadata-evalN/A

        \[\leadsto \frac{5}{6} + \left(\frac{\frac{1}{27} \cdot \frac{1}{t}}{t} - \frac{\color{blue}{\frac{2}{9}}}{t}\right) \]
      8. div-subN/A

        \[\leadsto \frac{5}{6} + \color{blue}{\frac{\frac{1}{27} \cdot \frac{1}{t} - \frac{2}{9}}{t}} \]
      9. +-lowering-+.f64N/A

        \[\leadsto \color{blue}{\frac{5}{6} + \frac{\frac{1}{27} \cdot \frac{1}{t} - \frac{2}{9}}{t}} \]
      10. /-lowering-/.f64N/A

        \[\leadsto \frac{5}{6} + \color{blue}{\frac{\frac{1}{27} \cdot \frac{1}{t} - \frac{2}{9}}{t}} \]
      11. sub-negN/A

        \[\leadsto \frac{5}{6} + \frac{\color{blue}{\frac{1}{27} \cdot \frac{1}{t} + \left(\mathsf{neg}\left(\frac{2}{9}\right)\right)}}{t} \]
      12. +-lowering-+.f64N/A

        \[\leadsto \frac{5}{6} + \frac{\color{blue}{\frac{1}{27} \cdot \frac{1}{t} + \left(\mathsf{neg}\left(\frac{2}{9}\right)\right)}}{t} \]
      13. associate-*r/N/A

        \[\leadsto \frac{5}{6} + \frac{\color{blue}{\frac{\frac{1}{27} \cdot 1}{t}} + \left(\mathsf{neg}\left(\frac{2}{9}\right)\right)}{t} \]
      14. metadata-evalN/A

        \[\leadsto \frac{5}{6} + \frac{\frac{\color{blue}{\frac{1}{27}}}{t} + \left(\mathsf{neg}\left(\frac{2}{9}\right)\right)}{t} \]
      15. /-lowering-/.f64N/A

        \[\leadsto \frac{5}{6} + \frac{\color{blue}{\frac{\frac{1}{27}}{t}} + \left(\mathsf{neg}\left(\frac{2}{9}\right)\right)}{t} \]
      16. metadata-eval99.1

        \[\leadsto 0.8333333333333334 + \frac{\frac{0.037037037037037035}{t} + \color{blue}{-0.2222222222222222}}{t} \]
    5. Simplified99.1%

      \[\leadsto \color{blue}{0.8333333333333334 + \frac{\frac{0.037037037037037035}{t} + -0.2222222222222222}{t}} \]
    6. Taylor expanded in t around 0

      \[\leadsto \frac{5}{6} + \color{blue}{\frac{\frac{1}{27} + \frac{-2}{9} \cdot t}{{t}^{2}}} \]
    7. Step-by-step derivation
      1. /-lowering-/.f64N/A

        \[\leadsto \frac{5}{6} + \color{blue}{\frac{\frac{1}{27} + \frac{-2}{9} \cdot t}{{t}^{2}}} \]
      2. +-commutativeN/A

        \[\leadsto \frac{5}{6} + \frac{\color{blue}{\frac{-2}{9} \cdot t + \frac{1}{27}}}{{t}^{2}} \]
      3. *-commutativeN/A

        \[\leadsto \frac{5}{6} + \frac{\color{blue}{t \cdot \frac{-2}{9}} + \frac{1}{27}}{{t}^{2}} \]
      4. accelerator-lowering-fma.f64N/A

        \[\leadsto \frac{5}{6} + \frac{\color{blue}{\mathsf{fma}\left(t, \frac{-2}{9}, \frac{1}{27}\right)}}{{t}^{2}} \]
      5. unpow2N/A

        \[\leadsto \frac{5}{6} + \frac{\mathsf{fma}\left(t, \frac{-2}{9}, \frac{1}{27}\right)}{\color{blue}{t \cdot t}} \]
      6. *-lowering-*.f6499.1

        \[\leadsto 0.8333333333333334 + \frac{\mathsf{fma}\left(t, -0.2222222222222222, 0.037037037037037035\right)}{\color{blue}{t \cdot t}} \]
    8. Simplified99.1%

      \[\leadsto 0.8333333333333334 + \color{blue}{\frac{\mathsf{fma}\left(t, -0.2222222222222222, 0.037037037037037035\right)}{t \cdot t}} \]
  3. Recombined 2 regimes into one program.
  4. Final simplification99.3%

    \[\leadsto \begin{array}{l} \mathbf{if}\;\left(2 + \frac{\frac{2}{t}}{-1 + \frac{-1}{t}}\right) \cdot \left(2 + \frac{\frac{2}{t}}{-1 + \frac{-1}{t}}\right) \leq 0.2:\\ \;\;\;\;\mathsf{fma}\left(t, \mathsf{fma}\left(t \cdot t, -2 + t, t\right), 0.5\right)\\ \mathbf{else}:\\ \;\;\;\;0.8333333333333334 + \frac{\mathsf{fma}\left(t, -0.2222222222222222, 0.037037037037037035\right)}{t \cdot t}\\ \end{array} \]
  5. Add Preprocessing

Alternative 4: 99.3% accurate, 0.9× speedup?

\[\begin{array}{l} \\ \begin{array}{l} t_1 := 2 + \frac{\frac{2}{t}}{-1 + \frac{-1}{t}}\\ \mathbf{if}\;t\_1 \cdot t\_1 \leq 0.2:\\ \;\;\;\;\mathsf{fma}\left(t, \mathsf{fma}\left(t \cdot t, -2 + t, t\right), 0.5\right)\\ \mathbf{else}:\\ \;\;\;\;0.8333333333333334 + \frac{-0.2222222222222222}{t}\\ \end{array} \end{array} \]
(FPCore (t)
 :precision binary64
 (let* ((t_1 (+ 2.0 (/ (/ 2.0 t) (+ -1.0 (/ -1.0 t))))))
   (if (<= (* t_1 t_1) 0.2)
     (fma t (fma (* t t) (+ -2.0 t) t) 0.5)
     (+ 0.8333333333333334 (/ -0.2222222222222222 t)))))
double code(double t) {
	double t_1 = 2.0 + ((2.0 / t) / (-1.0 + (-1.0 / t)));
	double tmp;
	if ((t_1 * t_1) <= 0.2) {
		tmp = fma(t, fma((t * t), (-2.0 + t), t), 0.5);
	} else {
		tmp = 0.8333333333333334 + (-0.2222222222222222 / t);
	}
	return tmp;
}
function code(t)
	t_1 = Float64(2.0 + Float64(Float64(2.0 / t) / Float64(-1.0 + Float64(-1.0 / t))))
	tmp = 0.0
	if (Float64(t_1 * t_1) <= 0.2)
		tmp = fma(t, fma(Float64(t * t), Float64(-2.0 + t), t), 0.5);
	else
		tmp = Float64(0.8333333333333334 + Float64(-0.2222222222222222 / t));
	end
	return tmp
end
code[t_] := Block[{t$95$1 = N[(2.0 + N[(N[(2.0 / t), $MachinePrecision] / N[(-1.0 + N[(-1.0 / t), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]}, If[LessEqual[N[(t$95$1 * t$95$1), $MachinePrecision], 0.2], N[(t * N[(N[(t * t), $MachinePrecision] * N[(-2.0 + t), $MachinePrecision] + t), $MachinePrecision] + 0.5), $MachinePrecision], N[(0.8333333333333334 + N[(-0.2222222222222222 / t), $MachinePrecision]), $MachinePrecision]]]
\begin{array}{l}

\\
\begin{array}{l}
t_1 := 2 + \frac{\frac{2}{t}}{-1 + \frac{-1}{t}}\\
\mathbf{if}\;t\_1 \cdot t\_1 \leq 0.2:\\
\;\;\;\;\mathsf{fma}\left(t, \mathsf{fma}\left(t \cdot t, -2 + t, t\right), 0.5\right)\\

\mathbf{else}:\\
\;\;\;\;0.8333333333333334 + \frac{-0.2222222222222222}{t}\\


\end{array}
\end{array}
Derivation
  1. Split input into 2 regimes
  2. if (*.f64 (-.f64 #s(literal 2 binary64) (/.f64 (/.f64 #s(literal 2 binary64) t) (+.f64 #s(literal 1 binary64) (/.f64 #s(literal 1 binary64) t)))) (-.f64 #s(literal 2 binary64) (/.f64 (/.f64 #s(literal 2 binary64) t) (+.f64 #s(literal 1 binary64) (/.f64 #s(literal 1 binary64) t))))) < 0.20000000000000001

    1. Initial program 100.0%

      \[1 - \frac{1}{2 + \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right) \cdot \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right)} \]
    2. Add Preprocessing
    3. Taylor expanded in t around 0

      \[\leadsto \color{blue}{\frac{1}{2} + {t}^{2} \cdot \left(1 + t \cdot \left(t - 2\right)\right)} \]
    4. Step-by-step derivation
      1. +-commutativeN/A

        \[\leadsto \color{blue}{{t}^{2} \cdot \left(1 + t \cdot \left(t - 2\right)\right) + \frac{1}{2}} \]
      2. unpow2N/A

        \[\leadsto \color{blue}{\left(t \cdot t\right)} \cdot \left(1 + t \cdot \left(t - 2\right)\right) + \frac{1}{2} \]
      3. associate-*l*N/A

        \[\leadsto \color{blue}{t \cdot \left(t \cdot \left(1 + t \cdot \left(t - 2\right)\right)\right)} + \frac{1}{2} \]
      4. accelerator-lowering-fma.f64N/A

        \[\leadsto \color{blue}{\mathsf{fma}\left(t, t \cdot \left(1 + t \cdot \left(t - 2\right)\right), \frac{1}{2}\right)} \]
      5. +-commutativeN/A

        \[\leadsto \mathsf{fma}\left(t, t \cdot \color{blue}{\left(t \cdot \left(t - 2\right) + 1\right)}, \frac{1}{2}\right) \]
      6. distribute-lft-inN/A

        \[\leadsto \mathsf{fma}\left(t, \color{blue}{t \cdot \left(t \cdot \left(t - 2\right)\right) + t \cdot 1}, \frac{1}{2}\right) \]
      7. associate-*r*N/A

        \[\leadsto \mathsf{fma}\left(t, \color{blue}{\left(t \cdot t\right) \cdot \left(t - 2\right)} + t \cdot 1, \frac{1}{2}\right) \]
      8. unpow2N/A

        \[\leadsto \mathsf{fma}\left(t, \color{blue}{{t}^{2}} \cdot \left(t - 2\right) + t \cdot 1, \frac{1}{2}\right) \]
      9. *-rgt-identityN/A

        \[\leadsto \mathsf{fma}\left(t, {t}^{2} \cdot \left(t - 2\right) + \color{blue}{t}, \frac{1}{2}\right) \]
      10. accelerator-lowering-fma.f64N/A

        \[\leadsto \mathsf{fma}\left(t, \color{blue}{\mathsf{fma}\left({t}^{2}, t - 2, t\right)}, \frac{1}{2}\right) \]
      11. unpow2N/A

        \[\leadsto \mathsf{fma}\left(t, \mathsf{fma}\left(\color{blue}{t \cdot t}, t - 2, t\right), \frac{1}{2}\right) \]
      12. *-lowering-*.f64N/A

        \[\leadsto \mathsf{fma}\left(t, \mathsf{fma}\left(\color{blue}{t \cdot t}, t - 2, t\right), \frac{1}{2}\right) \]
      13. sub-negN/A

        \[\leadsto \mathsf{fma}\left(t, \mathsf{fma}\left(t \cdot t, \color{blue}{t + \left(\mathsf{neg}\left(2\right)\right)}, t\right), \frac{1}{2}\right) \]
      14. metadata-evalN/A

        \[\leadsto \mathsf{fma}\left(t, \mathsf{fma}\left(t \cdot t, t + \color{blue}{-2}, t\right), \frac{1}{2}\right) \]
      15. +-commutativeN/A

        \[\leadsto \mathsf{fma}\left(t, \mathsf{fma}\left(t \cdot t, \color{blue}{-2 + t}, t\right), \frac{1}{2}\right) \]
      16. +-lowering-+.f6499.5

        \[\leadsto \mathsf{fma}\left(t, \mathsf{fma}\left(t \cdot t, \color{blue}{-2 + t}, t\right), 0.5\right) \]
    5. Simplified99.5%

      \[\leadsto \color{blue}{\mathsf{fma}\left(t, \mathsf{fma}\left(t \cdot t, -2 + t, t\right), 0.5\right)} \]

    if 0.20000000000000001 < (*.f64 (-.f64 #s(literal 2 binary64) (/.f64 (/.f64 #s(literal 2 binary64) t) (+.f64 #s(literal 1 binary64) (/.f64 #s(literal 1 binary64) t)))) (-.f64 #s(literal 2 binary64) (/.f64 (/.f64 #s(literal 2 binary64) t) (+.f64 #s(literal 1 binary64) (/.f64 #s(literal 1 binary64) t)))))

    1. Initial program 100.0%

      \[1 - \frac{1}{2 + \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right) \cdot \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right)} \]
    2. Add Preprocessing
    3. Taylor expanded in t around inf

      \[\leadsto \color{blue}{\frac{5}{6} - \frac{2}{9} \cdot \frac{1}{t}} \]
    4. Step-by-step derivation
      1. sub-negN/A

        \[\leadsto \color{blue}{\frac{5}{6} + \left(\mathsf{neg}\left(\frac{2}{9} \cdot \frac{1}{t}\right)\right)} \]
      2. +-lowering-+.f64N/A

        \[\leadsto \color{blue}{\frac{5}{6} + \left(\mathsf{neg}\left(\frac{2}{9} \cdot \frac{1}{t}\right)\right)} \]
      3. associate-*r/N/A

        \[\leadsto \frac{5}{6} + \left(\mathsf{neg}\left(\color{blue}{\frac{\frac{2}{9} \cdot 1}{t}}\right)\right) \]
      4. metadata-evalN/A

        \[\leadsto \frac{5}{6} + \left(\mathsf{neg}\left(\frac{\color{blue}{\frac{2}{9}}}{t}\right)\right) \]
      5. distribute-neg-fracN/A

        \[\leadsto \frac{5}{6} + \color{blue}{\frac{\mathsf{neg}\left(\frac{2}{9}\right)}{t}} \]
      6. /-lowering-/.f64N/A

        \[\leadsto \frac{5}{6} + \color{blue}{\frac{\mathsf{neg}\left(\frac{2}{9}\right)}{t}} \]
      7. metadata-eval99.0

        \[\leadsto 0.8333333333333334 + \frac{\color{blue}{-0.2222222222222222}}{t} \]
    5. Simplified99.0%

      \[\leadsto \color{blue}{0.8333333333333334 + \frac{-0.2222222222222222}{t}} \]
  3. Recombined 2 regimes into one program.
  4. Final simplification99.2%

    \[\leadsto \begin{array}{l} \mathbf{if}\;\left(2 + \frac{\frac{2}{t}}{-1 + \frac{-1}{t}}\right) \cdot \left(2 + \frac{\frac{2}{t}}{-1 + \frac{-1}{t}}\right) \leq 0.2:\\ \;\;\;\;\mathsf{fma}\left(t, \mathsf{fma}\left(t \cdot t, -2 + t, t\right), 0.5\right)\\ \mathbf{else}:\\ \;\;\;\;0.8333333333333334 + \frac{-0.2222222222222222}{t}\\ \end{array} \]
  5. Add Preprocessing

Alternative 5: 99.2% accurate, 0.9× speedup?

\[\begin{array}{l} \\ \begin{array}{l} t_1 := 2 + \frac{\frac{2}{t}}{-1 + \frac{-1}{t}}\\ \mathbf{if}\;t\_1 \cdot t\_1 \leq 0.2:\\ \;\;\;\;\mathsf{fma}\left(t, \mathsf{fma}\left(-2, t \cdot t, t\right), 0.5\right)\\ \mathbf{else}:\\ \;\;\;\;0.8333333333333334 + \frac{-0.2222222222222222}{t}\\ \end{array} \end{array} \]
(FPCore (t)
 :precision binary64
 (let* ((t_1 (+ 2.0 (/ (/ 2.0 t) (+ -1.0 (/ -1.0 t))))))
   (if (<= (* t_1 t_1) 0.2)
     (fma t (fma -2.0 (* t t) t) 0.5)
     (+ 0.8333333333333334 (/ -0.2222222222222222 t)))))
double code(double t) {
	double t_1 = 2.0 + ((2.0 / t) / (-1.0 + (-1.0 / t)));
	double tmp;
	if ((t_1 * t_1) <= 0.2) {
		tmp = fma(t, fma(-2.0, (t * t), t), 0.5);
	} else {
		tmp = 0.8333333333333334 + (-0.2222222222222222 / t);
	}
	return tmp;
}
function code(t)
	t_1 = Float64(2.0 + Float64(Float64(2.0 / t) / Float64(-1.0 + Float64(-1.0 / t))))
	tmp = 0.0
	if (Float64(t_1 * t_1) <= 0.2)
		tmp = fma(t, fma(-2.0, Float64(t * t), t), 0.5);
	else
		tmp = Float64(0.8333333333333334 + Float64(-0.2222222222222222 / t));
	end
	return tmp
end
code[t_] := Block[{t$95$1 = N[(2.0 + N[(N[(2.0 / t), $MachinePrecision] / N[(-1.0 + N[(-1.0 / t), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]}, If[LessEqual[N[(t$95$1 * t$95$1), $MachinePrecision], 0.2], N[(t * N[(-2.0 * N[(t * t), $MachinePrecision] + t), $MachinePrecision] + 0.5), $MachinePrecision], N[(0.8333333333333334 + N[(-0.2222222222222222 / t), $MachinePrecision]), $MachinePrecision]]]
\begin{array}{l}

\\
\begin{array}{l}
t_1 := 2 + \frac{\frac{2}{t}}{-1 + \frac{-1}{t}}\\
\mathbf{if}\;t\_1 \cdot t\_1 \leq 0.2:\\
\;\;\;\;\mathsf{fma}\left(t, \mathsf{fma}\left(-2, t \cdot t, t\right), 0.5\right)\\

\mathbf{else}:\\
\;\;\;\;0.8333333333333334 + \frac{-0.2222222222222222}{t}\\


\end{array}
\end{array}
Derivation
  1. Split input into 2 regimes
  2. if (*.f64 (-.f64 #s(literal 2 binary64) (/.f64 (/.f64 #s(literal 2 binary64) t) (+.f64 #s(literal 1 binary64) (/.f64 #s(literal 1 binary64) t)))) (-.f64 #s(literal 2 binary64) (/.f64 (/.f64 #s(literal 2 binary64) t) (+.f64 #s(literal 1 binary64) (/.f64 #s(literal 1 binary64) t))))) < 0.20000000000000001

    1. Initial program 100.0%

      \[1 - \frac{1}{2 + \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right) \cdot \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right)} \]
    2. Add Preprocessing
    3. Taylor expanded in t around 0

      \[\leadsto \color{blue}{\frac{1}{2} + {t}^{2} \cdot \left(1 + -2 \cdot t\right)} \]
    4. Step-by-step derivation
      1. +-commutativeN/A

        \[\leadsto \color{blue}{{t}^{2} \cdot \left(1 + -2 \cdot t\right) + \frac{1}{2}} \]
      2. unpow2N/A

        \[\leadsto \color{blue}{\left(t \cdot t\right)} \cdot \left(1 + -2 \cdot t\right) + \frac{1}{2} \]
      3. associate-*l*N/A

        \[\leadsto \color{blue}{t \cdot \left(t \cdot \left(1 + -2 \cdot t\right)\right)} + \frac{1}{2} \]
      4. *-commutativeN/A

        \[\leadsto t \cdot \color{blue}{\left(\left(1 + -2 \cdot t\right) \cdot t\right)} + \frac{1}{2} \]
      5. accelerator-lowering-fma.f64N/A

        \[\leadsto \color{blue}{\mathsf{fma}\left(t, \left(1 + -2 \cdot t\right) \cdot t, \frac{1}{2}\right)} \]
      6. +-commutativeN/A

        \[\leadsto \mathsf{fma}\left(t, \color{blue}{\left(-2 \cdot t + 1\right)} \cdot t, \frac{1}{2}\right) \]
      7. distribute-lft1-inN/A

        \[\leadsto \mathsf{fma}\left(t, \color{blue}{\left(-2 \cdot t\right) \cdot t + t}, \frac{1}{2}\right) \]
      8. associate-*l*N/A

        \[\leadsto \mathsf{fma}\left(t, \color{blue}{-2 \cdot \left(t \cdot t\right)} + t, \frac{1}{2}\right) \]
      9. unpow2N/A

        \[\leadsto \mathsf{fma}\left(t, -2 \cdot \color{blue}{{t}^{2}} + t, \frac{1}{2}\right) \]
      10. accelerator-lowering-fma.f64N/A

        \[\leadsto \mathsf{fma}\left(t, \color{blue}{\mathsf{fma}\left(-2, {t}^{2}, t\right)}, \frac{1}{2}\right) \]
      11. unpow2N/A

        \[\leadsto \mathsf{fma}\left(t, \mathsf{fma}\left(-2, \color{blue}{t \cdot t}, t\right), \frac{1}{2}\right) \]
      12. *-lowering-*.f6499.4

        \[\leadsto \mathsf{fma}\left(t, \mathsf{fma}\left(-2, \color{blue}{t \cdot t}, t\right), 0.5\right) \]
    5. Simplified99.4%

      \[\leadsto \color{blue}{\mathsf{fma}\left(t, \mathsf{fma}\left(-2, t \cdot t, t\right), 0.5\right)} \]

    if 0.20000000000000001 < (*.f64 (-.f64 #s(literal 2 binary64) (/.f64 (/.f64 #s(literal 2 binary64) t) (+.f64 #s(literal 1 binary64) (/.f64 #s(literal 1 binary64) t)))) (-.f64 #s(literal 2 binary64) (/.f64 (/.f64 #s(literal 2 binary64) t) (+.f64 #s(literal 1 binary64) (/.f64 #s(literal 1 binary64) t)))))

    1. Initial program 100.0%

      \[1 - \frac{1}{2 + \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right) \cdot \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right)} \]
    2. Add Preprocessing
    3. Taylor expanded in t around inf

      \[\leadsto \color{blue}{\frac{5}{6} - \frac{2}{9} \cdot \frac{1}{t}} \]
    4. Step-by-step derivation
      1. sub-negN/A

        \[\leadsto \color{blue}{\frac{5}{6} + \left(\mathsf{neg}\left(\frac{2}{9} \cdot \frac{1}{t}\right)\right)} \]
      2. +-lowering-+.f64N/A

        \[\leadsto \color{blue}{\frac{5}{6} + \left(\mathsf{neg}\left(\frac{2}{9} \cdot \frac{1}{t}\right)\right)} \]
      3. associate-*r/N/A

        \[\leadsto \frac{5}{6} + \left(\mathsf{neg}\left(\color{blue}{\frac{\frac{2}{9} \cdot 1}{t}}\right)\right) \]
      4. metadata-evalN/A

        \[\leadsto \frac{5}{6} + \left(\mathsf{neg}\left(\frac{\color{blue}{\frac{2}{9}}}{t}\right)\right) \]
      5. distribute-neg-fracN/A

        \[\leadsto \frac{5}{6} + \color{blue}{\frac{\mathsf{neg}\left(\frac{2}{9}\right)}{t}} \]
      6. /-lowering-/.f64N/A

        \[\leadsto \frac{5}{6} + \color{blue}{\frac{\mathsf{neg}\left(\frac{2}{9}\right)}{t}} \]
      7. metadata-eval99.0

        \[\leadsto 0.8333333333333334 + \frac{\color{blue}{-0.2222222222222222}}{t} \]
    5. Simplified99.0%

      \[\leadsto \color{blue}{0.8333333333333334 + \frac{-0.2222222222222222}{t}} \]
  3. Recombined 2 regimes into one program.
  4. Final simplification99.2%

    \[\leadsto \begin{array}{l} \mathbf{if}\;\left(2 + \frac{\frac{2}{t}}{-1 + \frac{-1}{t}}\right) \cdot \left(2 + \frac{\frac{2}{t}}{-1 + \frac{-1}{t}}\right) \leq 0.2:\\ \;\;\;\;\mathsf{fma}\left(t, \mathsf{fma}\left(-2, t \cdot t, t\right), 0.5\right)\\ \mathbf{else}:\\ \;\;\;\;0.8333333333333334 + \frac{-0.2222222222222222}{t}\\ \end{array} \]
  5. Add Preprocessing

Alternative 6: 98.4% accurate, 1.0× speedup?

\[\begin{array}{l} \\ \begin{array}{l} t_1 := 2 + \frac{\frac{2}{t}}{-1 + \frac{-1}{t}}\\ \mathbf{if}\;\frac{1}{2 + t\_1 \cdot t\_1} \leq 0.34:\\ \;\;\;\;0.8333333333333334\\ \mathbf{else}:\\ \;\;\;\;0.5\\ \end{array} \end{array} \]
(FPCore (t)
 :precision binary64
 (let* ((t_1 (+ 2.0 (/ (/ 2.0 t) (+ -1.0 (/ -1.0 t))))))
   (if (<= (/ 1.0 (+ 2.0 (* t_1 t_1))) 0.34) 0.8333333333333334 0.5)))
double code(double t) {
	double t_1 = 2.0 + ((2.0 / t) / (-1.0 + (-1.0 / t)));
	double tmp;
	if ((1.0 / (2.0 + (t_1 * t_1))) <= 0.34) {
		tmp = 0.8333333333333334;
	} else {
		tmp = 0.5;
	}
	return tmp;
}
real(8) function code(t)
    real(8), intent (in) :: t
    real(8) :: t_1
    real(8) :: tmp
    t_1 = 2.0d0 + ((2.0d0 / t) / ((-1.0d0) + ((-1.0d0) / t)))
    if ((1.0d0 / (2.0d0 + (t_1 * t_1))) <= 0.34d0) then
        tmp = 0.8333333333333334d0
    else
        tmp = 0.5d0
    end if
    code = tmp
end function
public static double code(double t) {
	double t_1 = 2.0 + ((2.0 / t) / (-1.0 + (-1.0 / t)));
	double tmp;
	if ((1.0 / (2.0 + (t_1 * t_1))) <= 0.34) {
		tmp = 0.8333333333333334;
	} else {
		tmp = 0.5;
	}
	return tmp;
}
def code(t):
	t_1 = 2.0 + ((2.0 / t) / (-1.0 + (-1.0 / t)))
	tmp = 0
	if (1.0 / (2.0 + (t_1 * t_1))) <= 0.34:
		tmp = 0.8333333333333334
	else:
		tmp = 0.5
	return tmp
function code(t)
	t_1 = Float64(2.0 + Float64(Float64(2.0 / t) / Float64(-1.0 + Float64(-1.0 / t))))
	tmp = 0.0
	if (Float64(1.0 / Float64(2.0 + Float64(t_1 * t_1))) <= 0.34)
		tmp = 0.8333333333333334;
	else
		tmp = 0.5;
	end
	return tmp
end
function tmp_2 = code(t)
	t_1 = 2.0 + ((2.0 / t) / (-1.0 + (-1.0 / t)));
	tmp = 0.0;
	if ((1.0 / (2.0 + (t_1 * t_1))) <= 0.34)
		tmp = 0.8333333333333334;
	else
		tmp = 0.5;
	end
	tmp_2 = tmp;
end
code[t_] := Block[{t$95$1 = N[(2.0 + N[(N[(2.0 / t), $MachinePrecision] / N[(-1.0 + N[(-1.0 / t), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]}, If[LessEqual[N[(1.0 / N[(2.0 + N[(t$95$1 * t$95$1), $MachinePrecision]), $MachinePrecision]), $MachinePrecision], 0.34], 0.8333333333333334, 0.5]]
\begin{array}{l}

\\
\begin{array}{l}
t_1 := 2 + \frac{\frac{2}{t}}{-1 + \frac{-1}{t}}\\
\mathbf{if}\;\frac{1}{2 + t\_1 \cdot t\_1} \leq 0.34:\\
\;\;\;\;0.8333333333333334\\

\mathbf{else}:\\
\;\;\;\;0.5\\


\end{array}
\end{array}
Derivation
  1. Split input into 2 regimes
  2. if (/.f64 #s(literal 1 binary64) (+.f64 #s(literal 2 binary64) (*.f64 (-.f64 #s(literal 2 binary64) (/.f64 (/.f64 #s(literal 2 binary64) t) (+.f64 #s(literal 1 binary64) (/.f64 #s(literal 1 binary64) t)))) (-.f64 #s(literal 2 binary64) (/.f64 (/.f64 #s(literal 2 binary64) t) (+.f64 #s(literal 1 binary64) (/.f64 #s(literal 1 binary64) t))))))) < 0.340000000000000024

    1. Initial program 100.0%

      \[1 - \frac{1}{2 + \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right) \cdot \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right)} \]
    2. Add Preprocessing
    3. Taylor expanded in t around inf

      \[\leadsto \color{blue}{\frac{5}{6}} \]
    4. Step-by-step derivation
      1. Simplified97.7%

        \[\leadsto \color{blue}{0.8333333333333334} \]

      if 0.340000000000000024 < (/.f64 #s(literal 1 binary64) (+.f64 #s(literal 2 binary64) (*.f64 (-.f64 #s(literal 2 binary64) (/.f64 (/.f64 #s(literal 2 binary64) t) (+.f64 #s(literal 1 binary64) (/.f64 #s(literal 1 binary64) t)))) (-.f64 #s(literal 2 binary64) (/.f64 (/.f64 #s(literal 2 binary64) t) (+.f64 #s(literal 1 binary64) (/.f64 #s(literal 1 binary64) t)))))))

      1. Initial program 100.0%

        \[1 - \frac{1}{2 + \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right) \cdot \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right)} \]
      2. Add Preprocessing
      3. Taylor expanded in t around 0

        \[\leadsto \color{blue}{\frac{1}{2}} \]
      4. Step-by-step derivation
        1. Simplified99.0%

          \[\leadsto \color{blue}{0.5} \]
      5. Recombined 2 regimes into one program.
      6. Final simplification98.3%

        \[\leadsto \begin{array}{l} \mathbf{if}\;\frac{1}{2 + \left(2 + \frac{\frac{2}{t}}{-1 + \frac{-1}{t}}\right) \cdot \left(2 + \frac{\frac{2}{t}}{-1 + \frac{-1}{t}}\right)} \leq 0.34:\\ \;\;\;\;0.8333333333333334\\ \mathbf{else}:\\ \;\;\;\;0.5\\ \end{array} \]
      7. Add Preprocessing

      Alternative 7: 99.5% accurate, 1.2× speedup?

      \[\begin{array}{l} \\ \begin{array}{l} \mathbf{if}\;\frac{\frac{2}{t}}{1 + \frac{1}{t}} \leq 0.01:\\ \;\;\;\;0.8333333333333334 + \frac{\frac{0.037037037037037035 + \frac{0.04938271604938271}{t}}{t} + -0.2222222222222222}{t}\\ \mathbf{else}:\\ \;\;\;\;1 + \frac{-1}{\mathsf{fma}\left(\mathsf{fma}\left(t, \mathsf{fma}\left(t, \mathsf{fma}\left(t, -16, 12\right), -8\right), 4\right), t \cdot t, 2\right)}\\ \end{array} \end{array} \]
      (FPCore (t)
       :precision binary64
       (if (<= (/ (/ 2.0 t) (+ 1.0 (/ 1.0 t))) 0.01)
         (+
          0.8333333333333334
          (/
           (+
            (/ (+ 0.037037037037037035 (/ 0.04938271604938271 t)) t)
            -0.2222222222222222)
           t))
         (+
          1.0
          (/ -1.0 (fma (fma t (fma t (fma t -16.0 12.0) -8.0) 4.0) (* t t) 2.0)))))
      double code(double t) {
      	double tmp;
      	if (((2.0 / t) / (1.0 + (1.0 / t))) <= 0.01) {
      		tmp = 0.8333333333333334 + ((((0.037037037037037035 + (0.04938271604938271 / t)) / t) + -0.2222222222222222) / t);
      	} else {
      		tmp = 1.0 + (-1.0 / fma(fma(t, fma(t, fma(t, -16.0, 12.0), -8.0), 4.0), (t * t), 2.0));
      	}
      	return tmp;
      }
      
      function code(t)
      	tmp = 0.0
      	if (Float64(Float64(2.0 / t) / Float64(1.0 + Float64(1.0 / t))) <= 0.01)
      		tmp = Float64(0.8333333333333334 + Float64(Float64(Float64(Float64(0.037037037037037035 + Float64(0.04938271604938271 / t)) / t) + -0.2222222222222222) / t));
      	else
      		tmp = Float64(1.0 + Float64(-1.0 / fma(fma(t, fma(t, fma(t, -16.0, 12.0), -8.0), 4.0), Float64(t * t), 2.0)));
      	end
      	return tmp
      end
      
      code[t_] := If[LessEqual[N[(N[(2.0 / t), $MachinePrecision] / N[(1.0 + N[(1.0 / t), $MachinePrecision]), $MachinePrecision]), $MachinePrecision], 0.01], N[(0.8333333333333334 + N[(N[(N[(N[(0.037037037037037035 + N[(0.04938271604938271 / t), $MachinePrecision]), $MachinePrecision] / t), $MachinePrecision] + -0.2222222222222222), $MachinePrecision] / t), $MachinePrecision]), $MachinePrecision], N[(1.0 + N[(-1.0 / N[(N[(t * N[(t * N[(t * -16.0 + 12.0), $MachinePrecision] + -8.0), $MachinePrecision] + 4.0), $MachinePrecision] * N[(t * t), $MachinePrecision] + 2.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]]
      
      \begin{array}{l}
      
      \\
      \begin{array}{l}
      \mathbf{if}\;\frac{\frac{2}{t}}{1 + \frac{1}{t}} \leq 0.01:\\
      \;\;\;\;0.8333333333333334 + \frac{\frac{0.037037037037037035 + \frac{0.04938271604938271}{t}}{t} + -0.2222222222222222}{t}\\
      
      \mathbf{else}:\\
      \;\;\;\;1 + \frac{-1}{\mathsf{fma}\left(\mathsf{fma}\left(t, \mathsf{fma}\left(t, \mathsf{fma}\left(t, -16, 12\right), -8\right), 4\right), t \cdot t, 2\right)}\\
      
      
      \end{array}
      \end{array}
      
      Derivation
      1. Split input into 2 regimes
      2. if (/.f64 (/.f64 #s(literal 2 binary64) t) (+.f64 #s(literal 1 binary64) (/.f64 #s(literal 1 binary64) t))) < 0.0100000000000000002

        1. Initial program 100.0%

          \[1 - \frac{1}{2 + \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right) \cdot \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right)} \]
        2. Add Preprocessing
        3. Taylor expanded in t around inf

          \[\leadsto \color{blue}{\left(\frac{5}{6} + \left(\frac{\frac{1}{27}}{{t}^{2}} + \frac{4}{81} \cdot \frac{1}{{t}^{3}}\right)\right) - \frac{2}{9} \cdot \frac{1}{t}} \]
        4. Simplified99.8%

          \[\leadsto \color{blue}{0.8333333333333334 + \frac{\frac{0.037037037037037035 + \frac{0.04938271604938271}{t}}{t} + -0.2222222222222222}{t}} \]

        if 0.0100000000000000002 < (/.f64 (/.f64 #s(literal 2 binary64) t) (+.f64 #s(literal 1 binary64) (/.f64 #s(literal 1 binary64) t)))

        1. Initial program 100.0%

          \[1 - \frac{1}{2 + \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right) \cdot \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right)} \]
        2. Add Preprocessing
        3. Taylor expanded in t around 0

          \[\leadsto 1 - \frac{1}{2 + \color{blue}{{t}^{2} \cdot \left(4 + t \cdot \left(t \cdot \left(12 + -16 \cdot t\right) - 8\right)\right)}} \]
        4. Step-by-step derivation
          1. *-lowering-*.f64N/A

            \[\leadsto 1 - \frac{1}{2 + \color{blue}{{t}^{2} \cdot \left(4 + t \cdot \left(t \cdot \left(12 + -16 \cdot t\right) - 8\right)\right)}} \]
          2. unpow2N/A

            \[\leadsto 1 - \frac{1}{2 + \color{blue}{\left(t \cdot t\right)} \cdot \left(4 + t \cdot \left(t \cdot \left(12 + -16 \cdot t\right) - 8\right)\right)} \]
          3. *-lowering-*.f64N/A

            \[\leadsto 1 - \frac{1}{2 + \color{blue}{\left(t \cdot t\right)} \cdot \left(4 + t \cdot \left(t \cdot \left(12 + -16 \cdot t\right) - 8\right)\right)} \]
          4. +-commutativeN/A

            \[\leadsto 1 - \frac{1}{2 + \left(t \cdot t\right) \cdot \color{blue}{\left(t \cdot \left(t \cdot \left(12 + -16 \cdot t\right) - 8\right) + 4\right)}} \]
          5. accelerator-lowering-fma.f64N/A

            \[\leadsto 1 - \frac{1}{2 + \left(t \cdot t\right) \cdot \color{blue}{\mathsf{fma}\left(t, t \cdot \left(12 + -16 \cdot t\right) - 8, 4\right)}} \]
          6. sub-negN/A

            \[\leadsto 1 - \frac{1}{2 + \left(t \cdot t\right) \cdot \mathsf{fma}\left(t, \color{blue}{t \cdot \left(12 + -16 \cdot t\right) + \left(\mathsf{neg}\left(8\right)\right)}, 4\right)} \]
          7. metadata-evalN/A

            \[\leadsto 1 - \frac{1}{2 + \left(t \cdot t\right) \cdot \mathsf{fma}\left(t, t \cdot \left(12 + -16 \cdot t\right) + \color{blue}{-8}, 4\right)} \]
          8. accelerator-lowering-fma.f64N/A

            \[\leadsto 1 - \frac{1}{2 + \left(t \cdot t\right) \cdot \mathsf{fma}\left(t, \color{blue}{\mathsf{fma}\left(t, 12 + -16 \cdot t, -8\right)}, 4\right)} \]
          9. +-commutativeN/A

            \[\leadsto 1 - \frac{1}{2 + \left(t \cdot t\right) \cdot \mathsf{fma}\left(t, \mathsf{fma}\left(t, \color{blue}{-16 \cdot t + 12}, -8\right), 4\right)} \]
          10. *-commutativeN/A

            \[\leadsto 1 - \frac{1}{2 + \left(t \cdot t\right) \cdot \mathsf{fma}\left(t, \mathsf{fma}\left(t, \color{blue}{t \cdot -16} + 12, -8\right), 4\right)} \]
          11. accelerator-lowering-fma.f6498.9

            \[\leadsto 1 - \frac{1}{2 + \left(t \cdot t\right) \cdot \mathsf{fma}\left(t, \mathsf{fma}\left(t, \color{blue}{\mathsf{fma}\left(t, -16, 12\right)}, -8\right), 4\right)} \]
        5. Simplified98.9%

          \[\leadsto 1 - \frac{1}{2 + \color{blue}{\left(t \cdot t\right) \cdot \mathsf{fma}\left(t, \mathsf{fma}\left(t, \mathsf{fma}\left(t, -16, 12\right), -8\right), 4\right)}} \]
        6. Step-by-step derivation
          1. +-commutativeN/A

            \[\leadsto 1 - \frac{1}{\color{blue}{\left(t \cdot t\right) \cdot \left(t \cdot \left(t \cdot \left(t \cdot -16 + 12\right) + -8\right) + 4\right) + 2}} \]
          2. *-commutativeN/A

            \[\leadsto 1 - \frac{1}{\color{blue}{\left(t \cdot \left(t \cdot \left(t \cdot -16 + 12\right) + -8\right) + 4\right) \cdot \left(t \cdot t\right)} + 2} \]
          3. accelerator-lowering-fma.f64N/A

            \[\leadsto 1 - \frac{1}{\color{blue}{\mathsf{fma}\left(t \cdot \left(t \cdot \left(t \cdot -16 + 12\right) + -8\right) + 4, t \cdot t, 2\right)}} \]
          4. accelerator-lowering-fma.f64N/A

            \[\leadsto 1 - \frac{1}{\mathsf{fma}\left(\color{blue}{\mathsf{fma}\left(t, t \cdot \left(t \cdot -16 + 12\right) + -8, 4\right)}, t \cdot t, 2\right)} \]
          5. accelerator-lowering-fma.f64N/A

            \[\leadsto 1 - \frac{1}{\mathsf{fma}\left(\mathsf{fma}\left(t, \color{blue}{\mathsf{fma}\left(t, t \cdot -16 + 12, -8\right)}, 4\right), t \cdot t, 2\right)} \]
          6. accelerator-lowering-fma.f64N/A

            \[\leadsto 1 - \frac{1}{\mathsf{fma}\left(\mathsf{fma}\left(t, \mathsf{fma}\left(t, \color{blue}{\mathsf{fma}\left(t, -16, 12\right)}, -8\right), 4\right), t \cdot t, 2\right)} \]
          7. *-lowering-*.f6498.9

            \[\leadsto 1 - \frac{1}{\mathsf{fma}\left(\mathsf{fma}\left(t, \mathsf{fma}\left(t, \mathsf{fma}\left(t, -16, 12\right), -8\right), 4\right), \color{blue}{t \cdot t}, 2\right)} \]
        7. Applied egg-rr98.9%

          \[\leadsto 1 - \frac{1}{\color{blue}{\mathsf{fma}\left(\mathsf{fma}\left(t, \mathsf{fma}\left(t, \mathsf{fma}\left(t, -16, 12\right), -8\right), 4\right), t \cdot t, 2\right)}} \]
      3. Recombined 2 regimes into one program.
      4. Final simplification99.3%

        \[\leadsto \begin{array}{l} \mathbf{if}\;\frac{\frac{2}{t}}{1 + \frac{1}{t}} \leq 0.01:\\ \;\;\;\;0.8333333333333334 + \frac{\frac{0.037037037037037035 + \frac{0.04938271604938271}{t}}{t} + -0.2222222222222222}{t}\\ \mathbf{else}:\\ \;\;\;\;1 + \frac{-1}{\mathsf{fma}\left(\mathsf{fma}\left(t, \mathsf{fma}\left(t, \mathsf{fma}\left(t, -16, 12\right), -8\right), 4\right), t \cdot t, 2\right)}\\ \end{array} \]
      5. Add Preprocessing

      Alternative 8: 99.1% accurate, 1.8× speedup?

      \[\begin{array}{l} \\ \begin{array}{l} \mathbf{if}\;\frac{\frac{2}{t}}{1 + \frac{1}{t}} \leq 0.01:\\ \;\;\;\;0.8333333333333334 + \frac{-0.2222222222222222}{t}\\ \mathbf{else}:\\ \;\;\;\;\mathsf{fma}\left(t, t, 0.5\right)\\ \end{array} \end{array} \]
      (FPCore (t)
       :precision binary64
       (if (<= (/ (/ 2.0 t) (+ 1.0 (/ 1.0 t))) 0.01)
         (+ 0.8333333333333334 (/ -0.2222222222222222 t))
         (fma t t 0.5)))
      double code(double t) {
      	double tmp;
      	if (((2.0 / t) / (1.0 + (1.0 / t))) <= 0.01) {
      		tmp = 0.8333333333333334 + (-0.2222222222222222 / t);
      	} else {
      		tmp = fma(t, t, 0.5);
      	}
      	return tmp;
      }
      
      function code(t)
      	tmp = 0.0
      	if (Float64(Float64(2.0 / t) / Float64(1.0 + Float64(1.0 / t))) <= 0.01)
      		tmp = Float64(0.8333333333333334 + Float64(-0.2222222222222222 / t));
      	else
      		tmp = fma(t, t, 0.5);
      	end
      	return tmp
      end
      
      code[t_] := If[LessEqual[N[(N[(2.0 / t), $MachinePrecision] / N[(1.0 + N[(1.0 / t), $MachinePrecision]), $MachinePrecision]), $MachinePrecision], 0.01], N[(0.8333333333333334 + N[(-0.2222222222222222 / t), $MachinePrecision]), $MachinePrecision], N[(t * t + 0.5), $MachinePrecision]]
      
      \begin{array}{l}
      
      \\
      \begin{array}{l}
      \mathbf{if}\;\frac{\frac{2}{t}}{1 + \frac{1}{t}} \leq 0.01:\\
      \;\;\;\;0.8333333333333334 + \frac{-0.2222222222222222}{t}\\
      
      \mathbf{else}:\\
      \;\;\;\;\mathsf{fma}\left(t, t, 0.5\right)\\
      
      
      \end{array}
      \end{array}
      
      Derivation
      1. Split input into 2 regimes
      2. if (/.f64 (/.f64 #s(literal 2 binary64) t) (+.f64 #s(literal 1 binary64) (/.f64 #s(literal 1 binary64) t))) < 0.0100000000000000002

        1. Initial program 100.0%

          \[1 - \frac{1}{2 + \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right) \cdot \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right)} \]
        2. Add Preprocessing
        3. Taylor expanded in t around inf

          \[\leadsto \color{blue}{\frac{5}{6} - \frac{2}{9} \cdot \frac{1}{t}} \]
        4. Step-by-step derivation
          1. sub-negN/A

            \[\leadsto \color{blue}{\frac{5}{6} + \left(\mathsf{neg}\left(\frac{2}{9} \cdot \frac{1}{t}\right)\right)} \]
          2. +-lowering-+.f64N/A

            \[\leadsto \color{blue}{\frac{5}{6} + \left(\mathsf{neg}\left(\frac{2}{9} \cdot \frac{1}{t}\right)\right)} \]
          3. associate-*r/N/A

            \[\leadsto \frac{5}{6} + \left(\mathsf{neg}\left(\color{blue}{\frac{\frac{2}{9} \cdot 1}{t}}\right)\right) \]
          4. metadata-evalN/A

            \[\leadsto \frac{5}{6} + \left(\mathsf{neg}\left(\frac{\color{blue}{\frac{2}{9}}}{t}\right)\right) \]
          5. distribute-neg-fracN/A

            \[\leadsto \frac{5}{6} + \color{blue}{\frac{\mathsf{neg}\left(\frac{2}{9}\right)}{t}} \]
          6. /-lowering-/.f64N/A

            \[\leadsto \frac{5}{6} + \color{blue}{\frac{\mathsf{neg}\left(\frac{2}{9}\right)}{t}} \]
          7. metadata-eval99.6

            \[\leadsto 0.8333333333333334 + \frac{\color{blue}{-0.2222222222222222}}{t} \]
        5. Simplified99.6%

          \[\leadsto \color{blue}{0.8333333333333334 + \frac{-0.2222222222222222}{t}} \]

        if 0.0100000000000000002 < (/.f64 (/.f64 #s(literal 2 binary64) t) (+.f64 #s(literal 1 binary64) (/.f64 #s(literal 1 binary64) t)))

        1. Initial program 100.0%

          \[1 - \frac{1}{2 + \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right) \cdot \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right)} \]
        2. Add Preprocessing
        3. Taylor expanded in t around 0

          \[\leadsto \color{blue}{\frac{1}{2} + {t}^{2}} \]
        4. Step-by-step derivation
          1. +-commutativeN/A

            \[\leadsto \color{blue}{{t}^{2} + \frac{1}{2}} \]
          2. unpow2N/A

            \[\leadsto \color{blue}{t \cdot t} + \frac{1}{2} \]
          3. accelerator-lowering-fma.f6498.7

            \[\leadsto \color{blue}{\mathsf{fma}\left(t, t, 0.5\right)} \]
        5. Simplified98.7%

          \[\leadsto \color{blue}{\mathsf{fma}\left(t, t, 0.5\right)} \]
      3. Recombined 2 regimes into one program.
      4. Add Preprocessing

      Alternative 9: 100.0% accurate, 1.8× speedup?

      \[\begin{array}{l} \\ \begin{array}{l} t_1 := 2 + \frac{-2}{1 + t}\\ 1 + \frac{-1}{\mathsf{fma}\left(t\_1, t\_1, 2\right)} \end{array} \end{array} \]
      (FPCore (t)
       :precision binary64
       (let* ((t_1 (+ 2.0 (/ -2.0 (+ 1.0 t))))) (+ 1.0 (/ -1.0 (fma t_1 t_1 2.0)))))
      double code(double t) {
      	double t_1 = 2.0 + (-2.0 / (1.0 + t));
      	return 1.0 + (-1.0 / fma(t_1, t_1, 2.0));
      }
      
      function code(t)
      	t_1 = Float64(2.0 + Float64(-2.0 / Float64(1.0 + t)))
      	return Float64(1.0 + Float64(-1.0 / fma(t_1, t_1, 2.0)))
      end
      
      code[t_] := Block[{t$95$1 = N[(2.0 + N[(-2.0 / N[(1.0 + t), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]}, N[(1.0 + N[(-1.0 / N[(t$95$1 * t$95$1 + 2.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]]
      
      \begin{array}{l}
      
      \\
      \begin{array}{l}
      t_1 := 2 + \frac{-2}{1 + t}\\
      1 + \frac{-1}{\mathsf{fma}\left(t\_1, t\_1, 2\right)}
      \end{array}
      \end{array}
      
      Derivation
      1. Initial program 100.0%

        \[1 - \frac{1}{2 + \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right) \cdot \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right)} \]
      2. Add Preprocessing
      3. Step-by-step derivation
        1. flip--N/A

          \[\leadsto 1 - \frac{1}{2 + \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right) \cdot \color{blue}{\frac{2 \cdot 2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}} \cdot \frac{\frac{2}{t}}{1 + \frac{1}{t}}}{2 + \frac{\frac{2}{t}}{1 + \frac{1}{t}}}}} \]
        2. clear-numN/A

          \[\leadsto 1 - \frac{1}{2 + \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right) \cdot \color{blue}{\frac{1}{\frac{2 + \frac{\frac{2}{t}}{1 + \frac{1}{t}}}{2 \cdot 2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}} \cdot \frac{\frac{2}{t}}{1 + \frac{1}{t}}}}}} \]
        3. un-div-invN/A

          \[\leadsto 1 - \frac{1}{2 + \color{blue}{\frac{2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}}{\frac{2 + \frac{\frac{2}{t}}{1 + \frac{1}{t}}}{2 \cdot 2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}} \cdot \frac{\frac{2}{t}}{1 + \frac{1}{t}}}}}} \]
        4. /-lowering-/.f64N/A

          \[\leadsto 1 - \frac{1}{2 + \color{blue}{\frac{2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}}{\frac{2 + \frac{\frac{2}{t}}{1 + \frac{1}{t}}}{2 \cdot 2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}} \cdot \frac{\frac{2}{t}}{1 + \frac{1}{t}}}}}} \]
      4. Applied egg-rr100.0%

        \[\leadsto 1 - \frac{1}{2 + \color{blue}{\frac{2 - \frac{2}{\mathsf{fma}\left(t, \frac{1}{t}, t\right)}}{\frac{1}{2 - \frac{2}{\mathsf{fma}\left(t, \frac{1}{t}, t\right)}}}}} \]
      5. Step-by-step derivation
        1. +-commutativeN/A

          \[\leadsto 1 - \frac{1}{\color{blue}{\frac{2 - \frac{2}{t \cdot \frac{1}{t} + t}}{\frac{1}{2 - \frac{2}{t \cdot \frac{1}{t} + t}}} + 2}} \]
        2. associate-/r/N/A

          \[\leadsto 1 - \frac{1}{\color{blue}{\frac{2 - \frac{2}{t \cdot \frac{1}{t} + t}}{1} \cdot \left(2 - \frac{2}{t \cdot \frac{1}{t} + t}\right)} + 2} \]
        3. /-rgt-identityN/A

          \[\leadsto 1 - \frac{1}{\color{blue}{\left(2 - \frac{2}{t \cdot \frac{1}{t} + t}\right)} \cdot \left(2 - \frac{2}{t \cdot \frac{1}{t} + t}\right) + 2} \]
        4. accelerator-lowering-fma.f64N/A

          \[\leadsto 1 - \frac{1}{\color{blue}{\mathsf{fma}\left(2 - \frac{2}{t \cdot \frac{1}{t} + t}, 2 - \frac{2}{t \cdot \frac{1}{t} + t}, 2\right)}} \]
      6. Applied egg-rr100.0%

        \[\leadsto 1 - \frac{1}{\color{blue}{\mathsf{fma}\left(2 + \frac{-2}{t + 1}, 2 + \frac{-2}{t + 1}, 2\right)}} \]
      7. Final simplification100.0%

        \[\leadsto 1 + \frac{-1}{\mathsf{fma}\left(2 + \frac{-2}{1 + t}, 2 + \frac{-2}{1 + t}, 2\right)} \]
      8. Add Preprocessing

      Alternative 10: 98.6% accurate, 2.1× speedup?

      \[\begin{array}{l} \\ \begin{array}{l} \mathbf{if}\;\frac{\frac{2}{t}}{1 + \frac{1}{t}} \leq 0.01:\\ \;\;\;\;0.8333333333333334\\ \mathbf{else}:\\ \;\;\;\;\mathsf{fma}\left(t, t, 0.5\right)\\ \end{array} \end{array} \]
      (FPCore (t)
       :precision binary64
       (if (<= (/ (/ 2.0 t) (+ 1.0 (/ 1.0 t))) 0.01)
         0.8333333333333334
         (fma t t 0.5)))
      double code(double t) {
      	double tmp;
      	if (((2.0 / t) / (1.0 + (1.0 / t))) <= 0.01) {
      		tmp = 0.8333333333333334;
      	} else {
      		tmp = fma(t, t, 0.5);
      	}
      	return tmp;
      }
      
      function code(t)
      	tmp = 0.0
      	if (Float64(Float64(2.0 / t) / Float64(1.0 + Float64(1.0 / t))) <= 0.01)
      		tmp = 0.8333333333333334;
      	else
      		tmp = fma(t, t, 0.5);
      	end
      	return tmp
      end
      
      code[t_] := If[LessEqual[N[(N[(2.0 / t), $MachinePrecision] / N[(1.0 + N[(1.0 / t), $MachinePrecision]), $MachinePrecision]), $MachinePrecision], 0.01], 0.8333333333333334, N[(t * t + 0.5), $MachinePrecision]]
      
      \begin{array}{l}
      
      \\
      \begin{array}{l}
      \mathbf{if}\;\frac{\frac{2}{t}}{1 + \frac{1}{t}} \leq 0.01:\\
      \;\;\;\;0.8333333333333334\\
      
      \mathbf{else}:\\
      \;\;\;\;\mathsf{fma}\left(t, t, 0.5\right)\\
      
      
      \end{array}
      \end{array}
      
      Derivation
      1. Split input into 2 regimes
      2. if (/.f64 (/.f64 #s(literal 2 binary64) t) (+.f64 #s(literal 1 binary64) (/.f64 #s(literal 1 binary64) t))) < 0.0100000000000000002

        1. Initial program 100.0%

          \[1 - \frac{1}{2 + \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right) \cdot \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right)} \]
        2. Add Preprocessing
        3. Taylor expanded in t around inf

          \[\leadsto \color{blue}{\frac{5}{6}} \]
        4. Step-by-step derivation
          1. Simplified98.3%

            \[\leadsto \color{blue}{0.8333333333333334} \]

          if 0.0100000000000000002 < (/.f64 (/.f64 #s(literal 2 binary64) t) (+.f64 #s(literal 1 binary64) (/.f64 #s(literal 1 binary64) t)))

          1. Initial program 100.0%

            \[1 - \frac{1}{2 + \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right) \cdot \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right)} \]
          2. Add Preprocessing
          3. Taylor expanded in t around 0

            \[\leadsto \color{blue}{\frac{1}{2} + {t}^{2}} \]
          4. Step-by-step derivation
            1. +-commutativeN/A

              \[\leadsto \color{blue}{{t}^{2} + \frac{1}{2}} \]
            2. unpow2N/A

              \[\leadsto \color{blue}{t \cdot t} + \frac{1}{2} \]
            3. accelerator-lowering-fma.f6498.7

              \[\leadsto \color{blue}{\mathsf{fma}\left(t, t, 0.5\right)} \]
          5. Simplified98.7%

            \[\leadsto \color{blue}{\mathsf{fma}\left(t, t, 0.5\right)} \]
        5. Recombined 2 regimes into one program.
        6. Add Preprocessing

        Alternative 11: 59.5% accurate, 101.0× speedup?

        \[\begin{array}{l} \\ 0.5 \end{array} \]
        (FPCore (t) :precision binary64 0.5)
        double code(double t) {
        	return 0.5;
        }
        
        real(8) function code(t)
            real(8), intent (in) :: t
            code = 0.5d0
        end function
        
        public static double code(double t) {
        	return 0.5;
        }
        
        def code(t):
        	return 0.5
        
        function code(t)
        	return 0.5
        end
        
        function tmp = code(t)
        	tmp = 0.5;
        end
        
        code[t_] := 0.5
        
        \begin{array}{l}
        
        \\
        0.5
        \end{array}
        
        Derivation
        1. Initial program 100.0%

          \[1 - \frac{1}{2 + \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right) \cdot \left(2 - \frac{\frac{2}{t}}{1 + \frac{1}{t}}\right)} \]
        2. Add Preprocessing
        3. Taylor expanded in t around 0

          \[\leadsto \color{blue}{\frac{1}{2}} \]
        4. Step-by-step derivation
          1. Simplified58.7%

            \[\leadsto \color{blue}{0.5} \]
          2. Add Preprocessing

          Reproduce

          ?
          herbie shell --seed 2024204 
          (FPCore (t)
            :name "Kahan p13 Example 3"
            :precision binary64
            (- 1.0 (/ 1.0 (+ 2.0 (* (- 2.0 (/ (/ 2.0 t) (+ 1.0 (/ 1.0 t)))) (- 2.0 (/ (/ 2.0 t) (+ 1.0 (/ 1.0 t)))))))))