Given's Rotation SVD example, simplified

Percentage Accurate: 75.2% → 100.0%
Time: 8.0s
Alternatives: 16
Speedup: 1.9×

Specification

?
\[\begin{array}{l} \\ 1 - \sqrt{0.5 \cdot \left(1 + \frac{1}{\mathsf{hypot}\left(1, x\right)}\right)} \end{array} \]
(FPCore (x)
 :precision binary64
 (- 1.0 (sqrt (* 0.5 (+ 1.0 (/ 1.0 (hypot 1.0 x)))))))
double code(double x) {
	return 1.0 - sqrt((0.5 * (1.0 + (1.0 / hypot(1.0, x)))));
}
public static double code(double x) {
	return 1.0 - Math.sqrt((0.5 * (1.0 + (1.0 / Math.hypot(1.0, x)))));
}
def code(x):
	return 1.0 - math.sqrt((0.5 * (1.0 + (1.0 / math.hypot(1.0, x)))))
function code(x)
	return Float64(1.0 - sqrt(Float64(0.5 * Float64(1.0 + Float64(1.0 / hypot(1.0, x))))))
end
function tmp = code(x)
	tmp = 1.0 - sqrt((0.5 * (1.0 + (1.0 / hypot(1.0, x)))));
end
code[x_] := N[(1.0 - N[Sqrt[N[(0.5 * N[(1.0 + N[(1.0 / N[Sqrt[1.0 ^ 2 + x ^ 2], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}

\\
1 - \sqrt{0.5 \cdot \left(1 + \frac{1}{\mathsf{hypot}\left(1, x\right)}\right)}
\end{array}

Sampling outcomes in binary64 precision:

Local Percentage Accuracy vs ?

The average percentage accuracy by input value. Horizontal axis shows value of an input variable; the variable is choosen in the title. Vertical axis is accuracy; higher is better. Red represent the original program, while blue represents Herbie's suggestion. These can be toggled with buttons below the plot. The line is an average while dots represent individual samples.

Accuracy vs Speed?

Herbie found 16 alternatives:

AlternativeAccuracySpeedup
The accuracy (vertical axis) and speed (horizontal axis) of each alternatives. Up and to the right is better. The red square shows the initial program, and each blue circle shows an alternative.The line shows the best available speed-accuracy tradeoffs.

Initial Program: 75.2% accurate, 1.0× speedup?

\[\begin{array}{l} \\ 1 - \sqrt{0.5 \cdot \left(1 + \frac{1}{\mathsf{hypot}\left(1, x\right)}\right)} \end{array} \]
(FPCore (x)
 :precision binary64
 (- 1.0 (sqrt (* 0.5 (+ 1.0 (/ 1.0 (hypot 1.0 x)))))))
double code(double x) {
	return 1.0 - sqrt((0.5 * (1.0 + (1.0 / hypot(1.0, x)))));
}
public static double code(double x) {
	return 1.0 - Math.sqrt((0.5 * (1.0 + (1.0 / Math.hypot(1.0, x)))));
}
def code(x):
	return 1.0 - math.sqrt((0.5 * (1.0 + (1.0 / math.hypot(1.0, x)))))
function code(x)
	return Float64(1.0 - sqrt(Float64(0.5 * Float64(1.0 + Float64(1.0 / hypot(1.0, x))))))
end
function tmp = code(x)
	tmp = 1.0 - sqrt((0.5 * (1.0 + (1.0 / hypot(1.0, x)))));
end
code[x_] := N[(1.0 - N[Sqrt[N[(0.5 * N[(1.0 + N[(1.0 / N[Sqrt[1.0 ^ 2 + x ^ 2], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}

\\
1 - \sqrt{0.5 \cdot \left(1 + \frac{1}{\mathsf{hypot}\left(1, x\right)}\right)}
\end{array}

Alternative 1: 100.0% accurate, 0.3× speedup?

\[\begin{array}{l} \\ \begin{array}{l} t_0 := \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\\ \mathbf{if}\;\mathsf{hypot}\left(1, x\right) \leq 1.0005:\\ \;\;\;\;{x}^{2} \cdot \left(0.125 + {x}^{2} \cdot \left({x}^{2} \cdot \left(0.0673828125 + {x}^{2} \cdot -0.056243896484375\right) - 0.0859375\right)\right)\\ \mathbf{else}:\\ \;\;\;\;{\left({\left(\frac{0.5 - t\_0}{1 + \sqrt{0.5 + t\_0}}\right)}^{3}\right)}^{0.3333333333333333}\\ \end{array} \end{array} \]
(FPCore (x)
 :precision binary64
 (let* ((t_0 (/ 0.5 (hypot 1.0 x))))
   (if (<= (hypot 1.0 x) 1.0005)
     (*
      (pow x 2.0)
      (+
       0.125
       (*
        (pow x 2.0)
        (-
         (* (pow x 2.0) (+ 0.0673828125 (* (pow x 2.0) -0.056243896484375)))
         0.0859375))))
     (pow
      (pow (/ (- 0.5 t_0) (+ 1.0 (sqrt (+ 0.5 t_0)))) 3.0)
      0.3333333333333333))))
double code(double x) {
	double t_0 = 0.5 / hypot(1.0, x);
	double tmp;
	if (hypot(1.0, x) <= 1.0005) {
		tmp = pow(x, 2.0) * (0.125 + (pow(x, 2.0) * ((pow(x, 2.0) * (0.0673828125 + (pow(x, 2.0) * -0.056243896484375))) - 0.0859375)));
	} else {
		tmp = pow(pow(((0.5 - t_0) / (1.0 + sqrt((0.5 + t_0)))), 3.0), 0.3333333333333333);
	}
	return tmp;
}
public static double code(double x) {
	double t_0 = 0.5 / Math.hypot(1.0, x);
	double tmp;
	if (Math.hypot(1.0, x) <= 1.0005) {
		tmp = Math.pow(x, 2.0) * (0.125 + (Math.pow(x, 2.0) * ((Math.pow(x, 2.0) * (0.0673828125 + (Math.pow(x, 2.0) * -0.056243896484375))) - 0.0859375)));
	} else {
		tmp = Math.pow(Math.pow(((0.5 - t_0) / (1.0 + Math.sqrt((0.5 + t_0)))), 3.0), 0.3333333333333333);
	}
	return tmp;
}
def code(x):
	t_0 = 0.5 / math.hypot(1.0, x)
	tmp = 0
	if math.hypot(1.0, x) <= 1.0005:
		tmp = math.pow(x, 2.0) * (0.125 + (math.pow(x, 2.0) * ((math.pow(x, 2.0) * (0.0673828125 + (math.pow(x, 2.0) * -0.056243896484375))) - 0.0859375)))
	else:
		tmp = math.pow(math.pow(((0.5 - t_0) / (1.0 + math.sqrt((0.5 + t_0)))), 3.0), 0.3333333333333333)
	return tmp
function code(x)
	t_0 = Float64(0.5 / hypot(1.0, x))
	tmp = 0.0
	if (hypot(1.0, x) <= 1.0005)
		tmp = Float64((x ^ 2.0) * Float64(0.125 + Float64((x ^ 2.0) * Float64(Float64((x ^ 2.0) * Float64(0.0673828125 + Float64((x ^ 2.0) * -0.056243896484375))) - 0.0859375))));
	else
		tmp = (Float64(Float64(0.5 - t_0) / Float64(1.0 + sqrt(Float64(0.5 + t_0)))) ^ 3.0) ^ 0.3333333333333333;
	end
	return tmp
end
function tmp_2 = code(x)
	t_0 = 0.5 / hypot(1.0, x);
	tmp = 0.0;
	if (hypot(1.0, x) <= 1.0005)
		tmp = (x ^ 2.0) * (0.125 + ((x ^ 2.0) * (((x ^ 2.0) * (0.0673828125 + ((x ^ 2.0) * -0.056243896484375))) - 0.0859375)));
	else
		tmp = (((0.5 - t_0) / (1.0 + sqrt((0.5 + t_0)))) ^ 3.0) ^ 0.3333333333333333;
	end
	tmp_2 = tmp;
end
code[x_] := Block[{t$95$0 = N[(0.5 / N[Sqrt[1.0 ^ 2 + x ^ 2], $MachinePrecision]), $MachinePrecision]}, If[LessEqual[N[Sqrt[1.0 ^ 2 + x ^ 2], $MachinePrecision], 1.0005], N[(N[Power[x, 2.0], $MachinePrecision] * N[(0.125 + N[(N[Power[x, 2.0], $MachinePrecision] * N[(N[(N[Power[x, 2.0], $MachinePrecision] * N[(0.0673828125 + N[(N[Power[x, 2.0], $MachinePrecision] * -0.056243896484375), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] - 0.0859375), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision], N[Power[N[Power[N[(N[(0.5 - t$95$0), $MachinePrecision] / N[(1.0 + N[Sqrt[N[(0.5 + t$95$0), $MachinePrecision]], $MachinePrecision]), $MachinePrecision]), $MachinePrecision], 3.0], $MachinePrecision], 0.3333333333333333], $MachinePrecision]]]
\begin{array}{l}

\\
\begin{array}{l}
t_0 := \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\\
\mathbf{if}\;\mathsf{hypot}\left(1, x\right) \leq 1.0005:\\
\;\;\;\;{x}^{2} \cdot \left(0.125 + {x}^{2} \cdot \left({x}^{2} \cdot \left(0.0673828125 + {x}^{2} \cdot -0.056243896484375\right) - 0.0859375\right)\right)\\

\mathbf{else}:\\
\;\;\;\;{\left({\left(\frac{0.5 - t\_0}{1 + \sqrt{0.5 + t\_0}}\right)}^{3}\right)}^{0.3333333333333333}\\


\end{array}
\end{array}
Derivation
  1. Split input into 2 regimes
  2. if (hypot.f64 #s(literal 1 binary64) x) < 1.00049999999999994

    1. Initial program 52.5%

      \[1 - \sqrt{0.5 \cdot \left(1 + \frac{1}{\mathsf{hypot}\left(1, x\right)}\right)} \]
    2. Step-by-step derivation
      1. distribute-lft-in52.5%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5 \cdot 1 + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}}} \]
      2. metadata-eval52.5%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5} + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}} \]
      3. associate-*r/52.5%

        \[\leadsto 1 - \sqrt{0.5 + \color{blue}{\frac{0.5 \cdot 1}{\mathsf{hypot}\left(1, x\right)}}} \]
      4. metadata-eval52.5%

        \[\leadsto 1 - \sqrt{0.5 + \frac{\color{blue}{0.5}}{\mathsf{hypot}\left(1, x\right)}} \]
    3. Simplified52.5%

      \[\leadsto \color{blue}{1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    4. Add Preprocessing
    5. Taylor expanded in x around 0 100.0%

      \[\leadsto \color{blue}{{x}^{2} \cdot \left(0.125 + {x}^{2} \cdot \left({x}^{2} \cdot \left(0.0673828125 + -0.056243896484375 \cdot {x}^{2}\right) - 0.0859375\right)\right)} \]

    if 1.00049999999999994 < (hypot.f64 #s(literal 1 binary64) x)

    1. Initial program 98.1%

      \[1 - \sqrt{0.5 \cdot \left(1 + \frac{1}{\mathsf{hypot}\left(1, x\right)}\right)} \]
    2. Step-by-step derivation
      1. distribute-lft-in98.1%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5 \cdot 1 + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}}} \]
      2. metadata-eval98.1%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5} + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}} \]
      3. associate-*r/98.1%

        \[\leadsto 1 - \sqrt{0.5 + \color{blue}{\frac{0.5 \cdot 1}{\mathsf{hypot}\left(1, x\right)}}} \]
      4. metadata-eval98.1%

        \[\leadsto 1 - \sqrt{0.5 + \frac{\color{blue}{0.5}}{\mathsf{hypot}\left(1, x\right)}} \]
    3. Simplified98.1%

      \[\leadsto \color{blue}{1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    4. Add Preprocessing
    5. Step-by-step derivation
      1. add-cbrt-cube97.2%

        \[\leadsto \color{blue}{\sqrt[3]{\left(\left(1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}\right) \cdot \left(1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}\right)\right) \cdot \left(1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}\right)}} \]
      2. pow1/398.1%

        \[\leadsto \color{blue}{{\left(\left(\left(1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}\right) \cdot \left(1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}\right)\right) \cdot \left(1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}\right)\right)}^{0.3333333333333333}} \]
      3. pow398.1%

        \[\leadsto {\color{blue}{\left({\left(1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}\right)}^{3}\right)}}^{0.3333333333333333} \]
    6. Applied egg-rr98.1%

      \[\leadsto \color{blue}{{\left({\left(1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}\right)}^{3}\right)}^{0.3333333333333333}} \]
    7. Step-by-step derivation
      1. flip--98.1%

        \[\leadsto {\left({\color{blue}{\left(\frac{1 \cdot 1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}} \cdot \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}\right)}}^{3}\right)}^{0.3333333333333333} \]
      2. metadata-eval98.1%

        \[\leadsto {\left({\left(\frac{\color{blue}{1} - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}} \cdot \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}\right)}^{3}\right)}^{0.3333333333333333} \]
      3. add-sqr-sqrt99.7%

        \[\leadsto {\left({\left(\frac{1 - \color{blue}{\left(0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right)}}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}\right)}^{3}\right)}^{0.3333333333333333} \]
      4. associate--r+99.7%

        \[\leadsto {\left({\left(\frac{\color{blue}{\left(1 - 0.5\right) - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}\right)}^{3}\right)}^{0.3333333333333333} \]
      5. metadata-eval99.7%

        \[\leadsto {\left({\left(\frac{\color{blue}{0.5} - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}\right)}^{3}\right)}^{0.3333333333333333} \]
    8. Applied egg-rr99.7%

      \[\leadsto {\left({\color{blue}{\left(\frac{0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}\right)}}^{3}\right)}^{0.3333333333333333} \]
  3. Recombined 2 regimes into one program.
  4. Final simplification99.8%

    \[\leadsto \begin{array}{l} \mathbf{if}\;\mathsf{hypot}\left(1, x\right) \leq 1.0005:\\ \;\;\;\;{x}^{2} \cdot \left(0.125 + {x}^{2} \cdot \left({x}^{2} \cdot \left(0.0673828125 + {x}^{2} \cdot -0.056243896484375\right) - 0.0859375\right)\right)\\ \mathbf{else}:\\ \;\;\;\;{\left({\left(\frac{0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}\right)}^{3}\right)}^{0.3333333333333333}\\ \end{array} \]
  5. Add Preprocessing

Alternative 2: 100.0% accurate, 0.3× speedup?

\[\begin{array}{l} \\ \begin{array}{l} \mathbf{if}\;\mathsf{hypot}\left(1, x\right) \leq 1.0005:\\ \;\;\;\;{x}^{2} \cdot \left(0.125 + {x}^{2} \cdot \left({x}^{2} \cdot \left(0.0673828125 + {x}^{2} \cdot -0.056243896484375\right) - 0.0859375\right)\right)\\ \mathbf{else}:\\ \;\;\;\;\frac{1}{\frac{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{e^{\log \left(0.5 + \frac{-0.5}{\mathsf{hypot}\left(1, x\right)}\right)}}}\\ \end{array} \end{array} \]
(FPCore (x)
 :precision binary64
 (if (<= (hypot 1.0 x) 1.0005)
   (*
    (pow x 2.0)
    (+
     0.125
     (*
      (pow x 2.0)
      (-
       (* (pow x 2.0) (+ 0.0673828125 (* (pow x 2.0) -0.056243896484375)))
       0.0859375))))
   (/
    1.0
    (/
     (+ 1.0 (sqrt (+ 0.5 (/ 0.5 (hypot 1.0 x)))))
     (exp (log (+ 0.5 (/ -0.5 (hypot 1.0 x)))))))))
double code(double x) {
	double tmp;
	if (hypot(1.0, x) <= 1.0005) {
		tmp = pow(x, 2.0) * (0.125 + (pow(x, 2.0) * ((pow(x, 2.0) * (0.0673828125 + (pow(x, 2.0) * -0.056243896484375))) - 0.0859375)));
	} else {
		tmp = 1.0 / ((1.0 + sqrt((0.5 + (0.5 / hypot(1.0, x))))) / exp(log((0.5 + (-0.5 / hypot(1.0, x))))));
	}
	return tmp;
}
public static double code(double x) {
	double tmp;
	if (Math.hypot(1.0, x) <= 1.0005) {
		tmp = Math.pow(x, 2.0) * (0.125 + (Math.pow(x, 2.0) * ((Math.pow(x, 2.0) * (0.0673828125 + (Math.pow(x, 2.0) * -0.056243896484375))) - 0.0859375)));
	} else {
		tmp = 1.0 / ((1.0 + Math.sqrt((0.5 + (0.5 / Math.hypot(1.0, x))))) / Math.exp(Math.log((0.5 + (-0.5 / Math.hypot(1.0, x))))));
	}
	return tmp;
}
def code(x):
	tmp = 0
	if math.hypot(1.0, x) <= 1.0005:
		tmp = math.pow(x, 2.0) * (0.125 + (math.pow(x, 2.0) * ((math.pow(x, 2.0) * (0.0673828125 + (math.pow(x, 2.0) * -0.056243896484375))) - 0.0859375)))
	else:
		tmp = 1.0 / ((1.0 + math.sqrt((0.5 + (0.5 / math.hypot(1.0, x))))) / math.exp(math.log((0.5 + (-0.5 / math.hypot(1.0, x))))))
	return tmp
function code(x)
	tmp = 0.0
	if (hypot(1.0, x) <= 1.0005)
		tmp = Float64((x ^ 2.0) * Float64(0.125 + Float64((x ^ 2.0) * Float64(Float64((x ^ 2.0) * Float64(0.0673828125 + Float64((x ^ 2.0) * -0.056243896484375))) - 0.0859375))));
	else
		tmp = Float64(1.0 / Float64(Float64(1.0 + sqrt(Float64(0.5 + Float64(0.5 / hypot(1.0, x))))) / exp(log(Float64(0.5 + Float64(-0.5 / hypot(1.0, x)))))));
	end
	return tmp
end
function tmp_2 = code(x)
	tmp = 0.0;
	if (hypot(1.0, x) <= 1.0005)
		tmp = (x ^ 2.0) * (0.125 + ((x ^ 2.0) * (((x ^ 2.0) * (0.0673828125 + ((x ^ 2.0) * -0.056243896484375))) - 0.0859375)));
	else
		tmp = 1.0 / ((1.0 + sqrt((0.5 + (0.5 / hypot(1.0, x))))) / exp(log((0.5 + (-0.5 / hypot(1.0, x))))));
	end
	tmp_2 = tmp;
end
code[x_] := If[LessEqual[N[Sqrt[1.0 ^ 2 + x ^ 2], $MachinePrecision], 1.0005], N[(N[Power[x, 2.0], $MachinePrecision] * N[(0.125 + N[(N[Power[x, 2.0], $MachinePrecision] * N[(N[(N[Power[x, 2.0], $MachinePrecision] * N[(0.0673828125 + N[(N[Power[x, 2.0], $MachinePrecision] * -0.056243896484375), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] - 0.0859375), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision], N[(1.0 / N[(N[(1.0 + N[Sqrt[N[(0.5 + N[(0.5 / N[Sqrt[1.0 ^ 2 + x ^ 2], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]), $MachinePrecision] / N[Exp[N[Log[N[(0.5 + N[(-0.5 / N[Sqrt[1.0 ^ 2 + x ^ 2], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]]
\begin{array}{l}

\\
\begin{array}{l}
\mathbf{if}\;\mathsf{hypot}\left(1, x\right) \leq 1.0005:\\
\;\;\;\;{x}^{2} \cdot \left(0.125 + {x}^{2} \cdot \left({x}^{2} \cdot \left(0.0673828125 + {x}^{2} \cdot -0.056243896484375\right) - 0.0859375\right)\right)\\

\mathbf{else}:\\
\;\;\;\;\frac{1}{\frac{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{e^{\log \left(0.5 + \frac{-0.5}{\mathsf{hypot}\left(1, x\right)}\right)}}}\\


\end{array}
\end{array}
Derivation
  1. Split input into 2 regimes
  2. if (hypot.f64 #s(literal 1 binary64) x) < 1.00049999999999994

    1. Initial program 52.5%

      \[1 - \sqrt{0.5 \cdot \left(1 + \frac{1}{\mathsf{hypot}\left(1, x\right)}\right)} \]
    2. Step-by-step derivation
      1. distribute-lft-in52.5%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5 \cdot 1 + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}}} \]
      2. metadata-eval52.5%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5} + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}} \]
      3. associate-*r/52.5%

        \[\leadsto 1 - \sqrt{0.5 + \color{blue}{\frac{0.5 \cdot 1}{\mathsf{hypot}\left(1, x\right)}}} \]
      4. metadata-eval52.5%

        \[\leadsto 1 - \sqrt{0.5 + \frac{\color{blue}{0.5}}{\mathsf{hypot}\left(1, x\right)}} \]
    3. Simplified52.5%

      \[\leadsto \color{blue}{1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    4. Add Preprocessing
    5. Taylor expanded in x around 0 100.0%

      \[\leadsto \color{blue}{{x}^{2} \cdot \left(0.125 + {x}^{2} \cdot \left({x}^{2} \cdot \left(0.0673828125 + -0.056243896484375 \cdot {x}^{2}\right) - 0.0859375\right)\right)} \]

    if 1.00049999999999994 < (hypot.f64 #s(literal 1 binary64) x)

    1. Initial program 98.1%

      \[1 - \sqrt{0.5 \cdot \left(1 + \frac{1}{\mathsf{hypot}\left(1, x\right)}\right)} \]
    2. Step-by-step derivation
      1. distribute-lft-in98.1%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5 \cdot 1 + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}}} \]
      2. metadata-eval98.1%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5} + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}} \]
      3. associate-*r/98.1%

        \[\leadsto 1 - \sqrt{0.5 + \color{blue}{\frac{0.5 \cdot 1}{\mathsf{hypot}\left(1, x\right)}}} \]
      4. metadata-eval98.1%

        \[\leadsto 1 - \sqrt{0.5 + \frac{\color{blue}{0.5}}{\mathsf{hypot}\left(1, x\right)}} \]
    3. Simplified98.1%

      \[\leadsto \color{blue}{1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    4. Add Preprocessing
    5. Step-by-step derivation
      1. flip--98.1%

        \[\leadsto \color{blue}{\frac{1 \cdot 1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}} \cdot \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
      2. div-inv98.1%

        \[\leadsto \color{blue}{\left(1 \cdot 1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}} \cdot \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
      3. metadata-eval98.1%

        \[\leadsto \left(\color{blue}{1} - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}} \cdot \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
      4. add-sqr-sqrt99.6%

        \[\leadsto \left(1 - \color{blue}{\left(0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right)}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
      5. associate--r+99.7%

        \[\leadsto \color{blue}{\left(\left(1 - 0.5\right) - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right)} \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
      6. metadata-eval99.7%

        \[\leadsto \left(\color{blue}{0.5} - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    6. Applied egg-rr99.7%

      \[\leadsto \color{blue}{\left(0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
    7. Step-by-step derivation
      1. *-commutative99.7%

        \[\leadsto \color{blue}{\frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \cdot \left(0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right)} \]
      2. associate-/r/99.7%

        \[\leadsto \color{blue}{\frac{1}{\frac{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
    8. Simplified99.7%

      \[\leadsto \color{blue}{\frac{1}{\frac{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
    9. Step-by-step derivation
      1. metadata-eval99.7%

        \[\leadsto \frac{1}{\frac{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{\color{blue}{\left(1 - 0.5\right)} - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
      2. associate--r+99.6%

        \[\leadsto \frac{1}{\frac{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{\color{blue}{1 - \left(0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right)}}} \]
      3. add-exp-log99.6%

        \[\leadsto \frac{1}{\frac{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{\color{blue}{e^{\log \left(1 - \left(0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right)\right)}}}} \]
      4. associate--r+99.7%

        \[\leadsto \frac{1}{\frac{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{e^{\log \color{blue}{\left(\left(1 - 0.5\right) - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right)}}}} \]
      5. metadata-eval99.7%

        \[\leadsto \frac{1}{\frac{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{e^{\log \left(\color{blue}{0.5} - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right)}}} \]
      6. sub-neg99.7%

        \[\leadsto \frac{1}{\frac{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{e^{\log \color{blue}{\left(0.5 + \left(-\frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right)\right)}}}} \]
      7. distribute-neg-frac99.7%

        \[\leadsto \frac{1}{\frac{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{e^{\log \left(0.5 + \color{blue}{\frac{-0.5}{\mathsf{hypot}\left(1, x\right)}}\right)}}} \]
      8. metadata-eval99.7%

        \[\leadsto \frac{1}{\frac{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{e^{\log \left(0.5 + \frac{\color{blue}{-0.5}}{\mathsf{hypot}\left(1, x\right)}\right)}}} \]
    10. Applied egg-rr99.7%

      \[\leadsto \frac{1}{\frac{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{\color{blue}{e^{\log \left(0.5 + \frac{-0.5}{\mathsf{hypot}\left(1, x\right)}\right)}}}} \]
  3. Recombined 2 regimes into one program.
  4. Final simplification99.8%

    \[\leadsto \begin{array}{l} \mathbf{if}\;\mathsf{hypot}\left(1, x\right) \leq 1.0005:\\ \;\;\;\;{x}^{2} \cdot \left(0.125 + {x}^{2} \cdot \left({x}^{2} \cdot \left(0.0673828125 + {x}^{2} \cdot -0.056243896484375\right) - 0.0859375\right)\right)\\ \mathbf{else}:\\ \;\;\;\;\frac{1}{\frac{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{e^{\log \left(0.5 + \frac{-0.5}{\mathsf{hypot}\left(1, x\right)}\right)}}}\\ \end{array} \]
  5. Add Preprocessing

Alternative 3: 100.0% accurate, 0.4× speedup?

\[\begin{array}{l} \\ \begin{array}{l} t_0 := \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\\ \mathbf{if}\;\mathsf{hypot}\left(1, x\right) \leq 1.0005:\\ \;\;\;\;{x}^{2} \cdot \left(0.125 + {x}^{2} \cdot \left({x}^{2} \cdot \left(0.0673828125 + {x}^{2} \cdot -0.056243896484375\right) - 0.0859375\right)\right)\\ \mathbf{else}:\\ \;\;\;\;\frac{1}{\frac{1 + \sqrt{0.5 + t\_0}}{0.5 - t\_0}}\\ \end{array} \end{array} \]
(FPCore (x)
 :precision binary64
 (let* ((t_0 (/ 0.5 (hypot 1.0 x))))
   (if (<= (hypot 1.0 x) 1.0005)
     (*
      (pow x 2.0)
      (+
       0.125
       (*
        (pow x 2.0)
        (-
         (* (pow x 2.0) (+ 0.0673828125 (* (pow x 2.0) -0.056243896484375)))
         0.0859375))))
     (/ 1.0 (/ (+ 1.0 (sqrt (+ 0.5 t_0))) (- 0.5 t_0))))))
double code(double x) {
	double t_0 = 0.5 / hypot(1.0, x);
	double tmp;
	if (hypot(1.0, x) <= 1.0005) {
		tmp = pow(x, 2.0) * (0.125 + (pow(x, 2.0) * ((pow(x, 2.0) * (0.0673828125 + (pow(x, 2.0) * -0.056243896484375))) - 0.0859375)));
	} else {
		tmp = 1.0 / ((1.0 + sqrt((0.5 + t_0))) / (0.5 - t_0));
	}
	return tmp;
}
public static double code(double x) {
	double t_0 = 0.5 / Math.hypot(1.0, x);
	double tmp;
	if (Math.hypot(1.0, x) <= 1.0005) {
		tmp = Math.pow(x, 2.0) * (0.125 + (Math.pow(x, 2.0) * ((Math.pow(x, 2.0) * (0.0673828125 + (Math.pow(x, 2.0) * -0.056243896484375))) - 0.0859375)));
	} else {
		tmp = 1.0 / ((1.0 + Math.sqrt((0.5 + t_0))) / (0.5 - t_0));
	}
	return tmp;
}
def code(x):
	t_0 = 0.5 / math.hypot(1.0, x)
	tmp = 0
	if math.hypot(1.0, x) <= 1.0005:
		tmp = math.pow(x, 2.0) * (0.125 + (math.pow(x, 2.0) * ((math.pow(x, 2.0) * (0.0673828125 + (math.pow(x, 2.0) * -0.056243896484375))) - 0.0859375)))
	else:
		tmp = 1.0 / ((1.0 + math.sqrt((0.5 + t_0))) / (0.5 - t_0))
	return tmp
function code(x)
	t_0 = Float64(0.5 / hypot(1.0, x))
	tmp = 0.0
	if (hypot(1.0, x) <= 1.0005)
		tmp = Float64((x ^ 2.0) * Float64(0.125 + Float64((x ^ 2.0) * Float64(Float64((x ^ 2.0) * Float64(0.0673828125 + Float64((x ^ 2.0) * -0.056243896484375))) - 0.0859375))));
	else
		tmp = Float64(1.0 / Float64(Float64(1.0 + sqrt(Float64(0.5 + t_0))) / Float64(0.5 - t_0)));
	end
	return tmp
end
function tmp_2 = code(x)
	t_0 = 0.5 / hypot(1.0, x);
	tmp = 0.0;
	if (hypot(1.0, x) <= 1.0005)
		tmp = (x ^ 2.0) * (0.125 + ((x ^ 2.0) * (((x ^ 2.0) * (0.0673828125 + ((x ^ 2.0) * -0.056243896484375))) - 0.0859375)));
	else
		tmp = 1.0 / ((1.0 + sqrt((0.5 + t_0))) / (0.5 - t_0));
	end
	tmp_2 = tmp;
end
code[x_] := Block[{t$95$0 = N[(0.5 / N[Sqrt[1.0 ^ 2 + x ^ 2], $MachinePrecision]), $MachinePrecision]}, If[LessEqual[N[Sqrt[1.0 ^ 2 + x ^ 2], $MachinePrecision], 1.0005], N[(N[Power[x, 2.0], $MachinePrecision] * N[(0.125 + N[(N[Power[x, 2.0], $MachinePrecision] * N[(N[(N[Power[x, 2.0], $MachinePrecision] * N[(0.0673828125 + N[(N[Power[x, 2.0], $MachinePrecision] * -0.056243896484375), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] - 0.0859375), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision], N[(1.0 / N[(N[(1.0 + N[Sqrt[N[(0.5 + t$95$0), $MachinePrecision]], $MachinePrecision]), $MachinePrecision] / N[(0.5 - t$95$0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]]]
\begin{array}{l}

\\
\begin{array}{l}
t_0 := \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\\
\mathbf{if}\;\mathsf{hypot}\left(1, x\right) \leq 1.0005:\\
\;\;\;\;{x}^{2} \cdot \left(0.125 + {x}^{2} \cdot \left({x}^{2} \cdot \left(0.0673828125 + {x}^{2} \cdot -0.056243896484375\right) - 0.0859375\right)\right)\\

\mathbf{else}:\\
\;\;\;\;\frac{1}{\frac{1 + \sqrt{0.5 + t\_0}}{0.5 - t\_0}}\\


\end{array}
\end{array}
Derivation
  1. Split input into 2 regimes
  2. if (hypot.f64 #s(literal 1 binary64) x) < 1.00049999999999994

    1. Initial program 52.5%

      \[1 - \sqrt{0.5 \cdot \left(1 + \frac{1}{\mathsf{hypot}\left(1, x\right)}\right)} \]
    2. Step-by-step derivation
      1. distribute-lft-in52.5%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5 \cdot 1 + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}}} \]
      2. metadata-eval52.5%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5} + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}} \]
      3. associate-*r/52.5%

        \[\leadsto 1 - \sqrt{0.5 + \color{blue}{\frac{0.5 \cdot 1}{\mathsf{hypot}\left(1, x\right)}}} \]
      4. metadata-eval52.5%

        \[\leadsto 1 - \sqrt{0.5 + \frac{\color{blue}{0.5}}{\mathsf{hypot}\left(1, x\right)}} \]
    3. Simplified52.5%

      \[\leadsto \color{blue}{1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    4. Add Preprocessing
    5. Taylor expanded in x around 0 100.0%

      \[\leadsto \color{blue}{{x}^{2} \cdot \left(0.125 + {x}^{2} \cdot \left({x}^{2} \cdot \left(0.0673828125 + -0.056243896484375 \cdot {x}^{2}\right) - 0.0859375\right)\right)} \]

    if 1.00049999999999994 < (hypot.f64 #s(literal 1 binary64) x)

    1. Initial program 98.1%

      \[1 - \sqrt{0.5 \cdot \left(1 + \frac{1}{\mathsf{hypot}\left(1, x\right)}\right)} \]
    2. Step-by-step derivation
      1. distribute-lft-in98.1%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5 \cdot 1 + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}}} \]
      2. metadata-eval98.1%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5} + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}} \]
      3. associate-*r/98.1%

        \[\leadsto 1 - \sqrt{0.5 + \color{blue}{\frac{0.5 \cdot 1}{\mathsf{hypot}\left(1, x\right)}}} \]
      4. metadata-eval98.1%

        \[\leadsto 1 - \sqrt{0.5 + \frac{\color{blue}{0.5}}{\mathsf{hypot}\left(1, x\right)}} \]
    3. Simplified98.1%

      \[\leadsto \color{blue}{1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    4. Add Preprocessing
    5. Step-by-step derivation
      1. flip--98.1%

        \[\leadsto \color{blue}{\frac{1 \cdot 1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}} \cdot \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
      2. div-inv98.1%

        \[\leadsto \color{blue}{\left(1 \cdot 1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}} \cdot \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
      3. metadata-eval98.1%

        \[\leadsto \left(\color{blue}{1} - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}} \cdot \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
      4. add-sqr-sqrt99.6%

        \[\leadsto \left(1 - \color{blue}{\left(0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right)}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
      5. associate--r+99.7%

        \[\leadsto \color{blue}{\left(\left(1 - 0.5\right) - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right)} \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
      6. metadata-eval99.7%

        \[\leadsto \left(\color{blue}{0.5} - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    6. Applied egg-rr99.7%

      \[\leadsto \color{blue}{\left(0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
    7. Step-by-step derivation
      1. *-commutative99.7%

        \[\leadsto \color{blue}{\frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \cdot \left(0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right)} \]
      2. associate-/r/99.7%

        \[\leadsto \color{blue}{\frac{1}{\frac{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
    8. Simplified99.7%

      \[\leadsto \color{blue}{\frac{1}{\frac{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
  3. Recombined 2 regimes into one program.
  4. Final simplification99.8%

    \[\leadsto \begin{array}{l} \mathbf{if}\;\mathsf{hypot}\left(1, x\right) \leq 1.0005:\\ \;\;\;\;{x}^{2} \cdot \left(0.125 + {x}^{2} \cdot \left({x}^{2} \cdot \left(0.0673828125 + {x}^{2} \cdot -0.056243896484375\right) - 0.0859375\right)\right)\\ \mathbf{else}:\\ \;\;\;\;\frac{1}{\frac{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}\\ \end{array} \]
  5. Add Preprocessing

Alternative 4: 99.9% accurate, 0.5× speedup?

\[\begin{array}{l} \\ \begin{array}{l} t_0 := \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\\ \mathbf{if}\;\mathsf{hypot}\left(1, x\right) \leq 1.0005:\\ \;\;\;\;{x}^{2} \cdot \left(0.125 + \left(\left(1 + {x}^{2}\right) + -1\right) \cdot \left(0.0673828125 \cdot \left(x \cdot x\right) - 0.0859375\right)\right)\\ \mathbf{else}:\\ \;\;\;\;\frac{1}{\frac{1 + \sqrt{0.5 + t\_0}}{0.5 - t\_0}}\\ \end{array} \end{array} \]
(FPCore (x)
 :precision binary64
 (let* ((t_0 (/ 0.5 (hypot 1.0 x))))
   (if (<= (hypot 1.0 x) 1.0005)
     (*
      (pow x 2.0)
      (+
       0.125
       (*
        (+ (+ 1.0 (pow x 2.0)) -1.0)
        (- (* 0.0673828125 (* x x)) 0.0859375))))
     (/ 1.0 (/ (+ 1.0 (sqrt (+ 0.5 t_0))) (- 0.5 t_0))))))
double code(double x) {
	double t_0 = 0.5 / hypot(1.0, x);
	double tmp;
	if (hypot(1.0, x) <= 1.0005) {
		tmp = pow(x, 2.0) * (0.125 + (((1.0 + pow(x, 2.0)) + -1.0) * ((0.0673828125 * (x * x)) - 0.0859375)));
	} else {
		tmp = 1.0 / ((1.0 + sqrt((0.5 + t_0))) / (0.5 - t_0));
	}
	return tmp;
}
public static double code(double x) {
	double t_0 = 0.5 / Math.hypot(1.0, x);
	double tmp;
	if (Math.hypot(1.0, x) <= 1.0005) {
		tmp = Math.pow(x, 2.0) * (0.125 + (((1.0 + Math.pow(x, 2.0)) + -1.0) * ((0.0673828125 * (x * x)) - 0.0859375)));
	} else {
		tmp = 1.0 / ((1.0 + Math.sqrt((0.5 + t_0))) / (0.5 - t_0));
	}
	return tmp;
}
def code(x):
	t_0 = 0.5 / math.hypot(1.0, x)
	tmp = 0
	if math.hypot(1.0, x) <= 1.0005:
		tmp = math.pow(x, 2.0) * (0.125 + (((1.0 + math.pow(x, 2.0)) + -1.0) * ((0.0673828125 * (x * x)) - 0.0859375)))
	else:
		tmp = 1.0 / ((1.0 + math.sqrt((0.5 + t_0))) / (0.5 - t_0))
	return tmp
function code(x)
	t_0 = Float64(0.5 / hypot(1.0, x))
	tmp = 0.0
	if (hypot(1.0, x) <= 1.0005)
		tmp = Float64((x ^ 2.0) * Float64(0.125 + Float64(Float64(Float64(1.0 + (x ^ 2.0)) + -1.0) * Float64(Float64(0.0673828125 * Float64(x * x)) - 0.0859375))));
	else
		tmp = Float64(1.0 / Float64(Float64(1.0 + sqrt(Float64(0.5 + t_0))) / Float64(0.5 - t_0)));
	end
	return tmp
end
function tmp_2 = code(x)
	t_0 = 0.5 / hypot(1.0, x);
	tmp = 0.0;
	if (hypot(1.0, x) <= 1.0005)
		tmp = (x ^ 2.0) * (0.125 + (((1.0 + (x ^ 2.0)) + -1.0) * ((0.0673828125 * (x * x)) - 0.0859375)));
	else
		tmp = 1.0 / ((1.0 + sqrt((0.5 + t_0))) / (0.5 - t_0));
	end
	tmp_2 = tmp;
end
code[x_] := Block[{t$95$0 = N[(0.5 / N[Sqrt[1.0 ^ 2 + x ^ 2], $MachinePrecision]), $MachinePrecision]}, If[LessEqual[N[Sqrt[1.0 ^ 2 + x ^ 2], $MachinePrecision], 1.0005], N[(N[Power[x, 2.0], $MachinePrecision] * N[(0.125 + N[(N[(N[(1.0 + N[Power[x, 2.0], $MachinePrecision]), $MachinePrecision] + -1.0), $MachinePrecision] * N[(N[(0.0673828125 * N[(x * x), $MachinePrecision]), $MachinePrecision] - 0.0859375), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision], N[(1.0 / N[(N[(1.0 + N[Sqrt[N[(0.5 + t$95$0), $MachinePrecision]], $MachinePrecision]), $MachinePrecision] / N[(0.5 - t$95$0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]]]
\begin{array}{l}

\\
\begin{array}{l}
t_0 := \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\\
\mathbf{if}\;\mathsf{hypot}\left(1, x\right) \leq 1.0005:\\
\;\;\;\;{x}^{2} \cdot \left(0.125 + \left(\left(1 + {x}^{2}\right) + -1\right) \cdot \left(0.0673828125 \cdot \left(x \cdot x\right) - 0.0859375\right)\right)\\

\mathbf{else}:\\
\;\;\;\;\frac{1}{\frac{1 + \sqrt{0.5 + t\_0}}{0.5 - t\_0}}\\


\end{array}
\end{array}
Derivation
  1. Split input into 2 regimes
  2. if (hypot.f64 #s(literal 1 binary64) x) < 1.00049999999999994

    1. Initial program 52.5%

      \[1 - \sqrt{0.5 \cdot \left(1 + \frac{1}{\mathsf{hypot}\left(1, x\right)}\right)} \]
    2. Step-by-step derivation
      1. distribute-lft-in52.5%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5 \cdot 1 + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}}} \]
      2. metadata-eval52.5%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5} + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}} \]
      3. associate-*r/52.5%

        \[\leadsto 1 - \sqrt{0.5 + \color{blue}{\frac{0.5 \cdot 1}{\mathsf{hypot}\left(1, x\right)}}} \]
      4. metadata-eval52.5%

        \[\leadsto 1 - \sqrt{0.5 + \frac{\color{blue}{0.5}}{\mathsf{hypot}\left(1, x\right)}} \]
    3. Simplified52.5%

      \[\leadsto \color{blue}{1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    4. Add Preprocessing
    5. Taylor expanded in x around 0 99.9%

      \[\leadsto \color{blue}{{x}^{2} \cdot \left(0.125 + {x}^{2} \cdot \left(0.0673828125 \cdot {x}^{2} - 0.0859375\right)\right)} \]
    6. Step-by-step derivation
      1. unpow299.9%

        \[\leadsto {x}^{2} \cdot \left(0.125 + {x}^{2} \cdot \left(0.0673828125 \cdot \color{blue}{\left(x \cdot x\right)} - 0.0859375\right)\right) \]
    7. Applied egg-rr99.9%

      \[\leadsto {x}^{2} \cdot \left(0.125 + {x}^{2} \cdot \left(0.0673828125 \cdot \color{blue}{\left(x \cdot x\right)} - 0.0859375\right)\right) \]
    8. Step-by-step derivation
      1. expm1-log1p-u99.9%

        \[\leadsto {x}^{2} \cdot \left(0.125 + \color{blue}{\mathsf{expm1}\left(\mathsf{log1p}\left({x}^{2}\right)\right)} \cdot \left(0.0673828125 \cdot \left(x \cdot x\right) - 0.0859375\right)\right) \]
      2. expm1-undefine99.9%

        \[\leadsto {x}^{2} \cdot \left(0.125 + \color{blue}{\left(e^{\mathsf{log1p}\left({x}^{2}\right)} - 1\right)} \cdot \left(0.0673828125 \cdot \left(x \cdot x\right) - 0.0859375\right)\right) \]
      3. log1p-undefine99.9%

        \[\leadsto {x}^{2} \cdot \left(0.125 + \left(e^{\color{blue}{\log \left(1 + {x}^{2}\right)}} - 1\right) \cdot \left(0.0673828125 \cdot \left(x \cdot x\right) - 0.0859375\right)\right) \]
      4. add-exp-log99.9%

        \[\leadsto {x}^{2} \cdot \left(0.125 + \left(\color{blue}{\left(1 + {x}^{2}\right)} - 1\right) \cdot \left(0.0673828125 \cdot \left(x \cdot x\right) - 0.0859375\right)\right) \]
    9. Applied egg-rr99.9%

      \[\leadsto {x}^{2} \cdot \left(0.125 + \color{blue}{\left(\left(1 + {x}^{2}\right) - 1\right)} \cdot \left(0.0673828125 \cdot \left(x \cdot x\right) - 0.0859375\right)\right) \]

    if 1.00049999999999994 < (hypot.f64 #s(literal 1 binary64) x)

    1. Initial program 98.1%

      \[1 - \sqrt{0.5 \cdot \left(1 + \frac{1}{\mathsf{hypot}\left(1, x\right)}\right)} \]
    2. Step-by-step derivation
      1. distribute-lft-in98.1%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5 \cdot 1 + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}}} \]
      2. metadata-eval98.1%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5} + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}} \]
      3. associate-*r/98.1%

        \[\leadsto 1 - \sqrt{0.5 + \color{blue}{\frac{0.5 \cdot 1}{\mathsf{hypot}\left(1, x\right)}}} \]
      4. metadata-eval98.1%

        \[\leadsto 1 - \sqrt{0.5 + \frac{\color{blue}{0.5}}{\mathsf{hypot}\left(1, x\right)}} \]
    3. Simplified98.1%

      \[\leadsto \color{blue}{1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    4. Add Preprocessing
    5. Step-by-step derivation
      1. flip--98.1%

        \[\leadsto \color{blue}{\frac{1 \cdot 1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}} \cdot \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
      2. div-inv98.1%

        \[\leadsto \color{blue}{\left(1 \cdot 1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}} \cdot \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
      3. metadata-eval98.1%

        \[\leadsto \left(\color{blue}{1} - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}} \cdot \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
      4. add-sqr-sqrt99.6%

        \[\leadsto \left(1 - \color{blue}{\left(0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right)}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
      5. associate--r+99.7%

        \[\leadsto \color{blue}{\left(\left(1 - 0.5\right) - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right)} \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
      6. metadata-eval99.7%

        \[\leadsto \left(\color{blue}{0.5} - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    6. Applied egg-rr99.7%

      \[\leadsto \color{blue}{\left(0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
    7. Step-by-step derivation
      1. *-commutative99.7%

        \[\leadsto \color{blue}{\frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \cdot \left(0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right)} \]
      2. associate-/r/99.7%

        \[\leadsto \color{blue}{\frac{1}{\frac{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
    8. Simplified99.7%

      \[\leadsto \color{blue}{\frac{1}{\frac{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
  3. Recombined 2 regimes into one program.
  4. Final simplification99.8%

    \[\leadsto \begin{array}{l} \mathbf{if}\;\mathsf{hypot}\left(1, x\right) \leq 1.0005:\\ \;\;\;\;{x}^{2} \cdot \left(0.125 + \left(\left(1 + {x}^{2}\right) + -1\right) \cdot \left(0.0673828125 \cdot \left(x \cdot x\right) - 0.0859375\right)\right)\\ \mathbf{else}:\\ \;\;\;\;\frac{1}{\frac{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}\\ \end{array} \]
  5. Add Preprocessing

Alternative 5: 99.9% accurate, 0.5× speedup?

\[\begin{array}{l} \\ \begin{array}{l} t_0 := \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\\ \mathbf{if}\;\mathsf{hypot}\left(1, x\right) \leq 1.0005:\\ \;\;\;\;{x}^{2} \cdot \left(0.125 + \left(\left(1 + {x}^{2}\right) + -1\right) \cdot \left(0.0673828125 \cdot \left(x \cdot x\right) - 0.0859375\right)\right)\\ \mathbf{else}:\\ \;\;\;\;\frac{0.5 - t\_0}{1 + \sqrt{0.5 + t\_0}}\\ \end{array} \end{array} \]
(FPCore (x)
 :precision binary64
 (let* ((t_0 (/ 0.5 (hypot 1.0 x))))
   (if (<= (hypot 1.0 x) 1.0005)
     (*
      (pow x 2.0)
      (+
       0.125
       (*
        (+ (+ 1.0 (pow x 2.0)) -1.0)
        (- (* 0.0673828125 (* x x)) 0.0859375))))
     (/ (- 0.5 t_0) (+ 1.0 (sqrt (+ 0.5 t_0)))))))
double code(double x) {
	double t_0 = 0.5 / hypot(1.0, x);
	double tmp;
	if (hypot(1.0, x) <= 1.0005) {
		tmp = pow(x, 2.0) * (0.125 + (((1.0 + pow(x, 2.0)) + -1.0) * ((0.0673828125 * (x * x)) - 0.0859375)));
	} else {
		tmp = (0.5 - t_0) / (1.0 + sqrt((0.5 + t_0)));
	}
	return tmp;
}
public static double code(double x) {
	double t_0 = 0.5 / Math.hypot(1.0, x);
	double tmp;
	if (Math.hypot(1.0, x) <= 1.0005) {
		tmp = Math.pow(x, 2.0) * (0.125 + (((1.0 + Math.pow(x, 2.0)) + -1.0) * ((0.0673828125 * (x * x)) - 0.0859375)));
	} else {
		tmp = (0.5 - t_0) / (1.0 + Math.sqrt((0.5 + t_0)));
	}
	return tmp;
}
def code(x):
	t_0 = 0.5 / math.hypot(1.0, x)
	tmp = 0
	if math.hypot(1.0, x) <= 1.0005:
		tmp = math.pow(x, 2.0) * (0.125 + (((1.0 + math.pow(x, 2.0)) + -1.0) * ((0.0673828125 * (x * x)) - 0.0859375)))
	else:
		tmp = (0.5 - t_0) / (1.0 + math.sqrt((0.5 + t_0)))
	return tmp
function code(x)
	t_0 = Float64(0.5 / hypot(1.0, x))
	tmp = 0.0
	if (hypot(1.0, x) <= 1.0005)
		tmp = Float64((x ^ 2.0) * Float64(0.125 + Float64(Float64(Float64(1.0 + (x ^ 2.0)) + -1.0) * Float64(Float64(0.0673828125 * Float64(x * x)) - 0.0859375))));
	else
		tmp = Float64(Float64(0.5 - t_0) / Float64(1.0 + sqrt(Float64(0.5 + t_0))));
	end
	return tmp
end
function tmp_2 = code(x)
	t_0 = 0.5 / hypot(1.0, x);
	tmp = 0.0;
	if (hypot(1.0, x) <= 1.0005)
		tmp = (x ^ 2.0) * (0.125 + (((1.0 + (x ^ 2.0)) + -1.0) * ((0.0673828125 * (x * x)) - 0.0859375)));
	else
		tmp = (0.5 - t_0) / (1.0 + sqrt((0.5 + t_0)));
	end
	tmp_2 = tmp;
end
code[x_] := Block[{t$95$0 = N[(0.5 / N[Sqrt[1.0 ^ 2 + x ^ 2], $MachinePrecision]), $MachinePrecision]}, If[LessEqual[N[Sqrt[1.0 ^ 2 + x ^ 2], $MachinePrecision], 1.0005], N[(N[Power[x, 2.0], $MachinePrecision] * N[(0.125 + N[(N[(N[(1.0 + N[Power[x, 2.0], $MachinePrecision]), $MachinePrecision] + -1.0), $MachinePrecision] * N[(N[(0.0673828125 * N[(x * x), $MachinePrecision]), $MachinePrecision] - 0.0859375), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision], N[(N[(0.5 - t$95$0), $MachinePrecision] / N[(1.0 + N[Sqrt[N[(0.5 + t$95$0), $MachinePrecision]], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]]]
\begin{array}{l}

\\
\begin{array}{l}
t_0 := \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\\
\mathbf{if}\;\mathsf{hypot}\left(1, x\right) \leq 1.0005:\\
\;\;\;\;{x}^{2} \cdot \left(0.125 + \left(\left(1 + {x}^{2}\right) + -1\right) \cdot \left(0.0673828125 \cdot \left(x \cdot x\right) - 0.0859375\right)\right)\\

\mathbf{else}:\\
\;\;\;\;\frac{0.5 - t\_0}{1 + \sqrt{0.5 + t\_0}}\\


\end{array}
\end{array}
Derivation
  1. Split input into 2 regimes
  2. if (hypot.f64 #s(literal 1 binary64) x) < 1.00049999999999994

    1. Initial program 52.5%

      \[1 - \sqrt{0.5 \cdot \left(1 + \frac{1}{\mathsf{hypot}\left(1, x\right)}\right)} \]
    2. Step-by-step derivation
      1. distribute-lft-in52.5%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5 \cdot 1 + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}}} \]
      2. metadata-eval52.5%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5} + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}} \]
      3. associate-*r/52.5%

        \[\leadsto 1 - \sqrt{0.5 + \color{blue}{\frac{0.5 \cdot 1}{\mathsf{hypot}\left(1, x\right)}}} \]
      4. metadata-eval52.5%

        \[\leadsto 1 - \sqrt{0.5 + \frac{\color{blue}{0.5}}{\mathsf{hypot}\left(1, x\right)}} \]
    3. Simplified52.5%

      \[\leadsto \color{blue}{1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    4. Add Preprocessing
    5. Taylor expanded in x around 0 99.9%

      \[\leadsto \color{blue}{{x}^{2} \cdot \left(0.125 + {x}^{2} \cdot \left(0.0673828125 \cdot {x}^{2} - 0.0859375\right)\right)} \]
    6. Step-by-step derivation
      1. unpow299.9%

        \[\leadsto {x}^{2} \cdot \left(0.125 + {x}^{2} \cdot \left(0.0673828125 \cdot \color{blue}{\left(x \cdot x\right)} - 0.0859375\right)\right) \]
    7. Applied egg-rr99.9%

      \[\leadsto {x}^{2} \cdot \left(0.125 + {x}^{2} \cdot \left(0.0673828125 \cdot \color{blue}{\left(x \cdot x\right)} - 0.0859375\right)\right) \]
    8. Step-by-step derivation
      1. expm1-log1p-u99.9%

        \[\leadsto {x}^{2} \cdot \left(0.125 + \color{blue}{\mathsf{expm1}\left(\mathsf{log1p}\left({x}^{2}\right)\right)} \cdot \left(0.0673828125 \cdot \left(x \cdot x\right) - 0.0859375\right)\right) \]
      2. expm1-undefine99.9%

        \[\leadsto {x}^{2} \cdot \left(0.125 + \color{blue}{\left(e^{\mathsf{log1p}\left({x}^{2}\right)} - 1\right)} \cdot \left(0.0673828125 \cdot \left(x \cdot x\right) - 0.0859375\right)\right) \]
      3. log1p-undefine99.9%

        \[\leadsto {x}^{2} \cdot \left(0.125 + \left(e^{\color{blue}{\log \left(1 + {x}^{2}\right)}} - 1\right) \cdot \left(0.0673828125 \cdot \left(x \cdot x\right) - 0.0859375\right)\right) \]
      4. add-exp-log99.9%

        \[\leadsto {x}^{2} \cdot \left(0.125 + \left(\color{blue}{\left(1 + {x}^{2}\right)} - 1\right) \cdot \left(0.0673828125 \cdot \left(x \cdot x\right) - 0.0859375\right)\right) \]
    9. Applied egg-rr99.9%

      \[\leadsto {x}^{2} \cdot \left(0.125 + \color{blue}{\left(\left(1 + {x}^{2}\right) - 1\right)} \cdot \left(0.0673828125 \cdot \left(x \cdot x\right) - 0.0859375\right)\right) \]

    if 1.00049999999999994 < (hypot.f64 #s(literal 1 binary64) x)

    1. Initial program 98.1%

      \[1 - \sqrt{0.5 \cdot \left(1 + \frac{1}{\mathsf{hypot}\left(1, x\right)}\right)} \]
    2. Step-by-step derivation
      1. distribute-lft-in98.1%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5 \cdot 1 + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}}} \]
      2. metadata-eval98.1%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5} + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}} \]
      3. associate-*r/98.1%

        \[\leadsto 1 - \sqrt{0.5 + \color{blue}{\frac{0.5 \cdot 1}{\mathsf{hypot}\left(1, x\right)}}} \]
      4. metadata-eval98.1%

        \[\leadsto 1 - \sqrt{0.5 + \frac{\color{blue}{0.5}}{\mathsf{hypot}\left(1, x\right)}} \]
    3. Simplified98.1%

      \[\leadsto \color{blue}{1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    4. Add Preprocessing
    5. Step-by-step derivation
      1. flip--98.1%

        \[\leadsto {\left({\color{blue}{\left(\frac{1 \cdot 1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}} \cdot \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}\right)}}^{3}\right)}^{0.3333333333333333} \]
      2. metadata-eval98.1%

        \[\leadsto {\left({\left(\frac{\color{blue}{1} - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}} \cdot \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}\right)}^{3}\right)}^{0.3333333333333333} \]
      3. add-sqr-sqrt99.7%

        \[\leadsto {\left({\left(\frac{1 - \color{blue}{\left(0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right)}}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}\right)}^{3}\right)}^{0.3333333333333333} \]
      4. associate--r+99.7%

        \[\leadsto {\left({\left(\frac{\color{blue}{\left(1 - 0.5\right) - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}\right)}^{3}\right)}^{0.3333333333333333} \]
      5. metadata-eval99.7%

        \[\leadsto {\left({\left(\frac{\color{blue}{0.5} - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}\right)}^{3}\right)}^{0.3333333333333333} \]
    6. Applied egg-rr99.7%

      \[\leadsto \color{blue}{\frac{0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
  3. Recombined 2 regimes into one program.
  4. Final simplification99.8%

    \[\leadsto \begin{array}{l} \mathbf{if}\;\mathsf{hypot}\left(1, x\right) \leq 1.0005:\\ \;\;\;\;{x}^{2} \cdot \left(0.125 + \left(\left(1 + {x}^{2}\right) + -1\right) \cdot \left(0.0673828125 \cdot \left(x \cdot x\right) - 0.0859375\right)\right)\\ \mathbf{else}:\\ \;\;\;\;\frac{0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}\\ \end{array} \]
  5. Add Preprocessing

Alternative 6: 99.2% accurate, 0.6× speedup?

\[\begin{array}{l} \\ \begin{array}{l} \mathbf{if}\;\mathsf{hypot}\left(1, x\right) \leq 1.0005:\\ \;\;\;\;{x}^{2} \cdot \left(0.125 + \left(\left(1 + {x}^{2}\right) + -1\right) \cdot \left(0.0673828125 \cdot \left(x \cdot x\right) - 0.0859375\right)\right)\\ \mathbf{else}:\\ \;\;\;\;1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}\\ \end{array} \end{array} \]
(FPCore (x)
 :precision binary64
 (if (<= (hypot 1.0 x) 1.0005)
   (*
    (pow x 2.0)
    (+
     0.125
     (* (+ (+ 1.0 (pow x 2.0)) -1.0) (- (* 0.0673828125 (* x x)) 0.0859375))))
   (- 1.0 (sqrt (+ 0.5 (/ 0.5 (hypot 1.0 x)))))))
double code(double x) {
	double tmp;
	if (hypot(1.0, x) <= 1.0005) {
		tmp = pow(x, 2.0) * (0.125 + (((1.0 + pow(x, 2.0)) + -1.0) * ((0.0673828125 * (x * x)) - 0.0859375)));
	} else {
		tmp = 1.0 - sqrt((0.5 + (0.5 / hypot(1.0, x))));
	}
	return tmp;
}
public static double code(double x) {
	double tmp;
	if (Math.hypot(1.0, x) <= 1.0005) {
		tmp = Math.pow(x, 2.0) * (0.125 + (((1.0 + Math.pow(x, 2.0)) + -1.0) * ((0.0673828125 * (x * x)) - 0.0859375)));
	} else {
		tmp = 1.0 - Math.sqrt((0.5 + (0.5 / Math.hypot(1.0, x))));
	}
	return tmp;
}
def code(x):
	tmp = 0
	if math.hypot(1.0, x) <= 1.0005:
		tmp = math.pow(x, 2.0) * (0.125 + (((1.0 + math.pow(x, 2.0)) + -1.0) * ((0.0673828125 * (x * x)) - 0.0859375)))
	else:
		tmp = 1.0 - math.sqrt((0.5 + (0.5 / math.hypot(1.0, x))))
	return tmp
function code(x)
	tmp = 0.0
	if (hypot(1.0, x) <= 1.0005)
		tmp = Float64((x ^ 2.0) * Float64(0.125 + Float64(Float64(Float64(1.0 + (x ^ 2.0)) + -1.0) * Float64(Float64(0.0673828125 * Float64(x * x)) - 0.0859375))));
	else
		tmp = Float64(1.0 - sqrt(Float64(0.5 + Float64(0.5 / hypot(1.0, x)))));
	end
	return tmp
end
function tmp_2 = code(x)
	tmp = 0.0;
	if (hypot(1.0, x) <= 1.0005)
		tmp = (x ^ 2.0) * (0.125 + (((1.0 + (x ^ 2.0)) + -1.0) * ((0.0673828125 * (x * x)) - 0.0859375)));
	else
		tmp = 1.0 - sqrt((0.5 + (0.5 / hypot(1.0, x))));
	end
	tmp_2 = tmp;
end
code[x_] := If[LessEqual[N[Sqrt[1.0 ^ 2 + x ^ 2], $MachinePrecision], 1.0005], N[(N[Power[x, 2.0], $MachinePrecision] * N[(0.125 + N[(N[(N[(1.0 + N[Power[x, 2.0], $MachinePrecision]), $MachinePrecision] + -1.0), $MachinePrecision] * N[(N[(0.0673828125 * N[(x * x), $MachinePrecision]), $MachinePrecision] - 0.0859375), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision], N[(1.0 - N[Sqrt[N[(0.5 + N[(0.5 / N[Sqrt[1.0 ^ 2 + x ^ 2], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]), $MachinePrecision]]
\begin{array}{l}

\\
\begin{array}{l}
\mathbf{if}\;\mathsf{hypot}\left(1, x\right) \leq 1.0005:\\
\;\;\;\;{x}^{2} \cdot \left(0.125 + \left(\left(1 + {x}^{2}\right) + -1\right) \cdot \left(0.0673828125 \cdot \left(x \cdot x\right) - 0.0859375\right)\right)\\

\mathbf{else}:\\
\;\;\;\;1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}\\


\end{array}
\end{array}
Derivation
  1. Split input into 2 regimes
  2. if (hypot.f64 #s(literal 1 binary64) x) < 1.00049999999999994

    1. Initial program 52.5%

      \[1 - \sqrt{0.5 \cdot \left(1 + \frac{1}{\mathsf{hypot}\left(1, x\right)}\right)} \]
    2. Step-by-step derivation
      1. distribute-lft-in52.5%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5 \cdot 1 + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}}} \]
      2. metadata-eval52.5%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5} + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}} \]
      3. associate-*r/52.5%

        \[\leadsto 1 - \sqrt{0.5 + \color{blue}{\frac{0.5 \cdot 1}{\mathsf{hypot}\left(1, x\right)}}} \]
      4. metadata-eval52.5%

        \[\leadsto 1 - \sqrt{0.5 + \frac{\color{blue}{0.5}}{\mathsf{hypot}\left(1, x\right)}} \]
    3. Simplified52.5%

      \[\leadsto \color{blue}{1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    4. Add Preprocessing
    5. Taylor expanded in x around 0 99.9%

      \[\leadsto \color{blue}{{x}^{2} \cdot \left(0.125 + {x}^{2} \cdot \left(0.0673828125 \cdot {x}^{2} - 0.0859375\right)\right)} \]
    6. Step-by-step derivation
      1. unpow299.9%

        \[\leadsto {x}^{2} \cdot \left(0.125 + {x}^{2} \cdot \left(0.0673828125 \cdot \color{blue}{\left(x \cdot x\right)} - 0.0859375\right)\right) \]
    7. Applied egg-rr99.9%

      \[\leadsto {x}^{2} \cdot \left(0.125 + {x}^{2} \cdot \left(0.0673828125 \cdot \color{blue}{\left(x \cdot x\right)} - 0.0859375\right)\right) \]
    8. Step-by-step derivation
      1. expm1-log1p-u99.9%

        \[\leadsto {x}^{2} \cdot \left(0.125 + \color{blue}{\mathsf{expm1}\left(\mathsf{log1p}\left({x}^{2}\right)\right)} \cdot \left(0.0673828125 \cdot \left(x \cdot x\right) - 0.0859375\right)\right) \]
      2. expm1-undefine99.9%

        \[\leadsto {x}^{2} \cdot \left(0.125 + \color{blue}{\left(e^{\mathsf{log1p}\left({x}^{2}\right)} - 1\right)} \cdot \left(0.0673828125 \cdot \left(x \cdot x\right) - 0.0859375\right)\right) \]
      3. log1p-undefine99.9%

        \[\leadsto {x}^{2} \cdot \left(0.125 + \left(e^{\color{blue}{\log \left(1 + {x}^{2}\right)}} - 1\right) \cdot \left(0.0673828125 \cdot \left(x \cdot x\right) - 0.0859375\right)\right) \]
      4. add-exp-log99.9%

        \[\leadsto {x}^{2} \cdot \left(0.125 + \left(\color{blue}{\left(1 + {x}^{2}\right)} - 1\right) \cdot \left(0.0673828125 \cdot \left(x \cdot x\right) - 0.0859375\right)\right) \]
    9. Applied egg-rr99.9%

      \[\leadsto {x}^{2} \cdot \left(0.125 + \color{blue}{\left(\left(1 + {x}^{2}\right) - 1\right)} \cdot \left(0.0673828125 \cdot \left(x \cdot x\right) - 0.0859375\right)\right) \]

    if 1.00049999999999994 < (hypot.f64 #s(literal 1 binary64) x)

    1. Initial program 98.1%

      \[1 - \sqrt{0.5 \cdot \left(1 + \frac{1}{\mathsf{hypot}\left(1, x\right)}\right)} \]
    2. Step-by-step derivation
      1. distribute-lft-in98.1%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5 \cdot 1 + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}}} \]
      2. metadata-eval98.1%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5} + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}} \]
      3. associate-*r/98.1%

        \[\leadsto 1 - \sqrt{0.5 + \color{blue}{\frac{0.5 \cdot 1}{\mathsf{hypot}\left(1, x\right)}}} \]
      4. metadata-eval98.1%

        \[\leadsto 1 - \sqrt{0.5 + \frac{\color{blue}{0.5}}{\mathsf{hypot}\left(1, x\right)}} \]
    3. Simplified98.1%

      \[\leadsto \color{blue}{1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    4. Add Preprocessing
  3. Recombined 2 regimes into one program.
  4. Final simplification99.0%

    \[\leadsto \begin{array}{l} \mathbf{if}\;\mathsf{hypot}\left(1, x\right) \leq 1.0005:\\ \;\;\;\;{x}^{2} \cdot \left(0.125 + \left(\left(1 + {x}^{2}\right) + -1\right) \cdot \left(0.0673828125 \cdot \left(x \cdot x\right) - 0.0859375\right)\right)\\ \mathbf{else}:\\ \;\;\;\;1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}\\ \end{array} \]
  5. Add Preprocessing

Alternative 7: 99.2% accurate, 0.7× speedup?

\[\begin{array}{l} \\ \begin{array}{l} \mathbf{if}\;\mathsf{hypot}\left(1, x\right) \leq 1.0005:\\ \;\;\;\;{x}^{2} \cdot \left(0.125 + \left(x \cdot x\right) \cdot \left(0.0673828125 \cdot \left(x \cdot x\right) - 0.0859375\right)\right)\\ \mathbf{else}:\\ \;\;\;\;1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}\\ \end{array} \end{array} \]
(FPCore (x)
 :precision binary64
 (if (<= (hypot 1.0 x) 1.0005)
   (* (pow x 2.0) (+ 0.125 (* (* x x) (- (* 0.0673828125 (* x x)) 0.0859375))))
   (- 1.0 (sqrt (+ 0.5 (/ 0.5 (hypot 1.0 x)))))))
double code(double x) {
	double tmp;
	if (hypot(1.0, x) <= 1.0005) {
		tmp = pow(x, 2.0) * (0.125 + ((x * x) * ((0.0673828125 * (x * x)) - 0.0859375)));
	} else {
		tmp = 1.0 - sqrt((0.5 + (0.5 / hypot(1.0, x))));
	}
	return tmp;
}
public static double code(double x) {
	double tmp;
	if (Math.hypot(1.0, x) <= 1.0005) {
		tmp = Math.pow(x, 2.0) * (0.125 + ((x * x) * ((0.0673828125 * (x * x)) - 0.0859375)));
	} else {
		tmp = 1.0 - Math.sqrt((0.5 + (0.5 / Math.hypot(1.0, x))));
	}
	return tmp;
}
def code(x):
	tmp = 0
	if math.hypot(1.0, x) <= 1.0005:
		tmp = math.pow(x, 2.0) * (0.125 + ((x * x) * ((0.0673828125 * (x * x)) - 0.0859375)))
	else:
		tmp = 1.0 - math.sqrt((0.5 + (0.5 / math.hypot(1.0, x))))
	return tmp
function code(x)
	tmp = 0.0
	if (hypot(1.0, x) <= 1.0005)
		tmp = Float64((x ^ 2.0) * Float64(0.125 + Float64(Float64(x * x) * Float64(Float64(0.0673828125 * Float64(x * x)) - 0.0859375))));
	else
		tmp = Float64(1.0 - sqrt(Float64(0.5 + Float64(0.5 / hypot(1.0, x)))));
	end
	return tmp
end
function tmp_2 = code(x)
	tmp = 0.0;
	if (hypot(1.0, x) <= 1.0005)
		tmp = (x ^ 2.0) * (0.125 + ((x * x) * ((0.0673828125 * (x * x)) - 0.0859375)));
	else
		tmp = 1.0 - sqrt((0.5 + (0.5 / hypot(1.0, x))));
	end
	tmp_2 = tmp;
end
code[x_] := If[LessEqual[N[Sqrt[1.0 ^ 2 + x ^ 2], $MachinePrecision], 1.0005], N[(N[Power[x, 2.0], $MachinePrecision] * N[(0.125 + N[(N[(x * x), $MachinePrecision] * N[(N[(0.0673828125 * N[(x * x), $MachinePrecision]), $MachinePrecision] - 0.0859375), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision], N[(1.0 - N[Sqrt[N[(0.5 + N[(0.5 / N[Sqrt[1.0 ^ 2 + x ^ 2], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]), $MachinePrecision]]
\begin{array}{l}

\\
\begin{array}{l}
\mathbf{if}\;\mathsf{hypot}\left(1, x\right) \leq 1.0005:\\
\;\;\;\;{x}^{2} \cdot \left(0.125 + \left(x \cdot x\right) \cdot \left(0.0673828125 \cdot \left(x \cdot x\right) - 0.0859375\right)\right)\\

\mathbf{else}:\\
\;\;\;\;1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}\\


\end{array}
\end{array}
Derivation
  1. Split input into 2 regimes
  2. if (hypot.f64 #s(literal 1 binary64) x) < 1.00049999999999994

    1. Initial program 52.5%

      \[1 - \sqrt{0.5 \cdot \left(1 + \frac{1}{\mathsf{hypot}\left(1, x\right)}\right)} \]
    2. Step-by-step derivation
      1. distribute-lft-in52.5%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5 \cdot 1 + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}}} \]
      2. metadata-eval52.5%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5} + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}} \]
      3. associate-*r/52.5%

        \[\leadsto 1 - \sqrt{0.5 + \color{blue}{\frac{0.5 \cdot 1}{\mathsf{hypot}\left(1, x\right)}}} \]
      4. metadata-eval52.5%

        \[\leadsto 1 - \sqrt{0.5 + \frac{\color{blue}{0.5}}{\mathsf{hypot}\left(1, x\right)}} \]
    3. Simplified52.5%

      \[\leadsto \color{blue}{1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    4. Add Preprocessing
    5. Taylor expanded in x around 0 99.9%

      \[\leadsto \color{blue}{{x}^{2} \cdot \left(0.125 + {x}^{2} \cdot \left(0.0673828125 \cdot {x}^{2} - 0.0859375\right)\right)} \]
    6. Step-by-step derivation
      1. unpow299.9%

        \[\leadsto {x}^{2} \cdot \left(0.125 + {x}^{2} \cdot \left(0.0673828125 \cdot \color{blue}{\left(x \cdot x\right)} - 0.0859375\right)\right) \]
    7. Applied egg-rr99.9%

      \[\leadsto {x}^{2} \cdot \left(0.125 + {x}^{2} \cdot \left(0.0673828125 \cdot \color{blue}{\left(x \cdot x\right)} - 0.0859375\right)\right) \]
    8. Step-by-step derivation
      1. unpow299.9%

        \[\leadsto {x}^{2} \cdot \left(0.125 + {x}^{2} \cdot \left(0.0673828125 \cdot \color{blue}{\left(x \cdot x\right)} - 0.0859375\right)\right) \]
    9. Applied egg-rr99.9%

      \[\leadsto {x}^{2} \cdot \left(0.125 + \color{blue}{\left(x \cdot x\right)} \cdot \left(0.0673828125 \cdot \left(x \cdot x\right) - 0.0859375\right)\right) \]

    if 1.00049999999999994 < (hypot.f64 #s(literal 1 binary64) x)

    1. Initial program 98.1%

      \[1 - \sqrt{0.5 \cdot \left(1 + \frac{1}{\mathsf{hypot}\left(1, x\right)}\right)} \]
    2. Step-by-step derivation
      1. distribute-lft-in98.1%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5 \cdot 1 + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}}} \]
      2. metadata-eval98.1%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5} + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}} \]
      3. associate-*r/98.1%

        \[\leadsto 1 - \sqrt{0.5 + \color{blue}{\frac{0.5 \cdot 1}{\mathsf{hypot}\left(1, x\right)}}} \]
      4. metadata-eval98.1%

        \[\leadsto 1 - \sqrt{0.5 + \frac{\color{blue}{0.5}}{\mathsf{hypot}\left(1, x\right)}} \]
    3. Simplified98.1%

      \[\leadsto \color{blue}{1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    4. Add Preprocessing
  3. Recombined 2 regimes into one program.
  4. Add Preprocessing

Alternative 8: 98.8% accurate, 0.9× speedup?

\[\begin{array}{l} \\ \begin{array}{l} \mathbf{if}\;\mathsf{hypot}\left(1, x\right) \leq 2:\\ \;\;\;\;{x}^{2} \cdot \left(0.125 + \left(x \cdot x\right) \cdot \left(0.0673828125 \cdot \left(x \cdot x\right) - 0.0859375\right)\right)\\ \mathbf{else}:\\ \;\;\;\;\frac{0.5}{1 + \sqrt{0.5}}\\ \end{array} \end{array} \]
(FPCore (x)
 :precision binary64
 (if (<= (hypot 1.0 x) 2.0)
   (* (pow x 2.0) (+ 0.125 (* (* x x) (- (* 0.0673828125 (* x x)) 0.0859375))))
   (/ 0.5 (+ 1.0 (sqrt 0.5)))))
double code(double x) {
	double tmp;
	if (hypot(1.0, x) <= 2.0) {
		tmp = pow(x, 2.0) * (0.125 + ((x * x) * ((0.0673828125 * (x * x)) - 0.0859375)));
	} else {
		tmp = 0.5 / (1.0 + sqrt(0.5));
	}
	return tmp;
}
public static double code(double x) {
	double tmp;
	if (Math.hypot(1.0, x) <= 2.0) {
		tmp = Math.pow(x, 2.0) * (0.125 + ((x * x) * ((0.0673828125 * (x * x)) - 0.0859375)));
	} else {
		tmp = 0.5 / (1.0 + Math.sqrt(0.5));
	}
	return tmp;
}
def code(x):
	tmp = 0
	if math.hypot(1.0, x) <= 2.0:
		tmp = math.pow(x, 2.0) * (0.125 + ((x * x) * ((0.0673828125 * (x * x)) - 0.0859375)))
	else:
		tmp = 0.5 / (1.0 + math.sqrt(0.5))
	return tmp
function code(x)
	tmp = 0.0
	if (hypot(1.0, x) <= 2.0)
		tmp = Float64((x ^ 2.0) * Float64(0.125 + Float64(Float64(x * x) * Float64(Float64(0.0673828125 * Float64(x * x)) - 0.0859375))));
	else
		tmp = Float64(0.5 / Float64(1.0 + sqrt(0.5)));
	end
	return tmp
end
function tmp_2 = code(x)
	tmp = 0.0;
	if (hypot(1.0, x) <= 2.0)
		tmp = (x ^ 2.0) * (0.125 + ((x * x) * ((0.0673828125 * (x * x)) - 0.0859375)));
	else
		tmp = 0.5 / (1.0 + sqrt(0.5));
	end
	tmp_2 = tmp;
end
code[x_] := If[LessEqual[N[Sqrt[1.0 ^ 2 + x ^ 2], $MachinePrecision], 2.0], N[(N[Power[x, 2.0], $MachinePrecision] * N[(0.125 + N[(N[(x * x), $MachinePrecision] * N[(N[(0.0673828125 * N[(x * x), $MachinePrecision]), $MachinePrecision] - 0.0859375), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision], N[(0.5 / N[(1.0 + N[Sqrt[0.5], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]]
\begin{array}{l}

\\
\begin{array}{l}
\mathbf{if}\;\mathsf{hypot}\left(1, x\right) \leq 2:\\
\;\;\;\;{x}^{2} \cdot \left(0.125 + \left(x \cdot x\right) \cdot \left(0.0673828125 \cdot \left(x \cdot x\right) - 0.0859375\right)\right)\\

\mathbf{else}:\\
\;\;\;\;\frac{0.5}{1 + \sqrt{0.5}}\\


\end{array}
\end{array}
Derivation
  1. Split input into 2 regimes
  2. if (hypot.f64 #s(literal 1 binary64) x) < 2

    1. Initial program 53.6%

      \[1 - \sqrt{0.5 \cdot \left(1 + \frac{1}{\mathsf{hypot}\left(1, x\right)}\right)} \]
    2. Step-by-step derivation
      1. distribute-lft-in53.6%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5 \cdot 1 + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}}} \]
      2. metadata-eval53.6%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5} + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}} \]
      3. associate-*r/53.6%

        \[\leadsto 1 - \sqrt{0.5 + \color{blue}{\frac{0.5 \cdot 1}{\mathsf{hypot}\left(1, x\right)}}} \]
      4. metadata-eval53.6%

        \[\leadsto 1 - \sqrt{0.5 + \frac{\color{blue}{0.5}}{\mathsf{hypot}\left(1, x\right)}} \]
    3. Simplified53.6%

      \[\leadsto \color{blue}{1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    4. Add Preprocessing
    5. Taylor expanded in x around 0 98.6%

      \[\leadsto \color{blue}{{x}^{2} \cdot \left(0.125 + {x}^{2} \cdot \left(0.0673828125 \cdot {x}^{2} - 0.0859375\right)\right)} \]
    6. Step-by-step derivation
      1. unpow298.6%

        \[\leadsto {x}^{2} \cdot \left(0.125 + {x}^{2} \cdot \left(0.0673828125 \cdot \color{blue}{\left(x \cdot x\right)} - 0.0859375\right)\right) \]
    7. Applied egg-rr98.6%

      \[\leadsto {x}^{2} \cdot \left(0.125 + {x}^{2} \cdot \left(0.0673828125 \cdot \color{blue}{\left(x \cdot x\right)} - 0.0859375\right)\right) \]
    8. Step-by-step derivation
      1. unpow298.6%

        \[\leadsto {x}^{2} \cdot \left(0.125 + {x}^{2} \cdot \left(0.0673828125 \cdot \color{blue}{\left(x \cdot x\right)} - 0.0859375\right)\right) \]
    9. Applied egg-rr98.6%

      \[\leadsto {x}^{2} \cdot \left(0.125 + \color{blue}{\left(x \cdot x\right)} \cdot \left(0.0673828125 \cdot \left(x \cdot x\right) - 0.0859375\right)\right) \]

    if 2 < (hypot.f64 #s(literal 1 binary64) x)

    1. Initial program 98.5%

      \[1 - \sqrt{0.5 \cdot \left(1 + \frac{1}{\mathsf{hypot}\left(1, x\right)}\right)} \]
    2. Step-by-step derivation
      1. distribute-lft-in98.5%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5 \cdot 1 + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}}} \]
      2. metadata-eval98.5%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5} + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}} \]
      3. associate-*r/98.5%

        \[\leadsto 1 - \sqrt{0.5 + \color{blue}{\frac{0.5 \cdot 1}{\mathsf{hypot}\left(1, x\right)}}} \]
      4. metadata-eval98.5%

        \[\leadsto 1 - \sqrt{0.5 + \frac{\color{blue}{0.5}}{\mathsf{hypot}\left(1, x\right)}} \]
    3. Simplified98.5%

      \[\leadsto \color{blue}{1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    4. Add Preprocessing
    5. Step-by-step derivation
      1. flip--98.5%

        \[\leadsto \color{blue}{\frac{1 \cdot 1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}} \cdot \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
      2. div-inv98.5%

        \[\leadsto \color{blue}{\left(1 \cdot 1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}} \cdot \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
      3. metadata-eval98.5%

        \[\leadsto \left(\color{blue}{1} - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}} \cdot \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
      4. add-sqr-sqrt100.0%

        \[\leadsto \left(1 - \color{blue}{\left(0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right)}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
      5. associate--r+100.0%

        \[\leadsto \color{blue}{\left(\left(1 - 0.5\right) - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right)} \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
      6. metadata-eval100.0%

        \[\leadsto \left(\color{blue}{0.5} - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    6. Applied egg-rr100.0%

      \[\leadsto \color{blue}{\left(0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
    7. Step-by-step derivation
      1. *-commutative100.0%

        \[\leadsto \color{blue}{\frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \cdot \left(0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right)} \]
      2. associate-/r/100.0%

        \[\leadsto \color{blue}{\frac{1}{\frac{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
    8. Simplified100.0%

      \[\leadsto \color{blue}{\frac{1}{\frac{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
    9. Taylor expanded in x around inf 98.7%

      \[\leadsto \color{blue}{\frac{0.5}{1 + \sqrt{0.5}}} \]
  3. Recombined 2 regimes into one program.
  4. Add Preprocessing

Alternative 9: 98.5% accurate, 1.0× speedup?

\[\begin{array}{l} \\ \begin{array}{l} \mathbf{if}\;\mathsf{hypot}\left(1, x\right) \leq 2:\\ \;\;\;\;{x}^{2} \cdot 0.125\\ \mathbf{else}:\\ \;\;\;\;\frac{0.5}{1 + \sqrt{0.5}}\\ \end{array} \end{array} \]
(FPCore (x)
 :precision binary64
 (if (<= (hypot 1.0 x) 2.0) (* (pow x 2.0) 0.125) (/ 0.5 (+ 1.0 (sqrt 0.5)))))
double code(double x) {
	double tmp;
	if (hypot(1.0, x) <= 2.0) {
		tmp = pow(x, 2.0) * 0.125;
	} else {
		tmp = 0.5 / (1.0 + sqrt(0.5));
	}
	return tmp;
}
public static double code(double x) {
	double tmp;
	if (Math.hypot(1.0, x) <= 2.0) {
		tmp = Math.pow(x, 2.0) * 0.125;
	} else {
		tmp = 0.5 / (1.0 + Math.sqrt(0.5));
	}
	return tmp;
}
def code(x):
	tmp = 0
	if math.hypot(1.0, x) <= 2.0:
		tmp = math.pow(x, 2.0) * 0.125
	else:
		tmp = 0.5 / (1.0 + math.sqrt(0.5))
	return tmp
function code(x)
	tmp = 0.0
	if (hypot(1.0, x) <= 2.0)
		tmp = Float64((x ^ 2.0) * 0.125);
	else
		tmp = Float64(0.5 / Float64(1.0 + sqrt(0.5)));
	end
	return tmp
end
function tmp_2 = code(x)
	tmp = 0.0;
	if (hypot(1.0, x) <= 2.0)
		tmp = (x ^ 2.0) * 0.125;
	else
		tmp = 0.5 / (1.0 + sqrt(0.5));
	end
	tmp_2 = tmp;
end
code[x_] := If[LessEqual[N[Sqrt[1.0 ^ 2 + x ^ 2], $MachinePrecision], 2.0], N[(N[Power[x, 2.0], $MachinePrecision] * 0.125), $MachinePrecision], N[(0.5 / N[(1.0 + N[Sqrt[0.5], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]]
\begin{array}{l}

\\
\begin{array}{l}
\mathbf{if}\;\mathsf{hypot}\left(1, x\right) \leq 2:\\
\;\;\;\;{x}^{2} \cdot 0.125\\

\mathbf{else}:\\
\;\;\;\;\frac{0.5}{1 + \sqrt{0.5}}\\


\end{array}
\end{array}
Derivation
  1. Split input into 2 regimes
  2. if (hypot.f64 #s(literal 1 binary64) x) < 2

    1. Initial program 53.6%

      \[1 - \sqrt{0.5 \cdot \left(1 + \frac{1}{\mathsf{hypot}\left(1, x\right)}\right)} \]
    2. Step-by-step derivation
      1. distribute-lft-in53.6%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5 \cdot 1 + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}}} \]
      2. metadata-eval53.6%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5} + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}} \]
      3. associate-*r/53.6%

        \[\leadsto 1 - \sqrt{0.5 + \color{blue}{\frac{0.5 \cdot 1}{\mathsf{hypot}\left(1, x\right)}}} \]
      4. metadata-eval53.6%

        \[\leadsto 1 - \sqrt{0.5 + \frac{\color{blue}{0.5}}{\mathsf{hypot}\left(1, x\right)}} \]
    3. Simplified53.6%

      \[\leadsto \color{blue}{1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    4. Add Preprocessing
    5. Taylor expanded in x around 0 97.1%

      \[\leadsto \color{blue}{0.125 \cdot {x}^{2}} \]

    if 2 < (hypot.f64 #s(literal 1 binary64) x)

    1. Initial program 98.5%

      \[1 - \sqrt{0.5 \cdot \left(1 + \frac{1}{\mathsf{hypot}\left(1, x\right)}\right)} \]
    2. Step-by-step derivation
      1. distribute-lft-in98.5%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5 \cdot 1 + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}}} \]
      2. metadata-eval98.5%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5} + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}} \]
      3. associate-*r/98.5%

        \[\leadsto 1 - \sqrt{0.5 + \color{blue}{\frac{0.5 \cdot 1}{\mathsf{hypot}\left(1, x\right)}}} \]
      4. metadata-eval98.5%

        \[\leadsto 1 - \sqrt{0.5 + \frac{\color{blue}{0.5}}{\mathsf{hypot}\left(1, x\right)}} \]
    3. Simplified98.5%

      \[\leadsto \color{blue}{1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    4. Add Preprocessing
    5. Step-by-step derivation
      1. flip--98.5%

        \[\leadsto \color{blue}{\frac{1 \cdot 1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}} \cdot \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
      2. div-inv98.5%

        \[\leadsto \color{blue}{\left(1 \cdot 1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}} \cdot \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
      3. metadata-eval98.5%

        \[\leadsto \left(\color{blue}{1} - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}} \cdot \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
      4. add-sqr-sqrt100.0%

        \[\leadsto \left(1 - \color{blue}{\left(0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right)}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
      5. associate--r+100.0%

        \[\leadsto \color{blue}{\left(\left(1 - 0.5\right) - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right)} \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
      6. metadata-eval100.0%

        \[\leadsto \left(\color{blue}{0.5} - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    6. Applied egg-rr100.0%

      \[\leadsto \color{blue}{\left(0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
    7. Step-by-step derivation
      1. *-commutative100.0%

        \[\leadsto \color{blue}{\frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \cdot \left(0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right)} \]
      2. associate-/r/100.0%

        \[\leadsto \color{blue}{\frac{1}{\frac{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
    8. Simplified100.0%

      \[\leadsto \color{blue}{\frac{1}{\frac{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
    9. Taylor expanded in x around inf 98.7%

      \[\leadsto \color{blue}{\frac{0.5}{1 + \sqrt{0.5}}} \]
  3. Recombined 2 regimes into one program.
  4. Final simplification97.8%

    \[\leadsto \begin{array}{l} \mathbf{if}\;\mathsf{hypot}\left(1, x\right) \leq 2:\\ \;\;\;\;{x}^{2} \cdot 0.125\\ \mathbf{else}:\\ \;\;\;\;\frac{0.5}{1 + \sqrt{0.5}}\\ \end{array} \]
  5. Add Preprocessing

Alternative 10: 75.3% accurate, 1.9× speedup?

\[\begin{array}{l} \\ \begin{array}{l} \mathbf{if}\;x \leq 1.52:\\ \;\;\;\;{x}^{2} \cdot 0.125\\ \mathbf{else}:\\ \;\;\;\;1 - \sqrt{0.5}\\ \end{array} \end{array} \]
(FPCore (x)
 :precision binary64
 (if (<= x 1.52) (* (pow x 2.0) 0.125) (- 1.0 (sqrt 0.5))))
double code(double x) {
	double tmp;
	if (x <= 1.52) {
		tmp = pow(x, 2.0) * 0.125;
	} else {
		tmp = 1.0 - sqrt(0.5);
	}
	return tmp;
}
real(8) function code(x)
    real(8), intent (in) :: x
    real(8) :: tmp
    if (x <= 1.52d0) then
        tmp = (x ** 2.0d0) * 0.125d0
    else
        tmp = 1.0d0 - sqrt(0.5d0)
    end if
    code = tmp
end function
public static double code(double x) {
	double tmp;
	if (x <= 1.52) {
		tmp = Math.pow(x, 2.0) * 0.125;
	} else {
		tmp = 1.0 - Math.sqrt(0.5);
	}
	return tmp;
}
def code(x):
	tmp = 0
	if x <= 1.52:
		tmp = math.pow(x, 2.0) * 0.125
	else:
		tmp = 1.0 - math.sqrt(0.5)
	return tmp
function code(x)
	tmp = 0.0
	if (x <= 1.52)
		tmp = Float64((x ^ 2.0) * 0.125);
	else
		tmp = Float64(1.0 - sqrt(0.5));
	end
	return tmp
end
function tmp_2 = code(x)
	tmp = 0.0;
	if (x <= 1.52)
		tmp = (x ^ 2.0) * 0.125;
	else
		tmp = 1.0 - sqrt(0.5);
	end
	tmp_2 = tmp;
end
code[x_] := If[LessEqual[x, 1.52], N[(N[Power[x, 2.0], $MachinePrecision] * 0.125), $MachinePrecision], N[(1.0 - N[Sqrt[0.5], $MachinePrecision]), $MachinePrecision]]
\begin{array}{l}

\\
\begin{array}{l}
\mathbf{if}\;x \leq 1.52:\\
\;\;\;\;{x}^{2} \cdot 0.125\\

\mathbf{else}:\\
\;\;\;\;1 - \sqrt{0.5}\\


\end{array}
\end{array}
Derivation
  1. Split input into 2 regimes
  2. if x < 1.52

    1. Initial program 67.0%

      \[1 - \sqrt{0.5 \cdot \left(1 + \frac{1}{\mathsf{hypot}\left(1, x\right)}\right)} \]
    2. Step-by-step derivation
      1. distribute-lft-in67.0%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5 \cdot 1 + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}}} \]
      2. metadata-eval67.0%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5} + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}} \]
      3. associate-*r/67.0%

        \[\leadsto 1 - \sqrt{0.5 + \color{blue}{\frac{0.5 \cdot 1}{\mathsf{hypot}\left(1, x\right)}}} \]
      4. metadata-eval67.0%

        \[\leadsto 1 - \sqrt{0.5 + \frac{\color{blue}{0.5}}{\mathsf{hypot}\left(1, x\right)}} \]
    3. Simplified67.0%

      \[\leadsto \color{blue}{1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    4. Add Preprocessing
    5. Taylor expanded in x around 0 69.2%

      \[\leadsto \color{blue}{0.125 \cdot {x}^{2}} \]

    if 1.52 < x

    1. Initial program 98.5%

      \[1 - \sqrt{0.5 \cdot \left(1 + \frac{1}{\mathsf{hypot}\left(1, x\right)}\right)} \]
    2. Step-by-step derivation
      1. distribute-lft-in98.5%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5 \cdot 1 + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}}} \]
      2. metadata-eval98.5%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5} + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}} \]
      3. associate-*r/98.5%

        \[\leadsto 1 - \sqrt{0.5 + \color{blue}{\frac{0.5 \cdot 1}{\mathsf{hypot}\left(1, x\right)}}} \]
      4. metadata-eval98.5%

        \[\leadsto 1 - \sqrt{0.5 + \frac{\color{blue}{0.5}}{\mathsf{hypot}\left(1, x\right)}} \]
    3. Simplified98.5%

      \[\leadsto \color{blue}{1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    4. Add Preprocessing
    5. Taylor expanded in x around inf 97.9%

      \[\leadsto \color{blue}{1 - \sqrt{0.5}} \]
  3. Recombined 2 regimes into one program.
  4. Final simplification76.2%

    \[\leadsto \begin{array}{l} \mathbf{if}\;x \leq 1.52:\\ \;\;\;\;{x}^{2} \cdot 0.125\\ \mathbf{else}:\\ \;\;\;\;1 - \sqrt{0.5}\\ \end{array} \]
  5. Add Preprocessing

Alternative 11: 59.0% accurate, 1.9× speedup?

\[\begin{array}{l} \\ \begin{array}{l} \mathbf{if}\;x \leq 1.2:\\ \;\;\;\;\frac{1}{4 - \frac{4 + \frac{-4}{x}}{x}}\\ \mathbf{else}:\\ \;\;\;\;1 - \sqrt{0.5}\\ \end{array} \end{array} \]
(FPCore (x)
 :precision binary64
 (if (<= x 1.2) (/ 1.0 (- 4.0 (/ (+ 4.0 (/ -4.0 x)) x))) (- 1.0 (sqrt 0.5))))
double code(double x) {
	double tmp;
	if (x <= 1.2) {
		tmp = 1.0 / (4.0 - ((4.0 + (-4.0 / x)) / x));
	} else {
		tmp = 1.0 - sqrt(0.5);
	}
	return tmp;
}
real(8) function code(x)
    real(8), intent (in) :: x
    real(8) :: tmp
    if (x <= 1.2d0) then
        tmp = 1.0d0 / (4.0d0 - ((4.0d0 + ((-4.0d0) / x)) / x))
    else
        tmp = 1.0d0 - sqrt(0.5d0)
    end if
    code = tmp
end function
public static double code(double x) {
	double tmp;
	if (x <= 1.2) {
		tmp = 1.0 / (4.0 - ((4.0 + (-4.0 / x)) / x));
	} else {
		tmp = 1.0 - Math.sqrt(0.5);
	}
	return tmp;
}
def code(x):
	tmp = 0
	if x <= 1.2:
		tmp = 1.0 / (4.0 - ((4.0 + (-4.0 / x)) / x))
	else:
		tmp = 1.0 - math.sqrt(0.5)
	return tmp
function code(x)
	tmp = 0.0
	if (x <= 1.2)
		tmp = Float64(1.0 / Float64(4.0 - Float64(Float64(4.0 + Float64(-4.0 / x)) / x)));
	else
		tmp = Float64(1.0 - sqrt(0.5));
	end
	return tmp
end
function tmp_2 = code(x)
	tmp = 0.0;
	if (x <= 1.2)
		tmp = 1.0 / (4.0 - ((4.0 + (-4.0 / x)) / x));
	else
		tmp = 1.0 - sqrt(0.5);
	end
	tmp_2 = tmp;
end
code[x_] := If[LessEqual[x, 1.2], N[(1.0 / N[(4.0 - N[(N[(4.0 + N[(-4.0 / x), $MachinePrecision]), $MachinePrecision] / x), $MachinePrecision]), $MachinePrecision]), $MachinePrecision], N[(1.0 - N[Sqrt[0.5], $MachinePrecision]), $MachinePrecision]]
\begin{array}{l}

\\
\begin{array}{l}
\mathbf{if}\;x \leq 1.2:\\
\;\;\;\;\frac{1}{4 - \frac{4 + \frac{-4}{x}}{x}}\\

\mathbf{else}:\\
\;\;\;\;1 - \sqrt{0.5}\\


\end{array}
\end{array}
Derivation
  1. Split input into 2 regimes
  2. if x < 1.19999999999999996

    1. Initial program 67.0%

      \[1 - \sqrt{0.5 \cdot \left(1 + \frac{1}{\mathsf{hypot}\left(1, x\right)}\right)} \]
    2. Step-by-step derivation
      1. distribute-lft-in67.0%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5 \cdot 1 + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}}} \]
      2. metadata-eval67.0%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5} + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}} \]
      3. associate-*r/67.0%

        \[\leadsto 1 - \sqrt{0.5 + \color{blue}{\frac{0.5 \cdot 1}{\mathsf{hypot}\left(1, x\right)}}} \]
      4. metadata-eval67.0%

        \[\leadsto 1 - \sqrt{0.5 + \frac{\color{blue}{0.5}}{\mathsf{hypot}\left(1, x\right)}} \]
    3. Simplified67.0%

      \[\leadsto \color{blue}{1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    4. Add Preprocessing
    5. Step-by-step derivation
      1. flip--67.0%

        \[\leadsto \color{blue}{\frac{1 \cdot 1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}} \cdot \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
      2. div-inv67.0%

        \[\leadsto \color{blue}{\left(1 \cdot 1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}} \cdot \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
      3. metadata-eval67.0%

        \[\leadsto \left(\color{blue}{1} - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}} \cdot \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
      4. add-sqr-sqrt67.4%

        \[\leadsto \left(1 - \color{blue}{\left(0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right)}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
      5. associate--r+67.5%

        \[\leadsto \color{blue}{\left(\left(1 - 0.5\right) - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right)} \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
      6. metadata-eval67.5%

        \[\leadsto \left(\color{blue}{0.5} - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    6. Applied egg-rr67.5%

      \[\leadsto \color{blue}{\left(0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
    7. Step-by-step derivation
      1. *-commutative67.5%

        \[\leadsto \color{blue}{\frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \cdot \left(0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right)} \]
      2. associate-/r/67.5%

        \[\leadsto \color{blue}{\frac{1}{\frac{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
    8. Simplified67.5%

      \[\leadsto \color{blue}{\frac{1}{\frac{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
    9. Taylor expanded in x around 0 43.1%

      \[\leadsto \frac{1}{\frac{\color{blue}{2}}{0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    10. Taylor expanded in x around -inf 46.3%

      \[\leadsto \frac{1}{\color{blue}{4 + -1 \cdot \frac{4 - 4 \cdot \frac{1}{x}}{x}}} \]
    11. Step-by-step derivation
      1. mul-1-neg46.3%

        \[\leadsto \frac{1}{4 + \color{blue}{\left(-\frac{4 - 4 \cdot \frac{1}{x}}{x}\right)}} \]
      2. unsub-neg46.3%

        \[\leadsto \frac{1}{\color{blue}{4 - \frac{4 - 4 \cdot \frac{1}{x}}{x}}} \]
      3. sub-neg46.3%

        \[\leadsto \frac{1}{4 - \frac{\color{blue}{4 + \left(-4 \cdot \frac{1}{x}\right)}}{x}} \]
      4. associate-*r/46.3%

        \[\leadsto \frac{1}{4 - \frac{4 + \left(-\color{blue}{\frac{4 \cdot 1}{x}}\right)}{x}} \]
      5. metadata-eval46.3%

        \[\leadsto \frac{1}{4 - \frac{4 + \left(-\frac{\color{blue}{4}}{x}\right)}{x}} \]
      6. distribute-neg-frac46.3%

        \[\leadsto \frac{1}{4 - \frac{4 + \color{blue}{\frac{-4}{x}}}{x}} \]
      7. metadata-eval46.3%

        \[\leadsto \frac{1}{4 - \frac{4 + \frac{\color{blue}{-4}}{x}}{x}} \]
    12. Simplified46.3%

      \[\leadsto \frac{1}{\color{blue}{4 - \frac{4 + \frac{-4}{x}}{x}}} \]

    if 1.19999999999999996 < x

    1. Initial program 98.5%

      \[1 - \sqrt{0.5 \cdot \left(1 + \frac{1}{\mathsf{hypot}\left(1, x\right)}\right)} \]
    2. Step-by-step derivation
      1. distribute-lft-in98.5%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5 \cdot 1 + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}}} \]
      2. metadata-eval98.5%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5} + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}} \]
      3. associate-*r/98.5%

        \[\leadsto 1 - \sqrt{0.5 + \color{blue}{\frac{0.5 \cdot 1}{\mathsf{hypot}\left(1, x\right)}}} \]
      4. metadata-eval98.5%

        \[\leadsto 1 - \sqrt{0.5 + \frac{\color{blue}{0.5}}{\mathsf{hypot}\left(1, x\right)}} \]
    3. Simplified98.5%

      \[\leadsto \color{blue}{1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    4. Add Preprocessing
    5. Taylor expanded in x around inf 97.9%

      \[\leadsto \color{blue}{1 - \sqrt{0.5}} \]
  3. Recombined 2 regimes into one program.
  4. Add Preprocessing

Alternative 12: 32.9% accurate, 15.0× speedup?

\[\begin{array}{l} \\ \begin{array}{l} \mathbf{if}\;x \leq 1:\\ \;\;\;\;1 + \left(-1 - \left(x \cdot x\right) \cdot -0.125\right)\\ \mathbf{else}:\\ \;\;\;\;\frac{1}{4 + \frac{4}{x}}\\ \end{array} \end{array} \]
(FPCore (x)
 :precision binary64
 (if (<= x 1.0) (+ 1.0 (- -1.0 (* (* x x) -0.125))) (/ 1.0 (+ 4.0 (/ 4.0 x)))))
double code(double x) {
	double tmp;
	if (x <= 1.0) {
		tmp = 1.0 + (-1.0 - ((x * x) * -0.125));
	} else {
		tmp = 1.0 / (4.0 + (4.0 / x));
	}
	return tmp;
}
real(8) function code(x)
    real(8), intent (in) :: x
    real(8) :: tmp
    if (x <= 1.0d0) then
        tmp = 1.0d0 + ((-1.0d0) - ((x * x) * (-0.125d0)))
    else
        tmp = 1.0d0 / (4.0d0 + (4.0d0 / x))
    end if
    code = tmp
end function
public static double code(double x) {
	double tmp;
	if (x <= 1.0) {
		tmp = 1.0 + (-1.0 - ((x * x) * -0.125));
	} else {
		tmp = 1.0 / (4.0 + (4.0 / x));
	}
	return tmp;
}
def code(x):
	tmp = 0
	if x <= 1.0:
		tmp = 1.0 + (-1.0 - ((x * x) * -0.125))
	else:
		tmp = 1.0 / (4.0 + (4.0 / x))
	return tmp
function code(x)
	tmp = 0.0
	if (x <= 1.0)
		tmp = Float64(1.0 + Float64(-1.0 - Float64(Float64(x * x) * -0.125)));
	else
		tmp = Float64(1.0 / Float64(4.0 + Float64(4.0 / x)));
	end
	return tmp
end
function tmp_2 = code(x)
	tmp = 0.0;
	if (x <= 1.0)
		tmp = 1.0 + (-1.0 - ((x * x) * -0.125));
	else
		tmp = 1.0 / (4.0 + (4.0 / x));
	end
	tmp_2 = tmp;
end
code[x_] := If[LessEqual[x, 1.0], N[(1.0 + N[(-1.0 - N[(N[(x * x), $MachinePrecision] * -0.125), $MachinePrecision]), $MachinePrecision]), $MachinePrecision], N[(1.0 / N[(4.0 + N[(4.0 / x), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]]
\begin{array}{l}

\\
\begin{array}{l}
\mathbf{if}\;x \leq 1:\\
\;\;\;\;1 + \left(-1 - \left(x \cdot x\right) \cdot -0.125\right)\\

\mathbf{else}:\\
\;\;\;\;\frac{1}{4 + \frac{4}{x}}\\


\end{array}
\end{array}
Derivation
  1. Split input into 2 regimes
  2. if x < 1

    1. Initial program 67.0%

      \[1 - \sqrt{0.5 \cdot \left(1 + \frac{1}{\mathsf{hypot}\left(1, x\right)}\right)} \]
    2. Step-by-step derivation
      1. distribute-lft-in67.0%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5 \cdot 1 + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}}} \]
      2. metadata-eval67.0%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5} + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}} \]
      3. associate-*r/67.0%

        \[\leadsto 1 - \sqrt{0.5 + \color{blue}{\frac{0.5 \cdot 1}{\mathsf{hypot}\left(1, x\right)}}} \]
      4. metadata-eval67.0%

        \[\leadsto 1 - \sqrt{0.5 + \frac{\color{blue}{0.5}}{\mathsf{hypot}\left(1, x\right)}} \]
    3. Simplified67.0%

      \[\leadsto \color{blue}{1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    4. Add Preprocessing
    5. Taylor expanded in x around 0 37.4%

      \[\leadsto 1 - \color{blue}{\left(1 + -0.125 \cdot {x}^{2}\right)} \]
    6. Step-by-step derivation
      1. *-commutative37.4%

        \[\leadsto 1 - \left(1 + \color{blue}{{x}^{2} \cdot -0.125}\right) \]
    7. Simplified37.4%

      \[\leadsto 1 - \color{blue}{\left(1 + {x}^{2} \cdot -0.125\right)} \]
    8. Step-by-step derivation
      1. unpow270.2%

        \[\leadsto {x}^{2} \cdot \left(0.125 + {x}^{2} \cdot \left(0.0673828125 \cdot \color{blue}{\left(x \cdot x\right)} - 0.0859375\right)\right) \]
    9. Applied egg-rr37.4%

      \[\leadsto 1 - \left(1 + \color{blue}{\left(x \cdot x\right)} \cdot -0.125\right) \]

    if 1 < x

    1. Initial program 98.5%

      \[1 - \sqrt{0.5 \cdot \left(1 + \frac{1}{\mathsf{hypot}\left(1, x\right)}\right)} \]
    2. Step-by-step derivation
      1. distribute-lft-in98.5%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5 \cdot 1 + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}}} \]
      2. metadata-eval98.5%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5} + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}} \]
      3. associate-*r/98.5%

        \[\leadsto 1 - \sqrt{0.5 + \color{blue}{\frac{0.5 \cdot 1}{\mathsf{hypot}\left(1, x\right)}}} \]
      4. metadata-eval98.5%

        \[\leadsto 1 - \sqrt{0.5 + \frac{\color{blue}{0.5}}{\mathsf{hypot}\left(1, x\right)}} \]
    3. Simplified98.5%

      \[\leadsto \color{blue}{1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    4. Add Preprocessing
    5. Step-by-step derivation
      1. flip--98.5%

        \[\leadsto \color{blue}{\frac{1 \cdot 1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}} \cdot \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
      2. div-inv98.5%

        \[\leadsto \color{blue}{\left(1 \cdot 1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}} \cdot \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
      3. metadata-eval98.5%

        \[\leadsto \left(\color{blue}{1} - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}} \cdot \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
      4. add-sqr-sqrt100.0%

        \[\leadsto \left(1 - \color{blue}{\left(0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right)}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
      5. associate--r+100.0%

        \[\leadsto \color{blue}{\left(\left(1 - 0.5\right) - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right)} \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
      6. metadata-eval100.0%

        \[\leadsto \left(\color{blue}{0.5} - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    6. Applied egg-rr100.0%

      \[\leadsto \color{blue}{\left(0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
    7. Step-by-step derivation
      1. *-commutative100.0%

        \[\leadsto \color{blue}{\frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \cdot \left(0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right)} \]
      2. associate-/r/100.0%

        \[\leadsto \color{blue}{\frac{1}{\frac{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
    8. Simplified100.0%

      \[\leadsto \color{blue}{\frac{1}{\frac{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
    9. Taylor expanded in x around 0 22.7%

      \[\leadsto \frac{1}{\frac{\color{blue}{2}}{0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    10. Taylor expanded in x around inf 22.7%

      \[\leadsto \frac{1}{\color{blue}{4 + 4 \cdot \frac{1}{x}}} \]
    11. Step-by-step derivation
      1. associate-*r/22.7%

        \[\leadsto \frac{1}{4 + \color{blue}{\frac{4 \cdot 1}{x}}} \]
      2. metadata-eval22.7%

        \[\leadsto \frac{1}{4 + \frac{\color{blue}{4}}{x}} \]
    12. Simplified22.7%

      \[\leadsto \frac{1}{\color{blue}{4 + \frac{4}{x}}} \]
  3. Recombined 2 regimes into one program.
  4. Final simplification33.8%

    \[\leadsto \begin{array}{l} \mathbf{if}\;x \leq 1:\\ \;\;\;\;1 + \left(-1 - \left(x \cdot x\right) \cdot -0.125\right)\\ \mathbf{else}:\\ \;\;\;\;\frac{1}{4 + \frac{4}{x}}\\ \end{array} \]
  5. Add Preprocessing

Alternative 13: 32.5% accurate, 17.5× speedup?

\[\begin{array}{l} \\ \begin{array}{l} \mathbf{if}\;x \leq 5.6 \cdot 10^{-103}:\\ \;\;\;\;0\\ \mathbf{else}:\\ \;\;\;\;\frac{1}{4 + \frac{4}{x}}\\ \end{array} \end{array} \]
(FPCore (x)
 :precision binary64
 (if (<= x 5.6e-103) 0.0 (/ 1.0 (+ 4.0 (/ 4.0 x)))))
double code(double x) {
	double tmp;
	if (x <= 5.6e-103) {
		tmp = 0.0;
	} else {
		tmp = 1.0 / (4.0 + (4.0 / x));
	}
	return tmp;
}
real(8) function code(x)
    real(8), intent (in) :: x
    real(8) :: tmp
    if (x <= 5.6d-103) then
        tmp = 0.0d0
    else
        tmp = 1.0d0 / (4.0d0 + (4.0d0 / x))
    end if
    code = tmp
end function
public static double code(double x) {
	double tmp;
	if (x <= 5.6e-103) {
		tmp = 0.0;
	} else {
		tmp = 1.0 / (4.0 + (4.0 / x));
	}
	return tmp;
}
def code(x):
	tmp = 0
	if x <= 5.6e-103:
		tmp = 0.0
	else:
		tmp = 1.0 / (4.0 + (4.0 / x))
	return tmp
function code(x)
	tmp = 0.0
	if (x <= 5.6e-103)
		tmp = 0.0;
	else
		tmp = Float64(1.0 / Float64(4.0 + Float64(4.0 / x)));
	end
	return tmp
end
function tmp_2 = code(x)
	tmp = 0.0;
	if (x <= 5.6e-103)
		tmp = 0.0;
	else
		tmp = 1.0 / (4.0 + (4.0 / x));
	end
	tmp_2 = tmp;
end
code[x_] := If[LessEqual[x, 5.6e-103], 0.0, N[(1.0 / N[(4.0 + N[(4.0 / x), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]]
\begin{array}{l}

\\
\begin{array}{l}
\mathbf{if}\;x \leq 5.6 \cdot 10^{-103}:\\
\;\;\;\;0\\

\mathbf{else}:\\
\;\;\;\;\frac{1}{4 + \frac{4}{x}}\\


\end{array}
\end{array}
Derivation
  1. Split input into 2 regimes
  2. if x < 5.60000000000000046e-103

    1. Initial program 75.7%

      \[1 - \sqrt{0.5 \cdot \left(1 + \frac{1}{\mathsf{hypot}\left(1, x\right)}\right)} \]
    2. Step-by-step derivation
      1. distribute-lft-in75.7%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5 \cdot 1 + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}}} \]
      2. metadata-eval75.7%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5} + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}} \]
      3. associate-*r/75.7%

        \[\leadsto 1 - \sqrt{0.5 + \color{blue}{\frac{0.5 \cdot 1}{\mathsf{hypot}\left(1, x\right)}}} \]
      4. metadata-eval75.7%

        \[\leadsto 1 - \sqrt{0.5 + \frac{\color{blue}{0.5}}{\mathsf{hypot}\left(1, x\right)}} \]
    3. Simplified75.7%

      \[\leadsto \color{blue}{1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    4. Add Preprocessing
    5. Taylor expanded in x around 0 41.5%

      \[\leadsto 1 - \color{blue}{1} \]
    6. Step-by-step derivation
      1. metadata-eval41.5%

        \[\leadsto \color{blue}{0} \]
    7. Applied egg-rr41.5%

      \[\leadsto \color{blue}{0} \]

    if 5.60000000000000046e-103 < x

    1. Initial program 72.7%

      \[1 - \sqrt{0.5 \cdot \left(1 + \frac{1}{\mathsf{hypot}\left(1, x\right)}\right)} \]
    2. Step-by-step derivation
      1. distribute-lft-in72.7%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5 \cdot 1 + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}}} \]
      2. metadata-eval72.7%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5} + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}} \]
      3. associate-*r/72.7%

        \[\leadsto 1 - \sqrt{0.5 + \color{blue}{\frac{0.5 \cdot 1}{\mathsf{hypot}\left(1, x\right)}}} \]
      4. metadata-eval72.7%

        \[\leadsto 1 - \sqrt{0.5 + \frac{\color{blue}{0.5}}{\mathsf{hypot}\left(1, x\right)}} \]
    3. Simplified72.7%

      \[\leadsto \color{blue}{1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    4. Add Preprocessing
    5. Step-by-step derivation
      1. flip--72.7%

        \[\leadsto \color{blue}{\frac{1 \cdot 1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}} \cdot \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
      2. div-inv72.7%

        \[\leadsto \color{blue}{\left(1 \cdot 1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}} \cdot \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
      3. metadata-eval72.7%

        \[\leadsto \left(\color{blue}{1} - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}} \cdot \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
      4. add-sqr-sqrt73.8%

        \[\leadsto \left(1 - \color{blue}{\left(0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right)}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
      5. associate--r+73.8%

        \[\leadsto \color{blue}{\left(\left(1 - 0.5\right) - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right)} \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
      6. metadata-eval73.8%

        \[\leadsto \left(\color{blue}{0.5} - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    6. Applied egg-rr73.8%

      \[\leadsto \color{blue}{\left(0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
    7. Step-by-step derivation
      1. *-commutative73.8%

        \[\leadsto \color{blue}{\frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \cdot \left(0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right)} \]
      2. associate-/r/73.8%

        \[\leadsto \color{blue}{\frac{1}{\frac{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
    8. Simplified73.8%

      \[\leadsto \color{blue}{\frac{1}{\frac{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
    9. Taylor expanded in x around 0 19.4%

      \[\leadsto \frac{1}{\frac{\color{blue}{2}}{0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    10. Taylor expanded in x around inf 18.1%

      \[\leadsto \frac{1}{\color{blue}{4 + 4 \cdot \frac{1}{x}}} \]
    11. Step-by-step derivation
      1. associate-*r/18.1%

        \[\leadsto \frac{1}{4 + \color{blue}{\frac{4 \cdot 1}{x}}} \]
      2. metadata-eval18.1%

        \[\leadsto \frac{1}{4 + \frac{\color{blue}{4}}{x}} \]
    12. Simplified18.1%

      \[\leadsto \frac{1}{\color{blue}{4 + \frac{4}{x}}} \]
  3. Recombined 2 regimes into one program.
  4. Add Preprocessing

Alternative 14: 40.4% accurate, 19.1× speedup?

\[\begin{array}{l} \\ \frac{1}{4 - \frac{4 + \frac{-4}{x}}{x}} \end{array} \]
(FPCore (x) :precision binary64 (/ 1.0 (- 4.0 (/ (+ 4.0 (/ -4.0 x)) x))))
double code(double x) {
	return 1.0 / (4.0 - ((4.0 + (-4.0 / x)) / x));
}
real(8) function code(x)
    real(8), intent (in) :: x
    code = 1.0d0 / (4.0d0 - ((4.0d0 + ((-4.0d0) / x)) / x))
end function
public static double code(double x) {
	return 1.0 / (4.0 - ((4.0 + (-4.0 / x)) / x));
}
def code(x):
	return 1.0 / (4.0 - ((4.0 + (-4.0 / x)) / x))
function code(x)
	return Float64(1.0 / Float64(4.0 - Float64(Float64(4.0 + Float64(-4.0 / x)) / x)))
end
function tmp = code(x)
	tmp = 1.0 / (4.0 - ((4.0 + (-4.0 / x)) / x));
end
code[x_] := N[(1.0 / N[(4.0 - N[(N[(4.0 + N[(-4.0 / x), $MachinePrecision]), $MachinePrecision] / x), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}

\\
\frac{1}{4 - \frac{4 + \frac{-4}{x}}{x}}
\end{array}
Derivation
  1. Initial program 74.6%

    \[1 - \sqrt{0.5 \cdot \left(1 + \frac{1}{\mathsf{hypot}\left(1, x\right)}\right)} \]
  2. Step-by-step derivation
    1. distribute-lft-in74.6%

      \[\leadsto 1 - \sqrt{\color{blue}{0.5 \cdot 1 + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}}} \]
    2. metadata-eval74.6%

      \[\leadsto 1 - \sqrt{\color{blue}{0.5} + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}} \]
    3. associate-*r/74.6%

      \[\leadsto 1 - \sqrt{0.5 + \color{blue}{\frac{0.5 \cdot 1}{\mathsf{hypot}\left(1, x\right)}}} \]
    4. metadata-eval74.6%

      \[\leadsto 1 - \sqrt{0.5 + \frac{\color{blue}{0.5}}{\mathsf{hypot}\left(1, x\right)}} \]
  3. Simplified74.6%

    \[\leadsto \color{blue}{1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
  4. Add Preprocessing
  5. Step-by-step derivation
    1. flip--74.6%

      \[\leadsto \color{blue}{\frac{1 \cdot 1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}} \cdot \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
    2. div-inv74.6%

      \[\leadsto \color{blue}{\left(1 \cdot 1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}} \cdot \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
    3. metadata-eval74.6%

      \[\leadsto \left(\color{blue}{1} - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}} \cdot \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    4. add-sqr-sqrt75.3%

      \[\leadsto \left(1 - \color{blue}{\left(0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right)}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    5. associate--r+75.4%

      \[\leadsto \color{blue}{\left(\left(1 - 0.5\right) - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right)} \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    6. metadata-eval75.4%

      \[\leadsto \left(\color{blue}{0.5} - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
  6. Applied egg-rr75.4%

    \[\leadsto \color{blue}{\left(0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
  7. Step-by-step derivation
    1. *-commutative75.4%

      \[\leadsto \color{blue}{\frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \cdot \left(0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right)} \]
    2. associate-/r/75.4%

      \[\leadsto \color{blue}{\frac{1}{\frac{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
  8. Simplified75.4%

    \[\leadsto \color{blue}{\frac{1}{\frac{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
  9. Taylor expanded in x around 0 38.2%

    \[\leadsto \frac{1}{\frac{\color{blue}{2}}{0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
  10. Taylor expanded in x around -inf 40.6%

    \[\leadsto \frac{1}{\color{blue}{4 + -1 \cdot \frac{4 - 4 \cdot \frac{1}{x}}{x}}} \]
  11. Step-by-step derivation
    1. mul-1-neg40.6%

      \[\leadsto \frac{1}{4 + \color{blue}{\left(-\frac{4 - 4 \cdot \frac{1}{x}}{x}\right)}} \]
    2. unsub-neg40.6%

      \[\leadsto \frac{1}{\color{blue}{4 - \frac{4 - 4 \cdot \frac{1}{x}}{x}}} \]
    3. sub-neg40.6%

      \[\leadsto \frac{1}{4 - \frac{\color{blue}{4 + \left(-4 \cdot \frac{1}{x}\right)}}{x}} \]
    4. associate-*r/40.6%

      \[\leadsto \frac{1}{4 - \frac{4 + \left(-\color{blue}{\frac{4 \cdot 1}{x}}\right)}{x}} \]
    5. metadata-eval40.6%

      \[\leadsto \frac{1}{4 - \frac{4 + \left(-\frac{\color{blue}{4}}{x}\right)}{x}} \]
    6. distribute-neg-frac40.6%

      \[\leadsto \frac{1}{4 - \frac{4 + \color{blue}{\frac{-4}{x}}}{x}} \]
    7. metadata-eval40.6%

      \[\leadsto \frac{1}{4 - \frac{4 + \frac{\color{blue}{-4}}{x}}{x}} \]
  12. Simplified40.6%

    \[\leadsto \frac{1}{\color{blue}{4 - \frac{4 + \frac{-4}{x}}{x}}} \]
  13. Add Preprocessing

Alternative 15: 32.4% accurate, 34.9× speedup?

\[\begin{array}{l} \\ \begin{array}{l} \mathbf{if}\;x \leq 2.15 \cdot 10^{-77}:\\ \;\;\;\;0\\ \mathbf{else}:\\ \;\;\;\;0.25\\ \end{array} \end{array} \]
(FPCore (x) :precision binary64 (if (<= x 2.15e-77) 0.0 0.25))
double code(double x) {
	double tmp;
	if (x <= 2.15e-77) {
		tmp = 0.0;
	} else {
		tmp = 0.25;
	}
	return tmp;
}
real(8) function code(x)
    real(8), intent (in) :: x
    real(8) :: tmp
    if (x <= 2.15d-77) then
        tmp = 0.0d0
    else
        tmp = 0.25d0
    end if
    code = tmp
end function
public static double code(double x) {
	double tmp;
	if (x <= 2.15e-77) {
		tmp = 0.0;
	} else {
		tmp = 0.25;
	}
	return tmp;
}
def code(x):
	tmp = 0
	if x <= 2.15e-77:
		tmp = 0.0
	else:
		tmp = 0.25
	return tmp
function code(x)
	tmp = 0.0
	if (x <= 2.15e-77)
		tmp = 0.0;
	else
		tmp = 0.25;
	end
	return tmp
end
function tmp_2 = code(x)
	tmp = 0.0;
	if (x <= 2.15e-77)
		tmp = 0.0;
	else
		tmp = 0.25;
	end
	tmp_2 = tmp;
end
code[x_] := If[LessEqual[x, 2.15e-77], 0.0, 0.25]
\begin{array}{l}

\\
\begin{array}{l}
\mathbf{if}\;x \leq 2.15 \cdot 10^{-77}:\\
\;\;\;\;0\\

\mathbf{else}:\\
\;\;\;\;0.25\\


\end{array}
\end{array}
Derivation
  1. Split input into 2 regimes
  2. if x < 2.1500000000000001e-77

    1. Initial program 72.8%

      \[1 - \sqrt{0.5 \cdot \left(1 + \frac{1}{\mathsf{hypot}\left(1, x\right)}\right)} \]
    2. Step-by-step derivation
      1. distribute-lft-in72.8%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5 \cdot 1 + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}}} \]
      2. metadata-eval72.8%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5} + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}} \]
      3. associate-*r/72.8%

        \[\leadsto 1 - \sqrt{0.5 + \color{blue}{\frac{0.5 \cdot 1}{\mathsf{hypot}\left(1, x\right)}}} \]
      4. metadata-eval72.8%

        \[\leadsto 1 - \sqrt{0.5 + \frac{\color{blue}{0.5}}{\mathsf{hypot}\left(1, x\right)}} \]
    3. Simplified72.8%

      \[\leadsto \color{blue}{1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    4. Add Preprocessing
    5. Taylor expanded in x around 0 40.0%

      \[\leadsto 1 - \color{blue}{1} \]
    6. Step-by-step derivation
      1. metadata-eval40.0%

        \[\leadsto \color{blue}{0} \]
    7. Applied egg-rr40.0%

      \[\leadsto \color{blue}{0} \]

    if 2.1500000000000001e-77 < x

    1. Initial program 78.3%

      \[1 - \sqrt{0.5 \cdot \left(1 + \frac{1}{\mathsf{hypot}\left(1, x\right)}\right)} \]
    2. Step-by-step derivation
      1. distribute-lft-in78.3%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5 \cdot 1 + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}}} \]
      2. metadata-eval78.3%

        \[\leadsto 1 - \sqrt{\color{blue}{0.5} + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}} \]
      3. associate-*r/78.3%

        \[\leadsto 1 - \sqrt{0.5 + \color{blue}{\frac{0.5 \cdot 1}{\mathsf{hypot}\left(1, x\right)}}} \]
      4. metadata-eval78.3%

        \[\leadsto 1 - \sqrt{0.5 + \frac{\color{blue}{0.5}}{\mathsf{hypot}\left(1, x\right)}} \]
    3. Simplified78.3%

      \[\leadsto \color{blue}{1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    4. Add Preprocessing
    5. Step-by-step derivation
      1. flip--78.2%

        \[\leadsto \color{blue}{\frac{1 \cdot 1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}} \cdot \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
      2. div-inv78.2%

        \[\leadsto \color{blue}{\left(1 \cdot 1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}} \cdot \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
      3. metadata-eval78.2%

        \[\leadsto \left(\color{blue}{1} - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}} \cdot \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
      4. add-sqr-sqrt79.4%

        \[\leadsto \left(1 - \color{blue}{\left(0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right)}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
      5. associate--r+79.5%

        \[\leadsto \color{blue}{\left(\left(1 - 0.5\right) - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right)} \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
      6. metadata-eval79.5%

        \[\leadsto \left(\color{blue}{0.5} - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    6. Applied egg-rr79.5%

      \[\leadsto \color{blue}{\left(0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right) \cdot \frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
    7. Step-by-step derivation
      1. *-commutative79.5%

        \[\leadsto \color{blue}{\frac{1}{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \cdot \left(0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}\right)} \]
      2. associate-/r/79.5%

        \[\leadsto \color{blue}{\frac{1}{\frac{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
    8. Simplified79.5%

      \[\leadsto \color{blue}{\frac{1}{\frac{1 + \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}{0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}}} \]
    9. Taylor expanded in x around 0 20.6%

      \[\leadsto \frac{1}{\frac{\color{blue}{2}}{0.5 - \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
    10. Taylor expanded in x around inf 18.7%

      \[\leadsto \color{blue}{0.25} \]
  3. Recombined 2 regimes into one program.
  4. Add Preprocessing

Alternative 16: 27.3% accurate, 210.0× speedup?

\[\begin{array}{l} \\ 0 \end{array} \]
(FPCore (x) :precision binary64 0.0)
double code(double x) {
	return 0.0;
}
real(8) function code(x)
    real(8), intent (in) :: x
    code = 0.0d0
end function
public static double code(double x) {
	return 0.0;
}
def code(x):
	return 0.0
function code(x)
	return 0.0
end
function tmp = code(x)
	tmp = 0.0;
end
code[x_] := 0.0
\begin{array}{l}

\\
0
\end{array}
Derivation
  1. Initial program 74.6%

    \[1 - \sqrt{0.5 \cdot \left(1 + \frac{1}{\mathsf{hypot}\left(1, x\right)}\right)} \]
  2. Step-by-step derivation
    1. distribute-lft-in74.6%

      \[\leadsto 1 - \sqrt{\color{blue}{0.5 \cdot 1 + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}}} \]
    2. metadata-eval74.6%

      \[\leadsto 1 - \sqrt{\color{blue}{0.5} + 0.5 \cdot \frac{1}{\mathsf{hypot}\left(1, x\right)}} \]
    3. associate-*r/74.6%

      \[\leadsto 1 - \sqrt{0.5 + \color{blue}{\frac{0.5 \cdot 1}{\mathsf{hypot}\left(1, x\right)}}} \]
    4. metadata-eval74.6%

      \[\leadsto 1 - \sqrt{0.5 + \frac{\color{blue}{0.5}}{\mathsf{hypot}\left(1, x\right)}} \]
  3. Simplified74.6%

    \[\leadsto \color{blue}{1 - \sqrt{0.5 + \frac{0.5}{\mathsf{hypot}\left(1, x\right)}}} \]
  4. Add Preprocessing
  5. Taylor expanded in x around 0 27.8%

    \[\leadsto 1 - \color{blue}{1} \]
  6. Step-by-step derivation
    1. metadata-eval27.8%

      \[\leadsto \color{blue}{0} \]
  7. Applied egg-rr27.8%

    \[\leadsto \color{blue}{0} \]
  8. Add Preprocessing

Reproduce

?
herbie shell --seed 2024181 
(FPCore (x)
  :name "Given's Rotation SVD example, simplified"
  :precision binary64
  (- 1.0 (sqrt (* 0.5 (+ 1.0 (/ 1.0 (hypot 1.0 x)))))))