Jmat.Real.dawson

Percentage Accurate: 53.4% → 100.0%
Time: 12.1s
Alternatives: 9
Speedup: 23.0×

Specification

?
\[\begin{array}{l} \\ \begin{array}{l} t_0 := \left(x \cdot x\right) \cdot \left(x \cdot x\right)\\ t_1 := t\_0 \cdot \left(x \cdot x\right)\\ t_2 := t\_1 \cdot \left(x \cdot x\right)\\ t_3 := t\_2 \cdot \left(x \cdot x\right)\\ \frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot t\_0\right) + 0.0072644182 \cdot t\_1\right) + 0.0005064034 \cdot t\_2\right) + 0.0001789971 \cdot t\_3}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot t\_0\right) + 0.0694555761 \cdot t\_1\right) + 0.0140005442 \cdot t\_2\right) + 0.0008327945 \cdot t\_3\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(t\_3 \cdot \left(x \cdot x\right)\right)} \cdot x \end{array} \end{array} \]
(FPCore (x)
 :precision binary64
 (let* ((t_0 (* (* x x) (* x x)))
        (t_1 (* t_0 (* x x)))
        (t_2 (* t_1 (* x x)))
        (t_3 (* t_2 (* x x))))
   (*
    (/
     (+
      (+
       (+
        (+ (+ 1.0 (* 0.1049934947 (* x x))) (* 0.0424060604 t_0))
        (* 0.0072644182 t_1))
       (* 0.0005064034 t_2))
      (* 0.0001789971 t_3))
     (+
      (+
       (+
        (+
         (+ (+ 1.0 (* 0.7715471019 (* x x))) (* 0.2909738639 t_0))
         (* 0.0694555761 t_1))
        (* 0.0140005442 t_2))
       (* 0.0008327945 t_3))
      (* (* 2.0 0.0001789971) (* t_3 (* x x)))))
    x)))
double code(double x) {
	double t_0 = (x * x) * (x * x);
	double t_1 = t_0 * (x * x);
	double t_2 = t_1 * (x * x);
	double t_3 = t_2 * (x * x);
	return ((((((1.0 + (0.1049934947 * (x * x))) + (0.0424060604 * t_0)) + (0.0072644182 * t_1)) + (0.0005064034 * t_2)) + (0.0001789971 * t_3)) / ((((((1.0 + (0.7715471019 * (x * x))) + (0.2909738639 * t_0)) + (0.0694555761 * t_1)) + (0.0140005442 * t_2)) + (0.0008327945 * t_3)) + ((2.0 * 0.0001789971) * (t_3 * (x * x))))) * x;
}
module fmin_fmax_functions
    implicit none
    private
    public fmax
    public fmin

    interface fmax
        module procedure fmax88
        module procedure fmax44
        module procedure fmax84
        module procedure fmax48
    end interface
    interface fmin
        module procedure fmin88
        module procedure fmin44
        module procedure fmin84
        module procedure fmin48
    end interface
contains
    real(8) function fmax88(x, y) result (res)
        real(8), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(x, max(x, y), y /= y), x /= x)
    end function
    real(4) function fmax44(x, y) result (res)
        real(4), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(y, merge(x, max(x, y), y /= y), x /= x)
    end function
    real(8) function fmax84(x, y) result(res)
        real(8), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(dble(y), merge(x, max(x, dble(y)), y /= y), x /= x)
    end function
    real(8) function fmax48(x, y) result(res)
        real(4), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(dble(x), max(dble(x), y), y /= y), x /= x)
    end function
    real(8) function fmin88(x, y) result (res)
        real(8), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(x, min(x, y), y /= y), x /= x)
    end function
    real(4) function fmin44(x, y) result (res)
        real(4), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(y, merge(x, min(x, y), y /= y), x /= x)
    end function
    real(8) function fmin84(x, y) result(res)
        real(8), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(dble(y), merge(x, min(x, dble(y)), y /= y), x /= x)
    end function
    real(8) function fmin48(x, y) result(res)
        real(4), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(dble(x), min(dble(x), y), y /= y), x /= x)
    end function
end module

real(8) function code(x)
use fmin_fmax_functions
    real(8), intent (in) :: x
    real(8) :: t_0
    real(8) :: t_1
    real(8) :: t_2
    real(8) :: t_3
    t_0 = (x * x) * (x * x)
    t_1 = t_0 * (x * x)
    t_2 = t_1 * (x * x)
    t_3 = t_2 * (x * x)
    code = ((((((1.0d0 + (0.1049934947d0 * (x * x))) + (0.0424060604d0 * t_0)) + (0.0072644182d0 * t_1)) + (0.0005064034d0 * t_2)) + (0.0001789971d0 * t_3)) / ((((((1.0d0 + (0.7715471019d0 * (x * x))) + (0.2909738639d0 * t_0)) + (0.0694555761d0 * t_1)) + (0.0140005442d0 * t_2)) + (0.0008327945d0 * t_3)) + ((2.0d0 * 0.0001789971d0) * (t_3 * (x * x))))) * x
end function
public static double code(double x) {
	double t_0 = (x * x) * (x * x);
	double t_1 = t_0 * (x * x);
	double t_2 = t_1 * (x * x);
	double t_3 = t_2 * (x * x);
	return ((((((1.0 + (0.1049934947 * (x * x))) + (0.0424060604 * t_0)) + (0.0072644182 * t_1)) + (0.0005064034 * t_2)) + (0.0001789971 * t_3)) / ((((((1.0 + (0.7715471019 * (x * x))) + (0.2909738639 * t_0)) + (0.0694555761 * t_1)) + (0.0140005442 * t_2)) + (0.0008327945 * t_3)) + ((2.0 * 0.0001789971) * (t_3 * (x * x))))) * x;
}
def code(x):
	t_0 = (x * x) * (x * x)
	t_1 = t_0 * (x * x)
	t_2 = t_1 * (x * x)
	t_3 = t_2 * (x * x)
	return ((((((1.0 + (0.1049934947 * (x * x))) + (0.0424060604 * t_0)) + (0.0072644182 * t_1)) + (0.0005064034 * t_2)) + (0.0001789971 * t_3)) / ((((((1.0 + (0.7715471019 * (x * x))) + (0.2909738639 * t_0)) + (0.0694555761 * t_1)) + (0.0140005442 * t_2)) + (0.0008327945 * t_3)) + ((2.0 * 0.0001789971) * (t_3 * (x * x))))) * x
function code(x)
	t_0 = Float64(Float64(x * x) * Float64(x * x))
	t_1 = Float64(t_0 * Float64(x * x))
	t_2 = Float64(t_1 * Float64(x * x))
	t_3 = Float64(t_2 * Float64(x * x))
	return Float64(Float64(Float64(Float64(Float64(Float64(Float64(1.0 + Float64(0.1049934947 * Float64(x * x))) + Float64(0.0424060604 * t_0)) + Float64(0.0072644182 * t_1)) + Float64(0.0005064034 * t_2)) + Float64(0.0001789971 * t_3)) / Float64(Float64(Float64(Float64(Float64(Float64(1.0 + Float64(0.7715471019 * Float64(x * x))) + Float64(0.2909738639 * t_0)) + Float64(0.0694555761 * t_1)) + Float64(0.0140005442 * t_2)) + Float64(0.0008327945 * t_3)) + Float64(Float64(2.0 * 0.0001789971) * Float64(t_3 * Float64(x * x))))) * x)
end
function tmp = code(x)
	t_0 = (x * x) * (x * x);
	t_1 = t_0 * (x * x);
	t_2 = t_1 * (x * x);
	t_3 = t_2 * (x * x);
	tmp = ((((((1.0 + (0.1049934947 * (x * x))) + (0.0424060604 * t_0)) + (0.0072644182 * t_1)) + (0.0005064034 * t_2)) + (0.0001789971 * t_3)) / ((((((1.0 + (0.7715471019 * (x * x))) + (0.2909738639 * t_0)) + (0.0694555761 * t_1)) + (0.0140005442 * t_2)) + (0.0008327945 * t_3)) + ((2.0 * 0.0001789971) * (t_3 * (x * x))))) * x;
end
code[x_] := Block[{t$95$0 = N[(N[(x * x), $MachinePrecision] * N[(x * x), $MachinePrecision]), $MachinePrecision]}, Block[{t$95$1 = N[(t$95$0 * N[(x * x), $MachinePrecision]), $MachinePrecision]}, Block[{t$95$2 = N[(t$95$1 * N[(x * x), $MachinePrecision]), $MachinePrecision]}, Block[{t$95$3 = N[(t$95$2 * N[(x * x), $MachinePrecision]), $MachinePrecision]}, N[(N[(N[(N[(N[(N[(N[(1.0 + N[(0.1049934947 * N[(x * x), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] + N[(0.0424060604 * t$95$0), $MachinePrecision]), $MachinePrecision] + N[(0.0072644182 * t$95$1), $MachinePrecision]), $MachinePrecision] + N[(0.0005064034 * t$95$2), $MachinePrecision]), $MachinePrecision] + N[(0.0001789971 * t$95$3), $MachinePrecision]), $MachinePrecision] / N[(N[(N[(N[(N[(N[(1.0 + N[(0.7715471019 * N[(x * x), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] + N[(0.2909738639 * t$95$0), $MachinePrecision]), $MachinePrecision] + N[(0.0694555761 * t$95$1), $MachinePrecision]), $MachinePrecision] + N[(0.0140005442 * t$95$2), $MachinePrecision]), $MachinePrecision] + N[(0.0008327945 * t$95$3), $MachinePrecision]), $MachinePrecision] + N[(N[(2.0 * 0.0001789971), $MachinePrecision] * N[(t$95$3 * N[(x * x), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] * x), $MachinePrecision]]]]]
\begin{array}{l}

\\
\begin{array}{l}
t_0 := \left(x \cdot x\right) \cdot \left(x \cdot x\right)\\
t_1 := t\_0 \cdot \left(x \cdot x\right)\\
t_2 := t\_1 \cdot \left(x \cdot x\right)\\
t_3 := t\_2 \cdot \left(x \cdot x\right)\\
\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot t\_0\right) + 0.0072644182 \cdot t\_1\right) + 0.0005064034 \cdot t\_2\right) + 0.0001789971 \cdot t\_3}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot t\_0\right) + 0.0694555761 \cdot t\_1\right) + 0.0140005442 \cdot t\_2\right) + 0.0008327945 \cdot t\_3\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(t\_3 \cdot \left(x \cdot x\right)\right)} \cdot x
\end{array}
\end{array}

Sampling outcomes in binary64 precision:

Local Percentage Accuracy vs ?

The average percentage accuracy by input value. Horizontal axis shows value of an input variable; the variable is choosen in the title. Vertical axis is accuracy; higher is better. Red represent the original program, while blue represents Herbie's suggestion. These can be toggled with buttons below the plot. The line is an average while dots represent individual samples.

Accuracy vs Speed?

Herbie found 9 alternatives:

AlternativeAccuracySpeedup
The accuracy (vertical axis) and speed (horizontal axis) of each alternatives. Up and to the right is better. The red square shows the initial program, and each blue circle shows an alternative.The line shows the best available speed-accuracy tradeoffs.

Initial Program: 53.4% accurate, 1.0× speedup?

\[\begin{array}{l} \\ \begin{array}{l} t_0 := \left(x \cdot x\right) \cdot \left(x \cdot x\right)\\ t_1 := t\_0 \cdot \left(x \cdot x\right)\\ t_2 := t\_1 \cdot \left(x \cdot x\right)\\ t_3 := t\_2 \cdot \left(x \cdot x\right)\\ \frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot t\_0\right) + 0.0072644182 \cdot t\_1\right) + 0.0005064034 \cdot t\_2\right) + 0.0001789971 \cdot t\_3}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot t\_0\right) + 0.0694555761 \cdot t\_1\right) + 0.0140005442 \cdot t\_2\right) + 0.0008327945 \cdot t\_3\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(t\_3 \cdot \left(x \cdot x\right)\right)} \cdot x \end{array} \end{array} \]
(FPCore (x)
 :precision binary64
 (let* ((t_0 (* (* x x) (* x x)))
        (t_1 (* t_0 (* x x)))
        (t_2 (* t_1 (* x x)))
        (t_3 (* t_2 (* x x))))
   (*
    (/
     (+
      (+
       (+
        (+ (+ 1.0 (* 0.1049934947 (* x x))) (* 0.0424060604 t_0))
        (* 0.0072644182 t_1))
       (* 0.0005064034 t_2))
      (* 0.0001789971 t_3))
     (+
      (+
       (+
        (+
         (+ (+ 1.0 (* 0.7715471019 (* x x))) (* 0.2909738639 t_0))
         (* 0.0694555761 t_1))
        (* 0.0140005442 t_2))
       (* 0.0008327945 t_3))
      (* (* 2.0 0.0001789971) (* t_3 (* x x)))))
    x)))
double code(double x) {
	double t_0 = (x * x) * (x * x);
	double t_1 = t_0 * (x * x);
	double t_2 = t_1 * (x * x);
	double t_3 = t_2 * (x * x);
	return ((((((1.0 + (0.1049934947 * (x * x))) + (0.0424060604 * t_0)) + (0.0072644182 * t_1)) + (0.0005064034 * t_2)) + (0.0001789971 * t_3)) / ((((((1.0 + (0.7715471019 * (x * x))) + (0.2909738639 * t_0)) + (0.0694555761 * t_1)) + (0.0140005442 * t_2)) + (0.0008327945 * t_3)) + ((2.0 * 0.0001789971) * (t_3 * (x * x))))) * x;
}
module fmin_fmax_functions
    implicit none
    private
    public fmax
    public fmin

    interface fmax
        module procedure fmax88
        module procedure fmax44
        module procedure fmax84
        module procedure fmax48
    end interface
    interface fmin
        module procedure fmin88
        module procedure fmin44
        module procedure fmin84
        module procedure fmin48
    end interface
contains
    real(8) function fmax88(x, y) result (res)
        real(8), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(x, max(x, y), y /= y), x /= x)
    end function
    real(4) function fmax44(x, y) result (res)
        real(4), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(y, merge(x, max(x, y), y /= y), x /= x)
    end function
    real(8) function fmax84(x, y) result(res)
        real(8), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(dble(y), merge(x, max(x, dble(y)), y /= y), x /= x)
    end function
    real(8) function fmax48(x, y) result(res)
        real(4), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(dble(x), max(dble(x), y), y /= y), x /= x)
    end function
    real(8) function fmin88(x, y) result (res)
        real(8), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(x, min(x, y), y /= y), x /= x)
    end function
    real(4) function fmin44(x, y) result (res)
        real(4), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(y, merge(x, min(x, y), y /= y), x /= x)
    end function
    real(8) function fmin84(x, y) result(res)
        real(8), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(dble(y), merge(x, min(x, dble(y)), y /= y), x /= x)
    end function
    real(8) function fmin48(x, y) result(res)
        real(4), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(dble(x), min(dble(x), y), y /= y), x /= x)
    end function
end module

real(8) function code(x)
use fmin_fmax_functions
    real(8), intent (in) :: x
    real(8) :: t_0
    real(8) :: t_1
    real(8) :: t_2
    real(8) :: t_3
    t_0 = (x * x) * (x * x)
    t_1 = t_0 * (x * x)
    t_2 = t_1 * (x * x)
    t_3 = t_2 * (x * x)
    code = ((((((1.0d0 + (0.1049934947d0 * (x * x))) + (0.0424060604d0 * t_0)) + (0.0072644182d0 * t_1)) + (0.0005064034d0 * t_2)) + (0.0001789971d0 * t_3)) / ((((((1.0d0 + (0.7715471019d0 * (x * x))) + (0.2909738639d0 * t_0)) + (0.0694555761d0 * t_1)) + (0.0140005442d0 * t_2)) + (0.0008327945d0 * t_3)) + ((2.0d0 * 0.0001789971d0) * (t_3 * (x * x))))) * x
end function
public static double code(double x) {
	double t_0 = (x * x) * (x * x);
	double t_1 = t_0 * (x * x);
	double t_2 = t_1 * (x * x);
	double t_3 = t_2 * (x * x);
	return ((((((1.0 + (0.1049934947 * (x * x))) + (0.0424060604 * t_0)) + (0.0072644182 * t_1)) + (0.0005064034 * t_2)) + (0.0001789971 * t_3)) / ((((((1.0 + (0.7715471019 * (x * x))) + (0.2909738639 * t_0)) + (0.0694555761 * t_1)) + (0.0140005442 * t_2)) + (0.0008327945 * t_3)) + ((2.0 * 0.0001789971) * (t_3 * (x * x))))) * x;
}
def code(x):
	t_0 = (x * x) * (x * x)
	t_1 = t_0 * (x * x)
	t_2 = t_1 * (x * x)
	t_3 = t_2 * (x * x)
	return ((((((1.0 + (0.1049934947 * (x * x))) + (0.0424060604 * t_0)) + (0.0072644182 * t_1)) + (0.0005064034 * t_2)) + (0.0001789971 * t_3)) / ((((((1.0 + (0.7715471019 * (x * x))) + (0.2909738639 * t_0)) + (0.0694555761 * t_1)) + (0.0140005442 * t_2)) + (0.0008327945 * t_3)) + ((2.0 * 0.0001789971) * (t_3 * (x * x))))) * x
function code(x)
	t_0 = Float64(Float64(x * x) * Float64(x * x))
	t_1 = Float64(t_0 * Float64(x * x))
	t_2 = Float64(t_1 * Float64(x * x))
	t_3 = Float64(t_2 * Float64(x * x))
	return Float64(Float64(Float64(Float64(Float64(Float64(Float64(1.0 + Float64(0.1049934947 * Float64(x * x))) + Float64(0.0424060604 * t_0)) + Float64(0.0072644182 * t_1)) + Float64(0.0005064034 * t_2)) + Float64(0.0001789971 * t_3)) / Float64(Float64(Float64(Float64(Float64(Float64(1.0 + Float64(0.7715471019 * Float64(x * x))) + Float64(0.2909738639 * t_0)) + Float64(0.0694555761 * t_1)) + Float64(0.0140005442 * t_2)) + Float64(0.0008327945 * t_3)) + Float64(Float64(2.0 * 0.0001789971) * Float64(t_3 * Float64(x * x))))) * x)
end
function tmp = code(x)
	t_0 = (x * x) * (x * x);
	t_1 = t_0 * (x * x);
	t_2 = t_1 * (x * x);
	t_3 = t_2 * (x * x);
	tmp = ((((((1.0 + (0.1049934947 * (x * x))) + (0.0424060604 * t_0)) + (0.0072644182 * t_1)) + (0.0005064034 * t_2)) + (0.0001789971 * t_3)) / ((((((1.0 + (0.7715471019 * (x * x))) + (0.2909738639 * t_0)) + (0.0694555761 * t_1)) + (0.0140005442 * t_2)) + (0.0008327945 * t_3)) + ((2.0 * 0.0001789971) * (t_3 * (x * x))))) * x;
end
code[x_] := Block[{t$95$0 = N[(N[(x * x), $MachinePrecision] * N[(x * x), $MachinePrecision]), $MachinePrecision]}, Block[{t$95$1 = N[(t$95$0 * N[(x * x), $MachinePrecision]), $MachinePrecision]}, Block[{t$95$2 = N[(t$95$1 * N[(x * x), $MachinePrecision]), $MachinePrecision]}, Block[{t$95$3 = N[(t$95$2 * N[(x * x), $MachinePrecision]), $MachinePrecision]}, N[(N[(N[(N[(N[(N[(N[(1.0 + N[(0.1049934947 * N[(x * x), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] + N[(0.0424060604 * t$95$0), $MachinePrecision]), $MachinePrecision] + N[(0.0072644182 * t$95$1), $MachinePrecision]), $MachinePrecision] + N[(0.0005064034 * t$95$2), $MachinePrecision]), $MachinePrecision] + N[(0.0001789971 * t$95$3), $MachinePrecision]), $MachinePrecision] / N[(N[(N[(N[(N[(N[(1.0 + N[(0.7715471019 * N[(x * x), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] + N[(0.2909738639 * t$95$0), $MachinePrecision]), $MachinePrecision] + N[(0.0694555761 * t$95$1), $MachinePrecision]), $MachinePrecision] + N[(0.0140005442 * t$95$2), $MachinePrecision]), $MachinePrecision] + N[(0.0008327945 * t$95$3), $MachinePrecision]), $MachinePrecision] + N[(N[(2.0 * 0.0001789971), $MachinePrecision] * N[(t$95$3 * N[(x * x), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] * x), $MachinePrecision]]]]]
\begin{array}{l}

\\
\begin{array}{l}
t_0 := \left(x \cdot x\right) \cdot \left(x \cdot x\right)\\
t_1 := t\_0 \cdot \left(x \cdot x\right)\\
t_2 := t\_1 \cdot \left(x \cdot x\right)\\
t_3 := t\_2 \cdot \left(x \cdot x\right)\\
\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot t\_0\right) + 0.0072644182 \cdot t\_1\right) + 0.0005064034 \cdot t\_2\right) + 0.0001789971 \cdot t\_3}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot t\_0\right) + 0.0694555761 \cdot t\_1\right) + 0.0140005442 \cdot t\_2\right) + 0.0008327945 \cdot t\_3\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(t\_3 \cdot \left(x \cdot x\right)\right)} \cdot x
\end{array}
\end{array}

Alternative 1: 100.0% accurate, 1.0× speedup?

\[\begin{array}{l} x\_m = \left|x\right| \\ x\_s = \mathsf{copysign}\left(1, x\right) \\ \begin{array}{l} t_0 := \left(x\_m \cdot x\_m\right) \cdot \left(x\_m \cdot x\_m\right)\\ t_1 := t\_0 \cdot \left(x\_m \cdot x\_m\right)\\ t_2 := t\_1 \cdot \left(x\_m \cdot x\_m\right)\\ t_3 := t\_2 \cdot \left(x\_m \cdot x\_m\right)\\ x\_s \cdot \begin{array}{l} \mathbf{if}\;x\_m \leq 20000:\\ \;\;\;\;\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x\_m \cdot x\_m\right)\right) + 0.0424060604 \cdot t\_0\right) + 0.0072644182 \cdot t\_1\right) + 0.0005064034 \cdot t\_2\right) + 0.0001789971 \cdot t\_3}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x\_m \cdot x\_m\right)\right) + 0.2909738639 \cdot t\_0\right) + 0.0694555761 \cdot t\_1\right) + 0.0140005442 \cdot t\_2\right) + 0.0008327945 \cdot t\_3\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(t\_3 \cdot \left(x\_m \cdot x\_m\right)\right)} \cdot x\_m\\ \mathbf{else}:\\ \;\;\;\;\frac{\frac{0.2514179000665374}{x\_m \cdot x\_m} + 0.5}{x\_m}\\ \end{array} \end{array} \end{array} \]
x\_m = (fabs.f64 x)
x\_s = (copysign.f64 #s(literal 1 binary64) x)
(FPCore (x_s x_m)
 :precision binary64
 (let* ((t_0 (* (* x_m x_m) (* x_m x_m)))
        (t_1 (* t_0 (* x_m x_m)))
        (t_2 (* t_1 (* x_m x_m)))
        (t_3 (* t_2 (* x_m x_m))))
   (*
    x_s
    (if (<= x_m 20000.0)
      (*
       (/
        (+
         (+
          (+
           (+ (+ 1.0 (* 0.1049934947 (* x_m x_m))) (* 0.0424060604 t_0))
           (* 0.0072644182 t_1))
          (* 0.0005064034 t_2))
         (* 0.0001789971 t_3))
        (+
         (+
          (+
           (+
            (+ (+ 1.0 (* 0.7715471019 (* x_m x_m))) (* 0.2909738639 t_0))
            (* 0.0694555761 t_1))
           (* 0.0140005442 t_2))
          (* 0.0008327945 t_3))
         (* (* 2.0 0.0001789971) (* t_3 (* x_m x_m)))))
       x_m)
      (/ (+ (/ 0.2514179000665374 (* x_m x_m)) 0.5) x_m)))))
x\_m = fabs(x);
x\_s = copysign(1.0, x);
double code(double x_s, double x_m) {
	double t_0 = (x_m * x_m) * (x_m * x_m);
	double t_1 = t_0 * (x_m * x_m);
	double t_2 = t_1 * (x_m * x_m);
	double t_3 = t_2 * (x_m * x_m);
	double tmp;
	if (x_m <= 20000.0) {
		tmp = ((((((1.0 + (0.1049934947 * (x_m * x_m))) + (0.0424060604 * t_0)) + (0.0072644182 * t_1)) + (0.0005064034 * t_2)) + (0.0001789971 * t_3)) / ((((((1.0 + (0.7715471019 * (x_m * x_m))) + (0.2909738639 * t_0)) + (0.0694555761 * t_1)) + (0.0140005442 * t_2)) + (0.0008327945 * t_3)) + ((2.0 * 0.0001789971) * (t_3 * (x_m * x_m))))) * x_m;
	} else {
		tmp = ((0.2514179000665374 / (x_m * x_m)) + 0.5) / x_m;
	}
	return x_s * tmp;
}
x\_m =     private
x\_s =     private
module fmin_fmax_functions
    implicit none
    private
    public fmax
    public fmin

    interface fmax
        module procedure fmax88
        module procedure fmax44
        module procedure fmax84
        module procedure fmax48
    end interface
    interface fmin
        module procedure fmin88
        module procedure fmin44
        module procedure fmin84
        module procedure fmin48
    end interface
contains
    real(8) function fmax88(x, y) result (res)
        real(8), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(x, max(x, y), y /= y), x /= x)
    end function
    real(4) function fmax44(x, y) result (res)
        real(4), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(y, merge(x, max(x, y), y /= y), x /= x)
    end function
    real(8) function fmax84(x, y) result(res)
        real(8), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(dble(y), merge(x, max(x, dble(y)), y /= y), x /= x)
    end function
    real(8) function fmax48(x, y) result(res)
        real(4), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(dble(x), max(dble(x), y), y /= y), x /= x)
    end function
    real(8) function fmin88(x, y) result (res)
        real(8), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(x, min(x, y), y /= y), x /= x)
    end function
    real(4) function fmin44(x, y) result (res)
        real(4), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(y, merge(x, min(x, y), y /= y), x /= x)
    end function
    real(8) function fmin84(x, y) result(res)
        real(8), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(dble(y), merge(x, min(x, dble(y)), y /= y), x /= x)
    end function
    real(8) function fmin48(x, y) result(res)
        real(4), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(dble(x), min(dble(x), y), y /= y), x /= x)
    end function
end module

real(8) function code(x_s, x_m)
use fmin_fmax_functions
    real(8), intent (in) :: x_s
    real(8), intent (in) :: x_m
    real(8) :: t_0
    real(8) :: t_1
    real(8) :: t_2
    real(8) :: t_3
    real(8) :: tmp
    t_0 = (x_m * x_m) * (x_m * x_m)
    t_1 = t_0 * (x_m * x_m)
    t_2 = t_1 * (x_m * x_m)
    t_3 = t_2 * (x_m * x_m)
    if (x_m <= 20000.0d0) then
        tmp = ((((((1.0d0 + (0.1049934947d0 * (x_m * x_m))) + (0.0424060604d0 * t_0)) + (0.0072644182d0 * t_1)) + (0.0005064034d0 * t_2)) + (0.0001789971d0 * t_3)) / ((((((1.0d0 + (0.7715471019d0 * (x_m * x_m))) + (0.2909738639d0 * t_0)) + (0.0694555761d0 * t_1)) + (0.0140005442d0 * t_2)) + (0.0008327945d0 * t_3)) + ((2.0d0 * 0.0001789971d0) * (t_3 * (x_m * x_m))))) * x_m
    else
        tmp = ((0.2514179000665374d0 / (x_m * x_m)) + 0.5d0) / x_m
    end if
    code = x_s * tmp
end function
x\_m = Math.abs(x);
x\_s = Math.copySign(1.0, x);
public static double code(double x_s, double x_m) {
	double t_0 = (x_m * x_m) * (x_m * x_m);
	double t_1 = t_0 * (x_m * x_m);
	double t_2 = t_1 * (x_m * x_m);
	double t_3 = t_2 * (x_m * x_m);
	double tmp;
	if (x_m <= 20000.0) {
		tmp = ((((((1.0 + (0.1049934947 * (x_m * x_m))) + (0.0424060604 * t_0)) + (0.0072644182 * t_1)) + (0.0005064034 * t_2)) + (0.0001789971 * t_3)) / ((((((1.0 + (0.7715471019 * (x_m * x_m))) + (0.2909738639 * t_0)) + (0.0694555761 * t_1)) + (0.0140005442 * t_2)) + (0.0008327945 * t_3)) + ((2.0 * 0.0001789971) * (t_3 * (x_m * x_m))))) * x_m;
	} else {
		tmp = ((0.2514179000665374 / (x_m * x_m)) + 0.5) / x_m;
	}
	return x_s * tmp;
}
x\_m = math.fabs(x)
x\_s = math.copysign(1.0, x)
def code(x_s, x_m):
	t_0 = (x_m * x_m) * (x_m * x_m)
	t_1 = t_0 * (x_m * x_m)
	t_2 = t_1 * (x_m * x_m)
	t_3 = t_2 * (x_m * x_m)
	tmp = 0
	if x_m <= 20000.0:
		tmp = ((((((1.0 + (0.1049934947 * (x_m * x_m))) + (0.0424060604 * t_0)) + (0.0072644182 * t_1)) + (0.0005064034 * t_2)) + (0.0001789971 * t_3)) / ((((((1.0 + (0.7715471019 * (x_m * x_m))) + (0.2909738639 * t_0)) + (0.0694555761 * t_1)) + (0.0140005442 * t_2)) + (0.0008327945 * t_3)) + ((2.0 * 0.0001789971) * (t_3 * (x_m * x_m))))) * x_m
	else:
		tmp = ((0.2514179000665374 / (x_m * x_m)) + 0.5) / x_m
	return x_s * tmp
x\_m = abs(x)
x\_s = copysign(1.0, x)
function code(x_s, x_m)
	t_0 = Float64(Float64(x_m * x_m) * Float64(x_m * x_m))
	t_1 = Float64(t_0 * Float64(x_m * x_m))
	t_2 = Float64(t_1 * Float64(x_m * x_m))
	t_3 = Float64(t_2 * Float64(x_m * x_m))
	tmp = 0.0
	if (x_m <= 20000.0)
		tmp = Float64(Float64(Float64(Float64(Float64(Float64(Float64(1.0 + Float64(0.1049934947 * Float64(x_m * x_m))) + Float64(0.0424060604 * t_0)) + Float64(0.0072644182 * t_1)) + Float64(0.0005064034 * t_2)) + Float64(0.0001789971 * t_3)) / Float64(Float64(Float64(Float64(Float64(Float64(1.0 + Float64(0.7715471019 * Float64(x_m * x_m))) + Float64(0.2909738639 * t_0)) + Float64(0.0694555761 * t_1)) + Float64(0.0140005442 * t_2)) + Float64(0.0008327945 * t_3)) + Float64(Float64(2.0 * 0.0001789971) * Float64(t_3 * Float64(x_m * x_m))))) * x_m);
	else
		tmp = Float64(Float64(Float64(0.2514179000665374 / Float64(x_m * x_m)) + 0.5) / x_m);
	end
	return Float64(x_s * tmp)
end
x\_m = abs(x);
x\_s = sign(x) * abs(1.0);
function tmp_2 = code(x_s, x_m)
	t_0 = (x_m * x_m) * (x_m * x_m);
	t_1 = t_0 * (x_m * x_m);
	t_2 = t_1 * (x_m * x_m);
	t_3 = t_2 * (x_m * x_m);
	tmp = 0.0;
	if (x_m <= 20000.0)
		tmp = ((((((1.0 + (0.1049934947 * (x_m * x_m))) + (0.0424060604 * t_0)) + (0.0072644182 * t_1)) + (0.0005064034 * t_2)) + (0.0001789971 * t_3)) / ((((((1.0 + (0.7715471019 * (x_m * x_m))) + (0.2909738639 * t_0)) + (0.0694555761 * t_1)) + (0.0140005442 * t_2)) + (0.0008327945 * t_3)) + ((2.0 * 0.0001789971) * (t_3 * (x_m * x_m))))) * x_m;
	else
		tmp = ((0.2514179000665374 / (x_m * x_m)) + 0.5) / x_m;
	end
	tmp_2 = x_s * tmp;
end
x\_m = N[Abs[x], $MachinePrecision]
x\_s = N[With[{TMP1 = Abs[1.0], TMP2 = Sign[x]}, TMP1 * If[TMP2 == 0, 1, TMP2]], $MachinePrecision]
code[x$95$s_, x$95$m_] := Block[{t$95$0 = N[(N[(x$95$m * x$95$m), $MachinePrecision] * N[(x$95$m * x$95$m), $MachinePrecision]), $MachinePrecision]}, Block[{t$95$1 = N[(t$95$0 * N[(x$95$m * x$95$m), $MachinePrecision]), $MachinePrecision]}, Block[{t$95$2 = N[(t$95$1 * N[(x$95$m * x$95$m), $MachinePrecision]), $MachinePrecision]}, Block[{t$95$3 = N[(t$95$2 * N[(x$95$m * x$95$m), $MachinePrecision]), $MachinePrecision]}, N[(x$95$s * If[LessEqual[x$95$m, 20000.0], N[(N[(N[(N[(N[(N[(N[(1.0 + N[(0.1049934947 * N[(x$95$m * x$95$m), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] + N[(0.0424060604 * t$95$0), $MachinePrecision]), $MachinePrecision] + N[(0.0072644182 * t$95$1), $MachinePrecision]), $MachinePrecision] + N[(0.0005064034 * t$95$2), $MachinePrecision]), $MachinePrecision] + N[(0.0001789971 * t$95$3), $MachinePrecision]), $MachinePrecision] / N[(N[(N[(N[(N[(N[(1.0 + N[(0.7715471019 * N[(x$95$m * x$95$m), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] + N[(0.2909738639 * t$95$0), $MachinePrecision]), $MachinePrecision] + N[(0.0694555761 * t$95$1), $MachinePrecision]), $MachinePrecision] + N[(0.0140005442 * t$95$2), $MachinePrecision]), $MachinePrecision] + N[(0.0008327945 * t$95$3), $MachinePrecision]), $MachinePrecision] + N[(N[(2.0 * 0.0001789971), $MachinePrecision] * N[(t$95$3 * N[(x$95$m * x$95$m), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] * x$95$m), $MachinePrecision], N[(N[(N[(0.2514179000665374 / N[(x$95$m * x$95$m), $MachinePrecision]), $MachinePrecision] + 0.5), $MachinePrecision] / x$95$m), $MachinePrecision]]), $MachinePrecision]]]]]
\begin{array}{l}
x\_m = \left|x\right|
\\
x\_s = \mathsf{copysign}\left(1, x\right)

\\
\begin{array}{l}
t_0 := \left(x\_m \cdot x\_m\right) \cdot \left(x\_m \cdot x\_m\right)\\
t_1 := t\_0 \cdot \left(x\_m \cdot x\_m\right)\\
t_2 := t\_1 \cdot \left(x\_m \cdot x\_m\right)\\
t_3 := t\_2 \cdot \left(x\_m \cdot x\_m\right)\\
x\_s \cdot \begin{array}{l}
\mathbf{if}\;x\_m \leq 20000:\\
\;\;\;\;\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x\_m \cdot x\_m\right)\right) + 0.0424060604 \cdot t\_0\right) + 0.0072644182 \cdot t\_1\right) + 0.0005064034 \cdot t\_2\right) + 0.0001789971 \cdot t\_3}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x\_m \cdot x\_m\right)\right) + 0.2909738639 \cdot t\_0\right) + 0.0694555761 \cdot t\_1\right) + 0.0140005442 \cdot t\_2\right) + 0.0008327945 \cdot t\_3\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(t\_3 \cdot \left(x\_m \cdot x\_m\right)\right)} \cdot x\_m\\

\mathbf{else}:\\
\;\;\;\;\frac{\frac{0.2514179000665374}{x\_m \cdot x\_m} + 0.5}{x\_m}\\


\end{array}
\end{array}
\end{array}
Derivation
  1. Split input into 2 regimes
  2. if x < 2e4

    1. Initial program 69.4%

      \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
    2. Add Preprocessing

    if 2e4 < x

    1. Initial program 6.0%

      \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
    2. Add Preprocessing
    3. Taylor expanded in x around inf

      \[\leadsto \color{blue}{\frac{\frac{1}{2} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}}{x}} \]
    4. Step-by-step derivation
      1. lower-/.f64N/A

        \[\leadsto \color{blue}{\frac{\frac{1}{2} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}}{x}} \]
      2. +-commutativeN/A

        \[\leadsto \frac{\color{blue}{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} + \frac{1}{2}}}{x} \]
      3. lower-+.f64N/A

        \[\leadsto \frac{\color{blue}{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} + \frac{1}{2}}}{x} \]
      4. associate-*r/N/A

        \[\leadsto \frac{\color{blue}{\frac{\frac{600041}{2386628} \cdot 1}{{x}^{2}}} + \frac{1}{2}}{x} \]
      5. metadata-evalN/A

        \[\leadsto \frac{\frac{\color{blue}{\frac{600041}{2386628}}}{{x}^{2}} + \frac{1}{2}}{x} \]
      6. lower-/.f64N/A

        \[\leadsto \frac{\color{blue}{\frac{\frac{600041}{2386628}}{{x}^{2}}} + \frac{1}{2}}{x} \]
      7. unpow2N/A

        \[\leadsto \frac{\frac{\frac{600041}{2386628}}{\color{blue}{x \cdot x}} + \frac{1}{2}}{x} \]
      8. lower-*.f64100.0

        \[\leadsto \frac{\frac{0.2514179000665374}{\color{blue}{x \cdot x}} + 0.5}{x} \]
    5. Applied rewrites100.0%

      \[\leadsto \color{blue}{\frac{\frac{0.2514179000665374}{x \cdot x} + 0.5}{x}} \]
  3. Recombined 2 regimes into one program.
  4. Add Preprocessing

Alternative 2: 99.7% accurate, 2.4× speedup?

\[\begin{array}{l} x\_m = \left|x\right| \\ x\_s = \mathsf{copysign}\left(1, x\right) \\ x\_s \cdot \begin{array}{l} \mathbf{if}\;x\_m \leq 1.45:\\ \;\;\;\;\mathsf{fma}\left(\mathsf{fma}\left(x\_m \cdot x\_m, \left(x\_m \cdot x\_m\right) \cdot -0.0732490286039007, 0.265709700396151 \cdot \left(x\_m \cdot x\_m\right) - 0.6665536072\right), x\_m \cdot x\_m, 1\right) \cdot x\_m\\ \mathbf{else}:\\ \;\;\;\;\frac{\frac{\frac{-11.259630434457211}{x\_m \cdot x\_m} + -0.15298196345929074}{{x\_m}^{4}} - \left(\frac{0.2514179000665374}{x\_m \cdot x\_m} - -0.5\right)}{-x\_m}\\ \end{array} \end{array} \]
x\_m = (fabs.f64 x)
x\_s = (copysign.f64 #s(literal 1 binary64) x)
(FPCore (x_s x_m)
 :precision binary64
 (*
  x_s
  (if (<= x_m 1.45)
    (*
     (fma
      (fma
       (* x_m x_m)
       (* (* x_m x_m) -0.0732490286039007)
       (- (* 0.265709700396151 (* x_m x_m)) 0.6665536072))
      (* x_m x_m)
      1.0)
     x_m)
    (/
     (-
      (/
       (+ (/ -11.259630434457211 (* x_m x_m)) -0.15298196345929074)
       (pow x_m 4.0))
      (- (/ 0.2514179000665374 (* x_m x_m)) -0.5))
     (- x_m)))))
x\_m = fabs(x);
x\_s = copysign(1.0, x);
double code(double x_s, double x_m) {
	double tmp;
	if (x_m <= 1.45) {
		tmp = fma(fma((x_m * x_m), ((x_m * x_m) * -0.0732490286039007), ((0.265709700396151 * (x_m * x_m)) - 0.6665536072)), (x_m * x_m), 1.0) * x_m;
	} else {
		tmp = ((((-11.259630434457211 / (x_m * x_m)) + -0.15298196345929074) / pow(x_m, 4.0)) - ((0.2514179000665374 / (x_m * x_m)) - -0.5)) / -x_m;
	}
	return x_s * tmp;
}
x\_m = abs(x)
x\_s = copysign(1.0, x)
function code(x_s, x_m)
	tmp = 0.0
	if (x_m <= 1.45)
		tmp = Float64(fma(fma(Float64(x_m * x_m), Float64(Float64(x_m * x_m) * -0.0732490286039007), Float64(Float64(0.265709700396151 * Float64(x_m * x_m)) - 0.6665536072)), Float64(x_m * x_m), 1.0) * x_m);
	else
		tmp = Float64(Float64(Float64(Float64(Float64(-11.259630434457211 / Float64(x_m * x_m)) + -0.15298196345929074) / (x_m ^ 4.0)) - Float64(Float64(0.2514179000665374 / Float64(x_m * x_m)) - -0.5)) / Float64(-x_m));
	end
	return Float64(x_s * tmp)
end
x\_m = N[Abs[x], $MachinePrecision]
x\_s = N[With[{TMP1 = Abs[1.0], TMP2 = Sign[x]}, TMP1 * If[TMP2 == 0, 1, TMP2]], $MachinePrecision]
code[x$95$s_, x$95$m_] := N[(x$95$s * If[LessEqual[x$95$m, 1.45], N[(N[(N[(N[(x$95$m * x$95$m), $MachinePrecision] * N[(N[(x$95$m * x$95$m), $MachinePrecision] * -0.0732490286039007), $MachinePrecision] + N[(N[(0.265709700396151 * N[(x$95$m * x$95$m), $MachinePrecision]), $MachinePrecision] - 0.6665536072), $MachinePrecision]), $MachinePrecision] * N[(x$95$m * x$95$m), $MachinePrecision] + 1.0), $MachinePrecision] * x$95$m), $MachinePrecision], N[(N[(N[(N[(N[(-11.259630434457211 / N[(x$95$m * x$95$m), $MachinePrecision]), $MachinePrecision] + -0.15298196345929074), $MachinePrecision] / N[Power[x$95$m, 4.0], $MachinePrecision]), $MachinePrecision] - N[(N[(0.2514179000665374 / N[(x$95$m * x$95$m), $MachinePrecision]), $MachinePrecision] - -0.5), $MachinePrecision]), $MachinePrecision] / (-x$95$m)), $MachinePrecision]]), $MachinePrecision]
\begin{array}{l}
x\_m = \left|x\right|
\\
x\_s = \mathsf{copysign}\left(1, x\right)

\\
x\_s \cdot \begin{array}{l}
\mathbf{if}\;x\_m \leq 1.45:\\
\;\;\;\;\mathsf{fma}\left(\mathsf{fma}\left(x\_m \cdot x\_m, \left(x\_m \cdot x\_m\right) \cdot -0.0732490286039007, 0.265709700396151 \cdot \left(x\_m \cdot x\_m\right) - 0.6665536072\right), x\_m \cdot x\_m, 1\right) \cdot x\_m\\

\mathbf{else}:\\
\;\;\;\;\frac{\frac{\frac{-11.259630434457211}{x\_m \cdot x\_m} + -0.15298196345929074}{{x\_m}^{4}} - \left(\frac{0.2514179000665374}{x\_m \cdot x\_m} - -0.5\right)}{-x\_m}\\


\end{array}
\end{array}
Derivation
  1. Split input into 2 regimes
  2. if x < 1.44999999999999996

    1. Initial program 69.1%

      \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
    2. Add Preprocessing
    3. Taylor expanded in x around 0

      \[\leadsto \color{blue}{\left(1 + {x}^{2} \cdot \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} + \frac{-9156128575487588197208397249}{125000000000000000000000000000} \cdot {x}^{2}\right) - \frac{833192009}{1250000000}\right)\right)} \cdot x \]
    4. Step-by-step derivation
      1. fp-cancel-sign-sub-invN/A

        \[\leadsto \color{blue}{\left(1 - \left(\mathsf{neg}\left({x}^{2}\right)\right) \cdot \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} + \frac{-9156128575487588197208397249}{125000000000000000000000000000} \cdot {x}^{2}\right) - \frac{833192009}{1250000000}\right)\right)} \cdot x \]
      2. fp-cancel-sub-sign-invN/A

        \[\leadsto \color{blue}{\left(1 + \left(\mathsf{neg}\left(\left(\mathsf{neg}\left({x}^{2}\right)\right)\right)\right) \cdot \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} + \frac{-9156128575487588197208397249}{125000000000000000000000000000} \cdot {x}^{2}\right) - \frac{833192009}{1250000000}\right)\right)} \cdot x \]
      3. +-commutativeN/A

        \[\leadsto \color{blue}{\left(\left(\mathsf{neg}\left(\left(\mathsf{neg}\left({x}^{2}\right)\right)\right)\right) \cdot \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} + \frac{-9156128575487588197208397249}{125000000000000000000000000000} \cdot {x}^{2}\right) - \frac{833192009}{1250000000}\right) + 1\right)} \cdot x \]
      4. distribute-lft-neg-outN/A

        \[\leadsto \left(\color{blue}{\left(\mathsf{neg}\left(\left(\mathsf{neg}\left({x}^{2}\right)\right) \cdot \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} + \frac{-9156128575487588197208397249}{125000000000000000000000000000} \cdot {x}^{2}\right) - \frac{833192009}{1250000000}\right)\right)\right)} + 1\right) \cdot x \]
      5. distribute-lft-neg-outN/A

        \[\leadsto \left(\left(\mathsf{neg}\left(\color{blue}{\left(\mathsf{neg}\left({x}^{2} \cdot \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} + \frac{-9156128575487588197208397249}{125000000000000000000000000000} \cdot {x}^{2}\right) - \frac{833192009}{1250000000}\right)\right)\right)}\right)\right) + 1\right) \cdot x \]
      6. remove-double-negN/A

        \[\leadsto \left(\color{blue}{{x}^{2} \cdot \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} + \frac{-9156128575487588197208397249}{125000000000000000000000000000} \cdot {x}^{2}\right) - \frac{833192009}{1250000000}\right)} + 1\right) \cdot x \]
      7. *-commutativeN/A

        \[\leadsto \left(\color{blue}{\left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} + \frac{-9156128575487588197208397249}{125000000000000000000000000000} \cdot {x}^{2}\right) - \frac{833192009}{1250000000}\right) \cdot {x}^{2}} + 1\right) \cdot x \]
      8. lower-fma.f64N/A

        \[\leadsto \color{blue}{\mathsf{fma}\left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} + \frac{-9156128575487588197208397249}{125000000000000000000000000000} \cdot {x}^{2}\right) - \frac{833192009}{1250000000}, {x}^{2}, 1\right)} \cdot x \]
    5. Applied rewrites67.3%

      \[\leadsto \color{blue}{\mathsf{fma}\left(\mathsf{fma}\left(-0.0732490286039007, {x}^{4}, 0.265709700396151 \cdot \left(x \cdot x\right) - 0.6665536072\right), x \cdot x, 1\right)} \cdot x \]
    6. Step-by-step derivation
      1. Applied rewrites67.3%

        \[\leadsto \mathsf{fma}\left(\mathsf{fma}\left(x \cdot x, \left(x \cdot x\right) \cdot -0.0732490286039007, 0.265709700396151 \cdot \left(x \cdot x\right) - 0.6665536072\right), \color{blue}{x} \cdot x, 1\right) \cdot x \]

      if 1.44999999999999996 < x

      1. Initial program 9.6%

        \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
      2. Add Preprocessing
      3. Applied rewrites5.7%

        \[\leadsto \color{blue}{\left(\frac{\mathsf{fma}\left(0.0001789971, {x}^{10}, \mathsf{fma}\left(0.0005064034, {x}^{8}, \mathsf{fma}\left(0.0072644182, {x}^{6}, \mathsf{fma}\left(0.0424060604, {x}^{4}, \mathsf{fma}\left(0.1049934947, x \cdot x, 1\right)\right)\right)\right)\right)}{\mathsf{fma}\left(4.5880481982529165 \cdot 10^{-11}, {\left({x}^{12}\right)}^{3}, {\left(\mathsf{fma}\left(0.0008327945, {x}^{10}, \mathsf{fma}\left(0.0140005442, {x}^{8}, \mathsf{fma}\left(0.0694555761, {x}^{6}, \mathsf{fma}\left(0.2909738639, {x}^{4}, \mathsf{fma}\left(0.7715471019, x \cdot x, 1\right)\right)\right)\right)\right)\right)}^{3}\right)} \cdot \mathsf{fma}\left(0.0003579942 \cdot {x}^{12}, 0.0003579942 \cdot {x}^{12} - \mathsf{fma}\left(0.0008327945, {x}^{10}, \mathsf{fma}\left(0.0140005442, {x}^{8}, \mathsf{fma}\left(0.0694555761, {x}^{6}, \mathsf{fma}\left(0.2909738639, {x}^{4}, \mathsf{fma}\left(0.7715471019, x \cdot x, 1\right)\right)\right)\right)\right), {\left(\mathsf{fma}\left(0.0008327945, {x}^{10}, \mathsf{fma}\left(0.0140005442, {x}^{8}, \mathsf{fma}\left(0.0694555761, {x}^{6}, \mathsf{fma}\left(0.2909738639, {x}^{4}, \mathsf{fma}\left(0.7715471019, x \cdot x, 1\right)\right)\right)\right)\right)\right)}^{2}\right)\right)} \cdot x \]
      4. Taylor expanded in x around -inf

        \[\leadsto \color{blue}{-1 \cdot \frac{-1 \cdot \frac{\frac{1307076337763}{8543989815576} + \frac{344398180852034095277}{30586987988352776592} \cdot \frac{1}{{x}^{2}}}{{x}^{4}} - \left(\frac{1}{2} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}\right)}{x}} \]
      5. Applied rewrites98.5%

        \[\leadsto \color{blue}{\frac{\frac{\frac{-11.259630434457211}{x \cdot x} + -0.15298196345929074}{{x}^{4}} - \left(\frac{0.2514179000665374}{x \cdot x} - -0.5\right)}{-x}} \]
    7. Recombined 2 regimes into one program.
    8. Add Preprocessing

    Alternative 3: 99.7% accurate, 2.7× speedup?

    \[\begin{array}{l} x\_m = \left|x\right| \\ x\_s = \mathsf{copysign}\left(1, x\right) \\ x\_s \cdot \begin{array}{l} \mathbf{if}\;x\_m \leq 1.2:\\ \;\;\;\;\mathsf{fma}\left(\mathsf{fma}\left(x\_m \cdot x\_m, \left(x\_m \cdot x\_m\right) \cdot -0.0732490286039007, 0.265709700396151 \cdot \left(x\_m \cdot x\_m\right) - 0.6665536072\right), x\_m \cdot x\_m, 1\right) \cdot x\_m\\ \mathbf{else}:\\ \;\;\;\;\frac{\left(\frac{0.15298196345929074}{{x\_m}^{4}} + 0.5\right) - \frac{-0.2514179000665374}{x\_m \cdot x\_m}}{x\_m}\\ \end{array} \end{array} \]
    x\_m = (fabs.f64 x)
    x\_s = (copysign.f64 #s(literal 1 binary64) x)
    (FPCore (x_s x_m)
     :precision binary64
     (*
      x_s
      (if (<= x_m 1.2)
        (*
         (fma
          (fma
           (* x_m x_m)
           (* (* x_m x_m) -0.0732490286039007)
           (- (* 0.265709700396151 (* x_m x_m)) 0.6665536072))
          (* x_m x_m)
          1.0)
         x_m)
        (/
         (-
          (+ (/ 0.15298196345929074 (pow x_m 4.0)) 0.5)
          (/ -0.2514179000665374 (* x_m x_m)))
         x_m))))
    x\_m = fabs(x);
    x\_s = copysign(1.0, x);
    double code(double x_s, double x_m) {
    	double tmp;
    	if (x_m <= 1.2) {
    		tmp = fma(fma((x_m * x_m), ((x_m * x_m) * -0.0732490286039007), ((0.265709700396151 * (x_m * x_m)) - 0.6665536072)), (x_m * x_m), 1.0) * x_m;
    	} else {
    		tmp = (((0.15298196345929074 / pow(x_m, 4.0)) + 0.5) - (-0.2514179000665374 / (x_m * x_m))) / x_m;
    	}
    	return x_s * tmp;
    }
    
    x\_m = abs(x)
    x\_s = copysign(1.0, x)
    function code(x_s, x_m)
    	tmp = 0.0
    	if (x_m <= 1.2)
    		tmp = Float64(fma(fma(Float64(x_m * x_m), Float64(Float64(x_m * x_m) * -0.0732490286039007), Float64(Float64(0.265709700396151 * Float64(x_m * x_m)) - 0.6665536072)), Float64(x_m * x_m), 1.0) * x_m);
    	else
    		tmp = Float64(Float64(Float64(Float64(0.15298196345929074 / (x_m ^ 4.0)) + 0.5) - Float64(-0.2514179000665374 / Float64(x_m * x_m))) / x_m);
    	end
    	return Float64(x_s * tmp)
    end
    
    x\_m = N[Abs[x], $MachinePrecision]
    x\_s = N[With[{TMP1 = Abs[1.0], TMP2 = Sign[x]}, TMP1 * If[TMP2 == 0, 1, TMP2]], $MachinePrecision]
    code[x$95$s_, x$95$m_] := N[(x$95$s * If[LessEqual[x$95$m, 1.2], N[(N[(N[(N[(x$95$m * x$95$m), $MachinePrecision] * N[(N[(x$95$m * x$95$m), $MachinePrecision] * -0.0732490286039007), $MachinePrecision] + N[(N[(0.265709700396151 * N[(x$95$m * x$95$m), $MachinePrecision]), $MachinePrecision] - 0.6665536072), $MachinePrecision]), $MachinePrecision] * N[(x$95$m * x$95$m), $MachinePrecision] + 1.0), $MachinePrecision] * x$95$m), $MachinePrecision], N[(N[(N[(N[(0.15298196345929074 / N[Power[x$95$m, 4.0], $MachinePrecision]), $MachinePrecision] + 0.5), $MachinePrecision] - N[(-0.2514179000665374 / N[(x$95$m * x$95$m), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] / x$95$m), $MachinePrecision]]), $MachinePrecision]
    
    \begin{array}{l}
    x\_m = \left|x\right|
    \\
    x\_s = \mathsf{copysign}\left(1, x\right)
    
    \\
    x\_s \cdot \begin{array}{l}
    \mathbf{if}\;x\_m \leq 1.2:\\
    \;\;\;\;\mathsf{fma}\left(\mathsf{fma}\left(x\_m \cdot x\_m, \left(x\_m \cdot x\_m\right) \cdot -0.0732490286039007, 0.265709700396151 \cdot \left(x\_m \cdot x\_m\right) - 0.6665536072\right), x\_m \cdot x\_m, 1\right) \cdot x\_m\\
    
    \mathbf{else}:\\
    \;\;\;\;\frac{\left(\frac{0.15298196345929074}{{x\_m}^{4}} + 0.5\right) - \frac{-0.2514179000665374}{x\_m \cdot x\_m}}{x\_m}\\
    
    
    \end{array}
    \end{array}
    
    Derivation
    1. Split input into 2 regimes
    2. if x < 1.19999999999999996

      1. Initial program 69.1%

        \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
      2. Add Preprocessing
      3. Taylor expanded in x around 0

        \[\leadsto \color{blue}{\left(1 + {x}^{2} \cdot \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} + \frac{-9156128575487588197208397249}{125000000000000000000000000000} \cdot {x}^{2}\right) - \frac{833192009}{1250000000}\right)\right)} \cdot x \]
      4. Step-by-step derivation
        1. fp-cancel-sign-sub-invN/A

          \[\leadsto \color{blue}{\left(1 - \left(\mathsf{neg}\left({x}^{2}\right)\right) \cdot \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} + \frac{-9156128575487588197208397249}{125000000000000000000000000000} \cdot {x}^{2}\right) - \frac{833192009}{1250000000}\right)\right)} \cdot x \]
        2. fp-cancel-sub-sign-invN/A

          \[\leadsto \color{blue}{\left(1 + \left(\mathsf{neg}\left(\left(\mathsf{neg}\left({x}^{2}\right)\right)\right)\right) \cdot \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} + \frac{-9156128575487588197208397249}{125000000000000000000000000000} \cdot {x}^{2}\right) - \frac{833192009}{1250000000}\right)\right)} \cdot x \]
        3. +-commutativeN/A

          \[\leadsto \color{blue}{\left(\left(\mathsf{neg}\left(\left(\mathsf{neg}\left({x}^{2}\right)\right)\right)\right) \cdot \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} + \frac{-9156128575487588197208397249}{125000000000000000000000000000} \cdot {x}^{2}\right) - \frac{833192009}{1250000000}\right) + 1\right)} \cdot x \]
        4. distribute-lft-neg-outN/A

          \[\leadsto \left(\color{blue}{\left(\mathsf{neg}\left(\left(\mathsf{neg}\left({x}^{2}\right)\right) \cdot \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} + \frac{-9156128575487588197208397249}{125000000000000000000000000000} \cdot {x}^{2}\right) - \frac{833192009}{1250000000}\right)\right)\right)} + 1\right) \cdot x \]
        5. distribute-lft-neg-outN/A

          \[\leadsto \left(\left(\mathsf{neg}\left(\color{blue}{\left(\mathsf{neg}\left({x}^{2} \cdot \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} + \frac{-9156128575487588197208397249}{125000000000000000000000000000} \cdot {x}^{2}\right) - \frac{833192009}{1250000000}\right)\right)\right)}\right)\right) + 1\right) \cdot x \]
        6. remove-double-negN/A

          \[\leadsto \left(\color{blue}{{x}^{2} \cdot \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} + \frac{-9156128575487588197208397249}{125000000000000000000000000000} \cdot {x}^{2}\right) - \frac{833192009}{1250000000}\right)} + 1\right) \cdot x \]
        7. *-commutativeN/A

          \[\leadsto \left(\color{blue}{\left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} + \frac{-9156128575487588197208397249}{125000000000000000000000000000} \cdot {x}^{2}\right) - \frac{833192009}{1250000000}\right) \cdot {x}^{2}} + 1\right) \cdot x \]
        8. lower-fma.f64N/A

          \[\leadsto \color{blue}{\mathsf{fma}\left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} + \frac{-9156128575487588197208397249}{125000000000000000000000000000} \cdot {x}^{2}\right) - \frac{833192009}{1250000000}, {x}^{2}, 1\right)} \cdot x \]
      5. Applied rewrites67.3%

        \[\leadsto \color{blue}{\mathsf{fma}\left(\mathsf{fma}\left(-0.0732490286039007, {x}^{4}, 0.265709700396151 \cdot \left(x \cdot x\right) - 0.6665536072\right), x \cdot x, 1\right)} \cdot x \]
      6. Step-by-step derivation
        1. Applied rewrites67.3%

          \[\leadsto \mathsf{fma}\left(\mathsf{fma}\left(x \cdot x, \left(x \cdot x\right) \cdot -0.0732490286039007, 0.265709700396151 \cdot \left(x \cdot x\right) - 0.6665536072\right), \color{blue}{x} \cdot x, 1\right) \cdot x \]

        if 1.19999999999999996 < x

        1. Initial program 9.6%

          \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
        2. Add Preprocessing
        3. Taylor expanded in x around inf

          \[\leadsto \color{blue}{\frac{\frac{1}{2} + \left(\frac{\frac{1307076337763}{8543989815576}}{{x}^{4}} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}\right)}{x}} \]
        4. Step-by-step derivation
          1. lower-/.f64N/A

            \[\leadsto \color{blue}{\frac{\frac{1}{2} + \left(\frac{\frac{1307076337763}{8543989815576}}{{x}^{4}} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}\right)}{x}} \]
          2. associate-+r+N/A

            \[\leadsto \frac{\color{blue}{\left(\frac{1}{2} + \frac{\frac{1307076337763}{8543989815576}}{{x}^{4}}\right) + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}}}{x} \]
          3. fp-cancel-sign-sub-invN/A

            \[\leadsto \frac{\color{blue}{\left(\frac{1}{2} + \frac{\frac{1307076337763}{8543989815576}}{{x}^{4}}\right) - \left(\mathsf{neg}\left(\frac{600041}{2386628}\right)\right) \cdot \frac{1}{{x}^{2}}}}{x} \]
          4. lower--.f64N/A

            \[\leadsto \frac{\color{blue}{\left(\frac{1}{2} + \frac{\frac{1307076337763}{8543989815576}}{{x}^{4}}\right) - \left(\mathsf{neg}\left(\frac{600041}{2386628}\right)\right) \cdot \frac{1}{{x}^{2}}}}{x} \]
          5. +-commutativeN/A

            \[\leadsto \frac{\color{blue}{\left(\frac{\frac{1307076337763}{8543989815576}}{{x}^{4}} + \frac{1}{2}\right)} - \left(\mathsf{neg}\left(\frac{600041}{2386628}\right)\right) \cdot \frac{1}{{x}^{2}}}{x} \]
          6. lower-+.f64N/A

            \[\leadsto \frac{\color{blue}{\left(\frac{\frac{1307076337763}{8543989815576}}{{x}^{4}} + \frac{1}{2}\right)} - \left(\mathsf{neg}\left(\frac{600041}{2386628}\right)\right) \cdot \frac{1}{{x}^{2}}}{x} \]
          7. lower-/.f64N/A

            \[\leadsto \frac{\left(\color{blue}{\frac{\frac{1307076337763}{8543989815576}}{{x}^{4}}} + \frac{1}{2}\right) - \left(\mathsf{neg}\left(\frac{600041}{2386628}\right)\right) \cdot \frac{1}{{x}^{2}}}{x} \]
          8. lower-pow.f64N/A

            \[\leadsto \frac{\left(\frac{\frac{1307076337763}{8543989815576}}{\color{blue}{{x}^{4}}} + \frac{1}{2}\right) - \left(\mathsf{neg}\left(\frac{600041}{2386628}\right)\right) \cdot \frac{1}{{x}^{2}}}{x} \]
          9. associate-*r/N/A

            \[\leadsto \frac{\left(\frac{\frac{1307076337763}{8543989815576}}{{x}^{4}} + \frac{1}{2}\right) - \color{blue}{\frac{\left(\mathsf{neg}\left(\frac{600041}{2386628}\right)\right) \cdot 1}{{x}^{2}}}}{x} \]
          10. metadata-evalN/A

            \[\leadsto \frac{\left(\frac{\frac{1307076337763}{8543989815576}}{{x}^{4}} + \frac{1}{2}\right) - \frac{\color{blue}{\frac{-600041}{2386628}} \cdot 1}{{x}^{2}}}{x} \]
          11. metadata-evalN/A

            \[\leadsto \frac{\left(\frac{\frac{1307076337763}{8543989815576}}{{x}^{4}} + \frac{1}{2}\right) - \frac{\color{blue}{\frac{-600041}{2386628}}}{{x}^{2}}}{x} \]
          12. metadata-evalN/A

            \[\leadsto \frac{\left(\frac{\frac{1307076337763}{8543989815576}}{{x}^{4}} + \frac{1}{2}\right) - \frac{\color{blue}{\mathsf{neg}\left(\frac{600041}{2386628}\right)}}{{x}^{2}}}{x} \]
          13. lower-/.f64N/A

            \[\leadsto \frac{\left(\frac{\frac{1307076337763}{8543989815576}}{{x}^{4}} + \frac{1}{2}\right) - \color{blue}{\frac{\mathsf{neg}\left(\frac{600041}{2386628}\right)}{{x}^{2}}}}{x} \]
          14. metadata-evalN/A

            \[\leadsto \frac{\left(\frac{\frac{1307076337763}{8543989815576}}{{x}^{4}} + \frac{1}{2}\right) - \frac{\color{blue}{\frac{-600041}{2386628}}}{{x}^{2}}}{x} \]
          15. unpow2N/A

            \[\leadsto \frac{\left(\frac{\frac{1307076337763}{8543989815576}}{{x}^{4}} + \frac{1}{2}\right) - \frac{\frac{-600041}{2386628}}{\color{blue}{x \cdot x}}}{x} \]
          16. lower-*.f6498.2

            \[\leadsto \frac{\left(\frac{0.15298196345929074}{{x}^{4}} + 0.5\right) - \frac{-0.2514179000665374}{\color{blue}{x \cdot x}}}{x} \]
        5. Applied rewrites98.2%

          \[\leadsto \color{blue}{\frac{\left(\frac{0.15298196345929074}{{x}^{4}} + 0.5\right) - \frac{-0.2514179000665374}{x \cdot x}}{x}} \]
      7. Recombined 2 regimes into one program.
      8. Add Preprocessing

      Alternative 4: 99.6% accurate, 7.3× speedup?

      \[\begin{array}{l} x\_m = \left|x\right| \\ x\_s = \mathsf{copysign}\left(1, x\right) \\ x\_s \cdot \begin{array}{l} \mathbf{if}\;x\_m \leq 1.15:\\ \;\;\;\;\mathsf{fma}\left(\mathsf{fma}\left(x\_m \cdot x\_m, \left(x\_m \cdot x\_m\right) \cdot -0.0732490286039007, 0.265709700396151 \cdot \left(x\_m \cdot x\_m\right) - 0.6665536072\right), x\_m \cdot x\_m, 1\right) \cdot x\_m\\ \mathbf{else}:\\ \;\;\;\;\frac{\frac{0.2514179000665374}{x\_m \cdot x\_m} + 0.5}{x\_m}\\ \end{array} \end{array} \]
      x\_m = (fabs.f64 x)
      x\_s = (copysign.f64 #s(literal 1 binary64) x)
      (FPCore (x_s x_m)
       :precision binary64
       (*
        x_s
        (if (<= x_m 1.15)
          (*
           (fma
            (fma
             (* x_m x_m)
             (* (* x_m x_m) -0.0732490286039007)
             (- (* 0.265709700396151 (* x_m x_m)) 0.6665536072))
            (* x_m x_m)
            1.0)
           x_m)
          (/ (+ (/ 0.2514179000665374 (* x_m x_m)) 0.5) x_m))))
      x\_m = fabs(x);
      x\_s = copysign(1.0, x);
      double code(double x_s, double x_m) {
      	double tmp;
      	if (x_m <= 1.15) {
      		tmp = fma(fma((x_m * x_m), ((x_m * x_m) * -0.0732490286039007), ((0.265709700396151 * (x_m * x_m)) - 0.6665536072)), (x_m * x_m), 1.0) * x_m;
      	} else {
      		tmp = ((0.2514179000665374 / (x_m * x_m)) + 0.5) / x_m;
      	}
      	return x_s * tmp;
      }
      
      x\_m = abs(x)
      x\_s = copysign(1.0, x)
      function code(x_s, x_m)
      	tmp = 0.0
      	if (x_m <= 1.15)
      		tmp = Float64(fma(fma(Float64(x_m * x_m), Float64(Float64(x_m * x_m) * -0.0732490286039007), Float64(Float64(0.265709700396151 * Float64(x_m * x_m)) - 0.6665536072)), Float64(x_m * x_m), 1.0) * x_m);
      	else
      		tmp = Float64(Float64(Float64(0.2514179000665374 / Float64(x_m * x_m)) + 0.5) / x_m);
      	end
      	return Float64(x_s * tmp)
      end
      
      x\_m = N[Abs[x], $MachinePrecision]
      x\_s = N[With[{TMP1 = Abs[1.0], TMP2 = Sign[x]}, TMP1 * If[TMP2 == 0, 1, TMP2]], $MachinePrecision]
      code[x$95$s_, x$95$m_] := N[(x$95$s * If[LessEqual[x$95$m, 1.15], N[(N[(N[(N[(x$95$m * x$95$m), $MachinePrecision] * N[(N[(x$95$m * x$95$m), $MachinePrecision] * -0.0732490286039007), $MachinePrecision] + N[(N[(0.265709700396151 * N[(x$95$m * x$95$m), $MachinePrecision]), $MachinePrecision] - 0.6665536072), $MachinePrecision]), $MachinePrecision] * N[(x$95$m * x$95$m), $MachinePrecision] + 1.0), $MachinePrecision] * x$95$m), $MachinePrecision], N[(N[(N[(0.2514179000665374 / N[(x$95$m * x$95$m), $MachinePrecision]), $MachinePrecision] + 0.5), $MachinePrecision] / x$95$m), $MachinePrecision]]), $MachinePrecision]
      
      \begin{array}{l}
      x\_m = \left|x\right|
      \\
      x\_s = \mathsf{copysign}\left(1, x\right)
      
      \\
      x\_s \cdot \begin{array}{l}
      \mathbf{if}\;x\_m \leq 1.15:\\
      \;\;\;\;\mathsf{fma}\left(\mathsf{fma}\left(x\_m \cdot x\_m, \left(x\_m \cdot x\_m\right) \cdot -0.0732490286039007, 0.265709700396151 \cdot \left(x\_m \cdot x\_m\right) - 0.6665536072\right), x\_m \cdot x\_m, 1\right) \cdot x\_m\\
      
      \mathbf{else}:\\
      \;\;\;\;\frac{\frac{0.2514179000665374}{x\_m \cdot x\_m} + 0.5}{x\_m}\\
      
      
      \end{array}
      \end{array}
      
      Derivation
      1. Split input into 2 regimes
      2. if x < 1.1499999999999999

        1. Initial program 69.1%

          \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
        2. Add Preprocessing
        3. Taylor expanded in x around 0

          \[\leadsto \color{blue}{\left(1 + {x}^{2} \cdot \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} + \frac{-9156128575487588197208397249}{125000000000000000000000000000} \cdot {x}^{2}\right) - \frac{833192009}{1250000000}\right)\right)} \cdot x \]
        4. Step-by-step derivation
          1. fp-cancel-sign-sub-invN/A

            \[\leadsto \color{blue}{\left(1 - \left(\mathsf{neg}\left({x}^{2}\right)\right) \cdot \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} + \frac{-9156128575487588197208397249}{125000000000000000000000000000} \cdot {x}^{2}\right) - \frac{833192009}{1250000000}\right)\right)} \cdot x \]
          2. fp-cancel-sub-sign-invN/A

            \[\leadsto \color{blue}{\left(1 + \left(\mathsf{neg}\left(\left(\mathsf{neg}\left({x}^{2}\right)\right)\right)\right) \cdot \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} + \frac{-9156128575487588197208397249}{125000000000000000000000000000} \cdot {x}^{2}\right) - \frac{833192009}{1250000000}\right)\right)} \cdot x \]
          3. +-commutativeN/A

            \[\leadsto \color{blue}{\left(\left(\mathsf{neg}\left(\left(\mathsf{neg}\left({x}^{2}\right)\right)\right)\right) \cdot \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} + \frac{-9156128575487588197208397249}{125000000000000000000000000000} \cdot {x}^{2}\right) - \frac{833192009}{1250000000}\right) + 1\right)} \cdot x \]
          4. distribute-lft-neg-outN/A

            \[\leadsto \left(\color{blue}{\left(\mathsf{neg}\left(\left(\mathsf{neg}\left({x}^{2}\right)\right) \cdot \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} + \frac{-9156128575487588197208397249}{125000000000000000000000000000} \cdot {x}^{2}\right) - \frac{833192009}{1250000000}\right)\right)\right)} + 1\right) \cdot x \]
          5. distribute-lft-neg-outN/A

            \[\leadsto \left(\left(\mathsf{neg}\left(\color{blue}{\left(\mathsf{neg}\left({x}^{2} \cdot \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} + \frac{-9156128575487588197208397249}{125000000000000000000000000000} \cdot {x}^{2}\right) - \frac{833192009}{1250000000}\right)\right)\right)}\right)\right) + 1\right) \cdot x \]
          6. remove-double-negN/A

            \[\leadsto \left(\color{blue}{{x}^{2} \cdot \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} + \frac{-9156128575487588197208397249}{125000000000000000000000000000} \cdot {x}^{2}\right) - \frac{833192009}{1250000000}\right)} + 1\right) \cdot x \]
          7. *-commutativeN/A

            \[\leadsto \left(\color{blue}{\left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} + \frac{-9156128575487588197208397249}{125000000000000000000000000000} \cdot {x}^{2}\right) - \frac{833192009}{1250000000}\right) \cdot {x}^{2}} + 1\right) \cdot x \]
          8. lower-fma.f64N/A

            \[\leadsto \color{blue}{\mathsf{fma}\left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} + \frac{-9156128575487588197208397249}{125000000000000000000000000000} \cdot {x}^{2}\right) - \frac{833192009}{1250000000}, {x}^{2}, 1\right)} \cdot x \]
        5. Applied rewrites67.3%

          \[\leadsto \color{blue}{\mathsf{fma}\left(\mathsf{fma}\left(-0.0732490286039007, {x}^{4}, 0.265709700396151 \cdot \left(x \cdot x\right) - 0.6665536072\right), x \cdot x, 1\right)} \cdot x \]
        6. Step-by-step derivation
          1. Applied rewrites67.3%

            \[\leadsto \mathsf{fma}\left(\mathsf{fma}\left(x \cdot x, \left(x \cdot x\right) \cdot -0.0732490286039007, 0.265709700396151 \cdot \left(x \cdot x\right) - 0.6665536072\right), \color{blue}{x} \cdot x, 1\right) \cdot x \]

          if 1.1499999999999999 < x

          1. Initial program 9.6%

            \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
          2. Add Preprocessing
          3. Taylor expanded in x around inf

            \[\leadsto \color{blue}{\frac{\frac{1}{2} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}}{x}} \]
          4. Step-by-step derivation
            1. lower-/.f64N/A

              \[\leadsto \color{blue}{\frac{\frac{1}{2} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}}{x}} \]
            2. +-commutativeN/A

              \[\leadsto \frac{\color{blue}{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} + \frac{1}{2}}}{x} \]
            3. lower-+.f64N/A

              \[\leadsto \frac{\color{blue}{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} + \frac{1}{2}}}{x} \]
            4. associate-*r/N/A

              \[\leadsto \frac{\color{blue}{\frac{\frac{600041}{2386628} \cdot 1}{{x}^{2}}} + \frac{1}{2}}{x} \]
            5. metadata-evalN/A

              \[\leadsto \frac{\frac{\color{blue}{\frac{600041}{2386628}}}{{x}^{2}} + \frac{1}{2}}{x} \]
            6. lower-/.f64N/A

              \[\leadsto \frac{\color{blue}{\frac{\frac{600041}{2386628}}{{x}^{2}}} + \frac{1}{2}}{x} \]
            7. unpow2N/A

              \[\leadsto \frac{\frac{\frac{600041}{2386628}}{\color{blue}{x \cdot x}} + \frac{1}{2}}{x} \]
            8. lower-*.f6497.9

              \[\leadsto \frac{\frac{0.2514179000665374}{\color{blue}{x \cdot x}} + 0.5}{x} \]
          5. Applied rewrites97.9%

            \[\leadsto \color{blue}{\frac{\frac{0.2514179000665374}{x \cdot x} + 0.5}{x}} \]
        7. Recombined 2 regimes into one program.
        8. Add Preprocessing

        Alternative 5: 99.6% accurate, 11.2× speedup?

        \[\begin{array}{l} x\_m = \left|x\right| \\ x\_s = \mathsf{copysign}\left(1, x\right) \\ x\_s \cdot \begin{array}{l} \mathbf{if}\;x\_m \leq 1.1:\\ \;\;\;\;\mathsf{fma}\left(\left(\left(x\_m \cdot x\_m\right) \cdot 0.265709700396151 - 0.6665536072\right) \cdot \left(x\_m \cdot x\_m\right), x\_m, x\_m\right)\\ \mathbf{else}:\\ \;\;\;\;\frac{\frac{0.2514179000665374}{x\_m \cdot x\_m} + 0.5}{x\_m}\\ \end{array} \end{array} \]
        x\_m = (fabs.f64 x)
        x\_s = (copysign.f64 #s(literal 1 binary64) x)
        (FPCore (x_s x_m)
         :precision binary64
         (*
          x_s
          (if (<= x_m 1.1)
            (fma
             (* (- (* (* x_m x_m) 0.265709700396151) 0.6665536072) (* x_m x_m))
             x_m
             x_m)
            (/ (+ (/ 0.2514179000665374 (* x_m x_m)) 0.5) x_m))))
        x\_m = fabs(x);
        x\_s = copysign(1.0, x);
        double code(double x_s, double x_m) {
        	double tmp;
        	if (x_m <= 1.1) {
        		tmp = fma(((((x_m * x_m) * 0.265709700396151) - 0.6665536072) * (x_m * x_m)), x_m, x_m);
        	} else {
        		tmp = ((0.2514179000665374 / (x_m * x_m)) + 0.5) / x_m;
        	}
        	return x_s * tmp;
        }
        
        x\_m = abs(x)
        x\_s = copysign(1.0, x)
        function code(x_s, x_m)
        	tmp = 0.0
        	if (x_m <= 1.1)
        		tmp = fma(Float64(Float64(Float64(Float64(x_m * x_m) * 0.265709700396151) - 0.6665536072) * Float64(x_m * x_m)), x_m, x_m);
        	else
        		tmp = Float64(Float64(Float64(0.2514179000665374 / Float64(x_m * x_m)) + 0.5) / x_m);
        	end
        	return Float64(x_s * tmp)
        end
        
        x\_m = N[Abs[x], $MachinePrecision]
        x\_s = N[With[{TMP1 = Abs[1.0], TMP2 = Sign[x]}, TMP1 * If[TMP2 == 0, 1, TMP2]], $MachinePrecision]
        code[x$95$s_, x$95$m_] := N[(x$95$s * If[LessEqual[x$95$m, 1.1], N[(N[(N[(N[(N[(x$95$m * x$95$m), $MachinePrecision] * 0.265709700396151), $MachinePrecision] - 0.6665536072), $MachinePrecision] * N[(x$95$m * x$95$m), $MachinePrecision]), $MachinePrecision] * x$95$m + x$95$m), $MachinePrecision], N[(N[(N[(0.2514179000665374 / N[(x$95$m * x$95$m), $MachinePrecision]), $MachinePrecision] + 0.5), $MachinePrecision] / x$95$m), $MachinePrecision]]), $MachinePrecision]
        
        \begin{array}{l}
        x\_m = \left|x\right|
        \\
        x\_s = \mathsf{copysign}\left(1, x\right)
        
        \\
        x\_s \cdot \begin{array}{l}
        \mathbf{if}\;x\_m \leq 1.1:\\
        \;\;\;\;\mathsf{fma}\left(\left(\left(x\_m \cdot x\_m\right) \cdot 0.265709700396151 - 0.6665536072\right) \cdot \left(x\_m \cdot x\_m\right), x\_m, x\_m\right)\\
        
        \mathbf{else}:\\
        \;\;\;\;\frac{\frac{0.2514179000665374}{x\_m \cdot x\_m} + 0.5}{x\_m}\\
        
        
        \end{array}
        \end{array}
        
        Derivation
        1. Split input into 2 regimes
        2. if x < 1.1000000000000001

          1. Initial program 69.1%

            \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
          2. Add Preprocessing
          3. Taylor expanded in x around 0

            \[\leadsto \color{blue}{x \cdot \left(1 + {x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right)\right)} \]
          4. Step-by-step derivation
            1. +-commutativeN/A

              \[\leadsto x \cdot \color{blue}{\left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right) + 1\right)} \]
            2. distribute-lft-inN/A

              \[\leadsto \color{blue}{x \cdot \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right)\right) + x \cdot 1} \]
            3. fp-cancel-sign-sub-invN/A

              \[\leadsto \color{blue}{x \cdot \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right)\right) - \left(\mathsf{neg}\left(x\right)\right) \cdot 1} \]
            4. distribute-lft-neg-inN/A

              \[\leadsto x \cdot \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right)\right) - \color{blue}{\left(\mathsf{neg}\left(x \cdot 1\right)\right)} \]
            5. distribute-rgt-neg-inN/A

              \[\leadsto x \cdot \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right)\right) - \color{blue}{x \cdot \left(\mathsf{neg}\left(1\right)\right)} \]
            6. metadata-evalN/A

              \[\leadsto x \cdot \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right)\right) - x \cdot \color{blue}{-1} \]
            7. fp-cancel-sub-sign-invN/A

              \[\leadsto \color{blue}{x \cdot \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right)\right) + \left(\mathsf{neg}\left(x\right)\right) \cdot -1} \]
            8. *-commutativeN/A

              \[\leadsto \color{blue}{\left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right)\right) \cdot x} + \left(\mathsf{neg}\left(x\right)\right) \cdot -1 \]
            9. *-commutativeN/A

              \[\leadsto \color{blue}{\left(\left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right) \cdot {x}^{2}\right)} \cdot x + \left(\mathsf{neg}\left(x\right)\right) \cdot -1 \]
            10. associate-*l*N/A

              \[\leadsto \color{blue}{\left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right) \cdot \left({x}^{2} \cdot x\right)} + \left(\mathsf{neg}\left(x\right)\right) \cdot -1 \]
            11. *-commutativeN/A

              \[\leadsto \color{blue}{\left({x}^{2} \cdot x\right) \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right)} + \left(\mathsf{neg}\left(x\right)\right) \cdot -1 \]
            12. distribute-lft-neg-inN/A

              \[\leadsto \left({x}^{2} \cdot x\right) \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right) + \color{blue}{\left(\mathsf{neg}\left(x \cdot -1\right)\right)} \]
            13. distribute-rgt-neg-inN/A

              \[\leadsto \left({x}^{2} \cdot x\right) \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right) + \color{blue}{x \cdot \left(\mathsf{neg}\left(-1\right)\right)} \]
            14. metadata-evalN/A

              \[\leadsto \left({x}^{2} \cdot x\right) \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right) + x \cdot \color{blue}{1} \]
            15. *-rgt-identityN/A

              \[\leadsto \left({x}^{2} \cdot x\right) \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right) + \color{blue}{x} \]
            16. lower-fma.f64N/A

              \[\leadsto \color{blue}{\mathsf{fma}\left({x}^{2} \cdot x, \frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}, x\right)} \]
          5. Applied rewrites67.8%

            \[\leadsto \color{blue}{\mathsf{fma}\left({x}^{3}, 0.265709700396151 \cdot \left(x \cdot x\right) - 0.6665536072, x\right)} \]
          6. Step-by-step derivation
            1. Applied rewrites67.8%

              \[\leadsto \mathsf{fma}\left(\left(\left(x \cdot x\right) \cdot 0.265709700396151 - 0.6665536072\right) \cdot \left(x \cdot x\right), \color{blue}{x}, x\right) \]

            if 1.1000000000000001 < x

            1. Initial program 9.6%

              \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
            2. Add Preprocessing
            3. Taylor expanded in x around inf

              \[\leadsto \color{blue}{\frac{\frac{1}{2} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}}{x}} \]
            4. Step-by-step derivation
              1. lower-/.f64N/A

                \[\leadsto \color{blue}{\frac{\frac{1}{2} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}}{x}} \]
              2. +-commutativeN/A

                \[\leadsto \frac{\color{blue}{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} + \frac{1}{2}}}{x} \]
              3. lower-+.f64N/A

                \[\leadsto \frac{\color{blue}{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} + \frac{1}{2}}}{x} \]
              4. associate-*r/N/A

                \[\leadsto \frac{\color{blue}{\frac{\frac{600041}{2386628} \cdot 1}{{x}^{2}}} + \frac{1}{2}}{x} \]
              5. metadata-evalN/A

                \[\leadsto \frac{\frac{\color{blue}{\frac{600041}{2386628}}}{{x}^{2}} + \frac{1}{2}}{x} \]
              6. lower-/.f64N/A

                \[\leadsto \frac{\color{blue}{\frac{\frac{600041}{2386628}}{{x}^{2}}} + \frac{1}{2}}{x} \]
              7. unpow2N/A

                \[\leadsto \frac{\frac{\frac{600041}{2386628}}{\color{blue}{x \cdot x}} + \frac{1}{2}}{x} \]
              8. lower-*.f6497.9

                \[\leadsto \frac{\frac{0.2514179000665374}{\color{blue}{x \cdot x}} + 0.5}{x} \]
            5. Applied rewrites97.9%

              \[\leadsto \color{blue}{\frac{\frac{0.2514179000665374}{x \cdot x} + 0.5}{x}} \]
          7. Recombined 2 regimes into one program.
          8. Add Preprocessing

          Alternative 6: 99.4% accurate, 11.5× speedup?

          \[\begin{array}{l} x\_m = \left|x\right| \\ x\_s = \mathsf{copysign}\left(1, x\right) \\ x\_s \cdot \begin{array}{l} \mathbf{if}\;x\_m \leq 0.86:\\ \;\;\;\;\mathsf{fma}\left(\left(\left(x\_m \cdot x\_m\right) \cdot 0.265709700396151 - 0.6665536072\right) \cdot \left(x\_m \cdot x\_m\right), x\_m, x\_m\right)\\ \mathbf{else}:\\ \;\;\;\;\frac{0.5}{x\_m}\\ \end{array} \end{array} \]
          x\_m = (fabs.f64 x)
          x\_s = (copysign.f64 #s(literal 1 binary64) x)
          (FPCore (x_s x_m)
           :precision binary64
           (*
            x_s
            (if (<= x_m 0.86)
              (fma
               (* (- (* (* x_m x_m) 0.265709700396151) 0.6665536072) (* x_m x_m))
               x_m
               x_m)
              (/ 0.5 x_m))))
          x\_m = fabs(x);
          x\_s = copysign(1.0, x);
          double code(double x_s, double x_m) {
          	double tmp;
          	if (x_m <= 0.86) {
          		tmp = fma(((((x_m * x_m) * 0.265709700396151) - 0.6665536072) * (x_m * x_m)), x_m, x_m);
          	} else {
          		tmp = 0.5 / x_m;
          	}
          	return x_s * tmp;
          }
          
          x\_m = abs(x)
          x\_s = copysign(1.0, x)
          function code(x_s, x_m)
          	tmp = 0.0
          	if (x_m <= 0.86)
          		tmp = fma(Float64(Float64(Float64(Float64(x_m * x_m) * 0.265709700396151) - 0.6665536072) * Float64(x_m * x_m)), x_m, x_m);
          	else
          		tmp = Float64(0.5 / x_m);
          	end
          	return Float64(x_s * tmp)
          end
          
          x\_m = N[Abs[x], $MachinePrecision]
          x\_s = N[With[{TMP1 = Abs[1.0], TMP2 = Sign[x]}, TMP1 * If[TMP2 == 0, 1, TMP2]], $MachinePrecision]
          code[x$95$s_, x$95$m_] := N[(x$95$s * If[LessEqual[x$95$m, 0.86], N[(N[(N[(N[(N[(x$95$m * x$95$m), $MachinePrecision] * 0.265709700396151), $MachinePrecision] - 0.6665536072), $MachinePrecision] * N[(x$95$m * x$95$m), $MachinePrecision]), $MachinePrecision] * x$95$m + x$95$m), $MachinePrecision], N[(0.5 / x$95$m), $MachinePrecision]]), $MachinePrecision]
          
          \begin{array}{l}
          x\_m = \left|x\right|
          \\
          x\_s = \mathsf{copysign}\left(1, x\right)
          
          \\
          x\_s \cdot \begin{array}{l}
          \mathbf{if}\;x\_m \leq 0.86:\\
          \;\;\;\;\mathsf{fma}\left(\left(\left(x\_m \cdot x\_m\right) \cdot 0.265709700396151 - 0.6665536072\right) \cdot \left(x\_m \cdot x\_m\right), x\_m, x\_m\right)\\
          
          \mathbf{else}:\\
          \;\;\;\;\frac{0.5}{x\_m}\\
          
          
          \end{array}
          \end{array}
          
          Derivation
          1. Split input into 2 regimes
          2. if x < 0.859999999999999987

            1. Initial program 69.1%

              \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
            2. Add Preprocessing
            3. Taylor expanded in x around 0

              \[\leadsto \color{blue}{x \cdot \left(1 + {x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right)\right)} \]
            4. Step-by-step derivation
              1. +-commutativeN/A

                \[\leadsto x \cdot \color{blue}{\left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right) + 1\right)} \]
              2. distribute-lft-inN/A

                \[\leadsto \color{blue}{x \cdot \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right)\right) + x \cdot 1} \]
              3. fp-cancel-sign-sub-invN/A

                \[\leadsto \color{blue}{x \cdot \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right)\right) - \left(\mathsf{neg}\left(x\right)\right) \cdot 1} \]
              4. distribute-lft-neg-inN/A

                \[\leadsto x \cdot \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right)\right) - \color{blue}{\left(\mathsf{neg}\left(x \cdot 1\right)\right)} \]
              5. distribute-rgt-neg-inN/A

                \[\leadsto x \cdot \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right)\right) - \color{blue}{x \cdot \left(\mathsf{neg}\left(1\right)\right)} \]
              6. metadata-evalN/A

                \[\leadsto x \cdot \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right)\right) - x \cdot \color{blue}{-1} \]
              7. fp-cancel-sub-sign-invN/A

                \[\leadsto \color{blue}{x \cdot \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right)\right) + \left(\mathsf{neg}\left(x\right)\right) \cdot -1} \]
              8. *-commutativeN/A

                \[\leadsto \color{blue}{\left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right)\right) \cdot x} + \left(\mathsf{neg}\left(x\right)\right) \cdot -1 \]
              9. *-commutativeN/A

                \[\leadsto \color{blue}{\left(\left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right) \cdot {x}^{2}\right)} \cdot x + \left(\mathsf{neg}\left(x\right)\right) \cdot -1 \]
              10. associate-*l*N/A

                \[\leadsto \color{blue}{\left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right) \cdot \left({x}^{2} \cdot x\right)} + \left(\mathsf{neg}\left(x\right)\right) \cdot -1 \]
              11. *-commutativeN/A

                \[\leadsto \color{blue}{\left({x}^{2} \cdot x\right) \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right)} + \left(\mathsf{neg}\left(x\right)\right) \cdot -1 \]
              12. distribute-lft-neg-inN/A

                \[\leadsto \left({x}^{2} \cdot x\right) \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right) + \color{blue}{\left(\mathsf{neg}\left(x \cdot -1\right)\right)} \]
              13. distribute-rgt-neg-inN/A

                \[\leadsto \left({x}^{2} \cdot x\right) \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right) + \color{blue}{x \cdot \left(\mathsf{neg}\left(-1\right)\right)} \]
              14. metadata-evalN/A

                \[\leadsto \left({x}^{2} \cdot x\right) \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right) + x \cdot \color{blue}{1} \]
              15. *-rgt-identityN/A

                \[\leadsto \left({x}^{2} \cdot x\right) \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right) + \color{blue}{x} \]
              16. lower-fma.f64N/A

                \[\leadsto \color{blue}{\mathsf{fma}\left({x}^{2} \cdot x, \frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}, x\right)} \]
            5. Applied rewrites67.8%

              \[\leadsto \color{blue}{\mathsf{fma}\left({x}^{3}, 0.265709700396151 \cdot \left(x \cdot x\right) - 0.6665536072, x\right)} \]
            6. Step-by-step derivation
              1. Applied rewrites67.8%

                \[\leadsto \mathsf{fma}\left(\left(\left(x \cdot x\right) \cdot 0.265709700396151 - 0.6665536072\right) \cdot \left(x \cdot x\right), \color{blue}{x}, x\right) \]

              if 0.859999999999999987 < x

              1. Initial program 9.6%

                \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
              2. Add Preprocessing
              3. Taylor expanded in x around inf

                \[\leadsto \color{blue}{\frac{\frac{1}{2}}{x}} \]
              4. Step-by-step derivation
                1. lower-/.f6496.9

                  \[\leadsto \color{blue}{\frac{0.5}{x}} \]
              5. Applied rewrites96.9%

                \[\leadsto \color{blue}{\frac{0.5}{x}} \]
            7. Recombined 2 regimes into one program.
            8. Add Preprocessing

            Alternative 7: 99.2% accurate, 18.0× speedup?

            \[\begin{array}{l} x\_m = \left|x\right| \\ x\_s = \mathsf{copysign}\left(1, x\right) \\ x\_s \cdot \begin{array}{l} \mathbf{if}\;x\_m \leq 0.8:\\ \;\;\;\;\mathsf{fma}\left(x\_m \cdot x\_m, -0.6665536072, 1\right) \cdot x\_m\\ \mathbf{else}:\\ \;\;\;\;\frac{0.5}{x\_m}\\ \end{array} \end{array} \]
            x\_m = (fabs.f64 x)
            x\_s = (copysign.f64 #s(literal 1 binary64) x)
            (FPCore (x_s x_m)
             :precision binary64
             (*
              x_s
              (if (<= x_m 0.8) (* (fma (* x_m x_m) -0.6665536072 1.0) x_m) (/ 0.5 x_m))))
            x\_m = fabs(x);
            x\_s = copysign(1.0, x);
            double code(double x_s, double x_m) {
            	double tmp;
            	if (x_m <= 0.8) {
            		tmp = fma((x_m * x_m), -0.6665536072, 1.0) * x_m;
            	} else {
            		tmp = 0.5 / x_m;
            	}
            	return x_s * tmp;
            }
            
            x\_m = abs(x)
            x\_s = copysign(1.0, x)
            function code(x_s, x_m)
            	tmp = 0.0
            	if (x_m <= 0.8)
            		tmp = Float64(fma(Float64(x_m * x_m), -0.6665536072, 1.0) * x_m);
            	else
            		tmp = Float64(0.5 / x_m);
            	end
            	return Float64(x_s * tmp)
            end
            
            x\_m = N[Abs[x], $MachinePrecision]
            x\_s = N[With[{TMP1 = Abs[1.0], TMP2 = Sign[x]}, TMP1 * If[TMP2 == 0, 1, TMP2]], $MachinePrecision]
            code[x$95$s_, x$95$m_] := N[(x$95$s * If[LessEqual[x$95$m, 0.8], N[(N[(N[(x$95$m * x$95$m), $MachinePrecision] * -0.6665536072 + 1.0), $MachinePrecision] * x$95$m), $MachinePrecision], N[(0.5 / x$95$m), $MachinePrecision]]), $MachinePrecision]
            
            \begin{array}{l}
            x\_m = \left|x\right|
            \\
            x\_s = \mathsf{copysign}\left(1, x\right)
            
            \\
            x\_s \cdot \begin{array}{l}
            \mathbf{if}\;x\_m \leq 0.8:\\
            \;\;\;\;\mathsf{fma}\left(x\_m \cdot x\_m, -0.6665536072, 1\right) \cdot x\_m\\
            
            \mathbf{else}:\\
            \;\;\;\;\frac{0.5}{x\_m}\\
            
            
            \end{array}
            \end{array}
            
            Derivation
            1. Split input into 2 regimes
            2. if x < 0.80000000000000004

              1. Initial program 69.1%

                \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
              2. Add Preprocessing
              3. Taylor expanded in x around 0

                \[\leadsto \color{blue}{\left(1 + \frac{-833192009}{1250000000} \cdot {x}^{2}\right)} \cdot x \]
              4. Step-by-step derivation
                1. +-commutativeN/A

                  \[\leadsto \color{blue}{\left(\frac{-833192009}{1250000000} \cdot {x}^{2} + 1\right)} \cdot x \]
                2. *-commutativeN/A

                  \[\leadsto \left(\color{blue}{{x}^{2} \cdot \frac{-833192009}{1250000000}} + 1\right) \cdot x \]
                3. lower-fma.f64N/A

                  \[\leadsto \color{blue}{\mathsf{fma}\left({x}^{2}, \frac{-833192009}{1250000000}, 1\right)} \cdot x \]
                4. unpow2N/A

                  \[\leadsto \mathsf{fma}\left(\color{blue}{x \cdot x}, \frac{-833192009}{1250000000}, 1\right) \cdot x \]
                5. lower-*.f6467.2

                  \[\leadsto \mathsf{fma}\left(\color{blue}{x \cdot x}, -0.6665536072, 1\right) \cdot x \]
              5. Applied rewrites67.2%

                \[\leadsto \color{blue}{\mathsf{fma}\left(x \cdot x, -0.6665536072, 1\right)} \cdot x \]

              if 0.80000000000000004 < x

              1. Initial program 9.6%

                \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
              2. Add Preprocessing
              3. Taylor expanded in x around inf

                \[\leadsto \color{blue}{\frac{\frac{1}{2}}{x}} \]
              4. Step-by-step derivation
                1. lower-/.f6496.9

                  \[\leadsto \color{blue}{\frac{0.5}{x}} \]
              5. Applied rewrites96.9%

                \[\leadsto \color{blue}{\frac{0.5}{x}} \]
            3. Recombined 2 regimes into one program.
            4. Add Preprocessing

            Alternative 8: 98.9% accurate, 23.0× speedup?

            \[\begin{array}{l} x\_m = \left|x\right| \\ x\_s = \mathsf{copysign}\left(1, x\right) \\ x\_s \cdot \begin{array}{l} \mathbf{if}\;x\_m \leq 0.72:\\ \;\;\;\;1 \cdot x\_m\\ \mathbf{else}:\\ \;\;\;\;\frac{0.5}{x\_m}\\ \end{array} \end{array} \]
            x\_m = (fabs.f64 x)
            x\_s = (copysign.f64 #s(literal 1 binary64) x)
            (FPCore (x_s x_m)
             :precision binary64
             (* x_s (if (<= x_m 0.72) (* 1.0 x_m) (/ 0.5 x_m))))
            x\_m = fabs(x);
            x\_s = copysign(1.0, x);
            double code(double x_s, double x_m) {
            	double tmp;
            	if (x_m <= 0.72) {
            		tmp = 1.0 * x_m;
            	} else {
            		tmp = 0.5 / x_m;
            	}
            	return x_s * tmp;
            }
            
            x\_m =     private
            x\_s =     private
            module fmin_fmax_functions
                implicit none
                private
                public fmax
                public fmin
            
                interface fmax
                    module procedure fmax88
                    module procedure fmax44
                    module procedure fmax84
                    module procedure fmax48
                end interface
                interface fmin
                    module procedure fmin88
                    module procedure fmin44
                    module procedure fmin84
                    module procedure fmin48
                end interface
            contains
                real(8) function fmax88(x, y) result (res)
                    real(8), intent (in) :: x
                    real(8), intent (in) :: y
                    res = merge(y, merge(x, max(x, y), y /= y), x /= x)
                end function
                real(4) function fmax44(x, y) result (res)
                    real(4), intent (in) :: x
                    real(4), intent (in) :: y
                    res = merge(y, merge(x, max(x, y), y /= y), x /= x)
                end function
                real(8) function fmax84(x, y) result(res)
                    real(8), intent (in) :: x
                    real(4), intent (in) :: y
                    res = merge(dble(y), merge(x, max(x, dble(y)), y /= y), x /= x)
                end function
                real(8) function fmax48(x, y) result(res)
                    real(4), intent (in) :: x
                    real(8), intent (in) :: y
                    res = merge(y, merge(dble(x), max(dble(x), y), y /= y), x /= x)
                end function
                real(8) function fmin88(x, y) result (res)
                    real(8), intent (in) :: x
                    real(8), intent (in) :: y
                    res = merge(y, merge(x, min(x, y), y /= y), x /= x)
                end function
                real(4) function fmin44(x, y) result (res)
                    real(4), intent (in) :: x
                    real(4), intent (in) :: y
                    res = merge(y, merge(x, min(x, y), y /= y), x /= x)
                end function
                real(8) function fmin84(x, y) result(res)
                    real(8), intent (in) :: x
                    real(4), intent (in) :: y
                    res = merge(dble(y), merge(x, min(x, dble(y)), y /= y), x /= x)
                end function
                real(8) function fmin48(x, y) result(res)
                    real(4), intent (in) :: x
                    real(8), intent (in) :: y
                    res = merge(y, merge(dble(x), min(dble(x), y), y /= y), x /= x)
                end function
            end module
            
            real(8) function code(x_s, x_m)
            use fmin_fmax_functions
                real(8), intent (in) :: x_s
                real(8), intent (in) :: x_m
                real(8) :: tmp
                if (x_m <= 0.72d0) then
                    tmp = 1.0d0 * x_m
                else
                    tmp = 0.5d0 / x_m
                end if
                code = x_s * tmp
            end function
            
            x\_m = Math.abs(x);
            x\_s = Math.copySign(1.0, x);
            public static double code(double x_s, double x_m) {
            	double tmp;
            	if (x_m <= 0.72) {
            		tmp = 1.0 * x_m;
            	} else {
            		tmp = 0.5 / x_m;
            	}
            	return x_s * tmp;
            }
            
            x\_m = math.fabs(x)
            x\_s = math.copysign(1.0, x)
            def code(x_s, x_m):
            	tmp = 0
            	if x_m <= 0.72:
            		tmp = 1.0 * x_m
            	else:
            		tmp = 0.5 / x_m
            	return x_s * tmp
            
            x\_m = abs(x)
            x\_s = copysign(1.0, x)
            function code(x_s, x_m)
            	tmp = 0.0
            	if (x_m <= 0.72)
            		tmp = Float64(1.0 * x_m);
            	else
            		tmp = Float64(0.5 / x_m);
            	end
            	return Float64(x_s * tmp)
            end
            
            x\_m = abs(x);
            x\_s = sign(x) * abs(1.0);
            function tmp_2 = code(x_s, x_m)
            	tmp = 0.0;
            	if (x_m <= 0.72)
            		tmp = 1.0 * x_m;
            	else
            		tmp = 0.5 / x_m;
            	end
            	tmp_2 = x_s * tmp;
            end
            
            x\_m = N[Abs[x], $MachinePrecision]
            x\_s = N[With[{TMP1 = Abs[1.0], TMP2 = Sign[x]}, TMP1 * If[TMP2 == 0, 1, TMP2]], $MachinePrecision]
            code[x$95$s_, x$95$m_] := N[(x$95$s * If[LessEqual[x$95$m, 0.72], N[(1.0 * x$95$m), $MachinePrecision], N[(0.5 / x$95$m), $MachinePrecision]]), $MachinePrecision]
            
            \begin{array}{l}
            x\_m = \left|x\right|
            \\
            x\_s = \mathsf{copysign}\left(1, x\right)
            
            \\
            x\_s \cdot \begin{array}{l}
            \mathbf{if}\;x\_m \leq 0.72:\\
            \;\;\;\;1 \cdot x\_m\\
            
            \mathbf{else}:\\
            \;\;\;\;\frac{0.5}{x\_m}\\
            
            
            \end{array}
            \end{array}
            
            Derivation
            1. Split input into 2 regimes
            2. if x < 0.71999999999999997

              1. Initial program 69.1%

                \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
              2. Add Preprocessing
              3. Taylor expanded in x around 0

                \[\leadsto \color{blue}{1} \cdot x \]
              4. Step-by-step derivation
                1. Applied rewrites67.8%

                  \[\leadsto \color{blue}{1} \cdot x \]

                if 0.71999999999999997 < x

                1. Initial program 9.6%

                  \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
                2. Add Preprocessing
                3. Taylor expanded in x around inf

                  \[\leadsto \color{blue}{\frac{\frac{1}{2}}{x}} \]
                4. Step-by-step derivation
                  1. lower-/.f6496.9

                    \[\leadsto \color{blue}{\frac{0.5}{x}} \]
                5. Applied rewrites96.9%

                  \[\leadsto \color{blue}{\frac{0.5}{x}} \]
              5. Recombined 2 regimes into one program.
              6. Add Preprocessing

              Alternative 9: 50.7% accurate, 69.2× speedup?

              \[\begin{array}{l} x\_m = \left|x\right| \\ x\_s = \mathsf{copysign}\left(1, x\right) \\ x\_s \cdot \left(1 \cdot x\_m\right) \end{array} \]
              x\_m = (fabs.f64 x)
              x\_s = (copysign.f64 #s(literal 1 binary64) x)
              (FPCore (x_s x_m) :precision binary64 (* x_s (* 1.0 x_m)))
              x\_m = fabs(x);
              x\_s = copysign(1.0, x);
              double code(double x_s, double x_m) {
              	return x_s * (1.0 * x_m);
              }
              
              x\_m =     private
              x\_s =     private
              module fmin_fmax_functions
                  implicit none
                  private
                  public fmax
                  public fmin
              
                  interface fmax
                      module procedure fmax88
                      module procedure fmax44
                      module procedure fmax84
                      module procedure fmax48
                  end interface
                  interface fmin
                      module procedure fmin88
                      module procedure fmin44
                      module procedure fmin84
                      module procedure fmin48
                  end interface
              contains
                  real(8) function fmax88(x, y) result (res)
                      real(8), intent (in) :: x
                      real(8), intent (in) :: y
                      res = merge(y, merge(x, max(x, y), y /= y), x /= x)
                  end function
                  real(4) function fmax44(x, y) result (res)
                      real(4), intent (in) :: x
                      real(4), intent (in) :: y
                      res = merge(y, merge(x, max(x, y), y /= y), x /= x)
                  end function
                  real(8) function fmax84(x, y) result(res)
                      real(8), intent (in) :: x
                      real(4), intent (in) :: y
                      res = merge(dble(y), merge(x, max(x, dble(y)), y /= y), x /= x)
                  end function
                  real(8) function fmax48(x, y) result(res)
                      real(4), intent (in) :: x
                      real(8), intent (in) :: y
                      res = merge(y, merge(dble(x), max(dble(x), y), y /= y), x /= x)
                  end function
                  real(8) function fmin88(x, y) result (res)
                      real(8), intent (in) :: x
                      real(8), intent (in) :: y
                      res = merge(y, merge(x, min(x, y), y /= y), x /= x)
                  end function
                  real(4) function fmin44(x, y) result (res)
                      real(4), intent (in) :: x
                      real(4), intent (in) :: y
                      res = merge(y, merge(x, min(x, y), y /= y), x /= x)
                  end function
                  real(8) function fmin84(x, y) result(res)
                      real(8), intent (in) :: x
                      real(4), intent (in) :: y
                      res = merge(dble(y), merge(x, min(x, dble(y)), y /= y), x /= x)
                  end function
                  real(8) function fmin48(x, y) result(res)
                      real(4), intent (in) :: x
                      real(8), intent (in) :: y
                      res = merge(y, merge(dble(x), min(dble(x), y), y /= y), x /= x)
                  end function
              end module
              
              real(8) function code(x_s, x_m)
              use fmin_fmax_functions
                  real(8), intent (in) :: x_s
                  real(8), intent (in) :: x_m
                  code = x_s * (1.0d0 * x_m)
              end function
              
              x\_m = Math.abs(x);
              x\_s = Math.copySign(1.0, x);
              public static double code(double x_s, double x_m) {
              	return x_s * (1.0 * x_m);
              }
              
              x\_m = math.fabs(x)
              x\_s = math.copysign(1.0, x)
              def code(x_s, x_m):
              	return x_s * (1.0 * x_m)
              
              x\_m = abs(x)
              x\_s = copysign(1.0, x)
              function code(x_s, x_m)
              	return Float64(x_s * Float64(1.0 * x_m))
              end
              
              x\_m = abs(x);
              x\_s = sign(x) * abs(1.0);
              function tmp = code(x_s, x_m)
              	tmp = x_s * (1.0 * x_m);
              end
              
              x\_m = N[Abs[x], $MachinePrecision]
              x\_s = N[With[{TMP1 = Abs[1.0], TMP2 = Sign[x]}, TMP1 * If[TMP2 == 0, 1, TMP2]], $MachinePrecision]
              code[x$95$s_, x$95$m_] := N[(x$95$s * N[(1.0 * x$95$m), $MachinePrecision]), $MachinePrecision]
              
              \begin{array}{l}
              x\_m = \left|x\right|
              \\
              x\_s = \mathsf{copysign}\left(1, x\right)
              
              \\
              x\_s \cdot \left(1 \cdot x\_m\right)
              \end{array}
              
              Derivation
              1. Initial program 57.0%

                \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
              2. Add Preprocessing
              3. Taylor expanded in x around 0

                \[\leadsto \color{blue}{1} \cdot x \]
              4. Step-by-step derivation
                1. Applied rewrites54.9%

                  \[\leadsto \color{blue}{1} \cdot x \]
                2. Add Preprocessing

                Reproduce

                ?
                herbie shell --seed 2025006 
                (FPCore (x)
                  :name "Jmat.Real.dawson"
                  :precision binary64
                  (* (/ (+ (+ (+ (+ (+ 1.0 (* 0.1049934947 (* x x))) (* 0.0424060604 (* (* x x) (* x x)))) (* 0.0072644182 (* (* (* x x) (* x x)) (* x x)))) (* 0.0005064034 (* (* (* (* x x) (* x x)) (* x x)) (* x x)))) (* 0.0001789971 (* (* (* (* (* x x) (* x x)) (* x x)) (* x x)) (* x x)))) (+ (+ (+ (+ (+ (+ 1.0 (* 0.7715471019 (* x x))) (* 0.2909738639 (* (* x x) (* x x)))) (* 0.0694555761 (* (* (* x x) (* x x)) (* x x)))) (* 0.0140005442 (* (* (* (* x x) (* x x)) (* x x)) (* x x)))) (* 0.0008327945 (* (* (* (* (* x x) (* x x)) (* x x)) (* x x)) (* x x)))) (* (* 2.0 0.0001789971) (* (* (* (* (* (* x x) (* x x)) (* x x)) (* x x)) (* x x)) (* x x))))) x))