Jmat.Real.dawson

Percentage Accurate: 53.9% → 99.7%
Time: 5.8s
Alternatives: 9
Speedup: 15.4×

Specification

?
\[\begin{array}{l} t_0 := \left(x \cdot x\right) \cdot \left(x \cdot x\right)\\ t_1 := t\_0 \cdot \left(x \cdot x\right)\\ t_2 := t\_1 \cdot \left(x \cdot x\right)\\ t_3 := t\_2 \cdot \left(x \cdot x\right)\\ \frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot t\_0\right) + 0.0072644182 \cdot t\_1\right) + 0.0005064034 \cdot t\_2\right) + 0.0001789971 \cdot t\_3}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot t\_0\right) + 0.0694555761 \cdot t\_1\right) + 0.0140005442 \cdot t\_2\right) + 0.0008327945 \cdot t\_3\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(t\_3 \cdot \left(x \cdot x\right)\right)} \cdot x \end{array} \]
(FPCore (x)
 :precision binary64
 (let* ((t_0 (* (* x x) (* x x)))
        (t_1 (* t_0 (* x x)))
        (t_2 (* t_1 (* x x)))
        (t_3 (* t_2 (* x x))))
   (*
    (/
     (+
      (+
       (+
        (+ (+ 1.0 (* 0.1049934947 (* x x))) (* 0.0424060604 t_0))
        (* 0.0072644182 t_1))
       (* 0.0005064034 t_2))
      (* 0.0001789971 t_3))
     (+
      (+
       (+
        (+
         (+ (+ 1.0 (* 0.7715471019 (* x x))) (* 0.2909738639 t_0))
         (* 0.0694555761 t_1))
        (* 0.0140005442 t_2))
       (* 0.0008327945 t_3))
      (* (* 2.0 0.0001789971) (* t_3 (* x x)))))
    x)))
double code(double x) {
	double t_0 = (x * x) * (x * x);
	double t_1 = t_0 * (x * x);
	double t_2 = t_1 * (x * x);
	double t_3 = t_2 * (x * x);
	return ((((((1.0 + (0.1049934947 * (x * x))) + (0.0424060604 * t_0)) + (0.0072644182 * t_1)) + (0.0005064034 * t_2)) + (0.0001789971 * t_3)) / ((((((1.0 + (0.7715471019 * (x * x))) + (0.2909738639 * t_0)) + (0.0694555761 * t_1)) + (0.0140005442 * t_2)) + (0.0008327945 * t_3)) + ((2.0 * 0.0001789971) * (t_3 * (x * x))))) * x;
}
module fmin_fmax_functions
    implicit none
    private
    public fmax
    public fmin

    interface fmax
        module procedure fmax88
        module procedure fmax44
        module procedure fmax84
        module procedure fmax48
    end interface
    interface fmin
        module procedure fmin88
        module procedure fmin44
        module procedure fmin84
        module procedure fmin48
    end interface
contains
    real(8) function fmax88(x, y) result (res)
        real(8), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(x, max(x, y), y /= y), x /= x)
    end function
    real(4) function fmax44(x, y) result (res)
        real(4), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(y, merge(x, max(x, y), y /= y), x /= x)
    end function
    real(8) function fmax84(x, y) result(res)
        real(8), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(dble(y), merge(x, max(x, dble(y)), y /= y), x /= x)
    end function
    real(8) function fmax48(x, y) result(res)
        real(4), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(dble(x), max(dble(x), y), y /= y), x /= x)
    end function
    real(8) function fmin88(x, y) result (res)
        real(8), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(x, min(x, y), y /= y), x /= x)
    end function
    real(4) function fmin44(x, y) result (res)
        real(4), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(y, merge(x, min(x, y), y /= y), x /= x)
    end function
    real(8) function fmin84(x, y) result(res)
        real(8), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(dble(y), merge(x, min(x, dble(y)), y /= y), x /= x)
    end function
    real(8) function fmin48(x, y) result(res)
        real(4), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(dble(x), min(dble(x), y), y /= y), x /= x)
    end function
end module

real(8) function code(x)
use fmin_fmax_functions
    real(8), intent (in) :: x
    real(8) :: t_0
    real(8) :: t_1
    real(8) :: t_2
    real(8) :: t_3
    t_0 = (x * x) * (x * x)
    t_1 = t_0 * (x * x)
    t_2 = t_1 * (x * x)
    t_3 = t_2 * (x * x)
    code = ((((((1.0d0 + (0.1049934947d0 * (x * x))) + (0.0424060604d0 * t_0)) + (0.0072644182d0 * t_1)) + (0.0005064034d0 * t_2)) + (0.0001789971d0 * t_3)) / ((((((1.0d0 + (0.7715471019d0 * (x * x))) + (0.2909738639d0 * t_0)) + (0.0694555761d0 * t_1)) + (0.0140005442d0 * t_2)) + (0.0008327945d0 * t_3)) + ((2.0d0 * 0.0001789971d0) * (t_3 * (x * x))))) * x
end function
public static double code(double x) {
	double t_0 = (x * x) * (x * x);
	double t_1 = t_0 * (x * x);
	double t_2 = t_1 * (x * x);
	double t_3 = t_2 * (x * x);
	return ((((((1.0 + (0.1049934947 * (x * x))) + (0.0424060604 * t_0)) + (0.0072644182 * t_1)) + (0.0005064034 * t_2)) + (0.0001789971 * t_3)) / ((((((1.0 + (0.7715471019 * (x * x))) + (0.2909738639 * t_0)) + (0.0694555761 * t_1)) + (0.0140005442 * t_2)) + (0.0008327945 * t_3)) + ((2.0 * 0.0001789971) * (t_3 * (x * x))))) * x;
}
def code(x):
	t_0 = (x * x) * (x * x)
	t_1 = t_0 * (x * x)
	t_2 = t_1 * (x * x)
	t_3 = t_2 * (x * x)
	return ((((((1.0 + (0.1049934947 * (x * x))) + (0.0424060604 * t_0)) + (0.0072644182 * t_1)) + (0.0005064034 * t_2)) + (0.0001789971 * t_3)) / ((((((1.0 + (0.7715471019 * (x * x))) + (0.2909738639 * t_0)) + (0.0694555761 * t_1)) + (0.0140005442 * t_2)) + (0.0008327945 * t_3)) + ((2.0 * 0.0001789971) * (t_3 * (x * x))))) * x
function code(x)
	t_0 = Float64(Float64(x * x) * Float64(x * x))
	t_1 = Float64(t_0 * Float64(x * x))
	t_2 = Float64(t_1 * Float64(x * x))
	t_3 = Float64(t_2 * Float64(x * x))
	return Float64(Float64(Float64(Float64(Float64(Float64(Float64(1.0 + Float64(0.1049934947 * Float64(x * x))) + Float64(0.0424060604 * t_0)) + Float64(0.0072644182 * t_1)) + Float64(0.0005064034 * t_2)) + Float64(0.0001789971 * t_3)) / Float64(Float64(Float64(Float64(Float64(Float64(1.0 + Float64(0.7715471019 * Float64(x * x))) + Float64(0.2909738639 * t_0)) + Float64(0.0694555761 * t_1)) + Float64(0.0140005442 * t_2)) + Float64(0.0008327945 * t_3)) + Float64(Float64(2.0 * 0.0001789971) * Float64(t_3 * Float64(x * x))))) * x)
end
function tmp = code(x)
	t_0 = (x * x) * (x * x);
	t_1 = t_0 * (x * x);
	t_2 = t_1 * (x * x);
	t_3 = t_2 * (x * x);
	tmp = ((((((1.0 + (0.1049934947 * (x * x))) + (0.0424060604 * t_0)) + (0.0072644182 * t_1)) + (0.0005064034 * t_2)) + (0.0001789971 * t_3)) / ((((((1.0 + (0.7715471019 * (x * x))) + (0.2909738639 * t_0)) + (0.0694555761 * t_1)) + (0.0140005442 * t_2)) + (0.0008327945 * t_3)) + ((2.0 * 0.0001789971) * (t_3 * (x * x))))) * x;
end
code[x_] := Block[{t$95$0 = N[(N[(x * x), $MachinePrecision] * N[(x * x), $MachinePrecision]), $MachinePrecision]}, Block[{t$95$1 = N[(t$95$0 * N[(x * x), $MachinePrecision]), $MachinePrecision]}, Block[{t$95$2 = N[(t$95$1 * N[(x * x), $MachinePrecision]), $MachinePrecision]}, Block[{t$95$3 = N[(t$95$2 * N[(x * x), $MachinePrecision]), $MachinePrecision]}, N[(N[(N[(N[(N[(N[(N[(1.0 + N[(0.1049934947 * N[(x * x), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] + N[(0.0424060604 * t$95$0), $MachinePrecision]), $MachinePrecision] + N[(0.0072644182 * t$95$1), $MachinePrecision]), $MachinePrecision] + N[(0.0005064034 * t$95$2), $MachinePrecision]), $MachinePrecision] + N[(0.0001789971 * t$95$3), $MachinePrecision]), $MachinePrecision] / N[(N[(N[(N[(N[(N[(1.0 + N[(0.7715471019 * N[(x * x), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] + N[(0.2909738639 * t$95$0), $MachinePrecision]), $MachinePrecision] + N[(0.0694555761 * t$95$1), $MachinePrecision]), $MachinePrecision] + N[(0.0140005442 * t$95$2), $MachinePrecision]), $MachinePrecision] + N[(0.0008327945 * t$95$3), $MachinePrecision]), $MachinePrecision] + N[(N[(2.0 * 0.0001789971), $MachinePrecision] * N[(t$95$3 * N[(x * x), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] * x), $MachinePrecision]]]]]
\begin{array}{l}
t_0 := \left(x \cdot x\right) \cdot \left(x \cdot x\right)\\
t_1 := t\_0 \cdot \left(x \cdot x\right)\\
t_2 := t\_1 \cdot \left(x \cdot x\right)\\
t_3 := t\_2 \cdot \left(x \cdot x\right)\\
\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot t\_0\right) + 0.0072644182 \cdot t\_1\right) + 0.0005064034 \cdot t\_2\right) + 0.0001789971 \cdot t\_3}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot t\_0\right) + 0.0694555761 \cdot t\_1\right) + 0.0140005442 \cdot t\_2\right) + 0.0008327945 \cdot t\_3\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(t\_3 \cdot \left(x \cdot x\right)\right)} \cdot x
\end{array}

Local Percentage Accuracy vs ?

The average percentage accuracy by input value. Horizontal axis shows value of an input variable; the variable is choosen in the title. Vertical axis is accuracy; higher is better. Red represent the original program, while blue represents Herbie's suggestion. These can be toggled with buttons below the plot. The line is an average while dots represent individual samples.

Accuracy vs Speed?

Herbie found 9 alternatives:

AlternativeAccuracySpeedup
The accuracy (vertical axis) and speed (horizontal axis) of each alternatives. Up and to the right is better. The red square shows the initial program, and each blue circle shows an alternative.The line shows the best available speed-accuracy tradeoffs.

Initial Program: 53.9% accurate, 1.0× speedup?

\[\begin{array}{l} t_0 := \left(x \cdot x\right) \cdot \left(x \cdot x\right)\\ t_1 := t\_0 \cdot \left(x \cdot x\right)\\ t_2 := t\_1 \cdot \left(x \cdot x\right)\\ t_3 := t\_2 \cdot \left(x \cdot x\right)\\ \frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot t\_0\right) + 0.0072644182 \cdot t\_1\right) + 0.0005064034 \cdot t\_2\right) + 0.0001789971 \cdot t\_3}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot t\_0\right) + 0.0694555761 \cdot t\_1\right) + 0.0140005442 \cdot t\_2\right) + 0.0008327945 \cdot t\_3\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(t\_3 \cdot \left(x \cdot x\right)\right)} \cdot x \end{array} \]
(FPCore (x)
 :precision binary64
 (let* ((t_0 (* (* x x) (* x x)))
        (t_1 (* t_0 (* x x)))
        (t_2 (* t_1 (* x x)))
        (t_3 (* t_2 (* x x))))
   (*
    (/
     (+
      (+
       (+
        (+ (+ 1.0 (* 0.1049934947 (* x x))) (* 0.0424060604 t_0))
        (* 0.0072644182 t_1))
       (* 0.0005064034 t_2))
      (* 0.0001789971 t_3))
     (+
      (+
       (+
        (+
         (+ (+ 1.0 (* 0.7715471019 (* x x))) (* 0.2909738639 t_0))
         (* 0.0694555761 t_1))
        (* 0.0140005442 t_2))
       (* 0.0008327945 t_3))
      (* (* 2.0 0.0001789971) (* t_3 (* x x)))))
    x)))
double code(double x) {
	double t_0 = (x * x) * (x * x);
	double t_1 = t_0 * (x * x);
	double t_2 = t_1 * (x * x);
	double t_3 = t_2 * (x * x);
	return ((((((1.0 + (0.1049934947 * (x * x))) + (0.0424060604 * t_0)) + (0.0072644182 * t_1)) + (0.0005064034 * t_2)) + (0.0001789971 * t_3)) / ((((((1.0 + (0.7715471019 * (x * x))) + (0.2909738639 * t_0)) + (0.0694555761 * t_1)) + (0.0140005442 * t_2)) + (0.0008327945 * t_3)) + ((2.0 * 0.0001789971) * (t_3 * (x * x))))) * x;
}
module fmin_fmax_functions
    implicit none
    private
    public fmax
    public fmin

    interface fmax
        module procedure fmax88
        module procedure fmax44
        module procedure fmax84
        module procedure fmax48
    end interface
    interface fmin
        module procedure fmin88
        module procedure fmin44
        module procedure fmin84
        module procedure fmin48
    end interface
contains
    real(8) function fmax88(x, y) result (res)
        real(8), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(x, max(x, y), y /= y), x /= x)
    end function
    real(4) function fmax44(x, y) result (res)
        real(4), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(y, merge(x, max(x, y), y /= y), x /= x)
    end function
    real(8) function fmax84(x, y) result(res)
        real(8), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(dble(y), merge(x, max(x, dble(y)), y /= y), x /= x)
    end function
    real(8) function fmax48(x, y) result(res)
        real(4), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(dble(x), max(dble(x), y), y /= y), x /= x)
    end function
    real(8) function fmin88(x, y) result (res)
        real(8), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(x, min(x, y), y /= y), x /= x)
    end function
    real(4) function fmin44(x, y) result (res)
        real(4), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(y, merge(x, min(x, y), y /= y), x /= x)
    end function
    real(8) function fmin84(x, y) result(res)
        real(8), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(dble(y), merge(x, min(x, dble(y)), y /= y), x /= x)
    end function
    real(8) function fmin48(x, y) result(res)
        real(4), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(dble(x), min(dble(x), y), y /= y), x /= x)
    end function
end module

real(8) function code(x)
use fmin_fmax_functions
    real(8), intent (in) :: x
    real(8) :: t_0
    real(8) :: t_1
    real(8) :: t_2
    real(8) :: t_3
    t_0 = (x * x) * (x * x)
    t_1 = t_0 * (x * x)
    t_2 = t_1 * (x * x)
    t_3 = t_2 * (x * x)
    code = ((((((1.0d0 + (0.1049934947d0 * (x * x))) + (0.0424060604d0 * t_0)) + (0.0072644182d0 * t_1)) + (0.0005064034d0 * t_2)) + (0.0001789971d0 * t_3)) / ((((((1.0d0 + (0.7715471019d0 * (x * x))) + (0.2909738639d0 * t_0)) + (0.0694555761d0 * t_1)) + (0.0140005442d0 * t_2)) + (0.0008327945d0 * t_3)) + ((2.0d0 * 0.0001789971d0) * (t_3 * (x * x))))) * x
end function
public static double code(double x) {
	double t_0 = (x * x) * (x * x);
	double t_1 = t_0 * (x * x);
	double t_2 = t_1 * (x * x);
	double t_3 = t_2 * (x * x);
	return ((((((1.0 + (0.1049934947 * (x * x))) + (0.0424060604 * t_0)) + (0.0072644182 * t_1)) + (0.0005064034 * t_2)) + (0.0001789971 * t_3)) / ((((((1.0 + (0.7715471019 * (x * x))) + (0.2909738639 * t_0)) + (0.0694555761 * t_1)) + (0.0140005442 * t_2)) + (0.0008327945 * t_3)) + ((2.0 * 0.0001789971) * (t_3 * (x * x))))) * x;
}
def code(x):
	t_0 = (x * x) * (x * x)
	t_1 = t_0 * (x * x)
	t_2 = t_1 * (x * x)
	t_3 = t_2 * (x * x)
	return ((((((1.0 + (0.1049934947 * (x * x))) + (0.0424060604 * t_0)) + (0.0072644182 * t_1)) + (0.0005064034 * t_2)) + (0.0001789971 * t_3)) / ((((((1.0 + (0.7715471019 * (x * x))) + (0.2909738639 * t_0)) + (0.0694555761 * t_1)) + (0.0140005442 * t_2)) + (0.0008327945 * t_3)) + ((2.0 * 0.0001789971) * (t_3 * (x * x))))) * x
function code(x)
	t_0 = Float64(Float64(x * x) * Float64(x * x))
	t_1 = Float64(t_0 * Float64(x * x))
	t_2 = Float64(t_1 * Float64(x * x))
	t_3 = Float64(t_2 * Float64(x * x))
	return Float64(Float64(Float64(Float64(Float64(Float64(Float64(1.0 + Float64(0.1049934947 * Float64(x * x))) + Float64(0.0424060604 * t_0)) + Float64(0.0072644182 * t_1)) + Float64(0.0005064034 * t_2)) + Float64(0.0001789971 * t_3)) / Float64(Float64(Float64(Float64(Float64(Float64(1.0 + Float64(0.7715471019 * Float64(x * x))) + Float64(0.2909738639 * t_0)) + Float64(0.0694555761 * t_1)) + Float64(0.0140005442 * t_2)) + Float64(0.0008327945 * t_3)) + Float64(Float64(2.0 * 0.0001789971) * Float64(t_3 * Float64(x * x))))) * x)
end
function tmp = code(x)
	t_0 = (x * x) * (x * x);
	t_1 = t_0 * (x * x);
	t_2 = t_1 * (x * x);
	t_3 = t_2 * (x * x);
	tmp = ((((((1.0 + (0.1049934947 * (x * x))) + (0.0424060604 * t_0)) + (0.0072644182 * t_1)) + (0.0005064034 * t_2)) + (0.0001789971 * t_3)) / ((((((1.0 + (0.7715471019 * (x * x))) + (0.2909738639 * t_0)) + (0.0694555761 * t_1)) + (0.0140005442 * t_2)) + (0.0008327945 * t_3)) + ((2.0 * 0.0001789971) * (t_3 * (x * x))))) * x;
end
code[x_] := Block[{t$95$0 = N[(N[(x * x), $MachinePrecision] * N[(x * x), $MachinePrecision]), $MachinePrecision]}, Block[{t$95$1 = N[(t$95$0 * N[(x * x), $MachinePrecision]), $MachinePrecision]}, Block[{t$95$2 = N[(t$95$1 * N[(x * x), $MachinePrecision]), $MachinePrecision]}, Block[{t$95$3 = N[(t$95$2 * N[(x * x), $MachinePrecision]), $MachinePrecision]}, N[(N[(N[(N[(N[(N[(N[(1.0 + N[(0.1049934947 * N[(x * x), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] + N[(0.0424060604 * t$95$0), $MachinePrecision]), $MachinePrecision] + N[(0.0072644182 * t$95$1), $MachinePrecision]), $MachinePrecision] + N[(0.0005064034 * t$95$2), $MachinePrecision]), $MachinePrecision] + N[(0.0001789971 * t$95$3), $MachinePrecision]), $MachinePrecision] / N[(N[(N[(N[(N[(N[(1.0 + N[(0.7715471019 * N[(x * x), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] + N[(0.2909738639 * t$95$0), $MachinePrecision]), $MachinePrecision] + N[(0.0694555761 * t$95$1), $MachinePrecision]), $MachinePrecision] + N[(0.0140005442 * t$95$2), $MachinePrecision]), $MachinePrecision] + N[(0.0008327945 * t$95$3), $MachinePrecision]), $MachinePrecision] + N[(N[(2.0 * 0.0001789971), $MachinePrecision] * N[(t$95$3 * N[(x * x), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] * x), $MachinePrecision]]]]]
\begin{array}{l}
t_0 := \left(x \cdot x\right) \cdot \left(x \cdot x\right)\\
t_1 := t\_0 \cdot \left(x \cdot x\right)\\
t_2 := t\_1 \cdot \left(x \cdot x\right)\\
t_3 := t\_2 \cdot \left(x \cdot x\right)\\
\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot t\_0\right) + 0.0072644182 \cdot t\_1\right) + 0.0005064034 \cdot t\_2\right) + 0.0001789971 \cdot t\_3}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot t\_0\right) + 0.0694555761 \cdot t\_1\right) + 0.0140005442 \cdot t\_2\right) + 0.0008327945 \cdot t\_3\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(t\_3 \cdot \left(x \cdot x\right)\right)} \cdot x
\end{array}

Alternative 1: 99.7% accurate, 3.0× speedup?

\[\begin{array}{l} t_0 := {\left(\left|x\right|\right)}^{2}\\ t_1 := \left|x\right| \cdot \left|x\right|\\ \mathsf{copysign}\left(1, x\right) \cdot \begin{array}{l} \mathbf{if}\;\left|x\right| \leq 1.5:\\ \;\;\;\;\left(1 + t\_0 \cdot \left(t\_0 \cdot \left(0.265709700396151 + -0.0732490286039007 \cdot t\_0\right) - 0.6665536072\right)\right) \cdot \left|x\right|\\ \mathbf{else}:\\ \;\;\;\;\frac{\frac{\frac{\frac{0.15298196345929074}{t\_1} - -0.2514179000665374}{\left|x\right|} - \frac{-11.259630434457211}{\left(t\_1 \cdot \left|x\right|\right) \cdot t\_1}}{\left|x\right|} - -0.5}{\left|x\right|}\\ \end{array} \end{array} \]
(FPCore (x)
 :precision binary64
 (let* ((t_0 (pow (fabs x) 2.0)) (t_1 (* (fabs x) (fabs x))))
   (*
    (copysign 1.0 x)
    (if (<= (fabs x) 1.5)
      (*
       (+
        1.0
        (*
         t_0
         (-
          (* t_0 (+ 0.265709700396151 (* -0.0732490286039007 t_0)))
          0.6665536072)))
       (fabs x))
      (/
       (-
        (/
         (-
          (/ (- (/ 0.15298196345929074 t_1) -0.2514179000665374) (fabs x))
          (/ -11.259630434457211 (* (* t_1 (fabs x)) t_1)))
         (fabs x))
        -0.5)
       (fabs x))))))
double code(double x) {
	double t_0 = pow(fabs(x), 2.0);
	double t_1 = fabs(x) * fabs(x);
	double tmp;
	if (fabs(x) <= 1.5) {
		tmp = (1.0 + (t_0 * ((t_0 * (0.265709700396151 + (-0.0732490286039007 * t_0))) - 0.6665536072))) * fabs(x);
	} else {
		tmp = ((((((0.15298196345929074 / t_1) - -0.2514179000665374) / fabs(x)) - (-11.259630434457211 / ((t_1 * fabs(x)) * t_1))) / fabs(x)) - -0.5) / fabs(x);
	}
	return copysign(1.0, x) * tmp;
}
public static double code(double x) {
	double t_0 = Math.pow(Math.abs(x), 2.0);
	double t_1 = Math.abs(x) * Math.abs(x);
	double tmp;
	if (Math.abs(x) <= 1.5) {
		tmp = (1.0 + (t_0 * ((t_0 * (0.265709700396151 + (-0.0732490286039007 * t_0))) - 0.6665536072))) * Math.abs(x);
	} else {
		tmp = ((((((0.15298196345929074 / t_1) - -0.2514179000665374) / Math.abs(x)) - (-11.259630434457211 / ((t_1 * Math.abs(x)) * t_1))) / Math.abs(x)) - -0.5) / Math.abs(x);
	}
	return Math.copySign(1.0, x) * tmp;
}
def code(x):
	t_0 = math.pow(math.fabs(x), 2.0)
	t_1 = math.fabs(x) * math.fabs(x)
	tmp = 0
	if math.fabs(x) <= 1.5:
		tmp = (1.0 + (t_0 * ((t_0 * (0.265709700396151 + (-0.0732490286039007 * t_0))) - 0.6665536072))) * math.fabs(x)
	else:
		tmp = ((((((0.15298196345929074 / t_1) - -0.2514179000665374) / math.fabs(x)) - (-11.259630434457211 / ((t_1 * math.fabs(x)) * t_1))) / math.fabs(x)) - -0.5) / math.fabs(x)
	return math.copysign(1.0, x) * tmp
function code(x)
	t_0 = abs(x) ^ 2.0
	t_1 = Float64(abs(x) * abs(x))
	tmp = 0.0
	if (abs(x) <= 1.5)
		tmp = Float64(Float64(1.0 + Float64(t_0 * Float64(Float64(t_0 * Float64(0.265709700396151 + Float64(-0.0732490286039007 * t_0))) - 0.6665536072))) * abs(x));
	else
		tmp = Float64(Float64(Float64(Float64(Float64(Float64(Float64(0.15298196345929074 / t_1) - -0.2514179000665374) / abs(x)) - Float64(-11.259630434457211 / Float64(Float64(t_1 * abs(x)) * t_1))) / abs(x)) - -0.5) / abs(x));
	end
	return Float64(copysign(1.0, x) * tmp)
end
function tmp_2 = code(x)
	t_0 = abs(x) ^ 2.0;
	t_1 = abs(x) * abs(x);
	tmp = 0.0;
	if (abs(x) <= 1.5)
		tmp = (1.0 + (t_0 * ((t_0 * (0.265709700396151 + (-0.0732490286039007 * t_0))) - 0.6665536072))) * abs(x);
	else
		tmp = ((((((0.15298196345929074 / t_1) - -0.2514179000665374) / abs(x)) - (-11.259630434457211 / ((t_1 * abs(x)) * t_1))) / abs(x)) - -0.5) / abs(x);
	end
	tmp_2 = (sign(x) * abs(1.0)) * tmp;
end
code[x_] := Block[{t$95$0 = N[Power[N[Abs[x], $MachinePrecision], 2.0], $MachinePrecision]}, Block[{t$95$1 = N[(N[Abs[x], $MachinePrecision] * N[Abs[x], $MachinePrecision]), $MachinePrecision]}, N[(N[With[{TMP1 = Abs[1.0], TMP2 = Sign[x]}, TMP1 * If[TMP2 == 0, 1, TMP2]], $MachinePrecision] * If[LessEqual[N[Abs[x], $MachinePrecision], 1.5], N[(N[(1.0 + N[(t$95$0 * N[(N[(t$95$0 * N[(0.265709700396151 + N[(-0.0732490286039007 * t$95$0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] - 0.6665536072), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] * N[Abs[x], $MachinePrecision]), $MachinePrecision], N[(N[(N[(N[(N[(N[(N[(0.15298196345929074 / t$95$1), $MachinePrecision] - -0.2514179000665374), $MachinePrecision] / N[Abs[x], $MachinePrecision]), $MachinePrecision] - N[(-11.259630434457211 / N[(N[(t$95$1 * N[Abs[x], $MachinePrecision]), $MachinePrecision] * t$95$1), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] / N[Abs[x], $MachinePrecision]), $MachinePrecision] - -0.5), $MachinePrecision] / N[Abs[x], $MachinePrecision]), $MachinePrecision]]), $MachinePrecision]]]
\begin{array}{l}
t_0 := {\left(\left|x\right|\right)}^{2}\\
t_1 := \left|x\right| \cdot \left|x\right|\\
\mathsf{copysign}\left(1, x\right) \cdot \begin{array}{l}
\mathbf{if}\;\left|x\right| \leq 1.5:\\
\;\;\;\;\left(1 + t\_0 \cdot \left(t\_0 \cdot \left(0.265709700396151 + -0.0732490286039007 \cdot t\_0\right) - 0.6665536072\right)\right) \cdot \left|x\right|\\

\mathbf{else}:\\
\;\;\;\;\frac{\frac{\frac{\frac{0.15298196345929074}{t\_1} - -0.2514179000665374}{\left|x\right|} - \frac{-11.259630434457211}{\left(t\_1 \cdot \left|x\right|\right) \cdot t\_1}}{\left|x\right|} - -0.5}{\left|x\right|}\\


\end{array}
\end{array}
Derivation
  1. Split input into 2 regimes
  2. if x < 1.5

    1. Initial program 53.9%

      \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
    2. Taylor expanded in x around 0

      \[\leadsto \color{blue}{\left(1 + {x}^{2} \cdot \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} + \frac{-9156128575487588197208397249}{125000000000000000000000000000} \cdot {x}^{2}\right) - \frac{833192009}{1250000000}\right)\right)} \cdot x \]
    3. Step-by-step derivation
      1. lower-+.f64N/A

        \[\leadsto \left(1 + \color{blue}{{x}^{2} \cdot \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} + \frac{-9156128575487588197208397249}{125000000000000000000000000000} \cdot {x}^{2}\right) - \frac{833192009}{1250000000}\right)}\right) \cdot x \]
    4. Applied rewrites50.2%

      \[\leadsto \color{blue}{\left(1 + {x}^{2} \cdot \left({x}^{2} \cdot \left(0.265709700396151 + -0.0732490286039007 \cdot {x}^{2}\right) - 0.6665536072\right)\right)} \cdot x \]

    if 1.5 < x

    1. Initial program 53.9%

      \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
    2. Taylor expanded in x around inf

      \[\leadsto \color{blue}{\frac{\frac{1}{2} + \left(\frac{\frac{1307076337763}{8543989815576}}{{x}^{4}} + \left(\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} + \frac{344398180852034095277}{30586987988352776592} \cdot \frac{1}{{x}^{6}}\right)\right)}{x}} \]
    3. Step-by-step derivation
      1. Applied rewrites51.4%

        \[\leadsto \color{blue}{\frac{0.5 + \left(\frac{0.15298196345929074}{{x}^{4}} + \mathsf{fma}\left(0.2514179000665374, \frac{1}{{x}^{2}}, 11.259630434457211 \cdot \frac{1}{{x}^{6}}\right)\right)}{x}} \]
      2. Applied rewrites51.4%

        \[\leadsto \frac{\left(\left(\frac{0.15298196345929074}{\left(\left(x \cdot x\right) \cdot x\right) \cdot x} - \frac{-0.2514179000665374}{x \cdot x}\right) - \frac{-11.259630434457211}{\left(\left(x \cdot x\right) \cdot x\right) \cdot \left(\left(x \cdot x\right) \cdot x\right)}\right) - -0.5}{x} \]
      3. Step-by-step derivation
        1. lift--.f64N/A

          \[\leadsto \frac{\left(\left(\frac{\frac{1307076337763}{8543989815576}}{\left(\left(x \cdot x\right) \cdot x\right) \cdot x} - \frac{\frac{-600041}{2386628}}{x \cdot x}\right) - \frac{\frac{-344398180852034095277}{30586987988352776592}}{\left(\left(x \cdot x\right) \cdot x\right) \cdot \left(\left(x \cdot x\right) \cdot x\right)}\right) - \frac{-1}{2}}{x} \]
        2. lift--.f64N/A

          \[\leadsto \frac{\left(\left(\frac{\frac{1307076337763}{8543989815576}}{\left(\left(x \cdot x\right) \cdot x\right) \cdot x} - \frac{\frac{-600041}{2386628}}{x \cdot x}\right) - \frac{\frac{-344398180852034095277}{30586987988352776592}}{\left(\left(x \cdot x\right) \cdot x\right) \cdot \left(\left(x \cdot x\right) \cdot x\right)}\right) - \frac{-1}{2}}{x} \]
        3. associate--l-N/A

          \[\leadsto \frac{\left(\frac{\frac{1307076337763}{8543989815576}}{\left(\left(x \cdot x\right) \cdot x\right) \cdot x} - \frac{\frac{-600041}{2386628}}{x \cdot x}\right) - \left(\frac{\frac{-344398180852034095277}{30586987988352776592}}{\left(\left(x \cdot x\right) \cdot x\right) \cdot \left(\left(x \cdot x\right) \cdot x\right)} + \frac{-1}{2}\right)}{x} \]
        4. lower--.f64N/A

          \[\leadsto \frac{\left(\frac{\frac{1307076337763}{8543989815576}}{\left(\left(x \cdot x\right) \cdot x\right) \cdot x} - \frac{\frac{-600041}{2386628}}{x \cdot x}\right) - \left(\frac{\frac{-344398180852034095277}{30586987988352776592}}{\left(\left(x \cdot x\right) \cdot x\right) \cdot \left(\left(x \cdot x\right) \cdot x\right)} + \frac{-1}{2}\right)}{x} \]
      4. Applied rewrites51.4%

        \[\leadsto \frac{\frac{\frac{0.15298196345929074}{\left(x \cdot x\right) \cdot x} - \frac{-0.2514179000665374}{x}}{x} - \left(\frac{-11.259630434457211}{\left(\left(\left(\left(x \cdot x\right) \cdot x\right) \cdot x\right) \cdot x\right) \cdot x} - 0.5\right)}{x} \]
      5. Applied rewrites51.4%

        \[\leadsto \frac{\frac{\frac{\frac{0.15298196345929074}{x \cdot x} - -0.2514179000665374}{x} - \frac{-11.259630434457211}{\left(\left(x \cdot x\right) \cdot x\right) \cdot \left(x \cdot x\right)}}{x} - -0.5}{\color{blue}{x}} \]
    4. Recombined 2 regimes into one program.
    5. Add Preprocessing

    Alternative 2: 99.7% accurate, 4.0× speedup?

    \[\begin{array}{l} t_0 := \left|x\right| \cdot \left|x\right|\\ \mathsf{copysign}\left(1, x\right) \cdot \begin{array}{l} \mathbf{if}\;\left|x\right| \leq 1.5:\\ \;\;\;\;\mathsf{fma}\left(\mathsf{fma}\left(0.265709700396151, t\_0, -0.6665536072\right), t\_0, 1\right) \cdot \left|x\right|\\ \mathbf{else}:\\ \;\;\;\;\frac{\frac{\frac{\frac{0.15298196345929074}{t\_0} - -0.2514179000665374}{\left|x\right|} - \frac{-11.259630434457211}{\left(t\_0 \cdot \left|x\right|\right) \cdot t\_0}}{\left|x\right|} - -0.5}{\left|x\right|}\\ \end{array} \end{array} \]
    (FPCore (x)
     :precision binary64
     (let* ((t_0 (* (fabs x) (fabs x))))
       (*
        (copysign 1.0 x)
        (if (<= (fabs x) 1.5)
          (* (fma (fma 0.265709700396151 t_0 -0.6665536072) t_0 1.0) (fabs x))
          (/
           (-
            (/
             (-
              (/ (- (/ 0.15298196345929074 t_0) -0.2514179000665374) (fabs x))
              (/ -11.259630434457211 (* (* t_0 (fabs x)) t_0)))
             (fabs x))
            -0.5)
           (fabs x))))))
    double code(double x) {
    	double t_0 = fabs(x) * fabs(x);
    	double tmp;
    	if (fabs(x) <= 1.5) {
    		tmp = fma(fma(0.265709700396151, t_0, -0.6665536072), t_0, 1.0) * fabs(x);
    	} else {
    		tmp = ((((((0.15298196345929074 / t_0) - -0.2514179000665374) / fabs(x)) - (-11.259630434457211 / ((t_0 * fabs(x)) * t_0))) / fabs(x)) - -0.5) / fabs(x);
    	}
    	return copysign(1.0, x) * tmp;
    }
    
    function code(x)
    	t_0 = Float64(abs(x) * abs(x))
    	tmp = 0.0
    	if (abs(x) <= 1.5)
    		tmp = Float64(fma(fma(0.265709700396151, t_0, -0.6665536072), t_0, 1.0) * abs(x));
    	else
    		tmp = Float64(Float64(Float64(Float64(Float64(Float64(Float64(0.15298196345929074 / t_0) - -0.2514179000665374) / abs(x)) - Float64(-11.259630434457211 / Float64(Float64(t_0 * abs(x)) * t_0))) / abs(x)) - -0.5) / abs(x));
    	end
    	return Float64(copysign(1.0, x) * tmp)
    end
    
    code[x_] := Block[{t$95$0 = N[(N[Abs[x], $MachinePrecision] * N[Abs[x], $MachinePrecision]), $MachinePrecision]}, N[(N[With[{TMP1 = Abs[1.0], TMP2 = Sign[x]}, TMP1 * If[TMP2 == 0, 1, TMP2]], $MachinePrecision] * If[LessEqual[N[Abs[x], $MachinePrecision], 1.5], N[(N[(N[(0.265709700396151 * t$95$0 + -0.6665536072), $MachinePrecision] * t$95$0 + 1.0), $MachinePrecision] * N[Abs[x], $MachinePrecision]), $MachinePrecision], N[(N[(N[(N[(N[(N[(N[(0.15298196345929074 / t$95$0), $MachinePrecision] - -0.2514179000665374), $MachinePrecision] / N[Abs[x], $MachinePrecision]), $MachinePrecision] - N[(-11.259630434457211 / N[(N[(t$95$0 * N[Abs[x], $MachinePrecision]), $MachinePrecision] * t$95$0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] / N[Abs[x], $MachinePrecision]), $MachinePrecision] - -0.5), $MachinePrecision] / N[Abs[x], $MachinePrecision]), $MachinePrecision]]), $MachinePrecision]]
    
    \begin{array}{l}
    t_0 := \left|x\right| \cdot \left|x\right|\\
    \mathsf{copysign}\left(1, x\right) \cdot \begin{array}{l}
    \mathbf{if}\;\left|x\right| \leq 1.5:\\
    \;\;\;\;\mathsf{fma}\left(\mathsf{fma}\left(0.265709700396151, t\_0, -0.6665536072\right), t\_0, 1\right) \cdot \left|x\right|\\
    
    \mathbf{else}:\\
    \;\;\;\;\frac{\frac{\frac{\frac{0.15298196345929074}{t\_0} - -0.2514179000665374}{\left|x\right|} - \frac{-11.259630434457211}{\left(t\_0 \cdot \left|x\right|\right) \cdot t\_0}}{\left|x\right|} - -0.5}{\left|x\right|}\\
    
    
    \end{array}
    \end{array}
    
    Derivation
    1. Split input into 2 regimes
    2. if x < 1.5

      1. Initial program 53.9%

        \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
      2. Taylor expanded in x around 0

        \[\leadsto \color{blue}{\left(1 + {x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right)\right)} \cdot x \]
      3. Step-by-step derivation
        1. lower-+.f64N/A

          \[\leadsto \left(1 + \color{blue}{{x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right)}\right) \cdot x \]
        2. lower-*.f64N/A

          \[\leadsto \left(1 + {x}^{2} \cdot \color{blue}{\left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right)}\right) \cdot x \]
      4. Applied rewrites50.9%

        \[\leadsto \color{blue}{\left(1 + {x}^{2} \cdot \left(0.265709700396151 \cdot {x}^{2} - 0.6665536072\right)\right)} \cdot x \]
      5. Step-by-step derivation
        1. lift-+.f64N/A

          \[\leadsto \left(1 + \color{blue}{{x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right)}\right) \cdot x \]
        2. +-commutativeN/A

          \[\leadsto \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right) + \color{blue}{1}\right) \cdot x \]
      6. Applied rewrites50.9%

        \[\leadsto \color{blue}{\mathsf{fma}\left(\mathsf{fma}\left(0.265709700396151, x \cdot x, -0.6665536072\right), x \cdot x, 1\right)} \cdot x \]

      if 1.5 < x

      1. Initial program 53.9%

        \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
      2. Taylor expanded in x around inf

        \[\leadsto \color{blue}{\frac{\frac{1}{2} + \left(\frac{\frac{1307076337763}{8543989815576}}{{x}^{4}} + \left(\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} + \frac{344398180852034095277}{30586987988352776592} \cdot \frac{1}{{x}^{6}}\right)\right)}{x}} \]
      3. Step-by-step derivation
        1. Applied rewrites51.4%

          \[\leadsto \color{blue}{\frac{0.5 + \left(\frac{0.15298196345929074}{{x}^{4}} + \mathsf{fma}\left(0.2514179000665374, \frac{1}{{x}^{2}}, 11.259630434457211 \cdot \frac{1}{{x}^{6}}\right)\right)}{x}} \]
        2. Applied rewrites51.4%

          \[\leadsto \frac{\left(\left(\frac{0.15298196345929074}{\left(\left(x \cdot x\right) \cdot x\right) \cdot x} - \frac{-0.2514179000665374}{x \cdot x}\right) - \frac{-11.259630434457211}{\left(\left(x \cdot x\right) \cdot x\right) \cdot \left(\left(x \cdot x\right) \cdot x\right)}\right) - -0.5}{x} \]
        3. Step-by-step derivation
          1. lift--.f64N/A

            \[\leadsto \frac{\left(\left(\frac{\frac{1307076337763}{8543989815576}}{\left(\left(x \cdot x\right) \cdot x\right) \cdot x} - \frac{\frac{-600041}{2386628}}{x \cdot x}\right) - \frac{\frac{-344398180852034095277}{30586987988352776592}}{\left(\left(x \cdot x\right) \cdot x\right) \cdot \left(\left(x \cdot x\right) \cdot x\right)}\right) - \frac{-1}{2}}{x} \]
          2. lift--.f64N/A

            \[\leadsto \frac{\left(\left(\frac{\frac{1307076337763}{8543989815576}}{\left(\left(x \cdot x\right) \cdot x\right) \cdot x} - \frac{\frac{-600041}{2386628}}{x \cdot x}\right) - \frac{\frac{-344398180852034095277}{30586987988352776592}}{\left(\left(x \cdot x\right) \cdot x\right) \cdot \left(\left(x \cdot x\right) \cdot x\right)}\right) - \frac{-1}{2}}{x} \]
          3. associate--l-N/A

            \[\leadsto \frac{\left(\frac{\frac{1307076337763}{8543989815576}}{\left(\left(x \cdot x\right) \cdot x\right) \cdot x} - \frac{\frac{-600041}{2386628}}{x \cdot x}\right) - \left(\frac{\frac{-344398180852034095277}{30586987988352776592}}{\left(\left(x \cdot x\right) \cdot x\right) \cdot \left(\left(x \cdot x\right) \cdot x\right)} + \frac{-1}{2}\right)}{x} \]
          4. lower--.f64N/A

            \[\leadsto \frac{\left(\frac{\frac{1307076337763}{8543989815576}}{\left(\left(x \cdot x\right) \cdot x\right) \cdot x} - \frac{\frac{-600041}{2386628}}{x \cdot x}\right) - \left(\frac{\frac{-344398180852034095277}{30586987988352776592}}{\left(\left(x \cdot x\right) \cdot x\right) \cdot \left(\left(x \cdot x\right) \cdot x\right)} + \frac{-1}{2}\right)}{x} \]
        4. Applied rewrites51.4%

          \[\leadsto \frac{\frac{\frac{0.15298196345929074}{\left(x \cdot x\right) \cdot x} - \frac{-0.2514179000665374}{x}}{x} - \left(\frac{-11.259630434457211}{\left(\left(\left(\left(x \cdot x\right) \cdot x\right) \cdot x\right) \cdot x\right) \cdot x} - 0.5\right)}{x} \]
        5. Applied rewrites51.4%

          \[\leadsto \frac{\frac{\frac{\frac{0.15298196345929074}{x \cdot x} - -0.2514179000665374}{x} - \frac{-11.259630434457211}{\left(\left(x \cdot x\right) \cdot x\right) \cdot \left(x \cdot x\right)}}{x} - -0.5}{\color{blue}{x}} \]
      4. Recombined 2 regimes into one program.
      5. Add Preprocessing

      Alternative 3: 99.6% accurate, 6.0× speedup?

      \[\begin{array}{l} t_0 := \left|x\right| \cdot \left|x\right|\\ \mathsf{copysign}\left(1, x\right) \cdot \begin{array}{l} \mathbf{if}\;\left|x\right| \leq 1.15:\\ \;\;\;\;\mathsf{fma}\left(\mathsf{fma}\left(0.265709700396151, t\_0, -0.6665536072\right), t\_0, 1\right) \cdot \left|x\right|\\ \mathbf{else}:\\ \;\;\;\;\left(\frac{\frac{0.15298196345929074}{t\_0} - -0.2514179000665374}{t\_0} - -0.5\right) \cdot \frac{1}{\left|x\right|}\\ \end{array} \end{array} \]
      (FPCore (x)
       :precision binary64
       (let* ((t_0 (* (fabs x) (fabs x))))
         (*
          (copysign 1.0 x)
          (if (<= (fabs x) 1.15)
            (* (fma (fma 0.265709700396151 t_0 -0.6665536072) t_0 1.0) (fabs x))
            (*
             (- (/ (- (/ 0.15298196345929074 t_0) -0.2514179000665374) t_0) -0.5)
             (/ 1.0 (fabs x)))))))
      double code(double x) {
      	double t_0 = fabs(x) * fabs(x);
      	double tmp;
      	if (fabs(x) <= 1.15) {
      		tmp = fma(fma(0.265709700396151, t_0, -0.6665536072), t_0, 1.0) * fabs(x);
      	} else {
      		tmp = ((((0.15298196345929074 / t_0) - -0.2514179000665374) / t_0) - -0.5) * (1.0 / fabs(x));
      	}
      	return copysign(1.0, x) * tmp;
      }
      
      function code(x)
      	t_0 = Float64(abs(x) * abs(x))
      	tmp = 0.0
      	if (abs(x) <= 1.15)
      		tmp = Float64(fma(fma(0.265709700396151, t_0, -0.6665536072), t_0, 1.0) * abs(x));
      	else
      		tmp = Float64(Float64(Float64(Float64(Float64(0.15298196345929074 / t_0) - -0.2514179000665374) / t_0) - -0.5) * Float64(1.0 / abs(x)));
      	end
      	return Float64(copysign(1.0, x) * tmp)
      end
      
      code[x_] := Block[{t$95$0 = N[(N[Abs[x], $MachinePrecision] * N[Abs[x], $MachinePrecision]), $MachinePrecision]}, N[(N[With[{TMP1 = Abs[1.0], TMP2 = Sign[x]}, TMP1 * If[TMP2 == 0, 1, TMP2]], $MachinePrecision] * If[LessEqual[N[Abs[x], $MachinePrecision], 1.15], N[(N[(N[(0.265709700396151 * t$95$0 + -0.6665536072), $MachinePrecision] * t$95$0 + 1.0), $MachinePrecision] * N[Abs[x], $MachinePrecision]), $MachinePrecision], N[(N[(N[(N[(N[(0.15298196345929074 / t$95$0), $MachinePrecision] - -0.2514179000665374), $MachinePrecision] / t$95$0), $MachinePrecision] - -0.5), $MachinePrecision] * N[(1.0 / N[Abs[x], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]]), $MachinePrecision]]
      
      \begin{array}{l}
      t_0 := \left|x\right| \cdot \left|x\right|\\
      \mathsf{copysign}\left(1, x\right) \cdot \begin{array}{l}
      \mathbf{if}\;\left|x\right| \leq 1.15:\\
      \;\;\;\;\mathsf{fma}\left(\mathsf{fma}\left(0.265709700396151, t\_0, -0.6665536072\right), t\_0, 1\right) \cdot \left|x\right|\\
      
      \mathbf{else}:\\
      \;\;\;\;\left(\frac{\frac{0.15298196345929074}{t\_0} - -0.2514179000665374}{t\_0} - -0.5\right) \cdot \frac{1}{\left|x\right|}\\
      
      
      \end{array}
      \end{array}
      
      Derivation
      1. Split input into 2 regimes
      2. if x < 1.1499999999999999

        1. Initial program 53.9%

          \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
        2. Taylor expanded in x around 0

          \[\leadsto \color{blue}{\left(1 + {x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right)\right)} \cdot x \]
        3. Step-by-step derivation
          1. lower-+.f64N/A

            \[\leadsto \left(1 + \color{blue}{{x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right)}\right) \cdot x \]
          2. lower-*.f64N/A

            \[\leadsto \left(1 + {x}^{2} \cdot \color{blue}{\left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right)}\right) \cdot x \]
        4. Applied rewrites50.9%

          \[\leadsto \color{blue}{\left(1 + {x}^{2} \cdot \left(0.265709700396151 \cdot {x}^{2} - 0.6665536072\right)\right)} \cdot x \]
        5. Step-by-step derivation
          1. lift-+.f64N/A

            \[\leadsto \left(1 + \color{blue}{{x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right)}\right) \cdot x \]
          2. +-commutativeN/A

            \[\leadsto \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right) + \color{blue}{1}\right) \cdot x \]
        6. Applied rewrites50.9%

          \[\leadsto \color{blue}{\mathsf{fma}\left(\mathsf{fma}\left(0.265709700396151, x \cdot x, -0.6665536072\right), x \cdot x, 1\right)} \cdot x \]

        if 1.1499999999999999 < x

        1. Initial program 53.9%

          \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
        2. Taylor expanded in x around inf

          \[\leadsto \color{blue}{\frac{\frac{1}{2} + \left(\frac{\frac{1307076337763}{8543989815576}}{{x}^{4}} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}\right)}{x}} \]
        3. Step-by-step derivation
          1. lower-/.f64N/A

            \[\leadsto \frac{\frac{1}{2} + \left(\frac{\frac{1307076337763}{8543989815576}}{{x}^{4}} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}\right)}{\color{blue}{x}} \]
        4. Applied rewrites51.4%

          \[\leadsto \color{blue}{\frac{0.5 + \left(\frac{0.15298196345929074}{{x}^{4}} + 0.2514179000665374 \cdot \frac{1}{{x}^{2}}\right)}{x}} \]
        5. Step-by-step derivation
          1. lift-/.f64N/A

            \[\leadsto \frac{\frac{1}{2} + \left(\frac{\frac{1307076337763}{8543989815576}}{{x}^{4}} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}\right)}{\color{blue}{x}} \]
          2. mult-flipN/A

            \[\leadsto \left(\frac{1}{2} + \left(\frac{\frac{1307076337763}{8543989815576}}{{x}^{4}} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}\right)\right) \cdot \color{blue}{\frac{1}{x}} \]
          3. lower-*.f64N/A

            \[\leadsto \left(\frac{1}{2} + \left(\frac{\frac{1307076337763}{8543989815576}}{{x}^{4}} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}\right)\right) \cdot \color{blue}{\frac{1}{x}} \]
        6. Applied rewrites51.4%

          \[\leadsto \left(\left(\frac{0.15298196345929074}{\left(\left(x \cdot x\right) \cdot x\right) \cdot x} - \frac{-0.2514179000665374}{x \cdot x}\right) - -0.5\right) \cdot \color{blue}{\frac{1}{x}} \]
        7. Step-by-step derivation
          1. lift--.f64N/A

            \[\leadsto \left(\left(\frac{\frac{1307076337763}{8543989815576}}{\left(\left(x \cdot x\right) \cdot x\right) \cdot x} - \frac{\frac{-600041}{2386628}}{x \cdot x}\right) - \frac{-1}{2}\right) \cdot \frac{1}{x} \]
          2. lift-/.f64N/A

            \[\leadsto \left(\left(\frac{\frac{1307076337763}{8543989815576}}{\left(\left(x \cdot x\right) \cdot x\right) \cdot x} - \frac{\frac{-600041}{2386628}}{x \cdot x}\right) - \frac{-1}{2}\right) \cdot \frac{1}{x} \]
          3. lift-*.f64N/A

            \[\leadsto \left(\left(\frac{\frac{1307076337763}{8543989815576}}{\left(\left(x \cdot x\right) \cdot x\right) \cdot x} - \frac{\frac{-600041}{2386628}}{x \cdot x}\right) - \frac{-1}{2}\right) \cdot \frac{1}{x} \]
          4. associate-/r*N/A

            \[\leadsto \left(\left(\frac{\frac{\frac{1307076337763}{8543989815576}}{\left(x \cdot x\right) \cdot x}}{x} - \frac{\frac{-600041}{2386628}}{x \cdot x}\right) - \frac{-1}{2}\right) \cdot \frac{1}{x} \]
          5. lift-/.f64N/A

            \[\leadsto \left(\left(\frac{\frac{\frac{1307076337763}{8543989815576}}{\left(x \cdot x\right) \cdot x}}{x} - \frac{\frac{-600041}{2386628}}{x \cdot x}\right) - \frac{-1}{2}\right) \cdot \frac{1}{x} \]
          6. lift-*.f64N/A

            \[\leadsto \left(\left(\frac{\frac{\frac{1307076337763}{8543989815576}}{\left(x \cdot x\right) \cdot x}}{x} - \frac{\frac{-600041}{2386628}}{x \cdot x}\right) - \frac{-1}{2}\right) \cdot \frac{1}{x} \]
          7. associate-/r*N/A

            \[\leadsto \left(\left(\frac{\frac{\frac{1307076337763}{8543989815576}}{\left(x \cdot x\right) \cdot x}}{x} - \frac{\frac{\frac{-600041}{2386628}}{x}}{x}\right) - \frac{-1}{2}\right) \cdot \frac{1}{x} \]
          8. sub-divN/A

            \[\leadsto \left(\frac{\frac{\frac{1307076337763}{8543989815576}}{\left(x \cdot x\right) \cdot x} - \frac{\frac{-600041}{2386628}}{x}}{x} - \frac{-1}{2}\right) \cdot \frac{1}{x} \]
          9. lower-/.f64N/A

            \[\leadsto \left(\frac{\frac{\frac{1307076337763}{8543989815576}}{\left(x \cdot x\right) \cdot x} - \frac{\frac{-600041}{2386628}}{x}}{x} - \frac{-1}{2}\right) \cdot \frac{1}{x} \]
          10. lower--.f64N/A

            \[\leadsto \left(\frac{\frac{\frac{1307076337763}{8543989815576}}{\left(x \cdot x\right) \cdot x} - \frac{\frac{-600041}{2386628}}{x}}{x} - \frac{-1}{2}\right) \cdot \frac{1}{x} \]
          11. lower-/.f64N/A

            \[\leadsto \left(\frac{\frac{\frac{1307076337763}{8543989815576}}{\left(x \cdot x\right) \cdot x} - \frac{\frac{-600041}{2386628}}{x}}{x} - \frac{-1}{2}\right) \cdot \frac{1}{x} \]
          12. lower-/.f6451.4

            \[\leadsto \left(\frac{\frac{0.15298196345929074}{\left(x \cdot x\right) \cdot x} - \frac{-0.2514179000665374}{x}}{x} - -0.5\right) \cdot \frac{1}{x} \]
        8. Applied rewrites51.4%

          \[\leadsto \left(\frac{\frac{0.15298196345929074}{\left(x \cdot x\right) \cdot x} - \frac{-0.2514179000665374}{x}}{x} - -0.5\right) \cdot \frac{\color{blue}{1}}{x} \]
        9. Step-by-step derivation
          1. lift-/.f64N/A

            \[\leadsto \left(\frac{\frac{\frac{1307076337763}{8543989815576}}{\left(x \cdot x\right) \cdot x} - \frac{\frac{-600041}{2386628}}{x}}{x} - \frac{-1}{2}\right) \cdot \frac{1}{x} \]
          2. lift--.f64N/A

            \[\leadsto \left(\frac{\frac{\frac{1307076337763}{8543989815576}}{\left(x \cdot x\right) \cdot x} - \frac{\frac{-600041}{2386628}}{x}}{x} - \frac{-1}{2}\right) \cdot \frac{1}{x} \]
          3. lift-/.f64N/A

            \[\leadsto \left(\frac{\frac{\frac{1307076337763}{8543989815576}}{\left(x \cdot x\right) \cdot x} - \frac{\frac{-600041}{2386628}}{x}}{x} - \frac{-1}{2}\right) \cdot \frac{1}{x} \]
          4. sub-to-fractionN/A

            \[\leadsto \left(\frac{\frac{\frac{\frac{1307076337763}{8543989815576}}{\left(x \cdot x\right) \cdot x} \cdot x - \frac{-600041}{2386628}}{x}}{x} - \frac{-1}{2}\right) \cdot \frac{1}{x} \]
          5. associate-/l/N/A

            \[\leadsto \left(\frac{\frac{\frac{1307076337763}{8543989815576}}{\left(x \cdot x\right) \cdot x} \cdot x - \frac{-600041}{2386628}}{x \cdot x} - \frac{-1}{2}\right) \cdot \frac{1}{x} \]
          6. lower-/.f64N/A

            \[\leadsto \left(\frac{\frac{\frac{1307076337763}{8543989815576}}{\left(x \cdot x\right) \cdot x} \cdot x - \frac{-600041}{2386628}}{x \cdot x} - \frac{-1}{2}\right) \cdot \frac{1}{x} \]
        10. Applied rewrites51.4%

          \[\leadsto \left(\frac{\frac{0.15298196345929074}{x \cdot x} - -0.2514179000665374}{x \cdot x} - -0.5\right) \cdot \frac{1}{x} \]
      3. Recombined 2 regimes into one program.
      4. Add Preprocessing

      Alternative 4: 99.6% accurate, 6.5× speedup?

      \[\begin{array}{l} t_0 := \left|x\right| \cdot \left|x\right|\\ \mathsf{copysign}\left(1, x\right) \cdot \begin{array}{l} \mathbf{if}\;\left|x\right| \leq 1.15:\\ \;\;\;\;\mathsf{fma}\left(\mathsf{fma}\left(0.265709700396151, t\_0, -0.6665536072\right), t\_0, 1\right) \cdot \left|x\right|\\ \mathbf{else}:\\ \;\;\;\;\frac{\frac{\frac{0.15298196345929074}{t\_0} - -0.2514179000665374}{t\_0} - -0.5}{\left|x\right|}\\ \end{array} \end{array} \]
      (FPCore (x)
       :precision binary64
       (let* ((t_0 (* (fabs x) (fabs x))))
         (*
          (copysign 1.0 x)
          (if (<= (fabs x) 1.15)
            (* (fma (fma 0.265709700396151 t_0 -0.6665536072) t_0 1.0) (fabs x))
            (/
             (- (/ (- (/ 0.15298196345929074 t_0) -0.2514179000665374) t_0) -0.5)
             (fabs x))))))
      double code(double x) {
      	double t_0 = fabs(x) * fabs(x);
      	double tmp;
      	if (fabs(x) <= 1.15) {
      		tmp = fma(fma(0.265709700396151, t_0, -0.6665536072), t_0, 1.0) * fabs(x);
      	} else {
      		tmp = ((((0.15298196345929074 / t_0) - -0.2514179000665374) / t_0) - -0.5) / fabs(x);
      	}
      	return copysign(1.0, x) * tmp;
      }
      
      function code(x)
      	t_0 = Float64(abs(x) * abs(x))
      	tmp = 0.0
      	if (abs(x) <= 1.15)
      		tmp = Float64(fma(fma(0.265709700396151, t_0, -0.6665536072), t_0, 1.0) * abs(x));
      	else
      		tmp = Float64(Float64(Float64(Float64(Float64(0.15298196345929074 / t_0) - -0.2514179000665374) / t_0) - -0.5) / abs(x));
      	end
      	return Float64(copysign(1.0, x) * tmp)
      end
      
      code[x_] := Block[{t$95$0 = N[(N[Abs[x], $MachinePrecision] * N[Abs[x], $MachinePrecision]), $MachinePrecision]}, N[(N[With[{TMP1 = Abs[1.0], TMP2 = Sign[x]}, TMP1 * If[TMP2 == 0, 1, TMP2]], $MachinePrecision] * If[LessEqual[N[Abs[x], $MachinePrecision], 1.15], N[(N[(N[(0.265709700396151 * t$95$0 + -0.6665536072), $MachinePrecision] * t$95$0 + 1.0), $MachinePrecision] * N[Abs[x], $MachinePrecision]), $MachinePrecision], N[(N[(N[(N[(N[(0.15298196345929074 / t$95$0), $MachinePrecision] - -0.2514179000665374), $MachinePrecision] / t$95$0), $MachinePrecision] - -0.5), $MachinePrecision] / N[Abs[x], $MachinePrecision]), $MachinePrecision]]), $MachinePrecision]]
      
      \begin{array}{l}
      t_0 := \left|x\right| \cdot \left|x\right|\\
      \mathsf{copysign}\left(1, x\right) \cdot \begin{array}{l}
      \mathbf{if}\;\left|x\right| \leq 1.15:\\
      \;\;\;\;\mathsf{fma}\left(\mathsf{fma}\left(0.265709700396151, t\_0, -0.6665536072\right), t\_0, 1\right) \cdot \left|x\right|\\
      
      \mathbf{else}:\\
      \;\;\;\;\frac{\frac{\frac{0.15298196345929074}{t\_0} - -0.2514179000665374}{t\_0} - -0.5}{\left|x\right|}\\
      
      
      \end{array}
      \end{array}
      
      Derivation
      1. Split input into 2 regimes
      2. if x < 1.1499999999999999

        1. Initial program 53.9%

          \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
        2. Taylor expanded in x around 0

          \[\leadsto \color{blue}{\left(1 + {x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right)\right)} \cdot x \]
        3. Step-by-step derivation
          1. lower-+.f64N/A

            \[\leadsto \left(1 + \color{blue}{{x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right)}\right) \cdot x \]
          2. lower-*.f64N/A

            \[\leadsto \left(1 + {x}^{2} \cdot \color{blue}{\left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right)}\right) \cdot x \]
        4. Applied rewrites50.9%

          \[\leadsto \color{blue}{\left(1 + {x}^{2} \cdot \left(0.265709700396151 \cdot {x}^{2} - 0.6665536072\right)\right)} \cdot x \]
        5. Step-by-step derivation
          1. lift-+.f64N/A

            \[\leadsto \left(1 + \color{blue}{{x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right)}\right) \cdot x \]
          2. +-commutativeN/A

            \[\leadsto \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right) + \color{blue}{1}\right) \cdot x \]
        6. Applied rewrites50.9%

          \[\leadsto \color{blue}{\mathsf{fma}\left(\mathsf{fma}\left(0.265709700396151, x \cdot x, -0.6665536072\right), x \cdot x, 1\right)} \cdot x \]

        if 1.1499999999999999 < x

        1. Initial program 53.9%

          \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
        2. Taylor expanded in x around inf

          \[\leadsto \color{blue}{\frac{\frac{1}{2} + \left(\frac{\frac{1307076337763}{8543989815576}}{{x}^{4}} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}\right)}{x}} \]
        3. Step-by-step derivation
          1. lower-/.f64N/A

            \[\leadsto \frac{\frac{1}{2} + \left(\frac{\frac{1307076337763}{8543989815576}}{{x}^{4}} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}\right)}{\color{blue}{x}} \]
        4. Applied rewrites51.4%

          \[\leadsto \color{blue}{\frac{0.5 + \left(\frac{0.15298196345929074}{{x}^{4}} + 0.2514179000665374 \cdot \frac{1}{{x}^{2}}\right)}{x}} \]
        5. Step-by-step derivation
          1. lift-/.f64N/A

            \[\leadsto \frac{\frac{1}{2} + \left(\frac{\frac{1307076337763}{8543989815576}}{{x}^{4}} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}\right)}{\color{blue}{x}} \]
          2. mult-flipN/A

            \[\leadsto \left(\frac{1}{2} + \left(\frac{\frac{1307076337763}{8543989815576}}{{x}^{4}} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}\right)\right) \cdot \color{blue}{\frac{1}{x}} \]
          3. lower-*.f64N/A

            \[\leadsto \left(\frac{1}{2} + \left(\frac{\frac{1307076337763}{8543989815576}}{{x}^{4}} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}\right)\right) \cdot \color{blue}{\frac{1}{x}} \]
        6. Applied rewrites51.4%

          \[\leadsto \left(\left(\frac{0.15298196345929074}{\left(\left(x \cdot x\right) \cdot x\right) \cdot x} - \frac{-0.2514179000665374}{x \cdot x}\right) - -0.5\right) \cdot \color{blue}{\frac{1}{x}} \]
        7. Step-by-step derivation
          1. Applied rewrites51.4%

            \[\leadsto \color{blue}{\frac{\frac{\frac{0.15298196345929074}{\left(x \cdot x\right) \cdot x} - \frac{-0.2514179000665374}{x}}{x} - -0.5}{x}} \]
          2. Step-by-step derivation
            1. Applied rewrites51.4%

              \[\leadsto \frac{\frac{\frac{0.15298196345929074}{x \cdot x} - -0.2514179000665374}{x \cdot x} - -0.5}{x} \]
          3. Recombined 2 regimes into one program.
          4. Add Preprocessing

          Alternative 5: 99.6% accurate, 6.9× speedup?

          \[\begin{array}{l} t_0 := \left|x\right| \cdot \left|x\right|\\ \mathsf{copysign}\left(1, x\right) \cdot \begin{array}{l} \mathbf{if}\;\left|x\right| \leq 1.1:\\ \;\;\;\;\mathsf{fma}\left(\mathsf{fma}\left(0.265709700396151, t\_0, -0.6665536072\right), t\_0, 1\right) \cdot \left|x\right|\\ \mathbf{else}:\\ \;\;\;\;\frac{\frac{0.2514179000665374}{t\_0} - -0.5}{\left|x\right|}\\ \end{array} \end{array} \]
          (FPCore (x)
           :precision binary64
           (let* ((t_0 (* (fabs x) (fabs x))))
             (*
              (copysign 1.0 x)
              (if (<= (fabs x) 1.1)
                (* (fma (fma 0.265709700396151 t_0 -0.6665536072) t_0 1.0) (fabs x))
                (/ (- (/ 0.2514179000665374 t_0) -0.5) (fabs x))))))
          double code(double x) {
          	double t_0 = fabs(x) * fabs(x);
          	double tmp;
          	if (fabs(x) <= 1.1) {
          		tmp = fma(fma(0.265709700396151, t_0, -0.6665536072), t_0, 1.0) * fabs(x);
          	} else {
          		tmp = ((0.2514179000665374 / t_0) - -0.5) / fabs(x);
          	}
          	return copysign(1.0, x) * tmp;
          }
          
          function code(x)
          	t_0 = Float64(abs(x) * abs(x))
          	tmp = 0.0
          	if (abs(x) <= 1.1)
          		tmp = Float64(fma(fma(0.265709700396151, t_0, -0.6665536072), t_0, 1.0) * abs(x));
          	else
          		tmp = Float64(Float64(Float64(0.2514179000665374 / t_0) - -0.5) / abs(x));
          	end
          	return Float64(copysign(1.0, x) * tmp)
          end
          
          code[x_] := Block[{t$95$0 = N[(N[Abs[x], $MachinePrecision] * N[Abs[x], $MachinePrecision]), $MachinePrecision]}, N[(N[With[{TMP1 = Abs[1.0], TMP2 = Sign[x]}, TMP1 * If[TMP2 == 0, 1, TMP2]], $MachinePrecision] * If[LessEqual[N[Abs[x], $MachinePrecision], 1.1], N[(N[(N[(0.265709700396151 * t$95$0 + -0.6665536072), $MachinePrecision] * t$95$0 + 1.0), $MachinePrecision] * N[Abs[x], $MachinePrecision]), $MachinePrecision], N[(N[(N[(0.2514179000665374 / t$95$0), $MachinePrecision] - -0.5), $MachinePrecision] / N[Abs[x], $MachinePrecision]), $MachinePrecision]]), $MachinePrecision]]
          
          \begin{array}{l}
          t_0 := \left|x\right| \cdot \left|x\right|\\
          \mathsf{copysign}\left(1, x\right) \cdot \begin{array}{l}
          \mathbf{if}\;\left|x\right| \leq 1.1:\\
          \;\;\;\;\mathsf{fma}\left(\mathsf{fma}\left(0.265709700396151, t\_0, -0.6665536072\right), t\_0, 1\right) \cdot \left|x\right|\\
          
          \mathbf{else}:\\
          \;\;\;\;\frac{\frac{0.2514179000665374}{t\_0} - -0.5}{\left|x\right|}\\
          
          
          \end{array}
          \end{array}
          
          Derivation
          1. Split input into 2 regimes
          2. if x < 1.1000000000000001

            1. Initial program 53.9%

              \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
            2. Taylor expanded in x around 0

              \[\leadsto \color{blue}{\left(1 + {x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right)\right)} \cdot x \]
            3. Step-by-step derivation
              1. lower-+.f64N/A

                \[\leadsto \left(1 + \color{blue}{{x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right)}\right) \cdot x \]
              2. lower-*.f64N/A

                \[\leadsto \left(1 + {x}^{2} \cdot \color{blue}{\left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right)}\right) \cdot x \]
            4. Applied rewrites50.9%

              \[\leadsto \color{blue}{\left(1 + {x}^{2} \cdot \left(0.265709700396151 \cdot {x}^{2} - 0.6665536072\right)\right)} \cdot x \]
            5. Step-by-step derivation
              1. lift-+.f64N/A

                \[\leadsto \left(1 + \color{blue}{{x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right)}\right) \cdot x \]
              2. +-commutativeN/A

                \[\leadsto \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right) + \color{blue}{1}\right) \cdot x \]
            6. Applied rewrites50.9%

              \[\leadsto \color{blue}{\mathsf{fma}\left(\mathsf{fma}\left(0.265709700396151, x \cdot x, -0.6665536072\right), x \cdot x, 1\right)} \cdot x \]

            if 1.1000000000000001 < x

            1. Initial program 53.9%

              \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
            2. Taylor expanded in x around inf

              \[\leadsto \color{blue}{\frac{\frac{1}{2} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}}{x}} \]
            3. Step-by-step derivation
              1. lower-/.f64N/A

                \[\leadsto \frac{\frac{1}{2} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}}{\color{blue}{x}} \]
            4. Applied rewrites51.5%

              \[\leadsto \color{blue}{\frac{0.5 + 0.2514179000665374 \cdot \frac{1}{{x}^{2}}}{x}} \]
            5. Step-by-step derivation
              1. lift-+.f64N/A

                \[\leadsto \frac{\frac{1}{2} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}}{x} \]
              2. +-commutativeN/A

                \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} + \frac{1}{2}}{x} \]
              3. add-flipN/A

                \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
              4. lower--.f64N/A

                \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
              5. lift-*.f64N/A

                \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
              6. lift-/.f64N/A

                \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
              7. lift-pow.f64N/A

                \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
              8. pow2N/A

                \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{x \cdot x} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
              9. mult-flip-revN/A

                \[\leadsto \frac{\frac{\frac{600041}{2386628}}{x \cdot x} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
              10. lower-/.f64N/A

                \[\leadsto \frac{\frac{\frac{600041}{2386628}}{x \cdot x} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
              11. lower-*.f64N/A

                \[\leadsto \frac{\frac{\frac{600041}{2386628}}{x \cdot x} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
              12. metadata-eval51.5

                \[\leadsto \frac{\frac{0.2514179000665374}{x \cdot x} - -0.5}{x} \]
            6. Applied rewrites51.5%

              \[\leadsto \frac{\frac{0.2514179000665374}{x \cdot x} - -0.5}{\color{blue}{x}} \]
          3. Recombined 2 regimes into one program.
          4. Add Preprocessing

          Alternative 6: 99.5% accurate, 9.1× speedup?

          \[\begin{array}{l} t_0 := \left|x\right| \cdot \left|x\right|\\ \mathsf{copysign}\left(1, x\right) \cdot \begin{array}{l} \mathbf{if}\;\left|x\right| \leq 1.25:\\ \;\;\;\;\mathsf{fma}\left(t\_0, -0.6665536072, 1\right) \cdot \left|x\right|\\ \mathbf{else}:\\ \;\;\;\;\frac{\frac{0.2514179000665374}{t\_0} - -0.5}{\left|x\right|}\\ \end{array} \end{array} \]
          (FPCore (x)
           :precision binary64
           (let* ((t_0 (* (fabs x) (fabs x))))
             (*
              (copysign 1.0 x)
              (if (<= (fabs x) 1.25)
                (* (fma t_0 -0.6665536072 1.0) (fabs x))
                (/ (- (/ 0.2514179000665374 t_0) -0.5) (fabs x))))))
          double code(double x) {
          	double t_0 = fabs(x) * fabs(x);
          	double tmp;
          	if (fabs(x) <= 1.25) {
          		tmp = fma(t_0, -0.6665536072, 1.0) * fabs(x);
          	} else {
          		tmp = ((0.2514179000665374 / t_0) - -0.5) / fabs(x);
          	}
          	return copysign(1.0, x) * tmp;
          }
          
          function code(x)
          	t_0 = Float64(abs(x) * abs(x))
          	tmp = 0.0
          	if (abs(x) <= 1.25)
          		tmp = Float64(fma(t_0, -0.6665536072, 1.0) * abs(x));
          	else
          		tmp = Float64(Float64(Float64(0.2514179000665374 / t_0) - -0.5) / abs(x));
          	end
          	return Float64(copysign(1.0, x) * tmp)
          end
          
          code[x_] := Block[{t$95$0 = N[(N[Abs[x], $MachinePrecision] * N[Abs[x], $MachinePrecision]), $MachinePrecision]}, N[(N[With[{TMP1 = Abs[1.0], TMP2 = Sign[x]}, TMP1 * If[TMP2 == 0, 1, TMP2]], $MachinePrecision] * If[LessEqual[N[Abs[x], $MachinePrecision], 1.25], N[(N[(t$95$0 * -0.6665536072 + 1.0), $MachinePrecision] * N[Abs[x], $MachinePrecision]), $MachinePrecision], N[(N[(N[(0.2514179000665374 / t$95$0), $MachinePrecision] - -0.5), $MachinePrecision] / N[Abs[x], $MachinePrecision]), $MachinePrecision]]), $MachinePrecision]]
          
          \begin{array}{l}
          t_0 := \left|x\right| \cdot \left|x\right|\\
          \mathsf{copysign}\left(1, x\right) \cdot \begin{array}{l}
          \mathbf{if}\;\left|x\right| \leq 1.25:\\
          \;\;\;\;\mathsf{fma}\left(t\_0, -0.6665536072, 1\right) \cdot \left|x\right|\\
          
          \mathbf{else}:\\
          \;\;\;\;\frac{\frac{0.2514179000665374}{t\_0} - -0.5}{\left|x\right|}\\
          
          
          \end{array}
          \end{array}
          
          Derivation
          1. Split input into 2 regimes
          2. if x < 1.25

            1. Initial program 53.9%

              \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
            2. Taylor expanded in x around 0

              \[\leadsto \color{blue}{\left(1 + \frac{-833192009}{1250000000} \cdot {x}^{2}\right)} \cdot x \]
            3. Step-by-step derivation
              1. lower-+.f64N/A

                \[\leadsto \left(1 + \color{blue}{\frac{-833192009}{1250000000} \cdot {x}^{2}}\right) \cdot x \]
              2. lower-*.f64N/A

                \[\leadsto \left(1 + \frac{-833192009}{1250000000} \cdot \color{blue}{{x}^{2}}\right) \cdot x \]
              3. lower-pow.f6450.1

                \[\leadsto \left(1 + -0.6665536072 \cdot {x}^{\color{blue}{2}}\right) \cdot x \]
            4. Applied rewrites50.1%

              \[\leadsto \color{blue}{\left(1 + -0.6665536072 \cdot {x}^{2}\right)} \cdot x \]
            5. Step-by-step derivation
              1. lift-+.f64N/A

                \[\leadsto \left(1 + \color{blue}{\frac{-833192009}{1250000000} \cdot {x}^{2}}\right) \cdot x \]
              2. +-commutativeN/A

                \[\leadsto \left(\frac{-833192009}{1250000000} \cdot {x}^{2} + \color{blue}{1}\right) \cdot x \]
              3. lift-*.f64N/A

                \[\leadsto \left(\frac{-833192009}{1250000000} \cdot {x}^{2} + 1\right) \cdot x \]
              4. lift-pow.f64N/A

                \[\leadsto \left(\frac{-833192009}{1250000000} \cdot {x}^{2} + 1\right) \cdot x \]
              5. pow2N/A

                \[\leadsto \left(\frac{-833192009}{1250000000} \cdot \left(x \cdot x\right) + 1\right) \cdot x \]
              6. lift-*.f64N/A

                \[\leadsto \left(\frac{-833192009}{1250000000} \cdot \left(x \cdot x\right) + 1\right) \cdot x \]
              7. *-commutativeN/A

                \[\leadsto \left(\left(x \cdot x\right) \cdot \frac{-833192009}{1250000000} + 1\right) \cdot x \]
              8. lower-fma.f6450.1

                \[\leadsto \mathsf{fma}\left(x \cdot x, \color{blue}{-0.6665536072}, 1\right) \cdot x \]
            6. Applied rewrites50.1%

              \[\leadsto \mathsf{fma}\left(x \cdot x, \color{blue}{-0.6665536072}, 1\right) \cdot x \]

            if 1.25 < x

            1. Initial program 53.9%

              \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
            2. Taylor expanded in x around inf

              \[\leadsto \color{blue}{\frac{\frac{1}{2} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}}{x}} \]
            3. Step-by-step derivation
              1. lower-/.f64N/A

                \[\leadsto \frac{\frac{1}{2} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}}{\color{blue}{x}} \]
            4. Applied rewrites51.5%

              \[\leadsto \color{blue}{\frac{0.5 + 0.2514179000665374 \cdot \frac{1}{{x}^{2}}}{x}} \]
            5. Step-by-step derivation
              1. lift-+.f64N/A

                \[\leadsto \frac{\frac{1}{2} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}}{x} \]
              2. +-commutativeN/A

                \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} + \frac{1}{2}}{x} \]
              3. add-flipN/A

                \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
              4. lower--.f64N/A

                \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
              5. lift-*.f64N/A

                \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
              6. lift-/.f64N/A

                \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
              7. lift-pow.f64N/A

                \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
              8. pow2N/A

                \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{x \cdot x} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
              9. mult-flip-revN/A

                \[\leadsto \frac{\frac{\frac{600041}{2386628}}{x \cdot x} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
              10. lower-/.f64N/A

                \[\leadsto \frac{\frac{\frac{600041}{2386628}}{x \cdot x} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
              11. lower-*.f64N/A

                \[\leadsto \frac{\frac{\frac{600041}{2386628}}{x \cdot x} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
              12. metadata-eval51.5

                \[\leadsto \frac{\frac{0.2514179000665374}{x \cdot x} - -0.5}{x} \]
            6. Applied rewrites51.5%

              \[\leadsto \frac{\frac{0.2514179000665374}{x \cdot x} - -0.5}{\color{blue}{x}} \]
          3. Recombined 2 regimes into one program.
          4. Add Preprocessing

          Alternative 7: 99.2% accurate, 9.6× speedup?

          \[\mathsf{copysign}\left(1, x\right) \cdot \begin{array}{l} \mathbf{if}\;\left|x\right| \leq 1.25:\\ \;\;\;\;\mathsf{fma}\left(\left|x\right| \cdot \left|x\right|, -0.6665536072, 1\right) \cdot \left|x\right|\\ \mathbf{else}:\\ \;\;\;\;\frac{0.5}{\left|x\right|}\\ \end{array} \]
          (FPCore (x)
           :precision binary64
           (*
            (copysign 1.0 x)
            (if (<= (fabs x) 1.25)
              (* (fma (* (fabs x) (fabs x)) -0.6665536072 1.0) (fabs x))
              (/ 0.5 (fabs x)))))
          double code(double x) {
          	double tmp;
          	if (fabs(x) <= 1.25) {
          		tmp = fma((fabs(x) * fabs(x)), -0.6665536072, 1.0) * fabs(x);
          	} else {
          		tmp = 0.5 / fabs(x);
          	}
          	return copysign(1.0, x) * tmp;
          }
          
          function code(x)
          	tmp = 0.0
          	if (abs(x) <= 1.25)
          		tmp = Float64(fma(Float64(abs(x) * abs(x)), -0.6665536072, 1.0) * abs(x));
          	else
          		tmp = Float64(0.5 / abs(x));
          	end
          	return Float64(copysign(1.0, x) * tmp)
          end
          
          code[x_] := N[(N[With[{TMP1 = Abs[1.0], TMP2 = Sign[x]}, TMP1 * If[TMP2 == 0, 1, TMP2]], $MachinePrecision] * If[LessEqual[N[Abs[x], $MachinePrecision], 1.25], N[(N[(N[(N[Abs[x], $MachinePrecision] * N[Abs[x], $MachinePrecision]), $MachinePrecision] * -0.6665536072 + 1.0), $MachinePrecision] * N[Abs[x], $MachinePrecision]), $MachinePrecision], N[(0.5 / N[Abs[x], $MachinePrecision]), $MachinePrecision]]), $MachinePrecision]
          
          \mathsf{copysign}\left(1, x\right) \cdot \begin{array}{l}
          \mathbf{if}\;\left|x\right| \leq 1.25:\\
          \;\;\;\;\mathsf{fma}\left(\left|x\right| \cdot \left|x\right|, -0.6665536072, 1\right) \cdot \left|x\right|\\
          
          \mathbf{else}:\\
          \;\;\;\;\frac{0.5}{\left|x\right|}\\
          
          
          \end{array}
          
          Derivation
          1. Split input into 2 regimes
          2. if x < 1.25

            1. Initial program 53.9%

              \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
            2. Taylor expanded in x around 0

              \[\leadsto \color{blue}{\left(1 + \frac{-833192009}{1250000000} \cdot {x}^{2}\right)} \cdot x \]
            3. Step-by-step derivation
              1. lower-+.f64N/A

                \[\leadsto \left(1 + \color{blue}{\frac{-833192009}{1250000000} \cdot {x}^{2}}\right) \cdot x \]
              2. lower-*.f64N/A

                \[\leadsto \left(1 + \frac{-833192009}{1250000000} \cdot \color{blue}{{x}^{2}}\right) \cdot x \]
              3. lower-pow.f6450.1

                \[\leadsto \left(1 + -0.6665536072 \cdot {x}^{\color{blue}{2}}\right) \cdot x \]
            4. Applied rewrites50.1%

              \[\leadsto \color{blue}{\left(1 + -0.6665536072 \cdot {x}^{2}\right)} \cdot x \]
            5. Step-by-step derivation
              1. lift-+.f64N/A

                \[\leadsto \left(1 + \color{blue}{\frac{-833192009}{1250000000} \cdot {x}^{2}}\right) \cdot x \]
              2. +-commutativeN/A

                \[\leadsto \left(\frac{-833192009}{1250000000} \cdot {x}^{2} + \color{blue}{1}\right) \cdot x \]
              3. lift-*.f64N/A

                \[\leadsto \left(\frac{-833192009}{1250000000} \cdot {x}^{2} + 1\right) \cdot x \]
              4. lift-pow.f64N/A

                \[\leadsto \left(\frac{-833192009}{1250000000} \cdot {x}^{2} + 1\right) \cdot x \]
              5. pow2N/A

                \[\leadsto \left(\frac{-833192009}{1250000000} \cdot \left(x \cdot x\right) + 1\right) \cdot x \]
              6. lift-*.f64N/A

                \[\leadsto \left(\frac{-833192009}{1250000000} \cdot \left(x \cdot x\right) + 1\right) \cdot x \]
              7. *-commutativeN/A

                \[\leadsto \left(\left(x \cdot x\right) \cdot \frac{-833192009}{1250000000} + 1\right) \cdot x \]
              8. lower-fma.f6450.1

                \[\leadsto \mathsf{fma}\left(x \cdot x, \color{blue}{-0.6665536072}, 1\right) \cdot x \]
            6. Applied rewrites50.1%

              \[\leadsto \mathsf{fma}\left(x \cdot x, \color{blue}{-0.6665536072}, 1\right) \cdot x \]

            if 1.25 < x

            1. Initial program 53.9%

              \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
            2. Taylor expanded in x around inf

              \[\leadsto \color{blue}{\frac{\frac{1}{2}}{x}} \]
            3. Step-by-step derivation
              1. lower-/.f6451.7

                \[\leadsto \frac{0.5}{\color{blue}{x}} \]
            4. Applied rewrites51.7%

              \[\leadsto \color{blue}{\frac{0.5}{x}} \]
          3. Recombined 2 regimes into one program.
          4. Add Preprocessing

          Alternative 8: 98.9% accurate, 15.4× speedup?

          \[\mathsf{copysign}\left(1, x\right) \cdot \begin{array}{l} \mathbf{if}\;\left|x\right| \leq 0.72:\\ \;\;\;\;\left|x\right|\\ \mathbf{else}:\\ \;\;\;\;\frac{0.5}{\left|x\right|}\\ \end{array} \]
          (FPCore (x)
           :precision binary64
           (* (copysign 1.0 x) (if (<= (fabs x) 0.72) (fabs x) (/ 0.5 (fabs x)))))
          double code(double x) {
          	double tmp;
          	if (fabs(x) <= 0.72) {
          		tmp = fabs(x);
          	} else {
          		tmp = 0.5 / fabs(x);
          	}
          	return copysign(1.0, x) * tmp;
          }
          
          public static double code(double x) {
          	double tmp;
          	if (Math.abs(x) <= 0.72) {
          		tmp = Math.abs(x);
          	} else {
          		tmp = 0.5 / Math.abs(x);
          	}
          	return Math.copySign(1.0, x) * tmp;
          }
          
          def code(x):
          	tmp = 0
          	if math.fabs(x) <= 0.72:
          		tmp = math.fabs(x)
          	else:
          		tmp = 0.5 / math.fabs(x)
          	return math.copysign(1.0, x) * tmp
          
          function code(x)
          	tmp = 0.0
          	if (abs(x) <= 0.72)
          		tmp = abs(x);
          	else
          		tmp = Float64(0.5 / abs(x));
          	end
          	return Float64(copysign(1.0, x) * tmp)
          end
          
          function tmp_2 = code(x)
          	tmp = 0.0;
          	if (abs(x) <= 0.72)
          		tmp = abs(x);
          	else
          		tmp = 0.5 / abs(x);
          	end
          	tmp_2 = (sign(x) * abs(1.0)) * tmp;
          end
          
          code[x_] := N[(N[With[{TMP1 = Abs[1.0], TMP2 = Sign[x]}, TMP1 * If[TMP2 == 0, 1, TMP2]], $MachinePrecision] * If[LessEqual[N[Abs[x], $MachinePrecision], 0.72], N[Abs[x], $MachinePrecision], N[(0.5 / N[Abs[x], $MachinePrecision]), $MachinePrecision]]), $MachinePrecision]
          
          \mathsf{copysign}\left(1, x\right) \cdot \begin{array}{l}
          \mathbf{if}\;\left|x\right| \leq 0.72:\\
          \;\;\;\;\left|x\right|\\
          
          \mathbf{else}:\\
          \;\;\;\;\frac{0.5}{\left|x\right|}\\
          
          
          \end{array}
          
          Derivation
          1. Split input into 2 regimes
          2. if x < 0.71999999999999997

            1. Initial program 53.9%

              \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
            2. Taylor expanded in x around 0

              \[\leadsto \color{blue}{x} \]
            3. Step-by-step derivation
              1. Applied rewrites51.1%

                \[\leadsto \color{blue}{x} \]

              if 0.71999999999999997 < x

              1. Initial program 53.9%

                \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
              2. Taylor expanded in x around inf

                \[\leadsto \color{blue}{\frac{\frac{1}{2}}{x}} \]
              3. Step-by-step derivation
                1. lower-/.f6451.7

                  \[\leadsto \frac{0.5}{\color{blue}{x}} \]
              4. Applied rewrites51.7%

                \[\leadsto \color{blue}{\frac{0.5}{x}} \]
            4. Recombined 2 regimes into one program.
            5. Add Preprocessing

            Alternative 9: 51.1% accurate, 253.1× speedup?

            \[x \]
            (FPCore (x) :precision binary64 x)
            double code(double x) {
            	return x;
            }
            
            module fmin_fmax_functions
                implicit none
                private
                public fmax
                public fmin
            
                interface fmax
                    module procedure fmax88
                    module procedure fmax44
                    module procedure fmax84
                    module procedure fmax48
                end interface
                interface fmin
                    module procedure fmin88
                    module procedure fmin44
                    module procedure fmin84
                    module procedure fmin48
                end interface
            contains
                real(8) function fmax88(x, y) result (res)
                    real(8), intent (in) :: x
                    real(8), intent (in) :: y
                    res = merge(y, merge(x, max(x, y), y /= y), x /= x)
                end function
                real(4) function fmax44(x, y) result (res)
                    real(4), intent (in) :: x
                    real(4), intent (in) :: y
                    res = merge(y, merge(x, max(x, y), y /= y), x /= x)
                end function
                real(8) function fmax84(x, y) result(res)
                    real(8), intent (in) :: x
                    real(4), intent (in) :: y
                    res = merge(dble(y), merge(x, max(x, dble(y)), y /= y), x /= x)
                end function
                real(8) function fmax48(x, y) result(res)
                    real(4), intent (in) :: x
                    real(8), intent (in) :: y
                    res = merge(y, merge(dble(x), max(dble(x), y), y /= y), x /= x)
                end function
                real(8) function fmin88(x, y) result (res)
                    real(8), intent (in) :: x
                    real(8), intent (in) :: y
                    res = merge(y, merge(x, min(x, y), y /= y), x /= x)
                end function
                real(4) function fmin44(x, y) result (res)
                    real(4), intent (in) :: x
                    real(4), intent (in) :: y
                    res = merge(y, merge(x, min(x, y), y /= y), x /= x)
                end function
                real(8) function fmin84(x, y) result(res)
                    real(8), intent (in) :: x
                    real(4), intent (in) :: y
                    res = merge(dble(y), merge(x, min(x, dble(y)), y /= y), x /= x)
                end function
                real(8) function fmin48(x, y) result(res)
                    real(4), intent (in) :: x
                    real(8), intent (in) :: y
                    res = merge(y, merge(dble(x), min(dble(x), y), y /= y), x /= x)
                end function
            end module
            
            real(8) function code(x)
            use fmin_fmax_functions
                real(8), intent (in) :: x
                code = x
            end function
            
            public static double code(double x) {
            	return x;
            }
            
            def code(x):
            	return x
            
            function code(x)
            	return x
            end
            
            function tmp = code(x)
            	tmp = x;
            end
            
            code[x_] := x
            
            x
            
            Derivation
            1. Initial program 53.9%

              \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
            2. Taylor expanded in x around 0

              \[\leadsto \color{blue}{x} \]
            3. Step-by-step derivation
              1. Applied rewrites51.1%

                \[\leadsto \color{blue}{x} \]
              2. Add Preprocessing

              Reproduce

              ?
              herbie shell --seed 2025180 
              (FPCore (x)
                :name "Jmat.Real.dawson"
                :precision binary64
                (* (/ (+ (+ (+ (+ (+ 1.0 (* 0.1049934947 (* x x))) (* 0.0424060604 (* (* x x) (* x x)))) (* 0.0072644182 (* (* (* x x) (* x x)) (* x x)))) (* 0.0005064034 (* (* (* (* x x) (* x x)) (* x x)) (* x x)))) (* 0.0001789971 (* (* (* (* (* x x) (* x x)) (* x x)) (* x x)) (* x x)))) (+ (+ (+ (+ (+ (+ 1.0 (* 0.7715471019 (* x x))) (* 0.2909738639 (* (* x x) (* x x)))) (* 0.0694555761 (* (* (* x x) (* x x)) (* x x)))) (* 0.0140005442 (* (* (* (* x x) (* x x)) (* x x)) (* x x)))) (* 0.0008327945 (* (* (* (* (* x x) (* x x)) (* x x)) (* x x)) (* x x)))) (* (* 2.0 0.0001789971) (* (* (* (* (* (* x x) (* x x)) (* x x)) (* x x)) (* x x)) (* x x))))) x))