Jmat.Real.dawson

Percentage Accurate: 53.7% → 99.7%
Time: 6.1s
Alternatives: 8
Speedup: 15.4×

Specification

?
\[\begin{array}{l} t_0 := \left(x \cdot x\right) \cdot \left(x \cdot x\right)\\ t_1 := t\_0 \cdot \left(x \cdot x\right)\\ t_2 := t\_1 \cdot \left(x \cdot x\right)\\ t_3 := t\_2 \cdot \left(x \cdot x\right)\\ \frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot t\_0\right) + 0.0072644182 \cdot t\_1\right) + 0.0005064034 \cdot t\_2\right) + 0.0001789971 \cdot t\_3}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot t\_0\right) + 0.0694555761 \cdot t\_1\right) + 0.0140005442 \cdot t\_2\right) + 0.0008327945 \cdot t\_3\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(t\_3 \cdot \left(x \cdot x\right)\right)} \cdot x \end{array} \]
(FPCore (x)
 :precision binary64
 (let* ((t_0 (* (* x x) (* x x)))
        (t_1 (* t_0 (* x x)))
        (t_2 (* t_1 (* x x)))
        (t_3 (* t_2 (* x x))))
   (*
    (/
     (+
      (+
       (+
        (+ (+ 1.0 (* 0.1049934947 (* x x))) (* 0.0424060604 t_0))
        (* 0.0072644182 t_1))
       (* 0.0005064034 t_2))
      (* 0.0001789971 t_3))
     (+
      (+
       (+
        (+
         (+ (+ 1.0 (* 0.7715471019 (* x x))) (* 0.2909738639 t_0))
         (* 0.0694555761 t_1))
        (* 0.0140005442 t_2))
       (* 0.0008327945 t_3))
      (* (* 2.0 0.0001789971) (* t_3 (* x x)))))
    x)))
double code(double x) {
	double t_0 = (x * x) * (x * x);
	double t_1 = t_0 * (x * x);
	double t_2 = t_1 * (x * x);
	double t_3 = t_2 * (x * x);
	return ((((((1.0 + (0.1049934947 * (x * x))) + (0.0424060604 * t_0)) + (0.0072644182 * t_1)) + (0.0005064034 * t_2)) + (0.0001789971 * t_3)) / ((((((1.0 + (0.7715471019 * (x * x))) + (0.2909738639 * t_0)) + (0.0694555761 * t_1)) + (0.0140005442 * t_2)) + (0.0008327945 * t_3)) + ((2.0 * 0.0001789971) * (t_3 * (x * x))))) * x;
}
module fmin_fmax_functions
    implicit none
    private
    public fmax
    public fmin

    interface fmax
        module procedure fmax88
        module procedure fmax44
        module procedure fmax84
        module procedure fmax48
    end interface
    interface fmin
        module procedure fmin88
        module procedure fmin44
        module procedure fmin84
        module procedure fmin48
    end interface
contains
    real(8) function fmax88(x, y) result (res)
        real(8), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(x, max(x, y), y /= y), x /= x)
    end function
    real(4) function fmax44(x, y) result (res)
        real(4), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(y, merge(x, max(x, y), y /= y), x /= x)
    end function
    real(8) function fmax84(x, y) result(res)
        real(8), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(dble(y), merge(x, max(x, dble(y)), y /= y), x /= x)
    end function
    real(8) function fmax48(x, y) result(res)
        real(4), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(dble(x), max(dble(x), y), y /= y), x /= x)
    end function
    real(8) function fmin88(x, y) result (res)
        real(8), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(x, min(x, y), y /= y), x /= x)
    end function
    real(4) function fmin44(x, y) result (res)
        real(4), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(y, merge(x, min(x, y), y /= y), x /= x)
    end function
    real(8) function fmin84(x, y) result(res)
        real(8), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(dble(y), merge(x, min(x, dble(y)), y /= y), x /= x)
    end function
    real(8) function fmin48(x, y) result(res)
        real(4), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(dble(x), min(dble(x), y), y /= y), x /= x)
    end function
end module

real(8) function code(x)
use fmin_fmax_functions
    real(8), intent (in) :: x
    real(8) :: t_0
    real(8) :: t_1
    real(8) :: t_2
    real(8) :: t_3
    t_0 = (x * x) * (x * x)
    t_1 = t_0 * (x * x)
    t_2 = t_1 * (x * x)
    t_3 = t_2 * (x * x)
    code = ((((((1.0d0 + (0.1049934947d0 * (x * x))) + (0.0424060604d0 * t_0)) + (0.0072644182d0 * t_1)) + (0.0005064034d0 * t_2)) + (0.0001789971d0 * t_3)) / ((((((1.0d0 + (0.7715471019d0 * (x * x))) + (0.2909738639d0 * t_0)) + (0.0694555761d0 * t_1)) + (0.0140005442d0 * t_2)) + (0.0008327945d0 * t_3)) + ((2.0d0 * 0.0001789971d0) * (t_3 * (x * x))))) * x
end function
public static double code(double x) {
	double t_0 = (x * x) * (x * x);
	double t_1 = t_0 * (x * x);
	double t_2 = t_1 * (x * x);
	double t_3 = t_2 * (x * x);
	return ((((((1.0 + (0.1049934947 * (x * x))) + (0.0424060604 * t_0)) + (0.0072644182 * t_1)) + (0.0005064034 * t_2)) + (0.0001789971 * t_3)) / ((((((1.0 + (0.7715471019 * (x * x))) + (0.2909738639 * t_0)) + (0.0694555761 * t_1)) + (0.0140005442 * t_2)) + (0.0008327945 * t_3)) + ((2.0 * 0.0001789971) * (t_3 * (x * x))))) * x;
}
def code(x):
	t_0 = (x * x) * (x * x)
	t_1 = t_0 * (x * x)
	t_2 = t_1 * (x * x)
	t_3 = t_2 * (x * x)
	return ((((((1.0 + (0.1049934947 * (x * x))) + (0.0424060604 * t_0)) + (0.0072644182 * t_1)) + (0.0005064034 * t_2)) + (0.0001789971 * t_3)) / ((((((1.0 + (0.7715471019 * (x * x))) + (0.2909738639 * t_0)) + (0.0694555761 * t_1)) + (0.0140005442 * t_2)) + (0.0008327945 * t_3)) + ((2.0 * 0.0001789971) * (t_3 * (x * x))))) * x
function code(x)
	t_0 = Float64(Float64(x * x) * Float64(x * x))
	t_1 = Float64(t_0 * Float64(x * x))
	t_2 = Float64(t_1 * Float64(x * x))
	t_3 = Float64(t_2 * Float64(x * x))
	return Float64(Float64(Float64(Float64(Float64(Float64(Float64(1.0 + Float64(0.1049934947 * Float64(x * x))) + Float64(0.0424060604 * t_0)) + Float64(0.0072644182 * t_1)) + Float64(0.0005064034 * t_2)) + Float64(0.0001789971 * t_3)) / Float64(Float64(Float64(Float64(Float64(Float64(1.0 + Float64(0.7715471019 * Float64(x * x))) + Float64(0.2909738639 * t_0)) + Float64(0.0694555761 * t_1)) + Float64(0.0140005442 * t_2)) + Float64(0.0008327945 * t_3)) + Float64(Float64(2.0 * 0.0001789971) * Float64(t_3 * Float64(x * x))))) * x)
end
function tmp = code(x)
	t_0 = (x * x) * (x * x);
	t_1 = t_0 * (x * x);
	t_2 = t_1 * (x * x);
	t_3 = t_2 * (x * x);
	tmp = ((((((1.0 + (0.1049934947 * (x * x))) + (0.0424060604 * t_0)) + (0.0072644182 * t_1)) + (0.0005064034 * t_2)) + (0.0001789971 * t_3)) / ((((((1.0 + (0.7715471019 * (x * x))) + (0.2909738639 * t_0)) + (0.0694555761 * t_1)) + (0.0140005442 * t_2)) + (0.0008327945 * t_3)) + ((2.0 * 0.0001789971) * (t_3 * (x * x))))) * x;
end
code[x_] := Block[{t$95$0 = N[(N[(x * x), $MachinePrecision] * N[(x * x), $MachinePrecision]), $MachinePrecision]}, Block[{t$95$1 = N[(t$95$0 * N[(x * x), $MachinePrecision]), $MachinePrecision]}, Block[{t$95$2 = N[(t$95$1 * N[(x * x), $MachinePrecision]), $MachinePrecision]}, Block[{t$95$3 = N[(t$95$2 * N[(x * x), $MachinePrecision]), $MachinePrecision]}, N[(N[(N[(N[(N[(N[(N[(1.0 + N[(0.1049934947 * N[(x * x), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] + N[(0.0424060604 * t$95$0), $MachinePrecision]), $MachinePrecision] + N[(0.0072644182 * t$95$1), $MachinePrecision]), $MachinePrecision] + N[(0.0005064034 * t$95$2), $MachinePrecision]), $MachinePrecision] + N[(0.0001789971 * t$95$3), $MachinePrecision]), $MachinePrecision] / N[(N[(N[(N[(N[(N[(1.0 + N[(0.7715471019 * N[(x * x), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] + N[(0.2909738639 * t$95$0), $MachinePrecision]), $MachinePrecision] + N[(0.0694555761 * t$95$1), $MachinePrecision]), $MachinePrecision] + N[(0.0140005442 * t$95$2), $MachinePrecision]), $MachinePrecision] + N[(0.0008327945 * t$95$3), $MachinePrecision]), $MachinePrecision] + N[(N[(2.0 * 0.0001789971), $MachinePrecision] * N[(t$95$3 * N[(x * x), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] * x), $MachinePrecision]]]]]
\begin{array}{l}
t_0 := \left(x \cdot x\right) \cdot \left(x \cdot x\right)\\
t_1 := t\_0 \cdot \left(x \cdot x\right)\\
t_2 := t\_1 \cdot \left(x \cdot x\right)\\
t_3 := t\_2 \cdot \left(x \cdot x\right)\\
\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot t\_0\right) + 0.0072644182 \cdot t\_1\right) + 0.0005064034 \cdot t\_2\right) + 0.0001789971 \cdot t\_3}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot t\_0\right) + 0.0694555761 \cdot t\_1\right) + 0.0140005442 \cdot t\_2\right) + 0.0008327945 \cdot t\_3\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(t\_3 \cdot \left(x \cdot x\right)\right)} \cdot x
\end{array}

Local Percentage Accuracy vs ?

The average percentage accuracy by input value. Horizontal axis shows value of an input variable; the variable is choosen in the title. Vertical axis is accuracy; higher is better. Red represent the original program, while blue represents Herbie's suggestion. These can be toggled with buttons below the plot. The line is an average while dots represent individual samples.

Accuracy vs Speed?

Herbie found 8 alternatives:

AlternativeAccuracySpeedup
The accuracy (vertical axis) and speed (horizontal axis) of each alternatives. Up and to the right is better. The red square shows the initial program, and each blue circle shows an alternative.The line shows the best available speed-accuracy tradeoffs.

Initial Program: 53.7% accurate, 1.0× speedup?

\[\begin{array}{l} t_0 := \left(x \cdot x\right) \cdot \left(x \cdot x\right)\\ t_1 := t\_0 \cdot \left(x \cdot x\right)\\ t_2 := t\_1 \cdot \left(x \cdot x\right)\\ t_3 := t\_2 \cdot \left(x \cdot x\right)\\ \frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot t\_0\right) + 0.0072644182 \cdot t\_1\right) + 0.0005064034 \cdot t\_2\right) + 0.0001789971 \cdot t\_3}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot t\_0\right) + 0.0694555761 \cdot t\_1\right) + 0.0140005442 \cdot t\_2\right) + 0.0008327945 \cdot t\_3\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(t\_3 \cdot \left(x \cdot x\right)\right)} \cdot x \end{array} \]
(FPCore (x)
 :precision binary64
 (let* ((t_0 (* (* x x) (* x x)))
        (t_1 (* t_0 (* x x)))
        (t_2 (* t_1 (* x x)))
        (t_3 (* t_2 (* x x))))
   (*
    (/
     (+
      (+
       (+
        (+ (+ 1.0 (* 0.1049934947 (* x x))) (* 0.0424060604 t_0))
        (* 0.0072644182 t_1))
       (* 0.0005064034 t_2))
      (* 0.0001789971 t_3))
     (+
      (+
       (+
        (+
         (+ (+ 1.0 (* 0.7715471019 (* x x))) (* 0.2909738639 t_0))
         (* 0.0694555761 t_1))
        (* 0.0140005442 t_2))
       (* 0.0008327945 t_3))
      (* (* 2.0 0.0001789971) (* t_3 (* x x)))))
    x)))
double code(double x) {
	double t_0 = (x * x) * (x * x);
	double t_1 = t_0 * (x * x);
	double t_2 = t_1 * (x * x);
	double t_3 = t_2 * (x * x);
	return ((((((1.0 + (0.1049934947 * (x * x))) + (0.0424060604 * t_0)) + (0.0072644182 * t_1)) + (0.0005064034 * t_2)) + (0.0001789971 * t_3)) / ((((((1.0 + (0.7715471019 * (x * x))) + (0.2909738639 * t_0)) + (0.0694555761 * t_1)) + (0.0140005442 * t_2)) + (0.0008327945 * t_3)) + ((2.0 * 0.0001789971) * (t_3 * (x * x))))) * x;
}
module fmin_fmax_functions
    implicit none
    private
    public fmax
    public fmin

    interface fmax
        module procedure fmax88
        module procedure fmax44
        module procedure fmax84
        module procedure fmax48
    end interface
    interface fmin
        module procedure fmin88
        module procedure fmin44
        module procedure fmin84
        module procedure fmin48
    end interface
contains
    real(8) function fmax88(x, y) result (res)
        real(8), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(x, max(x, y), y /= y), x /= x)
    end function
    real(4) function fmax44(x, y) result (res)
        real(4), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(y, merge(x, max(x, y), y /= y), x /= x)
    end function
    real(8) function fmax84(x, y) result(res)
        real(8), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(dble(y), merge(x, max(x, dble(y)), y /= y), x /= x)
    end function
    real(8) function fmax48(x, y) result(res)
        real(4), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(dble(x), max(dble(x), y), y /= y), x /= x)
    end function
    real(8) function fmin88(x, y) result (res)
        real(8), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(x, min(x, y), y /= y), x /= x)
    end function
    real(4) function fmin44(x, y) result (res)
        real(4), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(y, merge(x, min(x, y), y /= y), x /= x)
    end function
    real(8) function fmin84(x, y) result(res)
        real(8), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(dble(y), merge(x, min(x, dble(y)), y /= y), x /= x)
    end function
    real(8) function fmin48(x, y) result(res)
        real(4), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(dble(x), min(dble(x), y), y /= y), x /= x)
    end function
end module

real(8) function code(x)
use fmin_fmax_functions
    real(8), intent (in) :: x
    real(8) :: t_0
    real(8) :: t_1
    real(8) :: t_2
    real(8) :: t_3
    t_0 = (x * x) * (x * x)
    t_1 = t_0 * (x * x)
    t_2 = t_1 * (x * x)
    t_3 = t_2 * (x * x)
    code = ((((((1.0d0 + (0.1049934947d0 * (x * x))) + (0.0424060604d0 * t_0)) + (0.0072644182d0 * t_1)) + (0.0005064034d0 * t_2)) + (0.0001789971d0 * t_3)) / ((((((1.0d0 + (0.7715471019d0 * (x * x))) + (0.2909738639d0 * t_0)) + (0.0694555761d0 * t_1)) + (0.0140005442d0 * t_2)) + (0.0008327945d0 * t_3)) + ((2.0d0 * 0.0001789971d0) * (t_3 * (x * x))))) * x
end function
public static double code(double x) {
	double t_0 = (x * x) * (x * x);
	double t_1 = t_0 * (x * x);
	double t_2 = t_1 * (x * x);
	double t_3 = t_2 * (x * x);
	return ((((((1.0 + (0.1049934947 * (x * x))) + (0.0424060604 * t_0)) + (0.0072644182 * t_1)) + (0.0005064034 * t_2)) + (0.0001789971 * t_3)) / ((((((1.0 + (0.7715471019 * (x * x))) + (0.2909738639 * t_0)) + (0.0694555761 * t_1)) + (0.0140005442 * t_2)) + (0.0008327945 * t_3)) + ((2.0 * 0.0001789971) * (t_3 * (x * x))))) * x;
}
def code(x):
	t_0 = (x * x) * (x * x)
	t_1 = t_0 * (x * x)
	t_2 = t_1 * (x * x)
	t_3 = t_2 * (x * x)
	return ((((((1.0 + (0.1049934947 * (x * x))) + (0.0424060604 * t_0)) + (0.0072644182 * t_1)) + (0.0005064034 * t_2)) + (0.0001789971 * t_3)) / ((((((1.0 + (0.7715471019 * (x * x))) + (0.2909738639 * t_0)) + (0.0694555761 * t_1)) + (0.0140005442 * t_2)) + (0.0008327945 * t_3)) + ((2.0 * 0.0001789971) * (t_3 * (x * x))))) * x
function code(x)
	t_0 = Float64(Float64(x * x) * Float64(x * x))
	t_1 = Float64(t_0 * Float64(x * x))
	t_2 = Float64(t_1 * Float64(x * x))
	t_3 = Float64(t_2 * Float64(x * x))
	return Float64(Float64(Float64(Float64(Float64(Float64(Float64(1.0 + Float64(0.1049934947 * Float64(x * x))) + Float64(0.0424060604 * t_0)) + Float64(0.0072644182 * t_1)) + Float64(0.0005064034 * t_2)) + Float64(0.0001789971 * t_3)) / Float64(Float64(Float64(Float64(Float64(Float64(1.0 + Float64(0.7715471019 * Float64(x * x))) + Float64(0.2909738639 * t_0)) + Float64(0.0694555761 * t_1)) + Float64(0.0140005442 * t_2)) + Float64(0.0008327945 * t_3)) + Float64(Float64(2.0 * 0.0001789971) * Float64(t_3 * Float64(x * x))))) * x)
end
function tmp = code(x)
	t_0 = (x * x) * (x * x);
	t_1 = t_0 * (x * x);
	t_2 = t_1 * (x * x);
	t_3 = t_2 * (x * x);
	tmp = ((((((1.0 + (0.1049934947 * (x * x))) + (0.0424060604 * t_0)) + (0.0072644182 * t_1)) + (0.0005064034 * t_2)) + (0.0001789971 * t_3)) / ((((((1.0 + (0.7715471019 * (x * x))) + (0.2909738639 * t_0)) + (0.0694555761 * t_1)) + (0.0140005442 * t_2)) + (0.0008327945 * t_3)) + ((2.0 * 0.0001789971) * (t_3 * (x * x))))) * x;
end
code[x_] := Block[{t$95$0 = N[(N[(x * x), $MachinePrecision] * N[(x * x), $MachinePrecision]), $MachinePrecision]}, Block[{t$95$1 = N[(t$95$0 * N[(x * x), $MachinePrecision]), $MachinePrecision]}, Block[{t$95$2 = N[(t$95$1 * N[(x * x), $MachinePrecision]), $MachinePrecision]}, Block[{t$95$3 = N[(t$95$2 * N[(x * x), $MachinePrecision]), $MachinePrecision]}, N[(N[(N[(N[(N[(N[(N[(1.0 + N[(0.1049934947 * N[(x * x), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] + N[(0.0424060604 * t$95$0), $MachinePrecision]), $MachinePrecision] + N[(0.0072644182 * t$95$1), $MachinePrecision]), $MachinePrecision] + N[(0.0005064034 * t$95$2), $MachinePrecision]), $MachinePrecision] + N[(0.0001789971 * t$95$3), $MachinePrecision]), $MachinePrecision] / N[(N[(N[(N[(N[(N[(1.0 + N[(0.7715471019 * N[(x * x), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] + N[(0.2909738639 * t$95$0), $MachinePrecision]), $MachinePrecision] + N[(0.0694555761 * t$95$1), $MachinePrecision]), $MachinePrecision] + N[(0.0140005442 * t$95$2), $MachinePrecision]), $MachinePrecision] + N[(0.0008327945 * t$95$3), $MachinePrecision]), $MachinePrecision] + N[(N[(2.0 * 0.0001789971), $MachinePrecision] * N[(t$95$3 * N[(x * x), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] * x), $MachinePrecision]]]]]
\begin{array}{l}
t_0 := \left(x \cdot x\right) \cdot \left(x \cdot x\right)\\
t_1 := t\_0 \cdot \left(x \cdot x\right)\\
t_2 := t\_1 \cdot \left(x \cdot x\right)\\
t_3 := t\_2 \cdot \left(x \cdot x\right)\\
\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot t\_0\right) + 0.0072644182 \cdot t\_1\right) + 0.0005064034 \cdot t\_2\right) + 0.0001789971 \cdot t\_3}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot t\_0\right) + 0.0694555761 \cdot t\_1\right) + 0.0140005442 \cdot t\_2\right) + 0.0008327945 \cdot t\_3\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(t\_3 \cdot \left(x \cdot x\right)\right)} \cdot x
\end{array}

Alternative 1: 99.7% accurate, 1.2× speedup?

\[\begin{array}{l} t_0 := \left|x\right| \cdot \left|x\right|\\ t_1 := t\_0 \cdot \left|x\right|\\ t_2 := {t\_0}^{5}\\ \mathsf{copysign}\left(1, x\right) \cdot \begin{array}{l} \mathbf{if}\;\left|x\right| \leq 1.35:\\ \;\;\;\;\mathsf{fma}\left(t\_2, 0.0001789971, \mathsf{fma}\left(t\_0, \mathsf{fma}\left(\left|x\right|, \mathsf{fma}\left(0.0072644182 \cdot \left|x\right|, t\_0, \left(\left(t\_1 \cdot 0.0005064034\right) \cdot \left|x\right|\right) \cdot \left|x\right|\right), \mathsf{fma}\left(t\_0, 0.0424060604, 0.1049934947\right)\right), 1\right)\right) \cdot \frac{\left|x\right|}{\mathsf{fma}\left({t\_0}^{6}, 0.0003579942, \mathsf{fma}\left(t\_2, 0.0008327945, \mathsf{fma}\left(\left|x\right|, \mathsf{fma}\left(t\_0, \mathsf{fma}\left(0.0694555761, \left|x\right|, t\_1 \cdot 0.0140005442\right), \left|x\right| \cdot 0.2909738639\right), 0.7715471019\right) \cdot t\_0\right) - -1\right)}\\ \mathbf{else}:\\ \;\;\;\;\frac{\frac{0.2514179000665374}{t\_0} - -0.5}{\left|x\right|}\\ \end{array} \end{array} \]
(FPCore (x)
 :precision binary64
 (let* ((t_0 (* (fabs x) (fabs x))) (t_1 (* t_0 (fabs x))) (t_2 (pow t_0 5.0)))
   (*
    (copysign 1.0 x)
    (if (<= (fabs x) 1.35)
      (*
       (fma
        t_2
        0.0001789971
        (fma
         t_0
         (fma
          (fabs x)
          (fma
           (* 0.0072644182 (fabs x))
           t_0
           (* (* (* t_1 0.0005064034) (fabs x)) (fabs x)))
          (fma t_0 0.0424060604 0.1049934947))
         1.0))
       (/
        (fabs x)
        (fma
         (pow t_0 6.0)
         0.0003579942
         (-
          (fma
           t_2
           0.0008327945
           (*
            (fma
             (fabs x)
             (fma
              t_0
              (fma 0.0694555761 (fabs x) (* t_1 0.0140005442))
              (* (fabs x) 0.2909738639))
             0.7715471019)
            t_0))
          -1.0))))
      (/ (- (/ 0.2514179000665374 t_0) -0.5) (fabs x))))))
double code(double x) {
	double t_0 = fabs(x) * fabs(x);
	double t_1 = t_0 * fabs(x);
	double t_2 = pow(t_0, 5.0);
	double tmp;
	if (fabs(x) <= 1.35) {
		tmp = fma(t_2, 0.0001789971, fma(t_0, fma(fabs(x), fma((0.0072644182 * fabs(x)), t_0, (((t_1 * 0.0005064034) * fabs(x)) * fabs(x))), fma(t_0, 0.0424060604, 0.1049934947)), 1.0)) * (fabs(x) / fma(pow(t_0, 6.0), 0.0003579942, (fma(t_2, 0.0008327945, (fma(fabs(x), fma(t_0, fma(0.0694555761, fabs(x), (t_1 * 0.0140005442)), (fabs(x) * 0.2909738639)), 0.7715471019) * t_0)) - -1.0)));
	} else {
		tmp = ((0.2514179000665374 / t_0) - -0.5) / fabs(x);
	}
	return copysign(1.0, x) * tmp;
}
function code(x)
	t_0 = Float64(abs(x) * abs(x))
	t_1 = Float64(t_0 * abs(x))
	t_2 = t_0 ^ 5.0
	tmp = 0.0
	if (abs(x) <= 1.35)
		tmp = Float64(fma(t_2, 0.0001789971, fma(t_0, fma(abs(x), fma(Float64(0.0072644182 * abs(x)), t_0, Float64(Float64(Float64(t_1 * 0.0005064034) * abs(x)) * abs(x))), fma(t_0, 0.0424060604, 0.1049934947)), 1.0)) * Float64(abs(x) / fma((t_0 ^ 6.0), 0.0003579942, Float64(fma(t_2, 0.0008327945, Float64(fma(abs(x), fma(t_0, fma(0.0694555761, abs(x), Float64(t_1 * 0.0140005442)), Float64(abs(x) * 0.2909738639)), 0.7715471019) * t_0)) - -1.0))));
	else
		tmp = Float64(Float64(Float64(0.2514179000665374 / t_0) - -0.5) / abs(x));
	end
	return Float64(copysign(1.0, x) * tmp)
end
code[x_] := Block[{t$95$0 = N[(N[Abs[x], $MachinePrecision] * N[Abs[x], $MachinePrecision]), $MachinePrecision]}, Block[{t$95$1 = N[(t$95$0 * N[Abs[x], $MachinePrecision]), $MachinePrecision]}, Block[{t$95$2 = N[Power[t$95$0, 5.0], $MachinePrecision]}, N[(N[With[{TMP1 = Abs[1.0], TMP2 = Sign[x]}, TMP1 * If[TMP2 == 0, 1, TMP2]], $MachinePrecision] * If[LessEqual[N[Abs[x], $MachinePrecision], 1.35], N[(N[(t$95$2 * 0.0001789971 + N[(t$95$0 * N[(N[Abs[x], $MachinePrecision] * N[(N[(0.0072644182 * N[Abs[x], $MachinePrecision]), $MachinePrecision] * t$95$0 + N[(N[(N[(t$95$1 * 0.0005064034), $MachinePrecision] * N[Abs[x], $MachinePrecision]), $MachinePrecision] * N[Abs[x], $MachinePrecision]), $MachinePrecision]), $MachinePrecision] + N[(t$95$0 * 0.0424060604 + 0.1049934947), $MachinePrecision]), $MachinePrecision] + 1.0), $MachinePrecision]), $MachinePrecision] * N[(N[Abs[x], $MachinePrecision] / N[(N[Power[t$95$0, 6.0], $MachinePrecision] * 0.0003579942 + N[(N[(t$95$2 * 0.0008327945 + N[(N[(N[Abs[x], $MachinePrecision] * N[(t$95$0 * N[(0.0694555761 * N[Abs[x], $MachinePrecision] + N[(t$95$1 * 0.0140005442), $MachinePrecision]), $MachinePrecision] + N[(N[Abs[x], $MachinePrecision] * 0.2909738639), $MachinePrecision]), $MachinePrecision] + 0.7715471019), $MachinePrecision] * t$95$0), $MachinePrecision]), $MachinePrecision] - -1.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision], N[(N[(N[(0.2514179000665374 / t$95$0), $MachinePrecision] - -0.5), $MachinePrecision] / N[Abs[x], $MachinePrecision]), $MachinePrecision]]), $MachinePrecision]]]]
\begin{array}{l}
t_0 := \left|x\right| \cdot \left|x\right|\\
t_1 := t\_0 \cdot \left|x\right|\\
t_2 := {t\_0}^{5}\\
\mathsf{copysign}\left(1, x\right) \cdot \begin{array}{l}
\mathbf{if}\;\left|x\right| \leq 1.35:\\
\;\;\;\;\mathsf{fma}\left(t\_2, 0.0001789971, \mathsf{fma}\left(t\_0, \mathsf{fma}\left(\left|x\right|, \mathsf{fma}\left(0.0072644182 \cdot \left|x\right|, t\_0, \left(\left(t\_1 \cdot 0.0005064034\right) \cdot \left|x\right|\right) \cdot \left|x\right|\right), \mathsf{fma}\left(t\_0, 0.0424060604, 0.1049934947\right)\right), 1\right)\right) \cdot \frac{\left|x\right|}{\mathsf{fma}\left({t\_0}^{6}, 0.0003579942, \mathsf{fma}\left(t\_2, 0.0008327945, \mathsf{fma}\left(\left|x\right|, \mathsf{fma}\left(t\_0, \mathsf{fma}\left(0.0694555761, \left|x\right|, t\_1 \cdot 0.0140005442\right), \left|x\right| \cdot 0.2909738639\right), 0.7715471019\right) \cdot t\_0\right) - -1\right)}\\

\mathbf{else}:\\
\;\;\;\;\frac{\frac{0.2514179000665374}{t\_0} - -0.5}{\left|x\right|}\\


\end{array}
\end{array}
Derivation
  1. Split input into 2 regimes
  2. if x < 1.3500000000000001

    1. Initial program 53.7%

      \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
    2. Applied rewrites53.7%

      \[\leadsto \color{blue}{\mathsf{fma}\left({\left(x \cdot x\right)}^{5}, 0.0001789971, \mathsf{fma}\left(x \cdot x, \mathsf{fma}\left(0.0072644182 \cdot \left(\left(x \cdot x\right) \cdot x\right), x, \left(0.0005064034 \cdot \left(\left(x \cdot x\right) \cdot x\right)\right) \cdot \left(\left(x \cdot x\right) \cdot x\right)\right), \mathsf{fma}\left(x \cdot x, 0.1049934947 + 0.0424060604 \cdot \left(x \cdot x\right), 1\right)\right)\right) \cdot \frac{x}{\mathsf{fma}\left({\left(x \cdot x\right)}^{6}, 0.0003579942, \mathsf{fma}\left(0.0008327945, {\left(x \cdot x\right)}^{5}, \mathsf{fma}\left(x \cdot x, \mathsf{fma}\left(0.0694555761 \cdot \left(\left(x \cdot x\right) \cdot x\right), x, \left(0.0140005442 \cdot \left(\left(x \cdot x\right) \cdot x\right)\right) \cdot \left(\left(x \cdot x\right) \cdot x\right)\right), \mathsf{fma}\left(x \cdot x, 0.7715471019 + 0.2909738639 \cdot \left(x \cdot x\right), 1\right)\right)\right)\right)}} \]
    3. Applied rewrites53.7%

      \[\leadsto \color{blue}{\mathsf{fma}\left({\left(x \cdot x\right)}^{5}, 0.0001789971, \mathsf{fma}\left(x \cdot x, \mathsf{fma}\left(x, \mathsf{fma}\left(0.0072644182 \cdot x, x \cdot x, \left(\left(\left(\left(x \cdot x\right) \cdot x\right) \cdot 0.0005064034\right) \cdot x\right) \cdot x\right), \mathsf{fma}\left(x \cdot x, 0.0424060604, 0.1049934947\right)\right), 1\right)\right) \cdot \frac{x}{\mathsf{fma}\left({\left(x \cdot x\right)}^{6}, 0.0003579942, \mathsf{fma}\left({\left(x \cdot x\right)}^{5}, 0.0008327945, \mathsf{fma}\left(x \cdot x, \mathsf{fma}\left(x, \mathsf{fma}\left(0.0694555761 \cdot x, x \cdot x, \left(\left(\left(\left(x \cdot x\right) \cdot x\right) \cdot 0.0140005442\right) \cdot x\right) \cdot x\right), \mathsf{fma}\left(x \cdot x, 0.2909738639, 0.7715471019\right)\right), 1\right)\right)\right)}} \]
    4. Applied rewrites53.7%

      \[\leadsto \mathsf{fma}\left({\left(x \cdot x\right)}^{5}, 0.0001789971, \mathsf{fma}\left(x \cdot x, \mathsf{fma}\left(x, \mathsf{fma}\left(0.0072644182 \cdot x, x \cdot x, \left(\left(\left(\left(x \cdot x\right) \cdot x\right) \cdot 0.0005064034\right) \cdot x\right) \cdot x\right), \mathsf{fma}\left(x \cdot x, 0.0424060604, 0.1049934947\right)\right), 1\right)\right) \cdot \frac{x}{\mathsf{fma}\left({\left(x \cdot x\right)}^{6}, 0.0003579942, \color{blue}{\mathsf{fma}\left({\left(x \cdot x\right)}^{5}, 0.0008327945, \mathsf{fma}\left(x, \mathsf{fma}\left(x \cdot x, \mathsf{fma}\left(0.0694555761, x, \left(\left(x \cdot x\right) \cdot x\right) \cdot 0.0140005442\right), x \cdot 0.2909738639\right), 0.7715471019\right) \cdot \left(x \cdot x\right)\right) - -1}\right)} \]

    if 1.3500000000000001 < x

    1. Initial program 53.7%

      \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
    2. Taylor expanded in x around inf

      \[\leadsto \color{blue}{\frac{\frac{1}{2} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}}{x}} \]
    3. Step-by-step derivation
      1. lower-/.f64N/A

        \[\leadsto \frac{\frac{1}{2} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}}{\color{blue}{x}} \]
    4. Applied rewrites52.0%

      \[\leadsto \color{blue}{\frac{0.5 + 0.2514179000665374 \cdot \frac{1}{{x}^{2}}}{x}} \]
    5. Step-by-step derivation
      1. lift-+.f64N/A

        \[\leadsto \frac{\frac{1}{2} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}}{x} \]
      2. +-commutativeN/A

        \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} + \frac{1}{2}}{x} \]
      3. add-flipN/A

        \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      4. lower--.f64N/A

        \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      5. lift-*.f64N/A

        \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      6. lift-/.f64N/A

        \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      7. lift-pow.f64N/A

        \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      8. pow2N/A

        \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{x \cdot x} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      9. mult-flip-revN/A

        \[\leadsto \frac{\frac{\frac{600041}{2386628}}{x \cdot x} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      10. lower-/.f64N/A

        \[\leadsto \frac{\frac{\frac{600041}{2386628}}{x \cdot x} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      11. lower-*.f64N/A

        \[\leadsto \frac{\frac{\frac{600041}{2386628}}{x \cdot x} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      12. metadata-eval52.0%

        \[\leadsto \frac{\frac{0.2514179000665374}{x \cdot x} - -0.5}{x} \]
    6. Applied rewrites52.0%

      \[\leadsto \frac{\frac{0.2514179000665374}{x \cdot x} - -0.5}{\color{blue}{x}} \]
  3. Recombined 2 regimes into one program.
  4. Add Preprocessing

Alternative 2: 99.7% accurate, 1.3× speedup?

\[\begin{array}{l} t_0 := \left|x\right| \cdot \left|x\right|\\ t_1 := t\_0 \cdot \left|x\right|\\ t_2 := {t\_0}^{5}\\ \mathsf{copysign}\left(1, x\right) \cdot \begin{array}{l} \mathbf{if}\;\left|x\right| \leq 1.35:\\ \;\;\;\;\mathsf{fma}\left(t\_2, 0.0001789971, \mathsf{fma}\left(\mathsf{fma}\left(\left|x\right|, \mathsf{fma}\left(t\_0, \mathsf{fma}\left(0.0072644182, \left|x\right|, t\_1 \cdot 0.0005064034\right), \left|x\right| \cdot 0.0424060604\right), 0.1049934947\right), t\_0, 1\right)\right) \cdot \frac{\left|x\right|}{\mathsf{fma}\left({t\_0}^{6}, 0.0003579942, \mathsf{fma}\left(t\_2, 0.0008327945, \mathsf{fma}\left(\mathsf{fma}\left(\left|x\right|, \mathsf{fma}\left(t\_0, \mathsf{fma}\left(0.0694555761, \left|x\right|, t\_1 \cdot 0.0140005442\right), \left|x\right| \cdot 0.2909738639\right), 0.7715471019\right), t\_0, 1\right)\right)\right)}\\ \mathbf{else}:\\ \;\;\;\;\frac{\frac{0.2514179000665374}{t\_0} - -0.5}{\left|x\right|}\\ \end{array} \end{array} \]
(FPCore (x)
 :precision binary64
 (let* ((t_0 (* (fabs x) (fabs x))) (t_1 (* t_0 (fabs x))) (t_2 (pow t_0 5.0)))
   (*
    (copysign 1.0 x)
    (if (<= (fabs x) 1.35)
      (*
       (fma
        t_2
        0.0001789971
        (fma
         (fma
          (fabs x)
          (fma
           t_0
           (fma 0.0072644182 (fabs x) (* t_1 0.0005064034))
           (* (fabs x) 0.0424060604))
          0.1049934947)
         t_0
         1.0))
       (/
        (fabs x)
        (fma
         (pow t_0 6.0)
         0.0003579942
         (fma
          t_2
          0.0008327945
          (fma
           (fma
            (fabs x)
            (fma
             t_0
             (fma 0.0694555761 (fabs x) (* t_1 0.0140005442))
             (* (fabs x) 0.2909738639))
            0.7715471019)
           t_0
           1.0)))))
      (/ (- (/ 0.2514179000665374 t_0) -0.5) (fabs x))))))
double code(double x) {
	double t_0 = fabs(x) * fabs(x);
	double t_1 = t_0 * fabs(x);
	double t_2 = pow(t_0, 5.0);
	double tmp;
	if (fabs(x) <= 1.35) {
		tmp = fma(t_2, 0.0001789971, fma(fma(fabs(x), fma(t_0, fma(0.0072644182, fabs(x), (t_1 * 0.0005064034)), (fabs(x) * 0.0424060604)), 0.1049934947), t_0, 1.0)) * (fabs(x) / fma(pow(t_0, 6.0), 0.0003579942, fma(t_2, 0.0008327945, fma(fma(fabs(x), fma(t_0, fma(0.0694555761, fabs(x), (t_1 * 0.0140005442)), (fabs(x) * 0.2909738639)), 0.7715471019), t_0, 1.0))));
	} else {
		tmp = ((0.2514179000665374 / t_0) - -0.5) / fabs(x);
	}
	return copysign(1.0, x) * tmp;
}
function code(x)
	t_0 = Float64(abs(x) * abs(x))
	t_1 = Float64(t_0 * abs(x))
	t_2 = t_0 ^ 5.0
	tmp = 0.0
	if (abs(x) <= 1.35)
		tmp = Float64(fma(t_2, 0.0001789971, fma(fma(abs(x), fma(t_0, fma(0.0072644182, abs(x), Float64(t_1 * 0.0005064034)), Float64(abs(x) * 0.0424060604)), 0.1049934947), t_0, 1.0)) * Float64(abs(x) / fma((t_0 ^ 6.0), 0.0003579942, fma(t_2, 0.0008327945, fma(fma(abs(x), fma(t_0, fma(0.0694555761, abs(x), Float64(t_1 * 0.0140005442)), Float64(abs(x) * 0.2909738639)), 0.7715471019), t_0, 1.0)))));
	else
		tmp = Float64(Float64(Float64(0.2514179000665374 / t_0) - -0.5) / abs(x));
	end
	return Float64(copysign(1.0, x) * tmp)
end
code[x_] := Block[{t$95$0 = N[(N[Abs[x], $MachinePrecision] * N[Abs[x], $MachinePrecision]), $MachinePrecision]}, Block[{t$95$1 = N[(t$95$0 * N[Abs[x], $MachinePrecision]), $MachinePrecision]}, Block[{t$95$2 = N[Power[t$95$0, 5.0], $MachinePrecision]}, N[(N[With[{TMP1 = Abs[1.0], TMP2 = Sign[x]}, TMP1 * If[TMP2 == 0, 1, TMP2]], $MachinePrecision] * If[LessEqual[N[Abs[x], $MachinePrecision], 1.35], N[(N[(t$95$2 * 0.0001789971 + N[(N[(N[Abs[x], $MachinePrecision] * N[(t$95$0 * N[(0.0072644182 * N[Abs[x], $MachinePrecision] + N[(t$95$1 * 0.0005064034), $MachinePrecision]), $MachinePrecision] + N[(N[Abs[x], $MachinePrecision] * 0.0424060604), $MachinePrecision]), $MachinePrecision] + 0.1049934947), $MachinePrecision] * t$95$0 + 1.0), $MachinePrecision]), $MachinePrecision] * N[(N[Abs[x], $MachinePrecision] / N[(N[Power[t$95$0, 6.0], $MachinePrecision] * 0.0003579942 + N[(t$95$2 * 0.0008327945 + N[(N[(N[Abs[x], $MachinePrecision] * N[(t$95$0 * N[(0.0694555761 * N[Abs[x], $MachinePrecision] + N[(t$95$1 * 0.0140005442), $MachinePrecision]), $MachinePrecision] + N[(N[Abs[x], $MachinePrecision] * 0.2909738639), $MachinePrecision]), $MachinePrecision] + 0.7715471019), $MachinePrecision] * t$95$0 + 1.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision], N[(N[(N[(0.2514179000665374 / t$95$0), $MachinePrecision] - -0.5), $MachinePrecision] / N[Abs[x], $MachinePrecision]), $MachinePrecision]]), $MachinePrecision]]]]
\begin{array}{l}
t_0 := \left|x\right| \cdot \left|x\right|\\
t_1 := t\_0 \cdot \left|x\right|\\
t_2 := {t\_0}^{5}\\
\mathsf{copysign}\left(1, x\right) \cdot \begin{array}{l}
\mathbf{if}\;\left|x\right| \leq 1.35:\\
\;\;\;\;\mathsf{fma}\left(t\_2, 0.0001789971, \mathsf{fma}\left(\mathsf{fma}\left(\left|x\right|, \mathsf{fma}\left(t\_0, \mathsf{fma}\left(0.0072644182, \left|x\right|, t\_1 \cdot 0.0005064034\right), \left|x\right| \cdot 0.0424060604\right), 0.1049934947\right), t\_0, 1\right)\right) \cdot \frac{\left|x\right|}{\mathsf{fma}\left({t\_0}^{6}, 0.0003579942, \mathsf{fma}\left(t\_2, 0.0008327945, \mathsf{fma}\left(\mathsf{fma}\left(\left|x\right|, \mathsf{fma}\left(t\_0, \mathsf{fma}\left(0.0694555761, \left|x\right|, t\_1 \cdot 0.0140005442\right), \left|x\right| \cdot 0.2909738639\right), 0.7715471019\right), t\_0, 1\right)\right)\right)}\\

\mathbf{else}:\\
\;\;\;\;\frac{\frac{0.2514179000665374}{t\_0} - -0.5}{\left|x\right|}\\


\end{array}
\end{array}
Derivation
  1. Split input into 2 regimes
  2. if x < 1.3500000000000001

    1. Initial program 53.7%

      \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
    2. Applied rewrites53.7%

      \[\leadsto \color{blue}{\mathsf{fma}\left({\left(x \cdot x\right)}^{5}, 0.0001789971, \mathsf{fma}\left(x \cdot x, \mathsf{fma}\left(0.0072644182 \cdot \left(\left(x \cdot x\right) \cdot x\right), x, \left(0.0005064034 \cdot \left(\left(x \cdot x\right) \cdot x\right)\right) \cdot \left(\left(x \cdot x\right) \cdot x\right)\right), \mathsf{fma}\left(x \cdot x, 0.1049934947 + 0.0424060604 \cdot \left(x \cdot x\right), 1\right)\right)\right) \cdot \frac{x}{\mathsf{fma}\left({\left(x \cdot x\right)}^{6}, 0.0003579942, \mathsf{fma}\left(0.0008327945, {\left(x \cdot x\right)}^{5}, \mathsf{fma}\left(x \cdot x, \mathsf{fma}\left(0.0694555761 \cdot \left(\left(x \cdot x\right) \cdot x\right), x, \left(0.0140005442 \cdot \left(\left(x \cdot x\right) \cdot x\right)\right) \cdot \left(\left(x \cdot x\right) \cdot x\right)\right), \mathsf{fma}\left(x \cdot x, 0.7715471019 + 0.2909738639 \cdot \left(x \cdot x\right), 1\right)\right)\right)\right)}} \]
    3. Applied rewrites53.7%

      \[\leadsto \color{blue}{\mathsf{fma}\left({\left(x \cdot x\right)}^{5}, 0.0001789971, \mathsf{fma}\left(x \cdot x, \mathsf{fma}\left(x, \mathsf{fma}\left(0.0072644182 \cdot x, x \cdot x, \left(\left(\left(\left(x \cdot x\right) \cdot x\right) \cdot 0.0005064034\right) \cdot x\right) \cdot x\right), \mathsf{fma}\left(x \cdot x, 0.0424060604, 0.1049934947\right)\right), 1\right)\right) \cdot \frac{x}{\mathsf{fma}\left({\left(x \cdot x\right)}^{6}, 0.0003579942, \mathsf{fma}\left({\left(x \cdot x\right)}^{5}, 0.0008327945, \mathsf{fma}\left(x \cdot x, \mathsf{fma}\left(x, \mathsf{fma}\left(0.0694555761 \cdot x, x \cdot x, \left(\left(\left(\left(x \cdot x\right) \cdot x\right) \cdot 0.0140005442\right) \cdot x\right) \cdot x\right), \mathsf{fma}\left(x \cdot x, 0.2909738639, 0.7715471019\right)\right), 1\right)\right)\right)}} \]
    4. Applied rewrites53.7%

      \[\leadsto \color{blue}{\mathsf{fma}\left({\left(x \cdot x\right)}^{5}, 0.0001789971, \mathsf{fma}\left(\mathsf{fma}\left(x, \mathsf{fma}\left(x \cdot x, \mathsf{fma}\left(0.0072644182, x, \left(\left(x \cdot x\right) \cdot x\right) \cdot 0.0005064034\right), x \cdot 0.0424060604\right), 0.1049934947\right), x \cdot x, 1\right)\right) \cdot \frac{x}{\mathsf{fma}\left({\left(x \cdot x\right)}^{6}, 0.0003579942, \mathsf{fma}\left({\left(x \cdot x\right)}^{5}, 0.0008327945, \mathsf{fma}\left(\mathsf{fma}\left(x, \mathsf{fma}\left(x \cdot x, \mathsf{fma}\left(0.0694555761, x, \left(\left(x \cdot x\right) \cdot x\right) \cdot 0.0140005442\right), x \cdot 0.2909738639\right), 0.7715471019\right), x \cdot x, 1\right)\right)\right)}} \]

    if 1.3500000000000001 < x

    1. Initial program 53.7%

      \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
    2. Taylor expanded in x around inf

      \[\leadsto \color{blue}{\frac{\frac{1}{2} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}}{x}} \]
    3. Step-by-step derivation
      1. lower-/.f64N/A

        \[\leadsto \frac{\frac{1}{2} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}}{\color{blue}{x}} \]
    4. Applied rewrites52.0%

      \[\leadsto \color{blue}{\frac{0.5 + 0.2514179000665374 \cdot \frac{1}{{x}^{2}}}{x}} \]
    5. Step-by-step derivation
      1. lift-+.f64N/A

        \[\leadsto \frac{\frac{1}{2} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}}{x} \]
      2. +-commutativeN/A

        \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} + \frac{1}{2}}{x} \]
      3. add-flipN/A

        \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      4. lower--.f64N/A

        \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      5. lift-*.f64N/A

        \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      6. lift-/.f64N/A

        \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      7. lift-pow.f64N/A

        \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      8. pow2N/A

        \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{x \cdot x} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      9. mult-flip-revN/A

        \[\leadsto \frac{\frac{\frac{600041}{2386628}}{x \cdot x} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      10. lower-/.f64N/A

        \[\leadsto \frac{\frac{\frac{600041}{2386628}}{x \cdot x} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      11. lower-*.f64N/A

        \[\leadsto \frac{\frac{\frac{600041}{2386628}}{x \cdot x} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      12. metadata-eval52.0%

        \[\leadsto \frac{\frac{0.2514179000665374}{x \cdot x} - -0.5}{x} \]
    6. Applied rewrites52.0%

      \[\leadsto \frac{\frac{0.2514179000665374}{x \cdot x} - -0.5}{\color{blue}{x}} \]
  3. Recombined 2 regimes into one program.
  4. Add Preprocessing

Alternative 3: 99.6% accurate, 5.4× speedup?

\[\begin{array}{l} t_0 := \left|x\right| \cdot \left|x\right|\\ \mathsf{copysign}\left(1, x\right) \cdot \begin{array}{l} \mathbf{if}\;\left|x\right| \leq 1.5:\\ \;\;\;\;\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(-0.0732490286039007, t\_0, 0.265709700396151\right), t\_0, -0.6665536072\right) \cdot \left|x\right|, \left|x\right|, 1\right) \cdot \left|x\right|\\ \mathbf{else}:\\ \;\;\;\;\frac{\frac{0.2514179000665374}{t\_0} - -0.5}{\left|x\right|}\\ \end{array} \end{array} \]
(FPCore (x)
 :precision binary64
 (let* ((t_0 (* (fabs x) (fabs x))))
   (*
    (copysign 1.0 x)
    (if (<= (fabs x) 1.5)
      (*
       (fma
        (*
         (fma
          (fma -0.0732490286039007 t_0 0.265709700396151)
          t_0
          -0.6665536072)
         (fabs x))
        (fabs x)
        1.0)
       (fabs x))
      (/ (- (/ 0.2514179000665374 t_0) -0.5) (fabs x))))))
double code(double x) {
	double t_0 = fabs(x) * fabs(x);
	double tmp;
	if (fabs(x) <= 1.5) {
		tmp = fma((fma(fma(-0.0732490286039007, t_0, 0.265709700396151), t_0, -0.6665536072) * fabs(x)), fabs(x), 1.0) * fabs(x);
	} else {
		tmp = ((0.2514179000665374 / t_0) - -0.5) / fabs(x);
	}
	return copysign(1.0, x) * tmp;
}
function code(x)
	t_0 = Float64(abs(x) * abs(x))
	tmp = 0.0
	if (abs(x) <= 1.5)
		tmp = Float64(fma(Float64(fma(fma(-0.0732490286039007, t_0, 0.265709700396151), t_0, -0.6665536072) * abs(x)), abs(x), 1.0) * abs(x));
	else
		tmp = Float64(Float64(Float64(0.2514179000665374 / t_0) - -0.5) / abs(x));
	end
	return Float64(copysign(1.0, x) * tmp)
end
code[x_] := Block[{t$95$0 = N[(N[Abs[x], $MachinePrecision] * N[Abs[x], $MachinePrecision]), $MachinePrecision]}, N[(N[With[{TMP1 = Abs[1.0], TMP2 = Sign[x]}, TMP1 * If[TMP2 == 0, 1, TMP2]], $MachinePrecision] * If[LessEqual[N[Abs[x], $MachinePrecision], 1.5], N[(N[(N[(N[(N[(-0.0732490286039007 * t$95$0 + 0.265709700396151), $MachinePrecision] * t$95$0 + -0.6665536072), $MachinePrecision] * N[Abs[x], $MachinePrecision]), $MachinePrecision] * N[Abs[x], $MachinePrecision] + 1.0), $MachinePrecision] * N[Abs[x], $MachinePrecision]), $MachinePrecision], N[(N[(N[(0.2514179000665374 / t$95$0), $MachinePrecision] - -0.5), $MachinePrecision] / N[Abs[x], $MachinePrecision]), $MachinePrecision]]), $MachinePrecision]]
\begin{array}{l}
t_0 := \left|x\right| \cdot \left|x\right|\\
\mathsf{copysign}\left(1, x\right) \cdot \begin{array}{l}
\mathbf{if}\;\left|x\right| \leq 1.5:\\
\;\;\;\;\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(-0.0732490286039007, t\_0, 0.265709700396151\right), t\_0, -0.6665536072\right) \cdot \left|x\right|, \left|x\right|, 1\right) \cdot \left|x\right|\\

\mathbf{else}:\\
\;\;\;\;\frac{\frac{0.2514179000665374}{t\_0} - -0.5}{\left|x\right|}\\


\end{array}
\end{array}
Derivation
  1. Split input into 2 regimes
  2. if x < 1.5

    1. Initial program 53.7%

      \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
    2. Taylor expanded in x around 0

      \[\leadsto \color{blue}{\left(1 + {x}^{2} \cdot \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} + \frac{-9156128575487588197208397249}{125000000000000000000000000000} \cdot {x}^{2}\right) - \frac{833192009}{1250000000}\right)\right)} \cdot x \]
    3. Step-by-step derivation
      1. lower-+.f64N/A

        \[\leadsto \left(1 + \color{blue}{{x}^{2} \cdot \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} + \frac{-9156128575487588197208397249}{125000000000000000000000000000} \cdot {x}^{2}\right) - \frac{833192009}{1250000000}\right)}\right) \cdot x \]
    4. Applied rewrites49.6%

      \[\leadsto \color{blue}{\left(1 + {x}^{2} \cdot \left({x}^{2} \cdot \left(0.265709700396151 + -0.0732490286039007 \cdot {x}^{2}\right) - 0.6665536072\right)\right)} \cdot x \]
    5. Step-by-step derivation
      1. lift-+.f64N/A

        \[\leadsto \left(1 + \color{blue}{{x}^{2} \cdot \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} + \frac{-9156128575487588197208397249}{125000000000000000000000000000} \cdot {x}^{2}\right) - \frac{833192009}{1250000000}\right)}\right) \cdot x \]
    6. Applied rewrites49.6%

      \[\leadsto \color{blue}{\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(-0.0732490286039007, x \cdot x, 0.265709700396151\right), x \cdot x, -0.6665536072\right) \cdot x, x, 1\right)} \cdot x \]

    if 1.5 < x

    1. Initial program 53.7%

      \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
    2. Taylor expanded in x around inf

      \[\leadsto \color{blue}{\frac{\frac{1}{2} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}}{x}} \]
    3. Step-by-step derivation
      1. lower-/.f64N/A

        \[\leadsto \frac{\frac{1}{2} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}}{\color{blue}{x}} \]
    4. Applied rewrites52.0%

      \[\leadsto \color{blue}{\frac{0.5 + 0.2514179000665374 \cdot \frac{1}{{x}^{2}}}{x}} \]
    5. Step-by-step derivation
      1. lift-+.f64N/A

        \[\leadsto \frac{\frac{1}{2} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}}{x} \]
      2. +-commutativeN/A

        \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} + \frac{1}{2}}{x} \]
      3. add-flipN/A

        \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      4. lower--.f64N/A

        \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      5. lift-*.f64N/A

        \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      6. lift-/.f64N/A

        \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      7. lift-pow.f64N/A

        \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      8. pow2N/A

        \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{x \cdot x} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      9. mult-flip-revN/A

        \[\leadsto \frac{\frac{\frac{600041}{2386628}}{x \cdot x} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      10. lower-/.f64N/A

        \[\leadsto \frac{\frac{\frac{600041}{2386628}}{x \cdot x} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      11. lower-*.f64N/A

        \[\leadsto \frac{\frac{\frac{600041}{2386628}}{x \cdot x} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      12. metadata-eval52.0%

        \[\leadsto \frac{\frac{0.2514179000665374}{x \cdot x} - -0.5}{x} \]
    6. Applied rewrites52.0%

      \[\leadsto \frac{\frac{0.2514179000665374}{x \cdot x} - -0.5}{\color{blue}{x}} \]
  3. Recombined 2 regimes into one program.
  4. Add Preprocessing

Alternative 4: 99.6% accurate, 6.9× speedup?

\[\begin{array}{l} t_0 := \left|x\right| \cdot \left|x\right|\\ \mathsf{copysign}\left(1, x\right) \cdot \begin{array}{l} \mathbf{if}\;\left|x\right| \leq 1.1:\\ \;\;\;\;\mathsf{fma}\left(\mathsf{fma}\left(0.265709700396151, t\_0, -0.6665536072\right), t\_0, 1\right) \cdot \left|x\right|\\ \mathbf{else}:\\ \;\;\;\;\frac{\frac{0.2514179000665374}{t\_0} - -0.5}{\left|x\right|}\\ \end{array} \end{array} \]
(FPCore (x)
 :precision binary64
 (let* ((t_0 (* (fabs x) (fabs x))))
   (*
    (copysign 1.0 x)
    (if (<= (fabs x) 1.1)
      (* (fma (fma 0.265709700396151 t_0 -0.6665536072) t_0 1.0) (fabs x))
      (/ (- (/ 0.2514179000665374 t_0) -0.5) (fabs x))))))
double code(double x) {
	double t_0 = fabs(x) * fabs(x);
	double tmp;
	if (fabs(x) <= 1.1) {
		tmp = fma(fma(0.265709700396151, t_0, -0.6665536072), t_0, 1.0) * fabs(x);
	} else {
		tmp = ((0.2514179000665374 / t_0) - -0.5) / fabs(x);
	}
	return copysign(1.0, x) * tmp;
}
function code(x)
	t_0 = Float64(abs(x) * abs(x))
	tmp = 0.0
	if (abs(x) <= 1.1)
		tmp = Float64(fma(fma(0.265709700396151, t_0, -0.6665536072), t_0, 1.0) * abs(x));
	else
		tmp = Float64(Float64(Float64(0.2514179000665374 / t_0) - -0.5) / abs(x));
	end
	return Float64(copysign(1.0, x) * tmp)
end
code[x_] := Block[{t$95$0 = N[(N[Abs[x], $MachinePrecision] * N[Abs[x], $MachinePrecision]), $MachinePrecision]}, N[(N[With[{TMP1 = Abs[1.0], TMP2 = Sign[x]}, TMP1 * If[TMP2 == 0, 1, TMP2]], $MachinePrecision] * If[LessEqual[N[Abs[x], $MachinePrecision], 1.1], N[(N[(N[(0.265709700396151 * t$95$0 + -0.6665536072), $MachinePrecision] * t$95$0 + 1.0), $MachinePrecision] * N[Abs[x], $MachinePrecision]), $MachinePrecision], N[(N[(N[(0.2514179000665374 / t$95$0), $MachinePrecision] - -0.5), $MachinePrecision] / N[Abs[x], $MachinePrecision]), $MachinePrecision]]), $MachinePrecision]]
\begin{array}{l}
t_0 := \left|x\right| \cdot \left|x\right|\\
\mathsf{copysign}\left(1, x\right) \cdot \begin{array}{l}
\mathbf{if}\;\left|x\right| \leq 1.1:\\
\;\;\;\;\mathsf{fma}\left(\mathsf{fma}\left(0.265709700396151, t\_0, -0.6665536072\right), t\_0, 1\right) \cdot \left|x\right|\\

\mathbf{else}:\\
\;\;\;\;\frac{\frac{0.2514179000665374}{t\_0} - -0.5}{\left|x\right|}\\


\end{array}
\end{array}
Derivation
  1. Split input into 2 regimes
  2. if x < 1.1000000000000001

    1. Initial program 53.7%

      \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
    2. Taylor expanded in x around 0

      \[\leadsto \color{blue}{\left(1 + {x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right)\right)} \cdot x \]
    3. Step-by-step derivation
      1. lower-+.f64N/A

        \[\leadsto \left(1 + \color{blue}{{x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right)}\right) \cdot x \]
      2. lower-*.f64N/A

        \[\leadsto \left(1 + {x}^{2} \cdot \color{blue}{\left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right)}\right) \cdot x \]
    4. Applied rewrites50.4%

      \[\leadsto \color{blue}{\left(1 + {x}^{2} \cdot \left(0.265709700396151 \cdot {x}^{2} - 0.6665536072\right)\right)} \cdot x \]
    5. Step-by-step derivation
      1. lift-+.f64N/A

        \[\leadsto \left(1 + \color{blue}{{x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right)}\right) \cdot x \]
      2. +-commutativeN/A

        \[\leadsto \left({x}^{2} \cdot \left(\frac{3321371254951887171}{12500000000000000000} \cdot {x}^{2} - \frac{833192009}{1250000000}\right) + \color{blue}{1}\right) \cdot x \]
    6. Applied rewrites50.4%

      \[\leadsto \color{blue}{\mathsf{fma}\left(\mathsf{fma}\left(0.265709700396151, x \cdot x, -0.6665536072\right), x \cdot x, 1\right)} \cdot x \]

    if 1.1000000000000001 < x

    1. Initial program 53.7%

      \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
    2. Taylor expanded in x around inf

      \[\leadsto \color{blue}{\frac{\frac{1}{2} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}}{x}} \]
    3. Step-by-step derivation
      1. lower-/.f64N/A

        \[\leadsto \frac{\frac{1}{2} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}}{\color{blue}{x}} \]
    4. Applied rewrites52.0%

      \[\leadsto \color{blue}{\frac{0.5 + 0.2514179000665374 \cdot \frac{1}{{x}^{2}}}{x}} \]
    5. Step-by-step derivation
      1. lift-+.f64N/A

        \[\leadsto \frac{\frac{1}{2} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}}{x} \]
      2. +-commutativeN/A

        \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} + \frac{1}{2}}{x} \]
      3. add-flipN/A

        \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      4. lower--.f64N/A

        \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      5. lift-*.f64N/A

        \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      6. lift-/.f64N/A

        \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      7. lift-pow.f64N/A

        \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      8. pow2N/A

        \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{x \cdot x} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      9. mult-flip-revN/A

        \[\leadsto \frac{\frac{\frac{600041}{2386628}}{x \cdot x} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      10. lower-/.f64N/A

        \[\leadsto \frac{\frac{\frac{600041}{2386628}}{x \cdot x} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      11. lower-*.f64N/A

        \[\leadsto \frac{\frac{\frac{600041}{2386628}}{x \cdot x} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      12. metadata-eval52.0%

        \[\leadsto \frac{\frac{0.2514179000665374}{x \cdot x} - -0.5}{x} \]
    6. Applied rewrites52.0%

      \[\leadsto \frac{\frac{0.2514179000665374}{x \cdot x} - -0.5}{\color{blue}{x}} \]
  3. Recombined 2 regimes into one program.
  4. Add Preprocessing

Alternative 5: 99.5% accurate, 9.1× speedup?

\[\begin{array}{l} t_0 := \left|x\right| \cdot \left|x\right|\\ \mathsf{copysign}\left(1, x\right) \cdot \begin{array}{l} \mathbf{if}\;\left|x\right| \leq 1.2:\\ \;\;\;\;\mathsf{fma}\left(t\_0, -0.6665536072, 1\right) \cdot \left|x\right|\\ \mathbf{else}:\\ \;\;\;\;\frac{\frac{0.2514179000665374}{t\_0} - -0.5}{\left|x\right|}\\ \end{array} \end{array} \]
(FPCore (x)
 :precision binary64
 (let* ((t_0 (* (fabs x) (fabs x))))
   (*
    (copysign 1.0 x)
    (if (<= (fabs x) 1.2)
      (* (fma t_0 -0.6665536072 1.0) (fabs x))
      (/ (- (/ 0.2514179000665374 t_0) -0.5) (fabs x))))))
double code(double x) {
	double t_0 = fabs(x) * fabs(x);
	double tmp;
	if (fabs(x) <= 1.2) {
		tmp = fma(t_0, -0.6665536072, 1.0) * fabs(x);
	} else {
		tmp = ((0.2514179000665374 / t_0) - -0.5) / fabs(x);
	}
	return copysign(1.0, x) * tmp;
}
function code(x)
	t_0 = Float64(abs(x) * abs(x))
	tmp = 0.0
	if (abs(x) <= 1.2)
		tmp = Float64(fma(t_0, -0.6665536072, 1.0) * abs(x));
	else
		tmp = Float64(Float64(Float64(0.2514179000665374 / t_0) - -0.5) / abs(x));
	end
	return Float64(copysign(1.0, x) * tmp)
end
code[x_] := Block[{t$95$0 = N[(N[Abs[x], $MachinePrecision] * N[Abs[x], $MachinePrecision]), $MachinePrecision]}, N[(N[With[{TMP1 = Abs[1.0], TMP2 = Sign[x]}, TMP1 * If[TMP2 == 0, 1, TMP2]], $MachinePrecision] * If[LessEqual[N[Abs[x], $MachinePrecision], 1.2], N[(N[(t$95$0 * -0.6665536072 + 1.0), $MachinePrecision] * N[Abs[x], $MachinePrecision]), $MachinePrecision], N[(N[(N[(0.2514179000665374 / t$95$0), $MachinePrecision] - -0.5), $MachinePrecision] / N[Abs[x], $MachinePrecision]), $MachinePrecision]]), $MachinePrecision]]
\begin{array}{l}
t_0 := \left|x\right| \cdot \left|x\right|\\
\mathsf{copysign}\left(1, x\right) \cdot \begin{array}{l}
\mathbf{if}\;\left|x\right| \leq 1.2:\\
\;\;\;\;\mathsf{fma}\left(t\_0, -0.6665536072, 1\right) \cdot \left|x\right|\\

\mathbf{else}:\\
\;\;\;\;\frac{\frac{0.2514179000665374}{t\_0} - -0.5}{\left|x\right|}\\


\end{array}
\end{array}
Derivation
  1. Split input into 2 regimes
  2. if x < 1.2

    1. Initial program 53.7%

      \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
    2. Taylor expanded in x around 0

      \[\leadsto \color{blue}{\left(1 + \frac{-833192009}{1250000000} \cdot {x}^{2}\right)} \cdot x \]
    3. Step-by-step derivation
      1. lower-+.f64N/A

        \[\leadsto \left(1 + \color{blue}{\frac{-833192009}{1250000000} \cdot {x}^{2}}\right) \cdot x \]
      2. lower-*.f64N/A

        \[\leadsto \left(1 + \frac{-833192009}{1250000000} \cdot \color{blue}{{x}^{2}}\right) \cdot x \]
      3. lower-pow.f6449.5%

        \[\leadsto \left(1 + -0.6665536072 \cdot {x}^{\color{blue}{2}}\right) \cdot x \]
    4. Applied rewrites49.5%

      \[\leadsto \color{blue}{\left(1 + -0.6665536072 \cdot {x}^{2}\right)} \cdot x \]
    5. Step-by-step derivation
      1. lift-+.f64N/A

        \[\leadsto \left(1 + \color{blue}{\frac{-833192009}{1250000000} \cdot {x}^{2}}\right) \cdot x \]
      2. +-commutativeN/A

        \[\leadsto \left(\frac{-833192009}{1250000000} \cdot {x}^{2} + \color{blue}{1}\right) \cdot x \]
      3. lift-*.f64N/A

        \[\leadsto \left(\frac{-833192009}{1250000000} \cdot {x}^{2} + 1\right) \cdot x \]
      4. lift-pow.f64N/A

        \[\leadsto \left(\frac{-833192009}{1250000000} \cdot {x}^{2} + 1\right) \cdot x \]
      5. pow2N/A

        \[\leadsto \left(\frac{-833192009}{1250000000} \cdot \left(x \cdot x\right) + 1\right) \cdot x \]
      6. *-commutativeN/A

        \[\leadsto \left(\left(x \cdot x\right) \cdot \frac{-833192009}{1250000000} + 1\right) \cdot x \]
      7. lower-fma.f64N/A

        \[\leadsto \mathsf{fma}\left(x \cdot x, \color{blue}{\frac{-833192009}{1250000000}}, 1\right) \cdot x \]
      8. lower-*.f6449.5%

        \[\leadsto \mathsf{fma}\left(x \cdot x, -0.6665536072, 1\right) \cdot x \]
    6. Applied rewrites49.5%

      \[\leadsto \mathsf{fma}\left(x \cdot x, \color{blue}{-0.6665536072}, 1\right) \cdot x \]

    if 1.2 < x

    1. Initial program 53.7%

      \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
    2. Taylor expanded in x around inf

      \[\leadsto \color{blue}{\frac{\frac{1}{2} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}}{x}} \]
    3. Step-by-step derivation
      1. lower-/.f64N/A

        \[\leadsto \frac{\frac{1}{2} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}}{\color{blue}{x}} \]
    4. Applied rewrites52.0%

      \[\leadsto \color{blue}{\frac{0.5 + 0.2514179000665374 \cdot \frac{1}{{x}^{2}}}{x}} \]
    5. Step-by-step derivation
      1. lift-+.f64N/A

        \[\leadsto \frac{\frac{1}{2} + \frac{600041}{2386628} \cdot \frac{1}{{x}^{2}}}{x} \]
      2. +-commutativeN/A

        \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} + \frac{1}{2}}{x} \]
      3. add-flipN/A

        \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      4. lower--.f64N/A

        \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      5. lift-*.f64N/A

        \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      6. lift-/.f64N/A

        \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      7. lift-pow.f64N/A

        \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{{x}^{2}} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      8. pow2N/A

        \[\leadsto \frac{\frac{600041}{2386628} \cdot \frac{1}{x \cdot x} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      9. mult-flip-revN/A

        \[\leadsto \frac{\frac{\frac{600041}{2386628}}{x \cdot x} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      10. lower-/.f64N/A

        \[\leadsto \frac{\frac{\frac{600041}{2386628}}{x \cdot x} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      11. lower-*.f64N/A

        \[\leadsto \frac{\frac{\frac{600041}{2386628}}{x \cdot x} - \left(\mathsf{neg}\left(\frac{1}{2}\right)\right)}{x} \]
      12. metadata-eval52.0%

        \[\leadsto \frac{\frac{0.2514179000665374}{x \cdot x} - -0.5}{x} \]
    6. Applied rewrites52.0%

      \[\leadsto \frac{\frac{0.2514179000665374}{x \cdot x} - -0.5}{\color{blue}{x}} \]
  3. Recombined 2 regimes into one program.
  4. Add Preprocessing

Alternative 6: 99.2% accurate, 9.6× speedup?

\[\mathsf{copysign}\left(1, x\right) \cdot \begin{array}{l} \mathbf{if}\;\left|x\right| \leq 1.2:\\ \;\;\;\;\mathsf{fma}\left(\left|x\right| \cdot \left|x\right|, -0.6665536072, 1\right) \cdot \left|x\right|\\ \mathbf{else}:\\ \;\;\;\;\frac{0.5}{\left|x\right|}\\ \end{array} \]
(FPCore (x)
 :precision binary64
 (*
  (copysign 1.0 x)
  (if (<= (fabs x) 1.2)
    (* (fma (* (fabs x) (fabs x)) -0.6665536072 1.0) (fabs x))
    (/ 0.5 (fabs x)))))
double code(double x) {
	double tmp;
	if (fabs(x) <= 1.2) {
		tmp = fma((fabs(x) * fabs(x)), -0.6665536072, 1.0) * fabs(x);
	} else {
		tmp = 0.5 / fabs(x);
	}
	return copysign(1.0, x) * tmp;
}
function code(x)
	tmp = 0.0
	if (abs(x) <= 1.2)
		tmp = Float64(fma(Float64(abs(x) * abs(x)), -0.6665536072, 1.0) * abs(x));
	else
		tmp = Float64(0.5 / abs(x));
	end
	return Float64(copysign(1.0, x) * tmp)
end
code[x_] := N[(N[With[{TMP1 = Abs[1.0], TMP2 = Sign[x]}, TMP1 * If[TMP2 == 0, 1, TMP2]], $MachinePrecision] * If[LessEqual[N[Abs[x], $MachinePrecision], 1.2], N[(N[(N[(N[Abs[x], $MachinePrecision] * N[Abs[x], $MachinePrecision]), $MachinePrecision] * -0.6665536072 + 1.0), $MachinePrecision] * N[Abs[x], $MachinePrecision]), $MachinePrecision], N[(0.5 / N[Abs[x], $MachinePrecision]), $MachinePrecision]]), $MachinePrecision]
\mathsf{copysign}\left(1, x\right) \cdot \begin{array}{l}
\mathbf{if}\;\left|x\right| \leq 1.2:\\
\;\;\;\;\mathsf{fma}\left(\left|x\right| \cdot \left|x\right|, -0.6665536072, 1\right) \cdot \left|x\right|\\

\mathbf{else}:\\
\;\;\;\;\frac{0.5}{\left|x\right|}\\


\end{array}
Derivation
  1. Split input into 2 regimes
  2. if x < 1.2

    1. Initial program 53.7%

      \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
    2. Taylor expanded in x around 0

      \[\leadsto \color{blue}{\left(1 + \frac{-833192009}{1250000000} \cdot {x}^{2}\right)} \cdot x \]
    3. Step-by-step derivation
      1. lower-+.f64N/A

        \[\leadsto \left(1 + \color{blue}{\frac{-833192009}{1250000000} \cdot {x}^{2}}\right) \cdot x \]
      2. lower-*.f64N/A

        \[\leadsto \left(1 + \frac{-833192009}{1250000000} \cdot \color{blue}{{x}^{2}}\right) \cdot x \]
      3. lower-pow.f6449.5%

        \[\leadsto \left(1 + -0.6665536072 \cdot {x}^{\color{blue}{2}}\right) \cdot x \]
    4. Applied rewrites49.5%

      \[\leadsto \color{blue}{\left(1 + -0.6665536072 \cdot {x}^{2}\right)} \cdot x \]
    5. Step-by-step derivation
      1. lift-+.f64N/A

        \[\leadsto \left(1 + \color{blue}{\frac{-833192009}{1250000000} \cdot {x}^{2}}\right) \cdot x \]
      2. +-commutativeN/A

        \[\leadsto \left(\frac{-833192009}{1250000000} \cdot {x}^{2} + \color{blue}{1}\right) \cdot x \]
      3. lift-*.f64N/A

        \[\leadsto \left(\frac{-833192009}{1250000000} \cdot {x}^{2} + 1\right) \cdot x \]
      4. lift-pow.f64N/A

        \[\leadsto \left(\frac{-833192009}{1250000000} \cdot {x}^{2} + 1\right) \cdot x \]
      5. pow2N/A

        \[\leadsto \left(\frac{-833192009}{1250000000} \cdot \left(x \cdot x\right) + 1\right) \cdot x \]
      6. *-commutativeN/A

        \[\leadsto \left(\left(x \cdot x\right) \cdot \frac{-833192009}{1250000000} + 1\right) \cdot x \]
      7. lower-fma.f64N/A

        \[\leadsto \mathsf{fma}\left(x \cdot x, \color{blue}{\frac{-833192009}{1250000000}}, 1\right) \cdot x \]
      8. lower-*.f6449.5%

        \[\leadsto \mathsf{fma}\left(x \cdot x, -0.6665536072, 1\right) \cdot x \]
    6. Applied rewrites49.5%

      \[\leadsto \mathsf{fma}\left(x \cdot x, \color{blue}{-0.6665536072}, 1\right) \cdot x \]

    if 1.2 < x

    1. Initial program 53.7%

      \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
    2. Taylor expanded in x around inf

      \[\leadsto \color{blue}{\frac{\frac{1}{2}}{x}} \]
    3. Step-by-step derivation
      1. lower-/.f6452.2%

        \[\leadsto \frac{0.5}{\color{blue}{x}} \]
    4. Applied rewrites52.2%

      \[\leadsto \color{blue}{\frac{0.5}{x}} \]
  3. Recombined 2 regimes into one program.
  4. Add Preprocessing

Alternative 7: 99.0% accurate, 15.4× speedup?

\[\mathsf{copysign}\left(1, x\right) \cdot \begin{array}{l} \mathbf{if}\;\left|x\right| \leq 0.7:\\ \;\;\;\;\left|x\right|\\ \mathbf{else}:\\ \;\;\;\;\frac{0.5}{\left|x\right|}\\ \end{array} \]
(FPCore (x)
 :precision binary64
 (* (copysign 1.0 x) (if (<= (fabs x) 0.7) (fabs x) (/ 0.5 (fabs x)))))
double code(double x) {
	double tmp;
	if (fabs(x) <= 0.7) {
		tmp = fabs(x);
	} else {
		tmp = 0.5 / fabs(x);
	}
	return copysign(1.0, x) * tmp;
}
public static double code(double x) {
	double tmp;
	if (Math.abs(x) <= 0.7) {
		tmp = Math.abs(x);
	} else {
		tmp = 0.5 / Math.abs(x);
	}
	return Math.copySign(1.0, x) * tmp;
}
def code(x):
	tmp = 0
	if math.fabs(x) <= 0.7:
		tmp = math.fabs(x)
	else:
		tmp = 0.5 / math.fabs(x)
	return math.copysign(1.0, x) * tmp
function code(x)
	tmp = 0.0
	if (abs(x) <= 0.7)
		tmp = abs(x);
	else
		tmp = Float64(0.5 / abs(x));
	end
	return Float64(copysign(1.0, x) * tmp)
end
function tmp_2 = code(x)
	tmp = 0.0;
	if (abs(x) <= 0.7)
		tmp = abs(x);
	else
		tmp = 0.5 / abs(x);
	end
	tmp_2 = (sign(x) * abs(1.0)) * tmp;
end
code[x_] := N[(N[With[{TMP1 = Abs[1.0], TMP2 = Sign[x]}, TMP1 * If[TMP2 == 0, 1, TMP2]], $MachinePrecision] * If[LessEqual[N[Abs[x], $MachinePrecision], 0.7], N[Abs[x], $MachinePrecision], N[(0.5 / N[Abs[x], $MachinePrecision]), $MachinePrecision]]), $MachinePrecision]
\mathsf{copysign}\left(1, x\right) \cdot \begin{array}{l}
\mathbf{if}\;\left|x\right| \leq 0.7:\\
\;\;\;\;\left|x\right|\\

\mathbf{else}:\\
\;\;\;\;\frac{0.5}{\left|x\right|}\\


\end{array}
Derivation
  1. Split input into 2 regimes
  2. if x < 0.69999999999999996

    1. Initial program 53.7%

      \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
    2. Taylor expanded in x around 0

      \[\leadsto \color{blue}{x} \]
    3. Step-by-step derivation
      1. Applied rewrites50.6%

        \[\leadsto \color{blue}{x} \]

      if 0.69999999999999996 < x

      1. Initial program 53.7%

        \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
      2. Taylor expanded in x around inf

        \[\leadsto \color{blue}{\frac{\frac{1}{2}}{x}} \]
      3. Step-by-step derivation
        1. lower-/.f6452.2%

          \[\leadsto \frac{0.5}{\color{blue}{x}} \]
      4. Applied rewrites52.2%

        \[\leadsto \color{blue}{\frac{0.5}{x}} \]
    4. Recombined 2 regimes into one program.
    5. Add Preprocessing

    Alternative 8: 50.6% accurate, 253.1× speedup?

    \[x \]
    (FPCore (x) :precision binary64 x)
    double code(double x) {
    	return x;
    }
    
    module fmin_fmax_functions
        implicit none
        private
        public fmax
        public fmin
    
        interface fmax
            module procedure fmax88
            module procedure fmax44
            module procedure fmax84
            module procedure fmax48
        end interface
        interface fmin
            module procedure fmin88
            module procedure fmin44
            module procedure fmin84
            module procedure fmin48
        end interface
    contains
        real(8) function fmax88(x, y) result (res)
            real(8), intent (in) :: x
            real(8), intent (in) :: y
            res = merge(y, merge(x, max(x, y), y /= y), x /= x)
        end function
        real(4) function fmax44(x, y) result (res)
            real(4), intent (in) :: x
            real(4), intent (in) :: y
            res = merge(y, merge(x, max(x, y), y /= y), x /= x)
        end function
        real(8) function fmax84(x, y) result(res)
            real(8), intent (in) :: x
            real(4), intent (in) :: y
            res = merge(dble(y), merge(x, max(x, dble(y)), y /= y), x /= x)
        end function
        real(8) function fmax48(x, y) result(res)
            real(4), intent (in) :: x
            real(8), intent (in) :: y
            res = merge(y, merge(dble(x), max(dble(x), y), y /= y), x /= x)
        end function
        real(8) function fmin88(x, y) result (res)
            real(8), intent (in) :: x
            real(8), intent (in) :: y
            res = merge(y, merge(x, min(x, y), y /= y), x /= x)
        end function
        real(4) function fmin44(x, y) result (res)
            real(4), intent (in) :: x
            real(4), intent (in) :: y
            res = merge(y, merge(x, min(x, y), y /= y), x /= x)
        end function
        real(8) function fmin84(x, y) result(res)
            real(8), intent (in) :: x
            real(4), intent (in) :: y
            res = merge(dble(y), merge(x, min(x, dble(y)), y /= y), x /= x)
        end function
        real(8) function fmin48(x, y) result(res)
            real(4), intent (in) :: x
            real(8), intent (in) :: y
            res = merge(y, merge(dble(x), min(dble(x), y), y /= y), x /= x)
        end function
    end module
    
    real(8) function code(x)
    use fmin_fmax_functions
        real(8), intent (in) :: x
        code = x
    end function
    
    public static double code(double x) {
    	return x;
    }
    
    def code(x):
    	return x
    
    function code(x)
    	return x
    end
    
    function tmp = code(x)
    	tmp = x;
    end
    
    code[x_] := x
    
    x
    
    Derivation
    1. Initial program 53.7%

      \[\frac{\left(\left(\left(\left(1 + 0.1049934947 \cdot \left(x \cdot x\right)\right) + 0.0424060604 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0072644182 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0005064034 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0001789971 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)}{\left(\left(\left(\left(\left(1 + 0.7715471019 \cdot \left(x \cdot x\right)\right) + 0.2909738639 \cdot \left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0694555761 \cdot \left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0140005442 \cdot \left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + 0.0008327945 \cdot \left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)\right) + \left(2 \cdot 0.0001789971\right) \cdot \left(\left(\left(\left(\left(\left(x \cdot x\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right) \cdot \left(x \cdot x\right)\right)} \cdot x \]
    2. Taylor expanded in x around 0

      \[\leadsto \color{blue}{x} \]
    3. Step-by-step derivation
      1. Applied rewrites50.6%

        \[\leadsto \color{blue}{x} \]
      2. Add Preprocessing

      Reproduce

      ?
      herbie shell --seed 2025193 
      (FPCore (x)
        :name "Jmat.Real.dawson"
        :precision binary64
        (* (/ (+ (+ (+ (+ (+ 1.0 (* 0.1049934947 (* x x))) (* 0.0424060604 (* (* x x) (* x x)))) (* 0.0072644182 (* (* (* x x) (* x x)) (* x x)))) (* 0.0005064034 (* (* (* (* x x) (* x x)) (* x x)) (* x x)))) (* 0.0001789971 (* (* (* (* (* x x) (* x x)) (* x x)) (* x x)) (* x x)))) (+ (+ (+ (+ (+ (+ 1.0 (* 0.7715471019 (* x x))) (* 0.2909738639 (* (* x x) (* x x)))) (* 0.0694555761 (* (* (* x x) (* x x)) (* x x)))) (* 0.0140005442 (* (* (* (* x x) (* x x)) (* x x)) (* x x)))) (* 0.0008327945 (* (* (* (* (* x x) (* x x)) (* x x)) (* x x)) (* x x)))) (* (* 2.0 0.0001789971) (* (* (* (* (* (* x x) (* x x)) (* x x)) (* x x)) (* x x)) (* x x))))) x))