Linear.Quaternion:$ccos from linear-1.19.1.3

Percentage Accurate: 100.0% → 100.0%
Time: 5.4s
Alternatives: 13
Speedup: 1.0×

Specification

?
\[\begin{array}{l} \\ \sin x \cdot \frac{\sinh y}{y} \end{array} \]
(FPCore (x y) :precision binary64 (* (sin x) (/ (sinh y) y)))
double code(double x, double y) {
	return sin(x) * (sinh(y) / y);
}
module fmin_fmax_functions
    implicit none
    private
    public fmax
    public fmin

    interface fmax
        module procedure fmax88
        module procedure fmax44
        module procedure fmax84
        module procedure fmax48
    end interface
    interface fmin
        module procedure fmin88
        module procedure fmin44
        module procedure fmin84
        module procedure fmin48
    end interface
contains
    real(8) function fmax88(x, y) result (res)
        real(8), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(x, max(x, y), y /= y), x /= x)
    end function
    real(4) function fmax44(x, y) result (res)
        real(4), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(y, merge(x, max(x, y), y /= y), x /= x)
    end function
    real(8) function fmax84(x, y) result(res)
        real(8), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(dble(y), merge(x, max(x, dble(y)), y /= y), x /= x)
    end function
    real(8) function fmax48(x, y) result(res)
        real(4), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(dble(x), max(dble(x), y), y /= y), x /= x)
    end function
    real(8) function fmin88(x, y) result (res)
        real(8), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(x, min(x, y), y /= y), x /= x)
    end function
    real(4) function fmin44(x, y) result (res)
        real(4), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(y, merge(x, min(x, y), y /= y), x /= x)
    end function
    real(8) function fmin84(x, y) result(res)
        real(8), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(dble(y), merge(x, min(x, dble(y)), y /= y), x /= x)
    end function
    real(8) function fmin48(x, y) result(res)
        real(4), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(dble(x), min(dble(x), y), y /= y), x /= x)
    end function
end module

real(8) function code(x, y)
use fmin_fmax_functions
    real(8), intent (in) :: x
    real(8), intent (in) :: y
    code = sin(x) * (sinh(y) / y)
end function
public static double code(double x, double y) {
	return Math.sin(x) * (Math.sinh(y) / y);
}
def code(x, y):
	return math.sin(x) * (math.sinh(y) / y)
function code(x, y)
	return Float64(sin(x) * Float64(sinh(y) / y))
end
function tmp = code(x, y)
	tmp = sin(x) * (sinh(y) / y);
end
code[x_, y_] := N[(N[Sin[x], $MachinePrecision] * N[(N[Sinh[y], $MachinePrecision] / y), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}

\\
\sin x \cdot \frac{\sinh y}{y}
\end{array}

Sampling outcomes in binary64 precision:

Local Percentage Accuracy vs ?

The average percentage accuracy by input value. Horizontal axis shows value of an input variable; the variable is choosen in the title. Vertical axis is accuracy; higher is better. Red represent the original program, while blue represents Herbie's suggestion. These can be toggled with buttons below the plot. The line is an average while dots represent individual samples.

Accuracy vs Speed?

Herbie found 13 alternatives:

AlternativeAccuracySpeedup
The accuracy (vertical axis) and speed (horizontal axis) of each alternatives. Up and to the right is better. The red square shows the initial program, and each blue circle shows an alternative.The line shows the best available speed-accuracy tradeoffs.

Initial Program: 100.0% accurate, 1.0× speedup?

\[\begin{array}{l} \\ \sin x \cdot \frac{\sinh y}{y} \end{array} \]
(FPCore (x y) :precision binary64 (* (sin x) (/ (sinh y) y)))
double code(double x, double y) {
	return sin(x) * (sinh(y) / y);
}
module fmin_fmax_functions
    implicit none
    private
    public fmax
    public fmin

    interface fmax
        module procedure fmax88
        module procedure fmax44
        module procedure fmax84
        module procedure fmax48
    end interface
    interface fmin
        module procedure fmin88
        module procedure fmin44
        module procedure fmin84
        module procedure fmin48
    end interface
contains
    real(8) function fmax88(x, y) result (res)
        real(8), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(x, max(x, y), y /= y), x /= x)
    end function
    real(4) function fmax44(x, y) result (res)
        real(4), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(y, merge(x, max(x, y), y /= y), x /= x)
    end function
    real(8) function fmax84(x, y) result(res)
        real(8), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(dble(y), merge(x, max(x, dble(y)), y /= y), x /= x)
    end function
    real(8) function fmax48(x, y) result(res)
        real(4), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(dble(x), max(dble(x), y), y /= y), x /= x)
    end function
    real(8) function fmin88(x, y) result (res)
        real(8), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(x, min(x, y), y /= y), x /= x)
    end function
    real(4) function fmin44(x, y) result (res)
        real(4), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(y, merge(x, min(x, y), y /= y), x /= x)
    end function
    real(8) function fmin84(x, y) result(res)
        real(8), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(dble(y), merge(x, min(x, dble(y)), y /= y), x /= x)
    end function
    real(8) function fmin48(x, y) result(res)
        real(4), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(dble(x), min(dble(x), y), y /= y), x /= x)
    end function
end module

real(8) function code(x, y)
use fmin_fmax_functions
    real(8), intent (in) :: x
    real(8), intent (in) :: y
    code = sin(x) * (sinh(y) / y)
end function
public static double code(double x, double y) {
	return Math.sin(x) * (Math.sinh(y) / y);
}
def code(x, y):
	return math.sin(x) * (math.sinh(y) / y)
function code(x, y)
	return Float64(sin(x) * Float64(sinh(y) / y))
end
function tmp = code(x, y)
	tmp = sin(x) * (sinh(y) / y);
end
code[x_, y_] := N[(N[Sin[x], $MachinePrecision] * N[(N[Sinh[y], $MachinePrecision] / y), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}

\\
\sin x \cdot \frac{\sinh y}{y}
\end{array}

Alternative 1: 100.0% accurate, 1.0× speedup?

\[\begin{array}{l} \\ \sin x \cdot \frac{\sinh y}{y} \end{array} \]
(FPCore (x y) :precision binary64 (* (sin x) (/ (sinh y) y)))
double code(double x, double y) {
	return sin(x) * (sinh(y) / y);
}
module fmin_fmax_functions
    implicit none
    private
    public fmax
    public fmin

    interface fmax
        module procedure fmax88
        module procedure fmax44
        module procedure fmax84
        module procedure fmax48
    end interface
    interface fmin
        module procedure fmin88
        module procedure fmin44
        module procedure fmin84
        module procedure fmin48
    end interface
contains
    real(8) function fmax88(x, y) result (res)
        real(8), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(x, max(x, y), y /= y), x /= x)
    end function
    real(4) function fmax44(x, y) result (res)
        real(4), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(y, merge(x, max(x, y), y /= y), x /= x)
    end function
    real(8) function fmax84(x, y) result(res)
        real(8), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(dble(y), merge(x, max(x, dble(y)), y /= y), x /= x)
    end function
    real(8) function fmax48(x, y) result(res)
        real(4), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(dble(x), max(dble(x), y), y /= y), x /= x)
    end function
    real(8) function fmin88(x, y) result (res)
        real(8), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(x, min(x, y), y /= y), x /= x)
    end function
    real(4) function fmin44(x, y) result (res)
        real(4), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(y, merge(x, min(x, y), y /= y), x /= x)
    end function
    real(8) function fmin84(x, y) result(res)
        real(8), intent (in) :: x
        real(4), intent (in) :: y
        res = merge(dble(y), merge(x, min(x, dble(y)), y /= y), x /= x)
    end function
    real(8) function fmin48(x, y) result(res)
        real(4), intent (in) :: x
        real(8), intent (in) :: y
        res = merge(y, merge(dble(x), min(dble(x), y), y /= y), x /= x)
    end function
end module

real(8) function code(x, y)
use fmin_fmax_functions
    real(8), intent (in) :: x
    real(8), intent (in) :: y
    code = sin(x) * (sinh(y) / y)
end function
public static double code(double x, double y) {
	return Math.sin(x) * (Math.sinh(y) / y);
}
def code(x, y):
	return math.sin(x) * (math.sinh(y) / y)
function code(x, y)
	return Float64(sin(x) * Float64(sinh(y) / y))
end
function tmp = code(x, y)
	tmp = sin(x) * (sinh(y) / y);
end
code[x_, y_] := N[(N[Sin[x], $MachinePrecision] * N[(N[Sinh[y], $MachinePrecision] / y), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}

\\
\sin x \cdot \frac{\sinh y}{y}
\end{array}
Derivation
  1. Initial program 100.0%

    \[\sin x \cdot \frac{\sinh y}{y} \]
  2. Add Preprocessing
  3. Add Preprocessing

Alternative 2: 80.7% accurate, 0.4× speedup?

\[\begin{array}{l} \\ \begin{array}{l} t_0 := \sin x \cdot \frac{\sinh y}{y}\\ \mathbf{if}\;t\_0 \leq -\infty:\\ \;\;\;\;\mathsf{fma}\left(\left(x \cdot x\right) \cdot x, -0.16666666666666666, x\right) \cdot \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right)\\ \mathbf{elif}\;t\_0 \leq 1:\\ \;\;\;\;\sin x \cdot \mathsf{fma}\left(0.16666666666666666 \cdot y, y, 1\right)\\ \mathbf{else}:\\ \;\;\;\;\sin x \cdot \left(\left(\mathsf{fma}\left(y \cdot y, 0.008333333333333333, 0.16666666666666666\right) \cdot y\right) \cdot y\right)\\ \end{array} \end{array} \]
(FPCore (x y)
 :precision binary64
 (let* ((t_0 (* (sin x) (/ (sinh y) y))))
   (if (<= t_0 (- INFINITY))
     (*
      (fma (* (* x x) x) -0.16666666666666666 x)
      (fma (* y y) 0.16666666666666666 1.0))
     (if (<= t_0 1.0)
       (* (sin x) (fma (* 0.16666666666666666 y) y 1.0))
       (*
        (sin x)
        (* (* (fma (* y y) 0.008333333333333333 0.16666666666666666) y) y))))))
double code(double x, double y) {
	double t_0 = sin(x) * (sinh(y) / y);
	double tmp;
	if (t_0 <= -((double) INFINITY)) {
		tmp = fma(((x * x) * x), -0.16666666666666666, x) * fma((y * y), 0.16666666666666666, 1.0);
	} else if (t_0 <= 1.0) {
		tmp = sin(x) * fma((0.16666666666666666 * y), y, 1.0);
	} else {
		tmp = sin(x) * ((fma((y * y), 0.008333333333333333, 0.16666666666666666) * y) * y);
	}
	return tmp;
}
function code(x, y)
	t_0 = Float64(sin(x) * Float64(sinh(y) / y))
	tmp = 0.0
	if (t_0 <= Float64(-Inf))
		tmp = Float64(fma(Float64(Float64(x * x) * x), -0.16666666666666666, x) * fma(Float64(y * y), 0.16666666666666666, 1.0));
	elseif (t_0 <= 1.0)
		tmp = Float64(sin(x) * fma(Float64(0.16666666666666666 * y), y, 1.0));
	else
		tmp = Float64(sin(x) * Float64(Float64(fma(Float64(y * y), 0.008333333333333333, 0.16666666666666666) * y) * y));
	end
	return tmp
end
code[x_, y_] := Block[{t$95$0 = N[(N[Sin[x], $MachinePrecision] * N[(N[Sinh[y], $MachinePrecision] / y), $MachinePrecision]), $MachinePrecision]}, If[LessEqual[t$95$0, (-Infinity)], N[(N[(N[(N[(x * x), $MachinePrecision] * x), $MachinePrecision] * -0.16666666666666666 + x), $MachinePrecision] * N[(N[(y * y), $MachinePrecision] * 0.16666666666666666 + 1.0), $MachinePrecision]), $MachinePrecision], If[LessEqual[t$95$0, 1.0], N[(N[Sin[x], $MachinePrecision] * N[(N[(0.16666666666666666 * y), $MachinePrecision] * y + 1.0), $MachinePrecision]), $MachinePrecision], N[(N[Sin[x], $MachinePrecision] * N[(N[(N[(N[(y * y), $MachinePrecision] * 0.008333333333333333 + 0.16666666666666666), $MachinePrecision] * y), $MachinePrecision] * y), $MachinePrecision]), $MachinePrecision]]]]
\begin{array}{l}

\\
\begin{array}{l}
t_0 := \sin x \cdot \frac{\sinh y}{y}\\
\mathbf{if}\;t\_0 \leq -\infty:\\
\;\;\;\;\mathsf{fma}\left(\left(x \cdot x\right) \cdot x, -0.16666666666666666, x\right) \cdot \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right)\\

\mathbf{elif}\;t\_0 \leq 1:\\
\;\;\;\;\sin x \cdot \mathsf{fma}\left(0.16666666666666666 \cdot y, y, 1\right)\\

\mathbf{else}:\\
\;\;\;\;\sin x \cdot \left(\left(\mathsf{fma}\left(y \cdot y, 0.008333333333333333, 0.16666666666666666\right) \cdot y\right) \cdot y\right)\\


\end{array}
\end{array}
Derivation
  1. Split input into 3 regimes
  2. if (*.f64 (sin.f64 x) (/.f64 (sinh.f64 y) y)) < -inf.0

    1. Initial program 100.0%

      \[\sin x \cdot \frac{\sinh y}{y} \]
    2. Add Preprocessing
    3. Taylor expanded in y around 0

      \[\leadsto \sin x \cdot \color{blue}{\left(1 + \frac{1}{6} \cdot {y}^{2}\right)} \]
    4. Step-by-step derivation
      1. +-commutativeN/A

        \[\leadsto \sin x \cdot \color{blue}{\left(\frac{1}{6} \cdot {y}^{2} + 1\right)} \]
      2. *-commutativeN/A

        \[\leadsto \sin x \cdot \left(\color{blue}{{y}^{2} \cdot \frac{1}{6}} + 1\right) \]
      3. lower-fma.f64N/A

        \[\leadsto \sin x \cdot \color{blue}{\mathsf{fma}\left({y}^{2}, \frac{1}{6}, 1\right)} \]
      4. unpow2N/A

        \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{y \cdot y}, \frac{1}{6}, 1\right) \]
      5. lower-*.f6448.8

        \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{y \cdot y}, 0.16666666666666666, 1\right) \]
    5. Applied rewrites48.8%

      \[\leadsto \sin x \cdot \color{blue}{\mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right)} \]
    6. Taylor expanded in x around 0

      \[\leadsto \color{blue}{\left(x \cdot \left(1 + \frac{-1}{6} \cdot {x}^{2}\right)\right)} \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
    7. Step-by-step derivation
      1. +-commutativeN/A

        \[\leadsto \left(x \cdot \color{blue}{\left(\frac{-1}{6} \cdot {x}^{2} + 1\right)}\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
      2. distribute-lft-inN/A

        \[\leadsto \color{blue}{\left(x \cdot \left(\frac{-1}{6} \cdot {x}^{2}\right) + x \cdot 1\right)} \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
      3. *-commutativeN/A

        \[\leadsto \left(x \cdot \color{blue}{\left({x}^{2} \cdot \frac{-1}{6}\right)} + x \cdot 1\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
      4. associate-*r*N/A

        \[\leadsto \left(\color{blue}{\left(x \cdot {x}^{2}\right) \cdot \frac{-1}{6}} + x \cdot 1\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
      5. *-rgt-identityN/A

        \[\leadsto \left(\left(x \cdot {x}^{2}\right) \cdot \frac{-1}{6} + \color{blue}{x}\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
      6. lower-fma.f64N/A

        \[\leadsto \color{blue}{\mathsf{fma}\left(x \cdot {x}^{2}, \frac{-1}{6}, x\right)} \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
      7. unpow2N/A

        \[\leadsto \mathsf{fma}\left(x \cdot \color{blue}{\left(x \cdot x\right)}, \frac{-1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
      8. cube-unmultN/A

        \[\leadsto \mathsf{fma}\left(\color{blue}{{x}^{3}}, \frac{-1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
      9. metadata-evalN/A

        \[\leadsto \mathsf{fma}\left({x}^{\color{blue}{\left(2 + 1\right)}}, \frac{-1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
      10. lower-pow.f64N/A

        \[\leadsto \mathsf{fma}\left(\color{blue}{{x}^{\left(2 + 1\right)}}, \frac{-1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
      11. metadata-eval47.7

        \[\leadsto \mathsf{fma}\left({x}^{\color{blue}{3}}, -0.16666666666666666, x\right) \cdot \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right) \]
    8. Applied rewrites47.7%

      \[\leadsto \color{blue}{\mathsf{fma}\left({x}^{3}, -0.16666666666666666, x\right)} \cdot \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right) \]
    9. Step-by-step derivation
      1. Applied rewrites47.7%

        \[\leadsto \mathsf{fma}\left(\left(x \cdot x\right) \cdot x, -0.16666666666666666, x\right) \cdot \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right) \]

      if -inf.0 < (*.f64 (sin.f64 x) (/.f64 (sinh.f64 y) y)) < 1

      1. Initial program 100.0%

        \[\sin x \cdot \frac{\sinh y}{y} \]
      2. Add Preprocessing
      3. Taylor expanded in y around 0

        \[\leadsto \sin x \cdot \color{blue}{\left(1 + \frac{1}{6} \cdot {y}^{2}\right)} \]
      4. Step-by-step derivation
        1. +-commutativeN/A

          \[\leadsto \sin x \cdot \color{blue}{\left(\frac{1}{6} \cdot {y}^{2} + 1\right)} \]
        2. *-commutativeN/A

          \[\leadsto \sin x \cdot \left(\color{blue}{{y}^{2} \cdot \frac{1}{6}} + 1\right) \]
        3. lower-fma.f64N/A

          \[\leadsto \sin x \cdot \color{blue}{\mathsf{fma}\left({y}^{2}, \frac{1}{6}, 1\right)} \]
        4. unpow2N/A

          \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{y \cdot y}, \frac{1}{6}, 1\right) \]
        5. lower-*.f6497.6

          \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{y \cdot y}, 0.16666666666666666, 1\right) \]
      5. Applied rewrites97.6%

        \[\leadsto \sin x \cdot \color{blue}{\mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right)} \]
      6. Step-by-step derivation
        1. Applied rewrites97.6%

          \[\leadsto \sin x \cdot \mathsf{fma}\left(0.16666666666666666 \cdot y, \color{blue}{y}, 1\right) \]

        if 1 < (*.f64 (sin.f64 x) (/.f64 (sinh.f64 y) y))

        1. Initial program 100.0%

          \[\sin x \cdot \frac{\sinh y}{y} \]
        2. Add Preprocessing
        3. Taylor expanded in y around 0

          \[\leadsto \sin x \cdot \color{blue}{\left(1 + {y}^{2} \cdot \left(\frac{1}{6} + \frac{1}{120} \cdot {y}^{2}\right)\right)} \]
        4. Step-by-step derivation
          1. fp-cancel-sign-sub-invN/A

            \[\leadsto \sin x \cdot \color{blue}{\left(1 - \left(\mathsf{neg}\left({y}^{2}\right)\right) \cdot \left(\frac{1}{6} + \frac{1}{120} \cdot {y}^{2}\right)\right)} \]
          2. fp-cancel-sub-sign-invN/A

            \[\leadsto \sin x \cdot \color{blue}{\left(1 + \left(\mathsf{neg}\left(\left(\mathsf{neg}\left({y}^{2}\right)\right)\right)\right) \cdot \left(\frac{1}{6} + \frac{1}{120} \cdot {y}^{2}\right)\right)} \]
          3. +-commutativeN/A

            \[\leadsto \sin x \cdot \color{blue}{\left(\left(\mathsf{neg}\left(\left(\mathsf{neg}\left({y}^{2}\right)\right)\right)\right) \cdot \left(\frac{1}{6} + \frac{1}{120} \cdot {y}^{2}\right) + 1\right)} \]
          4. remove-double-negN/A

            \[\leadsto \sin x \cdot \left(\color{blue}{{y}^{2}} \cdot \left(\frac{1}{6} + \frac{1}{120} \cdot {y}^{2}\right) + 1\right) \]
          5. *-commutativeN/A

            \[\leadsto \sin x \cdot \left(\color{blue}{\left(\frac{1}{6} + \frac{1}{120} \cdot {y}^{2}\right) \cdot {y}^{2}} + 1\right) \]
          6. lower-fma.f64N/A

            \[\leadsto \sin x \cdot \color{blue}{\mathsf{fma}\left(\frac{1}{6} + \frac{1}{120} \cdot {y}^{2}, {y}^{2}, 1\right)} \]
          7. +-commutativeN/A

            \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{\frac{1}{120} \cdot {y}^{2} + \frac{1}{6}}, {y}^{2}, 1\right) \]
          8. *-commutativeN/A

            \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{{y}^{2} \cdot \frac{1}{120}} + \frac{1}{6}, {y}^{2}, 1\right) \]
          9. lower-fma.f64N/A

            \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{\mathsf{fma}\left({y}^{2}, \frac{1}{120}, \frac{1}{6}\right)}, {y}^{2}, 1\right) \]
          10. unpow2N/A

            \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(\color{blue}{y \cdot y}, \frac{1}{120}, \frac{1}{6}\right), {y}^{2}, 1\right) \]
          11. lower-*.f64N/A

            \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(\color{blue}{y \cdot y}, \frac{1}{120}, \frac{1}{6}\right), {y}^{2}, 1\right) \]
          12. unpow2N/A

            \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(y \cdot y, \frac{1}{120}, \frac{1}{6}\right), \color{blue}{y \cdot y}, 1\right) \]
          13. lower-*.f6478.7

            \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(y \cdot y, 0.008333333333333333, 0.16666666666666666\right), \color{blue}{y \cdot y}, 1\right) \]
        5. Applied rewrites78.7%

          \[\leadsto \sin x \cdot \color{blue}{\mathsf{fma}\left(\mathsf{fma}\left(y \cdot y, 0.008333333333333333, 0.16666666666666666\right), y \cdot y, 1\right)} \]
        6. Step-by-step derivation
          1. Applied rewrites78.7%

            \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(0.008333333333333333, y \cdot y, 0.16666666666666666\right) \cdot y, \color{blue}{y}, 1\right) \]
          2. Taylor expanded in y around inf

            \[\leadsto \sin x \cdot \left({y}^{4} \cdot \color{blue}{\left(\frac{1}{120} + \frac{1}{6} \cdot \frac{1}{{y}^{2}}\right)}\right) \]
          3. Step-by-step derivation
            1. Applied rewrites78.7%

              \[\leadsto \sin x \cdot \left(\left(\mathsf{fma}\left(y \cdot y, 0.008333333333333333, 0.16666666666666666\right) \cdot y\right) \cdot \color{blue}{y}\right) \]
          4. Recombined 3 regimes into one program.
          5. Add Preprocessing

          Alternative 3: 74.4% accurate, 0.4× speedup?

          \[\begin{array}{l} \\ \begin{array}{l} t_0 := \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right)\\ t_1 := \sin x \cdot \frac{\sinh y}{y}\\ \mathbf{if}\;t\_1 \leq -\infty:\\ \;\;\;\;\mathsf{fma}\left(\left(x \cdot x\right) \cdot x, -0.16666666666666666, x\right) \cdot t\_0\\ \mathbf{elif}\;t\_1 \leq 20000:\\ \;\;\;\;\sin x \cdot \mathsf{fma}\left(0.16666666666666666 \cdot y, y, 1\right)\\ \mathbf{else}:\\ \;\;\;\;\mathsf{fma}\left(\mathsf{fma}\left(x \cdot x, 0.008333333333333333, -0.16666666666666666\right) \cdot \left(x \cdot x\right), x, x\right) \cdot t\_0\\ \end{array} \end{array} \]
          (FPCore (x y)
           :precision binary64
           (let* ((t_0 (fma (* y y) 0.16666666666666666 1.0))
                  (t_1 (* (sin x) (/ (sinh y) y))))
             (if (<= t_1 (- INFINITY))
               (* (fma (* (* x x) x) -0.16666666666666666 x) t_0)
               (if (<= t_1 20000.0)
                 (* (sin x) (fma (* 0.16666666666666666 y) y 1.0))
                 (*
                  (fma
                   (* (fma (* x x) 0.008333333333333333 -0.16666666666666666) (* x x))
                   x
                   x)
                  t_0)))))
          double code(double x, double y) {
          	double t_0 = fma((y * y), 0.16666666666666666, 1.0);
          	double t_1 = sin(x) * (sinh(y) / y);
          	double tmp;
          	if (t_1 <= -((double) INFINITY)) {
          		tmp = fma(((x * x) * x), -0.16666666666666666, x) * t_0;
          	} else if (t_1 <= 20000.0) {
          		tmp = sin(x) * fma((0.16666666666666666 * y), y, 1.0);
          	} else {
          		tmp = fma((fma((x * x), 0.008333333333333333, -0.16666666666666666) * (x * x)), x, x) * t_0;
          	}
          	return tmp;
          }
          
          function code(x, y)
          	t_0 = fma(Float64(y * y), 0.16666666666666666, 1.0)
          	t_1 = Float64(sin(x) * Float64(sinh(y) / y))
          	tmp = 0.0
          	if (t_1 <= Float64(-Inf))
          		tmp = Float64(fma(Float64(Float64(x * x) * x), -0.16666666666666666, x) * t_0);
          	elseif (t_1 <= 20000.0)
          		tmp = Float64(sin(x) * fma(Float64(0.16666666666666666 * y), y, 1.0));
          	else
          		tmp = Float64(fma(Float64(fma(Float64(x * x), 0.008333333333333333, -0.16666666666666666) * Float64(x * x)), x, x) * t_0);
          	end
          	return tmp
          end
          
          code[x_, y_] := Block[{t$95$0 = N[(N[(y * y), $MachinePrecision] * 0.16666666666666666 + 1.0), $MachinePrecision]}, Block[{t$95$1 = N[(N[Sin[x], $MachinePrecision] * N[(N[Sinh[y], $MachinePrecision] / y), $MachinePrecision]), $MachinePrecision]}, If[LessEqual[t$95$1, (-Infinity)], N[(N[(N[(N[(x * x), $MachinePrecision] * x), $MachinePrecision] * -0.16666666666666666 + x), $MachinePrecision] * t$95$0), $MachinePrecision], If[LessEqual[t$95$1, 20000.0], N[(N[Sin[x], $MachinePrecision] * N[(N[(0.16666666666666666 * y), $MachinePrecision] * y + 1.0), $MachinePrecision]), $MachinePrecision], N[(N[(N[(N[(N[(x * x), $MachinePrecision] * 0.008333333333333333 + -0.16666666666666666), $MachinePrecision] * N[(x * x), $MachinePrecision]), $MachinePrecision] * x + x), $MachinePrecision] * t$95$0), $MachinePrecision]]]]]
          
          \begin{array}{l}
          
          \\
          \begin{array}{l}
          t_0 := \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right)\\
          t_1 := \sin x \cdot \frac{\sinh y}{y}\\
          \mathbf{if}\;t\_1 \leq -\infty:\\
          \;\;\;\;\mathsf{fma}\left(\left(x \cdot x\right) \cdot x, -0.16666666666666666, x\right) \cdot t\_0\\
          
          \mathbf{elif}\;t\_1 \leq 20000:\\
          \;\;\;\;\sin x \cdot \mathsf{fma}\left(0.16666666666666666 \cdot y, y, 1\right)\\
          
          \mathbf{else}:\\
          \;\;\;\;\mathsf{fma}\left(\mathsf{fma}\left(x \cdot x, 0.008333333333333333, -0.16666666666666666\right) \cdot \left(x \cdot x\right), x, x\right) \cdot t\_0\\
          
          
          \end{array}
          \end{array}
          
          Derivation
          1. Split input into 3 regimes
          2. if (*.f64 (sin.f64 x) (/.f64 (sinh.f64 y) y)) < -inf.0

            1. Initial program 100.0%

              \[\sin x \cdot \frac{\sinh y}{y} \]
            2. Add Preprocessing
            3. Taylor expanded in y around 0

              \[\leadsto \sin x \cdot \color{blue}{\left(1 + \frac{1}{6} \cdot {y}^{2}\right)} \]
            4. Step-by-step derivation
              1. +-commutativeN/A

                \[\leadsto \sin x \cdot \color{blue}{\left(\frac{1}{6} \cdot {y}^{2} + 1\right)} \]
              2. *-commutativeN/A

                \[\leadsto \sin x \cdot \left(\color{blue}{{y}^{2} \cdot \frac{1}{6}} + 1\right) \]
              3. lower-fma.f64N/A

                \[\leadsto \sin x \cdot \color{blue}{\mathsf{fma}\left({y}^{2}, \frac{1}{6}, 1\right)} \]
              4. unpow2N/A

                \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{y \cdot y}, \frac{1}{6}, 1\right) \]
              5. lower-*.f6448.8

                \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{y \cdot y}, 0.16666666666666666, 1\right) \]
            5. Applied rewrites48.8%

              \[\leadsto \sin x \cdot \color{blue}{\mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right)} \]
            6. Taylor expanded in x around 0

              \[\leadsto \color{blue}{\left(x \cdot \left(1 + \frac{-1}{6} \cdot {x}^{2}\right)\right)} \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
            7. Step-by-step derivation
              1. +-commutativeN/A

                \[\leadsto \left(x \cdot \color{blue}{\left(\frac{-1}{6} \cdot {x}^{2} + 1\right)}\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
              2. distribute-lft-inN/A

                \[\leadsto \color{blue}{\left(x \cdot \left(\frac{-1}{6} \cdot {x}^{2}\right) + x \cdot 1\right)} \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
              3. *-commutativeN/A

                \[\leadsto \left(x \cdot \color{blue}{\left({x}^{2} \cdot \frac{-1}{6}\right)} + x \cdot 1\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
              4. associate-*r*N/A

                \[\leadsto \left(\color{blue}{\left(x \cdot {x}^{2}\right) \cdot \frac{-1}{6}} + x \cdot 1\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
              5. *-rgt-identityN/A

                \[\leadsto \left(\left(x \cdot {x}^{2}\right) \cdot \frac{-1}{6} + \color{blue}{x}\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
              6. lower-fma.f64N/A

                \[\leadsto \color{blue}{\mathsf{fma}\left(x \cdot {x}^{2}, \frac{-1}{6}, x\right)} \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
              7. unpow2N/A

                \[\leadsto \mathsf{fma}\left(x \cdot \color{blue}{\left(x \cdot x\right)}, \frac{-1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
              8. cube-unmultN/A

                \[\leadsto \mathsf{fma}\left(\color{blue}{{x}^{3}}, \frac{-1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
              9. metadata-evalN/A

                \[\leadsto \mathsf{fma}\left({x}^{\color{blue}{\left(2 + 1\right)}}, \frac{-1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
              10. lower-pow.f64N/A

                \[\leadsto \mathsf{fma}\left(\color{blue}{{x}^{\left(2 + 1\right)}}, \frac{-1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
              11. metadata-eval47.7

                \[\leadsto \mathsf{fma}\left({x}^{\color{blue}{3}}, -0.16666666666666666, x\right) \cdot \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right) \]
            8. Applied rewrites47.7%

              \[\leadsto \color{blue}{\mathsf{fma}\left({x}^{3}, -0.16666666666666666, x\right)} \cdot \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right) \]
            9. Step-by-step derivation
              1. Applied rewrites47.7%

                \[\leadsto \mathsf{fma}\left(\left(x \cdot x\right) \cdot x, -0.16666666666666666, x\right) \cdot \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right) \]

              if -inf.0 < (*.f64 (sin.f64 x) (/.f64 (sinh.f64 y) y)) < 2e4

              1. Initial program 100.0%

                \[\sin x \cdot \frac{\sinh y}{y} \]
              2. Add Preprocessing
              3. Taylor expanded in y around 0

                \[\leadsto \sin x \cdot \color{blue}{\left(1 + \frac{1}{6} \cdot {y}^{2}\right)} \]
              4. Step-by-step derivation
                1. +-commutativeN/A

                  \[\leadsto \sin x \cdot \color{blue}{\left(\frac{1}{6} \cdot {y}^{2} + 1\right)} \]
                2. *-commutativeN/A

                  \[\leadsto \sin x \cdot \left(\color{blue}{{y}^{2} \cdot \frac{1}{6}} + 1\right) \]
                3. lower-fma.f64N/A

                  \[\leadsto \sin x \cdot \color{blue}{\mathsf{fma}\left({y}^{2}, \frac{1}{6}, 1\right)} \]
                4. unpow2N/A

                  \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{y \cdot y}, \frac{1}{6}, 1\right) \]
                5. lower-*.f6497.6

                  \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{y \cdot y}, 0.16666666666666666, 1\right) \]
              5. Applied rewrites97.6%

                \[\leadsto \sin x \cdot \color{blue}{\mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right)} \]
              6. Step-by-step derivation
                1. Applied rewrites97.6%

                  \[\leadsto \sin x \cdot \mathsf{fma}\left(0.16666666666666666 \cdot y, \color{blue}{y}, 1\right) \]

                if 2e4 < (*.f64 (sin.f64 x) (/.f64 (sinh.f64 y) y))

                1. Initial program 100.0%

                  \[\sin x \cdot \frac{\sinh y}{y} \]
                2. Add Preprocessing
                3. Taylor expanded in y around 0

                  \[\leadsto \sin x \cdot \color{blue}{\left(1 + \frac{1}{6} \cdot {y}^{2}\right)} \]
                4. Step-by-step derivation
                  1. +-commutativeN/A

                    \[\leadsto \sin x \cdot \color{blue}{\left(\frac{1}{6} \cdot {y}^{2} + 1\right)} \]
                  2. *-commutativeN/A

                    \[\leadsto \sin x \cdot \left(\color{blue}{{y}^{2} \cdot \frac{1}{6}} + 1\right) \]
                  3. lower-fma.f64N/A

                    \[\leadsto \sin x \cdot \color{blue}{\mathsf{fma}\left({y}^{2}, \frac{1}{6}, 1\right)} \]
                  4. unpow2N/A

                    \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{y \cdot y}, \frac{1}{6}, 1\right) \]
                  5. lower-*.f6457.6

                    \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{y \cdot y}, 0.16666666666666666, 1\right) \]
                5. Applied rewrites57.6%

                  \[\leadsto \sin x \cdot \color{blue}{\mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right)} \]
                6. Taylor expanded in x around 0

                  \[\leadsto \color{blue}{\left(x \cdot \left(1 + {x}^{2} \cdot \left(\frac{1}{120} \cdot {x}^{2} - \frac{1}{6}\right)\right)\right)} \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                7. Step-by-step derivation
                  1. +-commutativeN/A

                    \[\leadsto \left(x \cdot \color{blue}{\left({x}^{2} \cdot \left(\frac{1}{120} \cdot {x}^{2} - \frac{1}{6}\right) + 1\right)}\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                  2. distribute-lft-inN/A

                    \[\leadsto \color{blue}{\left(x \cdot \left({x}^{2} \cdot \left(\frac{1}{120} \cdot {x}^{2} - \frac{1}{6}\right)\right) + x \cdot 1\right)} \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                  3. associate-*r*N/A

                    \[\leadsto \left(\color{blue}{\left(x \cdot {x}^{2}\right) \cdot \left(\frac{1}{120} \cdot {x}^{2} - \frac{1}{6}\right)} + x \cdot 1\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                  4. *-rgt-identityN/A

                    \[\leadsto \left(\left(x \cdot {x}^{2}\right) \cdot \left(\frac{1}{120} \cdot {x}^{2} - \frac{1}{6}\right) + \color{blue}{x}\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                  5. lower-fma.f64N/A

                    \[\leadsto \color{blue}{\mathsf{fma}\left(x \cdot {x}^{2}, \frac{1}{120} \cdot {x}^{2} - \frac{1}{6}, x\right)} \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                  6. unpow2N/A

                    \[\leadsto \mathsf{fma}\left(x \cdot \color{blue}{\left(x \cdot x\right)}, \frac{1}{120} \cdot {x}^{2} - \frac{1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                  7. cube-unmultN/A

                    \[\leadsto \mathsf{fma}\left(\color{blue}{{x}^{3}}, \frac{1}{120} \cdot {x}^{2} - \frac{1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                  8. metadata-evalN/A

                    \[\leadsto \mathsf{fma}\left({x}^{\color{blue}{\left(2 + 1\right)}}, \frac{1}{120} \cdot {x}^{2} - \frac{1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                  9. lower-pow.f64N/A

                    \[\leadsto \mathsf{fma}\left(\color{blue}{{x}^{\left(2 + 1\right)}}, \frac{1}{120} \cdot {x}^{2} - \frac{1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                  10. metadata-evalN/A

                    \[\leadsto \mathsf{fma}\left({x}^{\color{blue}{3}}, \frac{1}{120} \cdot {x}^{2} - \frac{1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                  11. metadata-evalN/A

                    \[\leadsto \mathsf{fma}\left({x}^{3}, \frac{1}{120} \cdot {x}^{2} - \color{blue}{\frac{1}{6} \cdot 1}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                  12. fp-cancel-sub-sign-invN/A

                    \[\leadsto \mathsf{fma}\left({x}^{3}, \color{blue}{\frac{1}{120} \cdot {x}^{2} + \left(\mathsf{neg}\left(\frac{1}{6}\right)\right) \cdot 1}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                  13. metadata-evalN/A

                    \[\leadsto \mathsf{fma}\left({x}^{3}, \frac{1}{120} \cdot {x}^{2} + \color{blue}{\frac{-1}{6}} \cdot 1, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                  14. metadata-evalN/A

                    \[\leadsto \mathsf{fma}\left({x}^{3}, \frac{1}{120} \cdot {x}^{2} + \color{blue}{\frac{-1}{6}}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                  15. lower-fma.f64N/A

                    \[\leadsto \mathsf{fma}\left({x}^{3}, \color{blue}{\mathsf{fma}\left(\frac{1}{120}, {x}^{2}, \frac{-1}{6}\right)}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                  16. unpow2N/A

                    \[\leadsto \mathsf{fma}\left({x}^{3}, \mathsf{fma}\left(\frac{1}{120}, \color{blue}{x \cdot x}, \frac{-1}{6}\right), x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                  17. lower-*.f6451.5

                    \[\leadsto \mathsf{fma}\left({x}^{3}, \mathsf{fma}\left(0.008333333333333333, \color{blue}{x \cdot x}, -0.16666666666666666\right), x\right) \cdot \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right) \]
                8. Applied rewrites51.5%

                  \[\leadsto \color{blue}{\mathsf{fma}\left({x}^{3}, \mathsf{fma}\left(0.008333333333333333, x \cdot x, -0.16666666666666666\right), x\right)} \cdot \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right) \]
                9. Step-by-step derivation
                  1. Applied rewrites51.5%

                    \[\leadsto \mathsf{fma}\left(\mathsf{fma}\left(x \cdot x, 0.008333333333333333, -0.16666666666666666\right) \cdot \left(x \cdot x\right), \color{blue}{x}, x\right) \cdot \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right) \]
                10. Recombined 3 regimes into one program.
                11. Add Preprocessing

                Alternative 4: 74.1% accurate, 0.4× speedup?

                \[\begin{array}{l} \\ \begin{array}{l} t_0 := \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right)\\ t_1 := \sin x \cdot \frac{\sinh y}{y}\\ \mathbf{if}\;t\_1 \leq -\infty:\\ \;\;\;\;\mathsf{fma}\left(\left(x \cdot x\right) \cdot x, -0.16666666666666666, x\right) \cdot t\_0\\ \mathbf{elif}\;t\_1 \leq 1:\\ \;\;\;\;\sin x \cdot 1\\ \mathbf{else}:\\ \;\;\;\;\mathsf{fma}\left(\mathsf{fma}\left(x \cdot x, 0.008333333333333333, -0.16666666666666666\right) \cdot \left(x \cdot x\right), x, x\right) \cdot t\_0\\ \end{array} \end{array} \]
                (FPCore (x y)
                 :precision binary64
                 (let* ((t_0 (fma (* y y) 0.16666666666666666 1.0))
                        (t_1 (* (sin x) (/ (sinh y) y))))
                   (if (<= t_1 (- INFINITY))
                     (* (fma (* (* x x) x) -0.16666666666666666 x) t_0)
                     (if (<= t_1 1.0)
                       (* (sin x) 1.0)
                       (*
                        (fma
                         (* (fma (* x x) 0.008333333333333333 -0.16666666666666666) (* x x))
                         x
                         x)
                        t_0)))))
                double code(double x, double y) {
                	double t_0 = fma((y * y), 0.16666666666666666, 1.0);
                	double t_1 = sin(x) * (sinh(y) / y);
                	double tmp;
                	if (t_1 <= -((double) INFINITY)) {
                		tmp = fma(((x * x) * x), -0.16666666666666666, x) * t_0;
                	} else if (t_1 <= 1.0) {
                		tmp = sin(x) * 1.0;
                	} else {
                		tmp = fma((fma((x * x), 0.008333333333333333, -0.16666666666666666) * (x * x)), x, x) * t_0;
                	}
                	return tmp;
                }
                
                function code(x, y)
                	t_0 = fma(Float64(y * y), 0.16666666666666666, 1.0)
                	t_1 = Float64(sin(x) * Float64(sinh(y) / y))
                	tmp = 0.0
                	if (t_1 <= Float64(-Inf))
                		tmp = Float64(fma(Float64(Float64(x * x) * x), -0.16666666666666666, x) * t_0);
                	elseif (t_1 <= 1.0)
                		tmp = Float64(sin(x) * 1.0);
                	else
                		tmp = Float64(fma(Float64(fma(Float64(x * x), 0.008333333333333333, -0.16666666666666666) * Float64(x * x)), x, x) * t_0);
                	end
                	return tmp
                end
                
                code[x_, y_] := Block[{t$95$0 = N[(N[(y * y), $MachinePrecision] * 0.16666666666666666 + 1.0), $MachinePrecision]}, Block[{t$95$1 = N[(N[Sin[x], $MachinePrecision] * N[(N[Sinh[y], $MachinePrecision] / y), $MachinePrecision]), $MachinePrecision]}, If[LessEqual[t$95$1, (-Infinity)], N[(N[(N[(N[(x * x), $MachinePrecision] * x), $MachinePrecision] * -0.16666666666666666 + x), $MachinePrecision] * t$95$0), $MachinePrecision], If[LessEqual[t$95$1, 1.0], N[(N[Sin[x], $MachinePrecision] * 1.0), $MachinePrecision], N[(N[(N[(N[(N[(x * x), $MachinePrecision] * 0.008333333333333333 + -0.16666666666666666), $MachinePrecision] * N[(x * x), $MachinePrecision]), $MachinePrecision] * x + x), $MachinePrecision] * t$95$0), $MachinePrecision]]]]]
                
                \begin{array}{l}
                
                \\
                \begin{array}{l}
                t_0 := \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right)\\
                t_1 := \sin x \cdot \frac{\sinh y}{y}\\
                \mathbf{if}\;t\_1 \leq -\infty:\\
                \;\;\;\;\mathsf{fma}\left(\left(x \cdot x\right) \cdot x, -0.16666666666666666, x\right) \cdot t\_0\\
                
                \mathbf{elif}\;t\_1 \leq 1:\\
                \;\;\;\;\sin x \cdot 1\\
                
                \mathbf{else}:\\
                \;\;\;\;\mathsf{fma}\left(\mathsf{fma}\left(x \cdot x, 0.008333333333333333, -0.16666666666666666\right) \cdot \left(x \cdot x\right), x, x\right) \cdot t\_0\\
                
                
                \end{array}
                \end{array}
                
                Derivation
                1. Split input into 3 regimes
                2. if (*.f64 (sin.f64 x) (/.f64 (sinh.f64 y) y)) < -inf.0

                  1. Initial program 100.0%

                    \[\sin x \cdot \frac{\sinh y}{y} \]
                  2. Add Preprocessing
                  3. Taylor expanded in y around 0

                    \[\leadsto \sin x \cdot \color{blue}{\left(1 + \frac{1}{6} \cdot {y}^{2}\right)} \]
                  4. Step-by-step derivation
                    1. +-commutativeN/A

                      \[\leadsto \sin x \cdot \color{blue}{\left(\frac{1}{6} \cdot {y}^{2} + 1\right)} \]
                    2. *-commutativeN/A

                      \[\leadsto \sin x \cdot \left(\color{blue}{{y}^{2} \cdot \frac{1}{6}} + 1\right) \]
                    3. lower-fma.f64N/A

                      \[\leadsto \sin x \cdot \color{blue}{\mathsf{fma}\left({y}^{2}, \frac{1}{6}, 1\right)} \]
                    4. unpow2N/A

                      \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{y \cdot y}, \frac{1}{6}, 1\right) \]
                    5. lower-*.f6448.8

                      \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{y \cdot y}, 0.16666666666666666, 1\right) \]
                  5. Applied rewrites48.8%

                    \[\leadsto \sin x \cdot \color{blue}{\mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right)} \]
                  6. Taylor expanded in x around 0

                    \[\leadsto \color{blue}{\left(x \cdot \left(1 + \frac{-1}{6} \cdot {x}^{2}\right)\right)} \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                  7. Step-by-step derivation
                    1. +-commutativeN/A

                      \[\leadsto \left(x \cdot \color{blue}{\left(\frac{-1}{6} \cdot {x}^{2} + 1\right)}\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                    2. distribute-lft-inN/A

                      \[\leadsto \color{blue}{\left(x \cdot \left(\frac{-1}{6} \cdot {x}^{2}\right) + x \cdot 1\right)} \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                    3. *-commutativeN/A

                      \[\leadsto \left(x \cdot \color{blue}{\left({x}^{2} \cdot \frac{-1}{6}\right)} + x \cdot 1\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                    4. associate-*r*N/A

                      \[\leadsto \left(\color{blue}{\left(x \cdot {x}^{2}\right) \cdot \frac{-1}{6}} + x \cdot 1\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                    5. *-rgt-identityN/A

                      \[\leadsto \left(\left(x \cdot {x}^{2}\right) \cdot \frac{-1}{6} + \color{blue}{x}\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                    6. lower-fma.f64N/A

                      \[\leadsto \color{blue}{\mathsf{fma}\left(x \cdot {x}^{2}, \frac{-1}{6}, x\right)} \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                    7. unpow2N/A

                      \[\leadsto \mathsf{fma}\left(x \cdot \color{blue}{\left(x \cdot x\right)}, \frac{-1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                    8. cube-unmultN/A

                      \[\leadsto \mathsf{fma}\left(\color{blue}{{x}^{3}}, \frac{-1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                    9. metadata-evalN/A

                      \[\leadsto \mathsf{fma}\left({x}^{\color{blue}{\left(2 + 1\right)}}, \frac{-1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                    10. lower-pow.f64N/A

                      \[\leadsto \mathsf{fma}\left(\color{blue}{{x}^{\left(2 + 1\right)}}, \frac{-1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                    11. metadata-eval47.7

                      \[\leadsto \mathsf{fma}\left({x}^{\color{blue}{3}}, -0.16666666666666666, x\right) \cdot \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right) \]
                  8. Applied rewrites47.7%

                    \[\leadsto \color{blue}{\mathsf{fma}\left({x}^{3}, -0.16666666666666666, x\right)} \cdot \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right) \]
                  9. Step-by-step derivation
                    1. Applied rewrites47.7%

                      \[\leadsto \mathsf{fma}\left(\left(x \cdot x\right) \cdot x, -0.16666666666666666, x\right) \cdot \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right) \]

                    if -inf.0 < (*.f64 (sin.f64 x) (/.f64 (sinh.f64 y) y)) < 1

                    1. Initial program 100.0%

                      \[\sin x \cdot \frac{\sinh y}{y} \]
                    2. Add Preprocessing
                    3. Taylor expanded in y around 0

                      \[\leadsto \sin x \cdot \color{blue}{1} \]
                    4. Step-by-step derivation
                      1. Applied rewrites96.7%

                        \[\leadsto \sin x \cdot \color{blue}{1} \]

                      if 1 < (*.f64 (sin.f64 x) (/.f64 (sinh.f64 y) y))

                      1. Initial program 100.0%

                        \[\sin x \cdot \frac{\sinh y}{y} \]
                      2. Add Preprocessing
                      3. Taylor expanded in y around 0

                        \[\leadsto \sin x \cdot \color{blue}{\left(1 + \frac{1}{6} \cdot {y}^{2}\right)} \]
                      4. Step-by-step derivation
                        1. +-commutativeN/A

                          \[\leadsto \sin x \cdot \color{blue}{\left(\frac{1}{6} \cdot {y}^{2} + 1\right)} \]
                        2. *-commutativeN/A

                          \[\leadsto \sin x \cdot \left(\color{blue}{{y}^{2} \cdot \frac{1}{6}} + 1\right) \]
                        3. lower-fma.f64N/A

                          \[\leadsto \sin x \cdot \color{blue}{\mathsf{fma}\left({y}^{2}, \frac{1}{6}, 1\right)} \]
                        4. unpow2N/A

                          \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{y \cdot y}, \frac{1}{6}, 1\right) \]
                        5. lower-*.f6457.6

                          \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{y \cdot y}, 0.16666666666666666, 1\right) \]
                      5. Applied rewrites57.6%

                        \[\leadsto \sin x \cdot \color{blue}{\mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right)} \]
                      6. Taylor expanded in x around 0

                        \[\leadsto \color{blue}{\left(x \cdot \left(1 + {x}^{2} \cdot \left(\frac{1}{120} \cdot {x}^{2} - \frac{1}{6}\right)\right)\right)} \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                      7. Step-by-step derivation
                        1. +-commutativeN/A

                          \[\leadsto \left(x \cdot \color{blue}{\left({x}^{2} \cdot \left(\frac{1}{120} \cdot {x}^{2} - \frac{1}{6}\right) + 1\right)}\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                        2. distribute-lft-inN/A

                          \[\leadsto \color{blue}{\left(x \cdot \left({x}^{2} \cdot \left(\frac{1}{120} \cdot {x}^{2} - \frac{1}{6}\right)\right) + x \cdot 1\right)} \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                        3. associate-*r*N/A

                          \[\leadsto \left(\color{blue}{\left(x \cdot {x}^{2}\right) \cdot \left(\frac{1}{120} \cdot {x}^{2} - \frac{1}{6}\right)} + x \cdot 1\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                        4. *-rgt-identityN/A

                          \[\leadsto \left(\left(x \cdot {x}^{2}\right) \cdot \left(\frac{1}{120} \cdot {x}^{2} - \frac{1}{6}\right) + \color{blue}{x}\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                        5. lower-fma.f64N/A

                          \[\leadsto \color{blue}{\mathsf{fma}\left(x \cdot {x}^{2}, \frac{1}{120} \cdot {x}^{2} - \frac{1}{6}, x\right)} \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                        6. unpow2N/A

                          \[\leadsto \mathsf{fma}\left(x \cdot \color{blue}{\left(x \cdot x\right)}, \frac{1}{120} \cdot {x}^{2} - \frac{1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                        7. cube-unmultN/A

                          \[\leadsto \mathsf{fma}\left(\color{blue}{{x}^{3}}, \frac{1}{120} \cdot {x}^{2} - \frac{1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                        8. metadata-evalN/A

                          \[\leadsto \mathsf{fma}\left({x}^{\color{blue}{\left(2 + 1\right)}}, \frac{1}{120} \cdot {x}^{2} - \frac{1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                        9. lower-pow.f64N/A

                          \[\leadsto \mathsf{fma}\left(\color{blue}{{x}^{\left(2 + 1\right)}}, \frac{1}{120} \cdot {x}^{2} - \frac{1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                        10. metadata-evalN/A

                          \[\leadsto \mathsf{fma}\left({x}^{\color{blue}{3}}, \frac{1}{120} \cdot {x}^{2} - \frac{1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                        11. metadata-evalN/A

                          \[\leadsto \mathsf{fma}\left({x}^{3}, \frac{1}{120} \cdot {x}^{2} - \color{blue}{\frac{1}{6} \cdot 1}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                        12. fp-cancel-sub-sign-invN/A

                          \[\leadsto \mathsf{fma}\left({x}^{3}, \color{blue}{\frac{1}{120} \cdot {x}^{2} + \left(\mathsf{neg}\left(\frac{1}{6}\right)\right) \cdot 1}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                        13. metadata-evalN/A

                          \[\leadsto \mathsf{fma}\left({x}^{3}, \frac{1}{120} \cdot {x}^{2} + \color{blue}{\frac{-1}{6}} \cdot 1, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                        14. metadata-evalN/A

                          \[\leadsto \mathsf{fma}\left({x}^{3}, \frac{1}{120} \cdot {x}^{2} + \color{blue}{\frac{-1}{6}}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                        15. lower-fma.f64N/A

                          \[\leadsto \mathsf{fma}\left({x}^{3}, \color{blue}{\mathsf{fma}\left(\frac{1}{120}, {x}^{2}, \frac{-1}{6}\right)}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                        16. unpow2N/A

                          \[\leadsto \mathsf{fma}\left({x}^{3}, \mathsf{fma}\left(\frac{1}{120}, \color{blue}{x \cdot x}, \frac{-1}{6}\right), x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                        17. lower-*.f6451.5

                          \[\leadsto \mathsf{fma}\left({x}^{3}, \mathsf{fma}\left(0.008333333333333333, \color{blue}{x \cdot x}, -0.16666666666666666\right), x\right) \cdot \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right) \]
                      8. Applied rewrites51.5%

                        \[\leadsto \color{blue}{\mathsf{fma}\left({x}^{3}, \mathsf{fma}\left(0.008333333333333333, x \cdot x, -0.16666666666666666\right), x\right)} \cdot \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right) \]
                      9. Step-by-step derivation
                        1. Applied rewrites51.5%

                          \[\leadsto \mathsf{fma}\left(\mathsf{fma}\left(x \cdot x, 0.008333333333333333, -0.16666666666666666\right) \cdot \left(x \cdot x\right), \color{blue}{x}, x\right) \cdot \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right) \]
                      10. Recombined 3 regimes into one program.
                      11. Add Preprocessing

                      Alternative 5: 74.3% accurate, 0.6× speedup?

                      \[\begin{array}{l} \\ \begin{array}{l} \mathbf{if}\;\sin x \cdot \frac{\sinh y}{y} \leq -\infty:\\ \;\;\;\;\mathsf{fma}\left({x}^{3}, \mathsf{fma}\left(\mathsf{fma}\left(-0.0001984126984126984, x \cdot x, 0.008333333333333333\right), x \cdot x, -0.16666666666666666\right), x\right) \cdot 1\\ \mathbf{else}:\\ \;\;\;\;\sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(0.008333333333333333, y \cdot y, 0.16666666666666666\right) \cdot y, y, 1\right)\\ \end{array} \end{array} \]
                      (FPCore (x y)
                       :precision binary64
                       (if (<= (* (sin x) (/ (sinh y) y)) (- INFINITY))
                         (*
                          (fma
                           (pow x 3.0)
                           (fma
                            (fma -0.0001984126984126984 (* x x) 0.008333333333333333)
                            (* x x)
                            -0.16666666666666666)
                           x)
                          1.0)
                         (*
                          (sin x)
                          (fma (* (fma 0.008333333333333333 (* y y) 0.16666666666666666) y) y 1.0))))
                      double code(double x, double y) {
                      	double tmp;
                      	if ((sin(x) * (sinh(y) / y)) <= -((double) INFINITY)) {
                      		tmp = fma(pow(x, 3.0), fma(fma(-0.0001984126984126984, (x * x), 0.008333333333333333), (x * x), -0.16666666666666666), x) * 1.0;
                      	} else {
                      		tmp = sin(x) * fma((fma(0.008333333333333333, (y * y), 0.16666666666666666) * y), y, 1.0);
                      	}
                      	return tmp;
                      }
                      
                      function code(x, y)
                      	tmp = 0.0
                      	if (Float64(sin(x) * Float64(sinh(y) / y)) <= Float64(-Inf))
                      		tmp = Float64(fma((x ^ 3.0), fma(fma(-0.0001984126984126984, Float64(x * x), 0.008333333333333333), Float64(x * x), -0.16666666666666666), x) * 1.0);
                      	else
                      		tmp = Float64(sin(x) * fma(Float64(fma(0.008333333333333333, Float64(y * y), 0.16666666666666666) * y), y, 1.0));
                      	end
                      	return tmp
                      end
                      
                      code[x_, y_] := If[LessEqual[N[(N[Sin[x], $MachinePrecision] * N[(N[Sinh[y], $MachinePrecision] / y), $MachinePrecision]), $MachinePrecision], (-Infinity)], N[(N[(N[Power[x, 3.0], $MachinePrecision] * N[(N[(-0.0001984126984126984 * N[(x * x), $MachinePrecision] + 0.008333333333333333), $MachinePrecision] * N[(x * x), $MachinePrecision] + -0.16666666666666666), $MachinePrecision] + x), $MachinePrecision] * 1.0), $MachinePrecision], N[(N[Sin[x], $MachinePrecision] * N[(N[(N[(0.008333333333333333 * N[(y * y), $MachinePrecision] + 0.16666666666666666), $MachinePrecision] * y), $MachinePrecision] * y + 1.0), $MachinePrecision]), $MachinePrecision]]
                      
                      \begin{array}{l}
                      
                      \\
                      \begin{array}{l}
                      \mathbf{if}\;\sin x \cdot \frac{\sinh y}{y} \leq -\infty:\\
                      \;\;\;\;\mathsf{fma}\left({x}^{3}, \mathsf{fma}\left(\mathsf{fma}\left(-0.0001984126984126984, x \cdot x, 0.008333333333333333\right), x \cdot x, -0.16666666666666666\right), x\right) \cdot 1\\
                      
                      \mathbf{else}:\\
                      \;\;\;\;\sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(0.008333333333333333, y \cdot y, 0.16666666666666666\right) \cdot y, y, 1\right)\\
                      
                      
                      \end{array}
                      \end{array}
                      
                      Derivation
                      1. Split input into 2 regimes
                      2. if (*.f64 (sin.f64 x) (/.f64 (sinh.f64 y) y)) < -inf.0

                        1. Initial program 100.0%

                          \[\sin x \cdot \frac{\sinh y}{y} \]
                        2. Add Preprocessing
                        3. Taylor expanded in y around 0

                          \[\leadsto \sin x \cdot \color{blue}{1} \]
                        4. Step-by-step derivation
                          1. Applied rewrites2.7%

                            \[\leadsto \sin x \cdot \color{blue}{1} \]
                          2. Taylor expanded in x around 0

                            \[\leadsto \color{blue}{\left(x \cdot \left(1 + {x}^{2} \cdot \left({x}^{2} \cdot \left(\frac{1}{120} + \frac{-1}{5040} \cdot {x}^{2}\right) - \frac{1}{6}\right)\right)\right)} \cdot 1 \]
                          3. Step-by-step derivation
                            1. +-commutativeN/A

                              \[\leadsto \left(x \cdot \color{blue}{\left({x}^{2} \cdot \left({x}^{2} \cdot \left(\frac{1}{120} + \frac{-1}{5040} \cdot {x}^{2}\right) - \frac{1}{6}\right) + 1\right)}\right) \cdot 1 \]
                            2. distribute-lft-inN/A

                              \[\leadsto \color{blue}{\left(x \cdot \left({x}^{2} \cdot \left({x}^{2} \cdot \left(\frac{1}{120} + \frac{-1}{5040} \cdot {x}^{2}\right) - \frac{1}{6}\right)\right) + x \cdot 1\right)} \cdot 1 \]
                            3. associate-*r*N/A

                              \[\leadsto \left(\color{blue}{\left(x \cdot {x}^{2}\right) \cdot \left({x}^{2} \cdot \left(\frac{1}{120} + \frac{-1}{5040} \cdot {x}^{2}\right) - \frac{1}{6}\right)} + x \cdot 1\right) \cdot 1 \]
                            4. *-rgt-identityN/A

                              \[\leadsto \left(\left(x \cdot {x}^{2}\right) \cdot \left({x}^{2} \cdot \left(\frac{1}{120} + \frac{-1}{5040} \cdot {x}^{2}\right) - \frac{1}{6}\right) + \color{blue}{x}\right) \cdot 1 \]
                            5. lower-fma.f64N/A

                              \[\leadsto \color{blue}{\mathsf{fma}\left(x \cdot {x}^{2}, {x}^{2} \cdot \left(\frac{1}{120} + \frac{-1}{5040} \cdot {x}^{2}\right) - \frac{1}{6}, x\right)} \cdot 1 \]
                          4. Applied rewrites22.4%

                            \[\leadsto \color{blue}{\mathsf{fma}\left({x}^{3}, \mathsf{fma}\left(\mathsf{fma}\left(-0.0001984126984126984, x \cdot x, 0.008333333333333333\right), x \cdot x, -0.16666666666666666\right), x\right)} \cdot 1 \]

                          if -inf.0 < (*.f64 (sin.f64 x) (/.f64 (sinh.f64 y) y))

                          1. Initial program 100.0%

                            \[\sin x \cdot \frac{\sinh y}{y} \]
                          2. Add Preprocessing
                          3. Taylor expanded in y around 0

                            \[\leadsto \sin x \cdot \color{blue}{\left(1 + {y}^{2} \cdot \left(\frac{1}{6} + \frac{1}{120} \cdot {y}^{2}\right)\right)} \]
                          4. Step-by-step derivation
                            1. fp-cancel-sign-sub-invN/A

                              \[\leadsto \sin x \cdot \color{blue}{\left(1 - \left(\mathsf{neg}\left({y}^{2}\right)\right) \cdot \left(\frac{1}{6} + \frac{1}{120} \cdot {y}^{2}\right)\right)} \]
                            2. fp-cancel-sub-sign-invN/A

                              \[\leadsto \sin x \cdot \color{blue}{\left(1 + \left(\mathsf{neg}\left(\left(\mathsf{neg}\left({y}^{2}\right)\right)\right)\right) \cdot \left(\frac{1}{6} + \frac{1}{120} \cdot {y}^{2}\right)\right)} \]
                            3. +-commutativeN/A

                              \[\leadsto \sin x \cdot \color{blue}{\left(\left(\mathsf{neg}\left(\left(\mathsf{neg}\left({y}^{2}\right)\right)\right)\right) \cdot \left(\frac{1}{6} + \frac{1}{120} \cdot {y}^{2}\right) + 1\right)} \]
                            4. remove-double-negN/A

                              \[\leadsto \sin x \cdot \left(\color{blue}{{y}^{2}} \cdot \left(\frac{1}{6} + \frac{1}{120} \cdot {y}^{2}\right) + 1\right) \]
                            5. *-commutativeN/A

                              \[\leadsto \sin x \cdot \left(\color{blue}{\left(\frac{1}{6} + \frac{1}{120} \cdot {y}^{2}\right) \cdot {y}^{2}} + 1\right) \]
                            6. lower-fma.f64N/A

                              \[\leadsto \sin x \cdot \color{blue}{\mathsf{fma}\left(\frac{1}{6} + \frac{1}{120} \cdot {y}^{2}, {y}^{2}, 1\right)} \]
                            7. +-commutativeN/A

                              \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{\frac{1}{120} \cdot {y}^{2} + \frac{1}{6}}, {y}^{2}, 1\right) \]
                            8. *-commutativeN/A

                              \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{{y}^{2} \cdot \frac{1}{120}} + \frac{1}{6}, {y}^{2}, 1\right) \]
                            9. lower-fma.f64N/A

                              \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{\mathsf{fma}\left({y}^{2}, \frac{1}{120}, \frac{1}{6}\right)}, {y}^{2}, 1\right) \]
                            10. unpow2N/A

                              \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(\color{blue}{y \cdot y}, \frac{1}{120}, \frac{1}{6}\right), {y}^{2}, 1\right) \]
                            11. lower-*.f64N/A

                              \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(\color{blue}{y \cdot y}, \frac{1}{120}, \frac{1}{6}\right), {y}^{2}, 1\right) \]
                            12. unpow2N/A

                              \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(y \cdot y, \frac{1}{120}, \frac{1}{6}\right), \color{blue}{y \cdot y}, 1\right) \]
                            13. lower-*.f6489.9

                              \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(y \cdot y, 0.008333333333333333, 0.16666666666666666\right), \color{blue}{y \cdot y}, 1\right) \]
                          5. Applied rewrites89.9%

                            \[\leadsto \sin x \cdot \color{blue}{\mathsf{fma}\left(\mathsf{fma}\left(y \cdot y, 0.008333333333333333, 0.16666666666666666\right), y \cdot y, 1\right)} \]
                          6. Step-by-step derivation
                            1. Applied rewrites89.9%

                              \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(0.008333333333333333, y \cdot y, 0.16666666666666666\right) \cdot y, \color{blue}{y}, 1\right) \]
                          7. Recombined 2 regimes into one program.
                          8. Add Preprocessing

                          Alternative 6: 80.8% accurate, 0.6× speedup?

                          \[\begin{array}{l} \\ \begin{array}{l} \mathbf{if}\;\sin x \cdot \frac{\sinh y}{y} \leq -\infty:\\ \;\;\;\;\mathsf{fma}\left(\left(x \cdot x\right) \cdot x, -0.16666666666666666, x\right) \cdot \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right)\\ \mathbf{else}:\\ \;\;\;\;\sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(0.008333333333333333, y \cdot y, 0.16666666666666666\right) \cdot y, y, 1\right)\\ \end{array} \end{array} \]
                          (FPCore (x y)
                           :precision binary64
                           (if (<= (* (sin x) (/ (sinh y) y)) (- INFINITY))
                             (*
                              (fma (* (* x x) x) -0.16666666666666666 x)
                              (fma (* y y) 0.16666666666666666 1.0))
                             (*
                              (sin x)
                              (fma (* (fma 0.008333333333333333 (* y y) 0.16666666666666666) y) y 1.0))))
                          double code(double x, double y) {
                          	double tmp;
                          	if ((sin(x) * (sinh(y) / y)) <= -((double) INFINITY)) {
                          		tmp = fma(((x * x) * x), -0.16666666666666666, x) * fma((y * y), 0.16666666666666666, 1.0);
                          	} else {
                          		tmp = sin(x) * fma((fma(0.008333333333333333, (y * y), 0.16666666666666666) * y), y, 1.0);
                          	}
                          	return tmp;
                          }
                          
                          function code(x, y)
                          	tmp = 0.0
                          	if (Float64(sin(x) * Float64(sinh(y) / y)) <= Float64(-Inf))
                          		tmp = Float64(fma(Float64(Float64(x * x) * x), -0.16666666666666666, x) * fma(Float64(y * y), 0.16666666666666666, 1.0));
                          	else
                          		tmp = Float64(sin(x) * fma(Float64(fma(0.008333333333333333, Float64(y * y), 0.16666666666666666) * y), y, 1.0));
                          	end
                          	return tmp
                          end
                          
                          code[x_, y_] := If[LessEqual[N[(N[Sin[x], $MachinePrecision] * N[(N[Sinh[y], $MachinePrecision] / y), $MachinePrecision]), $MachinePrecision], (-Infinity)], N[(N[(N[(N[(x * x), $MachinePrecision] * x), $MachinePrecision] * -0.16666666666666666 + x), $MachinePrecision] * N[(N[(y * y), $MachinePrecision] * 0.16666666666666666 + 1.0), $MachinePrecision]), $MachinePrecision], N[(N[Sin[x], $MachinePrecision] * N[(N[(N[(0.008333333333333333 * N[(y * y), $MachinePrecision] + 0.16666666666666666), $MachinePrecision] * y), $MachinePrecision] * y + 1.0), $MachinePrecision]), $MachinePrecision]]
                          
                          \begin{array}{l}
                          
                          \\
                          \begin{array}{l}
                          \mathbf{if}\;\sin x \cdot \frac{\sinh y}{y} \leq -\infty:\\
                          \;\;\;\;\mathsf{fma}\left(\left(x \cdot x\right) \cdot x, -0.16666666666666666, x\right) \cdot \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right)\\
                          
                          \mathbf{else}:\\
                          \;\;\;\;\sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(0.008333333333333333, y \cdot y, 0.16666666666666666\right) \cdot y, y, 1\right)\\
                          
                          
                          \end{array}
                          \end{array}
                          
                          Derivation
                          1. Split input into 2 regimes
                          2. if (*.f64 (sin.f64 x) (/.f64 (sinh.f64 y) y)) < -inf.0

                            1. Initial program 100.0%

                              \[\sin x \cdot \frac{\sinh y}{y} \]
                            2. Add Preprocessing
                            3. Taylor expanded in y around 0

                              \[\leadsto \sin x \cdot \color{blue}{\left(1 + \frac{1}{6} \cdot {y}^{2}\right)} \]
                            4. Step-by-step derivation
                              1. +-commutativeN/A

                                \[\leadsto \sin x \cdot \color{blue}{\left(\frac{1}{6} \cdot {y}^{2} + 1\right)} \]
                              2. *-commutativeN/A

                                \[\leadsto \sin x \cdot \left(\color{blue}{{y}^{2} \cdot \frac{1}{6}} + 1\right) \]
                              3. lower-fma.f64N/A

                                \[\leadsto \sin x \cdot \color{blue}{\mathsf{fma}\left({y}^{2}, \frac{1}{6}, 1\right)} \]
                              4. unpow2N/A

                                \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{y \cdot y}, \frac{1}{6}, 1\right) \]
                              5. lower-*.f6448.8

                                \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{y \cdot y}, 0.16666666666666666, 1\right) \]
                            5. Applied rewrites48.8%

                              \[\leadsto \sin x \cdot \color{blue}{\mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right)} \]
                            6. Taylor expanded in x around 0

                              \[\leadsto \color{blue}{\left(x \cdot \left(1 + \frac{-1}{6} \cdot {x}^{2}\right)\right)} \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                            7. Step-by-step derivation
                              1. +-commutativeN/A

                                \[\leadsto \left(x \cdot \color{blue}{\left(\frac{-1}{6} \cdot {x}^{2} + 1\right)}\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                              2. distribute-lft-inN/A

                                \[\leadsto \color{blue}{\left(x \cdot \left(\frac{-1}{6} \cdot {x}^{2}\right) + x \cdot 1\right)} \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                              3. *-commutativeN/A

                                \[\leadsto \left(x \cdot \color{blue}{\left({x}^{2} \cdot \frac{-1}{6}\right)} + x \cdot 1\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                              4. associate-*r*N/A

                                \[\leadsto \left(\color{blue}{\left(x \cdot {x}^{2}\right) \cdot \frac{-1}{6}} + x \cdot 1\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                              5. *-rgt-identityN/A

                                \[\leadsto \left(\left(x \cdot {x}^{2}\right) \cdot \frac{-1}{6} + \color{blue}{x}\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                              6. lower-fma.f64N/A

                                \[\leadsto \color{blue}{\mathsf{fma}\left(x \cdot {x}^{2}, \frac{-1}{6}, x\right)} \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                              7. unpow2N/A

                                \[\leadsto \mathsf{fma}\left(x \cdot \color{blue}{\left(x \cdot x\right)}, \frac{-1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                              8. cube-unmultN/A

                                \[\leadsto \mathsf{fma}\left(\color{blue}{{x}^{3}}, \frac{-1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                              9. metadata-evalN/A

                                \[\leadsto \mathsf{fma}\left({x}^{\color{blue}{\left(2 + 1\right)}}, \frac{-1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                              10. lower-pow.f64N/A

                                \[\leadsto \mathsf{fma}\left(\color{blue}{{x}^{\left(2 + 1\right)}}, \frac{-1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                              11. metadata-eval47.7

                                \[\leadsto \mathsf{fma}\left({x}^{\color{blue}{3}}, -0.16666666666666666, x\right) \cdot \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right) \]
                            8. Applied rewrites47.7%

                              \[\leadsto \color{blue}{\mathsf{fma}\left({x}^{3}, -0.16666666666666666, x\right)} \cdot \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right) \]
                            9. Step-by-step derivation
                              1. Applied rewrites47.7%

                                \[\leadsto \mathsf{fma}\left(\left(x \cdot x\right) \cdot x, -0.16666666666666666, x\right) \cdot \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right) \]

                              if -inf.0 < (*.f64 (sin.f64 x) (/.f64 (sinh.f64 y) y))

                              1. Initial program 100.0%

                                \[\sin x \cdot \frac{\sinh y}{y} \]
                              2. Add Preprocessing
                              3. Taylor expanded in y around 0

                                \[\leadsto \sin x \cdot \color{blue}{\left(1 + {y}^{2} \cdot \left(\frac{1}{6} + \frac{1}{120} \cdot {y}^{2}\right)\right)} \]
                              4. Step-by-step derivation
                                1. fp-cancel-sign-sub-invN/A

                                  \[\leadsto \sin x \cdot \color{blue}{\left(1 - \left(\mathsf{neg}\left({y}^{2}\right)\right) \cdot \left(\frac{1}{6} + \frac{1}{120} \cdot {y}^{2}\right)\right)} \]
                                2. fp-cancel-sub-sign-invN/A

                                  \[\leadsto \sin x \cdot \color{blue}{\left(1 + \left(\mathsf{neg}\left(\left(\mathsf{neg}\left({y}^{2}\right)\right)\right)\right) \cdot \left(\frac{1}{6} + \frac{1}{120} \cdot {y}^{2}\right)\right)} \]
                                3. +-commutativeN/A

                                  \[\leadsto \sin x \cdot \color{blue}{\left(\left(\mathsf{neg}\left(\left(\mathsf{neg}\left({y}^{2}\right)\right)\right)\right) \cdot \left(\frac{1}{6} + \frac{1}{120} \cdot {y}^{2}\right) + 1\right)} \]
                                4. remove-double-negN/A

                                  \[\leadsto \sin x \cdot \left(\color{blue}{{y}^{2}} \cdot \left(\frac{1}{6} + \frac{1}{120} \cdot {y}^{2}\right) + 1\right) \]
                                5. *-commutativeN/A

                                  \[\leadsto \sin x \cdot \left(\color{blue}{\left(\frac{1}{6} + \frac{1}{120} \cdot {y}^{2}\right) \cdot {y}^{2}} + 1\right) \]
                                6. lower-fma.f64N/A

                                  \[\leadsto \sin x \cdot \color{blue}{\mathsf{fma}\left(\frac{1}{6} + \frac{1}{120} \cdot {y}^{2}, {y}^{2}, 1\right)} \]
                                7. +-commutativeN/A

                                  \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{\frac{1}{120} \cdot {y}^{2} + \frac{1}{6}}, {y}^{2}, 1\right) \]
                                8. *-commutativeN/A

                                  \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{{y}^{2} \cdot \frac{1}{120}} + \frac{1}{6}, {y}^{2}, 1\right) \]
                                9. lower-fma.f64N/A

                                  \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{\mathsf{fma}\left({y}^{2}, \frac{1}{120}, \frac{1}{6}\right)}, {y}^{2}, 1\right) \]
                                10. unpow2N/A

                                  \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(\color{blue}{y \cdot y}, \frac{1}{120}, \frac{1}{6}\right), {y}^{2}, 1\right) \]
                                11. lower-*.f64N/A

                                  \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(\color{blue}{y \cdot y}, \frac{1}{120}, \frac{1}{6}\right), {y}^{2}, 1\right) \]
                                12. unpow2N/A

                                  \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(y \cdot y, \frac{1}{120}, \frac{1}{6}\right), \color{blue}{y \cdot y}, 1\right) \]
                                13. lower-*.f6489.9

                                  \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(y \cdot y, 0.008333333333333333, 0.16666666666666666\right), \color{blue}{y \cdot y}, 1\right) \]
                              5. Applied rewrites89.9%

                                \[\leadsto \sin x \cdot \color{blue}{\mathsf{fma}\left(\mathsf{fma}\left(y \cdot y, 0.008333333333333333, 0.16666666666666666\right), y \cdot y, 1\right)} \]
                              6. Step-by-step derivation
                                1. Applied rewrites89.9%

                                  \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(0.008333333333333333, y \cdot y, 0.16666666666666666\right) \cdot y, \color{blue}{y}, 1\right) \]
                              7. Recombined 2 regimes into one program.
                              8. Add Preprocessing

                              Alternative 7: 80.3% accurate, 0.6× speedup?

                              \[\begin{array}{l} \\ \begin{array}{l} \mathbf{if}\;\sin x \cdot \frac{\sinh y}{y} \leq -\infty:\\ \;\;\;\;\mathsf{fma}\left(\left(x \cdot x\right) \cdot x, -0.16666666666666666, x\right) \cdot \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right)\\ \mathbf{else}:\\ \;\;\;\;\sin x \cdot \mathsf{fma}\left(\left(y \cdot y\right) \cdot 0.008333333333333333, y \cdot y, 1\right)\\ \end{array} \end{array} \]
                              (FPCore (x y)
                               :precision binary64
                               (if (<= (* (sin x) (/ (sinh y) y)) (- INFINITY))
                                 (*
                                  (fma (* (* x x) x) -0.16666666666666666 x)
                                  (fma (* y y) 0.16666666666666666 1.0))
                                 (* (sin x) (fma (* (* y y) 0.008333333333333333) (* y y) 1.0))))
                              double code(double x, double y) {
                              	double tmp;
                              	if ((sin(x) * (sinh(y) / y)) <= -((double) INFINITY)) {
                              		tmp = fma(((x * x) * x), -0.16666666666666666, x) * fma((y * y), 0.16666666666666666, 1.0);
                              	} else {
                              		tmp = sin(x) * fma(((y * y) * 0.008333333333333333), (y * y), 1.0);
                              	}
                              	return tmp;
                              }
                              
                              function code(x, y)
                              	tmp = 0.0
                              	if (Float64(sin(x) * Float64(sinh(y) / y)) <= Float64(-Inf))
                              		tmp = Float64(fma(Float64(Float64(x * x) * x), -0.16666666666666666, x) * fma(Float64(y * y), 0.16666666666666666, 1.0));
                              	else
                              		tmp = Float64(sin(x) * fma(Float64(Float64(y * y) * 0.008333333333333333), Float64(y * y), 1.0));
                              	end
                              	return tmp
                              end
                              
                              code[x_, y_] := If[LessEqual[N[(N[Sin[x], $MachinePrecision] * N[(N[Sinh[y], $MachinePrecision] / y), $MachinePrecision]), $MachinePrecision], (-Infinity)], N[(N[(N[(N[(x * x), $MachinePrecision] * x), $MachinePrecision] * -0.16666666666666666 + x), $MachinePrecision] * N[(N[(y * y), $MachinePrecision] * 0.16666666666666666 + 1.0), $MachinePrecision]), $MachinePrecision], N[(N[Sin[x], $MachinePrecision] * N[(N[(N[(y * y), $MachinePrecision] * 0.008333333333333333), $MachinePrecision] * N[(y * y), $MachinePrecision] + 1.0), $MachinePrecision]), $MachinePrecision]]
                              
                              \begin{array}{l}
                              
                              \\
                              \begin{array}{l}
                              \mathbf{if}\;\sin x \cdot \frac{\sinh y}{y} \leq -\infty:\\
                              \;\;\;\;\mathsf{fma}\left(\left(x \cdot x\right) \cdot x, -0.16666666666666666, x\right) \cdot \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right)\\
                              
                              \mathbf{else}:\\
                              \;\;\;\;\sin x \cdot \mathsf{fma}\left(\left(y \cdot y\right) \cdot 0.008333333333333333, y \cdot y, 1\right)\\
                              
                              
                              \end{array}
                              \end{array}
                              
                              Derivation
                              1. Split input into 2 regimes
                              2. if (*.f64 (sin.f64 x) (/.f64 (sinh.f64 y) y)) < -inf.0

                                1. Initial program 100.0%

                                  \[\sin x \cdot \frac{\sinh y}{y} \]
                                2. Add Preprocessing
                                3. Taylor expanded in y around 0

                                  \[\leadsto \sin x \cdot \color{blue}{\left(1 + \frac{1}{6} \cdot {y}^{2}\right)} \]
                                4. Step-by-step derivation
                                  1. +-commutativeN/A

                                    \[\leadsto \sin x \cdot \color{blue}{\left(\frac{1}{6} \cdot {y}^{2} + 1\right)} \]
                                  2. *-commutativeN/A

                                    \[\leadsto \sin x \cdot \left(\color{blue}{{y}^{2} \cdot \frac{1}{6}} + 1\right) \]
                                  3. lower-fma.f64N/A

                                    \[\leadsto \sin x \cdot \color{blue}{\mathsf{fma}\left({y}^{2}, \frac{1}{6}, 1\right)} \]
                                  4. unpow2N/A

                                    \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{y \cdot y}, \frac{1}{6}, 1\right) \]
                                  5. lower-*.f6448.8

                                    \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{y \cdot y}, 0.16666666666666666, 1\right) \]
                                5. Applied rewrites48.8%

                                  \[\leadsto \sin x \cdot \color{blue}{\mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right)} \]
                                6. Taylor expanded in x around 0

                                  \[\leadsto \color{blue}{\left(x \cdot \left(1 + \frac{-1}{6} \cdot {x}^{2}\right)\right)} \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                7. Step-by-step derivation
                                  1. +-commutativeN/A

                                    \[\leadsto \left(x \cdot \color{blue}{\left(\frac{-1}{6} \cdot {x}^{2} + 1\right)}\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                  2. distribute-lft-inN/A

                                    \[\leadsto \color{blue}{\left(x \cdot \left(\frac{-1}{6} \cdot {x}^{2}\right) + x \cdot 1\right)} \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                  3. *-commutativeN/A

                                    \[\leadsto \left(x \cdot \color{blue}{\left({x}^{2} \cdot \frac{-1}{6}\right)} + x \cdot 1\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                  4. associate-*r*N/A

                                    \[\leadsto \left(\color{blue}{\left(x \cdot {x}^{2}\right) \cdot \frac{-1}{6}} + x \cdot 1\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                  5. *-rgt-identityN/A

                                    \[\leadsto \left(\left(x \cdot {x}^{2}\right) \cdot \frac{-1}{6} + \color{blue}{x}\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                  6. lower-fma.f64N/A

                                    \[\leadsto \color{blue}{\mathsf{fma}\left(x \cdot {x}^{2}, \frac{-1}{6}, x\right)} \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                  7. unpow2N/A

                                    \[\leadsto \mathsf{fma}\left(x \cdot \color{blue}{\left(x \cdot x\right)}, \frac{-1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                  8. cube-unmultN/A

                                    \[\leadsto \mathsf{fma}\left(\color{blue}{{x}^{3}}, \frac{-1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                  9. metadata-evalN/A

                                    \[\leadsto \mathsf{fma}\left({x}^{\color{blue}{\left(2 + 1\right)}}, \frac{-1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                  10. lower-pow.f64N/A

                                    \[\leadsto \mathsf{fma}\left(\color{blue}{{x}^{\left(2 + 1\right)}}, \frac{-1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                  11. metadata-eval47.7

                                    \[\leadsto \mathsf{fma}\left({x}^{\color{blue}{3}}, -0.16666666666666666, x\right) \cdot \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right) \]
                                8. Applied rewrites47.7%

                                  \[\leadsto \color{blue}{\mathsf{fma}\left({x}^{3}, -0.16666666666666666, x\right)} \cdot \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right) \]
                                9. Step-by-step derivation
                                  1. Applied rewrites47.7%

                                    \[\leadsto \mathsf{fma}\left(\left(x \cdot x\right) \cdot x, -0.16666666666666666, x\right) \cdot \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right) \]

                                  if -inf.0 < (*.f64 (sin.f64 x) (/.f64 (sinh.f64 y) y))

                                  1. Initial program 100.0%

                                    \[\sin x \cdot \frac{\sinh y}{y} \]
                                  2. Add Preprocessing
                                  3. Taylor expanded in y around 0

                                    \[\leadsto \sin x \cdot \color{blue}{\left(1 + {y}^{2} \cdot \left(\frac{1}{6} + \frac{1}{120} \cdot {y}^{2}\right)\right)} \]
                                  4. Step-by-step derivation
                                    1. fp-cancel-sign-sub-invN/A

                                      \[\leadsto \sin x \cdot \color{blue}{\left(1 - \left(\mathsf{neg}\left({y}^{2}\right)\right) \cdot \left(\frac{1}{6} + \frac{1}{120} \cdot {y}^{2}\right)\right)} \]
                                    2. fp-cancel-sub-sign-invN/A

                                      \[\leadsto \sin x \cdot \color{blue}{\left(1 + \left(\mathsf{neg}\left(\left(\mathsf{neg}\left({y}^{2}\right)\right)\right)\right) \cdot \left(\frac{1}{6} + \frac{1}{120} \cdot {y}^{2}\right)\right)} \]
                                    3. +-commutativeN/A

                                      \[\leadsto \sin x \cdot \color{blue}{\left(\left(\mathsf{neg}\left(\left(\mathsf{neg}\left({y}^{2}\right)\right)\right)\right) \cdot \left(\frac{1}{6} + \frac{1}{120} \cdot {y}^{2}\right) + 1\right)} \]
                                    4. remove-double-negN/A

                                      \[\leadsto \sin x \cdot \left(\color{blue}{{y}^{2}} \cdot \left(\frac{1}{6} + \frac{1}{120} \cdot {y}^{2}\right) + 1\right) \]
                                    5. *-commutativeN/A

                                      \[\leadsto \sin x \cdot \left(\color{blue}{\left(\frac{1}{6} + \frac{1}{120} \cdot {y}^{2}\right) \cdot {y}^{2}} + 1\right) \]
                                    6. lower-fma.f64N/A

                                      \[\leadsto \sin x \cdot \color{blue}{\mathsf{fma}\left(\frac{1}{6} + \frac{1}{120} \cdot {y}^{2}, {y}^{2}, 1\right)} \]
                                    7. +-commutativeN/A

                                      \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{\frac{1}{120} \cdot {y}^{2} + \frac{1}{6}}, {y}^{2}, 1\right) \]
                                    8. *-commutativeN/A

                                      \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{{y}^{2} \cdot \frac{1}{120}} + \frac{1}{6}, {y}^{2}, 1\right) \]
                                    9. lower-fma.f64N/A

                                      \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{\mathsf{fma}\left({y}^{2}, \frac{1}{120}, \frac{1}{6}\right)}, {y}^{2}, 1\right) \]
                                    10. unpow2N/A

                                      \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(\color{blue}{y \cdot y}, \frac{1}{120}, \frac{1}{6}\right), {y}^{2}, 1\right) \]
                                    11. lower-*.f64N/A

                                      \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(\color{blue}{y \cdot y}, \frac{1}{120}, \frac{1}{6}\right), {y}^{2}, 1\right) \]
                                    12. unpow2N/A

                                      \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(y \cdot y, \frac{1}{120}, \frac{1}{6}\right), \color{blue}{y \cdot y}, 1\right) \]
                                    13. lower-*.f6489.9

                                      \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(y \cdot y, 0.008333333333333333, 0.16666666666666666\right), \color{blue}{y \cdot y}, 1\right) \]
                                  5. Applied rewrites89.9%

                                    \[\leadsto \sin x \cdot \color{blue}{\mathsf{fma}\left(\mathsf{fma}\left(y \cdot y, 0.008333333333333333, 0.16666666666666666\right), y \cdot y, 1\right)} \]
                                  6. Taylor expanded in y around inf

                                    \[\leadsto \sin x \cdot \mathsf{fma}\left(\frac{1}{120} \cdot {y}^{2}, \color{blue}{y} \cdot y, 1\right) \]
                                  7. Step-by-step derivation
                                    1. Applied rewrites89.2%

                                      \[\leadsto \sin x \cdot \mathsf{fma}\left(\left(y \cdot y\right) \cdot 0.008333333333333333, \color{blue}{y} \cdot y, 1\right) \]
                                  8. Recombined 2 regimes into one program.
                                  9. Add Preprocessing

                                  Alternative 8: 50.8% accurate, 1.4× speedup?

                                  \[\begin{array}{l} \\ \begin{array}{l} t_0 := \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right)\\ \mathbf{if}\;\sin x \leq -0.01:\\ \;\;\;\;\mathsf{fma}\left(\left(x \cdot x\right) \cdot x, -0.16666666666666666, x\right) \cdot t\_0\\ \mathbf{else}:\\ \;\;\;\;\mathsf{fma}\left(\mathsf{fma}\left(x \cdot x, 0.008333333333333333, -0.16666666666666666\right) \cdot \left(x \cdot x\right), x, x\right) \cdot t\_0\\ \end{array} \end{array} \]
                                  (FPCore (x y)
                                   :precision binary64
                                   (let* ((t_0 (fma (* y y) 0.16666666666666666 1.0)))
                                     (if (<= (sin x) -0.01)
                                       (* (fma (* (* x x) x) -0.16666666666666666 x) t_0)
                                       (*
                                        (fma
                                         (* (fma (* x x) 0.008333333333333333 -0.16666666666666666) (* x x))
                                         x
                                         x)
                                        t_0))))
                                  double code(double x, double y) {
                                  	double t_0 = fma((y * y), 0.16666666666666666, 1.0);
                                  	double tmp;
                                  	if (sin(x) <= -0.01) {
                                  		tmp = fma(((x * x) * x), -0.16666666666666666, x) * t_0;
                                  	} else {
                                  		tmp = fma((fma((x * x), 0.008333333333333333, -0.16666666666666666) * (x * x)), x, x) * t_0;
                                  	}
                                  	return tmp;
                                  }
                                  
                                  function code(x, y)
                                  	t_0 = fma(Float64(y * y), 0.16666666666666666, 1.0)
                                  	tmp = 0.0
                                  	if (sin(x) <= -0.01)
                                  		tmp = Float64(fma(Float64(Float64(x * x) * x), -0.16666666666666666, x) * t_0);
                                  	else
                                  		tmp = Float64(fma(Float64(fma(Float64(x * x), 0.008333333333333333, -0.16666666666666666) * Float64(x * x)), x, x) * t_0);
                                  	end
                                  	return tmp
                                  end
                                  
                                  code[x_, y_] := Block[{t$95$0 = N[(N[(y * y), $MachinePrecision] * 0.16666666666666666 + 1.0), $MachinePrecision]}, If[LessEqual[N[Sin[x], $MachinePrecision], -0.01], N[(N[(N[(N[(x * x), $MachinePrecision] * x), $MachinePrecision] * -0.16666666666666666 + x), $MachinePrecision] * t$95$0), $MachinePrecision], N[(N[(N[(N[(N[(x * x), $MachinePrecision] * 0.008333333333333333 + -0.16666666666666666), $MachinePrecision] * N[(x * x), $MachinePrecision]), $MachinePrecision] * x + x), $MachinePrecision] * t$95$0), $MachinePrecision]]]
                                  
                                  \begin{array}{l}
                                  
                                  \\
                                  \begin{array}{l}
                                  t_0 := \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right)\\
                                  \mathbf{if}\;\sin x \leq -0.01:\\
                                  \;\;\;\;\mathsf{fma}\left(\left(x \cdot x\right) \cdot x, -0.16666666666666666, x\right) \cdot t\_0\\
                                  
                                  \mathbf{else}:\\
                                  \;\;\;\;\mathsf{fma}\left(\mathsf{fma}\left(x \cdot x, 0.008333333333333333, -0.16666666666666666\right) \cdot \left(x \cdot x\right), x, x\right) \cdot t\_0\\
                                  
                                  
                                  \end{array}
                                  \end{array}
                                  
                                  Derivation
                                  1. Split input into 2 regimes
                                  2. if (sin.f64 x) < -0.0100000000000000002

                                    1. Initial program 100.0%

                                      \[\sin x \cdot \frac{\sinh y}{y} \]
                                    2. Add Preprocessing
                                    3. Taylor expanded in y around 0

                                      \[\leadsto \sin x \cdot \color{blue}{\left(1 + \frac{1}{6} \cdot {y}^{2}\right)} \]
                                    4. Step-by-step derivation
                                      1. +-commutativeN/A

                                        \[\leadsto \sin x \cdot \color{blue}{\left(\frac{1}{6} \cdot {y}^{2} + 1\right)} \]
                                      2. *-commutativeN/A

                                        \[\leadsto \sin x \cdot \left(\color{blue}{{y}^{2} \cdot \frac{1}{6}} + 1\right) \]
                                      3. lower-fma.f64N/A

                                        \[\leadsto \sin x \cdot \color{blue}{\mathsf{fma}\left({y}^{2}, \frac{1}{6}, 1\right)} \]
                                      4. unpow2N/A

                                        \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{y \cdot y}, \frac{1}{6}, 1\right) \]
                                      5. lower-*.f6473.4

                                        \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{y \cdot y}, 0.16666666666666666, 1\right) \]
                                    5. Applied rewrites73.4%

                                      \[\leadsto \sin x \cdot \color{blue}{\mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right)} \]
                                    6. Taylor expanded in x around 0

                                      \[\leadsto \color{blue}{\left(x \cdot \left(1 + \frac{-1}{6} \cdot {x}^{2}\right)\right)} \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                    7. Step-by-step derivation
                                      1. +-commutativeN/A

                                        \[\leadsto \left(x \cdot \color{blue}{\left(\frac{-1}{6} \cdot {x}^{2} + 1\right)}\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                      2. distribute-lft-inN/A

                                        \[\leadsto \color{blue}{\left(x \cdot \left(\frac{-1}{6} \cdot {x}^{2}\right) + x \cdot 1\right)} \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                      3. *-commutativeN/A

                                        \[\leadsto \left(x \cdot \color{blue}{\left({x}^{2} \cdot \frac{-1}{6}\right)} + x \cdot 1\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                      4. associate-*r*N/A

                                        \[\leadsto \left(\color{blue}{\left(x \cdot {x}^{2}\right) \cdot \frac{-1}{6}} + x \cdot 1\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                      5. *-rgt-identityN/A

                                        \[\leadsto \left(\left(x \cdot {x}^{2}\right) \cdot \frac{-1}{6} + \color{blue}{x}\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                      6. lower-fma.f64N/A

                                        \[\leadsto \color{blue}{\mathsf{fma}\left(x \cdot {x}^{2}, \frac{-1}{6}, x\right)} \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                      7. unpow2N/A

                                        \[\leadsto \mathsf{fma}\left(x \cdot \color{blue}{\left(x \cdot x\right)}, \frac{-1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                      8. cube-unmultN/A

                                        \[\leadsto \mathsf{fma}\left(\color{blue}{{x}^{3}}, \frac{-1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                      9. metadata-evalN/A

                                        \[\leadsto \mathsf{fma}\left({x}^{\color{blue}{\left(2 + 1\right)}}, \frac{-1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                      10. lower-pow.f64N/A

                                        \[\leadsto \mathsf{fma}\left(\color{blue}{{x}^{\left(2 + 1\right)}}, \frac{-1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                      11. metadata-eval25.1

                                        \[\leadsto \mathsf{fma}\left({x}^{\color{blue}{3}}, -0.16666666666666666, x\right) \cdot \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right) \]
                                    8. Applied rewrites25.1%

                                      \[\leadsto \color{blue}{\mathsf{fma}\left({x}^{3}, -0.16666666666666666, x\right)} \cdot \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right) \]
                                    9. Step-by-step derivation
                                      1. Applied rewrites25.1%

                                        \[\leadsto \mathsf{fma}\left(\left(x \cdot x\right) \cdot x, -0.16666666666666666, x\right) \cdot \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right) \]

                                      if -0.0100000000000000002 < (sin.f64 x)

                                      1. Initial program 100.0%

                                        \[\sin x \cdot \frac{\sinh y}{y} \]
                                      2. Add Preprocessing
                                      3. Taylor expanded in y around 0

                                        \[\leadsto \sin x \cdot \color{blue}{\left(1 + \frac{1}{6} \cdot {y}^{2}\right)} \]
                                      4. Step-by-step derivation
                                        1. +-commutativeN/A

                                          \[\leadsto \sin x \cdot \color{blue}{\left(\frac{1}{6} \cdot {y}^{2} + 1\right)} \]
                                        2. *-commutativeN/A

                                          \[\leadsto \sin x \cdot \left(\color{blue}{{y}^{2} \cdot \frac{1}{6}} + 1\right) \]
                                        3. lower-fma.f64N/A

                                          \[\leadsto \sin x \cdot \color{blue}{\mathsf{fma}\left({y}^{2}, \frac{1}{6}, 1\right)} \]
                                        4. unpow2N/A

                                          \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{y \cdot y}, \frac{1}{6}, 1\right) \]
                                        5. lower-*.f6471.6

                                          \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{y \cdot y}, 0.16666666666666666, 1\right) \]
                                      5. Applied rewrites71.6%

                                        \[\leadsto \sin x \cdot \color{blue}{\mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right)} \]
                                      6. Taylor expanded in x around 0

                                        \[\leadsto \color{blue}{\left(x \cdot \left(1 + {x}^{2} \cdot \left(\frac{1}{120} \cdot {x}^{2} - \frac{1}{6}\right)\right)\right)} \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                      7. Step-by-step derivation
                                        1. +-commutativeN/A

                                          \[\leadsto \left(x \cdot \color{blue}{\left({x}^{2} \cdot \left(\frac{1}{120} \cdot {x}^{2} - \frac{1}{6}\right) + 1\right)}\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                        2. distribute-lft-inN/A

                                          \[\leadsto \color{blue}{\left(x \cdot \left({x}^{2} \cdot \left(\frac{1}{120} \cdot {x}^{2} - \frac{1}{6}\right)\right) + x \cdot 1\right)} \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                        3. associate-*r*N/A

                                          \[\leadsto \left(\color{blue}{\left(x \cdot {x}^{2}\right) \cdot \left(\frac{1}{120} \cdot {x}^{2} - \frac{1}{6}\right)} + x \cdot 1\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                        4. *-rgt-identityN/A

                                          \[\leadsto \left(\left(x \cdot {x}^{2}\right) \cdot \left(\frac{1}{120} \cdot {x}^{2} - \frac{1}{6}\right) + \color{blue}{x}\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                        5. lower-fma.f64N/A

                                          \[\leadsto \color{blue}{\mathsf{fma}\left(x \cdot {x}^{2}, \frac{1}{120} \cdot {x}^{2} - \frac{1}{6}, x\right)} \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                        6. unpow2N/A

                                          \[\leadsto \mathsf{fma}\left(x \cdot \color{blue}{\left(x \cdot x\right)}, \frac{1}{120} \cdot {x}^{2} - \frac{1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                        7. cube-unmultN/A

                                          \[\leadsto \mathsf{fma}\left(\color{blue}{{x}^{3}}, \frac{1}{120} \cdot {x}^{2} - \frac{1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                        8. metadata-evalN/A

                                          \[\leadsto \mathsf{fma}\left({x}^{\color{blue}{\left(2 + 1\right)}}, \frac{1}{120} \cdot {x}^{2} - \frac{1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                        9. lower-pow.f64N/A

                                          \[\leadsto \mathsf{fma}\left(\color{blue}{{x}^{\left(2 + 1\right)}}, \frac{1}{120} \cdot {x}^{2} - \frac{1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                        10. metadata-evalN/A

                                          \[\leadsto \mathsf{fma}\left({x}^{\color{blue}{3}}, \frac{1}{120} \cdot {x}^{2} - \frac{1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                        11. metadata-evalN/A

                                          \[\leadsto \mathsf{fma}\left({x}^{3}, \frac{1}{120} \cdot {x}^{2} - \color{blue}{\frac{1}{6} \cdot 1}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                        12. fp-cancel-sub-sign-invN/A

                                          \[\leadsto \mathsf{fma}\left({x}^{3}, \color{blue}{\frac{1}{120} \cdot {x}^{2} + \left(\mathsf{neg}\left(\frac{1}{6}\right)\right) \cdot 1}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                        13. metadata-evalN/A

                                          \[\leadsto \mathsf{fma}\left({x}^{3}, \frac{1}{120} \cdot {x}^{2} + \color{blue}{\frac{-1}{6}} \cdot 1, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                        14. metadata-evalN/A

                                          \[\leadsto \mathsf{fma}\left({x}^{3}, \frac{1}{120} \cdot {x}^{2} + \color{blue}{\frac{-1}{6}}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                        15. lower-fma.f64N/A

                                          \[\leadsto \mathsf{fma}\left({x}^{3}, \color{blue}{\mathsf{fma}\left(\frac{1}{120}, {x}^{2}, \frac{-1}{6}\right)}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                        16. unpow2N/A

                                          \[\leadsto \mathsf{fma}\left({x}^{3}, \mathsf{fma}\left(\frac{1}{120}, \color{blue}{x \cdot x}, \frac{-1}{6}\right), x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                        17. lower-*.f6453.9

                                          \[\leadsto \mathsf{fma}\left({x}^{3}, \mathsf{fma}\left(0.008333333333333333, \color{blue}{x \cdot x}, -0.16666666666666666\right), x\right) \cdot \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right) \]
                                      8. Applied rewrites53.9%

                                        \[\leadsto \color{blue}{\mathsf{fma}\left({x}^{3}, \mathsf{fma}\left(0.008333333333333333, x \cdot x, -0.16666666666666666\right), x\right)} \cdot \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right) \]
                                      9. Step-by-step derivation
                                        1. Applied rewrites53.9%

                                          \[\leadsto \mathsf{fma}\left(\mathsf{fma}\left(x \cdot x, 0.008333333333333333, -0.16666666666666666\right) \cdot \left(x \cdot x\right), \color{blue}{x}, x\right) \cdot \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right) \]
                                      10. Recombined 2 regimes into one program.
                                      11. Add Preprocessing

                                      Alternative 9: 91.9% accurate, 1.6× speedup?

                                      \[\begin{array}{l} \\ \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(0.0001984126984126984, y \cdot y, 0.008333333333333333\right) \cdot y, y, 0.16666666666666666\right) \cdot y, y, 1\right) \end{array} \]
                                      (FPCore (x y)
                                       :precision binary64
                                       (*
                                        (sin x)
                                        (fma
                                         (*
                                          (fma
                                           (* (fma 0.0001984126984126984 (* y y) 0.008333333333333333) y)
                                           y
                                           0.16666666666666666)
                                          y)
                                         y
                                         1.0)))
                                      double code(double x, double y) {
                                      	return sin(x) * fma((fma((fma(0.0001984126984126984, (y * y), 0.008333333333333333) * y), y, 0.16666666666666666) * y), y, 1.0);
                                      }
                                      
                                      function code(x, y)
                                      	return Float64(sin(x) * fma(Float64(fma(Float64(fma(0.0001984126984126984, Float64(y * y), 0.008333333333333333) * y), y, 0.16666666666666666) * y), y, 1.0))
                                      end
                                      
                                      code[x_, y_] := N[(N[Sin[x], $MachinePrecision] * N[(N[(N[(N[(N[(0.0001984126984126984 * N[(y * y), $MachinePrecision] + 0.008333333333333333), $MachinePrecision] * y), $MachinePrecision] * y + 0.16666666666666666), $MachinePrecision] * y), $MachinePrecision] * y + 1.0), $MachinePrecision]), $MachinePrecision]
                                      
                                      \begin{array}{l}
                                      
                                      \\
                                      \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(0.0001984126984126984, y \cdot y, 0.008333333333333333\right) \cdot y, y, 0.16666666666666666\right) \cdot y, y, 1\right)
                                      \end{array}
                                      
                                      Derivation
                                      1. Initial program 100.0%

                                        \[\sin x \cdot \frac{\sinh y}{y} \]
                                      2. Add Preprocessing
                                      3. Taylor expanded in y around 0

                                        \[\leadsto \sin x \cdot \color{blue}{\left(1 + {y}^{2} \cdot \left(\frac{1}{6} + {y}^{2} \cdot \left(\frac{1}{120} + \frac{1}{5040} \cdot {y}^{2}\right)\right)\right)} \]
                                      4. Step-by-step derivation
                                        1. fp-cancel-sign-sub-invN/A

                                          \[\leadsto \sin x \cdot \color{blue}{\left(1 - \left(\mathsf{neg}\left({y}^{2}\right)\right) \cdot \left(\frac{1}{6} + {y}^{2} \cdot \left(\frac{1}{120} + \frac{1}{5040} \cdot {y}^{2}\right)\right)\right)} \]
                                        2. fp-cancel-sub-sign-invN/A

                                          \[\leadsto \sin x \cdot \color{blue}{\left(1 + \left(\mathsf{neg}\left(\left(\mathsf{neg}\left({y}^{2}\right)\right)\right)\right) \cdot \left(\frac{1}{6} + {y}^{2} \cdot \left(\frac{1}{120} + \frac{1}{5040} \cdot {y}^{2}\right)\right)\right)} \]
                                        3. +-commutativeN/A

                                          \[\leadsto \sin x \cdot \color{blue}{\left(\left(\mathsf{neg}\left(\left(\mathsf{neg}\left({y}^{2}\right)\right)\right)\right) \cdot \left(\frac{1}{6} + {y}^{2} \cdot \left(\frac{1}{120} + \frac{1}{5040} \cdot {y}^{2}\right)\right) + 1\right)} \]
                                        4. remove-double-negN/A

                                          \[\leadsto \sin x \cdot \left(\color{blue}{{y}^{2}} \cdot \left(\frac{1}{6} + {y}^{2} \cdot \left(\frac{1}{120} + \frac{1}{5040} \cdot {y}^{2}\right)\right) + 1\right) \]
                                        5. *-commutativeN/A

                                          \[\leadsto \sin x \cdot \left(\color{blue}{\left(\frac{1}{6} + {y}^{2} \cdot \left(\frac{1}{120} + \frac{1}{5040} \cdot {y}^{2}\right)\right) \cdot {y}^{2}} + 1\right) \]
                                        6. lower-fma.f64N/A

                                          \[\leadsto \sin x \cdot \color{blue}{\mathsf{fma}\left(\frac{1}{6} + {y}^{2} \cdot \left(\frac{1}{120} + \frac{1}{5040} \cdot {y}^{2}\right), {y}^{2}, 1\right)} \]
                                        7. +-commutativeN/A

                                          \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{{y}^{2} \cdot \left(\frac{1}{120} + \frac{1}{5040} \cdot {y}^{2}\right) + \frac{1}{6}}, {y}^{2}, 1\right) \]
                                        8. *-commutativeN/A

                                          \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{\left(\frac{1}{120} + \frac{1}{5040} \cdot {y}^{2}\right) \cdot {y}^{2}} + \frac{1}{6}, {y}^{2}, 1\right) \]
                                        9. lower-fma.f64N/A

                                          \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{\mathsf{fma}\left(\frac{1}{120} + \frac{1}{5040} \cdot {y}^{2}, {y}^{2}, \frac{1}{6}\right)}, {y}^{2}, 1\right) \]
                                        10. +-commutativeN/A

                                          \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(\color{blue}{\frac{1}{5040} \cdot {y}^{2} + \frac{1}{120}}, {y}^{2}, \frac{1}{6}\right), {y}^{2}, 1\right) \]
                                        11. lower-fma.f64N/A

                                          \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(\color{blue}{\mathsf{fma}\left(\frac{1}{5040}, {y}^{2}, \frac{1}{120}\right)}, {y}^{2}, \frac{1}{6}\right), {y}^{2}, 1\right) \]
                                        12. unpow2N/A

                                          \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{5040}, \color{blue}{y \cdot y}, \frac{1}{120}\right), {y}^{2}, \frac{1}{6}\right), {y}^{2}, 1\right) \]
                                        13. lower-*.f64N/A

                                          \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{5040}, \color{blue}{y \cdot y}, \frac{1}{120}\right), {y}^{2}, \frac{1}{6}\right), {y}^{2}, 1\right) \]
                                        14. unpow2N/A

                                          \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{5040}, y \cdot y, \frac{1}{120}\right), \color{blue}{y \cdot y}, \frac{1}{6}\right), {y}^{2}, 1\right) \]
                                        15. lower-*.f64N/A

                                          \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{5040}, y \cdot y, \frac{1}{120}\right), \color{blue}{y \cdot y}, \frac{1}{6}\right), {y}^{2}, 1\right) \]
                                        16. unpow2N/A

                                          \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{5040}, y \cdot y, \frac{1}{120}\right), y \cdot y, \frac{1}{6}\right), \color{blue}{y \cdot y}, 1\right) \]
                                        17. lower-*.f6490.9

                                          \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(0.0001984126984126984, y \cdot y, 0.008333333333333333\right), y \cdot y, 0.16666666666666666\right), \color{blue}{y \cdot y}, 1\right) \]
                                      5. Applied rewrites90.9%

                                        \[\leadsto \sin x \cdot \color{blue}{\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(0.0001984126984126984, y \cdot y, 0.008333333333333333\right), y \cdot y, 0.16666666666666666\right), y \cdot y, 1\right)} \]
                                      6. Step-by-step derivation
                                        1. Applied rewrites90.9%

                                          \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(y \cdot y, 0.0001984126984126984, 0.008333333333333333\right) \cdot y, y, 0.16666666666666666\right), \color{blue}{y} \cdot y, 1\right) \]
                                        2. Step-by-step derivation
                                          1. Applied rewrites90.9%

                                            \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(0.0001984126984126984, y \cdot y, 0.008333333333333333\right) \cdot y, y, 0.16666666666666666\right) \cdot y, \color{blue}{y}, 1\right) \]
                                          2. Add Preprocessing

                                          Alternative 10: 91.7% accurate, 1.6× speedup?

                                          \[\begin{array}{l} \\ \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(0.0001984126984126984 \cdot \left(y \cdot y\right), y \cdot y, 0.16666666666666666\right), y \cdot y, 1\right) \end{array} \]
                                          (FPCore (x y)
                                           :precision binary64
                                           (*
                                            (sin x)
                                            (fma
                                             (fma (* 0.0001984126984126984 (* y y)) (* y y) 0.16666666666666666)
                                             (* y y)
                                             1.0)))
                                          double code(double x, double y) {
                                          	return sin(x) * fma(fma((0.0001984126984126984 * (y * y)), (y * y), 0.16666666666666666), (y * y), 1.0);
                                          }
                                          
                                          function code(x, y)
                                          	return Float64(sin(x) * fma(fma(Float64(0.0001984126984126984 * Float64(y * y)), Float64(y * y), 0.16666666666666666), Float64(y * y), 1.0))
                                          end
                                          
                                          code[x_, y_] := N[(N[Sin[x], $MachinePrecision] * N[(N[(N[(0.0001984126984126984 * N[(y * y), $MachinePrecision]), $MachinePrecision] * N[(y * y), $MachinePrecision] + 0.16666666666666666), $MachinePrecision] * N[(y * y), $MachinePrecision] + 1.0), $MachinePrecision]), $MachinePrecision]
                                          
                                          \begin{array}{l}
                                          
                                          \\
                                          \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(0.0001984126984126984 \cdot \left(y \cdot y\right), y \cdot y, 0.16666666666666666\right), y \cdot y, 1\right)
                                          \end{array}
                                          
                                          Derivation
                                          1. Initial program 100.0%

                                            \[\sin x \cdot \frac{\sinh y}{y} \]
                                          2. Add Preprocessing
                                          3. Taylor expanded in y around 0

                                            \[\leadsto \sin x \cdot \color{blue}{\left(1 + {y}^{2} \cdot \left(\frac{1}{6} + {y}^{2} \cdot \left(\frac{1}{120} + \frac{1}{5040} \cdot {y}^{2}\right)\right)\right)} \]
                                          4. Step-by-step derivation
                                            1. fp-cancel-sign-sub-invN/A

                                              \[\leadsto \sin x \cdot \color{blue}{\left(1 - \left(\mathsf{neg}\left({y}^{2}\right)\right) \cdot \left(\frac{1}{6} + {y}^{2} \cdot \left(\frac{1}{120} + \frac{1}{5040} \cdot {y}^{2}\right)\right)\right)} \]
                                            2. fp-cancel-sub-sign-invN/A

                                              \[\leadsto \sin x \cdot \color{blue}{\left(1 + \left(\mathsf{neg}\left(\left(\mathsf{neg}\left({y}^{2}\right)\right)\right)\right) \cdot \left(\frac{1}{6} + {y}^{2} \cdot \left(\frac{1}{120} + \frac{1}{5040} \cdot {y}^{2}\right)\right)\right)} \]
                                            3. +-commutativeN/A

                                              \[\leadsto \sin x \cdot \color{blue}{\left(\left(\mathsf{neg}\left(\left(\mathsf{neg}\left({y}^{2}\right)\right)\right)\right) \cdot \left(\frac{1}{6} + {y}^{2} \cdot \left(\frac{1}{120} + \frac{1}{5040} \cdot {y}^{2}\right)\right) + 1\right)} \]
                                            4. remove-double-negN/A

                                              \[\leadsto \sin x \cdot \left(\color{blue}{{y}^{2}} \cdot \left(\frac{1}{6} + {y}^{2} \cdot \left(\frac{1}{120} + \frac{1}{5040} \cdot {y}^{2}\right)\right) + 1\right) \]
                                            5. *-commutativeN/A

                                              \[\leadsto \sin x \cdot \left(\color{blue}{\left(\frac{1}{6} + {y}^{2} \cdot \left(\frac{1}{120} + \frac{1}{5040} \cdot {y}^{2}\right)\right) \cdot {y}^{2}} + 1\right) \]
                                            6. lower-fma.f64N/A

                                              \[\leadsto \sin x \cdot \color{blue}{\mathsf{fma}\left(\frac{1}{6} + {y}^{2} \cdot \left(\frac{1}{120} + \frac{1}{5040} \cdot {y}^{2}\right), {y}^{2}, 1\right)} \]
                                            7. +-commutativeN/A

                                              \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{{y}^{2} \cdot \left(\frac{1}{120} + \frac{1}{5040} \cdot {y}^{2}\right) + \frac{1}{6}}, {y}^{2}, 1\right) \]
                                            8. *-commutativeN/A

                                              \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{\left(\frac{1}{120} + \frac{1}{5040} \cdot {y}^{2}\right) \cdot {y}^{2}} + \frac{1}{6}, {y}^{2}, 1\right) \]
                                            9. lower-fma.f64N/A

                                              \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{\mathsf{fma}\left(\frac{1}{120} + \frac{1}{5040} \cdot {y}^{2}, {y}^{2}, \frac{1}{6}\right)}, {y}^{2}, 1\right) \]
                                            10. +-commutativeN/A

                                              \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(\color{blue}{\frac{1}{5040} \cdot {y}^{2} + \frac{1}{120}}, {y}^{2}, \frac{1}{6}\right), {y}^{2}, 1\right) \]
                                            11. lower-fma.f64N/A

                                              \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(\color{blue}{\mathsf{fma}\left(\frac{1}{5040}, {y}^{2}, \frac{1}{120}\right)}, {y}^{2}, \frac{1}{6}\right), {y}^{2}, 1\right) \]
                                            12. unpow2N/A

                                              \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{5040}, \color{blue}{y \cdot y}, \frac{1}{120}\right), {y}^{2}, \frac{1}{6}\right), {y}^{2}, 1\right) \]
                                            13. lower-*.f64N/A

                                              \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{5040}, \color{blue}{y \cdot y}, \frac{1}{120}\right), {y}^{2}, \frac{1}{6}\right), {y}^{2}, 1\right) \]
                                            14. unpow2N/A

                                              \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{5040}, y \cdot y, \frac{1}{120}\right), \color{blue}{y \cdot y}, \frac{1}{6}\right), {y}^{2}, 1\right) \]
                                            15. lower-*.f64N/A

                                              \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{5040}, y \cdot y, \frac{1}{120}\right), \color{blue}{y \cdot y}, \frac{1}{6}\right), {y}^{2}, 1\right) \]
                                            16. unpow2N/A

                                              \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{5040}, y \cdot y, \frac{1}{120}\right), y \cdot y, \frac{1}{6}\right), \color{blue}{y \cdot y}, 1\right) \]
                                            17. lower-*.f6490.9

                                              \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(0.0001984126984126984, y \cdot y, 0.008333333333333333\right), y \cdot y, 0.16666666666666666\right), \color{blue}{y \cdot y}, 1\right) \]
                                          5. Applied rewrites90.9%

                                            \[\leadsto \sin x \cdot \color{blue}{\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(0.0001984126984126984, y \cdot y, 0.008333333333333333\right), y \cdot y, 0.16666666666666666\right), y \cdot y, 1\right)} \]
                                          6. Taylor expanded in y around inf

                                            \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{5040} \cdot {y}^{2}, y \cdot y, \frac{1}{6}\right), y \cdot y, 1\right) \]
                                          7. Step-by-step derivation
                                            1. Applied rewrites90.7%

                                              \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(0.0001984126984126984 \cdot \left(y \cdot y\right), y \cdot y, 0.16666666666666666\right), y \cdot y, 1\right) \]
                                            2. Add Preprocessing

                                            Alternative 11: 91.4% accurate, 1.6× speedup?

                                            \[\begin{array}{l} \\ \sin x \cdot \mathsf{fma}\left(\left(\mathsf{fma}\left(y \cdot y, 0.0001984126984126984, 0.008333333333333333\right) \cdot y\right) \cdot y, y \cdot y, 1\right) \end{array} \]
                                            (FPCore (x y)
                                             :precision binary64
                                             (*
                                              (sin x)
                                              (fma
                                               (* (* (fma (* y y) 0.0001984126984126984 0.008333333333333333) y) y)
                                               (* y y)
                                               1.0)))
                                            double code(double x, double y) {
                                            	return sin(x) * fma(((fma((y * y), 0.0001984126984126984, 0.008333333333333333) * y) * y), (y * y), 1.0);
                                            }
                                            
                                            function code(x, y)
                                            	return Float64(sin(x) * fma(Float64(Float64(fma(Float64(y * y), 0.0001984126984126984, 0.008333333333333333) * y) * y), Float64(y * y), 1.0))
                                            end
                                            
                                            code[x_, y_] := N[(N[Sin[x], $MachinePrecision] * N[(N[(N[(N[(N[(y * y), $MachinePrecision] * 0.0001984126984126984 + 0.008333333333333333), $MachinePrecision] * y), $MachinePrecision] * y), $MachinePrecision] * N[(y * y), $MachinePrecision] + 1.0), $MachinePrecision]), $MachinePrecision]
                                            
                                            \begin{array}{l}
                                            
                                            \\
                                            \sin x \cdot \mathsf{fma}\left(\left(\mathsf{fma}\left(y \cdot y, 0.0001984126984126984, 0.008333333333333333\right) \cdot y\right) \cdot y, y \cdot y, 1\right)
                                            \end{array}
                                            
                                            Derivation
                                            1. Initial program 100.0%

                                              \[\sin x \cdot \frac{\sinh y}{y} \]
                                            2. Add Preprocessing
                                            3. Taylor expanded in y around 0

                                              \[\leadsto \sin x \cdot \color{blue}{\left(1 + {y}^{2} \cdot \left(\frac{1}{6} + {y}^{2} \cdot \left(\frac{1}{120} + \frac{1}{5040} \cdot {y}^{2}\right)\right)\right)} \]
                                            4. Step-by-step derivation
                                              1. fp-cancel-sign-sub-invN/A

                                                \[\leadsto \sin x \cdot \color{blue}{\left(1 - \left(\mathsf{neg}\left({y}^{2}\right)\right) \cdot \left(\frac{1}{6} + {y}^{2} \cdot \left(\frac{1}{120} + \frac{1}{5040} \cdot {y}^{2}\right)\right)\right)} \]
                                              2. fp-cancel-sub-sign-invN/A

                                                \[\leadsto \sin x \cdot \color{blue}{\left(1 + \left(\mathsf{neg}\left(\left(\mathsf{neg}\left({y}^{2}\right)\right)\right)\right) \cdot \left(\frac{1}{6} + {y}^{2} \cdot \left(\frac{1}{120} + \frac{1}{5040} \cdot {y}^{2}\right)\right)\right)} \]
                                              3. +-commutativeN/A

                                                \[\leadsto \sin x \cdot \color{blue}{\left(\left(\mathsf{neg}\left(\left(\mathsf{neg}\left({y}^{2}\right)\right)\right)\right) \cdot \left(\frac{1}{6} + {y}^{2} \cdot \left(\frac{1}{120} + \frac{1}{5040} \cdot {y}^{2}\right)\right) + 1\right)} \]
                                              4. remove-double-negN/A

                                                \[\leadsto \sin x \cdot \left(\color{blue}{{y}^{2}} \cdot \left(\frac{1}{6} + {y}^{2} \cdot \left(\frac{1}{120} + \frac{1}{5040} \cdot {y}^{2}\right)\right) + 1\right) \]
                                              5. *-commutativeN/A

                                                \[\leadsto \sin x \cdot \left(\color{blue}{\left(\frac{1}{6} + {y}^{2} \cdot \left(\frac{1}{120} + \frac{1}{5040} \cdot {y}^{2}\right)\right) \cdot {y}^{2}} + 1\right) \]
                                              6. lower-fma.f64N/A

                                                \[\leadsto \sin x \cdot \color{blue}{\mathsf{fma}\left(\frac{1}{6} + {y}^{2} \cdot \left(\frac{1}{120} + \frac{1}{5040} \cdot {y}^{2}\right), {y}^{2}, 1\right)} \]
                                              7. +-commutativeN/A

                                                \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{{y}^{2} \cdot \left(\frac{1}{120} + \frac{1}{5040} \cdot {y}^{2}\right) + \frac{1}{6}}, {y}^{2}, 1\right) \]
                                              8. *-commutativeN/A

                                                \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{\left(\frac{1}{120} + \frac{1}{5040} \cdot {y}^{2}\right) \cdot {y}^{2}} + \frac{1}{6}, {y}^{2}, 1\right) \]
                                              9. lower-fma.f64N/A

                                                \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{\mathsf{fma}\left(\frac{1}{120} + \frac{1}{5040} \cdot {y}^{2}, {y}^{2}, \frac{1}{6}\right)}, {y}^{2}, 1\right) \]
                                              10. +-commutativeN/A

                                                \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(\color{blue}{\frac{1}{5040} \cdot {y}^{2} + \frac{1}{120}}, {y}^{2}, \frac{1}{6}\right), {y}^{2}, 1\right) \]
                                              11. lower-fma.f64N/A

                                                \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(\color{blue}{\mathsf{fma}\left(\frac{1}{5040}, {y}^{2}, \frac{1}{120}\right)}, {y}^{2}, \frac{1}{6}\right), {y}^{2}, 1\right) \]
                                              12. unpow2N/A

                                                \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{5040}, \color{blue}{y \cdot y}, \frac{1}{120}\right), {y}^{2}, \frac{1}{6}\right), {y}^{2}, 1\right) \]
                                              13. lower-*.f64N/A

                                                \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{5040}, \color{blue}{y \cdot y}, \frac{1}{120}\right), {y}^{2}, \frac{1}{6}\right), {y}^{2}, 1\right) \]
                                              14. unpow2N/A

                                                \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{5040}, y \cdot y, \frac{1}{120}\right), \color{blue}{y \cdot y}, \frac{1}{6}\right), {y}^{2}, 1\right) \]
                                              15. lower-*.f64N/A

                                                \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{5040}, y \cdot y, \frac{1}{120}\right), \color{blue}{y \cdot y}, \frac{1}{6}\right), {y}^{2}, 1\right) \]
                                              16. unpow2N/A

                                                \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{5040}, y \cdot y, \frac{1}{120}\right), y \cdot y, \frac{1}{6}\right), \color{blue}{y \cdot y}, 1\right) \]
                                              17. lower-*.f6490.9

                                                \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(0.0001984126984126984, y \cdot y, 0.008333333333333333\right), y \cdot y, 0.16666666666666666\right), \color{blue}{y \cdot y}, 1\right) \]
                                            5. Applied rewrites90.9%

                                              \[\leadsto \sin x \cdot \color{blue}{\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(0.0001984126984126984, y \cdot y, 0.008333333333333333\right), y \cdot y, 0.16666666666666666\right), y \cdot y, 1\right)} \]
                                            6. Step-by-step derivation
                                              1. Applied rewrites90.9%

                                                \[\leadsto \sin x \cdot \mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(y \cdot y, 0.0001984126984126984, 0.008333333333333333\right) \cdot y, y, 0.16666666666666666\right), \color{blue}{y} \cdot y, 1\right) \]
                                              2. Taylor expanded in y around inf

                                                \[\leadsto \sin x \cdot \mathsf{fma}\left({y}^{4} \cdot \left(\frac{1}{5040} + \frac{1}{120} \cdot \frac{1}{{y}^{2}}\right), \color{blue}{y} \cdot y, 1\right) \]
                                              3. Step-by-step derivation
                                                1. Applied rewrites90.3%

                                                  \[\leadsto \sin x \cdot \mathsf{fma}\left(\left(\mathsf{fma}\left(y \cdot y, 0.0001984126984126984, 0.008333333333333333\right) \cdot y\right) \cdot y, \color{blue}{y} \cdot y, 1\right) \]
                                                2. Add Preprocessing

                                                Alternative 12: 50.3% accurate, 6.6× speedup?

                                                \[\begin{array}{l} \\ \mathsf{fma}\left(\left(x \cdot x\right) \cdot x, -0.16666666666666666, x\right) \cdot \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right) \end{array} \]
                                                (FPCore (x y)
                                                 :precision binary64
                                                 (*
                                                  (fma (* (* x x) x) -0.16666666666666666 x)
                                                  (fma (* y y) 0.16666666666666666 1.0)))
                                                double code(double x, double y) {
                                                	return fma(((x * x) * x), -0.16666666666666666, x) * fma((y * y), 0.16666666666666666, 1.0);
                                                }
                                                
                                                function code(x, y)
                                                	return Float64(fma(Float64(Float64(x * x) * x), -0.16666666666666666, x) * fma(Float64(y * y), 0.16666666666666666, 1.0))
                                                end
                                                
                                                code[x_, y_] := N[(N[(N[(N[(x * x), $MachinePrecision] * x), $MachinePrecision] * -0.16666666666666666 + x), $MachinePrecision] * N[(N[(y * y), $MachinePrecision] * 0.16666666666666666 + 1.0), $MachinePrecision]), $MachinePrecision]
                                                
                                                \begin{array}{l}
                                                
                                                \\
                                                \mathsf{fma}\left(\left(x \cdot x\right) \cdot x, -0.16666666666666666, x\right) \cdot \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right)
                                                \end{array}
                                                
                                                Derivation
                                                1. Initial program 100.0%

                                                  \[\sin x \cdot \frac{\sinh y}{y} \]
                                                2. Add Preprocessing
                                                3. Taylor expanded in y around 0

                                                  \[\leadsto \sin x \cdot \color{blue}{\left(1 + \frac{1}{6} \cdot {y}^{2}\right)} \]
                                                4. Step-by-step derivation
                                                  1. +-commutativeN/A

                                                    \[\leadsto \sin x \cdot \color{blue}{\left(\frac{1}{6} \cdot {y}^{2} + 1\right)} \]
                                                  2. *-commutativeN/A

                                                    \[\leadsto \sin x \cdot \left(\color{blue}{{y}^{2} \cdot \frac{1}{6}} + 1\right) \]
                                                  3. lower-fma.f64N/A

                                                    \[\leadsto \sin x \cdot \color{blue}{\mathsf{fma}\left({y}^{2}, \frac{1}{6}, 1\right)} \]
                                                  4. unpow2N/A

                                                    \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{y \cdot y}, \frac{1}{6}, 1\right) \]
                                                  5. lower-*.f6472.1

                                                    \[\leadsto \sin x \cdot \mathsf{fma}\left(\color{blue}{y \cdot y}, 0.16666666666666666, 1\right) \]
                                                5. Applied rewrites72.1%

                                                  \[\leadsto \sin x \cdot \color{blue}{\mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right)} \]
                                                6. Taylor expanded in x around 0

                                                  \[\leadsto \color{blue}{\left(x \cdot \left(1 + \frac{-1}{6} \cdot {x}^{2}\right)\right)} \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                                7. Step-by-step derivation
                                                  1. +-commutativeN/A

                                                    \[\leadsto \left(x \cdot \color{blue}{\left(\frac{-1}{6} \cdot {x}^{2} + 1\right)}\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                                  2. distribute-lft-inN/A

                                                    \[\leadsto \color{blue}{\left(x \cdot \left(\frac{-1}{6} \cdot {x}^{2}\right) + x \cdot 1\right)} \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                                  3. *-commutativeN/A

                                                    \[\leadsto \left(x \cdot \color{blue}{\left({x}^{2} \cdot \frac{-1}{6}\right)} + x \cdot 1\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                                  4. associate-*r*N/A

                                                    \[\leadsto \left(\color{blue}{\left(x \cdot {x}^{2}\right) \cdot \frac{-1}{6}} + x \cdot 1\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                                  5. *-rgt-identityN/A

                                                    \[\leadsto \left(\left(x \cdot {x}^{2}\right) \cdot \frac{-1}{6} + \color{blue}{x}\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                                  6. lower-fma.f64N/A

                                                    \[\leadsto \color{blue}{\mathsf{fma}\left(x \cdot {x}^{2}, \frac{-1}{6}, x\right)} \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                                  7. unpow2N/A

                                                    \[\leadsto \mathsf{fma}\left(x \cdot \color{blue}{\left(x \cdot x\right)}, \frac{-1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                                  8. cube-unmultN/A

                                                    \[\leadsto \mathsf{fma}\left(\color{blue}{{x}^{3}}, \frac{-1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                                  9. metadata-evalN/A

                                                    \[\leadsto \mathsf{fma}\left({x}^{\color{blue}{\left(2 + 1\right)}}, \frac{-1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                                  10. lower-pow.f64N/A

                                                    \[\leadsto \mathsf{fma}\left(\color{blue}{{x}^{\left(2 + 1\right)}}, \frac{-1}{6}, x\right) \cdot \mathsf{fma}\left(y \cdot y, \frac{1}{6}, 1\right) \]
                                                  11. metadata-eval46.2

                                                    \[\leadsto \mathsf{fma}\left({x}^{\color{blue}{3}}, -0.16666666666666666, x\right) \cdot \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right) \]
                                                8. Applied rewrites46.2%

                                                  \[\leadsto \color{blue}{\mathsf{fma}\left({x}^{3}, -0.16666666666666666, x\right)} \cdot \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right) \]
                                                9. Step-by-step derivation
                                                  1. Applied rewrites46.2%

                                                    \[\leadsto \mathsf{fma}\left(\left(x \cdot x\right) \cdot x, -0.16666666666666666, x\right) \cdot \mathsf{fma}\left(y \cdot y, 0.16666666666666666, 1\right) \]
                                                  2. Add Preprocessing

                                                  Alternative 13: 34.9% accurate, 9.9× speedup?

                                                  \[\begin{array}{l} \\ \mathsf{fma}\left(x \cdot x, x \cdot -0.16666666666666666, x\right) \cdot 1 \end{array} \]
                                                  (FPCore (x y)
                                                   :precision binary64
                                                   (* (fma (* x x) (* x -0.16666666666666666) x) 1.0))
                                                  double code(double x, double y) {
                                                  	return fma((x * x), (x * -0.16666666666666666), x) * 1.0;
                                                  }
                                                  
                                                  function code(x, y)
                                                  	return Float64(fma(Float64(x * x), Float64(x * -0.16666666666666666), x) * 1.0)
                                                  end
                                                  
                                                  code[x_, y_] := N[(N[(N[(x * x), $MachinePrecision] * N[(x * -0.16666666666666666), $MachinePrecision] + x), $MachinePrecision] * 1.0), $MachinePrecision]
                                                  
                                                  \begin{array}{l}
                                                  
                                                  \\
                                                  \mathsf{fma}\left(x \cdot x, x \cdot -0.16666666666666666, x\right) \cdot 1
                                                  \end{array}
                                                  
                                                  Derivation
                                                  1. Initial program 100.0%

                                                    \[\sin x \cdot \frac{\sinh y}{y} \]
                                                  2. Add Preprocessing
                                                  3. Taylor expanded in y around 0

                                                    \[\leadsto \sin x \cdot \color{blue}{1} \]
                                                  4. Step-by-step derivation
                                                    1. Applied rewrites42.3%

                                                      \[\leadsto \sin x \cdot \color{blue}{1} \]
                                                    2. Taylor expanded in x around 0

                                                      \[\leadsto \color{blue}{\left(x \cdot \left(1 + \frac{-1}{6} \cdot {x}^{2}\right)\right)} \cdot 1 \]
                                                    3. Step-by-step derivation
                                                      1. +-commutativeN/A

                                                        \[\leadsto \left(x \cdot \color{blue}{\left(\frac{-1}{6} \cdot {x}^{2} + 1\right)}\right) \cdot 1 \]
                                                      2. distribute-lft-inN/A

                                                        \[\leadsto \color{blue}{\left(x \cdot \left(\frac{-1}{6} \cdot {x}^{2}\right) + x \cdot 1\right)} \cdot 1 \]
                                                      3. *-commutativeN/A

                                                        \[\leadsto \left(x \cdot \color{blue}{\left({x}^{2} \cdot \frac{-1}{6}\right)} + x \cdot 1\right) \cdot 1 \]
                                                      4. associate-*r*N/A

                                                        \[\leadsto \left(\color{blue}{\left(x \cdot {x}^{2}\right) \cdot \frac{-1}{6}} + x \cdot 1\right) \cdot 1 \]
                                                      5. *-rgt-identityN/A

                                                        \[\leadsto \left(\left(x \cdot {x}^{2}\right) \cdot \frac{-1}{6} + \color{blue}{x}\right) \cdot 1 \]
                                                      6. lower-fma.f64N/A

                                                        \[\leadsto \color{blue}{\mathsf{fma}\left(x \cdot {x}^{2}, \frac{-1}{6}, x\right)} \cdot 1 \]
                                                      7. unpow2N/A

                                                        \[\leadsto \mathsf{fma}\left(x \cdot \color{blue}{\left(x \cdot x\right)}, \frac{-1}{6}, x\right) \cdot 1 \]
                                                      8. cube-unmultN/A

                                                        \[\leadsto \mathsf{fma}\left(\color{blue}{{x}^{3}}, \frac{-1}{6}, x\right) \cdot 1 \]
                                                      9. metadata-evalN/A

                                                        \[\leadsto \mathsf{fma}\left({x}^{\color{blue}{\left(2 + 1\right)}}, \frac{-1}{6}, x\right) \cdot 1 \]
                                                      10. lower-pow.f64N/A

                                                        \[\leadsto \mathsf{fma}\left(\color{blue}{{x}^{\left(2 + 1\right)}}, \frac{-1}{6}, x\right) \cdot 1 \]
                                                      11. metadata-eval28.0

                                                        \[\leadsto \mathsf{fma}\left({x}^{\color{blue}{3}}, -0.16666666666666666, x\right) \cdot 1 \]
                                                    4. Applied rewrites28.0%

                                                      \[\leadsto \color{blue}{\mathsf{fma}\left({x}^{3}, -0.16666666666666666, x\right)} \cdot 1 \]
                                                    5. Step-by-step derivation
                                                      1. Applied rewrites28.0%

                                                        \[\leadsto \mathsf{fma}\left(x \cdot x, \color{blue}{x \cdot -0.16666666666666666}, x\right) \cdot 1 \]
                                                      2. Add Preprocessing

                                                      Reproduce

                                                      ?
                                                      herbie shell --seed 2025017 
                                                      (FPCore (x y)
                                                        :name "Linear.Quaternion:$ccos from linear-1.19.1.3"
                                                        :precision binary64
                                                        (* (sin x) (/ (sinh y) y)))