Hyperbolic sine

Percentage Accurate: 53.3% → 100.0%
Time: 7.7s
Alternatives: 12
Speedup: 9.9×

Specification

?
\[\begin{array}{l} \\ \frac{e^{x} - e^{-x}}{2} \end{array} \]
(FPCore (x) :precision binary64 (/ (- (exp x) (exp (- x))) 2.0))
double code(double x) {
	return (exp(x) - exp(-x)) / 2.0;
}
real(8) function code(x)
    real(8), intent (in) :: x
    code = (exp(x) - exp(-x)) / 2.0d0
end function
public static double code(double x) {
	return (Math.exp(x) - Math.exp(-x)) / 2.0;
}
def code(x):
	return (math.exp(x) - math.exp(-x)) / 2.0
function code(x)
	return Float64(Float64(exp(x) - exp(Float64(-x))) / 2.0)
end
function tmp = code(x)
	tmp = (exp(x) - exp(-x)) / 2.0;
end
code[x_] := N[(N[(N[Exp[x], $MachinePrecision] - N[Exp[(-x)], $MachinePrecision]), $MachinePrecision] / 2.0), $MachinePrecision]
\begin{array}{l}

\\
\frac{e^{x} - e^{-x}}{2}
\end{array}

Sampling outcomes in binary64 precision:

Local Percentage Accuracy vs ?

The average percentage accuracy by input value. Horizontal axis shows value of an input variable; the variable is choosen in the title. Vertical axis is accuracy; higher is better. Red represent the original program, while blue represents Herbie's suggestion. These can be toggled with buttons below the plot. The line is an average while dots represent individual samples.

Accuracy vs Speed?

Herbie found 12 alternatives:

AlternativeAccuracySpeedup
The accuracy (vertical axis) and speed (horizontal axis) of each alternatives. Up and to the right is better. The red square shows the initial program, and each blue circle shows an alternative.The line shows the best available speed-accuracy tradeoffs.

Initial Program: 53.3% accurate, 1.0× speedup?

\[\begin{array}{l} \\ \frac{e^{x} - e^{-x}}{2} \end{array} \]
(FPCore (x) :precision binary64 (/ (- (exp x) (exp (- x))) 2.0))
double code(double x) {
	return (exp(x) - exp(-x)) / 2.0;
}
real(8) function code(x)
    real(8), intent (in) :: x
    code = (exp(x) - exp(-x)) / 2.0d0
end function
public static double code(double x) {
	return (Math.exp(x) - Math.exp(-x)) / 2.0;
}
def code(x):
	return (math.exp(x) - math.exp(-x)) / 2.0
function code(x)
	return Float64(Float64(exp(x) - exp(Float64(-x))) / 2.0)
end
function tmp = code(x)
	tmp = (exp(x) - exp(-x)) / 2.0;
end
code[x_] := N[(N[(N[Exp[x], $MachinePrecision] - N[Exp[(-x)], $MachinePrecision]), $MachinePrecision] / 2.0), $MachinePrecision]
\begin{array}{l}

\\
\frac{e^{x} - e^{-x}}{2}
\end{array}

Alternative 1: 100.0% accurate, 2.1× speedup?

\[\begin{array}{l} \\ \sinh x \end{array} \]
(FPCore (x) :precision binary64 (sinh x))
double code(double x) {
	return sinh(x);
}
real(8) function code(x)
    real(8), intent (in) :: x
    code = sinh(x)
end function
public static double code(double x) {
	return Math.sinh(x);
}
def code(x):
	return math.sinh(x)
function code(x)
	return sinh(x)
end
function tmp = code(x)
	tmp = sinh(x);
end
code[x_] := N[Sinh[x], $MachinePrecision]
\begin{array}{l}

\\
\sinh x
\end{array}
Derivation
  1. Initial program 57.9%

    \[\frac{e^{x} - e^{-x}}{2} \]
  2. Add Preprocessing
  3. Step-by-step derivation
    1. lift--.f64N/A

      \[\leadsto \frac{\color{blue}{e^{x} - e^{-x}}}{2} \]
    2. lift-exp.f64N/A

      \[\leadsto \frac{\color{blue}{e^{x}} - e^{-x}}{2} \]
    3. lift-exp.f64N/A

      \[\leadsto \frac{e^{x} - \color{blue}{e^{-x}}}{2} \]
    4. lift-neg.f64N/A

      \[\leadsto \frac{e^{x} - e^{\color{blue}{\mathsf{neg}\left(x\right)}}}{2} \]
    5. sinh-undefN/A

      \[\leadsto \frac{\color{blue}{2 \cdot \sinh x}}{2} \]
    6. *-commutativeN/A

      \[\leadsto \frac{\color{blue}{\sinh x \cdot 2}}{2} \]
    7. lower-*.f64N/A

      \[\leadsto \frac{\color{blue}{\sinh x \cdot 2}}{2} \]
    8. lower-sinh.f64100.0

      \[\leadsto \frac{\color{blue}{\sinh x} \cdot 2}{2} \]
  4. Applied rewrites100.0%

    \[\leadsto \frac{\color{blue}{\sinh x \cdot 2}}{2} \]
  5. Step-by-step derivation
    1. lift-/.f64N/A

      \[\leadsto \color{blue}{\frac{\sinh x \cdot 2}{2}} \]
    2. lift-*.f64N/A

      \[\leadsto \frac{\color{blue}{\sinh x \cdot 2}}{2} \]
    3. associate-/l*N/A

      \[\leadsto \color{blue}{\sinh x \cdot \frac{2}{2}} \]
    4. metadata-evalN/A

      \[\leadsto \sinh x \cdot \color{blue}{1} \]
    5. *-rgt-identity100.0

      \[\leadsto \color{blue}{\sinh x} \]
  6. Applied rewrites100.0%

    \[\leadsto \color{blue}{\sinh x} \]
  7. Add Preprocessing

Alternative 2: 87.1% accurate, 0.9× speedup?

\[\begin{array}{l} \\ \begin{array}{l} \mathbf{if}\;e^{x} - e^{-x} \leq 0.002:\\ \;\;\;\;\mathsf{fma}\left(0.3333333333333333 \cdot \left(x \cdot x\right), x, 2 \cdot x\right) \cdot 0.5\\ \mathbf{else}:\\ \;\;\;\;\left(\left(\left(\mathsf{fma}\left(0.016666666666666666, x \cdot x, 0.3333333333333333\right) \cdot x\right) \cdot x\right) \cdot x\right) \cdot 0.5\\ \end{array} \end{array} \]
(FPCore (x)
 :precision binary64
 (if (<= (- (exp x) (exp (- x))) 0.002)
   (* (fma (* 0.3333333333333333 (* x x)) x (* 2.0 x)) 0.5)
   (*
    (* (* (* (fma 0.016666666666666666 (* x x) 0.3333333333333333) x) x) x)
    0.5)))
double code(double x) {
	double tmp;
	if ((exp(x) - exp(-x)) <= 0.002) {
		tmp = fma((0.3333333333333333 * (x * x)), x, (2.0 * x)) * 0.5;
	} else {
		tmp = (((fma(0.016666666666666666, (x * x), 0.3333333333333333) * x) * x) * x) * 0.5;
	}
	return tmp;
}
function code(x)
	tmp = 0.0
	if (Float64(exp(x) - exp(Float64(-x))) <= 0.002)
		tmp = Float64(fma(Float64(0.3333333333333333 * Float64(x * x)), x, Float64(2.0 * x)) * 0.5);
	else
		tmp = Float64(Float64(Float64(Float64(fma(0.016666666666666666, Float64(x * x), 0.3333333333333333) * x) * x) * x) * 0.5);
	end
	return tmp
end
code[x_] := If[LessEqual[N[(N[Exp[x], $MachinePrecision] - N[Exp[(-x)], $MachinePrecision]), $MachinePrecision], 0.002], N[(N[(N[(0.3333333333333333 * N[(x * x), $MachinePrecision]), $MachinePrecision] * x + N[(2.0 * x), $MachinePrecision]), $MachinePrecision] * 0.5), $MachinePrecision], N[(N[(N[(N[(N[(0.016666666666666666 * N[(x * x), $MachinePrecision] + 0.3333333333333333), $MachinePrecision] * x), $MachinePrecision] * x), $MachinePrecision] * x), $MachinePrecision] * 0.5), $MachinePrecision]]
\begin{array}{l}

\\
\begin{array}{l}
\mathbf{if}\;e^{x} - e^{-x} \leq 0.002:\\
\;\;\;\;\mathsf{fma}\left(0.3333333333333333 \cdot \left(x \cdot x\right), x, 2 \cdot x\right) \cdot 0.5\\

\mathbf{else}:\\
\;\;\;\;\left(\left(\left(\mathsf{fma}\left(0.016666666666666666, x \cdot x, 0.3333333333333333\right) \cdot x\right) \cdot x\right) \cdot x\right) \cdot 0.5\\


\end{array}
\end{array}
Derivation
  1. Split input into 2 regimes
  2. if (-.f64 (exp.f64 x) (exp.f64 (neg.f64 x))) < 2e-3

    1. Initial program 44.4%

      \[\frac{e^{x} - e^{-x}}{2} \]
    2. Add Preprocessing
    3. Taylor expanded in x around 0

      \[\leadsto \frac{\color{blue}{x \cdot \left(2 + \frac{1}{3} \cdot {x}^{2}\right)}}{2} \]
    4. Step-by-step derivation
      1. *-commutativeN/A

        \[\leadsto \frac{\color{blue}{\left(2 + \frac{1}{3} \cdot {x}^{2}\right) \cdot x}}{2} \]
      2. lower-*.f64N/A

        \[\leadsto \frac{\color{blue}{\left(2 + \frac{1}{3} \cdot {x}^{2}\right) \cdot x}}{2} \]
      3. +-commutativeN/A

        \[\leadsto \frac{\color{blue}{\left(\frac{1}{3} \cdot {x}^{2} + 2\right)} \cdot x}{2} \]
      4. *-commutativeN/A

        \[\leadsto \frac{\left(\color{blue}{{x}^{2} \cdot \frac{1}{3}} + 2\right) \cdot x}{2} \]
      5. lower-fma.f64N/A

        \[\leadsto \frac{\color{blue}{\mathsf{fma}\left({x}^{2}, \frac{1}{3}, 2\right)} \cdot x}{2} \]
      6. unpow2N/A

        \[\leadsto \frac{\mathsf{fma}\left(\color{blue}{x \cdot x}, \frac{1}{3}, 2\right) \cdot x}{2} \]
      7. lower-*.f6485.3

        \[\leadsto \frac{\mathsf{fma}\left(\color{blue}{x \cdot x}, 0.3333333333333333, 2\right) \cdot x}{2} \]
    5. Applied rewrites85.3%

      \[\leadsto \frac{\color{blue}{\mathsf{fma}\left(x \cdot x, 0.3333333333333333, 2\right) \cdot x}}{2} \]
    6. Step-by-step derivation
      1. Applied rewrites85.4%

        \[\leadsto \frac{\mathsf{fma}\left(0.3333333333333333 \cdot \left(x \cdot x\right), \color{blue}{x}, 2 \cdot x\right)}{2} \]
      2. Step-by-step derivation
        1. lift-/.f64N/A

          \[\leadsto \color{blue}{\frac{\mathsf{fma}\left(\frac{1}{3} \cdot \left(x \cdot x\right), x, 2 \cdot x\right)}{2}} \]
        2. div-invN/A

          \[\leadsto \color{blue}{\mathsf{fma}\left(\frac{1}{3} \cdot \left(x \cdot x\right), x, 2 \cdot x\right) \cdot \frac{1}{2}} \]
        3. metadata-evalN/A

          \[\leadsto \mathsf{fma}\left(\frac{1}{3} \cdot \left(x \cdot x\right), x, 2 \cdot x\right) \cdot \color{blue}{\frac{1}{2}} \]
        4. lower-*.f6485.4

          \[\leadsto \color{blue}{\mathsf{fma}\left(0.3333333333333333 \cdot \left(x \cdot x\right), x, 2 \cdot x\right) \cdot 0.5} \]
      3. Applied rewrites85.3%

        \[\leadsto \color{blue}{\left(\mathsf{fma}\left(x \cdot x, 0.3333333333333333, 2\right) \cdot x\right) \cdot 0.5} \]
      4. Step-by-step derivation
        1. Applied rewrites85.4%

          \[\leadsto \mathsf{fma}\left(0.3333333333333333 \cdot \left(x \cdot x\right), \color{blue}{x}, 2 \cdot x\right) \cdot 0.5 \]

        if 2e-3 < (-.f64 (exp.f64 x) (exp.f64 (neg.f64 x)))

        1. Initial program 100.0%

          \[\frac{e^{x} - e^{-x}}{2} \]
        2. Add Preprocessing
        3. Taylor expanded in x around 0

          \[\leadsto \frac{\color{blue}{x \cdot \left(2 + {x}^{2} \cdot \left(\frac{1}{3} + \frac{1}{60} \cdot {x}^{2}\right)\right)}}{2} \]
        4. Step-by-step derivation
          1. *-commutativeN/A

            \[\leadsto \frac{\color{blue}{\left(2 + {x}^{2} \cdot \left(\frac{1}{3} + \frac{1}{60} \cdot {x}^{2}\right)\right) \cdot x}}{2} \]
          2. lower-*.f64N/A

            \[\leadsto \frac{\color{blue}{\left(2 + {x}^{2} \cdot \left(\frac{1}{3} + \frac{1}{60} \cdot {x}^{2}\right)\right) \cdot x}}{2} \]
          3. +-commutativeN/A

            \[\leadsto \frac{\color{blue}{\left({x}^{2} \cdot \left(\frac{1}{3} + \frac{1}{60} \cdot {x}^{2}\right) + 2\right)} \cdot x}{2} \]
          4. *-commutativeN/A

            \[\leadsto \frac{\left(\color{blue}{\left(\frac{1}{3} + \frac{1}{60} \cdot {x}^{2}\right) \cdot {x}^{2}} + 2\right) \cdot x}{2} \]
          5. lower-fma.f64N/A

            \[\leadsto \frac{\color{blue}{\mathsf{fma}\left(\frac{1}{3} + \frac{1}{60} \cdot {x}^{2}, {x}^{2}, 2\right)} \cdot x}{2} \]
          6. +-commutativeN/A

            \[\leadsto \frac{\mathsf{fma}\left(\color{blue}{\frac{1}{60} \cdot {x}^{2} + \frac{1}{3}}, {x}^{2}, 2\right) \cdot x}{2} \]
          7. lower-fma.f64N/A

            \[\leadsto \frac{\mathsf{fma}\left(\color{blue}{\mathsf{fma}\left(\frac{1}{60}, {x}^{2}, \frac{1}{3}\right)}, {x}^{2}, 2\right) \cdot x}{2} \]
          8. unpow2N/A

            \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{60}, \color{blue}{x \cdot x}, \frac{1}{3}\right), {x}^{2}, 2\right) \cdot x}{2} \]
          9. lower-*.f64N/A

            \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{60}, \color{blue}{x \cdot x}, \frac{1}{3}\right), {x}^{2}, 2\right) \cdot x}{2} \]
          10. unpow2N/A

            \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{60}, x \cdot x, \frac{1}{3}\right), \color{blue}{x \cdot x}, 2\right) \cdot x}{2} \]
          11. lower-*.f6484.7

            \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(0.016666666666666666, x \cdot x, 0.3333333333333333\right), \color{blue}{x \cdot x}, 2\right) \cdot x}{2} \]
        5. Applied rewrites84.7%

          \[\leadsto \frac{\color{blue}{\mathsf{fma}\left(\mathsf{fma}\left(0.016666666666666666, x \cdot x, 0.3333333333333333\right), x \cdot x, 2\right) \cdot x}}{2} \]
        6. Step-by-step derivation
          1. lift-/.f64N/A

            \[\leadsto \color{blue}{\frac{\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{60}, x \cdot x, \frac{1}{3}\right), x \cdot x, 2\right) \cdot x}{2}} \]
          2. div-invN/A

            \[\leadsto \color{blue}{\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{60}, x \cdot x, \frac{1}{3}\right), x \cdot x, 2\right) \cdot x\right) \cdot \frac{1}{2}} \]
          3. lower-*.f64N/A

            \[\leadsto \color{blue}{\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{60}, x \cdot x, \frac{1}{3}\right), x \cdot x, 2\right) \cdot x\right) \cdot \frac{1}{2}} \]
          4. metadata-eval84.7

            \[\leadsto \left(\mathsf{fma}\left(\mathsf{fma}\left(0.016666666666666666, x \cdot x, 0.3333333333333333\right), x \cdot x, 2\right) \cdot x\right) \cdot \color{blue}{0.5} \]
        7. Applied rewrites84.7%

          \[\leadsto \color{blue}{\left(\mathsf{fma}\left(\mathsf{fma}\left(0.016666666666666666, x \cdot x, 0.3333333333333333\right), x \cdot x, 2\right) \cdot x\right) \cdot 0.5} \]
        8. Taylor expanded in x around inf

          \[\leadsto \left(\left({x}^{4} \cdot \left(\frac{1}{60} + \frac{1}{3} \cdot \frac{1}{{x}^{2}}\right)\right) \cdot x\right) \cdot \frac{1}{2} \]
        9. Step-by-step derivation
          1. Applied rewrites84.7%

            \[\leadsto \left(\left(\left(\mathsf{fma}\left(0.016666666666666666, x \cdot x, 0.3333333333333333\right) \cdot x\right) \cdot x\right) \cdot x\right) \cdot 0.5 \]
        10. Recombined 2 regimes into one program.
        11. Add Preprocessing

        Alternative 3: 93.5% accurate, 3.9× speedup?

        \[\begin{array}{l} \\ \frac{\mathsf{fma}\left(\left(\mathsf{fma}\left(\mathsf{fma}\left(0.0003968253968253968, x \cdot x, 0.016666666666666666\right), x \cdot x, 0.3333333333333333\right) \cdot x\right) \cdot x, x, 2 \cdot x\right)}{2} \end{array} \]
        (FPCore (x)
         :precision binary64
         (/
          (fma
           (*
            (*
             (fma
              (fma 0.0003968253968253968 (* x x) 0.016666666666666666)
              (* x x)
              0.3333333333333333)
             x)
            x)
           x
           (* 2.0 x))
          2.0))
        double code(double x) {
        	return fma(((fma(fma(0.0003968253968253968, (x * x), 0.016666666666666666), (x * x), 0.3333333333333333) * x) * x), x, (2.0 * x)) / 2.0;
        }
        
        function code(x)
        	return Float64(fma(Float64(Float64(fma(fma(0.0003968253968253968, Float64(x * x), 0.016666666666666666), Float64(x * x), 0.3333333333333333) * x) * x), x, Float64(2.0 * x)) / 2.0)
        end
        
        code[x_] := N[(N[(N[(N[(N[(N[(0.0003968253968253968 * N[(x * x), $MachinePrecision] + 0.016666666666666666), $MachinePrecision] * N[(x * x), $MachinePrecision] + 0.3333333333333333), $MachinePrecision] * x), $MachinePrecision] * x), $MachinePrecision] * x + N[(2.0 * x), $MachinePrecision]), $MachinePrecision] / 2.0), $MachinePrecision]
        
        \begin{array}{l}
        
        \\
        \frac{\mathsf{fma}\left(\left(\mathsf{fma}\left(\mathsf{fma}\left(0.0003968253968253968, x \cdot x, 0.016666666666666666\right), x \cdot x, 0.3333333333333333\right) \cdot x\right) \cdot x, x, 2 \cdot x\right)}{2}
        \end{array}
        
        Derivation
        1. Initial program 57.9%

          \[\frac{e^{x} - e^{-x}}{2} \]
        2. Add Preprocessing
        3. Taylor expanded in x around 0

          \[\leadsto \frac{\color{blue}{x \cdot \left(2 + {x}^{2} \cdot \left(\frac{1}{3} + {x}^{2} \cdot \left(\frac{1}{60} + \frac{1}{2520} \cdot {x}^{2}\right)\right)\right)}}{2} \]
        4. Step-by-step derivation
          1. *-commutativeN/A

            \[\leadsto \frac{\color{blue}{\left(2 + {x}^{2} \cdot \left(\frac{1}{3} + {x}^{2} \cdot \left(\frac{1}{60} + \frac{1}{2520} \cdot {x}^{2}\right)\right)\right) \cdot x}}{2} \]
          2. lower-*.f64N/A

            \[\leadsto \frac{\color{blue}{\left(2 + {x}^{2} \cdot \left(\frac{1}{3} + {x}^{2} \cdot \left(\frac{1}{60} + \frac{1}{2520} \cdot {x}^{2}\right)\right)\right) \cdot x}}{2} \]
          3. +-commutativeN/A

            \[\leadsto \frac{\color{blue}{\left({x}^{2} \cdot \left(\frac{1}{3} + {x}^{2} \cdot \left(\frac{1}{60} + \frac{1}{2520} \cdot {x}^{2}\right)\right) + 2\right)} \cdot x}{2} \]
          4. *-commutativeN/A

            \[\leadsto \frac{\left(\color{blue}{\left(\frac{1}{3} + {x}^{2} \cdot \left(\frac{1}{60} + \frac{1}{2520} \cdot {x}^{2}\right)\right) \cdot {x}^{2}} + 2\right) \cdot x}{2} \]
          5. lower-fma.f64N/A

            \[\leadsto \frac{\color{blue}{\mathsf{fma}\left(\frac{1}{3} + {x}^{2} \cdot \left(\frac{1}{60} + \frac{1}{2520} \cdot {x}^{2}\right), {x}^{2}, 2\right)} \cdot x}{2} \]
          6. +-commutativeN/A

            \[\leadsto \frac{\mathsf{fma}\left(\color{blue}{{x}^{2} \cdot \left(\frac{1}{60} + \frac{1}{2520} \cdot {x}^{2}\right) + \frac{1}{3}}, {x}^{2}, 2\right) \cdot x}{2} \]
          7. *-commutativeN/A

            \[\leadsto \frac{\mathsf{fma}\left(\color{blue}{\left(\frac{1}{60} + \frac{1}{2520} \cdot {x}^{2}\right) \cdot {x}^{2}} + \frac{1}{3}, {x}^{2}, 2\right) \cdot x}{2} \]
          8. lower-fma.f64N/A

            \[\leadsto \frac{\mathsf{fma}\left(\color{blue}{\mathsf{fma}\left(\frac{1}{60} + \frac{1}{2520} \cdot {x}^{2}, {x}^{2}, \frac{1}{3}\right)}, {x}^{2}, 2\right) \cdot x}{2} \]
          9. +-commutativeN/A

            \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(\color{blue}{\frac{1}{2520} \cdot {x}^{2} + \frac{1}{60}}, {x}^{2}, \frac{1}{3}\right), {x}^{2}, 2\right) \cdot x}{2} \]
          10. lower-fma.f64N/A

            \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(\color{blue}{\mathsf{fma}\left(\frac{1}{2520}, {x}^{2}, \frac{1}{60}\right)}, {x}^{2}, \frac{1}{3}\right), {x}^{2}, 2\right) \cdot x}{2} \]
          11. unpow2N/A

            \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{2520}, \color{blue}{x \cdot x}, \frac{1}{60}\right), {x}^{2}, \frac{1}{3}\right), {x}^{2}, 2\right) \cdot x}{2} \]
          12. lower-*.f64N/A

            \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{2520}, \color{blue}{x \cdot x}, \frac{1}{60}\right), {x}^{2}, \frac{1}{3}\right), {x}^{2}, 2\right) \cdot x}{2} \]
          13. unpow2N/A

            \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{2520}, x \cdot x, \frac{1}{60}\right), \color{blue}{x \cdot x}, \frac{1}{3}\right), {x}^{2}, 2\right) \cdot x}{2} \]
          14. lower-*.f64N/A

            \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{2520}, x \cdot x, \frac{1}{60}\right), \color{blue}{x \cdot x}, \frac{1}{3}\right), {x}^{2}, 2\right) \cdot x}{2} \]
          15. unpow2N/A

            \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{2520}, x \cdot x, \frac{1}{60}\right), x \cdot x, \frac{1}{3}\right), \color{blue}{x \cdot x}, 2\right) \cdot x}{2} \]
          16. lower-*.f6491.7

            \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(0.0003968253968253968, x \cdot x, 0.016666666666666666\right), x \cdot x, 0.3333333333333333\right), \color{blue}{x \cdot x}, 2\right) \cdot x}{2} \]
        5. Applied rewrites91.7%

          \[\leadsto \frac{\color{blue}{\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(0.0003968253968253968, x \cdot x, 0.016666666666666666\right), x \cdot x, 0.3333333333333333\right), x \cdot x, 2\right) \cdot x}}{2} \]
        6. Step-by-step derivation
          1. Applied rewrites91.7%

            \[\leadsto \frac{\mathsf{fma}\left(\left(\mathsf{fma}\left(\mathsf{fma}\left(0.0003968253968253968, x \cdot x, 0.016666666666666666\right), x \cdot x, 0.3333333333333333\right) \cdot x\right) \cdot x, \color{blue}{x}, 2 \cdot x\right)}{2} \]
          2. Add Preprocessing

          Alternative 4: 93.5% accurate, 4.9× speedup?

          \[\begin{array}{l} \\ 0.5 \cdot \left(\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(x \cdot x, 0.0003968253968253968, 0.016666666666666666\right), x \cdot x, 0.3333333333333333\right) \cdot x, x, 2\right) \cdot x\right) \end{array} \]
          (FPCore (x)
           :precision binary64
           (*
            0.5
            (*
             (fma
              (*
               (fma
                (fma (* x x) 0.0003968253968253968 0.016666666666666666)
                (* x x)
                0.3333333333333333)
               x)
              x
              2.0)
             x)))
          double code(double x) {
          	return 0.5 * (fma((fma(fma((x * x), 0.0003968253968253968, 0.016666666666666666), (x * x), 0.3333333333333333) * x), x, 2.0) * x);
          }
          
          function code(x)
          	return Float64(0.5 * Float64(fma(Float64(fma(fma(Float64(x * x), 0.0003968253968253968, 0.016666666666666666), Float64(x * x), 0.3333333333333333) * x), x, 2.0) * x))
          end
          
          code[x_] := N[(0.5 * N[(N[(N[(N[(N[(N[(x * x), $MachinePrecision] * 0.0003968253968253968 + 0.016666666666666666), $MachinePrecision] * N[(x * x), $MachinePrecision] + 0.3333333333333333), $MachinePrecision] * x), $MachinePrecision] * x + 2.0), $MachinePrecision] * x), $MachinePrecision]), $MachinePrecision]
          
          \begin{array}{l}
          
          \\
          0.5 \cdot \left(\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(x \cdot x, 0.0003968253968253968, 0.016666666666666666\right), x \cdot x, 0.3333333333333333\right) \cdot x, x, 2\right) \cdot x\right)
          \end{array}
          
          Derivation
          1. Initial program 57.9%

            \[\frac{e^{x} - e^{-x}}{2} \]
          2. Add Preprocessing
          3. Taylor expanded in x around 0

            \[\leadsto \frac{\color{blue}{x \cdot \left(2 + {x}^{2} \cdot \left(\frac{1}{3} + {x}^{2} \cdot \left(\frac{1}{60} + \frac{1}{2520} \cdot {x}^{2}\right)\right)\right)}}{2} \]
          4. Step-by-step derivation
            1. *-commutativeN/A

              \[\leadsto \frac{\color{blue}{\left(2 + {x}^{2} \cdot \left(\frac{1}{3} + {x}^{2} \cdot \left(\frac{1}{60} + \frac{1}{2520} \cdot {x}^{2}\right)\right)\right) \cdot x}}{2} \]
            2. lower-*.f64N/A

              \[\leadsto \frac{\color{blue}{\left(2 + {x}^{2} \cdot \left(\frac{1}{3} + {x}^{2} \cdot \left(\frac{1}{60} + \frac{1}{2520} \cdot {x}^{2}\right)\right)\right) \cdot x}}{2} \]
            3. +-commutativeN/A

              \[\leadsto \frac{\color{blue}{\left({x}^{2} \cdot \left(\frac{1}{3} + {x}^{2} \cdot \left(\frac{1}{60} + \frac{1}{2520} \cdot {x}^{2}\right)\right) + 2\right)} \cdot x}{2} \]
            4. *-commutativeN/A

              \[\leadsto \frac{\left(\color{blue}{\left(\frac{1}{3} + {x}^{2} \cdot \left(\frac{1}{60} + \frac{1}{2520} \cdot {x}^{2}\right)\right) \cdot {x}^{2}} + 2\right) \cdot x}{2} \]
            5. lower-fma.f64N/A

              \[\leadsto \frac{\color{blue}{\mathsf{fma}\left(\frac{1}{3} + {x}^{2} \cdot \left(\frac{1}{60} + \frac{1}{2520} \cdot {x}^{2}\right), {x}^{2}, 2\right)} \cdot x}{2} \]
            6. +-commutativeN/A

              \[\leadsto \frac{\mathsf{fma}\left(\color{blue}{{x}^{2} \cdot \left(\frac{1}{60} + \frac{1}{2520} \cdot {x}^{2}\right) + \frac{1}{3}}, {x}^{2}, 2\right) \cdot x}{2} \]
            7. *-commutativeN/A

              \[\leadsto \frac{\mathsf{fma}\left(\color{blue}{\left(\frac{1}{60} + \frac{1}{2520} \cdot {x}^{2}\right) \cdot {x}^{2}} + \frac{1}{3}, {x}^{2}, 2\right) \cdot x}{2} \]
            8. lower-fma.f64N/A

              \[\leadsto \frac{\mathsf{fma}\left(\color{blue}{\mathsf{fma}\left(\frac{1}{60} + \frac{1}{2520} \cdot {x}^{2}, {x}^{2}, \frac{1}{3}\right)}, {x}^{2}, 2\right) \cdot x}{2} \]
            9. +-commutativeN/A

              \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(\color{blue}{\frac{1}{2520} \cdot {x}^{2} + \frac{1}{60}}, {x}^{2}, \frac{1}{3}\right), {x}^{2}, 2\right) \cdot x}{2} \]
            10. lower-fma.f64N/A

              \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(\color{blue}{\mathsf{fma}\left(\frac{1}{2520}, {x}^{2}, \frac{1}{60}\right)}, {x}^{2}, \frac{1}{3}\right), {x}^{2}, 2\right) \cdot x}{2} \]
            11. unpow2N/A

              \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{2520}, \color{blue}{x \cdot x}, \frac{1}{60}\right), {x}^{2}, \frac{1}{3}\right), {x}^{2}, 2\right) \cdot x}{2} \]
            12. lower-*.f64N/A

              \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{2520}, \color{blue}{x \cdot x}, \frac{1}{60}\right), {x}^{2}, \frac{1}{3}\right), {x}^{2}, 2\right) \cdot x}{2} \]
            13. unpow2N/A

              \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{2520}, x \cdot x, \frac{1}{60}\right), \color{blue}{x \cdot x}, \frac{1}{3}\right), {x}^{2}, 2\right) \cdot x}{2} \]
            14. lower-*.f64N/A

              \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{2520}, x \cdot x, \frac{1}{60}\right), \color{blue}{x \cdot x}, \frac{1}{3}\right), {x}^{2}, 2\right) \cdot x}{2} \]
            15. unpow2N/A

              \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{2520}, x \cdot x, \frac{1}{60}\right), x \cdot x, \frac{1}{3}\right), \color{blue}{x \cdot x}, 2\right) \cdot x}{2} \]
            16. lower-*.f6491.7

              \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(0.0003968253968253968, x \cdot x, 0.016666666666666666\right), x \cdot x, 0.3333333333333333\right), \color{blue}{x \cdot x}, 2\right) \cdot x}{2} \]
          5. Applied rewrites91.7%

            \[\leadsto \frac{\color{blue}{\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(0.0003968253968253968, x \cdot x, 0.016666666666666666\right), x \cdot x, 0.3333333333333333\right), x \cdot x, 2\right) \cdot x}}{2} \]
          6. Step-by-step derivation
            1. Applied rewrites91.7%

              \[\leadsto \frac{\mathsf{fma}\left(\left(\mathsf{fma}\left(\mathsf{fma}\left(0.0003968253968253968, x \cdot x, 0.016666666666666666\right), x \cdot x, 0.3333333333333333\right) \cdot x\right) \cdot x, \color{blue}{x}, 2 \cdot x\right)}{2} \]
            2. Step-by-step derivation
              1. lift-/.f64N/A

                \[\leadsto \color{blue}{\frac{\mathsf{fma}\left(\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{2520}, x \cdot x, \frac{1}{60}\right), x \cdot x, \frac{1}{3}\right) \cdot x\right) \cdot x, x, 2 \cdot x\right)}{2}} \]
              2. div-invN/A

                \[\leadsto \color{blue}{\mathsf{fma}\left(\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{2520}, x \cdot x, \frac{1}{60}\right), x \cdot x, \frac{1}{3}\right) \cdot x\right) \cdot x, x, 2 \cdot x\right) \cdot \frac{1}{2}} \]
            3. Applied rewrites91.7%

              \[\leadsto \color{blue}{\left(\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(x \cdot x, 0.0003968253968253968, 0.016666666666666666\right), x \cdot x, 0.3333333333333333\right) \cdot x, x, 2\right) \cdot x\right) \cdot 0.5} \]
            4. Final simplification91.7%

              \[\leadsto 0.5 \cdot \left(\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(x \cdot x, 0.0003968253968253968, 0.016666666666666666\right), x \cdot x, 0.3333333333333333\right) \cdot x, x, 2\right) \cdot x\right) \]
            5. Add Preprocessing

            Alternative 5: 93.0% accurate, 5.0× speedup?

            \[\begin{array}{l} \\ \left(\mathsf{fma}\left(\left(\mathsf{fma}\left(x \cdot x, 0.0003968253968253968, 0.016666666666666666\right) \cdot x\right) \cdot x, x \cdot x, 2\right) \cdot x\right) \cdot 0.5 \end{array} \]
            (FPCore (x)
             :precision binary64
             (*
              (*
               (fma
                (* (* (fma (* x x) 0.0003968253968253968 0.016666666666666666) x) x)
                (* x x)
                2.0)
               x)
              0.5))
            double code(double x) {
            	return (fma(((fma((x * x), 0.0003968253968253968, 0.016666666666666666) * x) * x), (x * x), 2.0) * x) * 0.5;
            }
            
            function code(x)
            	return Float64(Float64(fma(Float64(Float64(fma(Float64(x * x), 0.0003968253968253968, 0.016666666666666666) * x) * x), Float64(x * x), 2.0) * x) * 0.5)
            end
            
            code[x_] := N[(N[(N[(N[(N[(N[(N[(x * x), $MachinePrecision] * 0.0003968253968253968 + 0.016666666666666666), $MachinePrecision] * x), $MachinePrecision] * x), $MachinePrecision] * N[(x * x), $MachinePrecision] + 2.0), $MachinePrecision] * x), $MachinePrecision] * 0.5), $MachinePrecision]
            
            \begin{array}{l}
            
            \\
            \left(\mathsf{fma}\left(\left(\mathsf{fma}\left(x \cdot x, 0.0003968253968253968, 0.016666666666666666\right) \cdot x\right) \cdot x, x \cdot x, 2\right) \cdot x\right) \cdot 0.5
            \end{array}
            
            Derivation
            1. Initial program 57.9%

              \[\frac{e^{x} - e^{-x}}{2} \]
            2. Add Preprocessing
            3. Taylor expanded in x around 0

              \[\leadsto \frac{\color{blue}{x \cdot \left(2 + {x}^{2} \cdot \left(\frac{1}{3} + {x}^{2} \cdot \left(\frac{1}{60} + \frac{1}{2520} \cdot {x}^{2}\right)\right)\right)}}{2} \]
            4. Step-by-step derivation
              1. *-commutativeN/A

                \[\leadsto \frac{\color{blue}{\left(2 + {x}^{2} \cdot \left(\frac{1}{3} + {x}^{2} \cdot \left(\frac{1}{60} + \frac{1}{2520} \cdot {x}^{2}\right)\right)\right) \cdot x}}{2} \]
              2. lower-*.f64N/A

                \[\leadsto \frac{\color{blue}{\left(2 + {x}^{2} \cdot \left(\frac{1}{3} + {x}^{2} \cdot \left(\frac{1}{60} + \frac{1}{2520} \cdot {x}^{2}\right)\right)\right) \cdot x}}{2} \]
              3. +-commutativeN/A

                \[\leadsto \frac{\color{blue}{\left({x}^{2} \cdot \left(\frac{1}{3} + {x}^{2} \cdot \left(\frac{1}{60} + \frac{1}{2520} \cdot {x}^{2}\right)\right) + 2\right)} \cdot x}{2} \]
              4. *-commutativeN/A

                \[\leadsto \frac{\left(\color{blue}{\left(\frac{1}{3} + {x}^{2} \cdot \left(\frac{1}{60} + \frac{1}{2520} \cdot {x}^{2}\right)\right) \cdot {x}^{2}} + 2\right) \cdot x}{2} \]
              5. lower-fma.f64N/A

                \[\leadsto \frac{\color{blue}{\mathsf{fma}\left(\frac{1}{3} + {x}^{2} \cdot \left(\frac{1}{60} + \frac{1}{2520} \cdot {x}^{2}\right), {x}^{2}, 2\right)} \cdot x}{2} \]
              6. +-commutativeN/A

                \[\leadsto \frac{\mathsf{fma}\left(\color{blue}{{x}^{2} \cdot \left(\frac{1}{60} + \frac{1}{2520} \cdot {x}^{2}\right) + \frac{1}{3}}, {x}^{2}, 2\right) \cdot x}{2} \]
              7. *-commutativeN/A

                \[\leadsto \frac{\mathsf{fma}\left(\color{blue}{\left(\frac{1}{60} + \frac{1}{2520} \cdot {x}^{2}\right) \cdot {x}^{2}} + \frac{1}{3}, {x}^{2}, 2\right) \cdot x}{2} \]
              8. lower-fma.f64N/A

                \[\leadsto \frac{\mathsf{fma}\left(\color{blue}{\mathsf{fma}\left(\frac{1}{60} + \frac{1}{2520} \cdot {x}^{2}, {x}^{2}, \frac{1}{3}\right)}, {x}^{2}, 2\right) \cdot x}{2} \]
              9. +-commutativeN/A

                \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(\color{blue}{\frac{1}{2520} \cdot {x}^{2} + \frac{1}{60}}, {x}^{2}, \frac{1}{3}\right), {x}^{2}, 2\right) \cdot x}{2} \]
              10. lower-fma.f64N/A

                \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(\color{blue}{\mathsf{fma}\left(\frac{1}{2520}, {x}^{2}, \frac{1}{60}\right)}, {x}^{2}, \frac{1}{3}\right), {x}^{2}, 2\right) \cdot x}{2} \]
              11. unpow2N/A

                \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{2520}, \color{blue}{x \cdot x}, \frac{1}{60}\right), {x}^{2}, \frac{1}{3}\right), {x}^{2}, 2\right) \cdot x}{2} \]
              12. lower-*.f64N/A

                \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{2520}, \color{blue}{x \cdot x}, \frac{1}{60}\right), {x}^{2}, \frac{1}{3}\right), {x}^{2}, 2\right) \cdot x}{2} \]
              13. unpow2N/A

                \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{2520}, x \cdot x, \frac{1}{60}\right), \color{blue}{x \cdot x}, \frac{1}{3}\right), {x}^{2}, 2\right) \cdot x}{2} \]
              14. lower-*.f64N/A

                \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{2520}, x \cdot x, \frac{1}{60}\right), \color{blue}{x \cdot x}, \frac{1}{3}\right), {x}^{2}, 2\right) \cdot x}{2} \]
              15. unpow2N/A

                \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{2520}, x \cdot x, \frac{1}{60}\right), x \cdot x, \frac{1}{3}\right), \color{blue}{x \cdot x}, 2\right) \cdot x}{2} \]
              16. lower-*.f6491.7

                \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(0.0003968253968253968, x \cdot x, 0.016666666666666666\right), x \cdot x, 0.3333333333333333\right), \color{blue}{x \cdot x}, 2\right) \cdot x}{2} \]
            5. Applied rewrites91.7%

              \[\leadsto \frac{\color{blue}{\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(0.0003968253968253968, x \cdot x, 0.016666666666666666\right), x \cdot x, 0.3333333333333333\right), x \cdot x, 2\right) \cdot x}}{2} \]
            6. Taylor expanded in x around inf

              \[\leadsto \frac{\mathsf{fma}\left({x}^{4} \cdot \left(\frac{1}{2520} + \frac{1}{60} \cdot \frac{1}{{x}^{2}}\right), x \cdot x, 2\right) \cdot x}{2} \]
            7. Step-by-step derivation
              1. Applied rewrites91.2%

                \[\leadsto \frac{\mathsf{fma}\left(\left(\mathsf{fma}\left(0.0003968253968253968, x \cdot x, 0.016666666666666666\right) \cdot x\right) \cdot x, x \cdot x, 2\right) \cdot x}{2} \]
              2. Step-by-step derivation
                1. lift-/.f64N/A

                  \[\leadsto \color{blue}{\frac{\mathsf{fma}\left(\left(\mathsf{fma}\left(\frac{1}{2520}, x \cdot x, \frac{1}{60}\right) \cdot x\right) \cdot x, x \cdot x, 2\right) \cdot x}{2}} \]
                2. div-invN/A

                  \[\leadsto \color{blue}{\left(\mathsf{fma}\left(\left(\mathsf{fma}\left(\frac{1}{2520}, x \cdot x, \frac{1}{60}\right) \cdot x\right) \cdot x, x \cdot x, 2\right) \cdot x\right) \cdot \frac{1}{2}} \]
                3. metadata-evalN/A

                  \[\leadsto \left(\mathsf{fma}\left(\left(\mathsf{fma}\left(\frac{1}{2520}, x \cdot x, \frac{1}{60}\right) \cdot x\right) \cdot x, x \cdot x, 2\right) \cdot x\right) \cdot \color{blue}{\frac{1}{2}} \]
                4. lower-*.f6491.2

                  \[\leadsto \color{blue}{\left(\mathsf{fma}\left(\left(\mathsf{fma}\left(0.0003968253968253968, x \cdot x, 0.016666666666666666\right) \cdot x\right) \cdot x, x \cdot x, 2\right) \cdot x\right) \cdot 0.5} \]
              3. Applied rewrites91.2%

                \[\leadsto \color{blue}{\left(\mathsf{fma}\left(\left(\mathsf{fma}\left(x \cdot x, 0.0003968253968253968, 0.016666666666666666\right) \cdot x\right) \cdot x, x \cdot x, 2\right) \cdot x\right) \cdot 0.5} \]
              4. Add Preprocessing

              Alternative 6: 90.5% accurate, 5.7× speedup?

              \[\begin{array}{l} \\ \mathsf{fma}\left(\left(\mathsf{fma}\left(0.016666666666666666, x \cdot x, 0.3333333333333333\right) \cdot x\right) \cdot x, x, 2 \cdot x\right) \cdot 0.5 \end{array} \]
              (FPCore (x)
               :precision binary64
               (*
                (fma
                 (* (* (fma 0.016666666666666666 (* x x) 0.3333333333333333) x) x)
                 x
                 (* 2.0 x))
                0.5))
              double code(double x) {
              	return fma(((fma(0.016666666666666666, (x * x), 0.3333333333333333) * x) * x), x, (2.0 * x)) * 0.5;
              }
              
              function code(x)
              	return Float64(fma(Float64(Float64(fma(0.016666666666666666, Float64(x * x), 0.3333333333333333) * x) * x), x, Float64(2.0 * x)) * 0.5)
              end
              
              code[x_] := N[(N[(N[(N[(N[(0.016666666666666666 * N[(x * x), $MachinePrecision] + 0.3333333333333333), $MachinePrecision] * x), $MachinePrecision] * x), $MachinePrecision] * x + N[(2.0 * x), $MachinePrecision]), $MachinePrecision] * 0.5), $MachinePrecision]
              
              \begin{array}{l}
              
              \\
              \mathsf{fma}\left(\left(\mathsf{fma}\left(0.016666666666666666, x \cdot x, 0.3333333333333333\right) \cdot x\right) \cdot x, x, 2 \cdot x\right) \cdot 0.5
              \end{array}
              
              Derivation
              1. Initial program 57.9%

                \[\frac{e^{x} - e^{-x}}{2} \]
              2. Add Preprocessing
              3. Taylor expanded in x around 0

                \[\leadsto \frac{\color{blue}{x \cdot \left(2 + {x}^{2} \cdot \left(\frac{1}{3} + \frac{1}{60} \cdot {x}^{2}\right)\right)}}{2} \]
              4. Step-by-step derivation
                1. *-commutativeN/A

                  \[\leadsto \frac{\color{blue}{\left(2 + {x}^{2} \cdot \left(\frac{1}{3} + \frac{1}{60} \cdot {x}^{2}\right)\right) \cdot x}}{2} \]
                2. lower-*.f64N/A

                  \[\leadsto \frac{\color{blue}{\left(2 + {x}^{2} \cdot \left(\frac{1}{3} + \frac{1}{60} \cdot {x}^{2}\right)\right) \cdot x}}{2} \]
                3. +-commutativeN/A

                  \[\leadsto \frac{\color{blue}{\left({x}^{2} \cdot \left(\frac{1}{3} + \frac{1}{60} \cdot {x}^{2}\right) + 2\right)} \cdot x}{2} \]
                4. *-commutativeN/A

                  \[\leadsto \frac{\left(\color{blue}{\left(\frac{1}{3} + \frac{1}{60} \cdot {x}^{2}\right) \cdot {x}^{2}} + 2\right) \cdot x}{2} \]
                5. lower-fma.f64N/A

                  \[\leadsto \frac{\color{blue}{\mathsf{fma}\left(\frac{1}{3} + \frac{1}{60} \cdot {x}^{2}, {x}^{2}, 2\right)} \cdot x}{2} \]
                6. +-commutativeN/A

                  \[\leadsto \frac{\mathsf{fma}\left(\color{blue}{\frac{1}{60} \cdot {x}^{2} + \frac{1}{3}}, {x}^{2}, 2\right) \cdot x}{2} \]
                7. lower-fma.f64N/A

                  \[\leadsto \frac{\mathsf{fma}\left(\color{blue}{\mathsf{fma}\left(\frac{1}{60}, {x}^{2}, \frac{1}{3}\right)}, {x}^{2}, 2\right) \cdot x}{2} \]
                8. unpow2N/A

                  \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{60}, \color{blue}{x \cdot x}, \frac{1}{3}\right), {x}^{2}, 2\right) \cdot x}{2} \]
                9. lower-*.f64N/A

                  \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{60}, \color{blue}{x \cdot x}, \frac{1}{3}\right), {x}^{2}, 2\right) \cdot x}{2} \]
                10. unpow2N/A

                  \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{60}, x \cdot x, \frac{1}{3}\right), \color{blue}{x \cdot x}, 2\right) \cdot x}{2} \]
                11. lower-*.f6488.6

                  \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(0.016666666666666666, x \cdot x, 0.3333333333333333\right), \color{blue}{x \cdot x}, 2\right) \cdot x}{2} \]
              5. Applied rewrites88.6%

                \[\leadsto \frac{\color{blue}{\mathsf{fma}\left(\mathsf{fma}\left(0.016666666666666666, x \cdot x, 0.3333333333333333\right), x \cdot x, 2\right) \cdot x}}{2} \]
              6. Step-by-step derivation
                1. lift-/.f64N/A

                  \[\leadsto \color{blue}{\frac{\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{60}, x \cdot x, \frac{1}{3}\right), x \cdot x, 2\right) \cdot x}{2}} \]
                2. div-invN/A

                  \[\leadsto \color{blue}{\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{60}, x \cdot x, \frac{1}{3}\right), x \cdot x, 2\right) \cdot x\right) \cdot \frac{1}{2}} \]
                3. lower-*.f64N/A

                  \[\leadsto \color{blue}{\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{60}, x \cdot x, \frac{1}{3}\right), x \cdot x, 2\right) \cdot x\right) \cdot \frac{1}{2}} \]
                4. metadata-eval88.6

                  \[\leadsto \left(\mathsf{fma}\left(\mathsf{fma}\left(0.016666666666666666, x \cdot x, 0.3333333333333333\right), x \cdot x, 2\right) \cdot x\right) \cdot \color{blue}{0.5} \]
              7. Applied rewrites88.6%

                \[\leadsto \color{blue}{\left(\mathsf{fma}\left(\mathsf{fma}\left(0.016666666666666666, x \cdot x, 0.3333333333333333\right), x \cdot x, 2\right) \cdot x\right) \cdot 0.5} \]
              8. Step-by-step derivation
                1. Applied rewrites88.6%

                  \[\leadsto \mathsf{fma}\left(\left(\mathsf{fma}\left(0.016666666666666666, x \cdot x, 0.3333333333333333\right) \cdot x\right) \cdot x, \color{blue}{x}, 2 \cdot x\right) \cdot 0.5 \]
                2. Add Preprocessing

                Alternative 7: 90.5% accurate, 6.6× speedup?

                \[\begin{array}{l} \\ \left(\mathsf{fma}\left(\mathsf{fma}\left(0.016666666666666666, x \cdot x, 0.3333333333333333\right) \cdot x, x, 2\right) \cdot x\right) \cdot 0.5 \end{array} \]
                (FPCore (x)
                 :precision binary64
                 (*
                  (* (fma (* (fma 0.016666666666666666 (* x x) 0.3333333333333333) x) x 2.0) x)
                  0.5))
                double code(double x) {
                	return (fma((fma(0.016666666666666666, (x * x), 0.3333333333333333) * x), x, 2.0) * x) * 0.5;
                }
                
                function code(x)
                	return Float64(Float64(fma(Float64(fma(0.016666666666666666, Float64(x * x), 0.3333333333333333) * x), x, 2.0) * x) * 0.5)
                end
                
                code[x_] := N[(N[(N[(N[(N[(0.016666666666666666 * N[(x * x), $MachinePrecision] + 0.3333333333333333), $MachinePrecision] * x), $MachinePrecision] * x + 2.0), $MachinePrecision] * x), $MachinePrecision] * 0.5), $MachinePrecision]
                
                \begin{array}{l}
                
                \\
                \left(\mathsf{fma}\left(\mathsf{fma}\left(0.016666666666666666, x \cdot x, 0.3333333333333333\right) \cdot x, x, 2\right) \cdot x\right) \cdot 0.5
                \end{array}
                
                Derivation
                1. Initial program 57.9%

                  \[\frac{e^{x} - e^{-x}}{2} \]
                2. Add Preprocessing
                3. Taylor expanded in x around 0

                  \[\leadsto \frac{\color{blue}{x \cdot \left(2 + {x}^{2} \cdot \left(\frac{1}{3} + \frac{1}{60} \cdot {x}^{2}\right)\right)}}{2} \]
                4. Step-by-step derivation
                  1. *-commutativeN/A

                    \[\leadsto \frac{\color{blue}{\left(2 + {x}^{2} \cdot \left(\frac{1}{3} + \frac{1}{60} \cdot {x}^{2}\right)\right) \cdot x}}{2} \]
                  2. lower-*.f64N/A

                    \[\leadsto \frac{\color{blue}{\left(2 + {x}^{2} \cdot \left(\frac{1}{3} + \frac{1}{60} \cdot {x}^{2}\right)\right) \cdot x}}{2} \]
                  3. +-commutativeN/A

                    \[\leadsto \frac{\color{blue}{\left({x}^{2} \cdot \left(\frac{1}{3} + \frac{1}{60} \cdot {x}^{2}\right) + 2\right)} \cdot x}{2} \]
                  4. *-commutativeN/A

                    \[\leadsto \frac{\left(\color{blue}{\left(\frac{1}{3} + \frac{1}{60} \cdot {x}^{2}\right) \cdot {x}^{2}} + 2\right) \cdot x}{2} \]
                  5. lower-fma.f64N/A

                    \[\leadsto \frac{\color{blue}{\mathsf{fma}\left(\frac{1}{3} + \frac{1}{60} \cdot {x}^{2}, {x}^{2}, 2\right)} \cdot x}{2} \]
                  6. +-commutativeN/A

                    \[\leadsto \frac{\mathsf{fma}\left(\color{blue}{\frac{1}{60} \cdot {x}^{2} + \frac{1}{3}}, {x}^{2}, 2\right) \cdot x}{2} \]
                  7. lower-fma.f64N/A

                    \[\leadsto \frac{\mathsf{fma}\left(\color{blue}{\mathsf{fma}\left(\frac{1}{60}, {x}^{2}, \frac{1}{3}\right)}, {x}^{2}, 2\right) \cdot x}{2} \]
                  8. unpow2N/A

                    \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{60}, \color{blue}{x \cdot x}, \frac{1}{3}\right), {x}^{2}, 2\right) \cdot x}{2} \]
                  9. lower-*.f64N/A

                    \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{60}, \color{blue}{x \cdot x}, \frac{1}{3}\right), {x}^{2}, 2\right) \cdot x}{2} \]
                  10. unpow2N/A

                    \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{60}, x \cdot x, \frac{1}{3}\right), \color{blue}{x \cdot x}, 2\right) \cdot x}{2} \]
                  11. lower-*.f6488.6

                    \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(0.016666666666666666, x \cdot x, 0.3333333333333333\right), \color{blue}{x \cdot x}, 2\right) \cdot x}{2} \]
                5. Applied rewrites88.6%

                  \[\leadsto \frac{\color{blue}{\mathsf{fma}\left(\mathsf{fma}\left(0.016666666666666666, x \cdot x, 0.3333333333333333\right), x \cdot x, 2\right) \cdot x}}{2} \]
                6. Step-by-step derivation
                  1. Applied rewrites88.6%

                    \[\leadsto \frac{\mathsf{fma}\left(\left(\mathsf{fma}\left(0.016666666666666666, x \cdot x, 0.3333333333333333\right) \cdot x\right) \cdot x, \color{blue}{x}, 2 \cdot x\right)}{2} \]
                  2. Step-by-step derivation
                    1. lift-/.f64N/A

                      \[\leadsto \color{blue}{\frac{\mathsf{fma}\left(\left(\mathsf{fma}\left(\frac{1}{60}, x \cdot x, \frac{1}{3}\right) \cdot x\right) \cdot x, x, 2 \cdot x\right)}{2}} \]
                    2. div-invN/A

                      \[\leadsto \color{blue}{\mathsf{fma}\left(\left(\mathsf{fma}\left(\frac{1}{60}, x \cdot x, \frac{1}{3}\right) \cdot x\right) \cdot x, x, 2 \cdot x\right) \cdot \frac{1}{2}} \]
                    3. metadata-evalN/A

                      \[\leadsto \mathsf{fma}\left(\left(\mathsf{fma}\left(\frac{1}{60}, x \cdot x, \frac{1}{3}\right) \cdot x\right) \cdot x, x, 2 \cdot x\right) \cdot \color{blue}{\frac{1}{2}} \]
                    4. lower-*.f6488.6

                      \[\leadsto \color{blue}{\mathsf{fma}\left(\left(\mathsf{fma}\left(0.016666666666666666, x \cdot x, 0.3333333333333333\right) \cdot x\right) \cdot x, x, 2 \cdot x\right) \cdot 0.5} \]
                  3. Applied rewrites88.6%

                    \[\leadsto \color{blue}{\left(\mathsf{fma}\left(\mathsf{fma}\left(0.016666666666666666, x \cdot x, 0.3333333333333333\right) \cdot x, x, 2\right) \cdot x\right) \cdot 0.5} \]
                  4. Add Preprocessing

                  Alternative 8: 90.1% accurate, 6.8× speedup?

                  \[\begin{array}{l} \\ \left(\mathsf{fma}\left(0.016666666666666666 \cdot \left(x \cdot x\right), x \cdot x, 2\right) \cdot x\right) \cdot 0.5 \end{array} \]
                  (FPCore (x)
                   :precision binary64
                   (* (* (fma (* 0.016666666666666666 (* x x)) (* x x) 2.0) x) 0.5))
                  double code(double x) {
                  	return (fma((0.016666666666666666 * (x * x)), (x * x), 2.0) * x) * 0.5;
                  }
                  
                  function code(x)
                  	return Float64(Float64(fma(Float64(0.016666666666666666 * Float64(x * x)), Float64(x * x), 2.0) * x) * 0.5)
                  end
                  
                  code[x_] := N[(N[(N[(N[(0.016666666666666666 * N[(x * x), $MachinePrecision]), $MachinePrecision] * N[(x * x), $MachinePrecision] + 2.0), $MachinePrecision] * x), $MachinePrecision] * 0.5), $MachinePrecision]
                  
                  \begin{array}{l}
                  
                  \\
                  \left(\mathsf{fma}\left(0.016666666666666666 \cdot \left(x \cdot x\right), x \cdot x, 2\right) \cdot x\right) \cdot 0.5
                  \end{array}
                  
                  Derivation
                  1. Initial program 57.9%

                    \[\frac{e^{x} - e^{-x}}{2} \]
                  2. Add Preprocessing
                  3. Taylor expanded in x around 0

                    \[\leadsto \frac{\color{blue}{x \cdot \left(2 + {x}^{2} \cdot \left(\frac{1}{3} + \frac{1}{60} \cdot {x}^{2}\right)\right)}}{2} \]
                  4. Step-by-step derivation
                    1. *-commutativeN/A

                      \[\leadsto \frac{\color{blue}{\left(2 + {x}^{2} \cdot \left(\frac{1}{3} + \frac{1}{60} \cdot {x}^{2}\right)\right) \cdot x}}{2} \]
                    2. lower-*.f64N/A

                      \[\leadsto \frac{\color{blue}{\left(2 + {x}^{2} \cdot \left(\frac{1}{3} + \frac{1}{60} \cdot {x}^{2}\right)\right) \cdot x}}{2} \]
                    3. +-commutativeN/A

                      \[\leadsto \frac{\color{blue}{\left({x}^{2} \cdot \left(\frac{1}{3} + \frac{1}{60} \cdot {x}^{2}\right) + 2\right)} \cdot x}{2} \]
                    4. *-commutativeN/A

                      \[\leadsto \frac{\left(\color{blue}{\left(\frac{1}{3} + \frac{1}{60} \cdot {x}^{2}\right) \cdot {x}^{2}} + 2\right) \cdot x}{2} \]
                    5. lower-fma.f64N/A

                      \[\leadsto \frac{\color{blue}{\mathsf{fma}\left(\frac{1}{3} + \frac{1}{60} \cdot {x}^{2}, {x}^{2}, 2\right)} \cdot x}{2} \]
                    6. +-commutativeN/A

                      \[\leadsto \frac{\mathsf{fma}\left(\color{blue}{\frac{1}{60} \cdot {x}^{2} + \frac{1}{3}}, {x}^{2}, 2\right) \cdot x}{2} \]
                    7. lower-fma.f64N/A

                      \[\leadsto \frac{\mathsf{fma}\left(\color{blue}{\mathsf{fma}\left(\frac{1}{60}, {x}^{2}, \frac{1}{3}\right)}, {x}^{2}, 2\right) \cdot x}{2} \]
                    8. unpow2N/A

                      \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{60}, \color{blue}{x \cdot x}, \frac{1}{3}\right), {x}^{2}, 2\right) \cdot x}{2} \]
                    9. lower-*.f64N/A

                      \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{60}, \color{blue}{x \cdot x}, \frac{1}{3}\right), {x}^{2}, 2\right) \cdot x}{2} \]
                    10. unpow2N/A

                      \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{60}, x \cdot x, \frac{1}{3}\right), \color{blue}{x \cdot x}, 2\right) \cdot x}{2} \]
                    11. lower-*.f6488.6

                      \[\leadsto \frac{\mathsf{fma}\left(\mathsf{fma}\left(0.016666666666666666, x \cdot x, 0.3333333333333333\right), \color{blue}{x \cdot x}, 2\right) \cdot x}{2} \]
                  5. Applied rewrites88.6%

                    \[\leadsto \frac{\color{blue}{\mathsf{fma}\left(\mathsf{fma}\left(0.016666666666666666, x \cdot x, 0.3333333333333333\right), x \cdot x, 2\right) \cdot x}}{2} \]
                  6. Step-by-step derivation
                    1. lift-/.f64N/A

                      \[\leadsto \color{blue}{\frac{\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{60}, x \cdot x, \frac{1}{3}\right), x \cdot x, 2\right) \cdot x}{2}} \]
                    2. div-invN/A

                      \[\leadsto \color{blue}{\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{60}, x \cdot x, \frac{1}{3}\right), x \cdot x, 2\right) \cdot x\right) \cdot \frac{1}{2}} \]
                    3. lower-*.f64N/A

                      \[\leadsto \color{blue}{\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{1}{60}, x \cdot x, \frac{1}{3}\right), x \cdot x, 2\right) \cdot x\right) \cdot \frac{1}{2}} \]
                    4. metadata-eval88.6

                      \[\leadsto \left(\mathsf{fma}\left(\mathsf{fma}\left(0.016666666666666666, x \cdot x, 0.3333333333333333\right), x \cdot x, 2\right) \cdot x\right) \cdot \color{blue}{0.5} \]
                  7. Applied rewrites88.6%

                    \[\leadsto \color{blue}{\left(\mathsf{fma}\left(\mathsf{fma}\left(0.016666666666666666, x \cdot x, 0.3333333333333333\right), x \cdot x, 2\right) \cdot x\right) \cdot 0.5} \]
                  8. Taylor expanded in x around inf

                    \[\leadsto \left(\mathsf{fma}\left(\frac{1}{60} \cdot {x}^{2}, x \cdot x, 2\right) \cdot x\right) \cdot \frac{1}{2} \]
                  9. Step-by-step derivation
                    1. Applied rewrites88.2%

                      \[\leadsto \left(\mathsf{fma}\left(0.016666666666666666 \cdot \left(x \cdot x\right), x \cdot x, 2\right) \cdot x\right) \cdot 0.5 \]
                    2. Add Preprocessing

                    Alternative 9: 68.9% accurate, 8.0× speedup?

                    \[\begin{array}{l} \\ \begin{array}{l} \mathbf{if}\;x \leq 2.4:\\ \;\;\;\;\left(2 \cdot x\right) \cdot 0.5\\ \mathbf{else}:\\ \;\;\;\;\left(\left(0.3333333333333333 \cdot \left(x \cdot x\right)\right) \cdot x\right) \cdot 0.5\\ \end{array} \end{array} \]
                    (FPCore (x)
                     :precision binary64
                     (if (<= x 2.4)
                       (* (* 2.0 x) 0.5)
                       (* (* (* 0.3333333333333333 (* x x)) x) 0.5)))
                    double code(double x) {
                    	double tmp;
                    	if (x <= 2.4) {
                    		tmp = (2.0 * x) * 0.5;
                    	} else {
                    		tmp = ((0.3333333333333333 * (x * x)) * x) * 0.5;
                    	}
                    	return tmp;
                    }
                    
                    real(8) function code(x)
                        real(8), intent (in) :: x
                        real(8) :: tmp
                        if (x <= 2.4d0) then
                            tmp = (2.0d0 * x) * 0.5d0
                        else
                            tmp = ((0.3333333333333333d0 * (x * x)) * x) * 0.5d0
                        end if
                        code = tmp
                    end function
                    
                    public static double code(double x) {
                    	double tmp;
                    	if (x <= 2.4) {
                    		tmp = (2.0 * x) * 0.5;
                    	} else {
                    		tmp = ((0.3333333333333333 * (x * x)) * x) * 0.5;
                    	}
                    	return tmp;
                    }
                    
                    def code(x):
                    	tmp = 0
                    	if x <= 2.4:
                    		tmp = (2.0 * x) * 0.5
                    	else:
                    		tmp = ((0.3333333333333333 * (x * x)) * x) * 0.5
                    	return tmp
                    
                    function code(x)
                    	tmp = 0.0
                    	if (x <= 2.4)
                    		tmp = Float64(Float64(2.0 * x) * 0.5);
                    	else
                    		tmp = Float64(Float64(Float64(0.3333333333333333 * Float64(x * x)) * x) * 0.5);
                    	end
                    	return tmp
                    end
                    
                    function tmp_2 = code(x)
                    	tmp = 0.0;
                    	if (x <= 2.4)
                    		tmp = (2.0 * x) * 0.5;
                    	else
                    		tmp = ((0.3333333333333333 * (x * x)) * x) * 0.5;
                    	end
                    	tmp_2 = tmp;
                    end
                    
                    code[x_] := If[LessEqual[x, 2.4], N[(N[(2.0 * x), $MachinePrecision] * 0.5), $MachinePrecision], N[(N[(N[(0.3333333333333333 * N[(x * x), $MachinePrecision]), $MachinePrecision] * x), $MachinePrecision] * 0.5), $MachinePrecision]]
                    
                    \begin{array}{l}
                    
                    \\
                    \begin{array}{l}
                    \mathbf{if}\;x \leq 2.4:\\
                    \;\;\;\;\left(2 \cdot x\right) \cdot 0.5\\
                    
                    \mathbf{else}:\\
                    \;\;\;\;\left(\left(0.3333333333333333 \cdot \left(x \cdot x\right)\right) \cdot x\right) \cdot 0.5\\
                    
                    
                    \end{array}
                    \end{array}
                    
                    Derivation
                    1. Split input into 2 regimes
                    2. if x < 2.39999999999999991

                      1. Initial program 44.4%

                        \[\frac{e^{x} - e^{-x}}{2} \]
                      2. Add Preprocessing
                      3. Taylor expanded in x around 0

                        \[\leadsto \frac{\color{blue}{2 \cdot x}}{2} \]
                      4. Step-by-step derivation
                        1. *-commutativeN/A

                          \[\leadsto \frac{\color{blue}{x \cdot 2}}{2} \]
                        2. lower-*.f6462.2

                          \[\leadsto \frac{\color{blue}{x \cdot 2}}{2} \]
                      5. Applied rewrites62.2%

                        \[\leadsto \frac{\color{blue}{x \cdot 2}}{2} \]
                      6. Step-by-step derivation
                        1. lift-/.f64N/A

                          \[\leadsto \color{blue}{\frac{x \cdot 2}{2}} \]
                        2. div-invN/A

                          \[\leadsto \color{blue}{\left(x \cdot 2\right) \cdot \frac{1}{2}} \]
                        3. lower-*.f64N/A

                          \[\leadsto \color{blue}{\left(x \cdot 2\right) \cdot \frac{1}{2}} \]
                        4. metadata-evalN/A

                          \[\leadsto \mathsf{Rewrite=>}\left(lower-*.f64, \left(2 \cdot x\right)\right) \cdot \color{blue}{\frac{1}{2}} \]
                      7. Applied rewrites62.2%

                        \[\leadsto \color{blue}{\left(2 \cdot x\right) \cdot 0.5} \]

                      if 2.39999999999999991 < x

                      1. Initial program 100.0%

                        \[\frac{e^{x} - e^{-x}}{2} \]
                      2. Add Preprocessing
                      3. Taylor expanded in x around 0

                        \[\leadsto \frac{\color{blue}{x \cdot \left(2 + \frac{1}{3} \cdot {x}^{2}\right)}}{2} \]
                      4. Step-by-step derivation
                        1. *-commutativeN/A

                          \[\leadsto \frac{\color{blue}{\left(2 + \frac{1}{3} \cdot {x}^{2}\right) \cdot x}}{2} \]
                        2. lower-*.f64N/A

                          \[\leadsto \frac{\color{blue}{\left(2 + \frac{1}{3} \cdot {x}^{2}\right) \cdot x}}{2} \]
                        3. +-commutativeN/A

                          \[\leadsto \frac{\color{blue}{\left(\frac{1}{3} \cdot {x}^{2} + 2\right)} \cdot x}{2} \]
                        4. *-commutativeN/A

                          \[\leadsto \frac{\left(\color{blue}{{x}^{2} \cdot \frac{1}{3}} + 2\right) \cdot x}{2} \]
                        5. lower-fma.f64N/A

                          \[\leadsto \frac{\color{blue}{\mathsf{fma}\left({x}^{2}, \frac{1}{3}, 2\right)} \cdot x}{2} \]
                        6. unpow2N/A

                          \[\leadsto \frac{\mathsf{fma}\left(\color{blue}{x \cdot x}, \frac{1}{3}, 2\right) \cdot x}{2} \]
                        7. lower-*.f6471.1

                          \[\leadsto \frac{\mathsf{fma}\left(\color{blue}{x \cdot x}, 0.3333333333333333, 2\right) \cdot x}{2} \]
                      5. Applied rewrites71.1%

                        \[\leadsto \frac{\color{blue}{\mathsf{fma}\left(x \cdot x, 0.3333333333333333, 2\right) \cdot x}}{2} \]
                      6. Step-by-step derivation
                        1. Applied rewrites71.1%

                          \[\leadsto \frac{\mathsf{fma}\left(0.3333333333333333 \cdot \left(x \cdot x\right), \color{blue}{x}, 2 \cdot x\right)}{2} \]
                        2. Step-by-step derivation
                          1. lift-/.f64N/A

                            \[\leadsto \color{blue}{\frac{\mathsf{fma}\left(\frac{1}{3} \cdot \left(x \cdot x\right), x, 2 \cdot x\right)}{2}} \]
                          2. div-invN/A

                            \[\leadsto \color{blue}{\mathsf{fma}\left(\frac{1}{3} \cdot \left(x \cdot x\right), x, 2 \cdot x\right) \cdot \frac{1}{2}} \]
                          3. metadata-evalN/A

                            \[\leadsto \mathsf{fma}\left(\frac{1}{3} \cdot \left(x \cdot x\right), x, 2 \cdot x\right) \cdot \color{blue}{\frac{1}{2}} \]
                          4. lower-*.f6471.1

                            \[\leadsto \color{blue}{\mathsf{fma}\left(0.3333333333333333 \cdot \left(x \cdot x\right), x, 2 \cdot x\right) \cdot 0.5} \]
                        3. Applied rewrites71.1%

                          \[\leadsto \color{blue}{\left(\mathsf{fma}\left(x \cdot x, 0.3333333333333333, 2\right) \cdot x\right) \cdot 0.5} \]
                        4. Taylor expanded in x around inf

                          \[\leadsto \left(\left(\frac{1}{3} \cdot {x}^{2}\right) \cdot x\right) \cdot \frac{1}{2} \]
                        5. Step-by-step derivation
                          1. Applied rewrites71.1%

                            \[\leadsto \left(\left(\left(x \cdot x\right) \cdot 0.3333333333333333\right) \cdot x\right) \cdot 0.5 \]
                        6. Recombined 2 regimes into one program.
                        7. Final simplification64.4%

                          \[\leadsto \begin{array}{l} \mathbf{if}\;x \leq 2.4:\\ \;\;\;\;\left(2 \cdot x\right) \cdot 0.5\\ \mathbf{else}:\\ \;\;\;\;\left(\left(0.3333333333333333 \cdot \left(x \cdot x\right)\right) \cdot x\right) \cdot 0.5\\ \end{array} \]
                        8. Add Preprocessing

                        Alternative 10: 84.3% accurate, 8.0× speedup?

                        \[\begin{array}{l} \\ \mathsf{fma}\left(0.3333333333333333 \cdot \left(x \cdot x\right), x, 2 \cdot x\right) \cdot 0.5 \end{array} \]
                        (FPCore (x)
                         :precision binary64
                         (* (fma (* 0.3333333333333333 (* x x)) x (* 2.0 x)) 0.5))
                        double code(double x) {
                        	return fma((0.3333333333333333 * (x * x)), x, (2.0 * x)) * 0.5;
                        }
                        
                        function code(x)
                        	return Float64(fma(Float64(0.3333333333333333 * Float64(x * x)), x, Float64(2.0 * x)) * 0.5)
                        end
                        
                        code[x_] := N[(N[(N[(0.3333333333333333 * N[(x * x), $MachinePrecision]), $MachinePrecision] * x + N[(2.0 * x), $MachinePrecision]), $MachinePrecision] * 0.5), $MachinePrecision]
                        
                        \begin{array}{l}
                        
                        \\
                        \mathsf{fma}\left(0.3333333333333333 \cdot \left(x \cdot x\right), x, 2 \cdot x\right) \cdot 0.5
                        \end{array}
                        
                        Derivation
                        1. Initial program 57.9%

                          \[\frac{e^{x} - e^{-x}}{2} \]
                        2. Add Preprocessing
                        3. Taylor expanded in x around 0

                          \[\leadsto \frac{\color{blue}{x \cdot \left(2 + \frac{1}{3} \cdot {x}^{2}\right)}}{2} \]
                        4. Step-by-step derivation
                          1. *-commutativeN/A

                            \[\leadsto \frac{\color{blue}{\left(2 + \frac{1}{3} \cdot {x}^{2}\right) \cdot x}}{2} \]
                          2. lower-*.f64N/A

                            \[\leadsto \frac{\color{blue}{\left(2 + \frac{1}{3} \cdot {x}^{2}\right) \cdot x}}{2} \]
                          3. +-commutativeN/A

                            \[\leadsto \frac{\color{blue}{\left(\frac{1}{3} \cdot {x}^{2} + 2\right)} \cdot x}{2} \]
                          4. *-commutativeN/A

                            \[\leadsto \frac{\left(\color{blue}{{x}^{2} \cdot \frac{1}{3}} + 2\right) \cdot x}{2} \]
                          5. lower-fma.f64N/A

                            \[\leadsto \frac{\color{blue}{\mathsf{fma}\left({x}^{2}, \frac{1}{3}, 2\right)} \cdot x}{2} \]
                          6. unpow2N/A

                            \[\leadsto \frac{\mathsf{fma}\left(\color{blue}{x \cdot x}, \frac{1}{3}, 2\right) \cdot x}{2} \]
                          7. lower-*.f6481.9

                            \[\leadsto \frac{\mathsf{fma}\left(\color{blue}{x \cdot x}, 0.3333333333333333, 2\right) \cdot x}{2} \]
                        5. Applied rewrites81.9%

                          \[\leadsto \frac{\color{blue}{\mathsf{fma}\left(x \cdot x, 0.3333333333333333, 2\right) \cdot x}}{2} \]
                        6. Step-by-step derivation
                          1. Applied rewrites81.9%

                            \[\leadsto \frac{\mathsf{fma}\left(0.3333333333333333 \cdot \left(x \cdot x\right), \color{blue}{x}, 2 \cdot x\right)}{2} \]
                          2. Step-by-step derivation
                            1. lift-/.f64N/A

                              \[\leadsto \color{blue}{\frac{\mathsf{fma}\left(\frac{1}{3} \cdot \left(x \cdot x\right), x, 2 \cdot x\right)}{2}} \]
                            2. div-invN/A

                              \[\leadsto \color{blue}{\mathsf{fma}\left(\frac{1}{3} \cdot \left(x \cdot x\right), x, 2 \cdot x\right) \cdot \frac{1}{2}} \]
                            3. metadata-evalN/A

                              \[\leadsto \mathsf{fma}\left(\frac{1}{3} \cdot \left(x \cdot x\right), x, 2 \cdot x\right) \cdot \color{blue}{\frac{1}{2}} \]
                            4. lower-*.f6481.9

                              \[\leadsto \color{blue}{\mathsf{fma}\left(0.3333333333333333 \cdot \left(x \cdot x\right), x, 2 \cdot x\right) \cdot 0.5} \]
                          3. Applied rewrites81.9%

                            \[\leadsto \color{blue}{\left(\mathsf{fma}\left(x \cdot x, 0.3333333333333333, 2\right) \cdot x\right) \cdot 0.5} \]
                          4. Step-by-step derivation
                            1. Applied rewrites81.9%

                              \[\leadsto \mathsf{fma}\left(0.3333333333333333 \cdot \left(x \cdot x\right), \color{blue}{x}, 2 \cdot x\right) \cdot 0.5 \]
                            2. Add Preprocessing

                            Alternative 11: 84.3% accurate, 9.9× speedup?

                            \[\begin{array}{l} \\ \left(\mathsf{fma}\left(0.3333333333333333, x \cdot x, 2\right) \cdot x\right) \cdot 0.5 \end{array} \]
                            (FPCore (x)
                             :precision binary64
                             (* (* (fma 0.3333333333333333 (* x x) 2.0) x) 0.5))
                            double code(double x) {
                            	return (fma(0.3333333333333333, (x * x), 2.0) * x) * 0.5;
                            }
                            
                            function code(x)
                            	return Float64(Float64(fma(0.3333333333333333, Float64(x * x), 2.0) * x) * 0.5)
                            end
                            
                            code[x_] := N[(N[(N[(0.3333333333333333 * N[(x * x), $MachinePrecision] + 2.0), $MachinePrecision] * x), $MachinePrecision] * 0.5), $MachinePrecision]
                            
                            \begin{array}{l}
                            
                            \\
                            \left(\mathsf{fma}\left(0.3333333333333333, x \cdot x, 2\right) \cdot x\right) \cdot 0.5
                            \end{array}
                            
                            Derivation
                            1. Initial program 57.9%

                              \[\frac{e^{x} - e^{-x}}{2} \]
                            2. Add Preprocessing
                            3. Taylor expanded in x around 0

                              \[\leadsto \frac{\color{blue}{x \cdot \left(2 + \frac{1}{3} \cdot {x}^{2}\right)}}{2} \]
                            4. Step-by-step derivation
                              1. *-commutativeN/A

                                \[\leadsto \frac{\color{blue}{\left(2 + \frac{1}{3} \cdot {x}^{2}\right) \cdot x}}{2} \]
                              2. lower-*.f64N/A

                                \[\leadsto \frac{\color{blue}{\left(2 + \frac{1}{3} \cdot {x}^{2}\right) \cdot x}}{2} \]
                              3. +-commutativeN/A

                                \[\leadsto \frac{\color{blue}{\left(\frac{1}{3} \cdot {x}^{2} + 2\right)} \cdot x}{2} \]
                              4. *-commutativeN/A

                                \[\leadsto \frac{\left(\color{blue}{{x}^{2} \cdot \frac{1}{3}} + 2\right) \cdot x}{2} \]
                              5. lower-fma.f64N/A

                                \[\leadsto \frac{\color{blue}{\mathsf{fma}\left({x}^{2}, \frac{1}{3}, 2\right)} \cdot x}{2} \]
                              6. unpow2N/A

                                \[\leadsto \frac{\mathsf{fma}\left(\color{blue}{x \cdot x}, \frac{1}{3}, 2\right) \cdot x}{2} \]
                              7. lower-*.f6481.9

                                \[\leadsto \frac{\mathsf{fma}\left(\color{blue}{x \cdot x}, 0.3333333333333333, 2\right) \cdot x}{2} \]
                            5. Applied rewrites81.9%

                              \[\leadsto \frac{\color{blue}{\mathsf{fma}\left(x \cdot x, 0.3333333333333333, 2\right) \cdot x}}{2} \]
                            6. Step-by-step derivation
                              1. lift-/.f64N/A

                                \[\leadsto \color{blue}{\frac{\mathsf{fma}\left(x \cdot x, \frac{1}{3}, 2\right) \cdot x}{2}} \]
                              2. div-invN/A

                                \[\leadsto \color{blue}{\left(\mathsf{fma}\left(x \cdot x, \frac{1}{3}, 2\right) \cdot x\right) \cdot \frac{1}{2}} \]
                              3. lower-*.f64N/A

                                \[\leadsto \color{blue}{\left(\mathsf{fma}\left(x \cdot x, \frac{1}{3}, 2\right) \cdot x\right) \cdot \frac{1}{2}} \]
                              4. metadata-evalN/A

                                \[\leadsto \left(\mathsf{Rewrite=>}\left(lower-fma.f64, \left(\mathsf{fma}\left(\frac{1}{3}, x \cdot x, 2\right)\right)\right) \cdot x\right) \cdot \color{blue}{\frac{1}{2}} \]
                            7. Applied rewrites81.9%

                              \[\leadsto \color{blue}{\left(\mathsf{fma}\left(0.3333333333333333, x \cdot x, 2\right) \cdot x\right) \cdot 0.5} \]
                            8. Add Preprocessing

                            Alternative 12: 53.2% accurate, 19.7× speedup?

                            \[\begin{array}{l} \\ \left(2 \cdot x\right) \cdot 0.5 \end{array} \]
                            (FPCore (x) :precision binary64 (* (* 2.0 x) 0.5))
                            double code(double x) {
                            	return (2.0 * x) * 0.5;
                            }
                            
                            real(8) function code(x)
                                real(8), intent (in) :: x
                                code = (2.0d0 * x) * 0.5d0
                            end function
                            
                            public static double code(double x) {
                            	return (2.0 * x) * 0.5;
                            }
                            
                            def code(x):
                            	return (2.0 * x) * 0.5
                            
                            function code(x)
                            	return Float64(Float64(2.0 * x) * 0.5)
                            end
                            
                            function tmp = code(x)
                            	tmp = (2.0 * x) * 0.5;
                            end
                            
                            code[x_] := N[(N[(2.0 * x), $MachinePrecision] * 0.5), $MachinePrecision]
                            
                            \begin{array}{l}
                            
                            \\
                            \left(2 \cdot x\right) \cdot 0.5
                            \end{array}
                            
                            Derivation
                            1. Initial program 57.9%

                              \[\frac{e^{x} - e^{-x}}{2} \]
                            2. Add Preprocessing
                            3. Taylor expanded in x around 0

                              \[\leadsto \frac{\color{blue}{2 \cdot x}}{2} \]
                            4. Step-by-step derivation
                              1. *-commutativeN/A

                                \[\leadsto \frac{\color{blue}{x \cdot 2}}{2} \]
                              2. lower-*.f6448.6

                                \[\leadsto \frac{\color{blue}{x \cdot 2}}{2} \]
                            5. Applied rewrites48.6%

                              \[\leadsto \frac{\color{blue}{x \cdot 2}}{2} \]
                            6. Step-by-step derivation
                              1. lift-/.f64N/A

                                \[\leadsto \color{blue}{\frac{x \cdot 2}{2}} \]
                              2. div-invN/A

                                \[\leadsto \color{blue}{\left(x \cdot 2\right) \cdot \frac{1}{2}} \]
                              3. lower-*.f64N/A

                                \[\leadsto \color{blue}{\left(x \cdot 2\right) \cdot \frac{1}{2}} \]
                              4. metadata-evalN/A

                                \[\leadsto \mathsf{Rewrite=>}\left(lower-*.f64, \left(2 \cdot x\right)\right) \cdot \color{blue}{\frac{1}{2}} \]
                            7. Applied rewrites48.6%

                              \[\leadsto \color{blue}{\left(2 \cdot x\right) \cdot 0.5} \]
                            8. Add Preprocessing

                            Reproduce

                            ?
                            herbie shell --seed 2024270 
                            (FPCore (x)
                              :name "Hyperbolic sine"
                              :precision binary64
                              (/ (- (exp x) (exp (- x))) 2.0))