Hyperbolic arc-(co)tangent

Percentage Accurate: 8.7% → 100.0%
Time: 9.8s
Alternatives: 7
Speedup: 12.2×

Specification

?
\[\begin{array}{l} \\ \frac{1}{2} \cdot \log \left(\frac{1 + x}{1 - x}\right) \end{array} \]
(FPCore (x) :precision binary64 (* (/ 1.0 2.0) (log (/ (+ 1.0 x) (- 1.0 x)))))
double code(double x) {
	return (1.0 / 2.0) * log(((1.0 + x) / (1.0 - x)));
}
real(8) function code(x)
    real(8), intent (in) :: x
    code = (1.0d0 / 2.0d0) * log(((1.0d0 + x) / (1.0d0 - x)))
end function
public static double code(double x) {
	return (1.0 / 2.0) * Math.log(((1.0 + x) / (1.0 - x)));
}
def code(x):
	return (1.0 / 2.0) * math.log(((1.0 + x) / (1.0 - x)))
function code(x)
	return Float64(Float64(1.0 / 2.0) * log(Float64(Float64(1.0 + x) / Float64(1.0 - x))))
end
function tmp = code(x)
	tmp = (1.0 / 2.0) * log(((1.0 + x) / (1.0 - x)));
end
code[x_] := N[(N[(1.0 / 2.0), $MachinePrecision] * N[Log[N[(N[(1.0 + x), $MachinePrecision] / N[(1.0 - x), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}

\\
\frac{1}{2} \cdot \log \left(\frac{1 + x}{1 - x}\right)
\end{array}

Sampling outcomes in binary64 precision:

Local Percentage Accuracy vs ?

The average percentage accuracy by input value. Horizontal axis shows value of an input variable; the variable is choosen in the title. Vertical axis is accuracy; higher is better. Red represent the original program, while blue represents Herbie's suggestion. These can be toggled with buttons below the plot. The line is an average while dots represent individual samples.

Accuracy vs Speed?

Herbie found 7 alternatives:

AlternativeAccuracySpeedup
The accuracy (vertical axis) and speed (horizontal axis) of each alternatives. Up and to the right is better. The red square shows the initial program, and each blue circle shows an alternative.The line shows the best available speed-accuracy tradeoffs.

Initial Program: 8.7% accurate, 1.0× speedup?

\[\begin{array}{l} \\ \frac{1}{2} \cdot \log \left(\frac{1 + x}{1 - x}\right) \end{array} \]
(FPCore (x) :precision binary64 (* (/ 1.0 2.0) (log (/ (+ 1.0 x) (- 1.0 x)))))
double code(double x) {
	return (1.0 / 2.0) * log(((1.0 + x) / (1.0 - x)));
}
real(8) function code(x)
    real(8), intent (in) :: x
    code = (1.0d0 / 2.0d0) * log(((1.0d0 + x) / (1.0d0 - x)))
end function
public static double code(double x) {
	return (1.0 / 2.0) * Math.log(((1.0 + x) / (1.0 - x)));
}
def code(x):
	return (1.0 / 2.0) * math.log(((1.0 + x) / (1.0 - x)))
function code(x)
	return Float64(Float64(1.0 / 2.0) * log(Float64(Float64(1.0 + x) / Float64(1.0 - x))))
end
function tmp = code(x)
	tmp = (1.0 / 2.0) * log(((1.0 + x) / (1.0 - x)));
end
code[x_] := N[(N[(1.0 / 2.0), $MachinePrecision] * N[Log[N[(N[(1.0 + x), $MachinePrecision] / N[(1.0 - x), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}

\\
\frac{1}{2} \cdot \log \left(\frac{1 + x}{1 - x}\right)
\end{array}

Alternative 1: 100.0% accurate, 0.6× speedup?

\[\begin{array}{l} \\ 0.5 \cdot \left(2 \cdot \mathsf{log1p}\left(x\right) - \mathsf{log1p}\left(\left(-x\right) \cdot x\right)\right) \end{array} \]
(FPCore (x)
 :precision binary64
 (* 0.5 (- (* 2.0 (log1p x)) (log1p (* (- x) x)))))
double code(double x) {
	return 0.5 * ((2.0 * log1p(x)) - log1p((-x * x)));
}
public static double code(double x) {
	return 0.5 * ((2.0 * Math.log1p(x)) - Math.log1p((-x * x)));
}
def code(x):
	return 0.5 * ((2.0 * math.log1p(x)) - math.log1p((-x * x)))
function code(x)
	return Float64(0.5 * Float64(Float64(2.0 * log1p(x)) - log1p(Float64(Float64(-x) * x))))
end
code[x_] := N[(0.5 * N[(N[(2.0 * N[Log[1 + x], $MachinePrecision]), $MachinePrecision] - N[Log[1 + N[((-x) * x), $MachinePrecision]], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}

\\
0.5 \cdot \left(2 \cdot \mathsf{log1p}\left(x\right) - \mathsf{log1p}\left(\left(-x\right) \cdot x\right)\right)
\end{array}
Derivation
  1. Initial program 9.2%

    \[\frac{1}{2} \cdot \log \left(\frac{1 + x}{1 - x}\right) \]
  2. Add Preprocessing
  3. Applied rewrites100.0%

    \[\leadsto \frac{1}{2} \cdot \color{blue}{\left(2 \cdot \mathsf{log1p}\left(x\right) - \mathsf{log1p}\left(\left(-x\right) \cdot x\right)\right)} \]
  4. Step-by-step derivation
    1. lift-*.f64N/A

      \[\leadsto \color{blue}{\frac{1}{2} \cdot \left(2 \cdot \mathsf{log1p}\left(x\right) - \mathsf{log1p}\left(\left(-x\right) \cdot x\right)\right)} \]
    2. *-commutativeN/A

      \[\leadsto \color{blue}{\left(2 \cdot \mathsf{log1p}\left(x\right) - \mathsf{log1p}\left(\left(-x\right) \cdot x\right)\right) \cdot \frac{1}{2}} \]
    3. lower-*.f64100.0

      \[\leadsto \color{blue}{\left(2 \cdot \mathsf{log1p}\left(x\right) - \mathsf{log1p}\left(\left(-x\right) \cdot x\right)\right) \cdot \frac{1}{2}} \]
    4. lift-*.f64N/A

      \[\leadsto \left(\color{blue}{2 \cdot \mathsf{log1p}\left(x\right)} - \mathsf{log1p}\left(\left(-x\right) \cdot x\right)\right) \cdot \frac{1}{2} \]
    5. *-commutativeN/A

      \[\leadsto \left(\color{blue}{\mathsf{log1p}\left(x\right) \cdot 2} - \mathsf{log1p}\left(\left(-x\right) \cdot x\right)\right) \cdot \frac{1}{2} \]
    6. lower-*.f64100.0

      \[\leadsto \left(\color{blue}{\mathsf{log1p}\left(x\right) \cdot 2} - \mathsf{log1p}\left(\left(-x\right) \cdot x\right)\right) \cdot \frac{1}{2} \]
    7. lift-/.f64N/A

      \[\leadsto \left(\mathsf{log1p}\left(x\right) \cdot 2 - \mathsf{log1p}\left(\left(-x\right) \cdot x\right)\right) \cdot \color{blue}{\frac{1}{2}} \]
    8. metadata-eval100.0

      \[\leadsto \left(\mathsf{log1p}\left(x\right) \cdot 2 - \mathsf{log1p}\left(\left(-x\right) \cdot x\right)\right) \cdot \color{blue}{0.5} \]
  5. Applied rewrites100.0%

    \[\leadsto \color{blue}{\left(\mathsf{log1p}\left(x\right) \cdot 2 - \mathsf{log1p}\left(\left(-x\right) \cdot x\right)\right) \cdot 0.5} \]
  6. Final simplification100.0%

    \[\leadsto 0.5 \cdot \left(2 \cdot \mathsf{log1p}\left(x\right) - \mathsf{log1p}\left(\left(-x\right) \cdot x\right)\right) \]
  7. Add Preprocessing

Alternative 2: 100.0% accurate, 0.6× speedup?

\[\begin{array}{l} \\ \left(\mathsf{log1p}\left(x\right) - \mathsf{log1p}\left(-x\right)\right) \cdot 0.5 \end{array} \]
(FPCore (x) :precision binary64 (* (- (log1p x) (log1p (- x))) 0.5))
double code(double x) {
	return (log1p(x) - log1p(-x)) * 0.5;
}
public static double code(double x) {
	return (Math.log1p(x) - Math.log1p(-x)) * 0.5;
}
def code(x):
	return (math.log1p(x) - math.log1p(-x)) * 0.5
function code(x)
	return Float64(Float64(log1p(x) - log1p(Float64(-x))) * 0.5)
end
code[x_] := N[(N[(N[Log[1 + x], $MachinePrecision] - N[Log[1 + (-x)], $MachinePrecision]), $MachinePrecision] * 0.5), $MachinePrecision]
\begin{array}{l}

\\
\left(\mathsf{log1p}\left(x\right) - \mathsf{log1p}\left(-x\right)\right) \cdot 0.5
\end{array}
Derivation
  1. Initial program 9.2%

    \[\frac{1}{2} \cdot \log \left(\frac{1 + x}{1 - x}\right) \]
  2. Add Preprocessing
  3. Step-by-step derivation
    1. lift-/.f64N/A

      \[\leadsto \color{blue}{\frac{1}{2}} \cdot \log \left(\frac{1 + x}{1 - x}\right) \]
    2. metadata-eval9.2

      \[\leadsto \color{blue}{0.5} \cdot \log \left(\frac{1 + x}{1 - x}\right) \]
  4. Applied rewrites9.2%

    \[\leadsto \color{blue}{0.5} \cdot \log \left(\frac{1 + x}{1 - x}\right) \]
  5. Step-by-step derivation
    1. lift-*.f64N/A

      \[\leadsto \color{blue}{\frac{1}{2} \cdot \log \left(\frac{1 + x}{1 - x}\right)} \]
    2. *-commutativeN/A

      \[\leadsto \color{blue}{\log \left(\frac{1 + x}{1 - x}\right) \cdot \frac{1}{2}} \]
    3. lower-*.f649.2

      \[\leadsto \color{blue}{\log \left(\frac{1 + x}{1 - x}\right) \cdot 0.5} \]
    4. lift-log.f64N/A

      \[\leadsto \color{blue}{\log \left(\frac{1 + x}{1 - x}\right)} \cdot \frac{1}{2} \]
    5. lift-/.f64N/A

      \[\leadsto \log \color{blue}{\left(\frac{1 + x}{1 - x}\right)} \cdot \frac{1}{2} \]
    6. log-divN/A

      \[\leadsto \color{blue}{\left(\log \left(1 + x\right) - \log \left(1 - x\right)\right)} \cdot \frac{1}{2} \]
    7. lift-+.f64N/A

      \[\leadsto \left(\log \color{blue}{\left(1 + x\right)} - \log \left(1 - x\right)\right) \cdot \frac{1}{2} \]
    8. lift-log1p.f64N/A

      \[\leadsto \left(\color{blue}{\mathsf{log1p}\left(x\right)} - \log \left(1 - x\right)\right) \cdot \frac{1}{2} \]
    9. lower--.f64N/A

      \[\leadsto \color{blue}{\left(\mathsf{log1p}\left(x\right) - \log \left(1 - x\right)\right)} \cdot \frac{1}{2} \]
    10. lift--.f64N/A

      \[\leadsto \left(\mathsf{log1p}\left(x\right) - \log \color{blue}{\left(1 - x\right)}\right) \cdot \frac{1}{2} \]
    11. sub-negN/A

      \[\leadsto \left(\mathsf{log1p}\left(x\right) - \log \color{blue}{\left(1 + \left(\mathsf{neg}\left(x\right)\right)\right)}\right) \cdot \frac{1}{2} \]
    12. lift-neg.f64N/A

      \[\leadsto \left(\mathsf{log1p}\left(x\right) - \log \left(1 + \color{blue}{\left(-x\right)}\right)\right) \cdot \frac{1}{2} \]
    13. lower-log1p.f64100.0

      \[\leadsto \left(\mathsf{log1p}\left(x\right) - \color{blue}{\mathsf{log1p}\left(-x\right)}\right) \cdot 0.5 \]
  6. Applied rewrites100.0%

    \[\leadsto \color{blue}{\left(\mathsf{log1p}\left(x\right) - \mathsf{log1p}\left(-x\right)\right) \cdot 0.5} \]
  7. Add Preprocessing

Alternative 3: 99.8% accurate, 0.9× speedup?

\[\begin{array}{l} \\ \left(2 \cdot \mathsf{log1p}\left(x\right) - \left(\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(-0.25 \cdot x, x, -0.3333333333333333\right), x \cdot x, -0.5\right), x \cdot x, -1\right) \cdot x\right) \cdot x\right) \cdot 0.5 \end{array} \]
(FPCore (x)
 :precision binary64
 (*
  (-
   (* 2.0 (log1p x))
   (*
    (*
     (fma
      (fma (fma (* -0.25 x) x -0.3333333333333333) (* x x) -0.5)
      (* x x)
      -1.0)
     x)
    x))
  0.5))
double code(double x) {
	return ((2.0 * log1p(x)) - ((fma(fma(fma((-0.25 * x), x, -0.3333333333333333), (x * x), -0.5), (x * x), -1.0) * x) * x)) * 0.5;
}
function code(x)
	return Float64(Float64(Float64(2.0 * log1p(x)) - Float64(Float64(fma(fma(fma(Float64(-0.25 * x), x, -0.3333333333333333), Float64(x * x), -0.5), Float64(x * x), -1.0) * x) * x)) * 0.5)
end
code[x_] := N[(N[(N[(2.0 * N[Log[1 + x], $MachinePrecision]), $MachinePrecision] - N[(N[(N[(N[(N[(N[(-0.25 * x), $MachinePrecision] * x + -0.3333333333333333), $MachinePrecision] * N[(x * x), $MachinePrecision] + -0.5), $MachinePrecision] * N[(x * x), $MachinePrecision] + -1.0), $MachinePrecision] * x), $MachinePrecision] * x), $MachinePrecision]), $MachinePrecision] * 0.5), $MachinePrecision]
\begin{array}{l}

\\
\left(2 \cdot \mathsf{log1p}\left(x\right) - \left(\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(-0.25 \cdot x, x, -0.3333333333333333\right), x \cdot x, -0.5\right), x \cdot x, -1\right) \cdot x\right) \cdot x\right) \cdot 0.5
\end{array}
Derivation
  1. Initial program 9.2%

    \[\frac{1}{2} \cdot \log \left(\frac{1 + x}{1 - x}\right) \]
  2. Add Preprocessing
  3. Applied rewrites100.0%

    \[\leadsto \frac{1}{2} \cdot \color{blue}{\left(2 \cdot \mathsf{log1p}\left(x\right) - \mathsf{log1p}\left(\left(-x\right) \cdot x\right)\right)} \]
  4. Step-by-step derivation
    1. lift-*.f64N/A

      \[\leadsto \color{blue}{\frac{1}{2} \cdot \left(2 \cdot \mathsf{log1p}\left(x\right) - \mathsf{log1p}\left(\left(-x\right) \cdot x\right)\right)} \]
    2. *-commutativeN/A

      \[\leadsto \color{blue}{\left(2 \cdot \mathsf{log1p}\left(x\right) - \mathsf{log1p}\left(\left(-x\right) \cdot x\right)\right) \cdot \frac{1}{2}} \]
    3. lower-*.f64100.0

      \[\leadsto \color{blue}{\left(2 \cdot \mathsf{log1p}\left(x\right) - \mathsf{log1p}\left(\left(-x\right) \cdot x\right)\right) \cdot \frac{1}{2}} \]
    4. lift-*.f64N/A

      \[\leadsto \left(\color{blue}{2 \cdot \mathsf{log1p}\left(x\right)} - \mathsf{log1p}\left(\left(-x\right) \cdot x\right)\right) \cdot \frac{1}{2} \]
    5. *-commutativeN/A

      \[\leadsto \left(\color{blue}{\mathsf{log1p}\left(x\right) \cdot 2} - \mathsf{log1p}\left(\left(-x\right) \cdot x\right)\right) \cdot \frac{1}{2} \]
    6. lower-*.f64100.0

      \[\leadsto \left(\color{blue}{\mathsf{log1p}\left(x\right) \cdot 2} - \mathsf{log1p}\left(\left(-x\right) \cdot x\right)\right) \cdot \frac{1}{2} \]
    7. lift-/.f64N/A

      \[\leadsto \left(\mathsf{log1p}\left(x\right) \cdot 2 - \mathsf{log1p}\left(\left(-x\right) \cdot x\right)\right) \cdot \color{blue}{\frac{1}{2}} \]
    8. metadata-eval100.0

      \[\leadsto \left(\mathsf{log1p}\left(x\right) \cdot 2 - \mathsf{log1p}\left(\left(-x\right) \cdot x\right)\right) \cdot \color{blue}{0.5} \]
  5. Applied rewrites100.0%

    \[\leadsto \color{blue}{\left(\mathsf{log1p}\left(x\right) \cdot 2 - \mathsf{log1p}\left(\left(-x\right) \cdot x\right)\right) \cdot 0.5} \]
  6. Taylor expanded in x around 0

    \[\leadsto \left(\mathsf{log1p}\left(x\right) \cdot 2 - \color{blue}{{x}^{2} \cdot \left({x}^{2} \cdot \left({x}^{2} \cdot \left(\frac{-1}{4} \cdot {x}^{2} - \frac{1}{3}\right) - \frac{1}{2}\right) - 1\right)}\right) \cdot \frac{1}{2} \]
  7. Step-by-step derivation
    1. *-commutativeN/A

      \[\leadsto \left(\mathsf{log1p}\left(x\right) \cdot 2 - \color{blue}{\left({x}^{2} \cdot \left({x}^{2} \cdot \left(\frac{-1}{4} \cdot {x}^{2} - \frac{1}{3}\right) - \frac{1}{2}\right) - 1\right) \cdot {x}^{2}}\right) \cdot \frac{1}{2} \]
    2. unpow2N/A

      \[\leadsto \left(\mathsf{log1p}\left(x\right) \cdot 2 - \left({x}^{2} \cdot \left({x}^{2} \cdot \left(\frac{-1}{4} \cdot {x}^{2} - \frac{1}{3}\right) - \frac{1}{2}\right) - 1\right) \cdot \color{blue}{\left(x \cdot x\right)}\right) \cdot \frac{1}{2} \]
    3. associate-*r*N/A

      \[\leadsto \left(\mathsf{log1p}\left(x\right) \cdot 2 - \color{blue}{\left(\left({x}^{2} \cdot \left({x}^{2} \cdot \left(\frac{-1}{4} \cdot {x}^{2} - \frac{1}{3}\right) - \frac{1}{2}\right) - 1\right) \cdot x\right) \cdot x}\right) \cdot \frac{1}{2} \]
    4. lower-*.f64N/A

      \[\leadsto \left(\mathsf{log1p}\left(x\right) \cdot 2 - \color{blue}{\left(\left({x}^{2} \cdot \left({x}^{2} \cdot \left(\frac{-1}{4} \cdot {x}^{2} - \frac{1}{3}\right) - \frac{1}{2}\right) - 1\right) \cdot x\right) \cdot x}\right) \cdot \frac{1}{2} \]
  8. Applied rewrites99.5%

    \[\leadsto \left(\mathsf{log1p}\left(x\right) \cdot 2 - \color{blue}{\left(\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(-0.25 \cdot x, x, -0.3333333333333333\right), x \cdot x, -0.5\right), x \cdot x, -1\right) \cdot x\right) \cdot x}\right) \cdot 0.5 \]
  9. Final simplification99.5%

    \[\leadsto \left(2 \cdot \mathsf{log1p}\left(x\right) - \left(\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(-0.25 \cdot x, x, -0.3333333333333333\right), x \cdot x, -0.5\right), x \cdot x, -1\right) \cdot x\right) \cdot x\right) \cdot 0.5 \]
  10. Add Preprocessing

Alternative 4: 99.7% accurate, 3.0× speedup?

\[\begin{array}{l} \\ \left(\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(0.2857142857142857, x \cdot x, 0.4\right), x \cdot x, 0.6666666666666666\right), x \cdot x, 2\right) \cdot x\right) \cdot 0.5 \end{array} \]
(FPCore (x)
 :precision binary64
 (*
  (*
   (fma
    (fma (fma 0.2857142857142857 (* x x) 0.4) (* x x) 0.6666666666666666)
    (* x x)
    2.0)
   x)
  0.5))
double code(double x) {
	return (fma(fma(fma(0.2857142857142857, (x * x), 0.4), (x * x), 0.6666666666666666), (x * x), 2.0) * x) * 0.5;
}
function code(x)
	return Float64(Float64(fma(fma(fma(0.2857142857142857, Float64(x * x), 0.4), Float64(x * x), 0.6666666666666666), Float64(x * x), 2.0) * x) * 0.5)
end
code[x_] := N[(N[(N[(N[(N[(0.2857142857142857 * N[(x * x), $MachinePrecision] + 0.4), $MachinePrecision] * N[(x * x), $MachinePrecision] + 0.6666666666666666), $MachinePrecision] * N[(x * x), $MachinePrecision] + 2.0), $MachinePrecision] * x), $MachinePrecision] * 0.5), $MachinePrecision]
\begin{array}{l}

\\
\left(\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(0.2857142857142857, x \cdot x, 0.4\right), x \cdot x, 0.6666666666666666\right), x \cdot x, 2\right) \cdot x\right) \cdot 0.5
\end{array}
Derivation
  1. Initial program 9.2%

    \[\frac{1}{2} \cdot \log \left(\frac{1 + x}{1 - x}\right) \]
  2. Add Preprocessing
  3. Applied rewrites100.0%

    \[\leadsto \frac{1}{2} \cdot \color{blue}{\left(2 \cdot \mathsf{log1p}\left(x\right) - \mathsf{log1p}\left(\left(-x\right) \cdot x\right)\right)} \]
  4. Step-by-step derivation
    1. lift-*.f64N/A

      \[\leadsto \color{blue}{\frac{1}{2} \cdot \left(2 \cdot \mathsf{log1p}\left(x\right) - \mathsf{log1p}\left(\left(-x\right) \cdot x\right)\right)} \]
    2. *-commutativeN/A

      \[\leadsto \color{blue}{\left(2 \cdot \mathsf{log1p}\left(x\right) - \mathsf{log1p}\left(\left(-x\right) \cdot x\right)\right) \cdot \frac{1}{2}} \]
    3. lower-*.f64100.0

      \[\leadsto \color{blue}{\left(2 \cdot \mathsf{log1p}\left(x\right) - \mathsf{log1p}\left(\left(-x\right) \cdot x\right)\right) \cdot \frac{1}{2}} \]
    4. lift-*.f64N/A

      \[\leadsto \left(\color{blue}{2 \cdot \mathsf{log1p}\left(x\right)} - \mathsf{log1p}\left(\left(-x\right) \cdot x\right)\right) \cdot \frac{1}{2} \]
    5. *-commutativeN/A

      \[\leadsto \left(\color{blue}{\mathsf{log1p}\left(x\right) \cdot 2} - \mathsf{log1p}\left(\left(-x\right) \cdot x\right)\right) \cdot \frac{1}{2} \]
    6. lower-*.f64100.0

      \[\leadsto \left(\color{blue}{\mathsf{log1p}\left(x\right) \cdot 2} - \mathsf{log1p}\left(\left(-x\right) \cdot x\right)\right) \cdot \frac{1}{2} \]
    7. lift-/.f64N/A

      \[\leadsto \left(\mathsf{log1p}\left(x\right) \cdot 2 - \mathsf{log1p}\left(\left(-x\right) \cdot x\right)\right) \cdot \color{blue}{\frac{1}{2}} \]
    8. metadata-eval100.0

      \[\leadsto \left(\mathsf{log1p}\left(x\right) \cdot 2 - \mathsf{log1p}\left(\left(-x\right) \cdot x\right)\right) \cdot \color{blue}{0.5} \]
  5. Applied rewrites100.0%

    \[\leadsto \color{blue}{\left(\mathsf{log1p}\left(x\right) \cdot 2 - \mathsf{log1p}\left(\left(-x\right) \cdot x\right)\right) \cdot 0.5} \]
  6. Taylor expanded in x around 0

    \[\leadsto \color{blue}{\left(x \cdot \left(2 + {x}^{2} \cdot \left(\frac{2}{3} + {x}^{2} \cdot \left(\frac{2}{5} + \frac{2}{7} \cdot {x}^{2}\right)\right)\right)\right)} \cdot \frac{1}{2} \]
  7. Step-by-step derivation
    1. *-commutativeN/A

      \[\leadsto \color{blue}{\left(\left(2 + {x}^{2} \cdot \left(\frac{2}{3} + {x}^{2} \cdot \left(\frac{2}{5} + \frac{2}{7} \cdot {x}^{2}\right)\right)\right) \cdot x\right)} \cdot \frac{1}{2} \]
    2. lower-*.f64N/A

      \[\leadsto \color{blue}{\left(\left(2 + {x}^{2} \cdot \left(\frac{2}{3} + {x}^{2} \cdot \left(\frac{2}{5} + \frac{2}{7} \cdot {x}^{2}\right)\right)\right) \cdot x\right)} \cdot \frac{1}{2} \]
    3. +-commutativeN/A

      \[\leadsto \left(\color{blue}{\left({x}^{2} \cdot \left(\frac{2}{3} + {x}^{2} \cdot \left(\frac{2}{5} + \frac{2}{7} \cdot {x}^{2}\right)\right) + 2\right)} \cdot x\right) \cdot \frac{1}{2} \]
    4. *-commutativeN/A

      \[\leadsto \left(\left(\color{blue}{\left(\frac{2}{3} + {x}^{2} \cdot \left(\frac{2}{5} + \frac{2}{7} \cdot {x}^{2}\right)\right) \cdot {x}^{2}} + 2\right) \cdot x\right) \cdot \frac{1}{2} \]
    5. lower-fma.f64N/A

      \[\leadsto \left(\color{blue}{\mathsf{fma}\left(\frac{2}{3} + {x}^{2} \cdot \left(\frac{2}{5} + \frac{2}{7} \cdot {x}^{2}\right), {x}^{2}, 2\right)} \cdot x\right) \cdot \frac{1}{2} \]
    6. +-commutativeN/A

      \[\leadsto \left(\mathsf{fma}\left(\color{blue}{{x}^{2} \cdot \left(\frac{2}{5} + \frac{2}{7} \cdot {x}^{2}\right) + \frac{2}{3}}, {x}^{2}, 2\right) \cdot x\right) \cdot \frac{1}{2} \]
    7. *-commutativeN/A

      \[\leadsto \left(\mathsf{fma}\left(\color{blue}{\left(\frac{2}{5} + \frac{2}{7} \cdot {x}^{2}\right) \cdot {x}^{2}} + \frac{2}{3}, {x}^{2}, 2\right) \cdot x\right) \cdot \frac{1}{2} \]
    8. lower-fma.f64N/A

      \[\leadsto \left(\mathsf{fma}\left(\color{blue}{\mathsf{fma}\left(\frac{2}{5} + \frac{2}{7} \cdot {x}^{2}, {x}^{2}, \frac{2}{3}\right)}, {x}^{2}, 2\right) \cdot x\right) \cdot \frac{1}{2} \]
    9. +-commutativeN/A

      \[\leadsto \left(\mathsf{fma}\left(\mathsf{fma}\left(\color{blue}{\frac{2}{7} \cdot {x}^{2} + \frac{2}{5}}, {x}^{2}, \frac{2}{3}\right), {x}^{2}, 2\right) \cdot x\right) \cdot \frac{1}{2} \]
    10. lower-fma.f64N/A

      \[\leadsto \left(\mathsf{fma}\left(\mathsf{fma}\left(\color{blue}{\mathsf{fma}\left(\frac{2}{7}, {x}^{2}, \frac{2}{5}\right)}, {x}^{2}, \frac{2}{3}\right), {x}^{2}, 2\right) \cdot x\right) \cdot \frac{1}{2} \]
    11. unpow2N/A

      \[\leadsto \left(\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{2}{7}, \color{blue}{x \cdot x}, \frac{2}{5}\right), {x}^{2}, \frac{2}{3}\right), {x}^{2}, 2\right) \cdot x\right) \cdot \frac{1}{2} \]
    12. lower-*.f64N/A

      \[\leadsto \left(\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{2}{7}, \color{blue}{x \cdot x}, \frac{2}{5}\right), {x}^{2}, \frac{2}{3}\right), {x}^{2}, 2\right) \cdot x\right) \cdot \frac{1}{2} \]
    13. unpow2N/A

      \[\leadsto \left(\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{2}{7}, x \cdot x, \frac{2}{5}\right), \color{blue}{x \cdot x}, \frac{2}{3}\right), {x}^{2}, 2\right) \cdot x\right) \cdot \frac{1}{2} \]
    14. lower-*.f64N/A

      \[\leadsto \left(\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{2}{7}, x \cdot x, \frac{2}{5}\right), \color{blue}{x \cdot x}, \frac{2}{3}\right), {x}^{2}, 2\right) \cdot x\right) \cdot \frac{1}{2} \]
    15. unpow2N/A

      \[\leadsto \left(\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{2}{7}, x \cdot x, \frac{2}{5}\right), x \cdot x, \frac{2}{3}\right), \color{blue}{x \cdot x}, 2\right) \cdot x\right) \cdot \frac{1}{2} \]
    16. lower-*.f6499.5

      \[\leadsto \left(\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(0.2857142857142857, x \cdot x, 0.4\right), x \cdot x, 0.6666666666666666\right), \color{blue}{x \cdot x}, 2\right) \cdot x\right) \cdot 0.5 \]
  8. Applied rewrites99.5%

    \[\leadsto \color{blue}{\left(\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(0.2857142857142857, x \cdot x, 0.4\right), x \cdot x, 0.6666666666666666\right), x \cdot x, 2\right) \cdot x\right)} \cdot 0.5 \]
  9. Add Preprocessing

Alternative 5: 99.6% accurate, 4.1× speedup?

\[\begin{array}{l} \\ \left(\mathsf{fma}\left(\mathsf{fma}\left(x \cdot x, 0.4, 0.6666666666666666\right), x \cdot x, 2\right) \cdot x\right) \cdot 0.5 \end{array} \]
(FPCore (x)
 :precision binary64
 (* (* (fma (fma (* x x) 0.4 0.6666666666666666) (* x x) 2.0) x) 0.5))
double code(double x) {
	return (fma(fma((x * x), 0.4, 0.6666666666666666), (x * x), 2.0) * x) * 0.5;
}
function code(x)
	return Float64(Float64(fma(fma(Float64(x * x), 0.4, 0.6666666666666666), Float64(x * x), 2.0) * x) * 0.5)
end
code[x_] := N[(N[(N[(N[(N[(x * x), $MachinePrecision] * 0.4 + 0.6666666666666666), $MachinePrecision] * N[(x * x), $MachinePrecision] + 2.0), $MachinePrecision] * x), $MachinePrecision] * 0.5), $MachinePrecision]
\begin{array}{l}

\\
\left(\mathsf{fma}\left(\mathsf{fma}\left(x \cdot x, 0.4, 0.6666666666666666\right), x \cdot x, 2\right) \cdot x\right) \cdot 0.5
\end{array}
Derivation
  1. Initial program 9.2%

    \[\frac{1}{2} \cdot \log \left(\frac{1 + x}{1 - x}\right) \]
  2. Add Preprocessing
  3. Taylor expanded in x around 0

    \[\leadsto \frac{1}{2} \cdot \color{blue}{\left(x \cdot \left(2 + {x}^{2} \cdot \left(\frac{2}{3} + \frac{2}{5} \cdot {x}^{2}\right)\right)\right)} \]
  4. Step-by-step derivation
    1. *-commutativeN/A

      \[\leadsto \frac{1}{2} \cdot \color{blue}{\left(\left(2 + {x}^{2} \cdot \left(\frac{2}{3} + \frac{2}{5} \cdot {x}^{2}\right)\right) \cdot x\right)} \]
    2. lower-*.f64N/A

      \[\leadsto \frac{1}{2} \cdot \color{blue}{\left(\left(2 + {x}^{2} \cdot \left(\frac{2}{3} + \frac{2}{5} \cdot {x}^{2}\right)\right) \cdot x\right)} \]
    3. +-commutativeN/A

      \[\leadsto \frac{1}{2} \cdot \left(\color{blue}{\left({x}^{2} \cdot \left(\frac{2}{3} + \frac{2}{5} \cdot {x}^{2}\right) + 2\right)} \cdot x\right) \]
    4. *-commutativeN/A

      \[\leadsto \frac{1}{2} \cdot \left(\left(\color{blue}{\left(\frac{2}{3} + \frac{2}{5} \cdot {x}^{2}\right) \cdot {x}^{2}} + 2\right) \cdot x\right) \]
    5. lower-fma.f64N/A

      \[\leadsto \frac{1}{2} \cdot \left(\color{blue}{\mathsf{fma}\left(\frac{2}{3} + \frac{2}{5} \cdot {x}^{2}, {x}^{2}, 2\right)} \cdot x\right) \]
    6. +-commutativeN/A

      \[\leadsto \frac{1}{2} \cdot \left(\mathsf{fma}\left(\color{blue}{\frac{2}{5} \cdot {x}^{2} + \frac{2}{3}}, {x}^{2}, 2\right) \cdot x\right) \]
    7. lower-fma.f64N/A

      \[\leadsto \frac{1}{2} \cdot \left(\mathsf{fma}\left(\color{blue}{\mathsf{fma}\left(\frac{2}{5}, {x}^{2}, \frac{2}{3}\right)}, {x}^{2}, 2\right) \cdot x\right) \]
    8. unpow2N/A

      \[\leadsto \frac{1}{2} \cdot \left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{2}{5}, \color{blue}{x \cdot x}, \frac{2}{3}\right), {x}^{2}, 2\right) \cdot x\right) \]
    9. lower-*.f64N/A

      \[\leadsto \frac{1}{2} \cdot \left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{2}{5}, \color{blue}{x \cdot x}, \frac{2}{3}\right), {x}^{2}, 2\right) \cdot x\right) \]
    10. unpow2N/A

      \[\leadsto \frac{1}{2} \cdot \left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{2}{5}, x \cdot x, \frac{2}{3}\right), \color{blue}{x \cdot x}, 2\right) \cdot x\right) \]
    11. lower-*.f6499.4

      \[\leadsto \frac{1}{2} \cdot \left(\mathsf{fma}\left(\mathsf{fma}\left(0.4, x \cdot x, 0.6666666666666666\right), \color{blue}{x \cdot x}, 2\right) \cdot x\right) \]
  5. Applied rewrites99.4%

    \[\leadsto \frac{1}{2} \cdot \color{blue}{\left(\mathsf{fma}\left(\mathsf{fma}\left(0.4, x \cdot x, 0.6666666666666666\right), x \cdot x, 2\right) \cdot x\right)} \]
  6. Step-by-step derivation
    1. lift-*.f64N/A

      \[\leadsto \color{blue}{\frac{1}{2} \cdot \left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{2}{5}, x \cdot x, \frac{2}{3}\right), x \cdot x, 2\right) \cdot x\right)} \]
    2. *-commutativeN/A

      \[\leadsto \color{blue}{\left(\mathsf{fma}\left(\mathsf{fma}\left(\frac{2}{5}, x \cdot x, \frac{2}{3}\right), x \cdot x, 2\right) \cdot x\right) \cdot \frac{1}{2}} \]
    3. lower-*.f6499.4

      \[\leadsto \color{blue}{\left(\mathsf{fma}\left(\mathsf{fma}\left(0.4, x \cdot x, 0.6666666666666666\right), x \cdot x, 2\right) \cdot x\right) \cdot \frac{1}{2}} \]
    4. lift-/.f64N/A

      \[\leadsto \left(\mathsf{fma}\left(\mathsf{Rewrite=>}\left(lower-fma.f64, \left(\mathsf{fma}\left(x \cdot x, \frac{2}{5}, \frac{2}{3}\right)\right)\right), x \cdot x, 2\right) \cdot x\right) \cdot \color{blue}{\frac{1}{2}} \]
    5. metadata-evalN/A

      \[\leadsto \left(\mathsf{fma}\left(\mathsf{Rewrite=>}\left(lower-fma.f64, \left(\mathsf{fma}\left(x \cdot x, \frac{2}{5}, \frac{2}{3}\right)\right)\right), x \cdot x, 2\right) \cdot x\right) \cdot \color{blue}{\frac{1}{2}} \]
  7. Applied rewrites99.4%

    \[\leadsto \color{blue}{\left(\mathsf{fma}\left(\mathsf{fma}\left(x \cdot x, 0.4, 0.6666666666666666\right), x \cdot x, 2\right) \cdot x\right) \cdot 0.5} \]
  8. Add Preprocessing

Alternative 6: 99.4% accurate, 7.9× speedup?

\[\begin{array}{l} \\ \mathsf{fma}\left(\left(x \cdot x\right) \cdot x, 0.3333333333333333, x\right) \end{array} \]
(FPCore (x) :precision binary64 (fma (* (* x x) x) 0.3333333333333333 x))
double code(double x) {
	return fma(((x * x) * x), 0.3333333333333333, x);
}
function code(x)
	return fma(Float64(Float64(x * x) * x), 0.3333333333333333, x)
end
code[x_] := N[(N[(N[(x * x), $MachinePrecision] * x), $MachinePrecision] * 0.3333333333333333 + x), $MachinePrecision]
\begin{array}{l}

\\
\mathsf{fma}\left(\left(x \cdot x\right) \cdot x, 0.3333333333333333, x\right)
\end{array}
Derivation
  1. Initial program 9.2%

    \[\frac{1}{2} \cdot \log \left(\frac{1 + x}{1 - x}\right) \]
  2. Add Preprocessing
  3. Taylor expanded in x around 0

    \[\leadsto \color{blue}{x \cdot \left(1 + \frac{1}{3} \cdot {x}^{2}\right)} \]
  4. Step-by-step derivation
    1. +-commutativeN/A

      \[\leadsto x \cdot \color{blue}{\left(\frac{1}{3} \cdot {x}^{2} + 1\right)} \]
    2. distribute-lft-inN/A

      \[\leadsto \color{blue}{x \cdot \left(\frac{1}{3} \cdot {x}^{2}\right) + x \cdot 1} \]
    3. *-commutativeN/A

      \[\leadsto x \cdot \color{blue}{\left({x}^{2} \cdot \frac{1}{3}\right)} + x \cdot 1 \]
    4. associate-*r*N/A

      \[\leadsto \color{blue}{\left(x \cdot {x}^{2}\right) \cdot \frac{1}{3}} + x \cdot 1 \]
    5. unpow2N/A

      \[\leadsto \left(x \cdot \color{blue}{\left(x \cdot x\right)}\right) \cdot \frac{1}{3} + x \cdot 1 \]
    6. cube-multN/A

      \[\leadsto \color{blue}{{x}^{3}} \cdot \frac{1}{3} + x \cdot 1 \]
    7. *-rgt-identityN/A

      \[\leadsto {x}^{3} \cdot \frac{1}{3} + \color{blue}{x} \]
    8. lower-fma.f64N/A

      \[\leadsto \color{blue}{\mathsf{fma}\left({x}^{3}, \frac{1}{3}, x\right)} \]
    9. lower-pow.f6499.3

      \[\leadsto \mathsf{fma}\left(\color{blue}{{x}^{3}}, 0.3333333333333333, x\right) \]
  5. Applied rewrites99.3%

    \[\leadsto \color{blue}{\mathsf{fma}\left({x}^{3}, 0.3333333333333333, x\right)} \]
  6. Step-by-step derivation
    1. Applied rewrites99.3%

      \[\leadsto \mathsf{fma}\left(\left(x \cdot x\right) \cdot x, 0.3333333333333333, x\right) \]
    2. Add Preprocessing

    Alternative 7: 98.9% accurate, 12.2× speedup?

    \[\begin{array}{l} \\ \left(2 \cdot x\right) \cdot 0.5 \end{array} \]
    (FPCore (x) :precision binary64 (* (* 2.0 x) 0.5))
    double code(double x) {
    	return (2.0 * x) * 0.5;
    }
    
    real(8) function code(x)
        real(8), intent (in) :: x
        code = (2.0d0 * x) * 0.5d0
    end function
    
    public static double code(double x) {
    	return (2.0 * x) * 0.5;
    }
    
    def code(x):
    	return (2.0 * x) * 0.5
    
    function code(x)
    	return Float64(Float64(2.0 * x) * 0.5)
    end
    
    function tmp = code(x)
    	tmp = (2.0 * x) * 0.5;
    end
    
    code[x_] := N[(N[(2.0 * x), $MachinePrecision] * 0.5), $MachinePrecision]
    
    \begin{array}{l}
    
    \\
    \left(2 \cdot x\right) \cdot 0.5
    \end{array}
    
    Derivation
    1. Initial program 9.2%

      \[\frac{1}{2} \cdot \log \left(\frac{1 + x}{1 - x}\right) \]
    2. Add Preprocessing
    3. Taylor expanded in x around 0

      \[\leadsto \frac{1}{2} \cdot \color{blue}{\left(2 \cdot x\right)} \]
    4. Step-by-step derivation
      1. lower-*.f6498.8

        \[\leadsto \frac{1}{2} \cdot \color{blue}{\left(2 \cdot x\right)} \]
    5. Applied rewrites98.8%

      \[\leadsto \frac{1}{2} \cdot \color{blue}{\left(2 \cdot x\right)} \]
    6. Step-by-step derivation
      1. lift-*.f64N/A

        \[\leadsto \color{blue}{\frac{1}{2} \cdot \left(2 \cdot x\right)} \]
      2. *-commutativeN/A

        \[\leadsto \color{blue}{\left(2 \cdot x\right) \cdot \frac{1}{2}} \]
      3. lower-*.f6498.8

        \[\leadsto \color{blue}{\left(2 \cdot x\right) \cdot \frac{1}{2}} \]
      4. lift-/.f64N/A

        \[\leadsto \mathsf{Rewrite=>}\left(lower-*.f64, \left(x \cdot 2\right)\right) \cdot \color{blue}{\frac{1}{2}} \]
      5. metadata-evalN/A

        \[\leadsto \mathsf{Rewrite=>}\left(lower-*.f64, \left(x \cdot 2\right)\right) \cdot \color{blue}{\frac{1}{2}} \]
    7. Applied rewrites98.8%

      \[\leadsto \color{blue}{\left(x \cdot 2\right) \cdot 0.5} \]
    8. Final simplification98.8%

      \[\leadsto \left(2 \cdot x\right) \cdot 0.5 \]
    9. Add Preprocessing

    Reproduce

    ?
    herbie shell --seed 2024284 
    (FPCore (x)
      :name "Hyperbolic arc-(co)tangent"
      :precision binary64
      (* (/ 1.0 2.0) (log (/ (+ 1.0 x) (- 1.0 x)))))