
(FPCore (x) :precision binary32 (atanh x))
float code(float x) {
return atanhf(x);
}
function code(x) return atanh(x) end
function tmp = code(x) tmp = atanh(x); end
\begin{array}{l}
\\
\tanh^{-1} x
\end{array}
Sampling outcomes in binary32 precision:
Herbie found 7 alternatives:
| Alternative | Accuracy | Speedup |
|---|
(FPCore (x) :precision binary32 (* 0.5 (log1p (/ (* 2.0 x) (- 1.0 x)))))
float code(float x) {
return 0.5f * log1pf(((2.0f * x) / (1.0f - x)));
}
function code(x) return Float32(Float32(0.5) * log1p(Float32(Float32(Float32(2.0) * x) / Float32(Float32(1.0) - x)))) end
\begin{array}{l}
\\
0.5 \cdot \mathsf{log1p}\left(\frac{2 \cdot x}{1 - x}\right)
\end{array}
(FPCore (x) :precision binary32 (* 0.5 (log1p (/ (* 2.0 (fma x x x)) (fma x (- x) 1.0)))))
float code(float x) {
return 0.5f * log1pf(((2.0f * fmaf(x, x, x)) / fmaf(x, -x, 1.0f)));
}
function code(x) return Float32(Float32(0.5) * log1p(Float32(Float32(Float32(2.0) * fma(x, x, x)) / fma(x, Float32(-x), Float32(1.0))))) end
\begin{array}{l}
\\
0.5 \cdot \mathsf{log1p}\left(\frac{2 \cdot \mathsf{fma}\left(x, x, x\right)}{\mathsf{fma}\left(x, -x, 1\right)}\right)
\end{array}
Initial program 99.8%
flip--N/A
associate-/r/N/A
associate-*l/N/A
/-lowering-/.f32N/A
associate-*l*N/A
+-commutativeN/A
distribute-rgt-outN/A
*-lowering-*.f32N/A
*-lft-identityN/A
accelerator-lowering-fma.f32N/A
metadata-evalN/A
sub-negN/A
+-commutativeN/A
distribute-rgt-neg-inN/A
accelerator-lowering-fma.f32N/A
neg-lowering-neg.f3299.9
Applied egg-rr99.9%
(FPCore (x) :precision binary32 (* 0.5 (log1p (/ (* 2.0 x) (- 1.0 x)))))
float code(float x) {
return 0.5f * log1pf(((2.0f * x) / (1.0f - x)));
}
function code(x) return Float32(Float32(0.5) * log1p(Float32(Float32(Float32(2.0) * x) / Float32(Float32(1.0) - x)))) end
\begin{array}{l}
\\
0.5 \cdot \mathsf{log1p}\left(\frac{2 \cdot x}{1 - x}\right)
\end{array}
Initial program 99.8%
(FPCore (x) :precision binary32 (fma (fma (* x x) (fma x (* x 0.14285714285714285) 0.2) 0.3333333333333333) (* x (* x x)) x))
float code(float x) {
return fmaf(fmaf((x * x), fmaf(x, (x * 0.14285714285714285f), 0.2f), 0.3333333333333333f), (x * (x * x)), x);
}
function code(x) return fma(fma(Float32(x * x), fma(x, Float32(x * Float32(0.14285714285714285)), Float32(0.2)), Float32(0.3333333333333333)), Float32(x * Float32(x * x)), x) end
\begin{array}{l}
\\
\mathsf{fma}\left(\mathsf{fma}\left(x \cdot x, \mathsf{fma}\left(x, x \cdot 0.14285714285714285, 0.2\right), 0.3333333333333333\right), x \cdot \left(x \cdot x\right), x\right)
\end{array}
Initial program 99.8%
Taylor expanded in x around 0
*-commutativeN/A
+-commutativeN/A
distribute-lft1-inN/A
*-commutativeN/A
associate-*l*N/A
unpow2N/A
unpow3N/A
accelerator-lowering-fma.f32N/A
Simplified99.4%
(FPCore (x) :precision binary32 (+ x (* (* x (* x x)) (fma (* x x) 0.2 0.3333333333333333))))
float code(float x) {
return x + ((x * (x * x)) * fmaf((x * x), 0.2f, 0.3333333333333333f));
}
function code(x) return Float32(x + Float32(Float32(x * Float32(x * x)) * fma(Float32(x * x), Float32(0.2), Float32(0.3333333333333333)))) end
\begin{array}{l}
\\
x + \left(x \cdot \left(x \cdot x\right)\right) \cdot \mathsf{fma}\left(x \cdot x, 0.2, 0.3333333333333333\right)
\end{array}
Initial program 99.8%
Taylor expanded in x around 0
distribute-rgt-inN/A
*-lft-identityN/A
+-commutativeN/A
*-commutativeN/A
associate-*l*N/A
unpow2N/A
unpow3N/A
accelerator-lowering-fma.f32N/A
+-commutativeN/A
*-commutativeN/A
accelerator-lowering-fma.f32N/A
unpow2N/A
*-lowering-*.f32N/A
cube-multN/A
unpow2N/A
*-lowering-*.f32N/A
unpow2N/A
*-lowering-*.f3299.1
Simplified99.1%
+-lowering-+.f32N/A
*-commutativeN/A
*-lowering-*.f32N/A
*-lowering-*.f32N/A
*-lowering-*.f32N/A
accelerator-lowering-fma.f32N/A
*-lowering-*.f3299.1
Applied egg-rr99.1%
Final simplification99.1%
(FPCore (x) :precision binary32 (fma (fma (* x x) 0.2 0.3333333333333333) (* x (* x x)) x))
float code(float x) {
return fmaf(fmaf((x * x), 0.2f, 0.3333333333333333f), (x * (x * x)), x);
}
function code(x) return fma(fma(Float32(x * x), Float32(0.2), Float32(0.3333333333333333)), Float32(x * Float32(x * x)), x) end
\begin{array}{l}
\\
\mathsf{fma}\left(\mathsf{fma}\left(x \cdot x, 0.2, 0.3333333333333333\right), x \cdot \left(x \cdot x\right), x\right)
\end{array}
Initial program 99.8%
Taylor expanded in x around 0
distribute-rgt-inN/A
*-lft-identityN/A
+-commutativeN/A
*-commutativeN/A
associate-*l*N/A
unpow2N/A
unpow3N/A
accelerator-lowering-fma.f32N/A
+-commutativeN/A
*-commutativeN/A
accelerator-lowering-fma.f32N/A
unpow2N/A
*-lowering-*.f32N/A
cube-multN/A
unpow2N/A
*-lowering-*.f32N/A
unpow2N/A
*-lowering-*.f3299.1
Simplified99.1%
(FPCore (x) :precision binary32 (fma 0.3333333333333333 (* x (* x x)) x))
float code(float x) {
return fmaf(0.3333333333333333f, (x * (x * x)), x);
}
function code(x) return fma(Float32(0.3333333333333333), Float32(x * Float32(x * x)), x) end
\begin{array}{l}
\\
\mathsf{fma}\left(0.3333333333333333, x \cdot \left(x \cdot x\right), x\right)
\end{array}
Initial program 99.8%
Taylor expanded in x around 0
distribute-rgt-inN/A
*-lft-identityN/A
+-commutativeN/A
associate-*l*N/A
unpow2N/A
unpow3N/A
accelerator-lowering-fma.f32N/A
cube-multN/A
unpow2N/A
*-lowering-*.f32N/A
unpow2N/A
*-lowering-*.f3298.5
Simplified98.5%
(FPCore (x) :precision binary32 x)
float code(float x) {
return x;
}
real(4) function code(x)
real(4), intent (in) :: x
code = x
end function
function code(x) return x end
function tmp = code(x) tmp = x; end
\begin{array}{l}
\\
x
\end{array}
Initial program 99.8%
Taylor expanded in x around 0
Simplified96.9%
herbie shell --seed 2024205
(FPCore (x)
:name "Rust f32::atanh"
:precision binary32
(* 0.5 (log1p (/ (* 2.0 x) (- 1.0 x)))))