| Alternative 1 | |
|---|---|
| Accuracy | 98.9% |
| Cost | 9092 |
(FPCore (wj x) :precision binary64 (- wj (/ (- (* wj (exp wj)) x) (+ (exp wj) (* wj (exp wj))))))
(FPCore (wj x)
:precision binary64
(let* ((t_0 (* wj (exp wj))))
(if (<= (+ wj (/ (- x t_0) (+ (exp wj) t_0))) 2e-10)
(+
(*
(pow wj 3.0)
(+
(* x -0.6666666666666666)
(+ (* x 3.0) (+ -1.0 (* -2.0 (+ (* x 4.0) (* x -1.5)))))))
(+ (* (* wj wj) (+ 1.0 (* x 2.5))) (+ x (* -2.0 (* wj x)))))
(fma (- (/ x (exp wj)) wj) (/ 1.0 (+ wj 1.0)) wj))))double code(double wj, double x) {
return wj - (((wj * exp(wj)) - x) / (exp(wj) + (wj * exp(wj))));
}
double code(double wj, double x) {
double t_0 = wj * exp(wj);
double tmp;
if ((wj + ((x - t_0) / (exp(wj) + t_0))) <= 2e-10) {
tmp = (pow(wj, 3.0) * ((x * -0.6666666666666666) + ((x * 3.0) + (-1.0 + (-2.0 * ((x * 4.0) + (x * -1.5))))))) + (((wj * wj) * (1.0 + (x * 2.5))) + (x + (-2.0 * (wj * x))));
} else {
tmp = fma(((x / exp(wj)) - wj), (1.0 / (wj + 1.0)), wj);
}
return tmp;
}
function code(wj, x) return Float64(wj - Float64(Float64(Float64(wj * exp(wj)) - x) / Float64(exp(wj) + Float64(wj * exp(wj))))) end
function code(wj, x) t_0 = Float64(wj * exp(wj)) tmp = 0.0 if (Float64(wj + Float64(Float64(x - t_0) / Float64(exp(wj) + t_0))) <= 2e-10) tmp = Float64(Float64((wj ^ 3.0) * Float64(Float64(x * -0.6666666666666666) + Float64(Float64(x * 3.0) + Float64(-1.0 + Float64(-2.0 * Float64(Float64(x * 4.0) + Float64(x * -1.5))))))) + Float64(Float64(Float64(wj * wj) * Float64(1.0 + Float64(x * 2.5))) + Float64(x + Float64(-2.0 * Float64(wj * x))))); else tmp = fma(Float64(Float64(x / exp(wj)) - wj), Float64(1.0 / Float64(wj + 1.0)), wj); end return tmp end
code[wj_, x_] := N[(wj - N[(N[(N[(wj * N[Exp[wj], $MachinePrecision]), $MachinePrecision] - x), $MachinePrecision] / N[(N[Exp[wj], $MachinePrecision] + N[(wj * N[Exp[wj], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
code[wj_, x_] := Block[{t$95$0 = N[(wj * N[Exp[wj], $MachinePrecision]), $MachinePrecision]}, If[LessEqual[N[(wj + N[(N[(x - t$95$0), $MachinePrecision] / N[(N[Exp[wj], $MachinePrecision] + t$95$0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision], 2e-10], N[(N[(N[Power[wj, 3.0], $MachinePrecision] * N[(N[(x * -0.6666666666666666), $MachinePrecision] + N[(N[(x * 3.0), $MachinePrecision] + N[(-1.0 + N[(-2.0 * N[(N[(x * 4.0), $MachinePrecision] + N[(x * -1.5), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] + N[(N[(N[(wj * wj), $MachinePrecision] * N[(1.0 + N[(x * 2.5), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] + N[(x + N[(-2.0 * N[(wj * x), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision], N[(N[(N[(x / N[Exp[wj], $MachinePrecision]), $MachinePrecision] - wj), $MachinePrecision] * N[(1.0 / N[(wj + 1.0), $MachinePrecision]), $MachinePrecision] + wj), $MachinePrecision]]]
wj - \frac{wj \cdot e^{wj} - x}{e^{wj} + wj \cdot e^{wj}}
\begin{array}{l}
t_0 := wj \cdot e^{wj}\\
\mathbf{if}\;wj + \frac{x - t_0}{e^{wj} + t_0} \leq 2 \cdot 10^{-10}:\\
\;\;\;\;{wj}^{3} \cdot \left(x \cdot -0.6666666666666666 + \left(x \cdot 3 + \left(-1 + -2 \cdot \left(x \cdot 4 + x \cdot -1.5\right)\right)\right)\right) + \left(\left(wj \cdot wj\right) \cdot \left(1 + x \cdot 2.5\right) + \left(x + -2 \cdot \left(wj \cdot x\right)\right)\right)\\
\mathbf{else}:\\
\;\;\;\;\mathsf{fma}\left(\frac{x}{e^{wj}} - wj, \frac{1}{wj + 1}, wj\right)\\
\end{array}
| Original | 78.7% |
|---|---|
| Target | 79.5% |
| Herbie | 99.3% |
if (-.f64 wj (/.f64 (-.f64 (*.f64 wj (exp.f64 wj)) x) (+.f64 (exp.f64 wj) (*.f64 wj (exp.f64 wj))))) < 2.00000000000000007e-10Initial program 72.1%
Simplified72.1%
[Start]72.1 | \[ wj - \frac{wj \cdot e^{wj} - x}{e^{wj} + wj \cdot e^{wj}}
\] |
|---|---|
sub-neg [=>]72.1 | \[ \color{blue}{wj + \left(-\frac{wj \cdot e^{wj} - x}{e^{wj} + wj \cdot e^{wj}}\right)}
\] |
neg-mul-1 [=>]72.1 | \[ wj + \color{blue}{-1 \cdot \frac{wj \cdot e^{wj} - x}{e^{wj} + wj \cdot e^{wj}}}
\] |
*-commutative [=>]72.1 | \[ wj + \color{blue}{\frac{wj \cdot e^{wj} - x}{e^{wj} + wj \cdot e^{wj}} \cdot -1}
\] |
*-commutative [<=]72.1 | \[ wj + \color{blue}{-1 \cdot \frac{wj \cdot e^{wj} - x}{e^{wj} + wj \cdot e^{wj}}}
\] |
neg-mul-1 [<=]72.1 | \[ wj + \color{blue}{\left(-\frac{wj \cdot e^{wj} - x}{e^{wj} + wj \cdot e^{wj}}\right)}
\] |
neg-sub0 [=>]72.1 | \[ wj + \color{blue}{\left(0 - \frac{wj \cdot e^{wj} - x}{e^{wj} + wj \cdot e^{wj}}\right)}
\] |
div-sub [=>]72.1 | \[ wj + \left(0 - \color{blue}{\left(\frac{wj \cdot e^{wj}}{e^{wj} + wj \cdot e^{wj}} - \frac{x}{e^{wj} + wj \cdot e^{wj}}\right)}\right)
\] |
associate--r- [=>]72.1 | \[ wj + \color{blue}{\left(\left(0 - \frac{wj \cdot e^{wj}}{e^{wj} + wj \cdot e^{wj}}\right) + \frac{x}{e^{wj} + wj \cdot e^{wj}}\right)}
\] |
+-commutative [=>]72.1 | \[ wj + \color{blue}{\left(\frac{x}{e^{wj} + wj \cdot e^{wj}} + \left(0 - \frac{wj \cdot e^{wj}}{e^{wj} + wj \cdot e^{wj}}\right)\right)}
\] |
sub0-neg [=>]72.1 | \[ wj + \left(\frac{x}{e^{wj} + wj \cdot e^{wj}} + \color{blue}{\left(-\frac{wj \cdot e^{wj}}{e^{wj} + wj \cdot e^{wj}}\right)}\right)
\] |
sub-neg [<=]72.1 | \[ wj + \color{blue}{\left(\frac{x}{e^{wj} + wj \cdot e^{wj}} - \frac{wj \cdot e^{wj}}{e^{wj} + wj \cdot e^{wj}}\right)}
\] |
Taylor expanded in wj around 0 99.1%
Taylor expanded in x around 0 99.1%
Simplified99.1%
[Start]99.1 | \[ -1 \cdot \left(\left(0.6666666666666666 \cdot x + \left(-3 \cdot x + \left(1 + -2 \cdot \left(-4 \cdot x + 1.5 \cdot x\right)\right)\right)\right) \cdot {wj}^{3}\right) + \left(\left({wj}^{2} + 2.5 \cdot \left({wj}^{2} \cdot x\right)\right) + \left(-2 \cdot \left(wj \cdot x\right) + x\right)\right)
\] |
|---|---|
unpow2 [=>]99.1 | \[ -1 \cdot \left(\left(0.6666666666666666 \cdot x + \left(-3 \cdot x + \left(1 + -2 \cdot \left(-4 \cdot x + 1.5 \cdot x\right)\right)\right)\right) \cdot {wj}^{3}\right) + \left(\left(\color{blue}{wj \cdot wj} + 2.5 \cdot \left({wj}^{2} \cdot x\right)\right) + \left(-2 \cdot \left(wj \cdot x\right) + x\right)\right)
\] |
*-commutative [=>]99.1 | \[ -1 \cdot \left(\left(0.6666666666666666 \cdot x + \left(-3 \cdot x + \left(1 + -2 \cdot \left(-4 \cdot x + 1.5 \cdot x\right)\right)\right)\right) \cdot {wj}^{3}\right) + \left(\left(wj \cdot wj + \color{blue}{\left({wj}^{2} \cdot x\right) \cdot 2.5}\right) + \left(-2 \cdot \left(wj \cdot x\right) + x\right)\right)
\] |
unpow2 [=>]99.1 | \[ -1 \cdot \left(\left(0.6666666666666666 \cdot x + \left(-3 \cdot x + \left(1 + -2 \cdot \left(-4 \cdot x + 1.5 \cdot x\right)\right)\right)\right) \cdot {wj}^{3}\right) + \left(\left(wj \cdot wj + \left(\color{blue}{\left(wj \cdot wj\right)} \cdot x\right) \cdot 2.5\right) + \left(-2 \cdot \left(wj \cdot x\right) + x\right)\right)
\] |
associate-*r* [<=]99.1 | \[ -1 \cdot \left(\left(0.6666666666666666 \cdot x + \left(-3 \cdot x + \left(1 + -2 \cdot \left(-4 \cdot x + 1.5 \cdot x\right)\right)\right)\right) \cdot {wj}^{3}\right) + \left(\left(wj \cdot wj + \color{blue}{\left(wj \cdot wj\right) \cdot \left(x \cdot 2.5\right)}\right) + \left(-2 \cdot \left(wj \cdot x\right) + x\right)\right)
\] |
*-rgt-identity [<=]99.1 | \[ -1 \cdot \left(\left(0.6666666666666666 \cdot x + \left(-3 \cdot x + \left(1 + -2 \cdot \left(-4 \cdot x + 1.5 \cdot x\right)\right)\right)\right) \cdot {wj}^{3}\right) + \left(\left(\color{blue}{\left(wj \cdot wj\right) \cdot 1} + \left(wj \cdot wj\right) \cdot \left(x \cdot 2.5\right)\right) + \left(-2 \cdot \left(wj \cdot x\right) + x\right)\right)
\] |
distribute-lft-in [<=]99.1 | \[ -1 \cdot \left(\left(0.6666666666666666 \cdot x + \left(-3 \cdot x + \left(1 + -2 \cdot \left(-4 \cdot x + 1.5 \cdot x\right)\right)\right)\right) \cdot {wj}^{3}\right) + \left(\color{blue}{\left(wj \cdot wj\right) \cdot \left(1 + x \cdot 2.5\right)} + \left(-2 \cdot \left(wj \cdot x\right) + x\right)\right)
\] |
*-commutative [=>]99.1 | \[ -1 \cdot \left(\left(0.6666666666666666 \cdot x + \left(-3 \cdot x + \left(1 + -2 \cdot \left(-4 \cdot x + 1.5 \cdot x\right)\right)\right)\right) \cdot {wj}^{3}\right) + \left(\left(wj \cdot wj\right) \cdot \left(1 + \color{blue}{2.5 \cdot x}\right) + \left(-2 \cdot \left(wj \cdot x\right) + x\right)\right)
\] |
if 2.00000000000000007e-10 < (-.f64 wj (/.f64 (-.f64 (*.f64 wj (exp.f64 wj)) x) (+.f64 (exp.f64 wj) (*.f64 wj (exp.f64 wj))))) Initial program 96.4%
Simplified99.6%
[Start]96.4 | \[ wj - \frac{wj \cdot e^{wj} - x}{e^{wj} + wj \cdot e^{wj}}
\] |
|---|---|
sub-neg [=>]96.4 | \[ \color{blue}{wj + \left(-\frac{wj \cdot e^{wj} - x}{e^{wj} + wj \cdot e^{wj}}\right)}
\] |
neg-mul-1 [=>]96.4 | \[ wj + \color{blue}{-1 \cdot \frac{wj \cdot e^{wj} - x}{e^{wj} + wj \cdot e^{wj}}}
\] |
*-commutative [=>]96.4 | \[ wj + \color{blue}{\frac{wj \cdot e^{wj} - x}{e^{wj} + wj \cdot e^{wj}} \cdot -1}
\] |
*-commutative [<=]96.4 | \[ wj + \color{blue}{-1 \cdot \frac{wj \cdot e^{wj} - x}{e^{wj} + wj \cdot e^{wj}}}
\] |
neg-mul-1 [<=]96.4 | \[ wj + \color{blue}{\left(-\frac{wj \cdot e^{wj} - x}{e^{wj} + wj \cdot e^{wj}}\right)}
\] |
neg-sub0 [=>]96.4 | \[ wj + \color{blue}{\left(0 - \frac{wj \cdot e^{wj} - x}{e^{wj} + wj \cdot e^{wj}}\right)}
\] |
div-sub [=>]96.4 | \[ wj + \left(0 - \color{blue}{\left(\frac{wj \cdot e^{wj}}{e^{wj} + wj \cdot e^{wj}} - \frac{x}{e^{wj} + wj \cdot e^{wj}}\right)}\right)
\] |
associate--r- [=>]96.4 | \[ wj + \color{blue}{\left(\left(0 - \frac{wj \cdot e^{wj}}{e^{wj} + wj \cdot e^{wj}}\right) + \frac{x}{e^{wj} + wj \cdot e^{wj}}\right)}
\] |
+-commutative [=>]96.4 | \[ wj + \color{blue}{\left(\frac{x}{e^{wj} + wj \cdot e^{wj}} + \left(0 - \frac{wj \cdot e^{wj}}{e^{wj} + wj \cdot e^{wj}}\right)\right)}
\] |
sub0-neg [=>]96.4 | \[ wj + \left(\frac{x}{e^{wj} + wj \cdot e^{wj}} + \color{blue}{\left(-\frac{wj \cdot e^{wj}}{e^{wj} + wj \cdot e^{wj}}\right)}\right)
\] |
sub-neg [<=]96.4 | \[ wj + \color{blue}{\left(\frac{x}{e^{wj} + wj \cdot e^{wj}} - \frac{wj \cdot e^{wj}}{e^{wj} + wj \cdot e^{wj}}\right)}
\] |
Applied egg-rr99.6%
Final simplification99.3%
| Alternative 1 | |
|---|---|
| Accuracy | 98.9% |
| Cost | 9092 |
| Alternative 2 | |
|---|---|
| Accuracy | 98.5% |
| Cost | 7812 |
| Alternative 3 | |
|---|---|
| Accuracy | 99.0% |
| Cost | 7812 |
| Alternative 4 | |
|---|---|
| Accuracy | 98.6% |
| Cost | 7492 |
| Alternative 5 | |
|---|---|
| Accuracy | 88.4% |
| Cost | 7236 |
| Alternative 6 | |
|---|---|
| Accuracy | 86.7% |
| Cost | 6848 |
| Alternative 7 | |
|---|---|
| Accuracy | 86.3% |
| Cost | 960 |
| Alternative 8 | |
|---|---|
| Accuracy | 86.0% |
| Cost | 448 |
| Alternative 9 | |
|---|---|
| Accuracy | 86.0% |
| Cost | 448 |
| Alternative 10 | |
|---|---|
| Accuracy | 4.3% |
| Cost | 64 |
| Alternative 11 | |
|---|---|
| Accuracy | 85.5% |
| Cost | 64 |
herbie shell --seed 2023129
(FPCore (wj x)
:name "Jmat.Real.lambertw, newton loop step"
:precision binary64
:herbie-target
(- wj (- (/ wj (+ wj 1.0)) (/ x (+ (exp wj) (* wj (exp wj))))))
(- wj (/ (- (* wj (exp wj)) x) (+ (exp wj) (* wj (exp wj))))))