| Alternative 1 | |
|---|---|
| Accuracy | 98.5% |
| Cost | 7049 |
\[\begin{array}{l}
\mathbf{if}\;N \leq -1 \lor \neg \left(N \leq 1\right):\\
\;\;\;\;\tan^{-1}_* \frac{1}{N + N \cdot N}\\
\mathbf{else}:\\
\;\;\;\;\tan^{-1}_* \frac{1}{1 + N}\\
\end{array}
\]
(FPCore (N) :precision binary64 (- (atan (+ N 1.0)) (atan N)))
(FPCore (N) :precision binary64 (atan2 1.0 (+ 1.0 (+ N (* N N)))))
double code(double N) {
return atan((N + 1.0)) - atan(N);
}
double code(double N) {
return atan2(1.0, (1.0 + (N + (N * N))));
}
real(8) function code(n)
real(8), intent (in) :: n
code = atan((n + 1.0d0)) - atan(n)
end function
real(8) function code(n)
real(8), intent (in) :: n
code = atan2(1.0d0, (1.0d0 + (n + (n * n))))
end function
public static double code(double N) {
return Math.atan((N + 1.0)) - Math.atan(N);
}
public static double code(double N) {
return Math.atan2(1.0, (1.0 + (N + (N * N))));
}
def code(N): return math.atan((N + 1.0)) - math.atan(N)
def code(N): return math.atan2(1.0, (1.0 + (N + (N * N))))
function code(N) return Float64(atan(Float64(N + 1.0)) - atan(N)) end
function code(N) return atan(1.0, Float64(1.0 + Float64(N + Float64(N * N)))) end
function tmp = code(N) tmp = atan((N + 1.0)) - atan(N); end
function tmp = code(N) tmp = atan2(1.0, (1.0 + (N + (N * N)))); end
code[N_] := N[(N[ArcTan[N[(N + 1.0), $MachinePrecision]], $MachinePrecision] - N[ArcTan[N], $MachinePrecision]), $MachinePrecision]
code[N_] := N[ArcTan[1.0 / N[(1.0 + N[(N + N[(N * N), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]
\tan^{-1} \left(N + 1\right) - \tan^{-1} N
\tan^{-1}_* \frac{1}{1 + \left(N + N \cdot N\right)}
Results
| Original | 75.2% |
|---|---|
| Target | 99.4% |
| Herbie | 99.4% |
Initial program 79.5%
Applied egg-rr82.0%
[Start]79.5 | \[ \tan^{-1} \left(N + 1\right) - \tan^{-1} N
\] |
|---|---|
diff-atan [=>]81.9 | \[ \color{blue}{\tan^{-1}_* \frac{\left(N + 1\right) - N}{1 + \left(N + 1\right) \cdot N}}
\] |
associate--l+ [=>]82.0 | \[ \tan^{-1}_* \frac{\color{blue}{N + \left(1 - N\right)}}{1 + \left(N + 1\right) \cdot N}
\] |
+-commutative [=>]82.0 | \[ \tan^{-1}_* \frac{N + \left(1 - N\right)}{\color{blue}{\left(N + 1\right) \cdot N + 1}}
\] |
*-commutative [=>]82.0 | \[ \tan^{-1}_* \frac{N + \left(1 - N\right)}{\color{blue}{N \cdot \left(N + 1\right)} + 1}
\] |
fma-def [=>]82.0 | \[ \tan^{-1}_* \frac{N + \left(1 - N\right)}{\color{blue}{\mathsf{fma}\left(N, N + 1, 1\right)}}
\] |
Simplified99.5%
[Start]82.0 | \[ \tan^{-1}_* \frac{N + \left(1 - N\right)}{\mathsf{fma}\left(N, N + 1, 1\right)}
\] |
|---|---|
+-commutative [=>]82.0 | \[ \tan^{-1}_* \frac{\color{blue}{\left(1 - N\right) + N}}{\mathsf{fma}\left(N, N + 1, 1\right)}
\] |
associate-+l- [=>]99.5 | \[ \tan^{-1}_* \frac{\color{blue}{1 - \left(N - N\right)}}{\mathsf{fma}\left(N, N + 1, 1\right)}
\] |
+-inverses [=>]99.5 | \[ \tan^{-1}_* \frac{1 - \color{blue}{0}}{\mathsf{fma}\left(N, N + 1, 1\right)}
\] |
metadata-eval [=>]99.5 | \[ \tan^{-1}_* \frac{\color{blue}{1}}{\mathsf{fma}\left(N, N + 1, 1\right)}
\] |
+-commutative [=>]99.5 | \[ \tan^{-1}_* \frac{1}{\mathsf{fma}\left(N, \color{blue}{1 + N}, 1\right)}
\] |
Applied egg-rr99.5%
[Start]99.5 | \[ \tan^{-1}_* \frac{1}{\mathsf{fma}\left(N, 1 + N, 1\right)}
\] |
|---|---|
fma-udef [=>]99.5 | \[ \tan^{-1}_* \frac{1}{\color{blue}{N \cdot \left(1 + N\right) + 1}}
\] |
distribute-rgt-in [=>]99.5 | \[ \tan^{-1}_* \frac{1}{\color{blue}{\left(1 \cdot N + N \cdot N\right)} + 1}
\] |
*-un-lft-identity [<=]99.5 | \[ \tan^{-1}_* \frac{1}{\left(\color{blue}{N} + N \cdot N\right) + 1}
\] |
Final simplification99.5%
| Alternative 1 | |
|---|---|
| Accuracy | 98.5% |
| Cost | 7049 |
| Alternative 2 | |
|---|---|
| Accuracy | 97.5% |
| Cost | 6921 |
| Alternative 3 | |
|---|---|
| Accuracy | 97.9% |
| Cost | 6921 |
| Alternative 4 | |
|---|---|
| Accuracy | 50.7% |
| Cost | 6528 |
herbie shell --seed 2023157
(FPCore (N)
:name "2atan (example 3.5)"
:precision binary64
:herbie-target
(atan (/ 1.0 (+ 1.0 (* N (+ N 1.0)))))
(- (atan (+ N 1.0)) (atan N)))