?

Average Error: 14.7 → 0.4
Time: 10.0s
Precision: binary64
Cost: 6912

?

\[\tan^{-1} \left(N + 1\right) - \tan^{-1} N \]
\[\tan^{-1}_* \frac{1}{\left(1 + N\right) + N \cdot N} \]
(FPCore (N) :precision binary64 (- (atan (+ N 1.0)) (atan N)))
(FPCore (N) :precision binary64 (atan2 1.0 (+ (+ 1.0 N) (* N N))))
double code(double N) {
	return atan((N + 1.0)) - atan(N);
}
double code(double N) {
	return atan2(1.0, ((1.0 + N) + (N * N)));
}
real(8) function code(n)
    real(8), intent (in) :: n
    code = atan((n + 1.0d0)) - atan(n)
end function
real(8) function code(n)
    real(8), intent (in) :: n
    code = atan2(1.0d0, ((1.0d0 + n) + (n * n)))
end function
public static double code(double N) {
	return Math.atan((N + 1.0)) - Math.atan(N);
}
public static double code(double N) {
	return Math.atan2(1.0, ((1.0 + N) + (N * N)));
}
def code(N):
	return math.atan((N + 1.0)) - math.atan(N)
def code(N):
	return math.atan2(1.0, ((1.0 + N) + (N * N)))
function code(N)
	return Float64(atan(Float64(N + 1.0)) - atan(N))
end
function code(N)
	return atan(1.0, Float64(Float64(1.0 + N) + Float64(N * N)))
end
function tmp = code(N)
	tmp = atan((N + 1.0)) - atan(N);
end
function tmp = code(N)
	tmp = atan2(1.0, ((1.0 + N) + (N * N)));
end
code[N_] := N[(N[ArcTan[N[(N + 1.0), $MachinePrecision]], $MachinePrecision] - N[ArcTan[N], $MachinePrecision]), $MachinePrecision]
code[N_] := N[ArcTan[1.0 / N[(N[(1.0 + N), $MachinePrecision] + N[(N * N), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]
\tan^{-1} \left(N + 1\right) - \tan^{-1} N
\tan^{-1}_* \frac{1}{\left(1 + N\right) + N \cdot N}

Error?

Try it out?

Your Program's Arguments

Results

Enter valid numbers for all inputs

Target

Original14.7
Target0.4
Herbie0.4
\[\tan^{-1} \left(\frac{1}{1 + N \cdot \left(N + 1\right)}\right) \]

Derivation?

  1. Initial program 14.7

    \[\tan^{-1} \left(N + 1\right) - \tan^{-1} N \]
  2. Applied egg-rr15.2

    \[\leadsto \color{blue}{\log \left(e^{\tan^{-1} \left(N + 1\right) - \tan^{-1} N}\right)} \]
  3. Applied egg-rr0.4

    \[\leadsto \color{blue}{\tan^{-1}_* \frac{1 + \left(N - N\right)}{1 + \left(N + N \cdot N\right)}} \]
  4. Simplified0.4

    \[\leadsto \color{blue}{\tan^{-1}_* \frac{1}{\left(N + 1\right) + N \cdot N}} \]
    Proof

    [Start]0.4

    \[ \tan^{-1}_* \frac{1 + \left(N - N\right)}{1 + \left(N + N \cdot N\right)} \]

    +-commutative [=>]0.4

    \[ \tan^{-1}_* \frac{\color{blue}{\left(N - N\right) + 1}}{1 + \left(N + N \cdot N\right)} \]

    +-inverses [=>]0.4

    \[ \tan^{-1}_* \frac{\color{blue}{0} + 1}{1 + \left(N + N \cdot N\right)} \]

    metadata-eval [=>]0.4

    \[ \tan^{-1}_* \frac{\color{blue}{1}}{1 + \left(N + N \cdot N\right)} \]

    associate-+r+ [=>]0.4

    \[ \tan^{-1}_* \frac{1}{\color{blue}{\left(1 + N\right) + N \cdot N}} \]

    +-commutative [<=]0.4

    \[ \tan^{-1}_* \frac{1}{\color{blue}{\left(N + 1\right)} + N \cdot N} \]
  5. Final simplification0.4

    \[\leadsto \tan^{-1}_* \frac{1}{\left(1 + N\right) + N \cdot N} \]

Alternatives

Alternative 1
Error0.9
Cost7049
\[\begin{array}{l} \mathbf{if}\;N \leq -1 \lor \neg \left(N \leq 1\right):\\ \;\;\;\;\tan^{-1}_* \frac{1}{N + N \cdot N}\\ \mathbf{else}:\\ \;\;\;\;\tan^{-1}_* \frac{1}{1 + N}\\ \end{array} \]
Alternative 2
Error1.5
Cost6921
\[\begin{array}{l} \mathbf{if}\;N \leq -1 \lor \neg \left(N \leq 1\right):\\ \;\;\;\;\tan^{-1}_* \frac{1}{N \cdot N}\\ \mathbf{else}:\\ \;\;\;\;\tan^{-1}_* \frac{1}{1}\\ \end{array} \]
Alternative 3
Error1.2
Cost6921
\[\begin{array}{l} \mathbf{if}\;N \leq -0.62 \lor \neg \left(N \leq 1.6\right):\\ \;\;\;\;\tan^{-1}_* \frac{1}{N \cdot N}\\ \mathbf{else}:\\ \;\;\;\;\tan^{-1}_* \frac{1}{1 + N}\\ \end{array} \]
Alternative 4
Error30.9
Cost6528
\[\tan^{-1}_* \frac{1}{1} \]

Error

Reproduce?

herbie shell --seed 2023039 
(FPCore (N)
  :name "2atan (example 3.5)"
  :precision binary64

  :herbie-target
  (atan (/ 1.0 (+ 1.0 (* N (+ N 1.0)))))

  (- (atan (+ N 1.0)) (atan N)))