| Alternative 1 | |
|---|---|
| Error | 0.4 |
| Cost | 6912 |
\[0.3333333333333333 \cdot x + 0.022222222222222223 \cdot {x}^{3}
\]
(FPCore (x) :precision binary64 (- (/ 1.0 x) (/ 1.0 (tan x))))
(FPCore (x) :precision binary64 (+ (* 0.3333333333333333 x) (+ (* 0.022222222222222223 (pow x 3.0)) (* 0.0021164021164021165 (pow x 5.0)))))
double code(double x) {
return (1.0 / x) - (1.0 / tan(x));
}
double code(double x) {
return (0.3333333333333333 * x) + ((0.022222222222222223 * pow(x, 3.0)) + (0.0021164021164021165 * pow(x, 5.0)));
}
real(8) function code(x)
real(8), intent (in) :: x
code = (1.0d0 / x) - (1.0d0 / tan(x))
end function
real(8) function code(x)
real(8), intent (in) :: x
code = (0.3333333333333333d0 * x) + ((0.022222222222222223d0 * (x ** 3.0d0)) + (0.0021164021164021165d0 * (x ** 5.0d0)))
end function
public static double code(double x) {
return (1.0 / x) - (1.0 / Math.tan(x));
}
public static double code(double x) {
return (0.3333333333333333 * x) + ((0.022222222222222223 * Math.pow(x, 3.0)) + (0.0021164021164021165 * Math.pow(x, 5.0)));
}
def code(x): return (1.0 / x) - (1.0 / math.tan(x))
def code(x): return (0.3333333333333333 * x) + ((0.022222222222222223 * math.pow(x, 3.0)) + (0.0021164021164021165 * math.pow(x, 5.0)))
function code(x) return Float64(Float64(1.0 / x) - Float64(1.0 / tan(x))) end
function code(x) return Float64(Float64(0.3333333333333333 * x) + Float64(Float64(0.022222222222222223 * (x ^ 3.0)) + Float64(0.0021164021164021165 * (x ^ 5.0)))) end
function tmp = code(x) tmp = (1.0 / x) - (1.0 / tan(x)); end
function tmp = code(x) tmp = (0.3333333333333333 * x) + ((0.022222222222222223 * (x ^ 3.0)) + (0.0021164021164021165 * (x ^ 5.0))); end
code[x_] := N[(N[(1.0 / x), $MachinePrecision] - N[(1.0 / N[Tan[x], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
code[x_] := N[(N[(0.3333333333333333 * x), $MachinePrecision] + N[(N[(0.022222222222222223 * N[Power[x, 3.0], $MachinePrecision]), $MachinePrecision] + N[(0.0021164021164021165 * N[Power[x, 5.0], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\frac{1}{x} - \frac{1}{\tan x}
0.3333333333333333 \cdot x + \left(0.022222222222222223 \cdot {x}^{3} + 0.0021164021164021165 \cdot {x}^{5}\right)
Results
| Original | 59.9 |
|---|---|
| Target | 0.1 |
| Herbie | 0.3 |
Initial program 59.9
Taylor expanded in x around 0 0.3
Simplified0.3
[Start]0.3 | \[ 0.3333333333333333 \cdot x + \left(0.0021164021164021165 \cdot {x}^{5} + 0.022222222222222223 \cdot {x}^{3}\right)
\] |
|---|---|
rational.json-simplify-1 [=>]0.3 | \[ 0.3333333333333333 \cdot x + \color{blue}{\left(0.022222222222222223 \cdot {x}^{3} + 0.0021164021164021165 \cdot {x}^{5}\right)}
\] |
Final simplification0.3
| Alternative 1 | |
|---|---|
| Error | 0.4 |
| Cost | 6912 |
| Alternative 2 | |
|---|---|
| Error | 0.7 |
| Cost | 192 |
herbie shell --seed 2023075
(FPCore (x)
:name "invcot (example 3.9)"
:precision binary64
:pre (and (< -0.026 x) (< x 0.026))
:herbie-target
(if (< (fabs x) 0.026) (* (/ x 3.0) (+ 1.0 (/ (* x x) 15.0))) (- (/ 1.0 x) (/ 1.0 (tan x))))
(- (/ 1.0 x) (/ 1.0 (tan x))))