
(FPCore (N) :precision binary64 (- (log (+ N 1.0)) (log N)))
double code(double N) {
return log((N + 1.0)) - log(N);
}
real(8) function code(n)
real(8), intent (in) :: n
code = log((n + 1.0d0)) - log(n)
end function
public static double code(double N) {
return Math.log((N + 1.0)) - Math.log(N);
}
def code(N): return math.log((N + 1.0)) - math.log(N)
function code(N) return Float64(log(Float64(N + 1.0)) - log(N)) end
function tmp = code(N) tmp = log((N + 1.0)) - log(N); end
code[N_] := N[(N[Log[N[(N + 1.0), $MachinePrecision]], $MachinePrecision] - N[Log[N], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\log \left(N + 1\right) - \log N
\end{array}
Sampling outcomes in binary64 precision:
Herbie found 7 alternatives:
| Alternative | Accuracy | Speedup |
|---|
(FPCore (N) :precision binary64 (- (log (+ N 1.0)) (log N)))
double code(double N) {
return log((N + 1.0)) - log(N);
}
real(8) function code(n)
real(8), intent (in) :: n
code = log((n + 1.0d0)) - log(n)
end function
public static double code(double N) {
return Math.log((N + 1.0)) - Math.log(N);
}
def code(N): return math.log((N + 1.0)) - math.log(N)
function code(N) return Float64(log(Float64(N + 1.0)) - log(N)) end
function tmp = code(N) tmp = log((N + 1.0)) - log(N); end
code[N_] := N[(N[Log[N[(N + 1.0), $MachinePrecision]], $MachinePrecision] - N[Log[N], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\log \left(N + 1\right) - \log N
\end{array}
(FPCore (N) :precision binary64 (log1p (pow N -1.0)))
double code(double N) {
return log1p(pow(N, -1.0));
}
public static double code(double N) {
return Math.log1p(Math.pow(N, -1.0));
}
def code(N): return math.log1p(math.pow(N, -1.0))
function code(N) return log1p((N ^ -1.0)) end
code[N_] := N[Log[1 + N[Power[N, -1.0], $MachinePrecision]], $MachinePrecision]
\begin{array}{l}
\\
\mathsf{log1p}\left({N}^{-1}\right)
\end{array}
Initial program 23.0%
lift--.f64N/A
lift-log.f64N/A
lift-log.f64N/A
diff-logN/A
lift-+.f64N/A
div-addN/A
*-inversesN/A
lower-log1p.f64N/A
inv-powN/A
lower-pow.f6499.8
Applied rewrites99.8%
lift-pow.f64N/A
inv-powN/A
lift-/.f6499.9
Applied rewrites99.9%
Final simplification99.8%
(FPCore (N) :precision binary64 (pow N -1.0))
double code(double N) {
return pow(N, -1.0);
}
real(8) function code(n)
real(8), intent (in) :: n
code = n ** (-1.0d0)
end function
public static double code(double N) {
return Math.pow(N, -1.0);
}
def code(N): return math.pow(N, -1.0)
function code(N) return N ^ -1.0 end
function tmp = code(N) tmp = N ^ -1.0; end
code[N_] := N[Power[N, -1.0], $MachinePrecision]
\begin{array}{l}
\\
{N}^{-1}
\end{array}
Initial program 23.0%
Taylor expanded in N around inf
lower-/.f6485.1
Applied rewrites85.1%
Final simplification85.1%
(FPCore (N) :precision binary64 (/ (- (/ (- (/ (+ (/ -0.25 N) 0.3333333333333333) N) 0.5) N) -1.0) N))
double code(double N) {
return ((((((-0.25 / N) + 0.3333333333333333) / N) - 0.5) / N) - -1.0) / N;
}
real(8) function code(n)
real(8), intent (in) :: n
code = (((((((-0.25d0) / n) + 0.3333333333333333d0) / n) - 0.5d0) / n) - (-1.0d0)) / n
end function
public static double code(double N) {
return ((((((-0.25 / N) + 0.3333333333333333) / N) - 0.5) / N) - -1.0) / N;
}
def code(N): return ((((((-0.25 / N) + 0.3333333333333333) / N) - 0.5) / N) - -1.0) / N
function code(N) return Float64(Float64(Float64(Float64(Float64(Float64(Float64(-0.25 / N) + 0.3333333333333333) / N) - 0.5) / N) - -1.0) / N) end
function tmp = code(N) tmp = ((((((-0.25 / N) + 0.3333333333333333) / N) - 0.5) / N) - -1.0) / N; end
code[N_] := N[(N[(N[(N[(N[(N[(N[(-0.25 / N), $MachinePrecision] + 0.3333333333333333), $MachinePrecision] / N), $MachinePrecision] - 0.5), $MachinePrecision] / N), $MachinePrecision] - -1.0), $MachinePrecision] / N), $MachinePrecision]
\begin{array}{l}
\\
\frac{\frac{\frac{\frac{-0.25}{N} + 0.3333333333333333}{N} - 0.5}{N} - -1}{N}
\end{array}
Initial program 23.0%
Taylor expanded in N around inf
Applied rewrites95.8%
(FPCore (N) :precision binary64 (/ (- (/ (- (/ 0.3333333333333333 N) 0.5) N) -1.0) N))
double code(double N) {
return ((((0.3333333333333333 / N) - 0.5) / N) - -1.0) / N;
}
real(8) function code(n)
real(8), intent (in) :: n
code = ((((0.3333333333333333d0 / n) - 0.5d0) / n) - (-1.0d0)) / n
end function
public static double code(double N) {
return ((((0.3333333333333333 / N) - 0.5) / N) - -1.0) / N;
}
def code(N): return ((((0.3333333333333333 / N) - 0.5) / N) - -1.0) / N
function code(N) return Float64(Float64(Float64(Float64(Float64(0.3333333333333333 / N) - 0.5) / N) - -1.0) / N) end
function tmp = code(N) tmp = ((((0.3333333333333333 / N) - 0.5) / N) - -1.0) / N; end
code[N_] := N[(N[(N[(N[(N[(0.3333333333333333 / N), $MachinePrecision] - 0.5), $MachinePrecision] / N), $MachinePrecision] - -1.0), $MachinePrecision] / N), $MachinePrecision]
\begin{array}{l}
\\
\frac{\frac{\frac{0.3333333333333333}{N} - 0.5}{N} - -1}{N}
\end{array}
Initial program 23.0%
Taylor expanded in N around inf
Applied rewrites94.7%
(FPCore (N) :precision binary64 (/ (* (- N) N) (* (- -0.5 N) (* N N))))
double code(double N) {
return (-N * N) / ((-0.5 - N) * (N * N));
}
real(8) function code(n)
real(8), intent (in) :: n
code = (-n * n) / (((-0.5d0) - n) * (n * n))
end function
public static double code(double N) {
return (-N * N) / ((-0.5 - N) * (N * N));
}
def code(N): return (-N * N) / ((-0.5 - N) * (N * N))
function code(N) return Float64(Float64(Float64(-N) * N) / Float64(Float64(-0.5 - N) * Float64(N * N))) end
function tmp = code(N) tmp = (-N * N) / ((-0.5 - N) * (N * N)); end
code[N_] := N[(N[((-N) * N), $MachinePrecision] / N[(N[(-0.5 - N), $MachinePrecision] * N[(N * N), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\frac{\left(-N\right) \cdot N}{\left(-0.5 - N\right) \cdot \left(N \cdot N\right)}
\end{array}
Initial program 23.0%
Taylor expanded in N around -inf
metadata-evalN/A
times-fracN/A
*-lft-identityN/A
associate-/r*N/A
lower-/.f64N/A
div-subN/A
metadata-evalN/A
distribute-neg-frac2N/A
distribute-neg-fracN/A
distribute-lft-neg-inN/A
/-rgt-identityN/A
metadata-evalN/A
lower--.f64N/A
metadata-evalN/A
associate-*r/N/A
metadata-evalN/A
lower-/.f6492.3
Applied rewrites92.3%
Taylor expanded in N around 0
Applied rewrites92.1%
Applied rewrites92.0%
Taylor expanded in N around inf
Applied rewrites92.6%
(FPCore (N) :precision binary64 (/ (- (/ -0.5 N) -1.0) N))
double code(double N) {
return ((-0.5 / N) - -1.0) / N;
}
real(8) function code(n)
real(8), intent (in) :: n
code = (((-0.5d0) / n) - (-1.0d0)) / n
end function
public static double code(double N) {
return ((-0.5 / N) - -1.0) / N;
}
def code(N): return ((-0.5 / N) - -1.0) / N
function code(N) return Float64(Float64(Float64(-0.5 / N) - -1.0) / N) end
function tmp = code(N) tmp = ((-0.5 / N) - -1.0) / N; end
code[N_] := N[(N[(N[(-0.5 / N), $MachinePrecision] - -1.0), $MachinePrecision] / N), $MachinePrecision]
\begin{array}{l}
\\
\frac{\frac{-0.5}{N} - -1}{N}
\end{array}
Initial program 23.0%
Taylor expanded in N around -inf
metadata-evalN/A
times-fracN/A
*-lft-identityN/A
associate-/r*N/A
lower-/.f64N/A
div-subN/A
metadata-evalN/A
distribute-neg-frac2N/A
distribute-neg-fracN/A
distribute-lft-neg-inN/A
/-rgt-identityN/A
metadata-evalN/A
lower--.f64N/A
metadata-evalN/A
associate-*r/N/A
metadata-evalN/A
lower-/.f6492.3
Applied rewrites92.3%
(FPCore (N) :precision binary64 (/ (+ -0.5 N) (* N N)))
double code(double N) {
return (-0.5 + N) / (N * N);
}
real(8) function code(n)
real(8), intent (in) :: n
code = ((-0.5d0) + n) / (n * n)
end function
public static double code(double N) {
return (-0.5 + N) / (N * N);
}
def code(N): return (-0.5 + N) / (N * N)
function code(N) return Float64(Float64(-0.5 + N) / Float64(N * N)) end
function tmp = code(N) tmp = (-0.5 + N) / (N * N); end
code[N_] := N[(N[(-0.5 + N), $MachinePrecision] / N[(N * N), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\frac{-0.5 + N}{N \cdot N}
\end{array}
Initial program 23.0%
Taylor expanded in N around -inf
metadata-evalN/A
times-fracN/A
*-lft-identityN/A
associate-/r*N/A
lower-/.f64N/A
div-subN/A
metadata-evalN/A
distribute-neg-frac2N/A
distribute-neg-fracN/A
distribute-lft-neg-inN/A
/-rgt-identityN/A
metadata-evalN/A
lower--.f64N/A
metadata-evalN/A
associate-*r/N/A
metadata-evalN/A
lower-/.f6492.3
Applied rewrites92.3%
Taylor expanded in N around 0
Applied rewrites92.1%
(FPCore (N) :precision binary64 (+ (+ (+ (/ 1.0 N) (/ -1.0 (* 2.0 (pow N 2.0)))) (/ 1.0 (* 3.0 (pow N 3.0)))) (/ -1.0 (* 4.0 (pow N 4.0)))))
double code(double N) {
return (((1.0 / N) + (-1.0 / (2.0 * pow(N, 2.0)))) + (1.0 / (3.0 * pow(N, 3.0)))) + (-1.0 / (4.0 * pow(N, 4.0)));
}
real(8) function code(n)
real(8), intent (in) :: n
code = (((1.0d0 / n) + ((-1.0d0) / (2.0d0 * (n ** 2.0d0)))) + (1.0d0 / (3.0d0 * (n ** 3.0d0)))) + ((-1.0d0) / (4.0d0 * (n ** 4.0d0)))
end function
public static double code(double N) {
return (((1.0 / N) + (-1.0 / (2.0 * Math.pow(N, 2.0)))) + (1.0 / (3.0 * Math.pow(N, 3.0)))) + (-1.0 / (4.0 * Math.pow(N, 4.0)));
}
def code(N): return (((1.0 / N) + (-1.0 / (2.0 * math.pow(N, 2.0)))) + (1.0 / (3.0 * math.pow(N, 3.0)))) + (-1.0 / (4.0 * math.pow(N, 4.0)))
function code(N) return Float64(Float64(Float64(Float64(1.0 / N) + Float64(-1.0 / Float64(2.0 * (N ^ 2.0)))) + Float64(1.0 / Float64(3.0 * (N ^ 3.0)))) + Float64(-1.0 / Float64(4.0 * (N ^ 4.0)))) end
function tmp = code(N) tmp = (((1.0 / N) + (-1.0 / (2.0 * (N ^ 2.0)))) + (1.0 / (3.0 * (N ^ 3.0)))) + (-1.0 / (4.0 * (N ^ 4.0))); end
code[N_] := N[(N[(N[(N[(1.0 / N), $MachinePrecision] + N[(-1.0 / N[(2.0 * N[Power[N, 2.0], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] + N[(1.0 / N[(3.0 * N[Power[N, 3.0], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] + N[(-1.0 / N[(4.0 * N[Power[N, 4.0], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\left(\left(\frac{1}{N} + \frac{-1}{2 \cdot {N}^{2}}\right) + \frac{1}{3 \cdot {N}^{3}}\right) + \frac{-1}{4 \cdot {N}^{4}}
\end{array}
herbie shell --seed 2024339
(FPCore (N)
:name "2log (problem 3.3.6)"
:precision binary64
:pre (and (> N 1.0) (< N 1e+40))
:alt
(! :herbie-platform default (+ (/ 1 N) (/ -1 (* 2 (pow N 2))) (/ 1 (* 3 (pow N 3))) (/ -1 (* 4 (pow N 4)))))
(- (log (+ N 1.0)) (log N)))