
(FPCore (x eps) :precision binary64 (- (cos (+ x eps)) (cos x)))
double code(double x, double eps) {
return cos((x + eps)) - cos(x);
}
real(8) function code(x, eps)
real(8), intent (in) :: x
real(8), intent (in) :: eps
code = cos((x + eps)) - cos(x)
end function
public static double code(double x, double eps) {
return Math.cos((x + eps)) - Math.cos(x);
}
def code(x, eps): return math.cos((x + eps)) - math.cos(x)
function code(x, eps) return Float64(cos(Float64(x + eps)) - cos(x)) end
function tmp = code(x, eps) tmp = cos((x + eps)) - cos(x); end
code[x_, eps_] := N[(N[Cos[N[(x + eps), $MachinePrecision]], $MachinePrecision] - N[Cos[x], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\cos \left(x + \varepsilon\right) - \cos x
\end{array}
Sampling outcomes in binary64 precision:
Herbie found 14 alternatives:
| Alternative | Accuracy | Speedup |
|---|
(FPCore (x eps) :precision binary64 (- (cos (+ x eps)) (cos x)))
double code(double x, double eps) {
return cos((x + eps)) - cos(x);
}
real(8) function code(x, eps)
real(8), intent (in) :: x
real(8), intent (in) :: eps
code = cos((x + eps)) - cos(x)
end function
public static double code(double x, double eps) {
return Math.cos((x + eps)) - Math.cos(x);
}
def code(x, eps): return math.cos((x + eps)) - math.cos(x)
function code(x, eps) return Float64(cos(Float64(x + eps)) - cos(x)) end
function tmp = code(x, eps) tmp = cos((x + eps)) - cos(x); end
code[x_, eps_] := N[(N[Cos[N[(x + eps), $MachinePrecision]], $MachinePrecision] - N[Cos[x], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\cos \left(x + \varepsilon\right) - \cos x
\end{array}
(FPCore (x eps) :precision binary64 (* -2.0 (* (sin (* eps 0.5)) (sin (* (fma 2.0 x eps) 0.5)))))
double code(double x, double eps) {
return -2.0 * (sin((eps * 0.5)) * sin((fma(2.0, x, eps) * 0.5)));
}
function code(x, eps) return Float64(-2.0 * Float64(sin(Float64(eps * 0.5)) * sin(Float64(fma(2.0, x, eps) * 0.5)))) end
code[x_, eps_] := N[(-2.0 * N[(N[Sin[N[(eps * 0.5), $MachinePrecision]], $MachinePrecision] * N[Sin[N[(N[(2.0 * x + eps), $MachinePrecision] * 0.5), $MachinePrecision]], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
-2 \cdot \left(\sin \left(\varepsilon \cdot 0.5\right) \cdot \sin \left(\mathsf{fma}\left(2, x, \varepsilon\right) \cdot 0.5\right)\right)
\end{array}
Initial program 53.7%
lift--.f64N/A
lift-cos.f64N/A
lift-cos.f64N/A
diff-cosN/A
*-commutativeN/A
lower-*.f64N/A
Applied rewrites99.8%
Final simplification99.8%
(FPCore (x eps) :precision binary64 (* (* (* (fma (* eps eps) -0.020833333333333332 0.5) eps) (sin (* (fma 2.0 x eps) 0.5))) -2.0))
double code(double x, double eps) {
return ((fma((eps * eps), -0.020833333333333332, 0.5) * eps) * sin((fma(2.0, x, eps) * 0.5))) * -2.0;
}
function code(x, eps) return Float64(Float64(Float64(fma(Float64(eps * eps), -0.020833333333333332, 0.5) * eps) * sin(Float64(fma(2.0, x, eps) * 0.5))) * -2.0) end
code[x_, eps_] := N[(N[(N[(N[(N[(eps * eps), $MachinePrecision] * -0.020833333333333332 + 0.5), $MachinePrecision] * eps), $MachinePrecision] * N[Sin[N[(N[(2.0 * x + eps), $MachinePrecision] * 0.5), $MachinePrecision]], $MachinePrecision]), $MachinePrecision] * -2.0), $MachinePrecision]
\begin{array}{l}
\\
\left(\left(\mathsf{fma}\left(\varepsilon \cdot \varepsilon, -0.020833333333333332, 0.5\right) \cdot \varepsilon\right) \cdot \sin \left(\mathsf{fma}\left(2, x, \varepsilon\right) \cdot 0.5\right)\right) \cdot -2
\end{array}
Initial program 53.7%
lift--.f64N/A
lift-cos.f64N/A
lift-cos.f64N/A
diff-cosN/A
*-commutativeN/A
lower-*.f64N/A
Applied rewrites99.8%
Taylor expanded in eps around 0
*-commutativeN/A
lower-*.f64N/A
+-commutativeN/A
*-commutativeN/A
lower-fma.f64N/A
unpow2N/A
lower-*.f6499.4
Applied rewrites99.4%
Final simplification99.4%
(FPCore (x eps) :precision binary64 (* (* (* eps 0.5) (sin (* (fma 2.0 x eps) 0.5))) -2.0))
double code(double x, double eps) {
return ((eps * 0.5) * sin((fma(2.0, x, eps) * 0.5))) * -2.0;
}
function code(x, eps) return Float64(Float64(Float64(eps * 0.5) * sin(Float64(fma(2.0, x, eps) * 0.5))) * -2.0) end
code[x_, eps_] := N[(N[(N[(eps * 0.5), $MachinePrecision] * N[Sin[N[(N[(2.0 * x + eps), $MachinePrecision] * 0.5), $MachinePrecision]], $MachinePrecision]), $MachinePrecision] * -2.0), $MachinePrecision]
\begin{array}{l}
\\
\left(\left(\varepsilon \cdot 0.5\right) \cdot \sin \left(\mathsf{fma}\left(2, x, \varepsilon\right) \cdot 0.5\right)\right) \cdot -2
\end{array}
Initial program 53.7%
lift--.f64N/A
lift-cos.f64N/A
lift-cos.f64N/A
diff-cosN/A
*-commutativeN/A
lower-*.f64N/A
Applied rewrites99.8%
Taylor expanded in eps around 0
*-commutativeN/A
lower-*.f6499.3
Applied rewrites99.3%
Final simplification99.3%
(FPCore (x eps)
:precision binary64
(*
(fma
(fma
(fma
(fma -0.027777777777777776 (* eps eps) 0.16666666666666666)
x
(* (fma (* eps eps) -0.020833333333333332 0.25) eps))
x
(fma (* 0.16666666666666666 eps) eps -1.0))
x
(* (fma (* eps eps) 0.041666666666666664 -0.5) eps))
eps))
double code(double x, double eps) {
return fma(fma(fma(fma(-0.027777777777777776, (eps * eps), 0.16666666666666666), x, (fma((eps * eps), -0.020833333333333332, 0.25) * eps)), x, fma((0.16666666666666666 * eps), eps, -1.0)), x, (fma((eps * eps), 0.041666666666666664, -0.5) * eps)) * eps;
}
function code(x, eps) return Float64(fma(fma(fma(fma(-0.027777777777777776, Float64(eps * eps), 0.16666666666666666), x, Float64(fma(Float64(eps * eps), -0.020833333333333332, 0.25) * eps)), x, fma(Float64(0.16666666666666666 * eps), eps, -1.0)), x, Float64(fma(Float64(eps * eps), 0.041666666666666664, -0.5) * eps)) * eps) end
code[x_, eps_] := N[(N[(N[(N[(N[(-0.027777777777777776 * N[(eps * eps), $MachinePrecision] + 0.16666666666666666), $MachinePrecision] * x + N[(N[(N[(eps * eps), $MachinePrecision] * -0.020833333333333332 + 0.25), $MachinePrecision] * eps), $MachinePrecision]), $MachinePrecision] * x + N[(N[(0.16666666666666666 * eps), $MachinePrecision] * eps + -1.0), $MachinePrecision]), $MachinePrecision] * x + N[(N[(N[(eps * eps), $MachinePrecision] * 0.041666666666666664 + -0.5), $MachinePrecision] * eps), $MachinePrecision]), $MachinePrecision] * eps), $MachinePrecision]
\begin{array}{l}
\\
\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(-0.027777777777777776, \varepsilon \cdot \varepsilon, 0.16666666666666666\right), x, \mathsf{fma}\left(\varepsilon \cdot \varepsilon, -0.020833333333333332, 0.25\right) \cdot \varepsilon\right), x, \mathsf{fma}\left(0.16666666666666666 \cdot \varepsilon, \varepsilon, -1\right)\right), x, \mathsf{fma}\left(\varepsilon \cdot \varepsilon, 0.041666666666666664, -0.5\right) \cdot \varepsilon\right) \cdot \varepsilon
\end{array}
Initial program 53.7%
Taylor expanded in eps around 0
*-commutativeN/A
lower-*.f64N/A
Applied rewrites99.4%
Taylor expanded in x around 0
Applied rewrites99.0%
(FPCore (x eps)
:precision binary64
(fma
(* (fma (* eps eps) 0.041666666666666664 -0.5) eps)
eps
(*
(fma
(* (fma 0.16666666666666666 x (* 0.25 eps)) eps)
x
(* (fma (* 0.16666666666666666 eps) eps -1.0) eps))
x)))
double code(double x, double eps) {
return fma((fma((eps * eps), 0.041666666666666664, -0.5) * eps), eps, (fma((fma(0.16666666666666666, x, (0.25 * eps)) * eps), x, (fma((0.16666666666666666 * eps), eps, -1.0) * eps)) * x));
}
function code(x, eps) return fma(Float64(fma(Float64(eps * eps), 0.041666666666666664, -0.5) * eps), eps, Float64(fma(Float64(fma(0.16666666666666666, x, Float64(0.25 * eps)) * eps), x, Float64(fma(Float64(0.16666666666666666 * eps), eps, -1.0) * eps)) * x)) end
code[x_, eps_] := N[(N[(N[(N[(eps * eps), $MachinePrecision] * 0.041666666666666664 + -0.5), $MachinePrecision] * eps), $MachinePrecision] * eps + N[(N[(N[(N[(0.16666666666666666 * x + N[(0.25 * eps), $MachinePrecision]), $MachinePrecision] * eps), $MachinePrecision] * x + N[(N[(N[(0.16666666666666666 * eps), $MachinePrecision] * eps + -1.0), $MachinePrecision] * eps), $MachinePrecision]), $MachinePrecision] * x), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\mathsf{fma}\left(\mathsf{fma}\left(\varepsilon \cdot \varepsilon, 0.041666666666666664, -0.5\right) \cdot \varepsilon, \varepsilon, \mathsf{fma}\left(\mathsf{fma}\left(0.16666666666666666, x, 0.25 \cdot \varepsilon\right) \cdot \varepsilon, x, \mathsf{fma}\left(0.16666666666666666 \cdot \varepsilon, \varepsilon, -1\right) \cdot \varepsilon\right) \cdot x\right)
\end{array}
Initial program 53.7%
Taylor expanded in eps around 0
*-commutativeN/A
lower-*.f64N/A
Applied rewrites99.4%
Taylor expanded in x around 0
Applied rewrites99.0%
Taylor expanded in eps around 0
Applied rewrites99.0%
(FPCore (x eps) :precision binary64 (fma (* (fma (* eps eps) 0.041666666666666664 -0.5) eps) eps (* (* (fma (* 0.16666666666666666 x) x (fma (* eps x) 0.25 -1.0)) eps) x)))
double code(double x, double eps) {
return fma((fma((eps * eps), 0.041666666666666664, -0.5) * eps), eps, ((fma((0.16666666666666666 * x), x, fma((eps * x), 0.25, -1.0)) * eps) * x));
}
function code(x, eps) return fma(Float64(fma(Float64(eps * eps), 0.041666666666666664, -0.5) * eps), eps, Float64(Float64(fma(Float64(0.16666666666666666 * x), x, fma(Float64(eps * x), 0.25, -1.0)) * eps) * x)) end
code[x_, eps_] := N[(N[(N[(N[(eps * eps), $MachinePrecision] * 0.041666666666666664 + -0.5), $MachinePrecision] * eps), $MachinePrecision] * eps + N[(N[(N[(N[(0.16666666666666666 * x), $MachinePrecision] * x + N[(N[(eps * x), $MachinePrecision] * 0.25 + -1.0), $MachinePrecision]), $MachinePrecision] * eps), $MachinePrecision] * x), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\mathsf{fma}\left(\mathsf{fma}\left(\varepsilon \cdot \varepsilon, 0.041666666666666664, -0.5\right) \cdot \varepsilon, \varepsilon, \left(\mathsf{fma}\left(0.16666666666666666 \cdot x, x, \mathsf{fma}\left(\varepsilon \cdot x, 0.25, -1\right)\right) \cdot \varepsilon\right) \cdot x\right)
\end{array}
Initial program 53.7%
Taylor expanded in eps around 0
*-commutativeN/A
lower-*.f64N/A
Applied rewrites99.4%
Taylor expanded in x around 0
Applied rewrites99.0%
Taylor expanded in eps around 0
Applied rewrites98.9%
(FPCore (x eps) :precision binary64 (fma (* (fma (* eps eps) 0.041666666666666664 -0.5) eps) eps (* (* (fma (* x x) 0.16666666666666666 -1.0) eps) x)))
double code(double x, double eps) {
return fma((fma((eps * eps), 0.041666666666666664, -0.5) * eps), eps, ((fma((x * x), 0.16666666666666666, -1.0) * eps) * x));
}
function code(x, eps) return fma(Float64(fma(Float64(eps * eps), 0.041666666666666664, -0.5) * eps), eps, Float64(Float64(fma(Float64(x * x), 0.16666666666666666, -1.0) * eps) * x)) end
code[x_, eps_] := N[(N[(N[(N[(eps * eps), $MachinePrecision] * 0.041666666666666664 + -0.5), $MachinePrecision] * eps), $MachinePrecision] * eps + N[(N[(N[(N[(x * x), $MachinePrecision] * 0.16666666666666666 + -1.0), $MachinePrecision] * eps), $MachinePrecision] * x), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\mathsf{fma}\left(\mathsf{fma}\left(\varepsilon \cdot \varepsilon, 0.041666666666666664, -0.5\right) \cdot \varepsilon, \varepsilon, \left(\mathsf{fma}\left(x \cdot x, 0.16666666666666666, -1\right) \cdot \varepsilon\right) \cdot x\right)
\end{array}
Initial program 53.7%
Taylor expanded in eps around 0
*-commutativeN/A
lower-*.f64N/A
Applied rewrites99.4%
Taylor expanded in x around 0
Applied rewrites99.0%
Taylor expanded in eps around 0
Applied rewrites98.7%
(FPCore (x eps) :precision binary64 (fma (- eps) x (* (* (fma (* eps eps) 0.041666666666666664 -0.5) eps) eps)))
double code(double x, double eps) {
return fma(-eps, x, ((fma((eps * eps), 0.041666666666666664, -0.5) * eps) * eps));
}
function code(x, eps) return fma(Float64(-eps), x, Float64(Float64(fma(Float64(eps * eps), 0.041666666666666664, -0.5) * eps) * eps)) end
code[x_, eps_] := N[((-eps) * x + N[(N[(N[(N[(eps * eps), $MachinePrecision] * 0.041666666666666664 + -0.5), $MachinePrecision] * eps), $MachinePrecision] * eps), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\mathsf{fma}\left(-\varepsilon, x, \left(\mathsf{fma}\left(\varepsilon \cdot \varepsilon, 0.041666666666666664, -0.5\right) \cdot \varepsilon\right) \cdot \varepsilon\right)
\end{array}
Initial program 53.7%
Taylor expanded in eps around 0
*-commutativeN/A
lower-*.f64N/A
Applied rewrites99.4%
Taylor expanded in x around 0
Applied rewrites98.4%
Taylor expanded in eps around 0
Applied rewrites98.4%
(FPCore (x eps) :precision binary64 (* (fma (fma (* 0.25 x) x -0.5) eps (- x)) eps))
double code(double x, double eps) {
return fma(fma((0.25 * x), x, -0.5), eps, -x) * eps;
}
function code(x, eps) return Float64(fma(fma(Float64(0.25 * x), x, -0.5), eps, Float64(-x)) * eps) end
code[x_, eps_] := N[(N[(N[(N[(0.25 * x), $MachinePrecision] * x + -0.5), $MachinePrecision] * eps + (-x)), $MachinePrecision] * eps), $MachinePrecision]
\begin{array}{l}
\\
\mathsf{fma}\left(\mathsf{fma}\left(0.25 \cdot x, x, -0.5\right), \varepsilon, -x\right) \cdot \varepsilon
\end{array}
Initial program 53.7%
Taylor expanded in eps around 0
*-commutativeN/A
lower-*.f64N/A
Applied rewrites99.4%
Taylor expanded in x around 0
Applied rewrites98.3%
Taylor expanded in eps around 0
Applied rewrites98.3%
(FPCore (x eps) :precision binary64 (* (fma (* (* x x) eps) 0.16666666666666666 (- eps)) x))
double code(double x, double eps) {
return fma(((x * x) * eps), 0.16666666666666666, -eps) * x;
}
function code(x, eps) return Float64(fma(Float64(Float64(x * x) * eps), 0.16666666666666666, Float64(-eps)) * x) end
code[x_, eps_] := N[(N[(N[(N[(x * x), $MachinePrecision] * eps), $MachinePrecision] * 0.16666666666666666 + (-eps)), $MachinePrecision] * x), $MachinePrecision]
\begin{array}{l}
\\
\mathsf{fma}\left(\left(x \cdot x\right) \cdot \varepsilon, 0.16666666666666666, -\varepsilon\right) \cdot x
\end{array}
Initial program 53.7%
Taylor expanded in eps around 0
associate-*r*N/A
lower-*.f64N/A
mul-1-negN/A
lower-neg.f64N/A
lower-sin.f6479.3
Applied rewrites79.3%
Taylor expanded in x around 0
Applied rewrites79.1%
(FPCore (x eps) :precision binary64 (* (* (fma (* x x) 0.16666666666666666 -1.0) eps) x))
double code(double x, double eps) {
return (fma((x * x), 0.16666666666666666, -1.0) * eps) * x;
}
function code(x, eps) return Float64(Float64(fma(Float64(x * x), 0.16666666666666666, -1.0) * eps) * x) end
code[x_, eps_] := N[(N[(N[(N[(x * x), $MachinePrecision] * 0.16666666666666666 + -1.0), $MachinePrecision] * eps), $MachinePrecision] * x), $MachinePrecision]
\begin{array}{l}
\\
\left(\mathsf{fma}\left(x \cdot x, 0.16666666666666666, -1\right) \cdot \varepsilon\right) \cdot x
\end{array}
Initial program 53.7%
Taylor expanded in eps around 0
associate-*r*N/A
lower-*.f64N/A
mul-1-negN/A
lower-neg.f64N/A
lower-sin.f6479.3
Applied rewrites79.3%
Taylor expanded in x around 0
Applied rewrites79.1%
Taylor expanded in eps around 0
Applied rewrites79.1%
(FPCore (x eps) :precision binary64 (* (* (fma (* 0.16666666666666666 x) x -1.0) x) eps))
double code(double x, double eps) {
return (fma((0.16666666666666666 * x), x, -1.0) * x) * eps;
}
function code(x, eps) return Float64(Float64(fma(Float64(0.16666666666666666 * x), x, -1.0) * x) * eps) end
code[x_, eps_] := N[(N[(N[(N[(0.16666666666666666 * x), $MachinePrecision] * x + -1.0), $MachinePrecision] * x), $MachinePrecision] * eps), $MachinePrecision]
\begin{array}{l}
\\
\left(\mathsf{fma}\left(0.16666666666666666 \cdot x, x, -1\right) \cdot x\right) \cdot \varepsilon
\end{array}
Initial program 53.7%
Taylor expanded in eps around 0
associate-*r*N/A
lower-*.f64N/A
mul-1-negN/A
lower-neg.f64N/A
lower-sin.f6479.3
Applied rewrites79.3%
Taylor expanded in x around 0
Applied rewrites79.1%
Taylor expanded in x around inf
Applied rewrites79.1%
Taylor expanded in eps around 0
Applied rewrites79.1%
(FPCore (x eps) :precision binary64 (* (- x) eps))
double code(double x, double eps) {
return -x * eps;
}
real(8) function code(x, eps)
real(8), intent (in) :: x
real(8), intent (in) :: eps
code = -x * eps
end function
public static double code(double x, double eps) {
return -x * eps;
}
def code(x, eps): return -x * eps
function code(x, eps) return Float64(Float64(-x) * eps) end
function tmp = code(x, eps) tmp = -x * eps; end
code[x_, eps_] := N[((-x) * eps), $MachinePrecision]
\begin{array}{l}
\\
\left(-x\right) \cdot \varepsilon
\end{array}
Initial program 53.7%
Taylor expanded in eps around 0
associate-*r*N/A
lower-*.f64N/A
mul-1-negN/A
lower-neg.f64N/A
lower-sin.f6479.3
Applied rewrites79.3%
Taylor expanded in x around 0
Applied rewrites79.0%
(FPCore (x eps) :precision binary64 (- 1.0 1.0))
double code(double x, double eps) {
return 1.0 - 1.0;
}
real(8) function code(x, eps)
real(8), intent (in) :: x
real(8), intent (in) :: eps
code = 1.0d0 - 1.0d0
end function
public static double code(double x, double eps) {
return 1.0 - 1.0;
}
def code(x, eps): return 1.0 - 1.0
function code(x, eps) return Float64(1.0 - 1.0) end
function tmp = code(x, eps) tmp = 1.0 - 1.0; end
code[x_, eps_] := N[(1.0 - 1.0), $MachinePrecision]
\begin{array}{l}
\\
1 - 1
\end{array}
Initial program 53.7%
Taylor expanded in x around 0
lower--.f64N/A
lower-cos.f6452.6
Applied rewrites52.6%
Taylor expanded in eps around 0
Applied rewrites52.5%
(FPCore (x eps) :precision binary64 (pow (cbrt (* (* -2.0 (sin (* 0.5 (fma 2.0 x eps)))) (sin (* 0.5 eps)))) 3.0))
double code(double x, double eps) {
return pow(cbrt(((-2.0 * sin((0.5 * fma(2.0, x, eps)))) * sin((0.5 * eps)))), 3.0);
}
function code(x, eps) return cbrt(Float64(Float64(-2.0 * sin(Float64(0.5 * fma(2.0, x, eps)))) * sin(Float64(0.5 * eps)))) ^ 3.0 end
code[x_, eps_] := N[Power[N[Power[N[(N[(-2.0 * N[Sin[N[(0.5 * N[(2.0 * x + eps), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]), $MachinePrecision] * N[Sin[N[(0.5 * eps), $MachinePrecision]], $MachinePrecision]), $MachinePrecision], 1/3], $MachinePrecision], 3.0], $MachinePrecision]
\begin{array}{l}
\\
{\left(\sqrt[3]{\left(-2 \cdot \sin \left(0.5 \cdot \mathsf{fma}\left(2, x, \varepsilon\right)\right)\right) \cdot \sin \left(0.5 \cdot \varepsilon\right)}\right)}^{3}
\end{array}
herbie shell --seed 2024278
(FPCore (x eps)
:name "2cos (problem 3.3.5)"
:precision binary64
:pre (and (and (and (<= -10000.0 x) (<= x 10000.0)) (< (* 1e-16 (fabs x)) eps)) (< eps (fabs x)))
:alt
(! :herbie-platform default (pow (cbrt (* -2 (sin (* 1/2 (fma 2 x eps))) (sin (* 1/2 eps)))) 3))
(- (cos (+ x eps)) (cos x)))