
(FPCore (x eps) :precision binary64 (- (cos (+ x eps)) (cos x)))
double code(double x, double eps) {
return cos((x + eps)) - cos(x);
}
real(8) function code(x, eps)
real(8), intent (in) :: x
real(8), intent (in) :: eps
code = cos((x + eps)) - cos(x)
end function
public static double code(double x, double eps) {
return Math.cos((x + eps)) - Math.cos(x);
}
def code(x, eps): return math.cos((x + eps)) - math.cos(x)
function code(x, eps) return Float64(cos(Float64(x + eps)) - cos(x)) end
function tmp = code(x, eps) tmp = cos((x + eps)) - cos(x); end
code[x_, eps_] := N[(N[Cos[N[(x + eps), $MachinePrecision]], $MachinePrecision] - N[Cos[x], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\cos \left(x + \varepsilon\right) - \cos x
\end{array}
Sampling outcomes in binary64 precision:
Herbie found 18 alternatives:
| Alternative | Accuracy | Speedup |
|---|
(FPCore (x eps) :precision binary64 (- (cos (+ x eps)) (cos x)))
double code(double x, double eps) {
return cos((x + eps)) - cos(x);
}
real(8) function code(x, eps)
real(8), intent (in) :: x
real(8), intent (in) :: eps
code = cos((x + eps)) - cos(x)
end function
public static double code(double x, double eps) {
return Math.cos((x + eps)) - Math.cos(x);
}
def code(x, eps): return math.cos((x + eps)) - math.cos(x)
function code(x, eps) return Float64(cos(Float64(x + eps)) - cos(x)) end
function tmp = code(x, eps) tmp = cos((x + eps)) - cos(x); end
code[x_, eps_] := N[(N[Cos[N[(x + eps), $MachinePrecision]], $MachinePrecision] - N[Cos[x], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\cos \left(x + \varepsilon\right) - \cos x
\end{array}
(FPCore (x eps) :precision binary64 (fma (* (sin x) (+ -1.0 (* 0.16666666666666666 (* eps eps)))) eps (* (* eps (* eps -0.5)) (cos x))))
double code(double x, double eps) {
return fma((sin(x) * (-1.0 + (0.16666666666666666 * (eps * eps)))), eps, ((eps * (eps * -0.5)) * cos(x)));
}
function code(x, eps) return fma(Float64(sin(x) * Float64(-1.0 + Float64(0.16666666666666666 * Float64(eps * eps)))), eps, Float64(Float64(eps * Float64(eps * -0.5)) * cos(x))) end
code[x_, eps_] := N[(N[(N[Sin[x], $MachinePrecision] * N[(-1.0 + N[(0.16666666666666666 * N[(eps * eps), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] * eps + N[(N[(eps * N[(eps * -0.5), $MachinePrecision]), $MachinePrecision] * N[Cos[x], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\mathsf{fma}\left(\sin x \cdot \left(-1 + 0.16666666666666666 \cdot \left(\varepsilon \cdot \varepsilon\right)\right), \varepsilon, \left(\varepsilon \cdot \left(\varepsilon \cdot -0.5\right)\right) \cdot \cos x\right)
\end{array}
Initial program 55.0%
Taylor expanded in eps around 0
*-lowering-*.f64N/A
sub-negN/A
distribute-lft-inN/A
associate-+l+N/A
*-commutativeN/A
+-lowering-+.f64N/A
*-lowering-*.f64N/A
*-lowering-*.f64N/A
cos-lowering-cos.f64N/A
*-commutativeN/A
associate-*r*N/A
associate-*r*N/A
neg-mul-1N/A
Simplified99.8%
+-commutativeN/A
distribute-rgt-inN/A
fma-defineN/A
fma-lowering-fma.f64N/A
*-lowering-*.f64N/A
sin-lowering-sin.f64N/A
+-lowering-+.f64N/A
associate-*r*N/A
*-commutativeN/A
*-lowering-*.f64N/A
*-lowering-*.f64N/A
*-commutativeN/A
associate-*r*N/A
associate-*r*N/A
*-lowering-*.f64N/A
*-lowering-*.f64N/A
*-lowering-*.f64N/A
cos-lowering-cos.f64100.0%
Applied egg-rr100.0%
Final simplification100.0%
(FPCore (x eps) :precision binary64 (* eps (+ (* eps (* -0.5 (cos x))) (* (sin x) (+ -1.0 (* eps (* 0.16666666666666666 eps)))))))
double code(double x, double eps) {
return eps * ((eps * (-0.5 * cos(x))) + (sin(x) * (-1.0 + (eps * (0.16666666666666666 * eps)))));
}
real(8) function code(x, eps)
real(8), intent (in) :: x
real(8), intent (in) :: eps
code = eps * ((eps * ((-0.5d0) * cos(x))) + (sin(x) * ((-1.0d0) + (eps * (0.16666666666666666d0 * eps)))))
end function
public static double code(double x, double eps) {
return eps * ((eps * (-0.5 * Math.cos(x))) + (Math.sin(x) * (-1.0 + (eps * (0.16666666666666666 * eps)))));
}
def code(x, eps): return eps * ((eps * (-0.5 * math.cos(x))) + (math.sin(x) * (-1.0 + (eps * (0.16666666666666666 * eps)))))
function code(x, eps) return Float64(eps * Float64(Float64(eps * Float64(-0.5 * cos(x))) + Float64(sin(x) * Float64(-1.0 + Float64(eps * Float64(0.16666666666666666 * eps)))))) end
function tmp = code(x, eps) tmp = eps * ((eps * (-0.5 * cos(x))) + (sin(x) * (-1.0 + (eps * (0.16666666666666666 * eps))))); end
code[x_, eps_] := N[(eps * N[(N[(eps * N[(-0.5 * N[Cos[x], $MachinePrecision]), $MachinePrecision]), $MachinePrecision] + N[(N[Sin[x], $MachinePrecision] * N[(-1.0 + N[(eps * N[(0.16666666666666666 * eps), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\varepsilon \cdot \left(\varepsilon \cdot \left(-0.5 \cdot \cos x\right) + \sin x \cdot \left(-1 + \varepsilon \cdot \left(0.16666666666666666 \cdot \varepsilon\right)\right)\right)
\end{array}
Initial program 55.0%
Taylor expanded in eps around 0
*-lowering-*.f64N/A
sub-negN/A
distribute-lft-inN/A
associate-+l+N/A
*-commutativeN/A
+-lowering-+.f64N/A
*-lowering-*.f64N/A
*-lowering-*.f64N/A
cos-lowering-cos.f64N/A
*-commutativeN/A
associate-*r*N/A
associate-*r*N/A
neg-mul-1N/A
Simplified99.8%
Final simplification99.8%
(FPCore (x eps) :precision binary64 (* eps (- (* eps (* -0.5 (cos x))) (sin x))))
double code(double x, double eps) {
return eps * ((eps * (-0.5 * cos(x))) - sin(x));
}
real(8) function code(x, eps)
real(8), intent (in) :: x
real(8), intent (in) :: eps
code = eps * ((eps * ((-0.5d0) * cos(x))) - sin(x))
end function
public static double code(double x, double eps) {
return eps * ((eps * (-0.5 * Math.cos(x))) - Math.sin(x));
}
def code(x, eps): return eps * ((eps * (-0.5 * math.cos(x))) - math.sin(x))
function code(x, eps) return Float64(eps * Float64(Float64(eps * Float64(-0.5 * cos(x))) - sin(x))) end
function tmp = code(x, eps) tmp = eps * ((eps * (-0.5 * cos(x))) - sin(x)); end
code[x_, eps_] := N[(eps * N[(N[(eps * N[(-0.5 * N[Cos[x], $MachinePrecision]), $MachinePrecision]), $MachinePrecision] - N[Sin[x], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\varepsilon \cdot \left(\varepsilon \cdot \left(-0.5 \cdot \cos x\right) - \sin x\right)
\end{array}
Initial program 55.0%
Taylor expanded in eps around 0
*-lowering-*.f64N/A
*-commutativeN/A
associate-*r*N/A
*-commutativeN/A
--lowering--.f64N/A
*-lowering-*.f64N/A
*-lowering-*.f64N/A
cos-lowering-cos.f64N/A
sin-lowering-sin.f6499.6%
Simplified99.6%
(FPCore (x eps) :precision binary64 (* eps (+ (* eps -0.5) (* (sin x) (+ -1.0 (* eps (* 0.16666666666666666 eps)))))))
double code(double x, double eps) {
return eps * ((eps * -0.5) + (sin(x) * (-1.0 + (eps * (0.16666666666666666 * eps)))));
}
real(8) function code(x, eps)
real(8), intent (in) :: x
real(8), intent (in) :: eps
code = eps * ((eps * (-0.5d0)) + (sin(x) * ((-1.0d0) + (eps * (0.16666666666666666d0 * eps)))))
end function
public static double code(double x, double eps) {
return eps * ((eps * -0.5) + (Math.sin(x) * (-1.0 + (eps * (0.16666666666666666 * eps)))));
}
def code(x, eps): return eps * ((eps * -0.5) + (math.sin(x) * (-1.0 + (eps * (0.16666666666666666 * eps)))))
function code(x, eps) return Float64(eps * Float64(Float64(eps * -0.5) + Float64(sin(x) * Float64(-1.0 + Float64(eps * Float64(0.16666666666666666 * eps)))))) end
function tmp = code(x, eps) tmp = eps * ((eps * -0.5) + (sin(x) * (-1.0 + (eps * (0.16666666666666666 * eps))))); end
code[x_, eps_] := N[(eps * N[(N[(eps * -0.5), $MachinePrecision] + N[(N[Sin[x], $MachinePrecision] * N[(-1.0 + N[(eps * N[(0.16666666666666666 * eps), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\varepsilon \cdot \left(\varepsilon \cdot -0.5 + \sin x \cdot \left(-1 + \varepsilon \cdot \left(0.16666666666666666 \cdot \varepsilon\right)\right)\right)
\end{array}
Initial program 55.0%
Taylor expanded in eps around 0
*-lowering-*.f64N/A
sub-negN/A
distribute-lft-inN/A
associate-+l+N/A
*-commutativeN/A
+-lowering-+.f64N/A
*-lowering-*.f64N/A
*-lowering-*.f64N/A
cos-lowering-cos.f64N/A
*-commutativeN/A
associate-*r*N/A
associate-*r*N/A
neg-mul-1N/A
Simplified99.8%
Taylor expanded in x around 0
*-commutativeN/A
*-lowering-*.f6498.7%
Simplified98.7%
Final simplification98.7%
(FPCore (x eps) :precision binary64 (* eps (- (* eps -0.5) (sin x))))
double code(double x, double eps) {
return eps * ((eps * -0.5) - sin(x));
}
real(8) function code(x, eps)
real(8), intent (in) :: x
real(8), intent (in) :: eps
code = eps * ((eps * (-0.5d0)) - sin(x))
end function
public static double code(double x, double eps) {
return eps * ((eps * -0.5) - Math.sin(x));
}
def code(x, eps): return eps * ((eps * -0.5) - math.sin(x))
function code(x, eps) return Float64(eps * Float64(Float64(eps * -0.5) - sin(x))) end
function tmp = code(x, eps) tmp = eps * ((eps * -0.5) - sin(x)); end
code[x_, eps_] := N[(eps * N[(N[(eps * -0.5), $MachinePrecision] - N[Sin[x], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\varepsilon \cdot \left(\varepsilon \cdot -0.5 - \sin x\right)
\end{array}
Initial program 55.0%
Taylor expanded in eps around 0
*-lowering-*.f64N/A
sub-negN/A
distribute-lft-inN/A
associate-+l+N/A
*-commutativeN/A
+-lowering-+.f64N/A
*-lowering-*.f64N/A
*-lowering-*.f64N/A
cos-lowering-cos.f64N/A
*-commutativeN/A
associate-*r*N/A
associate-*r*N/A
neg-mul-1N/A
Simplified99.8%
Taylor expanded in x around 0
*-commutativeN/A
*-lowering-*.f6498.7%
Simplified98.7%
Taylor expanded in eps around 0
*-lowering-*.f64N/A
+-commutativeN/A
mul-1-negN/A
unsub-negN/A
--lowering--.f64N/A
*-commutativeN/A
*-lowering-*.f64N/A
sin-lowering-sin.f6498.7%
Simplified98.7%
(FPCore (x eps)
:precision binary64
(*
eps
(+
(* eps (+ -0.5 (* x (* x 0.25))))
(*
(+ -1.0 (* eps (* 0.16666666666666666 eps)))
(*
x
(+
1.0
(*
x
(*
x
(+
-0.16666666666666666
(*
(* x x)
(+ 0.008333333333333333 (* (* x x) -0.0001984126984126984))))))))))))
double code(double x, double eps) {
return eps * ((eps * (-0.5 + (x * (x * 0.25)))) + ((-1.0 + (eps * (0.16666666666666666 * eps))) * (x * (1.0 + (x * (x * (-0.16666666666666666 + ((x * x) * (0.008333333333333333 + ((x * x) * -0.0001984126984126984))))))))));
}
real(8) function code(x, eps)
real(8), intent (in) :: x
real(8), intent (in) :: eps
code = eps * ((eps * ((-0.5d0) + (x * (x * 0.25d0)))) + (((-1.0d0) + (eps * (0.16666666666666666d0 * eps))) * (x * (1.0d0 + (x * (x * ((-0.16666666666666666d0) + ((x * x) * (0.008333333333333333d0 + ((x * x) * (-0.0001984126984126984d0)))))))))))
end function
public static double code(double x, double eps) {
return eps * ((eps * (-0.5 + (x * (x * 0.25)))) + ((-1.0 + (eps * (0.16666666666666666 * eps))) * (x * (1.0 + (x * (x * (-0.16666666666666666 + ((x * x) * (0.008333333333333333 + ((x * x) * -0.0001984126984126984))))))))));
}
def code(x, eps): return eps * ((eps * (-0.5 + (x * (x * 0.25)))) + ((-1.0 + (eps * (0.16666666666666666 * eps))) * (x * (1.0 + (x * (x * (-0.16666666666666666 + ((x * x) * (0.008333333333333333 + ((x * x) * -0.0001984126984126984))))))))))
function code(x, eps) return Float64(eps * Float64(Float64(eps * Float64(-0.5 + Float64(x * Float64(x * 0.25)))) + Float64(Float64(-1.0 + Float64(eps * Float64(0.16666666666666666 * eps))) * Float64(x * Float64(1.0 + Float64(x * Float64(x * Float64(-0.16666666666666666 + Float64(Float64(x * x) * Float64(0.008333333333333333 + Float64(Float64(x * x) * -0.0001984126984126984))))))))))) end
function tmp = code(x, eps) tmp = eps * ((eps * (-0.5 + (x * (x * 0.25)))) + ((-1.0 + (eps * (0.16666666666666666 * eps))) * (x * (1.0 + (x * (x * (-0.16666666666666666 + ((x * x) * (0.008333333333333333 + ((x * x) * -0.0001984126984126984)))))))))); end
code[x_, eps_] := N[(eps * N[(N[(eps * N[(-0.5 + N[(x * N[(x * 0.25), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] + N[(N[(-1.0 + N[(eps * N[(0.16666666666666666 * eps), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] * N[(x * N[(1.0 + N[(x * N[(x * N[(-0.16666666666666666 + N[(N[(x * x), $MachinePrecision] * N[(0.008333333333333333 + N[(N[(x * x), $MachinePrecision] * -0.0001984126984126984), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\varepsilon \cdot \left(\varepsilon \cdot \left(-0.5 + x \cdot \left(x \cdot 0.25\right)\right) + \left(-1 + \varepsilon \cdot \left(0.16666666666666666 \cdot \varepsilon\right)\right) \cdot \left(x \cdot \left(1 + x \cdot \left(x \cdot \left(-0.16666666666666666 + \left(x \cdot x\right) \cdot \left(0.008333333333333333 + \left(x \cdot x\right) \cdot -0.0001984126984126984\right)\right)\right)\right)\right)\right)
\end{array}
Initial program 55.0%
Taylor expanded in eps around 0
*-lowering-*.f64N/A
sub-negN/A
distribute-lft-inN/A
associate-+l+N/A
*-commutativeN/A
+-lowering-+.f64N/A
*-lowering-*.f64N/A
*-lowering-*.f64N/A
cos-lowering-cos.f64N/A
*-commutativeN/A
associate-*r*N/A
associate-*r*N/A
neg-mul-1N/A
Simplified99.8%
Taylor expanded in x around 0
*-commutativeN/A
associate-*r*N/A
distribute-rgt-outN/A
+-commutativeN/A
metadata-evalN/A
sub-negN/A
*-lowering-*.f64N/A
sub-negN/A
metadata-evalN/A
+-commutativeN/A
+-lowering-+.f64N/A
*-commutativeN/A
unpow2N/A
associate-*l*N/A
*-lowering-*.f64N/A
*-lowering-*.f6498.4%
Simplified98.4%
Taylor expanded in x around 0
*-lowering-*.f64N/A
+-lowering-+.f64N/A
unpow2N/A
associate-*l*N/A
*-lowering-*.f64N/A
*-lowering-*.f64N/A
sub-negN/A
metadata-evalN/A
+-commutativeN/A
+-lowering-+.f64N/A
*-lowering-*.f64N/A
unpow2N/A
*-lowering-*.f64N/A
+-lowering-+.f64N/A
*-commutativeN/A
*-lowering-*.f64N/A
unpow2N/A
*-lowering-*.f6497.7%
Simplified97.7%
Final simplification97.7%
(FPCore (x eps)
:precision binary64
(*
eps
(+
(* eps (+ -0.5 (* x (* x 0.25))))
(*
(+ -1.0 (* eps (* 0.16666666666666666 eps)))
(*
x
(+
1.0
(*
x
(* x (+ -0.16666666666666666 (* (* x x) 0.008333333333333333))))))))))
double code(double x, double eps) {
return eps * ((eps * (-0.5 + (x * (x * 0.25)))) + ((-1.0 + (eps * (0.16666666666666666 * eps))) * (x * (1.0 + (x * (x * (-0.16666666666666666 + ((x * x) * 0.008333333333333333))))))));
}
real(8) function code(x, eps)
real(8), intent (in) :: x
real(8), intent (in) :: eps
code = eps * ((eps * ((-0.5d0) + (x * (x * 0.25d0)))) + (((-1.0d0) + (eps * (0.16666666666666666d0 * eps))) * (x * (1.0d0 + (x * (x * ((-0.16666666666666666d0) + ((x * x) * 0.008333333333333333d0))))))))
end function
public static double code(double x, double eps) {
return eps * ((eps * (-0.5 + (x * (x * 0.25)))) + ((-1.0 + (eps * (0.16666666666666666 * eps))) * (x * (1.0 + (x * (x * (-0.16666666666666666 + ((x * x) * 0.008333333333333333))))))));
}
def code(x, eps): return eps * ((eps * (-0.5 + (x * (x * 0.25)))) + ((-1.0 + (eps * (0.16666666666666666 * eps))) * (x * (1.0 + (x * (x * (-0.16666666666666666 + ((x * x) * 0.008333333333333333))))))))
function code(x, eps) return Float64(eps * Float64(Float64(eps * Float64(-0.5 + Float64(x * Float64(x * 0.25)))) + Float64(Float64(-1.0 + Float64(eps * Float64(0.16666666666666666 * eps))) * Float64(x * Float64(1.0 + Float64(x * Float64(x * Float64(-0.16666666666666666 + Float64(Float64(x * x) * 0.008333333333333333))))))))) end
function tmp = code(x, eps) tmp = eps * ((eps * (-0.5 + (x * (x * 0.25)))) + ((-1.0 + (eps * (0.16666666666666666 * eps))) * (x * (1.0 + (x * (x * (-0.16666666666666666 + ((x * x) * 0.008333333333333333)))))))); end
code[x_, eps_] := N[(eps * N[(N[(eps * N[(-0.5 + N[(x * N[(x * 0.25), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] + N[(N[(-1.0 + N[(eps * N[(0.16666666666666666 * eps), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] * N[(x * N[(1.0 + N[(x * N[(x * N[(-0.16666666666666666 + N[(N[(x * x), $MachinePrecision] * 0.008333333333333333), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\varepsilon \cdot \left(\varepsilon \cdot \left(-0.5 + x \cdot \left(x \cdot 0.25\right)\right) + \left(-1 + \varepsilon \cdot \left(0.16666666666666666 \cdot \varepsilon\right)\right) \cdot \left(x \cdot \left(1 + x \cdot \left(x \cdot \left(-0.16666666666666666 + \left(x \cdot x\right) \cdot 0.008333333333333333\right)\right)\right)\right)\right)
\end{array}
Initial program 55.0%
Taylor expanded in eps around 0
*-lowering-*.f64N/A
sub-negN/A
distribute-lft-inN/A
associate-+l+N/A
*-commutativeN/A
+-lowering-+.f64N/A
*-lowering-*.f64N/A
*-lowering-*.f64N/A
cos-lowering-cos.f64N/A
*-commutativeN/A
associate-*r*N/A
associate-*r*N/A
neg-mul-1N/A
Simplified99.8%
Taylor expanded in x around 0
*-commutativeN/A
associate-*r*N/A
distribute-rgt-outN/A
+-commutativeN/A
metadata-evalN/A
sub-negN/A
*-lowering-*.f64N/A
sub-negN/A
metadata-evalN/A
+-commutativeN/A
+-lowering-+.f64N/A
*-commutativeN/A
unpow2N/A
associate-*l*N/A
*-lowering-*.f64N/A
*-lowering-*.f6498.4%
Simplified98.4%
Taylor expanded in x around 0
*-lowering-*.f64N/A
+-lowering-+.f64N/A
unpow2N/A
associate-*l*N/A
*-lowering-*.f64N/A
*-lowering-*.f64N/A
sub-negN/A
metadata-evalN/A
+-commutativeN/A
+-lowering-+.f64N/A
*-commutativeN/A
*-lowering-*.f64N/A
unpow2N/A
*-lowering-*.f6497.4%
Simplified97.4%
Final simplification97.4%
(FPCore (x eps)
:precision binary64
(*
eps
(+
(* eps -0.5)
(*
x
(+
-1.0
(+
(* eps (+ (* 0.16666666666666666 eps) (* x 0.25)))
(*
(+ -1.0 (* eps (* 0.16666666666666666 eps)))
(* x (* x -0.16666666666666666)))))))))
double code(double x, double eps) {
return eps * ((eps * -0.5) + (x * (-1.0 + ((eps * ((0.16666666666666666 * eps) + (x * 0.25))) + ((-1.0 + (eps * (0.16666666666666666 * eps))) * (x * (x * -0.16666666666666666)))))));
}
real(8) function code(x, eps)
real(8), intent (in) :: x
real(8), intent (in) :: eps
code = eps * ((eps * (-0.5d0)) + (x * ((-1.0d0) + ((eps * ((0.16666666666666666d0 * eps) + (x * 0.25d0))) + (((-1.0d0) + (eps * (0.16666666666666666d0 * eps))) * (x * (x * (-0.16666666666666666d0))))))))
end function
public static double code(double x, double eps) {
return eps * ((eps * -0.5) + (x * (-1.0 + ((eps * ((0.16666666666666666 * eps) + (x * 0.25))) + ((-1.0 + (eps * (0.16666666666666666 * eps))) * (x * (x * -0.16666666666666666)))))));
}
def code(x, eps): return eps * ((eps * -0.5) + (x * (-1.0 + ((eps * ((0.16666666666666666 * eps) + (x * 0.25))) + ((-1.0 + (eps * (0.16666666666666666 * eps))) * (x * (x * -0.16666666666666666)))))))
function code(x, eps) return Float64(eps * Float64(Float64(eps * -0.5) + Float64(x * Float64(-1.0 + Float64(Float64(eps * Float64(Float64(0.16666666666666666 * eps) + Float64(x * 0.25))) + Float64(Float64(-1.0 + Float64(eps * Float64(0.16666666666666666 * eps))) * Float64(x * Float64(x * -0.16666666666666666)))))))) end
function tmp = code(x, eps) tmp = eps * ((eps * -0.5) + (x * (-1.0 + ((eps * ((0.16666666666666666 * eps) + (x * 0.25))) + ((-1.0 + (eps * (0.16666666666666666 * eps))) * (x * (x * -0.16666666666666666))))))); end
code[x_, eps_] := N[(eps * N[(N[(eps * -0.5), $MachinePrecision] + N[(x * N[(-1.0 + N[(N[(eps * N[(N[(0.16666666666666666 * eps), $MachinePrecision] + N[(x * 0.25), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] + N[(N[(-1.0 + N[(eps * N[(0.16666666666666666 * eps), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] * N[(x * N[(x * -0.16666666666666666), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\varepsilon \cdot \left(\varepsilon \cdot -0.5 + x \cdot \left(-1 + \left(\varepsilon \cdot \left(0.16666666666666666 \cdot \varepsilon + x \cdot 0.25\right) + \left(-1 + \varepsilon \cdot \left(0.16666666666666666 \cdot \varepsilon\right)\right) \cdot \left(x \cdot \left(x \cdot -0.16666666666666666\right)\right)\right)\right)\right)
\end{array}
Initial program 55.0%
Taylor expanded in eps around 0
*-lowering-*.f64N/A
sub-negN/A
distribute-lft-inN/A
associate-+l+N/A
*-commutativeN/A
+-lowering-+.f64N/A
*-lowering-*.f64N/A
*-lowering-*.f64N/A
cos-lowering-cos.f64N/A
*-commutativeN/A
associate-*r*N/A
associate-*r*N/A
neg-mul-1N/A
Simplified99.8%
Taylor expanded in x around 0
Simplified97.3%
Final simplification97.3%
(FPCore (x eps)
:precision binary64
(*
eps
(+
(* eps (+ -0.5 (* x (* x 0.25))))
(*
(+ -1.0 (* eps (* 0.16666666666666666 eps)))
(* x (+ 1.0 (* -0.16666666666666666 (* x x))))))))
double code(double x, double eps) {
return eps * ((eps * (-0.5 + (x * (x * 0.25)))) + ((-1.0 + (eps * (0.16666666666666666 * eps))) * (x * (1.0 + (-0.16666666666666666 * (x * x))))));
}
real(8) function code(x, eps)
real(8), intent (in) :: x
real(8), intent (in) :: eps
code = eps * ((eps * ((-0.5d0) + (x * (x * 0.25d0)))) + (((-1.0d0) + (eps * (0.16666666666666666d0 * eps))) * (x * (1.0d0 + ((-0.16666666666666666d0) * (x * x))))))
end function
public static double code(double x, double eps) {
return eps * ((eps * (-0.5 + (x * (x * 0.25)))) + ((-1.0 + (eps * (0.16666666666666666 * eps))) * (x * (1.0 + (-0.16666666666666666 * (x * x))))));
}
def code(x, eps): return eps * ((eps * (-0.5 + (x * (x * 0.25)))) + ((-1.0 + (eps * (0.16666666666666666 * eps))) * (x * (1.0 + (-0.16666666666666666 * (x * x))))))
function code(x, eps) return Float64(eps * Float64(Float64(eps * Float64(-0.5 + Float64(x * Float64(x * 0.25)))) + Float64(Float64(-1.0 + Float64(eps * Float64(0.16666666666666666 * eps))) * Float64(x * Float64(1.0 + Float64(-0.16666666666666666 * Float64(x * x))))))) end
function tmp = code(x, eps) tmp = eps * ((eps * (-0.5 + (x * (x * 0.25)))) + ((-1.0 + (eps * (0.16666666666666666 * eps))) * (x * (1.0 + (-0.16666666666666666 * (x * x)))))); end
code[x_, eps_] := N[(eps * N[(N[(eps * N[(-0.5 + N[(x * N[(x * 0.25), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] + N[(N[(-1.0 + N[(eps * N[(0.16666666666666666 * eps), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] * N[(x * N[(1.0 + N[(-0.16666666666666666 * N[(x * x), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\varepsilon \cdot \left(\varepsilon \cdot \left(-0.5 + x \cdot \left(x \cdot 0.25\right)\right) + \left(-1 + \varepsilon \cdot \left(0.16666666666666666 \cdot \varepsilon\right)\right) \cdot \left(x \cdot \left(1 + -0.16666666666666666 \cdot \left(x \cdot x\right)\right)\right)\right)
\end{array}
Initial program 55.0%
Taylor expanded in eps around 0
*-lowering-*.f64N/A
sub-negN/A
distribute-lft-inN/A
associate-+l+N/A
*-commutativeN/A
+-lowering-+.f64N/A
*-lowering-*.f64N/A
*-lowering-*.f64N/A
cos-lowering-cos.f64N/A
*-commutativeN/A
associate-*r*N/A
associate-*r*N/A
neg-mul-1N/A
Simplified99.8%
Taylor expanded in x around 0
*-commutativeN/A
associate-*r*N/A
distribute-rgt-outN/A
+-commutativeN/A
metadata-evalN/A
sub-negN/A
*-lowering-*.f64N/A
sub-negN/A
metadata-evalN/A
+-commutativeN/A
+-lowering-+.f64N/A
*-commutativeN/A
unpow2N/A
associate-*l*N/A
*-lowering-*.f64N/A
*-lowering-*.f6498.4%
Simplified98.4%
Taylor expanded in x around 0
*-lowering-*.f64N/A
+-lowering-+.f64N/A
*-lowering-*.f64N/A
unpow2N/A
*-lowering-*.f6497.3%
Simplified97.3%
Final simplification97.3%
(FPCore (x eps)
:precision binary64
(*
eps
(+
(* eps -0.5)
(*
x
(*
(+ -1.0 (* 0.16666666666666666 (* eps eps)))
(+ 1.0 (* -0.16666666666666666 (* x x))))))))
double code(double x, double eps) {
return eps * ((eps * -0.5) + (x * ((-1.0 + (0.16666666666666666 * (eps * eps))) * (1.0 + (-0.16666666666666666 * (x * x))))));
}
real(8) function code(x, eps)
real(8), intent (in) :: x
real(8), intent (in) :: eps
code = eps * ((eps * (-0.5d0)) + (x * (((-1.0d0) + (0.16666666666666666d0 * (eps * eps))) * (1.0d0 + ((-0.16666666666666666d0) * (x * x))))))
end function
public static double code(double x, double eps) {
return eps * ((eps * -0.5) + (x * ((-1.0 + (0.16666666666666666 * (eps * eps))) * (1.0 + (-0.16666666666666666 * (x * x))))));
}
def code(x, eps): return eps * ((eps * -0.5) + (x * ((-1.0 + (0.16666666666666666 * (eps * eps))) * (1.0 + (-0.16666666666666666 * (x * x))))))
function code(x, eps) return Float64(eps * Float64(Float64(eps * -0.5) + Float64(x * Float64(Float64(-1.0 + Float64(0.16666666666666666 * Float64(eps * eps))) * Float64(1.0 + Float64(-0.16666666666666666 * Float64(x * x))))))) end
function tmp = code(x, eps) tmp = eps * ((eps * -0.5) + (x * ((-1.0 + (0.16666666666666666 * (eps * eps))) * (1.0 + (-0.16666666666666666 * (x * x)))))); end
code[x_, eps_] := N[(eps * N[(N[(eps * -0.5), $MachinePrecision] + N[(x * N[(N[(-1.0 + N[(0.16666666666666666 * N[(eps * eps), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] * N[(1.0 + N[(-0.16666666666666666 * N[(x * x), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\varepsilon \cdot \left(\varepsilon \cdot -0.5 + x \cdot \left(\left(-1 + 0.16666666666666666 \cdot \left(\varepsilon \cdot \varepsilon\right)\right) \cdot \left(1 + -0.16666666666666666 \cdot \left(x \cdot x\right)\right)\right)\right)
\end{array}
Initial program 55.0%
Taylor expanded in eps around 0
*-lowering-*.f64N/A
sub-negN/A
distribute-lft-inN/A
associate-+l+N/A
*-commutativeN/A
+-lowering-+.f64N/A
*-lowering-*.f64N/A
*-lowering-*.f64N/A
cos-lowering-cos.f64N/A
*-commutativeN/A
associate-*r*N/A
associate-*r*N/A
neg-mul-1N/A
Simplified99.8%
Taylor expanded in x around 0
*-commutativeN/A
*-lowering-*.f6498.7%
Simplified98.7%
Taylor expanded in x around 0
+-lowering-+.f64N/A
*-commutativeN/A
*-lowering-*.f64N/A
*-lowering-*.f64N/A
associate--l+N/A
associate-*r*N/A
distribute-lft1-inN/A
*-lowering-*.f64N/A
+-lowering-+.f64N/A
*-lowering-*.f64N/A
unpow2N/A
*-lowering-*.f64N/A
sub-negN/A
metadata-evalN/A
+-commutativeN/A
+-lowering-+.f64N/A
*-lowering-*.f64N/A
unpow2N/A
*-lowering-*.f6497.3%
Simplified97.3%
Final simplification97.3%
(FPCore (x eps) :precision binary64 (* x (+ (* eps (+ -1.0 (* 0.16666666666666666 (* eps eps)))) (/ (* (* eps eps) -0.5) x))))
double code(double x, double eps) {
return x * ((eps * (-1.0 + (0.16666666666666666 * (eps * eps)))) + (((eps * eps) * -0.5) / x));
}
real(8) function code(x, eps)
real(8), intent (in) :: x
real(8), intent (in) :: eps
code = x * ((eps * ((-1.0d0) + (0.16666666666666666d0 * (eps * eps)))) + (((eps * eps) * (-0.5d0)) / x))
end function
public static double code(double x, double eps) {
return x * ((eps * (-1.0 + (0.16666666666666666 * (eps * eps)))) + (((eps * eps) * -0.5) / x));
}
def code(x, eps): return x * ((eps * (-1.0 + (0.16666666666666666 * (eps * eps)))) + (((eps * eps) * -0.5) / x))
function code(x, eps) return Float64(x * Float64(Float64(eps * Float64(-1.0 + Float64(0.16666666666666666 * Float64(eps * eps)))) + Float64(Float64(Float64(eps * eps) * -0.5) / x))) end
function tmp = code(x, eps) tmp = x * ((eps * (-1.0 + (0.16666666666666666 * (eps * eps)))) + (((eps * eps) * -0.5) / x)); end
code[x_, eps_] := N[(x * N[(N[(eps * N[(-1.0 + N[(0.16666666666666666 * N[(eps * eps), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] + N[(N[(N[(eps * eps), $MachinePrecision] * -0.5), $MachinePrecision] / x), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
x \cdot \left(\varepsilon \cdot \left(-1 + 0.16666666666666666 \cdot \left(\varepsilon \cdot \varepsilon\right)\right) + \frac{\left(\varepsilon \cdot \varepsilon\right) \cdot -0.5}{x}\right)
\end{array}
Initial program 55.0%
Taylor expanded in x around 0
+-commutativeN/A
associate--l+N/A
+-commutativeN/A
mul-1-negN/A
sub-negN/A
--lowering--.f64N/A
sub-negN/A
metadata-evalN/A
+-lowering-+.f64N/A
cos-lowering-cos.f64N/A
*-lowering-*.f64N/A
sin-lowering-sin.f6478.4%
Simplified78.4%
Taylor expanded in eps around 0
*-lowering-*.f64N/A
--lowering--.f64N/A
*-lowering-*.f64N/A
sub-negN/A
metadata-evalN/A
+-commutativeN/A
+-lowering-+.f64N/A
*-lowering-*.f64N/A
*-lowering-*.f6496.8%
Simplified96.8%
Taylor expanded in x around inf
*-lowering-*.f64N/A
+-commutativeN/A
+-lowering-+.f64N/A
*-lowering-*.f64N/A
sub-negN/A
metadata-evalN/A
+-lowering-+.f64N/A
*-commutativeN/A
*-lowering-*.f64N/A
unpow2N/A
*-lowering-*.f64N/A
associate-*r/N/A
/-lowering-/.f64N/A
*-lowering-*.f64N/A
unpow2N/A
*-lowering-*.f6496.8%
Simplified96.8%
Final simplification96.8%
(FPCore (x eps) :precision binary64 (* eps (+ (* eps -0.5) (* x (+ -1.0 (* eps (+ (* 0.16666666666666666 eps) (* x 0.25))))))))
double code(double x, double eps) {
return eps * ((eps * -0.5) + (x * (-1.0 + (eps * ((0.16666666666666666 * eps) + (x * 0.25))))));
}
real(8) function code(x, eps)
real(8), intent (in) :: x
real(8), intent (in) :: eps
code = eps * ((eps * (-0.5d0)) + (x * ((-1.0d0) + (eps * ((0.16666666666666666d0 * eps) + (x * 0.25d0))))))
end function
public static double code(double x, double eps) {
return eps * ((eps * -0.5) + (x * (-1.0 + (eps * ((0.16666666666666666 * eps) + (x * 0.25))))));
}
def code(x, eps): return eps * ((eps * -0.5) + (x * (-1.0 + (eps * ((0.16666666666666666 * eps) + (x * 0.25))))))
function code(x, eps) return Float64(eps * Float64(Float64(eps * -0.5) + Float64(x * Float64(-1.0 + Float64(eps * Float64(Float64(0.16666666666666666 * eps) + Float64(x * 0.25))))))) end
function tmp = code(x, eps) tmp = eps * ((eps * -0.5) + (x * (-1.0 + (eps * ((0.16666666666666666 * eps) + (x * 0.25)))))); end
code[x_, eps_] := N[(eps * N[(N[(eps * -0.5), $MachinePrecision] + N[(x * N[(-1.0 + N[(eps * N[(N[(0.16666666666666666 * eps), $MachinePrecision] + N[(x * 0.25), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\varepsilon \cdot \left(\varepsilon \cdot -0.5 + x \cdot \left(-1 + \varepsilon \cdot \left(0.16666666666666666 \cdot \varepsilon + x \cdot 0.25\right)\right)\right)
\end{array}
Initial program 55.0%
Taylor expanded in eps around 0
*-lowering-*.f64N/A
sub-negN/A
distribute-lft-inN/A
associate-+l+N/A
*-commutativeN/A
+-lowering-+.f64N/A
*-lowering-*.f64N/A
*-lowering-*.f64N/A
cos-lowering-cos.f64N/A
*-commutativeN/A
associate-*r*N/A
associate-*r*N/A
neg-mul-1N/A
Simplified99.8%
Taylor expanded in x around 0
+-lowering-+.f64N/A
*-commutativeN/A
*-lowering-*.f64N/A
*-lowering-*.f64N/A
sub-negN/A
metadata-evalN/A
+-commutativeN/A
+-lowering-+.f64N/A
unpow2N/A
associate-*r*N/A
*-commutativeN/A
associate-*r*N/A
distribute-rgt-outN/A
*-lowering-*.f64N/A
+-lowering-+.f64N/A
*-commutativeN/A
*-lowering-*.f64N/A
*-lowering-*.f6496.8%
Simplified96.8%
Final simplification96.8%
(FPCore (x eps)
:precision binary64
(*
eps
(-
(*
eps
(+
-0.5
(* eps (+ (* eps 0.041666666666666664) (* x 0.16666666666666666)))))
x)))
double code(double x, double eps) {
return eps * ((eps * (-0.5 + (eps * ((eps * 0.041666666666666664) + (x * 0.16666666666666666))))) - x);
}
real(8) function code(x, eps)
real(8), intent (in) :: x
real(8), intent (in) :: eps
code = eps * ((eps * ((-0.5d0) + (eps * ((eps * 0.041666666666666664d0) + (x * 0.16666666666666666d0))))) - x)
end function
public static double code(double x, double eps) {
return eps * ((eps * (-0.5 + (eps * ((eps * 0.041666666666666664) + (x * 0.16666666666666666))))) - x);
}
def code(x, eps): return eps * ((eps * (-0.5 + (eps * ((eps * 0.041666666666666664) + (x * 0.16666666666666666))))) - x)
function code(x, eps) return Float64(eps * Float64(Float64(eps * Float64(-0.5 + Float64(eps * Float64(Float64(eps * 0.041666666666666664) + Float64(x * 0.16666666666666666))))) - x)) end
function tmp = code(x, eps) tmp = eps * ((eps * (-0.5 + (eps * ((eps * 0.041666666666666664) + (x * 0.16666666666666666))))) - x); end
code[x_, eps_] := N[(eps * N[(N[(eps * N[(-0.5 + N[(eps * N[(N[(eps * 0.041666666666666664), $MachinePrecision] + N[(x * 0.16666666666666666), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] - x), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\varepsilon \cdot \left(\varepsilon \cdot \left(-0.5 + \varepsilon \cdot \left(\varepsilon \cdot 0.041666666666666664 + x \cdot 0.16666666666666666\right)\right) - x\right)
\end{array}
Initial program 55.0%
Taylor expanded in x around 0
+-commutativeN/A
associate--l+N/A
+-commutativeN/A
mul-1-negN/A
sub-negN/A
--lowering--.f64N/A
sub-negN/A
metadata-evalN/A
+-lowering-+.f64N/A
cos-lowering-cos.f64N/A
*-lowering-*.f64N/A
sin-lowering-sin.f6478.4%
Simplified78.4%
Taylor expanded in eps around 0
*-lowering-*.f64N/A
--lowering--.f64N/A
*-lowering-*.f64N/A
sub-negN/A
metadata-evalN/A
+-commutativeN/A
+-lowering-+.f64N/A
*-lowering-*.f64N/A
cancel-sign-sub-invN/A
+-lowering-+.f64N/A
*-commutativeN/A
*-lowering-*.f64N/A
metadata-evalN/A
*-commutativeN/A
*-lowering-*.f6496.8%
Simplified96.8%
(FPCore (x eps) :precision binary64 (* eps (- (* eps (+ -0.5 (* 0.16666666666666666 (* x eps)))) x)))
double code(double x, double eps) {
return eps * ((eps * (-0.5 + (0.16666666666666666 * (x * eps)))) - x);
}
real(8) function code(x, eps)
real(8), intent (in) :: x
real(8), intent (in) :: eps
code = eps * ((eps * ((-0.5d0) + (0.16666666666666666d0 * (x * eps)))) - x)
end function
public static double code(double x, double eps) {
return eps * ((eps * (-0.5 + (0.16666666666666666 * (x * eps)))) - x);
}
def code(x, eps): return eps * ((eps * (-0.5 + (0.16666666666666666 * (x * eps)))) - x)
function code(x, eps) return Float64(eps * Float64(Float64(eps * Float64(-0.5 + Float64(0.16666666666666666 * Float64(x * eps)))) - x)) end
function tmp = code(x, eps) tmp = eps * ((eps * (-0.5 + (0.16666666666666666 * (x * eps)))) - x); end
code[x_, eps_] := N[(eps * N[(N[(eps * N[(-0.5 + N[(0.16666666666666666 * N[(x * eps), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] - x), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\varepsilon \cdot \left(\varepsilon \cdot \left(-0.5 + 0.16666666666666666 \cdot \left(x \cdot \varepsilon\right)\right) - x\right)
\end{array}
Initial program 55.0%
Taylor expanded in x around 0
+-commutativeN/A
associate--l+N/A
+-commutativeN/A
mul-1-negN/A
sub-negN/A
--lowering--.f64N/A
sub-negN/A
metadata-evalN/A
+-lowering-+.f64N/A
cos-lowering-cos.f64N/A
*-lowering-*.f64N/A
sin-lowering-sin.f6478.4%
Simplified78.4%
Taylor expanded in eps around 0
*-lowering-*.f64N/A
--lowering--.f64N/A
*-lowering-*.f64N/A
sub-negN/A
metadata-evalN/A
+-commutativeN/A
+-lowering-+.f64N/A
*-lowering-*.f64N/A
*-lowering-*.f6496.8%
Simplified96.8%
Final simplification96.8%
(FPCore (x eps) :precision binary64 (* eps (- (* eps -0.5) x)))
double code(double x, double eps) {
return eps * ((eps * -0.5) - x);
}
real(8) function code(x, eps)
real(8), intent (in) :: x
real(8), intent (in) :: eps
code = eps * ((eps * (-0.5d0)) - x)
end function
public static double code(double x, double eps) {
return eps * ((eps * -0.5) - x);
}
def code(x, eps): return eps * ((eps * -0.5) - x)
function code(x, eps) return Float64(eps * Float64(Float64(eps * -0.5) - x)) end
function tmp = code(x, eps) tmp = eps * ((eps * -0.5) - x); end
code[x_, eps_] := N[(eps * N[(N[(eps * -0.5), $MachinePrecision] - x), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\varepsilon \cdot \left(\varepsilon \cdot -0.5 - x\right)
\end{array}
Initial program 55.0%
Taylor expanded in x around 0
+-commutativeN/A
associate--l+N/A
+-commutativeN/A
mul-1-negN/A
sub-negN/A
--lowering--.f64N/A
sub-negN/A
metadata-evalN/A
+-lowering-+.f64N/A
cos-lowering-cos.f64N/A
*-lowering-*.f64N/A
sin-lowering-sin.f6478.4%
Simplified78.4%
Taylor expanded in eps around 0
*-lowering-*.f64N/A
--lowering--.f64N/A
*-commutativeN/A
*-lowering-*.f6496.8%
Simplified96.8%
(FPCore (x eps) :precision binary64 (- 0.0 (* x eps)))
double code(double x, double eps) {
return 0.0 - (x * eps);
}
real(8) function code(x, eps)
real(8), intent (in) :: x
real(8), intent (in) :: eps
code = 0.0d0 - (x * eps)
end function
public static double code(double x, double eps) {
return 0.0 - (x * eps);
}
def code(x, eps): return 0.0 - (x * eps)
function code(x, eps) return Float64(0.0 - Float64(x * eps)) end
function tmp = code(x, eps) tmp = 0.0 - (x * eps); end
code[x_, eps_] := N[(0.0 - N[(x * eps), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
0 - x \cdot \varepsilon
\end{array}
Initial program 55.0%
Taylor expanded in eps around 0
associate-*r*N/A
*-commutativeN/A
*-lowering-*.f64N/A
sin-lowering-sin.f64N/A
mul-1-negN/A
neg-sub0N/A
--lowering--.f6479.9%
Simplified79.9%
Taylor expanded in x around 0
Simplified78.3%
Final simplification78.3%
(FPCore (x eps) :precision binary64 (* eps (* eps -0.5)))
double code(double x, double eps) {
return eps * (eps * -0.5);
}
real(8) function code(x, eps)
real(8), intent (in) :: x
real(8), intent (in) :: eps
code = eps * (eps * (-0.5d0))
end function
public static double code(double x, double eps) {
return eps * (eps * -0.5);
}
def code(x, eps): return eps * (eps * -0.5)
function code(x, eps) return Float64(eps * Float64(eps * -0.5)) end
function tmp = code(x, eps) tmp = eps * (eps * -0.5); end
code[x_, eps_] := N[(eps * N[(eps * -0.5), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\varepsilon \cdot \left(\varepsilon \cdot -0.5\right)
\end{array}
Initial program 55.0%
Taylor expanded in eps around 0
*-lowering-*.f64N/A
sub-negN/A
distribute-lft-inN/A
associate-+l+N/A
*-commutativeN/A
+-lowering-+.f64N/A
*-lowering-*.f64N/A
*-lowering-*.f64N/A
cos-lowering-cos.f64N/A
*-commutativeN/A
associate-*r*N/A
associate-*r*N/A
neg-mul-1N/A
Simplified99.8%
Taylor expanded in x around 0
*-commutativeN/A
unpow2N/A
associate-*l*N/A
*-commutativeN/A
*-lowering-*.f64N/A
*-commutativeN/A
*-lowering-*.f6455.0%
Simplified55.0%
(FPCore (x eps) :precision binary64 0.0)
double code(double x, double eps) {
return 0.0;
}
real(8) function code(x, eps)
real(8), intent (in) :: x
real(8), intent (in) :: eps
code = 0.0d0
end function
public static double code(double x, double eps) {
return 0.0;
}
def code(x, eps): return 0.0
function code(x, eps) return 0.0 end
function tmp = code(x, eps) tmp = 0.0; end
code[x_, eps_] := 0.0
\begin{array}{l}
\\
0
\end{array}
Initial program 55.0%
Taylor expanded in x around 0
sub-negN/A
metadata-evalN/A
+-lowering-+.f64N/A
cos-lowering-cos.f6453.5%
Simplified53.5%
Taylor expanded in eps around 0
Simplified53.4%
metadata-eval53.4%
Applied egg-rr53.4%
(FPCore (x eps) :precision binary64 (* (* -2.0 (sin (+ x (/ eps 2.0)))) (sin (/ eps 2.0))))
double code(double x, double eps) {
return (-2.0 * sin((x + (eps / 2.0)))) * sin((eps / 2.0));
}
real(8) function code(x, eps)
real(8), intent (in) :: x
real(8), intent (in) :: eps
code = ((-2.0d0) * sin((x + (eps / 2.0d0)))) * sin((eps / 2.0d0))
end function
public static double code(double x, double eps) {
return (-2.0 * Math.sin((x + (eps / 2.0)))) * Math.sin((eps / 2.0));
}
def code(x, eps): return (-2.0 * math.sin((x + (eps / 2.0)))) * math.sin((eps / 2.0))
function code(x, eps) return Float64(Float64(-2.0 * sin(Float64(x + Float64(eps / 2.0)))) * sin(Float64(eps / 2.0))) end
function tmp = code(x, eps) tmp = (-2.0 * sin((x + (eps / 2.0)))) * sin((eps / 2.0)); end
code[x_, eps_] := N[(N[(-2.0 * N[Sin[N[(x + N[(eps / 2.0), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]), $MachinePrecision] * N[Sin[N[(eps / 2.0), $MachinePrecision]], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\left(-2 \cdot \sin \left(x + \frac{\varepsilon}{2}\right)\right) \cdot \sin \left(\frac{\varepsilon}{2}\right)
\end{array}
herbie shell --seed 2024150
(FPCore (x eps)
:name "2cos (problem 3.3.5)"
:precision binary64
:pre (and (and (and (<= -10000.0 x) (<= x 10000.0)) (< (* 1e-16 (fabs x)) eps)) (< eps (fabs x)))
:alt
(! :herbie-platform default (* -2 (sin (+ x (/ eps 2))) (sin (/ eps 2))))
(- (cos (+ x eps)) (cos x)))