
(FPCore (x) :precision binary64 (* (* x x) (- 3.0 (* x 2.0))))
double code(double x) {
return (x * x) * (3.0 - (x * 2.0));
}
real(8) function code(x)
real(8), intent (in) :: x
code = (x * x) * (3.0d0 - (x * 2.0d0))
end function
public static double code(double x) {
return (x * x) * (3.0 - (x * 2.0));
}
def code(x): return (x * x) * (3.0 - (x * 2.0))
function code(x) return Float64(Float64(x * x) * Float64(3.0 - Float64(x * 2.0))) end
function tmp = code(x) tmp = (x * x) * (3.0 - (x * 2.0)); end
code[x_] := N[(N[(x * x), $MachinePrecision] * N[(3.0 - N[(x * 2.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\left(x \cdot x\right) \cdot \left(3 - x \cdot 2\right)
\end{array}
Sampling outcomes in binary64 precision:
Herbie found 4 alternatives:
| Alternative | Accuracy | Speedup |
|---|
(FPCore (x) :precision binary64 (* (* x x) (- 3.0 (* x 2.0))))
double code(double x) {
return (x * x) * (3.0 - (x * 2.0));
}
real(8) function code(x)
real(8), intent (in) :: x
code = (x * x) * (3.0d0 - (x * 2.0d0))
end function
public static double code(double x) {
return (x * x) * (3.0 - (x * 2.0));
}
def code(x): return (x * x) * (3.0 - (x * 2.0))
function code(x) return Float64(Float64(x * x) * Float64(3.0 - Float64(x * 2.0))) end
function tmp = code(x) tmp = (x * x) * (3.0 - (x * 2.0)); end
code[x_] := N[(N[(x * x), $MachinePrecision] * N[(3.0 - N[(x * 2.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\left(x \cdot x\right) \cdot \left(3 - x \cdot 2\right)
\end{array}
(FPCore (x) :precision binary64 (* (* (fma -2.0 x 3.0) x) x))
double code(double x) {
return (fma(-2.0, x, 3.0) * x) * x;
}
function code(x) return Float64(Float64(fma(-2.0, x, 3.0) * x) * x) end
code[x_] := N[(N[(N[(-2.0 * x + 3.0), $MachinePrecision] * x), $MachinePrecision] * x), $MachinePrecision]
\begin{array}{l}
\\
\left(\mathsf{fma}\left(-2, x, 3\right) \cdot x\right) \cdot x
\end{array}
Initial program 99.7%
lift-*.f64N/A
lift-*.f64N/A
associate-*l*N/A
*-commutativeN/A
lower-*.f64N/A
*-commutativeN/A
lower-*.f6499.8
lift--.f64N/A
sub-negN/A
+-commutativeN/A
lift-*.f64N/A
*-commutativeN/A
distribute-lft-neg-inN/A
lower-fma.f64N/A
metadata-eval99.8
Applied rewrites99.8%
(FPCore (x)
:precision binary64
(let* ((t_0 (* (* x x) (- 3.0 (* x 2.0)))))
(if (or (<= t_0 -200.0) (not (<= t_0 1.0)))
(* (* x x) (* -2.0 x))
(* (* 3.0 x) x))))
double code(double x) {
double t_0 = (x * x) * (3.0 - (x * 2.0));
double tmp;
if ((t_0 <= -200.0) || !(t_0 <= 1.0)) {
tmp = (x * x) * (-2.0 * x);
} else {
tmp = (3.0 * x) * x;
}
return tmp;
}
real(8) function code(x)
real(8), intent (in) :: x
real(8) :: t_0
real(8) :: tmp
t_0 = (x * x) * (3.0d0 - (x * 2.0d0))
if ((t_0 <= (-200.0d0)) .or. (.not. (t_0 <= 1.0d0))) then
tmp = (x * x) * ((-2.0d0) * x)
else
tmp = (3.0d0 * x) * x
end if
code = tmp
end function
public static double code(double x) {
double t_0 = (x * x) * (3.0 - (x * 2.0));
double tmp;
if ((t_0 <= -200.0) || !(t_0 <= 1.0)) {
tmp = (x * x) * (-2.0 * x);
} else {
tmp = (3.0 * x) * x;
}
return tmp;
}
def code(x): t_0 = (x * x) * (3.0 - (x * 2.0)) tmp = 0 if (t_0 <= -200.0) or not (t_0 <= 1.0): tmp = (x * x) * (-2.0 * x) else: tmp = (3.0 * x) * x return tmp
function code(x) t_0 = Float64(Float64(x * x) * Float64(3.0 - Float64(x * 2.0))) tmp = 0.0 if ((t_0 <= -200.0) || !(t_0 <= 1.0)) tmp = Float64(Float64(x * x) * Float64(-2.0 * x)); else tmp = Float64(Float64(3.0 * x) * x); end return tmp end
function tmp_2 = code(x) t_0 = (x * x) * (3.0 - (x * 2.0)); tmp = 0.0; if ((t_0 <= -200.0) || ~((t_0 <= 1.0))) tmp = (x * x) * (-2.0 * x); else tmp = (3.0 * x) * x; end tmp_2 = tmp; end
code[x_] := Block[{t$95$0 = N[(N[(x * x), $MachinePrecision] * N[(3.0 - N[(x * 2.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]}, If[Or[LessEqual[t$95$0, -200.0], N[Not[LessEqual[t$95$0, 1.0]], $MachinePrecision]], N[(N[(x * x), $MachinePrecision] * N[(-2.0 * x), $MachinePrecision]), $MachinePrecision], N[(N[(3.0 * x), $MachinePrecision] * x), $MachinePrecision]]]
\begin{array}{l}
\\
\begin{array}{l}
t_0 := \left(x \cdot x\right) \cdot \left(3 - x \cdot 2\right)\\
\mathbf{if}\;t\_0 \leq -200 \lor \neg \left(t\_0 \leq 1\right):\\
\;\;\;\;\left(x \cdot x\right) \cdot \left(-2 \cdot x\right)\\
\mathbf{else}:\\
\;\;\;\;\left(3 \cdot x\right) \cdot x\\
\end{array}
\end{array}
if (*.f64 (*.f64 x x) (-.f64 #s(literal 3 binary64) (*.f64 x #s(literal 2 binary64)))) < -200 or 1 < (*.f64 (*.f64 x x) (-.f64 #s(literal 3 binary64) (*.f64 x #s(literal 2 binary64)))) Initial program 99.8%
Taylor expanded in x around inf
lower-*.f6497.3
Applied rewrites97.3%
if -200 < (*.f64 (*.f64 x x) (-.f64 #s(literal 3 binary64) (*.f64 x #s(literal 2 binary64)))) < 1Initial program 99.7%
Taylor expanded in x around 0
Applied rewrites97.2%
lift-*.f64N/A
lift-*.f64N/A
associate-*l*N/A
*-commutativeN/A
lower-*.f64N/A
*-commutativeN/A
lower-*.f6497.3
Applied rewrites97.3%
Final simplification97.3%
(FPCore (x) :precision binary64 (* (* 3.0 x) x))
double code(double x) {
return (3.0 * x) * x;
}
real(8) function code(x)
real(8), intent (in) :: x
code = (3.0d0 * x) * x
end function
public static double code(double x) {
return (3.0 * x) * x;
}
def code(x): return (3.0 * x) * x
function code(x) return Float64(Float64(3.0 * x) * x) end
function tmp = code(x) tmp = (3.0 * x) * x; end
code[x_] := N[(N[(3.0 * x), $MachinePrecision] * x), $MachinePrecision]
\begin{array}{l}
\\
\left(3 \cdot x\right) \cdot x
\end{array}
Initial program 99.7%
Taylor expanded in x around 0
Applied rewrites61.1%
lift-*.f64N/A
lift-*.f64N/A
associate-*l*N/A
*-commutativeN/A
lower-*.f64N/A
*-commutativeN/A
lower-*.f6461.2
Applied rewrites61.2%
(FPCore (x) :precision binary64 (* (* x x) 3.0))
double code(double x) {
return (x * x) * 3.0;
}
real(8) function code(x)
real(8), intent (in) :: x
code = (x * x) * 3.0d0
end function
public static double code(double x) {
return (x * x) * 3.0;
}
def code(x): return (x * x) * 3.0
function code(x) return Float64(Float64(x * x) * 3.0) end
function tmp = code(x) tmp = (x * x) * 3.0; end
code[x_] := N[(N[(x * x), $MachinePrecision] * 3.0), $MachinePrecision]
\begin{array}{l}
\\
\left(x \cdot x\right) \cdot 3
\end{array}
Initial program 99.7%
Taylor expanded in x around 0
Applied rewrites61.1%
(FPCore (x) :precision binary64 (* x (* x (- 3.0 (* x 2.0)))))
double code(double x) {
return x * (x * (3.0 - (x * 2.0)));
}
real(8) function code(x)
real(8), intent (in) :: x
code = x * (x * (3.0d0 - (x * 2.0d0)))
end function
public static double code(double x) {
return x * (x * (3.0 - (x * 2.0)));
}
def code(x): return x * (x * (3.0 - (x * 2.0)))
function code(x) return Float64(x * Float64(x * Float64(3.0 - Float64(x * 2.0)))) end
function tmp = code(x) tmp = x * (x * (3.0 - (x * 2.0))); end
code[x_] := N[(x * N[(x * N[(3.0 - N[(x * 2.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
x \cdot \left(x \cdot \left(3 - x \cdot 2\right)\right)
\end{array}
herbie shell --seed 2024313
(FPCore (x)
:name "Data.Spline.Key:interpolateKeys from smoothie-0.4.0.2"
:precision binary64
:alt
(! :herbie-platform default (* x (* x (- 3 (* x 2)))))
(* (* x x) (- 3.0 (* x 2.0))))