
(FPCore (x) :precision binary64 (sqrt (* (* 2.0 x) x)))
double code(double x) {
return sqrt(((2.0 * x) * x));
}
real(8) function code(x)
real(8), intent (in) :: x
code = sqrt(((2.0d0 * x) * x))
end function
public static double code(double x) {
return Math.sqrt(((2.0 * x) * x));
}
def code(x): return math.sqrt(((2.0 * x) * x))
function code(x) return sqrt(Float64(Float64(2.0 * x) * x)) end
function tmp = code(x) tmp = sqrt(((2.0 * x) * x)); end
code[x_] := N[Sqrt[N[(N[(2.0 * x), $MachinePrecision] * x), $MachinePrecision]], $MachinePrecision]
\begin{array}{l}
\\
\sqrt{\left(2 \cdot x\right) \cdot x}
\end{array}
Sampling outcomes in binary64 precision:
Herbie found 4 alternatives:
| Alternative | Accuracy | Speedup |
|---|
(FPCore (x) :precision binary64 (sqrt (* (* 2.0 x) x)))
double code(double x) {
return sqrt(((2.0 * x) * x));
}
real(8) function code(x)
real(8), intent (in) :: x
code = sqrt(((2.0d0 * x) * x))
end function
public static double code(double x) {
return Math.sqrt(((2.0 * x) * x));
}
def code(x): return math.sqrt(((2.0 * x) * x))
function code(x) return sqrt(Float64(Float64(2.0 * x) * x)) end
function tmp = code(x) tmp = sqrt(((2.0 * x) * x)); end
code[x_] := N[Sqrt[N[(N[(2.0 * x), $MachinePrecision] * x), $MachinePrecision]], $MachinePrecision]
\begin{array}{l}
\\
\sqrt{\left(2 \cdot x\right) \cdot x}
\end{array}
x_m = (fabs.f64 x) (FPCore (x_m) :precision binary64 (* (pow 2.0 0.375) (* x_m (pow 2.0 0.125))))
x_m = fabs(x);
double code(double x_m) {
return pow(2.0, 0.375) * (x_m * pow(2.0, 0.125));
}
x_m = abs(x)
real(8) function code(x_m)
real(8), intent (in) :: x_m
code = (2.0d0 ** 0.375d0) * (x_m * (2.0d0 ** 0.125d0))
end function
x_m = Math.abs(x);
public static double code(double x_m) {
return Math.pow(2.0, 0.375) * (x_m * Math.pow(2.0, 0.125));
}
x_m = math.fabs(x) def code(x_m): return math.pow(2.0, 0.375) * (x_m * math.pow(2.0, 0.125))
x_m = abs(x) function code(x_m) return Float64((2.0 ^ 0.375) * Float64(x_m * (2.0 ^ 0.125))) end
x_m = abs(x); function tmp = code(x_m) tmp = (2.0 ^ 0.375) * (x_m * (2.0 ^ 0.125)); end
x_m = N[Abs[x], $MachinePrecision] code[x$95$m_] := N[(N[Power[2.0, 0.375], $MachinePrecision] * N[(x$95$m * N[Power[2.0, 0.125], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
x_m = \left|x\right|
\\
{2}^{0.375} \cdot \left(x\_m \cdot {2}^{0.125}\right)
\end{array}
Initial program 61.4%
lift-sqrt.f64N/A
pow1/2N/A
lift-*.f64N/A
unpow-prod-downN/A
lower-*.f64N/A
pow1/2N/A
lower-sqrt.f64N/A
pow1/2N/A
lower-sqrt.f6445.4
Applied rewrites45.4%
lift-*.f64N/A
lift-sqrt.f64N/A
lift-*.f64N/A
sqrt-prodN/A
lift-sqrt.f64N/A
lift-sqrt.f64N/A
associate-*l*N/A
lift-sqrt.f64N/A
pow1/2N/A
metadata-evalN/A
pow-powN/A
lift-pow.f64N/A
pow2N/A
lift-sqrt.f64N/A
lift-sqrt.f64N/A
rem-square-sqrtN/A
associate-*r*N/A
lift-pow.f64N/A
metadata-evalN/A
pow-sqrN/A
lift-pow.f64N/A
lift-pow.f64N/A
associate-*l*N/A
associate-*r*N/A
lower-*.f64N/A
Applied rewrites46.8%
x_m = (fabs.f64 x) (FPCore (x_m) :precision binary64 (* (pow x_m 0.75) (sqrt (* 2.0 (sqrt x_m)))))
x_m = fabs(x);
double code(double x_m) {
return pow(x_m, 0.75) * sqrt((2.0 * sqrt(x_m)));
}
x_m = abs(x)
real(8) function code(x_m)
real(8), intent (in) :: x_m
code = (x_m ** 0.75d0) * sqrt((2.0d0 * sqrt(x_m)))
end function
x_m = Math.abs(x);
public static double code(double x_m) {
return Math.pow(x_m, 0.75) * Math.sqrt((2.0 * Math.sqrt(x_m)));
}
x_m = math.fabs(x) def code(x_m): return math.pow(x_m, 0.75) * math.sqrt((2.0 * math.sqrt(x_m)))
x_m = abs(x) function code(x_m) return Float64((x_m ^ 0.75) * sqrt(Float64(2.0 * sqrt(x_m)))) end
x_m = abs(x); function tmp = code(x_m) tmp = (x_m ^ 0.75) * sqrt((2.0 * sqrt(x_m))); end
x_m = N[Abs[x], $MachinePrecision] code[x$95$m_] := N[(N[Power[x$95$m, 0.75], $MachinePrecision] * N[Sqrt[N[(2.0 * N[Sqrt[x$95$m], $MachinePrecision]), $MachinePrecision]], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
x_m = \left|x\right|
\\
{x\_m}^{0.75} \cdot \sqrt{2 \cdot \sqrt{x\_m}}
\end{array}
Initial program 61.4%
lift-sqrt.f64N/A
lift-*.f64N/A
lift-*.f64N/A
associate-*l*N/A
*-commutativeN/A
rem-square-sqrtN/A
pow1/2N/A
pow1/2N/A
associate-*l*N/A
*-commutativeN/A
associate-*l*N/A
sqrt-prodN/A
rem-square-sqrtN/A
sqrt-prodN/A
pow1/2N/A
pow-prod-downN/A
sqrt-pow1N/A
pow-prod-downN/A
unpow-prod-downN/A
sqr-powN/A
*-commutativeN/A
Applied rewrites33.2%
lift-sqrt.f64N/A
pow1/2N/A
lift-*.f64N/A
*-commutativeN/A
lift-sqrt.f64N/A
pow1/2N/A
pow-plusN/A
pow-powN/A
lower-pow.f64N/A
metadata-evalN/A
metadata-eval45.5
Applied rewrites45.5%
Final simplification45.5%
x_m = (fabs.f64 x) (FPCore (x_m) :precision binary64 (* (sqrt x_m) (sqrt (* 2.0 x_m))))
x_m = fabs(x);
double code(double x_m) {
return sqrt(x_m) * sqrt((2.0 * x_m));
}
x_m = abs(x)
real(8) function code(x_m)
real(8), intent (in) :: x_m
code = sqrt(x_m) * sqrt((2.0d0 * x_m))
end function
x_m = Math.abs(x);
public static double code(double x_m) {
return Math.sqrt(x_m) * Math.sqrt((2.0 * x_m));
}
x_m = math.fabs(x) def code(x_m): return math.sqrt(x_m) * math.sqrt((2.0 * x_m))
x_m = abs(x) function code(x_m) return Float64(sqrt(x_m) * sqrt(Float64(2.0 * x_m))) end
x_m = abs(x); function tmp = code(x_m) tmp = sqrt(x_m) * sqrt((2.0 * x_m)); end
x_m = N[Abs[x], $MachinePrecision] code[x$95$m_] := N[(N[Sqrt[x$95$m], $MachinePrecision] * N[Sqrt[N[(2.0 * x$95$m), $MachinePrecision]], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
x_m = \left|x\right|
\\
\sqrt{x\_m} \cdot \sqrt{2 \cdot x\_m}
\end{array}
Initial program 61.4%
lift-sqrt.f64N/A
pow1/2N/A
lift-*.f64N/A
unpow-prod-downN/A
lower-*.f64N/A
pow1/2N/A
lower-sqrt.f64N/A
pow1/2N/A
lower-sqrt.f6445.4
Applied rewrites45.4%
Final simplification45.4%
x_m = (fabs.f64 x) (FPCore (x_m) :precision binary64 (* x_m (sqrt 2.0)))
x_m = fabs(x);
double code(double x_m) {
return x_m * sqrt(2.0);
}
x_m = abs(x)
real(8) function code(x_m)
real(8), intent (in) :: x_m
code = x_m * sqrt(2.0d0)
end function
x_m = Math.abs(x);
public static double code(double x_m) {
return x_m * Math.sqrt(2.0);
}
x_m = math.fabs(x) def code(x_m): return x_m * math.sqrt(2.0)
x_m = abs(x) function code(x_m) return Float64(x_m * sqrt(2.0)) end
x_m = abs(x); function tmp = code(x_m) tmp = x_m * sqrt(2.0); end
x_m = N[Abs[x], $MachinePrecision] code[x$95$m_] := N[(x$95$m * N[Sqrt[2.0], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
x_m = \left|x\right|
\\
x\_m \cdot \sqrt{2}
\end{array}
Initial program 61.4%
lift-sqrt.f64N/A
pow1/2N/A
lift-*.f64N/A
lift-*.f64N/A
associate-*l*N/A
unpow-prod-downN/A
pow1/2N/A
sqrt-prodN/A
rem-square-sqrtN/A
lower-*.f64N/A
pow1/2N/A
lower-sqrt.f6446.7
Applied rewrites46.7%
Final simplification46.7%
herbie shell --seed 2024237
(FPCore (x)
:name "sqrt B (should all be same)"
:precision binary64
(sqrt (* (* 2.0 x) x)))