
(FPCore (v) :precision binary64 (* (* (/ (sqrt 2.0) 4.0) (sqrt (- 1.0 (* 3.0 (* v v))))) (- 1.0 (* v v))))
double code(double v) {
return ((sqrt(2.0) / 4.0) * sqrt((1.0 - (3.0 * (v * v))))) * (1.0 - (v * v));
}
real(8) function code(v)
real(8), intent (in) :: v
code = ((sqrt(2.0d0) / 4.0d0) * sqrt((1.0d0 - (3.0d0 * (v * v))))) * (1.0d0 - (v * v))
end function
public static double code(double v) {
return ((Math.sqrt(2.0) / 4.0) * Math.sqrt((1.0 - (3.0 * (v * v))))) * (1.0 - (v * v));
}
def code(v): return ((math.sqrt(2.0) / 4.0) * math.sqrt((1.0 - (3.0 * (v * v))))) * (1.0 - (v * v))
function code(v) return Float64(Float64(Float64(sqrt(2.0) / 4.0) * sqrt(Float64(1.0 - Float64(3.0 * Float64(v * v))))) * Float64(1.0 - Float64(v * v))) end
function tmp = code(v) tmp = ((sqrt(2.0) / 4.0) * sqrt((1.0 - (3.0 * (v * v))))) * (1.0 - (v * v)); end
code[v_] := N[(N[(N[(N[Sqrt[2.0], $MachinePrecision] / 4.0), $MachinePrecision] * N[Sqrt[N[(1.0 - N[(3.0 * N[(v * v), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]), $MachinePrecision] * N[(1.0 - N[(v * v), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\left(\frac{\sqrt{2}}{4} \cdot \sqrt{1 - 3 \cdot \left(v \cdot v\right)}\right) \cdot \left(1 - v \cdot v\right)
\end{array}
Sampling outcomes in binary64 precision:
Herbie found 4 alternatives:
| Alternative | Accuracy | Speedup |
|---|
(FPCore (v) :precision binary64 (* (* (/ (sqrt 2.0) 4.0) (sqrt (- 1.0 (* 3.0 (* v v))))) (- 1.0 (* v v))))
double code(double v) {
return ((sqrt(2.0) / 4.0) * sqrt((1.0 - (3.0 * (v * v))))) * (1.0 - (v * v));
}
real(8) function code(v)
real(8), intent (in) :: v
code = ((sqrt(2.0d0) / 4.0d0) * sqrt((1.0d0 - (3.0d0 * (v * v))))) * (1.0d0 - (v * v))
end function
public static double code(double v) {
return ((Math.sqrt(2.0) / 4.0) * Math.sqrt((1.0 - (3.0 * (v * v))))) * (1.0 - (v * v));
}
def code(v): return ((math.sqrt(2.0) / 4.0) * math.sqrt((1.0 - (3.0 * (v * v))))) * (1.0 - (v * v))
function code(v) return Float64(Float64(Float64(sqrt(2.0) / 4.0) * sqrt(Float64(1.0 - Float64(3.0 * Float64(v * v))))) * Float64(1.0 - Float64(v * v))) end
function tmp = code(v) tmp = ((sqrt(2.0) / 4.0) * sqrt((1.0 - (3.0 * (v * v))))) * (1.0 - (v * v)); end
code[v_] := N[(N[(N[(N[Sqrt[2.0], $MachinePrecision] / 4.0), $MachinePrecision] * N[Sqrt[N[(1.0 - N[(3.0 * N[(v * v), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]), $MachinePrecision] * N[(1.0 - N[(v * v), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\left(\frac{\sqrt{2}}{4} \cdot \sqrt{1 - 3 \cdot \left(v \cdot v\right)}\right) \cdot \left(1 - v \cdot v\right)
\end{array}
(FPCore (v) :precision binary64 (/ (* (sqrt (fma v (* v -6.0) 2.0)) -0.25) (/ 1.0 (fma v v -1.0))))
double code(double v) {
return (sqrt(fma(v, (v * -6.0), 2.0)) * -0.25) / (1.0 / fma(v, v, -1.0));
}
function code(v) return Float64(Float64(sqrt(fma(v, Float64(v * -6.0), 2.0)) * -0.25) / Float64(1.0 / fma(v, v, -1.0))) end
code[v_] := N[(N[(N[Sqrt[N[(v * N[(v * -6.0), $MachinePrecision] + 2.0), $MachinePrecision]], $MachinePrecision] * -0.25), $MachinePrecision] / N[(1.0 / N[(v * v + -1.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\frac{\sqrt{\mathsf{fma}\left(v, v \cdot -6, 2\right)} \cdot -0.25}{\frac{1}{\mathsf{fma}\left(v, v, -1\right)}}
\end{array}
Initial program 100.0%
div-invN/A
*-lowering-*.f64N/A
sqrt-lowering-sqrt.f64N/A
metadata-eval100.0
Applied egg-rr100.0%
*-commutativeN/A
*-lowering-*.f64N/A
--lowering--.f64N/A
*-lowering-*.f64N/A
*-commutativeN/A
associate-*l*N/A
*-lowering-*.f64N/A
metadata-evalN/A
*-commutativeN/A
*-rgt-identityN/A
sqrt-unprodN/A
sqrt-lowering-sqrt.f64N/A
*-lowering-*.f64N/A
metadata-evalN/A
sub-negN/A
*-rgt-identityN/A
*-commutativeN/A
Applied egg-rr100.0%
*-commutativeN/A
*-commutativeN/A
distribute-rgt-inN/A
metadata-evalN/A
associate-*l*N/A
metadata-evalN/A
associate-*r*N/A
*-commutativeN/A
/-rgt-identityN/A
associate-/r/N/A
frac-2negN/A
/-lowering-/.f64N/A
Applied egg-rr100.0%
(FPCore (v) :precision binary64 (* (* (sqrt (fma v (* v -6.0) 2.0)) -0.25) (fma v v -1.0)))
double code(double v) {
return (sqrt(fma(v, (v * -6.0), 2.0)) * -0.25) * fma(v, v, -1.0);
}
function code(v) return Float64(Float64(sqrt(fma(v, Float64(v * -6.0), 2.0)) * -0.25) * fma(v, v, -1.0)) end
code[v_] := N[(N[(N[Sqrt[N[(v * N[(v * -6.0), $MachinePrecision] + 2.0), $MachinePrecision]], $MachinePrecision] * -0.25), $MachinePrecision] * N[(v * v + -1.0), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\left(\sqrt{\mathsf{fma}\left(v, v \cdot -6, 2\right)} \cdot -0.25\right) \cdot \mathsf{fma}\left(v, v, -1\right)
\end{array}
Initial program 100.0%
div-invN/A
*-lowering-*.f64N/A
sqrt-lowering-sqrt.f64N/A
metadata-eval100.0
Applied egg-rr100.0%
*-commutativeN/A
*-lowering-*.f64N/A
--lowering--.f64N/A
*-lowering-*.f64N/A
*-commutativeN/A
associate-*l*N/A
*-lowering-*.f64N/A
metadata-evalN/A
*-commutativeN/A
*-rgt-identityN/A
sqrt-unprodN/A
sqrt-lowering-sqrt.f64N/A
*-lowering-*.f64N/A
metadata-evalN/A
sub-negN/A
*-rgt-identityN/A
*-commutativeN/A
Applied egg-rr100.0%
*-commutativeN/A
associate-*l*N/A
distribute-rgt-inN/A
metadata-evalN/A
associate-*l*N/A
metadata-evalN/A
associate-*r*N/A
associate-*l*N/A
/-rgt-identityN/A
frac-2negN/A
metadata-evalN/A
div-invN/A
metadata-evalN/A
Applied egg-rr100.0%
Final simplification100.0%
(FPCore (v) :precision binary64 (* (sqrt 2.0) (fma (* v v) -0.625 0.25)))
double code(double v) {
return sqrt(2.0) * fma((v * v), -0.625, 0.25);
}
function code(v) return Float64(sqrt(2.0) * fma(Float64(v * v), -0.625, 0.25)) end
code[v_] := N[(N[Sqrt[2.0], $MachinePrecision] * N[(N[(v * v), $MachinePrecision] * -0.625 + 0.25), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\sqrt{2} \cdot \mathsf{fma}\left(v \cdot v, -0.625, 0.25\right)
\end{array}
Initial program 100.0%
Applied egg-rr100.0%
Taylor expanded in v around 0
+-commutativeN/A
*-commutativeN/A
associate-*r*N/A
*-commutativeN/A
distribute-rgt-outN/A
metadata-evalN/A
associate-*l*N/A
metadata-evalN/A
associate-*l*N/A
*-commutativeN/A
distribute-lft-outN/A
*-lowering-*.f64N/A
sqrt-lowering-sqrt.f64N/A
*-commutativeN/A
accelerator-lowering-fma.f64N/A
Simplified99.8%
(FPCore (v) :precision binary64 (* (sqrt 2.0) 0.25))
double code(double v) {
return sqrt(2.0) * 0.25;
}
real(8) function code(v)
real(8), intent (in) :: v
code = sqrt(2.0d0) * 0.25d0
end function
public static double code(double v) {
return Math.sqrt(2.0) * 0.25;
}
def code(v): return math.sqrt(2.0) * 0.25
function code(v) return Float64(sqrt(2.0) * 0.25) end
function tmp = code(v) tmp = sqrt(2.0) * 0.25; end
code[v_] := N[(N[Sqrt[2.0], $MachinePrecision] * 0.25), $MachinePrecision]
\begin{array}{l}
\\
\sqrt{2} \cdot 0.25
\end{array}
Initial program 100.0%
Taylor expanded in v around 0
*-lowering-*.f64N/A
sqrt-lowering-sqrt.f6499.4
Simplified99.4%
Final simplification99.4%
herbie shell --seed 2024204
(FPCore (v)
:name "Falkner and Boettcher, Appendix B, 2"
:precision binary64
(* (* (/ (sqrt 2.0) 4.0) (sqrt (- 1.0 (* 3.0 (* v v))))) (- 1.0 (* v v))))