
(FPCore (x)
:precision binary64
(let* ((t_0 (* (* (fabs x) (fabs x)) (fabs x)))
(t_1 (* (* t_0 (fabs x)) (fabs x))))
(fabs
(*
(/ 1.0 (sqrt PI))
(+
(+ (+ (* 2.0 (fabs x)) (* (/ 2.0 3.0) t_0)) (* (/ 1.0 5.0) t_1))
(* (/ 1.0 21.0) (* (* t_1 (fabs x)) (fabs x))))))))
double code(double x) {
double t_0 = (fabs(x) * fabs(x)) * fabs(x);
double t_1 = (t_0 * fabs(x)) * fabs(x);
return fabs(((1.0 / sqrt(((double) M_PI))) * ((((2.0 * fabs(x)) + ((2.0 / 3.0) * t_0)) + ((1.0 / 5.0) * t_1)) + ((1.0 / 21.0) * ((t_1 * fabs(x)) * fabs(x))))));
}
public static double code(double x) {
double t_0 = (Math.abs(x) * Math.abs(x)) * Math.abs(x);
double t_1 = (t_0 * Math.abs(x)) * Math.abs(x);
return Math.abs(((1.0 / Math.sqrt(Math.PI)) * ((((2.0 * Math.abs(x)) + ((2.0 / 3.0) * t_0)) + ((1.0 / 5.0) * t_1)) + ((1.0 / 21.0) * ((t_1 * Math.abs(x)) * Math.abs(x))))));
}
def code(x): t_0 = (math.fabs(x) * math.fabs(x)) * math.fabs(x) t_1 = (t_0 * math.fabs(x)) * math.fabs(x) return math.fabs(((1.0 / math.sqrt(math.pi)) * ((((2.0 * math.fabs(x)) + ((2.0 / 3.0) * t_0)) + ((1.0 / 5.0) * t_1)) + ((1.0 / 21.0) * ((t_1 * math.fabs(x)) * math.fabs(x))))))
function code(x) t_0 = Float64(Float64(abs(x) * abs(x)) * abs(x)) t_1 = Float64(Float64(t_0 * abs(x)) * abs(x)) return abs(Float64(Float64(1.0 / sqrt(pi)) * Float64(Float64(Float64(Float64(2.0 * abs(x)) + Float64(Float64(2.0 / 3.0) * t_0)) + Float64(Float64(1.0 / 5.0) * t_1)) + Float64(Float64(1.0 / 21.0) * Float64(Float64(t_1 * abs(x)) * abs(x)))))) end
function tmp = code(x) t_0 = (abs(x) * abs(x)) * abs(x); t_1 = (t_0 * abs(x)) * abs(x); tmp = abs(((1.0 / sqrt(pi)) * ((((2.0 * abs(x)) + ((2.0 / 3.0) * t_0)) + ((1.0 / 5.0) * t_1)) + ((1.0 / 21.0) * ((t_1 * abs(x)) * abs(x)))))); end
code[x_] := Block[{t$95$0 = N[(N[(N[Abs[x], $MachinePrecision] * N[Abs[x], $MachinePrecision]), $MachinePrecision] * N[Abs[x], $MachinePrecision]), $MachinePrecision]}, Block[{t$95$1 = N[(N[(t$95$0 * N[Abs[x], $MachinePrecision]), $MachinePrecision] * N[Abs[x], $MachinePrecision]), $MachinePrecision]}, N[Abs[N[(N[(1.0 / N[Sqrt[Pi], $MachinePrecision]), $MachinePrecision] * N[(N[(N[(N[(2.0 * N[Abs[x], $MachinePrecision]), $MachinePrecision] + N[(N[(2.0 / 3.0), $MachinePrecision] * t$95$0), $MachinePrecision]), $MachinePrecision] + N[(N[(1.0 / 5.0), $MachinePrecision] * t$95$1), $MachinePrecision]), $MachinePrecision] + N[(N[(1.0 / 21.0), $MachinePrecision] * N[(N[(t$95$1 * N[Abs[x], $MachinePrecision]), $MachinePrecision] * N[Abs[x], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]]]
\begin{array}{l}
\\
\begin{array}{l}
t_0 := \left(\left|x\right| \cdot \left|x\right|\right) \cdot \left|x\right|\\
t_1 := \left(t_0 \cdot \left|x\right|\right) \cdot \left|x\right|\\
\left|\frac{1}{\sqrt{\pi}} \cdot \left(\left(\left(2 \cdot \left|x\right| + \frac{2}{3} \cdot t_0\right) + \frac{1}{5} \cdot t_1\right) + \frac{1}{21} \cdot \left(\left(t_1 \cdot \left|x\right|\right) \cdot \left|x\right|\right)\right)\right|
\end{array}
\end{array}
Sampling outcomes in binary64 precision:
Herbie found 12 alternatives:
| Alternative | Accuracy | Speedup |
|---|
(FPCore (x)
:precision binary64
(let* ((t_0 (* (* (fabs x) (fabs x)) (fabs x)))
(t_1 (* (* t_0 (fabs x)) (fabs x))))
(fabs
(*
(/ 1.0 (sqrt PI))
(+
(+ (+ (* 2.0 (fabs x)) (* (/ 2.0 3.0) t_0)) (* (/ 1.0 5.0) t_1))
(* (/ 1.0 21.0) (* (* t_1 (fabs x)) (fabs x))))))))
double code(double x) {
double t_0 = (fabs(x) * fabs(x)) * fabs(x);
double t_1 = (t_0 * fabs(x)) * fabs(x);
return fabs(((1.0 / sqrt(((double) M_PI))) * ((((2.0 * fabs(x)) + ((2.0 / 3.0) * t_0)) + ((1.0 / 5.0) * t_1)) + ((1.0 / 21.0) * ((t_1 * fabs(x)) * fabs(x))))));
}
public static double code(double x) {
double t_0 = (Math.abs(x) * Math.abs(x)) * Math.abs(x);
double t_1 = (t_0 * Math.abs(x)) * Math.abs(x);
return Math.abs(((1.0 / Math.sqrt(Math.PI)) * ((((2.0 * Math.abs(x)) + ((2.0 / 3.0) * t_0)) + ((1.0 / 5.0) * t_1)) + ((1.0 / 21.0) * ((t_1 * Math.abs(x)) * Math.abs(x))))));
}
def code(x): t_0 = (math.fabs(x) * math.fabs(x)) * math.fabs(x) t_1 = (t_0 * math.fabs(x)) * math.fabs(x) return math.fabs(((1.0 / math.sqrt(math.pi)) * ((((2.0 * math.fabs(x)) + ((2.0 / 3.0) * t_0)) + ((1.0 / 5.0) * t_1)) + ((1.0 / 21.0) * ((t_1 * math.fabs(x)) * math.fabs(x))))))
function code(x) t_0 = Float64(Float64(abs(x) * abs(x)) * abs(x)) t_1 = Float64(Float64(t_0 * abs(x)) * abs(x)) return abs(Float64(Float64(1.0 / sqrt(pi)) * Float64(Float64(Float64(Float64(2.0 * abs(x)) + Float64(Float64(2.0 / 3.0) * t_0)) + Float64(Float64(1.0 / 5.0) * t_1)) + Float64(Float64(1.0 / 21.0) * Float64(Float64(t_1 * abs(x)) * abs(x)))))) end
function tmp = code(x) t_0 = (abs(x) * abs(x)) * abs(x); t_1 = (t_0 * abs(x)) * abs(x); tmp = abs(((1.0 / sqrt(pi)) * ((((2.0 * abs(x)) + ((2.0 / 3.0) * t_0)) + ((1.0 / 5.0) * t_1)) + ((1.0 / 21.0) * ((t_1 * abs(x)) * abs(x)))))); end
code[x_] := Block[{t$95$0 = N[(N[(N[Abs[x], $MachinePrecision] * N[Abs[x], $MachinePrecision]), $MachinePrecision] * N[Abs[x], $MachinePrecision]), $MachinePrecision]}, Block[{t$95$1 = N[(N[(t$95$0 * N[Abs[x], $MachinePrecision]), $MachinePrecision] * N[Abs[x], $MachinePrecision]), $MachinePrecision]}, N[Abs[N[(N[(1.0 / N[Sqrt[Pi], $MachinePrecision]), $MachinePrecision] * N[(N[(N[(N[(2.0 * N[Abs[x], $MachinePrecision]), $MachinePrecision] + N[(N[(2.0 / 3.0), $MachinePrecision] * t$95$0), $MachinePrecision]), $MachinePrecision] + N[(N[(1.0 / 5.0), $MachinePrecision] * t$95$1), $MachinePrecision]), $MachinePrecision] + N[(N[(1.0 / 21.0), $MachinePrecision] * N[(N[(t$95$1 * N[Abs[x], $MachinePrecision]), $MachinePrecision] * N[Abs[x], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]]]
\begin{array}{l}
\\
\begin{array}{l}
t_0 := \left(\left|x\right| \cdot \left|x\right|\right) \cdot \left|x\right|\\
t_1 := \left(t_0 \cdot \left|x\right|\right) \cdot \left|x\right|\\
\left|\frac{1}{\sqrt{\pi}} \cdot \left(\left(\left(2 \cdot \left|x\right| + \frac{2}{3} \cdot t_0\right) + \frac{1}{5} \cdot t_1\right) + \frac{1}{21} \cdot \left(\left(t_1 \cdot \left|x\right|\right) \cdot \left|x\right|\right)\right)\right|
\end{array}
\end{array}
(FPCore (x)
:precision binary64
(let* ((t_0 (* (fabs x) (* x x))) (t_1 (* (fabs x) (* (fabs x) t_0))))
(fabs
(*
(/ 1.0 (sqrt PI))
(+
(+ (+ (* 2.0 (fabs x)) (* 0.6666666666666666 t_0)) (* 0.2 t_1))
(* 0.047619047619047616 (* (fabs x) (* (fabs x) t_1))))))))
double code(double x) {
double t_0 = fabs(x) * (x * x);
double t_1 = fabs(x) * (fabs(x) * t_0);
return fabs(((1.0 / sqrt(((double) M_PI))) * ((((2.0 * fabs(x)) + (0.6666666666666666 * t_0)) + (0.2 * t_1)) + (0.047619047619047616 * (fabs(x) * (fabs(x) * t_1))))));
}
public static double code(double x) {
double t_0 = Math.abs(x) * (x * x);
double t_1 = Math.abs(x) * (Math.abs(x) * t_0);
return Math.abs(((1.0 / Math.sqrt(Math.PI)) * ((((2.0 * Math.abs(x)) + (0.6666666666666666 * t_0)) + (0.2 * t_1)) + (0.047619047619047616 * (Math.abs(x) * (Math.abs(x) * t_1))))));
}
def code(x): t_0 = math.fabs(x) * (x * x) t_1 = math.fabs(x) * (math.fabs(x) * t_0) return math.fabs(((1.0 / math.sqrt(math.pi)) * ((((2.0 * math.fabs(x)) + (0.6666666666666666 * t_0)) + (0.2 * t_1)) + (0.047619047619047616 * (math.fabs(x) * (math.fabs(x) * t_1))))))
function code(x) t_0 = Float64(abs(x) * Float64(x * x)) t_1 = Float64(abs(x) * Float64(abs(x) * t_0)) return abs(Float64(Float64(1.0 / sqrt(pi)) * Float64(Float64(Float64(Float64(2.0 * abs(x)) + Float64(0.6666666666666666 * t_0)) + Float64(0.2 * t_1)) + Float64(0.047619047619047616 * Float64(abs(x) * Float64(abs(x) * t_1)))))) end
function tmp = code(x) t_0 = abs(x) * (x * x); t_1 = abs(x) * (abs(x) * t_0); tmp = abs(((1.0 / sqrt(pi)) * ((((2.0 * abs(x)) + (0.6666666666666666 * t_0)) + (0.2 * t_1)) + (0.047619047619047616 * (abs(x) * (abs(x) * t_1)))))); end
code[x_] := Block[{t$95$0 = N[(N[Abs[x], $MachinePrecision] * N[(x * x), $MachinePrecision]), $MachinePrecision]}, Block[{t$95$1 = N[(N[Abs[x], $MachinePrecision] * N[(N[Abs[x], $MachinePrecision] * t$95$0), $MachinePrecision]), $MachinePrecision]}, N[Abs[N[(N[(1.0 / N[Sqrt[Pi], $MachinePrecision]), $MachinePrecision] * N[(N[(N[(N[(2.0 * N[Abs[x], $MachinePrecision]), $MachinePrecision] + N[(0.6666666666666666 * t$95$0), $MachinePrecision]), $MachinePrecision] + N[(0.2 * t$95$1), $MachinePrecision]), $MachinePrecision] + N[(0.047619047619047616 * N[(N[Abs[x], $MachinePrecision] * N[(N[Abs[x], $MachinePrecision] * t$95$1), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]]]
\begin{array}{l}
\\
\begin{array}{l}
t_0 := \left|x\right| \cdot \left(x \cdot x\right)\\
t_1 := \left|x\right| \cdot \left(\left|x\right| \cdot t_0\right)\\
\left|\frac{1}{\sqrt{\pi}} \cdot \left(\left(\left(2 \cdot \left|x\right| + 0.6666666666666666 \cdot t_0\right) + 0.2 \cdot t_1\right) + 0.047619047619047616 \cdot \left(\left|x\right| \cdot \left(\left|x\right| \cdot t_1\right)\right)\right)\right|
\end{array}
\end{array}
Initial program 99.8%
Final simplification99.8%
(FPCore (x)
:precision binary64
(let* ((t_0 (* (fabs x) (* x x))) (t_1 (* (* x x) t_0)))
(fabs
(*
(/ 1.0 (sqrt PI))
(+
(+ (fma 2.0 (fabs x) (* 0.6666666666666666 t_0)) (* 0.2 t_1))
(* 0.047619047619047616 (* (* x x) t_1)))))))
double code(double x) {
double t_0 = fabs(x) * (x * x);
double t_1 = (x * x) * t_0;
return fabs(((1.0 / sqrt(((double) M_PI))) * ((fma(2.0, fabs(x), (0.6666666666666666 * t_0)) + (0.2 * t_1)) + (0.047619047619047616 * ((x * x) * t_1)))));
}
function code(x) t_0 = Float64(abs(x) * Float64(x * x)) t_1 = Float64(Float64(x * x) * t_0) return abs(Float64(Float64(1.0 / sqrt(pi)) * Float64(Float64(fma(2.0, abs(x), Float64(0.6666666666666666 * t_0)) + Float64(0.2 * t_1)) + Float64(0.047619047619047616 * Float64(Float64(x * x) * t_1))))) end
code[x_] := Block[{t$95$0 = N[(N[Abs[x], $MachinePrecision] * N[(x * x), $MachinePrecision]), $MachinePrecision]}, Block[{t$95$1 = N[(N[(x * x), $MachinePrecision] * t$95$0), $MachinePrecision]}, N[Abs[N[(N[(1.0 / N[Sqrt[Pi], $MachinePrecision]), $MachinePrecision] * N[(N[(N[(2.0 * N[Abs[x], $MachinePrecision] + N[(0.6666666666666666 * t$95$0), $MachinePrecision]), $MachinePrecision] + N[(0.2 * t$95$1), $MachinePrecision]), $MachinePrecision] + N[(0.047619047619047616 * N[(N[(x * x), $MachinePrecision] * t$95$1), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]]]
\begin{array}{l}
\\
\begin{array}{l}
t_0 := \left|x\right| \cdot \left(x \cdot x\right)\\
t_1 := \left(x \cdot x\right) \cdot t_0\\
\left|\frac{1}{\sqrt{\pi}} \cdot \left(\left(\mathsf{fma}\left(2, \left|x\right|, 0.6666666666666666 \cdot t_0\right) + 0.2 \cdot t_1\right) + 0.047619047619047616 \cdot \left(\left(x \cdot x\right) \cdot t_1\right)\right)\right|
\end{array}
\end{array}
Initial program 99.8%
Simplified99.8%
Final simplification99.8%
(FPCore (x)
:precision binary64
(if (<= (fabs x) 2.0)
(*
x
(/
(fma 0.2 (pow x 4.0) (+ 2.0 (* 0.6666666666666666 (pow x 2.0))))
(sqrt PI)))
(fabs
(*
(sqrt (/ 1.0 PI))
(*
(fabs x)
(fma 0.2 (pow x 4.0) (* 0.047619047619047616 (pow x 6.0))))))))
double code(double x) {
double tmp;
if (fabs(x) <= 2.0) {
tmp = x * (fma(0.2, pow(x, 4.0), (2.0 + (0.6666666666666666 * pow(x, 2.0)))) / sqrt(((double) M_PI)));
} else {
tmp = fabs((sqrt((1.0 / ((double) M_PI))) * (fabs(x) * fma(0.2, pow(x, 4.0), (0.047619047619047616 * pow(x, 6.0))))));
}
return tmp;
}
function code(x) tmp = 0.0 if (abs(x) <= 2.0) tmp = Float64(x * Float64(fma(0.2, (x ^ 4.0), Float64(2.0 + Float64(0.6666666666666666 * (x ^ 2.0)))) / sqrt(pi))); else tmp = abs(Float64(sqrt(Float64(1.0 / pi)) * Float64(abs(x) * fma(0.2, (x ^ 4.0), Float64(0.047619047619047616 * (x ^ 6.0)))))); end return tmp end
code[x_] := If[LessEqual[N[Abs[x], $MachinePrecision], 2.0], N[(x * N[(N[(0.2 * N[Power[x, 4.0], $MachinePrecision] + N[(2.0 + N[(0.6666666666666666 * N[Power[x, 2.0], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] / N[Sqrt[Pi], $MachinePrecision]), $MachinePrecision]), $MachinePrecision], N[Abs[N[(N[Sqrt[N[(1.0 / Pi), $MachinePrecision]], $MachinePrecision] * N[(N[Abs[x], $MachinePrecision] * N[(0.2 * N[Power[x, 4.0], $MachinePrecision] + N[(0.047619047619047616 * N[Power[x, 6.0], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]]
\begin{array}{l}
\\
\begin{array}{l}
\mathbf{if}\;\left|x\right| \leq 2:\\
\;\;\;\;x \cdot \frac{\mathsf{fma}\left(0.2, {x}^{4}, 2 + 0.6666666666666666 \cdot {x}^{2}\right)}{\sqrt{\pi}}\\
\mathbf{else}:\\
\;\;\;\;\left|\sqrt{\frac{1}{\pi}} \cdot \left(\left|x\right| \cdot \mathsf{fma}\left(0.2, {x}^{4}, 0.047619047619047616 \cdot {x}^{6}\right)\right)\right|\\
\end{array}
\end{array}
if (fabs.f64 x) < 2Initial program 99.8%
Simplified99.2%
Taylor expanded in x around 0 98.1%
div-inv98.8%
add-sqr-sqrt48.6%
fabs-sqr48.6%
add-sqr-sqrt50.7%
add-sqr-sqrt50.7%
fabs-sqr50.7%
add-sqr-sqrt50.7%
clear-num50.7%
fma-def50.7%
pow250.7%
Applied egg-rr50.7%
fma-udef50.7%
Applied egg-rr50.7%
if 2 < (fabs.f64 x) Initial program 99.9%
Simplified99.8%
Taylor expanded in x around inf 99.6%
Simplified99.6%
Final simplification66.2%
(FPCore (x)
:precision binary64
(fabs
(*
(/ (fabs x) (sqrt PI))
(+
(+ (* 0.2 (pow x 4.0)) (* 0.047619047619047616 (pow x 6.0)))
(fma 0.6666666666666666 (* x x) 2.0)))))
double code(double x) {
return fabs(((fabs(x) / sqrt(((double) M_PI))) * (((0.2 * pow(x, 4.0)) + (0.047619047619047616 * pow(x, 6.0))) + fma(0.6666666666666666, (x * x), 2.0))));
}
function code(x) return abs(Float64(Float64(abs(x) / sqrt(pi)) * Float64(Float64(Float64(0.2 * (x ^ 4.0)) + Float64(0.047619047619047616 * (x ^ 6.0))) + fma(0.6666666666666666, Float64(x * x), 2.0)))) end
code[x_] := N[Abs[N[(N[(N[Abs[x], $MachinePrecision] / N[Sqrt[Pi], $MachinePrecision]), $MachinePrecision] * N[(N[(N[(0.2 * N[Power[x, 4.0], $MachinePrecision]), $MachinePrecision] + N[(0.047619047619047616 * N[Power[x, 6.0], $MachinePrecision]), $MachinePrecision]), $MachinePrecision] + N[(0.6666666666666666 * N[(x * x), $MachinePrecision] + 2.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]
\begin{array}{l}
\\
\left|\frac{\left|x\right|}{\sqrt{\pi}} \cdot \left(\left(0.2 \cdot {x}^{4} + 0.047619047619047616 \cdot {x}^{6}\right) + \mathsf{fma}\left(0.6666666666666666, x \cdot x, 2\right)\right)\right|
\end{array}
Initial program 99.8%
Simplified99.4%
fma-udef99.4%
Applied egg-rr99.4%
Final simplification99.4%
(FPCore (x)
:precision binary64
(if (<= (fabs x) 2.0)
(*
x
(/
(fma 0.2 (pow x 4.0) (+ 2.0 (* 0.6666666666666666 (pow x 2.0))))
(sqrt PI)))
(/
(fabs x)
(fabs (* (sqrt PI) (+ (/ 21.0 (pow x 6.0)) (/ -88.2 (pow x 8.0))))))))
double code(double x) {
double tmp;
if (fabs(x) <= 2.0) {
tmp = x * (fma(0.2, pow(x, 4.0), (2.0 + (0.6666666666666666 * pow(x, 2.0)))) / sqrt(((double) M_PI)));
} else {
tmp = fabs(x) / fabs((sqrt(((double) M_PI)) * ((21.0 / pow(x, 6.0)) + (-88.2 / pow(x, 8.0)))));
}
return tmp;
}
function code(x) tmp = 0.0 if (abs(x) <= 2.0) tmp = Float64(x * Float64(fma(0.2, (x ^ 4.0), Float64(2.0 + Float64(0.6666666666666666 * (x ^ 2.0)))) / sqrt(pi))); else tmp = Float64(abs(x) / abs(Float64(sqrt(pi) * Float64(Float64(21.0 / (x ^ 6.0)) + Float64(-88.2 / (x ^ 8.0)))))); end return tmp end
code[x_] := If[LessEqual[N[Abs[x], $MachinePrecision], 2.0], N[(x * N[(N[(0.2 * N[Power[x, 4.0], $MachinePrecision] + N[(2.0 + N[(0.6666666666666666 * N[Power[x, 2.0], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] / N[Sqrt[Pi], $MachinePrecision]), $MachinePrecision]), $MachinePrecision], N[(N[Abs[x], $MachinePrecision] / N[Abs[N[(N[Sqrt[Pi], $MachinePrecision] * N[(N[(21.0 / N[Power[x, 6.0], $MachinePrecision]), $MachinePrecision] + N[(-88.2 / N[Power[x, 8.0], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]), $MachinePrecision]]
\begin{array}{l}
\\
\begin{array}{l}
\mathbf{if}\;\left|x\right| \leq 2:\\
\;\;\;\;x \cdot \frac{\mathsf{fma}\left(0.2, {x}^{4}, 2 + 0.6666666666666666 \cdot {x}^{2}\right)}{\sqrt{\pi}}\\
\mathbf{else}:\\
\;\;\;\;\frac{\left|x\right|}{\left|\sqrt{\pi} \cdot \left(\frac{21}{{x}^{6}} + \frac{-88.2}{{x}^{8}}\right)\right|}\\
\end{array}
\end{array}
if (fabs.f64 x) < 2Initial program 99.8%
Simplified99.2%
Taylor expanded in x around 0 98.1%
div-inv98.8%
add-sqr-sqrt48.6%
fabs-sqr48.6%
add-sqr-sqrt50.7%
add-sqr-sqrt50.7%
fabs-sqr50.7%
add-sqr-sqrt50.7%
clear-num50.7%
fma-def50.7%
pow250.7%
Applied egg-rr50.7%
fma-udef50.7%
Applied egg-rr50.7%
if 2 < (fabs.f64 x) Initial program 99.9%
Simplified99.9%
Taylor expanded in x around inf 99.5%
+-commutative99.5%
associate-*r*99.6%
associate-*r*99.6%
distribute-rgt-out99.6%
associate-*r/99.6%
metadata-eval99.6%
associate-*r/99.6%
metadata-eval99.6%
Simplified99.6%
Final simplification66.2%
(FPCore (x)
:precision binary64
(if (<= (fabs x) 2.0)
(*
x
(/
(fma 0.2 (pow x 4.0) (+ 2.0 (* 0.6666666666666666 (pow x 2.0))))
(sqrt PI)))
(* (/ 0.047619047619047616 (sqrt PI)) (/ x (pow x -6.0)))))
double code(double x) {
double tmp;
if (fabs(x) <= 2.0) {
tmp = x * (fma(0.2, pow(x, 4.0), (2.0 + (0.6666666666666666 * pow(x, 2.0)))) / sqrt(((double) M_PI)));
} else {
tmp = (0.047619047619047616 / sqrt(((double) M_PI))) * (x / pow(x, -6.0));
}
return tmp;
}
function code(x) tmp = 0.0 if (abs(x) <= 2.0) tmp = Float64(x * Float64(fma(0.2, (x ^ 4.0), Float64(2.0 + Float64(0.6666666666666666 * (x ^ 2.0)))) / sqrt(pi))); else tmp = Float64(Float64(0.047619047619047616 / sqrt(pi)) * Float64(x / (x ^ -6.0))); end return tmp end
code[x_] := If[LessEqual[N[Abs[x], $MachinePrecision], 2.0], N[(x * N[(N[(0.2 * N[Power[x, 4.0], $MachinePrecision] + N[(2.0 + N[(0.6666666666666666 * N[Power[x, 2.0], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] / N[Sqrt[Pi], $MachinePrecision]), $MachinePrecision]), $MachinePrecision], N[(N[(0.047619047619047616 / N[Sqrt[Pi], $MachinePrecision]), $MachinePrecision] * N[(x / N[Power[x, -6.0], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]]
\begin{array}{l}
\\
\begin{array}{l}
\mathbf{if}\;\left|x\right| \leq 2:\\
\;\;\;\;x \cdot \frac{\mathsf{fma}\left(0.2, {x}^{4}, 2 + 0.6666666666666666 \cdot {x}^{2}\right)}{\sqrt{\pi}}\\
\mathbf{else}:\\
\;\;\;\;\frac{0.047619047619047616}{\sqrt{\pi}} \cdot \frac{x}{{x}^{-6}}\\
\end{array}
\end{array}
if (fabs.f64 x) < 2Initial program 99.8%
Simplified99.2%
Taylor expanded in x around 0 98.1%
div-inv98.8%
add-sqr-sqrt48.6%
fabs-sqr48.6%
add-sqr-sqrt50.7%
add-sqr-sqrt50.7%
fabs-sqr50.7%
add-sqr-sqrt50.7%
clear-num50.7%
fma-def50.7%
pow250.7%
Applied egg-rr50.7%
fma-udef50.7%
Applied egg-rr50.7%
if 2 < (fabs.f64 x) Initial program 99.9%
Simplified99.9%
Taylor expanded in x around inf 99.2%
add-sqr-sqrt0.0%
fabs-sqr0.0%
add-sqr-sqrt0.0%
fabs-sqr0.0%
add-sqr-sqrt0.1%
*-un-lft-identity0.1%
add-sqr-sqrt0.1%
times-frac0.1%
metadata-eval0.1%
*-commutative0.1%
pow-flip0.1%
metadata-eval0.1%
Applied egg-rr0.1%
expm1-log1p-u0.0%
expm1-udef0.0%
associate-*r/0.0%
times-frac0.0%
Applied egg-rr0.0%
expm1-def0.0%
expm1-log1p0.1%
Simplified0.1%
Final simplification34.7%
(FPCore (x)
:precision binary64
(let* ((t_0 (sqrt (/ 1.0 PI))))
(if (<= (fabs x) 2.0)
(+ (* 0.6666666666666666 (* t_0 (pow x 3.0))) (* 2.0 (* x t_0)))
(* (/ 0.047619047619047616 (sqrt PI)) (/ x (pow x -6.0))))))
double code(double x) {
double t_0 = sqrt((1.0 / ((double) M_PI)));
double tmp;
if (fabs(x) <= 2.0) {
tmp = (0.6666666666666666 * (t_0 * pow(x, 3.0))) + (2.0 * (x * t_0));
} else {
tmp = (0.047619047619047616 / sqrt(((double) M_PI))) * (x / pow(x, -6.0));
}
return tmp;
}
public static double code(double x) {
double t_0 = Math.sqrt((1.0 / Math.PI));
double tmp;
if (Math.abs(x) <= 2.0) {
tmp = (0.6666666666666666 * (t_0 * Math.pow(x, 3.0))) + (2.0 * (x * t_0));
} else {
tmp = (0.047619047619047616 / Math.sqrt(Math.PI)) * (x / Math.pow(x, -6.0));
}
return tmp;
}
def code(x): t_0 = math.sqrt((1.0 / math.pi)) tmp = 0 if math.fabs(x) <= 2.0: tmp = (0.6666666666666666 * (t_0 * math.pow(x, 3.0))) + (2.0 * (x * t_0)) else: tmp = (0.047619047619047616 / math.sqrt(math.pi)) * (x / math.pow(x, -6.0)) return tmp
function code(x) t_0 = sqrt(Float64(1.0 / pi)) tmp = 0.0 if (abs(x) <= 2.0) tmp = Float64(Float64(0.6666666666666666 * Float64(t_0 * (x ^ 3.0))) + Float64(2.0 * Float64(x * t_0))); else tmp = Float64(Float64(0.047619047619047616 / sqrt(pi)) * Float64(x / (x ^ -6.0))); end return tmp end
function tmp_2 = code(x) t_0 = sqrt((1.0 / pi)); tmp = 0.0; if (abs(x) <= 2.0) tmp = (0.6666666666666666 * (t_0 * (x ^ 3.0))) + (2.0 * (x * t_0)); else tmp = (0.047619047619047616 / sqrt(pi)) * (x / (x ^ -6.0)); end tmp_2 = tmp; end
code[x_] := Block[{t$95$0 = N[Sqrt[N[(1.0 / Pi), $MachinePrecision]], $MachinePrecision]}, If[LessEqual[N[Abs[x], $MachinePrecision], 2.0], N[(N[(0.6666666666666666 * N[(t$95$0 * N[Power[x, 3.0], $MachinePrecision]), $MachinePrecision]), $MachinePrecision] + N[(2.0 * N[(x * t$95$0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision], N[(N[(0.047619047619047616 / N[Sqrt[Pi], $MachinePrecision]), $MachinePrecision] * N[(x / N[Power[x, -6.0], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]]]
\begin{array}{l}
\\
\begin{array}{l}
t_0 := \sqrt{\frac{1}{\pi}}\\
\mathbf{if}\;\left|x\right| \leq 2:\\
\;\;\;\;0.6666666666666666 \cdot \left(t_0 \cdot {x}^{3}\right) + 2 \cdot \left(x \cdot t_0\right)\\
\mathbf{else}:\\
\;\;\;\;\frac{0.047619047619047616}{\sqrt{\pi}} \cdot \frac{x}{{x}^{-6}}\\
\end{array}
\end{array}
if (fabs.f64 x) < 2Initial program 99.8%
Simplified99.2%
Taylor expanded in x around 0 98.1%
div-inv98.8%
add-sqr-sqrt48.6%
fabs-sqr48.6%
add-sqr-sqrt50.7%
add-sqr-sqrt50.7%
fabs-sqr50.7%
add-sqr-sqrt50.7%
clear-num50.7%
fma-def50.7%
pow250.7%
Applied egg-rr50.7%
Taylor expanded in x around 0 50.6%
if 2 < (fabs.f64 x) Initial program 99.9%
Simplified99.9%
Taylor expanded in x around inf 99.2%
add-sqr-sqrt0.0%
fabs-sqr0.0%
add-sqr-sqrt0.0%
fabs-sqr0.0%
add-sqr-sqrt0.1%
*-un-lft-identity0.1%
add-sqr-sqrt0.1%
times-frac0.1%
metadata-eval0.1%
*-commutative0.1%
pow-flip0.1%
metadata-eval0.1%
Applied egg-rr0.1%
expm1-log1p-u0.0%
expm1-udef0.0%
associate-*r/0.0%
times-frac0.0%
Applied egg-rr0.0%
expm1-def0.0%
expm1-log1p0.1%
Simplified0.1%
Final simplification34.7%
(FPCore (x) :precision binary64 (if (<= (fabs x) 2.0) (* (sqrt (/ 1.0 PI)) (+ (* 2.0 x) (* 0.6666666666666666 (pow x 3.0)))) (* (/ 0.047619047619047616 (sqrt PI)) (/ x (pow x -6.0)))))
double code(double x) {
double tmp;
if (fabs(x) <= 2.0) {
tmp = sqrt((1.0 / ((double) M_PI))) * ((2.0 * x) + (0.6666666666666666 * pow(x, 3.0)));
} else {
tmp = (0.047619047619047616 / sqrt(((double) M_PI))) * (x / pow(x, -6.0));
}
return tmp;
}
public static double code(double x) {
double tmp;
if (Math.abs(x) <= 2.0) {
tmp = Math.sqrt((1.0 / Math.PI)) * ((2.0 * x) + (0.6666666666666666 * Math.pow(x, 3.0)));
} else {
tmp = (0.047619047619047616 / Math.sqrt(Math.PI)) * (x / Math.pow(x, -6.0));
}
return tmp;
}
def code(x): tmp = 0 if math.fabs(x) <= 2.0: tmp = math.sqrt((1.0 / math.pi)) * ((2.0 * x) + (0.6666666666666666 * math.pow(x, 3.0))) else: tmp = (0.047619047619047616 / math.sqrt(math.pi)) * (x / math.pow(x, -6.0)) return tmp
function code(x) tmp = 0.0 if (abs(x) <= 2.0) tmp = Float64(sqrt(Float64(1.0 / pi)) * Float64(Float64(2.0 * x) + Float64(0.6666666666666666 * (x ^ 3.0)))); else tmp = Float64(Float64(0.047619047619047616 / sqrt(pi)) * Float64(x / (x ^ -6.0))); end return tmp end
function tmp_2 = code(x) tmp = 0.0; if (abs(x) <= 2.0) tmp = sqrt((1.0 / pi)) * ((2.0 * x) + (0.6666666666666666 * (x ^ 3.0))); else tmp = (0.047619047619047616 / sqrt(pi)) * (x / (x ^ -6.0)); end tmp_2 = tmp; end
code[x_] := If[LessEqual[N[Abs[x], $MachinePrecision], 2.0], N[(N[Sqrt[N[(1.0 / Pi), $MachinePrecision]], $MachinePrecision] * N[(N[(2.0 * x), $MachinePrecision] + N[(0.6666666666666666 * N[Power[x, 3.0], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision], N[(N[(0.047619047619047616 / N[Sqrt[Pi], $MachinePrecision]), $MachinePrecision] * N[(x / N[Power[x, -6.0], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]]
\begin{array}{l}
\\
\begin{array}{l}
\mathbf{if}\;\left|x\right| \leq 2:\\
\;\;\;\;\sqrt{\frac{1}{\pi}} \cdot \left(2 \cdot x + 0.6666666666666666 \cdot {x}^{3}\right)\\
\mathbf{else}:\\
\;\;\;\;\frac{0.047619047619047616}{\sqrt{\pi}} \cdot \frac{x}{{x}^{-6}}\\
\end{array}
\end{array}
if (fabs.f64 x) < 2Initial program 99.8%
Simplified99.2%
Taylor expanded in x around 0 98.1%
div-inv98.8%
add-sqr-sqrt48.6%
fabs-sqr48.6%
add-sqr-sqrt50.7%
add-sqr-sqrt50.7%
fabs-sqr50.7%
add-sqr-sqrt50.7%
clear-num50.7%
fma-def50.7%
pow250.7%
Applied egg-rr50.7%
Taylor expanded in x around 0 50.6%
+-commutative50.6%
associate-*r*50.6%
*-commutative50.6%
associate-*r*50.6%
distribute-rgt-out50.6%
*-commutative50.6%
Simplified50.6%
if 2 < (fabs.f64 x) Initial program 99.9%
Simplified99.9%
Taylor expanded in x around inf 99.2%
add-sqr-sqrt0.0%
fabs-sqr0.0%
add-sqr-sqrt0.0%
fabs-sqr0.0%
add-sqr-sqrt0.1%
*-un-lft-identity0.1%
add-sqr-sqrt0.1%
times-frac0.1%
metadata-eval0.1%
*-commutative0.1%
pow-flip0.1%
metadata-eval0.1%
Applied egg-rr0.1%
expm1-log1p-u0.0%
expm1-udef0.0%
associate-*r/0.0%
times-frac0.0%
Applied egg-rr0.0%
expm1-def0.0%
expm1-log1p0.1%
Simplified0.1%
Final simplification34.7%
(FPCore (x) :precision binary64 (if (<= x 1.85) (* x (/ 2.0 (sqrt PI))) (* 0.047619047619047616 (/ x (* (sqrt PI) (pow x -6.0))))))
double code(double x) {
double tmp;
if (x <= 1.85) {
tmp = x * (2.0 / sqrt(((double) M_PI)));
} else {
tmp = 0.047619047619047616 * (x / (sqrt(((double) M_PI)) * pow(x, -6.0)));
}
return tmp;
}
public static double code(double x) {
double tmp;
if (x <= 1.85) {
tmp = x * (2.0 / Math.sqrt(Math.PI));
} else {
tmp = 0.047619047619047616 * (x / (Math.sqrt(Math.PI) * Math.pow(x, -6.0)));
}
return tmp;
}
def code(x): tmp = 0 if x <= 1.85: tmp = x * (2.0 / math.sqrt(math.pi)) else: tmp = 0.047619047619047616 * (x / (math.sqrt(math.pi) * math.pow(x, -6.0))) return tmp
function code(x) tmp = 0.0 if (x <= 1.85) tmp = Float64(x * Float64(2.0 / sqrt(pi))); else tmp = Float64(0.047619047619047616 * Float64(x / Float64(sqrt(pi) * (x ^ -6.0)))); end return tmp end
function tmp_2 = code(x) tmp = 0.0; if (x <= 1.85) tmp = x * (2.0 / sqrt(pi)); else tmp = 0.047619047619047616 * (x / (sqrt(pi) * (x ^ -6.0))); end tmp_2 = tmp; end
code[x_] := If[LessEqual[x, 1.85], N[(x * N[(2.0 / N[Sqrt[Pi], $MachinePrecision]), $MachinePrecision]), $MachinePrecision], N[(0.047619047619047616 * N[(x / N[(N[Sqrt[Pi], $MachinePrecision] * N[Power[x, -6.0], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]]
\begin{array}{l}
\\
\begin{array}{l}
\mathbf{if}\;x \leq 1.85:\\
\;\;\;\;x \cdot \frac{2}{\sqrt{\pi}}\\
\mathbf{else}:\\
\;\;\;\;0.047619047619047616 \cdot \frac{x}{\sqrt{\pi} \cdot {x}^{-6}}\\
\end{array}
\end{array}
if x < 1.8500000000000001Initial program 99.8%
Simplified99.4%
Taylor expanded in x around 0 68.6%
*-commutative68.6%
Simplified68.6%
Taylor expanded in x around 0 68.6%
fabs-neg68.6%
*-commutative68.6%
fabs-div68.6%
neg-mul-168.6%
*-commutative68.6%
times-frac68.6%
metadata-eval68.6%
metadata-eval68.6%
distribute-lft-neg-in68.6%
fabs-neg68.6%
rem-square-sqrt32.8%
fabs-sqr32.8%
rem-square-sqrt34.3%
*-commutative34.3%
metadata-eval34.3%
times-frac34.3%
*-rgt-identity34.3%
*-commutative34.3%
associate-/r*34.3%
Simplified34.3%
associate-/l/34.3%
*-un-lft-identity34.3%
*-commutative34.3%
times-frac34.3%
metadata-eval34.3%
rem-log-exp4.3%
*-un-lft-identity4.3%
log-prod4.3%
metadata-eval4.3%
rem-log-exp34.3%
metadata-eval34.3%
times-frac34.3%
*-un-lft-identity34.3%
*-commutative34.3%
Applied egg-rr34.3%
+-lft-identity34.3%
*-rgt-identity34.3%
times-frac34.3%
metadata-eval34.3%
associate-*l/34.3%
associate-*r/34.5%
Simplified34.5%
if 1.8500000000000001 < x Initial program 99.8%
Simplified99.4%
Taylor expanded in x around inf 35.3%
add-sqr-sqrt1.9%
fabs-sqr1.9%
add-sqr-sqrt1.9%
fabs-sqr1.9%
add-sqr-sqrt3.8%
*-un-lft-identity3.8%
add-sqr-sqrt3.8%
times-frac3.8%
metadata-eval3.8%
*-commutative3.8%
pow-flip3.8%
metadata-eval3.8%
Applied egg-rr3.8%
Final simplification34.5%
(FPCore (x) :precision binary64 (if (<= x 1.85) (* x (/ 2.0 (sqrt PI))) (* (/ 0.047619047619047616 (sqrt PI)) (/ x (pow x -6.0)))))
double code(double x) {
double tmp;
if (x <= 1.85) {
tmp = x * (2.0 / sqrt(((double) M_PI)));
} else {
tmp = (0.047619047619047616 / sqrt(((double) M_PI))) * (x / pow(x, -6.0));
}
return tmp;
}
public static double code(double x) {
double tmp;
if (x <= 1.85) {
tmp = x * (2.0 / Math.sqrt(Math.PI));
} else {
tmp = (0.047619047619047616 / Math.sqrt(Math.PI)) * (x / Math.pow(x, -6.0));
}
return tmp;
}
def code(x): tmp = 0 if x <= 1.85: tmp = x * (2.0 / math.sqrt(math.pi)) else: tmp = (0.047619047619047616 / math.sqrt(math.pi)) * (x / math.pow(x, -6.0)) return tmp
function code(x) tmp = 0.0 if (x <= 1.85) tmp = Float64(x * Float64(2.0 / sqrt(pi))); else tmp = Float64(Float64(0.047619047619047616 / sqrt(pi)) * Float64(x / (x ^ -6.0))); end return tmp end
function tmp_2 = code(x) tmp = 0.0; if (x <= 1.85) tmp = x * (2.0 / sqrt(pi)); else tmp = (0.047619047619047616 / sqrt(pi)) * (x / (x ^ -6.0)); end tmp_2 = tmp; end
code[x_] := If[LessEqual[x, 1.85], N[(x * N[(2.0 / N[Sqrt[Pi], $MachinePrecision]), $MachinePrecision]), $MachinePrecision], N[(N[(0.047619047619047616 / N[Sqrt[Pi], $MachinePrecision]), $MachinePrecision] * N[(x / N[Power[x, -6.0], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]]
\begin{array}{l}
\\
\begin{array}{l}
\mathbf{if}\;x \leq 1.85:\\
\;\;\;\;x \cdot \frac{2}{\sqrt{\pi}}\\
\mathbf{else}:\\
\;\;\;\;\frac{0.047619047619047616}{\sqrt{\pi}} \cdot \frac{x}{{x}^{-6}}\\
\end{array}
\end{array}
if x < 1.8500000000000001Initial program 99.8%
Simplified99.4%
Taylor expanded in x around 0 68.6%
*-commutative68.6%
Simplified68.6%
Taylor expanded in x around 0 68.6%
fabs-neg68.6%
*-commutative68.6%
fabs-div68.6%
neg-mul-168.6%
*-commutative68.6%
times-frac68.6%
metadata-eval68.6%
metadata-eval68.6%
distribute-lft-neg-in68.6%
fabs-neg68.6%
rem-square-sqrt32.8%
fabs-sqr32.8%
rem-square-sqrt34.3%
*-commutative34.3%
metadata-eval34.3%
times-frac34.3%
*-rgt-identity34.3%
*-commutative34.3%
associate-/r*34.3%
Simplified34.3%
associate-/l/34.3%
*-un-lft-identity34.3%
*-commutative34.3%
times-frac34.3%
metadata-eval34.3%
rem-log-exp4.3%
*-un-lft-identity4.3%
log-prod4.3%
metadata-eval4.3%
rem-log-exp34.3%
metadata-eval34.3%
times-frac34.3%
*-un-lft-identity34.3%
*-commutative34.3%
Applied egg-rr34.3%
+-lft-identity34.3%
*-rgt-identity34.3%
times-frac34.3%
metadata-eval34.3%
associate-*l/34.3%
associate-*r/34.5%
Simplified34.5%
if 1.8500000000000001 < x Initial program 99.8%
Simplified99.4%
Taylor expanded in x around inf 35.3%
add-sqr-sqrt1.9%
fabs-sqr1.9%
add-sqr-sqrt1.9%
fabs-sqr1.9%
add-sqr-sqrt3.8%
*-un-lft-identity3.8%
add-sqr-sqrt3.8%
times-frac3.8%
metadata-eval3.8%
*-commutative3.8%
pow-flip3.8%
metadata-eval3.8%
Applied egg-rr3.8%
expm1-log1p-u3.7%
expm1-udef3.7%
associate-*r/3.7%
times-frac3.7%
Applied egg-rr3.7%
expm1-def3.7%
expm1-log1p3.8%
Simplified3.8%
Final simplification34.5%
(FPCore (x) :precision binary64 (if (<= x 2e-81) (* x (/ 2.0 (sqrt PI))) (sqrt (* x (* x (/ 4.0 PI))))))
double code(double x) {
double tmp;
if (x <= 2e-81) {
tmp = x * (2.0 / sqrt(((double) M_PI)));
} else {
tmp = sqrt((x * (x * (4.0 / ((double) M_PI)))));
}
return tmp;
}
public static double code(double x) {
double tmp;
if (x <= 2e-81) {
tmp = x * (2.0 / Math.sqrt(Math.PI));
} else {
tmp = Math.sqrt((x * (x * (4.0 / Math.PI))));
}
return tmp;
}
def code(x): tmp = 0 if x <= 2e-81: tmp = x * (2.0 / math.sqrt(math.pi)) else: tmp = math.sqrt((x * (x * (4.0 / math.pi)))) return tmp
function code(x) tmp = 0.0 if (x <= 2e-81) tmp = Float64(x * Float64(2.0 / sqrt(pi))); else tmp = sqrt(Float64(x * Float64(x * Float64(4.0 / pi)))); end return tmp end
function tmp_2 = code(x) tmp = 0.0; if (x <= 2e-81) tmp = x * (2.0 / sqrt(pi)); else tmp = sqrt((x * (x * (4.0 / pi)))); end tmp_2 = tmp; end
code[x_] := If[LessEqual[x, 2e-81], N[(x * N[(2.0 / N[Sqrt[Pi], $MachinePrecision]), $MachinePrecision]), $MachinePrecision], N[Sqrt[N[(x * N[(x * N[(4.0 / Pi), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]]
\begin{array}{l}
\\
\begin{array}{l}
\mathbf{if}\;x \leq 2 \cdot 10^{-81}:\\
\;\;\;\;x \cdot \frac{2}{\sqrt{\pi}}\\
\mathbf{else}:\\
\;\;\;\;\sqrt{x \cdot \left(x \cdot \frac{4}{\pi}\right)}\\
\end{array}
\end{array}
if x < 1.9999999999999999e-81Initial program 99.9%
Simplified99.4%
Taylor expanded in x around 0 66.1%
*-commutative66.1%
Simplified66.1%
Taylor expanded in x around 0 66.1%
fabs-neg66.1%
*-commutative66.1%
fabs-div66.1%
neg-mul-166.1%
*-commutative66.1%
times-frac66.1%
metadata-eval66.1%
metadata-eval66.1%
distribute-lft-neg-in66.1%
fabs-neg66.1%
rem-square-sqrt26.7%
fabs-sqr26.7%
rem-square-sqrt28.4%
*-commutative28.4%
metadata-eval28.4%
times-frac28.4%
*-rgt-identity28.4%
*-commutative28.4%
associate-/r*28.4%
Simplified28.4%
associate-/l/28.4%
*-un-lft-identity28.4%
*-commutative28.4%
times-frac28.4%
metadata-eval28.4%
rem-log-exp3.7%
*-un-lft-identity3.7%
log-prod3.7%
metadata-eval3.7%
rem-log-exp28.4%
metadata-eval28.4%
times-frac28.4%
*-un-lft-identity28.4%
*-commutative28.4%
Applied egg-rr28.4%
+-lft-identity28.4%
*-rgt-identity28.4%
times-frac28.4%
metadata-eval28.4%
associate-*l/28.4%
associate-*r/28.6%
Simplified28.6%
if 1.9999999999999999e-81 < x Initial program 99.7%
Simplified99.3%
Taylor expanded in x around 0 94.4%
*-commutative94.4%
Simplified94.4%
add-sqr-sqrt94.3%
sqrt-unprod94.4%
div-fabs94.4%
div-fabs94.4%
sqr-abs94.4%
frac-times94.3%
pow294.3%
swap-sqr94.3%
add-sqr-sqrt94.6%
metadata-eval94.6%
Applied egg-rr94.6%
div-inv94.6%
unpow294.6%
associate-*l*94.8%
*-commutative94.8%
associate-/r*94.8%
metadata-eval94.8%
Applied egg-rr94.8%
Final simplification34.5%
(FPCore (x) :precision binary64 (* x (/ 2.0 (sqrt PI))))
double code(double x) {
return x * (2.0 / sqrt(((double) M_PI)));
}
public static double code(double x) {
return x * (2.0 / Math.sqrt(Math.PI));
}
def code(x): return x * (2.0 / math.sqrt(math.pi))
function code(x) return Float64(x * Float64(2.0 / sqrt(pi))) end
function tmp = code(x) tmp = x * (2.0 / sqrt(pi)); end
code[x_] := N[(x * N[(2.0 / N[Sqrt[Pi], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
x \cdot \frac{2}{\sqrt{\pi}}
\end{array}
Initial program 99.8%
Simplified99.4%
Taylor expanded in x around 0 68.6%
*-commutative68.6%
Simplified68.6%
Taylor expanded in x around 0 68.6%
fabs-neg68.6%
*-commutative68.6%
fabs-div68.6%
neg-mul-168.6%
*-commutative68.6%
times-frac68.6%
metadata-eval68.6%
metadata-eval68.6%
distribute-lft-neg-in68.6%
fabs-neg68.6%
rem-square-sqrt32.8%
fabs-sqr32.8%
rem-square-sqrt34.3%
*-commutative34.3%
metadata-eval34.3%
times-frac34.3%
*-rgt-identity34.3%
*-commutative34.3%
associate-/r*34.3%
Simplified34.3%
associate-/l/34.3%
*-un-lft-identity34.3%
*-commutative34.3%
times-frac34.3%
metadata-eval34.3%
rem-log-exp4.3%
*-un-lft-identity4.3%
log-prod4.3%
metadata-eval4.3%
rem-log-exp34.3%
metadata-eval34.3%
times-frac34.3%
*-un-lft-identity34.3%
*-commutative34.3%
Applied egg-rr34.3%
+-lft-identity34.3%
*-rgt-identity34.3%
times-frac34.3%
metadata-eval34.3%
associate-*l/34.3%
associate-*r/34.5%
Simplified34.5%
Final simplification34.5%
herbie shell --seed 2024021
(FPCore (x)
:name "Jmat.Real.erfi, branch x less than or equal to 0.5"
:precision binary64
:pre (<= x 0.5)
(fabs (* (/ 1.0 (sqrt PI)) (+ (+ (+ (* 2.0 (fabs x)) (* (/ 2.0 3.0) (* (* (fabs x) (fabs x)) (fabs x)))) (* (/ 1.0 5.0) (* (* (* (* (fabs x) (fabs x)) (fabs x)) (fabs x)) (fabs x)))) (* (/ 1.0 21.0) (* (* (* (* (* (* (fabs x) (fabs x)) (fabs x)) (fabs x)) (fabs x)) (fabs x)) (fabs x)))))))