
(FPCore (x)
:precision binary64
(let* ((t_0 (/ 1.0 (fabs x)))
(t_1 (* (* t_0 t_0) t_0))
(t_2 (* (* t_1 t_0) t_0)))
(*
(* (/ 1.0 (sqrt PI)) (exp (* (fabs x) (fabs x))))
(+
(+ (+ t_0 (* (/ 1.0 2.0) t_1)) (* (/ 3.0 4.0) t_2))
(* (/ 15.0 8.0) (* (* t_2 t_0) t_0))))))
double code(double x) {
double t_0 = 1.0 / fabs(x);
double t_1 = (t_0 * t_0) * t_0;
double t_2 = (t_1 * t_0) * t_0;
return ((1.0 / sqrt(((double) M_PI))) * exp((fabs(x) * fabs(x)))) * (((t_0 + ((1.0 / 2.0) * t_1)) + ((3.0 / 4.0) * t_2)) + ((15.0 / 8.0) * ((t_2 * t_0) * t_0)));
}
public static double code(double x) {
double t_0 = 1.0 / Math.abs(x);
double t_1 = (t_0 * t_0) * t_0;
double t_2 = (t_1 * t_0) * t_0;
return ((1.0 / Math.sqrt(Math.PI)) * Math.exp((Math.abs(x) * Math.abs(x)))) * (((t_0 + ((1.0 / 2.0) * t_1)) + ((3.0 / 4.0) * t_2)) + ((15.0 / 8.0) * ((t_2 * t_0) * t_0)));
}
def code(x): t_0 = 1.0 / math.fabs(x) t_1 = (t_0 * t_0) * t_0 t_2 = (t_1 * t_0) * t_0 return ((1.0 / math.sqrt(math.pi)) * math.exp((math.fabs(x) * math.fabs(x)))) * (((t_0 + ((1.0 / 2.0) * t_1)) + ((3.0 / 4.0) * t_2)) + ((15.0 / 8.0) * ((t_2 * t_0) * t_0)))
function code(x) t_0 = Float64(1.0 / abs(x)) t_1 = Float64(Float64(t_0 * t_0) * t_0) t_2 = Float64(Float64(t_1 * t_0) * t_0) return Float64(Float64(Float64(1.0 / sqrt(pi)) * exp(Float64(abs(x) * abs(x)))) * Float64(Float64(Float64(t_0 + Float64(Float64(1.0 / 2.0) * t_1)) + Float64(Float64(3.0 / 4.0) * t_2)) + Float64(Float64(15.0 / 8.0) * Float64(Float64(t_2 * t_0) * t_0)))) end
function tmp = code(x) t_0 = 1.0 / abs(x); t_1 = (t_0 * t_0) * t_0; t_2 = (t_1 * t_0) * t_0; tmp = ((1.0 / sqrt(pi)) * exp((abs(x) * abs(x)))) * (((t_0 + ((1.0 / 2.0) * t_1)) + ((3.0 / 4.0) * t_2)) + ((15.0 / 8.0) * ((t_2 * t_0) * t_0))); end
code[x_] := Block[{t$95$0 = N[(1.0 / N[Abs[x], $MachinePrecision]), $MachinePrecision]}, Block[{t$95$1 = N[(N[(t$95$0 * t$95$0), $MachinePrecision] * t$95$0), $MachinePrecision]}, Block[{t$95$2 = N[(N[(t$95$1 * t$95$0), $MachinePrecision] * t$95$0), $MachinePrecision]}, N[(N[(N[(1.0 / N[Sqrt[Pi], $MachinePrecision]), $MachinePrecision] * N[Exp[N[(N[Abs[x], $MachinePrecision] * N[Abs[x], $MachinePrecision]), $MachinePrecision]], $MachinePrecision]), $MachinePrecision] * N[(N[(N[(t$95$0 + N[(N[(1.0 / 2.0), $MachinePrecision] * t$95$1), $MachinePrecision]), $MachinePrecision] + N[(N[(3.0 / 4.0), $MachinePrecision] * t$95$2), $MachinePrecision]), $MachinePrecision] + N[(N[(15.0 / 8.0), $MachinePrecision] * N[(N[(t$95$2 * t$95$0), $MachinePrecision] * t$95$0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]]]]
\begin{array}{l}
\\
\begin{array}{l}
t_0 := \frac{1}{\left|x\right|}\\
t_1 := \left(t\_0 \cdot t\_0\right) \cdot t\_0\\
t_2 := \left(t\_1 \cdot t\_0\right) \cdot t\_0\\
\left(\frac{1}{\sqrt{\pi}} \cdot e^{\left|x\right| \cdot \left|x\right|}\right) \cdot \left(\left(\left(t\_0 + \frac{1}{2} \cdot t\_1\right) + \frac{3}{4} \cdot t\_2\right) + \frac{15}{8} \cdot \left(\left(t\_2 \cdot t\_0\right) \cdot t\_0\right)\right)
\end{array}
\end{array}
Sampling outcomes in binary64 precision:
Herbie found 10 alternatives:
| Alternative | Accuracy | Speedup |
|---|
(FPCore (x)
:precision binary64
(let* ((t_0 (/ 1.0 (fabs x)))
(t_1 (* (* t_0 t_0) t_0))
(t_2 (* (* t_1 t_0) t_0)))
(*
(* (/ 1.0 (sqrt PI)) (exp (* (fabs x) (fabs x))))
(+
(+ (+ t_0 (* (/ 1.0 2.0) t_1)) (* (/ 3.0 4.0) t_2))
(* (/ 15.0 8.0) (* (* t_2 t_0) t_0))))))
double code(double x) {
double t_0 = 1.0 / fabs(x);
double t_1 = (t_0 * t_0) * t_0;
double t_2 = (t_1 * t_0) * t_0;
return ((1.0 / sqrt(((double) M_PI))) * exp((fabs(x) * fabs(x)))) * (((t_0 + ((1.0 / 2.0) * t_1)) + ((3.0 / 4.0) * t_2)) + ((15.0 / 8.0) * ((t_2 * t_0) * t_0)));
}
public static double code(double x) {
double t_0 = 1.0 / Math.abs(x);
double t_1 = (t_0 * t_0) * t_0;
double t_2 = (t_1 * t_0) * t_0;
return ((1.0 / Math.sqrt(Math.PI)) * Math.exp((Math.abs(x) * Math.abs(x)))) * (((t_0 + ((1.0 / 2.0) * t_1)) + ((3.0 / 4.0) * t_2)) + ((15.0 / 8.0) * ((t_2 * t_0) * t_0)));
}
def code(x): t_0 = 1.0 / math.fabs(x) t_1 = (t_0 * t_0) * t_0 t_2 = (t_1 * t_0) * t_0 return ((1.0 / math.sqrt(math.pi)) * math.exp((math.fabs(x) * math.fabs(x)))) * (((t_0 + ((1.0 / 2.0) * t_1)) + ((3.0 / 4.0) * t_2)) + ((15.0 / 8.0) * ((t_2 * t_0) * t_0)))
function code(x) t_0 = Float64(1.0 / abs(x)) t_1 = Float64(Float64(t_0 * t_0) * t_0) t_2 = Float64(Float64(t_1 * t_0) * t_0) return Float64(Float64(Float64(1.0 / sqrt(pi)) * exp(Float64(abs(x) * abs(x)))) * Float64(Float64(Float64(t_0 + Float64(Float64(1.0 / 2.0) * t_1)) + Float64(Float64(3.0 / 4.0) * t_2)) + Float64(Float64(15.0 / 8.0) * Float64(Float64(t_2 * t_0) * t_0)))) end
function tmp = code(x) t_0 = 1.0 / abs(x); t_1 = (t_0 * t_0) * t_0; t_2 = (t_1 * t_0) * t_0; tmp = ((1.0 / sqrt(pi)) * exp((abs(x) * abs(x)))) * (((t_0 + ((1.0 / 2.0) * t_1)) + ((3.0 / 4.0) * t_2)) + ((15.0 / 8.0) * ((t_2 * t_0) * t_0))); end
code[x_] := Block[{t$95$0 = N[(1.0 / N[Abs[x], $MachinePrecision]), $MachinePrecision]}, Block[{t$95$1 = N[(N[(t$95$0 * t$95$0), $MachinePrecision] * t$95$0), $MachinePrecision]}, Block[{t$95$2 = N[(N[(t$95$1 * t$95$0), $MachinePrecision] * t$95$0), $MachinePrecision]}, N[(N[(N[(1.0 / N[Sqrt[Pi], $MachinePrecision]), $MachinePrecision] * N[Exp[N[(N[Abs[x], $MachinePrecision] * N[Abs[x], $MachinePrecision]), $MachinePrecision]], $MachinePrecision]), $MachinePrecision] * N[(N[(N[(t$95$0 + N[(N[(1.0 / 2.0), $MachinePrecision] * t$95$1), $MachinePrecision]), $MachinePrecision] + N[(N[(3.0 / 4.0), $MachinePrecision] * t$95$2), $MachinePrecision]), $MachinePrecision] + N[(N[(15.0 / 8.0), $MachinePrecision] * N[(N[(t$95$2 * t$95$0), $MachinePrecision] * t$95$0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]]]]
\begin{array}{l}
\\
\begin{array}{l}
t_0 := \frac{1}{\left|x\right|}\\
t_1 := \left(t\_0 \cdot t\_0\right) \cdot t\_0\\
t_2 := \left(t\_1 \cdot t\_0\right) \cdot t\_0\\
\left(\frac{1}{\sqrt{\pi}} \cdot e^{\left|x\right| \cdot \left|x\right|}\right) \cdot \left(\left(\left(t\_0 + \frac{1}{2} \cdot t\_1\right) + \frac{3}{4} \cdot t\_2\right) + \frac{15}{8} \cdot \left(\left(t\_2 \cdot t\_0\right) \cdot t\_0\right)\right)
\end{array}
\end{array}
(FPCore (x)
:precision binary64
(let* ((t_0 (/ 1.0 (fabs x))) (t_1 (pow t_0 3.0)))
(*
(/ (pow (exp x) x) (expm1 (log1p (sqrt PI))))
(fma
1.875
(* t_1 (pow x -4.0))
(fma 0.75 (/ (pow x -4.0) (fabs x)) (fma 0.5 t_1 t_0))))))
double code(double x) {
double t_0 = 1.0 / fabs(x);
double t_1 = pow(t_0, 3.0);
return (pow(exp(x), x) / expm1(log1p(sqrt(((double) M_PI))))) * fma(1.875, (t_1 * pow(x, -4.0)), fma(0.75, (pow(x, -4.0) / fabs(x)), fma(0.5, t_1, t_0)));
}
function code(x) t_0 = Float64(1.0 / abs(x)) t_1 = t_0 ^ 3.0 return Float64(Float64((exp(x) ^ x) / expm1(log1p(sqrt(pi)))) * fma(1.875, Float64(t_1 * (x ^ -4.0)), fma(0.75, Float64((x ^ -4.0) / abs(x)), fma(0.5, t_1, t_0)))) end
code[x_] := Block[{t$95$0 = N[(1.0 / N[Abs[x], $MachinePrecision]), $MachinePrecision]}, Block[{t$95$1 = N[Power[t$95$0, 3.0], $MachinePrecision]}, N[(N[(N[Power[N[Exp[x], $MachinePrecision], x], $MachinePrecision] / N[(Exp[N[Log[1 + N[Sqrt[Pi], $MachinePrecision]], $MachinePrecision]] - 1), $MachinePrecision]), $MachinePrecision] * N[(1.875 * N[(t$95$1 * N[Power[x, -4.0], $MachinePrecision]), $MachinePrecision] + N[(0.75 * N[(N[Power[x, -4.0], $MachinePrecision] / N[Abs[x], $MachinePrecision]), $MachinePrecision] + N[(0.5 * t$95$1 + t$95$0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]]]
\begin{array}{l}
\\
\begin{array}{l}
t_0 := \frac{1}{\left|x\right|}\\
t_1 := {t\_0}^{3}\\
\frac{{\left(e^{x}\right)}^{x}}{\mathsf{expm1}\left(\mathsf{log1p}\left(\sqrt{\pi}\right)\right)} \cdot \mathsf{fma}\left(1.875, t\_1 \cdot {x}^{-4}, \mathsf{fma}\left(0.75, \frac{{x}^{-4}}{\left|x\right|}, \mathsf{fma}\left(0.5, t\_1, t\_0\right)\right)\right)
\end{array}
\end{array}
Initial program 100.0%
Simplified100.0%
expm1-log1p-u100.0%
Applied egg-rr100.0%
expm1-log1p-u100.0%
expm1-udef100.0%
Applied egg-rr100.0%
expm1-def100.0%
expm1-log1p100.0%
unpow2100.0%
pow-sqr100.0%
metadata-eval100.0%
Simplified100.0%
expm1-log1p-u100.0%
expm1-udef100.0%
Applied egg-rr100.0%
expm1-def100.0%
expm1-log1p100.0%
unpow2100.0%
pow-sqr100.0%
metadata-eval100.0%
Simplified100.0%
Final simplification100.0%
(FPCore (x)
:precision binary64
(*
(pow (exp x) x)
(*
(sqrt (/ 1.0 PI))
(+
(/ 0.5 (pow x 3.0))
(+
(/ 0.75 (pow x 5.0))
(+ (expm1 (log1p (* 1.875 (pow x -7.0)))) (/ 1.0 x)))))))
double code(double x) {
return pow(exp(x), x) * (sqrt((1.0 / ((double) M_PI))) * ((0.5 / pow(x, 3.0)) + ((0.75 / pow(x, 5.0)) + (expm1(log1p((1.875 * pow(x, -7.0)))) + (1.0 / x)))));
}
public static double code(double x) {
return Math.pow(Math.exp(x), x) * (Math.sqrt((1.0 / Math.PI)) * ((0.5 / Math.pow(x, 3.0)) + ((0.75 / Math.pow(x, 5.0)) + (Math.expm1(Math.log1p((1.875 * Math.pow(x, -7.0)))) + (1.0 / x)))));
}
def code(x): return math.pow(math.exp(x), x) * (math.sqrt((1.0 / math.pi)) * ((0.5 / math.pow(x, 3.0)) + ((0.75 / math.pow(x, 5.0)) + (math.expm1(math.log1p((1.875 * math.pow(x, -7.0)))) + (1.0 / x)))))
function code(x) return Float64((exp(x) ^ x) * Float64(sqrt(Float64(1.0 / pi)) * Float64(Float64(0.5 / (x ^ 3.0)) + Float64(Float64(0.75 / (x ^ 5.0)) + Float64(expm1(log1p(Float64(1.875 * (x ^ -7.0)))) + Float64(1.0 / x)))))) end
code[x_] := N[(N[Power[N[Exp[x], $MachinePrecision], x], $MachinePrecision] * N[(N[Sqrt[N[(1.0 / Pi), $MachinePrecision]], $MachinePrecision] * N[(N[(0.5 / N[Power[x, 3.0], $MachinePrecision]), $MachinePrecision] + N[(N[(0.75 / N[Power[x, 5.0], $MachinePrecision]), $MachinePrecision] + N[(N[(Exp[N[Log[1 + N[(1.875 * N[Power[x, -7.0], $MachinePrecision]), $MachinePrecision]], $MachinePrecision]] - 1), $MachinePrecision] + N[(1.0 / x), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
{\left(e^{x}\right)}^{x} \cdot \left(\sqrt{\frac{1}{\pi}} \cdot \left(\frac{0.5}{{x}^{3}} + \left(\frac{0.75}{{x}^{5}} + \left(\mathsf{expm1}\left(\mathsf{log1p}\left(1.875 \cdot {x}^{-7}\right)\right) + \frac{1}{x}\right)\right)\right)\right)
\end{array}
Initial program 100.0%
Simplified100.0%
Taylor expanded in x around 0 100.0%
associate-*r*100.0%
*-commutative100.0%
distribute-rgt-out100.0%
Simplified100.0%
exp-prod100.0%
Applied egg-rr100.0%
div-inv100.0%
pow-flip100.0%
metadata-eval100.0%
expm1-log1p-u100.0%
Applied egg-rr100.0%
Final simplification100.0%
(FPCore (x)
:precision binary64
(*
(pow (exp x) x)
(*
(sqrt (/ 1.0 PI))
(+
(/ 0.5 (pow x 3.0))
(+ (/ 0.75 (pow x 5.0)) (+ (/ 1.0 x) (/ 1.875 (pow x 7.0))))))))
double code(double x) {
return pow(exp(x), x) * (sqrt((1.0 / ((double) M_PI))) * ((0.5 / pow(x, 3.0)) + ((0.75 / pow(x, 5.0)) + ((1.0 / x) + (1.875 / pow(x, 7.0))))));
}
public static double code(double x) {
return Math.pow(Math.exp(x), x) * (Math.sqrt((1.0 / Math.PI)) * ((0.5 / Math.pow(x, 3.0)) + ((0.75 / Math.pow(x, 5.0)) + ((1.0 / x) + (1.875 / Math.pow(x, 7.0))))));
}
def code(x): return math.pow(math.exp(x), x) * (math.sqrt((1.0 / math.pi)) * ((0.5 / math.pow(x, 3.0)) + ((0.75 / math.pow(x, 5.0)) + ((1.0 / x) + (1.875 / math.pow(x, 7.0))))))
function code(x) return Float64((exp(x) ^ x) * Float64(sqrt(Float64(1.0 / pi)) * Float64(Float64(0.5 / (x ^ 3.0)) + Float64(Float64(0.75 / (x ^ 5.0)) + Float64(Float64(1.0 / x) + Float64(1.875 / (x ^ 7.0))))))) end
function tmp = code(x) tmp = (exp(x) ^ x) * (sqrt((1.0 / pi)) * ((0.5 / (x ^ 3.0)) + ((0.75 / (x ^ 5.0)) + ((1.0 / x) + (1.875 / (x ^ 7.0)))))); end
code[x_] := N[(N[Power[N[Exp[x], $MachinePrecision], x], $MachinePrecision] * N[(N[Sqrt[N[(1.0 / Pi), $MachinePrecision]], $MachinePrecision] * N[(N[(0.5 / N[Power[x, 3.0], $MachinePrecision]), $MachinePrecision] + N[(N[(0.75 / N[Power[x, 5.0], $MachinePrecision]), $MachinePrecision] + N[(N[(1.0 / x), $MachinePrecision] + N[(1.875 / N[Power[x, 7.0], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
{\left(e^{x}\right)}^{x} \cdot \left(\sqrt{\frac{1}{\pi}} \cdot \left(\frac{0.5}{{x}^{3}} + \left(\frac{0.75}{{x}^{5}} + \left(\frac{1}{x} + \frac{1.875}{{x}^{7}}\right)\right)\right)\right)
\end{array}
Initial program 100.0%
Simplified100.0%
Taylor expanded in x around 0 100.0%
associate-*r*100.0%
*-commutative100.0%
distribute-rgt-out100.0%
Simplified100.0%
exp-prod100.0%
Applied egg-rr100.0%
Final simplification100.0%
(FPCore (x)
:precision binary64
(*
(*
(sqrt (/ 1.0 PI))
(+
(/ 0.5 (pow x 3.0))
(+ (/ 0.75 (pow x 5.0)) (+ (/ 1.0 x) (/ 1.875 (pow x 7.0))))))
(exp (* x x))))
double code(double x) {
return (sqrt((1.0 / ((double) M_PI))) * ((0.5 / pow(x, 3.0)) + ((0.75 / pow(x, 5.0)) + ((1.0 / x) + (1.875 / pow(x, 7.0)))))) * exp((x * x));
}
public static double code(double x) {
return (Math.sqrt((1.0 / Math.PI)) * ((0.5 / Math.pow(x, 3.0)) + ((0.75 / Math.pow(x, 5.0)) + ((1.0 / x) + (1.875 / Math.pow(x, 7.0)))))) * Math.exp((x * x));
}
def code(x): return (math.sqrt((1.0 / math.pi)) * ((0.5 / math.pow(x, 3.0)) + ((0.75 / math.pow(x, 5.0)) + ((1.0 / x) + (1.875 / math.pow(x, 7.0)))))) * math.exp((x * x))
function code(x) return Float64(Float64(sqrt(Float64(1.0 / pi)) * Float64(Float64(0.5 / (x ^ 3.0)) + Float64(Float64(0.75 / (x ^ 5.0)) + Float64(Float64(1.0 / x) + Float64(1.875 / (x ^ 7.0)))))) * exp(Float64(x * x))) end
function tmp = code(x) tmp = (sqrt((1.0 / pi)) * ((0.5 / (x ^ 3.0)) + ((0.75 / (x ^ 5.0)) + ((1.0 / x) + (1.875 / (x ^ 7.0)))))) * exp((x * x)); end
code[x_] := N[(N[(N[Sqrt[N[(1.0 / Pi), $MachinePrecision]], $MachinePrecision] * N[(N[(0.5 / N[Power[x, 3.0], $MachinePrecision]), $MachinePrecision] + N[(N[(0.75 / N[Power[x, 5.0], $MachinePrecision]), $MachinePrecision] + N[(N[(1.0 / x), $MachinePrecision] + N[(1.875 / N[Power[x, 7.0], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] * N[Exp[N[(x * x), $MachinePrecision]], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\left(\sqrt{\frac{1}{\pi}} \cdot \left(\frac{0.5}{{x}^{3}} + \left(\frac{0.75}{{x}^{5}} + \left(\frac{1}{x} + \frac{1.875}{{x}^{7}}\right)\right)\right)\right) \cdot e^{x \cdot x}
\end{array}
Initial program 100.0%
Simplified100.0%
Taylor expanded in x around 0 100.0%
associate-*r*100.0%
*-commutative100.0%
distribute-rgt-out100.0%
Simplified100.0%
Final simplification100.0%
(FPCore (x) :precision binary64 (* (exp (* x x)) (+ (/ (pow PI -0.5) x) (* (sqrt (/ 1.0 PI)) (+ (/ 0.5 (pow x 3.0)) (/ 0.75 (pow x 5.0)))))))
double code(double x) {
return exp((x * x)) * ((pow(((double) M_PI), -0.5) / x) + (sqrt((1.0 / ((double) M_PI))) * ((0.5 / pow(x, 3.0)) + (0.75 / pow(x, 5.0)))));
}
public static double code(double x) {
return Math.exp((x * x)) * ((Math.pow(Math.PI, -0.5) / x) + (Math.sqrt((1.0 / Math.PI)) * ((0.5 / Math.pow(x, 3.0)) + (0.75 / Math.pow(x, 5.0)))));
}
def code(x): return math.exp((x * x)) * ((math.pow(math.pi, -0.5) / x) + (math.sqrt((1.0 / math.pi)) * ((0.5 / math.pow(x, 3.0)) + (0.75 / math.pow(x, 5.0)))))
function code(x) return Float64(exp(Float64(x * x)) * Float64(Float64((pi ^ -0.5) / x) + Float64(sqrt(Float64(1.0 / pi)) * Float64(Float64(0.5 / (x ^ 3.0)) + Float64(0.75 / (x ^ 5.0)))))) end
function tmp = code(x) tmp = exp((x * x)) * (((pi ^ -0.5) / x) + (sqrt((1.0 / pi)) * ((0.5 / (x ^ 3.0)) + (0.75 / (x ^ 5.0))))); end
code[x_] := N[(N[Exp[N[(x * x), $MachinePrecision]], $MachinePrecision] * N[(N[(N[Power[Pi, -0.5], $MachinePrecision] / x), $MachinePrecision] + N[(N[Sqrt[N[(1.0 / Pi), $MachinePrecision]], $MachinePrecision] * N[(N[(0.5 / N[Power[x, 3.0], $MachinePrecision]), $MachinePrecision] + N[(0.75 / N[Power[x, 5.0], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
e^{x \cdot x} \cdot \left(\frac{{\pi}^{-0.5}}{x} + \sqrt{\frac{1}{\pi}} \cdot \left(\frac{0.5}{{x}^{3}} + \frac{0.75}{{x}^{5}}\right)\right)
\end{array}
Initial program 100.0%
Simplified100.0%
Taylor expanded in x around 0 100.0%
associate-*r*100.0%
*-commutative100.0%
distribute-rgt-out100.0%
Simplified100.0%
Taylor expanded in x around inf 99.2%
associate-+r+99.2%
+-commutative99.2%
associate-*l/99.2%
*-lft-identity99.2%
associate-*r*99.2%
associate-*r*99.2%
distribute-rgt-out99.2%
associate-*r/99.2%
metadata-eval99.2%
associate-*r/99.2%
metadata-eval99.2%
Simplified99.2%
expm1-log1p-u99.2%
expm1-udef32.0%
div-inv32.0%
div-inv32.0%
inv-pow32.0%
sqrt-pow132.0%
metadata-eval32.0%
Applied egg-rr32.0%
expm1-def99.2%
expm1-log1p99.2%
Simplified99.2%
Final simplification99.2%
(FPCore (x) :precision binary64 (* (exp (* x x)) (* (sqrt (/ 1.0 PI)) (+ (/ 0.75 (pow x 5.0)) (+ (/ 1.0 x) (/ 1.875 (pow x 7.0)))))))
double code(double x) {
return exp((x * x)) * (sqrt((1.0 / ((double) M_PI))) * ((0.75 / pow(x, 5.0)) + ((1.0 / x) + (1.875 / pow(x, 7.0)))));
}
public static double code(double x) {
return Math.exp((x * x)) * (Math.sqrt((1.0 / Math.PI)) * ((0.75 / Math.pow(x, 5.0)) + ((1.0 / x) + (1.875 / Math.pow(x, 7.0)))));
}
def code(x): return math.exp((x * x)) * (math.sqrt((1.0 / math.pi)) * ((0.75 / math.pow(x, 5.0)) + ((1.0 / x) + (1.875 / math.pow(x, 7.0)))))
function code(x) return Float64(exp(Float64(x * x)) * Float64(sqrt(Float64(1.0 / pi)) * Float64(Float64(0.75 / (x ^ 5.0)) + Float64(Float64(1.0 / x) + Float64(1.875 / (x ^ 7.0)))))) end
function tmp = code(x) tmp = exp((x * x)) * (sqrt((1.0 / pi)) * ((0.75 / (x ^ 5.0)) + ((1.0 / x) + (1.875 / (x ^ 7.0))))); end
code[x_] := N[(N[Exp[N[(x * x), $MachinePrecision]], $MachinePrecision] * N[(N[Sqrt[N[(1.0 / Pi), $MachinePrecision]], $MachinePrecision] * N[(N[(0.75 / N[Power[x, 5.0], $MachinePrecision]), $MachinePrecision] + N[(N[(1.0 / x), $MachinePrecision] + N[(1.875 / N[Power[x, 7.0], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
e^{x \cdot x} \cdot \left(\sqrt{\frac{1}{\pi}} \cdot \left(\frac{0.75}{{x}^{5}} + \left(\frac{1}{x} + \frac{1.875}{{x}^{7}}\right)\right)\right)
\end{array}
Initial program 100.0%
Simplified100.0%
Taylor expanded in x around inf 99.1%
associate-+r+99.1%
+-commutative99.1%
associate-*r/99.1%
metadata-eval99.1%
rem-square-sqrt99.1%
fabs-sqr99.1%
rem-square-sqrt99.1%
associate-*r/99.1%
metadata-eval99.1%
rem-square-sqrt99.1%
fabs-sqr99.1%
rem-square-sqrt99.1%
rem-square-sqrt99.1%
fabs-sqr99.1%
rem-square-sqrt99.1%
Simplified99.1%
Final simplification99.1%
(FPCore (x) :precision binary64 (* (exp (* x x)) (* (sqrt (/ 1.0 PI)) (+ (/ 0.5 (pow x 3.0)) (/ 1.0 x)))))
double code(double x) {
return exp((x * x)) * (sqrt((1.0 / ((double) M_PI))) * ((0.5 / pow(x, 3.0)) + (1.0 / x)));
}
public static double code(double x) {
return Math.exp((x * x)) * (Math.sqrt((1.0 / Math.PI)) * ((0.5 / Math.pow(x, 3.0)) + (1.0 / x)));
}
def code(x): return math.exp((x * x)) * (math.sqrt((1.0 / math.pi)) * ((0.5 / math.pow(x, 3.0)) + (1.0 / x)))
function code(x) return Float64(exp(Float64(x * x)) * Float64(sqrt(Float64(1.0 / pi)) * Float64(Float64(0.5 / (x ^ 3.0)) + Float64(1.0 / x)))) end
function tmp = code(x) tmp = exp((x * x)) * (sqrt((1.0 / pi)) * ((0.5 / (x ^ 3.0)) + (1.0 / x))); end
code[x_] := N[(N[Exp[N[(x * x), $MachinePrecision]], $MachinePrecision] * N[(N[Sqrt[N[(1.0 / Pi), $MachinePrecision]], $MachinePrecision] * N[(N[(0.5 / N[Power[x, 3.0], $MachinePrecision]), $MachinePrecision] + N[(1.0 / x), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
e^{x \cdot x} \cdot \left(\sqrt{\frac{1}{\pi}} \cdot \left(\frac{0.5}{{x}^{3}} + \frac{1}{x}\right)\right)
\end{array}
Initial program 100.0%
Simplified100.0%
Taylor expanded in x around 0 100.0%
associate-*r*100.0%
*-commutative100.0%
distribute-rgt-out100.0%
Simplified100.0%
Taylor expanded in x around inf 99.1%
associate-*r*99.1%
distribute-rgt-out99.1%
associate-*r/99.1%
metadata-eval99.1%
Simplified99.1%
Final simplification99.1%
(FPCore (x) :precision binary64 (* (exp (* x x)) (/ (sqrt (/ 1.0 PI)) x)))
double code(double x) {
return exp((x * x)) * (sqrt((1.0 / ((double) M_PI))) / x);
}
public static double code(double x) {
return Math.exp((x * x)) * (Math.sqrt((1.0 / Math.PI)) / x);
}
def code(x): return math.exp((x * x)) * (math.sqrt((1.0 / math.pi)) / x)
function code(x) return Float64(exp(Float64(x * x)) * Float64(sqrt(Float64(1.0 / pi)) / x)) end
function tmp = code(x) tmp = exp((x * x)) * (sqrt((1.0 / pi)) / x); end
code[x_] := N[(N[Exp[N[(x * x), $MachinePrecision]], $MachinePrecision] * N[(N[Sqrt[N[(1.0 / Pi), $MachinePrecision]], $MachinePrecision] / x), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
e^{x \cdot x} \cdot \frac{\sqrt{\frac{1}{\pi}}}{x}
\end{array}
Initial program 100.0%
Simplified100.0%
Taylor expanded in x around 0 100.0%
associate-*r*100.0%
*-commutative100.0%
distribute-rgt-out100.0%
Simplified100.0%
Taylor expanded in x around inf 99.1%
associate-*l/99.1%
*-lft-identity99.1%
Simplified99.1%
Final simplification99.1%
(FPCore (x) :precision binary64 (/ 0.5 (* x (sqrt PI))))
double code(double x) {
return 0.5 / (x * sqrt(((double) M_PI)));
}
public static double code(double x) {
return 0.5 / (x * Math.sqrt(Math.PI));
}
def code(x): return 0.5 / (x * math.sqrt(math.pi))
function code(x) return Float64(0.5 / Float64(x * sqrt(pi))) end
function tmp = code(x) tmp = 0.5 / (x * sqrt(pi)); end
code[x_] := N[(0.5 / N[(x * N[Sqrt[Pi], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\frac{0.5}{x \cdot \sqrt{\pi}}
\end{array}
Initial program 100.0%
Simplified100.0%
Taylor expanded in x around 0 31.8%
associate-*r*31.8%
*-commutative31.8%
associate-*r/31.8%
metadata-eval31.8%
rem-square-sqrt31.8%
fabs-sqr31.8%
rem-square-sqrt31.8%
pow-plus31.8%
metadata-eval31.8%
Simplified31.8%
Taylor expanded in x around 0 2.3%
+-commutative2.3%
associate-*r*2.3%
associate-*r*2.3%
distribute-rgt-out2.3%
associate-*r/2.3%
metadata-eval2.3%
associate-*r/2.3%
metadata-eval2.3%
Simplified2.3%
Taylor expanded in x around inf 2.3%
associate-*r*2.3%
associate-*r/2.3%
metadata-eval2.3%
Simplified2.3%
*-commutative2.3%
sqrt-div2.3%
metadata-eval2.3%
frac-times2.3%
metadata-eval2.3%
Applied egg-rr2.3%
Final simplification2.3%
(FPCore (x) :precision binary64 (/ (sqrt (/ 1.0 PI)) x))
double code(double x) {
return sqrt((1.0 / ((double) M_PI))) / x;
}
public static double code(double x) {
return Math.sqrt((1.0 / Math.PI)) / x;
}
def code(x): return math.sqrt((1.0 / math.pi)) / x
function code(x) return Float64(sqrt(Float64(1.0 / pi)) / x) end
function tmp = code(x) tmp = sqrt((1.0 / pi)) / x; end
code[x_] := N[(N[Sqrt[N[(1.0 / Pi), $MachinePrecision]], $MachinePrecision] / x), $MachinePrecision]
\begin{array}{l}
\\
\frac{\sqrt{\frac{1}{\pi}}}{x}
\end{array}
Initial program 100.0%
Simplified100.0%
Taylor expanded in x around 0 100.0%
associate-*r*100.0%
*-commutative100.0%
distribute-rgt-out100.0%
Simplified100.0%
Taylor expanded in x around inf 99.1%
associate-*r*99.1%
distribute-rgt-out99.1%
associate-*r/99.1%
metadata-eval99.1%
Simplified99.1%
Taylor expanded in x around 0 2.4%
Taylor expanded in x around inf 2.3%
associate-*l/2.3%
*-lft-identity2.3%
Simplified2.3%
Final simplification2.3%
herbie shell --seed 2024035
(FPCore (x)
:name "Jmat.Real.erfi, branch x greater than or equal to 5"
:precision binary64
:pre (>= x 0.5)
(* (* (/ 1.0 (sqrt PI)) (exp (* (fabs x) (fabs x)))) (+ (+ (+ (/ 1.0 (fabs x)) (* (/ 1.0 2.0) (* (* (/ 1.0 (fabs x)) (/ 1.0 (fabs x))) (/ 1.0 (fabs x))))) (* (/ 3.0 4.0) (* (* (* (* (/ 1.0 (fabs x)) (/ 1.0 (fabs x))) (/ 1.0 (fabs x))) (/ 1.0 (fabs x))) (/ 1.0 (fabs x))))) (* (/ 15.0 8.0) (* (* (* (* (* (* (/ 1.0 (fabs x)) (/ 1.0 (fabs x))) (/ 1.0 (fabs x))) (/ 1.0 (fabs x))) (/ 1.0 (fabs x))) (/ 1.0 (fabs x))) (/ 1.0 (fabs x)))))))