
(FPCore (x) :precision binary64 (* (cos x) (exp (* 10.0 (* x x)))))
double code(double x) {
return cos(x) * exp((10.0 * (x * x)));
}
real(8) function code(x)
real(8), intent (in) :: x
code = cos(x) * exp((10.0d0 * (x * x)))
end function
public static double code(double x) {
return Math.cos(x) * Math.exp((10.0 * (x * x)));
}
def code(x): return math.cos(x) * math.exp((10.0 * (x * x)))
function code(x) return Float64(cos(x) * exp(Float64(10.0 * Float64(x * x)))) end
function tmp = code(x) tmp = cos(x) * exp((10.0 * (x * x))); end
code[x_] := N[(N[Cos[x], $MachinePrecision] * N[Exp[N[(10.0 * N[(x * x), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\cos x \cdot e^{10 \cdot \left(x \cdot x\right)}
\end{array}
Sampling outcomes in binary64 precision:
Herbie found 19 alternatives:
| Alternative | Accuracy | Speedup |
|---|
(FPCore (x) :precision binary64 (* (cos x) (exp (* 10.0 (* x x)))))
double code(double x) {
return cos(x) * exp((10.0 * (x * x)));
}
real(8) function code(x)
real(8), intent (in) :: x
code = cos(x) * exp((10.0d0 * (x * x)))
end function
public static double code(double x) {
return Math.cos(x) * Math.exp((10.0 * (x * x)));
}
def code(x): return math.cos(x) * math.exp((10.0 * (x * x)))
function code(x) return Float64(cos(x) * exp(Float64(10.0 * Float64(x * x)))) end
function tmp = code(x) tmp = cos(x) * exp((10.0 * (x * x))); end
code[x_] := N[(N[Cos[x], $MachinePrecision] * N[Exp[N[(10.0 * N[(x * x), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\cos x \cdot e^{10 \cdot \left(x \cdot x\right)}
\end{array}
(FPCore (x) :precision binary64 (* (pow (pow (exp 20.0) x) (* 0.5 x)) (cos x)))
double code(double x) {
return pow(pow(exp(20.0), x), (0.5 * x)) * cos(x);
}
real(8) function code(x)
real(8), intent (in) :: x
code = ((exp(20.0d0) ** x) ** (0.5d0 * x)) * cos(x)
end function
public static double code(double x) {
return Math.pow(Math.pow(Math.exp(20.0), x), (0.5 * x)) * Math.cos(x);
}
def code(x): return math.pow(math.pow(math.exp(20.0), x), (0.5 * x)) * math.cos(x)
function code(x) return Float64(((exp(20.0) ^ x) ^ Float64(0.5 * x)) * cos(x)) end
function tmp = code(x) tmp = ((exp(20.0) ^ x) ^ (0.5 * x)) * cos(x); end
code[x_] := N[(N[Power[N[Power[N[Exp[20.0], $MachinePrecision], x], $MachinePrecision], N[(0.5 * x), $MachinePrecision]], $MachinePrecision] * N[Cos[x], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
{\left({\left(e^{20}\right)}^{x}\right)}^{\left(0.5 \cdot x\right)} \cdot \cos x
\end{array}
Initial program 94.4%
lift-exp.f64N/A
lift-*.f64N/A
exp-prodN/A
sqr-powN/A
pow-prod-downN/A
lift-*.f64N/A
associate-/l*N/A
pow-unpowN/A
lower-pow.f64N/A
Applied rewrites96.7%
lift-pow.f64N/A
lift-exp.f64N/A
pow-expN/A
*-commutativeN/A
pow-expN/A
lift-exp.f64N/A
lift-pow.f6499.4
Applied rewrites99.4%
Final simplification99.4%
(FPCore (x) :precision binary64 (* (pow (pow (exp -20.0) x) (* -0.5 x)) (cos x)))
double code(double x) {
return pow(pow(exp(-20.0), x), (-0.5 * x)) * cos(x);
}
real(8) function code(x)
real(8), intent (in) :: x
code = ((exp((-20.0d0)) ** x) ** ((-0.5d0) * x)) * cos(x)
end function
public static double code(double x) {
return Math.pow(Math.pow(Math.exp(-20.0), x), (-0.5 * x)) * Math.cos(x);
}
def code(x): return math.pow(math.pow(math.exp(-20.0), x), (-0.5 * x)) * math.cos(x)
function code(x) return Float64(((exp(-20.0) ^ x) ^ Float64(-0.5 * x)) * cos(x)) end
function tmp = code(x) tmp = ((exp(-20.0) ^ x) ^ (-0.5 * x)) * cos(x); end
code[x_] := N[(N[Power[N[Power[N[Exp[-20.0], $MachinePrecision], x], $MachinePrecision], N[(-0.5 * x), $MachinePrecision]], $MachinePrecision] * N[Cos[x], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
{\left({\left(e^{-20}\right)}^{x}\right)}^{\left(-0.5 \cdot x\right)} \cdot \cos x
\end{array}
Initial program 94.4%
lift-exp.f64N/A
lift-*.f64N/A
exp-prodN/A
sqr-powN/A
pow-prod-downN/A
lift-*.f64N/A
associate-/l*N/A
pow-unpowN/A
lower-pow.f64N/A
Applied rewrites96.7%
lift-pow.f64N/A
lift-exp.f64N/A
pow-expN/A
*-commutativeN/A
pow-expN/A
lift-exp.f64N/A
lift-pow.f6499.4
Applied rewrites99.4%
lift-pow.f64N/A
lift-pow.f64N/A
pow-powN/A
lift-*.f64N/A
*-commutativeN/A
associate-*r*N/A
lift-*.f64N/A
metadata-evalN/A
distribute-rgt-neg-inN/A
distribute-lft-neg-outN/A
lift-*.f64N/A
distribute-lft-neg-outN/A
lift-neg.f64N/A
lift-*.f64N/A
lift-*.f64N/A
associate-*l*N/A
*-commutativeN/A
lift-*.f64N/A
pow-powN/A
lower-pow.f64N/A
Applied rewrites99.3%
Final simplification99.3%
(FPCore (x) :precision binary64 (* (pow (pow (exp -10.0) x) (- x)) (cos x)))
double code(double x) {
return pow(pow(exp(-10.0), x), -x) * cos(x);
}
real(8) function code(x)
real(8), intent (in) :: x
code = ((exp((-10.0d0)) ** x) ** -x) * cos(x)
end function
public static double code(double x) {
return Math.pow(Math.pow(Math.exp(-10.0), x), -x) * Math.cos(x);
}
def code(x): return math.pow(math.pow(math.exp(-10.0), x), -x) * math.cos(x)
function code(x) return Float64(((exp(-10.0) ^ x) ^ Float64(-x)) * cos(x)) end
function tmp = code(x) tmp = ((exp(-10.0) ^ x) ^ -x) * cos(x); end
code[x_] := N[(N[Power[N[Power[N[Exp[-10.0], $MachinePrecision], x], $MachinePrecision], (-x)], $MachinePrecision] * N[Cos[x], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
{\left({\left(e^{-10}\right)}^{x}\right)}^{\left(-x\right)} \cdot \cos x
\end{array}
Initial program 94.4%
lift-exp.f64N/A
lift-*.f64N/A
exp-prodN/A
sqr-powN/A
pow-prod-upN/A
flip-+N/A
+-inversesN/A
+-inversesN/A
metadata-evalN/A
associate-/l/N/A
+-inversesN/A
+-inversesN/A
flip-+N/A
count-2N/A
*-commutativeN/A
associate-*l/N/A
associate-/l*N/A
metadata-evalN/A
*-rgt-identityN/A
frac-2negN/A
distribute-frac-negN/A
Applied rewrites95.3%
lift-/.f64N/A
lift-pow.f64N/A
lift-*.f64N/A
pow-unpowN/A
pow-flipN/A
lift-neg.f64N/A
lower-pow.f64N/A
lift-neg.f64N/A
neg-mul-1N/A
pow-unpowN/A
lower-pow.f64N/A
pow-to-expN/A
lower-exp.f64N/A
lift-exp.f64N/A
rem-log-expN/A
metadata-eval98.1
Applied rewrites98.1%
Final simplification98.1%
(FPCore (x) :precision binary64 (* (pow (pow (exp 10.0) x) x) (cos x)))
double code(double x) {
return pow(pow(exp(10.0), x), x) * cos(x);
}
real(8) function code(x)
real(8), intent (in) :: x
code = ((exp(10.0d0) ** x) ** x) * cos(x)
end function
public static double code(double x) {
return Math.pow(Math.pow(Math.exp(10.0), x), x) * Math.cos(x);
}
def code(x): return math.pow(math.pow(math.exp(10.0), x), x) * math.cos(x)
function code(x) return Float64(((exp(10.0) ^ x) ^ x) * cos(x)) end
function tmp = code(x) tmp = ((exp(10.0) ^ x) ^ x) * cos(x); end
code[x_] := N[(N[Power[N[Power[N[Exp[10.0], $MachinePrecision], x], $MachinePrecision], x], $MachinePrecision] * N[Cos[x], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
{\left({\left(e^{10}\right)}^{x}\right)}^{x} \cdot \cos x
\end{array}
Initial program 94.4%
lift-exp.f64N/A
lift-*.f64N/A
exp-prodN/A
lift-*.f64N/A
pow-unpowN/A
lower-pow.f64N/A
lower-pow.f64N/A
lower-exp.f6497.8
Applied rewrites97.8%
Final simplification97.8%
(FPCore (x) :precision binary64 (* (pow (pow (exp x) 10.0) x) (cos x)))
double code(double x) {
return pow(pow(exp(x), 10.0), x) * cos(x);
}
real(8) function code(x)
real(8), intent (in) :: x
code = ((exp(x) ** 10.0d0) ** x) * cos(x)
end function
public static double code(double x) {
return Math.pow(Math.pow(Math.exp(x), 10.0), x) * Math.cos(x);
}
def code(x): return math.pow(math.pow(math.exp(x), 10.0), x) * math.cos(x)
function code(x) return Float64(((exp(x) ^ 10.0) ^ x) * cos(x)) end
function tmp = code(x) tmp = ((exp(x) ^ 10.0) ^ x) * cos(x); end
code[x_] := N[(N[Power[N[Power[N[Exp[x], $MachinePrecision], 10.0], $MachinePrecision], x], $MachinePrecision] * N[Cos[x], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
{\left({\left(e^{x}\right)}^{10}\right)}^{x} \cdot \cos x
\end{array}
Initial program 94.4%
Taylor expanded in x around inf
unpow2N/A
associate-*r*N/A
exp-prodN/A
lower-pow.f64N/A
*-commutativeN/A
exp-prodN/A
lower-pow.f64N/A
lower-exp.f6496.7
Applied rewrites96.7%
Final simplification96.7%
(FPCore (x) :precision binary64 (* (pow (exp (* 10.0 x)) x) (cos x)))
double code(double x) {
return pow(exp((10.0 * x)), x) * cos(x);
}
real(8) function code(x)
real(8), intent (in) :: x
code = (exp((10.0d0 * x)) ** x) * cos(x)
end function
public static double code(double x) {
return Math.pow(Math.exp((10.0 * x)), x) * Math.cos(x);
}
def code(x): return math.pow(math.exp((10.0 * x)), x) * math.cos(x)
function code(x) return Float64((exp(Float64(10.0 * x)) ^ x) * cos(x)) end
function tmp = code(x) tmp = (exp((10.0 * x)) ^ x) * cos(x); end
code[x_] := N[(N[Power[N[Exp[N[(10.0 * x), $MachinePrecision]], $MachinePrecision], x], $MachinePrecision] * N[Cos[x], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
{\left(e^{10 \cdot x}\right)}^{x} \cdot \cos x
\end{array}
Initial program 94.4%
lift-exp.f64N/A
lift-*.f64N/A
exp-prodN/A
sqr-powN/A
pow-prod-upN/A
flip-+N/A
+-inversesN/A
+-inversesN/A
metadata-evalN/A
associate-/l/N/A
+-inversesN/A
+-inversesN/A
flip-+N/A
count-2N/A
*-commutativeN/A
associate-*l/N/A
associate-/l*N/A
metadata-evalN/A
*-rgt-identityN/A
frac-2negN/A
distribute-frac-negN/A
Applied rewrites95.3%
lift-/.f64N/A
lift-pow.f64N/A
pow-flipN/A
lift-*.f64N/A
lift-neg.f64N/A
distribute-lft-neg-outN/A
lift-*.f64N/A
remove-double-negN/A
lift-exp.f64N/A
exp-prodN/A
lift-*.f64N/A
associate-*r*N/A
*-commutativeN/A
pow-expN/A
lift-exp.f64N/A
pow-powN/A
metadata-evalN/A
lower-pow.f64N/A
metadata-evalN/A
lower-pow.f6496.7
Applied rewrites96.7%
rem-exp-logN/A
lift-pow.f64N/A
log-powN/A
metadata-evalN/A
rem-log-expN/A
lift-exp.f64N/A
lift-exp.f64N/A
rem-log-expN/A
associate-*r*N/A
*-commutativeN/A
metadata-evalN/A
div-invN/A
pow-to-expN/A
sqrt-pow1N/A
lift-pow.f64N/A
unpow1/2N/A
pow-to-expN/A
lower-exp.f64N/A
lift-pow.f64N/A
pow-to-expN/A
rem-log-expN/A
associate-*r*N/A
*-commutativeN/A
associate-*r*N/A
Applied rewrites95.4%
Final simplification95.4%
(FPCore (x) :precision binary64 (* (pow (exp 10.0) (* x x)) (cos x)))
double code(double x) {
return pow(exp(10.0), (x * x)) * cos(x);
}
real(8) function code(x)
real(8), intent (in) :: x
code = (exp(10.0d0) ** (x * x)) * cos(x)
end function
public static double code(double x) {
return Math.pow(Math.exp(10.0), (x * x)) * Math.cos(x);
}
def code(x): return math.pow(math.exp(10.0), (x * x)) * math.cos(x)
function code(x) return Float64((exp(10.0) ^ Float64(x * x)) * cos(x)) end
function tmp = code(x) tmp = (exp(10.0) ^ (x * x)) * cos(x); end
code[x_] := N[(N[Power[N[Exp[10.0], $MachinePrecision], N[(x * x), $MachinePrecision]], $MachinePrecision] * N[Cos[x], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
{\left(e^{10}\right)}^{\left(x \cdot x\right)} \cdot \cos x
\end{array}
Initial program 94.4%
lift-exp.f64N/A
lift-*.f64N/A
exp-prodN/A
lower-pow.f64N/A
lower-exp.f6495.2
Applied rewrites95.2%
Final simplification95.2%
(FPCore (x) :precision binary64 (* (sqrt (exp (* 2.0 (* (* x x) 10.0)))) (cos x)))
double code(double x) {
return sqrt(exp((2.0 * ((x * x) * 10.0)))) * cos(x);
}
real(8) function code(x)
real(8), intent (in) :: x
code = sqrt(exp((2.0d0 * ((x * x) * 10.0d0)))) * cos(x)
end function
public static double code(double x) {
return Math.sqrt(Math.exp((2.0 * ((x * x) * 10.0)))) * Math.cos(x);
}
def code(x): return math.sqrt(math.exp((2.0 * ((x * x) * 10.0)))) * math.cos(x)
function code(x) return Float64(sqrt(exp(Float64(2.0 * Float64(Float64(x * x) * 10.0)))) * cos(x)) end
function tmp = code(x) tmp = sqrt(exp((2.0 * ((x * x) * 10.0)))) * cos(x); end
code[x_] := N[(N[Sqrt[N[Exp[N[(2.0 * N[(N[(x * x), $MachinePrecision] * 10.0), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]], $MachinePrecision] * N[Cos[x], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\sqrt{e^{2 \cdot \left(\left(x \cdot x\right) \cdot 10\right)}} \cdot \cos x
\end{array}
Initial program 94.4%
lift-exp.f64N/A
lift-*.f64N/A
exp-prodN/A
sqr-powN/A
pow-prod-downN/A
frac-2negN/A
div-invN/A
pow-unpowN/A
lower-pow.f64N/A
Applied rewrites95.2%
Taylor expanded in x around inf
lower-sqrt.f64N/A
rec-expN/A
unpow2N/A
associate-*r*N/A
distribute-lft-neg-inN/A
exp-prodN/A
lower-pow.f64N/A
distribute-lft-neg-inN/A
metadata-evalN/A
exp-prodN/A
lower-pow.f64N/A
lower-exp.f6499.4
Applied rewrites99.4%
Applied rewrites94.4%
Final simplification94.4%
(FPCore (x) :precision binary64 (* (/ 1.0 (exp (* (* x x) -10.0))) (cos x)))
double code(double x) {
return (1.0 / exp(((x * x) * -10.0))) * cos(x);
}
real(8) function code(x)
real(8), intent (in) :: x
code = (1.0d0 / exp(((x * x) * (-10.0d0)))) * cos(x)
end function
public static double code(double x) {
return (1.0 / Math.exp(((x * x) * -10.0))) * Math.cos(x);
}
def code(x): return (1.0 / math.exp(((x * x) * -10.0))) * math.cos(x)
function code(x) return Float64(Float64(1.0 / exp(Float64(Float64(x * x) * -10.0))) * cos(x)) end
function tmp = code(x) tmp = (1.0 / exp(((x * x) * -10.0))) * cos(x); end
code[x_] := N[(N[(1.0 / N[Exp[N[(N[(x * x), $MachinePrecision] * -10.0), $MachinePrecision]], $MachinePrecision]), $MachinePrecision] * N[Cos[x], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\frac{1}{e^{\left(x \cdot x\right) \cdot -10}} \cdot \cos x
\end{array}
Initial program 94.4%
lift-exp.f64N/A
lift-*.f64N/A
exp-prodN/A
sqr-powN/A
pow-prod-upN/A
flip-+N/A
+-inversesN/A
+-inversesN/A
metadata-evalN/A
associate-/l/N/A
+-inversesN/A
+-inversesN/A
flip-+N/A
count-2N/A
*-commutativeN/A
associate-*l/N/A
associate-/l*N/A
metadata-evalN/A
*-rgt-identityN/A
frac-2negN/A
distribute-frac-negN/A
Applied rewrites95.3%
lift-pow.f64N/A
lift-exp.f64N/A
pow-expN/A
lower-exp.f64N/A
lift-*.f64N/A
lift-neg.f64N/A
distribute-lft-neg-outN/A
lift-*.f64N/A
neg-mul-1N/A
associate-*r*N/A
rem-log-expN/A
lift-exp.f64N/A
lower-*.f64N/A
lift-exp.f64N/A
rem-log-expN/A
metadata-eval94.4
Applied rewrites94.4%
Final simplification94.4%
(FPCore (x) :precision binary64 (* (exp (* (* x x) 10.0)) (cos x)))
double code(double x) {
return exp(((x * x) * 10.0)) * cos(x);
}
real(8) function code(x)
real(8), intent (in) :: x
code = exp(((x * x) * 10.0d0)) * cos(x)
end function
public static double code(double x) {
return Math.exp(((x * x) * 10.0)) * Math.cos(x);
}
def code(x): return math.exp(((x * x) * 10.0)) * math.cos(x)
function code(x) return Float64(exp(Float64(Float64(x * x) * 10.0)) * cos(x)) end
function tmp = code(x) tmp = exp(((x * x) * 10.0)) * cos(x); end
code[x_] := N[(N[Exp[N[(N[(x * x), $MachinePrecision] * 10.0), $MachinePrecision]], $MachinePrecision] * N[Cos[x], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
e^{\left(x \cdot x\right) \cdot 10} \cdot \cos x
\end{array}
Initial program 94.4%
Final simplification94.4%
(FPCore (x)
:precision binary64
(*
(+
(*
(fma (* (fma -0.001388888888888889 (* x x) 0.041666666666666664) x) x -0.5)
(* x x))
1.0)
(exp (* (* x x) 10.0))))
double code(double x) {
return ((fma((fma(-0.001388888888888889, (x * x), 0.041666666666666664) * x), x, -0.5) * (x * x)) + 1.0) * exp(((x * x) * 10.0));
}
function code(x) return Float64(Float64(Float64(fma(Float64(fma(-0.001388888888888889, Float64(x * x), 0.041666666666666664) * x), x, -0.5) * Float64(x * x)) + 1.0) * exp(Float64(Float64(x * x) * 10.0))) end
code[x_] := N[(N[(N[(N[(N[(N[(-0.001388888888888889 * N[(x * x), $MachinePrecision] + 0.041666666666666664), $MachinePrecision] * x), $MachinePrecision] * x + -0.5), $MachinePrecision] * N[(x * x), $MachinePrecision]), $MachinePrecision] + 1.0), $MachinePrecision] * N[Exp[N[(N[(x * x), $MachinePrecision] * 10.0), $MachinePrecision]], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\left(\mathsf{fma}\left(\mathsf{fma}\left(-0.001388888888888889, x \cdot x, 0.041666666666666664\right) \cdot x, x, -0.5\right) \cdot \left(x \cdot x\right) + 1\right) \cdot e^{\left(x \cdot x\right) \cdot 10}
\end{array}
Initial program 94.4%
Taylor expanded in x around 0
+-commutativeN/A
*-commutativeN/A
lower-fma.f64N/A
sub-negN/A
metadata-evalN/A
*-commutativeN/A
lower-fma.f64N/A
+-commutativeN/A
lower-fma.f64N/A
unpow2N/A
lower-*.f64N/A
unpow2N/A
lower-*.f64N/A
unpow2N/A
lower-*.f6427.5
Applied rewrites27.5%
Applied rewrites27.5%
Applied rewrites27.5%
Final simplification27.5%
(FPCore (x)
:precision binary64
(*
(fma
x
(*
(fma (* (fma -0.001388888888888889 (* x x) 0.041666666666666664) x) x -0.5)
x)
1.0)
(exp (* (* x x) 10.0))))
double code(double x) {
return fma(x, (fma((fma(-0.001388888888888889, (x * x), 0.041666666666666664) * x), x, -0.5) * x), 1.0) * exp(((x * x) * 10.0));
}
function code(x) return Float64(fma(x, Float64(fma(Float64(fma(-0.001388888888888889, Float64(x * x), 0.041666666666666664) * x), x, -0.5) * x), 1.0) * exp(Float64(Float64(x * x) * 10.0))) end
code[x_] := N[(N[(x * N[(N[(N[(N[(-0.001388888888888889 * N[(x * x), $MachinePrecision] + 0.041666666666666664), $MachinePrecision] * x), $MachinePrecision] * x + -0.5), $MachinePrecision] * x), $MachinePrecision] + 1.0), $MachinePrecision] * N[Exp[N[(N[(x * x), $MachinePrecision] * 10.0), $MachinePrecision]], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\mathsf{fma}\left(x, \mathsf{fma}\left(\mathsf{fma}\left(-0.001388888888888889, x \cdot x, 0.041666666666666664\right) \cdot x, x, -0.5\right) \cdot x, 1\right) \cdot e^{\left(x \cdot x\right) \cdot 10}
\end{array}
Initial program 94.4%
Taylor expanded in x around 0
+-commutativeN/A
*-commutativeN/A
lower-fma.f64N/A
sub-negN/A
metadata-evalN/A
*-commutativeN/A
lower-fma.f64N/A
+-commutativeN/A
lower-fma.f64N/A
unpow2N/A
lower-*.f64N/A
unpow2N/A
lower-*.f64N/A
unpow2N/A
lower-*.f6427.5
Applied rewrites27.5%
Applied rewrites27.5%
Applied rewrites27.5%
Final simplification27.5%
(FPCore (x) :precision binary64 (* (fma (fma 0.041666666666666664 (* x x) -0.5) (* x x) 1.0) (exp (* (* x x) 10.0))))
double code(double x) {
return fma(fma(0.041666666666666664, (x * x), -0.5), (x * x), 1.0) * exp(((x * x) * 10.0));
}
function code(x) return Float64(fma(fma(0.041666666666666664, Float64(x * x), -0.5), Float64(x * x), 1.0) * exp(Float64(Float64(x * x) * 10.0))) end
code[x_] := N[(N[(N[(0.041666666666666664 * N[(x * x), $MachinePrecision] + -0.5), $MachinePrecision] * N[(x * x), $MachinePrecision] + 1.0), $MachinePrecision] * N[Exp[N[(N[(x * x), $MachinePrecision] * 10.0), $MachinePrecision]], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\mathsf{fma}\left(\mathsf{fma}\left(0.041666666666666664, x \cdot x, -0.5\right), x \cdot x, 1\right) \cdot e^{\left(x \cdot x\right) \cdot 10}
\end{array}
Initial program 94.4%
Taylor expanded in x around 0
+-commutativeN/A
*-commutativeN/A
lower-fma.f64N/A
sub-negN/A
metadata-evalN/A
lower-fma.f64N/A
unpow2N/A
lower-*.f64N/A
unpow2N/A
lower-*.f6421.3
Applied rewrites21.3%
Final simplification21.3%
(FPCore (x) :precision binary64 (* (fma -0.5 (* x x) 1.0) (exp (* (* x x) 10.0))))
double code(double x) {
return fma(-0.5, (x * x), 1.0) * exp(((x * x) * 10.0));
}
function code(x) return Float64(fma(-0.5, Float64(x * x), 1.0) * exp(Float64(Float64(x * x) * 10.0))) end
code[x_] := N[(N[(-0.5 * N[(x * x), $MachinePrecision] + 1.0), $MachinePrecision] * N[Exp[N[(N[(x * x), $MachinePrecision] * 10.0), $MachinePrecision]], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\mathsf{fma}\left(-0.5, x \cdot x, 1\right) \cdot e^{\left(x \cdot x\right) \cdot 10}
\end{array}
Initial program 94.4%
Taylor expanded in x around 0
+-commutativeN/A
lower-fma.f64N/A
unpow2N/A
lower-*.f6418.2
Applied rewrites18.2%
Final simplification18.2%
(FPCore (x) :precision binary64 (* (fma (fma (fma 166.66666666666666 (* x x) 50.0) (* x x) 10.0) (* x x) 1.0) (fma -0.5 (* x x) 1.0)))
double code(double x) {
return fma(fma(fma(166.66666666666666, (x * x), 50.0), (x * x), 10.0), (x * x), 1.0) * fma(-0.5, (x * x), 1.0);
}
function code(x) return Float64(fma(fma(fma(166.66666666666666, Float64(x * x), 50.0), Float64(x * x), 10.0), Float64(x * x), 1.0) * fma(-0.5, Float64(x * x), 1.0)) end
code[x_] := N[(N[(N[(N[(166.66666666666666 * N[(x * x), $MachinePrecision] + 50.0), $MachinePrecision] * N[(x * x), $MachinePrecision] + 10.0), $MachinePrecision] * N[(x * x), $MachinePrecision] + 1.0), $MachinePrecision] * N[(-0.5 * N[(x * x), $MachinePrecision] + 1.0), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(166.66666666666666, x \cdot x, 50\right), x \cdot x, 10\right), x \cdot x, 1\right) \cdot \mathsf{fma}\left(-0.5, x \cdot x, 1\right)
\end{array}
Initial program 94.4%
lift-exp.f64N/A
lift-*.f64N/A
exp-prodN/A
sqr-powN/A
pow-prod-downN/A
frac-2negN/A
div-invN/A
pow-unpowN/A
lower-pow.f64N/A
Applied rewrites95.2%
Taylor expanded in x around 0
Applied rewrites9.6%
Taylor expanded in x around 0
+-commutativeN/A
metadata-evalN/A
metadata-evalN/A
*-inversesN/A
distribute-frac-neg2N/A
associate-/l*N/A
associate-*l/N/A
distribute-neg-frac2N/A
pow-sqrN/A
metadata-evalN/A
*-commutativeN/A
lower-fma.f64N/A
Applied rewrites9.7%
Taylor expanded in x around 0
+-commutativeN/A
*-commutativeN/A
lower-fma.f64N/A
+-commutativeN/A
*-commutativeN/A
lower-fma.f64N/A
+-commutativeN/A
lower-fma.f64N/A
unpow2N/A
lower-*.f64N/A
unpow2N/A
lower-*.f64N/A
unpow2N/A
lower-*.f6410.3
Applied rewrites10.3%
Final simplification10.3%
(FPCore (x) :precision binary64 (* (fma (fma 50.0 (* x x) 10.0) (* x x) 1.0) (fma -0.5 (* x x) 1.0)))
double code(double x) {
return fma(fma(50.0, (x * x), 10.0), (x * x), 1.0) * fma(-0.5, (x * x), 1.0);
}
function code(x) return Float64(fma(fma(50.0, Float64(x * x), 10.0), Float64(x * x), 1.0) * fma(-0.5, Float64(x * x), 1.0)) end
code[x_] := N[(N[(N[(50.0 * N[(x * x), $MachinePrecision] + 10.0), $MachinePrecision] * N[(x * x), $MachinePrecision] + 1.0), $MachinePrecision] * N[(-0.5 * N[(x * x), $MachinePrecision] + 1.0), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\mathsf{fma}\left(\mathsf{fma}\left(50, x \cdot x, 10\right), x \cdot x, 1\right) \cdot \mathsf{fma}\left(-0.5, x \cdot x, 1\right)
\end{array}
Initial program 94.4%
lift-exp.f64N/A
lift-*.f64N/A
exp-prodN/A
sqr-powN/A
pow-prod-downN/A
frac-2negN/A
div-invN/A
pow-unpowN/A
lower-pow.f64N/A
Applied rewrites95.2%
Taylor expanded in x around 0
Applied rewrites9.6%
Taylor expanded in x around 0
+-commutativeN/A
metadata-evalN/A
metadata-evalN/A
*-inversesN/A
distribute-frac-neg2N/A
associate-/l*N/A
associate-*l/N/A
distribute-neg-frac2N/A
pow-sqrN/A
metadata-evalN/A
*-commutativeN/A
lower-fma.f64N/A
Applied rewrites9.7%
Taylor expanded in x around 0
+-commutativeN/A
*-commutativeN/A
lower-fma.f64N/A
+-commutativeN/A
lower-fma.f64N/A
unpow2N/A
lower-*.f64N/A
unpow2N/A
lower-*.f6410.1
Applied rewrites10.1%
Final simplification10.1%
(FPCore (x) :precision binary64 (* (fma 10.0 (* x x) 1.0) (fma -0.5 (* x x) 1.0)))
double code(double x) {
return fma(10.0, (x * x), 1.0) * fma(-0.5, (x * x), 1.0);
}
function code(x) return Float64(fma(10.0, Float64(x * x), 1.0) * fma(-0.5, Float64(x * x), 1.0)) end
code[x_] := N[(N[(10.0 * N[(x * x), $MachinePrecision] + 1.0), $MachinePrecision] * N[(-0.5 * N[(x * x), $MachinePrecision] + 1.0), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\mathsf{fma}\left(10, x \cdot x, 1\right) \cdot \mathsf{fma}\left(-0.5, x \cdot x, 1\right)
\end{array}
Initial program 94.4%
lift-exp.f64N/A
lift-*.f64N/A
exp-prodN/A
sqr-powN/A
pow-prod-downN/A
frac-2negN/A
div-invN/A
pow-unpowN/A
lower-pow.f64N/A
Applied rewrites95.2%
Taylor expanded in x around 0
Applied rewrites9.6%
Taylor expanded in x around 0
+-commutativeN/A
metadata-evalN/A
metadata-evalN/A
*-inversesN/A
distribute-frac-neg2N/A
associate-/l*N/A
associate-*l/N/A
distribute-neg-frac2N/A
pow-sqrN/A
metadata-evalN/A
*-commutativeN/A
lower-fma.f64N/A
Applied rewrites9.7%
Taylor expanded in x around 0
+-commutativeN/A
lower-fma.f64N/A
unpow2N/A
lower-*.f649.9
Applied rewrites9.9%
Final simplification9.9%
(FPCore (x) :precision binary64 (* 1.0 (* (* x x) -0.5)))
double code(double x) {
return 1.0 * ((x * x) * -0.5);
}
real(8) function code(x)
real(8), intent (in) :: x
code = 1.0d0 * ((x * x) * (-0.5d0))
end function
public static double code(double x) {
return 1.0 * ((x * x) * -0.5);
}
def code(x): return 1.0 * ((x * x) * -0.5)
function code(x) return Float64(1.0 * Float64(Float64(x * x) * -0.5)) end
function tmp = code(x) tmp = 1.0 * ((x * x) * -0.5); end
code[x_] := N[(1.0 * N[(N[(x * x), $MachinePrecision] * -0.5), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
1 \cdot \left(\left(x \cdot x\right) \cdot -0.5\right)
\end{array}
Initial program 94.4%
lift-exp.f64N/A
lift-*.f64N/A
exp-prodN/A
sqr-powN/A
pow-prod-downN/A
frac-2negN/A
div-invN/A
pow-unpowN/A
lower-pow.f64N/A
Applied rewrites95.2%
Taylor expanded in x around 0
Applied rewrites9.6%
Taylor expanded in x around 0
+-commutativeN/A
metadata-evalN/A
metadata-evalN/A
*-inversesN/A
distribute-frac-neg2N/A
associate-/l*N/A
associate-*l/N/A
distribute-neg-frac2N/A
pow-sqrN/A
metadata-evalN/A
*-commutativeN/A
lower-fma.f64N/A
Applied rewrites9.7%
Taylor expanded in x around inf
Applied rewrites9.7%
Final simplification9.7%
(FPCore (x) :precision binary64 1.0)
double code(double x) {
return 1.0;
}
real(8) function code(x)
real(8), intent (in) :: x
code = 1.0d0
end function
public static double code(double x) {
return 1.0;
}
def code(x): return 1.0
function code(x) return 1.0 end
function tmp = code(x) tmp = 1.0; end
code[x_] := 1.0
\begin{array}{l}
\\
1
\end{array}
Initial program 94.4%
Taylor expanded in x around 0
Applied rewrites1.5%
herbie shell --seed 2024271
(FPCore (x)
:name "ENA, Section 1.4, Exercise 1"
:precision binary64
:pre (and (<= 1.99 x) (<= x 2.01))
(* (cos x) (exp (* 10.0 (* x x)))))