
(FPCore (alpha beta i) :precision binary64 (let* ((t_0 (+ (+ alpha beta) (* 2.0 i)))) (/ (+ (/ (/ (* (+ alpha beta) (- beta alpha)) t_0) (+ t_0 2.0)) 1.0) 2.0)))
double code(double alpha, double beta, double i) {
double t_0 = (alpha + beta) + (2.0 * i);
return (((((alpha + beta) * (beta - alpha)) / t_0) / (t_0 + 2.0)) + 1.0) / 2.0;
}
real(8) function code(alpha, beta, i)
real(8), intent (in) :: alpha
real(8), intent (in) :: beta
real(8), intent (in) :: i
real(8) :: t_0
t_0 = (alpha + beta) + (2.0d0 * i)
code = (((((alpha + beta) * (beta - alpha)) / t_0) / (t_0 + 2.0d0)) + 1.0d0) / 2.0d0
end function
public static double code(double alpha, double beta, double i) {
double t_0 = (alpha + beta) + (2.0 * i);
return (((((alpha + beta) * (beta - alpha)) / t_0) / (t_0 + 2.0)) + 1.0) / 2.0;
}
def code(alpha, beta, i): t_0 = (alpha + beta) + (2.0 * i) return (((((alpha + beta) * (beta - alpha)) / t_0) / (t_0 + 2.0)) + 1.0) / 2.0
function code(alpha, beta, i) t_0 = Float64(Float64(alpha + beta) + Float64(2.0 * i)) return Float64(Float64(Float64(Float64(Float64(Float64(alpha + beta) * Float64(beta - alpha)) / t_0) / Float64(t_0 + 2.0)) + 1.0) / 2.0) end
function tmp = code(alpha, beta, i) t_0 = (alpha + beta) + (2.0 * i); tmp = (((((alpha + beta) * (beta - alpha)) / t_0) / (t_0 + 2.0)) + 1.0) / 2.0; end
code[alpha_, beta_, i_] := Block[{t$95$0 = N[(N[(alpha + beta), $MachinePrecision] + N[(2.0 * i), $MachinePrecision]), $MachinePrecision]}, N[(N[(N[(N[(N[(N[(alpha + beta), $MachinePrecision] * N[(beta - alpha), $MachinePrecision]), $MachinePrecision] / t$95$0), $MachinePrecision] / N[(t$95$0 + 2.0), $MachinePrecision]), $MachinePrecision] + 1.0), $MachinePrecision] / 2.0), $MachinePrecision]]
\begin{array}{l}
\\
\begin{array}{l}
t_0 := \left(\alpha + \beta\right) + 2 \cdot i\\
\frac{\frac{\frac{\left(\alpha + \beta\right) \cdot \left(\beta - \alpha\right)}{t\_0}}{t\_0 + 2} + 1}{2}
\end{array}
\end{array}
Sampling outcomes in binary64 precision:
Herbie found 14 alternatives:
| Alternative | Accuracy | Speedup |
|---|
(FPCore (alpha beta i) :precision binary64 (let* ((t_0 (+ (+ alpha beta) (* 2.0 i)))) (/ (+ (/ (/ (* (+ alpha beta) (- beta alpha)) t_0) (+ t_0 2.0)) 1.0) 2.0)))
double code(double alpha, double beta, double i) {
double t_0 = (alpha + beta) + (2.0 * i);
return (((((alpha + beta) * (beta - alpha)) / t_0) / (t_0 + 2.0)) + 1.0) / 2.0;
}
real(8) function code(alpha, beta, i)
real(8), intent (in) :: alpha
real(8), intent (in) :: beta
real(8), intent (in) :: i
real(8) :: t_0
t_0 = (alpha + beta) + (2.0d0 * i)
code = (((((alpha + beta) * (beta - alpha)) / t_0) / (t_0 + 2.0d0)) + 1.0d0) / 2.0d0
end function
public static double code(double alpha, double beta, double i) {
double t_0 = (alpha + beta) + (2.0 * i);
return (((((alpha + beta) * (beta - alpha)) / t_0) / (t_0 + 2.0)) + 1.0) / 2.0;
}
def code(alpha, beta, i): t_0 = (alpha + beta) + (2.0 * i) return (((((alpha + beta) * (beta - alpha)) / t_0) / (t_0 + 2.0)) + 1.0) / 2.0
function code(alpha, beta, i) t_0 = Float64(Float64(alpha + beta) + Float64(2.0 * i)) return Float64(Float64(Float64(Float64(Float64(Float64(alpha + beta) * Float64(beta - alpha)) / t_0) / Float64(t_0 + 2.0)) + 1.0) / 2.0) end
function tmp = code(alpha, beta, i) t_0 = (alpha + beta) + (2.0 * i); tmp = (((((alpha + beta) * (beta - alpha)) / t_0) / (t_0 + 2.0)) + 1.0) / 2.0; end
code[alpha_, beta_, i_] := Block[{t$95$0 = N[(N[(alpha + beta), $MachinePrecision] + N[(2.0 * i), $MachinePrecision]), $MachinePrecision]}, N[(N[(N[(N[(N[(N[(alpha + beta), $MachinePrecision] * N[(beta - alpha), $MachinePrecision]), $MachinePrecision] / t$95$0), $MachinePrecision] / N[(t$95$0 + 2.0), $MachinePrecision]), $MachinePrecision] + 1.0), $MachinePrecision] / 2.0), $MachinePrecision]]
\begin{array}{l}
\\
\begin{array}{l}
t_0 := \left(\alpha + \beta\right) + 2 \cdot i\\
\frac{\frac{\frac{\left(\alpha + \beta\right) \cdot \left(\beta - \alpha\right)}{t\_0}}{t\_0 + 2} + 1}{2}
\end{array}
\end{array}
(FPCore (alpha beta i)
:precision binary64
(let* ((t_0 (+ (+ alpha beta) (* 2.0 i)))
(t_1 (+ 2.0 (fma 2.0 beta (* i 4.0)))))
(if (<= (/ (/ (* (+ alpha beta) (- beta alpha)) t_0) (+ 2.0 t_0)) -0.5)
(/
(/
(-
(+ (* beta 0.0) (/ (pow beta 2.0) alpha))
(fma
-1.0
t_1
(fma
-1.0
(* (+ 2.0 (fma 2.0 i beta)) (/ (fma 2.0 i beta) alpha))
(* t_1 (/ (+ (* beta 0.0) t_1) alpha)))))
alpha)
2.0)
(/
(pow
(pow
(fma
(- beta alpha)
(/
(/ (+ alpha beta) (+ beta (fma 2.0 i alpha)))
(+ (+ alpha beta) (fma 2.0 i 2.0)))
1.0)
3.0)
0.3333333333333333)
2.0))))
double code(double alpha, double beta, double i) {
double t_0 = (alpha + beta) + (2.0 * i);
double t_1 = 2.0 + fma(2.0, beta, (i * 4.0));
double tmp;
if (((((alpha + beta) * (beta - alpha)) / t_0) / (2.0 + t_0)) <= -0.5) {
tmp = ((((beta * 0.0) + (pow(beta, 2.0) / alpha)) - fma(-1.0, t_1, fma(-1.0, ((2.0 + fma(2.0, i, beta)) * (fma(2.0, i, beta) / alpha)), (t_1 * (((beta * 0.0) + t_1) / alpha))))) / alpha) / 2.0;
} else {
tmp = pow(pow(fma((beta - alpha), (((alpha + beta) / (beta + fma(2.0, i, alpha))) / ((alpha + beta) + fma(2.0, i, 2.0))), 1.0), 3.0), 0.3333333333333333) / 2.0;
}
return tmp;
}
function code(alpha, beta, i) t_0 = Float64(Float64(alpha + beta) + Float64(2.0 * i)) t_1 = Float64(2.0 + fma(2.0, beta, Float64(i * 4.0))) tmp = 0.0 if (Float64(Float64(Float64(Float64(alpha + beta) * Float64(beta - alpha)) / t_0) / Float64(2.0 + t_0)) <= -0.5) tmp = Float64(Float64(Float64(Float64(Float64(beta * 0.0) + Float64((beta ^ 2.0) / alpha)) - fma(-1.0, t_1, fma(-1.0, Float64(Float64(2.0 + fma(2.0, i, beta)) * Float64(fma(2.0, i, beta) / alpha)), Float64(t_1 * Float64(Float64(Float64(beta * 0.0) + t_1) / alpha))))) / alpha) / 2.0); else tmp = Float64(((fma(Float64(beta - alpha), Float64(Float64(Float64(alpha + beta) / Float64(beta + fma(2.0, i, alpha))) / Float64(Float64(alpha + beta) + fma(2.0, i, 2.0))), 1.0) ^ 3.0) ^ 0.3333333333333333) / 2.0); end return tmp end
code[alpha_, beta_, i_] := Block[{t$95$0 = N[(N[(alpha + beta), $MachinePrecision] + N[(2.0 * i), $MachinePrecision]), $MachinePrecision]}, Block[{t$95$1 = N[(2.0 + N[(2.0 * beta + N[(i * 4.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]}, If[LessEqual[N[(N[(N[(N[(alpha + beta), $MachinePrecision] * N[(beta - alpha), $MachinePrecision]), $MachinePrecision] / t$95$0), $MachinePrecision] / N[(2.0 + t$95$0), $MachinePrecision]), $MachinePrecision], -0.5], N[(N[(N[(N[(N[(beta * 0.0), $MachinePrecision] + N[(N[Power[beta, 2.0], $MachinePrecision] / alpha), $MachinePrecision]), $MachinePrecision] - N[(-1.0 * t$95$1 + N[(-1.0 * N[(N[(2.0 + N[(2.0 * i + beta), $MachinePrecision]), $MachinePrecision] * N[(N[(2.0 * i + beta), $MachinePrecision] / alpha), $MachinePrecision]), $MachinePrecision] + N[(t$95$1 * N[(N[(N[(beta * 0.0), $MachinePrecision] + t$95$1), $MachinePrecision] / alpha), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] / alpha), $MachinePrecision] / 2.0), $MachinePrecision], N[(N[Power[N[Power[N[(N[(beta - alpha), $MachinePrecision] * N[(N[(N[(alpha + beta), $MachinePrecision] / N[(beta + N[(2.0 * i + alpha), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] / N[(N[(alpha + beta), $MachinePrecision] + N[(2.0 * i + 2.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] + 1.0), $MachinePrecision], 3.0], $MachinePrecision], 0.3333333333333333], $MachinePrecision] / 2.0), $MachinePrecision]]]]
\begin{array}{l}
\\
\begin{array}{l}
t_0 := \left(\alpha + \beta\right) + 2 \cdot i\\
t_1 := 2 + \mathsf{fma}\left(2, \beta, i \cdot 4\right)\\
\mathbf{if}\;\frac{\frac{\left(\alpha + \beta\right) \cdot \left(\beta - \alpha\right)}{t\_0}}{2 + t\_0} \leq -0.5:\\
\;\;\;\;\frac{\frac{\left(\beta \cdot 0 + \frac{{\beta}^{2}}{\alpha}\right) - \mathsf{fma}\left(-1, t\_1, \mathsf{fma}\left(-1, \left(2 + \mathsf{fma}\left(2, i, \beta\right)\right) \cdot \frac{\mathsf{fma}\left(2, i, \beta\right)}{\alpha}, t\_1 \cdot \frac{\beta \cdot 0 + t\_1}{\alpha}\right)\right)}{\alpha}}{2}\\
\mathbf{else}:\\
\;\;\;\;\frac{{\left({\left(\mathsf{fma}\left(\beta - \alpha, \frac{\frac{\alpha + \beta}{\beta + \mathsf{fma}\left(2, i, \alpha\right)}}{\left(\alpha + \beta\right) + \mathsf{fma}\left(2, i, 2\right)}, 1\right)\right)}^{3}\right)}^{0.3333333333333333}}{2}\\
\end{array}
\end{array}
if (/.f64 (/.f64 (*.f64 (+.f64 alpha beta) (-.f64 beta alpha)) (+.f64 (+.f64 alpha beta) (*.f64 #s(literal 2 binary64) i))) (+.f64 (+.f64 (+.f64 alpha beta) (*.f64 #s(literal 2 binary64) i)) #s(literal 2 binary64))) < -0.5Initial program 5.5%
Simplified16.1%
Taylor expanded in alpha around inf 79.0%
Simplified94.0%
if -0.5 < (/.f64 (/.f64 (*.f64 (+.f64 alpha beta) (-.f64 beta alpha)) (+.f64 (+.f64 alpha beta) (*.f64 #s(literal 2 binary64) i))) (+.f64 (+.f64 (+.f64 alpha beta) (*.f64 #s(literal 2 binary64) i)) #s(literal 2 binary64))) Initial program 78.2%
Simplified100.0%
Applied egg-rr100.0%
Final simplification99.0%
(FPCore (alpha beta i)
:precision binary64
(let* ((t_0 (+ beta (* 2.0 i)))
(t_1 (fma 2.0 beta (* i 4.0)))
(t_2 (+ 2.0 t_1))
(t_3 (+ (+ alpha beta) (* 2.0 i))))
(if (<= (/ (/ (* (+ alpha beta) (- beta alpha)) t_3) (+ 2.0 t_3)) -0.5)
(/
(/
(+
(/ (pow beta 2.0) alpha)
(-
(+ (- t_1 -2.0) (* (/ t_0 alpha) (- t_0 -2.0)))
(* t_2 (/ t_2 alpha))))
alpha)
2.0)
(/
(pow
(pow
(fma
(- beta alpha)
(/
(/ (+ alpha beta) (+ beta (fma 2.0 i alpha)))
(+ (+ alpha beta) (fma 2.0 i 2.0)))
1.0)
3.0)
0.3333333333333333)
2.0))))
double code(double alpha, double beta, double i) {
double t_0 = beta + (2.0 * i);
double t_1 = fma(2.0, beta, (i * 4.0));
double t_2 = 2.0 + t_1;
double t_3 = (alpha + beta) + (2.0 * i);
double tmp;
if (((((alpha + beta) * (beta - alpha)) / t_3) / (2.0 + t_3)) <= -0.5) {
tmp = (((pow(beta, 2.0) / alpha) + (((t_1 - -2.0) + ((t_0 / alpha) * (t_0 - -2.0))) - (t_2 * (t_2 / alpha)))) / alpha) / 2.0;
} else {
tmp = pow(pow(fma((beta - alpha), (((alpha + beta) / (beta + fma(2.0, i, alpha))) / ((alpha + beta) + fma(2.0, i, 2.0))), 1.0), 3.0), 0.3333333333333333) / 2.0;
}
return tmp;
}
function code(alpha, beta, i) t_0 = Float64(beta + Float64(2.0 * i)) t_1 = fma(2.0, beta, Float64(i * 4.0)) t_2 = Float64(2.0 + t_1) t_3 = Float64(Float64(alpha + beta) + Float64(2.0 * i)) tmp = 0.0 if (Float64(Float64(Float64(Float64(alpha + beta) * Float64(beta - alpha)) / t_3) / Float64(2.0 + t_3)) <= -0.5) tmp = Float64(Float64(Float64(Float64((beta ^ 2.0) / alpha) + Float64(Float64(Float64(t_1 - -2.0) + Float64(Float64(t_0 / alpha) * Float64(t_0 - -2.0))) - Float64(t_2 * Float64(t_2 / alpha)))) / alpha) / 2.0); else tmp = Float64(((fma(Float64(beta - alpha), Float64(Float64(Float64(alpha + beta) / Float64(beta + fma(2.0, i, alpha))) / Float64(Float64(alpha + beta) + fma(2.0, i, 2.0))), 1.0) ^ 3.0) ^ 0.3333333333333333) / 2.0); end return tmp end
code[alpha_, beta_, i_] := Block[{t$95$0 = N[(beta + N[(2.0 * i), $MachinePrecision]), $MachinePrecision]}, Block[{t$95$1 = N[(2.0 * beta + N[(i * 4.0), $MachinePrecision]), $MachinePrecision]}, Block[{t$95$2 = N[(2.0 + t$95$1), $MachinePrecision]}, Block[{t$95$3 = N[(N[(alpha + beta), $MachinePrecision] + N[(2.0 * i), $MachinePrecision]), $MachinePrecision]}, If[LessEqual[N[(N[(N[(N[(alpha + beta), $MachinePrecision] * N[(beta - alpha), $MachinePrecision]), $MachinePrecision] / t$95$3), $MachinePrecision] / N[(2.0 + t$95$3), $MachinePrecision]), $MachinePrecision], -0.5], N[(N[(N[(N[(N[Power[beta, 2.0], $MachinePrecision] / alpha), $MachinePrecision] + N[(N[(N[(t$95$1 - -2.0), $MachinePrecision] + N[(N[(t$95$0 / alpha), $MachinePrecision] * N[(t$95$0 - -2.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] - N[(t$95$2 * N[(t$95$2 / alpha), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] / alpha), $MachinePrecision] / 2.0), $MachinePrecision], N[(N[Power[N[Power[N[(N[(beta - alpha), $MachinePrecision] * N[(N[(N[(alpha + beta), $MachinePrecision] / N[(beta + N[(2.0 * i + alpha), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] / N[(N[(alpha + beta), $MachinePrecision] + N[(2.0 * i + 2.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] + 1.0), $MachinePrecision], 3.0], $MachinePrecision], 0.3333333333333333], $MachinePrecision] / 2.0), $MachinePrecision]]]]]]
\begin{array}{l}
\\
\begin{array}{l}
t_0 := \beta + 2 \cdot i\\
t_1 := \mathsf{fma}\left(2, \beta, i \cdot 4\right)\\
t_2 := 2 + t\_1\\
t_3 := \left(\alpha + \beta\right) + 2 \cdot i\\
\mathbf{if}\;\frac{\frac{\left(\alpha + \beta\right) \cdot \left(\beta - \alpha\right)}{t\_3}}{2 + t\_3} \leq -0.5:\\
\;\;\;\;\frac{\frac{\frac{{\beta}^{2}}{\alpha} + \left(\left(\left(t\_1 - -2\right) + \frac{t\_0}{\alpha} \cdot \left(t\_0 - -2\right)\right) - t\_2 \cdot \frac{t\_2}{\alpha}\right)}{\alpha}}{2}\\
\mathbf{else}:\\
\;\;\;\;\frac{{\left({\left(\mathsf{fma}\left(\beta - \alpha, \frac{\frac{\alpha + \beta}{\beta + \mathsf{fma}\left(2, i, \alpha\right)}}{\left(\alpha + \beta\right) + \mathsf{fma}\left(2, i, 2\right)}, 1\right)\right)}^{3}\right)}^{0.3333333333333333}}{2}\\
\end{array}
\end{array}
if (/.f64 (/.f64 (*.f64 (+.f64 alpha beta) (-.f64 beta alpha)) (+.f64 (+.f64 alpha beta) (*.f64 #s(literal 2 binary64) i))) (+.f64 (+.f64 (+.f64 alpha beta) (*.f64 #s(literal 2 binary64) i)) #s(literal 2 binary64))) < -0.5Initial program 5.5%
Simplified16.1%
associate-*r/5.5%
*-commutative5.5%
fma-undefine5.5%
+-commutative5.5%
fma-define5.5%
associate-+r+5.5%
associate-/l/4.6%
+-commutative4.6%
associate-+r+4.6%
times-frac16.1%
+-commutative16.1%
+-commutative16.1%
fma-define16.1%
Applied egg-rr16.1%
log1p-expm1-u16.1%
Applied egg-rr16.1%
Taylor expanded in alpha around inf 79.0%
Simplified94.0%
if -0.5 < (/.f64 (/.f64 (*.f64 (+.f64 alpha beta) (-.f64 beta alpha)) (+.f64 (+.f64 alpha beta) (*.f64 #s(literal 2 binary64) i))) (+.f64 (+.f64 (+.f64 alpha beta) (*.f64 #s(literal 2 binary64) i)) #s(literal 2 binary64))) Initial program 78.2%
Simplified100.0%
Applied egg-rr100.0%
Final simplification99.0%
(FPCore (alpha beta i)
:precision binary64
(let* ((t_0 (+ beta (* 2.0 i)))
(t_1 (fma 2.0 beta (* i 4.0)))
(t_2 (+ 2.0 t_1))
(t_3 (+ (+ alpha beta) (* 2.0 i))))
(if (<= (/ (/ (* (+ alpha beta) (- beta alpha)) t_3) (+ 2.0 t_3)) -0.5)
(/
(/
(+
(/ (pow beta 2.0) alpha)
(-
(+ (- t_1 -2.0) (* (/ t_0 alpha) (- t_0 -2.0)))
(* t_2 (/ t_2 alpha))))
alpha)
2.0)
(/
(+
1.0
(/
(* (- beta alpha) (/ (+ alpha beta) (fma 2.0 i (+ alpha beta))))
(+ alpha (+ beta (fma 2.0 i 2.0)))))
2.0))))
double code(double alpha, double beta, double i) {
double t_0 = beta + (2.0 * i);
double t_1 = fma(2.0, beta, (i * 4.0));
double t_2 = 2.0 + t_1;
double t_3 = (alpha + beta) + (2.0 * i);
double tmp;
if (((((alpha + beta) * (beta - alpha)) / t_3) / (2.0 + t_3)) <= -0.5) {
tmp = (((pow(beta, 2.0) / alpha) + (((t_1 - -2.0) + ((t_0 / alpha) * (t_0 - -2.0))) - (t_2 * (t_2 / alpha)))) / alpha) / 2.0;
} else {
tmp = (1.0 + (((beta - alpha) * ((alpha + beta) / fma(2.0, i, (alpha + beta)))) / (alpha + (beta + fma(2.0, i, 2.0))))) / 2.0;
}
return tmp;
}
function code(alpha, beta, i) t_0 = Float64(beta + Float64(2.0 * i)) t_1 = fma(2.0, beta, Float64(i * 4.0)) t_2 = Float64(2.0 + t_1) t_3 = Float64(Float64(alpha + beta) + Float64(2.0 * i)) tmp = 0.0 if (Float64(Float64(Float64(Float64(alpha + beta) * Float64(beta - alpha)) / t_3) / Float64(2.0 + t_3)) <= -0.5) tmp = Float64(Float64(Float64(Float64((beta ^ 2.0) / alpha) + Float64(Float64(Float64(t_1 - -2.0) + Float64(Float64(t_0 / alpha) * Float64(t_0 - -2.0))) - Float64(t_2 * Float64(t_2 / alpha)))) / alpha) / 2.0); else tmp = Float64(Float64(1.0 + Float64(Float64(Float64(beta - alpha) * Float64(Float64(alpha + beta) / fma(2.0, i, Float64(alpha + beta)))) / Float64(alpha + Float64(beta + fma(2.0, i, 2.0))))) / 2.0); end return tmp end
code[alpha_, beta_, i_] := Block[{t$95$0 = N[(beta + N[(2.0 * i), $MachinePrecision]), $MachinePrecision]}, Block[{t$95$1 = N[(2.0 * beta + N[(i * 4.0), $MachinePrecision]), $MachinePrecision]}, Block[{t$95$2 = N[(2.0 + t$95$1), $MachinePrecision]}, Block[{t$95$3 = N[(N[(alpha + beta), $MachinePrecision] + N[(2.0 * i), $MachinePrecision]), $MachinePrecision]}, If[LessEqual[N[(N[(N[(N[(alpha + beta), $MachinePrecision] * N[(beta - alpha), $MachinePrecision]), $MachinePrecision] / t$95$3), $MachinePrecision] / N[(2.0 + t$95$3), $MachinePrecision]), $MachinePrecision], -0.5], N[(N[(N[(N[(N[Power[beta, 2.0], $MachinePrecision] / alpha), $MachinePrecision] + N[(N[(N[(t$95$1 - -2.0), $MachinePrecision] + N[(N[(t$95$0 / alpha), $MachinePrecision] * N[(t$95$0 - -2.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] - N[(t$95$2 * N[(t$95$2 / alpha), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] / alpha), $MachinePrecision] / 2.0), $MachinePrecision], N[(N[(1.0 + N[(N[(N[(beta - alpha), $MachinePrecision] * N[(N[(alpha + beta), $MachinePrecision] / N[(2.0 * i + N[(alpha + beta), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] / N[(alpha + N[(beta + N[(2.0 * i + 2.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] / 2.0), $MachinePrecision]]]]]]
\begin{array}{l}
\\
\begin{array}{l}
t_0 := \beta + 2 \cdot i\\
t_1 := \mathsf{fma}\left(2, \beta, i \cdot 4\right)\\
t_2 := 2 + t\_1\\
t_3 := \left(\alpha + \beta\right) + 2 \cdot i\\
\mathbf{if}\;\frac{\frac{\left(\alpha + \beta\right) \cdot \left(\beta - \alpha\right)}{t\_3}}{2 + t\_3} \leq -0.5:\\
\;\;\;\;\frac{\frac{\frac{{\beta}^{2}}{\alpha} + \left(\left(\left(t\_1 - -2\right) + \frac{t\_0}{\alpha} \cdot \left(t\_0 - -2\right)\right) - t\_2 \cdot \frac{t\_2}{\alpha}\right)}{\alpha}}{2}\\
\mathbf{else}:\\
\;\;\;\;\frac{1 + \frac{\left(\beta - \alpha\right) \cdot \frac{\alpha + \beta}{\mathsf{fma}\left(2, i, \alpha + \beta\right)}}{\alpha + \left(\beta + \mathsf{fma}\left(2, i, 2\right)\right)}}{2}\\
\end{array}
\end{array}
if (/.f64 (/.f64 (*.f64 (+.f64 alpha beta) (-.f64 beta alpha)) (+.f64 (+.f64 alpha beta) (*.f64 #s(literal 2 binary64) i))) (+.f64 (+.f64 (+.f64 alpha beta) (*.f64 #s(literal 2 binary64) i)) #s(literal 2 binary64))) < -0.5Initial program 5.5%
Simplified16.1%
associate-*r/5.5%
*-commutative5.5%
fma-undefine5.5%
+-commutative5.5%
fma-define5.5%
associate-+r+5.5%
associate-/l/4.6%
+-commutative4.6%
associate-+r+4.6%
times-frac16.1%
+-commutative16.1%
+-commutative16.1%
fma-define16.1%
Applied egg-rr16.1%
log1p-expm1-u16.1%
Applied egg-rr16.1%
Taylor expanded in alpha around inf 79.0%
Simplified94.0%
if -0.5 < (/.f64 (/.f64 (*.f64 (+.f64 alpha beta) (-.f64 beta alpha)) (+.f64 (+.f64 alpha beta) (*.f64 #s(literal 2 binary64) i))) (+.f64 (+.f64 (+.f64 alpha beta) (*.f64 #s(literal 2 binary64) i)) #s(literal 2 binary64))) Initial program 78.2%
Simplified100.0%
Final simplification99.0%
(FPCore (alpha beta i)
:precision binary64
(let* ((t_0 (+ (+ alpha beta) (* 2.0 i))))
(if (<= (/ (/ (* (+ alpha beta) (- beta alpha)) t_0) (+ 2.0 t_0)) -0.5)
(/ (+ (* 2.0 (/ beta alpha)) (/ (+ 2.0 (* i 4.0)) alpha)) 2.0)
(/
(+
1.0
(/
(* (- beta alpha) (/ (+ alpha beta) (fma 2.0 i (+ alpha beta))))
(+ alpha (+ beta (fma 2.0 i 2.0)))))
2.0))))
double code(double alpha, double beta, double i) {
double t_0 = (alpha + beta) + (2.0 * i);
double tmp;
if (((((alpha + beta) * (beta - alpha)) / t_0) / (2.0 + t_0)) <= -0.5) {
tmp = ((2.0 * (beta / alpha)) + ((2.0 + (i * 4.0)) / alpha)) / 2.0;
} else {
tmp = (1.0 + (((beta - alpha) * ((alpha + beta) / fma(2.0, i, (alpha + beta)))) / (alpha + (beta + fma(2.0, i, 2.0))))) / 2.0;
}
return tmp;
}
function code(alpha, beta, i) t_0 = Float64(Float64(alpha + beta) + Float64(2.0 * i)) tmp = 0.0 if (Float64(Float64(Float64(Float64(alpha + beta) * Float64(beta - alpha)) / t_0) / Float64(2.0 + t_0)) <= -0.5) tmp = Float64(Float64(Float64(2.0 * Float64(beta / alpha)) + Float64(Float64(2.0 + Float64(i * 4.0)) / alpha)) / 2.0); else tmp = Float64(Float64(1.0 + Float64(Float64(Float64(beta - alpha) * Float64(Float64(alpha + beta) / fma(2.0, i, Float64(alpha + beta)))) / Float64(alpha + Float64(beta + fma(2.0, i, 2.0))))) / 2.0); end return tmp end
code[alpha_, beta_, i_] := Block[{t$95$0 = N[(N[(alpha + beta), $MachinePrecision] + N[(2.0 * i), $MachinePrecision]), $MachinePrecision]}, If[LessEqual[N[(N[(N[(N[(alpha + beta), $MachinePrecision] * N[(beta - alpha), $MachinePrecision]), $MachinePrecision] / t$95$0), $MachinePrecision] / N[(2.0 + t$95$0), $MachinePrecision]), $MachinePrecision], -0.5], N[(N[(N[(2.0 * N[(beta / alpha), $MachinePrecision]), $MachinePrecision] + N[(N[(2.0 + N[(i * 4.0), $MachinePrecision]), $MachinePrecision] / alpha), $MachinePrecision]), $MachinePrecision] / 2.0), $MachinePrecision], N[(N[(1.0 + N[(N[(N[(beta - alpha), $MachinePrecision] * N[(N[(alpha + beta), $MachinePrecision] / N[(2.0 * i + N[(alpha + beta), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] / N[(alpha + N[(beta + N[(2.0 * i + 2.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] / 2.0), $MachinePrecision]]]
\begin{array}{l}
\\
\begin{array}{l}
t_0 := \left(\alpha + \beta\right) + 2 \cdot i\\
\mathbf{if}\;\frac{\frac{\left(\alpha + \beta\right) \cdot \left(\beta - \alpha\right)}{t\_0}}{2 + t\_0} \leq -0.5:\\
\;\;\;\;\frac{2 \cdot \frac{\beta}{\alpha} + \frac{2 + i \cdot 4}{\alpha}}{2}\\
\mathbf{else}:\\
\;\;\;\;\frac{1 + \frac{\left(\beta - \alpha\right) \cdot \frac{\alpha + \beta}{\mathsf{fma}\left(2, i, \alpha + \beta\right)}}{\alpha + \left(\beta + \mathsf{fma}\left(2, i, 2\right)\right)}}{2}\\
\end{array}
\end{array}
if (/.f64 (/.f64 (*.f64 (+.f64 alpha beta) (-.f64 beta alpha)) (+.f64 (+.f64 alpha beta) (*.f64 #s(literal 2 binary64) i))) (+.f64 (+.f64 (+.f64 alpha beta) (*.f64 #s(literal 2 binary64) i)) #s(literal 2 binary64))) < -0.5Initial program 5.5%
Simplified16.1%
Taylor expanded in alpha around inf 10.4%
Taylor expanded in beta around 0 91.2%
mul-1-neg91.2%
*-commutative91.2%
Simplified91.2%
if -0.5 < (/.f64 (/.f64 (*.f64 (+.f64 alpha beta) (-.f64 beta alpha)) (+.f64 (+.f64 alpha beta) (*.f64 #s(literal 2 binary64) i))) (+.f64 (+.f64 (+.f64 alpha beta) (*.f64 #s(literal 2 binary64) i)) #s(literal 2 binary64))) Initial program 78.2%
Simplified100.0%
Final simplification98.5%
(FPCore (alpha beta i)
:precision binary64
(let* ((t_0 (+ (+ alpha beta) (* 2.0 i))))
(if (<= (/ (/ (* (+ alpha beta) (- beta alpha)) t_0) (+ 2.0 t_0)) -0.5)
(/ (+ (* 2.0 (/ beta alpha)) (/ (+ 2.0 (* i 4.0)) alpha)) 2.0)
(/
(+
1.0
(*
(/ (+ alpha beta) (+ (+ alpha beta) (fma 2.0 i 2.0)))
(/ (- beta alpha) (+ beta (fma 2.0 i alpha)))))
2.0))))
double code(double alpha, double beta, double i) {
double t_0 = (alpha + beta) + (2.0 * i);
double tmp;
if (((((alpha + beta) * (beta - alpha)) / t_0) / (2.0 + t_0)) <= -0.5) {
tmp = ((2.0 * (beta / alpha)) + ((2.0 + (i * 4.0)) / alpha)) / 2.0;
} else {
tmp = (1.0 + (((alpha + beta) / ((alpha + beta) + fma(2.0, i, 2.0))) * ((beta - alpha) / (beta + fma(2.0, i, alpha))))) / 2.0;
}
return tmp;
}
function code(alpha, beta, i) t_0 = Float64(Float64(alpha + beta) + Float64(2.0 * i)) tmp = 0.0 if (Float64(Float64(Float64(Float64(alpha + beta) * Float64(beta - alpha)) / t_0) / Float64(2.0 + t_0)) <= -0.5) tmp = Float64(Float64(Float64(2.0 * Float64(beta / alpha)) + Float64(Float64(2.0 + Float64(i * 4.0)) / alpha)) / 2.0); else tmp = Float64(Float64(1.0 + Float64(Float64(Float64(alpha + beta) / Float64(Float64(alpha + beta) + fma(2.0, i, 2.0))) * Float64(Float64(beta - alpha) / Float64(beta + fma(2.0, i, alpha))))) / 2.0); end return tmp end
code[alpha_, beta_, i_] := Block[{t$95$0 = N[(N[(alpha + beta), $MachinePrecision] + N[(2.0 * i), $MachinePrecision]), $MachinePrecision]}, If[LessEqual[N[(N[(N[(N[(alpha + beta), $MachinePrecision] * N[(beta - alpha), $MachinePrecision]), $MachinePrecision] / t$95$0), $MachinePrecision] / N[(2.0 + t$95$0), $MachinePrecision]), $MachinePrecision], -0.5], N[(N[(N[(2.0 * N[(beta / alpha), $MachinePrecision]), $MachinePrecision] + N[(N[(2.0 + N[(i * 4.0), $MachinePrecision]), $MachinePrecision] / alpha), $MachinePrecision]), $MachinePrecision] / 2.0), $MachinePrecision], N[(N[(1.0 + N[(N[(N[(alpha + beta), $MachinePrecision] / N[(N[(alpha + beta), $MachinePrecision] + N[(2.0 * i + 2.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] * N[(N[(beta - alpha), $MachinePrecision] / N[(beta + N[(2.0 * i + alpha), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] / 2.0), $MachinePrecision]]]
\begin{array}{l}
\\
\begin{array}{l}
t_0 := \left(\alpha + \beta\right) + 2 \cdot i\\
\mathbf{if}\;\frac{\frac{\left(\alpha + \beta\right) \cdot \left(\beta - \alpha\right)}{t\_0}}{2 + t\_0} \leq -0.5:\\
\;\;\;\;\frac{2 \cdot \frac{\beta}{\alpha} + \frac{2 + i \cdot 4}{\alpha}}{2}\\
\mathbf{else}:\\
\;\;\;\;\frac{1 + \frac{\alpha + \beta}{\left(\alpha + \beta\right) + \mathsf{fma}\left(2, i, 2\right)} \cdot \frac{\beta - \alpha}{\beta + \mathsf{fma}\left(2, i, \alpha\right)}}{2}\\
\end{array}
\end{array}
if (/.f64 (/.f64 (*.f64 (+.f64 alpha beta) (-.f64 beta alpha)) (+.f64 (+.f64 alpha beta) (*.f64 #s(literal 2 binary64) i))) (+.f64 (+.f64 (+.f64 alpha beta) (*.f64 #s(literal 2 binary64) i)) #s(literal 2 binary64))) < -0.5Initial program 5.5%
Simplified16.1%
Taylor expanded in alpha around inf 10.4%
Taylor expanded in beta around 0 91.2%
mul-1-neg91.2%
*-commutative91.2%
Simplified91.2%
if -0.5 < (/.f64 (/.f64 (*.f64 (+.f64 alpha beta) (-.f64 beta alpha)) (+.f64 (+.f64 alpha beta) (*.f64 #s(literal 2 binary64) i))) (+.f64 (+.f64 (+.f64 alpha beta) (*.f64 #s(literal 2 binary64) i)) #s(literal 2 binary64))) Initial program 78.2%
Simplified100.0%
associate-*r/78.2%
*-commutative78.2%
fma-undefine78.2%
+-commutative78.2%
fma-define78.2%
associate-+r+78.2%
associate-/l/77.2%
+-commutative77.2%
associate-+r+77.2%
times-frac100.0%
+-commutative100.0%
+-commutative100.0%
fma-define100.0%
Applied egg-rr100.0%
Final simplification98.5%
(FPCore (alpha beta i)
:precision binary64
(let* ((t_0 (+ (+ alpha beta) (* 2.0 i))))
(if (<= (/ (/ (* (+ alpha beta) (- beta alpha)) t_0) (+ 2.0 t_0)) -0.5)
(/ (+ (* 2.0 (/ beta alpha)) (/ (+ 2.0 (* i 4.0)) alpha)) 2.0)
(/
(+
1.0
(/
(* (- beta alpha) (/ beta (+ beta (* 2.0 i))))
(+ alpha (+ beta (fma 2.0 i 2.0)))))
2.0))))
double code(double alpha, double beta, double i) {
double t_0 = (alpha + beta) + (2.0 * i);
double tmp;
if (((((alpha + beta) * (beta - alpha)) / t_0) / (2.0 + t_0)) <= -0.5) {
tmp = ((2.0 * (beta / alpha)) + ((2.0 + (i * 4.0)) / alpha)) / 2.0;
} else {
tmp = (1.0 + (((beta - alpha) * (beta / (beta + (2.0 * i)))) / (alpha + (beta + fma(2.0, i, 2.0))))) / 2.0;
}
return tmp;
}
function code(alpha, beta, i) t_0 = Float64(Float64(alpha + beta) + Float64(2.0 * i)) tmp = 0.0 if (Float64(Float64(Float64(Float64(alpha + beta) * Float64(beta - alpha)) / t_0) / Float64(2.0 + t_0)) <= -0.5) tmp = Float64(Float64(Float64(2.0 * Float64(beta / alpha)) + Float64(Float64(2.0 + Float64(i * 4.0)) / alpha)) / 2.0); else tmp = Float64(Float64(1.0 + Float64(Float64(Float64(beta - alpha) * Float64(beta / Float64(beta + Float64(2.0 * i)))) / Float64(alpha + Float64(beta + fma(2.0, i, 2.0))))) / 2.0); end return tmp end
code[alpha_, beta_, i_] := Block[{t$95$0 = N[(N[(alpha + beta), $MachinePrecision] + N[(2.0 * i), $MachinePrecision]), $MachinePrecision]}, If[LessEqual[N[(N[(N[(N[(alpha + beta), $MachinePrecision] * N[(beta - alpha), $MachinePrecision]), $MachinePrecision] / t$95$0), $MachinePrecision] / N[(2.0 + t$95$0), $MachinePrecision]), $MachinePrecision], -0.5], N[(N[(N[(2.0 * N[(beta / alpha), $MachinePrecision]), $MachinePrecision] + N[(N[(2.0 + N[(i * 4.0), $MachinePrecision]), $MachinePrecision] / alpha), $MachinePrecision]), $MachinePrecision] / 2.0), $MachinePrecision], N[(N[(1.0 + N[(N[(N[(beta - alpha), $MachinePrecision] * N[(beta / N[(beta + N[(2.0 * i), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] / N[(alpha + N[(beta + N[(2.0 * i + 2.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] / 2.0), $MachinePrecision]]]
\begin{array}{l}
\\
\begin{array}{l}
t_0 := \left(\alpha + \beta\right) + 2 \cdot i\\
\mathbf{if}\;\frac{\frac{\left(\alpha + \beta\right) \cdot \left(\beta - \alpha\right)}{t\_0}}{2 + t\_0} \leq -0.5:\\
\;\;\;\;\frac{2 \cdot \frac{\beta}{\alpha} + \frac{2 + i \cdot 4}{\alpha}}{2}\\
\mathbf{else}:\\
\;\;\;\;\frac{1 + \frac{\left(\beta - \alpha\right) \cdot \frac{\beta}{\beta + 2 \cdot i}}{\alpha + \left(\beta + \mathsf{fma}\left(2, i, 2\right)\right)}}{2}\\
\end{array}
\end{array}
if (/.f64 (/.f64 (*.f64 (+.f64 alpha beta) (-.f64 beta alpha)) (+.f64 (+.f64 alpha beta) (*.f64 #s(literal 2 binary64) i))) (+.f64 (+.f64 (+.f64 alpha beta) (*.f64 #s(literal 2 binary64) i)) #s(literal 2 binary64))) < -0.5Initial program 5.5%
Simplified16.1%
Taylor expanded in alpha around inf 10.4%
Taylor expanded in beta around 0 91.2%
mul-1-neg91.2%
*-commutative91.2%
Simplified91.2%
if -0.5 < (/.f64 (/.f64 (*.f64 (+.f64 alpha beta) (-.f64 beta alpha)) (+.f64 (+.f64 alpha beta) (*.f64 #s(literal 2 binary64) i))) (+.f64 (+.f64 (+.f64 alpha beta) (*.f64 #s(literal 2 binary64) i)) #s(literal 2 binary64))) Initial program 78.2%
Simplified100.0%
Taylor expanded in alpha around 0 99.2%
Final simplification97.9%
(FPCore (alpha beta i)
:precision binary64
(let* ((t_0 (+ (+ alpha beta) (* 2.0 i)))
(t_1 (+ 2.0 t_0))
(t_2 (/ (/ (* (+ alpha beta) (- beta alpha)) t_0) t_1)))
(if (<= t_2 -0.5)
(/ (+ (* 2.0 (/ beta alpha)) (/ (+ 2.0 (* i 4.0)) alpha)) 2.0)
(if (<= t_2 1.0)
(/ (- 1.0 (/ (/ (* (+ alpha beta) (- alpha beta)) t_0) t_1)) 2.0)
(/ (- 2.0 (/ (* alpha 2.0) beta)) 2.0)))))
double code(double alpha, double beta, double i) {
double t_0 = (alpha + beta) + (2.0 * i);
double t_1 = 2.0 + t_0;
double t_2 = (((alpha + beta) * (beta - alpha)) / t_0) / t_1;
double tmp;
if (t_2 <= -0.5) {
tmp = ((2.0 * (beta / alpha)) + ((2.0 + (i * 4.0)) / alpha)) / 2.0;
} else if (t_2 <= 1.0) {
tmp = (1.0 - ((((alpha + beta) * (alpha - beta)) / t_0) / t_1)) / 2.0;
} else {
tmp = (2.0 - ((alpha * 2.0) / beta)) / 2.0;
}
return tmp;
}
real(8) function code(alpha, beta, i)
real(8), intent (in) :: alpha
real(8), intent (in) :: beta
real(8), intent (in) :: i
real(8) :: t_0
real(8) :: t_1
real(8) :: t_2
real(8) :: tmp
t_0 = (alpha + beta) + (2.0d0 * i)
t_1 = 2.0d0 + t_0
t_2 = (((alpha + beta) * (beta - alpha)) / t_0) / t_1
if (t_2 <= (-0.5d0)) then
tmp = ((2.0d0 * (beta / alpha)) + ((2.0d0 + (i * 4.0d0)) / alpha)) / 2.0d0
else if (t_2 <= 1.0d0) then
tmp = (1.0d0 - ((((alpha + beta) * (alpha - beta)) / t_0) / t_1)) / 2.0d0
else
tmp = (2.0d0 - ((alpha * 2.0d0) / beta)) / 2.0d0
end if
code = tmp
end function
public static double code(double alpha, double beta, double i) {
double t_0 = (alpha + beta) + (2.0 * i);
double t_1 = 2.0 + t_0;
double t_2 = (((alpha + beta) * (beta - alpha)) / t_0) / t_1;
double tmp;
if (t_2 <= -0.5) {
tmp = ((2.0 * (beta / alpha)) + ((2.0 + (i * 4.0)) / alpha)) / 2.0;
} else if (t_2 <= 1.0) {
tmp = (1.0 - ((((alpha + beta) * (alpha - beta)) / t_0) / t_1)) / 2.0;
} else {
tmp = (2.0 - ((alpha * 2.0) / beta)) / 2.0;
}
return tmp;
}
def code(alpha, beta, i): t_0 = (alpha + beta) + (2.0 * i) t_1 = 2.0 + t_0 t_2 = (((alpha + beta) * (beta - alpha)) / t_0) / t_1 tmp = 0 if t_2 <= -0.5: tmp = ((2.0 * (beta / alpha)) + ((2.0 + (i * 4.0)) / alpha)) / 2.0 elif t_2 <= 1.0: tmp = (1.0 - ((((alpha + beta) * (alpha - beta)) / t_0) / t_1)) / 2.0 else: tmp = (2.0 - ((alpha * 2.0) / beta)) / 2.0 return tmp
function code(alpha, beta, i) t_0 = Float64(Float64(alpha + beta) + Float64(2.0 * i)) t_1 = Float64(2.0 + t_0) t_2 = Float64(Float64(Float64(Float64(alpha + beta) * Float64(beta - alpha)) / t_0) / t_1) tmp = 0.0 if (t_2 <= -0.5) tmp = Float64(Float64(Float64(2.0 * Float64(beta / alpha)) + Float64(Float64(2.0 + Float64(i * 4.0)) / alpha)) / 2.0); elseif (t_2 <= 1.0) tmp = Float64(Float64(1.0 - Float64(Float64(Float64(Float64(alpha + beta) * Float64(alpha - beta)) / t_0) / t_1)) / 2.0); else tmp = Float64(Float64(2.0 - Float64(Float64(alpha * 2.0) / beta)) / 2.0); end return tmp end
function tmp_2 = code(alpha, beta, i) t_0 = (alpha + beta) + (2.0 * i); t_1 = 2.0 + t_0; t_2 = (((alpha + beta) * (beta - alpha)) / t_0) / t_1; tmp = 0.0; if (t_2 <= -0.5) tmp = ((2.0 * (beta / alpha)) + ((2.0 + (i * 4.0)) / alpha)) / 2.0; elseif (t_2 <= 1.0) tmp = (1.0 - ((((alpha + beta) * (alpha - beta)) / t_0) / t_1)) / 2.0; else tmp = (2.0 - ((alpha * 2.0) / beta)) / 2.0; end tmp_2 = tmp; end
code[alpha_, beta_, i_] := Block[{t$95$0 = N[(N[(alpha + beta), $MachinePrecision] + N[(2.0 * i), $MachinePrecision]), $MachinePrecision]}, Block[{t$95$1 = N[(2.0 + t$95$0), $MachinePrecision]}, Block[{t$95$2 = N[(N[(N[(N[(alpha + beta), $MachinePrecision] * N[(beta - alpha), $MachinePrecision]), $MachinePrecision] / t$95$0), $MachinePrecision] / t$95$1), $MachinePrecision]}, If[LessEqual[t$95$2, -0.5], N[(N[(N[(2.0 * N[(beta / alpha), $MachinePrecision]), $MachinePrecision] + N[(N[(2.0 + N[(i * 4.0), $MachinePrecision]), $MachinePrecision] / alpha), $MachinePrecision]), $MachinePrecision] / 2.0), $MachinePrecision], If[LessEqual[t$95$2, 1.0], N[(N[(1.0 - N[(N[(N[(N[(alpha + beta), $MachinePrecision] * N[(alpha - beta), $MachinePrecision]), $MachinePrecision] / t$95$0), $MachinePrecision] / t$95$1), $MachinePrecision]), $MachinePrecision] / 2.0), $MachinePrecision], N[(N[(2.0 - N[(N[(alpha * 2.0), $MachinePrecision] / beta), $MachinePrecision]), $MachinePrecision] / 2.0), $MachinePrecision]]]]]]
\begin{array}{l}
\\
\begin{array}{l}
t_0 := \left(\alpha + \beta\right) + 2 \cdot i\\
t_1 := 2 + t\_0\\
t_2 := \frac{\frac{\left(\alpha + \beta\right) \cdot \left(\beta - \alpha\right)}{t\_0}}{t\_1}\\
\mathbf{if}\;t\_2 \leq -0.5:\\
\;\;\;\;\frac{2 \cdot \frac{\beta}{\alpha} + \frac{2 + i \cdot 4}{\alpha}}{2}\\
\mathbf{elif}\;t\_2 \leq 1:\\
\;\;\;\;\frac{1 - \frac{\frac{\left(\alpha + \beta\right) \cdot \left(\alpha - \beta\right)}{t\_0}}{t\_1}}{2}\\
\mathbf{else}:\\
\;\;\;\;\frac{2 - \frac{\alpha \cdot 2}{\beta}}{2}\\
\end{array}
\end{array}
if (/.f64 (/.f64 (*.f64 (+.f64 alpha beta) (-.f64 beta alpha)) (+.f64 (+.f64 alpha beta) (*.f64 #s(literal 2 binary64) i))) (+.f64 (+.f64 (+.f64 alpha beta) (*.f64 #s(literal 2 binary64) i)) #s(literal 2 binary64))) < -0.5Initial program 5.5%
Simplified16.1%
Taylor expanded in alpha around inf 10.4%
Taylor expanded in beta around 0 91.2%
mul-1-neg91.2%
*-commutative91.2%
Simplified91.2%
if -0.5 < (/.f64 (/.f64 (*.f64 (+.f64 alpha beta) (-.f64 beta alpha)) (+.f64 (+.f64 alpha beta) (*.f64 #s(literal 2 binary64) i))) (+.f64 (+.f64 (+.f64 alpha beta) (*.f64 #s(literal 2 binary64) i)) #s(literal 2 binary64))) < 1Initial program 100.0%
if 1 < (/.f64 (/.f64 (*.f64 (+.f64 alpha beta) (-.f64 beta alpha)) (+.f64 (+.f64 alpha beta) (*.f64 #s(literal 2 binary64) i))) (+.f64 (+.f64 (+.f64 alpha beta) (*.f64 #s(literal 2 binary64) i)) #s(literal 2 binary64))) Initial program 3.1%
associate-/l/0.0%
associate-+l+0.0%
+-commutative0.0%
associate-+l+0.0%
Simplified0.0%
Taylor expanded in i around 0 0.0%
*-commutative0.0%
Simplified0.0%
Taylor expanded in beta around inf 94.9%
mul-1-neg94.9%
*-commutative94.9%
Simplified94.9%
Taylor expanded in alpha around inf 94.9%
associate-*r/94.9%
*-commutative94.9%
Simplified94.9%
Final simplification97.6%
(FPCore (alpha beta i)
:precision binary64
(if (<= beta 1.12e+144)
(/
(+
1.0
(/
(* (+ alpha beta) (- beta alpha))
(* (+ (+ alpha beta) (+ 2.0 (* 2.0 i))) (+ beta (+ alpha (* 2.0 i))))))
2.0)
(/ (+ 1.0 (/ (- beta alpha) (+ (+ alpha beta) 2.0))) 2.0)))
double code(double alpha, double beta, double i) {
double tmp;
if (beta <= 1.12e+144) {
tmp = (1.0 + (((alpha + beta) * (beta - alpha)) / (((alpha + beta) + (2.0 + (2.0 * i))) * (beta + (alpha + (2.0 * i)))))) / 2.0;
} else {
tmp = (1.0 + ((beta - alpha) / ((alpha + beta) + 2.0))) / 2.0;
}
return tmp;
}
real(8) function code(alpha, beta, i)
real(8), intent (in) :: alpha
real(8), intent (in) :: beta
real(8), intent (in) :: i
real(8) :: tmp
if (beta <= 1.12d+144) then
tmp = (1.0d0 + (((alpha + beta) * (beta - alpha)) / (((alpha + beta) + (2.0d0 + (2.0d0 * i))) * (beta + (alpha + (2.0d0 * i)))))) / 2.0d0
else
tmp = (1.0d0 + ((beta - alpha) / ((alpha + beta) + 2.0d0))) / 2.0d0
end if
code = tmp
end function
public static double code(double alpha, double beta, double i) {
double tmp;
if (beta <= 1.12e+144) {
tmp = (1.0 + (((alpha + beta) * (beta - alpha)) / (((alpha + beta) + (2.0 + (2.0 * i))) * (beta + (alpha + (2.0 * i)))))) / 2.0;
} else {
tmp = (1.0 + ((beta - alpha) / ((alpha + beta) + 2.0))) / 2.0;
}
return tmp;
}
def code(alpha, beta, i): tmp = 0 if beta <= 1.12e+144: tmp = (1.0 + (((alpha + beta) * (beta - alpha)) / (((alpha + beta) + (2.0 + (2.0 * i))) * (beta + (alpha + (2.0 * i)))))) / 2.0 else: tmp = (1.0 + ((beta - alpha) / ((alpha + beta) + 2.0))) / 2.0 return tmp
function code(alpha, beta, i) tmp = 0.0 if (beta <= 1.12e+144) tmp = Float64(Float64(1.0 + Float64(Float64(Float64(alpha + beta) * Float64(beta - alpha)) / Float64(Float64(Float64(alpha + beta) + Float64(2.0 + Float64(2.0 * i))) * Float64(beta + Float64(alpha + Float64(2.0 * i)))))) / 2.0); else tmp = Float64(Float64(1.0 + Float64(Float64(beta - alpha) / Float64(Float64(alpha + beta) + 2.0))) / 2.0); end return tmp end
function tmp_2 = code(alpha, beta, i) tmp = 0.0; if (beta <= 1.12e+144) tmp = (1.0 + (((alpha + beta) * (beta - alpha)) / (((alpha + beta) + (2.0 + (2.0 * i))) * (beta + (alpha + (2.0 * i)))))) / 2.0; else tmp = (1.0 + ((beta - alpha) / ((alpha + beta) + 2.0))) / 2.0; end tmp_2 = tmp; end
code[alpha_, beta_, i_] := If[LessEqual[beta, 1.12e+144], N[(N[(1.0 + N[(N[(N[(alpha + beta), $MachinePrecision] * N[(beta - alpha), $MachinePrecision]), $MachinePrecision] / N[(N[(N[(alpha + beta), $MachinePrecision] + N[(2.0 + N[(2.0 * i), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] * N[(beta + N[(alpha + N[(2.0 * i), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] / 2.0), $MachinePrecision], N[(N[(1.0 + N[(N[(beta - alpha), $MachinePrecision] / N[(N[(alpha + beta), $MachinePrecision] + 2.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] / 2.0), $MachinePrecision]]
\begin{array}{l}
\\
\begin{array}{l}
\mathbf{if}\;\beta \leq 1.12 \cdot 10^{+144}:\\
\;\;\;\;\frac{1 + \frac{\left(\alpha + \beta\right) \cdot \left(\beta - \alpha\right)}{\left(\left(\alpha + \beta\right) + \left(2 + 2 \cdot i\right)\right) \cdot \left(\beta + \left(\alpha + 2 \cdot i\right)\right)}}{2}\\
\mathbf{else}:\\
\;\;\;\;\frac{1 + \frac{\beta - \alpha}{\left(\alpha + \beta\right) + 2}}{2}\\
\end{array}
\end{array}
if beta < 1.11999999999999999e144Initial program 80.7%
associate-/l/80.3%
associate-+l+80.3%
+-commutative80.3%
associate-+l+80.3%
Simplified80.3%
if 1.11999999999999999e144 < beta Initial program 5.0%
Simplified98.5%
Taylor expanded in i around 0 93.6%
Final simplification82.9%
(FPCore (alpha beta i) :precision binary64 (if (<= alpha 1.3e+143) (/ (+ 1.0 (/ beta (+ beta 2.0))) 2.0) (/ (+ (* 2.0 (/ beta alpha)) (/ (+ 2.0 (* i 4.0)) alpha)) 2.0)))
double code(double alpha, double beta, double i) {
double tmp;
if (alpha <= 1.3e+143) {
tmp = (1.0 + (beta / (beta + 2.0))) / 2.0;
} else {
tmp = ((2.0 * (beta / alpha)) + ((2.0 + (i * 4.0)) / alpha)) / 2.0;
}
return tmp;
}
real(8) function code(alpha, beta, i)
real(8), intent (in) :: alpha
real(8), intent (in) :: beta
real(8), intent (in) :: i
real(8) :: tmp
if (alpha <= 1.3d+143) then
tmp = (1.0d0 + (beta / (beta + 2.0d0))) / 2.0d0
else
tmp = ((2.0d0 * (beta / alpha)) + ((2.0d0 + (i * 4.0d0)) / alpha)) / 2.0d0
end if
code = tmp
end function
public static double code(double alpha, double beta, double i) {
double tmp;
if (alpha <= 1.3e+143) {
tmp = (1.0 + (beta / (beta + 2.0))) / 2.0;
} else {
tmp = ((2.0 * (beta / alpha)) + ((2.0 + (i * 4.0)) / alpha)) / 2.0;
}
return tmp;
}
def code(alpha, beta, i): tmp = 0 if alpha <= 1.3e+143: tmp = (1.0 + (beta / (beta + 2.0))) / 2.0 else: tmp = ((2.0 * (beta / alpha)) + ((2.0 + (i * 4.0)) / alpha)) / 2.0 return tmp
function code(alpha, beta, i) tmp = 0.0 if (alpha <= 1.3e+143) tmp = Float64(Float64(1.0 + Float64(beta / Float64(beta + 2.0))) / 2.0); else tmp = Float64(Float64(Float64(2.0 * Float64(beta / alpha)) + Float64(Float64(2.0 + Float64(i * 4.0)) / alpha)) / 2.0); end return tmp end
function tmp_2 = code(alpha, beta, i) tmp = 0.0; if (alpha <= 1.3e+143) tmp = (1.0 + (beta / (beta + 2.0))) / 2.0; else tmp = ((2.0 * (beta / alpha)) + ((2.0 + (i * 4.0)) / alpha)) / 2.0; end tmp_2 = tmp; end
code[alpha_, beta_, i_] := If[LessEqual[alpha, 1.3e+143], N[(N[(1.0 + N[(beta / N[(beta + 2.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] / 2.0), $MachinePrecision], N[(N[(N[(2.0 * N[(beta / alpha), $MachinePrecision]), $MachinePrecision] + N[(N[(2.0 + N[(i * 4.0), $MachinePrecision]), $MachinePrecision] / alpha), $MachinePrecision]), $MachinePrecision] / 2.0), $MachinePrecision]]
\begin{array}{l}
\\
\begin{array}{l}
\mathbf{if}\;\alpha \leq 1.3 \cdot 10^{+143}:\\
\;\;\;\;\frac{1 + \frac{\beta}{\beta + 2}}{2}\\
\mathbf{else}:\\
\;\;\;\;\frac{2 \cdot \frac{\beta}{\alpha} + \frac{2 + i \cdot 4}{\alpha}}{2}\\
\end{array}
\end{array}
if alpha < 1.2999999999999999e143Initial program 76.0%
associate-/l/75.4%
associate-+l+75.4%
+-commutative75.4%
associate-+l+75.4%
Simplified75.4%
Taylor expanded in i around 0 62.8%
*-commutative62.8%
Simplified62.8%
Taylor expanded in alpha around 0 84.1%
if 1.2999999999999999e143 < alpha Initial program 4.3%
Simplified33.2%
Taylor expanded in alpha around inf 8.8%
Taylor expanded in beta around 0 73.9%
mul-1-neg73.9%
*-commutative73.9%
Simplified73.9%
Final simplification82.7%
(FPCore (alpha beta i) :precision binary64 (if (<= i 3.1e+108) (/ (+ 1.0 (/ (- beta alpha) (+ (+ alpha beta) 2.0))) 2.0) 0.5))
double code(double alpha, double beta, double i) {
double tmp;
if (i <= 3.1e+108) {
tmp = (1.0 + ((beta - alpha) / ((alpha + beta) + 2.0))) / 2.0;
} else {
tmp = 0.5;
}
return tmp;
}
real(8) function code(alpha, beta, i)
real(8), intent (in) :: alpha
real(8), intent (in) :: beta
real(8), intent (in) :: i
real(8) :: tmp
if (i <= 3.1d+108) then
tmp = (1.0d0 + ((beta - alpha) / ((alpha + beta) + 2.0d0))) / 2.0d0
else
tmp = 0.5d0
end if
code = tmp
end function
public static double code(double alpha, double beta, double i) {
double tmp;
if (i <= 3.1e+108) {
tmp = (1.0 + ((beta - alpha) / ((alpha + beta) + 2.0))) / 2.0;
} else {
tmp = 0.5;
}
return tmp;
}
def code(alpha, beta, i): tmp = 0 if i <= 3.1e+108: tmp = (1.0 + ((beta - alpha) / ((alpha + beta) + 2.0))) / 2.0 else: tmp = 0.5 return tmp
function code(alpha, beta, i) tmp = 0.0 if (i <= 3.1e+108) tmp = Float64(Float64(1.0 + Float64(Float64(beta - alpha) / Float64(Float64(alpha + beta) + 2.0))) / 2.0); else tmp = 0.5; end return tmp end
function tmp_2 = code(alpha, beta, i) tmp = 0.0; if (i <= 3.1e+108) tmp = (1.0 + ((beta - alpha) / ((alpha + beta) + 2.0))) / 2.0; else tmp = 0.5; end tmp_2 = tmp; end
code[alpha_, beta_, i_] := If[LessEqual[i, 3.1e+108], N[(N[(1.0 + N[(N[(beta - alpha), $MachinePrecision] / N[(N[(alpha + beta), $MachinePrecision] + 2.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] / 2.0), $MachinePrecision], 0.5]
\begin{array}{l}
\\
\begin{array}{l}
\mathbf{if}\;i \leq 3.1 \cdot 10^{+108}:\\
\;\;\;\;\frac{1 + \frac{\beta - \alpha}{\left(\alpha + \beta\right) + 2}}{2}\\
\mathbf{else}:\\
\;\;\;\;0.5\\
\end{array}
\end{array}
if i < 3.1000000000000001e108Initial program 61.4%
Simplified83.5%
Taylor expanded in i around 0 80.9%
if 3.1000000000000001e108 < i Initial program 75.8%
Simplified91.0%
Taylor expanded in i around inf 82.7%
Final simplification81.5%
(FPCore (alpha beta i) :precision binary64 (if (<= alpha 8.5e+146) (/ (+ 1.0 (/ beta (+ beta 2.0))) 2.0) (/ (/ (+ 2.0 (* i 4.0)) alpha) 2.0)))
double code(double alpha, double beta, double i) {
double tmp;
if (alpha <= 8.5e+146) {
tmp = (1.0 + (beta / (beta + 2.0))) / 2.0;
} else {
tmp = ((2.0 + (i * 4.0)) / alpha) / 2.0;
}
return tmp;
}
real(8) function code(alpha, beta, i)
real(8), intent (in) :: alpha
real(8), intent (in) :: beta
real(8), intent (in) :: i
real(8) :: tmp
if (alpha <= 8.5d+146) then
tmp = (1.0d0 + (beta / (beta + 2.0d0))) / 2.0d0
else
tmp = ((2.0d0 + (i * 4.0d0)) / alpha) / 2.0d0
end if
code = tmp
end function
public static double code(double alpha, double beta, double i) {
double tmp;
if (alpha <= 8.5e+146) {
tmp = (1.0 + (beta / (beta + 2.0))) / 2.0;
} else {
tmp = ((2.0 + (i * 4.0)) / alpha) / 2.0;
}
return tmp;
}
def code(alpha, beta, i): tmp = 0 if alpha <= 8.5e+146: tmp = (1.0 + (beta / (beta + 2.0))) / 2.0 else: tmp = ((2.0 + (i * 4.0)) / alpha) / 2.0 return tmp
function code(alpha, beta, i) tmp = 0.0 if (alpha <= 8.5e+146) tmp = Float64(Float64(1.0 + Float64(beta / Float64(beta + 2.0))) / 2.0); else tmp = Float64(Float64(Float64(2.0 + Float64(i * 4.0)) / alpha) / 2.0); end return tmp end
function tmp_2 = code(alpha, beta, i) tmp = 0.0; if (alpha <= 8.5e+146) tmp = (1.0 + (beta / (beta + 2.0))) / 2.0; else tmp = ((2.0 + (i * 4.0)) / alpha) / 2.0; end tmp_2 = tmp; end
code[alpha_, beta_, i_] := If[LessEqual[alpha, 8.5e+146], N[(N[(1.0 + N[(beta / N[(beta + 2.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] / 2.0), $MachinePrecision], N[(N[(N[(2.0 + N[(i * 4.0), $MachinePrecision]), $MachinePrecision] / alpha), $MachinePrecision] / 2.0), $MachinePrecision]]
\begin{array}{l}
\\
\begin{array}{l}
\mathbf{if}\;\alpha \leq 8.5 \cdot 10^{+146}:\\
\;\;\;\;\frac{1 + \frac{\beta}{\beta + 2}}{2}\\
\mathbf{else}:\\
\;\;\;\;\frac{\frac{2 + i \cdot 4}{\alpha}}{2}\\
\end{array}
\end{array}
if alpha < 8.5e146Initial program 76.0%
associate-/l/75.4%
associate-+l+75.4%
+-commutative75.4%
associate-+l+75.4%
Simplified75.4%
Taylor expanded in i around 0 62.8%
*-commutative62.8%
Simplified62.8%
Taylor expanded in alpha around 0 84.1%
if 8.5e146 < alpha Initial program 4.3%
Simplified32.4%
Taylor expanded in beta around 0 16.3%
Taylor expanded in alpha around inf 63.2%
*-commutative63.2%
Simplified63.2%
Final simplification81.2%
(FPCore (alpha beta i) :precision binary64 (if (<= i 7e+107) (/ (+ 1.0 (/ beta (+ beta 2.0))) 2.0) 0.5))
double code(double alpha, double beta, double i) {
double tmp;
if (i <= 7e+107) {
tmp = (1.0 + (beta / (beta + 2.0))) / 2.0;
} else {
tmp = 0.5;
}
return tmp;
}
real(8) function code(alpha, beta, i)
real(8), intent (in) :: alpha
real(8), intent (in) :: beta
real(8), intent (in) :: i
real(8) :: tmp
if (i <= 7d+107) then
tmp = (1.0d0 + (beta / (beta + 2.0d0))) / 2.0d0
else
tmp = 0.5d0
end if
code = tmp
end function
public static double code(double alpha, double beta, double i) {
double tmp;
if (i <= 7e+107) {
tmp = (1.0 + (beta / (beta + 2.0))) / 2.0;
} else {
tmp = 0.5;
}
return tmp;
}
def code(alpha, beta, i): tmp = 0 if i <= 7e+107: tmp = (1.0 + (beta / (beta + 2.0))) / 2.0 else: tmp = 0.5 return tmp
function code(alpha, beta, i) tmp = 0.0 if (i <= 7e+107) tmp = Float64(Float64(1.0 + Float64(beta / Float64(beta + 2.0))) / 2.0); else tmp = 0.5; end return tmp end
function tmp_2 = code(alpha, beta, i) tmp = 0.0; if (i <= 7e+107) tmp = (1.0 + (beta / (beta + 2.0))) / 2.0; else tmp = 0.5; end tmp_2 = tmp; end
code[alpha_, beta_, i_] := If[LessEqual[i, 7e+107], N[(N[(1.0 + N[(beta / N[(beta + 2.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] / 2.0), $MachinePrecision], 0.5]
\begin{array}{l}
\\
\begin{array}{l}
\mathbf{if}\;i \leq 7 \cdot 10^{+107}:\\
\;\;\;\;\frac{1 + \frac{\beta}{\beta + 2}}{2}\\
\mathbf{else}:\\
\;\;\;\;0.5\\
\end{array}
\end{array}
if i < 6.9999999999999995e107Initial program 61.4%
associate-/l/60.5%
associate-+l+60.5%
+-commutative60.5%
associate-+l+60.5%
Simplified60.5%
Taylor expanded in i around 0 57.9%
*-commutative57.9%
Simplified57.9%
Taylor expanded in alpha around 0 79.3%
if 6.9999999999999995e107 < i Initial program 75.8%
Simplified91.0%
Taylor expanded in i around inf 82.7%
Final simplification80.3%
(FPCore (alpha beta i) :precision binary64 (if (<= beta 2.3e+55) 0.5 1.0))
double code(double alpha, double beta, double i) {
double tmp;
if (beta <= 2.3e+55) {
tmp = 0.5;
} else {
tmp = 1.0;
}
return tmp;
}
real(8) function code(alpha, beta, i)
real(8), intent (in) :: alpha
real(8), intent (in) :: beta
real(8), intent (in) :: i
real(8) :: tmp
if (beta <= 2.3d+55) then
tmp = 0.5d0
else
tmp = 1.0d0
end if
code = tmp
end function
public static double code(double alpha, double beta, double i) {
double tmp;
if (beta <= 2.3e+55) {
tmp = 0.5;
} else {
tmp = 1.0;
}
return tmp;
}
def code(alpha, beta, i): tmp = 0 if beta <= 2.3e+55: tmp = 0.5 else: tmp = 1.0 return tmp
function code(alpha, beta, i) tmp = 0.0 if (beta <= 2.3e+55) tmp = 0.5; else tmp = 1.0; end return tmp end
function tmp_2 = code(alpha, beta, i) tmp = 0.0; if (beta <= 2.3e+55) tmp = 0.5; else tmp = 1.0; end tmp_2 = tmp; end
code[alpha_, beta_, i_] := If[LessEqual[beta, 2.3e+55], 0.5, 1.0]
\begin{array}{l}
\\
\begin{array}{l}
\mathbf{if}\;\beta \leq 2.3 \cdot 10^{+55}:\\
\;\;\;\;0.5\\
\mathbf{else}:\\
\;\;\;\;1\\
\end{array}
\end{array}
if beta < 2.29999999999999987e55Initial program 79.8%
Simplified81.6%
Taylor expanded in i around inf 74.3%
if 2.29999999999999987e55 < beta Initial program 34.4%
Simplified95.7%
Taylor expanded in beta around inf 79.2%
Final simplification75.8%
(FPCore (alpha beta i) :precision binary64 0.5)
double code(double alpha, double beta, double i) {
return 0.5;
}
real(8) function code(alpha, beta, i)
real(8), intent (in) :: alpha
real(8), intent (in) :: beta
real(8), intent (in) :: i
code = 0.5d0
end function
public static double code(double alpha, double beta, double i) {
return 0.5;
}
def code(alpha, beta, i): return 0.5
function code(alpha, beta, i) return 0.5 end
function tmp = code(alpha, beta, i) tmp = 0.5; end
code[alpha_, beta_, i_] := 0.5
\begin{array}{l}
\\
0.5
\end{array}
Initial program 66.0%
Simplified85.9%
Taylor expanded in i around inf 61.3%
Final simplification61.3%
herbie shell --seed 2024102
(FPCore (alpha beta i)
:name "Octave 3.8, jcobi/2"
:precision binary64
:pre (and (and (> alpha -1.0) (> beta -1.0)) (> i 0.0))
(/ (+ (/ (/ (* (+ alpha beta) (- beta alpha)) (+ (+ alpha beta) (* 2.0 i))) (+ (+ (+ alpha beta) (* 2.0 i)) 2.0)) 1.0) 2.0))