
(FPCore (alpha beta) :precision binary64 (/ (+ (/ (- beta alpha) (+ (+ alpha beta) 2.0)) 1.0) 2.0))
double code(double alpha, double beta) {
return (((beta - alpha) / ((alpha + beta) + 2.0)) + 1.0) / 2.0;
}
real(8) function code(alpha, beta)
real(8), intent (in) :: alpha
real(8), intent (in) :: beta
code = (((beta - alpha) / ((alpha + beta) + 2.0d0)) + 1.0d0) / 2.0d0
end function
public static double code(double alpha, double beta) {
return (((beta - alpha) / ((alpha + beta) + 2.0)) + 1.0) / 2.0;
}
def code(alpha, beta): return (((beta - alpha) / ((alpha + beta) + 2.0)) + 1.0) / 2.0
function code(alpha, beta) return Float64(Float64(Float64(Float64(beta - alpha) / Float64(Float64(alpha + beta) + 2.0)) + 1.0) / 2.0) end
function tmp = code(alpha, beta) tmp = (((beta - alpha) / ((alpha + beta) + 2.0)) + 1.0) / 2.0; end
code[alpha_, beta_] := N[(N[(N[(N[(beta - alpha), $MachinePrecision] / N[(N[(alpha + beta), $MachinePrecision] + 2.0), $MachinePrecision]), $MachinePrecision] + 1.0), $MachinePrecision] / 2.0), $MachinePrecision]
\begin{array}{l}
\\
\frac{\frac{\beta - \alpha}{\left(\alpha + \beta\right) + 2} + 1}{2}
\end{array}
Sampling outcomes in binary64 precision:
Herbie found 10 alternatives:
| Alternative | Accuracy | Speedup |
|---|
(FPCore (alpha beta) :precision binary64 (/ (+ (/ (- beta alpha) (+ (+ alpha beta) 2.0)) 1.0) 2.0))
double code(double alpha, double beta) {
return (((beta - alpha) / ((alpha + beta) + 2.0)) + 1.0) / 2.0;
}
real(8) function code(alpha, beta)
real(8), intent (in) :: alpha
real(8), intent (in) :: beta
code = (((beta - alpha) / ((alpha + beta) + 2.0d0)) + 1.0d0) / 2.0d0
end function
public static double code(double alpha, double beta) {
return (((beta - alpha) / ((alpha + beta) + 2.0)) + 1.0) / 2.0;
}
def code(alpha, beta): return (((beta - alpha) / ((alpha + beta) + 2.0)) + 1.0) / 2.0
function code(alpha, beta) return Float64(Float64(Float64(Float64(beta - alpha) / Float64(Float64(alpha + beta) + 2.0)) + 1.0) / 2.0) end
function tmp = code(alpha, beta) tmp = (((beta - alpha) / ((alpha + beta) + 2.0)) + 1.0) / 2.0; end
code[alpha_, beta_] := N[(N[(N[(N[(beta - alpha), $MachinePrecision] / N[(N[(alpha + beta), $MachinePrecision] + 2.0), $MachinePrecision]), $MachinePrecision] + 1.0), $MachinePrecision] / 2.0), $MachinePrecision]
\begin{array}{l}
\\
\frac{\frac{\beta - \alpha}{\left(\alpha + \beta\right) + 2} + 1}{2}
\end{array}
(FPCore (alpha beta)
:precision binary64
(let* ((t_0 (+ beta (+ alpha 2.0))))
(if (<= (/ (- beta alpha) (+ (+ beta alpha) 2.0)) -0.999995)
(/
(+
(* (- -2.0 beta) (/ (+ beta (+ beta 2.0)) (pow alpha 2.0)))
(/ (+ beta (- beta -2.0)) alpha))
2.0)
(/ (- (/ beta t_0) (fma alpha (/ 1.0 t_0) -1.0)) 2.0))))
double code(double alpha, double beta) {
double t_0 = beta + (alpha + 2.0);
double tmp;
if (((beta - alpha) / ((beta + alpha) + 2.0)) <= -0.999995) {
tmp = (((-2.0 - beta) * ((beta + (beta + 2.0)) / pow(alpha, 2.0))) + ((beta + (beta - -2.0)) / alpha)) / 2.0;
} else {
tmp = ((beta / t_0) - fma(alpha, (1.0 / t_0), -1.0)) / 2.0;
}
return tmp;
}
function code(alpha, beta) t_0 = Float64(beta + Float64(alpha + 2.0)) tmp = 0.0 if (Float64(Float64(beta - alpha) / Float64(Float64(beta + alpha) + 2.0)) <= -0.999995) tmp = Float64(Float64(Float64(Float64(-2.0 - beta) * Float64(Float64(beta + Float64(beta + 2.0)) / (alpha ^ 2.0))) + Float64(Float64(beta + Float64(beta - -2.0)) / alpha)) / 2.0); else tmp = Float64(Float64(Float64(beta / t_0) - fma(alpha, Float64(1.0 / t_0), -1.0)) / 2.0); end return tmp end
code[alpha_, beta_] := Block[{t$95$0 = N[(beta + N[(alpha + 2.0), $MachinePrecision]), $MachinePrecision]}, If[LessEqual[N[(N[(beta - alpha), $MachinePrecision] / N[(N[(beta + alpha), $MachinePrecision] + 2.0), $MachinePrecision]), $MachinePrecision], -0.999995], N[(N[(N[(N[(-2.0 - beta), $MachinePrecision] * N[(N[(beta + N[(beta + 2.0), $MachinePrecision]), $MachinePrecision] / N[Power[alpha, 2.0], $MachinePrecision]), $MachinePrecision]), $MachinePrecision] + N[(N[(beta + N[(beta - -2.0), $MachinePrecision]), $MachinePrecision] / alpha), $MachinePrecision]), $MachinePrecision] / 2.0), $MachinePrecision], N[(N[(N[(beta / t$95$0), $MachinePrecision] - N[(alpha * N[(1.0 / t$95$0), $MachinePrecision] + -1.0), $MachinePrecision]), $MachinePrecision] / 2.0), $MachinePrecision]]]
\begin{array}{l}
\\
\begin{array}{l}
t_0 := \beta + \left(\alpha + 2\right)\\
\mathbf{if}\;\frac{\beta - \alpha}{\left(\beta + \alpha\right) + 2} \leq -0.999995:\\
\;\;\;\;\frac{\left(-2 - \beta\right) \cdot \frac{\beta + \left(\beta + 2\right)}{{\alpha}^{2}} + \frac{\beta + \left(\beta - -2\right)}{\alpha}}{2}\\
\mathbf{else}:\\
\;\;\;\;\frac{\frac{\beta}{t\_0} - \mathsf{fma}\left(\alpha, \frac{1}{t\_0}, -1\right)}{2}\\
\end{array}
\end{array}
if (/.f64 (-.f64 beta alpha) (+.f64 (+.f64 alpha beta) 2)) < -0.99999499999999997Initial program 6.6%
Taylor expanded in alpha around -inf 97.2%
mul-1-neg97.2%
unsub-neg97.2%
Simplified100.0%
if -0.99999499999999997 < (/.f64 (-.f64 beta alpha) (+.f64 (+.f64 alpha beta) 2)) Initial program 99.9%
div-sub99.9%
associate-+l-99.9%
+-commutative99.9%
associate-+r+99.9%
+-commutative99.9%
associate-+r+99.9%
Applied egg-rr99.9%
div-inv99.9%
fma-neg99.9%
metadata-eval99.9%
Applied egg-rr99.9%
Final simplification99.9%
(FPCore (alpha beta)
:precision binary64
(let* ((t_0 (+ beta (+ alpha 2.0))))
(if (<= (/ (- beta alpha) (+ (+ beta alpha) 2.0)) -0.999995)
(/ (/ (+ 2.0 (* beta 2.0)) alpha) 2.0)
(/ (- (/ beta t_0) (fma alpha (/ 1.0 t_0) -1.0)) 2.0))))
double code(double alpha, double beta) {
double t_0 = beta + (alpha + 2.0);
double tmp;
if (((beta - alpha) / ((beta + alpha) + 2.0)) <= -0.999995) {
tmp = ((2.0 + (beta * 2.0)) / alpha) / 2.0;
} else {
tmp = ((beta / t_0) - fma(alpha, (1.0 / t_0), -1.0)) / 2.0;
}
return tmp;
}
function code(alpha, beta) t_0 = Float64(beta + Float64(alpha + 2.0)) tmp = 0.0 if (Float64(Float64(beta - alpha) / Float64(Float64(beta + alpha) + 2.0)) <= -0.999995) tmp = Float64(Float64(Float64(2.0 + Float64(beta * 2.0)) / alpha) / 2.0); else tmp = Float64(Float64(Float64(beta / t_0) - fma(alpha, Float64(1.0 / t_0), -1.0)) / 2.0); end return tmp end
code[alpha_, beta_] := Block[{t$95$0 = N[(beta + N[(alpha + 2.0), $MachinePrecision]), $MachinePrecision]}, If[LessEqual[N[(N[(beta - alpha), $MachinePrecision] / N[(N[(beta + alpha), $MachinePrecision] + 2.0), $MachinePrecision]), $MachinePrecision], -0.999995], N[(N[(N[(2.0 + N[(beta * 2.0), $MachinePrecision]), $MachinePrecision] / alpha), $MachinePrecision] / 2.0), $MachinePrecision], N[(N[(N[(beta / t$95$0), $MachinePrecision] - N[(alpha * N[(1.0 / t$95$0), $MachinePrecision] + -1.0), $MachinePrecision]), $MachinePrecision] / 2.0), $MachinePrecision]]]
\begin{array}{l}
\\
\begin{array}{l}
t_0 := \beta + \left(\alpha + 2\right)\\
\mathbf{if}\;\frac{\beta - \alpha}{\left(\beta + \alpha\right) + 2} \leq -0.999995:\\
\;\;\;\;\frac{\frac{2 + \beta \cdot 2}{\alpha}}{2}\\
\mathbf{else}:\\
\;\;\;\;\frac{\frac{\beta}{t\_0} - \mathsf{fma}\left(\alpha, \frac{1}{t\_0}, -1\right)}{2}\\
\end{array}
\end{array}
if (/.f64 (-.f64 beta alpha) (+.f64 (+.f64 alpha beta) 2)) < -0.99999499999999997Initial program 6.6%
Taylor expanded in alpha around inf 99.5%
if -0.99999499999999997 < (/.f64 (-.f64 beta alpha) (+.f64 (+.f64 alpha beta) 2)) Initial program 99.9%
div-sub99.9%
associate-+l-99.9%
+-commutative99.9%
associate-+r+99.9%
+-commutative99.9%
associate-+r+99.9%
Applied egg-rr99.9%
div-inv99.9%
fma-neg99.9%
metadata-eval99.9%
Applied egg-rr99.9%
Final simplification99.8%
(FPCore (alpha beta)
:precision binary64
(let* ((t_0 (/ (/ 2.0 alpha) 2.0)) (t_1 (/ (- 1.0 (* alpha 0.5)) 2.0)))
(if (<= beta 6.4e-167)
t_1
(if (<= beta 4.6e-156)
t_0
(if (<= beta 1.68e-74)
t_1
(if (<= beta 3.9e-50) t_0 (if (<= beta 1.4) t_1 1.0)))))))
double code(double alpha, double beta) {
double t_0 = (2.0 / alpha) / 2.0;
double t_1 = (1.0 - (alpha * 0.5)) / 2.0;
double tmp;
if (beta <= 6.4e-167) {
tmp = t_1;
} else if (beta <= 4.6e-156) {
tmp = t_0;
} else if (beta <= 1.68e-74) {
tmp = t_1;
} else if (beta <= 3.9e-50) {
tmp = t_0;
} else if (beta <= 1.4) {
tmp = t_1;
} else {
tmp = 1.0;
}
return tmp;
}
real(8) function code(alpha, beta)
real(8), intent (in) :: alpha
real(8), intent (in) :: beta
real(8) :: t_0
real(8) :: t_1
real(8) :: tmp
t_0 = (2.0d0 / alpha) / 2.0d0
t_1 = (1.0d0 - (alpha * 0.5d0)) / 2.0d0
if (beta <= 6.4d-167) then
tmp = t_1
else if (beta <= 4.6d-156) then
tmp = t_0
else if (beta <= 1.68d-74) then
tmp = t_1
else if (beta <= 3.9d-50) then
tmp = t_0
else if (beta <= 1.4d0) then
tmp = t_1
else
tmp = 1.0d0
end if
code = tmp
end function
public static double code(double alpha, double beta) {
double t_0 = (2.0 / alpha) / 2.0;
double t_1 = (1.0 - (alpha * 0.5)) / 2.0;
double tmp;
if (beta <= 6.4e-167) {
tmp = t_1;
} else if (beta <= 4.6e-156) {
tmp = t_0;
} else if (beta <= 1.68e-74) {
tmp = t_1;
} else if (beta <= 3.9e-50) {
tmp = t_0;
} else if (beta <= 1.4) {
tmp = t_1;
} else {
tmp = 1.0;
}
return tmp;
}
def code(alpha, beta): t_0 = (2.0 / alpha) / 2.0 t_1 = (1.0 - (alpha * 0.5)) / 2.0 tmp = 0 if beta <= 6.4e-167: tmp = t_1 elif beta <= 4.6e-156: tmp = t_0 elif beta <= 1.68e-74: tmp = t_1 elif beta <= 3.9e-50: tmp = t_0 elif beta <= 1.4: tmp = t_1 else: tmp = 1.0 return tmp
function code(alpha, beta) t_0 = Float64(Float64(2.0 / alpha) / 2.0) t_1 = Float64(Float64(1.0 - Float64(alpha * 0.5)) / 2.0) tmp = 0.0 if (beta <= 6.4e-167) tmp = t_1; elseif (beta <= 4.6e-156) tmp = t_0; elseif (beta <= 1.68e-74) tmp = t_1; elseif (beta <= 3.9e-50) tmp = t_0; elseif (beta <= 1.4) tmp = t_1; else tmp = 1.0; end return tmp end
function tmp_2 = code(alpha, beta) t_0 = (2.0 / alpha) / 2.0; t_1 = (1.0 - (alpha * 0.5)) / 2.0; tmp = 0.0; if (beta <= 6.4e-167) tmp = t_1; elseif (beta <= 4.6e-156) tmp = t_0; elseif (beta <= 1.68e-74) tmp = t_1; elseif (beta <= 3.9e-50) tmp = t_0; elseif (beta <= 1.4) tmp = t_1; else tmp = 1.0; end tmp_2 = tmp; end
code[alpha_, beta_] := Block[{t$95$0 = N[(N[(2.0 / alpha), $MachinePrecision] / 2.0), $MachinePrecision]}, Block[{t$95$1 = N[(N[(1.0 - N[(alpha * 0.5), $MachinePrecision]), $MachinePrecision] / 2.0), $MachinePrecision]}, If[LessEqual[beta, 6.4e-167], t$95$1, If[LessEqual[beta, 4.6e-156], t$95$0, If[LessEqual[beta, 1.68e-74], t$95$1, If[LessEqual[beta, 3.9e-50], t$95$0, If[LessEqual[beta, 1.4], t$95$1, 1.0]]]]]]]
\begin{array}{l}
\\
\begin{array}{l}
t_0 := \frac{\frac{2}{\alpha}}{2}\\
t_1 := \frac{1 - \alpha \cdot 0.5}{2}\\
\mathbf{if}\;\beta \leq 6.4 \cdot 10^{-167}:\\
\;\;\;\;t\_1\\
\mathbf{elif}\;\beta \leq 4.6 \cdot 10^{-156}:\\
\;\;\;\;t\_0\\
\mathbf{elif}\;\beta \leq 1.68 \cdot 10^{-74}:\\
\;\;\;\;t\_1\\
\mathbf{elif}\;\beta \leq 3.9 \cdot 10^{-50}:\\
\;\;\;\;t\_0\\
\mathbf{elif}\;\beta \leq 1.4:\\
\;\;\;\;t\_1\\
\mathbf{else}:\\
\;\;\;\;1\\
\end{array}
\end{array}
if beta < 6.4000000000000003e-167 or 4.5999999999999999e-156 < beta < 1.6800000000000001e-74 or 3.90000000000000021e-50 < beta < 1.3999999999999999Initial program 71.5%
Taylor expanded in beta around 0 70.5%
+-commutative70.5%
Simplified70.5%
Taylor expanded in alpha around 0 67.3%
*-commutative67.3%
Simplified67.3%
if 6.4000000000000003e-167 < beta < 4.5999999999999999e-156 or 1.6800000000000001e-74 < beta < 3.90000000000000021e-50Initial program 14.9%
Taylor expanded in beta around 0 14.9%
+-commutative14.9%
Simplified14.9%
Taylor expanded in alpha around inf 91.7%
if 1.3999999999999999 < beta Initial program 84.7%
+-commutative84.7%
sub-neg84.7%
+-commutative84.7%
neg-sub084.7%
associate-+l-84.7%
sub0-neg84.7%
distribute-frac-neg84.7%
+-commutative84.7%
sub-neg84.7%
div-sub84.7%
metadata-eval84.7%
+-commutative84.7%
associate-/l/84.7%
sub-neg84.7%
remove-double-neg84.7%
distribute-neg-out84.7%
neg-mul-184.7%
*-commutative84.7%
Simplified84.6%
Taylor expanded in beta around inf 82.4%
Final simplification73.9%
(FPCore (alpha beta)
:precision binary64
(let* ((t_0 (/ (/ 2.0 alpha) 2.0)) (t_1 (/ (- 1.0 (* alpha 0.5)) 2.0)))
(if (<= beta 1.48e-167)
t_1
(if (<= beta 4.5e-156)
t_0
(if (<= beta 2e-74)
t_1
(if (<= beta 1e-49)
t_0
(if (<= beta 1.35) t_1 (/ (- 2.0 (/ 2.0 beta)) 2.0))))))))
double code(double alpha, double beta) {
double t_0 = (2.0 / alpha) / 2.0;
double t_1 = (1.0 - (alpha * 0.5)) / 2.0;
double tmp;
if (beta <= 1.48e-167) {
tmp = t_1;
} else if (beta <= 4.5e-156) {
tmp = t_0;
} else if (beta <= 2e-74) {
tmp = t_1;
} else if (beta <= 1e-49) {
tmp = t_0;
} else if (beta <= 1.35) {
tmp = t_1;
} else {
tmp = (2.0 - (2.0 / beta)) / 2.0;
}
return tmp;
}
real(8) function code(alpha, beta)
real(8), intent (in) :: alpha
real(8), intent (in) :: beta
real(8) :: t_0
real(8) :: t_1
real(8) :: tmp
t_0 = (2.0d0 / alpha) / 2.0d0
t_1 = (1.0d0 - (alpha * 0.5d0)) / 2.0d0
if (beta <= 1.48d-167) then
tmp = t_1
else if (beta <= 4.5d-156) then
tmp = t_0
else if (beta <= 2d-74) then
tmp = t_1
else if (beta <= 1d-49) then
tmp = t_0
else if (beta <= 1.35d0) then
tmp = t_1
else
tmp = (2.0d0 - (2.0d0 / beta)) / 2.0d0
end if
code = tmp
end function
public static double code(double alpha, double beta) {
double t_0 = (2.0 / alpha) / 2.0;
double t_1 = (1.0 - (alpha * 0.5)) / 2.0;
double tmp;
if (beta <= 1.48e-167) {
tmp = t_1;
} else if (beta <= 4.5e-156) {
tmp = t_0;
} else if (beta <= 2e-74) {
tmp = t_1;
} else if (beta <= 1e-49) {
tmp = t_0;
} else if (beta <= 1.35) {
tmp = t_1;
} else {
tmp = (2.0 - (2.0 / beta)) / 2.0;
}
return tmp;
}
def code(alpha, beta): t_0 = (2.0 / alpha) / 2.0 t_1 = (1.0 - (alpha * 0.5)) / 2.0 tmp = 0 if beta <= 1.48e-167: tmp = t_1 elif beta <= 4.5e-156: tmp = t_0 elif beta <= 2e-74: tmp = t_1 elif beta <= 1e-49: tmp = t_0 elif beta <= 1.35: tmp = t_1 else: tmp = (2.0 - (2.0 / beta)) / 2.0 return tmp
function code(alpha, beta) t_0 = Float64(Float64(2.0 / alpha) / 2.0) t_1 = Float64(Float64(1.0 - Float64(alpha * 0.5)) / 2.0) tmp = 0.0 if (beta <= 1.48e-167) tmp = t_1; elseif (beta <= 4.5e-156) tmp = t_0; elseif (beta <= 2e-74) tmp = t_1; elseif (beta <= 1e-49) tmp = t_0; elseif (beta <= 1.35) tmp = t_1; else tmp = Float64(Float64(2.0 - Float64(2.0 / beta)) / 2.0); end return tmp end
function tmp_2 = code(alpha, beta) t_0 = (2.0 / alpha) / 2.0; t_1 = (1.0 - (alpha * 0.5)) / 2.0; tmp = 0.0; if (beta <= 1.48e-167) tmp = t_1; elseif (beta <= 4.5e-156) tmp = t_0; elseif (beta <= 2e-74) tmp = t_1; elseif (beta <= 1e-49) tmp = t_0; elseif (beta <= 1.35) tmp = t_1; else tmp = (2.0 - (2.0 / beta)) / 2.0; end tmp_2 = tmp; end
code[alpha_, beta_] := Block[{t$95$0 = N[(N[(2.0 / alpha), $MachinePrecision] / 2.0), $MachinePrecision]}, Block[{t$95$1 = N[(N[(1.0 - N[(alpha * 0.5), $MachinePrecision]), $MachinePrecision] / 2.0), $MachinePrecision]}, If[LessEqual[beta, 1.48e-167], t$95$1, If[LessEqual[beta, 4.5e-156], t$95$0, If[LessEqual[beta, 2e-74], t$95$1, If[LessEqual[beta, 1e-49], t$95$0, If[LessEqual[beta, 1.35], t$95$1, N[(N[(2.0 - N[(2.0 / beta), $MachinePrecision]), $MachinePrecision] / 2.0), $MachinePrecision]]]]]]]]
\begin{array}{l}
\\
\begin{array}{l}
t_0 := \frac{\frac{2}{\alpha}}{2}\\
t_1 := \frac{1 - \alpha \cdot 0.5}{2}\\
\mathbf{if}\;\beta \leq 1.48 \cdot 10^{-167}:\\
\;\;\;\;t\_1\\
\mathbf{elif}\;\beta \leq 4.5 \cdot 10^{-156}:\\
\;\;\;\;t\_0\\
\mathbf{elif}\;\beta \leq 2 \cdot 10^{-74}:\\
\;\;\;\;t\_1\\
\mathbf{elif}\;\beta \leq 10^{-49}:\\
\;\;\;\;t\_0\\
\mathbf{elif}\;\beta \leq 1.35:\\
\;\;\;\;t\_1\\
\mathbf{else}:\\
\;\;\;\;\frac{2 - \frac{2}{\beta}}{2}\\
\end{array}
\end{array}
if beta < 1.47999999999999999e-167 or 4.49999999999999986e-156 < beta < 1.99999999999999992e-74 or 9.99999999999999936e-50 < beta < 1.3500000000000001Initial program 71.5%
Taylor expanded in beta around 0 70.5%
+-commutative70.5%
Simplified70.5%
Taylor expanded in alpha around 0 67.3%
*-commutative67.3%
Simplified67.3%
if 1.47999999999999999e-167 < beta < 4.49999999999999986e-156 or 1.99999999999999992e-74 < beta < 9.99999999999999936e-50Initial program 14.9%
Taylor expanded in beta around 0 14.9%
+-commutative14.9%
Simplified14.9%
Taylor expanded in alpha around inf 91.7%
if 1.3500000000000001 < beta Initial program 84.7%
Taylor expanded in beta around inf 82.1%
mul-1-neg82.1%
unsub-neg82.1%
*-commutative82.1%
Simplified82.1%
Taylor expanded in alpha around 0 82.5%
Final simplification74.0%
(FPCore (alpha beta)
:precision binary64
(let* ((t_0 (+ beta (+ alpha 2.0))))
(if (<= (/ (- beta alpha) (+ (+ beta alpha) 2.0)) -0.999995)
(/ (/ (+ 2.0 (* beta 2.0)) alpha) 2.0)
(/ (+ (/ beta t_0) (- 1.0 (/ alpha t_0))) 2.0))))
double code(double alpha, double beta) {
double t_0 = beta + (alpha + 2.0);
double tmp;
if (((beta - alpha) / ((beta + alpha) + 2.0)) <= -0.999995) {
tmp = ((2.0 + (beta * 2.0)) / alpha) / 2.0;
} else {
tmp = ((beta / t_0) + (1.0 - (alpha / t_0))) / 2.0;
}
return tmp;
}
real(8) function code(alpha, beta)
real(8), intent (in) :: alpha
real(8), intent (in) :: beta
real(8) :: t_0
real(8) :: tmp
t_0 = beta + (alpha + 2.0d0)
if (((beta - alpha) / ((beta + alpha) + 2.0d0)) <= (-0.999995d0)) then
tmp = ((2.0d0 + (beta * 2.0d0)) / alpha) / 2.0d0
else
tmp = ((beta / t_0) + (1.0d0 - (alpha / t_0))) / 2.0d0
end if
code = tmp
end function
public static double code(double alpha, double beta) {
double t_0 = beta + (alpha + 2.0);
double tmp;
if (((beta - alpha) / ((beta + alpha) + 2.0)) <= -0.999995) {
tmp = ((2.0 + (beta * 2.0)) / alpha) / 2.0;
} else {
tmp = ((beta / t_0) + (1.0 - (alpha / t_0))) / 2.0;
}
return tmp;
}
def code(alpha, beta): t_0 = beta + (alpha + 2.0) tmp = 0 if ((beta - alpha) / ((beta + alpha) + 2.0)) <= -0.999995: tmp = ((2.0 + (beta * 2.0)) / alpha) / 2.0 else: tmp = ((beta / t_0) + (1.0 - (alpha / t_0))) / 2.0 return tmp
function code(alpha, beta) t_0 = Float64(beta + Float64(alpha + 2.0)) tmp = 0.0 if (Float64(Float64(beta - alpha) / Float64(Float64(beta + alpha) + 2.0)) <= -0.999995) tmp = Float64(Float64(Float64(2.0 + Float64(beta * 2.0)) / alpha) / 2.0); else tmp = Float64(Float64(Float64(beta / t_0) + Float64(1.0 - Float64(alpha / t_0))) / 2.0); end return tmp end
function tmp_2 = code(alpha, beta) t_0 = beta + (alpha + 2.0); tmp = 0.0; if (((beta - alpha) / ((beta + alpha) + 2.0)) <= -0.999995) tmp = ((2.0 + (beta * 2.0)) / alpha) / 2.0; else tmp = ((beta / t_0) + (1.0 - (alpha / t_0))) / 2.0; end tmp_2 = tmp; end
code[alpha_, beta_] := Block[{t$95$0 = N[(beta + N[(alpha + 2.0), $MachinePrecision]), $MachinePrecision]}, If[LessEqual[N[(N[(beta - alpha), $MachinePrecision] / N[(N[(beta + alpha), $MachinePrecision] + 2.0), $MachinePrecision]), $MachinePrecision], -0.999995], N[(N[(N[(2.0 + N[(beta * 2.0), $MachinePrecision]), $MachinePrecision] / alpha), $MachinePrecision] / 2.0), $MachinePrecision], N[(N[(N[(beta / t$95$0), $MachinePrecision] + N[(1.0 - N[(alpha / t$95$0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] / 2.0), $MachinePrecision]]]
\begin{array}{l}
\\
\begin{array}{l}
t_0 := \beta + \left(\alpha + 2\right)\\
\mathbf{if}\;\frac{\beta - \alpha}{\left(\beta + \alpha\right) + 2} \leq -0.999995:\\
\;\;\;\;\frac{\frac{2 + \beta \cdot 2}{\alpha}}{2}\\
\mathbf{else}:\\
\;\;\;\;\frac{\frac{\beta}{t\_0} + \left(1 - \frac{\alpha}{t\_0}\right)}{2}\\
\end{array}
\end{array}
if (/.f64 (-.f64 beta alpha) (+.f64 (+.f64 alpha beta) 2)) < -0.99999499999999997Initial program 6.6%
Taylor expanded in alpha around inf 99.5%
if -0.99999499999999997 < (/.f64 (-.f64 beta alpha) (+.f64 (+.f64 alpha beta) 2)) Initial program 99.9%
div-sub99.9%
associate-+l-99.9%
+-commutative99.9%
associate-+r+99.9%
+-commutative99.9%
associate-+r+99.9%
Applied egg-rr99.9%
Final simplification99.8%
(FPCore (alpha beta) :precision binary64 (if (<= (/ (- beta alpha) (+ (+ beta alpha) 2.0)) -0.999995) (/ (/ (+ 2.0 (* beta 2.0)) alpha) 2.0) (+ 0.5 (* (/ -0.5 (+ beta (+ alpha 2.0))) (- alpha beta)))))
double code(double alpha, double beta) {
double tmp;
if (((beta - alpha) / ((beta + alpha) + 2.0)) <= -0.999995) {
tmp = ((2.0 + (beta * 2.0)) / alpha) / 2.0;
} else {
tmp = 0.5 + ((-0.5 / (beta + (alpha + 2.0))) * (alpha - beta));
}
return tmp;
}
real(8) function code(alpha, beta)
real(8), intent (in) :: alpha
real(8), intent (in) :: beta
real(8) :: tmp
if (((beta - alpha) / ((beta + alpha) + 2.0d0)) <= (-0.999995d0)) then
tmp = ((2.0d0 + (beta * 2.0d0)) / alpha) / 2.0d0
else
tmp = 0.5d0 + (((-0.5d0) / (beta + (alpha + 2.0d0))) * (alpha - beta))
end if
code = tmp
end function
public static double code(double alpha, double beta) {
double tmp;
if (((beta - alpha) / ((beta + alpha) + 2.0)) <= -0.999995) {
tmp = ((2.0 + (beta * 2.0)) / alpha) / 2.0;
} else {
tmp = 0.5 + ((-0.5 / (beta + (alpha + 2.0))) * (alpha - beta));
}
return tmp;
}
def code(alpha, beta): tmp = 0 if ((beta - alpha) / ((beta + alpha) + 2.0)) <= -0.999995: tmp = ((2.0 + (beta * 2.0)) / alpha) / 2.0 else: tmp = 0.5 + ((-0.5 / (beta + (alpha + 2.0))) * (alpha - beta)) return tmp
function code(alpha, beta) tmp = 0.0 if (Float64(Float64(beta - alpha) / Float64(Float64(beta + alpha) + 2.0)) <= -0.999995) tmp = Float64(Float64(Float64(2.0 + Float64(beta * 2.0)) / alpha) / 2.0); else tmp = Float64(0.5 + Float64(Float64(-0.5 / Float64(beta + Float64(alpha + 2.0))) * Float64(alpha - beta))); end return tmp end
function tmp_2 = code(alpha, beta) tmp = 0.0; if (((beta - alpha) / ((beta + alpha) + 2.0)) <= -0.999995) tmp = ((2.0 + (beta * 2.0)) / alpha) / 2.0; else tmp = 0.5 + ((-0.5 / (beta + (alpha + 2.0))) * (alpha - beta)); end tmp_2 = tmp; end
code[alpha_, beta_] := If[LessEqual[N[(N[(beta - alpha), $MachinePrecision] / N[(N[(beta + alpha), $MachinePrecision] + 2.0), $MachinePrecision]), $MachinePrecision], -0.999995], N[(N[(N[(2.0 + N[(beta * 2.0), $MachinePrecision]), $MachinePrecision] / alpha), $MachinePrecision] / 2.0), $MachinePrecision], N[(0.5 + N[(N[(-0.5 / N[(beta + N[(alpha + 2.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] * N[(alpha - beta), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]]
\begin{array}{l}
\\
\begin{array}{l}
\mathbf{if}\;\frac{\beta - \alpha}{\left(\beta + \alpha\right) + 2} \leq -0.999995:\\
\;\;\;\;\frac{\frac{2 + \beta \cdot 2}{\alpha}}{2}\\
\mathbf{else}:\\
\;\;\;\;0.5 + \frac{-0.5}{\beta + \left(\alpha + 2\right)} \cdot \left(\alpha - \beta\right)\\
\end{array}
\end{array}
if (/.f64 (-.f64 beta alpha) (+.f64 (+.f64 alpha beta) 2)) < -0.99999499999999997Initial program 6.6%
Taylor expanded in alpha around inf 99.5%
if -0.99999499999999997 < (/.f64 (-.f64 beta alpha) (+.f64 (+.f64 alpha beta) 2)) Initial program 99.9%
+-commutative99.9%
sub-neg99.9%
+-commutative99.9%
neg-sub099.9%
associate-+l-99.9%
sub0-neg99.9%
distribute-frac-neg99.9%
+-commutative99.9%
sub-neg99.9%
div-sub99.9%
metadata-eval99.9%
+-commutative99.9%
associate-/l/99.9%
sub-neg99.9%
remove-double-neg99.9%
distribute-neg-out99.9%
neg-mul-199.9%
*-commutative99.9%
Simplified99.9%
Final simplification99.8%
(FPCore (alpha beta) :precision binary64 (if (<= alpha 2.6e+14) (/ (+ 1.0 (/ beta (+ beta 2.0))) 2.0) (/ (/ (+ beta 2.0) alpha) 2.0)))
double code(double alpha, double beta) {
double tmp;
if (alpha <= 2.6e+14) {
tmp = (1.0 + (beta / (beta + 2.0))) / 2.0;
} else {
tmp = ((beta + 2.0) / alpha) / 2.0;
}
return tmp;
}
real(8) function code(alpha, beta)
real(8), intent (in) :: alpha
real(8), intent (in) :: beta
real(8) :: tmp
if (alpha <= 2.6d+14) then
tmp = (1.0d0 + (beta / (beta + 2.0d0))) / 2.0d0
else
tmp = ((beta + 2.0d0) / alpha) / 2.0d0
end if
code = tmp
end function
public static double code(double alpha, double beta) {
double tmp;
if (alpha <= 2.6e+14) {
tmp = (1.0 + (beta / (beta + 2.0))) / 2.0;
} else {
tmp = ((beta + 2.0) / alpha) / 2.0;
}
return tmp;
}
def code(alpha, beta): tmp = 0 if alpha <= 2.6e+14: tmp = (1.0 + (beta / (beta + 2.0))) / 2.0 else: tmp = ((beta + 2.0) / alpha) / 2.0 return tmp
function code(alpha, beta) tmp = 0.0 if (alpha <= 2.6e+14) tmp = Float64(Float64(1.0 + Float64(beta / Float64(beta + 2.0))) / 2.0); else tmp = Float64(Float64(Float64(beta + 2.0) / alpha) / 2.0); end return tmp end
function tmp_2 = code(alpha, beta) tmp = 0.0; if (alpha <= 2.6e+14) tmp = (1.0 + (beta / (beta + 2.0))) / 2.0; else tmp = ((beta + 2.0) / alpha) / 2.0; end tmp_2 = tmp; end
code[alpha_, beta_] := If[LessEqual[alpha, 2.6e+14], N[(N[(1.0 + N[(beta / N[(beta + 2.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] / 2.0), $MachinePrecision], N[(N[(N[(beta + 2.0), $MachinePrecision] / alpha), $MachinePrecision] / 2.0), $MachinePrecision]]
\begin{array}{l}
\\
\begin{array}{l}
\mathbf{if}\;\alpha \leq 2.6 \cdot 10^{+14}:\\
\;\;\;\;\frac{1 + \frac{\beta}{\beta + 2}}{2}\\
\mathbf{else}:\\
\;\;\;\;\frac{\frac{\beta + 2}{\alpha}}{2}\\
\end{array}
\end{array}
if alpha < 2.6e14Initial program 99.6%
Taylor expanded in alpha around 0 97.0%
if 2.6e14 < alpha Initial program 22.3%
div-sub22.3%
associate-+l-24.9%
+-commutative24.9%
associate-+r+24.9%
+-commutative24.9%
associate-+r+24.9%
Applied egg-rr24.9%
sub-neg24.9%
flip-+24.9%
metadata-eval24.9%
metadata-eval24.9%
metadata-eval24.9%
sub-neg24.9%
pow224.9%
metadata-eval24.9%
metadata-eval24.9%
Applied egg-rr24.9%
Taylor expanded in alpha around inf 83.8%
mul-1-neg83.8%
distribute-neg-frac283.8%
Simplified83.8%
Taylor expanded in alpha around 0 68.5%
Final simplification87.4%
(FPCore (alpha beta) :precision binary64 (if (<= alpha 9.5e+14) (/ (+ 1.0 (/ beta (+ beta 2.0))) 2.0) (/ (/ (+ 2.0 (* beta 2.0)) alpha) 2.0)))
double code(double alpha, double beta) {
double tmp;
if (alpha <= 9.5e+14) {
tmp = (1.0 + (beta / (beta + 2.0))) / 2.0;
} else {
tmp = ((2.0 + (beta * 2.0)) / alpha) / 2.0;
}
return tmp;
}
real(8) function code(alpha, beta)
real(8), intent (in) :: alpha
real(8), intent (in) :: beta
real(8) :: tmp
if (alpha <= 9.5d+14) then
tmp = (1.0d0 + (beta / (beta + 2.0d0))) / 2.0d0
else
tmp = ((2.0d0 + (beta * 2.0d0)) / alpha) / 2.0d0
end if
code = tmp
end function
public static double code(double alpha, double beta) {
double tmp;
if (alpha <= 9.5e+14) {
tmp = (1.0 + (beta / (beta + 2.0))) / 2.0;
} else {
tmp = ((2.0 + (beta * 2.0)) / alpha) / 2.0;
}
return tmp;
}
def code(alpha, beta): tmp = 0 if alpha <= 9.5e+14: tmp = (1.0 + (beta / (beta + 2.0))) / 2.0 else: tmp = ((2.0 + (beta * 2.0)) / alpha) / 2.0 return tmp
function code(alpha, beta) tmp = 0.0 if (alpha <= 9.5e+14) tmp = Float64(Float64(1.0 + Float64(beta / Float64(beta + 2.0))) / 2.0); else tmp = Float64(Float64(Float64(2.0 + Float64(beta * 2.0)) / alpha) / 2.0); end return tmp end
function tmp_2 = code(alpha, beta) tmp = 0.0; if (alpha <= 9.5e+14) tmp = (1.0 + (beta / (beta + 2.0))) / 2.0; else tmp = ((2.0 + (beta * 2.0)) / alpha) / 2.0; end tmp_2 = tmp; end
code[alpha_, beta_] := If[LessEqual[alpha, 9.5e+14], N[(N[(1.0 + N[(beta / N[(beta + 2.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] / 2.0), $MachinePrecision], N[(N[(N[(2.0 + N[(beta * 2.0), $MachinePrecision]), $MachinePrecision] / alpha), $MachinePrecision] / 2.0), $MachinePrecision]]
\begin{array}{l}
\\
\begin{array}{l}
\mathbf{if}\;\alpha \leq 9.5 \cdot 10^{+14}:\\
\;\;\;\;\frac{1 + \frac{\beta}{\beta + 2}}{2}\\
\mathbf{else}:\\
\;\;\;\;\frac{\frac{2 + \beta \cdot 2}{\alpha}}{2}\\
\end{array}
\end{array}
if alpha < 9.5e14Initial program 99.6%
Taylor expanded in alpha around 0 97.0%
if 9.5e14 < alpha Initial program 22.3%
Taylor expanded in alpha around inf 83.8%
Final simplification92.6%
(FPCore (alpha beta) :precision binary64 (if (<= alpha 1.9e+14) 1.0 (/ (/ 2.0 alpha) 2.0)))
double code(double alpha, double beta) {
double tmp;
if (alpha <= 1.9e+14) {
tmp = 1.0;
} else {
tmp = (2.0 / alpha) / 2.0;
}
return tmp;
}
real(8) function code(alpha, beta)
real(8), intent (in) :: alpha
real(8), intent (in) :: beta
real(8) :: tmp
if (alpha <= 1.9d+14) then
tmp = 1.0d0
else
tmp = (2.0d0 / alpha) / 2.0d0
end if
code = tmp
end function
public static double code(double alpha, double beta) {
double tmp;
if (alpha <= 1.9e+14) {
tmp = 1.0;
} else {
tmp = (2.0 / alpha) / 2.0;
}
return tmp;
}
def code(alpha, beta): tmp = 0 if alpha <= 1.9e+14: tmp = 1.0 else: tmp = (2.0 / alpha) / 2.0 return tmp
function code(alpha, beta) tmp = 0.0 if (alpha <= 1.9e+14) tmp = 1.0; else tmp = Float64(Float64(2.0 / alpha) / 2.0); end return tmp end
function tmp_2 = code(alpha, beta) tmp = 0.0; if (alpha <= 1.9e+14) tmp = 1.0; else tmp = (2.0 / alpha) / 2.0; end tmp_2 = tmp; end
code[alpha_, beta_] := If[LessEqual[alpha, 1.9e+14], 1.0, N[(N[(2.0 / alpha), $MachinePrecision] / 2.0), $MachinePrecision]]
\begin{array}{l}
\\
\begin{array}{l}
\mathbf{if}\;\alpha \leq 1.9 \cdot 10^{+14}:\\
\;\;\;\;1\\
\mathbf{else}:\\
\;\;\;\;\frac{\frac{2}{\alpha}}{2}\\
\end{array}
\end{array}
if alpha < 1.9e14Initial program 99.6%
+-commutative99.6%
sub-neg99.6%
+-commutative99.6%
neg-sub099.6%
associate-+l-99.6%
sub0-neg99.6%
distribute-frac-neg99.6%
+-commutative99.6%
sub-neg99.6%
div-sub99.6%
metadata-eval99.6%
+-commutative99.6%
associate-/l/99.6%
sub-neg99.6%
remove-double-neg99.6%
distribute-neg-out99.6%
neg-mul-199.6%
*-commutative99.6%
Simplified99.6%
Taylor expanded in beta around inf 48.4%
if 1.9e14 < alpha Initial program 22.3%
Taylor expanded in beta around 0 5.2%
+-commutative5.2%
Simplified5.2%
Taylor expanded in alpha around inf 66.5%
Final simplification54.5%
(FPCore (alpha beta) :precision binary64 1.0)
double code(double alpha, double beta) {
return 1.0;
}
real(8) function code(alpha, beta)
real(8), intent (in) :: alpha
real(8), intent (in) :: beta
code = 1.0d0
end function
public static double code(double alpha, double beta) {
return 1.0;
}
def code(alpha, beta): return 1.0
function code(alpha, beta) return 1.0 end
function tmp = code(alpha, beta) tmp = 1.0; end
code[alpha_, beta_] := 1.0
\begin{array}{l}
\\
1
\end{array}
Initial program 73.6%
+-commutative73.6%
sub-neg73.6%
+-commutative73.6%
neg-sub073.6%
associate-+l-73.6%
sub0-neg73.6%
distribute-frac-neg73.6%
+-commutative73.6%
sub-neg73.6%
div-sub73.6%
metadata-eval73.6%
+-commutative73.6%
associate-/l/73.6%
sub-neg73.6%
remove-double-neg73.6%
distribute-neg-out73.6%
neg-mul-173.6%
*-commutative73.6%
Simplified73.7%
Taylor expanded in beta around inf 38.8%
Final simplification38.8%
herbie shell --seed 2024046
(FPCore (alpha beta)
:name "Octave 3.8, jcobi/1"
:precision binary64
:pre (and (> alpha -1.0) (> beta -1.0))
(/ (+ (/ (- beta alpha) (+ (+ alpha beta) 2.0)) 1.0) 2.0))