Average Error: 3.6 → 0.1
Time: 6.1s
Precision: binary64
\[\alpha > -1 \land \beta > -1\]
\[\frac{\frac{\frac{\left(\left(\alpha + \beta\right) + \beta \cdot \alpha\right) + 1}{\left(\alpha + \beta\right) + 2 \cdot 1}}{\left(\alpha + \beta\right) + 2 \cdot 1}}{\left(\left(\alpha + \beta\right) + 2 \cdot 1\right) + 1} \]
\[\begin{array}{l} t_0 := \alpha + \left(\beta + 2\right)\\ \sqrt{{\left(\frac{\beta + \left(\alpha + 3\right)}{\alpha + 1}\right)}^{-2}} \cdot \frac{\frac{\beta + 1}{t_0}}{t_0} \end{array} \]
(FPCore (alpha beta)
 :precision binary64
 (/
  (/
   (/ (+ (+ (+ alpha beta) (* beta alpha)) 1.0) (+ (+ alpha beta) (* 2.0 1.0)))
   (+ (+ alpha beta) (* 2.0 1.0)))
  (+ (+ (+ alpha beta) (* 2.0 1.0)) 1.0)))
(FPCore (alpha beta)
 :precision binary64
 (let* ((t_0 (+ alpha (+ beta 2.0))))
   (*
    (sqrt (pow (/ (+ beta (+ alpha 3.0)) (+ alpha 1.0)) -2.0))
    (/ (/ (+ beta 1.0) t_0) t_0))))
double code(double alpha, double beta) {
	return (((((alpha + beta) + (beta * alpha)) + 1.0) / ((alpha + beta) + (2.0 * 1.0))) / ((alpha + beta) + (2.0 * 1.0))) / (((alpha + beta) + (2.0 * 1.0)) + 1.0);
}
double code(double alpha, double beta) {
	double t_0 = alpha + (beta + 2.0);
	return sqrt(pow(((beta + (alpha + 3.0)) / (alpha + 1.0)), -2.0)) * (((beta + 1.0) / t_0) / t_0);
}
real(8) function code(alpha, beta)
    real(8), intent (in) :: alpha
    real(8), intent (in) :: beta
    code = (((((alpha + beta) + (beta * alpha)) + 1.0d0) / ((alpha + beta) + (2.0d0 * 1.0d0))) / ((alpha + beta) + (2.0d0 * 1.0d0))) / (((alpha + beta) + (2.0d0 * 1.0d0)) + 1.0d0)
end function
real(8) function code(alpha, beta)
    real(8), intent (in) :: alpha
    real(8), intent (in) :: beta
    real(8) :: t_0
    t_0 = alpha + (beta + 2.0d0)
    code = sqrt((((beta + (alpha + 3.0d0)) / (alpha + 1.0d0)) ** (-2.0d0))) * (((beta + 1.0d0) / t_0) / t_0)
end function
public static double code(double alpha, double beta) {
	return (((((alpha + beta) + (beta * alpha)) + 1.0) / ((alpha + beta) + (2.0 * 1.0))) / ((alpha + beta) + (2.0 * 1.0))) / (((alpha + beta) + (2.0 * 1.0)) + 1.0);
}
public static double code(double alpha, double beta) {
	double t_0 = alpha + (beta + 2.0);
	return Math.sqrt(Math.pow(((beta + (alpha + 3.0)) / (alpha + 1.0)), -2.0)) * (((beta + 1.0) / t_0) / t_0);
}
def code(alpha, beta):
	return (((((alpha + beta) + (beta * alpha)) + 1.0) / ((alpha + beta) + (2.0 * 1.0))) / ((alpha + beta) + (2.0 * 1.0))) / (((alpha + beta) + (2.0 * 1.0)) + 1.0)
def code(alpha, beta):
	t_0 = alpha + (beta + 2.0)
	return math.sqrt(math.pow(((beta + (alpha + 3.0)) / (alpha + 1.0)), -2.0)) * (((beta + 1.0) / t_0) / t_0)
function code(alpha, beta)
	return Float64(Float64(Float64(Float64(Float64(Float64(alpha + beta) + Float64(beta * alpha)) + 1.0) / Float64(Float64(alpha + beta) + Float64(2.0 * 1.0))) / Float64(Float64(alpha + beta) + Float64(2.0 * 1.0))) / Float64(Float64(Float64(alpha + beta) + Float64(2.0 * 1.0)) + 1.0))
end
function code(alpha, beta)
	t_0 = Float64(alpha + Float64(beta + 2.0))
	return Float64(sqrt((Float64(Float64(beta + Float64(alpha + 3.0)) / Float64(alpha + 1.0)) ^ -2.0)) * Float64(Float64(Float64(beta + 1.0) / t_0) / t_0))
end
function tmp = code(alpha, beta)
	tmp = (((((alpha + beta) + (beta * alpha)) + 1.0) / ((alpha + beta) + (2.0 * 1.0))) / ((alpha + beta) + (2.0 * 1.0))) / (((alpha + beta) + (2.0 * 1.0)) + 1.0);
end
function tmp = code(alpha, beta)
	t_0 = alpha + (beta + 2.0);
	tmp = sqrt((((beta + (alpha + 3.0)) / (alpha + 1.0)) ^ -2.0)) * (((beta + 1.0) / t_0) / t_0);
end
code[alpha_, beta_] := N[(N[(N[(N[(N[(N[(alpha + beta), $MachinePrecision] + N[(beta * alpha), $MachinePrecision]), $MachinePrecision] + 1.0), $MachinePrecision] / N[(N[(alpha + beta), $MachinePrecision] + N[(2.0 * 1.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] / N[(N[(alpha + beta), $MachinePrecision] + N[(2.0 * 1.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] / N[(N[(N[(alpha + beta), $MachinePrecision] + N[(2.0 * 1.0), $MachinePrecision]), $MachinePrecision] + 1.0), $MachinePrecision]), $MachinePrecision]
code[alpha_, beta_] := Block[{t$95$0 = N[(alpha + N[(beta + 2.0), $MachinePrecision]), $MachinePrecision]}, N[(N[Sqrt[N[Power[N[(N[(beta + N[(alpha + 3.0), $MachinePrecision]), $MachinePrecision] / N[(alpha + 1.0), $MachinePrecision]), $MachinePrecision], -2.0], $MachinePrecision]], $MachinePrecision] * N[(N[(N[(beta + 1.0), $MachinePrecision] / t$95$0), $MachinePrecision] / t$95$0), $MachinePrecision]), $MachinePrecision]]
\frac{\frac{\frac{\left(\left(\alpha + \beta\right) + \beta \cdot \alpha\right) + 1}{\left(\alpha + \beta\right) + 2 \cdot 1}}{\left(\alpha + \beta\right) + 2 \cdot 1}}{\left(\left(\alpha + \beta\right) + 2 \cdot 1\right) + 1}
\begin{array}{l}
t_0 := \alpha + \left(\beta + 2\right)\\
\sqrt{{\left(\frac{\beta + \left(\alpha + 3\right)}{\alpha + 1}\right)}^{-2}} \cdot \frac{\frac{\beta + 1}{t_0}}{t_0}
\end{array}

Error

Bits error versus alpha

Bits error versus beta

Try it out

Your Program's Arguments

Results

Enter valid numbers for all inputs

Derivation

  1. Initial program 3.6

    \[\frac{\frac{\frac{\left(\left(\alpha + \beta\right) + \beta \cdot \alpha\right) + 1}{\left(\alpha + \beta\right) + 2 \cdot 1}}{\left(\alpha + \beta\right) + 2 \cdot 1}}{\left(\left(\alpha + \beta\right) + 2 \cdot 1\right) + 1} \]
  2. Simplified2.2

    \[\leadsto \color{blue}{\frac{\alpha + 1}{\beta + \left(\alpha + 3\right)} \cdot \frac{\beta + 1}{\left(\alpha + \left(\beta + 2\right)\right) \cdot \left(\alpha + \left(\beta + 2\right)\right)}} \]
  3. Applied egg-rr0.3

    \[\leadsto \color{blue}{\frac{\frac{\beta + 1}{\alpha + \left(\beta + 2\right)}}{\frac{\beta + \left(\alpha + 3\right)}{1 + \alpha} \cdot \left(\alpha + \left(\beta + 2\right)\right)}} \]
  4. Applied egg-rr0.1

    \[\leadsto \color{blue}{\frac{1}{\frac{\left(\beta + \alpha\right) + 3}{1 + \alpha}} \cdot \frac{\frac{\beta + 1}{\alpha + \left(\beta + 2\right)}}{\alpha + \left(\beta + 2\right)}} \]
  5. Applied egg-rr0.1

    \[\leadsto \color{blue}{\sqrt{{\left(\frac{\beta + \left(\alpha + 3\right)}{1 + \alpha}\right)}^{-2}}} \cdot \frac{\frac{\beta + 1}{\alpha + \left(\beta + 2\right)}}{\alpha + \left(\beta + 2\right)} \]
  6. Final simplification0.1

    \[\leadsto \sqrt{{\left(\frac{\beta + \left(\alpha + 3\right)}{\alpha + 1}\right)}^{-2}} \cdot \frac{\frac{\beta + 1}{\alpha + \left(\beta + 2\right)}}{\alpha + \left(\beta + 2\right)} \]

Reproduce

herbie shell --seed 2022165 
(FPCore (alpha beta)
  :name "Octave 3.8, jcobi/3"
  :precision binary64
  :pre (and (> alpha -1.0) (> beta -1.0))
  (/ (/ (/ (+ (+ (+ alpha beta) (* beta alpha)) 1.0) (+ (+ alpha beta) (* 2.0 1.0))) (+ (+ alpha beta) (* 2.0 1.0))) (+ (+ (+ alpha beta) (* 2.0 1.0)) 1.0)))