
(FPCore (x) :precision binary64 (- (sqrt (+ 1.0 x)) (sqrt (- 1.0 x))))
double code(double x) {
return sqrt((1.0 + x)) - sqrt((1.0 - x));
}
real(8) function code(x)
real(8), intent (in) :: x
code = sqrt((1.0d0 + x)) - sqrt((1.0d0 - x))
end function
public static double code(double x) {
return Math.sqrt((1.0 + x)) - Math.sqrt((1.0 - x));
}
def code(x): return math.sqrt((1.0 + x)) - math.sqrt((1.0 - x))
function code(x) return Float64(sqrt(Float64(1.0 + x)) - sqrt(Float64(1.0 - x))) end
function tmp = code(x) tmp = sqrt((1.0 + x)) - sqrt((1.0 - x)); end
code[x_] := N[(N[Sqrt[N[(1.0 + x), $MachinePrecision]], $MachinePrecision] - N[Sqrt[N[(1.0 - x), $MachinePrecision]], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\sqrt{1 + x} - \sqrt{1 - x}
\end{array}
Sampling outcomes in binary64 precision:
Herbie found 5 alternatives:
| Alternative | Accuracy | Speedup |
|---|
(FPCore (x) :precision binary64 (- (sqrt (+ 1.0 x)) (sqrt (- 1.0 x))))
double code(double x) {
return sqrt((1.0 + x)) - sqrt((1.0 - x));
}
real(8) function code(x)
real(8), intent (in) :: x
code = sqrt((1.0d0 + x)) - sqrt((1.0d0 - x))
end function
public static double code(double x) {
return Math.sqrt((1.0 + x)) - Math.sqrt((1.0 - x));
}
def code(x): return math.sqrt((1.0 + x)) - math.sqrt((1.0 - x))
function code(x) return Float64(sqrt(Float64(1.0 + x)) - sqrt(Float64(1.0 - x))) end
function tmp = code(x) tmp = sqrt((1.0 + x)) - sqrt((1.0 - x)); end
code[x_] := N[(N[Sqrt[N[(1.0 + x), $MachinePrecision]], $MachinePrecision] - N[Sqrt[N[(1.0 - x), $MachinePrecision]], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\sqrt{1 + x} - \sqrt{1 - x}
\end{array}
(FPCore (x) :precision binary64 (+ x (* x (* (* x x) (+ 0.125 (* x (* x (+ 0.0546875 (* (* x x) 0.0322265625)))))))))
double code(double x) {
return x + (x * ((x * x) * (0.125 + (x * (x * (0.0546875 + ((x * x) * 0.0322265625)))))));
}
real(8) function code(x)
real(8), intent (in) :: x
code = x + (x * ((x * x) * (0.125d0 + (x * (x * (0.0546875d0 + ((x * x) * 0.0322265625d0)))))))
end function
public static double code(double x) {
return x + (x * ((x * x) * (0.125 + (x * (x * (0.0546875 + ((x * x) * 0.0322265625)))))));
}
def code(x): return x + (x * ((x * x) * (0.125 + (x * (x * (0.0546875 + ((x * x) * 0.0322265625)))))))
function code(x) return Float64(x + Float64(x * Float64(Float64(x * x) * Float64(0.125 + Float64(x * Float64(x * Float64(0.0546875 + Float64(Float64(x * x) * 0.0322265625)))))))) end
function tmp = code(x) tmp = x + (x * ((x * x) * (0.125 + (x * (x * (0.0546875 + ((x * x) * 0.0322265625))))))); end
code[x_] := N[(x + N[(x * N[(N[(x * x), $MachinePrecision] * N[(0.125 + N[(x * N[(x * N[(0.0546875 + N[(N[(x * x), $MachinePrecision] * 0.0322265625), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
x + x \cdot \left(\left(x \cdot x\right) \cdot \left(0.125 + x \cdot \left(x \cdot \left(0.0546875 + \left(x \cdot x\right) \cdot 0.0322265625\right)\right)\right)\right)
\end{array}
Initial program 7.3%
Taylor expanded in x around 0
*-lowering-*.f64N/A
+-lowering-+.f64N/A
*-lowering-*.f64N/A
unpow2N/A
*-lowering-*.f64N/A
+-lowering-+.f64N/A
unpow2N/A
associate-*l*N/A
*-lowering-*.f64N/A
*-lowering-*.f64N/A
+-lowering-+.f64N/A
*-commutativeN/A
*-lowering-*.f64N/A
unpow2N/A
*-lowering-*.f64100.0%
Simplified100.0%
+-commutativeN/A
distribute-rgt-inN/A
*-lft-identityN/A
+-lowering-+.f64N/A
Applied egg-rr100.0%
Final simplification100.0%
(FPCore (x) :precision binary64 (* x (+ (* (* x x) (+ 0.125 (* x (* x (+ 0.0546875 (* (* x x) 0.0322265625)))))) 1.0)))
double code(double x) {
return x * (((x * x) * (0.125 + (x * (x * (0.0546875 + ((x * x) * 0.0322265625)))))) + 1.0);
}
real(8) function code(x)
real(8), intent (in) :: x
code = x * (((x * x) * (0.125d0 + (x * (x * (0.0546875d0 + ((x * x) * 0.0322265625d0)))))) + 1.0d0)
end function
public static double code(double x) {
return x * (((x * x) * (0.125 + (x * (x * (0.0546875 + ((x * x) * 0.0322265625)))))) + 1.0);
}
def code(x): return x * (((x * x) * (0.125 + (x * (x * (0.0546875 + ((x * x) * 0.0322265625)))))) + 1.0)
function code(x) return Float64(x * Float64(Float64(Float64(x * x) * Float64(0.125 + Float64(x * Float64(x * Float64(0.0546875 + Float64(Float64(x * x) * 0.0322265625)))))) + 1.0)) end
function tmp = code(x) tmp = x * (((x * x) * (0.125 + (x * (x * (0.0546875 + ((x * x) * 0.0322265625)))))) + 1.0); end
code[x_] := N[(x * N[(N[(N[(x * x), $MachinePrecision] * N[(0.125 + N[(x * N[(x * N[(0.0546875 + N[(N[(x * x), $MachinePrecision] * 0.0322265625), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] + 1.0), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
x \cdot \left(\left(x \cdot x\right) \cdot \left(0.125 + x \cdot \left(x \cdot \left(0.0546875 + \left(x \cdot x\right) \cdot 0.0322265625\right)\right)\right) + 1\right)
\end{array}
Initial program 7.3%
Taylor expanded in x around 0
*-lowering-*.f64N/A
+-lowering-+.f64N/A
*-lowering-*.f64N/A
unpow2N/A
*-lowering-*.f64N/A
+-lowering-+.f64N/A
unpow2N/A
associate-*l*N/A
*-lowering-*.f64N/A
*-lowering-*.f64N/A
+-lowering-+.f64N/A
*-commutativeN/A
*-lowering-*.f64N/A
unpow2N/A
*-lowering-*.f64100.0%
Simplified100.0%
Final simplification100.0%
(FPCore (x) :precision binary64 (* x (+ 1.0 (* (* x x) (+ 0.125 (* (* x x) 0.0546875))))))
double code(double x) {
return x * (1.0 + ((x * x) * (0.125 + ((x * x) * 0.0546875))));
}
real(8) function code(x)
real(8), intent (in) :: x
code = x * (1.0d0 + ((x * x) * (0.125d0 + ((x * x) * 0.0546875d0))))
end function
public static double code(double x) {
return x * (1.0 + ((x * x) * (0.125 + ((x * x) * 0.0546875))));
}
def code(x): return x * (1.0 + ((x * x) * (0.125 + ((x * x) * 0.0546875))))
function code(x) return Float64(x * Float64(1.0 + Float64(Float64(x * x) * Float64(0.125 + Float64(Float64(x * x) * 0.0546875))))) end
function tmp = code(x) tmp = x * (1.0 + ((x * x) * (0.125 + ((x * x) * 0.0546875)))); end
code[x_] := N[(x * N[(1.0 + N[(N[(x * x), $MachinePrecision] * N[(0.125 + N[(N[(x * x), $MachinePrecision] * 0.0546875), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
x \cdot \left(1 + \left(x \cdot x\right) \cdot \left(0.125 + \left(x \cdot x\right) \cdot 0.0546875\right)\right)
\end{array}
Initial program 7.3%
Taylor expanded in x around 0
*-lowering-*.f64N/A
+-lowering-+.f64N/A
*-lowering-*.f64N/A
unpow2N/A
*-lowering-*.f64N/A
+-lowering-+.f64N/A
*-commutativeN/A
*-lowering-*.f64N/A
unpow2N/A
*-lowering-*.f6499.9%
Simplified99.9%
(FPCore (x) :precision binary64 (* x (+ 1.0 (* (* x x) 0.125))))
double code(double x) {
return x * (1.0 + ((x * x) * 0.125));
}
real(8) function code(x)
real(8), intent (in) :: x
code = x * (1.0d0 + ((x * x) * 0.125d0))
end function
public static double code(double x) {
return x * (1.0 + ((x * x) * 0.125));
}
def code(x): return x * (1.0 + ((x * x) * 0.125))
function code(x) return Float64(x * Float64(1.0 + Float64(Float64(x * x) * 0.125))) end
function tmp = code(x) tmp = x * (1.0 + ((x * x) * 0.125)); end
code[x_] := N[(x * N[(1.0 + N[(N[(x * x), $MachinePrecision] * 0.125), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
x \cdot \left(1 + \left(x \cdot x\right) \cdot 0.125\right)
\end{array}
Initial program 7.3%
Taylor expanded in x around 0
*-lowering-*.f64N/A
+-lowering-+.f64N/A
*-lowering-*.f64N/A
unpow2N/A
*-lowering-*.f6499.7%
Simplified99.7%
Final simplification99.7%
(FPCore (x) :precision binary64 x)
double code(double x) {
return x;
}
real(8) function code(x)
real(8), intent (in) :: x
code = x
end function
public static double code(double x) {
return x;
}
def code(x): return x
function code(x) return x end
function tmp = code(x) tmp = x; end
code[x_] := x
\begin{array}{l}
\\
x
\end{array}
Initial program 7.3%
Taylor expanded in x around 0
Simplified99.4%
(FPCore (x) :precision binary64 (/ (* 2.0 x) (+ (sqrt (+ 1.0 x)) (sqrt (- 1.0 x)))))
double code(double x) {
return (2.0 * x) / (sqrt((1.0 + x)) + sqrt((1.0 - x)));
}
real(8) function code(x)
real(8), intent (in) :: x
code = (2.0d0 * x) / (sqrt((1.0d0 + x)) + sqrt((1.0d0 - x)))
end function
public static double code(double x) {
return (2.0 * x) / (Math.sqrt((1.0 + x)) + Math.sqrt((1.0 - x)));
}
def code(x): return (2.0 * x) / (math.sqrt((1.0 + x)) + math.sqrt((1.0 - x)))
function code(x) return Float64(Float64(2.0 * x) / Float64(sqrt(Float64(1.0 + x)) + sqrt(Float64(1.0 - x)))) end
function tmp = code(x) tmp = (2.0 * x) / (sqrt((1.0 + x)) + sqrt((1.0 - x))); end
code[x_] := N[(N[(2.0 * x), $MachinePrecision] / N[(N[Sqrt[N[(1.0 + x), $MachinePrecision]], $MachinePrecision] + N[Sqrt[N[(1.0 - x), $MachinePrecision]], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\frac{2 \cdot x}{\sqrt{1 + x} + \sqrt{1 - x}}
\end{array}
herbie shell --seed 2024161
(FPCore (x)
:name "bug333 (missed optimization)"
:precision binary64
:pre (and (<= -1.0 x) (<= x 1.0))
:alt
(! :herbie-platform default (/ (* 2 x) (+ (sqrt (+ 1 x)) (sqrt (- 1 x)))))
(- (sqrt (+ 1.0 x)) (sqrt (- 1.0 x))))