
(FPCore (c_p c_n t s)
:precision binary64
(let* ((t_1 (/ 1.0 (+ 1.0 (exp (- t))))) (t_2 (/ 1.0 (+ 1.0 (exp (- s))))))
(/
(* (pow t_2 c_p) (pow (- 1.0 t_2) c_n))
(* (pow t_1 c_p) (pow (- 1.0 t_1) c_n)))))
double code(double c_p, double c_n, double t, double s) {
double t_1 = 1.0 / (1.0 + exp(-t));
double t_2 = 1.0 / (1.0 + exp(-s));
return (pow(t_2, c_p) * pow((1.0 - t_2), c_n)) / (pow(t_1, c_p) * pow((1.0 - t_1), c_n));
}
module fmin_fmax_functions
implicit none
private
public fmax
public fmin
interface fmax
module procedure fmax88
module procedure fmax44
module procedure fmax84
module procedure fmax48
end interface
interface fmin
module procedure fmin88
module procedure fmin44
module procedure fmin84
module procedure fmin48
end interface
contains
real(8) function fmax88(x, y) result (res)
real(8), intent (in) :: x
real(8), intent (in) :: y
res = merge(y, merge(x, max(x, y), y /= y), x /= x)
end function
real(4) function fmax44(x, y) result (res)
real(4), intent (in) :: x
real(4), intent (in) :: y
res = merge(y, merge(x, max(x, y), y /= y), x /= x)
end function
real(8) function fmax84(x, y) result(res)
real(8), intent (in) :: x
real(4), intent (in) :: y
res = merge(dble(y), merge(x, max(x, dble(y)), y /= y), x /= x)
end function
real(8) function fmax48(x, y) result(res)
real(4), intent (in) :: x
real(8), intent (in) :: y
res = merge(y, merge(dble(x), max(dble(x), y), y /= y), x /= x)
end function
real(8) function fmin88(x, y) result (res)
real(8), intent (in) :: x
real(8), intent (in) :: y
res = merge(y, merge(x, min(x, y), y /= y), x /= x)
end function
real(4) function fmin44(x, y) result (res)
real(4), intent (in) :: x
real(4), intent (in) :: y
res = merge(y, merge(x, min(x, y), y /= y), x /= x)
end function
real(8) function fmin84(x, y) result(res)
real(8), intent (in) :: x
real(4), intent (in) :: y
res = merge(dble(y), merge(x, min(x, dble(y)), y /= y), x /= x)
end function
real(8) function fmin48(x, y) result(res)
real(4), intent (in) :: x
real(8), intent (in) :: y
res = merge(y, merge(dble(x), min(dble(x), y), y /= y), x /= x)
end function
end module
real(8) function code(c_p, c_n, t, s)
use fmin_fmax_functions
real(8), intent (in) :: c_p
real(8), intent (in) :: c_n
real(8), intent (in) :: t
real(8), intent (in) :: s
real(8) :: t_1
real(8) :: t_2
t_1 = 1.0d0 / (1.0d0 + exp(-t))
t_2 = 1.0d0 / (1.0d0 + exp(-s))
code = ((t_2 ** c_p) * ((1.0d0 - t_2) ** c_n)) / ((t_1 ** c_p) * ((1.0d0 - t_1) ** c_n))
end function
public static double code(double c_p, double c_n, double t, double s) {
double t_1 = 1.0 / (1.0 + Math.exp(-t));
double t_2 = 1.0 / (1.0 + Math.exp(-s));
return (Math.pow(t_2, c_p) * Math.pow((1.0 - t_2), c_n)) / (Math.pow(t_1, c_p) * Math.pow((1.0 - t_1), c_n));
}
def code(c_p, c_n, t, s): t_1 = 1.0 / (1.0 + math.exp(-t)) t_2 = 1.0 / (1.0 + math.exp(-s)) return (math.pow(t_2, c_p) * math.pow((1.0 - t_2), c_n)) / (math.pow(t_1, c_p) * math.pow((1.0 - t_1), c_n))
function code(c_p, c_n, t, s) t_1 = Float64(1.0 / Float64(1.0 + exp(Float64(-t)))) t_2 = Float64(1.0 / Float64(1.0 + exp(Float64(-s)))) return Float64(Float64((t_2 ^ c_p) * (Float64(1.0 - t_2) ^ c_n)) / Float64((t_1 ^ c_p) * (Float64(1.0 - t_1) ^ c_n))) end
function tmp = code(c_p, c_n, t, s) t_1 = 1.0 / (1.0 + exp(-t)); t_2 = 1.0 / (1.0 + exp(-s)); tmp = ((t_2 ^ c_p) * ((1.0 - t_2) ^ c_n)) / ((t_1 ^ c_p) * ((1.0 - t_1) ^ c_n)); end
code[c$95$p_, c$95$n_, t_, s_] := Block[{t$95$1 = N[(1.0 / N[(1.0 + N[Exp[(-t)], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]}, Block[{t$95$2 = N[(1.0 / N[(1.0 + N[Exp[(-s)], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]}, N[(N[(N[Power[t$95$2, c$95$p], $MachinePrecision] * N[Power[N[(1.0 - t$95$2), $MachinePrecision], c$95$n], $MachinePrecision]), $MachinePrecision] / N[(N[Power[t$95$1, c$95$p], $MachinePrecision] * N[Power[N[(1.0 - t$95$1), $MachinePrecision], c$95$n], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]]]
\begin{array}{l}
\\
\begin{array}{l}
t_1 := \frac{1}{1 + e^{-t}}\\
t_2 := \frac{1}{1 + e^{-s}}\\
\frac{{t\_2}^{c\_p} \cdot {\left(1 - t\_2\right)}^{c\_n}}{{t\_1}^{c\_p} \cdot {\left(1 - t\_1\right)}^{c\_n}}
\end{array}
\end{array}
Sampling outcomes in binary64 precision:
Herbie found 9 alternatives:
| Alternative | Accuracy | Speedup |
|---|
(FPCore (c_p c_n t s)
:precision binary64
(let* ((t_1 (/ 1.0 (+ 1.0 (exp (- t))))) (t_2 (/ 1.0 (+ 1.0 (exp (- s))))))
(/
(* (pow t_2 c_p) (pow (- 1.0 t_2) c_n))
(* (pow t_1 c_p) (pow (- 1.0 t_1) c_n)))))
double code(double c_p, double c_n, double t, double s) {
double t_1 = 1.0 / (1.0 + exp(-t));
double t_2 = 1.0 / (1.0 + exp(-s));
return (pow(t_2, c_p) * pow((1.0 - t_2), c_n)) / (pow(t_1, c_p) * pow((1.0 - t_1), c_n));
}
module fmin_fmax_functions
implicit none
private
public fmax
public fmin
interface fmax
module procedure fmax88
module procedure fmax44
module procedure fmax84
module procedure fmax48
end interface
interface fmin
module procedure fmin88
module procedure fmin44
module procedure fmin84
module procedure fmin48
end interface
contains
real(8) function fmax88(x, y) result (res)
real(8), intent (in) :: x
real(8), intent (in) :: y
res = merge(y, merge(x, max(x, y), y /= y), x /= x)
end function
real(4) function fmax44(x, y) result (res)
real(4), intent (in) :: x
real(4), intent (in) :: y
res = merge(y, merge(x, max(x, y), y /= y), x /= x)
end function
real(8) function fmax84(x, y) result(res)
real(8), intent (in) :: x
real(4), intent (in) :: y
res = merge(dble(y), merge(x, max(x, dble(y)), y /= y), x /= x)
end function
real(8) function fmax48(x, y) result(res)
real(4), intent (in) :: x
real(8), intent (in) :: y
res = merge(y, merge(dble(x), max(dble(x), y), y /= y), x /= x)
end function
real(8) function fmin88(x, y) result (res)
real(8), intent (in) :: x
real(8), intent (in) :: y
res = merge(y, merge(x, min(x, y), y /= y), x /= x)
end function
real(4) function fmin44(x, y) result (res)
real(4), intent (in) :: x
real(4), intent (in) :: y
res = merge(y, merge(x, min(x, y), y /= y), x /= x)
end function
real(8) function fmin84(x, y) result(res)
real(8), intent (in) :: x
real(4), intent (in) :: y
res = merge(dble(y), merge(x, min(x, dble(y)), y /= y), x /= x)
end function
real(8) function fmin48(x, y) result(res)
real(4), intent (in) :: x
real(8), intent (in) :: y
res = merge(y, merge(dble(x), min(dble(x), y), y /= y), x /= x)
end function
end module
real(8) function code(c_p, c_n, t, s)
use fmin_fmax_functions
real(8), intent (in) :: c_p
real(8), intent (in) :: c_n
real(8), intent (in) :: t
real(8), intent (in) :: s
real(8) :: t_1
real(8) :: t_2
t_1 = 1.0d0 / (1.0d0 + exp(-t))
t_2 = 1.0d0 / (1.0d0 + exp(-s))
code = ((t_2 ** c_p) * ((1.0d0 - t_2) ** c_n)) / ((t_1 ** c_p) * ((1.0d0 - t_1) ** c_n))
end function
public static double code(double c_p, double c_n, double t, double s) {
double t_1 = 1.0 / (1.0 + Math.exp(-t));
double t_2 = 1.0 / (1.0 + Math.exp(-s));
return (Math.pow(t_2, c_p) * Math.pow((1.0 - t_2), c_n)) / (Math.pow(t_1, c_p) * Math.pow((1.0 - t_1), c_n));
}
def code(c_p, c_n, t, s): t_1 = 1.0 / (1.0 + math.exp(-t)) t_2 = 1.0 / (1.0 + math.exp(-s)) return (math.pow(t_2, c_p) * math.pow((1.0 - t_2), c_n)) / (math.pow(t_1, c_p) * math.pow((1.0 - t_1), c_n))
function code(c_p, c_n, t, s) t_1 = Float64(1.0 / Float64(1.0 + exp(Float64(-t)))) t_2 = Float64(1.0 / Float64(1.0 + exp(Float64(-s)))) return Float64(Float64((t_2 ^ c_p) * (Float64(1.0 - t_2) ^ c_n)) / Float64((t_1 ^ c_p) * (Float64(1.0 - t_1) ^ c_n))) end
function tmp = code(c_p, c_n, t, s) t_1 = 1.0 / (1.0 + exp(-t)); t_2 = 1.0 / (1.0 + exp(-s)); tmp = ((t_2 ^ c_p) * ((1.0 - t_2) ^ c_n)) / ((t_1 ^ c_p) * ((1.0 - t_1) ^ c_n)); end
code[c$95$p_, c$95$n_, t_, s_] := Block[{t$95$1 = N[(1.0 / N[(1.0 + N[Exp[(-t)], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]}, Block[{t$95$2 = N[(1.0 / N[(1.0 + N[Exp[(-s)], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]}, N[(N[(N[Power[t$95$2, c$95$p], $MachinePrecision] * N[Power[N[(1.0 - t$95$2), $MachinePrecision], c$95$n], $MachinePrecision]), $MachinePrecision] / N[(N[Power[t$95$1, c$95$p], $MachinePrecision] * N[Power[N[(1.0 - t$95$1), $MachinePrecision], c$95$n], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]]]
\begin{array}{l}
\\
\begin{array}{l}
t_1 := \frac{1}{1 + e^{-t}}\\
t_2 := \frac{1}{1 + e^{-s}}\\
\frac{{t\_2}^{c\_p} \cdot {\left(1 - t\_2\right)}^{c\_n}}{{t\_1}^{c\_p} \cdot {\left(1 - t\_1\right)}^{c\_n}}
\end{array}
\end{array}
(FPCore (c_p c_n t s)
:precision binary64
(let* ((t_1 (exp (- s))) (t_2 (log1p (exp (- t)))))
(if (<= (- s) 500000000.0)
(exp
(*
(- (log (- 1.0 (exp (- (log1p t_1))))) (log (- 1.0 (exp (- t_2)))))
c_n))
(/ (pow (+ t_1 1.0) (- c_p)) (fma t_2 (- c_p) 1.0)))))
double code(double c_p, double c_n, double t, double s) {
double t_1 = exp(-s);
double t_2 = log1p(exp(-t));
double tmp;
if (-s <= 500000000.0) {
tmp = exp(((log((1.0 - exp(-log1p(t_1)))) - log((1.0 - exp(-t_2)))) * c_n));
} else {
tmp = pow((t_1 + 1.0), -c_p) / fma(t_2, -c_p, 1.0);
}
return tmp;
}
function code(c_p, c_n, t, s) t_1 = exp(Float64(-s)) t_2 = log1p(exp(Float64(-t))) tmp = 0.0 if (Float64(-s) <= 500000000.0) tmp = exp(Float64(Float64(log(Float64(1.0 - exp(Float64(-log1p(t_1))))) - log(Float64(1.0 - exp(Float64(-t_2))))) * c_n)); else tmp = Float64((Float64(t_1 + 1.0) ^ Float64(-c_p)) / fma(t_2, Float64(-c_p), 1.0)); end return tmp end
code[c$95$p_, c$95$n_, t_, s_] := Block[{t$95$1 = N[Exp[(-s)], $MachinePrecision]}, Block[{t$95$2 = N[Log[1 + N[Exp[(-t)], $MachinePrecision]], $MachinePrecision]}, If[LessEqual[(-s), 500000000.0], N[Exp[N[(N[(N[Log[N[(1.0 - N[Exp[(-N[Log[1 + t$95$1], $MachinePrecision])], $MachinePrecision]), $MachinePrecision]], $MachinePrecision] - N[Log[N[(1.0 - N[Exp[(-t$95$2)], $MachinePrecision]), $MachinePrecision]], $MachinePrecision]), $MachinePrecision] * c$95$n), $MachinePrecision]], $MachinePrecision], N[(N[Power[N[(t$95$1 + 1.0), $MachinePrecision], (-c$95$p)], $MachinePrecision] / N[(t$95$2 * (-c$95$p) + 1.0), $MachinePrecision]), $MachinePrecision]]]]
\begin{array}{l}
\\
\begin{array}{l}
t_1 := e^{-s}\\
t_2 := \mathsf{log1p}\left(e^{-t}\right)\\
\mathbf{if}\;-s \leq 500000000:\\
\;\;\;\;e^{\left(\log \left(1 - e^{-\mathsf{log1p}\left(t\_1\right)}\right) - \log \left(1 - e^{-t\_2}\right)\right) \cdot c\_n}\\
\mathbf{else}:\\
\;\;\;\;\frac{{\left(t\_1 + 1\right)}^{\left(-c\_p\right)}}{\mathsf{fma}\left(t\_2, -c\_p, 1\right)}\\
\end{array}
\end{array}
if (neg.f64 s) < 5e8Initial program 92.8%
Applied rewrites97.2%
Taylor expanded in c_p around 0
Applied rewrites99.1%
if 5e8 < (neg.f64 s) Initial program 28.6%
Taylor expanded in c_n around 0
lower-/.f64N/A
lower-pow.f64N/A
lower-/.f64N/A
+-commutativeN/A
lower-+.f64N/A
lower-exp.f64N/A
lower-neg.f64N/A
lower-pow.f64N/A
lower-/.f64N/A
+-commutativeN/A
lower-+.f64N/A
lower-exp.f64N/A
lower-neg.f6471.4
Applied rewrites71.4%
Taylor expanded in c_p around 0
Applied rewrites100.0%
Applied rewrites100.0%
(FPCore (c_p c_n t s)
:precision binary64
(let* ((t_1 (log1p (exp (- s))))
(t_2
(exp
(-
(fma (log (- 1.0 (exp (- t_1)))) c_n (* (- c_p) t_1))
(fma (- c_p) (log 2.0) (* (log 0.5) c_n))))))
(fma (* t_2 (* -0.5 (- c_n c_p))) (- t) t_2)))
double code(double c_p, double c_n, double t, double s) {
double t_1 = log1p(exp(-s));
double t_2 = exp((fma(log((1.0 - exp(-t_1))), c_n, (-c_p * t_1)) - fma(-c_p, log(2.0), (log(0.5) * c_n))));
return fma((t_2 * (-0.5 * (c_n - c_p))), -t, t_2);
}
function code(c_p, c_n, t, s) t_1 = log1p(exp(Float64(-s))) t_2 = exp(Float64(fma(log(Float64(1.0 - exp(Float64(-t_1)))), c_n, Float64(Float64(-c_p) * t_1)) - fma(Float64(-c_p), log(2.0), Float64(log(0.5) * c_n)))) return fma(Float64(t_2 * Float64(-0.5 * Float64(c_n - c_p))), Float64(-t), t_2) end
code[c$95$p_, c$95$n_, t_, s_] := Block[{t$95$1 = N[Log[1 + N[Exp[(-s)], $MachinePrecision]], $MachinePrecision]}, Block[{t$95$2 = N[Exp[N[(N[(N[Log[N[(1.0 - N[Exp[(-t$95$1)], $MachinePrecision]), $MachinePrecision]], $MachinePrecision] * c$95$n + N[((-c$95$p) * t$95$1), $MachinePrecision]), $MachinePrecision] - N[((-c$95$p) * N[Log[2.0], $MachinePrecision] + N[(N[Log[0.5], $MachinePrecision] * c$95$n), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]}, N[(N[(t$95$2 * N[(-0.5 * N[(c$95$n - c$95$p), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] * (-t) + t$95$2), $MachinePrecision]]]
\begin{array}{l}
\\
\begin{array}{l}
t_1 := \mathsf{log1p}\left(e^{-s}\right)\\
t_2 := e^{\mathsf{fma}\left(\log \left(1 - e^{-t\_1}\right), c\_n, \left(-c\_p\right) \cdot t\_1\right) - \mathsf{fma}\left(-c\_p, \log 2, \log 0.5 \cdot c\_n\right)}\\
\mathsf{fma}\left(t\_2 \cdot \left(-0.5 \cdot \left(c\_n - c\_p\right)\right), -t, t\_2\right)
\end{array}
\end{array}
Initial program 91.1%
Applied rewrites96.1%
Taylor expanded in t around 0
Applied rewrites98.2%
Final simplification98.2%
(FPCore (c_p c_n t s)
:precision binary64
(let* ((t_1 (exp (- t)))
(t_2 (pow (+ 1.0 t_1) -1.0))
(t_3 (exp (- s)))
(t_4 (pow (+ 1.0 t_3) -1.0)))
(if (<=
(/
(* (pow t_4 c_p) (pow (- 1.0 t_4) c_n))
(* (pow t_2 c_p) (pow (- 1.0 t_2) c_n)))
INFINITY)
(/ (pow (+ t_3 1.0) (- c_p)) (pow 0.5 c_p))
(/ (pow 2.0 (- c_p)) (fma (log1p t_1) (- c_p) 1.0)))))
double code(double c_p, double c_n, double t, double s) {
double t_1 = exp(-t);
double t_2 = pow((1.0 + t_1), -1.0);
double t_3 = exp(-s);
double t_4 = pow((1.0 + t_3), -1.0);
double tmp;
if (((pow(t_4, c_p) * pow((1.0 - t_4), c_n)) / (pow(t_2, c_p) * pow((1.0 - t_2), c_n))) <= ((double) INFINITY)) {
tmp = pow((t_3 + 1.0), -c_p) / pow(0.5, c_p);
} else {
tmp = pow(2.0, -c_p) / fma(log1p(t_1), -c_p, 1.0);
}
return tmp;
}
function code(c_p, c_n, t, s) t_1 = exp(Float64(-t)) t_2 = Float64(1.0 + t_1) ^ -1.0 t_3 = exp(Float64(-s)) t_4 = Float64(1.0 + t_3) ^ -1.0 tmp = 0.0 if (Float64(Float64((t_4 ^ c_p) * (Float64(1.0 - t_4) ^ c_n)) / Float64((t_2 ^ c_p) * (Float64(1.0 - t_2) ^ c_n))) <= Inf) tmp = Float64((Float64(t_3 + 1.0) ^ Float64(-c_p)) / (0.5 ^ c_p)); else tmp = Float64((2.0 ^ Float64(-c_p)) / fma(log1p(t_1), Float64(-c_p), 1.0)); end return tmp end
code[c$95$p_, c$95$n_, t_, s_] := Block[{t$95$1 = N[Exp[(-t)], $MachinePrecision]}, Block[{t$95$2 = N[Power[N[(1.0 + t$95$1), $MachinePrecision], -1.0], $MachinePrecision]}, Block[{t$95$3 = N[Exp[(-s)], $MachinePrecision]}, Block[{t$95$4 = N[Power[N[(1.0 + t$95$3), $MachinePrecision], -1.0], $MachinePrecision]}, If[LessEqual[N[(N[(N[Power[t$95$4, c$95$p], $MachinePrecision] * N[Power[N[(1.0 - t$95$4), $MachinePrecision], c$95$n], $MachinePrecision]), $MachinePrecision] / N[(N[Power[t$95$2, c$95$p], $MachinePrecision] * N[Power[N[(1.0 - t$95$2), $MachinePrecision], c$95$n], $MachinePrecision]), $MachinePrecision]), $MachinePrecision], Infinity], N[(N[Power[N[(t$95$3 + 1.0), $MachinePrecision], (-c$95$p)], $MachinePrecision] / N[Power[0.5, c$95$p], $MachinePrecision]), $MachinePrecision], N[(N[Power[2.0, (-c$95$p)], $MachinePrecision] / N[(N[Log[1 + t$95$1], $MachinePrecision] * (-c$95$p) + 1.0), $MachinePrecision]), $MachinePrecision]]]]]]
\begin{array}{l}
\\
\begin{array}{l}
t_1 := e^{-t}\\
t_2 := {\left(1 + t\_1\right)}^{-1}\\
t_3 := e^{-s}\\
t_4 := {\left(1 + t\_3\right)}^{-1}\\
\mathbf{if}\;\frac{{t\_4}^{c\_p} \cdot {\left(1 - t\_4\right)}^{c\_n}}{{t\_2}^{c\_p} \cdot {\left(1 - t\_2\right)}^{c\_n}} \leq \infty:\\
\;\;\;\;\frac{{\left(t\_3 + 1\right)}^{\left(-c\_p\right)}}{{0.5}^{c\_p}}\\
\mathbf{else}:\\
\;\;\;\;\frac{{2}^{\left(-c\_p\right)}}{\mathsf{fma}\left(\mathsf{log1p}\left(t\_1\right), -c\_p, 1\right)}\\
\end{array}
\end{array}
if (/.f64 (*.f64 (pow.f64 (/.f64 #s(literal 1 binary64) (+.f64 #s(literal 1 binary64) (exp.f64 (neg.f64 s)))) c_p) (pow.f64 (-.f64 #s(literal 1 binary64) (/.f64 #s(literal 1 binary64) (+.f64 #s(literal 1 binary64) (exp.f64 (neg.f64 s))))) c_n)) (*.f64 (pow.f64 (/.f64 #s(literal 1 binary64) (+.f64 #s(literal 1 binary64) (exp.f64 (neg.f64 t)))) c_p) (pow.f64 (-.f64 #s(literal 1 binary64) (/.f64 #s(literal 1 binary64) (+.f64 #s(literal 1 binary64) (exp.f64 (neg.f64 t))))) c_n))) < +inf.0Initial program 98.8%
Taylor expanded in c_n around 0
lower-/.f64N/A
lower-pow.f64N/A
lower-/.f64N/A
+-commutativeN/A
lower-+.f64N/A
lower-exp.f64N/A
lower-neg.f64N/A
lower-pow.f64N/A
lower-/.f64N/A
+-commutativeN/A
lower-+.f64N/A
lower-exp.f64N/A
lower-neg.f6498.8
Applied rewrites98.8%
Taylor expanded in c_p around 0
Applied rewrites96.5%
Applied rewrites96.5%
Taylor expanded in t around 0
Applied rewrites98.9%
if +inf.0 < (/.f64 (*.f64 (pow.f64 (/.f64 #s(literal 1 binary64) (+.f64 #s(literal 1 binary64) (exp.f64 (neg.f64 s)))) c_p) (pow.f64 (-.f64 #s(literal 1 binary64) (/.f64 #s(literal 1 binary64) (+.f64 #s(literal 1 binary64) (exp.f64 (neg.f64 s))))) c_n)) (*.f64 (pow.f64 (/.f64 #s(literal 1 binary64) (+.f64 #s(literal 1 binary64) (exp.f64 (neg.f64 t)))) c_p) (pow.f64 (-.f64 #s(literal 1 binary64) (/.f64 #s(literal 1 binary64) (+.f64 #s(literal 1 binary64) (exp.f64 (neg.f64 t))))) c_n))) Initial program 0.0%
Taylor expanded in c_n around 0
lower-/.f64N/A
lower-pow.f64N/A
lower-/.f64N/A
+-commutativeN/A
lower-+.f64N/A
lower-exp.f64N/A
lower-neg.f64N/A
lower-pow.f64N/A
lower-/.f64N/A
+-commutativeN/A
lower-+.f64N/A
lower-exp.f64N/A
lower-neg.f6430.2
Applied rewrites30.2%
Taylor expanded in c_p around 0
Applied rewrites55.4%
Applied rewrites55.4%
Taylor expanded in s around 0
Applied rewrites55.5%
Final simplification95.5%
(FPCore (c_p c_n t s)
:precision binary64
(let* ((t_1 (exp (- t))) (t_2 (+ (exp (- s)) 1.0)))
(if (<= s -750000000.0)
(/ (pow t_2 (- c_p)) (fma (log1p t_1) (- c_p) 1.0))
(/
(pow (- 1.0 (pow t_2 -1.0)) c_n)
(pow (- 1.0 (pow (+ t_1 1.0) -1.0)) c_n)))))
double code(double c_p, double c_n, double t, double s) {
double t_1 = exp(-t);
double t_2 = exp(-s) + 1.0;
double tmp;
if (s <= -750000000.0) {
tmp = pow(t_2, -c_p) / fma(log1p(t_1), -c_p, 1.0);
} else {
tmp = pow((1.0 - pow(t_2, -1.0)), c_n) / pow((1.0 - pow((t_1 + 1.0), -1.0)), c_n);
}
return tmp;
}
function code(c_p, c_n, t, s) t_1 = exp(Float64(-t)) t_2 = Float64(exp(Float64(-s)) + 1.0) tmp = 0.0 if (s <= -750000000.0) tmp = Float64((t_2 ^ Float64(-c_p)) / fma(log1p(t_1), Float64(-c_p), 1.0)); else tmp = Float64((Float64(1.0 - (t_2 ^ -1.0)) ^ c_n) / (Float64(1.0 - (Float64(t_1 + 1.0) ^ -1.0)) ^ c_n)); end return tmp end
code[c$95$p_, c$95$n_, t_, s_] := Block[{t$95$1 = N[Exp[(-t)], $MachinePrecision]}, Block[{t$95$2 = N[(N[Exp[(-s)], $MachinePrecision] + 1.0), $MachinePrecision]}, If[LessEqual[s, -750000000.0], N[(N[Power[t$95$2, (-c$95$p)], $MachinePrecision] / N[(N[Log[1 + t$95$1], $MachinePrecision] * (-c$95$p) + 1.0), $MachinePrecision]), $MachinePrecision], N[(N[Power[N[(1.0 - N[Power[t$95$2, -1.0], $MachinePrecision]), $MachinePrecision], c$95$n], $MachinePrecision] / N[Power[N[(1.0 - N[Power[N[(t$95$1 + 1.0), $MachinePrecision], -1.0], $MachinePrecision]), $MachinePrecision], c$95$n], $MachinePrecision]), $MachinePrecision]]]]
\begin{array}{l}
\\
\begin{array}{l}
t_1 := e^{-t}\\
t_2 := e^{-s} + 1\\
\mathbf{if}\;s \leq -750000000:\\
\;\;\;\;\frac{{t\_2}^{\left(-c\_p\right)}}{\mathsf{fma}\left(\mathsf{log1p}\left(t\_1\right), -c\_p, 1\right)}\\
\mathbf{else}:\\
\;\;\;\;\frac{{\left(1 - {t\_2}^{-1}\right)}^{c\_n}}{{\left(1 - {\left(t\_1 + 1\right)}^{-1}\right)}^{c\_n}}\\
\end{array}
\end{array}
if s < -7.5e8Initial program 28.6%
Taylor expanded in c_n around 0
lower-/.f64N/A
lower-pow.f64N/A
lower-/.f64N/A
+-commutativeN/A
lower-+.f64N/A
lower-exp.f64N/A
lower-neg.f64N/A
lower-pow.f64N/A
lower-/.f64N/A
+-commutativeN/A
lower-+.f64N/A
lower-exp.f64N/A
lower-neg.f6471.4
Applied rewrites71.4%
Taylor expanded in c_p around 0
Applied rewrites100.0%
Applied rewrites100.0%
if -7.5e8 < s Initial program 92.8%
Taylor expanded in c_p around 0
lower-/.f64N/A
lower-pow.f64N/A
lower--.f64N/A
lower-/.f64N/A
+-commutativeN/A
lower-+.f64N/A
lower-exp.f64N/A
lower-neg.f64N/A
lower-pow.f64N/A
Applied rewrites96.0%
Final simplification96.1%
(FPCore (c_p c_n t s)
:precision binary64
(let* ((t_1 (exp (- s))))
(if (<= s -750000000.0)
(/ (pow (+ t_1 1.0) (- c_p)) (fma (log1p (exp (- t))) (- c_p) 1.0))
(exp (* c_n (log (/ (- 1.0 (exp (- (log1p t_1)))) (- 1.0 0.5))))))))
double code(double c_p, double c_n, double t, double s) {
double t_1 = exp(-s);
double tmp;
if (s <= -750000000.0) {
tmp = pow((t_1 + 1.0), -c_p) / fma(log1p(exp(-t)), -c_p, 1.0);
} else {
tmp = exp((c_n * log(((1.0 - exp(-log1p(t_1))) / (1.0 - 0.5)))));
}
return tmp;
}
function code(c_p, c_n, t, s) t_1 = exp(Float64(-s)) tmp = 0.0 if (s <= -750000000.0) tmp = Float64((Float64(t_1 + 1.0) ^ Float64(-c_p)) / fma(log1p(exp(Float64(-t))), Float64(-c_p), 1.0)); else tmp = exp(Float64(c_n * log(Float64(Float64(1.0 - exp(Float64(-log1p(t_1)))) / Float64(1.0 - 0.5))))); end return tmp end
code[c$95$p_, c$95$n_, t_, s_] := Block[{t$95$1 = N[Exp[(-s)], $MachinePrecision]}, If[LessEqual[s, -750000000.0], N[(N[Power[N[(t$95$1 + 1.0), $MachinePrecision], (-c$95$p)], $MachinePrecision] / N[(N[Log[1 + N[Exp[(-t)], $MachinePrecision]], $MachinePrecision] * (-c$95$p) + 1.0), $MachinePrecision]), $MachinePrecision], N[Exp[N[(c$95$n * N[Log[N[(N[(1.0 - N[Exp[(-N[Log[1 + t$95$1], $MachinePrecision])], $MachinePrecision]), $MachinePrecision] / N[(1.0 - 0.5), $MachinePrecision]), $MachinePrecision]], $MachinePrecision]), $MachinePrecision]], $MachinePrecision]]]
\begin{array}{l}
\\
\begin{array}{l}
t_1 := e^{-s}\\
\mathbf{if}\;s \leq -750000000:\\
\;\;\;\;\frac{{\left(t\_1 + 1\right)}^{\left(-c\_p\right)}}{\mathsf{fma}\left(\mathsf{log1p}\left(e^{-t}\right), -c\_p, 1\right)}\\
\mathbf{else}:\\
\;\;\;\;e^{c\_n \cdot \log \left(\frac{1 - e^{-\mathsf{log1p}\left(t\_1\right)}}{1 - 0.5}\right)}\\
\end{array}
\end{array}
if s < -7.5e8Initial program 28.6%
Taylor expanded in c_n around 0
lower-/.f64N/A
lower-pow.f64N/A
lower-/.f64N/A
+-commutativeN/A
lower-+.f64N/A
lower-exp.f64N/A
lower-neg.f64N/A
lower-pow.f64N/A
lower-/.f64N/A
+-commutativeN/A
lower-+.f64N/A
lower-exp.f64N/A
lower-neg.f6471.4
Applied rewrites71.4%
Taylor expanded in c_p around 0
Applied rewrites100.0%
Applied rewrites100.0%
if -7.5e8 < s Initial program 92.8%
Taylor expanded in c_p around 0
lower-/.f64N/A
lower-pow.f64N/A
lower--.f64N/A
lower-/.f64N/A
+-commutativeN/A
lower-+.f64N/A
lower-exp.f64N/A
lower-neg.f64N/A
lower-pow.f64N/A
Applied rewrites96.0%
Taylor expanded in t around 0
Applied rewrites95.3%
Applied rewrites95.8%
Applied rewrites98.5%
(FPCore (c_p c_n t s) :precision binary64 (if (<= s -2700000.0) (/ (pow (+ (exp (- s)) 1.0) (- c_p)) (fma (log1p (exp (- t))) (- c_p) 1.0)) 1.0))
double code(double c_p, double c_n, double t, double s) {
double tmp;
if (s <= -2700000.0) {
tmp = pow((exp(-s) + 1.0), -c_p) / fma(log1p(exp(-t)), -c_p, 1.0);
} else {
tmp = 1.0;
}
return tmp;
}
function code(c_p, c_n, t, s) tmp = 0.0 if (s <= -2700000.0) tmp = Float64((Float64(exp(Float64(-s)) + 1.0) ^ Float64(-c_p)) / fma(log1p(exp(Float64(-t))), Float64(-c_p), 1.0)); else tmp = 1.0; end return tmp end
code[c$95$p_, c$95$n_, t_, s_] := If[LessEqual[s, -2700000.0], N[(N[Power[N[(N[Exp[(-s)], $MachinePrecision] + 1.0), $MachinePrecision], (-c$95$p)], $MachinePrecision] / N[(N[Log[1 + N[Exp[(-t)], $MachinePrecision]], $MachinePrecision] * (-c$95$p) + 1.0), $MachinePrecision]), $MachinePrecision], 1.0]
\begin{array}{l}
\\
\begin{array}{l}
\mathbf{if}\;s \leq -2700000:\\
\;\;\;\;\frac{{\left(e^{-s} + 1\right)}^{\left(-c\_p\right)}}{\mathsf{fma}\left(\mathsf{log1p}\left(e^{-t}\right), -c\_p, 1\right)}\\
\mathbf{else}:\\
\;\;\;\;1\\
\end{array}
\end{array}
if s < -2.7e6Initial program 33.3%
Taylor expanded in c_n around 0
lower-/.f64N/A
lower-pow.f64N/A
lower-/.f64N/A
+-commutativeN/A
lower-+.f64N/A
lower-exp.f64N/A
lower-neg.f64N/A
lower-pow.f64N/A
lower-/.f64N/A
+-commutativeN/A
lower-+.f64N/A
lower-exp.f64N/A
lower-neg.f6466.8
Applied rewrites66.8%
Taylor expanded in c_p around 0
Applied rewrites89.1%
Applied rewrites89.1%
if -2.7e6 < s Initial program 93.2%
Taylor expanded in c_n around 0
lower-/.f64N/A
lower-pow.f64N/A
lower-/.f64N/A
+-commutativeN/A
lower-+.f64N/A
lower-exp.f64N/A
lower-neg.f64N/A
lower-pow.f64N/A
lower-/.f64N/A
+-commutativeN/A
lower-+.f64N/A
lower-exp.f64N/A
lower-neg.f6494.4
Applied rewrites94.4%
Taylor expanded in c_p around 0
Applied rewrites96.1%
(FPCore (c_p c_n t s) :precision binary64 (/ (pow (pow (fma (- (* 0.5 s) 1.0) s 2.0) -1.0) c_p) (pow (fma 0.25 t 0.5) c_p)))
double code(double c_p, double c_n, double t, double s) {
return pow(pow(fma(((0.5 * s) - 1.0), s, 2.0), -1.0), c_p) / pow(fma(0.25, t, 0.5), c_p);
}
function code(c_p, c_n, t, s) return Float64(((fma(Float64(Float64(0.5 * s) - 1.0), s, 2.0) ^ -1.0) ^ c_p) / (fma(0.25, t, 0.5) ^ c_p)) end
code[c$95$p_, c$95$n_, t_, s_] := N[(N[Power[N[Power[N[(N[(N[(0.5 * s), $MachinePrecision] - 1.0), $MachinePrecision] * s + 2.0), $MachinePrecision], -1.0], $MachinePrecision], c$95$p], $MachinePrecision] / N[Power[N[(0.25 * t + 0.5), $MachinePrecision], c$95$p], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\frac{{\left({\left(\mathsf{fma}\left(0.5 \cdot s - 1, s, 2\right)\right)}^{-1}\right)}^{c\_p}}{{\left(\mathsf{fma}\left(0.25, t, 0.5\right)\right)}^{c\_p}}
\end{array}
Initial program 91.1%
Taylor expanded in c_n around 0
lower-/.f64N/A
lower-pow.f64N/A
lower-/.f64N/A
+-commutativeN/A
lower-+.f64N/A
lower-exp.f64N/A
lower-neg.f64N/A
lower-pow.f64N/A
lower-/.f64N/A
+-commutativeN/A
lower-+.f64N/A
lower-exp.f64N/A
lower-neg.f6493.4
Applied rewrites93.4%
Taylor expanded in t around 0
Applied rewrites93.0%
Taylor expanded in s around 0
Applied rewrites93.4%
Final simplification93.4%
(FPCore (c_p c_n t s) :precision binary64 (if (<= s -100000.0) (/ (pow (pow (- 2.0 s) -1.0) c_p) (pow (fma 0.25 t 0.5) c_p)) 1.0))
double code(double c_p, double c_n, double t, double s) {
double tmp;
if (s <= -100000.0) {
tmp = pow(pow((2.0 - s), -1.0), c_p) / pow(fma(0.25, t, 0.5), c_p);
} else {
tmp = 1.0;
}
return tmp;
}
function code(c_p, c_n, t, s) tmp = 0.0 if (s <= -100000.0) tmp = Float64(((Float64(2.0 - s) ^ -1.0) ^ c_p) / (fma(0.25, t, 0.5) ^ c_p)); else tmp = 1.0; end return tmp end
code[c$95$p_, c$95$n_, t_, s_] := If[LessEqual[s, -100000.0], N[(N[Power[N[Power[N[(2.0 - s), $MachinePrecision], -1.0], $MachinePrecision], c$95$p], $MachinePrecision] / N[Power[N[(0.25 * t + 0.5), $MachinePrecision], c$95$p], $MachinePrecision]), $MachinePrecision], 1.0]
\begin{array}{l}
\\
\begin{array}{l}
\mathbf{if}\;s \leq -100000:\\
\;\;\;\;\frac{{\left({\left(2 - s\right)}^{-1}\right)}^{c\_p}}{{\left(\mathsf{fma}\left(0.25, t, 0.5\right)\right)}^{c\_p}}\\
\mathbf{else}:\\
\;\;\;\;1\\
\end{array}
\end{array}
if s < -1e5Initial program 33.3%
Taylor expanded in c_n around 0
lower-/.f64N/A
lower-pow.f64N/A
lower-/.f64N/A
+-commutativeN/A
lower-+.f64N/A
lower-exp.f64N/A
lower-neg.f64N/A
lower-pow.f64N/A
lower-/.f64N/A
+-commutativeN/A
lower-+.f64N/A
lower-exp.f64N/A
lower-neg.f6466.8
Applied rewrites66.8%
Taylor expanded in t around 0
Applied rewrites66.8%
Taylor expanded in s around 0
Applied rewrites56.3%
if -1e5 < s Initial program 93.2%
Taylor expanded in c_n around 0
lower-/.f64N/A
lower-pow.f64N/A
lower-/.f64N/A
+-commutativeN/A
lower-+.f64N/A
lower-exp.f64N/A
lower-neg.f64N/A
lower-pow.f64N/A
lower-/.f64N/A
+-commutativeN/A
lower-+.f64N/A
lower-exp.f64N/A
lower-neg.f6494.4
Applied rewrites94.4%
Taylor expanded in c_p around 0
Applied rewrites96.1%
Final simplification94.7%
(FPCore (c_p c_n t s) :precision binary64 1.0)
double code(double c_p, double c_n, double t, double s) {
return 1.0;
}
module fmin_fmax_functions
implicit none
private
public fmax
public fmin
interface fmax
module procedure fmax88
module procedure fmax44
module procedure fmax84
module procedure fmax48
end interface
interface fmin
module procedure fmin88
module procedure fmin44
module procedure fmin84
module procedure fmin48
end interface
contains
real(8) function fmax88(x, y) result (res)
real(8), intent (in) :: x
real(8), intent (in) :: y
res = merge(y, merge(x, max(x, y), y /= y), x /= x)
end function
real(4) function fmax44(x, y) result (res)
real(4), intent (in) :: x
real(4), intent (in) :: y
res = merge(y, merge(x, max(x, y), y /= y), x /= x)
end function
real(8) function fmax84(x, y) result(res)
real(8), intent (in) :: x
real(4), intent (in) :: y
res = merge(dble(y), merge(x, max(x, dble(y)), y /= y), x /= x)
end function
real(8) function fmax48(x, y) result(res)
real(4), intent (in) :: x
real(8), intent (in) :: y
res = merge(y, merge(dble(x), max(dble(x), y), y /= y), x /= x)
end function
real(8) function fmin88(x, y) result (res)
real(8), intent (in) :: x
real(8), intent (in) :: y
res = merge(y, merge(x, min(x, y), y /= y), x /= x)
end function
real(4) function fmin44(x, y) result (res)
real(4), intent (in) :: x
real(4), intent (in) :: y
res = merge(y, merge(x, min(x, y), y /= y), x /= x)
end function
real(8) function fmin84(x, y) result(res)
real(8), intent (in) :: x
real(4), intent (in) :: y
res = merge(dble(y), merge(x, min(x, dble(y)), y /= y), x /= x)
end function
real(8) function fmin48(x, y) result(res)
real(4), intent (in) :: x
real(8), intent (in) :: y
res = merge(y, merge(dble(x), min(dble(x), y), y /= y), x /= x)
end function
end module
real(8) function code(c_p, c_n, t, s)
use fmin_fmax_functions
real(8), intent (in) :: c_p
real(8), intent (in) :: c_n
real(8), intent (in) :: t
real(8), intent (in) :: s
code = 1.0d0
end function
public static double code(double c_p, double c_n, double t, double s) {
return 1.0;
}
def code(c_p, c_n, t, s): return 1.0
function code(c_p, c_n, t, s) return 1.0 end
function tmp = code(c_p, c_n, t, s) tmp = 1.0; end
code[c$95$p_, c$95$n_, t_, s_] := 1.0
\begin{array}{l}
\\
1
\end{array}
Initial program 91.1%
Taylor expanded in c_n around 0
lower-/.f64N/A
lower-pow.f64N/A
lower-/.f64N/A
+-commutativeN/A
lower-+.f64N/A
lower-exp.f64N/A
lower-neg.f64N/A
lower-pow.f64N/A
lower-/.f64N/A
+-commutativeN/A
lower-+.f64N/A
lower-exp.f64N/A
lower-neg.f6493.4
Applied rewrites93.4%
Taylor expanded in c_p around 0
Applied rewrites92.9%
(FPCore (c_p c_n t s) :precision binary64 (* (pow (/ (+ 1.0 (exp (- t))) (+ 1.0 (exp (- s)))) c_p) (pow (/ (+ 1.0 (exp t)) (+ 1.0 (exp s))) c_n)))
double code(double c_p, double c_n, double t, double s) {
return pow(((1.0 + exp(-t)) / (1.0 + exp(-s))), c_p) * pow(((1.0 + exp(t)) / (1.0 + exp(s))), c_n);
}
module fmin_fmax_functions
implicit none
private
public fmax
public fmin
interface fmax
module procedure fmax88
module procedure fmax44
module procedure fmax84
module procedure fmax48
end interface
interface fmin
module procedure fmin88
module procedure fmin44
module procedure fmin84
module procedure fmin48
end interface
contains
real(8) function fmax88(x, y) result (res)
real(8), intent (in) :: x
real(8), intent (in) :: y
res = merge(y, merge(x, max(x, y), y /= y), x /= x)
end function
real(4) function fmax44(x, y) result (res)
real(4), intent (in) :: x
real(4), intent (in) :: y
res = merge(y, merge(x, max(x, y), y /= y), x /= x)
end function
real(8) function fmax84(x, y) result(res)
real(8), intent (in) :: x
real(4), intent (in) :: y
res = merge(dble(y), merge(x, max(x, dble(y)), y /= y), x /= x)
end function
real(8) function fmax48(x, y) result(res)
real(4), intent (in) :: x
real(8), intent (in) :: y
res = merge(y, merge(dble(x), max(dble(x), y), y /= y), x /= x)
end function
real(8) function fmin88(x, y) result (res)
real(8), intent (in) :: x
real(8), intent (in) :: y
res = merge(y, merge(x, min(x, y), y /= y), x /= x)
end function
real(4) function fmin44(x, y) result (res)
real(4), intent (in) :: x
real(4), intent (in) :: y
res = merge(y, merge(x, min(x, y), y /= y), x /= x)
end function
real(8) function fmin84(x, y) result(res)
real(8), intent (in) :: x
real(4), intent (in) :: y
res = merge(dble(y), merge(x, min(x, dble(y)), y /= y), x /= x)
end function
real(8) function fmin48(x, y) result(res)
real(4), intent (in) :: x
real(8), intent (in) :: y
res = merge(y, merge(dble(x), min(dble(x), y), y /= y), x /= x)
end function
end module
real(8) function code(c_p, c_n, t, s)
use fmin_fmax_functions
real(8), intent (in) :: c_p
real(8), intent (in) :: c_n
real(8), intent (in) :: t
real(8), intent (in) :: s
code = (((1.0d0 + exp(-t)) / (1.0d0 + exp(-s))) ** c_p) * (((1.0d0 + exp(t)) / (1.0d0 + exp(s))) ** c_n)
end function
public static double code(double c_p, double c_n, double t, double s) {
return Math.pow(((1.0 + Math.exp(-t)) / (1.0 + Math.exp(-s))), c_p) * Math.pow(((1.0 + Math.exp(t)) / (1.0 + Math.exp(s))), c_n);
}
def code(c_p, c_n, t, s): return math.pow(((1.0 + math.exp(-t)) / (1.0 + math.exp(-s))), c_p) * math.pow(((1.0 + math.exp(t)) / (1.0 + math.exp(s))), c_n)
function code(c_p, c_n, t, s) return Float64((Float64(Float64(1.0 + exp(Float64(-t))) / Float64(1.0 + exp(Float64(-s)))) ^ c_p) * (Float64(Float64(1.0 + exp(t)) / Float64(1.0 + exp(s))) ^ c_n)) end
function tmp = code(c_p, c_n, t, s) tmp = (((1.0 + exp(-t)) / (1.0 + exp(-s))) ^ c_p) * (((1.0 + exp(t)) / (1.0 + exp(s))) ^ c_n); end
code[c$95$p_, c$95$n_, t_, s_] := N[(N[Power[N[(N[(1.0 + N[Exp[(-t)], $MachinePrecision]), $MachinePrecision] / N[(1.0 + N[Exp[(-s)], $MachinePrecision]), $MachinePrecision]), $MachinePrecision], c$95$p], $MachinePrecision] * N[Power[N[(N[(1.0 + N[Exp[t], $MachinePrecision]), $MachinePrecision] / N[(1.0 + N[Exp[s], $MachinePrecision]), $MachinePrecision]), $MachinePrecision], c$95$n], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
{\left(\frac{1 + e^{-t}}{1 + e^{-s}}\right)}^{c\_p} \cdot {\left(\frac{1 + e^{t}}{1 + e^{s}}\right)}^{c\_n}
\end{array}
herbie shell --seed 2024360
(FPCore (c_p c_n t s)
:name "Harley's example"
:precision binary64
:pre (and (< 0.0 c_p) (< 0.0 c_n))
:alt
(! :herbie-platform default (* (pow (/ (+ 1 (exp (- t))) (+ 1 (exp (- s)))) c_p) (pow (/ (+ 1 (exp t)) (+ 1 (exp s))) c_n)))
(/ (* (pow (/ 1.0 (+ 1.0 (exp (- s)))) c_p) (pow (- 1.0 (/ 1.0 (+ 1.0 (exp (- s))))) c_n)) (* (pow (/ 1.0 (+ 1.0 (exp (- t)))) c_p) (pow (- 1.0 (/ 1.0 (+ 1.0 (exp (- t))))) c_n))))