
(FPCore (u v) :precision binary32 (+ 1.0 (* v (log (+ u (* (- 1.0 u) (exp (/ -2.0 v))))))))
float code(float u, float v) {
return 1.0f + (v * logf((u + ((1.0f - u) * expf((-2.0f / v))))));
}
real(4) function code(u, v)
real(4), intent (in) :: u
real(4), intent (in) :: v
code = 1.0e0 + (v * log((u + ((1.0e0 - u) * exp(((-2.0e0) / v))))))
end function
function code(u, v) return Float32(Float32(1.0) + Float32(v * log(Float32(u + Float32(Float32(Float32(1.0) - u) * exp(Float32(Float32(-2.0) / v))))))) end
function tmp = code(u, v) tmp = single(1.0) + (v * log((u + ((single(1.0) - u) * exp((single(-2.0) / v)))))); end
\begin{array}{l}
\\
1 + v \cdot \log \left(u + \left(1 - u\right) \cdot e^{\frac{-2}{v}}\right)
\end{array}
Sampling outcomes in binary32 precision:
Herbie found 5 alternatives:
| Alternative | Accuracy | Speedup |
|---|
(FPCore (u v) :precision binary32 (+ 1.0 (* v (log (+ u (* (- 1.0 u) (exp (/ -2.0 v))))))))
float code(float u, float v) {
return 1.0f + (v * logf((u + ((1.0f - u) * expf((-2.0f / v))))));
}
real(4) function code(u, v)
real(4), intent (in) :: u
real(4), intent (in) :: v
code = 1.0e0 + (v * log((u + ((1.0e0 - u) * exp(((-2.0e0) / v))))))
end function
function code(u, v) return Float32(Float32(1.0) + Float32(v * log(Float32(u + Float32(Float32(Float32(1.0) - u) * exp(Float32(Float32(-2.0) / v))))))) end
function tmp = code(u, v) tmp = single(1.0) + (v * log((u + ((single(1.0) - u) * exp((single(-2.0) / v)))))); end
\begin{array}{l}
\\
1 + v \cdot \log \left(u + \left(1 - u\right) \cdot e^{\frac{-2}{v}}\right)
\end{array}
(FPCore (u v) :precision binary32 (+ 1.0 (* v (log (+ u (* (- 1.0 u) (exp (/ -2.0 v))))))))
float code(float u, float v) {
return 1.0f + (v * logf((u + ((1.0f - u) * expf((-2.0f / v))))));
}
real(4) function code(u, v)
real(4), intent (in) :: u
real(4), intent (in) :: v
code = 1.0e0 + (v * log((u + ((1.0e0 - u) * exp(((-2.0e0) / v))))))
end function
function code(u, v) return Float32(Float32(1.0) + Float32(v * log(Float32(u + Float32(Float32(Float32(1.0) - u) * exp(Float32(Float32(-2.0) / v))))))) end
function tmp = code(u, v) tmp = single(1.0) + (v * log((u + ((single(1.0) - u) * exp((single(-2.0) / v)))))); end
\begin{array}{l}
\\
1 + v \cdot \log \left(u + \left(1 - u\right) \cdot e^{\frac{-2}{v}}\right)
\end{array}
Initial program 99.6%
(FPCore (u v) :precision binary32 (+ 1.0 (* (log (fma (exp (/ -2.0 v)) (- 1.0 u) u)) v)))
float code(float u, float v) {
return 1.0f + (logf(fmaf(expf((-2.0f / v)), (1.0f - u), u)) * v);
}
function code(u, v) return Float32(Float32(1.0) + Float32(log(fma(exp(Float32(Float32(-2.0) / v)), Float32(Float32(1.0) - u), u)) * v)) end
\begin{array}{l}
\\
1 + \log \left(\mathsf{fma}\left(e^{\frac{-2}{v}}, 1 - u, u\right)\right) \cdot v
\end{array}
Initial program 99.6%
lift-*.f32N/A
*-commutativeN/A
lower-*.f3299.6
lift-+.f32N/A
+-commutativeN/A
lift-*.f32N/A
*-commutativeN/A
lower-fma.f3293.7
Applied rewrites93.9%
(FPCore (u v) :precision binary32 (+ 1.0 (* v (log (* u (- 1.0 (exp (/ -2.0 v))))))))
float code(float u, float v) {
return 1.0f + (v * logf((u * (1.0f - expf((-2.0f / v))))));
}
real(4) function code(u, v)
real(4), intent (in) :: u
real(4), intent (in) :: v
code = 1.0e0 + (v * log((u * (1.0e0 - exp(((-2.0e0) / v))))))
end function
function code(u, v) return Float32(Float32(1.0) + Float32(v * log(Float32(u * Float32(Float32(1.0) - exp(Float32(Float32(-2.0) / v))))))) end
function tmp = code(u, v) tmp = single(1.0) + (v * log((u * (single(1.0) - exp((single(-2.0) / v)))))); end
\begin{array}{l}
\\
1 + v \cdot \log \left(u \cdot \left(1 - e^{\frac{-2}{v}}\right)\right)
\end{array}
Initial program 99.6%
Taylor expanded in v around inf
associate-+r+N/A
fp-cancel-sign-sub-invN/A
metadata-evalN/A
associate-*r/N/A
associate-+l-N/A
fp-cancel-sub-sign-invN/A
metadata-evalN/A
unpow2N/A
associate-/r*N/A
associate-/l*N/A
div-addN/A
+-commutativeN/A
lower--.f32N/A
Applied rewrites43.3%
Taylor expanded in u around inf
lower-*.f32N/A
lower-+.f32N/A
mul-1-negN/A
lower-neg.f32N/A
lower-exp.f32N/A
lower-/.f3295.3
Applied rewrites95.3%
Final simplification95.3%
(FPCore (u v) :precision binary32 1.0)
float code(float u, float v) {
return 1.0f;
}
real(4) function code(u, v)
real(4), intent (in) :: u
real(4), intent (in) :: v
code = 1.0e0
end function
function code(u, v) return Float32(1.0) end
function tmp = code(u, v) tmp = single(1.0); end
\begin{array}{l}
\\
1
\end{array}
Initial program 99.6%
Taylor expanded in v around 0
Applied rewrites89.3%
(FPCore (u v) :precision binary32 -1.0)
float code(float u, float v) {
return -1.0f;
}
real(4) function code(u, v)
real(4), intent (in) :: u
real(4), intent (in) :: v
code = -1.0e0
end function
function code(u, v) return Float32(-1.0) end
function tmp = code(u, v) tmp = single(-1.0); end
\begin{array}{l}
\\
-1
\end{array}
Initial program 99.6%
Taylor expanded in u around 0
Applied rewrites4.8%
herbie shell --seed 2024337
(FPCore (u v)
:name "HairBSDF, sample_f, cosTheta"
:precision binary32
:pre (and (and (<= 1e-5 u) (<= u 1.0)) (and (<= 0.0 v) (<= v 109.746574)))
(+ 1.0 (* v (log (+ u (* (- 1.0 u) (exp (/ -2.0 v))))))))