
(FPCore (e v) :precision binary64 (/ (* e (sin v)) (+ 1.0 (* e (cos v)))))
double code(double e, double v) {
return (e * sin(v)) / (1.0 + (e * cos(v)));
}
real(8) function code(e, v)
real(8), intent (in) :: e
real(8), intent (in) :: v
code = (e * sin(v)) / (1.0d0 + (e * cos(v)))
end function
public static double code(double e, double v) {
return (e * Math.sin(v)) / (1.0 + (e * Math.cos(v)));
}
def code(e, v): return (e * math.sin(v)) / (1.0 + (e * math.cos(v)))
function code(e, v) return Float64(Float64(e * sin(v)) / Float64(1.0 + Float64(e * cos(v)))) end
function tmp = code(e, v) tmp = (e * sin(v)) / (1.0 + (e * cos(v))); end
code[e_, v_] := N[(N[(e * N[Sin[v], $MachinePrecision]), $MachinePrecision] / N[(1.0 + N[(e * N[Cos[v], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\frac{e \cdot \sin v}{1 + e \cdot \cos v}
\end{array}
Sampling outcomes in binary64 precision:
Herbie found 14 alternatives:
| Alternative | Accuracy | Speedup |
|---|
(FPCore (e v) :precision binary64 (/ (* e (sin v)) (+ 1.0 (* e (cos v)))))
double code(double e, double v) {
return (e * sin(v)) / (1.0 + (e * cos(v)));
}
real(8) function code(e, v)
real(8), intent (in) :: e
real(8), intent (in) :: v
code = (e * sin(v)) / (1.0d0 + (e * cos(v)))
end function
public static double code(double e, double v) {
return (e * Math.sin(v)) / (1.0 + (e * Math.cos(v)));
}
def code(e, v): return (e * math.sin(v)) / (1.0 + (e * math.cos(v)))
function code(e, v) return Float64(Float64(e * sin(v)) / Float64(1.0 + Float64(e * cos(v)))) end
function tmp = code(e, v) tmp = (e * sin(v)) / (1.0 + (e * cos(v))); end
code[e_, v_] := N[(N[(e * N[Sin[v], $MachinePrecision]), $MachinePrecision] / N[(1.0 + N[(e * N[Cos[v], $MachinePrecision]), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\frac{e \cdot \sin v}{1 + e \cdot \cos v}
\end{array}
(FPCore (e v) :precision binary64 (* (sin v) (/ e (fma (cos v) e 1.0))))
double code(double e, double v) {
return sin(v) * (e / fma(cos(v), e, 1.0));
}
function code(e, v) return Float64(sin(v) * Float64(e / fma(cos(v), e, 1.0))) end
code[e_, v_] := N[(N[Sin[v], $MachinePrecision] * N[(e / N[(N[Cos[v], $MachinePrecision] * e + 1.0), $MachinePrecision]), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\sin v \cdot \frac{e}{\mathsf{fma}\left(\cos v, e, 1\right)}
\end{array}
Initial program 99.8%
lift-/.f64N/A
lift-*.f64N/A
*-commutativeN/A
associate-/l*N/A
*-commutativeN/A
lower-*.f64N/A
lower-/.f6499.8
lift-+.f64N/A
+-commutativeN/A
lift-*.f64N/A
*-commutativeN/A
lower-fma.f6499.8
Applied rewrites99.8%
Final simplification99.8%
(FPCore (e v) :precision binary64 (* (fma (* (- e) (cos v)) e e) (sin v)))
double code(double e, double v) {
return fma((-e * cos(v)), e, e) * sin(v);
}
function code(e, v) return Float64(fma(Float64(Float64(-e) * cos(v)), e, e) * sin(v)) end
code[e_, v_] := N[(N[(N[((-e) * N[Cos[v], $MachinePrecision]), $MachinePrecision] * e + e), $MachinePrecision] * N[Sin[v], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\mathsf{fma}\left(\left(-e\right) \cdot \cos v, e, e\right) \cdot \sin v
\end{array}
Initial program 99.8%
lift-/.f64N/A
lift-*.f64N/A
*-commutativeN/A
associate-/l*N/A
*-commutativeN/A
lower-*.f64N/A
lower-/.f6499.8
lift-+.f64N/A
+-commutativeN/A
lift-*.f64N/A
*-commutativeN/A
lower-fma.f6499.8
Applied rewrites99.8%
Taylor expanded in e around 0
*-commutativeN/A
lower-*.f64N/A
+-commutativeN/A
associate-*r*N/A
*-commutativeN/A
lower-fma.f64N/A
lower-cos.f64N/A
mul-1-negN/A
lower-neg.f6499.5
Applied rewrites99.5%
Applied rewrites99.5%
(FPCore (e v) :precision binary64 (* (* (fma (cos v) (- e) 1.0) e) (sin v)))
double code(double e, double v) {
return (fma(cos(v), -e, 1.0) * e) * sin(v);
}
function code(e, v) return Float64(Float64(fma(cos(v), Float64(-e), 1.0) * e) * sin(v)) end
code[e_, v_] := N[(N[(N[(N[Cos[v], $MachinePrecision] * (-e) + 1.0), $MachinePrecision] * e), $MachinePrecision] * N[Sin[v], $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\left(\mathsf{fma}\left(\cos v, -e, 1\right) \cdot e\right) \cdot \sin v
\end{array}
Initial program 99.8%
lift-/.f64N/A
lift-*.f64N/A
*-commutativeN/A
associate-/l*N/A
*-commutativeN/A
lower-*.f64N/A
lower-/.f6499.8
lift-+.f64N/A
+-commutativeN/A
lift-*.f64N/A
*-commutativeN/A
lower-fma.f6499.8
Applied rewrites99.8%
Taylor expanded in e around 0
*-commutativeN/A
lower-*.f64N/A
+-commutativeN/A
associate-*r*N/A
*-commutativeN/A
lower-fma.f64N/A
lower-cos.f64N/A
mul-1-negN/A
lower-neg.f6499.5
Applied rewrites99.5%
(FPCore (e v) :precision binary64 (* (* (sin v) e) (fma (- e) (cos v) 1.0)))
double code(double e, double v) {
return (sin(v) * e) * fma(-e, cos(v), 1.0);
}
function code(e, v) return Float64(Float64(sin(v) * e) * fma(Float64(-e), cos(v), 1.0)) end
code[e_, v_] := N[(N[(N[Sin[v], $MachinePrecision] * e), $MachinePrecision] * N[((-e) * N[Cos[v], $MachinePrecision] + 1.0), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\left(\sin v \cdot e\right) \cdot \mathsf{fma}\left(-e, \cos v, 1\right)
\end{array}
Initial program 99.8%
Taylor expanded in e around 0
distribute-lft-inN/A
*-commutativeN/A
mul-1-negN/A
associate-*r*N/A
distribute-lft-neg-inN/A
associate-*l*N/A
*-commutativeN/A
distribute-rgt1-inN/A
lower-*.f64N/A
distribute-lft-neg-inN/A
mul-1-negN/A
lower-fma.f64N/A
mul-1-negN/A
lower-neg.f64N/A
lower-cos.f64N/A
*-commutativeN/A
lower-*.f64N/A
lower-sin.f6499.4
Applied rewrites99.4%
Final simplification99.4%
(FPCore (e v) :precision binary64 (* (/ (sin v) (+ 1.0 e)) e))
double code(double e, double v) {
return (sin(v) / (1.0 + e)) * e;
}
real(8) function code(e, v)
real(8), intent (in) :: e
real(8), intent (in) :: v
code = (sin(v) / (1.0d0 + e)) * e
end function
public static double code(double e, double v) {
return (Math.sin(v) / (1.0 + e)) * e;
}
def code(e, v): return (math.sin(v) / (1.0 + e)) * e
function code(e, v) return Float64(Float64(sin(v) / Float64(1.0 + e)) * e) end
function tmp = code(e, v) tmp = (sin(v) / (1.0 + e)) * e; end
code[e_, v_] := N[(N[(N[Sin[v], $MachinePrecision] / N[(1.0 + e), $MachinePrecision]), $MachinePrecision] * e), $MachinePrecision]
\begin{array}{l}
\\
\frac{\sin v}{1 + e} \cdot e
\end{array}
Initial program 99.8%
lift-/.f64N/A
lift-*.f64N/A
*-commutativeN/A
associate-/l*N/A
*-commutativeN/A
lower-*.f64N/A
lower-/.f6499.8
lift-+.f64N/A
+-commutativeN/A
lift-*.f64N/A
*-commutativeN/A
lower-fma.f6499.8
Applied rewrites99.8%
lift-*.f64N/A
lift-/.f64N/A
associate-*l/N/A
lift-sin.f64N/A
lower-/.f64N/A
lift-sin.f64N/A
*-commutativeN/A
lift-*.f6499.8
Applied rewrites99.8%
Taylor expanded in v around 0
lower-+.f6498.9
Applied rewrites98.9%
lift-/.f64N/A
lift-*.f64N/A
*-commutativeN/A
associate-/l*N/A
lower-*.f64N/A
lower-/.f6498.9
Applied rewrites98.9%
Final simplification98.9%
(FPCore (e v) :precision binary64 (* (- 1.0 e) (* (sin v) e)))
double code(double e, double v) {
return (1.0 - e) * (sin(v) * e);
}
real(8) function code(e, v)
real(8), intent (in) :: e
real(8), intent (in) :: v
code = (1.0d0 - e) * (sin(v) * e)
end function
public static double code(double e, double v) {
return (1.0 - e) * (Math.sin(v) * e);
}
def code(e, v): return (1.0 - e) * (math.sin(v) * e)
function code(e, v) return Float64(Float64(1.0 - e) * Float64(sin(v) * e)) end
function tmp = code(e, v) tmp = (1.0 - e) * (sin(v) * e); end
code[e_, v_] := N[(N[(1.0 - e), $MachinePrecision] * N[(N[Sin[v], $MachinePrecision] * e), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\left(1 - e\right) \cdot \left(\sin v \cdot e\right)
\end{array}
Initial program 99.8%
Taylor expanded in e around 0
distribute-lft-inN/A
*-commutativeN/A
mul-1-negN/A
associate-*r*N/A
distribute-lft-neg-inN/A
associate-*l*N/A
*-commutativeN/A
distribute-rgt1-inN/A
lower-*.f64N/A
distribute-lft-neg-inN/A
mul-1-negN/A
lower-fma.f64N/A
mul-1-negN/A
lower-neg.f64N/A
lower-cos.f64N/A
*-commutativeN/A
lower-*.f64N/A
lower-sin.f6499.4
Applied rewrites99.4%
Taylor expanded in v around 0
Applied rewrites98.9%
(FPCore (e v) :precision binary64 (* (sin v) e))
double code(double e, double v) {
return sin(v) * e;
}
real(8) function code(e, v)
real(8), intent (in) :: e
real(8), intent (in) :: v
code = sin(v) * e
end function
public static double code(double e, double v) {
return Math.sin(v) * e;
}
def code(e, v): return math.sin(v) * e
function code(e, v) return Float64(sin(v) * e) end
function tmp = code(e, v) tmp = sin(v) * e; end
code[e_, v_] := N[(N[Sin[v], $MachinePrecision] * e), $MachinePrecision]
\begin{array}{l}
\\
\sin v \cdot e
\end{array}
Initial program 99.8%
Taylor expanded in e around 0
*-commutativeN/A
lower-*.f64N/A
lower-sin.f6498.6
Applied rewrites98.6%
(FPCore (e v)
:precision binary64
(let* ((t_0 (fma -0.5 e (fma 0.16666666666666666 e 0.16666666666666666))))
(/
e
(/
(fma
(fma
(fma
0.041666666666666664
e
(fma
t_0
0.16666666666666666
(fma -0.008333333333333333 e -0.008333333333333333)))
(* v v)
t_0)
(* v v)
(+ 1.0 e))
v))))
double code(double e, double v) {
double t_0 = fma(-0.5, e, fma(0.16666666666666666, e, 0.16666666666666666));
return e / (fma(fma(fma(0.041666666666666664, e, fma(t_0, 0.16666666666666666, fma(-0.008333333333333333, e, -0.008333333333333333))), (v * v), t_0), (v * v), (1.0 + e)) / v);
}
function code(e, v) t_0 = fma(-0.5, e, fma(0.16666666666666666, e, 0.16666666666666666)) return Float64(e / Float64(fma(fma(fma(0.041666666666666664, e, fma(t_0, 0.16666666666666666, fma(-0.008333333333333333, e, -0.008333333333333333))), Float64(v * v), t_0), Float64(v * v), Float64(1.0 + e)) / v)) end
code[e_, v_] := Block[{t$95$0 = N[(-0.5 * e + N[(0.16666666666666666 * e + 0.16666666666666666), $MachinePrecision]), $MachinePrecision]}, N[(e / N[(N[(N[(N[(0.041666666666666664 * e + N[(t$95$0 * 0.16666666666666666 + N[(-0.008333333333333333 * e + -0.008333333333333333), $MachinePrecision]), $MachinePrecision]), $MachinePrecision] * N[(v * v), $MachinePrecision] + t$95$0), $MachinePrecision] * N[(v * v), $MachinePrecision] + N[(1.0 + e), $MachinePrecision]), $MachinePrecision] / v), $MachinePrecision]), $MachinePrecision]]
\begin{array}{l}
\\
\begin{array}{l}
t_0 := \mathsf{fma}\left(-0.5, e, \mathsf{fma}\left(0.16666666666666666, e, 0.16666666666666666\right)\right)\\
\frac{e}{\frac{\mathsf{fma}\left(\mathsf{fma}\left(\mathsf{fma}\left(0.041666666666666664, e, \mathsf{fma}\left(t\_0, 0.16666666666666666, \mathsf{fma}\left(-0.008333333333333333, e, -0.008333333333333333\right)\right)\right), v \cdot v, t\_0\right), v \cdot v, 1 + e\right)}{v}}
\end{array}
\end{array}
Initial program 99.8%
lift-/.f64N/A
lift-*.f64N/A
associate-/l*N/A
clear-numN/A
un-div-invN/A
lower-/.f64N/A
lower-/.f6499.7
lift-+.f64N/A
+-commutativeN/A
lift-*.f64N/A
*-commutativeN/A
lower-fma.f6499.7
Applied rewrites99.7%
Taylor expanded in v around 0
Applied rewrites55.4%
(FPCore (e v)
:precision binary64
(/
e
(/
(fma
(fma -0.5 e (fma 0.16666666666666666 e 0.16666666666666666))
(* v v)
(+ 1.0 e))
v)))
double code(double e, double v) {
return e / (fma(fma(-0.5, e, fma(0.16666666666666666, e, 0.16666666666666666)), (v * v), (1.0 + e)) / v);
}
function code(e, v) return Float64(e / Float64(fma(fma(-0.5, e, fma(0.16666666666666666, e, 0.16666666666666666)), Float64(v * v), Float64(1.0 + e)) / v)) end
code[e_, v_] := N[(e / N[(N[(N[(-0.5 * e + N[(0.16666666666666666 * e + 0.16666666666666666), $MachinePrecision]), $MachinePrecision] * N[(v * v), $MachinePrecision] + N[(1.0 + e), $MachinePrecision]), $MachinePrecision] / v), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\frac{e}{\frac{\mathsf{fma}\left(\mathsf{fma}\left(-0.5, e, \mathsf{fma}\left(0.16666666666666666, e, 0.16666666666666666\right)\right), v \cdot v, 1 + e\right)}{v}}
\end{array}
Initial program 99.8%
lift-/.f64N/A
lift-*.f64N/A
associate-/l*N/A
clear-numN/A
un-div-invN/A
lower-/.f64N/A
lower-/.f6499.7
lift-+.f64N/A
+-commutativeN/A
lift-*.f64N/A
*-commutativeN/A
lower-fma.f6499.7
Applied rewrites99.7%
Taylor expanded in v around 0
lower-/.f64N/A
associate-+r+N/A
+-commutativeN/A
*-commutativeN/A
lower-fma.f64N/A
sub-negN/A
lower-fma.f64N/A
distribute-lft-neg-inN/A
metadata-evalN/A
+-commutativeN/A
distribute-lft-inN/A
metadata-evalN/A
lower-fma.f64N/A
unpow2N/A
lower-*.f64N/A
lower-+.f6455.4
Applied rewrites55.4%
(FPCore (e v) :precision binary64 (* (/ e (+ 1.0 e)) v))
double code(double e, double v) {
return (e / (1.0 + e)) * v;
}
real(8) function code(e, v)
real(8), intent (in) :: e
real(8), intent (in) :: v
code = (e / (1.0d0 + e)) * v
end function
public static double code(double e, double v) {
return (e / (1.0 + e)) * v;
}
def code(e, v): return (e / (1.0 + e)) * v
function code(e, v) return Float64(Float64(e / Float64(1.0 + e)) * v) end
function tmp = code(e, v) tmp = (e / (1.0 + e)) * v; end
code[e_, v_] := N[(N[(e / N[(1.0 + e), $MachinePrecision]), $MachinePrecision] * v), $MachinePrecision]
\begin{array}{l}
\\
\frac{e}{1 + e} \cdot v
\end{array}
Initial program 99.8%
Taylor expanded in v around 0
associate-*l/N/A
lower-*.f64N/A
lower-/.f64N/A
+-commutativeN/A
lower-+.f6454.2
Applied rewrites54.2%
Final simplification54.2%
(FPCore (e v) :precision binary64 (* (fma (- (* v e) v) e v) e))
double code(double e, double v) {
return fma(((v * e) - v), e, v) * e;
}
function code(e, v) return Float64(fma(Float64(Float64(v * e) - v), e, v) * e) end
code[e_, v_] := N[(N[(N[(N[(v * e), $MachinePrecision] - v), $MachinePrecision] * e + v), $MachinePrecision] * e), $MachinePrecision]
\begin{array}{l}
\\
\mathsf{fma}\left(v \cdot e - v, e, v\right) \cdot e
\end{array}
Initial program 99.8%
Taylor expanded in v around 0
associate-*l/N/A
lower-*.f64N/A
lower-/.f64N/A
+-commutativeN/A
lower-+.f6454.2
Applied rewrites54.2%
Taylor expanded in e around 0
Applied rewrites54.2%
(FPCore (e v) :precision binary64 (fma v e (* (* (- v) e) e)))
double code(double e, double v) {
return fma(v, e, ((-v * e) * e));
}
function code(e, v) return fma(v, e, Float64(Float64(Float64(-v) * e) * e)) end
code[e_, v_] := N[(v * e + N[(N[((-v) * e), $MachinePrecision] * e), $MachinePrecision]), $MachinePrecision]
\begin{array}{l}
\\
\mathsf{fma}\left(v, e, \left(\left(-v\right) \cdot e\right) \cdot e\right)
\end{array}
Initial program 99.8%
Taylor expanded in v around 0
associate-*l/N/A
lower-*.f64N/A
lower-/.f64N/A
+-commutativeN/A
lower-+.f6454.2
Applied rewrites54.2%
Applied rewrites54.1%
Taylor expanded in e around 0
Applied rewrites54.2%
Applied rewrites54.2%
(FPCore (e v) :precision binary64 (* (- v (* v e)) e))
double code(double e, double v) {
return (v - (v * e)) * e;
}
real(8) function code(e, v)
real(8), intent (in) :: e
real(8), intent (in) :: v
code = (v - (v * e)) * e
end function
public static double code(double e, double v) {
return (v - (v * e)) * e;
}
def code(e, v): return (v - (v * e)) * e
function code(e, v) return Float64(Float64(v - Float64(v * e)) * e) end
function tmp = code(e, v) tmp = (v - (v * e)) * e; end
code[e_, v_] := N[(N[(v - N[(v * e), $MachinePrecision]), $MachinePrecision] * e), $MachinePrecision]
\begin{array}{l}
\\
\left(v - v \cdot e\right) \cdot e
\end{array}
Initial program 99.8%
Taylor expanded in v around 0
associate-*l/N/A
lower-*.f64N/A
lower-/.f64N/A
+-commutativeN/A
lower-+.f6454.2
Applied rewrites54.2%
Applied rewrites54.1%
Taylor expanded in e around 0
Applied rewrites54.2%
(FPCore (e v) :precision binary64 (* v e))
double code(double e, double v) {
return v * e;
}
real(8) function code(e, v)
real(8), intent (in) :: e
real(8), intent (in) :: v
code = v * e
end function
public static double code(double e, double v) {
return v * e;
}
def code(e, v): return v * e
function code(e, v) return Float64(v * e) end
function tmp = code(e, v) tmp = v * e; end
code[e_, v_] := N[(v * e), $MachinePrecision]
\begin{array}{l}
\\
v \cdot e
\end{array}
Initial program 99.8%
Taylor expanded in v around 0
associate-*l/N/A
lower-*.f64N/A
lower-/.f64N/A
+-commutativeN/A
lower-+.f6454.2
Applied rewrites54.2%
Taylor expanded in e around 0
Applied rewrites53.9%
herbie shell --seed 2024235
(FPCore (e v)
:name "Trigonometry A"
:precision binary64
:pre (and (<= 0.0 e) (<= e 1.0))
(/ (* e (sin v)) (+ 1.0 (* e (cos v)))))