Simplified0.5
\[\leadsto \color{blue}{\mathsf{fma}\left(\varepsilon, 1 + \frac{{\sin x}^{2}}{{\cos x}^{2}}, \left(\varepsilon \cdot \varepsilon\right) \cdot \left(\left(\frac{\sin x}{\cos x} + {\left(\frac{\sin x}{\cos x}\right)}^{3}\right) + \left(0.3333333333333333 - \left(\frac{{\sin x}^{2}}{{\cos x}^{2}} \cdot -1.3333333333333333 - \frac{{\sin x}^{4}}{{\cos x}^{4}}\right)\right) \cdot \varepsilon\right)\right)}
\]
Proof
(fma.f64 eps (+.f64 1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2))) (*.f64 (*.f64 eps eps) (+.f64 (+.f64 (/.f64 (sin.f64 x) (cos.f64 x)) (pow.f64 (/.f64 (sin.f64 x) (cos.f64 x)) 3)) (*.f64 (-.f64 1/3 (-.f64 (*.f64 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)) -4/3) (/.f64 (pow.f64 (sin.f64 x) 4) (pow.f64 (cos.f64 x) 4)))) eps)))): 0 points increase in error, 0 points decrease in error
(fma.f64 eps (+.f64 1 (Rewrite<= *-lft-identity_binary64 (*.f64 1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2))))) (*.f64 (*.f64 eps eps) (+.f64 (+.f64 (/.f64 (sin.f64 x) (cos.f64 x)) (pow.f64 (/.f64 (sin.f64 x) (cos.f64 x)) 3)) (*.f64 (-.f64 1/3 (-.f64 (*.f64 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)) -4/3) (/.f64 (pow.f64 (sin.f64 x) 4) (pow.f64 (cos.f64 x) 4)))) eps)))): 0 points increase in error, 0 points decrease in error
(fma.f64 eps (+.f64 1 (*.f64 (Rewrite<= metadata-eval (neg.f64 -1)) (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))) (*.f64 (*.f64 eps eps) (+.f64 (+.f64 (/.f64 (sin.f64 x) (cos.f64 x)) (pow.f64 (/.f64 (sin.f64 x) (cos.f64 x)) 3)) (*.f64 (-.f64 1/3 (-.f64 (*.f64 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)) -4/3) (/.f64 (pow.f64 (sin.f64 x) 4) (pow.f64 (cos.f64 x) 4)))) eps)))): 0 points increase in error, 0 points decrease in error
(fma.f64 eps (Rewrite<= cancel-sign-sub-inv_binary64 (-.f64 1 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2))))) (*.f64 (*.f64 eps eps) (+.f64 (+.f64 (/.f64 (sin.f64 x) (cos.f64 x)) (pow.f64 (/.f64 (sin.f64 x) (cos.f64 x)) 3)) (*.f64 (-.f64 1/3 (-.f64 (*.f64 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)) -4/3) (/.f64 (pow.f64 (sin.f64 x) 4) (pow.f64 (cos.f64 x) 4)))) eps)))): 0 points increase in error, 0 points decrease in error
(fma.f64 eps (-.f64 1 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))) (*.f64 (Rewrite<= unpow2_binary64 (pow.f64 eps 2)) (+.f64 (+.f64 (/.f64 (sin.f64 x) (cos.f64 x)) (pow.f64 (/.f64 (sin.f64 x) (cos.f64 x)) 3)) (*.f64 (-.f64 1/3 (-.f64 (*.f64 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)) -4/3) (/.f64 (pow.f64 (sin.f64 x) 4) (pow.f64 (cos.f64 x) 4)))) eps)))): 0 points increase in error, 0 points decrease in error
(fma.f64 eps (-.f64 1 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))) (*.f64 (pow.f64 eps 2) (+.f64 (+.f64 (/.f64 (sin.f64 x) (cos.f64 x)) (Rewrite<= cube-unmult_binary64 (*.f64 (/.f64 (sin.f64 x) (cos.f64 x)) (*.f64 (/.f64 (sin.f64 x) (cos.f64 x)) (/.f64 (sin.f64 x) (cos.f64 x)))))) (*.f64 (-.f64 1/3 (-.f64 (*.f64 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)) -4/3) (/.f64 (pow.f64 (sin.f64 x) 4) (pow.f64 (cos.f64 x) 4)))) eps)))): 0 points increase in error, 0 points decrease in error
(fma.f64 eps (-.f64 1 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))) (*.f64 (pow.f64 eps 2) (+.f64 (+.f64 (/.f64 (sin.f64 x) (cos.f64 x)) (*.f64 (/.f64 (sin.f64 x) (cos.f64 x)) (Rewrite<= times-frac_binary64 (/.f64 (*.f64 (sin.f64 x) (sin.f64 x)) (*.f64 (cos.f64 x) (cos.f64 x)))))) (*.f64 (-.f64 1/3 (-.f64 (*.f64 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)) -4/3) (/.f64 (pow.f64 (sin.f64 x) 4) (pow.f64 (cos.f64 x) 4)))) eps)))): 0 points increase in error, 0 points decrease in error
(fma.f64 eps (-.f64 1 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))) (*.f64 (pow.f64 eps 2) (+.f64 (+.f64 (/.f64 (sin.f64 x) (cos.f64 x)) (*.f64 (/.f64 (sin.f64 x) (cos.f64 x)) (/.f64 (Rewrite<= unpow2_binary64 (pow.f64 (sin.f64 x) 2)) (*.f64 (cos.f64 x) (cos.f64 x))))) (*.f64 (-.f64 1/3 (-.f64 (*.f64 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)) -4/3) (/.f64 (pow.f64 (sin.f64 x) 4) (pow.f64 (cos.f64 x) 4)))) eps)))): 0 points increase in error, 0 points decrease in error
(fma.f64 eps (-.f64 1 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))) (*.f64 (pow.f64 eps 2) (+.f64 (+.f64 (/.f64 (sin.f64 x) (cos.f64 x)) (*.f64 (/.f64 (sin.f64 x) (cos.f64 x)) (/.f64 (pow.f64 (sin.f64 x) 2) (Rewrite<= unpow2_binary64 (pow.f64 (cos.f64 x) 2))))) (*.f64 (-.f64 1/3 (-.f64 (*.f64 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)) -4/3) (/.f64 (pow.f64 (sin.f64 x) 4) (pow.f64 (cos.f64 x) 4)))) eps)))): 0 points increase in error, 0 points decrease in error
(fma.f64 eps (-.f64 1 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))) (*.f64 (pow.f64 eps 2) (+.f64 (+.f64 (/.f64 (sin.f64 x) (cos.f64 x)) (Rewrite<= times-frac_binary64 (/.f64 (*.f64 (sin.f64 x) (pow.f64 (sin.f64 x) 2)) (*.f64 (cos.f64 x) (pow.f64 (cos.f64 x) 2))))) (*.f64 (-.f64 1/3 (-.f64 (*.f64 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)) -4/3) (/.f64 (pow.f64 (sin.f64 x) 4) (pow.f64 (cos.f64 x) 4)))) eps)))): 0 points increase in error, 0 points decrease in error
(fma.f64 eps (-.f64 1 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))) (*.f64 (pow.f64 eps 2) (+.f64 (+.f64 (/.f64 (sin.f64 x) (cos.f64 x)) (/.f64 (*.f64 (sin.f64 x) (Rewrite=> unpow2_binary64 (*.f64 (sin.f64 x) (sin.f64 x)))) (*.f64 (cos.f64 x) (pow.f64 (cos.f64 x) 2)))) (*.f64 (-.f64 1/3 (-.f64 (*.f64 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)) -4/3) (/.f64 (pow.f64 (sin.f64 x) 4) (pow.f64 (cos.f64 x) 4)))) eps)))): 0 points increase in error, 0 points decrease in error
(fma.f64 eps (-.f64 1 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))) (*.f64 (pow.f64 eps 2) (+.f64 (+.f64 (/.f64 (sin.f64 x) (cos.f64 x)) (/.f64 (Rewrite<= cube-mult_binary64 (pow.f64 (sin.f64 x) 3)) (*.f64 (cos.f64 x) (pow.f64 (cos.f64 x) 2)))) (*.f64 (-.f64 1/3 (-.f64 (*.f64 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)) -4/3) (/.f64 (pow.f64 (sin.f64 x) 4) (pow.f64 (cos.f64 x) 4)))) eps)))): 0 points increase in error, 0 points decrease in error
(fma.f64 eps (-.f64 1 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))) (*.f64 (pow.f64 eps 2) (+.f64 (+.f64 (/.f64 (sin.f64 x) (cos.f64 x)) (/.f64 (pow.f64 (sin.f64 x) 3) (*.f64 (cos.f64 x) (Rewrite=> unpow2_binary64 (*.f64 (cos.f64 x) (cos.f64 x)))))) (*.f64 (-.f64 1/3 (-.f64 (*.f64 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)) -4/3) (/.f64 (pow.f64 (sin.f64 x) 4) (pow.f64 (cos.f64 x) 4)))) eps)))): 0 points increase in error, 0 points decrease in error
(fma.f64 eps (-.f64 1 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))) (*.f64 (pow.f64 eps 2) (+.f64 (+.f64 (/.f64 (sin.f64 x) (cos.f64 x)) (/.f64 (pow.f64 (sin.f64 x) 3) (Rewrite<= cube-mult_binary64 (pow.f64 (cos.f64 x) 3)))) (*.f64 (-.f64 1/3 (-.f64 (*.f64 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)) -4/3) (/.f64 (pow.f64 (sin.f64 x) 4) (pow.f64 (cos.f64 x) 4)))) eps)))): 0 points increase in error, 0 points decrease in error
(fma.f64 eps (-.f64 1 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))) (*.f64 (pow.f64 eps 2) (+.f64 (Rewrite<= +-commutative_binary64 (+.f64 (/.f64 (pow.f64 (sin.f64 x) 3) (pow.f64 (cos.f64 x) 3)) (/.f64 (sin.f64 x) (cos.f64 x)))) (*.f64 (-.f64 1/3 (-.f64 (*.f64 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)) -4/3) (/.f64 (pow.f64 (sin.f64 x) 4) (pow.f64 (cos.f64 x) 4)))) eps)))): 0 points increase in error, 0 points decrease in error
(fma.f64 eps (-.f64 1 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))) (*.f64 (pow.f64 eps 2) (+.f64 (+.f64 (/.f64 (pow.f64 (sin.f64 x) 3) (pow.f64 (cos.f64 x) 3)) (/.f64 (sin.f64 x) (cos.f64 x))) (*.f64 (-.f64 1/3 (-.f64 (*.f64 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)) (Rewrite<= metadata-eval (+.f64 -1 -1/3))) (/.f64 (pow.f64 (sin.f64 x) 4) (pow.f64 (cos.f64 x) 4)))) eps)))): 0 points increase in error, 0 points decrease in error
(fma.f64 eps (-.f64 1 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))) (*.f64 (pow.f64 eps 2) (+.f64 (+.f64 (/.f64 (pow.f64 (sin.f64 x) 3) (pow.f64 (cos.f64 x) 3)) (/.f64 (sin.f64 x) (cos.f64 x))) (*.f64 (-.f64 1/3 (-.f64 (Rewrite<= distribute-rgt-out_binary64 (+.f64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2))) (*.f64 -1/3 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2))))) (/.f64 (pow.f64 (sin.f64 x) 4) (pow.f64 (cos.f64 x) 4)))) eps)))): 0 points increase in error, 0 points decrease in error
(fma.f64 eps (-.f64 1 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))) (*.f64 (pow.f64 eps 2) (+.f64 (+.f64 (/.f64 (pow.f64 (sin.f64 x) 3) (pow.f64 (cos.f64 x) 3)) (/.f64 (sin.f64 x) (cos.f64 x))) (*.f64 (-.f64 1/3 (Rewrite<= unsub-neg_binary64 (+.f64 (+.f64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2))) (*.f64 -1/3 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))) (neg.f64 (/.f64 (pow.f64 (sin.f64 x) 4) (pow.f64 (cos.f64 x) 4)))))) eps)))): 0 points increase in error, 0 points decrease in error
(fma.f64 eps (-.f64 1 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))) (*.f64 (pow.f64 eps 2) (+.f64 (+.f64 (/.f64 (pow.f64 (sin.f64 x) 3) (pow.f64 (cos.f64 x) 3)) (/.f64 (sin.f64 x) (cos.f64 x))) (*.f64 (-.f64 1/3 (+.f64 (+.f64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2))) (*.f64 -1/3 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))) (Rewrite<= mul-1-neg_binary64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 4) (pow.f64 (cos.f64 x) 4)))))) eps)))): 0 points increase in error, 0 points decrease in error
(fma.f64 eps (-.f64 1 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))) (*.f64 (pow.f64 eps 2) (+.f64 (+.f64 (/.f64 (pow.f64 (sin.f64 x) 3) (pow.f64 (cos.f64 x) 3)) (/.f64 (sin.f64 x) (cos.f64 x))) (*.f64 (-.f64 1/3 (Rewrite<= associate-+r+_binary64 (+.f64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2))) (+.f64 (*.f64 -1/3 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2))) (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 4) (pow.f64 (cos.f64 x) 4))))))) eps)))): 2 points increase in error, 0 points decrease in error
(fma.f64 eps (-.f64 1 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))) (*.f64 (pow.f64 eps 2) (+.f64 (+.f64 (/.f64 (pow.f64 (sin.f64 x) 3) (pow.f64 (cos.f64 x) 3)) (/.f64 (sin.f64 x) (cos.f64 x))) (*.f64 (-.f64 1/3 (+.f64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2))) (Rewrite<= +-commutative_binary64 (+.f64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 4) (pow.f64 (cos.f64 x) 4))) (*.f64 -1/3 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2))))))) eps)))): 0 points increase in error, 0 points decrease in error
(fma.f64 eps (-.f64 1 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))) (Rewrite<= distribute-rgt-out_binary64 (+.f64 (*.f64 (+.f64 (/.f64 (pow.f64 (sin.f64 x) 3) (pow.f64 (cos.f64 x) 3)) (/.f64 (sin.f64 x) (cos.f64 x))) (pow.f64 eps 2)) (*.f64 (*.f64 (-.f64 1/3 (+.f64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2))) (+.f64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 4) (pow.f64 (cos.f64 x) 4))) (*.f64 -1/3 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))))) eps) (pow.f64 eps 2))))): 0 points increase in error, 0 points decrease in error
(fma.f64 eps (-.f64 1 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))) (+.f64 (*.f64 (Rewrite<= remove-double-neg_binary64 (neg.f64 (neg.f64 (+.f64 (/.f64 (pow.f64 (sin.f64 x) 3) (pow.f64 (cos.f64 x) 3)) (/.f64 (sin.f64 x) (cos.f64 x)))))) (pow.f64 eps 2)) (*.f64 (*.f64 (-.f64 1/3 (+.f64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2))) (+.f64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 4) (pow.f64 (cos.f64 x) 4))) (*.f64 -1/3 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))))) eps) (pow.f64 eps 2)))): 0 points increase in error, 0 points decrease in error
(fma.f64 eps (-.f64 1 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))) (+.f64 (*.f64 (neg.f64 (Rewrite<= mul-1-neg_binary64 (*.f64 -1 (+.f64 (/.f64 (pow.f64 (sin.f64 x) 3) (pow.f64 (cos.f64 x) 3)) (/.f64 (sin.f64 x) (cos.f64 x)))))) (pow.f64 eps 2)) (*.f64 (*.f64 (-.f64 1/3 (+.f64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2))) (+.f64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 4) (pow.f64 (cos.f64 x) 4))) (*.f64 -1/3 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))))) eps) (pow.f64 eps 2)))): 0 points increase in error, 0 points decrease in error
(fma.f64 eps (-.f64 1 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))) (+.f64 (*.f64 (neg.f64 (Rewrite<= distribute-lft-out_binary64 (+.f64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 3) (pow.f64 (cos.f64 x) 3))) (*.f64 -1 (/.f64 (sin.f64 x) (cos.f64 x)))))) (pow.f64 eps 2)) (*.f64 (*.f64 (-.f64 1/3 (+.f64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2))) (+.f64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 4) (pow.f64 (cos.f64 x) 4))) (*.f64 -1/3 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))))) eps) (pow.f64 eps 2)))): 0 points increase in error, 0 points decrease in error
(fma.f64 eps (-.f64 1 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))) (+.f64 (Rewrite<= distribute-lft-neg-in_binary64 (neg.f64 (*.f64 (+.f64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 3) (pow.f64 (cos.f64 x) 3))) (*.f64 -1 (/.f64 (sin.f64 x) (cos.f64 x)))) (pow.f64 eps 2)))) (*.f64 (*.f64 (-.f64 1/3 (+.f64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2))) (+.f64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 4) (pow.f64 (cos.f64 x) 4))) (*.f64 -1/3 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))))) eps) (pow.f64 eps 2)))): 0 points increase in error, 0 points decrease in error
(fma.f64 eps (-.f64 1 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))) (+.f64 (neg.f64 (Rewrite<= *-commutative_binary64 (*.f64 (pow.f64 eps 2) (+.f64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 3) (pow.f64 (cos.f64 x) 3))) (*.f64 -1 (/.f64 (sin.f64 x) (cos.f64 x))))))) (*.f64 (*.f64 (-.f64 1/3 (+.f64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2))) (+.f64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 4) (pow.f64 (cos.f64 x) 4))) (*.f64 -1/3 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))))) eps) (pow.f64 eps 2)))): 0 points increase in error, 0 points decrease in error
(fma.f64 eps (-.f64 1 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))) (+.f64 (Rewrite<= mul-1-neg_binary64 (*.f64 -1 (*.f64 (pow.f64 eps 2) (+.f64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 3) (pow.f64 (cos.f64 x) 3))) (*.f64 -1 (/.f64 (sin.f64 x) (cos.f64 x))))))) (*.f64 (*.f64 (-.f64 1/3 (+.f64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2))) (+.f64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 4) (pow.f64 (cos.f64 x) 4))) (*.f64 -1/3 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))))) eps) (pow.f64 eps 2)))): 0 points increase in error, 0 points decrease in error
(fma.f64 eps (-.f64 1 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))) (+.f64 (*.f64 -1 (*.f64 (pow.f64 eps 2) (+.f64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 3) (pow.f64 (cos.f64 x) 3))) (*.f64 -1 (/.f64 (sin.f64 x) (cos.f64 x)))))) (Rewrite<= associate-*r*_binary64 (*.f64 (-.f64 1/3 (+.f64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2))) (+.f64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 4) (pow.f64 (cos.f64 x) 4))) (*.f64 -1/3 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))))) (*.f64 eps (pow.f64 eps 2)))))): 13 points increase in error, 10 points decrease in error
(fma.f64 eps (-.f64 1 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))) (+.f64 (*.f64 -1 (*.f64 (pow.f64 eps 2) (+.f64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 3) (pow.f64 (cos.f64 x) 3))) (*.f64 -1 (/.f64 (sin.f64 x) (cos.f64 x)))))) (*.f64 (-.f64 1/3 (+.f64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2))) (+.f64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 4) (pow.f64 (cos.f64 x) 4))) (*.f64 -1/3 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))))) (*.f64 eps (Rewrite=> unpow2_binary64 (*.f64 eps eps)))))): 0 points increase in error, 0 points decrease in error
(fma.f64 eps (-.f64 1 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))) (+.f64 (*.f64 -1 (*.f64 (pow.f64 eps 2) (+.f64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 3) (pow.f64 (cos.f64 x) 3))) (*.f64 -1 (/.f64 (sin.f64 x) (cos.f64 x)))))) (*.f64 (-.f64 1/3 (+.f64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2))) (+.f64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 4) (pow.f64 (cos.f64 x) 4))) (*.f64 -1/3 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))))) (Rewrite<= cube-mult_binary64 (pow.f64 eps 3))))): 2 points increase in error, 6 points decrease in error
(fma.f64 eps (-.f64 1 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))) (+.f64 (*.f64 -1 (*.f64 (pow.f64 eps 2) (+.f64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 3) (pow.f64 (cos.f64 x) 3))) (*.f64 -1 (/.f64 (sin.f64 x) (cos.f64 x)))))) (Rewrite<= *-commutative_binary64 (*.f64 (pow.f64 eps 3) (-.f64 1/3 (+.f64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2))) (+.f64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 4) (pow.f64 (cos.f64 x) 4))) (*.f64 -1/3 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))))))))): 0 points increase in error, 0 points decrease in error
(Rewrite<= fma-def_binary64 (+.f64 (*.f64 eps (-.f64 1 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2))))) (+.f64 (*.f64 -1 (*.f64 (pow.f64 eps 2) (+.f64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 3) (pow.f64 (cos.f64 x) 3))) (*.f64 -1 (/.f64 (sin.f64 x) (cos.f64 x)))))) (*.f64 (pow.f64 eps 3) (-.f64 1/3 (+.f64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2))) (+.f64 (*.f64 -1 (/.f64 (pow.f64 (sin.f64 x) 4) (pow.f64 (cos.f64 x) 4))) (*.f64 -1/3 (/.f64 (pow.f64 (sin.f64 x) 2) (pow.f64 (cos.f64 x) 2)))))))))): 0 points increase in error, 1 points decrease in error