(*.f64 (/.f64 (pow.f64 (sin.f64 k) 2) l) (/.f64 (*.f64 k (*.f64 k t)) (*.f64 (cos.f64 k) l))): 0 points increase in error, 0 points decrease in error
(*.f64 (/.f64 (pow.f64 (sin.f64 k) 2) l) (/.f64 (Rewrite<= associate-*l*_binary64 (*.f64 (*.f64 k k) t)) (*.f64 (cos.f64 k) l))): 34 points increase in error, 8 points decrease in error
(*.f64 (/.f64 (pow.f64 (sin.f64 k) 2) l) (/.f64 (*.f64 (Rewrite<= unpow2_binary64 (pow.f64 k 2)) t) (*.f64 (cos.f64 k) l))): 0 points increase in error, 0 points decrease in error
(*.f64 (/.f64 (pow.f64 (sin.f64 k) 2) l) (/.f64 (Rewrite<= *-commutative_binary64 (*.f64 t (pow.f64 k 2))) (*.f64 (cos.f64 k) l))): 0 points increase in error, 0 points decrease in error
(*.f64 (/.f64 (pow.f64 (sin.f64 k) 2) l) (/.f64 (*.f64 t (pow.f64 k 2)) (Rewrite=> *-commutative_binary64 (*.f64 l (cos.f64 k))))): 0 points increase in error, 0 points decrease in error
(Rewrite<= times-frac_binary64 (/.f64 (*.f64 (pow.f64 (sin.f64 k) 2) (*.f64 t (pow.f64 k 2))) (*.f64 l (*.f64 l (cos.f64 k))))): 50 points increase in error, 10 points decrease in error
(/.f64 (Rewrite<= associate-*l*_binary64 (*.f64 (*.f64 (pow.f64 (sin.f64 k) 2) t) (pow.f64 k 2))) (*.f64 l (*.f64 l (cos.f64 k)))): 8 points increase in error, 8 points decrease in error
(/.f64 (Rewrite<= *-commutative_binary64 (*.f64 (pow.f64 k 2) (*.f64 (pow.f64 (sin.f64 k) 2) t))) (*.f64 l (*.f64 l (cos.f64 k)))): 0 points increase in error, 0 points decrease in error
(/.f64 (*.f64 (pow.f64 k 2) (*.f64 (pow.f64 (sin.f64 k) 2) t)) (Rewrite<= associate-*l*_binary64 (*.f64 (*.f64 l l) (cos.f64 k)))): 8 points increase in error, 9 points decrease in error
(/.f64 (*.f64 (pow.f64 k 2) (*.f64 (pow.f64 (sin.f64 k) 2) t)) (*.f64 (Rewrite<= unpow2_binary64 (pow.f64 l 2)) (cos.f64 k))): 0 points increase in error, 0 points decrease in error
(/.f64 (*.f64 (pow.f64 k 2) (*.f64 (pow.f64 (sin.f64 k) 2) t)) (Rewrite<= *-commutative_binary64 (*.f64 (cos.f64 k) (pow.f64 l 2)))): 0 points increase in error, 0 points decrease in error
(*.f64 (/.f64 (pow.f64 (sin.f64 k) 2) l) (/.f64 (*.f64 k (*.f64 k t)) (*.f64 (cos.f64 k) l))): 0 points increase in error, 0 points decrease in error
(*.f64 (/.f64 (pow.f64 (sin.f64 k) 2) l) (/.f64 (Rewrite<= associate-*l*_binary64 (*.f64 (*.f64 k k) t)) (*.f64 (cos.f64 k) l))): 34 points increase in error, 8 points decrease in error
(*.f64 (/.f64 (pow.f64 (sin.f64 k) 2) l) (/.f64 (*.f64 (Rewrite<= unpow2_binary64 (pow.f64 k 2)) t) (*.f64 (cos.f64 k) l))): 0 points increase in error, 0 points decrease in error
(*.f64 (/.f64 (pow.f64 (sin.f64 k) 2) l) (/.f64 (Rewrite<= *-commutative_binary64 (*.f64 t (pow.f64 k 2))) (*.f64 (cos.f64 k) l))): 0 points increase in error, 0 points decrease in error
(*.f64 (/.f64 (pow.f64 (sin.f64 k) 2) l) (/.f64 (*.f64 t (pow.f64 k 2)) (Rewrite=> *-commutative_binary64 (*.f64 l (cos.f64 k))))): 0 points increase in error, 0 points decrease in error
(Rewrite<= times-frac_binary64 (/.f64 (*.f64 (pow.f64 (sin.f64 k) 2) (*.f64 t (pow.f64 k 2))) (*.f64 l (*.f64 l (cos.f64 k))))): 50 points increase in error, 10 points decrease in error
(/.f64 (Rewrite<= associate-*l*_binary64 (*.f64 (*.f64 (pow.f64 (sin.f64 k) 2) t) (pow.f64 k 2))) (*.f64 l (*.f64 l (cos.f64 k)))): 8 points increase in error, 8 points decrease in error
(/.f64 (Rewrite<= *-commutative_binary64 (*.f64 (pow.f64 k 2) (*.f64 (pow.f64 (sin.f64 k) 2) t))) (*.f64 l (*.f64 l (cos.f64 k)))): 0 points increase in error, 0 points decrease in error
(/.f64 (*.f64 (pow.f64 k 2) (*.f64 (pow.f64 (sin.f64 k) 2) t)) (Rewrite<= associate-*l*_binary64 (*.f64 (*.f64 l l) (cos.f64 k)))): 8 points increase in error, 9 points decrease in error
(/.f64 (*.f64 (pow.f64 k 2) (*.f64 (pow.f64 (sin.f64 k) 2) t)) (*.f64 (Rewrite<= unpow2_binary64 (pow.f64 l 2)) (cos.f64 k))): 0 points increase in error, 0 points decrease in error
(/.f64 (*.f64 (pow.f64 k 2) (*.f64 (pow.f64 (sin.f64 k) 2) t)) (Rewrite<= *-commutative_binary64 (*.f64 (cos.f64 k) (pow.f64 l 2)))): 0 points increase in error, 0 points decrease in error
(*.f64 (/.f64 (pow.f64 (sin.f64 k) 2) l) (/.f64 (*.f64 k (*.f64 k t)) (*.f64 (cos.f64 k) l))): 0 points increase in error, 0 points decrease in error
(*.f64 (/.f64 (pow.f64 (sin.f64 k) 2) l) (/.f64 (Rewrite<= associate-*l*_binary64 (*.f64 (*.f64 k k) t)) (*.f64 (cos.f64 k) l))): 34 points increase in error, 8 points decrease in error
(*.f64 (/.f64 (pow.f64 (sin.f64 k) 2) l) (/.f64 (*.f64 (Rewrite<= unpow2_binary64 (pow.f64 k 2)) t) (*.f64 (cos.f64 k) l))): 0 points increase in error, 0 points decrease in error
(*.f64 (/.f64 (pow.f64 (sin.f64 k) 2) l) (/.f64 (Rewrite<= *-commutative_binary64 (*.f64 t (pow.f64 k 2))) (*.f64 (cos.f64 k) l))): 0 points increase in error, 0 points decrease in error
(*.f64 (/.f64 (pow.f64 (sin.f64 k) 2) l) (/.f64 (*.f64 t (pow.f64 k 2)) (Rewrite=> *-commutative_binary64 (*.f64 l (cos.f64 k))))): 0 points increase in error, 0 points decrease in error
(Rewrite<= times-frac_binary64 (/.f64 (*.f64 (pow.f64 (sin.f64 k) 2) (*.f64 t (pow.f64 k 2))) (*.f64 l (*.f64 l (cos.f64 k))))): 50 points increase in error, 10 points decrease in error
(/.f64 (Rewrite<= associate-*l*_binary64 (*.f64 (*.f64 (pow.f64 (sin.f64 k) 2) t) (pow.f64 k 2))) (*.f64 l (*.f64 l (cos.f64 k)))): 8 points increase in error, 8 points decrease in error
(/.f64 (Rewrite<= *-commutative_binary64 (*.f64 (pow.f64 k 2) (*.f64 (pow.f64 (sin.f64 k) 2) t))) (*.f64 l (*.f64 l (cos.f64 k)))): 0 points increase in error, 0 points decrease in error
(/.f64 (*.f64 (pow.f64 k 2) (*.f64 (pow.f64 (sin.f64 k) 2) t)) (Rewrite<= associate-*l*_binary64 (*.f64 (*.f64 l l) (cos.f64 k)))): 8 points increase in error, 9 points decrease in error
(/.f64 (*.f64 (pow.f64 k 2) (*.f64 (pow.f64 (sin.f64 k) 2) t)) (*.f64 (Rewrite<= unpow2_binary64 (pow.f64 l 2)) (cos.f64 k))): 0 points increase in error, 0 points decrease in error
(/.f64 (*.f64 (pow.f64 k 2) (*.f64 (pow.f64 (sin.f64 k) 2) t)) (Rewrite<= *-commutative_binary64 (*.f64 (cos.f64 k) (pow.f64 l 2)))): 0 points increase in error, 0 points decrease in error