\section{Conclusion}
We implemented three approaches to solving the Tetris problem: Cross Entropy, REINFORCE and Natural Policy Gradient. For all of our methods, we achieved about 30,000 lines cleared on average. While this is pretty good performance, we expected to be clearing millions of lines on average. Given more time to think and compute, we would try incorporating some new features and increasing the number of times the simulation is run in Cross Entropy in order to reduce the variance in our estimate of the reward function.