This document provides an overview of multi-dimensional RNNs and some architectural issues and recent results related to them. It begins with an introduction to RNNs compared to feedforward neural networks, and solutions like LSTM and GRU to address the vanishing gradient problem. It then discusses several generalizations of the simple RNN architecture, including directionality with BRNN/BLSTM, dimensionality with MDRNN/MDLSTM, and directionality + dimensionality with MDMDRNN. It also covers hierarchical subsampling with HSRNN. The document concludes by summarizing some recent examples that apply these ideas, such as 2D LSTM for scene labeling, as well as new ideas like ReNet, PyraMiD-LSTM, and Grid LSTM.