Need Help in Understanding RNN input sequences vs number of cells used in the layer

An LSTM layer consists of a bunch of stacked LSTM cells.

You feed the 1st word to the 1st lstm cell. The 1st cell feeds the outputs to the next LSTM cell connected to it. This process continues till the last LSTM cell from where outputs is taken for feeding to the next layer or interpretation. In short, you can consider this entire arrangement as a single LSTM cell. Now that you get this perspective, the rest is the same as using an LSTM cell for multiple timesteps.

1 Like