C5W3A1 - 2.1 - Attention Mechanism - Some mistakes in the architecture explanation

hi, in the notebook of C5W3A1, I think there a couple of things wrong in section 2.1:

  • in the architecture, the plot in the left panel shows the outputs, but we have y<1> twice.
  • The explanation reads The post-attention LSTM at time 't' only takes the hidden state ๐‘ โŸจ๐‘กโŸฉ and cell state ๐‘โŸจ๐‘กโŸฉ as input. but it should actually be ๐‘ โŸจ๐‘ก-1โŸฉ and ๐‘โŸจ๐‘ก-1โŸฉ.

Hope that helps!

Thanks

Thanks for the careful proofreading! I will file a bug report with these items.

1 Like