This concerns a possible typo?
“”""
Forward pass for the entire Transformer
Arguments:
input_sentence – Tensor of shape (batch_size, input_seq_len, fully_connected_dim)
An array of the indexes of the words in the input sentence
output_sentence – Tensor of shape (batch_size, target_seq_len, fully_connected_dim)
An array of the indexes of the words in the output sentence
“”""
I thought the input should be 2-D [batch, max_sentence_length], why is fully_connected_dim involved in input_sentence. I also read the unit_test and
sentence_lang_a = np.array([[2, 1, 4, 3, 0]])
the argument passed as input_sentence is 2-D and whats expected.
Is this a typo?