# BLOCK 1
# calculate self-attention and return attention scores as attn_weights_block1 (~1 line)
attn1, attn_weights_block1 = self.mha1(x, x, x, look_ahead_mask, return_attention_scores=True)
# BLOCK 2
# calculate self-attention using the Q from the first block and K and V from the encoder output.
# Return attention scores as attn_weights_block2 (~1 line)
attn2, attn_weights_block2 = self.mha2(out1, enc_output, enc_output, padding_mask, return_attention_scores=True)