Always mention Assignment too in your header or explanation.
You need to check how you have recalled self.mha1? which you will use to get the correct values for attn_w_b1
Read these instructions
Block 1 is a multi-head attention layer with a residual connection, and look-ahead mask. Like in the EncoderLayer, Dropout is defined within the multi-head attention layer.
The first two blocks are fairly similar to the EncoderLayer except you will return attention_scores when computing self-attention
so make sure you have return_attention_scores=True, look_ahead_mask
Also usual mistake in this assignment is learner using training recall which is not required here for attn_w_b1
hope you are not mixing with block1 and block2. Re-read the instructions again and again, point by point, see if you have followed it as per instructions.