Course 2, Week 3, compute_total_loss(logits, labels)

There’s another issue that they don’t really discuss in much detail here which is arguably quite a bit more subtle: that’s the fact that they did not include the activation function at the output layer in forward propagation. If you were searching the forums for threads about this section, you probably also found some discussing the from_logits issue, e.g. this one.