In week 1, we’re using Tensorflow and numpy to implement simple neural networks.
When using Tensorflow, I’m assuming that Tensorflow does a lot “magic” under the hood to compute the weights and biases.
When the layers are manually implemented using numpy, we’re not computing the cost function and minimizing it for each neuron in each layer using gradient descent, why?
I think it is just not the goal for the Numpy lab to train a model. Training a multi-layer neural network from scratch (without Tensorflow) is not covered in this specialization, but in the Deep Learning Specialization.
What about in the CoffeeRoastingNumPy, is there supposed to be Gradient descent but we are plugging in optimal w & b values so the code doesn’t show that? More of a query to make sure my understanding is correct.
Assuming for a real NN, we would also have it do gradient descent and tell us the optimal w & b values