Consider the below statement from the optional lab [C2_W1_Lab02_CoffeeRoasting_TF]:
"Letβs examine the weights and biases Tensorflow has instantiated. The weights π
should be of size (number of features in input, number of units in the layer) while the bias π size should match the number of units in the layer:
-
In the first layer with 3 units, we expect W to have a size of (2,3) and π
should have 3 elements. -
In the second layer with 1 unit, we expect W to have a size of (3,1) and π
should have 1 element."
In the first layer, there are 3 neurons and the input X has 2 features (Temperature and Duration). So, this makes sense for first layer, where 3 neurons with 2 feature each for weight and 3 values for bias, which is 3 * 2 + 3 => 9
However, in the second layer, where there is only one neuron, not sure how the number of features became 3 here (as it was 2 in the first layer).
Please explain more on this.