Hi guys,
In Course 2, week 1, Grad Check, Exercise 2 - backward propagation:
Doesn’t this method just return x? If so, there is no difference in the returned values of J+ and J- and J? In other words, what is the effect of theta?
Hi, @Baraa.
backward_propagation
is not used to compute J_plus
or J_minus
.
What you are doing is using the definition of the derivative to approximate the gradient (with J_plus
and J_minus
) and comparing it with the gradient returned by backward_propagation
. If backward_propagation
was implemented correctly, both values should be very close.
The parameter theta
has no effect in backward_propagation
because it does not appear in the derivative of J with respect to theta (it may for a different cost function).
Did you complete the assignment?
Thanks @nramon. Apparently, I was confused about how gradient checking is working. I am about to complete the assignment. Your help is really appreciated.
Happy to help, @Baraa. Good luck with the assignment!
Thank you! @nramon …