I think the algorithm fails if we guess w , b at local maxima
refer this link
For most simple regression systems, the cost function is known to be convex. So a local maximum is not possible.
1 Like
May I encounter cost function in other models that may cause the gradient descent algo to converge to a local maxima?
How to prevent it in that case?
Thx.
If any of the initial weight values cause the gradients to be non-zero, you won’t start at a maximum, and the gradients will lead away from the maximum on the first iteration.
You would have to be extremely unlucky to choose initial weight values that had all of the gradients at exactly zero.
In practice, you can safely ignore this as a potential issue.
1 Like