In Week 3 → Why do you need Non-Linear Activation Functions?
I have one question → I think ReLU is still a linear function of input parameters, because all it is doing is just discarding some values i.e. ones which are negative. But overall it will be still a linear function of input parameters. Please suggest.