Hi,
I am currently taking the Supervised Machine Learning: Regression and Classification course on Coursera. In the Week 1 content, the cost function and its derivatives are explained using the error term as (y^−y)2 [predicted y minus y whole squared]. However, in my textbook and most online references, the error is represented as (y−y^)2.
Could you please clarify why the course uses (y^−y)2 instead of (y−y^)2, and whether there is any difference between the two?
Thank you.