In the regularization lecture, professor Andrew Ng introduces two different notations for L1 regularization.

To be on the same page, notice that this formula applies for the logistic regression where

`w`

is a column vector consisting of weights between each input neuron and the output layer (which consists of one neuron if we are dealing with binary classification).
Now, my confusion arises from the fact that LHS of the equation is not equal to the RHS of the equation. In LHS, we sum the absolute values of each weight and then multiply with `(lambda)/2m`

.

However, in the RHS we calculate the magnitude of the vector `w`

, which is equivalent to summing squares of each weight and then taking the square root of the sum. Obviously LHS and RHS are not the same and they donâ€™t necessarily give the same results.

I strongly believe LHS is the real L1 regularization.

1- Is there really a mistake at this point about the notation in RHS or do i miss something?

2- Can you confirm that LHS is the one that is called L1 regularization.

Thanks a lot.