prof Andrew in the course teaches how to implement regularization into both normal regression and logistic regression but he did not mention how to do it if you are using libs like sklearn or so and this is the real-world scenario and any one help on how to use it in any library , better if sklearn.

note: when I researched I found they are either scaling all data or using lasso or ridge or another types but not what prof Andrew do is there a way to implement what he says using sklearn or other alternatives

This MLS only teaches the L2 regularization, and teaches it in the context of solving it with gradient descent. In sklearn, Ridge, Lasso, and ElasticNet do similar things but not using gradient descent:

name

L1 regularization type available

use gradient descent?

sklearn.linear_model.SGDRegressor

L2 / L1 / ElasticNet

Yes

sklearn.linear_model.Lasso

L1

No

sklearn.linear_model.Ridge

L2

No

sklearn.linear_model.ElasticNet

L2 / L1 / ElasticNet

No

So, the closest tool to our MLS is sklearn.linear_model.SGDRegressor, but it still has one major difference: MLS teaches batch gradient descent, whereas SGDRegressor uses stochastic gradient descent, so don’t expect them to produce the same result. If you just want more similar results, then Ridge should be the easiest one.

All info above can be found on sklearn’s documentation, so you can make a similar table for logistic regression if you want to, and you may start from here for what’s available.