During the lecture Andrew states: As a rule of thumb, when performing feature scaling, you might want to aim for getting the features to range from maybe anywhere around negative one to somewhere around plus one for each feature x. **But these values, negative one and plus one can be a little bit loose**. If the features range from negative three to plus three or negative 0.3 to plus 0.3, all of these are completely okay.
What is the issue with feature scaling to [-1;1]?
Why this range is considered as a loose one?
When you reference a lecture, it’s helpful if you give both the lecture title and the time mark.
Andrew lectures in a very intuitive way, rather than assuming the audience has a high degree of math skills.
In this case, what he means by “these values can be a little bit loose” is that -1 and +1 aren’t hard limits. You can use any range of small negative values that are symmetric around zero.