Laurence Moroney recommends the courses of Andrew Ng to learn more about Hyperparameter tuning in the end of the course. Though, he taught many. Do you have a specific recommendation?
Considering where the last assignment of the course left off, I feel this another essential skill to master. I am aware of a broad set of opportunities to optimize a model, potentially. Though I am missing a clear framework (/ best practices) in terms of structuring and prioritizing this task well.
Some example questions:
- Should I first optimize the lr and then the batch size, and then the optimizer, and then, …
- At what stage do I focus on optimizing the nn architecture, rather than the hyper parameters?
- How can I optimally employ tools such as keras tuner, at what stage? Should I rather optimize one hyperparameter at a time, or a lot in conjunction (due to interdependencies)?
- How to make the best of own sanity vs hyperparameter search tools?
Clearly, not everything has a black or white/ universally true answer. But I am sure, there is a lot one could learn from other’s experience and well-proven routines in this regard!
Cheers,
Steffen