Hi everyone,
I’m currently working on a deep learning project as part of Week 3 of the course. I have a model that performs reasonably well on the training data, but its accuracy on validation/test data is lower than I’d like. My dataset is limited, so I cannot collect more data to improve performance.
I’d like advice on practical ways to increase model accuracy without adding new data. I’ve read about several potential strategies, but I’m not sure how to apply them effectively. Some approaches I’m considering include:
-
Data augmentation: Creating variations of existing data, but I’m unsure which techniques work best for my type of data.
-
Regularization: Using dropout, L1/L2 penalties, or batch normalization to prevent overfitting.
-
Hyperparameter tuning: Adjusting learning rate, batch size, number of layers, or activation functions.
-
Transfer learning: Leveraging pre-trained models to improve performance on small datasets.
-
Ensemble methods: Combining predictions from multiple models.
I’m looking for guidance on which of these methods tend to have the most impact in practice, and any tips on implementing them effectively. Examples, references, or personal experiences would be really helpful!
Thank you in advance for your advice.