Hi,
In C2W2 we learn, among many other things, how to augment image data in order to avoid over fitting. In the Assignment exercise, we use the cats and dogs data to test the previously learnt principles. The accuracy I get after 15 epochs and using the same model is:
- Without augmentation: Training = 98% and Validation = 89%
- With augmentation: Training = 76% and Validation = 82%
Given the above results, why should we augment the training set since the results degrade on the validation set? In the end the result on the validation is what counts no?
Could it be that, due to the augmentation, the training set becomes more general than the test set? The model tries to capture this generalization without success resulting in a degradation of the results on the validation set.
Could a solution be to augment the validation set in the same way as the training set? Most information I find on the internet however says that augmenting the validation set is very uncommon.
Any thoughts?
Best regards,
Wouter