The pre-trained ResNets50 model in Lab 1 Week 2 has a putative accuracy of 95%, but it does not perform well on my own pictures. For example, this is the output for the image shown
the model also fails with other images. A “one” sign is classified as class 0.
After I trained the model for 12 epochs (reaching accuracy on tests of 97%) the performance on my own images was equally poor.
The elementary logistic model of week 1 does much better. !!>??
Answering my own question. After I progressed to the ResNets50 model (Lab, Week 2), I applied it to a problem of my own (a categorical with 10 classes). The performance was poor, both in training and testing, until I removed a few of the deepest layers in the model, reducing the number of parameters from 25 million to about 10 million. With the smaller (shallower) model, I got to 80% accuracy in training, a bit worse in testing. In short: one of the reasons these models fail is what Prof. Ng defined as “overfitting”. Reducing the number of “neurons”/parameters might be a way to improve performance.