1. Try to overfit your network on much smaller data and for many epochs without augmenting first (Link)
2. Explore variations of the network (Link)
3. Change layers number and units number.
4. Change batch size.
5. Add dropout layer. (Link)
6. Verify that your code is bug free There's a saying among writers that "All writing is re-writing" -- that is, the greater part of writing is revising.
7. Keep a Logbook of Experiments
8. Look at individual layers, Build a simpler model first (Link)
9. Use nn.BCEWithLogitsLoss as the criterion instead of nn.BCELoss and pass the raw logits to it by removing the sigmoid. (Link)
10. In colab plus GPU + HighRAM worked better
Even after 100+Epochs doesn't seem promising.
Github code is a good start but needs to relearn, Start again :)
Keep Learning!!!
No comments:
Post a Comment