Using Step dropout to improve accuracy of modified NIN network on CIFAR 100
Adaptive Dropout: This is an approach which i tried for carrying out step by step increase of dropout rate. The value of dropout increase with number of epoch. It reaches the maximum value of 0.5 at the last epoch File:
Non-Adaptive Dropout: The value of dropout is fixed and does not change File:
Curriculum dropout (annealing) is unofficial implementation of the paper
https://arxiv.org/abs/1703.06229
Official implementation can be found in the following link https://github.com/pmorerio/curriculum-dropout
Curriculum (line drop) This is an approach that I tried having properties similar to curriculum annealing and adaptive dropout
| Technique | train_loss | valid_loss | accuracy | error_rate | top_k_accuracy |
|---|---|---|---|---|---|
| Adaptive | 2.445148 | 1.571715 | 0.5796 | 0.4204 | 0.8477 |
| Non-Adaptive | 2.478166 | 1.655046 | 0.5577 | 0.4423 | 0.8336 |
| Curriculum (Annealing) | 2.206255 | 1.5655344 | 0.5837 | 0.4163 | 0.8473 |
| Curriculum (Line drop) | 2.461889 | 1.599115 | 0.5712 | 0.4288 | 0.8442 |
Testing is done on modified NIN architecture. Results are shown only for it. Further analysis is under progress Try it and kindly share your suggestions and views
Next Steps: To Setup Curriculum dropout