Summary of Paper “WHAT’S IN A LOSS FUNCTION FOR IMAGE CLASSIFICATION?”
- Different losses and regularizers achieve broadly similar accuracies. Although the acuuracy differences are larger enough to be meaningful in some contexts, the largerst difference is still < 1.5%.
- The choice of loss function affects representations in only the last few layers of the network, suggesting inherent limitations to what can be achieved by manipulating the loss.
- Different objectives (losses/regularizers) lead to substantially different penultimate layer representations. Claa separation is an important factor that distinguishes these different penultimate layer representations, ans show that it is inversely related to transferability of representation to other tasks.
References
- Paper “WHAT’S IN A LOSS FUNCTION FOR IMAGE CLASSIFICATION?” https://arxiv.org/abs/2010.16402
- https://neptune.ai/blog/pytorch-loss-functions?utm_source=facebook&utm_medium=post-in-group&utm_campaign=blog-pytorch-loss-functions&fbclid=IwAR13rfPMXDJQdTRBboUrycW8W-qIrevFbsvS4ppy00Da40hqJRK0kG9aNaU