The Calibration Generalization Gap

This paper was accepted at the Workshop on Distribution-Free Uncertainty Quantification at ICML 2022.
Calibration is a fundamental property of a good predictive model: it requires that the model predicts correctly in proportion to its confidence. Modern neural networks, however, provide no strong guarantees on their calibration— and can be either poorly calibrated or well-calibrated depending on the setting. It is currently unclear which factors contribute to good calibration (architecture, data augmentation, overparameterization, etc), though various claims exist in the literature. We…Apple Machine Learning Research