Stop Measuring Calibration When Humans Disagree

Open Access
Authors
Publication date 2022
Host editors
  • Y. Goldberg
  • Z. Kozareva
  • Y. Zhang
Book title Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing
Book subtitle December 7-11, 2022, Abu Dhabi, United Arab Emirates
Event The 2022 Conference on Empirical Methods in Natural Language Processing
Pages (from-to) 1892–1915
Number of pages 24
Publisher Association for Computational Linguistics
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
  • Interfacultary Research - Institute for Logic, Language and Computation (ILLC)
Abstract
Calibration is a popular framework to evaluate whether a classifier knows when it does not know - i.e., its predictive probabilities are a good indication of how likely a prediction is to be correct. Correctness is commonly estimated against the human majority class. Recently, calibration to human majority has been measured on tasks where humans inherently disagree about which class applies. We show that measuring calibration to human majority given inherent disagreements is theoretically problematic, demonstrate this empirically on the ChaosNLI dataset, and derive several instance-level measures of calibration that capture key statistical properties of human judgements - including class frequency, ranking and entropy.
Document type Conference contribution
Language English
Published at https://doi.org/10.18653/v1/2022.emnlp-main.124
Other links https://github.com/jsbaan/calibration-on-disagreement-data
Downloads
2022.emnlp-main.124 (Final published version)
Permalink to this page
Back