Labeling confidence for uncertainty-aware histology image classification

Comput Med Imaging Graph. 2023 Jul:107:102231. doi: 10.1016/j.compmedimag.2023.102231. Epub 2023 Apr 11.

Abstract

Deep learning-based models applied to digital pathology require large, curated datasets with high-quality (HQ) annotations to perform correctly. In many cases, recruiting expert pathologists to annotate large databases is not feasible, and it is necessary to collect additional labeled data with varying label qualities, e.g., pathologists-in-training (henceforth, non-expert annotators). Learning from datasets with noisy labels is more challenging in medical applications since medical imaging datasets tend to have instance-dependent noise and suffer from high inter/intra-observer variability. In this paper, we design an uncertainty-driven labeling strategy with which we generate soft labels from 10 non-expert annotators for multi-class skin cancer classification. Based on this soft annotation, we propose an uncertainty estimation-based framework to handle these noisy labels. This framework is based on a novel formulation using a dual-branch min-max entropy calibration to penalize inexact labels during the training. Comprehensive experiments demonstrate the promising performance of our labeling strategy. Results show a consistent improvement by using soft labels with standard cross-entropy loss during training (∼4.0% F1-score) and increases when calibrating the model with the proposed min-max entropy calibration (∼6.6% F1-score). These improvements are produced at negligible cost, both in terms of annotation and calculation.

Keywords: Digital pathology; Model calibration; Non-expert annotators; Uncertainty estimation.

Publication types

  • Research Support, Non-U.S. Gov't

MeSH terms

  • Calibration
  • Databases, Factual
  • Entropy
  • Histological Techniques*
  • Uncertainty