Digital Stereotypes in HMI-The Influence of Feature Quantity Distribution in Deep Learning Models Training

Sensors (Basel). 2022 Sep 6;22(18):6739. doi: 10.3390/s22186739.

Abstract

This paper proposes a concept of Digital Stereotypes, observed during research on quantitative overrepresentation of one class over others, and its impact on the results of the training of Deep Learning models. The real-life observed data classes are rarely of the same size, and the intuition of presenting multiple examples of one class and then showing a few counterexamples may be very misleading in multimodal classification. Deep Learning models, when taught with overrepresentation, may produce incorrect inferring results, similar to stereotypes. The generic idea of stereotypes seems to be helpful for categorisation from the training point of view, but it has a negative influence on the inferring result. Authors evaluate a large dataset in various scenarios: overrepresentation of one or two classes, underrepresentation of some classes, and same-size (trimmed) classes. The presented research can be applied to any multiclassification applications, but it may be especially important in AI, where the classification, uncertainty and building new knowledge overlap. This paper presents specific 'decreases in accuracy' observed within multiclassification of unleveled datasets. The 'decreases in accuracy', named by the authors 'stereotypes', can also bring an inspiring insight into other fields and applications, not only multimodal sentiment analysis.

Keywords: Industry 5.0; artificial intelligence; cognitive technologies; digital stereotypes; feature measurement; machine learning.

MeSH terms

  • Artificial Intelligence*
  • Deep Learning*

Grants and funding

This research received no external funding.