Denoising Autoencoders for Overgeneralization in Neural Networks

IEEE Trans Pattern Anal Mach Intell. 2020 Apr;42(4):998-1004. doi: 10.1109/TPAMI.2019.2909876. Epub 2019 Apr 9.

Abstract

Despite recent developments that allowed neural networks to achieve impressive performance on a variety of applications, these models are intrinsically affected by the problem of overgeneralization, due to their partitioning of the full input space into the fixed set of target classes used during training. Thus it is possible for novel inputs belonging to categories unknown during training or even completely unrecognizable to humans to fool the system into classifying them as one of the known classes, even with a high degree of confidence. This problem can lead to security problems in critical applications, and is closely linked to open set recognition and 1-class recognition. This paper presents a novel way to compute a confidence score using the reconstruction error of denoising autoencoders and shows how it can correctly identify the regions of the input space close to the training distribution. The proposed solution is tested on benchmarks of 'fooling', open set recognition and 1-class recognition constructed from the MNIST and Fashion-MNIST datasets.

Publication types

  • Research Support, Non-U.S. Gov't