Locally linear attributes of ReLU neural networks

Front Artif Intell. 2023 Nov 23:6:1255192. doi: 10.3389/frai.2023.1255192. eCollection 2023.

Abstract

A ReLU neural network functions as a continuous piecewise linear map from an input space to an output space. The weights in the neural network determine a partitioning of the input space into convex polytopes, where each polytope is associated with a distinct affine mapping. The structure of this partitioning, together with the affine map attached to each polytope, can be analyzed to investigate the behavior of the associated neural network. We investigate simple problems to build intuition on how these regions act and both how they can potentially be reduced in number and how similar structures occur across different networks. To validate these intuitions, we apply them to networks trained on MNIST to demonstrate similarity between those networks and the potential for them to be reduced in complexity.

Keywords: Jacobian matrices; ReLU; linear mapping; linearization; neural networks; polyhedral decomposition.

Grants and funding

This study is partially supported by the DARPA Geometries of Learning Program under Award No. HR00112290074.