The many-body expansion combined with neural networks

J Chem Phys. 2017 Jan 7;146(1):014106. doi: 10.1063/1.4973380.

Abstract

Fragmentation methods such as the many-body expansion (MBE) are a common strategy to model large systems by partitioning energies into a hierarchy of decreasingly significant contributions. The number of calculations required for chemical accuracy is still prohibitively expensive for the ab initio MBE to compete with force field approximations for applications beyond single-point energies. Alongside the MBE, empirical models of ab initio potential energy surfaces have improved, especially non-linear models based on neural networks (NNs) which can reproduce ab initio potential energy surfaces rapidly and accurately. Although they are fast, NNs suffer from their own curse of dimensionality; they must be trained on a representative sample of chemical space. In this paper we examine the synergy of the MBE and NN's and explore their complementarity. The MBE offers a systematic way to treat systems of arbitrary size while reducing the scaling problem of large systems. NN's reduce, by a factor in excess of 106, the computational overhead of the MBE and reproduce the accuracy of ab initio calculations without specialized force fields. We show that for a small molecule extended system like methanol, accuracy can be achieved with drastically different chemical embeddings. To assess this we test a new chemical embedding which can be inverted to predict molecules with desired properties. We also provide our open-source code for the neural network many-body expansion, Tensormol.