Revisiting tests for neglected nonlinearity using artificial neural networks

Neural Comput. 2011 May;23(5):1133-86. doi: 10.1162/NECO_a_00117. Epub 2011 Feb 7.

Abstract

Tests for regression neglected nonlinearity based on artificial neural networks (ANNs) have so far been studied by separately analyzing the two ways in which the null of regression linearity can hold. This implies that the asymptotic behavior of general ANN-based tests for neglected nonlinearity is still an open question. Here we analyze a convenient ANN-based quasi-likelihood ratio statistic for testing neglected nonlinearity, paying careful attention to both components of the null. We derive the asymptotic null distribution under each component separately and analyze their interaction. Somewhat remarkably, it turns out that the previously known asymptotic null distribution for the type 1 case still applies, but under somewhat stronger conditions than previously recognized. We present Monte Carlo experiments corroborating our theoretical results and showing that standard methods can yield misleading inference when our new, stronger regularity conditions are violated.

Publication types

  • Research Support, Non-U.S. Gov't

MeSH terms

  • Algorithms
  • Artificial Intelligence*
  • Computer Simulation / standards
  • Mathematical Concepts
  • Models, Theoretical
  • Monte Carlo Method
  • Neural Networks, Computer*
  • Nonlinear Dynamics*
  • Regression Analysis