Dependency distance: A new perspective on syntactic patterns in natural languages

Phys Life Rev. 2017 Jul:21:171-193. doi: 10.1016/j.plrev.2017.03.002. Epub 2017 Mar 27.

Abstract

Dependency distance, measured by the linear distance between two syntactically related words in a sentence, is generally held as an important index of memory burden and an indicator of syntactic difficulty. Since this constraint of memory is common for all human beings, there may well be a universal preference for dependency distance minimization (DDM) for the sake of reducing memory burden. This human-driven language universal is supported by big data analyses of various corpora that consistently report shorter overall dependency distance in natural languages than in artificial random languages and long-tailed distributions featuring a majority of short dependencies and a minority of long ones. Human languages, as complex systems, seem to have evolved to come up with diverse syntactic patterns under the universal pressure for dependency distance minimization. However, there always exist a small number of long-distance dependencies in natural languages, which may reflect some other biological or functional constraints. Language system may adapt itself to these sporadic long-distance dependencies. It is these universal constraints that have shaped such a rich diversity of syntactic patterns in human languages.

Keywords: Dependency distance; Language universal; Syntactic patterns.

Publication types

  • Review

MeSH terms

  • Humans
  • Linguistics*