Revisiting the case for explicit syntactic information in language models

Ariya Rastrow, Sanjeev Khudanpur, Mark Dredze

Research output: Contribution to conferencePaperpeer-review

Abstract

Statistical language models used in deployed systems for speech recognition, machine translation and other human language technologies are almost exclusively n-gram models. They are regarded as linguistically naïve, but estimating them from any amount of text, large or small, is straightforward. Furthermore, they have doggedly matched or outperformed numerous competing proposals for syntactically well-motivated models. This unusual resilience of n-grams, as well as their weaknesses, are examined here. It is demonstrated that n-grams are good word-predictors, even linguistically speaking, in a large majority of word-positions, and it is suggested that to improve over n-grams, one must explore syntax-aware (or other) language models that focus on positions where n-grams are weak.

Conference

Conference2012 Workshop on Will We Ever Really Replace the N-gram Model? On the Future of Language Modeling for HLT, WLM 2012 at the 2012 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2012
Country/TerritoryCanada
CityMontreal
Period6/8/12 → …

ASJC Scopus subject areas

  • Linguistics and Language
  • Language and Linguistics
  • Computer Science Applications

Fingerprint

Dive into the research topics of 'Revisiting the case for explicit syntactic information in language models'. Together they form a unique fingerprint.

Cite this