Abstract
Statistical language models used in deployed systems for speech recognition, machine translation and other human language technologies are almost exclusively n-gram models. They are regarded as linguistically naïve, but estimating them from any amount of text, large or small, is straightforward. Furthermore, they have doggedly matched or outperformed numerous competing proposals for syntactically well-motivated models. This unusual resilience of n-grams, as well as their weaknesses, are examined here. It is demonstrated that n-grams are good word-predictors, even linguistically speaking, in a large majority of word-positions, and it is suggested that to improve over n-grams, one must explore syntax-aware (or other) language models that focus on positions where n-grams are weak.
Original language | English (US) |
---|---|
Pages | 50-58 |
Number of pages | 9 |
State | Published - 2012 |
Externally published | Yes |
Event | 2012 Workshop on Will We Ever Really Replace the N-gram Model? On the Future of Language Modeling for HLT, WLM 2012 at the 2012 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2012 - Montreal, Canada Duration: Jun 8 2012 → … |
Conference
Conference | 2012 Workshop on Will We Ever Really Replace the N-gram Model? On the Future of Language Modeling for HLT, WLM 2012 at the 2012 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2012 |
---|---|
Country/Territory | Canada |
City | Montreal |
Period | 6/8/12 → … |
ASJC Scopus subject areas
- Linguistics and Language
- Language and Linguistics
- Computer Science Applications