LEARNING SEMANTICS-ENHANCED LANGUAGE MODELS APPLIED TO UNSUEPRVISED WSD
- Los Alamos National Laboratory
An N-gram language model aims at capturing statistical syntactic word order information from corpora. Although the concept of language models has been applied extensively to handle a variety of NLP problems with reasonable success, the standard model does not incorporate semantic information, and consequently limits its applicability to semantic problems such as word sense disambiguation. We propose a framework that integrates semantic information into the language model schema, allowing a system to exploit both syntactic and semantic information to address NLP problems. Furthermore, acknowledging the limited availability of semantically annotated data, we discuss how the proposed model can be learned without annotated training examples. Finally, we report on a case study showing how the semantics-enhanced language model can be applied to unsupervised word sense disambiguation with promising results.
- Research Organization:
- Los Alamos National Laboratory (LANL), Los Alamos, NM (United States)
- Sponsoring Organization:
- USDOE National Nuclear Security Administration (NNSA)
- DOE Contract Number:
- AC52-06NA25396
- OSTI ID:
- 985889
- Report Number(s):
- LA-UR-07-0618; TRN: US201017%%67
- Resource Relation:
- Conference: ASSOCIATION FOR COMPUTATIONAL LINGUISTICS ANNUAL MEETING ; 200706 ; PRAGUE
- Country of Publication:
- United States
- Language:
- English
Similar Records
Computationally Efficient Learning of Quality Controlled Word Embeddings for Natural Language Processing
Experiments in automatic word class and word sense identification for information retrieval