Abstract

Unsupervised word representations are very useful in NLP tasks both as inputs to learning algorithms and as extra word features in NLP systems. However, most of these models are built with only local context and one representation per word. This is problematic because words are often polysemous and global context can also provide useful information for learning word meanings. We present a new neural network architecture which 1) learns word embeddings that better capture the semantics of words by incorporating both local and global document context, and 2) accounts for homonymy and polysemy by learning multiple embeddings per word. We introduce a new dataset with human judgments on pairs of words in sentential context, and evaluate our model on it, showing that our model outperforms competitive baselines and other neural language models. 1 1

Keywords

PolysemyComputer scienceWord (group theory)Natural language processingArtificial intelligenceContext (archaeology)Semantics (computer science)Context modelRepresentation (politics)SemEvalArtificial neural networkLinguistics

Affiliated Institutions

Related Publications

Word Space

Representations for semantic information about words are necessary for many applications of neural networks in natural language processing. This paper describes an efficient, co...

1992 Neural Information Processing Systems 212 citations

Publication Info

Year
2012
Type
article
Pages
873-882
Citations
1099
Access
Closed

External Links

Citation Metrics

1099
OpenAlex

Cite This

Eric Huang, Richard Socher, Christopher D. Manning et al. (2012). Improving Word Representations via Global Context and Multiple Word Prototypes. , 873-882.