A gloss composition and context clustering based distributed word sense representation model

Tao Chen, Ruifeng Xu*, Yulan He, Xuan Wang

*Corresponding author for this work

Research output: Contribution to journalArticle

Abstract

In recent years, there has been an increasing interest in learning a distributed representation of word sense. Traditional context clustering based models usually require careful tuning of model parameters, and typically perform worse on infrequent word senses. This paper presents a novel approach which addresses these limitations by first initializing the word sense embeddings through learning sentence-level embeddings from WordNet glosses using a convolutional neural networks. The initialized word sense embeddings are used by a context clustering based model to generate the distributed representations of word senses. Our learned representations outperform the publicly available embeddings on half of the metrics in the word similarity task, 6 out of 13 sub tasks in the analogical reasoning task, and gives the best overall accuracy in the word sense effect classification task, which shows the effectiveness of our proposed distributed distribution learning model.

Original languageEnglish
Pages (from-to)6007-6024
Number of pages18
JournalEntropy
Volume17
Issue number9
DOIs
Publication statusPublished - 27 Aug 2015

Bibliographical note

This is an open access article distributed under the Creative Commons Attribution License which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Keywords

  • distributed representation
  • lexical semantic compositionality
  • natural language processing
  • word sense disambiguation

Fingerprint Dive into the research topics of 'A gloss composition and context clustering based distributed word sense representation model'. Together they form a unique fingerprint.

  • Cite this