Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) 2018
DOI: 10.18653/v1/p18-1230
|View full text |Cite
|
Sign up to set email alerts
|

Incorporating Glosses into Neural Word Sense Disambiguation

Abstract: Word Sense Disambiguation (WSD) aims to identify the correct meaning of polysemous words in the particular context. Lexical resources like WordNet which are proved to be of great help for WSD in the knowledge-based methods. However, previous neural networks for WSD always rely on massive labeled data (context), ignoring lexical resources like glosses (sense definitions). In this paper, we integrate the context and glosses of the target word into a unified framework in order to make full use of both labeled dat… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
62
0

Year Published

2018
2018
2021
2021

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 86 publications
(63 citation statements)
references
References 20 publications
1
62
0
Order By: Relevance
“…Bi-LSTM +att.+LEX+P OS (Raganato et al, 2017a) is a multi-task learning framework for WSD, POS tagging, and LEX with self-attention mechanism, which converts WSD to a sequence learning task. GAS ext (Luo et al, 2018b) is a variant of GAS which is a gloss-augmented variant of the memory network by extending gloss knowledge. CAN s and HCAN (Luo et al, 2018a) are sentence-level and hierarchical co-attention neural network models which leverage gloss knowledge.…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…Bi-LSTM +att.+LEX+P OS (Raganato et al, 2017a) is a multi-task learning framework for WSD, POS tagging, and LEX with self-attention mechanism, which converts WSD to a sequence learning task. GAS ext (Luo et al, 2018b) is a variant of GAS which is a gloss-augmented variant of the memory network by extending gloss knowledge. CAN s and HCAN (Luo et al, 2018a) are sentence-level and hierarchical co-attention neural network models which leverage gloss knowledge.…”
Section: Resultsmentioning
confidence: 99%
“…For a fair comparison, we use the benchmark datasets proposed by Raganato et al (2017b) which include five standard all-words fine-grained WSD datasets from the Senseval and SemEval competitions: Senseval-2 (SE2), Senseval-3 (SE3), SemEval-2007 (SE07), SemEval-2013 (SE13) and SemEval-2015 (SE15). Following Luo et al (2018a), Luo et al (2018b) and Raganato et al (2017a), we choose SE07, the smallest among these test sets, as the development set.…”
Section: Evaluation Datasetsmentioning
confidence: 99%
“…Addressing the issue of scarce annotations, recent works have proposed methods for using resources from knowledge-based approaches. Luo et al (2018a) and Luo et al (2018b) combine information from glosses present in WordNet, with NLMs based on BiLSTMs, through memory networks and co-attention mechanisms, respectively. Vial et al (2018) follows Raganato et al (2017b)'s BiLSTM method, but leverages the semantic network to strategically reduce the set of senses required for disambiguating words.…”
Section: Wsd State-of-the-artmentioning
confidence: 99%
“…8. Some methods can also be categorized as hybrid, as they make use of both sense-annotated corpora and knowledge resources, e.g., the gloss-augmented model of Luo et al (2018).…”
Section: Word Sense Disambiguationmentioning
confidence: 99%