Proceedings of the Third Workshop on Representation Learning for NLP 2018
DOI: 10.18653/v1/w18-3003
|View full text |Cite
|
Sign up to set email alerts
|

Extrofitting: Enriching Word Representation and its Vector Space with Semantic Lexicons

Abstract: We propose post-processing method for enriching not only word representation but also its vector space using semantic lexicons, which we call extrofitting. The method consists of 3 steps as follows: (i) Expanding 1 or more dimension(s) on all the word vectors, filling with their representative value. (ii) Transferring semantic knowledge by averaging each representative values of synonyms and filling them in the expanded dimension(s). These two steps make representations of the synonyms close together. (iii) Pr… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2020
2020
2021
2021

Publication Types

Select...
3
2

Relationship

1
4

Authors

Journals

citations
Cited by 5 publications
(3 citation statements)
references
References 20 publications
0
3
0
Order By: Relevance
“…We use GloVe (Pennington et al, 2014) as pretrained embeddings. To increase model performance, we apply a word vector post-processing method called extrofitting (Jo and Choi, 2018). We prepare 3 topic classification datasets; DBpedia ontology (DBpedia) (Lehmann et al, 2015), YahooAnswers (Yahoo) (Chang et al, 2008), AG-News.…”
Section: Methodsmentioning
confidence: 99%
“…We use GloVe (Pennington et al, 2014) as pretrained embeddings. To increase model performance, we apply a word vector post-processing method called extrofitting (Jo and Choi, 2018). We prepare 3 topic classification datasets; DBpedia ontology (DBpedia) (Lehmann et al, 2015), YahooAnswers (Yahoo) (Chang et al, 2008), AG-News.…”
Section: Methodsmentioning
confidence: 99%
“…Joint specialization models (Yu and Dredze, 2014;Kiela et al, 2015;Liu et al, 2015;Osborne et al, 2016;Nguyen et al, 2017, inter alia) jointly train word embedding models from scratch and enforce the external constraints with an auxiliary objective. On the other hand, retrofitting models are postprocessors that fine-tune pretrained word embeddings by gauging pairwise distances according to the external constraints (Faruqui et al, 2015;Wieting et al, 2015;Mrkšić et al, 2016;Mrkšić et al, 2017;Jo and Choi, 2018;Lengerich et al, 2018).…”
Section: Specialization For Semantic Similaritymentioning
confidence: 99%
“…Post-hoc Approaches. In the post-hoc approach, pre-trained word vectors such as GloVe (Pennington et al, 2014), Word2Vec (Mikolov et al, 2013), FastText (Bojanowski et al, 2017), or Paragram (Wieting et al, 2015) are fine-tuned to endow them with lexical relational information (Faruqui et al, 2015;Rothe and Schütze, 2015;Wieting et al, 2015;Mrkšić et al, 2016Jo, 2018;Jo and Choi, 2018;Glavaš and Vulić, 2018). In this paper, we primarily discuss LEXSUB as a post-hoc model.…”
Section: Related Workmentioning
confidence: 99%