Proceedings of the 5th Workshop on Representation Learning for NLP 2020
DOI: 10.18653/v1/2020.repl4nlp-1.22
|View full text |Cite
|
Sign up to set email alerts
|

Evaluating Compositionality of Sentence Representation Models

Abstract: We evaluate the compositionality of generalpurpose sentence encoders by proposing two metrics to quantify compositional understanding capability of sentence encoders. We introduce a novel metric, Polarity Sensitivity Scoring (PSS), which utilizes sentiment perturbations as a proxy for measuring compositionality. We then compare results from PSS with those obtained via our proposed extension of a metric called Tree Reconstruction Error (TRE) (Andreas, 2019) where compositionality is evaluated by measuring how w… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
6
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(6 citation statements)
references
References 14 publications
0
6
0
Order By: Relevance
“…When employing the metric, one should define an appropriate distance function (δ) and define fη parametrised by η. Andreas illustrates the TRE's versatility by instantiating it for three scenarios: to investigate whether image representations are similar to composed image attributes, whether phrase embeddings are similar to the vector addition of their components, and whether generalisation accuracy in a reference game positively correlates with TRE. Bhathena et al (2020) present two methods based on TRE to obtain compositionality ratings for sentiment trees, referred to as tree impurity and weighted node switching that express the difference between the sentiment label of the root and the other nodes in the tree. Zheng and Jiang (2022) ranked examples of sentiment analysis based on the extent to which neural models should memorise examples in order to capture their target correctly.…”
Section: Related Workmentioning
confidence: 99%
“…When employing the metric, one should define an appropriate distance function (δ) and define fη parametrised by η. Andreas illustrates the TRE's versatility by instantiating it for three scenarios: to investigate whether image representations are similar to composed image attributes, whether phrase embeddings are similar to the vector addition of their components, and whether generalisation accuracy in a reference game positively correlates with TRE. Bhathena et al (2020) present two methods based on TRE to obtain compositionality ratings for sentiment trees, referred to as tree impurity and weighted node switching that express the difference between the sentiment label of the root and the other nodes in the tree. Zheng and Jiang (2022) ranked examples of sentiment analysis based on the extent to which neural models should memorise examples in order to capture their target correctly.…”
Section: Related Workmentioning
confidence: 99%
“…Firstly, existing work derive a single embedding for the entire query. The problem of representing longer text inputs is being actively researched in the community and remains an open problem [7,8,6]. This means that specific details or nested subqueries of the query may be omitted or not represented properly -getting lost in the embedding.…”
Section: Neural Models For Semantic Code Searchmentioning
confidence: 99%
“…Despite impressive results on SCS, current neural approaches are far from satisfactory in dealing with a wide range of natural-language queries, especially on ones with compositional language structure. Encoding longer text into a dense vector is an open problem for neural language models, as neural networks are not believed to be extracting systematic rules from data [7,8,6]. Not only does this a) affect the performance, but it can b) drastically reduce a model's value for the users, because compositional queries such as "Check that directory does not exist before creating it" require performing multi-step reasoning on code.…”
Section: Introductionmentioning
confidence: 99%
“…Phrase and sentence composition has drawn frequent attention in analysis of neural models, often focusing on analysis of internal representations and downstream task behavior (Ettinger et al, 2018;Conneau et al, 2019;Nandakumar et al, 2019;Yu and Ettinger, 2020;Bhathena et al, 2020;Mu and Andreas, 2020;Andreas, 2019). Some work investigates compositionality via constructing linguistic (Keysers et al, 2019) and non-linguistic (Liška et al, 2018;Hupkes et al, 2018;Baan et al, 2019) and Ettinger (2020).…”
Section: Related Workmentioning
confidence: 99%