2023
DOI: 10.48550/arxiv.2302.07232
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

A Psycholinguistic Analysis of BERT's Representations of Compounds

Abstract: This work studies the semantic representations learned by BERT for compounds, that is, expressions such as sunlight or bodyguard. We build on recent studies that explore semantic information in Transformers at the word level and test whether BERT aligns with human semantic intuitions when dealing with expressions (e.g., sunlight) whose overall meaning depends-to a various extent-on the semantics of the constituent words (sun, light). We leverage a dataset that includes human judgments on two psycholinguistic m… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

1
3
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(4 citation statements)
references
References 25 publications
1
3
0
Order By: Relevance
“…We thus do observe that different layers in BERT's architecture capitalise on and highlight different pieces of information extracted from co-occurrence. Even if BERT learns static representations, therefore, deriving prototypes at mid layers through averaging seems to capture semantic associations better, as evaluated using human ratings, in line with previous evidence that lexical semantics is best captured around layer 8 (Buijtelaar & Pezzelle, 2023;Rogers et al, 2020). On the contrary, non-contextualised prototypes' poor performance at deeper layers, i.e.…”
Section: Semantic Relatedness and Similarity Normssupporting
confidence: 76%
See 3 more Smart Citations
“…We thus do observe that different layers in BERT's architecture capitalise on and highlight different pieces of information extracted from co-occurrence. Even if BERT learns static representations, therefore, deriving prototypes at mid layers through averaging seems to capture semantic associations better, as evaluated using human ratings, in line with previous evidence that lexical semantics is best captured around layer 8 (Buijtelaar & Pezzelle, 2023;Rogers et al, 2020). On the contrary, non-contextualised prototypes' poor performance at deeper layers, i.e.…”
Section: Semantic Relatedness and Similarity Normssupporting
confidence: 76%
“…This core, moreover, can shifts following predictable and systematic context effects. We highlight that static and context-dependent representations co-exist in the same embedding space: in this, BERT offers a viable single implementation of both, and holds promise for analysing the interplay and interplay between prototypes and exemplars in a learning model (Buijtelaar & Pezzelle, 2023;Madabushi et al, 2022).…”
Section: Discussionmentioning
confidence: 99%
See 2 more Smart Citations