Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Langua 2022
DOI: 10.18653/v1/2022.naacl-main.327
|View full text |Cite
|
Sign up to set email alerts
|

What company do words keep? Revisiting the distributional semantics of J.R. Firth & Zellig Harris

Abstract: The power of word embeddings is attributed to the linguistic theory that similar words will appear in similar contexts. This idea is specifically invoked by noting that "you shall know a word by the company it keeps," a quote from British linguist J.R. Firth who, along with his American colleague Zellig Harris, is often credited with the invention of "distributional semantics." While both Firth and Harris are cited in all major NLP textbooks and many foundational papers, the content and differences between the… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 7 publications
(4 citation statements)
references
References 15 publications
0
4
0
Order By: Relevance
“…Distributional semantics methods allow us to quantify word meanings without specifying semantic features (e.g., that a door and gate are both "movable barriers"; Rogers & McClelland, 2004;Miller, 1995). Distributional techniques presuppose that two words that typically occur in similar (distributional) contexts are probably similar in meaning and thus have similar vector representations (Brunila & LaViolette, 2022). The similarity in meaning between two words is approximated by some measure of distance (e.g., Euclidean distance or cosine similarity) between the vectors, with more distant vectors being less similar in meaning.…”
Section: The Cloze Task and Large Language Models (Llms)mentioning
confidence: 99%
“…Distributional semantics methods allow us to quantify word meanings without specifying semantic features (e.g., that a door and gate are both "movable barriers"; Rogers & McClelland, 2004;Miller, 1995). Distributional techniques presuppose that two words that typically occur in similar (distributional) contexts are probably similar in meaning and thus have similar vector representations (Brunila & LaViolette, 2022). The similarity in meaning between two words is approximated by some measure of distance (e.g., Euclidean distance or cosine similarity) between the vectors, with more distant vectors being less similar in meaning.…”
Section: The Cloze Task and Large Language Models (Llms)mentioning
confidence: 99%
“…Similar advances occurred in NLP, which refers to the branch of AI Zygon focused on developing software to understand and generate natural language (such as text and spoken language). NLP built upon a shift in understanding "meaning" in computational linguistics from depending upon symbol references to deriving from word cooccurrence and shared contexts (Brunila and LaViolette 2022). This shift to associative and distributional theories of semantics (Firth 1957;Harris 1968;Sahlgren 2008) facilitated statistical methods for processing and understanding natural language and enabled significant improvements in modeling language and performing language-related tasks (Sejnowski 2020;Goldberg 2016;Bommasani et al 2022).…”
Section: Paradigms For Ai Researchmentioning
confidence: 99%
“…The distributional semantics (Harris, 1954;Firth, 1957) suggests that "similar words will appear in similar contexts" (Brunila and LaViolette, 2022). This implies that the conditional probability distribution p(•|w) represents the meaning of a word w.…”
Section: Kl Divergence Measures Information Gainmentioning
confidence: 99%