Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics 2019
DOI: 10.18653/v1/p19-1071
|View full text |Cite
|
Sign up to set email alerts
|

SP-10K: A Large-scale Evaluation Set for Selectional Preference Acquisition

Abstract: Selectional Preference (SP) is a commonly observed language phenomenon and proved to be useful in many natural language processing tasks. To provide a better evaluation method for SP models, we introduce SP-10K, a largescale evaluation set that provides human ratings for the plausibility of 10,000 SP pairs over five SP relations, covering 2,500 most frequent verbs, nouns, and adjectives in American English. Three representative SP acquisition methods based on pseudo-disambiguation are evaluated with SP-10K. To… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
43
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
3
3
1

Relationship

2
5

Authors

Journals

citations
Cited by 29 publications
(45 citation statements)
references
References 19 publications
2
43
0
Order By: Relevance
“…We create two training sets based on separate corpora: first, we parse English Wikipedia using the StanfordNLP neural pipeline (Qi et al, 2018) and extract attested s-v-o triples. Wikipedia has led to relatively good results for selectional preference (Zhang et al, 2019), and in total we extract 6 million unique triples with a cumulative 10 million occurrences. Second, we use the NELL (Carlson et al, 2010) dataset of 604 million s-v-o triples extracted from the dependency parsed ClueWeb09 dataset.…”
Section: Learning From Textmentioning
confidence: 99%
“…We create two training sets based on separate corpora: first, we parse English Wikipedia using the StanfordNLP neural pipeline (Qi et al, 2018) and extract attested s-v-o triples. Wikipedia has led to relatively good results for selectional preference (Zhang et al, 2019), and in total we extract 6 million unique triples with a cumulative 10 million occurrences. Second, we use the NELL (Carlson et al, 2010) dataset of 604 million s-v-o triples extracted from the dependency parsed ClueWeb09 dataset.…”
Section: Learning From Textmentioning
confidence: 99%
“…Besides that, we also compare with the selectional preference (SP) based method [46]. Following the original setting, two resources (human annotation and Posterior Probability) of SP knowledge are considered and we denote them as SP (human) and SP (PP) respectively 15 .…”
Section: Baseline Methodsmentioning
confidence: 99%
“…Given one sentence 'The dog is chasing the cat, it barks loudly', we can correctly resolve 'it' to 'dog' rather than 'cat' because 'dog barks' appears 12,247 times in ASER, while 'cat barks' never appears. This is usually called selectional preference [43], which has recently been evaluated in a larger scale in [46]. ASER naturally reflects human's selectional preference for many kinds of syntactic patterns.…”
Section: # Eventuality # Relation # R Typesmentioning
confidence: 99%
“…• Distributional Similarity (DS) (Erk et al, 2010), a method that uses the similarity of the SP Evaluation Set #W #P Keller (Keller and Lapata, 2003) 571 360 SP-10K (Zhang et al, 2019a) 2,500 6,000 embedding of the target argument and average embedding of observed golden arguments in the corpus to predict the preference strength.…”
Section: Baselinesmentioning
confidence: 99%
“…Thus, in this section, we evaluate different SP acquisition methods with ground truth (human labeled datasets). Two representative datasets, Keller (Keller and Lapata, 2003) and SP-10K (Zhang et al, 2019a) word pairs are provided. For each of the word pairs, the datasets also provide the annotated plausibility score of how likely a preference exists between that word pair under the corresponding SP relation.…”
Section: Selectional Preference Acquisitionmentioning
confidence: 99%