2015
DOI: 10.1162/tacl_a_00154
|View full text |Cite
|
Sign up to set email alerts
|

Plato: A Selective Context Model for Entity Resolution

Abstract: We present Plato, a probabilistic model for entity resolution that includes a novel approach for handling noisy or uninformative features, and supplements labeled training data derived from Wikipedia with a very large unlabeled text corpus. Training and inference in the proposed model can easily be distributed across many servers, allowing it to scale to over 107 entities. We evaluate Plato on three standard datasets for entity resolution. Our approach achieves the best results to-date on TAC KBP 2011 and is h… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
82
0

Year Published

2017
2017
2021
2021

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 59 publications
(82 citation statements)
references
References 13 publications
0
82
0
Order By: Relevance
“…for each i ∈ {1, ..., n} (Bunescu and Paşca, 2006;Lazic et al, 2015;Yamada et al, 2017). A global model, besides using local context within Ψ(e i , c i ), takes into account entity coherency.…”
Section: Local and Global Modelsmentioning
confidence: 99%
See 1 more Smart Citation
“…for each i ∈ {1, ..., n} (Bunescu and Paşca, 2006;Lazic et al, 2015;Yamada et al, 2017). A global model, besides using local context within Ψ(e i , c i ), takes into account entity coherency.…”
Section: Local and Global Modelsmentioning
confidence: 99%
“…NEL methods typically consider only coreference, relying either on off-the-shelf systems or some simple heuristics (Lazic et al, 2015), and exploit them in a pipeline fashion, though some (e.g., Cheng and Roth (2013); Ren et al (2017)) additionally exploit a range of syntactic-semantic relations such as apposition and possessives. Another line of work ignores relations altogether and models the predicted sequence of KB entities as a bag (Globerson et al, 2016;Yamada et al, 2016;Ganea and Hofmann, 2017).…”
Section: Introductionmentioning
confidence: 99%
“…(2011a)). This is also true of the only other method which, like ours, uses a combination of Wikipedia data and unlabeled texts (Lazic et al, 2015). We will refer to approaches using this form of supervision, including our approach, as Wikipedia-based linkers.…”
Section: Introductionmentioning
confidence: 84%
“…The resulting model is significantly less accurate than the one which used unlabeled documents. The score difference is larger (Chisholm and Hachey, 2015) 84.9 ------Wiki + unlab (Lazic et al, 2015) 86. (Chisholm and Hachey, 2015) 88.7 ------Fully-supervised (Wiki + AIDA CoNLL train) (Guo and Barbosa, 2016) 89.0 92 87 88 77 84.5 85.7 (Globerson et al, 2016) 91.0 ------ (Yamada et al, 2016) 91. for AIDA-CoNLL test set than for the other 5 test sets.…”
Section: Analysis and Ablationsmentioning
confidence: 99%
“…Though we experimented with local models, the local-global distinction is largely orthogonal as we can directly integrate coherence modeling components in our DL approach. Different types of supervision have been considered in previous work: full supervision (Yamada et al, 2017;Ganea and Hofmann, 2017;Le and Titov, 2018), using combinations of labeled and unlabeled data (Lazic et al, 2015), and even distant supervision (Fan et al, 2015). The approach of Fan et al (2015) is heavily Wikipediabased: they rely on a heuristic mapping from Freebase entities to Wikipedia entities, and learn features from Wikipedia articles.…”
Section: Related Workmentioning
confidence: 99%