2013
DOI: 10.14778/2536360.2536374
|View full text |Cite
|
Sign up to set email alerts
|

Reducing uncertainty of schema matching via crowdsourcing

Abstract: Schema matching is a central challenge for data integration systems. Automated tools are often uncertain about schema matchings they suggest, and this uncertainty is inherent since it arises from the inability of the schema to fully capture the semantics of the represented data. Human common sense can often help. Inspired by the popularity and the success of easily accessible crowdsourcing platforms, we explore the use of crowdsourcing to reduce the uncertainty of schema matching. Since it is typical to ask si… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
50
0

Year Published

2014
2014
2023
2023

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 82 publications
(50 citation statements)
references
References 14 publications
0
50
0
Order By: Relevance
“…Multi-user scenarios include CrowdMap [19] for ontology matching, ZenCrowd [8] for entity linking, and Zhang et al [23] for database schema matching, which use crowdsourcing on a web platform.…”
Section: Multi-user Feedbackmentioning
confidence: 99%
“…Multi-user scenarios include CrowdMap [19] for ontology matching, ZenCrowd [8] for entity linking, and Zhang et al [23] for database schema matching, which use crowdsourcing on a web platform.…”
Section: Multi-user Feedbackmentioning
confidence: 99%
“…For example, in [22], the crowdsourcing is employed to validate the search results of automated image search on mobile devices. In [45], the authors leveraged the user CAPTCHAs inputs in web forms to recognize difficult words that cannot 14 Similar to our work, the authors of [46] also make use of crowdsourcing to validate the correspondences and reduce their uncertainty. However, they only focus on a pair-wise matching and using entropy-based decision strategy to maximize the uncertainty reduction at a single validation step.…”
Section: Crowdsourcingmentioning
confidence: 97%
“…Clearly, to be able to assign annotator j ∈ A to some task in her k-th iteration, she must have been assigned to some other k − 1 tasks in the past. We assume p jk to be known, as inferrable, e.g., from historical data or from the worker's performance on ad-hoc verification tasks, as, e.g., in [28,8]. Indeed, if we consider the perspective of a platform owner, it is reasonable to assume the availability of worker statistics for each task type supported by the platform, as done in [18].…”
Section: Preliminariesmentioning
confidence: 99%
“…However, differently from previous works, in this paper we study the problem of assigning annotators to tasks under the assumption that the annotators' reliability could change depending on their workload, as a result of, e.g., fatigue and learning. In this regard, a crowdsourcing platform may use all historic data and trends from previous sessions about its workers to provide a reliable estimate of the typical accuracy variations experienced by each annotator, as successfully done by several works in the literature [28,8,18]. These estimates allow us to devise task assignment policies whose attention to changes in annotators' accuracies proves extremely beneficial in terms of reduction of the classification error, as we will better illustrate in the remainder of the paper.…”
Section: Introductionmentioning
confidence: 99%