2012
DOI: 10.1007/978-3-642-30284-8_15
|View full text |Cite
|
Sign up to set email alerts
|

Unsupervised Learning of Link Discovery Configuration

Abstract: Discovering links between overlapping datasets on the Web is generally realised through the use of fuzzy similarity measures. Configuring such measures is often a non-trivial task that depends on the domain, ontological schemas, and formatting conventions in data. Existing solutions either rely on the user's knowledge of the data and the domain or on the use of machine learning to discover these parameters based on training data. In this paper, we present a novel approach to tackle the issue of data linking wh… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
67
0
2

Year Published

2013
2013
2021
2021

Publication Types

Select...
6
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 77 publications
(69 citation statements)
references
References 14 publications
0
67
0
2
Order By: Relevance
“…Thus, learning link specifications can be combined with our approach. For example, specification templates can be used to seed genetic programming algorithms [10] such as to accelerate their convergence. In addition, knowing which template to use can help when choosing the right deterministic model (Boolean classifier, linear classifier) as well as its initialization for these models [8].…”
Section: Discussionmentioning
confidence: 99%
See 2 more Smart Citations
“…Thus, learning link specifications can be combined with our approach. For example, specification templates can be used to seed genetic programming algorithms [10] such as to accelerate their convergence. In addition, knowing which template to use can help when choosing the right deterministic model (Boolean classifier, linear classifier) as well as its initialization for these models [8].…”
Section: Discussionmentioning
confidence: 99%
“…In the case of link discovery, X = S × T while Y = {+1, −1} with f (x i ) = +1 if ρ(s, t) and f (x i ) = −1 in all other cases. Finding the function f for link discovery tasks is generally very costly, as it requires either (mostly manually) labeled training data [9] or a significant amount of computation [10]. The idea behind transfer learning (also coined knowledge transfer) [11] can be broadly described as follows: Given other machine learning tasks t with known or unknown classification functions f that are somehow "related" to f , use the functions f or the domain knowledge available for determining f (i.e., transfer the knowledge from the tasks t ) to improve the process of finding (f, Y).…”
Section: B Transfer Learningmentioning
confidence: 99%
See 1 more Smart Citation
“…These approaches are often adapted to some set of data. In some cases, this type of rule is automatically discovered [14,3,2,7]. Other approaches are based on logical rules [8,17] that are generated automatically using the semantics of the keys or functional properties.…”
Section: State Of the Artmentioning
confidence: 99%
“…Standard blocking approaches were implemented in the first versions of SILK and later replaced with MultiBlock [9], a lossless multidimensional blocking technique. KnoFuss [22] also implements blocking techniques to achieve acceptable runtimes. Further LD frameworks have been participated in the ontology alignment evaluation initiative [4].…”
Section: Related Workmentioning
confidence: 99%