2020
DOI: 10.1007/978-3-030-59833-4_1
|View full text |Cite
|
Sign up to set email alerts
|

The New DBpedia Release Cycle: Increasing Agility and Efficiency in Knowledge Extraction Workflows

Abstract: Since its inception in 2007, DBpedia has been constantly releasing open data in RDF, extracted from various Wikimedia projects using a complex software system called the DBpedia Information Extraction Framework (DIEF). For the past 12 years, the software received a plethora of extensions by the community, which positively affected the size and data quality. Due to the increase in size and complexity, the release process was facing huge delays (from 12 to 17 months cycle), thus impacting the agility of the deve… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 10 publications
(4 citation statements)
references
References 12 publications
0
4
0
Order By: Relevance
“…We started with DBpedia, a popular and large knowledge base comprising billions of triples extracted from Wikipedia, texts, and other sources [ 20 , 21 ]. We analyzed DBpedia to identify possible instances or subclasses for the HITO classes.…”
Section: Methodsmentioning
confidence: 99%
“…We started with DBpedia, a popular and large knowledge base comprising billions of triples extracted from Wikipedia, texts, and other sources [ 20 , 21 ]. We analyzed DBpedia to identify possible instances or subclasses for the HITO classes.…”
Section: Methodsmentioning
confidence: 99%
“…We run experiments on a server equipped with 96 cores and 900GB RAM running Ubuntu 18.04. All the resources used in the reported experimental study are publicly available 13 .…”
Section: ) Implementationmentioning
confidence: 99%
“…We choose the mentioned KGs because of: 1) the richness of the used KGs and how frequent they are updated. A new release for DBpedia is created every month [13]. Wikidata is updated daily with further contributions from the crowd 2 .…”
Section: Introductionmentioning
confidence: 99%
“…Our motivation comes from the growth and incompleteness nature of real-world KGs. For example, the release bot of DBpedia [13] extracts about 21 billion new triples per month [10], and Wikidata [30] releases data dumps in a weekly cycle. 4 The new entities and triples bring about new alignment to be found and provide new clues for correcting the previous alignment.…”
Section: Introductionmentioning
confidence: 99%