2019
DOI: 10.3390/info10040137
|View full text |Cite
|
Sign up to set email alerts
|

Data Consistency Theory and Case Study for Scientific Big Data

Abstract: Big data technique is a series of novel technologies to deal with large amounts of data from various sources. Unfortunately, it is inevitable that the data from different sources conflict with each other from the aspects of format, semantics, and value. To solve the problem of conflicts, the paper proposes data consistency theory for scientific big data, including the basic concepts, properties, and quantitative evaluation method. Data consistency can be divided into different grades as complete consistency, s… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
9
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
8
1
1

Relationship

0
10

Authors

Journals

citations
Cited by 22 publications
(9 citation statements)
references
References 26 publications
0
9
0
Order By: Relevance
“…It has to be stressed that for the data export, the operating system used (Windows or Mac) was taken into account, through appropriate selections provided by the VOSviewer interface. An important task of the refinements relates to the consistency of the records in the databases [49] obtained in the previous stage, as part of the minimum data cleaning procedures (data cleaning Rahm and Do [50]). Because of the fact that the database was a simple one, with a single table, we noticed that this consistency refered, in principle, to the completion of all table attributes with complete and correct information.…”
Section: Refining By Filtering the Initial Resultsmentioning
confidence: 99%
“…It has to be stressed that for the data export, the operating system used (Windows or Mac) was taken into account, through appropriate selections provided by the VOSviewer interface. An important task of the refinements relates to the consistency of the records in the databases [49] obtained in the previous stage, as part of the minimum data cleaning procedures (data cleaning Rahm and Do [50]). Because of the fact that the database was a simple one, with a single table, we noticed that this consistency refered, in principle, to the completion of all table attributes with complete and correct information.…”
Section: Refining By Filtering the Initial Resultsmentioning
confidence: 99%
“…Therefore, more complete quality control steps and automatic DQ checking rules should be embedded throughout the entire scientific data manufacturing process. Taking the stage of data processing, for example, we can integrate scientific data from different sources with the help of metadata and standard terminology which could benefit the data consistency and interdisciplinary data sharing (Pasquetto et al , 2019; Shi et al , 2019). Therefore, for quality assurance, it is necessary for researchers to establish a complete DMP to assure scientific DQ before constructing an IP-Map.…”
Section: Strategies For Improving Scientific Data Qualitymentioning
confidence: 99%
“…Consistency of data refers to the concept that the same data stored in separate places or separate time points still match, meaning contradictory conlusions cannot be derived from the given data. 48 For example, can be ensured that archived/backedup/repository-deposited information can be kept up to date?…”
Section: Can Meta-information Offer Sufficient Contextual Information...mentioning
confidence: 99%