2024
DOI: 10.1037/met0000526
|View full text |Cite
|
Sign up to set email alerts
|

Using synthetic data to improve the reproducibility of statistical results in psychological research.

Abstract: In recent years, psychological research has faced a credibility crisis, and open data are often regarded as an important step toward a more reproducible psychological science. However, privacy concerns are among the main reasons that prevent data sharing. Synthetic data procedures, which are based on the multiple imputation (MI) approach to missing data, can be used to replace sensitive data with simulated values, which can be analyzed in place of the original data. One crucial requirement of this approach is … Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
12
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 12 publications
(12 citation statements)
references
References 66 publications
0
12
0
Order By: Relevance
“…Sixth, such publications should adhere to open sciences practices. Sophisticated methods of de‐identification and synthetic data procedures have been proposed that address legitimate privacy concerns while ensuring the ability to verify results (Grund et al, 2022; Walsh et al, 2018). If more studies meeting these requirements were available, a more dependable verdict concerning the predictive validity of self‐report measures of personality and its moderators were possible.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…Sixth, such publications should adhere to open sciences practices. Sophisticated methods of de‐identification and synthetic data procedures have been proposed that address legitimate privacy concerns while ensuring the ability to verify results (Grund et al, 2022; Walsh et al, 2018). If more studies meeting these requirements were available, a more dependable verdict concerning the predictive validity of self‐report measures of personality and its moderators were possible.…”
Section: Discussionmentioning
confidence: 99%
“…that address legitimate privacy concerns while ensuring the ability to verify results (Grund et al, 2022;Walsh et al, 2018). If more studies meeting these requirements were available, a more dependable verdict concerning the predictive validity of self-report measures of personality and its moderators were possible.…”
Section: Limitations and Future Directionsmentioning
confidence: 99%
“…The replication code for the empirical application can be found in Supplement S2 at https://osf.io/n5zm6/?view_only=086ea651bbea49bb8b2aae44e3971db8. As the original data can only be downloaded by individual researchers from the OECD websites and we are not allowed to make our processed PISA data publicly available, we created synthetic datasets (see Grund et al, 2022) that are very similar to the original datasets. The generation of the synthetic datasets is described in Appendix A.…”
Section: Methodsmentioning
confidence: 99%
“…To enable the replicability of the results of the empirical example by independent researchers, we created synthetic Datasets for the two PISA assessments that strongly resemble the original data. The principle of data generation relied on the approach of Jiang et al (2022), which was also investigated by Grund et al (2022). Synthetic datasets were produced at the country level for each of the three cognitive domains and the two PISA assessments, PISA 2006 andPISA 2009. To also represent the balanced incomplete block design in the synthetic datasets, we applied the synthesis model at the level of each administered booklet in the test.…”
Section: Appendix Appendix A: Generation Of Synthetic Data In the Emp...mentioning
confidence: 99%
See 1 more Smart Citation