2016
DOI: 10.2139/ssrn.2805205
|View full text |Cite
|
Sign up to set email alerts
|

What Predicts Stroop Performance? A Conditional Random Forest Approach

Abstract: An experimental science relies on solid and replicable results. The last few years have seen a rich discussion on the reliability and validity of psychological science and whether our experimental findings can falsify our existing theoretical models. Yet, concerns have also arisen that this movement may impede new theoretical developments. In this article, we reanalyze the data from a crowdsourced replication project that concluded that lab site did not matter as predictor for Stroop performance, and, therefor… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2016
2016
2023
2023

Publication Types

Select...
6

Relationship

4
2

Authors

Journals

citations
Cited by 7 publications
(3 citation statements)
references
References 51 publications
0
3
0
Order By: Relevance
“…When calculating RF models, three important parameters must be specified which can impact the stability of the results 70,71 : the number of trees (ntree), the number of predictors randomly selected from all predictors in the model used in each tree (mtry), and the computational starting point for the randomization (seed). Consistent with standard recommendations, we set ntree to 1000 [71][72][73] and mtry to one third of the total numbers of predictors (in our case with 8 predictors, this would be 3; 71,74 ) or the square root of the total number of predictors (in our case 3; 30,73,75 ). We thus calculated five different models with five different mtry parameters (mtry = 3-7) in order to analyze the stability of the model.…”
Section: Log Transformation Of Time Variablesmentioning
confidence: 99%
“…When calculating RF models, three important parameters must be specified which can impact the stability of the results 70,71 : the number of trees (ntree), the number of predictors randomly selected from all predictors in the model used in each tree (mtry), and the computational starting point for the randomization (seed). Consistent with standard recommendations, we set ntree to 1000 [71][72][73] and mtry to one third of the total numbers of predictors (in our case with 8 predictors, this would be 3; 71,74 ) or the square root of the total number of predictors (in our case 3; 30,73,75 ). We thus calculated five different models with five different mtry parameters (mtry = 3-7) in order to analyze the stability of the model.…”
Section: Log Transformation Of Time Variablesmentioning
confidence: 99%
“…As a last step, we analyzed the relationship between these variables using structural equation modeling to derive a model that is based on both theory and empirical data. Using the conditional random forest method for selecting variables to include in a model is beneficial because it (1) prevents overfitting of models (i.e., mistaking noise in the data for the real signal, see IJzerman et al, 2016) and (2) has much less problems with collinearity than traditional variable selection procedures (IJzerman et al, 2018).…”
Section: The Current Studiesmentioning
confidence: 99%
“…The outcome in the case of conditional random forests is a variable importance list. The importance list allows us to identify which are the best predictors of the variable of interest and which of the computed variables differ from random noise when predicting the variable of interest (see also [ 68 ]). In our case, the random forest allows us to select the variables to be included in the mediation analyses in the exploratory phase.…”
Section: Studymentioning
confidence: 99%