2018
DOI: 10.1101/467498
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Refining interaction search through signed iterative Random Forests

Abstract: Advances in supervised learning have enabled accurate prediction in biological systems governed by complex interactions among biomolecules. However, state-of-the-art predictive algorithms are typically "black-boxes," learning statistical interactions that are difficult to translate into testable hypotheses. The iterative Random Forest (iRF) algorithm took a step towards bridging this gap by providing a computationally tractable procedure to identify the stable, high-order feature interactions that drive the pr… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
34
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
6
1
1

Relationship

4
4

Authors

Journals

citations
Cited by 17 publications
(34 citation statements)
references
References 43 publications
0
34
0
Order By: Relevance
“…Interactions are important as ML models are often highly nonlinear and learn complex interactions between features. Methods exist to extract interactions from many ML models, including random forests (21,57,58) and neural networks (59,60). In the below example, the descriptive accuracy of random forests is increased by extracting Boolean interactions (a problem-relevant form of interpretation) from a trained model.…”
Section: Post Hoc Interpretabilitymentioning
confidence: 99%
See 1 more Smart Citation
“…Interactions are important as ML models are often highly nonlinear and learn complex interactions between features. Methods exist to extract interactions from many ML models, including random forests (21,57,58) and neural networks (59,60). In the below example, the descriptive accuracy of random forests is increased by extracting Boolean interactions (a problem-relevant form of interpretation) from a trained model.…”
Section: Post Hoc Interpretabilitymentioning
confidence: 99%
“…A previous line of work considers the problem of searching for biological interactions associated with important biological processes (21,57). To identify candidate biological interactions, the authors train a series of iteratively reweighted random forests (RFs) and search for stable combinations of features that frequently co-occur along the predictive RF decision paths.…”
Section: Post Hoc Interpretabilitymentioning
confidence: 99%
“…Dropout in neural networks is a form of algorithm perturbation that leverages stability to improve generalizability (43). Our previous work (34) stabilizes random forests to interpret decision rules in tree ensembles (34,44), which are perturbed using random feature selection (model perturbation) and bootstrap (data perturbation).…”
Section: D2 Data Perturbationmentioning
confidence: 99%
“…The iterative Random Forest algorithm (iRF), and corresponding iRF R package, take a step towards addressing these issues with a computationally tractable approach to search for important interactions in a fitted random forest ,Kumbier, Basu, Brown, Celniker, & Yu (2018). Our algorithm grows a series of feature weighted random forests (Amaratunga, Cabrera, & Lee, 2008) to perform soft regularization on the model based on predictive features.…”
Section: Discussionmentioning
confidence: 99%