2018
DOI: 10.1007/978-1-4939-8639-2_13
|View full text |Cite
|
Sign up to set email alerts
|

Selection of Informative Examples in Chemogenomic Datasets

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
16
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
7

Relationship

1
6

Authors

Journals

citations
Cited by 12 publications
(16 citation statements)
references
References 74 publications
0
16
0
Order By: Relevance
“…Chemogenomic active learning using random forest as the underlying estimator algorithm was employed similar to previous studies [15,16,34]. Based on previous studies [16,28], the number of trees to comprise a forest was set to 100.…”
Section: Methodsmentioning
confidence: 99%
“…Chemogenomic active learning using random forest as the underlying estimator algorithm was employed similar to previous studies [15,16,34]. Based on previous studies [16,28], the number of trees to comprise a forest was set to 100.…”
Section: Methodsmentioning
confidence: 99%
“…Many of the current, successful applications of AI such as image classification have required large training datasets, often in combination with data augmentation and model pre-training with weakly labeled data to capture the diversity of the input and obtain a model that generalizes well 46,47 . The number and nature of molecules required to build a predictive model for drug design is still undetermined 48,49 .…”
Section: Challenge 1: Generating and Obtaining Appropriate Datasetsmentioning
confidence: 99%
“…Indeed, not all data are equally valuable, and similar training entries may be considered redundant or uninformative from a machine-learning vantage point. As a consequence, more data does not necessarily translate into better predictive models; the performance of learning algorithms may actually improve if only informative, high-quality data are kept for training 102,103 . Therefore, active learning is primed to compress search spaces to the bare minimum, absorb relevant knowledge through fast feedback loops and design experiments 'on the fly' by leveraging lean models, i.e.…”
Section: Optimization Of Reaction Conditionsmentioning
confidence: 99%