2020
DOI: 10.1021/acs.jpca.0c00042
|View full text |Cite
|
Sign up to set email alerts
|

Simpler is Better: How Linear Prediction Tasks Improve Transfer Learning in Chemical Autoencoders

Abstract: Transfer learning is a subfield of machine learning that leverages proficiency in one or more prediction tasks to improve proficiency in a related task. For chemical property prediction, transfer learning models represent a promising approach for addressing the data scarcity limitations of many properties by utilizing potentially abundant data from one or more adjacent applications. Transfer learning models typically utilize a latent variable that is common to several prediction tasks and provides a mechanism … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

2
25
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
1
1

Relationship

1
5

Authors

Journals

citations
Cited by 27 publications
(27 citation statements)
references
References 33 publications
2
25
0
Order By: Relevance
“…In contrast to earlier joint training approaches that were not designed for scarce data, [25,29] or are based on pretraining using chemical properties and a multitask transfer learning approach for target properties that are scarce, [27,28] the current multi‐class approach finds a joint embedding with manual feature vectors which therefore finds the latent space using structure information instead of properties. The augmentation is thus done using the low‐level structural information while the properties between the small and large data sets remain distinct and possibly non‐overlapping.…”
Section: Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…In contrast to earlier joint training approaches that were not designed for scarce data, [25,29] or are based on pretraining using chemical properties and a multitask transfer learning approach for target properties that are scarce, [27,28] the current multi‐class approach finds a joint embedding with manual feature vectors which therefore finds the latent space using structure information instead of properties. The augmentation is thus done using the low‐level structural information while the properties between the small and large data sets remain distinct and possibly non‐overlapping.…”
Section: Methodsmentioning
confidence: 99%
“…It has been shown capable of learning underlying relationships from a diverse set of molecular data by letting multiple data tasks and domains interact adaptively while generating the joint embeddings. Joint training is also a type of joint embedding that has been successfully applied to improve deep learning‐based molecule generation and enable transfer learning [25–29] . Joint training incorporates a property prediction task into a variational autoencoder (VAE) [30] and has been shown to organize points in the VAE latent space, making the latent space amenable to inverse molecular design and optimization [25,29] .…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…We utilize the grammar variational autoencoder 37 (GVAE) to achieve generation of molecules with targeted properties, with the alteration of using a single linear predictor layer so that properties tend to vary linearly along the latent dimensions. 38 Training was conducted using the RMSprop algorithm with a learning rate of 0.001, which was set to decay by a factor of 0.3 in the case of a plateau in the validation loss. KL prediction task based on a linear prediction network ensures that those properties will vary linearly along the principal components of the latent encodings.…”
Section: Methodsmentioning
confidence: 99%
“…KL prediction task based on a linear prediction network ensures that those properties will vary linearly along the principal components of the latent encodings. 38 Compounds with specific properties can then be generated by targeting regions of the latent space based on univariate linear regression between the property of interest and the position along one of the principle components. In the case of multi-property models, correlation between the properties may lead to latent space organization not being exactly orthogonal.…”
Section: Methodsmentioning
confidence: 99%