2022
DOI: 10.1088/2632-2153/ac7ddc
|View full text |Cite
|
Sign up to set email alerts
|

Curiosity in exploring chemical spaces: intrinsic rewards for molecular reinforcement learning

Abstract: Computer aided design of molecules has the potential to disrupt the field of drug and material discovery. Machine learning and deep learning in particular, made big strides in recent years and promises to greatly benefit computer aided methods. Reinforcement learning is a particularly promising approach since it enables de novo molecule design, that is molecular design, without providing any prior knowledge. However, the search space is vast, and therefore any reinforcement learning agent needs to perform effi… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
32
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
7
2
1

Relationship

1
9

Authors

Journals

citations
Cited by 24 publications
(34 citation statements)
references
References 32 publications
0
32
0
Order By: Relevance
“…In drug discovery, experimental verification of molecular properties is very time-consuming and labor-intensive. Most DL generative models have not yet incorporated this challenge into the design process. ,, ,, , To make ChemistGA more aware of this constraint by design, we propose an augmented ChemistGA, called R-ChemistGA, which simulates the real environment as further elucidated below.…”
Section: Results and Discussionmentioning
confidence: 99%
“…In drug discovery, experimental verification of molecular properties is very time-consuming and labor-intensive. Most DL generative models have not yet incorporated this challenge into the design process. ,, ,, , To make ChemistGA more aware of this constraint by design, we propose an augmented ChemistGA, called R-ChemistGA, which simulates the real environment as further elucidated below.…”
Section: Results and Discussionmentioning
confidence: 99%
“…It then chooses actions that lead to situations it cannot predict well, thus maximizing its own understanding of the environment. It has been shown using curious agents in simulated virtual universes 124 and robot agents in real laboratories 84 that curiosity is an efficient exploration strategy. Alternative intrinsic rewards for artificial agents are ‘computational creativity’ 125 , 126 and ‘surprise’ 127 .…”
Section: Three Dimensions Of Computer-assisted Understandingmentioning
confidence: 99%
“…As previously mentioned, the challenging syntax of SMILES strings makes their construction difficult. For this reason, Thiede et al 91 worked with SELFIES strings, in which every combination of characters is valid and the substrings generated during the construction process can be directly interpreted. Optimisation is handled via PPO, while the reward function is defined by a combination of an extrinsic reward (based on the predicted properties of the molecule) and an intrinsic reward named curiosity to encourage increased exploration of the state space.…”
Section: Applications Of Reinforcement Learning In Chemistrymentioning
confidence: 99%