Optimization of Molecules via Deep Reinforcement Learning
About
We present a framework, which we call Molecule Deep $Q$-Networks (MolDQN), for molecule optimization by combining domain knowledge of chemistry and state-of-the-art reinforcement learning techniques (double $Q$-learning and randomized value functions). We directly define modifications on molecules, thereby ensuring 100\% chemical validity. Further, we operate without pre-training on any dataset to avoid possible bias from the choice of that set. Inspired by problems faced during medicinal chemistry lead optimization, we extend our model with multi-objective reinforcement learning, which maximizes drug-likeness while maintaining similarity to the original molecule. We further show the path through chemical space to achieve optimization for a molecule to understand how the model works.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Receptor Docking Affinity | TDC DRD3 (leaderboard) | Affinity Score-7 | 48 | |
| Molecular Property Optimization | ZINC 250K | logP6.96 | 30 | |
| Multi-Objective Optimization | ZINC-250k (logP-TPSA) | Hypervolume1.78e+3 | 30 | |
| de novo molecular design | GuacaMol goal-directed tasks | Osimertinib MPO Score0.699 | 23 | |
| Molecule Design Optimization | TDC DRD3 (test) | Best Score-7.62 | 19 | |
| structure-based drug design | protein targets (Set B) | Uniqueness76.5 | 14 | |
| structure-based drug design | Cross-Docked 2020 57 (test) | TOP-100 Score-6.287 | 14 | |
| Molecular Optimization | GuacaMol | Median Score 10.188 | 7 |