Interpretable Deep-Learning pK(a) Prediction for Small Molecule Drugs via Atomic Sensitivity Analysis

基于原子灵敏度分析的小分子药物可解释深度学习pK(a)预测

阅读:1

Abstract

Machine learning (ML) models now play a crucial role in predicting properties essential to drug development, such as a drug's logscale acid-dissociation constant (pK(a)). Despite recent architectural advances, these models often generalize poorly to novel compounds due to a scarcity of ground-truth data. Further, these models lack interpretability. To this end, with deliberate molecular embeddings, atomic-resolution information is accessible in chemical structures by observing the model response to atomic perturbations of an input molecule. Here, we present BCL-XpKa, a deep neural network (DNN)-based multitask classifier for pK(a) prediction that encodes local atomic environments through Mol2D descriptors. BCL-XpKa outputs a discrete distribution for each molecule, which stores the pK(a) prediction and the model's uncertainty for that molecule. BCL-XpKa generalizes well to novel small molecules. BCL-XpKa performs competitively with modern ML pK(a) predictors, outperforms several models in generalization tasks, and accurately models the effects of common molecular modifications on a molecule's ionizability. We then leverage BCL-XpKa's granular descriptor set and distribution-centered output through atomic sensitivity analysis (ASA), which decomposes a molecule's predicted pK(a) value into its respective atomic contributions without model retraining. ASA reveals that BCL-XpKa has implicitly learned high-resolution information about molecular substructures. We further demonstrate ASA's utility in structure preparation for protein-ligand docking by identifying ionization sites in 93.2% and 87.8% of complex small molecule acids and bases. We then applied ASA with BCL-XpKa to identify and optimize the physicochemical liabilities of a recently published KRAS-degrading PROTAC.

特别声明

1、本页面内容包含部分的内容是基于公开信息的合理引用;引用内容仅为补充信息,不代表本站立场。

2、若认为本页面引用内容涉及侵权,请及时与本站联系,我们将第一时间处理。

3、其他媒体/个人如需使用本页面原创内容,需注明“来源:[生知库]”并获得授权;使用引用内容的,需自行联系原作者获得许可。

4、投稿及合作请联系:info@biocloudy.com。