A central problem in computational biophysics is protein structure prediction, i.e., finding the optimal folding of a given amino acid sequence. This problem has been studied in a classical abstract model, the HP model, where the protein is modeled as a sequence of H (hydrophobic) and P (polar) amino acids on a lattice. The objective is to find conformations maximizing H-H contacts. It is known that even in this reduced setting, the problem is intractable (NP-hard). In this work, we apply deep reinforcement learning (DRL) to the two-dimensional HP model. We can obtain the conformations of best known energies for benchmark HP sequences with lengths from 20 to 50. Our DRL is based on a deep Q-network (DQN). We find that a DQN based on long short-term memory (LSTM) architecture greatly enhances the RL learning ability and significantly improves the search process. DRL can sample the state space efficiently, without the need of manual heuristics. Experimentally we show that it can find multiple distinct best-known solutions per trial. This study demonstrates the effectiveness of deep reinforcement learning in the HP model for protein folding.
翻译:计算生物物理的一个中心问题是蛋白质结构预测,即找到某个氨基酸序列的最佳折叠。这个问题已经在经典抽象模型HP模型中研究过,即HP模型,该模型的蛋白质以H(疏水)和P(polar)氨基酸的序列为模型,在薄饼上以H(湿)和P(P(polar)氨基酸为模型。目标是找到最大程度的H-H接触的符合性。众所周知,即使在这种降低的环境下,问题也是棘手的(NP-hard)。在这项工作中,我们对二维的HP模型进行深度强化学习(DRLL)。我们实验性地表明,在深度的Q网络上,我们能找到多种已知的HP序列的精密能量。我们发现,基于长期短期记忆(LSTM)结构的DQN可极大地提高RL学习能力,并大大改进搜索过程。DRL可以有效地对州空间进行取样,而不需要手动的超度。我们实验性地表明,在一次试验中,它能够找到多种已知的模型,在深度的模型中找到各种最明显的强化的模型。