Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Robust Optimization in Protein Fitness Landscapes Using Reinforcement Learning in Latent Space (2405.18986v1)

Published 29 May 2024 in cs.LG, q-bio.BM, and q-bio.QM

Abstract: Proteins are complex molecules responsible for different functions in nature. Enhancing the functionality of proteins and cellular fitness can significantly impact various industries. However, protein optimization using computational methods remains challenging, especially when starting from low-fitness sequences. We propose LatProtRL, an optimization method to efficiently traverse a latent space learned by an encoder-decoder leveraging a large protein LLM. To escape local optima, our optimization is modeled as a Markov decision process using reinforcement learning acting directly in latent space. We evaluate our approach on two important fitness optimization tasks, demonstrating its ability to achieve comparable or superior fitness over baseline methods. Our findings and in vitro evaluation show that the generated sequences can reach high-fitness regions, suggesting a substantial potential of LatProtRL in lab-in-the-loop scenarios.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (39)
  1. Protein structure and sequence generation with equivariant denoising diffusion probabilistic models. arXiv preprint arXiv:2205.15019, 2022.
  2. Model-based reinforcement learning for biological sequence design. In International conference on learning representations, 2019.
  3. Biological sequences design using batched bayesian optimization. 2019.
  4. Proteinbert: A universal deep-learning model of protein sequence and function. Bioinformatics, 38(8):2102–2110, 2022.
  5. Conditioning by adaptive sampling for robust design. In International conference on machine learning, pp.  773–782. PMLR, 2019.
  6. Deep diversification of an aav capsid protein by machine learning. Nature Biotechnology, 39(6):691–696, 2021.
  7. Relso: a transformer-based model for latent space optimization and generation of proteins. arXiv preprint arXiv:2201.09948, 2022.
  8. Flip: Benchmark tasks in fitness landscape inference for proteins. bioRxiv, pp.  2021–11, 2021.
  9. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805, 2018.
  10. First return, then explore. Nature, 590(7847):580–586, 2021.
  11. Protein discovery with discrete walk-jump sampling. arXiv preprint arXiv:2306.12360, 2023.
  12. Automatic chemical design using a data-driven continuous representation of molecules. ACS central science, 4(2):268–276, 2018.
  13. Gaussian error linear units (gelus). arXiv preprint arXiv:1606.08415, 2016.
  14. Stable baselines. https://github.com/hill-a/stable-baselines, 2018.
  15. The coming of age of de novo protein design. Nature, 537(7620):320–327, 2016.
  16. Biological sequence design with gflownets. In International Conference on Machine Learning, pp.  9786–9801. PMLR, 2022.
  17. Highly accurate protein structure prediction with alphafold. Nature, 596(7873):583–589, 2021.
  18. The nk model of rugged fitness landscapes and its application to maturation of the immune response. Journal of theoretical biology, 141(2):211–245, 1989.
  19. Bootstrapped training of score-conditioned generator for offline design of biological sequences. arXiv preprint arXiv:2306.03111, 2023.
  20. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
  21. Improving protein optimization with smoothed fitness landscapes. In The Twelfth International Conference on Learning Representations, 2023.
  22. Kruskal, J. B. Nonmetric multidimensional scaling: a numerical method. Psychometrika, 29(2):115–129, 1964.
  23. Language models of protein sequences at the scale of evolution enable accurate structure prediction. bioRxiv, 2022.
  24. Tranception: protein fitness prediction with autoregressive transformers and inference-time retrieval. In International Conference on Machine Learning, pp.  16990–17017. PMLR, 2022.
  25. Evaluating protein transfer learning with tape. Advances in neural information processing systems, 32, 2019.
  26. Proximal exploration for model-guided protein sequence design. In International Conference on Machine Learning, pp.  18520–18536. PMLR, 2022.
  27. Biological structure and function emerge from scaling unsupervised learning to 250 million protein sequences. Proceedings of the National Academy of Sciences, 118(15):e2016239118, 2021.
  28. Navigating the protein fitness landscape with gaussian processes. Proceedings of the National Academy of Sciences, 110(3):E193–E201, 2013.
  29. Local fitness landscape of the green fluorescent protein. Nature, 533(7603):397–401, 2016.
  30. Prediction of designer-recombinases for dna editing with generative deep learning. Nature Communications, 13(1):7966, 2022.
  31. Proximal policy optimization algorithms. arXiv preprint arXiv:1707.06347, 2017.
  32. Adalead: A simple and robust adaptive greedy search algorithm for sequence design. arXiv preprint arXiv:2010.02141, 2020.
  33. Accelerating bayesian optimization for biological sequence design with denoising autoencoders. arXiv preprint arXiv:2203.12742, 2022.
  34. Amortized bayesian optimization over discrete spaces. In Conference on Uncertainty in Artificial Intelligence, pp.  769–778. PMLR, 2020.
  35. Quantitative analyses of empirical fitness landscapes. Journal of Statistical Mechanics: Theory and Experiment, 2013(01):P01005, 2013.
  36. Black-box optimization for automated discovery. Accounts of Chemical Research, 54(6):1334–1346, 2021.
  37. Self-play reinforcement learning guides protein engineering. Nature Machine Intelligence, 5(8):845–860, 2023.
  38. [cls] token is all you need for zero-shot semantic segmentation. arXiv preprint arXiv:2304.06212, 2023.
  39. Machine-learning-guided directed evolution for protein engineering. Nature methods, 16(8):687–694, 2019.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Minji Lee (37 papers)
  2. Luiz Felipe Vecchietti (9 papers)
  3. Hyunkyu Jung (2 papers)
  4. Hyun Joo Ro (1 paper)
  5. Meeyoung Cha (63 papers)
  6. Ho Min Kim (3 papers)
Citations (1)

Summary

  • The paper introduces LatProtRL, a novel reinforcement learning framework that formulates protein optimization as a Markov Decision Process in latent space.
  • It employs an encoder-decoder model with pre-trained protein representations to efficiently explore vast combinatorial sequence spaces and enable multi-mutation optimizations.
  • Results on GFP and AAV benchmarks demonstrate that LatProtRL outperforms traditional evolutionary and Bayesian methods by achieving superior protein fitness.

Robust Optimization in Protein Fitness Landscapes Using Reinforcement Learning in Latent Space

The paper "Robust Optimization in Protein Fitness Landscapes Using Reinforcement Learning in Latent Space" presents a novel approach to protein optimization, a critical task with significant applications in biotechnology and therapeutics. The challenge tackled in this research is optimizing protein sequences to enhance desired functionalities, typically initiated from low-fitness sequences. This issue is compounded by the vast combinatorial space of potential protein sequences, making exhaustive experimentation impractical.

Methodology

The authors introduce a method named LatProtRL, which leverages reinforcement learning (RL) to navigate the latent space of protein sequences. LatProtRL's central innovation lies in its formulation of protein optimization as a Markov Decision Process (MDP), enabling the exploitation of RL techniques to escape local optima within the protein fitness landscape. This approach significantly differs from traditional sequence-based optimizations by mapping protein sequences into a latent representation space learned by an encoder-decoder model. The encoder-decoder model is based on large-scale pre-trained protein LLMs that generate compact, informative representations of protein sequences.

Key Components:

  • State Representation: Protein sequences are encoded into a low-dimensional latent space. This compression allows the RL agent to focus on meaningful mutations rather than exploring an unmanageably large sequence space.
  • Action Space: Instead of direct sequence mutations, actions are modeled as perturbations in the latent space. This enables the exploration of variants requiring multiple simultaneous mutations, facilitating more efficient traversal of the fitness landscape.
  • Optimization Loop: The method employs an on-policy RL algorithm, Proximal Policy Optimization (PPO), to train the policy that perturbs the latent representation, simulating evolutionary steps to maximize protein fitness.

Evaluation

LatProtRL's efficacy is demonstrated on two well-studied proteins: the Green Fluorescent Protein (GFP) and the adeno-associated virus (AAV) capsid proteins. These proteins serve as benchmarks due to the availability of extensive experimental data. The results show that LatProtRL can achieve fitness levels comparable or superior to baseline methods, including evolutionary strategies and Bayesian optimization techniques.

  • Numerical Results: In the GFP medium and hard tasks, LatProtRL significantly outperforms other methods in generating high-fitness sequences, with demonstrated ability to reach high-fitness regions that are proximal to experimentally validated high-function sequences.

Theoretical and Practical Implications

The research offers a compelling argument for using latent space exploration over traditional sequence-based methods in protein design. The decoupling of representation learning and optimization provides scalability and adaptability, which can be crucial for other biological systems with large, rugged fitness landscapes.

Theoretical Implications:

  • The latent space approach aligns with the increasing interest in representation learning, emphasizing the separation of feature extraction from downstream tasks, which could inspire similar methodologies in other domains of computational biology and beyond.

Practical Implications:

  • LatProtRL can facilitate the targeted evolution of proteins, offering a substantial potential in applications like enzyme engineering, drug design, and synthetic biology where lab-in-the-loop scenarios are becoming increasingly critical.

Future Directions

Future work could explore integrating structural prediction feedback (e.g., from tools like AlphaFold) into the RL loop, potentially improving the biological plausibility of proposed designs. Further enhancements to the encoder-decoder architecture could also be explored to boost decoding accuracy, addressing current limitations when dealing with longer protein sequences or those with insertions/deletions.

Overall, this paper provides an innovative approach to protein sequence optimization that efficiently navigates complex fitness landscapes, representing a meaningful step in computational biology applications aimed at harnessing evolutionary strategies through machine learning techniques.