这是indexloc提供的服务,不要输入任何密码

CodonBERT large language model for mRNA vaccines

  1. Sven Jager1
  1. 1Digital R&D, Sanofi, Cambridge, Massachusetts 02141, USA;
  2. 2mRNA Center of Excellence, Sanofi, Waltham, Massachusetts 02451, USA;
  3. 3mRNA Center of Excellence, Sanofi, 69280 Marcy L'Etoile, France
  1. 4 These authors contributed equally to this work.

  • Corresponding authors: zivbj{at}cs.cmu.edu, sven.jager{at}sanofi.com
  • Abstract

    mRNA-based vaccines and therapeutics are gaining popularity and usage across a wide range of conditions. One of the critical issues when designing such mRNAs is sequence optimization. Even small proteins or peptides can be encoded by an enormously large number of mRNAs. The actual mRNA sequence can have a large impact on several properties, including expression, stability, immunogenicity, and more. To enable the selection of an optimal sequence, we developed CodonBERT, a large language model (LLM) for mRNAs. Unlike prior models, CodonBERT uses codons as inputs, which enables it to learn better representations. CodonBERT was trained using more than 10 million mRNA sequences from a diverse set of organisms. The resulting model captures important biological concepts. CodonBERT can also be extended to perform prediction tasks for various mRNA properties. CodonBERT outperforms previous mRNA prediction methods, including on a new flu vaccine data set.

    Footnotes

    • Received December 15, 2023.
    • Accepted June 25, 2024.

    This article is distributed exclusively by Cold Spring Harbor Laboratory Press for the first six months after the full-issue publication date (see https://genome.cshlp.org/site/misc/terms.xhtml). After six months, it is available under a Creative Commons License (Attribution-NonCommercial 4.0 International), as described at http://creativecommons.org/licenses/by-nc/4.0/.

    | Table of Contents

    Preprint Server