Counting is almost all you need

Front Immunol. 2023 Jan 20:13:1031011. doi: 10.3389/fimmu.2022.1031011. eCollection 2022.

Abstract

The immune memory repertoire encodes the history of present and past infections and immunological attributes of the individual. As such, multiple methods were proposed to use T-cell receptor (TCR) repertoires to detect disease history. We here show that the counting method outperforms two leading algorithms. We then show that the counting can be further improved using a novel attention model to weigh the different TCRs. The attention model is based on the projection of TCRs using a Variational AutoEncoder (VAE). Both counting and attention algorithms predict better than current leading algorithms whether the host had CMV and its HLA alleles. As an intermediate solution between the complex attention model and the very simple counting model, we propose a new Graph Convolutional Network approach that obtains the accuracy of the attention model and the simplicity of the counting model. The code for the models used in the paper is provided at: https://github.com/louzounlab/CountingIsAlmostAllYouNeed.

Keywords: T cells; attention; graphs; immune repertoire; immunology; machine learning; repertoire classification.

Publication types

  • Research Support, Non-U.S. Gov't

MeSH terms

  • Algorithms*
  • Immunologic Memory
  • Receptors, Antigen, T-Cell* / genetics

Substances

  • Receptors, Antigen, T-Cell

Grants and funding

This work was funded by ISF grant 20/870 and by the BIU DSI Grant.