Abstract
Accurate identification of human leukocyte antigen (HLA) alleles is essential for various clinical and research applications, such as transplant matching and drug sensitivities. Recent advances in RNA-seq technology have made it possible to impute HLA types from high throughput sequencing data, spurring the development of a large number of computational HLA typing tools. However, the relative performance of these tools is unknown, limiting the ability for clinical and biomedical research to make informed choices regarding which tools to use. Here, we rigorously compare the performance of 9 HLA callers on 652 RNA-seq samples across 6 datasets with molecularly defined gold standard. We find that OptiType has the highest accuracy at both low and high resolution with an accuracy above 99%, followed by arcasHLA and seq2HLA with accuracies above 96%. Despite OptiType’s high accuracy, it is only capable of Class I predictions, thereby limiting its application to clinical procedures like transplantation requiring Class II predictions. Furthermore, our findings reveal significant variation in accuracy across each HLA locus, with HLA-A exhibiting the highest accuracy and HLA-DRB1 exhibiting the lowest accuracy. We also find that class II genes are generally more challenging to impute than class I genes, with most typing algorithms capable of making Class I predictions to >97% accuracy whereas the best Class II tool predicts with 94.2% accuracy. Moreover, we identify notable differences in the computational resources necessary to run each tool. We find that the most computationally expensive tools are OptiType and HLA-HD which require 105 and 102 times greater RAM and CPU, respectively, than the least computationally expensive tools, seq2HLA and RNA2HLA. Furthermore, all tools have decreased accuracy for African samples with respect to European samples at four digit resolution. We conclude that RNA-Seq HLA callers are capable of returning high-quality results, but the tools that offer a good balance between accuracy, consistency, and computational expensiveness are yet to be developed.
Introduction
The Human Leukocyte Antigen (HLA) complex, otherwise known as the human Major Histocompatibility Complex (MHC), is a 3.6 Mb segment of the human genome which encodes for cell surface glycoproteins involved in cell-mediated immunity and self/nonself recognition. There are two classes of the HLA system typically implicated in immunity: Class I, which includes the loci HLA-A, HLA-B, and HLA-C and Class II, which includes the loci HLA-DRB1 and HLA-DQB1 (Fig 1a). These five loci comprise the Classical HLA loci and are the most well documented in the entire region.
The HLA region is the most polymorphic region of the human genome with over 25,000 HLA alleles currently available in the Immuno Polymorphism Database ImMunoGeneTics (IPD-IMGT) HLA reference database1, 2. The high polymorphism within the locus has the biological function of allowing each individual to produce a dramatically distinct immune response. Therefore, accurate identification of HLA haplotypes is extremely relevant to immunology and the biomedical sciences. For instance, HLA typing is widely used in vaccinology3 and transplant medicine including organ and hematopoietic stem cell transplantations4. Furthermore, certain HLA variants are associated with pathogenesis of both infectious and autoimmune diseases5, 6, adverse reactions to certain drugs7, and cancer development8. Recent studies have also found that some HLA variations are associated with susceptibility, progression, and severity of SARS-CoV-2 infection9. US standards require an 8/8 match for all alleles within the HLA-A, HLA-B, HLA-C, and HLA-DQB1 for transplant donor selection, and most European standards require a more stringent 10/10 match for HLA-A, HLA-B, HLA-C, HLA-DRB1, and HLA-DQB1 alleles. Whether a donor’s HLA type constitutes an “HLA match” depends on the resolution of interest, with the most frequently used resolutions being either two-digit resolution (low resolution) or four-digit resolution (high resolution). High resolution typing is recommended by the National Marrow Donors Program and hence is of considerable clinical significance10.
Established HLA typing methods involve laboratory-based techniques such as amplification and quantification of HLA genes via sequence-specific oligonucleotide probes (SSOP) or sequence-specific primers (SSP), or via PCR followed by Sanger sequencing and comparison against an HLA reference database (SBT). SBT is the current gold standard method of typing, both it is still cost prohibitive and laborious to be effective for large-scale clinical and research applications9. More recently, developments in accessibility and throughput of NGS technology have spurred the development of HLA callers that impute HLA types from short read data. They use algorithms that can be divided into two general categories, alignment-based methods and assembly methods. Alignment-based methods align reads to an HLA reference containing a set of known allele sequences, then use that information to either build a probabilistic or optimization model to impute the HLA type. The less common assembly method links fragments into contigs which are then aligned to the reference genome to make the imputation of HLA type.
Despite the plethora of recently developed HLA callers, there is a lack of comprehensive and systematic benchmarking of RNA-seq-based HLA callers using large-scale and realist gold standards, limiting the ability for clinical and biomedical research to make informed choices regarding which tools to use. Furthermore, most existing benchmarking papers focus solely on whole-genome (WGS) or whole-exome (WES) data11–14, with the only existing RNA-seq study published seven years ago15. Since the previous benchmarking study, seven new HLA callers have been developed, and the accuracy of HLA callers on varying read length and ancestry is yet to be evaluated. To address these limitations, we evaluated the performance of 9 HLA callers across 652 RNA-seq samples with available gold standard HLA alleles. The tools we included in this analysis are HLAminer16, seq2HLA17, HLAforest18, PHLAT19, OptiType20, HLA-VBseq21, HLA-HD22, arcasHLA23, and RNA2HLA24, of which 4 are novel and not considered in the previous benchmarking study. All tools are capable of making predictions at all Class I and Class II classical HLA loci, with the exception of OptiType20, which only makes Class I predictions. In each case, we produced evaluation metrics of accuracy that is the percentage of correctly predicted alleles to determine the best performing tool. We also conducted comparisons of the performance of each HLA caller across the various HLA genes to detect the genes that are consistently miscalled by HLA calling methods. Furthermore, we examined the impact of sequencing parameters on the performance of each HLA caller and thoroughly assessed the computational resources required for each software. Additionally, we compared the performance of HLA methods across different ancestry groups and identified tools and trends in prediction accuracy and quality across samples of European and African ancestry. In summary, we conducted an extensive evaluation of all publicly accessible HLA callers, revealing valuable insights for researchers and clinicians to make an informed decision selecting the most appropriate RNA-Seq based HLA caller, offering a balance of consistency, performance, and scalability for their clinical and scientific needs.
Methods
RNA-Seq Data
We retrieved data from 6 publicly available RNA-seq datasets (ERP00194225, SRP29870426, ERP00010127, SRP09956828, SRP19849729, SRP09632930) to assess the performance of the tools, as detailed in Table 1. From the datasets we selected only the samples containing ground truth HLA types to be used for this study, amounting to a total of 652 samples collected from lymphoblastoid cell lines (LCLs) (n=552) and peripheral blood mononuclear cells (PMBCs) (n=100). Altogether, these datasets represent a diverse set of samples spanning 5 1000-genomes project populations (Northern Europeans in Utah, Finnish, British in England and Scotland, Tuscan in Italy, Yoruba), 4 sequencing platforms (Illumina HiSeq 2000, Illumina HiSeq 2500, Illumina Genome Analyzer II, Illumina NextSeq 550), read counts between 24.5 million and 337.8 million reads, and read lengths ranging from 36bp to 126bp.
We used the SRA-Toolkit v2.11.0 to download all samples as fastq files from the NCBI SRA archive, using the prefetch and fasterq-dump commands:
prefetch --option-file accession_list.txt, to download all samples in SRA format, where ‘accession_list.txt’ contains a list of the accession numbers of all samples separated by newlines (\n)
fasterq-dump --split-files sample.sra” on each sample (‘sample.sra’), to convert from sra to FASTQ format
We used STAR aligner v2.7.0e to align reads to the GRCh38 reference using the command ““STAR -- runThreadN 8 --genomeDir /path/to/genomeIndicesFile --readFilesIn path/to/sample1.fastq path/to/sample2.fastq --outSAMtype BAM SortedByCoordinate --outFileNamePrefix /path/to/outDir/prefix”. The ‘--genomeDir’ argument specifies the location of the file containing the genome indices, and ‘-- outSAMtype BAM SortedByCoordinate’ indicates the output files to be sorted BAM files. From the aligned BAM files, we extracted the chromosome 6 regions using Samtools v1.19 commands, with default parameters unless otherwise specified. We first used the Samtools index function to produce indexed BAI files from the BAM-format samples for fast random access. Then we then invoked Samtools view with default parameters to extract the reads aligning to chromosome 6 into output files. Finally, we invoked ‘samtools view’ with -b argument on each of the output files to convert them from SAM to BAM format. Finally, Bedtools v2.27.1 bamtofastq command was used to convert the BAM files to FASTQ files. After these steps we produced a pair of FASTQ samples for each of the 652 samples. These data served as the final inputs for all HLA callers.
Gold Standard HLA types
The gold standard HLA types of samples were determined via laboratory based methods summarized in Table 1. For dataset 1, HLA types were found via PCR amplification of the HLA-associated genes followed by Sanger sequencing-based typing of the exons (SBT method). Dataset 2 and 4’s HLA types were found using PCR-based sequence-specific primers (SSP method) which amplifies the specific regions of DNA using primers that are designed specifically to bind to and amplify particular alleles of HLA genes. Dataset 3’s HLA types were found using PCR-based sequence-specific oligonucleotide probes (SSOP method) which uses PCR to amplify a specific region of the HLA gene, and then uses sequence-specific oligonucleotide probes that will detect and differentiate specific alleles or variants in that region based on their complementary binding. Dataset 5 and 6 were created from cells from the HLA Class I-deficient B721.221 cell line transduced with one or two Class I HLA allele-expressing cDNA vectors.
We formatted the gold standard HLA types for each dataset into a standard format csv where rows represent the samples and columns represent the loci(s) of interest (HLA-A, HLA-B, HLA-C, HLA-DRB1, HLA-DQB1) for which gold standard allele data are available. The csv files have been made available at https://github.com/ydottie/hlaproject.
HLA definition and nomenclature
To specify each HLA allele type, we used the HLA nomenclature system defined by the World Health Organization Committee for Factors of the HLA System in 2010 (Fig 1b). This system uses a four-part naming system: the HLA Prefix, the gene, the field groups, and the suffix letter as shown in Figure 1b. Every allele begins with the HLA Prefix, “HLA.” Following the HLA prefix is the gene identifier, which denotes the gene or locus of the allele. Then lies the field groups which are divided into 4 fields of increasing resolution, separated by colons. Field 1 represents the allele group, or specifically what serological antigen type is present for this particular allele. Field 2, which is typically 2 or 3 digits, represents the specific HLA protein. Field 3 represents allele variants with synonymous or intronic differences. Field 4 designates whether the allele is a single or multiple nucleotide polymorphism in the non-coding region of the gene. Finally, the suffix letter denotes differences in expression of the allele as described in Table S3. Two-digit resolution is denoted by the HLA prefix, the locus, and the first field groups (i.e., HLA-A*02). Four-digit resolution is the clinically relevant resolution and is denoted by the HLA prefix, the locus, and the first 2 field groups (i.e., HLA-A*02:10). Two-digit resolution may also be referred to as low resolution, and four-digit resolution may be referred to as high resolution or clinically relevant resolution. For the purposes of this benchmarking study, we assessed performance at the two-digit and four-digit resolutions.
HLA caller choice and installation
We benchmarked all existing alignment-based HLA callers capable of imputing HLA types from RNA-seq data. With these prerequisites, we identified 12 tools to be included in this study, namely HLAminer16, seq2HLA17, HLAforest18, PHLAT19, OptiType20, HLA-VBseq21, HLA-HD22, arcasHLA23, RNA2HLA24, HLApersa31, HLAProfiler32, and HISAT-genotype18 (Table 2). We downloaded each tool from their respective repository onto the USC CARC Discovery Cluster, closely following all available documentation and recommendations for each tool. If testing suites were provided, we ran the tools on the tests and manually verified that the tools successfully completed the tests. Furthermore, we emailed the developers of each tool to verify our process of installation and for recommendations on best practices for installation and running. During the installation process we faced unresolvable issues in 3 tools (HLApersa31, HLAProfiler32, and HISAT-genotype18), and hence excluded these tools from the study. The errors can be found in Table S1.
Run HLA profiling methods on RNA-Seq samples
After installation, we ran each HLA caller by submitting bash scripted jobs to the SLURM job scheduler on the USC CARC Discovery cluster. The commands used to run each tool are summarized in Table 3. Although some tools are capable of imputation at greater resolutions than 4 digits, we considered only the 2-digit (low) and 4-digit (high) resolution due to the limitation that our datasets contain only gold standard types up to 4-digit resolution. We parsed each tool’s custom output files into the standard csv format which parallels the format of the gold standard datasets, with each row representing a different sample and each column representing a different loci (see “Gold Standard HLA types” subsection). We considered lack of prediction as an incorrect prediction.
Evaluating HLA predictions for accuracy
We evaluated the accuracy of each HLA tool in comparison to the gold standard using the following equation:
We assessed accuracy at 2 digit and 4 digit resolutions, with 2-digit resolution considering only the section of the HLA allele up to the first field and 4-digit resolution considering the section of the allele up to the second field (Figure 1). The numerator “number of alleles correctly predicted” is determined via a locus-by-locus comparison of the prediction with the gold standard. Because each locus contains 2 independent haplotypes, our algorithm assesses the 2 and 4 digit accuracy of both a parallel and cross-wise comparison between the 2 alleles of the gold standard and the prediction. Between the parallel and cross-wise comparison, we selected the option that maximizes both the 2 and 4 digit accuracies. We treated no-calls as missed-calls, with the exception of OptiType, in which all Class II loci were ignored. Furthermore, we stratified accuracy by class, locus and allele to examine for variations. We also examined the presence of specific allele haplotypes that are commonly mispredicted as a specific allele, followed by pairwise alignment of the aligned sequences with the hypothesis that these “frequent pairs” are the result of high sequence similarity.
Computational modification of RNA-Seq sample properties
We are currently working on using Seqtk v1.3 to produce simulated RNA-seq samples with truncated read lengths. We will perform this computational modification for 10 samples from each of datasets 1-4, and all samples of datasets 5-6 because these datasets contain less than 10 samples, for a total of 52 samples. So far, we have produced simulated samples for 12 samples only, but plan to complete the rest of the samples in the future. We first converted each sample from FASTQ to FASTA format using the command seqtk seq -a in.fq > out.fa. For each sample, we then iteratively invoked the seqtk trimfq -b 5 in.fa > out.fa, such that each iteration truncates 5 base pairs from the right of every read, stopping when reads become shorter than 36 bp. Each original sample is hence used as a template for multiple simulated samples, each with read lengths ranging from 36bp to the original read length, with a step size of 5bp. The final files were converted from fa back to fq format using the seqtk seq -a in.fa > out.fq command. All simulated samples were ran on the HLA caller tools with default and optimal parameters for accuracy analysis.
For analysis of coverage effect on accuracy (in progress), we will computationally modify the coverage via randomly selecting a subset of reads from our original RNA-seq samples to create new, lower-coverage samples. We will quantify coverage as a percentage of the original read count, subsampling from a coverage of 20% of the original read count to 100% of the read count with a step size of 5% of the read count for each sample. At each percentage level, we will use the seqtk subsample function to produce simulated lower coverage samples, ensuring that we use the same seed for both files from each paired- end sample. All samples will then be run on the HLA caller tools for accuracy analysis.
Assessing the effect of ancestral diversity on prediction quality
We used all samples in Dataset 1 (n=490) in order to evaluate differences in performances between European and African ancestry groups. For each caller, we determined the accuracy at each HLA locus separately for the European subset (n=423) and the African subset (n=67) of dataset 1, both at two and four-digit resolutions. After calculating accuracy, we used a chi-square test of independence to determine the statistical significance of the differences in accuracy across both ancestral groups. We also isolated the alleles to investigate the misclassification rate for each allele when looking at European versus African samples.
Assessing computational resources required to run HLA callers
For each tool, CPU time and RAM metrics were recorded to determine the computational performance. Each tool was run with 10 samples, 10 times each on the USC Center for Advanced Research Computing (USC CARC) with 1 node, 1 task and 4 CPUs per task. Metrics were collected using an inbuilt command from the cluster’s Slurm Workload Manager, called upon by adding the following line of code, ‘#SBATCH --mail-user=<E-mail address>’ to the sbatch script header, would email the user a full record of the execution of the script, containing information on the CPU time and the RAM usage. We averaged the 10 runs for each sample to produce 10 final values for each tool’s CPU and RAM, in order to minimize biases from resource availability in the shared cluster.
Results
ArcasHLA has the greatest overall accuracy across Class I and Class II loci
At two digit resolution, the accuracy varies from 14.9% to 99.4% for Class I predictions, and 7.4% to 98.4% accuracy for Class II predictions. At four digit resolution, the accuracy is lower, with class I accuracy ranging from 6.4% to 98.3% accuracy, while Class II predictions ranging from 0% to 93.4% accuracy (Fig 2a).
At both 2 digit and 4 digit resolutions, for Class I alleles only, OptiType has the best performance in terms of accuracy for Class I alleles with scores of 99.4% and 98.3% respectively. For Class II alleles, ArcasHLA has the best accuracy performance with scores of 98.4% and 93.4%. Notably, ArcasHLA is able to make both Class I and Class II predictions with consistently high accuracy, whereas OptiType is only able to make class I predictions. HLAMiner and HLAForest both struggle to make accurate predictions and have the worst performance for both Class I and Class II. These tools have a respective accuracy of 18.3% and 56.4% at two digit resolution, and 6.4% and 0% at four digit accuracy. Exact accuracy values of each tool from 0 (fully inaccurate) to 1 (fully accurate) are reported in Table S2a-c.
The HLA-B and HLA-DRB1 loci are most commonly mispredicted
We assessed each tool’s performance on each locus (A, B, C, DRB1, DQB1) to determine differences in accuracy across loci (Fig 2b). We find that Class II underperform compared to Class I, with the least accurate loci being HLA-DQB1 (70.1%) and HLA-DRB1 (75.5%) at four digits of resolution. Of the Class I loci, HLA-B performs the worst, with a 75.6% accuracy, whereas HLA-A and HLA-C both perform fairly well with accuracies of 89.3% and 88.8% respectively at four digits of significance.
We additionally evaluated the percentages at which each individual allele in the gold standard datasets is mispredicted (Fig 2c-d). Our results show that the Class II loci, particularly HLA-DQB1, is mispredicted at the highest ratio in comparison with alleles from other loci. Notably, we find that HLA-VBseq and HLAMiner are responsible for a majority of the Class II mispredictions, as removing these two tools from analysis decreased the significance of the difference in accuracies between Class I and Class II genes.
Computational expense varies greatly between HLA callers
We measured the CPU time and RAM usage of each caller for two samples per dataset to assess computational expensiveness (Fig 3a-b). We find great variation in the computational resources necessary to run the tools, with RAM ranging from the order of 100 kB to 100 GB, and CPU time ranging from approximately 10 minutes to 3 hours. While there is often a trade off between accuracy and computational resources required by each tool, we observe that RNA2HLA yielded the best balance between high accuracy (93.3%) at low computational expense (average 0.36 GB RAM amd 6.9 min CPU time), making it suitable for HLA typing with large scale studies. In contrast, the best accuracy tool, OptiType, requires an average of 14.2 GB RAM and 45.3 min CPU time per sample.
Read lengths do not exhibit large effect on prediction quality
From our preliminary analysis of 12 samples across Datasets 2 and 3, we find that increasing read length is associated with an increase in accuracy (Fig 3c) for all tools except HLAMiner. A read length of 36 has an average accuracy of 63%, a read length of 51 has an average accuracy of 67% and a read length of 76 has an average accuracy of 77%. In the future we will perform more in-depth analysis with 52 samples and smaller step sizes between the read lengths.
All HLA callers have higher accuracy on European than African samples
We used all samples from Dataset 1 (n=490) to evaluate differences in performance between Europe and African ancestry groups. For each caller, we determined the accuracy at each HLA locus separately for European (n=423) and African (n=67) subsets. We observe the greatest disparity in accuracy in HLAforest6 (p=0.019) and PHLAT7 (p=0.017). Furthermore, we find that differences in accuracy are more pronounced at four-digit resolution than they are at two-digit resolution (Fig 4a). At four digits of resolution, every single tool performed better on European than African ancestry groups. Although the disparity in some tools is not pronounced enough to be statistically significant, a binomial test demonstrates that the overall lower accuracy in African ancestry groups across all tools is significant (p=0.0039).
When we stratify the accuracy of each ancestry group by locus (Fig 4b), we find that there is some variability in whether European or African subsets have greater accuracy for Class I alleles, but a vast majority of tools seem to be better at making accurate predictions for the European group for Class II alleles. The greatest discrepancy in accuracy out of the Class II genes is HLA-DQB1 in which the African subset accuracy is 2.7% lower than the European subset. Less extreme but still significant, between the Class I genes the greatest discrepancies in accuracy are observed in HLA-B where the average African subset accuracy is 1.3% lower than the European subset. Furthermore, for HLA-A the difference in average accuracy is 1.3% and for HLA-DRB1 it is 1.0%. HLA-C is the only loci where the African subset outperformed the European subset on accuracy, with a difference of 0.3 percentage points. We then isolated alleles and plotted the Europe against Africa misclassification rate for each allele, finding a slight skew towards higher misclassification rates for the Africa samples. We also notice a significantly greater misclassification rate for Class II alleles for both Europe and Africa alleles, which disappears with the removal of the tools HLA-VBseq and HLAMiner, again verifying that these two tools are insufficient for making accurate Class II imputations.
Discussion
Our rigorous assessment of HLA callers highlights the advantages and limitations of computational HLA typing algorithms across several Class I and Class II loci and various sequencing parameters. Altogether, this study offers crucial, up-to-date information for researchers regarding appropriate choice of tool for an HLA typing from RNA-seq data. Furthermore, our study demonstrates the potential for RNA-seq based HLA typing tools to outperform traditional methods as a scalable and cost-effective method for imputing HLA typing. Crucially, our comparison of the performance of HLA callers across different ancestry groups highlights the pressing need to develop improved HLA typing algorithms and databases that can accurately account for the distinct genetic variations present in non-European populations.
In comparison to the previous RNA-seq benchmarking study15, we found highly consistent results with the same tools performing well in their study, OptiType and arcasHLA performing similarly well and HLA-VBseq struggling to make accurate imputations. We also found that more recently developed tools perform similarly well to prior tools, without many significant improvements to accuracy in the past few years. For instance, Optitype remains the most accurate tool for Class I genotypes. All tools performed much better on Class I than on Class II genes, with Class II accuracy being much lower compared to the current gold standard HLA typing methods.
Our study is the first to evaluate the ability for HLA callers to impute HLA types from specifically RNA-seq data. We find that all tools, except HLA-VBseq and HLAMiner, are capable of making Class I predictions to 4 digit (clinical) resolution to 82.6% -98.3% accuracy when compared with the gold standard method. We find that there are few tools that are able to make consistently accurate predictions across all loci and classes, and we believe that arcasHLA is the overall best tool as it performs with a consistently high 93.4% accuracy on Class I and Class II. In addition to arcasHLA, we also recommend RNA2HLA as a tool for high quality predictions at relatively low computational expense. Our study is also the first to explore the effect of read length and ancestry on the imputation quality of the callers. We found that read length does not have a substantial effect on accuracy. The tools are more accurate at making predictions for individuals of European ancestry than those of African ancestry.
In the future, we plan to complete the read length analysis and data collection for the CPU and RAM usage of the tools as we are currently missing some data. We also plan to determine the effect of coverage on the accuracy of the HLA typing algorithms. We also plan to determine the concordance in predictions across various samples, to evaluate whether the algorithms have the tendency to predict certain alleles. Once identifying patterns of misclassifications across tools, we also hope to analyze the sequence similarity of the allele with the misclassification to establish a hypothesis for why these errors occur. One final step we hope to complete is running tools with various custom parameters and arguments to determine the best set of parameters for each tool, rather than simply the default parameters like we are currently doing.
We envision that future HLA callers will accurately impute HLA genotypes for both Class I and Class II genes to 4-digit resolution, thereby offering clinicians and researchers a well performing and scalable option for HLA typing with WGS data. Current methods show an inability to make accurate Class II predictions, and discrepancies in prediction quality across different ancestral groups, and addressing these challenges will make HLA callers a viable choice for HLA typing studies. One potential limitation is that our CPU/RAM was collected from jobs submitted to a remote high performance computing cluster, and hence may not be completely controlled because it might be dependent on resource availability of the cluster at the time we ran each sample. We mitigated this limitation by running each sample 10 times and taking the average to be used for our analysis. We hope this study offers crucial information for researchers regarding appropriate choices of methods for conducting HLA typing with RNA-seq data.
Data Availability
The RNA-seq datasets used for this analysis are available under the following accession numbers on the NCBI SRA archive: ERP00194225, SRP29870426, ERP00010127, SRP09956828, SRP19849729, SRP09632930. All data required to produce the figures and analysis performed in this paper are freely available at https://github.com/ydottie/hlaproject.
Code Availability
All code required to produce the figures and analysis performed in this paper is under the Massachusetts Institute of Technology (MIT) license and is available at https://github.com/ydottie/hlaproject.
Acknowledgments
Thank you to Dr. Serghei Mangul (USC School of Pharmacy) for mentoring this project. Thank you to Ram Ayyala (PhD student in USC QCB department) for working jointly with me on the data collection for this project.