There are many reasons explaining why artificial intelligence has become a hot topic in many research fields, including biotechnology and life sciences. A key motivation is the increasing complexity and dimensionality of data, which are generated by increasingly accessible and cheaper technologies. This phenomenon has contributed to “big-data” becoming another critical aspect of modern life sciences. Big data refers to some of the characteristics of the datasets these new technologies can generate, such as:
Given the availability of these technologies at reasonable prices per sample, the generation of an increasing wealth of data shifts the real challenge to the dimensionality of the datasets. When we use the term “dimensionality” we refer to the number of variables (or features) describing each record in a dataset. This problem has been known for many years and regarded as the “curse of dimensionality”: it refers to the difficulty to represent or uncover the relationships between variables when the number of variables (i.e. dimensions) per data point increases. Especially in biology and life sciences, the increase in dimensionality corresponds to the higher number of descriptors necessary to characterise more complex phenomena, which in turn represents an obstacle in identifying meaningful trends or patterns in the data.
Finally, larger datasets require an additional effort in terms of data quality since the noise and potential bias impacting the analysis might increase with dimensionality, and might have a significant impact on the results.
These tools have therefore a huge potential to resolve unmet challenges, and the number of applications of deep learning to the analysis of omics data has exponentially increased during the past few years. For the scope of this review, we are particularly interested in a specific subset of these methods, i.e. explainable deep learning.
In this review we will focus on the role that explainable deep learning is having in providing novel insights, and we will highlight the impact this approach will have on a wide range of life science and omics data analyses.
Explainable DL and Genome Engineering
A variety of biotechnology settings, from healthcare improvement [
20,
21,
22] to food production [
23,
24], leverage artificial intelligence [
25]. Specifically, deep learning, originally inspired by biological models of computation and cognition in the human brain [
26], is rapidly replacing traditional AI algorithms in numerous fields, including gene therapy [
27,
28,
29,
30], drug research [
31], medical image analysis [
32,
33], and genetics [
11,
34]. The DL’s strength over other machine learning techniques, i.e. is its ability to detect and classify non-linear relationships in a dataset, is an invaluable asset in this area.
The clustered regularly interspaced short palindromic repeat (CRISPR)-based systems are among the most promising tools for genetic manipulation. The CRISPR toolbox and its applications have profoundly changed biological research, impacting agricultural practices and genomic medicine [
35]. Notably, this technology has a significant potential in addressing genetically determined Human diseases, because genome editing provides the means to correct genetic variants and modify specific gene functions, thus offering actionability for otherwise incurable pathologies. One of the most critical aspects of designing CRISPR intervention though, is the ability to control on-target versus off-target mutations: the implications are evident when CRISPR is used to plan medical interventions and gene therapies on Human beings. In this field, deep learning shows tremendous potential to help predicting CRISPR outcomes, by integrating high-dimensional heterogeneous data, such as sequence composition and secondary structure, thermodynamics, and chromatin accessibility within cell-specific contexts. Convolutional neural networks (CNNs) are particularly attractive solutions for the design of sgRNAs and predicting genome editing outcomes: they are able to perform feature extraction directly from sequence data via one-hot encoding, and offer improved interpretability of the biological data compared to state-of-the-art methods.
A wide array of deep-learning methods has therefore been developed to address these very challenging and outcome-determining aspects of genome engineering. For example, Xue and colleagues [
36] developed a CNN model to predict the sgRNA activity for the CRISPR-Cas9 system, from public experimentally validated sgRNA efficacy datasets in gene knockouts, covering several cell types, across five species. To capture the sgRNA motifs which determine the target activity of genome editing, a convolutional kernel visualisation strategy was implemented, assigning an importance score to nucleotide positions. A strong preference for guanine and a low preference for thymine were shown in different cell types. The nucleotide composition of the sgRNA seed region was then identified as the major driver for the formation of the guide-RNA-target-DNA heteroduplex. By deploying CNN coupled with explainability, the authors’ findings were consistent with what was already reported in the literature. Unlike Xue et al., Xiang et al. [
37] trained a CNN framework to predict the on-target gRNA efficacy, by integrating both sequence and thermodynamic data. In-house on-target gRNA activity data for 10,592 SpCas9 gRNAs and other publicly available datasets were used to train the model. The feature analysis unveiled two elements as key determinants for model predictions: the gRNA-DNA binding energy, and the two nucleotides proximal to the PAM, where G and A are favoured over C and T. In the same way, the CNN framework EpicasDL [
38] predicts the sgRNA on-target efficacy by integrating sequence and epigenetic data from 9 independent datasets. This approach uses different CRISPR technologies, including CRISPR interference and CRISPR activation to alter the gene expression, and CRISPRoff for heritable gene silencing. To discover the input sequence features associated with highly efficient epigenome editing, the authors explained their importance with saliency maps: they identified nucleosome positioning, RNA expression, and chromatin accessibility in the case of CRISPRoff and discovered specific sequence features in the case of CRISPRa. The results of the interpretation of the model were in accordance with findings reported before.
Another perspective to assess the safety of CRISPR intervention is to predict off-target effects: they are critical to clinical applications, as mentioned, but their prediction is more challenging. One of the key obstacles is represented by the mismatches between the gRNA spacer and the sequence of the potential off-target site [
39]. DeepCRISPR [
29] is a deep convolutional denoising neural network (DCDNN)-based autoencoder which meets the demand for both on- and off-target sgRNAs predictions. The authors confirmed via feature analysis that nucleotide preferences at specific positions, and open chromatin structures are preferred for sgRNA on-target design. Compared to previous approaches, the authors used the explainability of their models to build a new representation of the sgRNA off-target site, characterising better the boundaries of those mutations that can be tolerated and those that increase off-target activity.
Finally, some studies have deployed self-explaining deep learning models for CRISPR outcome predictions, in which the interpretability is performed by an attention module, intrinsic to the neural network [
40,
41,
42]. Among these, AttCRISPR [
40] is an ensemble of CNN and RNN methods trained on the DeepHF dataset [
43]. This model revealed Cas9 preference, when binding sgRNAs, of purines to pyrimidines, in agreement with previous investigations.
These applications show the relevance of AI in the field of gene editing: it holds the promise to advance the design of more precise therapeutics. In particular, CNNs coupled with post-hoc explanation models potentially allow better control of the outcome and precision of CRISPR gene editing, thus reducing the risks associated with this technology. We should also highlight a current limitation: most of the tools we described have been designed for CRISPR-Cas9 non-homologous end joining (NHEJ), therefore they might not be suitable for other CRISPR-based technologies.
Explainable DL and Drug Design
Drug research is one of the most promising biotech domains which may benefit from xAI in terms of costs and time saving [
44]. This area includes both the development of new drugs, as well as the repurposing of existing drugs for treating other conditions. Explainable AI can also play a pivotal role in the design of clinical trials, by providing unprecedented predictions for the correct assessment of eligible patients and identification of patient categories. Additionally, deep learning introduces new opportunities to anticipate adverse drug reactions (ADR) and aid correct assessments in view of drug approval applications. DL also increases the ability to integrate large-scale and high-dimensional patient-data, collected i.e. in the Electronic Health Records (EHR), or repositories such as the National Genomic Research Library realised by Genomics England [
45] and biobanks such as the UK Biobank [
46]. In the following paragraphs we will provide a few examples of these applications.
A recent announcement by the FDA’s Orphan Drug Designation demonstrates the huge potential xAI has in this area. A novel drug for idiopathic pulmonary fibrosis successfully concluded a Phase I clinical trial. The target for this therapeutic has been discovered with artificial intelligence and the novel drug, called INS018_055, has been designed using an AI model. The tool developed for target discovery (PandaOmics [
47]) is capable of integrating chemical, biological and medical data. The tool employed for drug design (Chemistry42 [
48]) was realised by combining different ML models, to increase its effectiveness.
Drug repurposing [
49] indicates the discovery of previously unknown drug-target interactions from approved or well-established clinical drugs. This activity nowadays plays an essential role in the availability of therapeutics for a wide range of conditions. The discovery leverages available databases, like Drugbank [
50] for drug-target interaction data, Pubchem [
51] for molecular information, CCLE [
52] collecting cancer cell line anticancer drug responses, and ChEMBL [
53], with chemical, bioactivity and genomic data. In this context, the model by Karimi et al. [
54] was trained on data from BindingDB [
55], STITCH [
56] and UniRef [
57]. It first implements a recurrent neural network model, to learn compound and protein representations. Next, a CNN is trained on the RNN outputs to predict the compound–protein affinity. The explainability of this architecture allowed the identification of pairwise drug-target interactions and was realised using an attention model. The effectiveness of this approach is demonstrated by the discovery of the antithrombotic effect of the DX-9065a compound, which was shown to selectively inhibit factor X (Xa) over Thrombin, forming a hydrogen bond with an aspartate (Asp189) of factor Xa.Rodríguez-Pérez et al. [
58] instead implemented multi-task DNNs (MT-DNNs) for predicting multi-target activity. Compounds and activity data were sourced from the ChEMBL database [
53]. The extended-connectivity fingerprint (ECFP4) was used as a molecular representation. A test compound was predicted to form potent interactions with growth factor receptor 2 kinase and serine/threonine Aurora-B kinase. The explainability provided by SHapley Additive exPlanations (SHAP) detected which molecular substructures on the ECFP4 representation contribute positively and negatively to the predictions of the multi-target activity.
Graph neural networks (GNNs) represent an increasing class of DNNs for deep learning in the life sciences and chemistry, probably due to their ability to learn directly from molecular graph representations, where nodes represent atoms and edges represent bonds connecting atoms. GNNs appear to work for compound activity prediction, a central task in drug discovery. However, few approaches are available to explain the rationale behind the model decisions. Pope et al. [
59] met this need by extending explainability methods as gradient-based saliency maps, Class Activation Mapping (CAM), and Excitation Backpropagation (EB) originally designed for CNNs, to GNNs. They used those explanation methods to identify functional groups of organic molecules within three different datasets. Among the investigated methods, Grad-CAM was the most suitable for explanations on molecular graphs. It unveiled functional groups, i.e. amides, trichloromethyl, sulfonamides, and aromatic structures, to be further experimentally validated. EdgeSHAPer [
60] is another example of a framework implementing GNNs and post-hoc explainability models. It introduces a novel approach to assess edge importance for GNNs predictions in chemical applications. EdgeSHAPer considers both directions for edges, each with its own contribution. Accordingly, the total contribution of an edge can be calculated by summing the EdgeSHAPer values for the two directions. The GCN model was applied to a compound classification task, to distinguish between dopamine D2 receptor ligands and other compounds. Then, the explainability model was used to map supporting and opposing edges on predicted test compounds.
Explainable DL is just beginning to show its utility in ADR prediction, and few cases of its application in this area are documented. Dey et al. [
61] developed a deep learning framework to predict ADRs using drug-ADR associations from the Side Effect Resource (SIDER) database [
62]. The model allows the identification of the input molecular substructures associated with ADRs, via feature analysis by an internal attention module. The explainability identified a particular substructure associated with aseptic necrosis, in 5 compounds, including Clobetasol. This discovered drug is not labelled in the SIDER database as associated with aseptic necrosis, but its long-term use is reported in a case study associated with necrosis. These findings suggest the model can predict ADRs, and discover the drug substructures that potentially play an important role in causing a specific or a group of ADRs.
Effective patient stratification is critical to the proper design of clinical trials. Deep learning can process heterogeneous data from EHRs and learn patient representations which can improve this phase. This offers the ability to predict clinical events and design more targeted pharmacological interventions. Prescience [
63] is a machine-learning-based method, which predicts intraoperative hypoxaemia events before they occur (the decrease in SpO2 <= 92%). Crucially, explainable methods are used to identify the patient- and surgery-specific factors that increase that risk. In this approach, the authors trained a gradient boosting machine model to get hypoxaemia predictions at the start of a procedure, as well as real-time predictions in the following period. Minute-by-minute hemodynamic and ventilation parameters were integrated with patient data from EHRs. Using SHAP for explainability, the authors were able to discover BMI and age as significant risk factors. Additionally, Prescience findings confirmed other clinical observations and the relationship between obesity and other comorbidities with adverse anaesthesiology outcomes.
These examples show the transformative effect deep learning and explainability are having on drug design, including clinical trials. Artificial intelligence holds the promise to change the way they work, thus providing a wide range of more effective and targeted therapeutics.
Explainable DL and Radiomics
Radiomics refers to the analysis of radiological images by using computational and statistical approaches. An image is converted into quantitative features, such as intensity, shape, volume, and texture of the region of interest (ROI), based on the assumption that such features can represent the ROI phenotype or an underlying molecular profile. The features can be extracted from different medical imaging techniques: computed tomography (CT), magnetic resonance imaging (MRI), and positron emission tomography (PET) [
64].
Explainable DL allows spatially resolved predictions directly from routine histopathology images of hematoxylin and eosin (H&E) stained samples. In oncology, these methods can infer genetic alterations of tissues, and predict patient survival and treatment response [
65]. For instance, Schmauch et al. [
66] trained HE2RNA, a multilayer perceptron (MLP) model, on TCGA samples with matched RNA-Seq and whole slide image (WSI) data, for the prediction of gene expression directly from WSI. The identification of WSI-specific regions which could predict gene expression levels was possible with explainability methods. The authors in this work used a virtual spatialization map (VSM). This approach allowed to distinguish the tiles containing T and B lymphocytes within a dataset of 86 CRC slides. Lu et al. [
67] presented Tumour Origin Assessment via Deep Learning (TOAD), an interpretable deep-learning model to assist pathologists for diagnoses of cancer of unknown primary origin. First, a CNN is trained on digitised WSIs from publicly available and in-house datasets: the model uses H&E slides from tumours with known primary origins to predict primary tumour or metastasis with its origin site. In a second step, an attention-based module provides explainability of the inputs and ranks regions in H&E slide on their relative importance. In this case, explainability allows the discovery of the tumour origin, for which additional tests would otherwise be required for diagnosis. In another study, [
68] a deep learning framework was developed to diagnose Barrett’s oesophagus (BE), the main precursor of esophageal adenocarcinoma, by analysis of Cytosponge-TFF3 pathology slides. Cytosponge-TFF3 patient data with paired pathology and endoscopy data were used to train a model. In this work, the saliency maps enable the identification of image features relevant to diagnosis, and whether they confirm the human classification. In this way, the visualisations generated by the explainability method, provide the pathologist with an additional level of validation for their classification and further quality control of the regions used to achieve their final assessment.
Most radiological applications implement visualisation or occlusion methods to focus on brain areas involved in degenerative pathologies like Alzheimer’s Disease (AD), by analysing MRI, CT or PET images. Yee et al. [
69] used FDG-PET images from the Alzheimer’s Disease Neuroimaging Initiative (ADNI) database (adni.loni.usc.edu), to train a 3D CNN and discriminate between patients affected by Dementia of the Alzheimer’s type (DAT) and stable normal controls. Explainability methods were used to highlight the regions of the brain that are most important to the classification task. To this goal, they employed an averaged probability score, from Guided Backpropagation (GB) and Gradient-weighted class activation mapping (GradCAM). The averaged maps provided tools for better classification of the patients, and highlighted regions known to be severely affected by DAT, including the posterior cingulate cortex, precuneus, and hippocampus. Similarly, Etminani et al. [
70] developed a 3D CNN that predicts the clinical diagnosis of AD, dementia with Lewy bodies (DLB), mild cognitive impairment due to Alzheimer’s disease (MCI-AD), and cognitively normal (CN), using fluorine 18 fluorodeoxyglucose PET (18F-FDG PET) from the European DLB (EDLB) Consortium [
71]. They used explainable methods to create heatmaps of the scans, and help clinicians characterise the clinical case. The occlusion sensitivity heatmaps pointed to the posterior and anterior cingulate cortex, as well as the temporal lobes as the regions which best discriminate patients and help diagnose those with AD.
Qiu et al. [
72] describe another application which implements explainability to identify brain relevant regions in AD diagnosis. They developed a CNN-based framework to classify individuals with normal cognition (NC), mild cognitive impairment (MCI), AD, and non-AD dementias (nADD), from volumetric MRI data. They used eight cohorts including both publicly available and in-house maintained data, to train the model. The authors leveraged the SHAP method, to link dementia predictions to brain locations from MRI images. SHAP values were negative in the hippocampal region in NC participants and nADD cases, but positive in participants with dementia: this confirmed the known role of the hippocampus in memory function, and hippocampal atrophy with AD-related aetiology. In nADD cases, the role of the lateral ventricles and frontal lobes was highlighted by SHAP values.
Finally, Poplin et al. [
73] optimised a deep learning model to predict cardiovascular risk factors, including age, gender and smoking status, not previously quantified from fundus images of patient, collected in UK Biobank [
46] and EyePACS [
74] datasets. To explain the outcomes of the DL architecture, they implemented a self-attention mechanism, which generated saliency maps: the visualisation was used to understand which anatomical features were most important for the neural network to reach its predictions. The explainable approach highlighted anatomical features like the optic disk and blood vessels which have been determinant to achieve the prediction of cardiovascular risk factors from the images.
CNNs have been initially developed in image analysis: this explains why they are the method of choice to process biomedical images. They have successfully fulfilled tasks such as tumour diagnosis, prediction of mutations and gene expression, classification of cancers of unknown primary origin. In most cases they achieved prediction accuracy otherwise very difficult for a human operator. The explainability methods have allowed DL to be safely employed within the clinical research, offering effective opportunities for assisted diagnostics while clinicians retained a crucial role in the assessment of medical images.
Explainable DL and genomics
In the search for variants that together explain complex traits in Humans, the vast majority of disease risk modelling is based on the idea that a disease risk is mostly determined by the sum of the effects of each single variant. This is grounded on the assumption that, even when the individual effect is very small, each variant will always show a marginal effect [
75,
76] which will be detected, provided a sufficiently large sample size has been achieved, to measure the effect.
With the introduction of polygenic risk scores, the contribution of variants which would not achieve genome-wide statistical significance has also been accounted for. The assumption of additive effects remains at the heart of this approach. Polygenic risk score models are essentially a linear combination of the effects of the variants selected for the model [
77,
78].
For decades, there has been a debate over the concept and the nature of “missing heritability” [
79]. This is defined as the fraction of phenotype variance which cannot be immediately explained by accounted genetic variation: the progressive transition from chip genotyping to whole-genome sequencing has allowed a massive increase in the resolution of our genome analyses and in the ability to capture genetic variation, including extremely rare variants and assess their contribution to phenotypes [
80]. This improvement in measuring genetic variation and its distribution across phenotypes allowed a more granular decomposition of heritability estimates as well as a better understanding of the contribution of different types of variants to heritability. Recently, a work by Wainschtein and colleagues has shown that part of the still missing heritability is due to rare variants in genomic regions with low linkage disequilibrium [
81]. Although important, this approach still relies on massive sample sizes in order to be able to capture extremely small effects, and cannot quite fill in the remaining gap of the missing heritability [
82]. Indeed, the analysis of very large population cohorts shows that rare coding variants collectively explain on average 1.3% of phenotypic variance [
83] and that rare-variants heritability is more concentrated on a specific subset of genes, mostly constrained, while common variants heritability is more polygenic [
83].
Non-linear effects do not appear to play a major role in the outcomes of these analyses: however, accounting for non-linear relationships seems to improve both prediction and power [
84]. Additionally, the discovery of variants contributing to phenotypes and their heterogeneity might be limited due to the widespread use of linear models, which can only achieve sufficient power with extremely large sample sizes to uncover variants with extremely small marginal effects, and are essentially unable to capture purely epistatic effects [
85,
86,
87].
Epistasis has two quite different components: nonspecific genetic interactions due to non-linear mapping between genotypes and phenotypes, as well as specific interactions between two or more genetic variants [
88]. In both cases, the departure from linearity is a central element, and dedicated methods are needed to model and capture this contribution. Detecting epistasis is a challenging task, and a quite daunting one from a computational point of view [
87,
89]: methods assessing interactions are severely limited in detecting higher order epistasis, which likely parallels the order of complexity observed in biological organisms [
90,
91].
Within this picture, it is quite clear that deep learning might offer a valid way forward to address the challenge:
- -
deep layers in a neural network increase the abstraction in the representation of the input data, using non-linear activation functions
- -
non-linear relationships in the data are therefore captured by a neural network, enhancing the ability to unmask different patterns in the data [
92]
- -
latent representations offer new insights into networks and interactions at different level in genomics, epigenomics and functional dimensions of the data [
93]
On top of this, the advancement of explainability methods adds the element of discovery we mentioned earlier, which offers unprecedented opportunities to tackle existing challenges in genomics and association to diseases [
93,
94].
Despite DNN’s potential to detect interactions, few studies are reported in literature attempting the identification of SNP epistasis from GWAS [
94,
95,
96]. Uppu and colleagues [
94] implemented a deep feed-forward network to detect two-locus interacting SNPs and disease association. The model was trained on simulated datasets, as well as an existing cohort of breast cancer patients and controls [
97]. The results predicted pairwise interactions occurring in five oestrogen-metabolism genes, potentially associated with the onset of breast cancer. While significant for the use of DL in unmasking epistasis, this work did not implement an explainable approach. Romagnoni et al. [
95] first addressed this issue by adopting a permutation feature importance (PFI) approach tailored to DNN models. The authors deployed a residual Dense NN with 3 hidden layers of 64 neurons (ResDN3) for the classification of Crohn’s Disease (CD) patients. The input of their model was a large dataset, Immunochip [
98], from CD patients and healthy controls, genotyped for more than 150 thousand genetic variants. Since the Immunochip is an unphased dataset, each SNP was encoded as 0-0, 0-1, 1-1, U-U (where U is an unknown allele). The authors then used the PFI explainability model to discover the regions harbouring the best-associated SNPs in CD patients. In this specific case the upstream selection of variants needed for the design of the Immunochip prevented the discovery of new polymorphisms, but confirmed the implication of a large number of SNPs in regions known to harbour associations with CD.
An additional example of the application of DNNs to process GWAS data is reported in Mieth et al. [
96]. The authors developed the DeepCOMBI model to classify individuals based on their SNPs: the architecture was trained on a synthetic GWAS dataset [
99]. As a post-hoc explainability method, the authors chose the layer-wise relevance propagation (LRP). This approach assigned a score to each variant in the input, based on the relevance played in the results of the classification. These scores can be used to rank the SNPs, and to select the most relevant subset for further multiple-hypothesis testing. Notably, the application of this approach to real-world WTCCC data [
100] allowed the discovery of two novel disease associations, rs10889923 for hypertension, rs4769283 for type 1 diabetes.
Recently, feature importance score methods have been applied to explain more complex epistatic interactions [
101]. Greenside et al. presented a new method called Deep Feature Interaction Maps (DFIM) to estimate interactions between all pairs of features in any one-hot encoded input DNA sequence. This approach, computed using DeepLIFT, approximates SHAP scores and produces a Feature Interaction Score (FIS). This value can be used to quantify pairwise epistasis. The method developed in this work was used to investigate interactions between motifs of co-binding TFs, and allowed the discovery of strong interactions between TAL1 and GATA1 motifs at short range (<20bp), but also for longer-range interactions, beyond 70bp.
Feature attribution methods also support the discovery of cis-regulatory patterns in the input DNA sequences, by estimating the contribution of each input feature: this approach can be applied both to single nucleotides as well as to motifs, by evaluating the prediction of a DNN. For example, DeepBind [
102] and DeepSEA [
103] have been among the first CNN methods applied to genomics data. DeepBind [
102] is a DNN tool for the prediction of the sequence specificity of DNA-and RNA-binding proteins. The model is trained on public PBM, RNAcompete, ChIP-seq and HT-SELEX experiment data. In this approach, the specificities of different sequences are visualised in a mutation map, which indicates how variations affect binding affinity and therefore have an impact on gene expression. The mutation map is an in-silico mutagenesis approach which estimates the weight of each nucleotide on the binding affinity. It can easily be translated into well-known motif plots, where the height of each nucleotide represents its importance. First, DeepBind was trained on ChIP-seq and HT-SELEX data. Then, a mutation map was applied to perturb the input and estimate the effect on the model output. This method demonstrated that variations at the TF binding site of SP1, in the low-density lipoprotein receptor (LDLR) gene promoter, disrupt the SP1 binding site, known to be associated with familial hypercholesterolemia. Similarly, the DeepSEA model was developed to predict the binding of chromatin proteins and histone marks on DNA sequences used as an input [
103]. It was trained by integrating TF binding, DHS and histone-mark data from the ENCODE [
104] and Roadmap Epigenomics projects [
105]. Like in the previous use-case, in-silico mutagenesis was then used as an input perturbation approach to explain the model. This method allowed the authors to discover which features of the input sequences impacted most on chromatin predictions. This model accurately predicted the C>T substitution in the locus rs4784227 associated with risk for breast cancer. Similarly, it allowed them to discover that a T>C transversion in a SNP associated with the α thalassemia creates a binding site for GATA1.
The use of explainable deep learning is not limited to the discovery of epistatic interactions, or novel variants affecting binding motifs, but has shown impressive results in other areas such as transcriptomics and spatial functional genomics.
Yap and colleagues [
106] trained a CNN to classify which tissue type a sample originates from, by analysing their transcriptome in the GTEx project [
107]. For a given sample, the gene-vector of TMM values was converted into a square matrix suitable for input to the CNN. Once the model was trained, they used SHAP to explain the model results. This approach identified which genes are likely to play a key role in determining each tissue type. Many of the most relevant genes were also identified in an independent dataset, HPA [
108], and by traditional differential expression (DE) analyses, while a small genes subset was SHAP-specific. Moreover, pathway analyses from SHAP and DE gene lists encompassed similar pathways for each tissue. These findings demonstrate the reliability of SHAP to recognise genes that are biologically relevant for tissue classification, and its ability to find novel insights by detecting more gene signatures, indistinguishable in DE analysis.
The availability of large amounts of data such as TF and histone modification ChIP-Seq, Hi-C and RNA-seq, allowed the use of DNN models to investigate the regulatory mechanisms underlying gene expression. To this aim, Zeng et al. [
109] trained their DeepExpression DNN model on chromatin [
110] and expression data [
104] from three human cell lines, to regress gene expression from promoter and enhancer profiles. The algorithm includes three modules: the first proximal promoter module receives one-hot encoded DNA sequences in promoter regions as input; the second distal enhancer–promoter interaction module receives 400-dimensional HiChIP enhancer–promoter interactions signals as input. The last joint module integrates the outputs of the first two, to produce a predicted gene expression signal. After the model was fully trained, based on the assumption that TF binding is needed to form the three-dimensional enhancer–promoter interactions, the authors demonstrated the interpretability of their model by identifying motifs learned from promoter and enhancer regions, in the first convolution layer of DeepExpression. In order to process the promoters, they converted filters in the first convolution layer into a position weight matrix (PWM), searched for activated positions, and then converted them into motifs. The CisModule [
111] was instead applied to visualise the motifs learned from enhancer sequences. In this way, they were able to match about 65% of motifs learned in promoters, and 92% of motifs from enhancers, to known Vertebrates motifs, in different cell lines.
Avsec and colleagues in 2021 [
112] provided another example of the potential these approaches offer to gain deeper insights on the mechanisms regulating gene expression. They developed the Enformer DNN model, for the prediction of gene expression and chromatin states using just one-hot-encoded DNA sequence data in both Humans and Mouse. The model was trained on the same targets and genomic intervals from Kelley et al. datasets [
113]. Contribution scores were used to provide explainability to Enformer and allow the functional interpretation of the input DNA sequences. Additionally, in-silico mutagenesis was used to add another layer of explainability and increase the resolution at the variant-level. Using this approach they identified the effects of a list of variants on different cell-type-specific gene expression. For example, in a test case Enformer predicted a reduced NLRC5 expression. The in-silico mutagenesis method revealed that the variant rs11644125 affects the known motif of the SP1 transcription factor motif, suggesting that the perturbation of SP1 binding in hematopoietic cells can alter NLRC5 expression.
In the approaches we have summarised, the encoding method used to convert DNA sequences into a suitable input for a neural network plays an important role. This choice affects which explainability method one can apply to the deep learning model, and can affect both the capability and nature of the patterns learned by the networks, as well as the type of insights post-hoc explainability can provide. We have represented some of the most frequently used encodings in
Figure 2, illustrating the explainability methods used in the above-described studies.
Deep learning has clearly already become a widely adopted tool in genomics, due to its ability to perform the integrated analysis of high-dimensional omics data. The downstream interpretation methods are mostly focused on identifying the impact of DNA and/or RNA sequence variations on protein binding, chromatin accessibility, and in the regulation of gene expression. The examples we have given demonstrate how effective they are in discovering the interactions between genes in pathways, the role and characteristics of regulatory regions, the interplay (and epistasis) between genomic variations.